Bartowski
bartowski
AI & ML interests
None yet
Organizations
bartowski's activity
Model doesn't seem to tokenize new lines in chat template?
4
#84 opened 4 days ago
by
bartowski
GGUF IQuants of TheDrummer/Smegmma-Deluxe-9B-v1?
2
#1 opened 1 day ago
by
Joseph717171
ValueError: Failed to load model from file: models\gemma-2-27b-it-Q3_K_M.gguf
2
#8 opened 3 days ago
by
Star81
"Experimental" = ZeroWw method.
37
#7 opened 7 days ago
by
ZeroWw
Problem Model
13
#1 opened 6 days ago
by
ClaudioItaly
Q6_K response quality diverges (in a bad way)
34
#4 opened 8 days ago
by
thethinkmachine
prompt template?
2
#1 opened 5 days ago
by
mirek190
Upload Qwen2ForCausalLM
2
#5 opened 7 days ago
by
Locutusque
GGUF Model
1
#2 opened 5 days ago
by
Pecc
FYI: likely broken 128k - maybe not, maybe only PPL
2
#1 opened 6 days ago
by
bartowski
Experimental = ZeroWw way.
2
#1 opened 6 days ago
by
ZeroWw
Uploaded GGUF and exl2 as Phi 3.1
5
#80 opened 7 days ago
by
bartowski
It seems the quants were created before the BPE pre-tokenizer fix?
4
#1 opened about 1 month ago
by
skruse
More gemma 2 llama.cpp merges, do they require GGUF regen again?
4
#6 opened 7 days ago
by
IHadToMakeAccount
Updated model should have a new version number
2
#79 opened 7 days ago
by
EwoutH
Template ? not able t use it.
7
#1 opened 9 days ago
by
NK-Spacewalk
Load gemma-2-27b-it-Q4_K_S.gguf from LMStudio always crashed my MacBook Pro M2
3
#4 opened 10 days ago
by
Million
ChatML template is not working
3
#2 opened 9 days ago
by
NK-Spacewalk
Llama.cpp fixes have been merged, requires gguf regen
3
#5 opened 9 days ago
by
RamoreRemora
Pretty good stuff!
1
#1 opened 10 days ago
by
Razrien
Prompt format <bos> not needed?
16
#3 opened 11 days ago
by
eamag
OOM when trying to load the Q8 in Oobabooga (I have 98GB VRAM)
2
#1 opened 9 days ago
by
AIGUYCONTENT
Hallucinations, misspellings etc. Something seems broken?
16
#10 opened 11 days ago
by
sam-paech
'LlamaCppModel' object has no attribute 'model'
10
#2 opened 11 days ago
by
DrNicefellow
Tokenizer was broken in the PR when these were made, now fixed(?)
2
#1 opened 11 days ago
by
Fizzarolli
Testing experimental quants
31
#2 opened 22 days ago
by
bartowski
Llama 3 8B Instruct - Q8 vs FP16 vs FP32
2
#8 opened 14 days ago
by
Hearcharted
Update README.md with license information
1
#1 opened 14 days ago
by
Chen-01AI
Update README.md
2
#1 opened 14 days ago
by
rombodawg
Update prompt template
1
#1 opened 14 days ago
by
rombodawg
Please post f16 quantization.
8
#1 opened about 2 months ago
by
ZeroWw
I made a mild oopsie.
4
#1 opened 15 days ago
by
Nitral-AI
Fix GGUF link
1
#1 opened 15 days ago
by
bartowski
Poor Chinese performance
1
#2 opened 15 days ago
by
yttria
Error loading model
14
#1 opened 21 days ago
by
smchapman54
Thank you for quants! 🤗
1
#1 opened 17 days ago
by
SicariusSicariiStuff
Create config.json
3
#7 opened 17 days ago
by
standingout
Please check these quantizations.
4
#40 opened 20 days ago
by
ZeroWw
Problem with LLM Studio
5
#2 opened 19 days ago
by
Techn0man1ac
LM studio Loading error
6
#1 opened 20 days ago
by
Vezora
EOS token is getting printed
9
#1 opened 21 days ago
by
migtissera