Help with convert-hf-to-gguf-update.py

#31
by Zibri - opened

I created Meta-Llama-3-8B-Instruct-f32-Q8_0.gguf to test it... but the tokenizer is messed up by the previous version...
how do I use convert-hf-to-gguf-update.py to fix it?

Zibri changed discussion title from Help. to Help with convert-hf-to-gguf-update.py

You need to convert again from the beginning.

You need to use the ...update.py script as explained in the Warning box, in the card, to get the llama-bpe tokenizer config files, then replace the ones in your original HF-Model files folder.

HF model β†’ outfile BF16 GGUF β†’ Quant

I recommend just using the lossless script from this repo, during the model download, replace the config files and it should work.

Adding myself here so I can reply more easily.

Currently you should be good to use the default configs with the recent models, since they should all have them correctly by now.

FantasiaFoundry changed discussion status to closed

Sign up or log in to comment