--- license: apache-2.0 --- # hwarnecke/SauerkrautLM-Nemo-12b-Instruct-Q6_K-GGUF This model was converted to GGUF format from [`VAGOsolutions/Llama-3.1-SauerkrautLM-70b-Instruct`](https://huggingface.co/VAGOsolutions/Llama-3.1-SauerkrautLM-70b-Instruct) using llama.cpp. Refer to the [original model card](https://huggingface.co/VAGOsolutions/Llama-3.1-SauerkrautLM-70b-Instruct) for more details on the model. Since HuggingFace supports files up to 50GB, the Q6_K quant is split into two files instead. You probably need to merge them again before you can use them. You can use llama.cpp for that. Use ```shell ./llama-gguf-split -h ``` to find out more about the function, after you installed llama.cpp.