--- title: README emoji: 📈 colorFrom: green colorTo: yellow sdk: gradio pinned: false --- # Welcome to Our Extreme Quantization Hub Here, we focus on models built with **extreme quantization techniques**. Our mission is to push the boundaries of this technology, making it accessible for the community and setting new standards for the field. --- ### 🚀 **Latest Releases**: 8B Models Fine-tuned on BitNet Architecture You can learn more about how we created the following models [in this blogpost](https://huggingface.co/blog/1_58_llm_extreme_quantization) - **[Llama3-8B-1.58-100B-tokens](https://huggingface.co/HF1BitLLM/Llama3-8B-1.58-100B-tokens)** *Fine-tuned on 100B tokens for maximum performance.* - **[Llama3-8B-1.58-Linear-10B-tokens](https://huggingface.co/HF1BitLLM/Llama3-8B-1.58-Linear-10B-tokens)** *Fine-tuned with a Linear Lambda scheduler on 10B tokens.* - **[Llama3-8B-1.58-Sigmoid-k100-10B-tokens](https://huggingface.co/HF1BitLLM/Llama3-8B-1.58-Sigmoid-k100-10B-tokens)** *Fine-tuned with a Simgoid Lambda scheduler with k=100 on 10B tokens.* --- Join us in the era of extreme quantization as we continue to push this technology forward !