library_name: transformers | |
license: apache-2.0 | |
base_model: | |
- Youlln/ECE-Qwen0.5B-FT-V2 | |
### Model Description | |
This model was fine-tuned on top of the Youlln/ECE-Qwen0.5B-FT-V2 to improve its performance for specific tasks. After fine-tuning, an 8-bit quantization technique was applied using the bitsandbytes library. This process reduced the model size and optimized inference speed while maintaining a good level of accuracy. The model is suitable for environments where memory and computational efficiency are critical, such as edge devices or applications requiring faster response times. | |
Quantization was selectively applied, and some layers remain in float16 to ensure precision in key computations, balancing efficiency and accuracy | |
- **Developed by:** Youri Lalain (@Youlln) | |
- **Organization:** ECE engineering school |