File size: 2,809 Bytes
1cae86c
 
 
 
 
 
 
b36cd2b
a3cbf7f
 
 
220e2bb
 
 
 
00d15ed
e8bd1d3
220e2bb
4822f90
b0a4f72
4822f90
 
 
a0f5ff6
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
---
license: other
license_name: sacla
license_link: >-
  https://huggingface.co/stabilityai/stable-diffusion-3.5-large-turbo/blob/main/LICENSE.md
base_model:
- stabilityai/stable-diffusion-3.5-large-turbo
base_model_relation: quantized
---
These models are made to work with [stable-diffusion.cpp](https://github.com/leejet/stable-diffusion.cpp) release [master-ac54e00](https://github.com/leejet/stable-diffusion.cpp/releases/tag/master-ac54e00) onwards. Support for other inference backends is not guarenteed.

Quantized using this PR https://github.com/leejet/stable-diffusion.cpp/pull/447

### Files:

- [sd3.5_large_turbo-q2_k_4_0.gguf](https://huggingface.co/stduhpf/SD3.5-Large-Turbo-GGUF-mixed-sdcpp/blob/main/sd3.5_large_turbo-q2_k_4_0.gguf): Smallest quantization yet. Use this if you can't afford anything bigger
- [sd3.5_large_turbo-q3_k_4_0.gguf](https://huggingface.co/stduhpf/SD3.5-Large-Turbo-GGUF-mixed-sdcpp/blob/main/sd3.5_large_turbo-q3_k_4_0.gguf): Smaller than q4_0, acceptable degradation.
- [sd3.5_large_turbo-q4_k_4_0.gguf](https://huggingface.co/stduhpf/SD3.5-Large-Turbo-GGUF-mixed-sdcpp/blob/main/sd3.5_large_turbo-q4_k_4_0.gguf): Exacty same size as q4_0, but with slightly less degradation. Recommended
- [sd3.5_large_turbo-q4_k_4_1.gguf](https://huggingface.co/stduhpf/SD3.5-Large-Turbo-GGUF-mixed-sdcpp/blob/main/sd3.5_large_turbo-q4_k_4_1.gguf): Smaller than q4_1, and with comparable degradation. Recommended
- [sd3.5_large_turbo-q4_k_5_0.gguf](https://huggingface.co/stduhpf/SD3.5-Large-Turbo-GGUF-mixed-sdcpp/blob/main/sd3.5_large_turbo-q4_k_5_0.gguf): Smaller than q5_0, and with comparable degradation. Recommended

### Outputs:

| Name               | Image                            | Image                              |
| ------------------ | -------------------------------- | ---------------------------------- |
| q2_k_4_0           | ![q2_k_4_0](Images/q2_k_4_0.png) | ![q2_k_4_0](Images/1_q2_k_4_0.png) |
| q3_k_4_0           | ![q3_k_4_0](Images/q3_k_4_0.png) | ![q3_k_4_0](Images/1_q3_k_4_0.png) |
| q4_0               | ![q4_0](Images/q4_0.png)         | ![q4_0](Images/1_q4_0.png)         |
| q4_k_4_0           | ![q4_k_4_0](Images/q4_k_4_0.png) | ![q4_k_4_0](Images/1_q4_k_4_0.png) |
| q4_k_4_1           | ![q4_k_4_1](Images/q4_k_4_1.png) | ![q4_k_4_1](Images/1_q4_k_4_1.png) |
| q4_1               | ![q4_1](Images/q4_1.png)         | ![q4_1](Images/1_q4_1.png)         |
| q4_k_5_0           | ![q4_k_5_0](Images/q4_k_5_0.png) | ![q4_k_5_0](Images/1_q4_k_5_0.png) |
| q5_0               | ![q5_0](Images/q5_0.png)         | ![q5_0](Images/1_q5_0.png)         |
| q8_0               | ![q8_0](Images/q8_0.png)         | ![q8_0](Images/1_q8_0.png)         |
| f16(sft)           | ![f16](Images/f16.png)           | ![f16](Images/1_f16.png)           |