stduhpf commited on
Commit
4bc0828
1 Parent(s): 985e0c2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -5
README.md CHANGED
@@ -7,13 +7,16 @@ base_model:
7
  - stabilityai/stable-diffusion-3.5-large-turbo
8
  base_model_relation: quantized
9
  ---
 
10
  These models are made to work with [stable-diffusion.cpp](https://github.com/leejet/stable-diffusion.cpp) release [master-ac54e00](https://github.com/leejet/stable-diffusion.cpp/releases/tag/master-ac54e00) onwards. Support for other inference backends is not guarenteed.
11
 
12
  Quantized using this PR https://github.com/leejet/stable-diffusion.cpp/pull/447
13
 
14
- ### Files:
15
 
16
- #### Mixed Types:
 
 
17
 
18
  - [sd3.5_large_turbo-q2_k_4_0.gguf](https://huggingface.co/stduhpf/SD3.5-Large-Turbo-GGUF-mixed-sdcpp/blob/main/sd3.5_large_turbo-q2_k_4_0.gguf): Smallest quantization yet. Use this if you can't afford anything bigger
19
  - [sd3.5_large_turbo-q3_k_4_0.gguf](https://huggingface.co/stduhpf/SD3.5-Large-Turbo-GGUF-mixed-sdcpp/blob/main/sd3.5_large_turbo-q3_k_4_0.gguf): Smaller than q4_0, acceptable degradation.
@@ -21,12 +24,12 @@ Quantized using this PR https://github.com/leejet/stable-diffusion.cpp/pull/447
21
  - [sd3.5_large_turbo-q4_k_4_1.gguf](https://huggingface.co/stduhpf/SD3.5-Large-Turbo-GGUF-mixed-sdcpp/blob/main/sd3.5_large_turbo-q4_k_4_1.gguf): Smaller than q4_1, and with comparable degradation. Recommended
22
  - [sd3.5_large_turbo-q4_k_5_0.gguf](https://huggingface.co/stduhpf/SD3.5-Large-Turbo-GGUF-mixed-sdcpp/blob/main/sd3.5_large_turbo-q4_k_5_0.gguf): Smaller than q5_0, and with comparable degradation. Recommended
23
 
24
- #### Legacy types:
25
 
26
  - [sd3.5_large_turbo-q4_0.gguf](https://huggingface.co/stduhpf/SD3.5-Large-Turbo-GGUF-mixed-sdcpp/blob/main/legacy/sd3.5_large_turbo-q4_0.gguf): Same size as q4_k_4_0, Not recommended (use q4_k_4_0 instead)
27
- (I wanted to upload more, but it's not working anymore, maybe i hit a rate limit)
28
 
29
- ### Outputs:
30
 
31
  | Name | Image | Image | Image |
32
  | ------------------ | -------------------------------- | ---------------------------------- | ---------------------------------- |
 
7
  - stabilityai/stable-diffusion-3.5-large-turbo
8
  base_model_relation: quantized
9
  ---
10
+ ## Overview
11
  These models are made to work with [stable-diffusion.cpp](https://github.com/leejet/stable-diffusion.cpp) release [master-ac54e00](https://github.com/leejet/stable-diffusion.cpp/releases/tag/master-ac54e00) onwards. Support for other inference backends is not guarenteed.
12
 
13
  Quantized using this PR https://github.com/leejet/stable-diffusion.cpp/pull/447
14
 
15
+ Normal K-quants are not working properly with SD3.5-Large models because over 90% of the weights are in tensors whose shape doesn't match the 256 superblock size of K-quants and can't be quantized. Mixing quantization types allow to take adventage of the better fidelity of k-quantto some extent while keeping the file size small.
16
 
17
+ ## Files:
18
+
19
+ ### Mixed Types:
20
 
21
  - [sd3.5_large_turbo-q2_k_4_0.gguf](https://huggingface.co/stduhpf/SD3.5-Large-Turbo-GGUF-mixed-sdcpp/blob/main/sd3.5_large_turbo-q2_k_4_0.gguf): Smallest quantization yet. Use this if you can't afford anything bigger
22
  - [sd3.5_large_turbo-q3_k_4_0.gguf](https://huggingface.co/stduhpf/SD3.5-Large-Turbo-GGUF-mixed-sdcpp/blob/main/sd3.5_large_turbo-q3_k_4_0.gguf): Smaller than q4_0, acceptable degradation.
 
24
  - [sd3.5_large_turbo-q4_k_4_1.gguf](https://huggingface.co/stduhpf/SD3.5-Large-Turbo-GGUF-mixed-sdcpp/blob/main/sd3.5_large_turbo-q4_k_4_1.gguf): Smaller than q4_1, and with comparable degradation. Recommended
25
  - [sd3.5_large_turbo-q4_k_5_0.gguf](https://huggingface.co/stduhpf/SD3.5-Large-Turbo-GGUF-mixed-sdcpp/blob/main/sd3.5_large_turbo-q4_k_5_0.gguf): Smaller than q5_0, and with comparable degradation. Recommended
26
 
27
+ ### Legacy types:
28
 
29
  - [sd3.5_large_turbo-q4_0.gguf](https://huggingface.co/stduhpf/SD3.5-Large-Turbo-GGUF-mixed-sdcpp/blob/main/legacy/sd3.5_large_turbo-q4_0.gguf): Same size as q4_k_4_0, Not recommended (use q4_k_4_0 instead)
30
+ - (I wanted to upload more, but it's not working anymore, maybe i hit a rate limit)
31
 
32
+ ## Outputs:
33
 
34
  | Name | Image | Image | Image |
35
  | ------------------ | -------------------------------- | ---------------------------------- | ---------------------------------- |