shaowenchen commited on
Commit
68bd6b9
1 Parent(s): 4433df2

improvement: readme

Browse files
Files changed (1) hide show
  1. README.md +18 -11
README.md CHANGED
@@ -12,7 +12,6 @@ quantized_by: shaowenchen
12
  tasks:
13
  - text2text-generation
14
  tags:
15
- - meta
16
  - gguf
17
  - llama
18
  - llama-2
@@ -42,20 +41,28 @@ tags:
42
  Usage:
43
 
44
  ```bash
45
- docker run --rm -it -p 8000:8000 -v /path/to/models:/models -e MODEL=/models/gguf-model-name.gguf hubimage/llama-cpp-python:latest
46
  ```
47
 
48
  ## Provided images
49
 
50
- | Name | Quant method | Size |
51
- | -------------------------------------------------- | ------------ | ------- |
52
- | `shaowenchen/colossal-llama-2-7b-base-gguf:Q2_K` | Q2_K | 3.68 GB |
53
- | `shaowenchen/colossal-llama-2-7b-base-gguf:Q3_K` | Q3_K | 4.16 GB |
54
- | `shaowenchen/colossal-llama-2-7b-base-gguf:Q3_K_L` | Q3_K_L | 4.46 GB |
55
- | `shaowenchen/colossal-llama-2-7b-base-gguf:Q3_K_S` | Q3_K_S | 3.81 GB |
56
- | `shaowenchen/colossal-llama-2-7b-base-gguf:Q4_0` | Q4_0 | 4.7 GB |
57
- | `shaowenchen/colossal-llama-2-7b-base-gguf:Q4_K` | Q4_K | 4.95 GB |
58
- | `shaowenchen/colossal-llama-2-7b-base-gguf:Q4_K_S` | Q4_K_S | 4.73 GB |
 
 
 
 
 
 
 
 
59
 
60
  Usage:
61
 
 
12
  tasks:
13
  - text2text-generation
14
  tags:
 
15
  - gguf
16
  - llama
17
  - llama-2
 
41
  Usage:
42
 
43
  ```bash
44
+ docker run --rm -it -p 8000:8000 -v /path/to/models:/models -e MODEL=/models/gguf-model-name.gguf shaowenchen/llama-cpp-python:0.2.6
45
  ```
46
 
47
  ## Provided images
48
 
49
+ | Name | Quant method | Compressed Size |
50
+ | -------------------------------------------------- | ------------ | --------------- |
51
+ | `shaowenchen/colossal-llama-2-7b-base-gguf:Q2_K` | Q2_K | 3.68 GB |
52
+ | `shaowenchen/colossal-llama-2-7b-base-gguf:Q3_K` | Q3_K | 4.16 GB |
53
+ | `shaowenchen/colossal-llama-2-7b-base-gguf:Q3_K_L` | Q3_K_L | 4.46 GB |
54
+ | `shaowenchen/colossal-llama-2-7b-base-gguf:Q3_K_S` | Q3_K_S | 3.81 GB |
55
+ | `shaowenchen/colossal-llama-2-7b-base-gguf:Q4_0` | Q4_0 | 4.7 GB |
56
+ | `shaowenchen/colossal-llama-2-7b-base-gguf:Q4_1` | Q4_1 | 5.1 GB |
57
+ | `shaowenchen/colossal-llama-2-7b-base-gguf:Q4_K` | Q4_K | 4.95 GB |
58
+ | `shaowenchen/colossal-llama-2-7b-base-gguf:Q4_K_S` | Q4_K_S | 4.73 GB |
59
+ | `shaowenchen/colossal-llama-2-7b-base-gguf:Q5_0` | Q5_0 | 5.3 GB |
60
+ | `shaowenchen/colossal-llama-2-7b-base-gguf:Q5_1` | Q5_1 | 5.7 GB |
61
+ | `shaowenchen/colossal-llama-2-7b-base-gguf:Q5_K` | Q5_K | 5.5 GB |
62
+ | `shaowenchen/colossal-llama-2-7b-base-gguf:Q5_K_S` | Q5_K_S | 5.3 GB |
63
+ | `shaowenchen/colossal-llama-2-7b-base-gguf:Q6_K` | Q6_K | 6.3 GB |
64
+ | `shaowenchen/colossal-llama-2-7b-base-gguf:Q8_0` | Q8_0 | 8.2 GB |
65
+ | `shaowenchen/colossal-llama-2-7b-base-gguf:full` | full | 14 GB |
66
 
67
  Usage:
68