TheBloke commited on
Commit
a693e20
1 Parent(s): 4c1ae8d

Initial GPTQ model commit

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -58,7 +58,7 @@ from auto_gptq import AutoGPTQForCausalLM, BaseQuantizeConfig
58
  import argparse
59
 
60
  model_name_or_path = "TheBloke/samantha-1.1-llama-33B-GPTQ"
61
- model_basename = "gptq_model-4bit-128g"
62
 
63
  use_triton = False
64
 
@@ -103,17 +103,17 @@ print(pipe(prompt_template)[0]['generated_text'])
103
 
104
  ## Provided files
105
 
106
- **gptq_model-4bit-128g.safetensors**
107
 
108
  This will work with AutoGPTQ and CUDA versions of GPTQ-for-LLaMa. There are reports of issues with Triton mode of recent GPTQ-for-LLaMa. If you have issues, please use AutoGPTQ instead.
109
 
110
- It was created with group_size 128 to increase inference accuracy, but without --act-order (desc_act) to increase compatibility and improve inference speed.
111
 
112
- * `gptq_model-4bit-128g.safetensors`
113
  * Works with AutoGPTQ in CUDA or Triton modes.
114
  * Works with GPTQ-for-LLaMa in CUDA mode. May have issues with GPTQ-for-LLaMa Triton mode.
115
  * Works with text-generation-webui, including one-click-installers.
116
- * Parameters: Groupsize = 128. Act Order / desc_act = False.
117
 
118
  <!-- footer start -->
119
  ## Discord
 
58
  import argparse
59
 
60
  model_name_or_path = "TheBloke/samantha-1.1-llama-33B-GPTQ"
61
+ model_basename = "samantha-1.1-llama-33b-GPTQ-4bit--1g.act.order"
62
 
63
  use_triton = False
64
 
 
103
 
104
  ## Provided files
105
 
106
+ **samantha-1.1-llama-33b-GPTQ-4bit--1g.act.order.safetensors**
107
 
108
  This will work with AutoGPTQ and CUDA versions of GPTQ-for-LLaMa. There are reports of issues with Triton mode of recent GPTQ-for-LLaMa. If you have issues, please use AutoGPTQ instead.
109
 
110
+ It was created without group_size to lower VRAM requirements, and with --act-order (desc_act) to boost inference accuracy as much as possible.
111
 
112
+ * `samantha-1.1-llama-33b-GPTQ-4bit--1g.act.order.safetensors`
113
  * Works with AutoGPTQ in CUDA or Triton modes.
114
  * Works with GPTQ-for-LLaMa in CUDA mode. May have issues with GPTQ-for-LLaMa Triton mode.
115
  * Works with text-generation-webui, including one-click-installers.
116
+ * Parameters: Groupsize = -1. Act Order / desc_act = True.
117
 
118
  <!-- footer start -->
119
  ## Discord