aalleavitch commited on
Commit
fe28c29
1 Parent(s): 6968119

Upload MistralForCausalLM

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "unsloth/Mistral-Small-Instruct-2409-bnb-4bit",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
@@ -18,7 +18,10 @@
18
  "num_key_value_heads": 8,
19
  "pad_token_id": 750,
20
  "quantization_config": {
 
 
21
  "bnb_4bit_compute_dtype": "bfloat16",
 
22
  "bnb_4bit_quant_type": "nf4",
23
  "bnb_4bit_use_double_quant": true,
24
  "llm_int8_enable_fp32_cpu_offload": false,
@@ -34,7 +37,7 @@
34
  "rope_theta": 1000000.0,
35
  "sliding_window": null,
36
  "tie_word_embeddings": false,
37
- "torch_dtype": "bfloat16",
38
  "transformers_version": "4.44.2",
39
  "unsloth_version": "2024.9.post4",
40
  "use_cache": true,
 
1
  {
2
+ "_name_or_path": "HobieLLM-merge4",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
 
18
  "num_key_value_heads": 8,
19
  "pad_token_id": 750,
20
  "quantization_config": {
21
+ "_load_in_4bit": true,
22
+ "_load_in_8bit": false,
23
  "bnb_4bit_compute_dtype": "bfloat16",
24
+ "bnb_4bit_quant_storage": "uint8",
25
  "bnb_4bit_quant_type": "nf4",
26
  "bnb_4bit_use_double_quant": true,
27
  "llm_int8_enable_fp32_cpu_offload": false,
 
37
  "rope_theta": 1000000.0,
38
  "sliding_window": null,
39
  "tie_word_embeddings": false,
40
+ "torch_dtype": "float16",
41
  "transformers_version": "4.44.2",
42
  "unsloth_version": "2024.9.post4",
43
  "use_cache": true,
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e0cb5938436da9b2db7dca6b99e74ddd95d9fea9352902ecc552c77a27a8768
3
- size 4979766316
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0a671f9e5a1ae768412cf25844bd9725f3f4324cba17b38a576962e93692981
3
+ size 4979766219
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b35a02483e36215c157d910ea9566c47787728bd5f0e3358ed4b8d6d684a3d0b
3
- size 4979665814
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5da0d662e51dcc09667e6daf6128c848866635321d69b35a2599f3e89d07fc6f
3
+ size 4979665734
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:74385fbb4d73f3f8b71545d82dfdcd51bd601ca74db158ebd0e0e36a89147783
3
- size 2116663382
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be539c4837dd4f14a7aa2f0555fdb2f915df5224006e40831618103a344d0d17
3
+ size 2116663356
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 12075801064
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00003-of-00003.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 12075800989
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00003-of-00003.safetensors",