flydust commited on
Commit
16a6ec1
1 Parent(s): 4b25c9d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -15
README.md CHANGED
@@ -1,6 +1,5 @@
1
  ---
2
  library_name: transformers
3
- license: other
4
  base_model: nvidia/Llama-3.1-Minitron-4B-Width-Base
5
  tags:
6
  - axolotl
@@ -8,6 +7,10 @@ tags:
8
  model-index:
9
  - name: MagpieLM-4B-SFT-v0.1
10
  results: []
 
 
 
 
11
  ---
12
 
13
  ![Magpie](https://cdn-uploads.huggingface.co/production/uploads/653df1323479e9ebbe3eb6cc/FWWILXrAGNwWr52aghV0S.png)
@@ -24,11 +27,7 @@ Codes: [https://github.com/magpie-align/magpie](https://github.com/magpie-align/
24
 
25
  *Model full name: Llama3.1-MagpieLM-4B-SFT-v0.1*
26
 
27
- This model is a fine-tuned version of [nvidia/Llama-3.1-Minitron-4B-Width-Base](https://huggingface.co/nvidia/Llama-3.1-Minitron-4B-Width-Base) on
28
- - [Magpie-Align/Magpie-Llama-3.1-Pro-MT-500K-v0.1](https://huggingface.co/datasets/Magpie-Align/Magpie-Llama-3.1-Pro-MT-500K-v0.1), and
29
- - [Magpie-Align/Magpie-Reasoning-150K](https://huggingface.co/datasets/Magpie-Align/Magpie-Reasoning-150K).
30
-
31
- ## Training procedure
32
 
33
  ## Training procedure
34
 
@@ -89,15 +88,9 @@ datasets:
89
  - path: flydust/Magpie-Llama-3-70B-300K-Gemma2-9B
90
  type: sharegpt
91
  conversation: llama3
92
- - path: flydust/Magpie-Reasoning-150K-Gemma2-9B
93
- type: sharegpt
94
- conversation: llama3
95
- - path: flydust/Magpie-100k-Gemma2-9B
96
- type: sharegpt
97
- conversation: llama3
98
  dataset_prepared_path: last_run_prepared
99
  val_set_size: 0.001
100
- output_dir: /data/zhangchen_xu/axolotl_out/Minitron-4B-Magpie-Gemma2-9B-550K
101
 
102
  sequence_len: 8192
103
  sample_packing: true
@@ -109,7 +102,7 @@ wandb_entity:
109
  wandb_watch:
110
  wandb_name: Llama3.1-MagpieLM-4B-SFT-v0.1
111
  wandb_log_model:
112
- hub_model_id: Magpie-Align/Llama3.1-MagpieLM-4B-SFT-v0.1
113
 
114
  gradient_accumulation_steps: 32
115
  micro_batch_size: 1
@@ -150,4 +143,4 @@ special_tokens:
150
 
151
  ```
152
 
153
- </details><br>
 
1
  ---
2
  library_name: transformers
 
3
  base_model: nvidia/Llama-3.1-Minitron-4B-Width-Base
4
  tags:
5
  - axolotl
 
7
  model-index:
8
  - name: MagpieLM-4B-SFT-v0.1
9
  results: []
10
+ datasets:
11
+ - Magpie-Align/MagpieLM-4B-SFT
12
+ language:
13
+ - en
14
  ---
15
 
16
  ![Magpie](https://cdn-uploads.huggingface.co/production/uploads/653df1323479e9ebbe3eb6cc/FWWILXrAGNwWr52aghV0S.png)
 
27
 
28
  *Model full name: Llama3.1-MagpieLM-4B-SFT-v0.1*
29
 
30
+ This model is a fine-tuned version of [nvidia/Llama-3.1-Minitron-4B-Width-Base](https://huggingface.co/nvidia/Llama-3.1-Minitron-4B-Width-Base) on [Magpie-Align/MagpieLM-4B-SFT](https://huggingface.co/datasets/Magpie-Align/Magpie-Align/MagpieLM-4B-SFT).
 
 
 
 
31
 
32
  ## Training procedure
33
 
 
88
  - path: flydust/Magpie-Llama-3-70B-300K-Gemma2-9B
89
  type: sharegpt
90
  conversation: llama3
 
 
 
 
 
 
91
  dataset_prepared_path: last_run_prepared
92
  val_set_size: 0.001
93
+ output_dir: axolotl_out/MagpieLM-4B-SFT-v0.1
94
 
95
  sequence_len: 8192
96
  sample_packing: true
 
102
  wandb_watch:
103
  wandb_name: Llama3.1-MagpieLM-4B-SFT-v0.1
104
  wandb_log_model:
105
+ hub_model_id: Magpie-Align/MagpieLM-4B-SFT-v0.1
106
 
107
  gradient_accumulation_steps: 32
108
  micro_batch_size: 1
 
143
 
144
  ```
145
 
146
+ </details><br>