Update README.md
Browse files
README.md
CHANGED
@@ -1,6 +1,5 @@
|
|
1 |
---
|
2 |
library_name: transformers
|
3 |
-
license: other
|
4 |
base_model: nvidia/Llama-3.1-Minitron-4B-Width-Base
|
5 |
tags:
|
6 |
- axolotl
|
@@ -8,6 +7,10 @@ tags:
|
|
8 |
model-index:
|
9 |
- name: MagpieLM-4B-SFT-v0.1
|
10 |
results: []
|
|
|
|
|
|
|
|
|
11 |
---
|
12 |
|
13 |
![Magpie](https://cdn-uploads.huggingface.co/production/uploads/653df1323479e9ebbe3eb6cc/FWWILXrAGNwWr52aghV0S.png)
|
@@ -24,11 +27,7 @@ Codes: [https://github.com/magpie-align/magpie](https://github.com/magpie-align/
|
|
24 |
|
25 |
*Model full name: Llama3.1-MagpieLM-4B-SFT-v0.1*
|
26 |
|
27 |
-
This model is a fine-tuned version of [nvidia/Llama-3.1-Minitron-4B-Width-Base](https://huggingface.co/nvidia/Llama-3.1-Minitron-4B-Width-Base) on
|
28 |
-
- [Magpie-Align/Magpie-Llama-3.1-Pro-MT-500K-v0.1](https://huggingface.co/datasets/Magpie-Align/Magpie-Llama-3.1-Pro-MT-500K-v0.1), and
|
29 |
-
- [Magpie-Align/Magpie-Reasoning-150K](https://huggingface.co/datasets/Magpie-Align/Magpie-Reasoning-150K).
|
30 |
-
|
31 |
-
## Training procedure
|
32 |
|
33 |
## Training procedure
|
34 |
|
@@ -89,15 +88,9 @@ datasets:
|
|
89 |
- path: flydust/Magpie-Llama-3-70B-300K-Gemma2-9B
|
90 |
type: sharegpt
|
91 |
conversation: llama3
|
92 |
-
- path: flydust/Magpie-Reasoning-150K-Gemma2-9B
|
93 |
-
type: sharegpt
|
94 |
-
conversation: llama3
|
95 |
-
- path: flydust/Magpie-100k-Gemma2-9B
|
96 |
-
type: sharegpt
|
97 |
-
conversation: llama3
|
98 |
dataset_prepared_path: last_run_prepared
|
99 |
val_set_size: 0.001
|
100 |
-
output_dir:
|
101 |
|
102 |
sequence_len: 8192
|
103 |
sample_packing: true
|
@@ -109,7 +102,7 @@ wandb_entity:
|
|
109 |
wandb_watch:
|
110 |
wandb_name: Llama3.1-MagpieLM-4B-SFT-v0.1
|
111 |
wandb_log_model:
|
112 |
-
hub_model_id: Magpie-Align/
|
113 |
|
114 |
gradient_accumulation_steps: 32
|
115 |
micro_batch_size: 1
|
@@ -150,4 +143,4 @@ special_tokens:
|
|
150 |
|
151 |
```
|
152 |
|
153 |
-
</details><br>
|
|
|
1 |
---
|
2 |
library_name: transformers
|
|
|
3 |
base_model: nvidia/Llama-3.1-Minitron-4B-Width-Base
|
4 |
tags:
|
5 |
- axolotl
|
|
|
7 |
model-index:
|
8 |
- name: MagpieLM-4B-SFT-v0.1
|
9 |
results: []
|
10 |
+
datasets:
|
11 |
+
- Magpie-Align/MagpieLM-4B-SFT
|
12 |
+
language:
|
13 |
+
- en
|
14 |
---
|
15 |
|
16 |
![Magpie](https://cdn-uploads.huggingface.co/production/uploads/653df1323479e9ebbe3eb6cc/FWWILXrAGNwWr52aghV0S.png)
|
|
|
27 |
|
28 |
*Model full name: Llama3.1-MagpieLM-4B-SFT-v0.1*
|
29 |
|
30 |
+
This model is a fine-tuned version of [nvidia/Llama-3.1-Minitron-4B-Width-Base](https://huggingface.co/nvidia/Llama-3.1-Minitron-4B-Width-Base) on [Magpie-Align/MagpieLM-4B-SFT](https://huggingface.co/datasets/Magpie-Align/Magpie-Align/MagpieLM-4B-SFT).
|
|
|
|
|
|
|
|
|
31 |
|
32 |
## Training procedure
|
33 |
|
|
|
88 |
- path: flydust/Magpie-Llama-3-70B-300K-Gemma2-9B
|
89 |
type: sharegpt
|
90 |
conversation: llama3
|
|
|
|
|
|
|
|
|
|
|
|
|
91 |
dataset_prepared_path: last_run_prepared
|
92 |
val_set_size: 0.001
|
93 |
+
output_dir: axolotl_out/MagpieLM-4B-SFT-v0.1
|
94 |
|
95 |
sequence_len: 8192
|
96 |
sample_packing: true
|
|
|
102 |
wandb_watch:
|
103 |
wandb_name: Llama3.1-MagpieLM-4B-SFT-v0.1
|
104 |
wandb_log_model:
|
105 |
+
hub_model_id: Magpie-Align/MagpieLM-4B-SFT-v0.1
|
106 |
|
107 |
gradient_accumulation_steps: 32
|
108 |
micro_batch_size: 1
|
|
|
143 |
|
144 |
```
|
145 |
|
146 |
+
</details><br>
|