flydust commited on
Commit
41d221d
1 Parent(s): de82fb5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +67 -77
README.md CHANGED
@@ -6,19 +6,75 @@ tags:
6
  - axolotl
7
  - generated_from_trainer
8
  model-index:
9
- - name: Llama-3.1-8B-Magpie-SFT-GMix-550K
10
  results: []
 
 
11
  ---
12
 
13
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
- should probably proofread and complete it, then remove this comment. -->
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
 
16
  [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
17
  <details><summary>See axolotl config</summary>
18
 
19
  axolotl version: `0.4.1`
20
  ```yaml
21
-
22
  base_model: meta-llama/Meta-Llama-3.1-8B
23
  model_type: LlamaForCausalLM
24
  tokenizer_type: AutoTokenizer
@@ -30,18 +86,13 @@ strict: false
30
  main_process_port: 0
31
 
32
  datasets:
33
- - path: flydust/Magpie-Llama-3-70B-300K-Gemma2-9B
34
  type: sharegpt
35
  conversation: llama3
36
- - path: flydust/Magpie-Reasoning-150K-Gemma2-9B
37
- type: sharegpt
38
- conversation: llama3
39
- - path: flydust/Magpie-100k-Gemma2-9B
40
- type: sharegpt
41
- conversation: llama3
42
- dataset_prepared_path: /data/zhangchen_xu/last_run_prepared
43
  val_set_size: 0.001
44
- output_dir: /data/zhangchen_xu/axolotl_out/Llama-3.1-8B-SFT-GMix-550K
45
 
46
  sequence_len: 8192
47
  sample_packing: true
@@ -51,9 +102,9 @@ pad_to_sequence_len: true
51
  wandb_project: SynDa
52
  wandb_entity:
53
  wandb_watch:
54
- wandb_name: Llama-3.1-8B-Mix-SFT-GMix-550K
55
  wandb_log_model:
56
- hub_model_id: Magpie-Align/Llama-3.1-8B-Magpie-SFT-GMix-550K
57
 
58
  gradient_accumulation_steps: 32
59
  micro_batch_size: 1
@@ -90,65 +141,4 @@ special_tokens:
90
  pad_token: <|end_of_text|>
91
 
92
  ```
93
-
94
- </details><br>
95
-
96
- # Llama-3.1-8B-Magpie-SFT-GMix-550K
97
-
98
- This model is a fine-tuned version of [meta-llama/Meta-Llama-3.1-8B](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B) on the None dataset.
99
- It achieves the following results on the evaluation set:
100
- - Loss: 0.4544
101
-
102
- ## Model description
103
-
104
- More information needed
105
-
106
- ## Intended uses & limitations
107
-
108
- More information needed
109
-
110
- ## Training and evaluation data
111
-
112
- More information needed
113
-
114
- ## Training procedure
115
-
116
- ### Training hyperparameters
117
-
118
- The following hyperparameters were used during training:
119
- - learning_rate: 2e-05
120
- - train_batch_size: 1
121
- - eval_batch_size: 1
122
- - seed: 42
123
- - distributed_type: multi-GPU
124
- - num_devices: 4
125
- - gradient_accumulation_steps: 32
126
- - total_train_batch_size: 128
127
- - total_eval_batch_size: 4
128
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
129
- - lr_scheduler_type: cosine
130
- - lr_scheduler_warmup_steps: 51
131
- - num_epochs: 2
132
-
133
- ### Training results
134
-
135
- | Training Loss | Epoch | Step | Validation Loss |
136
- |:-------------:|:------:|:----:|:---------------:|
137
- | 0.9311 | 0.0038 | 1 | 0.9847 |
138
- | 0.561 | 0.2015 | 53 | 0.5765 |
139
- | 0.4843 | 0.4030 | 106 | 0.5039 |
140
- | 0.4608 | 0.6045 | 159 | 0.4814 |
141
- | 0.4454 | 0.8060 | 212 | 0.4678 |
142
- | 0.4403 | 1.0075 | 265 | 0.4596 |
143
- | 0.3965 | 1.1938 | 318 | 0.4574 |
144
- | 0.3952 | 1.3953 | 371 | 0.4554 |
145
- | 0.3962 | 1.5968 | 424 | 0.4547 |
146
- | 0.3948 | 1.7983 | 477 | 0.4544 |
147
-
148
-
149
- ### Framework versions
150
-
151
- - Transformers 4.45.0.dev0
152
- - Pytorch 2.3.0+cu121
153
- - Datasets 2.19.1
154
- - Tokenizers 0.19.1
 
6
  - axolotl
7
  - generated_from_trainer
8
  model-index:
9
+ - name: MagpieLM-8B-SFT-v0.1
10
  results: []
11
+ datasets:
12
+ - Magpie-Align/MagpieLM-SFT-Data-v0.1
13
  ---
14
 
15
+ ![Magpie](https://cdn-uploads.huggingface.co/production/uploads/653df1323479e9ebbe3eb6cc/FWWILXrAGNwWr52aghV0S.png)
16
+
17
+ # 🐦 MagpieLM-8B-SFT-v0.1
18
+
19
+ Project Web: [https://magpie-align.github.io/](https://magpie-align.github.io/)
20
+
21
+ Arxiv Technical Report: [https://arxiv.org/abs/2406.08464](https://arxiv.org/abs/2406.08464)
22
+
23
+ Codes: [https://github.com/magpie-align/magpie](https://github.com/magpie-align/magpie)
24
+
25
+ ## About This Model
26
+
27
+ *Model full name: Llama3.1-MagpieLM-8B-SFT-v0.1*
28
+
29
+ This model is a fine-tuned version of [meta-llama/Meta-Llama-3.1-8B](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B) on [Magpie-Align/MagpieLM-SFT-Data-v0.1](https://huggingface.co/datasets/Magpie-Align/MagpieLM-SFT-Data-v0.1) dataset.
30
+
31
+ ## Training procedure
32
+
33
+ ### Training hyperparameters
34
+
35
+ The following hyperparameters were used during training:
36
+ - learning_rate: 2e-05
37
+ - train_batch_size: 1
38
+ - eval_batch_size: 1
39
+ - seed: 42
40
+ - distributed_type: multi-GPU
41
+ - num_devices: 4
42
+ - gradient_accumulation_steps: 32
43
+ - total_train_batch_size: 128
44
+ - total_eval_batch_size: 4
45
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
+ - lr_scheduler_type: cosine
47
+ - lr_scheduler_warmup_steps: 51
48
+ - num_epochs: 2
49
+
50
+ ### Training results
51
+
52
+ | Training Loss | Epoch | Step | Validation Loss |
53
+ |:-------------:|:------:|:----:|:---------------:|
54
+ | 0.9311 | 0.0038 | 1 | 0.9847 |
55
+ | 0.561 | 0.2015 | 53 | 0.5765 |
56
+ | 0.4843 | 0.4030 | 106 | 0.5039 |
57
+ | 0.4608 | 0.6045 | 159 | 0.4814 |
58
+ | 0.4454 | 0.8060 | 212 | 0.4678 |
59
+ | 0.4403 | 1.0075 | 265 | 0.4596 |
60
+ | 0.3965 | 1.1938 | 318 | 0.4574 |
61
+ | 0.3952 | 1.3953 | 371 | 0.4554 |
62
+ | 0.3962 | 1.5968 | 424 | 0.4547 |
63
+ | 0.3948 | 1.7983 | 477 | 0.4544 |
64
+
65
+
66
+ ### Framework versions
67
+
68
+ - Transformers 4.45.0.dev0
69
+ - Pytorch 2.3.0+cu121
70
+ - Datasets 2.19.1
71
+ - Tokenizers 0.19.1
72
 
73
  [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
74
  <details><summary>See axolotl config</summary>
75
 
76
  axolotl version: `0.4.1`
77
  ```yaml
 
78
  base_model: meta-llama/Meta-Llama-3.1-8B
79
  model_type: LlamaForCausalLM
80
  tokenizer_type: AutoTokenizer
 
86
  main_process_port: 0
87
 
88
  datasets:
89
+ - path: Magpie-Align/MagpieLM-SFT-Data-v0.1
90
  type: sharegpt
91
  conversation: llama3
92
+
93
+ dataset_prepared_path: last_run_prepared
 
 
 
 
 
94
  val_set_size: 0.001
95
+ output_dir: axolotl_out/MagpieLM-8B-SFT-v0.1
96
 
97
  sequence_len: 8192
98
  sample_packing: true
 
102
  wandb_project: SynDa
103
  wandb_entity:
104
  wandb_watch:
105
+ wandb_name: MagpieLM-8B-SFT-v0.1
106
  wandb_log_model:
107
+ hub_model_id: Magpie-Align/MagpieLM-8B-SFT-v0.1
108
 
109
  gradient_accumulation_steps: 32
110
  micro_batch_size: 1
 
141
  pad_token: <|end_of_text|>
142
 
143
  ```
144
+ </details><br>