quyanh commited on
Commit
d7334f3
1 Parent(s): 0306e62

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -18,7 +18,6 @@ base_model: Qwen/Qwen-7B
18
 
19
 
20
  - **Developed by:** [More Information Needed]
21
- - **Funded by [optional]:** [More Information Needed]
22
  - **Shared by [optional]:** [More Information Needed]
23
  - **Model type:** [More Information Needed]
24
  - **Language(s) (NLP):** [More Information Needed]
@@ -77,7 +76,7 @@ Use the code below to get started with the model.
77
 
78
  ### Training Data
79
 
80
- <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
 
82
  [More Information Needed]
83
 
@@ -108,7 +107,7 @@ Use the code below to get started with the model.
108
 
109
  #### Testing Data
110
 
111
- <!-- This should link to a Dataset Card if possible. -->
112
 
113
  [More Information Needed]
114
 
 
18
 
19
 
20
  - **Developed by:** [More Information Needed]
 
21
  - **Shared by [optional]:** [More Information Needed]
22
  - **Model type:** [More Information Needed]
23
  - **Language(s) (NLP):** [More Information Needed]
 
76
 
77
  ### Training Data
78
 
79
+ <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
80
 
81
  [More Information Needed]
82
 
 
107
 
108
  #### Testing Data
109
 
110
+ <!-- This should link to a Data Card if possible. -->
111
 
112
  [More Information Needed]
113
 
adapter_config.json CHANGED
@@ -8,7 +8,7 @@
8
  "init_lora_weights": true,
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
- "lora_alpha": 32,
12
  "lora_dropout": 0.05,
13
  "modules_to_save": null,
14
  "peft_type": "LORA",
@@ -17,13 +17,13 @@
17
  "revision": null,
18
  "target_modules": [
19
  "up_proj",
20
- "down_proj",
21
- "q_proj",
22
  "gate_proj",
 
 
23
  "o_proj",
24
- "v_proj",
25
- "k_proj",
26
- "lm_head"
27
  ],
28
  "task_type": "CAUSAL_LM"
29
  }
 
8
  "init_lora_weights": true,
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
+ "lora_alpha": 16,
12
  "lora_dropout": 0.05,
13
  "modules_to_save": null,
14
  "peft_type": "LORA",
 
17
  "revision": null,
18
  "target_modules": [
19
  "up_proj",
20
+ "k_proj",
21
+ "v_proj",
22
  "gate_proj",
23
+ "q_proj",
24
+ "lm_head",
25
  "o_proj",
26
+ "down_proj"
 
 
27
  ],
28
  "task_type": "CAUSAL_LM"
29
  }
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c9efca0a41b1fa0a7d2c73a7c0668d7fdd266dbc38b22073b3785ec1a03d107
3
  size 4994175
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e03393f30be6aabf595239c99982eb0834fc0e03b87b1be304f49b2e1adac248
3
  size 4994175
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:77e7334e269ed3f118133b4e9a64cbdd8962344cc1a926a089fc1f3a3aaa49ce
3
  size 2506975
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca6bf3778c5a5dbfd5eb20342f741f20dfcb3e7f387f57f5e2b7141f7576e46b
3
  size 2506975
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0eb21a3129c20b585a208f88d463aa0b3f24c0c1cd36c5d747317950a557c052
3
  size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ecd26e0c4b589f9cc8af39ffd4ca0913eaa05895d20110af3d5ab7e80d59877
3
  size 14575
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:37b925d914a427f9d777adc61a9630c02045830538343e941aba5fd083fe36b9
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:933395f7fd252489a558205952ea72f3d20cc20b5ccab7087d643266a68fdfff
3
  size 627
trainer_state.json CHANGED
@@ -1,75 +1,169 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.21447721179624665,
5
- "eval_steps": 20,
6
- "global_step": 80,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.05,
13
  "learning_rate": 1.9393939393939395e-05,
14
- "loss": 2.0524,
15
  "step": 20
16
  },
17
  {
18
- "epoch": 0.05,
19
- "eval_loss": 2.070744514465332,
20
- "eval_runtime": 124.4779,
21
- "eval_samples_per_second": 2.635,
22
- "eval_steps_per_second": 1.318,
23
- "step": 20
24
- },
25
- {
26
- "epoch": 0.11,
27
  "learning_rate": 1.8585858585858588e-05,
28
- "loss": 2.059,
29
  "step": 40
30
  },
31
  {
32
- "epoch": 0.11,
33
- "eval_loss": 2.068726062774658,
34
- "eval_runtime": 124.4655,
35
- "eval_samples_per_second": 2.635,
36
- "eval_steps_per_second": 1.318,
37
- "step": 40
38
- },
39
- {
40
- "epoch": 0.16,
41
  "learning_rate": 1.7777777777777777e-05,
42
- "loss": 2.0337,
43
  "step": 60
44
  },
45
  {
46
- "epoch": 0.16,
47
- "eval_loss": 2.066453218460083,
48
- "eval_runtime": 124.4729,
49
- "eval_samples_per_second": 2.635,
50
- "eval_steps_per_second": 1.318,
51
- "step": 60
52
- },
53
- {
54
- "epoch": 0.21,
55
  "learning_rate": 1.6969696969696972e-05,
56
- "loss": 2.0875,
57
  "step": 80
58
  },
59
  {
60
  "epoch": 0.21,
61
- "eval_loss": 2.063451051712036,
62
- "eval_runtime": 124.438,
63
- "eval_samples_per_second": 2.636,
64
- "eval_steps_per_second": 1.318,
65
- "step": 80
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
66
  }
67
  ],
68
  "logging_steps": 20,
69
  "max_steps": 500,
70
  "num_train_epochs": 2,
71
  "save_steps": 20,
72
- "total_flos": 3.037389329006592e+16,
73
  "trial_name": null,
74
  "trial_params": null
75
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.0621348911311737,
5
+ "eval_steps": 500,
6
+ "global_step": 500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.04,
13
  "learning_rate": 1.9393939393939395e-05,
14
+ "loss": 1.9406,
15
  "step": 20
16
  },
17
  {
18
+ "epoch": 0.08,
 
 
 
 
 
 
 
 
19
  "learning_rate": 1.8585858585858588e-05,
20
+ "loss": 1.9925,
21
  "step": 40
22
  },
23
  {
24
+ "epoch": 0.13,
 
 
 
 
 
 
 
 
25
  "learning_rate": 1.7777777777777777e-05,
26
+ "loss": 1.9952,
27
  "step": 60
28
  },
29
  {
30
+ "epoch": 0.17,
 
 
 
 
 
 
 
 
31
  "learning_rate": 1.6969696969696972e-05,
32
+ "loss": 1.9616,
33
  "step": 80
34
  },
35
  {
36
  "epoch": 0.21,
37
+ "learning_rate": 1.616161616161616e-05,
38
+ "loss": 2.0246,
39
+ "step": 100
40
+ },
41
+ {
42
+ "epoch": 0.25,
43
+ "learning_rate": 1.5353535353535354e-05,
44
+ "loss": 1.9661,
45
+ "step": 120
46
+ },
47
+ {
48
+ "epoch": 0.3,
49
+ "learning_rate": 1.4545454545454546e-05,
50
+ "loss": 1.9318,
51
+ "step": 140
52
+ },
53
+ {
54
+ "epoch": 0.34,
55
+ "learning_rate": 1.3737373737373739e-05,
56
+ "loss": 1.9104,
57
+ "step": 160
58
+ },
59
+ {
60
+ "epoch": 0.38,
61
+ "learning_rate": 1.2929292929292931e-05,
62
+ "loss": 1.9786,
63
+ "step": 180
64
+ },
65
+ {
66
+ "epoch": 0.42,
67
+ "learning_rate": 1.2121212121212122e-05,
68
+ "loss": 1.9611,
69
+ "step": 200
70
+ },
71
+ {
72
+ "epoch": 0.47,
73
+ "learning_rate": 1.1313131313131314e-05,
74
+ "loss": 1.9473,
75
+ "step": 220
76
+ },
77
+ {
78
+ "epoch": 0.51,
79
+ "learning_rate": 1.0505050505050507e-05,
80
+ "loss": 1.9428,
81
+ "step": 240
82
+ },
83
+ {
84
+ "epoch": 0.55,
85
+ "learning_rate": 9.696969696969698e-06,
86
+ "loss": 1.899,
87
+ "step": 260
88
+ },
89
+ {
90
+ "epoch": 0.59,
91
+ "learning_rate": 8.888888888888888e-06,
92
+ "loss": 1.9273,
93
+ "step": 280
94
+ },
95
+ {
96
+ "epoch": 0.64,
97
+ "learning_rate": 8.08080808080808e-06,
98
+ "loss": 1.9628,
99
+ "step": 300
100
+ },
101
+ {
102
+ "epoch": 0.68,
103
+ "learning_rate": 7.272727272727273e-06,
104
+ "loss": 1.9621,
105
+ "step": 320
106
+ },
107
+ {
108
+ "epoch": 0.72,
109
+ "learning_rate": 6.464646464646466e-06,
110
+ "loss": 1.949,
111
+ "step": 340
112
+ },
113
+ {
114
+ "epoch": 0.76,
115
+ "learning_rate": 5.656565656565657e-06,
116
+ "loss": 1.9522,
117
+ "step": 360
118
+ },
119
+ {
120
+ "epoch": 0.81,
121
+ "learning_rate": 4.848484848484849e-06,
122
+ "loss": 1.9618,
123
+ "step": 380
124
+ },
125
+ {
126
+ "epoch": 0.85,
127
+ "learning_rate": 4.04040404040404e-06,
128
+ "loss": 1.9464,
129
+ "step": 400
130
+ },
131
+ {
132
+ "epoch": 0.89,
133
+ "learning_rate": 3.232323232323233e-06,
134
+ "loss": 1.9332,
135
+ "step": 420
136
+ },
137
+ {
138
+ "epoch": 0.93,
139
+ "learning_rate": 2.4242424242424244e-06,
140
+ "loss": 1.9209,
141
+ "step": 440
142
+ },
143
+ {
144
+ "epoch": 0.98,
145
+ "learning_rate": 1.6161616161616164e-06,
146
+ "loss": 1.9663,
147
+ "step": 460
148
+ },
149
+ {
150
+ "epoch": 1.02,
151
+ "learning_rate": 8.080808080808082e-07,
152
+ "loss": 1.9469,
153
+ "step": 480
154
+ },
155
+ {
156
+ "epoch": 1.06,
157
+ "learning_rate": 0.0,
158
+ "loss": 1.947,
159
+ "step": 500
160
  }
161
  ],
162
  "logging_steps": 20,
163
  "max_steps": 500,
164
  "num_train_epochs": 2,
165
  "save_steps": 20,
166
+ "total_flos": 1.89836833062912e+17,
167
  "trial_name": null,
168
  "trial_params": null
169
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3180999a8d8b01de32f6df7932aa3def96692646eef435d7c634bd0d994fee56
3
  size 4027
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:314fddf569f46a3933ef800aa8a159837c4220c59256ba35a8c9aa88331d62c3
3
  size 4027