anyiwang commited on
Commit
b766b8a
1 Parent(s): f1cfad1

End of training

Browse files
README.md CHANGED
@@ -41,6 +41,11 @@ The following hyperparameters were used during training:
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: cosine
43
  - num_epochs: 5
 
 
 
 
 
44
 
45
  ### Framework versions
46
 
 
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: cosine
43
  - num_epochs: 5
44
+ - mixed_precision_training: Native AMP
45
+
46
+ ### Training results
47
+
48
+
49
 
50
  ### Framework versions
51
 
adapter_config.json CHANGED
@@ -19,13 +19,13 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "lm_head",
23
  "k_proj",
24
  "q_proj",
25
- "gate_proj",
26
- "up_proj",
27
  "o_proj",
28
  "v_proj",
 
29
  "down_proj"
30
  ],
31
  "task_type": "CAUSAL_LM"
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "gate_proj",
23
  "k_proj",
24
  "q_proj",
25
+ "lm_head",
 
26
  "o_proj",
27
  "v_proj",
28
+ "up_proj",
29
  "down_proj"
30
  ],
31
  "task_type": "CAUSAL_LM"
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:091d01db986b82b92baec6e38e1e57fcecacb667928926185a1d2a60ab9799c8
3
  size 694431312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a006b07568e179a86e4b40b4cb84fc5e6e8113c97350a1d40dbe76180c53f67
3
  size 694431312
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:22aa1809070ec5ae84dd334a6296b3604ca00810b74ceef5aaef78ce9a9a5d30
3
  size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5318028c0a68bbc6499df23837ecfc810922c90fd4962caa2bdba56df140162e
3
  size 4728