supermy commited on
Commit
8aceac5
1 Parent(s): 938a7ad

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +35 -34
README.md CHANGED
@@ -65,44 +65,45 @@ model = AutoModelForCausalLM.from_pretrained("supermy/couplet")
65
 
66
  bpe分词:"vocab_size"=50000
67
  ```
68
- [INFO|trainer.py:1608] 2022-11-29 16:00:16,391 >> ***** Running training *****
69
- [INFO|trainer.py:1609] 2022-11-29 16:00:16,391 >> Num examples = 249327
70
- [INFO|trainer.py:1610] 2022-11-29 16:00:16,391 >> Num Epochs = 38
71
- [INFO|trainer.py:1611] 2022-11-29 16:00:16,391 >> Instantaneous batch size per device = 96
72
- [INFO|trainer.py:1612] 2022-11-29 16:00:16,391 >> Total train batch size (w. parallel, distributed & accumulation) = 96
73
- [INFO|trainer.py:1613] 2022-11-29 16:00:16,391 >> Gradient Accumulation steps = 1
74
- [INFO|trainer.py:1614] 2022-11-29 16:00:16,391 >> Total optimization steps = 98724
75
- [INFO|trainer.py:1616] 2022-11-29 16:00:16,392 >> Number of trainable parameters = 124439808
76
-
77
- {'loss': 6.4109, 'learning_rate': 4.975031400672582e-05, 'epoch': 0.19}
78
- {'loss': 5.8476, 'learning_rate': 4.9497082776224627e-05, 'epoch': 0.38}
 
79
  ......
80
  ......
81
  ......
82
- {'loss': 3.4331, 'learning_rate': 1.3573193954864066e-07, 'epoch': 37.91}
83
- {'train_runtime': 65776.233, 'train_samples_per_second': 144.04, 'train_steps_per_second': 1.501, 'train_loss': 3.74187503763847, 'epoch': 38.0}
84
  ***** train metrics *****
85
- epoch = 38.0
86
- train_loss = 3.7419
87
- train_runtime = 18:16:16.23
88
- train_samples = 249327
89
- train_samples_per_second = 144.04
90
- train_steps_per_second = 1.501
91
- 11/30/2022 10:16:35 - INFO - __main__ - *** Evaluate ***
92
- [INFO|trainer.py:2929] 2022-11-30 10:16:35,902 >> ***** Running Evaluation *****
93
- [INFO|trainer.py:2931] 2022-11-30 10:16:35,902 >> Num examples = 1290
94
- [INFO|trainer.py:2934] 2022-11-30 10:16:35,902 >> Batch size = 96
95
- 100%|██████████| 14/14 [00:03<00:00, 4.13it/s]
96
- [INFO|modelcard.py:449] 2022-11-30 10:16:40,821 >> Dropping the following result as it does not have all the necessary fields:
97
- {'task': {'name': 'Causal Language Modeling', 'type': 'text-generation'}, 'metrics': [{'name': 'Accuracy', 'type': 'accuracy', 'value': 0.39426602682416634}]}
98
  ***** eval metrics *****
99
- epoch = 38.0
100
- eval_accuracy = 0.3943
101
- eval_loss = 3.546
102
- eval_runtime = 0:00:03.67
103
- eval_samples = 1290
104
- eval_samples_per_second = 351.199
105
- eval_steps_per_second = 3.811
106
- perplexity = 34.6733
107
 
108
  ```
 
65
 
66
  bpe分词:"vocab_size"=50000
67
  ```
68
+ [INFO|trainer.py:1608] 2022-11-30 12:51:36,357 >> ***** Running training *****
69
+ [INFO|trainer.py:1609] 2022-11-30 12:51:36,357 >> Num examples = 260926
70
+ [INFO|trainer.py:1610] 2022-11-30 12:51:36,357 >> Num Epochs = 81
71
+ [INFO|trainer.py:1611] 2022-11-30 12:51:36,357 >> Instantaneous batch size per device = 96
72
+ [INFO|trainer.py:1612] 2022-11-30 12:51:36,357 >> Total train batch size (w. parallel, distributed & accumulation) = 96
73
+ [INFO|trainer.py:1613] 2022-11-30 12:51:36,357 >> Gradient Accumulation steps = 1
74
+ [INFO|trainer.py:1614] 2022-11-30 12:51:36,357 >> Total optimization steps = 220158
75
+ [INFO|trainer.py:1616] 2022-11-30 12:51:36,358 >> Number of trainable parameters = 124439808
76
+
77
+ {'loss': 6.1104, 'learning_rate': 4.9888034956712906e-05, 'epoch': 0.18}
78
+ {'loss': 5.5855, 'learning_rate': 4.977448014607691e-05, 'epoch': 0.37}
79
+ {'loss': 5.3264, 'learning_rate': 4.966092533544091e-05, 'epoch': 0.55}
80
  ......
81
  ......
82
  ......
83
+ {'loss': 2.8539, 'learning_rate': 5.677740531799889e-08, 'epoch': 80.94}
84
+ {'train_runtime': 146835.0563, 'train_samples_per_second': 143.937, 'train_steps_per_second': 1.499, 'train_loss': 3.1762605669072217, 'epoch': 81.0}
85
  ***** train metrics *****
86
+ epoch = 81.0
87
+ train_loss = 3.1763
88
+ train_runtime = 1 day, 16:47:15.05
89
+ train_samples = 260926
90
+ train_samples_per_second = 143.937
91
+ train_steps_per_second = 1.499
92
+ 12/02/2022 05:38:54 - INFO - __main__ - *** Evaluate ***
93
+ [INFO|trainer.py:2929] 2022-12-02 05:38:54,688 >> ***** Running Evaluation *****
94
+ [INFO|trainer.py:2931] 2022-12-02 05:38:54,688 >> Num examples = 1350
95
+ [INFO|trainer.py:2934] 2022-12-02 05:38:54,688 >> Batch size = 96
96
+ 100%|██████████| 15/15 [00:03<00:00, 4.20it/s]
97
+ [INFO|modelcard.py:449] 2022-12-02 05:38:59,875 >> Dropping the following result as it does not have all the necessary fields:
98
+ {'task': {'name': 'Causal Language Modeling', 'type': 'text-generation'}, 'metrics': [{'name': 'Accuracy', 'type': 'accuracy', 'value': 0.4447501469723692}]}
99
  ***** eval metrics *****
100
+ epoch = 81.0
101
+ eval_accuracy = 0.4448
102
+ eval_loss = 3.2813
103
+ eval_runtime = 0:00:03.86
104
+ eval_samples = 1350
105
+ eval_samples_per_second = 349.505
106
+ eval_steps_per_second = 3.883
107
+ perplexity = 26.6108
108
 
109
  ```