uukuguy commited on
Commit
f7b1f87
1 Parent(s): f431c04

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +43 -37
README.md CHANGED
@@ -24,7 +24,7 @@ model-index:
24
  metrics:
25
  - name: pass@1
26
  type: pass@1
27
- value: 0.0
28
  verified: false
29
  ---
30
 
@@ -40,45 +40,11 @@ Total 201,981 samples.
40
  - TokenBender/python_eval_instruct_51k: “python” in output .40,309 samples
41
  - Spider: 8,659 samples
42
 
43
-
44
- | | |
45
- |------ | ------ |
46
- | lr | 2e-4 |
47
- | lr_scheduler_type | cosine |
48
- | weight_decay | 0.0 |
49
- | optim | paged_adamw_8bit |
50
- | flash_attention | True |
51
- | rerope | False |
52
- | max_new_tokens | 4096 |
53
- | num_train_epochs | 2 |
54
- | bits | 4 |
55
- | lora_r | 64 |
56
- | lora_alpha | 16 |
57
- | lora_dropout | 0.05 |
58
- | double_quant | True |
59
- | quant_type | nf4 |
60
- | dataset_format | airoboros |
61
- | mini_batch_size | 2 |
62
- | grandient_accumulation_steps | 32 |
63
- | bf16 | True |
64
-
65
- A100-40G x 4
66
-
67
- | | |
68
- |------ | ------ |
69
- | epoch | 2.0 |
70
- | etrain_loss | 0.4708 |
71
- | etrain_runtime | 12:12:53.64 |
72
- | etrain_samples_per_second | 9.002 |
73
- | etrain_steps_per_second | 0.07 |
74
- | eeval_loss | 0.4851 |
75
- | eeval_runtime | 0:00:10.31 |
76
- | eeval_samples_per_second | 19.385 |
77
- | eeval_steps_per_second | 4.846 |
78
 
79
  | Metric | Value |
80
  | --- | --- |
81
- | humaneval-python | |
82
 
83
  [Big Code Models Leaderboard](https://huggingface.co/spaces/bigcode/bigcode-models-leaderboard)
84
 
@@ -94,6 +60,9 @@ CodeLlama-13B-Python: 42.89
94
 
95
  CodeLlama-13B: 35.07
96
 
 
 
 
97
  [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
98
  | Metric | Value |
99
  | --- | --- |
@@ -104,3 +73,40 @@ CodeLlama-13B: 35.07
104
  | Average | |
105
 
106
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  metrics:
25
  - name: pass@1
26
  type: pass@1
27
+ value: 51.829
28
  verified: false
29
  ---
30
 
 
40
  - TokenBender/python_eval_instruct_51k: “python” in output .40,309 samples
41
  - Spider: 8,659 samples
42
 
43
+ ## HumanEval
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
44
 
45
  | Metric | Value |
46
  | --- | --- |
47
+ | humaneval-python | 51.829 |
48
 
49
  [Big Code Models Leaderboard](https://huggingface.co/spaces/bigcode/bigcode-models-leaderboard)
50
 
 
60
 
61
  CodeLlama-13B: 35.07
62
 
63
+
64
+ ## LM-Evaluation-Harness
65
+
66
  [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
67
  | Metric | Value |
68
  | --- | --- |
 
73
  | Average | |
74
 
75
 
76
+ ## Parameters
77
+
78
+ | | |
79
+ |------ | ------ |
80
+ | lr | 2e-4 |
81
+ | lr_scheduler_type | cosine |
82
+ | weight_decay | 0.0 |
83
+ | optim | paged_adamw_8bit |
84
+ | flash_attention | True |
85
+ | rerope | False |
86
+ | max_new_tokens | 4096 |
87
+ | num_train_epochs | 2 |
88
+ | bits | 4 |
89
+ | lora_r | 64 |
90
+ | lora_alpha | 16 |
91
+ | lora_dropout | 0.05 |
92
+ | double_quant | True |
93
+ | quant_type | nf4 |
94
+ | dataset_format | airoboros |
95
+ | mini_batch_size | 2 |
96
+ | grandient_accumulation_steps | 32 |
97
+ | bf16 | True |
98
+
99
+ A800-80G x 2
100
+
101
+ | | |
102
+ |------ | ------ |
103
+ | epoch | 2.0 |
104
+ | etrain_loss | 0.5891 |
105
+ | etrain_runtime | 19:24:49.43 |
106
+ | etrain_samples_per_second | 5.664 |
107
+ | etrain_steps_per_second | 0.044 |
108
+ | eeval_loss | 0.5872 |
109
+ | eeval_runtime | 0:00:15.59 |
110
+ | eeval_samples_per_second | 12.822 |
111
+ | eeval_steps_per_second | 6.411 |
112
+