abhi5hekjangid commited on
Commit
ea80f1d
1 Parent(s): 4968b9c

End of training

Browse files
README.md CHANGED
@@ -16,7 +16,9 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # phi-2-finetuned-abhishek
18
 
19
- This model is a fine-tuned version of [microsoft/phi-2](https://huggingface.co/microsoft/phi-2) on an unknown dataset.
 
 
20
 
21
  ## Model description
22
 
@@ -36,15 +38,25 @@ More information needed
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 0.0002
39
- - train_batch_size: 4
40
  - eval_batch_size: 8
41
  - seed: 42
 
 
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: cosine
44
- - training_steps: 2000
 
 
45
 
46
  ### Training results
47
 
 
 
 
 
 
 
48
 
49
 
50
  ### Framework versions
 
16
 
17
  # phi-2-finetuned-abhishek
18
 
19
+ This model is a fine-tuned version of [microsoft/phi-2](https://huggingface.co/microsoft/phi-2) on the None dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 0.9799
22
 
23
  ## Model description
24
 
 
38
 
39
  The following hyperparameters were used during training:
40
  - learning_rate: 0.0002
41
+ - train_batch_size: 1
42
  - eval_batch_size: 8
43
  - seed: 42
44
+ - gradient_accumulation_steps: 8
45
+ - total_train_batch_size: 8
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: cosine
48
+ - lr_scheduler_warmup_ratio: 0.3
49
+ - num_epochs: 4
50
+ - mixed_precision_training: Native AMP
51
 
52
  ### Training results
53
 
54
+ | Training Loss | Epoch | Step | Validation Loss |
55
+ |:-------------:|:-----:|:----:|:---------------:|
56
+ | 1.2648 | 1.0 | 779 | 1.1722 |
57
+ | 1.0878 | 2.0 | 1558 | 1.0711 |
58
+ | 0.9319 | 3.0 | 2338 | 0.9918 |
59
+ | 0.8719 | 4.0 | 3116 | 0.9799 |
60
 
61
 
62
  ### Framework versions
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "microsoft/phi-2",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -9,19 +9,19 @@
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "loftq_config": {},
12
- "lora_alpha": 64,
13
  "lora_dropout": 0.05,
14
  "megatron_config": null,
15
  "megatron_core": "megatron.core",
16
  "modules_to_save": null,
17
  "peft_type": "LORA",
18
- "r": 32,
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
22
  "Wqkv",
23
- "fc2",
24
- "fc1"
25
  ],
26
  "task_type": "CAUSAL_LM"
27
  }
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": null,
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "loftq_config": {},
12
+ "lora_alpha": 16,
13
  "lora_dropout": 0.05,
14
  "megatron_config": null,
15
  "megatron_core": "megatron.core",
16
  "modules_to_save": null,
17
  "peft_type": "LORA",
18
+ "r": 16,
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "fc1",
23
  "Wqkv",
24
+ "fc2"
 
25
  ],
26
  "task_type": "CAUSAL_LM"
27
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e800434cf63cd64d88ec4fb04496a72c4cfb8f8cfab932725eb9e3c44e49c46f
3
- size 104873776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f49a1ef4918b36c34af00759620a6b70fa4cdd41e038f01dc44f9890f16b976
3
+ size 52447112
runs/Jan24_08-21-41_89b7d5a0e94a/events.out.tfevents.1706084541.89b7d5a0e94a.858.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c642c7b3606404e7f5c6cbdded49d5f695bc16c3152b1ffd33772fdcc072828
3
+ size 5212
runs/Jan24_08-24-23_89b7d5a0e94a/events.out.tfevents.1706084671.89b7d5a0e94a.858.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5817fbf655b3843ba869774224dac38f103c8a685c8861fe756273be3975157
3
+ size 5674
runs/Jan24_08-29-57_89b7d5a0e94a/events.out.tfevents.1706085002.89b7d5a0e94a.858.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0945de58022b91c7c884cb8841da1770ed467e5cbe359d9b9184e00e08a1c0ce
3
+ size 5214
runs/Jan24_08-31-40_89b7d5a0e94a/events.out.tfevents.1706085102.89b7d5a0e94a.858.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab0f3f7e5e576d6ce8f6856e9b70bfce85f6296cfa83a8ee6cc379b442b2fca6
3
+ size 5055
runs/Jan24_08-32-40_89b7d5a0e94a/events.out.tfevents.1706085163.89b7d5a0e94a.858.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74a42b9ce220279014b06b3b04861281ece3153aad949685324e9099a4484353
3
+ size 25949
special_tokens_map.json CHANGED
@@ -13,7 +13,7 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": "<|endoftext|>",
17
  "unk_token": {
18
  "content": "<|endoftext|>",
19
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": " ",
17
  "unk_token": {
18
  "content": "<|endoftext|>",
19
  "lstrip": false,
tokenizer.json CHANGED
@@ -2,7 +2,7 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 1024,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 512,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
tokenizer_config.json CHANGED
@@ -318,7 +318,7 @@
318
  "clean_up_tokenization_spaces": true,
319
  "eos_token": "<|endoftext|>",
320
  "model_max_length": 2048,
321
- "pad_token": "<|endoftext|>",
322
  "tokenizer_class": "CodeGenTokenizer",
323
  "unk_token": "<|endoftext|>"
324
  }
 
318
  "clean_up_tokenization_spaces": true,
319
  "eos_token": "<|endoftext|>",
320
  "model_max_length": 2048,
321
+ "pad_token": " ",
322
  "tokenizer_class": "CodeGenTokenizer",
323
  "unk_token": "<|endoftext|>"
324
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc41746de4a3009ec01409f58e569d90e1adf855a253db41510236448bfb58c7
3
  size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46bd555d73b517291e231f99c32b47739d6609e277f33cfd905ae2a171d4ab4d
3
  size 4728