oz1115 commited on
Commit
d9c5183
1 Parent(s): 182d99e

oz1115/phi3_adapter

Browse files
README.md CHANGED
@@ -14,7 +14,7 @@ model-index:
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
16
 
17
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/gsitm/Phi3-mini-ft-python-code/runs/q22y9rcg)
18
  # phi-3-mini-LoRA
19
 
20
  This model is a fine-tuned version of [microsoft/Phi-3-mini-4k-instruct](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct) on an unknown dataset.
@@ -55,6 +55,6 @@ The following hyperparameters were used during training:
55
 
56
  - PEFT 0.11.1
57
  - Transformers 4.42.4
58
- - Pytorch 2.3.0+cu121
59
  - Datasets 2.20.0
60
  - Tokenizers 0.19.1
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
16
 
17
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/gsitm/Phi3-mini-ft-python-code/runs/4ipkuc1m)
18
  # phi-3-mini-LoRA
19
 
20
  This model is a fine-tuned version of [microsoft/Phi-3-mini-4k-instruct](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct) on an unknown dataset.
 
55
 
56
  - PEFT 0.11.1
57
  - Transformers 4.42.4
58
+ - Pytorch 2.3.1+cu121
59
  - Datasets 2.20.0
60
  - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "down_proj",
 
24
  "o_proj",
25
- "v_proj",
26
- "q_proj",
27
  "up_proj",
28
- "gate_proj",
29
- "k_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "k_proj",
24
+ "gate_proj",
25
  "o_proj",
26
+ "down_proj",
 
27
  "up_proj",
28
+ "v_proj",
29
+ "q_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:26edf52e3634dbc646044d179be2c06589d2bb5743cc3aaac0423d01bf23f75d
3
  size 17842848
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c003ca40159a6a6c58001ddd7b790a980e5db074f965e8e5dc6f7c087d80293
3
  size 17842848
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b5b1518ec136466ef05c807a7657cf246b028eeac97f922e834c454e81e46e08
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed50aadc2275e6f3f53808d4a15c23a22156a5e3e833a006951eaf63849511b7
3
  size 5368