oz1115 commited on
Commit
9d03a37
1 Parent(s): 3cb5b6e

oz1115/gemma_2b_it_ko_adapter

Browse files
README.md CHANGED
@@ -43,7 +43,7 @@ The following hyperparameters were used during training:
43
  - total_train_batch_size: 4
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
- - lr_scheduler_warmup_steps: 0.03
47
  - training_steps: 100
48
  - mixed_precision_training: Native AMP
49
 
@@ -53,8 +53,8 @@ The following hyperparameters were used during training:
53
 
54
  ### Framework versions
55
 
56
- - PEFT 0.9.0
57
- - Transformers 4.38.2
58
  - Pytorch 2.3.1+cu121
59
- - Datasets 2.18.0
60
- - Tokenizers 0.15.2
 
43
  - total_train_batch_size: 4
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
+ - lr_scheduler_warmup_steps: 3
47
  - training_steps: 100
48
  - mixed_precision_training: Native AMP
49
 
 
53
 
54
  ### Framework versions
55
 
56
+ - PEFT 0.12.0
57
+ - Transformers 4.43.4
58
  - Pytorch 2.3.1+cu121
59
+ - Datasets 2.20.0
60
+ - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -6,6 +6,7 @@
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
 
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "loftq_config": {},
@@ -19,13 +20,13 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "o_proj",
23
  "up_proj",
24
- "down_proj",
25
- "q_proj",
26
  "k_proj",
 
 
27
  "gate_proj",
28
- "v_proj"
 
29
  ],
30
  "task_type": "CAUSAL_LM",
31
  "use_dora": false,
 
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
9
+ "layer_replication": null,
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "up_proj",
 
 
24
  "k_proj",
25
+ "v_proj",
26
+ "q_proj",
27
  "gate_proj",
28
+ "down_proj",
29
+ "o_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e44ce263e6fd885f50d82ca515b9325375b43ee36ededb75acf161ce88bc2e41
3
- size 48
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90882106c502840a0e2fbe1f1fe83cce47769b9856403520435fa95ef5fecbd6
3
+ size 29450584
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6060511074d2fd48bd3d5248318fdcfeb03d03ec0d9d9c4ecfa3befc7016173c
3
- size 17518595
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36827d0e8282c6dfe284c3a59cae2b855cf498397ce22e8fd2559e5a14acb882
3
+ size 17518623
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:674d6f5b7e182e62f6c92f67f5060a24c0de8de28bba1face8b9537d423446cd
3
- size 4856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9a44ffd8187d932a01cde7e407be1557d5f34af8cab478d9bdd8a3a458d30ef
3
+ size 5368