SeanLee97 commited on
Commit
81db7c5
1 Parent(s): a44c6fa

Training in progress, step 500

Browse files
adapter_config.json CHANGED
@@ -9,26 +9,23 @@
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "loftq_config": {},
12
- "lora_alpha": 32,
13
  "lora_dropout": 0.1,
14
  "megatron_config": null,
15
  "megatron_core": "megatron.core",
16
- "modules_to_save": [
17
- "model.leaky_norm.weight",
18
- "model.leaky_norm.bias"
19
- ],
20
  "peft_type": "LORA",
21
- "r": 32,
22
  "rank_pattern": {},
23
  "revision": null,
24
  "target_modules": [
 
25
  "o_proj",
26
- "q_proj",
27
- "up_proj",
28
- "v_proj",
29
  "down_proj",
30
- "k_proj",
31
- "gate_proj"
 
 
32
  ],
33
  "task_type": "CAUSAL_LM",
34
  "use_rslora": false
 
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "loftq_config": {},
12
+ "lora_alpha": 16,
13
  "lora_dropout": 0.1,
14
  "megatron_config": null,
15
  "megatron_core": "megatron.core",
16
+ "modules_to_save": null,
 
 
 
17
  "peft_type": "LORA",
18
+ "r": 64,
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "k_proj",
23
  "o_proj",
 
 
 
24
  "down_proj",
25
+ "v_proj",
26
+ "gate_proj",
27
+ "q_proj",
28
+ "up_proj"
29
  ],
30
  "task_type": "CAUSAL_LM",
31
  "use_rslora": false
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e435f0aa476689f73a4533c0ed439a269faf4383a979cfce4c6c3bcbf39e055d
3
- size 319876032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aee0d090b0541bb5167188b9198a2ab5f7292290651b09ba6ea477a2d4c1d6a5
3
+ size 639691872
angle.config CHANGED
@@ -5,10 +5,19 @@
5
  "pooling_strategy": "cls",
6
  "lora_config_kwargs": {
7
  "task_type": "CAUSAL_LM",
8
- "r": 32,
9
- "lora_alpha": 32,
10
  "lora_dropout": 0.1,
11
- "bias": "none"
 
 
 
 
 
 
 
 
 
12
  },
13
  "apply_lora": 1
14
  }
 
5
  "pooling_strategy": "cls",
6
  "lora_config_kwargs": {
7
  "task_type": "CAUSAL_LM",
8
+ "r": 64,
9
+ "lora_alpha": 16,
10
  "lora_dropout": 0.1,
11
+ "bias": "none",
12
+ "target_modules": [
13
+ "v_proj",
14
+ "gate_proj",
15
+ "k_proj",
16
+ "down_proj",
17
+ "q_proj",
18
+ "up_proj",
19
+ "o_proj"
20
+ ]
21
  },
22
  "apply_lora": 1
23
  }
runs/Mar20_06-17-12_instance-20240226-074643/events.out.tfevents.1710915444.instance-20240226-074643.1721310.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c90b97ddcede90885b0cc6bc66fb578c855a60808b3a7b4cc8a07df808d628f6
3
- size 14486
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09aea37f7e2efe0e4e90d82ccafddd3f98f1bef1f72ae98bd53c3e6329f9d4a5
3
+ size 14840
runs/Mar20_13-21-43_instance-20240226-074643/events.out.tfevents.1710940908.instance-20240226-074643.1922101.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:092413b45d9f29c97a7d165bf3a6ec151e9f7430bac4f0043e8389335c9626eb
3
+ size 6257
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8a17c613d3e0e22264379b569a7ef021e7162704a72661d8fc663d1e63e01d8e
3
  size 5048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0462a9ded0b9349dfe5bea51e9ef06739149993b3576240578be5b4934a7ceae
3
  size 5048