liuylhf commited on
Commit
a58bce4
1 Parent(s): 2ae9c7d

Training in progress, step 210

Browse files
adapter_config.json CHANGED
@@ -19,10 +19,10 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "k_proj",
23
  "v_proj",
24
  "o_proj",
25
- "q_proj"
 
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_dora": false,
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
22
  "v_proj",
23
  "o_proj",
24
+ "q_proj",
25
+ "k_proj"
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b682eb31639f20779d103bc101fb9b5073fbd918e51f476cde04545a9856fbc2
3
  size 109086416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2838d91c914744c5d0b671db70347f932be367d0fbef20ac965efb6538d094ea
3
  size 109086416
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a2f2a5a3d76e7e978111652280ccca2d6c9a772c7c2c27236e59b51a91f0198d
3
  size 5688
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7e5a2052c04e540669e26e2cdecab171529f4d2389711ea7707e33ed2a48535
3
  size 5688