TeeA commited on
Commit
c1635a9
1 Parent(s): 0d9f4b2

Training in progress, step 14200

Browse files
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "down_proj",
24
- "v_proj",
25
- "q_proj",
26
- "up_proj",
27
  "k_proj",
28
  "o_proj",
29
- "gate_proj"
 
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "gate_proj",
 
 
 
24
  "k_proj",
25
  "o_proj",
26
+ "up_proj",
27
+ "down_proj",
28
+ "v_proj",
29
+ "q_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1d632cb2ed583ce38bece05e65a614435f7b6e96a5a1685d33c353b398bddaea
3
  size 1279323952
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee603a440c664217c1d050080d00ec863b00cb76fddc2fc47d50c17c6d680cfb
3
  size 1279323952
runs/Apr24_14-29-06_kao-dgxa-e10-u17/events.out.tfevents.1713965351.kao-dgxa-e10-u17.3864025.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:88efc3551d5d7e9a3654f9327f0f8379f3fd2ee5347bf7b0770360a463f391a0
3
- size 1067279
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ef84355a12f6b9728998bb1e16883be12fcee6cad7d1d20f9fa1387237168bf
3
+ size 1072343
runs/Apr25_13-59-40_kao-dgxa-e10-u17/events.out.tfevents.1714049984.kao-dgxa-e10-u17.1365672.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c51df1af0bbb2451880eca954c678d83f13520dfe26ccd078e5c8a71c988aae9
3
+ size 47975
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fb4b6829eb8c1ad52fa891cd2b1aa5be9322fb25279ffb21273b53e1a2e534dd
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afd013a64bdc62b2280c2cda4f8df0ee5251ef95ea278081d19bd1ca2f0e980a
3
  size 5176