azmoulai commited on
Commit
7dbc563
1 Parent(s): f0b82ec

Training in progress, step 93

Browse files
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "up_proj",
24
- "v_proj",
25
- "down_proj",
26
  "gate_proj",
 
 
27
  "o_proj",
28
  "k_proj",
29
- "q_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
 
23
  "gate_proj",
24
+ "q_proj",
25
+ "v_proj",
26
  "o_proj",
27
  "k_proj",
28
+ "up_proj",
29
+ "down_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:94e031f7b9a648e92b757ebc11f22a35610f7ea70cbcacd1dfcccd86846b8461
3
  size 45258384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4066c7dc2f49ecfd08396d382fa73d1e6c8aa0801ee5376f3060a9370ef8fce3
3
  size 45258384
runs/Jun15_07-03-54_c8ac5ac3e7ee/events.out.tfevents.1718435694.c8ac5ac3e7ee.796.12 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cacb67ee1435c39d9352cbc818cc248511b5c8cbd8406be958e7b08c2383b888
3
+ size 354
runs/Jun15_07-18-54_c8ac5ac3e7ee/events.out.tfevents.1718435935.c8ac5ac3e7ee.1293.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:587a0541c254fe7fc6bf227a6551119323844cd866f5e73d6ec7c88fd1eb7f75
3
+ size 6065
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3ca364a65f9dae26d421266308c5a960676463c7125aa62623b6569b81b329d8
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bece795c6cee461325b42344aa5c0016bdf6161ec6bd2717a846eb0fe0c7923
3
  size 5176