seyviour commited on
Commit
f32f0c6
1 Parent(s): 82e9394

Training in progress, epoch 0

Browse files
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "v_proj",
24
- "k_proj",
25
  "q_proj",
26
- "gate_proj",
27
- "down_proj",
28
  "o_proj",
29
- "up_proj"
 
 
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
23
  "q_proj",
 
 
24
  "o_proj",
25
+ "k_proj",
26
+ "down_proj",
27
+ "v_proj",
28
+ "up_proj",
29
+ "gate_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d9feafc35e91dd5bb7bc4dd4f0313de99a8f697a270c73941785ab5620e00c0
3
  size 45258384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3c59b7bdd2de37cbd24ee76462d8796aa5d175cb2e80dcc56bbac8bcc36eeeb
3
  size 45258384
runs/Nov12_16-15-00_t2/events.out.tfevents.1731453311.t2 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a4beb49a22d4df053cd93ceac03d279537eb3119783482ec9d3cb1d4e800e072
3
- size 6038
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45a6ac0f9ad6465c5efb5499052fff086422bacef1da2884f8d6dff41b8181d0
3
+ size 6309
runs/Nov12_17-36-40_t2/events.out.tfevents.1731458202.t2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6fd80ccc19f9bcca497324423717c06f438ca4117ef93c0dbc84270376bb0b1
3
+ size 5774
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca3b85fd6690616634b79dbf94138a1971eecbdc2be9357eed2ce71f3053f3b4
3
  size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ae3e37337e9c57a0e1b3d72562d190b36aec9522f5cebefc0bdd229393f1c4b
3
  size 5240