nextM commited on
Commit
34022c8
1 Parent(s): c40433f

Training in progress, epoch 0

Browse files
adapter_config.json CHANGED
@@ -20,8 +20,8 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "q_proj",
24
- "v_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "v_proj",
24
+ "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:93a7c2b3e180d4734804b48c6ad7facb478e1ad0742ccb4ec6ab4900302f91c1
3
  size 10107280
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:783d0747bc3df8b1d3674474b15cf6f9cb1a4d2fd6409ffaa3a47a5bd22c1ce3
3
  size 10107280
runs/Oct03_16-00-21_modal/events.out.tfevents.1727971222.modal.2.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:544742bbee595542dd866360d867926ee85f605826d429b953cb8d29a79994a1
3
+ size 5839
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d384b8a42115e902aaa7a4a36b4c8cbe89cc79328fdddaf841320cdcd3ac611d
3
  size 5560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:482e2809002604da6fe77b7aa61e9905f3b1b43c2452782472464112888d5481
3
  size 5560