k-r-l commited on
Commit
9be6010
1 Parent(s): 37c1d81

Training in progress, step 1

Browse files
adapter_config.json CHANGED
@@ -10,18 +10,18 @@
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
- "lora_alpha": 16,
14
  "lora_dropout": 0.25,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
- "r": 16,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "v_proj",
24
- "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
 
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
+ "lora_alpha": 8,
14
  "lora_dropout": 0.25,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
+ "r": 8,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "q_proj",
24
+ "v_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3ef30aaf06a4abcb662004f6125a1acc35f54d6c9d8d88285afbc8cc3a69240f
3
- size 27280152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2b0b804b4a0a92d0649de91e421dcf106fb6dae0b03e89266dd5cc8da6c9027
3
+ size 13648432
runs/Aug14_16-51-28_d35c9d528975/events.out.tfevents.1723654290.d35c9d528975.12415.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c4bcadf084cacc04e534f091c9a4e13af990f6b2cb4e3d6fc8949c913e8507a
3
+ size 6849
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c023621548c06beba85bc74f614a5433390ae629b639afa8513fdcc38901ab03
3
  size 5560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:575103b7ce33ea7859ea06e4782e5c4f7adb889bebb434ec2f6f6f8424947019
3
  size 5560