rafimumtaz commited on
Commit
dcd8c91
1 Parent(s): 640973c

Training in progress, step 186

Browse files
adapter_config.json CHANGED
@@ -20,8 +20,8 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "q_proj",
24
- "v_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "v_proj",
24
+ "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3328534f68e916d746e4fe7096185b34e6f519b8dc9fa8a3cbcebc71afd21628
3
  size 6824216
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d342299a5f2a8a0a77891fc908ec51008a636ff1ce732dbf0c3655cc53540800
3
  size 6824216
runs/Oct09_09-33-22_0d19356e166f/events.out.tfevents.1728466403.0d19356e166f.8727.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:feb141ec33c7b154859e88e93941148219fb620814d3ad2088f46b3817d8cd5a
3
+ size 5498
tokenizer.json CHANGED
@@ -1,21 +1,7 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 256,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
- "padding": {
10
- "strategy": {
11
- "Fixed": 256
12
- },
13
- "direction": "Right",
14
- "pad_to_multiple_of": null,
15
- "pad_id": 128001,
16
- "pad_type_id": 0,
17
- "pad_token": "<|end_of_text|>"
18
- },
19
  "added_tokens": [
20
  {
21
  "id": 128000,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 128000,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f1bea0f64a81355020a2f794acbef4077e35cc4a3653eb71964d25479a23f79f
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:771e921f7afc20a045d21c5d1318632287d44dc922d1394f742c1f603f644fda
3
  size 5176