mkopecki commited on
Commit
689afb6
1 Parent(s): 6bbb597

Training in progress, step 500

Browse files
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "v_proj",
24
- "up_proj",
25
  "q_proj",
 
26
  "o_proj",
27
- "k_proj",
28
  "down_proj",
29
- "gate_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "k_proj",
 
24
  "q_proj",
25
+ "up_proj",
26
  "o_proj",
 
27
  "down_proj",
28
+ "gate_proj",
29
+ "v_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7796a8b8b48c8a65dd69eaa720ca767ce8793f9c21b27bee282ae5a9c1fd9103
3
  size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:852d0f7f85615da468c797ef1ec1f1f8660fe222c480cbfee2ffb2f35e3cb625
3
  size 167832240
runs/Jul14_15-27-22_ml-cluster/events.out.tfevents.1720970846.ml-cluster ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c35b537b4fe060729f5b1a15ef0f29a9aafc548e3b6ba667cc56860656a7e10d
3
+ size 5702
runs/Jul14_15-30-21_ml-cluster/events.out.tfevents.1720971025.ml-cluster ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc559b59082c869748e1d9e482c0cedac80595444b17e0503ac4549446d2fb87
3
+ size 5702
special_tokens_map.json CHANGED
@@ -13,5 +13,11 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": "<|end_of_text|>"
 
 
 
 
 
 
17
  }
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
  }
tokenizer.json CHANGED
@@ -2311,6 +2311,15 @@
2311
  "rstrip": false,
2312
  "normalized": false,
2313
  "special": true
 
 
 
 
 
 
 
 
 
2314
  }
2315
  ],
2316
  "normalizer": null,
 
2311
  "rstrip": false,
2312
  "normalized": false,
2313
  "special": true
2314
+ },
2315
+ {
2316
+ "id": 128256,
2317
+ "content": "[PAD]",
2318
+ "single_word": false,
2319
+ "lstrip": false,
2320
+ "rstrip": false,
2321
+ "normalized": false,
2322
+ "special": true
2323
  }
2324
  ],
2325
  "normalizer": null,
tokenizer_config.json CHANGED
@@ -2047,6 +2047,14 @@
2047
  "rstrip": false,
2048
  "single_word": false,
2049
  "special": true
 
 
 
 
 
 
 
 
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
@@ -2057,6 +2065,6 @@
2057
  "attention_mask"
2058
  ],
2059
  "model_max_length": 1000000000000000019884624838656,
2060
- "pad_token": "<|end_of_text|>",
2061
  "tokenizer_class": "PreTrainedTokenizerFast"
2062
  }
 
2047
  "rstrip": false,
2048
  "single_word": false,
2049
  "special": true
2050
+ },
2051
+ "128256": {
2052
+ "content": "[PAD]",
2053
+ "lstrip": false,
2054
+ "normalized": false,
2055
+ "rstrip": false,
2056
+ "single_word": false,
2057
+ "special": true
2058
  }
2059
  },
2060
  "bos_token": "<|begin_of_text|>",
 
2065
  "attention_mask"
2066
  ],
2067
  "model_max_length": 1000000000000000019884624838656,
2068
+ "pad_token": "[PAD]",
2069
  "tokenizer_class": "PreTrainedTokenizerFast"
2070
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1ecda7d7300bb607ef09690b066dcf6aa108dfb72fad4082997b8645b12578b6
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e2008df619968f860bea0ef68062b4745c3ab9910343223e27015d96b4ec03d
3
  size 5432