mnoukhov commited on
Commit
df7a908
1 Parent(s): 1adee4f

mnoukhov/pythia410m-test-tldr

Browse files
README.md CHANGED
@@ -16,13 +16,13 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [mnoukhov/pythia410m-sft-tldr](https://huggingface.co/mnoukhov/pythia410m-sft-tldr) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 0.6867
20
- - Rewards/chosen: 0.0544
21
- - Rewards/rejected: 0.0391
22
- - Rewards/accuracies: 0.4922
23
- - Rewards/margins: 0.0154
24
- - Logps/rejected: -103.2449
25
- - Logps/chosen: -103.2449
26
  - Logps/ref Rejected: -82.9119
27
  - Logps/ref Chosen: -104.3332
28
 
@@ -56,14 +56,14 @@ The following hyperparameters were used during training:
56
 
57
  | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logps/ref Rejected | Logps/ref Chosen |
58
  |:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:------------------:|:----------------:|
59
- | No log | 0.5 | 4 | 0.6956 | 0.0594 | 0.0634 | 0.3828 | -0.0040 | -103.1458 | -103.1458 | -82.9119 | -104.3332 |
60
- | No log | 1.0 | 8 | 0.6867 | 0.0544 | 0.0391 | 0.4922 | 0.0154 | -103.2449 | -103.2449 | -82.9119 | -104.3332 |
61
 
62
 
63
  ### Framework versions
64
 
65
  - PEFT 0.10.0
66
- - Transformers 4.38.2
67
- - Pytorch 2.1.2+cu121
68
- - Datasets 2.17.0
69
- - Tokenizers 0.15.2
 
16
 
17
  This model is a fine-tuned version of [mnoukhov/pythia410m-sft-tldr](https://huggingface.co/mnoukhov/pythia410m-sft-tldr) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 0.6766
20
+ - Rewards/chosen: 0.0321
21
+ - Rewards/rejected: -0.0032
22
+ - Rewards/accuracies: 0.7656
23
+ - Rewards/margins: 0.0353
24
+ - Logps/rejected: -103.6910
25
+ - Logps/chosen: -103.6910
26
  - Logps/ref Rejected: -82.9119
27
  - Logps/ref Chosen: -104.3332
28
 
 
56
 
57
  | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logps/ref Rejected | Logps/ref Chosen |
58
  |:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:------------------:|:----------------:|
59
+ | No log | 0.5 | 4 | 0.6336 | 0.1889 | 0.0621 | 0.7656 | 0.1268 | -100.5552 | -100.5552 | -82.9119 | -104.3332 |
60
+ | No log | 1.0 | 8 | 0.6766 | 0.0321 | -0.0032 | 0.7656 | 0.0353 | -103.6910 | -103.6910 | -82.9119 | -104.3332 |
61
 
62
 
63
  ### Framework versions
64
 
65
  - PEFT 0.10.0
66
+ - Transformers 4.40.2
67
+ - Pytorch 2.2.1+cu121
68
+ - Datasets 2.19.1
69
+ - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -20,10 +20,10 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "query_key_value",
24
  "dense",
25
  "dense_h_to_4h",
26
- "dense_4h_to_h"
 
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "dense",
24
  "dense_h_to_4h",
25
+ "dense_4h_to_h",
26
+ "query_key_value"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:675ab8d15657e56da216f0b2e7cce6e21affa5ec77d9229f6f2a0dbb8f7f985c
3
  size 25192592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbe98ae4175ae9261c28a9e4b4495a94aa90c0878dcf4b6b27826d857287445b
3
  size 25192592
tokenizer.json CHANGED
@@ -239,10 +239,30 @@
239
  "use_regex": true
240
  },
241
  "post_processor": {
242
- "type": "ByteLevel",
243
- "add_prefix_space": false,
244
- "trim_offsets": true,
245
- "use_regex": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
246
  },
247
  "decoder": {
248
  "type": "ByteLevel",
@@ -258,6 +278,7 @@
258
  "end_of_word_suffix": null,
259
  "fuse_unk": false,
260
  "byte_fallback": false,
 
261
  "vocab": {
262
  "<|endoftext|>": 0,
263
  "<|padding|>": 1,
 
239
  "use_regex": true
240
  },
241
  "post_processor": {
242
+ "type": "TemplateProcessing",
243
+ "single": [
244
+ {
245
+ "Sequence": {
246
+ "id": "A",
247
+ "type_id": 0
248
+ }
249
+ }
250
+ ],
251
+ "pair": [
252
+ {
253
+ "Sequence": {
254
+ "id": "A",
255
+ "type_id": 0
256
+ }
257
+ },
258
+ {
259
+ "Sequence": {
260
+ "id": "B",
261
+ "type_id": 1
262
+ }
263
+ }
264
+ ],
265
+ "special_tokens": {}
266
  },
267
  "decoder": {
268
  "type": "ByteLevel",
 
278
  "end_of_word_suffix": null,
279
  "fuse_unk": false,
280
  "byte_fallback": false,
281
+ "ignore_merges": false,
282
  "vocab": {
283
  "<|endoftext|>": 0,
284
  "<|padding|>": 1,
tokenizer_config.json CHANGED
@@ -1,4 +1,6 @@
1
  {
 
 
2
  "add_prefix_space": false,
3
  "added_tokens_decoder": {
4
  "0": {
 
1
  {
2
+ "add_bos_token": false,
3
+ "add_eos_token": false,
4
  "add_prefix_space": false,
5
  "added_tokens_decoder": {
6
  "0": {
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7bf786461ba419274a353d2957dc06029f0eca66d77c6e6b9067a0f19f08adf4
3
- size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1053542ee39d77231244ed0f658cfb77d2771b4cfbf5216e8004482fed975a2b
3
+ size 5240