maxadmin commited on
Commit
e6ad4d4
1 Parent(s): a2b04d6

End of training

Browse files
README.md CHANGED
@@ -17,12 +17,12 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 2.1443
21
- - Rouge1: 27.8954
22
- - Rouge2: 7.9325
23
- - Rougel: 27.8954
24
- - Rougelsum: 27.0723
25
- - Gen Len: 12.5
26
 
27
  ## Model description
28
 
@@ -53,16 +53,16 @@ The following hyperparameters were used during training:
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
55
  |:-------------:|:-----:|:----:|:---------------:|:-------:|:------:|:-------:|:---------:|:-------:|
56
- | No log | 1.0 | 3 | 2.8565 | 15.4762 | 2.439 | 15.0794 | 15.4762 | 16.1667 |
57
- | No log | 2.0 | 6 | 2.5544 | 13.0291 | 2.439 | 12.6323 | 13.0291 | 16.1667 |
58
- | No log | 3.0 | 9 | 2.4527 | 14.6825 | 2.439 | 14.2857 | 14.2857 | 10.8333 |
59
- | No log | 4.0 | 12 | 2.3570 | 20.2381 | 2.439 | 19.8413 | 19.8413 | 11.1667 |
60
- | No log | 5.0 | 15 | 2.2745 | 27.1017 | 7.9325 | 27.1017 | 26.2787 | 11.3333 |
61
- | No log | 6.0 | 18 | 2.2170 | 27.8954 | 7.9325 | 27.8954 | 27.0723 | 13.1667 |
62
- | No log | 7.0 | 21 | 2.1860 | 27.8954 | 7.9325 | 27.8954 | 27.0723 | 12.6667 |
63
- | No log | 8.0 | 24 | 2.1568 | 27.8954 | 7.9325 | 27.8954 | 27.0723 | 12.5 |
64
- | No log | 9.0 | 27 | 2.1445 | 27.8954 | 7.9325 | 27.8954 | 27.0723 | 12.5 |
65
- | No log | 10.0 | 30 | 2.1443 | 27.8954 | 7.9325 | 27.8954 | 27.0723 | 12.5 |
66
 
67
 
68
  ### Framework versions
 
17
 
18
  This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 1.1056
21
+ - Rouge1: 59.8095
22
+ - Rouge2: 47.0
23
+ - Rougel: 59.8095
24
+ - Rougelsum: 59.8095
25
+ - Gen Len: 14.6
26
 
27
  ## Model description
28
 
 
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
55
  |:-------------:|:-----:|:----:|:---------------:|:-------:|:------:|:-------:|:---------:|:-------:|
56
+ | No log | 1.0 | 5 | 2.8921 | 15.7143 | 0.0 | 15.7143 | 15.0 | 7.2 |
57
+ | No log | 2.0 | 10 | 2.1172 | 19.0 | 0.0 | 19.1905 | 18.5714 | 8.2 |
58
+ | No log | 3.0 | 15 | 1.7513 | 33.7143 | 20.0 | 34.0 | 33.3333 | 7.8 |
59
+ | No log | 4.0 | 20 | 1.4905 | 46.7143 | 34.0 | 47.0476 | 46.7143 | 12.3 |
60
+ | No log | 5.0 | 25 | 1.3527 | 55.5714 | 39.0 | 55.4762 | 54.5714 | 12.9 |
61
+ | No log | 6.0 | 30 | 1.3376 | 58.0952 | 39.0 | 58.0952 | 57.4286 | 13.6 |
62
+ | No log | 7.0 | 35 | 1.2002 | 58.5714 | 39.0 | 58.5714 | 57.8095 | 13.3 |
63
+ | No log | 8.0 | 40 | 1.1349 | 55.0476 | 39.0 | 54.5714 | 54.5714 | 14.3 |
64
+ | No log | 9.0 | 45 | 1.1106 | 59.8095 | 47.0 | 59.8095 | 59.8095 | 14.6 |
65
+ | No log | 10.0 | 50 | 1.1056 | 59.8095 | 47.0 | 59.8095 | 59.8095 | 14.6 |
66
 
67
 
68
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4367079d068ae4a90dad689e370cea4267c9bd0aaf158078596d1a469e3e9783
3
  size 307867048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a4e757066a16cbdea075ad0690a8bb8b2c067cfbda800feed03940145b544cf
3
  size 307867048
runs/Dec06_09-23-29_7bb9bb6007a1/events.out.tfevents.1701854615.7bb9bb6007a1.626.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee2e817a71d1ec137accda5d5af51868878999409243928f6378c4d750deeb47
3
+ size 10789
special_tokens_map.json CHANGED
@@ -101,25 +101,7 @@
101
  "<extra_id_98>",
102
  "<extra_id_99>"
103
  ],
104
- "eos_token": {
105
- "content": "</s>",
106
- "lstrip": false,
107
- "normalized": false,
108
- "rstrip": false,
109
- "single_word": false
110
- },
111
- "pad_token": {
112
- "content": "<pad>",
113
- "lstrip": false,
114
- "normalized": false,
115
- "rstrip": false,
116
- "single_word": false
117
- },
118
- "unk_token": {
119
- "content": "<unk>",
120
- "lstrip": false,
121
- "normalized": false,
122
- "rstrip": false,
123
- "single_word": false
124
- }
125
  }
 
101
  "<extra_id_98>",
102
  "<extra_id_99>"
103
  ],
104
+ "eos_token": "</s>",
105
+ "pad_token": "<pad>",
106
+ "unk_token": "<unk>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
107
  }
tokenizer.json CHANGED
@@ -1,7 +1,21 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
4
- "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 48,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
+ "padding": {
10
+ "strategy": {
11
+ "Fixed": 48
12
+ },
13
+ "direction": "Right",
14
+ "pad_to_multiple_of": null,
15
+ "pad_id": 0,
16
+ "pad_type_id": 0,
17
+ "pad_token": "<pad>"
18
+ },
19
  "added_tokens": [
20
  {
21
  "id": 0,
tokenizer_config.json CHANGED
@@ -930,16 +930,9 @@
930
  "clean_up_tokenization_spaces": true,
931
  "eos_token": "</s>",
932
  "extra_ids": 100,
933
- "max_length": 134,
934
  "model_max_length": 512,
935
- "pad_to_multiple_of": null,
936
  "pad_token": "<pad>",
937
- "pad_token_type_id": 0,
938
- "padding_side": "right",
939
  "sp_model_kwargs": {},
940
- "stride": 0,
941
  "tokenizer_class": "T5Tokenizer",
942
- "truncation_side": "right",
943
- "truncation_strategy": "longest_first",
944
  "unk_token": "<unk>"
945
  }
 
930
  "clean_up_tokenization_spaces": true,
931
  "eos_token": "</s>",
932
  "extra_ids": 100,
 
933
  "model_max_length": 512,
 
934
  "pad_token": "<pad>",
 
 
935
  "sp_model_kwargs": {},
 
936
  "tokenizer_class": "T5Tokenizer",
 
 
937
  "unk_token": "<unk>"
938
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a161e0b6388ab1227fa88a6277b82c5f6d8ec1d1f75a018734741917d9db803
3
  size 4792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e6f79856928c7cafbd0d91f85bac2898e498cda748266dbeb800c555bb3eea8
3
  size 4792