maxadmin commited on
Commit
5362abf
1 Parent(s): f48320b

End of training

Browse files
README.md CHANGED
@@ -15,14 +15,14 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # maximo-t5-chat
17
 
18
- This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 1.9662
21
- - Rouge1: 40.75
22
- - Rouge2: 29.1975
23
- - Rougel: 37.9379
24
- - Rougelsum: 38.4814
25
- - Gen Len: 16.0
26
 
27
  ## Model description
28
 
@@ -47,26 +47,25 @@ The following hyperparameters were used during training:
47
  - seed: 42
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
- - num_epochs: 10
51
 
52
  ### Training results
53
 
54
- | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
55
- |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
56
- | No log | 1.0 | 2 | 2.3093 | 19.8026 | 2.37 | 17.9508 | 17.9508 | 8.25 |
57
- | No log | 2.0 | 4 | 2.1217 | 27.8825 | 7.1408 | 24.4377 | 25.5648 | 15.5 |
58
- | No log | 3.0 | 6 | 2.0575 | 43.9982 | 32.2234 | 40.0719 | 40.8556 | 15.75 |
59
- | No log | 4.0 | 8 | 2.0250 | 41.0247 | 30.6915 | 39.0215 | 40.4149 | 16.0 |
60
- | No log | 5.0 | 10 | 2.0107 | 40.751 | 30.1882 | 36.6149 | 38.1412 | 16.0 |
61
- | No log | 6.0 | 12 | 2.0020 | 38.2692 | 27.9512 | 34.5122 | 36.6315 | 16.0 |
62
- | No log | 7.0 | 14 | 1.9886 | 41.4727 | 29.7718 | 37.1689 | 37.2678 | 16.0 |
63
- | No log | 8.0 | 16 | 1.9752 | 41.4727 | 29.7718 | 37.1689 | 37.2678 | 16.0 |
64
- | No log | 9.0 | 18 | 1.9695 | 39.4309 | 29.1975 | 35.6826 | 37.226 | 16.0 |
65
- | No log | 10.0 | 20 | 1.9662 | 40.75 | 29.1975 | 37.9379 | 38.4814 | 16.0 |
66
 
67
 
68
  ### Framework versions
69
 
70
  - Transformers 4.35.2
71
  - Pytorch 2.1.0+cu118
 
72
  - Tokenizers 0.15.0
 
15
 
16
  # maximo-t5-chat
17
 
18
+ This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 2.2093
21
+ - Rouge1: 27.8954
22
+ - Rouge2: 7.9325
23
+ - Rougel: 27.8954
24
+ - Rougelsum: 27.0723
25
+ - Gen Len: 12.6667
26
 
27
  ## Model description
28
 
 
47
  - seed: 42
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
+ - num_epochs: 8
51
 
52
  ### Training results
53
 
54
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
55
+ |:-------------:|:-----:|:----:|:---------------:|:-------:|:------:|:-------:|:---------:|:-------:|
56
+ | No log | 1.0 | 3 | 2.8591 | 15.4762 | 2.439 | 15.0794 | 15.4762 | 16.1667 |
57
+ | No log | 2.0 | 6 | 2.5612 | 13.0291 | 2.439 | 12.6323 | 13.0291 | 16.1667 |
58
+ | No log | 3.0 | 9 | 2.4590 | 14.6825 | 2.439 | 14.2857 | 14.2857 | 10.6667 |
59
+ | No log | 4.0 | 12 | 2.3716 | 14.6825 | 2.439 | 14.2857 | 14.2857 | 12.0 |
60
+ | No log | 5.0 | 15 | 2.2986 | 19.4444 | 2.439 | 19.0476 | 19.0476 | 13.0 |
61
+ | No log | 6.0 | 18 | 2.2474 | 27.1017 | 7.9325 | 27.1017 | 26.2787 | 11.3333 |
62
+ | No log | 7.0 | 21 | 2.2220 | 27.8954 | 7.9325 | 27.8954 | 27.0723 | 12.6667 |
63
+ | No log | 8.0 | 24 | 2.2093 | 27.8954 | 7.9325 | 27.8954 | 27.0723 | 12.6667 |
 
 
64
 
65
 
66
  ### Framework versions
67
 
68
  - Transformers 4.35.2
69
  - Pytorch 2.1.0+cu118
70
+ - Datasets 2.15.0
71
  - Tokenizers 0.15.0
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:602284453f9e86b63e4f436b0becd589cd7398f061253c30e294c7aadb1c7d2a
3
  size 307867048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41200cec214955e004363b87de4d984852058b6bd738aeb7f24f0dd943b6bc67
3
  size 307867048
runs/Dec06_08-52-03_7bb9bb6007a1/events.out.tfevents.1701852743.7bb9bb6007a1.626.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2333b33ce80f7f3fc2f10caacbecdcca82634668a3639fa022ec01a642006545
3
+ size 9758
special_tokens_map.json CHANGED
@@ -101,25 +101,7 @@
101
  "<extra_id_98>",
102
  "<extra_id_99>"
103
  ],
104
- "eos_token": {
105
- "content": "</s>",
106
- "lstrip": false,
107
- "normalized": false,
108
- "rstrip": false,
109
- "single_word": false
110
- },
111
- "pad_token": {
112
- "content": "<pad>",
113
- "lstrip": false,
114
- "normalized": false,
115
- "rstrip": false,
116
- "single_word": false
117
- },
118
- "unk_token": {
119
- "content": "<unk>",
120
- "lstrip": false,
121
- "normalized": false,
122
- "rstrip": false,
123
- "single_word": false
124
- }
125
  }
 
101
  "<extra_id_98>",
102
  "<extra_id_99>"
103
  ],
104
+ "eos_token": "</s>",
105
+ "pad_token": "<pad>",
106
+ "unk_token": "<unk>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
107
  }
tokenizer.json CHANGED
@@ -1,7 +1,21 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
4
- "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 134,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
+ "padding": {
10
+ "strategy": {
11
+ "Fixed": 134
12
+ },
13
+ "direction": "Right",
14
+ "pad_to_multiple_of": null,
15
+ "pad_id": 0,
16
+ "pad_type_id": 0,
17
+ "pad_token": "<pad>"
18
+ },
19
  "added_tokens": [
20
  {
21
  "id": 0,
tokenizer_config.json CHANGED
@@ -930,16 +930,9 @@
930
  "clean_up_tokenization_spaces": true,
931
  "eos_token": "</s>",
932
  "extra_ids": 100,
933
- "max_length": 134,
934
  "model_max_length": 512,
935
- "pad_to_multiple_of": null,
936
  "pad_token": "<pad>",
937
- "pad_token_type_id": 0,
938
- "padding_side": "right",
939
  "sp_model_kwargs": {},
940
- "stride": 0,
941
  "tokenizer_class": "T5Tokenizer",
942
- "truncation_side": "right",
943
- "truncation_strategy": "longest_first",
944
  "unk_token": "<unk>"
945
  }
 
930
  "clean_up_tokenization_spaces": true,
931
  "eos_token": "</s>",
932
  "extra_ids": 100,
 
933
  "model_max_length": 512,
 
934
  "pad_token": "<pad>",
 
 
935
  "sp_model_kwargs": {},
 
936
  "tokenizer_class": "T5Tokenizer",
 
 
937
  "unk_token": "<unk>"
938
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c193ab31dda15c1846281b635b2135ef67cda13e731cc9a4cc7a77cd56f15ed9
3
- size 4792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:acbc58e0ec061e0c46176aaca4d853d3e1fe77ba3d9d25358bb1eca822836e4c
3
+ size 4728