Venkatesh4342 commited on
Commit
39e7c1b
1 Parent(s): d536a1b

Training complete!

Browse files
Files changed (3) hide show
  1. README.md +26 -19
  2. generation_config.json +1 -2
  3. pytorch_model.bin +1 -1
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: google/pegasus-cnn_dailymail
3
  tags:
4
  - generated_from_trainer
5
  datasets:
@@ -21,7 +21,7 @@ model-index:
21
  metrics:
22
  - name: Rouge1
23
  type: rouge
24
- value: 0.4616
25
  ---
26
 
27
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -29,14 +29,14 @@ should probably proofread and complete it, then remove this comment. -->
29
 
30
  # pegasus-samsum
31
 
32
- This model is a fine-tuned version of [google/pegasus-cnn_dailymail](https://huggingface.co/google/pegasus-cnn_dailymail) on the samsum dataset.
33
  It achieves the following results on the evaluation set:
34
- - Loss: 1.3665
35
- - Rouge1: 0.4616
36
- - Rouge2: 0.2275
37
- - Rougel: 0.3725
38
- - Rougelsum: 0.3738
39
- - Gen Len: 35.6667
40
 
41
  ## Model description
42
 
@@ -55,25 +55,32 @@ More information needed
55
  ### Training hyperparameters
56
 
57
  The following hyperparameters were used during training:
58
- - learning_rate: 5.750420024069848e-05
59
- - train_batch_size: 8
60
- - eval_batch_size: 8
61
  - seed: 42
62
- - gradient_accumulation_steps: 16
63
- - total_train_batch_size: 128
64
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
65
  - lr_scheduler_type: linear
66
  - lr_scheduler_warmup_steps: 500
67
- - num_epochs: 4
68
 
69
  ### Training results
70
 
71
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
72
  |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|
73
- | 2.2186 | 0.87 | 100 | 1.7567 | 0.3571 | 0.1437 | 0.2779 | 0.2797 | 46.7733 |
74
- | 1.7368 | 1.74 | 200 | 1.4933 | 0.4347 | 0.2053 | 0.3459 | 0.3461 | 35.4533 |
75
- | 1.6744 | 2.61 | 300 | 1.4059 | 0.4547 | 0.2179 | 0.3629 | 0.3634 | 35.68 |
76
- | 1.5978 | 3.47 | 400 | 1.3665 | 0.4616 | 0.2275 | 0.3725 | 0.3738 | 35.6667 |
 
 
 
 
 
 
 
77
 
78
 
79
  ### Framework versions
 
1
  ---
2
+ base_model: google/pegasus-large
3
  tags:
4
  - generated_from_trainer
5
  datasets:
 
21
  metrics:
22
  - name: Rouge1
23
  type: rouge
24
+ value: 0.4659
25
  ---
26
 
27
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
29
 
30
  # pegasus-samsum
31
 
32
+ This model is a fine-tuned version of [google/pegasus-large](https://huggingface.co/google/pegasus-large) on the samsum dataset.
33
  It achieves the following results on the evaluation set:
34
+ - Loss: 1.4091
35
+ - Rouge1: 0.4659
36
+ - Rouge2: 0.2345
37
+ - Rougel: 0.3946
38
+ - Rougelsum: 0.3951
39
+ - Gen Len: 17.7467
40
 
41
  ## Model description
42
 
 
55
  ### Training hyperparameters
56
 
57
  The following hyperparameters were used during training:
58
+ - learning_rate: 2e-05
59
+ - train_batch_size: 4
60
+ - eval_batch_size: 4
61
  - seed: 42
62
+ - gradient_accumulation_steps: 2
63
+ - total_train_batch_size: 8
64
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
65
  - lr_scheduler_type: linear
66
  - lr_scheduler_warmup_steps: 500
67
+ - num_epochs: 3
68
 
69
  ### Training results
70
 
71
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
72
  |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|
73
+ | 1.8025 | 0.27 | 500 | 1.4403 | 0.4466 | 0.2101 | 0.3832 | 0.3841 | 21.64 |
74
+ | 1.5936 | 0.54 | 1000 | 1.3766 | 0.4786 | 0.2374 | 0.4017 | 0.4013 | 21.24 |
75
+ | 1.5926 | 0.81 | 1500 | 1.3910 | 0.5118 | 0.2643 | 0.4282 | 0.4286 | 20.2267 |
76
+ | 1.5067 | 1.09 | 2000 | 1.4028 | 0.4982 | 0.261 | 0.4155 | 0.4157 | 20.4267 |
77
+ | 1.5712 | 1.36 | 2500 | 1.4236 | 0.4712 | 0.234 | 0.3964 | 0.3969 | 17.0 |
78
+ | 1.6177 | 1.63 | 3000 | 1.4151 | 0.4768 | 0.2382 | 0.4019 | 0.4022 | 16.28 |
79
+ | 1.6289 | 1.9 | 3500 | 1.4112 | 0.4744 | 0.2346 | 0.402 | 0.4033 | 17.0267 |
80
+ | 1.6326 | 2.17 | 4000 | 1.4096 | 0.4682 | 0.234 | 0.3985 | 0.3994 | 17.1333 |
81
+ | 1.5929 | 2.44 | 4500 | 1.4093 | 0.4637 | 0.2342 | 0.3939 | 0.3942 | 17.16 |
82
+ | 1.4351 | 2.72 | 5000 | 1.4090 | 0.4684 | 0.2346 | 0.3953 | 0.3955 | 17.8133 |
83
+ | 1.6445 | 2.99 | 5500 | 1.4091 | 0.4659 | 0.2345 | 0.3946 | 0.3951 | 17.7467 |
84
 
85
 
86
  ### Framework versions
generation_config.json CHANGED
@@ -4,8 +4,7 @@
4
  "eos_token_id": 1,
5
  "forced_eos_token_id": 1,
6
  "length_penalty": 0.8,
7
- "max_length": 128,
8
- "min_length": 32,
9
  "num_beams": 8,
10
  "pad_token_id": 0,
11
  "transformers_version": "4.33.1"
 
4
  "eos_token_id": 1,
5
  "forced_eos_token_id": 1,
6
  "length_penalty": 0.8,
7
+ "max_length": 256,
 
8
  "num_beams": 8,
9
  "pad_token_id": 0,
10
  "transformers_version": "4.33.1"
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a231eec02b23d1fd6b3262471a927754ba96c7bdc25d6ce79c6f7fb81505cde5
3
  size 2283804653
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73d0af06060e82a368bcafdbfc8eb5ba5ad6071968bc699e79006b1c515db840
3
  size 2283804653