--- language: en tags: - summarization model-index: - name: google/pegasus-cnn_dailymail results: - task: type: summarization name: Summarization dataset: name: cnn_dailymail type: cnn_dailymail config: 3.0.0 split: test metrics: - type: rouge value: 43.1513 name: ROUGE-1 verified: true verifyToken: eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiOTQyN2NhOTJmOTFjNjMwMmQxNWNkZTFlZDQ2NDM5MmVmYzRjZDY3YTg4YmE0NWVhZjhhNDg2ZTI0MmY2Yzc3NiIsInZlcnNpb24iOjF9.reFJeFkcBYYyYpc5QGKEtoJv0p3Vdt71dHiA294Ln8fubDYaxj0lVH5E9TotL8c1bIruXQbQXgT-jtQWyzWsAA - type: rouge value: 20.7332 name: ROUGE-2 verified: true verifyToken: eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiZjRkM2FmZjI5NjdiNzdkOTIyMDFjNDI3MjUyNWYxMzk1MDE3YjgyMTI0NWJlYTFlZmI0OWUwMDQxYzY4YmRhOCIsInZlcnNpb24iOjF9.9nZ1_2qP7zAKqBRLo_wZyhwVetlyxTemXvzVLfJVhFjzDS9iV6pH_5k7bRgvFVjnjBE6PQKszopAhIiveqYCCw - type: rouge value: 30.4613 name: ROUGE-L verified: true verifyToken: eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiZjhiNTY1NGM2ZDQyNTRiNzUxOGFjNjIyODVkYWQzM2Y4NTRjYzJhOGJmNDZiODczNzVlN2QwNzJmOGUwYWU4NyIsInZlcnNpb24iOjF9.QKgtVADI1yZQ3gCym_zhxzK9KJCezl6tCDxI9h_ST4u8xQpskNSFG_kOBu4tv0F9Hg2Q_AnMhdrNxgp6v4cPCg - type: rouge value: 36.7925 name: ROUGE-LSUM verified: true verifyToken: eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiZmJjMTE5OTc2NjdiZmIzMDVmMjFlYTc4MGI5ODA0M2RmZTg4OGE5NDIzZjZjNWY5NGU1OGUzYTlkYjdhYTdlYSIsInZlcnNpb24iOjF9.43kd7Mn1EJADFdFlzJArstuoIY_06wEQwLYFR8LAgJwyDBpeUZ9Nf6PNdt6xgOlgZlIsWn86rPnVP7VMQnA_Ag - type: loss value: 2.030170202255249 name: loss verified: true verifyToken: eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiOWQyODUyYWM3YzJhZTlkNTZmNjI1Yjg2NDYwYjJhMTZhNDE0MjlhYWY3ODk3NjA5OGNmYjRlZGUyYTM0NjNkZCIsInZlcnNpb24iOjF9.LDgNKKiKR78mwEwWD6656Ry76Y6C8f5Xq9edQ1sBzFtLE5KpMkRYrHbb1kzQva49tIEOpezVaYBf5yvwPSt2AQ - type: gen_len value: 78.0633 name: gen_len verified: true verifyToken: eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiNTk3NGJhNTc3NDliODA2Y2U4NWU5ODBlMjFmNzY2MTQ1OWVlYjU2NDU0NzIxMzAzYjczYjZjNzc2NDBkZGRiMCIsInZlcnNpb24iOjF9.07iOF20jsfcMuVhHXX0K4mg57FxFKY1NXtfFELY2qTFzbUPQuNCBXUfEF2vSTt0zV3MBCQEnzBNMz4a2ZJFaAw --- ### Pegasus Models See Docs: [here](https://huggingface.co/transformers/master/model_doc/pegasus.html) Original TF 1 code [here](https://github.com/google-research/pegasus) Authors: Jingqing Zhang, Yao Zhao, Mohammad Saleh and Peter J. Liu on Dec 18, 2019 Maintained by: [@sshleifer](https://twitter.com/sam_shleifer) Task: Summarization The following is copied from the authors' README. # Mixed & Stochastic Checkpoints We train a pegasus model with sampled gap sentence ratios on both C4 and HugeNews, and stochastically sample important sentences. The updated the results are reported in this table. | dataset | C4 | HugeNews | Mixed & Stochastic| | ---- | ---- | ---- | ----| | xsum | 45.20/22.06/36.99 | 47.21/24.56/39.25 | 47.60/24.83/39.64| | cnn_dailymail | 43.90/21.20/40.76 | 44.17/21.47/41.11 | 44.16/21.56/41.30| | newsroom | 45.07/33.39/41.28 | 45.15/33.51/41.33 | 45.98/34.20/42.18| | multi_news | 46.74/17.95/24.26 | 47.52/18.72/24.91 | 47.65/18.75/24.95| | gigaword | 38.75/19.96/36.14 | 39.12/19.86/36.24 | 39.65/20.47/36.76| | wikihow | 43.07/19.70/34.79 | 41.35/18.51/33.42 | 46.39/22.12/38.41 *| | reddit_tifu | 26.54/8.94/21.64 | 26.63/9.01/21.60 | 27.99/9.81/22.94| | big_patent | 53.63/33.16/42.25 | 53.41/32.89/42.07 | 52.29/33.08/41.66 *| | arxiv | 44.70/17.27/25.80 | 44.67/17.18/25.73 | 44.21/16.95/25.67| | pubmed | 45.49/19.90/27.69 | 45.09/19.56/27.42 | 45.97/20.15/28.25| | aeslc | 37.69/21.85/36.84 | 37.40/21.22/36.45 | 37.68/21.25/36.51| | billsum | 57.20/39.56/45.80 | 57.31/40.19/45.82 | 59.67/41.58/47.59| The "Mixed & Stochastic" model has the following changes: - trained on both C4 and HugeNews (dataset mixture is weighted by their number of examples). - trained for 1.5M instead of 500k (we observe slower convergence on pretraining perplexity). - the model uniformly sample a gap sentence ratio between 15% and 45%. - importance sentences are sampled using a 20% uniform noise to importance scores. - the sentencepiece tokenizer is updated to be able to encode newline character. (*) the numbers of wikihow and big_patent datasets are not comparable because of change in tokenization and data: - wikihow dataset contains newline characters which is useful for paragraph segmentation, the C4 and HugeNews model's sentencepiece tokenizer doesn't encode newline and loose this information. - we update the BigPatent dataset to preserve casing, some format cleanings are also changed, please refer to change in TFDS. The "Mixed & Stochastic" model has the following changes (from pegasus-large in the paper): trained on both C4 and HugeNews (dataset mixture is weighted by their number of examples). trained for 1.5M instead of 500k (we observe slower convergence on pretraining perplexity). the model uniformly sample a gap sentence ratio between 15% and 45%. importance sentences are sampled using a 20% uniform noise to importance scores. the sentencepiece tokenizer is updated to be able to encode newline character. Citation ``` @misc{zhang2019pegasus, title={PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization}, author={Jingqing Zhang and Yao Zhao and Mohammad Saleh and Peter J. Liu}, year={2019}, eprint={1912.08777}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```