Update README.md
Browse files
README.md
CHANGED
@@ -100,7 +100,8 @@ license: apache-2.0
|
|
100 |
|
101 |
# TL;DR
|
102 |
|
103 |
-
If you already know T5, FLAN-T5 is just better at everything
|
|
|
104 |
> Flan-PaLM 540B achieves state-of-the-art performance on several benchmarks, such as 75.2% on five-shot MMLU. We also publicly release Flan-T5 checkpoints,1 which achieve strong few-shot performance even compared to much larger models, such as PaLM 62B. Overall, instruction finetuning is a general method for improving the performance and usability of pretrained language models.
|
105 |
|
106 |
**Disclaimer**: Content from **this** model card has been written by the Hugging Face team, and parts of it were copy pasted from the [T5 model card](https://huggingface.co/t5-large).
|
|
|
100 |
|
101 |
# TL;DR
|
102 |
|
103 |
+
If you already know T5, FLAN-T5 is just better at everything.
|
104 |
+
As mentioned in the first few lines of the abstract :
|
105 |
> Flan-PaLM 540B achieves state-of-the-art performance on several benchmarks, such as 75.2% on five-shot MMLU. We also publicly release Flan-T5 checkpoints,1 which achieve strong few-shot performance even compared to much larger models, such as PaLM 62B. Overall, instruction finetuning is a general method for improving the performance and usability of pretrained language models.
|
106 |
|
107 |
**Disclaimer**: Content from **this** model card has been written by the Hugging Face team, and parts of it were copy pasted from the [T5 model card](https://huggingface.co/t5-large).
|