Update README.md
Browse files
README.md
CHANGED
@@ -118,7 +118,7 @@ Pass [other parameters related to beam search textgen](https://huggingface.co/bl
|
|
118 |
|
119 |
`kmfoda/booksum` dataset on HuggingFace - read [the original paper here](https://arxiv.org/abs/2105.08209). Summaries longer than 1024 LongT5 tokens were filtered out with the intent of preventing the model from learning to generate "partial" summaries.
|
120 |
|
121 |
-
|
122 |
|
123 |
## Training procedure
|
124 |
|
|
|
118 |
|
119 |
`kmfoda/booksum` dataset on HuggingFace - read [the original paper here](https://arxiv.org/abs/2105.08209). Summaries longer than 1024 LongT5 tokens were filtered out with the intent of preventing the model from learning to generate "partial" summaries.
|
120 |
|
121 |
+
_NOTE: early checkpoints of this model were trained on a "smaller" subsection of the dataset as it was filtered for summaries of **1024 characters**. This was subsequently caught and adjusted to **1024 tokens** and then trained further for at least five epochs._
|
122 |
|
123 |
## Training procedure
|
124 |
|