Bert2Bert-HunSum-1 / README.md
NyistMilan's picture
Added more information to README.
c0fd531
|
raw
history blame
No virus
1.35 kB
---
datasets:
- SZTAKI-HLT/HunSum-1
language:
- hu
metrics:
- rouge
pipeline_tag: text2text-generation
inference:
parameters:
num_beams: 5
length_penalty: 2
max_length: 128
no_repeat_ngram_size: 3
early_stopping: True
tags:
- hubert
- bert
- summarization
---
# Model Card for Bert2Bert-HunSum-1
The Bert2Bert-HunSum-1 is a Hungarian abstractive summarization model, which was trained on the [SZTAKI-HLT/HunSum-1 dataset](https://huggingface.co/datasets/SZTAKI-HLT/HunSum-1).
The model is based on [SZTAKI-HLT/hubert-base-cc](https://huggingface.co/SZTAKI-HLT/hubert-base-cc).
## Intended uses & limitations
- **Model type:** Text Summarization
- **Language(s) (NLP):** Hungarian
- **Resource(s) for more information:**
- [GitHub Repo](https://github.com/dorinapetra/summarization)
## Parameters
- **Batch Size:** 13
- **Learning Rate:** 5e-5
- **Weight Decay:** 0.01
- **Warmup Steps:** 16000
- **Epochs:** 15
- **no_repeat_ngram_size:** 3
- **num_beams:** 5
- **early_stopping:** True
## Results
| Metric | Value |
| :------------ | :------------------------------------------ |
| ROUGE-1 | 28.52 |
| ROUGE-2 | 10.35 |
| ROUGE-L | 20.07 |