Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
@@ -5,9 +5,9 @@ tags:
|
|
5 |
- abstractive
|
6 |
- hybrid
|
7 |
- multistep
|
|
|
8 |
datasets: dennlinger/eur-lex-sum
|
9 |
pipeline_tag: summarization
|
10 |
-
base_model: LongT5
|
11 |
model-index:
|
12 |
- name: BART
|
13 |
results:
|
@@ -27,7 +27,7 @@ model-index:
|
|
27 |
- type: BERTScore
|
28 |
value: 0.8496733237911203
|
29 |
- type: BARTScore
|
30 |
-
value: -
|
31 |
- type: BLANC
|
32 |
value: 0.17446891570320744
|
33 |
---
|
@@ -38,7 +38,7 @@ model-index:
|
|
38 |
---
|
39 |
### Model Description
|
40 |
|
41 |
-
This model is a fine-tuned version of LongT5. The research involves a multi-step summarization approach to long, legal documents. Many decisions in the renewables energy space are heavily dependent on regulations. But these regulations are often long and complicated. The proposed architecture first uses one or more extractive summarization steps to compress the source text, before the final summary is created by the abstractive summarization model. This fine-tuned abstractive model has been trained on a dataset, pre-processed through extractive summarization by
|
42 |
|
43 |
The dataset used by this model is the [EUR-lex-sum](https://huggingface.co/datasets/dennlinger/eur-lex-sum) dataset. The evaluation metrics can be found in the metadata of this model card.
|
44 |
This paper was introduced by the master thesis of Mika Sie at the University Utrecht in collaboration with Power2x. More information can be found in PAPER_LINK.
|
@@ -59,7 +59,7 @@ This paper was introduced by the master thesis of Mika Sie at the University Utr
|
|
59 |
---
|
60 |
### Direct Use
|
61 |
|
62 |
-
This model can be directly used for summarizing long, legal documents. However, it is recommended to first use an extractive summarization tool, such as
|
63 |
An example using the Huggingface pipeline could be:
|
64 |
|
65 |
```python
|
|
|
5 |
- abstractive
|
6 |
- hybrid
|
7 |
- multistep
|
8 |
+
base_model: LongT5
|
9 |
datasets: dennlinger/eur-lex-sum
|
10 |
pipeline_tag: summarization
|
|
|
11 |
model-index:
|
12 |
- name: BART
|
13 |
results:
|
|
|
27 |
- type: BERTScore
|
28 |
value: 0.8496733237911203
|
29 |
- type: BARTScore
|
30 |
+
value: -2.2194945887200546
|
31 |
- type: BLANC
|
32 |
value: 0.17446891570320744
|
33 |
---
|
|
|
38 |
---
|
39 |
### Model Description
|
40 |
|
41 |
+
This model is a fine-tuned version of LongT5. The research involves a multi-step summarization approach to long, legal documents. Many decisions in the renewables energy space are heavily dependent on regulations. But these regulations are often long and complicated. The proposed architecture first uses one or more extractive summarization steps to compress the source text, before the final summary is created by the abstractive summarization model. This fine-tuned abstractive model has been trained on a dataset, pre-processed through extractive summarization by No extractive model with No ratio ratio. The research has used multiple extractive-abstractive model combinations, which can be found on https://huggingface.co/MikaSie. To obtain optimal results, feed the model an extractive summary as input as it was designed this way!
|
42 |
|
43 |
The dataset used by this model is the [EUR-lex-sum](https://huggingface.co/datasets/dennlinger/eur-lex-sum) dataset. The evaluation metrics can be found in the metadata of this model card.
|
44 |
This paper was introduced by the master thesis of Mika Sie at the University Utrecht in collaboration with Power2x. More information can be found in PAPER_LINK.
|
|
|
59 |
---
|
60 |
### Direct Use
|
61 |
|
62 |
+
This model can be directly used for summarizing long, legal documents. However, it is recommended to first use an extractive summarization tool, such as No extractive model, to compress the source text before feeding it to this model. This model has been specifically designed to work with extractive summaries.
|
63 |
An example using the Huggingface pipeline could be:
|
64 |
|
65 |
```python
|