roemmele commited on
Commit
9f6021e
1 Parent(s): 8e81ae4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -11
README.md CHANGED
@@ -4,12 +4,13 @@ datasets:
4
  language:
5
  - en
6
  pipeline_tag: summarization
 
7
  ---
8
  # Model Card for roemmele/ablit-bart-base
9
 
10
  <!-- Provide a quick summary of what the model is/does. -->
11
 
12
- This model is initialized from facebook/bart-base. It has been fine-tuned on the AbLit dataset, which consists of abridged versions of books aligned with their original versions at the passage level. Given a text, the model generates an abridgement of the text based on what it has observed in AbLit. See the cited paper for more details.
13
 
14
  ## Model Details
15
 
@@ -20,15 +21,15 @@ This model is initialized from facebook/bart-base. It has been fine-tuned on the
20
  - **Developed by:** Language Weaver (Melissa Roemmele, Kyle Shaffer, Katrina Olsen, Yiyi Wang, and Steve DeNeefe)
21
  - **Model type:** Seq2SeqLM
22
  - **Language(s) (NLP):** English
23
- - **License:** [More Information Needed]
24
- - **Finetuned from model [optional]:** facebook/bart-base
25
 
26
- ### Model Sources [optional]
27
 
28
  <!-- Provide the basic links for the model. -->
29
 
30
  - **Repository:** [github.com/roemmele/AbLit](https://github.com/roemmele/AbLit)
31
- - **Paper [optional]:** [AbLit: A Resource for Analyzing and Generating Abridged Versions of English Literature](https://arxiv.org/pdf/2302.06579.pdf)
32
 
33
  ## Uses
34
 
@@ -81,12 +82,9 @@ data = load_dataset("roemmele/ablit", "chunks-10-sentences")
81
 
82
  We used the training script [here](https://github.com/huggingface/transformers/blob/main/examples/pytorch/summarization/run_summarization.py).
83
 
84
- Hyperparameter settings: We specified maximum length of 1024 for both the source (original passage) and target (abridged passage), and truncated all tokens beyond this limit. We evaluated each model on the AbLit development set after each epoch and concluded training when cross-entropy loss stopped decreasing. We used a batch size of 4. For all other hyperparameters we used the default values set by this script.
85
-
86
-
87
  #### Training Hyperparameters
88
 
89
- - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
90
 
91
  #### Speeds, Sizes, Times
92
 
@@ -109,7 +107,7 @@ The model obtained a ROUGE-L score of 0.78 on the AbLit test set. See the paper
109
 
110
  Our analysis shows that in comparison with human-authored abridgements, the model-generated abridgements tend to preserve more of the original text, suggesting it is challenging to learn what text can be removed while maintaining loyalty to the important parts of the original text.
111
 
112
- ## Citation [optional]
113
 
114
  <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
115
 
@@ -133,4 +131,4 @@ Melissa Roemmele
133
 
134
  ## Model Card Contact
135
 
136
 
4
  language:
5
  - en
6
  pipeline_tag: summarization
7
+ license: mit
8
  ---
9
  # Model Card for roemmele/ablit-bart-base
10
 
11
  <!-- Provide a quick summary of what the model is/does. -->
12
 
13
+ This model is initialized from facebook/bart-base. It has been fine-tuned on the AbLit dataset, which consists of abridged versions of books aligned with their original versions at the passage level. Given a text, the model generates an abridgement of the text based on what it has observed in AbLit. See the paper cited below for more details.
14
 
15
  ## Model Details
16
 
 
21
  - **Developed by:** Language Weaver (Melissa Roemmele, Kyle Shaffer, Katrina Olsen, Yiyi Wang, and Steve DeNeefe)
22
  - **Model type:** Seq2SeqLM
23
  - **Language(s) (NLP):** English
24
+ - **License:** mit
25
+ - **Finetuned from model:** [facebook/bart-base](https://huggingface.co/facebook/bart-base)
26
 
27
+ ### Model Sources
28
 
29
  <!-- Provide the basic links for the model. -->
30
 
31
  - **Repository:** [github.com/roemmele/AbLit](https://github.com/roemmele/AbLit)
32
+ - **Paper:** [AbLit: A Resource for Analyzing and Generating Abridged Versions of English Literature](https://arxiv.org/pdf/2302.06579.pdf)
33
 
34
  ## Uses
35
 
 
82
 
83
  We used the training script [here](https://github.com/huggingface/transformers/blob/main/examples/pytorch/summarization/run_summarization.py).
84
 
 
 
 
85
  #### Training Hyperparameters
86
 
87
+ We specified maximum length of 1024 for both the source (original passage) and target (abridged passage), and truncated all tokens beyond this limit. We evaluated each model on the AbLit development set after each epoch and concluded training when cross-entropy loss stopped decreasing. We used a batch size of 4. For all other hyperparameters we used the default values set by this script.
88
 
89
  #### Speeds, Sizes, Times
90
 
 
107
 
108
  Our analysis shows that in comparison with human-authored abridgements, the model-generated abridgements tend to preserve more of the original text, suggesting it is challenging to learn what text can be removed while maintaining loyalty to the important parts of the original text.
109
 
110
+ ## Citation
111
 
112
  <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
113
 
 
131
 
132
  ## Model Card Contact
133
 
134