Update README.md
Browse files
README.md
CHANGED
@@ -7,9 +7,6 @@ model-index:
|
|
7 |
results: []
|
8 |
---
|
9 |
|
10 |
-
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
11 |
-
should probably proofread and complete it, then remove this comment. -->
|
12 |
-
|
13 |
# bart-base-News_Summarization_CNN
|
14 |
|
15 |
This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on the None dataset.
|
@@ -18,17 +15,23 @@ It achieves the following results on the evaluation set:
|
|
18 |
|
19 |
## Model description
|
20 |
|
21 |
-
|
|
|
|
|
22 |
|
23 |
## Intended uses & limitations
|
24 |
|
25 |
-
|
26 |
|
27 |
## Training and evaluation data
|
28 |
|
29 |
More information needed
|
30 |
|
31 |
## Training procedure
|
|
|
|
|
|
|
|
|
32 |
|
33 |
### Training hyperparameters
|
34 |
|
@@ -46,11 +49,10 @@ The following hyperparameters were used during training:
|
|
46 |
|
47 |
### Training results
|
48 |
|
49 |
-
| Training Loss | Epoch | Step | Validation Loss |
|
50 |
-
|
51 |
-
| 0.7491 | 1.0 | 1089 | 0.1618 |
|
52 |
-
| 0.1641 | 2.0 | 2178 | 0.1603 |
|
53 |
-
|
54 |
|
55 |
### Framework versions
|
56 |
|
|
|
7 |
results: []
|
8 |
---
|
9 |
|
|
|
|
|
|
|
10 |
# bart-base-News_Summarization_CNN
|
11 |
|
12 |
This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on the None dataset.
|
|
|
15 |
|
16 |
## Model description
|
17 |
|
18 |
+
Using the dataset from the following link, I trained a text summarization model.
|
19 |
+
|
20 |
+
https://www.kaggle.com/datasets/hadasu92/cnn-articles-after-basic-cleaning
|
21 |
|
22 |
## Intended uses & limitations
|
23 |
|
24 |
+
I used this to improve my skillset. I thank all of authors of the different technologies and dataset(s) for their contributions that have this possible. I am not too worried about getting credit for my part, but make sure to properly cite the authors of the different technologies and dataset(s) as they absolutely deserve credit for their contributions.
|
25 |
|
26 |
## Training and evaluation data
|
27 |
|
28 |
More information needed
|
29 |
|
30 |
## Training procedure
|
31 |
+
CPU trained on all samples where the article length is less than 820 words and the summary length is no more than 52 words in length. Additionally, any sample that was missing a new article or summarization was removed. In all, 24,911 out of the possible 42,025 samples were used for training/testing/evaluation.
|
32 |
+
|
33 |
+
Here is the link to the code that was used to train this model:
|
34 |
+
https://github.com/DunnBC22/NLP_Projects/blob/main/Text%20Summarization/CNN%20News%20Text%20Summarization.ipynb
|
35 |
|
36 |
### Training hyperparameters
|
37 |
|
|
|
49 |
|
50 |
### Training results
|
51 |
|
52 |
+
| Training Loss | Epoch | Step | Validation Loss | rouge1 | rouge2 | rougeL | rougeLsum |
|
53 |
+
|:-------------:|:-----:|:----:|:---------------:|:--------:|:--------:|:----------:|:----------:|
|
54 |
+
| 0.7491 | 1.0 | 1089 | 0.1618 | N/A | N/A | N/A | N/A |
|
55 |
+
| 0.1641 | 2.0 | 2178 | 0.1603 | 0.834343 | 0.793822 | 0.823824 | 0.823778 |
|
|
|
56 |
|
57 |
### Framework versions
|
58 |
|