DunnBC22 commited on
Commit
35192d6
1 Parent(s): fc91e6e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -10
README.md CHANGED
@@ -7,9 +7,6 @@ model-index:
7
  results: []
8
  ---
9
 
10
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
- should probably proofread and complete it, then remove this comment. -->
12
-
13
  # bart-base-News_Summarization_CNN
14
 
15
  This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on the None dataset.
@@ -18,17 +15,23 @@ It achieves the following results on the evaluation set:
18
 
19
  ## Model description
20
 
21
- More information needed
 
 
22
 
23
  ## Intended uses & limitations
24
 
25
- More information needed
26
 
27
  ## Training and evaluation data
28
 
29
  More information needed
30
 
31
  ## Training procedure
 
 
 
 
32
 
33
  ### Training hyperparameters
34
 
@@ -46,11 +49,10 @@ The following hyperparameters were used during training:
46
 
47
  ### Training results
48
 
49
- | Training Loss | Epoch | Step | Validation Loss |
50
- |:-------------:|:-----:|:----:|:---------------:|
51
- | 0.7491 | 1.0 | 1089 | 0.1618 |
52
- | 0.1641 | 2.0 | 2178 | 0.1603 |
53
-
54
 
55
  ### Framework versions
56
 
 
7
  results: []
8
  ---
9
 
 
 
 
10
  # bart-base-News_Summarization_CNN
11
 
12
  This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on the None dataset.
 
15
 
16
  ## Model description
17
 
18
+ Using the dataset from the following link, I trained a text summarization model.
19
+
20
+ https://www.kaggle.com/datasets/hadasu92/cnn-articles-after-basic-cleaning
21
 
22
  ## Intended uses & limitations
23
 
24
+ I used this to improve my skillset. I thank all of authors of the different technologies and dataset(s) for their contributions that have this possible. I am not too worried about getting credit for my part, but make sure to properly cite the authors of the different technologies and dataset(s) as they absolutely deserve credit for their contributions.
25
 
26
  ## Training and evaluation data
27
 
28
  More information needed
29
 
30
  ## Training procedure
31
+ CPU trained on all samples where the article length is less than 820 words and the summary length is no more than 52 words in length. Additionally, any sample that was missing a new article or summarization was removed. In all, 24,911 out of the possible 42,025 samples were used for training/testing/evaluation.
32
+
33
+ Here is the link to the code that was used to train this model:
34
+ https://github.com/DunnBC22/NLP_Projects/blob/main/Text%20Summarization/CNN%20News%20Text%20Summarization.ipynb
35
 
36
  ### Training hyperparameters
37
 
 
49
 
50
  ### Training results
51
 
52
+ | Training Loss | Epoch | Step | Validation Loss | rouge1 | rouge2 | rougeL | rougeLsum |
53
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:--------:|:----------:|:----------:|
54
+ | 0.7491 | 1.0 | 1089 | 0.1618 | N/A | N/A | N/A | N/A |
55
+ | 0.1641 | 2.0 | 2178 | 0.1603 | 0.834343 | 0.793822 | 0.823824 | 0.823778 |
 
56
 
57
  ### Framework versions
58