LoicDL commited on
Commit
0bf703e
1 Parent(s): b19127c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +29 -2
README.md CHANGED
@@ -1,5 +1,5 @@
1
  <p align="center" style="margin:0;padding:0">
2
- <img src="https://huggingface.co/LoicDL/NewsBERTje-base/img/newsbertje-logo.png" alt="newsbertje banner" width="400" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
3
  </p>
4
 
5
  <div style="margin:auto; text-align:center">
@@ -9,7 +9,34 @@
9
 
10
 
11
  ## Model description
 
 
 
12
 
 
13
 
 
14
 
15
- ## Usage
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  <p align="center" style="margin:0;padding:0">
2
+ <img src="img/newsbertje-logo.png" alt="newsbertje banner" width="250" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
3
  </p>
4
 
5
  <div style="margin:auto; text-align:center">
 
9
 
10
 
11
  ## Model description
12
+ NewsBERTje is a domain-adaptated Dutch BERT-based model, aimed towards the processing of texts in the news domain.
13
+ We set up a domain-specific corpus consisting of 20 million tokens of news articles from online versions of Dutch (and Flemish) newspapers such as NOS, De Morgen, Het Nieuwsblad, Het Laatste Nieuws, De Standaard and Het Belang van Limburg, as well as articles published on the news website of the Flemish public broadcasting agency VRT News.
14
+ Then, the model was trained with the supplemental corpus for a total of 4 epochs keeping the same hyperparameter configuration that was used in the model's original pre-training (BERTje) .
15
 
16
+ ## Performance
17
 
18
+ The model's performance was benchmarked by evaluating it on a large number of tasks within the Dutch news domain such as news sentiment classification, news event prominence classification, sarcasm and partisanship detection and both coarse -and fine-grained news topic classification. NewsBERTje outperforms both other Dutch BERT models such as BERTje, RobBERTje and RobBERT and a series of generative LLMs (zero-shot settings) on each of these tasks.
19
 
20
+
21
+ | Model |Topic Classification (fine) | Topic CLassification (coarse) | Sarcasm Detection | Sentiment Classification | Prominence Classification | Partizanship Detection |
22
+ |:-------------:|:-----:|:----:|:---------------:|:-------------:|:-----:|:----:|
23
+ | BERTje | 92.9 | 82.2 | 92.9 |43.8 | 74.4 | 57.8 |
24
+ | RobBERT-2023 | 91.5 | 80.1 | 91.6 |40.7 | 69.7 | 41.7 |
25
+ | Llama 3.1-8B-Instruct | 81.5 | 64.7 | 78.4 |33.4 | 68.7 | 49.7 |
26
+ | GPT 3.5 Turbo | 84.6 | 49.5 | 51.1 |35.8 | 64.3 | 51.6 |
27
+ | GPT 4o | 92.6 | 64.8 | 82.6 | 42.6 | 70.8 | 58.7 |
28
+ | **NewsBERTje** | **93.8** | **83.6** | **93.8** | **49.2** | **73.8** | **61.0** |
29
+
30
+ ## Citation
31
+
32
+ If you use this model, please cite our work as:
33
+
34
+ ```
35
+ @inproceedings{de2024enhancing,
36
+ title={Enhancing Unrestricted Cross-Document Event Coreference with Graph Reconstruction Networks},
37
+ author={De Langhe, Loic and De Clercq, Orph{\'e}e and Hoste, Veronique},
38
+ booktitle={Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)},
39
+ pages={6122--6133},
40
+ year={2024}
41
+ }
42
+ ```