bgonzalezbustamante commited on
Commit
11ac6c9
1 Parent(s): 0d84fdf

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -8
README.md CHANGED
@@ -21,8 +21,8 @@ This is a fine-tuned roBERTa-base model trained using as a base model Twitter-ro
21
  The dataset comprises almost 5M data points from three Latin American protest events: (a) protests against the coronavirus and judicial reform measures in Argentina during August 2020; (b) protests against education budget cuts in Brazil in May 2019; and (c) the social outburst in Chile stemming from protests against the underground fare hike in October 2019. We are focusing on interactions in Spanish to elaborate a gold standard for digital interactions in this language, therefore, we prioritise Argentinian and Chilean data.
22
 
23
  - [GitHub repository](https://github.com/training-datalab/gold-standard-toxicity).
24
- - [Dataset on Zenodo](https://zenodo.org/doi/10.5281/zenodo.12574288).
25
- - [Reference paper](https://doi.org/10.48550/arXiv.2409.09741).
26
 
27
  **Labels: NONTOXIC and TOXIC.**
28
 
@@ -32,12 +32,10 @@ WIP
32
 
33
  ## Validation Metrics
34
 
35
- | Metric | Value |
36
- |---|---|
37
- | Accuracy | 0.790 |
38
- | Precision | 0.920 |
39
- | Reccall | 0.657 |
40
- | F1-Score | 0.767 |
41
 
42
  ## License
43
 
 
21
  The dataset comprises almost 5M data points from three Latin American protest events: (a) protests against the coronavirus and judicial reform measures in Argentina during August 2020; (b) protests against education budget cuts in Brazil in May 2019; and (c) the social outburst in Chile stemming from protests against the underground fare hike in October 2019. We are focusing on interactions in Spanish to elaborate a gold standard for digital interactions in this language, therefore, we prioritise Argentinian and Chilean data.
22
 
23
  - [GitHub repository](https://github.com/training-datalab/gold-standard-toxicity).
24
+ - [Dataset on Zenodo](zenodo.org/doi/10.5281/zenodo.12574288).
25
+ - [Reference paper](arxiv.org/abs/2409.09741)
26
 
27
  **Labels: NONTOXIC and TOXIC.**
28
 
 
32
 
33
  ## Validation Metrics
34
 
35
+ - Accuracy: 0.790
36
+ - Precision: 0.920
37
+ - Reccall: 0.657
38
+ - F1-Score: 0.767
 
 
39
 
40
  ## License
41