FactNews / README.md
eduagarcia's picture
Update README.md
683cbf3 verified
|
raw
history blame
6.18 kB
metadata
dataset_info:
  - config_name: bias_prediction
    features:
      - name: file
        dtype: string
      - name: id_sente
        dtype: string
      - name: id_article
        dtype: string
      - name: domain
        dtype: string
      - name: year
        dtype: string
      - name: sentences
        dtype: string
      - name: label
        dtype: int64
      - name: label_text
        dtype: string
    splits:
      - name: train
        num_bytes: 163041
        num_examples: 738
      - name: full_train
        num_bytes: 951010
        num_examples: 4403
      - name: test
        num_bytes: 384327
        num_examples: 1788
    download_size: 718605
    dataset_size: 1498378
  - config_name: factuality_prediction
    features:
      - name: file
        dtype: string
      - name: id_sente
        dtype: string
      - name: id_article
        dtype: string
      - name: domain
        dtype: string
      - name: year
        dtype: string
      - name: sentences
        dtype: string
      - name: label
        dtype: int64
      - name: label_text
        dtype: string
    splits:
      - name: train
        num_bytes: 606722
        num_examples: 2826
      - name: full_train
        num_bytes: 944929
        num_examples: 4403
      - name: test
        num_bytes: 381863
        num_examples: 1788
    download_size: 927856
    dataset_size: 1933514
  - config_name: original
    features:
      - name: file
        dtype: string
      - name: id_sente
        dtype: string
      - name: id_article
        dtype: string
      - name: domain
        dtype: string
      - name: year
        dtype: string
      - name: sentences
        dtype: string
      - name: classe
        dtype: int64
      - name: label_text
        dtype: string
    splits:
      - name: train
        num_bytes: 1317047
        num_examples: 6191
    download_size: 516651
    dataset_size: 1317047
configs:
  - config_name: bias_prediction
    data_files:
      - split: train
        path: bias_prediction/train-*
      - split: full_train
        path: bias_prediction/full_train-*
      - split: test
        path: bias_prediction/test-*
  - config_name: factuality_prediction
    data_files:
      - split: train
        path: factuality_prediction/train-*
      - split: full_train
        path: factuality_prediction/full_train-*
      - split: test
        path: factuality_prediction/test-*
  - config_name: original
    data_files:
      - split: train
        path: original/train-*
license: unknown
task_categories:
  - text-classification
language:
  - pt
  - por
pretty_name: FactNews
size_categories:
  - 1K<n<10K
multilinguality:
  - monolingual
language_creators:
  - found
annotations_creators:
  - expert-generated
tags:
  - subjectivity
  - mediabias
  - media-bias

Disclaimer

I am not the author of this dataset, this is a modified version of the FactCheck dataset on HuggingFace, the original data is made avaliable by Vargas et. al, 2023 and can be downloaded from the link: https://github.com/franciellevargas/FactNews

Modifications:

  • The "original" subset contains the unmodified original CSV
  • The subsets for the task of "bias_prediction" and "factuality_prediction" were splited between train (70%) AND test (30%) by randomly selecting sentences grouped by their id_article. This configuration difers from the authors, who made a 90%/10% 10-fold split on the papers.
  • Each task contains an unbalanced split (full-train) and the balanced-split (train)

Sentence-Level Annotated Dataset for Predicting Factuality of News and Bias of Media Outlets in Portuguese

Automated fact-checking and news credibility verification at scale require accurate prediction of news factuality and media bias. Here, we introduce a large sentence-level dataset, titled FactNews, composed of 6,191 sentences expertly annotated according to factuality and media bias definitions proposed by AllSides. We used the FactNews to assess the overall reliability of news sources by formulating two text classification problems for predicting sentence-level factuality of news reporting and bias of media outlets. Our experiments demonstrate that biased sentences present a higher number of words compared to factual sentences, besides having a predominance of emotions. Hence, the fine-grained analysis of subjectivity and impartiality of news articles showed promising results for predicting the reliability of the entire media outlet. Finally, due to the severity of fake news and political polarization in Brazil, and the lack of research for Portuguese, both dataset and baseline were proposed for Brazilian Portuguese. The following table describes in detail the FactNews labels, documents, and stories:

Factual Quotes Biased Total sentences Total news stories Total news documents
4,242 1,391 558 6,161 100 300

Sources:

  • Media 1: Folha de São Paulo
  • Media 2: Estadão
  • Media 3: O Globo

Paper Results:

Sentence-Level Media Bias Prediction (90%/10% 10-fold split)

  • 67% (F1-Score) by Fine-tuned mBert-case

Sentence-Level Factuality Prediction (90%/10% 10-fold split)

  • 88% (F1-Score) by Fine-tuned mBert-case

Citation

Vargas, F., Jaidka, K., Pardo, T.A.S., Benevenuto, F. (2023). Predicting Sentence-Level Factuality of News and Bias of Media Outlets. Proceedings of the 14th International Conference on Recent Advances in Natural Language Processing, pp.1197--1206. Varna, Bulgaria. Association for Computational Linguistics (ACL).

Bibtex

@inproceedings{vargas-etal-2023-predicting,
    title = "Predicting Sentence-Level Factuality of News and Bias of Media Outlets",
    author = "Vargas, Francielle  and
      Jaidka, Kokil  and
      Pardo, Thiago  and
      Benevenuto, Fabr{\'\i}cio",
    editor = "Mitkov, Ruslan  and
      Angelova, Galia",
    booktitle = "Proceedings of the 14th International Conference on Recent Advances in Natural Language Processing",
    month = sep,
    year = "2023",
    address = "Varna, Bulgaria",
    publisher = "INCOMA Ltd., Shoumen, Bulgaria",
    url = "https://aclanthology.org/2023.ranlp-1.127",
    pages = "1197--1206",
}

Dataset Description