|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
"""Being Right for Whose Right Reasons?""" |
|
|
|
import json |
|
import os |
|
import textwrap |
|
|
|
import datasets |
|
|
|
|
|
MAIN_CITATION = """\ |
|
@inproceedings{thorn-jakobsen-etal-2023-right, |
|
title = {Being Right for Whose Right Reasons?}, |
|
author = {Thorn Jakobsen, Terne Sasha and |
|
Cabello, Laura and |
|
S{\o}gaard, Anders}, |
|
booktitle = {Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, |
|
year = {2023}, |
|
publisher = {Association for Computational Linguistics}, |
|
url = {https://aclanthology.org/2023.acl-long.59}, |
|
doi = {10.18653/v1/2023.acl-long.59}, |
|
pages = {1033--1054} |
|
} |
|
""" |
|
|
|
_DESCRIPTION = """\ |
|
Explainability methods are used to benchmark |
|
the extent to which model predictions align |
|
with human rationales i.e., are 'right for the |
|
right reasons'. Previous work has failed to acknowledge, however, |
|
that what counts as a rationale is sometimes subjective. This paper |
|
presents what we think is a first of its kind, a |
|
collection of human rationale annotations augmented with the annotators demographic information. |
|
""" |
|
|
|
SST2_LABELS = ["negative", "positive", "no sentiment"] |
|
|
|
DYNASENT_LABELS = ["negative", "positive", "no sentiment"] |
|
|
|
MAIN_PATH = "https://huggingface.co/datasets/coastalcph/fair-rationales/resolve/main" |
|
|
|
|
|
class FairRationalesConfig(datasets.BuilderConfig): |
|
"""BuilderConfig for FairRationales.""" |
|
|
|
def __init__( |
|
self, |
|
name, |
|
url, |
|
data_url, |
|
attributes, |
|
citation, |
|
description, |
|
label_classes=None, |
|
label_classes_original=None, |
|
**kwargs, |
|
): |
|
"""BuilderConfig for FairRationales. |
|
Args: |
|
label_column: `string`, name of the column in the jsonl file corresponding |
|
to the label |
|
url: `string`, url for the original project |
|
data_url: `string`, url to download the zip file from |
|
data_file: `string`, filename for data set |
|
citation: `string`, citation for the data set |
|
url: `string`, url for information about the data set |
|
label_classes: `list[string]`, the list of classes if the label is |
|
categorical. If not provided, then the label will be of type |
|
`datasets.Value('float32')`. |
|
attributes: `List<string>`, names of the protected attributes |
|
**kwargs: keyword arguments forwarded to super. |
|
""" |
|
super(FairRationalesConfig, self).__init__(version=datasets.Version("1.0.0", ""), **kwargs) |
|
self.name = name |
|
self.label_classes = label_classes |
|
self.label_classes_original = label_classes_original |
|
self.attributes = attributes |
|
self.url = url |
|
self.data_url = data_url |
|
self.description = description |
|
self.citation = citation |
|
|
|
|
|
class FairRationales(datasets.GeneratorBasedBuilder): |
|
"""FairRationales: A multilingual benchmark for evaluating fairness in legal text processing. Version 1.0""" |
|
|
|
BUILDER_CONFIGS = [ |
|
FairRationalesConfig( |
|
name="sst2", |
|
description=textwrap.dedent( |
|
"""\ |
|
The Stanford Sentiment Treebank is a corpus with fully labeled parse trees that allows for a complete analysis of the compositional effects of sentiment in language. |
|
Binary classification experiments on full sentences (negative or somewhat negative vs somewhat positive or positive with neutral sentences discarded) refer to the dataset as SST-2 or SST binary. |
|
This is a subset of the original data where annotators were allowed to re-annotate an instance as neutral or "no sentiment" and provide rationales for it. |
|
Therefore, this is a ternary text classification task (hihgly unbalanced for the 'no sentiment' class). |
|
Given a sentence, the goal is to predict the sentiment it conveys (positive, neutral, no sentiment).""" |
|
), |
|
|
|
label_classes=SST2_LABELS, |
|
label_classes_original=["negative", "positive"], |
|
attributes=[ |
|
("group_id", ["BO", "BY", "WO", "WY", "LO", "LY"]), |
|
("sst2_id", datasets.Value("int32")), |
|
("sst2_split", datasets.Value("string")), |
|
], |
|
data_url=os.path.join(MAIN_PATH, "sst2.zip"), |
|
url="https://huggingface.co/datasets/sst2", |
|
citation=textwrap.dedent( |
|
"""\ |
|
@inproceedings{socher-etal-2013-recursive, |
|
title = "Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank", |
|
author = "Socher, Richard and |
|
Perelygin, Alex and |
|
Wu, Jean and |
|
Chuang, Jason and |
|
Manning, Christopher D. and |
|
Ng, Andrew and |
|
Potts, Christopher", |
|
booktitle = "Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing", |
|
month = oct, |
|
year = "2013", |
|
address = "Seattle, Washington, USA", |
|
publisher = "Association for Computational Linguistics", |
|
url = "https://www.aclweb.org/anthology/D13-1170", |
|
pages = "1631--1642", |
|
} |
|
}""" |
|
), |
|
), |
|
FairRationalesConfig( |
|
name="dynasent", |
|
description=textwrap.dedent( |
|
"""\ |
|
DynaSent is an English-language benchmark task for ternary (positive/negative/neutral) sentiment analysis. |
|
This is a subset of the original data where annotators were allowed to re-annotate an instance as neutral or "no sentiment" and provide rationales for it. |
|
Therefore, this is a ternary text classification task (hihgly unbalanced for the 'no sentiment' class). |
|
Given a sentence, the goal is to predict the sentiment it conveys (positive, neutral, no sentiment). |
|
""" |
|
), |
|
label_classes=DYNASENT_LABELS, |
|
label_classes_original=["negative", "positive"], |
|
attributes=[ |
|
("group_id", ["BO", "BY", "WO", "WY", "LO", "LY"]), |
|
], |
|
data_url=os.path.join(MAIN_PATH, "dynasent.zip"), |
|
url="https://huggingface.co/datasets/dynabench/dynasent", |
|
citation=textwrap.dedent( |
|
"""\ |
|
@article{potts-etal-2020-dynasent, |
|
title={{DynaSent}: A Dynamic Benchmark for Sentiment Analysis}, |
|
author={Potts, Christopher and Wu, Zhengxuan and Geiger, Atticus and Kiela, Douwe}, |
|
journal={arXiv preprint arXiv:2012.15349}, |
|
url={https://arxiv.org/abs/2012.15349}, |
|
year={2020} |
|
}""" |
|
), |
|
), |
|
FairRationalesConfig( |
|
name="cose", |
|
description=textwrap.dedent( |
|
"""\ |
|
Common Sense Explanations (CoS-E) allows for training language models to automatically |
|
generate explanations that can be used during training and inference in a novel |
|
Commonsense Auto-Generated Explanation (CAGE) framework. |
|
This is a subset of the original data where annotators were allowed to re-annotate the questions and provide rationales for it. |
|
This is a question-answering task with 1 correct answer out of 5 options. |
|
Given a question, the goal is to predict the right answer. |
|
""" |
|
), |
|
label_classes_original=["A", "B", "C", "D", "E"], |
|
attributes=[ |
|
("group_id", ["BO", "BY", "WO", "WY", "LO", "LY"]), |
|
], |
|
data_url=os.path.join(MAIN_PATH, "cose.zip"), |
|
url="https://huggingface.co/datasets/cos_e", |
|
citation=textwrap.dedent( |
|
"""\ |
|
@inproceedings{rajani2019explain, |
|
title = "Explain Yourself! Leveraging Language models for Commonsense Reasoning", |
|
author = "Rajani, Nazneen Fatema and |
|
McCann, Bryan and |
|
Xiong, Caiming and |
|
Socher, Richard", |
|
year="2019", |
|
booktitle = "Proceedings of the 2019 Conference of the Association for Computational Linguistics (ACL2019)", |
|
url ="https://arxiv.org/abs/1906.02361" |
|
} |
|
}""" |
|
), |
|
), |
|
] |
|
|
|
def _info(self): |
|
features = {"QID": datasets.Value("string"), |
|
"text_id": datasets.Value("int64"), |
|
"sentence": datasets.Value("string"), |
|
"label_index": datasets.Value("int64"), |
|
"original_label": datasets.ClassLabel(names=self.config.label_classes_original), |
|
"rationale": datasets.Value("string"), |
|
"rationale_index": datasets.Value("string"), |
|
"rationale_binary": datasets.Value("string"), |
|
"age": datasets.Value("int32"), |
|
"ethnicity": datasets.Value("string"), |
|
"originaldata_id": datasets.Value("string"), |
|
"annotator_ID": datasets.Value("int64"), |
|
"english_proficiency": datasets.Value("string"), |
|
"attentioncheck": datasets.Value("string"), |
|
"gender": datasets.Value("string"), |
|
"recruitment_age": datasets.Value("string"), |
|
"recruitment_ethnicity": datasets.Value("string") |
|
} |
|
if self.config.name == "cose": |
|
features["label"] = datasets.Value("string") |
|
else: |
|
features["label"] = datasets.ClassLabel(names=self.config.label_classes) |
|
for attribute_name, attribute_groups in self.config.attributes: |
|
if "sst2" not in attribute_name: |
|
features[attribute_name] = datasets.ClassLabel(names=attribute_groups) |
|
else: |
|
features[attribute_name] = attribute_groups |
|
return datasets.DatasetInfo( |
|
description=self.config.description, |
|
features=datasets.Features(features), |
|
homepage=self.config.url, |
|
citation=self.config.citation + "\n" + MAIN_CITATION, |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
data_dir = dl_manager.download_and_extract(self.config.data_url) |
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TRAIN, |
|
|
|
gen_kwargs={ |
|
"filepath": os.path.join(data_dir, self.config.name, "train.jsonl"), |
|
"split": "train" |
|
}, |
|
), |
|
] |
|
|
|
def _generate_examples(self, filepath, split): |
|
"""This function returns the examples in the raw (text) form.""" |
|
with open(filepath, encoding="utf-8") as f: |
|
for id_, row in enumerate(f): |
|
data = json.loads(row) |
|
example = { |
|
"sentence": data["sentence"], |
|
"label": data["label"], |
|
"text_id": data["text_id"], |
|
"QID": data["QID"], |
|
"label_index": data["label_index"], |
|
"original_label": data["original_label"], |
|
"rationale": data["rationale"], |
|
"rationale_index": data["rationale_index"], |
|
"rationale_binary": data["rationale_binary"], |
|
"age": data["age"], |
|
"recruitment_age": data["recruitment_age"], |
|
"ethnicity": data["ethnicity"], |
|
"recruitment_ethnicity": data["recruitment_ethnicity"], |
|
"gender": data["gender"], |
|
"originaldata_id": data["originaldata_id"], |
|
"annotator_ID": data["annotator_ID"], |
|
"english_proficiency": data["english_proficiency"], |
|
"attentioncheck": data["attentioncheck"] |
|
} |
|
for attribute_name, _ in self.config.attributes: |
|
example[attribute_name] = data[attribute_name] |
|
if self.config.name == "sst2": |
|
example["sst2_id"] = data["sst2_id"] |
|
example["sst2_split"] = data["sst2_split"] |
|
yield id_, example |