aae2 / aae2.py
ArneBinder's picture
fix again https://huggingface.co/datasets/pie/aae2/discussions/1 (#3)
3cfad37 verified
raw
history blame contribute delete
No virus
7.49 kB
import os
from typing import Dict
import pandas as pd
from pie_modules.document.processing import RegexPartitioner
from pytorch_ie.annotations import BinaryRelation
from pytorch_ie.documents import (
TextDocumentWithLabeledSpansAndBinaryRelations,
TextDocumentWithLabeledSpansBinaryRelationsAndLabeledPartitions,
)
from pie_datasets.builders import BratBuilder
from pie_datasets.builders.brat import BratConfig, BratDocumentWithMergedSpans
from pie_datasets.core.dataset import DocumentConvertersType
from pie_datasets.document.processing import Caster, Converter, Pipeline
def get_split_paths(url_split_ids: str, subdirectory: str) -> Dict[str, str]:
df_splits = pd.read_csv(url_split_ids, sep=";")
splits2ids = df_splits.groupby(df_splits["SET"]).agg(list).to_dict()["ID"]
return {
split.lower(): [os.path.join(subdirectory, split_id) for split_id in split_ids]
for split, split_ids in splits2ids.items()
}
URL = "https://github.com/ArneBinder/pie-datasets/raw/83fb46f904b13f335b6da3cce2fc7004d802ce4e/data/datasets/ArgumentAnnotatedEssays-2.0/brat-project-final.zip"
URL_SPLIT_IDS = "https://raw.githubusercontent.com/ArneBinder/pie-datasets/83fb46f904b13f335b6da3cce2fc7004d802ce4e/data/datasets/ArgumentAnnotatedEssays-2.0/train-test-split.csv"
SPLIT_PATHS = get_split_paths(URL_SPLIT_IDS, subdirectory="brat-project-final")
DEFAULT_ATTRIBUTIONS_TO_RELATIONS_DICT = {"For": "supports", "Against": "attacks"}
def convert_aae2_claim_attributions_to_relations(
document: BratDocumentWithMergedSpans,
method: str,
attributions_to_relations_mapping: Dict[str, str] = DEFAULT_ATTRIBUTIONS_TO_RELATIONS_DICT,
major_claim_label: str = "MajorClaim",
claim_label: str = "Claim",
semantically_same_label: str = "semantically_same",
) -> TextDocumentWithLabeledSpansAndBinaryRelations:
"""This function collects the attributions of Claims from BratDocumentWithMergedSpans, and
build new relations between MajorClaims and Claims based on these attributions in the following
way:
1) "connect_first":
Each Claim points to the first MajorClaim,
and the other MajorClaim(s) is labeled as semantically same as the first MajorClaim.
The number of new relations created are: NoOfMajorClaim - 1 + NoOfClaim.
2) "connect_all":
Each Claim points to every MajorClaim; creating many-to-many relations.
The number of new relations created are: NoOfMajorClaim x NoOfClaim.
The attributions are transformed into the relation labels as listed in
DEFAULT_ATTRIBUTIONS_TO_RELATIONS_DICT dictionary.
"""
document = document.copy()
new_document = TextDocumentWithLabeledSpansAndBinaryRelations(
text=document.text, id=document.id, metadata=document.metadata
)
# import from document
spans = document.spans.clear()
new_document.labeled_spans.extend(spans)
relations = document.relations.clear()
new_document.binary_relations.extend(relations)
claim_attributes = [
attribute
for attribute in document.span_attributes
if attribute.annotation.label == claim_label
]
# get all MajorClaims
# sorted by start position to ensure the first MajorClaim is really the first one that occurs in the text
major_claims = sorted(
[mc for mc in new_document.labeled_spans if mc.label == major_claim_label],
key=lambda span: span.start,
)
if method == "connect_first":
if len(major_claims) > 0:
first_major_claim = major_claims.pop(0)
# Add relation between Claims and first MajorClaim
for claim_attribute in claim_attributes:
new_relation = BinaryRelation(
head=claim_attribute.annotation,
tail=first_major_claim,
label=attributions_to_relations_mapping[claim_attribute.value],
)
new_document.binary_relations.append(new_relation)
# Add relations between MajorClaims
for majorclaim in major_claims:
new_relation = BinaryRelation(
head=majorclaim,
tail=first_major_claim,
label=semantically_same_label,
)
new_document.binary_relations.append(new_relation)
elif method == "connect_all":
for major_claim in major_claims:
for claim_attribute in claim_attributes:
new_relation = BinaryRelation(
head=claim_attribute.annotation,
tail=major_claim,
label=attributions_to_relations_mapping[claim_attribute.value],
)
new_document.binary_relations.append(new_relation)
else:
raise ValueError(f"unknown method: {method}")
return new_document
def get_common_pipeline_steps(conversion_method: str) -> dict:
return dict(
convert=Converter(
function=convert_aae2_claim_attributions_to_relations,
method=conversion_method,
),
)
class ArgumentAnnotatedEssaysV2Config(BratConfig):
def __init__(self, conversion_method: str, **kwargs):
"""BuilderConfig for ArgumentAnnotatedEssaysV2.
Args:
conversion_method: either "connect_first" or "connect_all", see convert_aae2_claim_attributions_to_relations
**kwargs: keyword arguments forwarded to super.
"""
super().__init__(merge_fragmented_spans=True, **kwargs)
self.conversion_method = conversion_method
class ArgumentAnnotatedEssaysV2(BratBuilder):
BUILDER_CONFIG_CLASS = ArgumentAnnotatedEssaysV2Config
BASE_DATASET_PATH = "DFKI-SLT/brat"
BASE_DATASET_REVISION = "bb8c37d84ddf2da1e691d226c55fef48fd8149b5"
# we need to add None to the list of dataset variants to support the default dataset variant
BASE_BUILDER_KWARGS_DICT = {
dataset_variant: {"url": URL, "split_paths": SPLIT_PATHS}
for dataset_variant in [BratBuilder.DEFAULT_CONFIG_NAME, None]
}
BUILDER_CONFIGS = [
ArgumentAnnotatedEssaysV2Config(
name=BratBuilder.DEFAULT_CONFIG_NAME,
conversion_method="connect_first",
),
]
DOCUMENT_TYPES = {
BratBuilder.DEFAULT_CONFIG_NAME: BratDocumentWithMergedSpans,
}
@property
def document_converters(self) -> DocumentConvertersType:
if self.config.name == "default" or None:
return {
TextDocumentWithLabeledSpansAndBinaryRelations: Pipeline(
**get_common_pipeline_steps(conversion_method=self.config.conversion_method)
),
TextDocumentWithLabeledSpansBinaryRelationsAndLabeledPartitions: Pipeline(
**get_common_pipeline_steps(conversion_method=self.config.conversion_method),
cast=Caster(
document_type=TextDocumentWithLabeledSpansBinaryRelationsAndLabeledPartitions
),
add_partitions=RegexPartitioner(
partition_layer_name="labeled_partitions",
default_partition_label="paragraph",
pattern="\n",
strip_whitespace=True,
verbose=False,
),
),
}
else:
raise ValueError(f"Unknown dataset variant: {self.config.name}")