Datasets:
Update README.md
Browse files
README.md
CHANGED
@@ -89,6 +89,8 @@ This dataset tests the capabilities of language models to correctly capture the
|
|
89 |
We used probabilitic soft logic to combine probabilistic statements expressed with WEP (WEP-Reasoning) and we also used the UNLI dataset (https://nlp.jhu.edu/unli/) to directly check whether models can detect the WEP matching human-annotated probabilities.
|
90 |
The dataset can be used as natural langauge inference data (context, premise, label) or multiple choice question answering (context,valid_hypothesis, invalid_hypothesis).
|
91 |
|
|
|
|
|
92 |
```bib
|
93 |
@article{sileo2022probing,
|
94 |
title={Probing neural language models for understanding of words of estimative probability},
|
@@ -97,4 +99,3 @@ The dataset can be used as natural langauge inference data (context, premise, la
|
|
97 |
year={2022}
|
98 |
}
|
99 |
```
|
100 |
-
Accepted at Starsem2023 (The 12th Joint Conference on Lexical and Computational Semantics)
|
|
|
89 |
We used probabilitic soft logic to combine probabilistic statements expressed with WEP (WEP-Reasoning) and we also used the UNLI dataset (https://nlp.jhu.edu/unli/) to directly check whether models can detect the WEP matching human-annotated probabilities.
|
90 |
The dataset can be used as natural langauge inference data (context, premise, label) or multiple choice question answering (context,valid_hypothesis, invalid_hypothesis).
|
91 |
|
92 |
+
|
93 |
+
Accepted at Starsem2023 (The 12th Joint Conference on Lexical and Computational Semantics). Temporary citation:
|
94 |
```bib
|
95 |
@article{sileo2022probing,
|
96 |
title={Probing neural language models for understanding of words of estimative probability},
|
|
|
99 |
year={2022}
|
100 |
}
|
101 |
```
|
|