abhi1nandy2
commited on
Commit
•
f5c692b
1
Parent(s):
2c80650
Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Refer to https://aclanthology.org/2021.findings-emnlp.392/ for the paper and https://sites.google.com/view/emanualqa/home for the project website
|
2 |
+
|
3 |
+
## Citation
|
4 |
+
|
5 |
+
Please cite the work if you would like to use it.
|
6 |
+
|
7 |
+
```
|
8 |
+
@inproceedings{nandy-etal-2021-question-answering,
|
9 |
+
title = "Question Answering over Electronic Devices: A New Benchmark Dataset and a Multi-Task Learning based {QA} Framework",
|
10 |
+
author = "Nandy, Abhilash and
|
11 |
+
Sharma, Soumya and
|
12 |
+
Maddhashiya, Shubham and
|
13 |
+
Sachdeva, Kapil and
|
14 |
+
Goyal, Pawan and
|
15 |
+
Ganguly, NIloy",
|
16 |
+
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2021",
|
17 |
+
month = nov,
|
18 |
+
year = "2021",
|
19 |
+
address = "Punta Cana, Dominican Republic",
|
20 |
+
publisher = "Association for Computational Linguistics",
|
21 |
+
url = "https://aclanthology.org/2021.findings-emnlp.392",
|
22 |
+
doi = "10.18653/v1/2021.findings-emnlp.392",
|
23 |
+
pages = "4600--4609",
|
24 |
+
abstract = "Answering questions asked from instructional corpora such as E-manuals, recipe books, etc., has been far less studied than open-domain factoid context-based question answering. This can be primarily attributed to the absence of standard benchmark datasets. In this paper, we meticulously create a large amount of data connected with E-manuals and develop a suitable algorithm to exploit it. We collect E-Manual Corpus, a huge corpus of 307,957 E-manuals, and pretrain RoBERTa on this large corpus. We create various benchmark QA datasets which include question answer pairs curated by experts based upon two E-manuals, real user questions from Community Question Answering Forum pertaining to E-manuals etc. We introduce EMQAP (E-Manual Question Answering Pipeline) that answers questions pertaining to electronics devices. Built upon the pretrained RoBERTa, it harbors a supervised multi-task learning framework which efficiently performs the dual tasks of identifying the section in the E-manual where the answer can be found and the exact answer span within that section. For E-Manual annotated question-answer pairs, we show an improvement of about 40{\%} in ROUGE-L F1 scores over most competitive baseline. We perform a detailed ablation study and establish the versatility of EMQAP across different circumstances. The code and datasets are shared at https://github.com/abhi1nandy2/EMNLP-2021-Findings, and the corresponding project website is https://sites.google.com/view/emanualqa/home.",
|
25 |
+
}
|
26 |
+
```
|