Commit
•
89db1c3
1
Parent(s):
0881747
Add evaluation results on the squad_v2 config of squad_v2 (#1)
Browse files- Add evaluation results on the squad_v2 config of squad_v2 (afaec4df4fb9b4cef1694dc2961e45dccb216da2)
Co-authored-by: Evaluation Bot <[email protected]>
README.md
CHANGED
@@ -1,9 +1,31 @@
|
|
1 |
---
|
2 |
widget:
|
3 |
-
- context: While deep and large pre-trained models are the state-of-the-art for various
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4 |
example_title: xtremedistil q1
|
5 |
text: What is XtremeDistil?
|
6 |
-
- context: While deep and large pre-trained models are the state-of-the-art for various
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
example_title: xtremedistil q2
|
8 |
text: On what is the model validated?
|
9 |
datasets:
|
@@ -13,6 +35,26 @@ metrics:
|
|
13 |
- exact
|
14 |
tags:
|
15 |
- question-answering
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
---
|
17 |
|
18 |
xtremedistil-l12-h384 trained on SQuAD 2.0
|
|
|
1 |
---
|
2 |
widget:
|
3 |
+
- context: While deep and large pre-trained models are the state-of-the-art for various
|
4 |
+
natural language processing tasks, their huge size poses significant challenges
|
5 |
+
for practical uses in resource constrained settings. Recent works in knowledge
|
6 |
+
distillation propose task-agnostic as well as task-specific methods to compress
|
7 |
+
these models, with task-specific ones often yielding higher compression rate.
|
8 |
+
In this work, we develop a new task-agnostic distillation framework XtremeDistilTransformers
|
9 |
+
that leverages the advantage of task-specific methods for learning a small universal
|
10 |
+
model that can be applied to arbitrary tasks and languages. To this end, we study
|
11 |
+
the transferability of several source tasks, augmentation resources and model
|
12 |
+
architecture for distillation. We evaluate our model performance on multiple tasks,
|
13 |
+
including the General Language Understanding Evaluation (GLUE) benchmark, SQuAD
|
14 |
+
question answering dataset and a massive multi-lingual NER dataset with 41 languages.
|
15 |
example_title: xtremedistil q1
|
16 |
text: What is XtremeDistil?
|
17 |
+
- context: While deep and large pre-trained models are the state-of-the-art for various
|
18 |
+
natural language processing tasks, their huge size poses significant challenges
|
19 |
+
for practical uses in resource constrained settings. Recent works in knowledge
|
20 |
+
distillation propose task-agnostic as well as task-specific methods to compress
|
21 |
+
these models, with task-specific ones often yielding higher compression rate.
|
22 |
+
In this work, we develop a new task-agnostic distillation framework XtremeDistilTransformers
|
23 |
+
that leverages the advantage of task-specific methods for learning a small universal
|
24 |
+
model that can be applied to arbitrary tasks and languages. To this end, we study
|
25 |
+
the transferability of several source tasks, augmentation resources and model
|
26 |
+
architecture for distillation. We evaluate our model performance on multiple tasks,
|
27 |
+
including the General Language Understanding Evaluation (GLUE) benchmark, SQuAD
|
28 |
+
question answering dataset and a massive multi-lingual NER dataset with 41 languages.
|
29 |
example_title: xtremedistil q2
|
30 |
text: On what is the model validated?
|
31 |
datasets:
|
|
|
35 |
- exact
|
36 |
tags:
|
37 |
- question-answering
|
38 |
+
model-index:
|
39 |
+
- name: nbroad/xdistil-l12-h384-squad2
|
40 |
+
results:
|
41 |
+
- task:
|
42 |
+
type: question-answering
|
43 |
+
name: Question Answering
|
44 |
+
dataset:
|
45 |
+
name: squad_v2
|
46 |
+
type: squad_v2
|
47 |
+
config: squad_v2
|
48 |
+
split: validation
|
49 |
+
metrics:
|
50 |
+
- name: Exact Match
|
51 |
+
type: exact_match
|
52 |
+
value: 75.4591
|
53 |
+
verified: true
|
54 |
+
- name: F1
|
55 |
+
type: f1
|
56 |
+
value: 79.3321
|
57 |
+
verified: true
|
58 |
---
|
59 |
|
60 |
xtremedistil-l12-h384 trained on SQuAD 2.0
|