nbroad HF staff autoevaluator HF staff commited on
Commit
89db1c3
1 Parent(s): 0881747

Add evaluation results on the squad_v2 config of squad_v2 (#1)

Browse files

- Add evaluation results on the squad_v2 config of squad_v2 (afaec4df4fb9b4cef1694dc2961e45dccb216da2)


Co-authored-by: Evaluation Bot <[email protected]>

Files changed (1) hide show
  1. README.md +44 -2
README.md CHANGED
@@ -1,9 +1,31 @@
1
  ---
2
  widget:
3
- - context: While deep and large pre-trained models are the state-of-the-art for various natural language processing tasks, their huge size poses significant challenges for practical uses in resource constrained settings. Recent works in knowledge distillation propose task-agnostic as well as task-specific methods to compress these models, with task-specific ones often yielding higher compression rate. In this work, we develop a new task-agnostic distillation framework XtremeDistilTransformers that leverages the advantage of task-specific methods for learning a small universal model that can be applied to arbitrary tasks and languages. To this end, we study the transferability of several source tasks, augmentation resources and model architecture for distillation. We evaluate our model performance on multiple tasks, including the General Language Understanding Evaluation (GLUE) benchmark, SQuAD question answering dataset and a massive multi-lingual NER dataset with 41 languages.
 
 
 
 
 
 
 
 
 
 
 
4
  example_title: xtremedistil q1
5
  text: What is XtremeDistil?
6
- - context: While deep and large pre-trained models are the state-of-the-art for various natural language processing tasks, their huge size poses significant challenges for practical uses in resource constrained settings. Recent works in knowledge distillation propose task-agnostic as well as task-specific methods to compress these models, with task-specific ones often yielding higher compression rate. In this work, we develop a new task-agnostic distillation framework XtremeDistilTransformers that leverages the advantage of task-specific methods for learning a small universal model that can be applied to arbitrary tasks and languages. To this end, we study the transferability of several source tasks, augmentation resources and model architecture for distillation. We evaluate our model performance on multiple tasks, including the General Language Understanding Evaluation (GLUE) benchmark, SQuAD question answering dataset and a massive multi-lingual NER dataset with 41 languages.
 
 
 
 
 
 
 
 
 
 
 
7
  example_title: xtremedistil q2
8
  text: On what is the model validated?
9
  datasets:
@@ -13,6 +35,26 @@ metrics:
13
  - exact
14
  tags:
15
  - question-answering
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  ---
17
 
18
  xtremedistil-l12-h384 trained on SQuAD 2.0
 
1
  ---
2
  widget:
3
+ - context: While deep and large pre-trained models are the state-of-the-art for various
4
+ natural language processing tasks, their huge size poses significant challenges
5
+ for practical uses in resource constrained settings. Recent works in knowledge
6
+ distillation propose task-agnostic as well as task-specific methods to compress
7
+ these models, with task-specific ones often yielding higher compression rate.
8
+ In this work, we develop a new task-agnostic distillation framework XtremeDistilTransformers
9
+ that leverages the advantage of task-specific methods for learning a small universal
10
+ model that can be applied to arbitrary tasks and languages. To this end, we study
11
+ the transferability of several source tasks, augmentation resources and model
12
+ architecture for distillation. We evaluate our model performance on multiple tasks,
13
+ including the General Language Understanding Evaluation (GLUE) benchmark, SQuAD
14
+ question answering dataset and a massive multi-lingual NER dataset with 41 languages.
15
  example_title: xtremedistil q1
16
  text: What is XtremeDistil?
17
+ - context: While deep and large pre-trained models are the state-of-the-art for various
18
+ natural language processing tasks, their huge size poses significant challenges
19
+ for practical uses in resource constrained settings. Recent works in knowledge
20
+ distillation propose task-agnostic as well as task-specific methods to compress
21
+ these models, with task-specific ones often yielding higher compression rate.
22
+ In this work, we develop a new task-agnostic distillation framework XtremeDistilTransformers
23
+ that leverages the advantage of task-specific methods for learning a small universal
24
+ model that can be applied to arbitrary tasks and languages. To this end, we study
25
+ the transferability of several source tasks, augmentation resources and model
26
+ architecture for distillation. We evaluate our model performance on multiple tasks,
27
+ including the General Language Understanding Evaluation (GLUE) benchmark, SQuAD
28
+ question answering dataset and a massive multi-lingual NER dataset with 41 languages.
29
  example_title: xtremedistil q2
30
  text: On what is the model validated?
31
  datasets:
 
35
  - exact
36
  tags:
37
  - question-answering
38
+ model-index:
39
+ - name: nbroad/xdistil-l12-h384-squad2
40
+ results:
41
+ - task:
42
+ type: question-answering
43
+ name: Question Answering
44
+ dataset:
45
+ name: squad_v2
46
+ type: squad_v2
47
+ config: squad_v2
48
+ split: validation
49
+ metrics:
50
+ - name: Exact Match
51
+ type: exact_match
52
+ value: 75.4591
53
+ verified: true
54
+ - name: F1
55
+ type: f1
56
+ value: 79.3321
57
+ verified: true
58
  ---
59
 
60
  xtremedistil-l12-h384 trained on SQuAD 2.0