Marissa commited on
Commit
7eda95c
1 Parent(s): 5fd13b8

Update model card

Browse files

This PR has a preliminary model card, open to any feedback! cc

@Ezi



@Meg



@Nazneen

Files changed (1) hide show
  1. README.md +177 -2
README.md CHANGED
@@ -12,5 +12,180 @@ license: apache-2.0
12
 
13
  # DistilBERT base uncased distilled SQuAD
14
 
15
- This model is a fine-tune checkpoint of [DistilBERT-base-uncased](https://huggingface.co/distilbert-base-uncased), fine-tuned using (a second step of) knowledge distillation on SQuAD v1.1.
16
- This model reaches a F1 score of 86.9 on the dev set (for comparison, Bert bert-base-uncased version reaches a F1 score of 88.5).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12
 
13
  # DistilBERT base uncased distilled SQuAD
14
 
15
+ ## Table of Contents
16
+ - [Model Details](#model-details)
17
+ - [How To Get Started With the Model](#how-to-get-started-with-the-model)
18
+ - [Uses](#uses)
19
+ - [Risks, Limitations and Biases](#risks-limitations-and-biases)
20
+ - [Training](#training)
21
+ - [Evaluation](#evaluation)
22
+ - [Environmental Impact](#environmental-impact)
23
+ - [Technical Specifications](#technical-specifications)
24
+ - [Citation Information](#citation-information)
25
+ - [Model Card Authors](#model-card-authors)
26
+
27
+ ## Model Details
28
+
29
+ **Model Description:** The DistilBERT model was proposed in the blog post [Smaller, faster, cheaper, lighter: Introducing DistilBERT, adistilled version of BERT](https://medium.com/huggingface/distilbert-8cf3380435b5), and the paper [DistilBERT, adistilled version of BERT: smaller, faster, cheaper and lighter](https://arxiv.org/abs/1910.01108). DistilBERT is a small, fast, cheap and light Transformer model trained by distilling BERT base. It has 40% less parameters than *bert-base-uncased*, runs 60% faster while preserving over 95% of BERT's performances as measured on the GLUE language understanding benchmark.
30
+
31
+ This model is a fine-tune checkpoint of [DistilBERT-base-uncased](https://huggingface.co/distilbert-base-uncased), fine-tuned using (a second step of) knowledge distillation on [SQuAD v1.1](https://huggingface.co/datasets/squad).
32
+
33
+ - **Developed by:** Hugging Face
34
+ - **Model Type:** Transformer-based language model
35
+ - **Language(s):** English
36
+ - **License:** Apache 2.0
37
+ - **Related Models:** [DistilBERT-base-uncased](https://huggingface.co/distilbert-base-uncased)
38
+ - **Resources for more information:**
39
+ - See [this repository](https://github.com/huggingface/transformers/tree/main/examples/research_projects/distillation) for more about Distil\* (a class of compressed models including this model)
40
+ - See [Sanh et al. (2019)](https://arxiv.org/abs/1910.01108) for more information about knowledge distillation and the training procedure
41
+
42
+ ## How to Get Started with the Model
43
+
44
+ Use the code below to get started with the model.
45
+
46
+ ```python
47
+ >>> from transformers import pipeline
48
+ >>> question_answerer = pipeline("question-answering", model='distilbert-base-uncased-distilled-squad')
49
+
50
+ >>> context = r"""
51
+ ... Extractive Question Answering is the task of extracting an answer from a text given a question. An example of a
52
+ ... question answering dataset is the SQuAD dataset, which is entirely based on that task. If you would like to fine-tune
53
+ ... a model on a SQuAD task, you may leverage the examples/pytorch/question-answering/run_squad.py script.
54
+ ... """
55
+
56
+ >>> result = question_answerer(question="What is a good example of a question answering dataset?", context=context)
57
+ >>> print(
58
+ ... f"Answer: '{result['answer']}', score: {round(result['score'], 4)}, start: {result['start']}, end: {result['end']}"
59
+ ...)
60
+
61
+ Answer: 'SQuAD dataset', score: 0.4704, start: 147, end: 160
62
+ ```
63
+
64
+ Here is how to use this model in PyTorch:
65
+
66
+ ```python
67
+ from transformers import DistilBertTokenizer, DistilBertModel
68
+ import torch
69
+ tokenizer = DistilBertTokenizer.from_pretrained('distilbert-base-uncased-distilled-squad')
70
+ model = DistilBertModel.from_pretrained('distilbert-base-uncased-distilled-squad')
71
+
72
+ question, text = "Who was Jim Henson?", "Jim Henson was a nice puppet"
73
+
74
+ inputs = tokenizer(question, text, return_tensors="pt")
75
+ with torch.no_grad():
76
+ outputs = model(**inputs)
77
+
78
+ print(outputs)
79
+ ```
80
+
81
+ And in TensorFlow:
82
+
83
+ ```python
84
+ from transformers import DistilBertTokenizer, TFDistilBertForQuestionAnswering
85
+ import tensorflow as tf
86
+
87
+ tokenizer = DistilBertTokenizer.from_pretrained("distilbert-base-uncased-distilled-squad")
88
+ model = TFDistilBertForQuestionAnswering.from_pretrained("distilbert-base-uncased-distilled-squad")
89
+
90
+ question, text = "Who was Jim Henson?", "Jim Henson was a nice puppet"
91
+
92
+ inputs = tokenizer(question, text, return_tensors="tf")
93
+ outputs = model(**inputs)
94
+
95
+ answer_start_index = int(tf.math.argmax(outputs.start_logits, axis=-1)[0])
96
+ answer_end_index = int(tf.math.argmax(outputs.end_logits, axis=-1)[0])
97
+
98
+ predict_answer_tokens = inputs.input_ids[0, answer_start_index : answer_end_index + 1]
99
+ tokenizer.decode(predict_answer_tokens)
100
+ ```
101
+
102
+ ## Uses
103
+
104
+ This model can be used for question answering.
105
+
106
+ #### Misuse and Out-of-scope Use
107
+
108
+ The model should not be used to intentionally create hostile or alienating environments for people. In addition, the model was not trained to be factual or true representations of people or events, and therefore using the model to generate such content is out-of-scope for the abilities of this model.
109
+
110
+ ## Risks, Limitations and Biases
111
+
112
+ **CONTENT WARNING: Readers should be aware that language generated by this model can be disturbing or offensive to some and can propagate historical and current stereotypes.**
113
+
114
+ Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)). Predictions generated by the model can include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups. For example:
115
+
116
+
117
+ ```python
118
+ >>> from transformers import pipeline
119
+ >>> question_answerer = pipeline("question-answering", model='distilbert-base-uncased-distilled-squad')
120
+
121
+ >>> context = r"""
122
+ ... Alice is sitting on the bench. Bob is sitting next to her.
123
+ ... """
124
+
125
+ >>> result = question_answerer(question="Who is the CEO?", context=context)
126
+ >>> print(
127
+ ... f"Answer: '{result['answer']}', score: {round(result['score'], 4)}, start: {result['start']}, end: {result['end']}"
128
+ ...)
129
+
130
+ Answer: 'Bob', score: 0.4183, start: 32, end: 35
131
+ ```
132
+
133
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model.
134
+
135
+ ## Training
136
+
137
+ #### Training Data
138
+
139
+ The [distilbert-base-uncased model](https://huggingface.co/distilbert-base-uncased) model describes it's training data as:
140
+
141
+ > DistilBERT pretrained on the same data as BERT, which is [BookCorpus](https://yknzhu.wixsite.com/mbweb), a dataset consisting of 11,038 unpublished books and [English Wikipedia](https://en.wikipedia.org/wiki/English_Wikipedia) (excluding lists, tables and headers).
142
+
143
+ To learn more about the SQuAD v1.1 dataset, see the [SQuAD v1.1 data card](https://huggingface.co/datasets/squad).
144
+
145
+ #### Training Procedure
146
+
147
+ ##### Preprocessing
148
+
149
+ See the [distilbert-base-uncased model card](https://huggingface.co/distilbert-base-uncased) for further details.
150
+
151
+ ##### Pretraining
152
+
153
+ See the [distilbert-base-uncased model card](https://huggingface.co/distilbert-base-uncased) for further details.
154
+
155
+ ## Evaluation
156
+
157
+ As discussed in the [model repository](https://github.com/huggingface/transformers/blob/main/examples/research_projects/distillation/README.md)
158
+
159
+ > This model reaches a F1 score of 86.9 on the [SQuAD v1.1] dev set (for comparison, Bert bert-base-uncased version reaches a F1 score of 88.5).
160
+
161
+ ## Environmental Impact
162
+
163
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). We present the hardware type and hours used based on the [associated paper](https://arxiv.org/pdf/1910.01108.pdf). Note that these details are just for training DistilBERT, not including the fine-tuning with SQuAD.
164
+
165
+ - **Hardware Type:** 8 16GB V100 GPUs
166
+ - **Hours used:** 90 hours
167
+ - **Cloud Provider:** Unknown
168
+ - **Compute Region:** Unknown
169
+ - **Carbon Emitted:** Unknown
170
+
171
+ ## Technical Specifications
172
+
173
+ See the [associated paper](https://arxiv.org/abs/1910.01108) for details on the modeling architecture, objective, compute infrastructure, and training details.
174
+
175
+ ## Citation Information
176
+
177
+ ```bibtex
178
+ @inproceedings{sanh2019distilbert,
179
+ title={DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter},
180
+ author={Sanh, Victor and Debut, Lysandre and Chaumond, Julien and Wolf, Thomas},
181
+ booktitle={NeurIPS EMC^2 Workshop},
182
+ year={2019}
183
+ }
184
+ ```
185
+
186
+ APA:
187
+ - Sanh, V., Debut, L., Chaumond, J., & Wolf, T. (2019). DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108.
188
+
189
+ ## Model Card Authors
190
+
191
+ This model card was written by the Hugging Face team.