Update README.md
#19
by
ninon64
- opened
README.md
CHANGED
@@ -32,7 +32,7 @@ with three objectives:
|
|
32 |
- Cosine embedding loss: the model was also trained to generate hidden states as close as possible as the BERT base
|
33 |
model.
|
34 |
|
35 |
-
This way, the model learns the same inner representation of the English language
|
36 |
faster for inference or downstream tasks.
|
37 |
|
38 |
## Intended uses & limitations
|
|
|
32 |
- Cosine embedding loss: the model was also trained to generate hidden states as close as possible as the BERT base
|
33 |
model.
|
34 |
|
35 |
+
This way, the model learns the same inner representation of the English language as its teacher model, while being
|
36 |
faster for inference or downstream tasks.
|
37 |
|
38 |
## Intended uses & limitations
|