mrm8488 commited on
Commit
ffaaea3
1 Parent(s): 704a03e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -3
README.md CHANGED
@@ -1,13 +1,18 @@
1
  ---
2
  language: es
3
  thumbnail: https://i.imgur.com/uxAvBfh.png
 
 
 
4
 
 
 
5
 
6
  ---
7
 
8
  ## ELECTRICIDAD: The Spanish Electra [Imgur](https://imgur.com/uxAvBfh)
9
 
10
- **ELECTRICIDAD** is a small Electra like model (discriminator in this case) trained on a + 20 GB of the [OSCAR](https://oscar-corpus.com/) Spanish corpus.
11
 
12
  As mentioned in the original [paper](https://openreview.net/pdf?id=r1xMH1BtvB):
13
  **ELECTRA** is a new method for self-supervised language representation learning. It can be used to pre-train transformer networks using relatively little compute. ELECTRA models are trained to distinguish "real" input tokens vs "fake" input tokens generated by another neural network, similar to the discriminator of a [GAN](https://arxiv.org/pdf/1406.2661.pdf). At small scale, ELECTRA achieves strong results even when trained on a single GPU. At large scale, ELECTRA achieves state-of-the-art results on the [SQuAD 2.0](https://rajpurkar.github.io/SQuAD-explorer/) dataset.
@@ -18,8 +23,8 @@ For a detailed description and experimental results, please refer the paper [ELE
18
 
19
  |Param| # Value|
20
  |-----|--------|
21
- |Layers| 12 |
22
- |Hidden |256 |
23
  |Params| 14M|
24
 
25
  ## Evaluation metrics (for discriminator) 🧾
 
1
  ---
2
  language: es
3
  thumbnail: https://i.imgur.com/uxAvBfh.png
4
+ tags:
5
+ - Spanish
6
+ - Electra
7
 
8
+ datasets:
9
+ - large_spanish_corpus
10
 
11
  ---
12
 
13
  ## ELECTRICIDAD: The Spanish Electra [Imgur](https://imgur.com/uxAvBfh)
14
 
15
+ **ELECTRICIDAD** is a small Electra like model (discriminator in this case) trained on on a [Large Spanish Corpus](https://github.com/josecannete/spanish-corpora) (aka BETO's corpus).
16
 
17
  As mentioned in the original [paper](https://openreview.net/pdf?id=r1xMH1BtvB):
18
  **ELECTRA** is a new method for self-supervised language representation learning. It can be used to pre-train transformer networks using relatively little compute. ELECTRA models are trained to distinguish "real" input tokens vs "fake" input tokens generated by another neural network, similar to the discriminator of a [GAN](https://arxiv.org/pdf/1406.2661.pdf). At small scale, ELECTRA achieves strong results even when trained on a single GPU. At large scale, ELECTRA achieves state-of-the-art results on the [SQuAD 2.0](https://rajpurkar.github.io/SQuAD-explorer/) dataset.
 
23
 
24
  |Param| # Value|
25
  |-----|--------|
26
+ |Layers|\t12 |
27
+ |Hidden |256 \t|
28
  |Params| 14M|
29
 
30
  ## Evaluation metrics (for discriminator) 🧾