mrm8488 commited on
Commit
322968f
1 Parent(s): b1e7fea

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -4
README.md CHANGED
@@ -1,13 +1,18 @@
1
  ---
2
  language: es
3
  thumbnail: https://i.imgur.com/uxAvBfh.png
 
 
 
4
 
 
 
5
 
6
  ---
7
 
8
  ## ELECTRICIDAD: The Spanish Electra [Imgur](https://imgur.com/uxAvBfh)
9
 
10
- **Electricidad-base-discriminator** (uncased) is a ```base``` Electra like model (discriminator in this case) trained on a + 20 GB of the [OSCAR](https://oscar-corpus.com/) Spanish corpus.
11
 
12
  As mentioned in the original [paper](https://openreview.net/pdf?id=r1xMH1BtvB):
13
  **ELECTRA** is a new method for self-supervised language representation learning. It can be used to pre-train transformer networks using relatively little compute. ELECTRA models are trained to distinguish "real" input tokens vs "fake" input tokens generated by another neural network, similar to the discriminator of a [GAN](https://arxiv.org/pdf/1406.2661.pdf). At small scale, ELECTRA achieves strong results even when trained on a single GPU. At large scale, ELECTRA achieves state-of-the-art results on the [SQuAD 2.0](https://rajpurkar.github.io/SQuAD-explorer/) dataset.
@@ -19,8 +24,8 @@ For a detailed description and experimental results, please refer the paper [ELE
19
 
20
  |Name| # Value|
21
  |-----|--------|
22
- |Layers| 12 |
23
- |Hidden |768 |
24
  |Params| 110M|
25
 
26
  ## Evaluation metrics (for discriminator) 🧾
@@ -70,7 +75,18 @@ As you can see there are **1s** in the places where the model detected a fake to
70
 
71
  [NER](https://huggingface.co/mrm8488/electricidad-base-finetuned-ner)
72
 
73
- [Paraphrase Identification](https://huggingface.co/mrm8488/RuPERTa-base-finetuned-pawsx-es)
 
 
 
 
 
 
 
 
 
 
 
74
 
75
 
76
 
 
1
  ---
2
  language: es
3
  thumbnail: https://i.imgur.com/uxAvBfh.png
4
+ tags:
5
+ - Spanish
6
+ - Electra
7
 
8
+ datasets:
9
+ -large_spanish_corpus
10
 
11
  ---
12
 
13
  ## ELECTRICIDAD: The Spanish Electra [Imgur](https://imgur.com/uxAvBfh)
14
 
15
+ **Electricidad-base-discriminator** (uncased) is a ```base``` Electra like model (discriminator in this case) trained on a [Large Spanish Corpus](https://github.com/josecannete/spanish-corpora) (aka BETO's corpus)
16
 
17
  As mentioned in the original [paper](https://openreview.net/pdf?id=r1xMH1BtvB):
18
  **ELECTRA** is a new method for self-supervised language representation learning. It can be used to pre-train transformer networks using relatively little compute. ELECTRA models are trained to distinguish "real" input tokens vs "fake" input tokens generated by another neural network, similar to the discriminator of a [GAN](https://arxiv.org/pdf/1406.2661.pdf). At small scale, ELECTRA achieves strong results even when trained on a single GPU. At large scale, ELECTRA achieves state-of-the-art results on the [SQuAD 2.0](https://rajpurkar.github.io/SQuAD-explorer/) dataset.
 
24
 
25
  |Name| # Value|
26
  |-----|--------|
27
+ |Layers|\t12 |
28
+ |Hidden |768 \t|
29
  |Params| 110M|
30
 
31
  ## Evaluation metrics (for discriminator) 🧾
 
75
 
76
  [NER](https://huggingface.co/mrm8488/electricidad-base-finetuned-ner)
77
 
78
+
79
+ ### Spanish LM model comparison 📊
80
+ | Dataset | Metric | RoBERTa-b | RoBERTa-l | BETO | mBERT | BERTIN | Electricidad-b |
81
+ |-------------|----------|-----------|-----------|--------|--------|--------|---------|
82
+ | UD-POS | F1 | 0.9907 | 0.9901 | 0.9900 | 0.9886 | 0.9904 | 0.9818 |
83
+ | Conll-NER | F1 | 0.8851 | 0.8772 | 0.8759 | 0.8691 | 0.8627 | 0.7954 |
84
+ | Capitel-POS | F1 | 0.9846 | 0.9851 | 0.9836 | 0.9839 | 0.9826 | 0.9816 |
85
+ | Capitel-NER | F1 | 0.8959 | 0.8998 | 0.8771 | 0.8810 | 0.8741 | 0.8035 |
86
+ | STS | Combined | 0.8423 | 0.8420 | 0.8216 | 0.8249 | 0.7822 | 0.8065 |
87
+ | MLDoc | Accuracy | 0.9595 | 0.9600 | 0.9650 | 0.9560 | 0.9673 | 0.9490 |
88
+ | PAWS-X | F1 | 0.9035 | 0.9000 | 0.8915 | 0.9020 | 0.8820 | **0.9045** |
89
+ | XNLI | Accuracy | 0.8016 | 0.7958 | 0.8130 | 0.7876 | 0.7864 | 0.7878 |
90
 
91
 
92