Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,54 @@
|
|
1 |
---
|
|
|
|
|
2 |
license: cc-by-4.0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
language:
|
3 |
+
- es
|
4 |
license: cc-by-4.0
|
5 |
+
tags:
|
6 |
+
- anglicisms # Example: audio
|
7 |
+
- loanwords # Example: automatic-speech-recognition
|
8 |
+
- borrowing # Example: speech
|
9 |
+
- codeswitching # Example to specify a library: allennlp
|
10 |
+
- flair
|
11 |
+
- token-classification
|
12 |
+
- sequence-tagger-model
|
13 |
+
datasets:
|
14 |
+
- coalas # Example: common_voice. Use dataset id from https://hf.co/datasets
|
15 |
+
widget:
|
16 |
+
- text: "Las fake news sobre la celebrity se reprodujeron por los 'mass media' en prime time."
|
17 |
+
- text: "Me gusta el cine noir y el anime."
|
18 |
+
- text: "Benching, estar en el banquillo de tu 'crush' mientras otro juega de titular."
|
19 |
+
- text: "Recetas de noviembre para el batch cooking."
|
20 |
+
- text: "Utilizaron técnicas de machine learning, big data o blockchain."
|
21 |
---
|
22 |
+
|
23 |
+
# anglicisms-spanish-flair-cs
|
24 |
+
|
25 |
+
This is a pretrained model for detecting unassimilated English lexical borrowings (a.k.a. anglicisms) on Spanish newswire. This model labels words of foreign origin (fundamentally from English) used in Spanish language, words such as *fake news*, *machine learning*, *smartwatch*, *influencer* or *streaming*.
|
26 |
+
|
27 |
+
The model is a BiLSTM-CRF model fed with [Transformer-based embeddings pretrained on codeswitched data](https://huggingface.co/sagorsarker/codeswitch-spaeng-lid-lince) along subword embeddings (BPE and character embeddings). The model was trained on the [COALAS](https://github.com/lirondos/coalas/) corpus for the task of detecting lexical borrowings.
|
28 |
+
|
29 |
+
The model considers two labels:
|
30 |
+
* ``ENG``: For English lexical borrowings (*smartphone*, *online*, *podcast*)
|
31 |
+
* ``OTHER``: For lexical borrowings from any other language (*boutique*, *anime*, *umami*)
|
32 |
+
|
33 |
+
The model uses BIO encoding to account for multitoken borrowings.
|
34 |
+
|
35 |
+
## Metrics (on the test set)
|
36 |
+
|
37 |
+
| LABEL | Precision | Recall | F1 |
|
38 |
+
|:-------:|:-----:|:-----:|:---------:|
|
39 |
+
| ALL | 90.14 | 81.79 | 85.76 |
|
40 |
+
| ENG | 90.16 | 84.34 | 87.16 |
|
41 |
+
| OTHER | 85.71 | 13.04 | 22.64 |
|
42 |
+
|
43 |
+
There is another [mBERT -based model](https://huggingface.co/lirondos/anglicisms-spanish-mbert) for this same task trained using the ``Transformers`` library. That model however produced worse results than this Flair-based model (F1 = 83.55).
|
44 |
+
|
45 |
+
## Dataset
|
46 |
+
This model was trained on [COALAS](https://github.com/lirondos/coalas/), a corpus of Spanish newswire annotated with unassimilated lexical borrowings. The corpus contains 370,000 tokens and includes various written media written in European Spanish. The test set was designed to be as difficult as possible: it covers sources and dates not seen in the training set, includes a high number of OOV words (92% of the borrowings in the test set are OOV) and is very borrowing-dense (20 borrowings per 1,000 tokens).
|
47 |
+
|
48 |
+
|Set | Tokens | ENG | OTHER | Unique |
|
49 |
+
|:-------:|:-----:|:-----:|:---------:|:---------:|
|
50 |
+
|Training |231,126 |1,493 | 28 |380 |
|
51 |
+
|Development |82,578 |306 |49 |316|
|
52 |
+
|Test |58,997 |1,239 |46 |987|
|
53 |
+
|Total |372,701 |3,038 |123 |1,683 |
|
54 |
+
|