Pedrada commited on
Commit
f2f1202
โ€ข
1 Parent(s): 43e6b8e

Remove abstract from reference

Browse files
Files changed (1) hide show
  1. README.md +1 -2
README.md CHANGED
@@ -107,8 +107,7 @@ Output:
107
  address = "Marseille, France",
108
  publisher = "European Language Resources Association",
109
  url = "https://aclanthology.org/2022.lrec-1.27",
110
- pages = "258--266",
111
- abstract = "Language models are ubiquitous in current NLP, and their multilingual capacity has recently attracted considerable attention. However, current analyses have almost exclusively focused on (multilingual variants of) standard benchmarks, and have relied on clean pre-training and task-specific corpora as multilingual signals. In this paper, we introduce XLM-T, a model to train and evaluate multilingual language models in Twitter. In this paper we provide: (1) a new strong multilingual baseline consisting of an XLM-R (Conneau et al. 2020) model pre-trained on millions of tweets in over thirty languages, alongside starter code to subsequently fine-tune on a target task; and (2) a set of unified sentiment analysis Twitter datasets in eight different languages and a XLM-T model trained on this dataset.",
112
  }
113
 
114
  ```
 
107
  address = "Marseille, France",
108
  publisher = "European Language Resources Association",
109
  url = "https://aclanthology.org/2022.lrec-1.27",
110
+ pages = "258--266"
 
111
  }
112
 
113
  ```