Spaces:
Runtime error
Runtime error
# RoBERTa: A Robustly Optimized BERT Pretraining Approach | |
https://arxiv.org/abs/1907.11692 | |
## Introduction | |
RoBERTa iterates on BERT's pretraining procedure, including training the model longer, with bigger batches over more data; removing the next sentence prediction objective; training on longer sequences; and dynamically changing the masking pattern applied to the training data. See the associated paper for more details. | |
### What's New: | |
- December 2020: German model (GottBERT) is available: [GottBERT](https://github.com/pytorch/fairseq/tree/main/examples/gottbert). | |
- January 2020: Italian model (UmBERTo) is available from Musixmatch Research: [UmBERTo](https://github.com/musixmatchresearch/umberto). | |
- November 2019: French model (CamemBERT) is available: [CamemBERT](https://github.com/pytorch/fairseq/tree/main/examples/camembert). | |
- November 2019: Multilingual encoder (XLM-RoBERTa) is available: [XLM-R](https://github.com/pytorch/fairseq/tree/main/examples/xlmr). | |
- September 2019: TensorFlow and TPU support via the [transformers library](https://github.com/huggingface/transformers). | |
- August 2019: RoBERTa is now supported in the [pytorch-transformers library](https://github.com/huggingface/pytorch-transformers). | |
- August 2019: Added [tutorial for finetuning on WinoGrande](https://github.com/pytorch/fairseq/tree/main/examples/roberta/wsc#roberta-training-on-winogrande-dataset). | |
- August 2019: Added [tutorial for pretraining RoBERTa using your own data](README.pretraining.md). | |
## Pre-trained models | |
Model | Description | # params | Download | |
---|---|---|--- | |
`roberta.base` | RoBERTa using the BERT-base architecture | 125M | [roberta.base.tar.gz](https://dl.fbaipublicfiles.com/fairseq/models/roberta.base.tar.gz) | |
`roberta.large` | RoBERTa using the BERT-large architecture | 355M | [roberta.large.tar.gz](https://dl.fbaipublicfiles.com/fairseq/models/roberta.large.tar.gz) | |
`roberta.large.mnli` | `roberta.large` finetuned on [MNLI](http://www.nyu.edu/projects/bowman/multinli) | 355M | [roberta.large.mnli.tar.gz](https://dl.fbaipublicfiles.com/fairseq/models/roberta.large.mnli.tar.gz) | |
`roberta.large.wsc` | `roberta.large` finetuned on [WSC](wsc/README.md) | 355M | [roberta.large.wsc.tar.gz](https://dl.fbaipublicfiles.com/fairseq/models/roberta.large.wsc.tar.gz) | |
## Results | |
**[GLUE (Wang et al., 2019)](https://gluebenchmark.com/)** | |
_(dev set, single model, single-task finetuning)_ | |
Model | MNLI | QNLI | QQP | RTE | SST-2 | MRPC | CoLA | STS-B | |
---|---|---|---|---|---|---|---|--- | |
`roberta.base` | 87.6 | 92.8 | 91.9 | 78.7 | 94.8 | 90.2 | 63.6 | 91.2 | |
`roberta.large` | 90.2 | 94.7 | 92.2 | 86.6 | 96.4 | 90.9 | 68.0 | 92.4 | |
`roberta.large.mnli` | 90.2 | - | - | - | - | - | - | - | |
**[SuperGLUE (Wang et al., 2019)](https://super.gluebenchmark.com/)** | |
_(dev set, single model, single-task finetuning)_ | |
Model | BoolQ | CB | COPA | MultiRC | RTE | WiC | WSC | |
---|---|---|---|---|---|---|--- | |
`roberta.large` | 86.9 | 98.2 | 94.0 | 85.7 | 89.5 | 75.6 | - | |
`roberta.large.wsc` | - | - | - | - | - | - | 91.3 | |
**[SQuAD (Rajpurkar et al., 2018)](https://rajpurkar.github.io/SQuAD-explorer/)** | |
_(dev set, no additional data used)_ | |
Model | SQuAD 1.1 EM/F1 | SQuAD 2.0 EM/F1 | |
---|---|--- | |
`roberta.large` | 88.9/94.6 | 86.5/89.4 | |
**[RACE (Lai et al., 2017)](http://www.qizhexie.com/data/RACE_leaderboard.html)** | |
_(test set)_ | |
Model | Accuracy | Middle | High | |
---|---|---|--- | |
`roberta.large` | 83.2 | 86.5 | 81.3 | |
**[HellaSwag (Zellers et al., 2019)](https://rowanzellers.com/hellaswag/)** | |
_(test set)_ | |
Model | Overall | In-domain | Zero-shot | ActivityNet | WikiHow | |
---|---|---|---|---|--- | |
`roberta.large` | 85.2 | 87.3 | 83.1 | 74.6 | 90.9 | |
**[Commonsense QA (Talmor et al., 2019)](https://www.tau-nlp.org/commonsenseqa)** | |
_(test set)_ | |
Model | Accuracy | |
---|--- | |
`roberta.large` (single model) | 72.1 | |
`roberta.large` (ensemble) | 72.5 | |
**[Winogrande (Sakaguchi et al., 2019)](https://arxiv.org/abs/1907.10641)** | |
_(test set)_ | |
Model | Accuracy | |
---|--- | |
`roberta.large` | 78.1 | |
**[XNLI (Conneau et al., 2018)](https://arxiv.org/abs/1809.05053)** | |
_(TRANSLATE-TEST)_ | |
Model | en | fr | es | de | el | bg | ru | tr | ar | vi | th | zh | hi | sw | ur | |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|--- | |
`roberta.large.mnli` | 91.3 | 82.91 | 84.27 | 81.24 | 81.74 | 83.13 | 78.28 | 76.79 | 76.64 | 74.17 | 74.05 | 77.5 | 70.9 | 66.65 | 66.81 | |
## Example usage | |
##### Load RoBERTa from torch.hub (PyTorch >= 1.1): | |
```python | |
import torch | |
roberta = torch.hub.load('pytorch/fairseq', 'roberta.large') | |
roberta.eval() # disable dropout (or leave in train mode to finetune) | |
``` | |
##### Load RoBERTa (for PyTorch 1.0 or custom models): | |
```python | |
# Download roberta.large model | |
wget https://dl.fbaipublicfiles.com/fairseq/models/roberta.large.tar.gz | |
tar -xzvf roberta.large.tar.gz | |
# Load the model in fairseq | |
from fairseq.models.roberta import RobertaModel | |
roberta = RobertaModel.from_pretrained('/path/to/roberta.large', checkpoint_file='model.pt') | |
roberta.eval() # disable dropout (or leave in train mode to finetune) | |
``` | |
##### Apply Byte-Pair Encoding (BPE) to input text: | |
```python | |
tokens = roberta.encode('Hello world!') | |
assert tokens.tolist() == [0, 31414, 232, 328, 2] | |
roberta.decode(tokens) # 'Hello world!' | |
``` | |
##### Extract features from RoBERTa: | |
```python | |
# Extract the last layer's features | |
last_layer_features = roberta.extract_features(tokens) | |
assert last_layer_features.size() == torch.Size([1, 5, 1024]) | |
# Extract all layer's features (layer 0 is the embedding layer) | |
all_layers = roberta.extract_features(tokens, return_all_hiddens=True) | |
assert len(all_layers) == 25 | |
assert torch.all(all_layers[-1] == last_layer_features) | |
``` | |
##### Use RoBERTa for sentence-pair classification tasks: | |
```python | |
# Download RoBERTa already finetuned for MNLI | |
roberta = torch.hub.load('pytorch/fairseq', 'roberta.large.mnli') | |
roberta.eval() # disable dropout for evaluation | |
# Encode a pair of sentences and make a prediction | |
tokens = roberta.encode('Roberta is a heavily optimized version of BERT.', 'Roberta is not very optimized.') | |
roberta.predict('mnli', tokens).argmax() # 0: contradiction | |
# Encode another pair of sentences | |
tokens = roberta.encode('Roberta is a heavily optimized version of BERT.', 'Roberta is based on BERT.') | |
roberta.predict('mnli', tokens).argmax() # 2: entailment | |
``` | |
##### Register a new (randomly initialized) classification head: | |
```python | |
roberta.register_classification_head('new_task', num_classes=3) | |
logprobs = roberta.predict('new_task', tokens) # tensor([[-1.1050, -1.0672, -1.1245]], grad_fn=<LogSoftmaxBackward>) | |
``` | |
##### Batched prediction: | |
```python | |
import torch | |
from fairseq.data.data_utils import collate_tokens | |
roberta = torch.hub.load('pytorch/fairseq', 'roberta.large.mnli') | |
roberta.eval() | |
batch_of_pairs = [ | |
['Roberta is a heavily optimized version of BERT.', 'Roberta is not very optimized.'], | |
['Roberta is a heavily optimized version of BERT.', 'Roberta is based on BERT.'], | |
['potatoes are awesome.', 'I like to run.'], | |
['Mars is very far from earth.', 'Mars is very close.'], | |
] | |
batch = collate_tokens( | |
[roberta.encode(pair[0], pair[1]) for pair in batch_of_pairs], pad_idx=1 | |
) | |
logprobs = roberta.predict('mnli', batch) | |
print(logprobs.argmax(dim=1)) | |
# tensor([0, 2, 1, 0]) | |
``` | |
##### Using the GPU: | |
```python | |
roberta.cuda() | |
roberta.predict('new_task', tokens) # tensor([[-1.1050, -1.0672, -1.1245]], device='cuda:0', grad_fn=<LogSoftmaxBackward>) | |
``` | |
## Advanced usage | |
#### Filling masks: | |
RoBERTa can be used to fill `<mask>` tokens in the input. Some examples from the | |
[Natural Questions dataset](https://ai.google.com/research/NaturalQuestions/): | |
```python | |
roberta.fill_mask('The first Star wars movie came out in <mask>', topk=3) | |
# [('The first Star wars movie came out in 1977', 0.9504708051681519, ' 1977'), ('The first Star wars movie came out in 1978', 0.009986862540245056, ' 1978'), ('The first Star wars movie came out in 1979', 0.009574787691235542, ' 1979')] | |
roberta.fill_mask('Vikram samvat calender is official in <mask>', topk=3) | |
# [('Vikram samvat calender is official in India', 0.21878819167613983, ' India'), ('Vikram samvat calender is official in Delhi', 0.08547237515449524, ' Delhi'), ('Vikram samvat calender is official in Gujarat', 0.07556215673685074, ' Gujarat')] | |
roberta.fill_mask('<mask> is the common currency of the European Union', topk=3) | |
# [('Euro is the common currency of the European Union', 0.9456493854522705, 'Euro'), ('euro is the common currency of the European Union', 0.025748178362846375, 'euro'), ('€ is the common currency of the European Union', 0.011183084920048714, '€')] | |
``` | |
#### Pronoun disambiguation (Winograd Schema Challenge): | |
RoBERTa can be used to disambiguate pronouns. First install spaCy and download the English-language model: | |
```bash | |
pip install spacy | |
python -m spacy download en_core_web_lg | |
``` | |
Next load the `roberta.large.wsc` model and call the `disambiguate_pronoun` | |
function. The pronoun should be surrounded by square brackets (`[]`) and the | |
query referent surrounded by underscores (`_`), or left blank to return the | |
predicted candidate text directly: | |
```python | |
roberta = torch.hub.load('pytorch/fairseq', 'roberta.large.wsc', user_dir='examples/roberta/wsc') | |
roberta.cuda() # use the GPU (optional) | |
roberta.disambiguate_pronoun('The _trophy_ would not fit in the brown suitcase because [it] was too big.') | |
# True | |
roberta.disambiguate_pronoun('The trophy would not fit in the brown _suitcase_ because [it] was too big.') | |
# False | |
roberta.disambiguate_pronoun('The city councilmen refused the demonstrators a permit because [they] feared violence.') | |
# 'The city councilmen' | |
roberta.disambiguate_pronoun('The city councilmen refused the demonstrators a permit because [they] advocated violence.') | |
# 'demonstrators' | |
``` | |
See the [RoBERTA Winograd Schema Challenge (WSC) README](wsc/README.md) for more details on how to train this model. | |
#### Extract features aligned to words: | |
By default RoBERTa outputs one feature vector per BPE token. You can instead | |
realign the features to match [spaCy's word-level tokenization](https://spacy.io/usage/linguistic-features#tokenization) | |
with the `extract_features_aligned_to_words` method. This will compute a | |
weighted average of the BPE-level features for each word and expose them in | |
spaCy's `Token.vector` attribute: | |
```python | |
doc = roberta.extract_features_aligned_to_words('I said, "hello RoBERTa."') | |
assert len(doc) == 10 | |
for tok in doc: | |
print('{:10}{} (...)'.format(str(tok), tok.vector[:5])) | |
# <s> tensor([-0.1316, -0.0386, -0.0832, -0.0477, 0.1943], grad_fn=<SliceBackward>) (...) | |
# I tensor([ 0.0559, 0.1541, -0.4832, 0.0880, 0.0120], grad_fn=<SliceBackward>) (...) | |
# said tensor([-0.1565, -0.0069, -0.8915, 0.0501, -0.0647], grad_fn=<SliceBackward>) (...) | |
# , tensor([-0.1318, -0.0387, -0.0834, -0.0477, 0.1944], grad_fn=<SliceBackward>) (...) | |
# " tensor([-0.0486, 0.1818, -0.3946, -0.0553, 0.0981], grad_fn=<SliceBackward>) (...) | |
# hello tensor([ 0.0079, 0.1799, -0.6204, -0.0777, -0.0923], grad_fn=<SliceBackward>) (...) | |
# RoBERTa tensor([-0.2339, -0.1184, -0.7343, -0.0492, 0.5829], grad_fn=<SliceBackward>) (...) | |
# . tensor([-0.1341, -0.1203, -0.1012, -0.0621, 0.1892], grad_fn=<SliceBackward>) (...) | |
# " tensor([-0.1341, -0.1203, -0.1012, -0.0621, 0.1892], grad_fn=<SliceBackward>) (...) | |
# </s> tensor([-0.0930, -0.0392, -0.0821, 0.0158, 0.0649], grad_fn=<SliceBackward>) (...) | |
``` | |
#### Evaluating the `roberta.large.mnli` model: | |
Example python code snippet to evaluate accuracy on the MNLI `dev_matched` set. | |
```python | |
label_map = {0: 'contradiction', 1: 'neutral', 2: 'entailment'} | |
ncorrect, nsamples = 0, 0 | |
roberta.cuda() | |
roberta.eval() | |
with open('glue_data/MNLI/dev_matched.tsv') as fin: | |
fin.readline() | |
for index, line in enumerate(fin): | |
tokens = line.strip().split('\t') | |
sent1, sent2, target = tokens[8], tokens[9], tokens[-1] | |
tokens = roberta.encode(sent1, sent2) | |
prediction = roberta.predict('mnli', tokens).argmax().item() | |
prediction_label = label_map[prediction] | |
ncorrect += int(prediction_label == target) | |
nsamples += 1 | |
print('| Accuracy: ', float(ncorrect)/float(nsamples)) | |
# Expected output: 0.9060 | |
``` | |
## Finetuning | |
- [Finetuning on GLUE](README.glue.md) | |
- [Finetuning on custom classification tasks (e.g., IMDB)](README.custom_classification.md) | |
- [Finetuning on Winograd Schema Challenge (WSC)](wsc/README.md) | |
- [Finetuning on Commonsense QA (CQA)](commonsense_qa/README.md) | |
## Pretraining using your own data | |
See the [tutorial for pretraining RoBERTa using your own data](README.pretraining.md). | |
## Citation | |
```bibtex | |
@article{liu2019roberta, | |
title = {RoBERTa: A Robustly Optimized BERT Pretraining Approach}, | |
author = {Yinhan Liu and Myle Ott and Naman Goyal and Jingfei Du and | |
Mandar Joshi and Danqi Chen and Omer Levy and Mike Lewis and | |
Luke Zettlemoyer and Veselin Stoyanov}, | |
journal={arXiv preprint arXiv:1907.11692}, | |
year = {2019}, | |
} | |
``` | |