princeton-nlp commited on
Commit
08e4470
1 Parent(s): dc3d269

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -1
README.md CHANGED
@@ -1,3 +1,8 @@
1
  ---
2
  inference: false
3
- ---
 
 
 
 
 
 
1
  ---
2
  inference: false
3
+ ---
4
+ This is a model checkpoint for ["Should You Mask 15% in Masked Language Modeling"](https://arxiv.org/abs/2202.08005) [(code)](https://github.com/princeton-nlp/DinkyTrain.git). We use pre layer norm, which is not supported by HuggingFace. To use our model, go to our [github repo](https://github.com/princeton-nlp/DinkyTrain.git), download our code, and import the RoBERTa class from `huggingface/modeling_roberta_prelayernorm.py`. For example,
5
+
6
+ ``` bash
7
+ from huggingface.modeling_roberta_prelayernorm import RobertaForMaskedLM, RobertaForSequenceClassification
8
+ ```