Transformers
PyTorch
English
bert
pretraining
Inference Endpoints
baskra commited on
Commit
39b6a9b
1 Parent(s): e687f70

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +28 -0
README.md CHANGED
@@ -1,3 +1,31 @@
1
  ---
 
2
  license: apache-2.0
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language: en
3
  license: apache-2.0
4
+ datasets:
5
+ - wikipedia
6
  ---
7
+
8
+ # BERT Large Uncased (CDA) - Counterfactual Data Augmentation
9
+
10
+ Pretrained model on English language using a masked language modeling (MLM) objective. It was introduced
11
+ in [this paper](https://arxiv.org/abs/1810.04805) and first released
12
+ in [this repository](https://github.com/google-research-datasets/Zari). The model is pre-trained from scratch over
13
+ Wikipedia. Word substitutions for data augmentation are determined using the word lists provided
14
+ at [corefBias](https://github.com/uclanlp/corefBias) ([Zhao et al. (2018)](https://arxiv.org/abs/1804.06876)).
15
+
16
+ Disclaimer: The team releasing BERT did not write a model card for this model so this model card has been written by
17
+ the FairNLP team.
18
+
19
+
20
+ ### BibTeX entry and citation info
21
+
22
+ ```
23
+ @misc{zari,
24
+ title={Measuring and Reducing Gendered Correlations in Pre-trained Models},
25
+ author={Kellie Webster and Xuezhi Wang and Ian Tenney and Alex Beutel and Emily Pitler and Ellie Pavlick and Jilin Chen and Slav Petrov},
26
+ year={2020},
27
+ eprint={2010.06032},
28
+ archivePrefix={arXiv},
29
+ primaryClass={cs.CL}
30
+ }
31
+ ```