rebeccaqian
commited on
Commit
•
582c272
1
Parent(s):
69c7778
Update README.md
Browse files
README.md
CHANGED
@@ -24,34 +24,7 @@ metrics:
|
|
24 |
- bleu
|
25 |
---
|
26 |
|
27 |
-
#
|
28 |
-
|
29 |
-
## Table of Contents
|
30 |
-
- [Table of Contents](#table-of-contents)
|
31 |
-
- [Model Description](#model-description)
|
32 |
-
- [Model Summary](#dataset-summary)
|
33 |
-
- [Languages](#languages)
|
34 |
-
- [Dataset Structure](#dataset-structure)
|
35 |
-
- [Data Instances](#data-instances)
|
36 |
-
- [Data Fields](#data-fields)
|
37 |
-
- [Data Splits](#data-splits)
|
38 |
-
- [Dataset Creation](#dataset-creation)
|
39 |
-
- [Curation Rationale](#curation-rationale)
|
40 |
-
- [Source Data](#source-data)
|
41 |
-
- [Annotations](#annotations)
|
42 |
-
- [Personal and Sensitive Information](#personal-and-sensitive-information)
|
43 |
-
- [Considerations for Using the Data](#considerations-for-using-the-data)
|
44 |
-
- [Social Impact of Dataset](#social-impact-of-dataset)
|
45 |
-
- [Discussion of Biases](#discussion-of-biases)
|
46 |
-
- [Other Known Limitations](#other-known-limitations)
|
47 |
-
- [Additional Information](#additional-information)
|
48 |
-
- [Dataset Curators](#dataset-curators)
|
49 |
-
- [Licensing Information](#licensing-information)
|
50 |
-
- [Citation Information](#citation-information)
|
51 |
-
- [Contributions](#contributions)
|
52 |
-
|
53 |
-
## Model Description
|
54 |
-
|
55 |
The perturber is a seq2seq controlled generation model that rewrites text along a specified demographic axis and attribute.
|
56 |
|
57 |
The perturber takes in (i) a source text snippet, (ii) a word in the snippet referring to a demographic group, and (iii) a new target demographic attribute, and generates a perturbed snippet that refers to the target demographic attribute, while preserving overall meaning.
|
@@ -61,14 +34,12 @@ The perturber takes in (i) a source text snippet, (ii) a word in the snippet ref
|
|
61 |
- **Point of Contact:** rebeccaqian@meta.com, ccross@meta.com, douwe@huggingface.co, adinawilliams@meta.com
|
62 |
- **License:** MIT
|
63 |
|
64 |
-
|
65 |
-
|
66 |
The perturber is a finetuned BART model (Lewis et al., 2020) with 24 layers, 1024 hidden size, 406M parameters, and 16 attention heads. To train the perturber in the original paper, we finetune BART on PANDA using the ParlAI library.
|
67 |
|
68 |
This model release is separately trained using the HuggingFace transformers library, with the same parameters as the ParlAI model.
|
69 |
|
70 |
### Uses
|
71 |
-
|
72 |
The perturber is intended for use by fairness researchers and engineers working on demographic debiasing applications. The perturber is a controllable generation model that given a word, target demographic attribute and input text, outputs text where the selected word and associated references are rewritten to the target demographic attribute. Control variables and the input text are separated by a <PERT_SEP> token.
|
73 |
|
74 |
## Examples
|
@@ -145,5 +116,4 @@ For an in-depth discussion of bias, risks and limitations, see the Limitations s
|
|
145 |
```
|
146 |
|
147 |
### Model Card Contact
|
148 |
-
|
149 |
Thanks to [@Rebecca-Qian](https://github.com/Rebecca-Qian) for adding this model.
|
|
|
24 |
- bleu
|
25 |
---
|
26 |
|
27 |
+
# The Perturber
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
28 |
The perturber is a seq2seq controlled generation model that rewrites text along a specified demographic axis and attribute.
|
29 |
|
30 |
The perturber takes in (i) a source text snippet, (ii) a word in the snippet referring to a demographic group, and (iii) a new target demographic attribute, and generates a perturbed snippet that refers to the target demographic attribute, while preserving overall meaning.
|
|
|
34 |
- **Point of Contact:** rebeccaqian@meta.com, ccross@meta.com, douwe@huggingface.co, adinawilliams@meta.com
|
35 |
- **License:** MIT
|
36 |
|
37 |
+
## Model Description
|
|
|
38 |
The perturber is a finetuned BART model (Lewis et al., 2020) with 24 layers, 1024 hidden size, 406M parameters, and 16 attention heads. To train the perturber in the original paper, we finetune BART on PANDA using the ParlAI library.
|
39 |
|
40 |
This model release is separately trained using the HuggingFace transformers library, with the same parameters as the ParlAI model.
|
41 |
|
42 |
### Uses
|
|
|
43 |
The perturber is intended for use by fairness researchers and engineers working on demographic debiasing applications. The perturber is a controllable generation model that given a word, target demographic attribute and input text, outputs text where the selected word and associated references are rewritten to the target demographic attribute. Control variables and the input text are separated by a <PERT_SEP> token.
|
44 |
|
45 |
## Examples
|
|
|
116 |
```
|
117 |
|
118 |
### Model Card Contact
|
|
|
119 |
Thanks to [@Rebecca-Qian](https://github.com/Rebecca-Qian) for adding this model.
|