Update README.md
Browse files
README.md
CHANGED
@@ -17,6 +17,9 @@ The model was trained on nearly 10 million hospital notes from the Amsterdam Uni
|
|
17 |
|
18 |
## Privacy
|
19 |
By anonymizing the training data we made sure the model did not learn any representative associations linked to names. Apart from the training data, the model's vocabulary was also anonymized. This ensures that the model can not predict any names in the generative fill-mask task.
|
|
|
|
|
|
|
20 |
|
21 |
## Authors
|
22 |
Stella Verkijk, Piek Vossen
|
|
|
17 |
|
18 |
## Privacy
|
19 |
By anonymizing the training data we made sure the model did not learn any representative associations linked to names. Apart from the training data, the model's vocabulary was also anonymized. This ensures that the model can not predict any names in the generative fill-mask task.
|
20 |
+
For more information on the model's anonymization see our publication:
|
21 |
+
|
22 |
+
Verkijk, S., & Vossen, P. (2022). Efficiently and thoroughly anonymizing a transformer language model for Dutch electronic health records: a two-step method. In Proceedings of the Thirteenth Language Resources and Evaluation Conference (pp. 1098-1103).
|
23 |
|
24 |
## Authors
|
25 |
Stella Verkijk, Piek Vossen
|