Update README.md
Browse files
README.md
CHANGED
@@ -1,8 +1,13 @@
|
|
|
|
|
|
|
|
|
|
1 |
# ClinicalBERT
|
2 |
|
3 |
<!-- Provide a quick summary of what the model is/does. -->
|
4 |
|
5 |
This model card describes the ClinicalBERT model, which was trained on a large multicenter dataset with a large corpus of 1.2B words of diverse diseases we constructed.
|
|
|
6 |
|
7 |
## Pretraining Data
|
8 |
|
@@ -11,11 +16,10 @@ For more details, see here.
|
|
11 |
|
12 |
## Model Pretraining
|
13 |
|
14 |
-
|
15 |
### Pretraining Procedures
|
16 |
-
|
|
|
17 |
The training code can be found [here](https://www.github.com/xxx) and the model was trained on four A100 GPU.
|
18 |
-
Model parameters were initialized with xxx.
|
19 |
|
20 |
### Pretraining Hyperparameters
|
21 |
|
@@ -29,8 +33,8 @@ All other default parameters were used (xxx).
|
|
29 |
Load the model via the transformers library:
|
30 |
```python
|
31 |
from transformers import AutoTokenizer, AutoModel
|
32 |
-
tokenizer = AutoTokenizer.from_pretrained("
|
33 |
-
model = AutoModel.from_pretrained("
|
34 |
|
35 |
```
|
36 |
|
@@ -40,6 +44,4 @@ Refer to the paper xxx.
|
|
40 |
|
41 |
## Questions?
|
42 |
|
43 |
-
Post a Github issue on the xxx repo or email xxx with any questions.
|
44 |
-
|
45 |
-
|
|
|
1 |
+
---
|
2 |
+
tags:
|
3 |
+
- medical
|
4 |
+
---
|
5 |
# ClinicalBERT
|
6 |
|
7 |
<!-- Provide a quick summary of what the model is/does. -->
|
8 |
|
9 |
This model card describes the ClinicalBERT model, which was trained on a large multicenter dataset with a large corpus of 1.2B words of diverse diseases we constructed.
|
10 |
+
We then utilized a very large corpus of EHRs from 3,136,266 pediatric outpatient visits to fine tune the base language model.
|
11 |
|
12 |
## Pretraining Data
|
13 |
|
|
|
16 |
|
17 |
## Model Pretraining
|
18 |
|
|
|
19 |
### Pretraining Procedures
|
20 |
+
The ClinicalBERT was initialized from BERT. Then the training followed the principle of masked language model, in which given a piece of text, we randomly replace some tokens by MASKs,
|
21 |
+
special tokens for masking, and then require the model to predict the original tokens via contextual text.
|
22 |
The training code can be found [here](https://www.github.com/xxx) and the model was trained on four A100 GPU.
|
|
|
23 |
|
24 |
### Pretraining Hyperparameters
|
25 |
|
|
|
33 |
Load the model via the transformers library:
|
34 |
```python
|
35 |
from transformers import AutoTokenizer, AutoModel
|
36 |
+
tokenizer = AutoTokenizer.from_pretrained("medicalai/ClinicalBERT")
|
37 |
+
model = AutoModel.from_pretrained("medicalai/ClinicalBERT")
|
38 |
|
39 |
```
|
40 |
|
|
|
44 |
|
45 |
## Questions?
|
46 |
|
47 |
+
Post a Github issue on the xxx repo or email xxx with any questions.
|
|
|
|