kimpty commited on
Commit
4ead442
·
1 Parent(s): cb302af

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -8
README.md CHANGED
@@ -1,8 +1,13 @@
 
 
 
 
1
  # ClinicalBERT
2
 
3
  <!-- Provide a quick summary of what the model is/does. -->
4
 
5
  This model card describes the ClinicalBERT model, which was trained on a large multicenter dataset with a large corpus of 1.2B words of diverse diseases we constructed.
 
6
 
7
  ## Pretraining Data
8
 
@@ -11,11 +16,10 @@ For more details, see here.
11
 
12
  ## Model Pretraining
13
 
14
-
15
  ### Pretraining Procedures
16
-
 
17
  The training code can be found [here](https://www.github.com/xxx) and the model was trained on four A100 GPU.
18
- Model parameters were initialized with xxx.
19
 
20
  ### Pretraining Hyperparameters
21
 
@@ -29,8 +33,8 @@ All other default parameters were used (xxx).
29
  Load the model via the transformers library:
30
  ```python
31
  from transformers import AutoTokenizer, AutoModel
32
- tokenizer = AutoTokenizer.from_pretrained("kimpty/ClinicalBERT")
33
- model = AutoModel.from_pretrained("kimpty/ClinicalBERT")
34
 
35
  ```
36
 
@@ -40,6 +44,4 @@ Refer to the paper xxx.
40
 
41
  ## Questions?
42
 
43
- Post a Github issue on the xxx repo or email xxx with any questions.
44
-
45
-
 
1
+ ---
2
+ tags:
3
+ - medical
4
+ ---
5
  # ClinicalBERT
6
 
7
  <!-- Provide a quick summary of what the model is/does. -->
8
 
9
  This model card describes the ClinicalBERT model, which was trained on a large multicenter dataset with a large corpus of 1.2B words of diverse diseases we constructed.
10
+ We then utilized a very large corpus of EHRs from 3,136,266 pediatric outpatient visits to fine tune the base language model.
11
 
12
  ## Pretraining Data
13
 
 
16
 
17
  ## Model Pretraining
18
 
 
19
  ### Pretraining Procedures
20
+ The ClinicalBERT was initialized from BERT. Then the training followed the principle of masked language model, in which given a piece of text, we randomly replace some tokens by MASKs,
21
+ special tokens for masking, and then require the model to predict the original tokens via contextual text.
22
  The training code can be found [here](https://www.github.com/xxx) and the model was trained on four A100 GPU.
 
23
 
24
  ### Pretraining Hyperparameters
25
 
 
33
  Load the model via the transformers library:
34
  ```python
35
  from transformers import AutoTokenizer, AutoModel
36
+ tokenizer = AutoTokenizer.from_pretrained("medicalai/ClinicalBERT")
37
+ model = AutoModel.from_pretrained("medicalai/ClinicalBERT")
38
 
39
  ```
40
 
 
44
 
45
  ## Questions?
46
 
47
+ Post a Github issue on the xxx repo or email xxx with any questions.