kimpty commited on
Commit
cb302af
1 Parent(s): 0f6826f

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +45 -0
README.md ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # ClinicalBERT
2
+
3
+ <!-- Provide a quick summary of what the model is/does. -->
4
+
5
+ This model card describes the ClinicalBERT model, which was trained on a large multicenter dataset with a large corpus of 1.2B words of diverse diseases we constructed.
6
+
7
+ ## Pretraining Data
8
+
9
+ The ClinicalBERT model was trained on a large multicenter dataset with a large corpus of 1.2B words of diverse diseases we constructed.
10
+ For more details, see here.
11
+
12
+ ## Model Pretraining
13
+
14
+
15
+ ### Pretraining Procedures
16
+
17
+ The training code can be found [here](https://www.github.com/xxx) and the model was trained on four A100 GPU.
18
+ Model parameters were initialized with xxx.
19
+
20
+ ### Pretraining Hyperparameters
21
+
22
+ We used a batch size of xx, a maximum sequence length of xx, and a learning rate of xx for pre-training our models.
23
+ The model was trained for xx steps.
24
+ The dup factor for duplicating input data with different masks was set to 5.
25
+ All other default parameters were used (xxx).
26
+
27
+ ## How to use the model
28
+
29
+ Load the model via the transformers library:
30
+ ```python
31
+ from transformers import AutoTokenizer, AutoModel
32
+ tokenizer = AutoTokenizer.from_pretrained("kimpty/ClinicalBERT")
33
+ model = AutoModel.from_pretrained("kimpty/ClinicalBERT")
34
+
35
+ ```
36
+
37
+ ## More Information
38
+
39
+ Refer to the paper xxx.
40
+
41
+ ## Questions?
42
+
43
+ Post a Github issue on the xxx repo or email xxx with any questions.
44
+
45
+