yarongef commited on
Commit
10d995a
1 Parent(s): 742defc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +55 -55
README.md CHANGED
@@ -1,56 +1,56 @@
1
- ---
2
- license: mit
3
- language: protein
4
- tags:
5
- - protein language model
6
- datasets:
7
- - Uniref50
8
- ---
9
-
10
- # DistilProtBert model
11
-
12
- Distilled protein language of [ProtBert](https://huggingface.co/Rostlab/prot_bert).
13
- In addition to cross entropy and cosine teacher-student losses, DistilProtBert was pretrained on a masked language modeling (MLM) objective and it only works with capital letter amino acids.
14
-
15
- # Model description
16
-
17
- DistilProtBert was pretrained on millions of proteins sequences.
18
-
19
- Few important differences between DistilProtBert model and the original ProtBert version are:
20
- 1. The size of the model
21
- 2. The size of the pretraining dataset
22
- 3. Time & hardware used for pretraining
23
-
24
- ## Intended uses & limitations
25
-
26
- The model could be used for protein feature extraction or to be fine-tuned on downstream tasks.
27
-
28
- ### How to use
29
-
30
- The model can be used the same as ProtBert.
31
-
32
- ## Training data
33
-
34
- DistilProtBert model was pretrained on [Uniref50](https://www.uniprot.org/downloads), a dataset consisting of ~43 million protein sequences (only sequences of length between 20 to 512 amino acids were used).
35
-
36
- # Pretraining procedure
37
-
38
- Preprocessing was done using ProtBert's tokenizer.
39
- The details of the masking procedure for each sequence followed the original Bert (as mentioned in [ProtBert](https://huggingface.co/Rostlab/prot_bert)).
40
-
41
- The model was pretrained on a single DGX cluster 3 epochs in total. local batch size was 16, the optimizer used was AdamW with a learning rate of 5e-5 and mixed precision settings.
42
-
43
- ## Evaluation results
44
-
45
- When fine-tuned on downstream tasks, this model achieves the following results:
46
-
47
- | Task/Dataset | secondary structure (3-states) | Membrane |
48
- |:-----:|:-----:|:-----:|
49
- | CASP12 | 72 | |
50
- | TS115 | 81 | |
51
- | CB513 | 79 | |
52
- | DeepLoc | | 86 |
53
-
54
- Distinguish between:
55
-
56
  ### BibTeX entry and citation info
 
1
+ ---
2
+ license: mit
3
+ language: protein
4
+ tags:
5
+ - protein language model
6
+ datasets:
7
+ - Uniref50
8
+ ---
9
+
10
+ # DistilProtBert model
11
+
12
+ Distilled version of [ProtBert](https://huggingface.co/Rostlab/prot_bert) model.
13
+ In addition to cross entropy and cosine teacher-student losses, DistilProtBert was pretrained on a masked language modeling (MLM) objective and it only works with capital letter amino acids.
14
+
15
+ # Model description
16
+
17
+ DistilProtBert was pretrained on millions of proteins sequences.
18
+
19
+ Few important differences between DistilProtBert model and the original ProtBert version are:
20
+ 1. The size of the model
21
+ 2. The size of the pretraining dataset
22
+ 3. Time & hardware used for pretraining
23
+
24
+ ## Intended uses & limitations
25
+
26
+ The model could be used for protein feature extraction or to be fine-tuned on downstream tasks.
27
+
28
+ ### How to use
29
+
30
+ The model can be used the same as ProtBert.
31
+
32
+ ## Training data
33
+
34
+ DistilProtBert model was pretrained on [Uniref50](https://www.uniprot.org/downloads), a dataset consisting of ~43 million protein sequences (only sequences of length between 20 to 512 amino acids were used).
35
+
36
+ # Pretraining procedure
37
+
38
+ Preprocessing was done using ProtBert's tokenizer.
39
+ The details of the masking procedure for each sequence followed the original Bert (as mentioned in [ProtBert](https://huggingface.co/Rostlab/prot_bert)).
40
+
41
+ The model was pretrained on a single DGX cluster 3 epochs in total. local batch size was 16, the optimizer used was AdamW with a learning rate of 5e-5 and mixed precision settings.
42
+
43
+ ## Evaluation results
44
+
45
+ When fine-tuned on downstream tasks, this model achieves the following results:
46
+
47
+ | Task/Dataset | secondary structure (3-states) | Membrane |
48
+ |:-----:|:-----:|:-----:|
49
+ | CASP12 | 72 | |
50
+ | TS115 | 81 | |
51
+ | CB513 | 79 | |
52
+ | DeepLoc | | 86 |
53
+
54
+ Distinguish between:
55
+
56
  ### BibTeX entry and citation info