Evan Komp commited on
Commit
42e5a73
1 Parent(s): e10138a

reset remote

Browse files
README.md DELETED
@@ -1,65 +0,0 @@
1
- ---
2
- tags:
3
- - generated_from_trainer
4
- metrics:
5
- - f1
6
- - accuracy
7
- model-index:
8
- - name: learn2therm
9
- results: []
10
- ---
11
-
12
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
- should probably proofread and complete it, then remove this comment. -->
14
-
15
- # learn2therm
16
-
17
- This model is a fine-tuned version of [Rostlab/prot_bert](https://huggingface.co/Rostlab/prot_bert) on the None dataset.
18
- It achieves the following results on the evaluation set:
19
- - Loss: 0.6942
20
- - F1: 0.0
21
- - Accuracy: 0.5125
22
- - Matthew: -0.0308
23
- - Cfm: [1025, 2, 973, 0]
24
-
25
- ## Model description
26
-
27
- More information needed
28
-
29
- ## Intended uses & limitations
30
-
31
- More information needed
32
-
33
- ## Training and evaluation data
34
-
35
- More information needed
36
-
37
- ## Training procedure
38
-
39
- ### Training hyperparameters
40
-
41
- The following hyperparameters were used during training:
42
- - learning_rate: 5e-05
43
- - train_batch_size: 32
44
- - eval_batch_size: 32
45
- - seed: 42
46
- - distributed_type: multi-GPU
47
- - num_devices: 2
48
- - gradient_accumulation_steps: 25
49
- - total_train_batch_size: 1600
50
- - total_eval_batch_size: 64
51
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
- - lr_scheduler_type: linear
53
- - num_epochs: 1
54
- - mixed_precision_training: Native AMP
55
-
56
- ### Training results
57
-
58
-
59
-
60
- ### Framework versions
61
-
62
- - Transformers 4.26.0
63
- - Pytorch 2.0.1
64
- - Datasets 2.12.0
65
- - Tokenizers 0.13.3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
config.json DELETED
@@ -1,26 +0,0 @@
1
- {
2
- "_name_or_path": "Rostlab/prot_bert",
3
- "architectures": [
4
- "BertForSequenceClassification"
5
- ],
6
- "attention_probs_dropout_prob": 0.1,
7
- "classifier_dropout": null,
8
- "hidden_act": "gelu",
9
- "hidden_dropout_prob": 0.1,
10
- "hidden_size": 1024,
11
- "initializer_range": 0.02,
12
- "intermediate_size": 4096,
13
- "layer_norm_eps": 1e-12,
14
- "max_position_embeddings": 40000,
15
- "model_type": "bert",
16
- "num_attention_heads": 16,
17
- "num_hidden_layers": 30,
18
- "pad_token_id": 0,
19
- "position_embedding_type": "absolute",
20
- "problem_type": "single_label_classification",
21
- "torch_dtype": "float32",
22
- "transformers_version": "4.26.0",
23
- "type_vocab_size": 2,
24
- "use_cache": true,
25
- "vocab_size": 30
26
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7f18ce42eb16a6d98d368c0fb0c9a39f338b2eecbdca21d5ab965794b25376e8
3
- size 1680220533
 
 
 
 
special_tokens_map.json DELETED
@@ -1,7 +0,0 @@
1
- {
2
- "cls_token": "[CLS]",
3
- "mask_token": "[MASK]",
4
- "pad_token": "[PAD]",
5
- "sep_token": "[SEP]",
6
- "unk_token": "[UNK]"
7
- }
 
 
 
 
 
 
 
 
tokenizer.json DELETED
@@ -1,194 +0,0 @@
1
- {
2
- "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 512,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
- "padding": {
10
- "strategy": {
11
- "Fixed": 512
12
- },
13
- "direction": "Right",
14
- "pad_to_multiple_of": null,
15
- "pad_id": 0,
16
- "pad_type_id": 0,
17
- "pad_token": "[PAD]"
18
- },
19
- "added_tokens": [
20
- {
21
- "id": 0,
22
- "content": "[PAD]",
23
- "single_word": false,
24
- "lstrip": false,
25
- "rstrip": false,
26
- "normalized": false,
27
- "special": true
28
- },
29
- {
30
- "id": 1,
31
- "content": "[UNK]",
32
- "single_word": false,
33
- "lstrip": false,
34
- "rstrip": false,
35
- "normalized": false,
36
- "special": true
37
- },
38
- {
39
- "id": 2,
40
- "content": "[CLS]",
41
- "single_word": false,
42
- "lstrip": false,
43
- "rstrip": false,
44
- "normalized": false,
45
- "special": true
46
- },
47
- {
48
- "id": 3,
49
- "content": "[SEP]",
50
- "single_word": false,
51
- "lstrip": false,
52
- "rstrip": false,
53
- "normalized": false,
54
- "special": true
55
- },
56
- {
57
- "id": 4,
58
- "content": "[MASK]",
59
- "single_word": false,
60
- "lstrip": false,
61
- "rstrip": false,
62
- "normalized": false,
63
- "special": true
64
- }
65
- ],
66
- "normalizer": {
67
- "type": "BertNormalizer",
68
- "clean_text": true,
69
- "handle_chinese_chars": true,
70
- "strip_accents": null,
71
- "lowercase": false
72
- },
73
- "pre_tokenizer": {
74
- "type": "BertPreTokenizer"
75
- },
76
- "post_processor": {
77
- "type": "TemplateProcessing",
78
- "single": [
79
- {
80
- "SpecialToken": {
81
- "id": "[CLS]",
82
- "type_id": 0
83
- }
84
- },
85
- {
86
- "Sequence": {
87
- "id": "A",
88
- "type_id": 0
89
- }
90
- },
91
- {
92
- "SpecialToken": {
93
- "id": "[SEP]",
94
- "type_id": 0
95
- }
96
- }
97
- ],
98
- "pair": [
99
- {
100
- "SpecialToken": {
101
- "id": "[CLS]",
102
- "type_id": 0
103
- }
104
- },
105
- {
106
- "Sequence": {
107
- "id": "A",
108
- "type_id": 0
109
- }
110
- },
111
- {
112
- "SpecialToken": {
113
- "id": "[SEP]",
114
- "type_id": 0
115
- }
116
- },
117
- {
118
- "Sequence": {
119
- "id": "B",
120
- "type_id": 1
121
- }
122
- },
123
- {
124
- "SpecialToken": {
125
- "id": "[SEP]",
126
- "type_id": 1
127
- }
128
- }
129
- ],
130
- "special_tokens": {
131
- "[CLS]": {
132
- "id": "[CLS]",
133
- "ids": [
134
- 2
135
- ],
136
- "tokens": [
137
- "[CLS]"
138
- ]
139
- },
140
- "[SEP]": {
141
- "id": "[SEP]",
142
- "ids": [
143
- 3
144
- ],
145
- "tokens": [
146
- "[SEP]"
147
- ]
148
- }
149
- }
150
- },
151
- "decoder": {
152
- "type": "WordPiece",
153
- "prefix": "##",
154
- "cleanup": true
155
- },
156
- "model": {
157
- "type": "WordPiece",
158
- "unk_token": "[UNK]",
159
- "continuing_subword_prefix": "##",
160
- "max_input_chars_per_word": 100,
161
- "vocab": {
162
- "[PAD]": 0,
163
- "[UNK]": 1,
164
- "[CLS]": 2,
165
- "[SEP]": 3,
166
- "[MASK]": 4,
167
- "L": 5,
168
- "A": 6,
169
- "G": 7,
170
- "V": 8,
171
- "E": 9,
172
- "S": 10,
173
- "I": 11,
174
- "K": 12,
175
- "R": 13,
176
- "D": 14,
177
- "T": 15,
178
- "P": 16,
179
- "N": 17,
180
- "Q": 18,
181
- "F": 19,
182
- "Y": 20,
183
- "M": 21,
184
- "H": 22,
185
- "C": 23,
186
- "W": 24,
187
- "X": 25,
188
- "U": 26,
189
- "B": 27,
190
- "Z": 28,
191
- "O": 29
192
- }
193
- }
194
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tokenizer_config.json DELETED
@@ -1,17 +0,0 @@
1
- {
2
- "cls_token": "[CLS]",
3
- "do_basic_tokenize": true,
4
- "do_lower_case": false,
5
- "full_tokenizer_file": null,
6
- "mask_token": "[MASK]",
7
- "model_max_length": 1000000000000000019884624838656,
8
- "name_or_path": "Rostlab/prot_bert",
9
- "never_split": null,
10
- "pad_token": "[PAD]",
11
- "sep_token": "[SEP]",
12
- "special_tokens_map_file": null,
13
- "strip_accents": null,
14
- "tokenize_chinese_chars": true,
15
- "tokenizer_class": "BertTokenizer",
16
- "unk_token": "[UNK]"
17
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:419aa73009c0ecc1eba835169dfc10cfa9702535731c4c34e1f9b7f19e068360
3
- size 3579
 
 
 
 
vocab.txt DELETED
@@ -1,30 +0,0 @@
1
- [PAD]
2
- [UNK]
3
- [CLS]
4
- [SEP]
5
- [MASK]
6
- L
7
- A
8
- G
9
- V
10
- E
11
- S
12
- I
13
- K
14
- R
15
- D
16
- T
17
- P
18
- N
19
- Q
20
- F
21
- Y
22
- M
23
- H
24
- C
25
- W
26
- X
27
- U
28
- B
29
- Z
30
- O