senfu commited on
Commit
1c6388f
1 Parent(s): 64d0649

Upload 9 files

Browse files
config.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/mnt/data/device-aware-bert/token_pruning/teachers/SST2",
3
+ "architectures": [
4
+ "ToPBertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "do_layer_distill": false,
9
+ "finetuning_task": "sst2",
10
+ "gradient_checkpointing": false,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "id2label": {
15
+ "0": 0,
16
+ "1": 1
17
+ },
18
+ "initializer_range": 0.02,
19
+ "intermediate_size": 3072,
20
+ "label2id": {
21
+ "0": 0,
22
+ "1": 1
23
+ },
24
+ "layer_norm_eps": 1e-12,
25
+ "max_position_embeddings": 512,
26
+ "model_type": "bert",
27
+ "num_attention_heads": 12,
28
+ "num_hidden_layers": 12,
29
+ "output_attentions": true,
30
+ "output_hidden_states": true,
31
+ "pad_token_id": 0,
32
+ "position_embedding_type": "absolute",
33
+ "pruned_heads": {
34
+ "0": [],
35
+ "1": [],
36
+ "2": [],
37
+ "3": [],
38
+ "4": [],
39
+ "5": [],
40
+ "6": [],
41
+ "7": [],
42
+ "8": [],
43
+ "9": [],
44
+ "10": [],
45
+ "11": []
46
+ },
47
+ "torch_dtype": "float32",
48
+ "transformers_version": "4.16.0",
49
+ "type_vocab_size": 2,
50
+ "use_cache": true,
51
+ "vocab_size": 30522
52
+ }
l0_module.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:528a5c844d68eb2f90729ddc985d0f7abc31814cc3e3763b171cc3d21ad5651c
3
+ size 6043
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37ff8f5ddc469179b489c9fa2c86400ba3fd1cc04215063db415a3dc1f6f7fc8
3
+ size 438019245
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
sst2-log.txt ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "/mnt/data/device-aware-bert/token_pruning/teachers/SST2", "tokenizer_class": "BertTokenizer"}
vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
zs.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:968d746dac5a8e6f41b3eab45616a3ac9d60f705a172f3588e1875d029050e72
3
+ size 1959