vazish commited on
Commit
4fb32a9
1 Parent(s): fd04c9c

vazish/paraphrase-multilingual-MiniLM-L12-v2

Browse files
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ unigram.json filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  library_name: transformers
3
  license: apache-2.0
4
- base_model: distilbert/distilbert-base-multilingual-cased
5
  tags:
6
  - generated_from_trainer
7
  metrics:
@@ -11,8 +11,6 @@ metrics:
11
  model-index:
12
  - name: fine-tuned-distilbert-autofill
13
  results: []
14
- datasets:
15
- - vazish/autofill_15_labels
16
  ---
17
 
18
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -20,42 +18,29 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  # fine-tuned-distilbert-autofill
22
 
23
- This model is a fine-tuned version of [distilbert/distilbert-base-multilingual-cased](https://huggingface.co/distilbert/distilbert-base-multilingual-cased) on the [autofill_15_labels](https://huggingface.co/datasets/vazish/autofill_15_labels) dataset.
24
  It achieves the following results on the evaluation set:
25
- - Loss: 0.2261
26
- - Precision: 0.9637
27
- - Recall: 0.9582
28
- - F1: 0.9601
29
- - Confusion Matrix: [[ 93 7 0 0 0 0 0 0 0 0 0 0 0 0
30
- 0]
31
- [ 9 49 0 0 0 0 0 0 0 0 0 0 0 0
32
- 0]
33
- [ 0 0 51 0 0 0 0 0 0 0 0 0 0 0
34
- 9]
35
- [ 0 0 0 33 0 0 0 0 1 1 0 0 0 0
36
- 0]
37
- [ 0 0 1 0 66 0 0 0 0 0 0 0 0 0
38
- 9]
39
- [ 0 0 0 0 0 53 0 0 1 1 0 0 0 0
40
- 0]
41
- [ 0 0 0 0 0 0 37 1 0 0 0 0 0 0
42
- 3]
43
- [ 0 0 0 0 0 0 4 35 0 0 0 0 0 0
44
- 2]
45
- [ 0 0 0 0 0 1 0 0 46 0 0 0 0 0
46
- 0]
47
- [ 0 0 0 0 0 0 0 0 0 32 0 0 0 0
48
- 0]
49
- [ 0 0 0 0 0 0 0 2 0 0 12 0 0 0
50
- 0]
51
- [ 0 0 0 0 0 0 0 1 0 0 1 16 0 0
52
- 0]
53
- [ 0 0 0 0 0 0 0 0 0 1 0 0 5 0
54
- 0]
55
- [ 0 0 0 0 0 0 0 0 0 1 0 0 0 5
56
- 0]
57
- [ 2 0 39 3 12 0 4 0 6 1 2 0 3 1
58
- 2427]]
59
 
60
  ## Model description
61
 
@@ -80,22 +65,97 @@ The following hyperparameters were used during training:
80
  - seed: 42
81
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
82
  - lr_scheduler_type: linear
83
- - num_epochs: 10
84
 
85
  ### Training results
86
 
87
- | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 |
88
- |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|
89
- | 0.902 | 1.0 | 598 | 0.2462 | 0.9202 | 0.9304 | 0.9208 |
90
- | 0.4122 | 2.0 | 1196 | 0.2071 | 0.9561 | 0.9456 | 0.9457 |
91
- | 0.2816 | 3.0 | 1794 | 0.2004 | 0.9566 | 0.9508 | 0.9518 |
92
- | 0.2272 | 4.0 | 2392 | 0.1930 | 0.9651 | 0.9595 | 0.9606 |
93
- | 0.1854 | 5.0 | 2990 | 0.2301 | 0.9593 | 0.9501 | 0.9529 |
94
- | 0.1064 | 6.0 | 3588 | 0.2330 | 0.9593 | 0.9518 | 0.9539 |
95
- | 0.0756 | 7.0 | 4186 | 0.2207 | 0.9641 | 0.9592 | 0.9606 |
96
- | 0.0815 | 8.0 | 4784 | 0.2203 | 0.9651 | 0.9615 | 0.9621 |
97
- | 0.0677 | 9.0 | 5382 | 0.2214 | 0.9636 | 0.9586 | 0.9602 |
98
- | 0.0579 | 10.0 | 5980 | 0.2261 | 0.9637 | 0.9582 | 0.9601 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
99
 
100
 
101
  ### Framework versions
@@ -103,4 +163,4 @@ The following hyperparameters were used during training:
103
  - Transformers 4.44.2
104
  - Pytorch 2.1.2
105
  - Datasets 2.19.2
106
- - Tokenizers 0.19.1
 
1
  ---
2
  library_name: transformers
3
  license: apache-2.0
4
+ base_model: sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2
5
  tags:
6
  - generated_from_trainer
7
  metrics:
 
11
  model-index:
12
  - name: fine-tuned-distilbert-autofill
13
  results: []
 
 
14
  ---
15
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
18
 
19
  # fine-tuned-distilbert-autofill
20
 
21
+ This model is a fine-tuned version of [sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2) on the None dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 0.2367
24
+ - Precision: 0.9484
25
+ - Recall: 0.9473
26
+ - F1: 0.9473
27
+ - Confusion Matrix: [[ 94 5 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0]
28
+ [ 14 44 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]
29
+ [ 0 0 130 0 0 0 0 0 0 0 0 0 0 0 0 0 9]
30
+ [ 0 0 0 33 0 0 0 0 0 0 0 0 0 0 2 0 0]
31
+ [ 0 0 2 0 64 0 0 0 0 0 3 0 0 0 0 0 7]
32
+ [ 0 0 0 0 0 53 0 0 0 0 0 0 0 0 2 0 0]
33
+ [ 0 0 0 0 0 0 37 1 0 0 0 0 0 0 0 0 3]
34
+ [ 0 0 0 0 0 0 4 35 0 0 0 0 0 0 0 0 2]
35
+ [ 1 0 0 0 0 1 0 0 43 0 0 0 0 0 2 0 0]
36
+ [ 0 0 0 0 0 0 0 0 0 31 0 0 0 0 1 0 0]
37
+ [ 0 0 0 0 2 0 0 2 0 0 10 0 0 0 0 0 0]
38
+ [ 0 0 0 0 0 0 0 1 0 0 1 16 0 0 0 0 0]
39
+ [ 0 0 0 0 0 0 0 0 0 1 0 0 5 0 0 0 0]
40
+ [ 0 0 0 0 0 0 0 0 0 0 0 0 0 6 0 0 0]
41
+ [ 0 0 0 1 0 0 1 0 0 0 0 0 0 0 73 0 0]
42
+ [ 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 3 0]
43
+ [ 1 0 9 1 4 0 0 0 2 0 2 0 1 2 1 0 977]]
 
 
 
 
 
 
 
 
 
 
 
 
 
44
 
45
  ## Model description
46
 
 
65
  - seed: 42
66
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
67
  - lr_scheduler_type: linear
68
+ - num_epochs: 5
69
 
70
  ### Training results
71
 
72
+ | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Confusion Matrix |
73
+ |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------:|
74
+ | 0.7726 | 1.0 | 987 | 0.3096 | 0.8920 | 0.9141 | 0.8988 | [[100 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]
75
+ [ 58 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]
76
+ [ 0 0 129 0 0 0 0 0 0 0 0 0 0 0 0 0 10]
77
+ [ 0 0 0 32 0 0 0 0 1 1 0 0 0 0 0 0 1]
78
+ [ 0 0 4 0 63 0 0 0 0 0 0 0 0 0 0 0 9]
79
+ [ 0 0 0 0 0 52 0 0 0 2 0 0 0 0 0 0 1]
80
+ [ 0 0 0 0 0 0 36 0 0 0 2 0 0 0 0 0 3]
81
+ [ 0 0 0 0 0 0 2 33 0 0 4 0 0 0 0 0 2]
82
+ [ 1 0 0 0 0 1 0 0 43 2 0 0 0 0 0 0 0]
83
+ [ 0 0 0 0 0 0 0 0 0 32 0 0 0 0 0 0 0]
84
+ [ 0 0 0 0 2 0 0 0 0 0 12 0 0 0 0 0 0]
85
+ [ 0 0 0 0 0 0 4 0 0 0 1 13 0 0 0 0 0]
86
+ [ 0 0 0 0 0 0 0 0 0 6 0 0 0 0 0 0 0]
87
+ [ 0 0 0 0 0 0 0 0 0 6 0 0 0 0 0 0 0]
88
+ [ 0 0 0 0 0 0 1 0 1 2 0 0 0 0 71 0 0]
89
+ [ 0 0 0 0 2 0 0 0 0 1 0 0 0 0 0 0 0]
90
+ [ 1 0 7 1 5 0 0 0 1 3 1 0 0 0 1 0 980]] |
91
+ | 0.2616 | 2.0 | 1974 | 0.2645 | 0.9356 | 0.9273 | 0.9179 | [[ 99 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0]
92
+ [ 43 7 5 0 0 0 0 0 0 0 0 0 0 0 0 0 3]
93
+ [ 0 0 128 0 0 0 0 0 0 0 0 0 0 0 0 0 11]
94
+ [ 0 0 0 33 0 0 0 0 1 0 0 0 0 0 0 0 1]
95
+ [ 0 0 0 0 64 0 0 0 0 0 0 0 0 0 0 0 12]
96
+ [ 0 0 0 1 0 53 0 0 0 0 0 0 0 0 0 0 1]
97
+ [ 0 0 0 0 0 0 36 2 0 0 0 0 0 0 0 0 3]
98
+ [ 0 0 0 0 0 0 3 36 0 0 0 0 0 0 0 0 2]
99
+ [ 1 0 0 0 0 2 0 0 43 0 0 0 0 0 0 0 1]
100
+ [ 0 0 0 1 0 0 0 0 0 31 0 0 0 0 0 0 0]
101
+ [ 0 0 0 0 2 0 0 3 0 0 9 0 0 0 0 0 0]
102
+ [ 0 0 0 0 0 0 1 3 0 0 1 13 0 0 0 0 0]
103
+ [ 0 0 0 0 0 0 0 0 0 1 0 0 5 0 0 0 0]
104
+ [ 0 0 0 0 0 0 0 0 0 0 0 0 0 6 0 0 0]
105
+ [ 0 0 0 1 0 0 1 0 1 1 0 0 0 0 71 0 0]
106
+ [ 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 3 0]
107
+ [ 1 0 6 1 3 0 0 0 1 0 2 0 1 2 1 0 982]] |
108
+ | 0.1814 | 3.0 | 2961 | 0.2332 | 0.9437 | 0.9422 | 0.9420 | [[ 94 5 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0]
109
+ [ 15 43 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]
110
+ [ 0 0 127 0 1 0 0 0 0 0 0 0 0 0 0 0 11]
111
+ [ 0 0 0 34 0 0 0 0 0 0 0 0 0 0 0 0 1]
112
+ [ 0 0 1 0 63 0 0 0 0 0 2 0 0 0 0 0 10]
113
+ [ 0 0 0 1 0 52 0 0 0 1 1 0 0 0 0 0 0]
114
+ [ 0 0 0 0 0 0 37 1 0 0 0 0 0 0 0 0 3]
115
+ [ 0 0 0 0 0 0 4 35 0 0 0 0 0 0 0 0 2]
116
+ [ 1 0 0 0 0 1 0 0 43 2 0 0 0 0 0 0 0]
117
+ [ 0 0 0 1 0 0 0 0 0 31 0 0 0 0 0 0 0]
118
+ [ 0 0 0 0 2 0 0 2 0 0 10 0 0 0 0 0 0]
119
+ [ 0 0 0 0 0 0 2 2 0 0 1 13 0 0 0 0 0]
120
+ [ 0 0 0 0 0 0 0 0 0 1 0 0 5 0 0 0 0]
121
+ [ 0 0 0 0 0 0 0 0 0 0 0 0 0 6 0 0 0]
122
+ [ 0 0 0 1 0 0 1 0 0 1 0 0 1 0 71 0 0]
123
+ [ 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 3 0]
124
+ [ 1 0 8 1 4 0 0 1 1 0 2 0 1 2 1 0 978]] |
125
+ | 0.1248 | 4.0 | 3948 | 0.2255 | 0.9501 | 0.9479 | 0.9482 | [[ 95 4 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0]
126
+ [ 13 45 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]
127
+ [ 0 0 130 0 0 0 0 0 0 0 0 0 0 0 0 0 9]
128
+ [ 0 0 0 33 0 0 0 0 0 0 0 0 0 0 2 0 0]
129
+ [ 0 0 2 0 65 0 0 0 0 0 4 0 0 0 0 0 5]
130
+ [ 0 0 0 0 0 52 0 0 0 0 1 0 0 0 2 0 0]
131
+ [ 0 0 0 0 0 0 38 0 0 0 0 0 0 0 0 0 3]
132
+ [ 0 0 0 0 0 0 5 34 0 0 0 0 0 0 0 0 2]
133
+ [ 1 0 0 0 0 1 0 0 43 0 0 0 0 0 2 0 0]
134
+ [ 0 0 0 0 0 0 0 0 1 30 0 0 0 0 1 0 0]
135
+ [ 0 0 0 0 2 0 2 0 0 0 10 0 0 0 0 0 0]
136
+ [ 0 0 0 0 0 0 0 1 0 0 1 16 0 0 0 0 0]
137
+ [ 0 0 0 0 0 0 0 0 0 1 0 0 5 0 0 0 0]
138
+ [ 0 0 0 0 0 0 0 0 0 0 0 0 0 6 0 0 0]
139
+ [ 0 0 0 1 0 0 1 0 0 0 0 0 0 0 73 0 0]
140
+ [ 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 3 0]
141
+ [ 1 0 9 1 4 0 0 0 2 0 2 0 1 2 1 0 977]] |
142
+ | 0.1032 | 5.0 | 4935 | 0.2367 | 0.9484 | 0.9473 | 0.9473 | [[ 94 5 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0]
143
+ [ 14 44 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0]
144
+ [ 0 0 130 0 0 0 0 0 0 0 0 0 0 0 0 0 9]
145
+ [ 0 0 0 33 0 0 0 0 0 0 0 0 0 0 2 0 0]
146
+ [ 0 0 2 0 64 0 0 0 0 0 3 0 0 0 0 0 7]
147
+ [ 0 0 0 0 0 53 0 0 0 0 0 0 0 0 2 0 0]
148
+ [ 0 0 0 0 0 0 37 1 0 0 0 0 0 0 0 0 3]
149
+ [ 0 0 0 0 0 0 4 35 0 0 0 0 0 0 0 0 2]
150
+ [ 1 0 0 0 0 1 0 0 43 0 0 0 0 0 2 0 0]
151
+ [ 0 0 0 0 0 0 0 0 0 31 0 0 0 0 1 0 0]
152
+ [ 0 0 0 0 2 0 0 2 0 0 10 0 0 0 0 0 0]
153
+ [ 0 0 0 0 0 0 0 1 0 0 1 16 0 0 0 0 0]
154
+ [ 0 0 0 0 0 0 0 0 0 1 0 0 5 0 0 0 0]
155
+ [ 0 0 0 0 0 0 0 0 0 0 0 0 0 6 0 0 0]
156
+ [ 0 0 0 1 0 0 1 0 0 0 0 0 0 0 73 0 0]
157
+ [ 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 3 0]
158
+ [ 1 0 9 1 4 0 0 0 2 0 2 0 1 2 1 0 977]] |
159
 
160
 
161
  ### Framework versions
 
163
  - Transformers 4.44.2
164
  - Pytorch 2.1.2
165
  - Datasets 2.19.2
166
+ - Tokenizers 0.19.1
config.json CHANGED
@@ -1,13 +1,14 @@
1
  {
2
- "_name_or_path": "distilbert/distilbert-base-multilingual-cased",
3
- "activation": "gelu",
4
  "architectures": [
5
- "DistilBertForSequenceClassification"
6
  ],
7
- "attention_dropout": 0.1,
8
- "dim": 768,
9
- "dropout": 0.1,
10
- "hidden_dim": 3072,
 
 
11
  "id2label": {
12
  "0": "LABEL_0",
13
  "1": "LABEL_1",
@@ -23,9 +24,12 @@
23
  "11": "LABEL_11",
24
  "12": "LABEL_12",
25
  "13": "LABEL_13",
26
- "14": "LABEL_14"
 
 
27
  },
28
  "initializer_range": 0.02,
 
29
  "label2id": {
30
  "LABEL_0": 0,
31
  "LABEL_1": 1,
@@ -34,6 +38,8 @@
34
  "LABEL_12": 12,
35
  "LABEL_13": 13,
36
  "LABEL_14": 14,
 
 
37
  "LABEL_2": 2,
38
  "LABEL_3": 3,
39
  "LABEL_4": 4,
@@ -43,18 +49,17 @@
43
  "LABEL_8": 8,
44
  "LABEL_9": 9
45
  },
 
46
  "max_position_embeddings": 512,
47
- "model_type": "distilbert",
48
- "n_heads": 12,
49
- "n_layers": 6,
50
- "output_past": true,
51
  "pad_token_id": 0,
 
52
  "problem_type": "single_label_classification",
53
- "qa_dropout": 0.1,
54
- "seq_classif_dropout": 0.2,
55
- "sinusoidal_pos_embds": false,
56
- "tie_weights_": true,
57
  "torch_dtype": "float32",
58
  "transformers_version": "4.44.2",
59
- "vocab_size": 119547
 
 
60
  }
 
1
  {
2
+ "_name_or_path": "sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2",
 
3
  "architectures": [
4
+ "BertForSequenceClassification"
5
  ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 384,
12
  "id2label": {
13
  "0": "LABEL_0",
14
  "1": "LABEL_1",
 
24
  "11": "LABEL_11",
25
  "12": "LABEL_12",
26
  "13": "LABEL_13",
27
+ "14": "LABEL_14",
28
+ "15": "LABEL_15",
29
+ "16": "LABEL_16"
30
  },
31
  "initializer_range": 0.02,
32
+ "intermediate_size": 1536,
33
  "label2id": {
34
  "LABEL_0": 0,
35
  "LABEL_1": 1,
 
38
  "LABEL_12": 12,
39
  "LABEL_13": 13,
40
  "LABEL_14": 14,
41
+ "LABEL_15": 15,
42
+ "LABEL_16": 16,
43
  "LABEL_2": 2,
44
  "LABEL_3": 3,
45
  "LABEL_4": 4,
 
49
  "LABEL_8": 8,
50
  "LABEL_9": 9
51
  },
52
+ "layer_norm_eps": 1e-12,
53
  "max_position_embeddings": 512,
54
+ "model_type": "bert",
55
+ "num_attention_heads": 12,
56
+ "num_hidden_layers": 12,
 
57
  "pad_token_id": 0,
58
+ "position_embedding_type": "absolute",
59
  "problem_type": "single_label_classification",
 
 
 
 
60
  "torch_dtype": "float32",
61
  "transformers_version": "4.44.2",
62
+ "type_vocab_size": 2,
63
+ "use_cache": true,
64
+ "vocab_size": 250037
65
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9ec8fa4b168d85eab3d8a71c9fe2f7a6991f3b3a9ba48307d85fb67d9c4e3b68
3
- size 541357364
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d42a3ba75adf65462c2a5132b58cd0419eba6423439d38ef0895d9ce20b34dd
3
+ size 470664764
special_tokens_map.json CHANGED
@@ -1,7 +1,51 @@
1
  {
2
- "cls_token": "[CLS]",
3
- "mask_token": "[MASK]",
4
- "pad_token": "[PAD]",
5
- "sep_token": "[SEP]",
6
- "unk_token": "[UNK]"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  }
 
1
  {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "<unk>",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,55 +1,64 @@
1
  {
2
  "added_tokens_decoder": {
3
  "0": {
4
- "content": "[PAD]",
5
  "lstrip": false,
6
  "normalized": false,
7
  "rstrip": false,
8
  "single_word": false,
9
  "special": true
10
  },
11
- "100": {
12
- "content": "[UNK]",
13
  "lstrip": false,
14
  "normalized": false,
15
  "rstrip": false,
16
  "single_word": false,
17
  "special": true
18
  },
19
- "101": {
20
- "content": "[CLS]",
21
  "lstrip": false,
22
  "normalized": false,
23
  "rstrip": false,
24
  "single_word": false,
25
  "special": true
26
  },
27
- "102": {
28
- "content": "[SEP]",
29
  "lstrip": false,
30
  "normalized": false,
31
  "rstrip": false,
32
  "single_word": false,
33
  "special": true
34
  },
35
- "103": {
36
- "content": "[MASK]",
37
- "lstrip": false,
38
  "normalized": false,
39
  "rstrip": false,
40
  "single_word": false,
41
  "special": true
42
  }
43
  },
 
44
  "clean_up_tokenization_spaces": true,
45
- "cls_token": "[CLS]",
46
- "do_lower_case": false,
47
- "mask_token": "[MASK]",
 
 
48
  "model_max_length": 512,
49
- "pad_token": "[PAD]",
50
- "sep_token": "[SEP]",
 
 
 
 
51
  "strip_accents": null,
52
  "tokenize_chinese_chars": true,
53
- "tokenizer_class": "DistilBertTokenizer",
54
- "unk_token": "[UNK]"
 
 
55
  }
 
1
  {
2
  "added_tokens_decoder": {
3
  "0": {
4
+ "content": "<s>",
5
  "lstrip": false,
6
  "normalized": false,
7
  "rstrip": false,
8
  "single_word": false,
9
  "special": true
10
  },
11
+ "1": {
12
+ "content": "<pad>",
13
  "lstrip": false,
14
  "normalized": false,
15
  "rstrip": false,
16
  "single_word": false,
17
  "special": true
18
  },
19
+ "2": {
20
+ "content": "</s>",
21
  "lstrip": false,
22
  "normalized": false,
23
  "rstrip": false,
24
  "single_word": false,
25
  "special": true
26
  },
27
+ "3": {
28
+ "content": "<unk>",
29
  "lstrip": false,
30
  "normalized": false,
31
  "rstrip": false,
32
  "single_word": false,
33
  "special": true
34
  },
35
+ "250001": {
36
+ "content": "<mask>",
37
+ "lstrip": true,
38
  "normalized": false,
39
  "rstrip": false,
40
  "single_word": false,
41
  "special": true
42
  }
43
  },
44
+ "bos_token": "<s>",
45
  "clean_up_tokenization_spaces": true,
46
+ "cls_token": "<s>",
47
+ "do_lower_case": true,
48
+ "eos_token": "</s>",
49
+ "mask_token": "<mask>",
50
+ "max_length": 128,
51
  "model_max_length": 512,
52
+ "pad_to_multiple_of": null,
53
+ "pad_token": "<pad>",
54
+ "pad_token_type_id": 0,
55
+ "padding_side": "right",
56
+ "sep_token": "</s>",
57
+ "stride": 0,
58
  "strip_accents": null,
59
  "tokenize_chinese_chars": true,
60
+ "tokenizer_class": "BertTokenizer",
61
+ "truncation_side": "right",
62
+ "truncation_strategy": "longest_first",
63
+ "unk_token": "<unk>"
64
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:291aa03153329af908c92a49b63a89dcf928a4a32e0ae2e64c63015ee3d6b40a
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76b894dfd5dfa4dd424ce8a12f04180f4166fafa716af6c53cc59dfe08c60790
3
  size 5176
unigram.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da145b5e7700ae40f16691ec32a0b1fdc1ee3298db22a31ea55f57a966c4a65d
3
+ size 14763260