kamikaze20
commited on
Commit
•
4000310
1
Parent(s):
dfb14d8
Upload folder using huggingface_hub
Browse files- checkpoint-2733/config.json +30 -0
- checkpoint-2733/model.safetensors +3 -0
- checkpoint-2733/optimizer.pt +3 -0
- checkpoint-2733/rng_state.pth +3 -0
- checkpoint-2733/scheduler.pt +3 -0
- checkpoint-2733/trainer_state.json +0 -0
- checkpoint-2733/training_args.bin +3 -0
- checkpoint-5466/config.json +30 -0
- checkpoint-5466/model.safetensors +3 -0
- checkpoint-5466/optimizer.pt +3 -0
- checkpoint-5466/rng_state.pth +3 -0
- checkpoint-5466/scheduler.pt +3 -0
- checkpoint-5466/trainer_state.json +0 -0
- checkpoint-5466/training_args.bin +3 -0
- config.json +30 -0
- merges.txt +0 -0
- model.safetensors +3 -0
- training_args.bin +3 -0
- vocab.json +0 -0
checkpoint-2733/config.json
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "distilbert-base-cased",
|
3 |
+
"activation": "gelu",
|
4 |
+
"architectures": [
|
5 |
+
"DistilBertForSequenceClassification"
|
6 |
+
],
|
7 |
+
"attention_dropout": 0.1,
|
8 |
+
"dim": 768,
|
9 |
+
"dropout": 0.1,
|
10 |
+
"hidden_dim": 3072,
|
11 |
+
"id2label": {
|
12 |
+
"0": "SAFE EMAIL",
|
13 |
+
"1": "PHISHING EMAIL"
|
14 |
+
},
|
15 |
+
"initializer_range": 0.02,
|
16 |
+
"max_position_embeddings": 512,
|
17 |
+
"model_type": "distilbert",
|
18 |
+
"n_heads": 12,
|
19 |
+
"n_layers": 6,
|
20 |
+
"output_past": true,
|
21 |
+
"pad_token_id": 0,
|
22 |
+
"problem_type": "single_label_classification",
|
23 |
+
"qa_dropout": 0.1,
|
24 |
+
"seq_classif_dropout": 0.2,
|
25 |
+
"sinusoidal_pos_embds": false,
|
26 |
+
"tie_weights_": true,
|
27 |
+
"torch_dtype": "float32",
|
28 |
+
"transformers_version": "4.42.3",
|
29 |
+
"vocab_size": 28996
|
30 |
+
}
|
checkpoint-2733/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:276ccf774a87410a40489fb30eed30f5779d8bdec849312f0a3fbf8c5d19f288
|
3 |
+
size 263144680
|
checkpoint-2733/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6c40ce5113a3027ef045f19748dc68ace9f700b7845a5708fe8686d0910d9aff
|
3 |
+
size 526351546
|
checkpoint-2733/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8113af33472187a41f747dae65cf3509668a803caafe57203ad4700e35941843
|
3 |
+
size 14244
|
checkpoint-2733/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1a0fa4736f88092d1e51765ae0882bc9cfcefaf33b5647bb67cdbd82151d3c95
|
3 |
+
size 1064
|
checkpoint-2733/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-2733/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:456eacf2cf0202dd95b4b727fe6e5853a883055cc0bdd4bfaf908a96de2e958c
|
3 |
+
size 5112
|
checkpoint-5466/config.json
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "distilbert-base-cased",
|
3 |
+
"activation": "gelu",
|
4 |
+
"architectures": [
|
5 |
+
"DistilBertForSequenceClassification"
|
6 |
+
],
|
7 |
+
"attention_dropout": 0.1,
|
8 |
+
"dim": 768,
|
9 |
+
"dropout": 0.1,
|
10 |
+
"hidden_dim": 3072,
|
11 |
+
"id2label": {
|
12 |
+
"0": "SAFE EMAIL",
|
13 |
+
"1": "PHISHING EMAIL"
|
14 |
+
},
|
15 |
+
"initializer_range": 0.02,
|
16 |
+
"max_position_embeddings": 512,
|
17 |
+
"model_type": "distilbert",
|
18 |
+
"n_heads": 12,
|
19 |
+
"n_layers": 6,
|
20 |
+
"output_past": true,
|
21 |
+
"pad_token_id": 0,
|
22 |
+
"problem_type": "single_label_classification",
|
23 |
+
"qa_dropout": 0.1,
|
24 |
+
"seq_classif_dropout": 0.2,
|
25 |
+
"sinusoidal_pos_embds": false,
|
26 |
+
"tie_weights_": true,
|
27 |
+
"torch_dtype": "float32",
|
28 |
+
"transformers_version": "4.42.3",
|
29 |
+
"vocab_size": 28996
|
30 |
+
}
|
checkpoint-5466/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d6d3b0ba5ae07ce6588734f2328217ad31cfe510a6f1441646a9ad78c3116a72
|
3 |
+
size 263144680
|
checkpoint-5466/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3e9fe29c55f7b210afe25de814803fd30014737f8b2ff7eddffa1d126883a3fb
|
3 |
+
size 526351546
|
checkpoint-5466/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:58efe8863288e0513279dc7793cc5e5ab71cad90e6c44738771caa3e66175bf2
|
3 |
+
size 14244
|
checkpoint-5466/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:852e420945546f8a871e922c84d1c4895469427bcb049640d8e38e6735ac520f
|
3 |
+
size 1064
|
checkpoint-5466/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-5466/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:456eacf2cf0202dd95b4b727fe6e5853a883055cc0bdd4bfaf908a96de2e958c
|
3 |
+
size 5112
|
config.json
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "distilbert-base-cased",
|
3 |
+
"activation": "gelu",
|
4 |
+
"architectures": [
|
5 |
+
"DistilBertForSequenceClassification"
|
6 |
+
],
|
7 |
+
"attention_dropout": 0.1,
|
8 |
+
"dim": 768,
|
9 |
+
"dropout": 0.1,
|
10 |
+
"hidden_dim": 3072,
|
11 |
+
"id2label": {
|
12 |
+
"0": "SAFE EMAIL",
|
13 |
+
"1": "PHISHING EMAIL"
|
14 |
+
},
|
15 |
+
"initializer_range": 0.02,
|
16 |
+
"max_position_embeddings": 512,
|
17 |
+
"model_type": "distilbert",
|
18 |
+
"n_heads": 12,
|
19 |
+
"n_layers": 6,
|
20 |
+
"output_past": true,
|
21 |
+
"pad_token_id": 0,
|
22 |
+
"problem_type": "single_label_classification",
|
23 |
+
"qa_dropout": 0.1,
|
24 |
+
"seq_classif_dropout": 0.2,
|
25 |
+
"sinusoidal_pos_embds": false,
|
26 |
+
"tie_weights_": true,
|
27 |
+
"torch_dtype": "float32",
|
28 |
+
"transformers_version": "4.42.3",
|
29 |
+
"vocab_size": 28996
|
30 |
+
}
|
merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:276ccf774a87410a40489fb30eed30f5779d8bdec849312f0a3fbf8c5d19f288
|
3 |
+
size 263144680
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:456eacf2cf0202dd95b4b727fe6e5853a883055cc0bdd4bfaf908a96de2e958c
|
3 |
+
size 5112
|
vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|