quanqnv19 commited on
Commit
a09a7b5
·
verified ·
1 Parent(s): de57d83

Upload 5 files

Browse files
Files changed (5) hide show
  1. bpe.codes +0 -0
  2. config.json +40 -0
  3. pytorch_model.bin +3 -0
  4. tokenizer.json +0 -0
  5. vocab.txt +0 -0
bpe.codes ADDED
The diff for this file is too large to render. See raw diff
 
config.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/content/drive/MyDrive/thesis/models/ecomm-review-phobert-base-5epochs",
3
+ "architectures": [
4
+ "RobertaForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "gradient_checkpointing": false,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "id2label": {
15
+ "0": "CLEAN",
16
+ "1": "OFFENSIVE",
17
+ "2": "HATE"
18
+ },
19
+ "initializer_range": 0.02,
20
+ "intermediate_size": 3072,
21
+ "label2id": {
22
+ "CLEAN": 0,
23
+ "OFFENSIVE": 2,
24
+ "HATE": 1
25
+ },
26
+ "layer_norm_eps": 1e-05,
27
+ "max_position_embeddings": 258,
28
+ "model_type": "roberta",
29
+ "num_attention_heads": 12,
30
+ "num_hidden_layers": 12,
31
+ "pad_token_id": 1,
32
+ "position_embedding_type": "absolute",
33
+ "problem_type": "single_label_classification",
34
+ "tokenizer_class": "PhobertTokenizer",
35
+ "torch_dtype": "float32",
36
+ "transformers_version": "4.18.0",
37
+ "type_vocab_size": 1,
38
+ "use_cache": true,
39
+ "vocab_size": 64001
40
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43db4d9d76dea05b6bd1d35d434a0a94c3f1b79b81572bdc66a88e6efe9f7c81
3
+ size 540090200
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
vocab.txt ADDED
The diff for this file is too large to render. See raw diff