sohamtiwari3120 commited on
Commit
8dec46d
1 Parent(s): d585867

Training in progress, epoch 1

Browse files
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
config.json CHANGED
@@ -1,11 +1,9 @@
1
  {
2
- "_name_or_path": "bert-base-uncased",
3
  "architectures": [
4
- "BertForTokenClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
- "classifier_dropout": null,
8
- "gradient_checkpointing": false,
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
11
  "hidden_size": 768,
@@ -45,16 +43,27 @@
45
  "I-TaskName": 12,
46
  "O": 0
47
  },
48
- "layer_norm_eps": 1e-12,
49
  "max_position_embeddings": 512,
50
- "model_type": "bert",
 
 
51
  "num_attention_heads": 12,
52
  "num_hidden_layers": 12,
53
  "pad_token_id": 0,
54
- "position_embedding_type": "absolute",
 
 
 
 
 
 
 
 
 
 
55
  "torch_dtype": "float32",
56
  "transformers_version": "4.23.1",
57
- "type_vocab_size": 2,
58
- "use_cache": true,
59
- "vocab_size": 30522
60
  }
 
1
  {
2
+ "_name_or_path": "microsoft/deberta-v3-base",
3
  "architectures": [
4
+ "DebertaV2ForTokenClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
 
 
7
  "hidden_act": "gelu",
8
  "hidden_dropout_prob": 0.1,
9
  "hidden_size": 768,
 
43
  "I-TaskName": 12,
44
  "O": 0
45
  },
46
+ "layer_norm_eps": 1e-07,
47
  "max_position_embeddings": 512,
48
+ "max_relative_positions": -1,
49
+ "model_type": "deberta-v2",
50
+ "norm_rel_ebd": "layer_norm",
51
  "num_attention_heads": 12,
52
  "num_hidden_layers": 12,
53
  "pad_token_id": 0,
54
+ "pooler_dropout": 0,
55
+ "pooler_hidden_act": "gelu",
56
+ "pooler_hidden_size": 768,
57
+ "pos_att_type": [
58
+ "p2c",
59
+ "c2p"
60
+ ],
61
+ "position_biased_input": false,
62
+ "position_buckets": 256,
63
+ "relative_attention": true,
64
+ "share_att_key": true,
65
  "torch_dtype": "float32",
66
  "transformers_version": "4.23.1",
67
+ "type_vocab_size": 0,
68
+ "vocab_size": 128100
 
69
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e4d15d595efafa800453610c5f561aa4a9899c79971a687c4b3013d10bd1330
3
- size 435683121
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3caac61944e37ff1fe421707daf1fc313870544851fdc48063c0e89081442f0b
3
+ size 735443887
special_tokens_map.json CHANGED
@@ -1,5 +1,7 @@
1
  {
 
2
  "cls_token": "[CLS]",
 
3
  "mask_token": "[MASK]",
4
  "pad_token": "[PAD]",
5
  "sep_token": "[SEP]",
 
1
  {
2
+ "bos_token": "[CLS]",
3
  "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
  "mask_token": "[MASK]",
6
  "pad_token": "[PAD]",
7
  "sep_token": "[SEP]",
spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,15 +1,17 @@
1
  {
2
  "add_prefix_space": true,
 
3
  "cls_token": "[CLS]",
4
- "do_lower_case": true,
 
5
  "mask_token": "[MASK]",
6
- "model_max_length": 512,
7
- "name_or_path": "bert-base-uncased",
8
  "pad_token": "[PAD]",
9
  "sep_token": "[SEP]",
 
10
  "special_tokens_map_file": null,
11
- "strip_accents": null,
12
- "tokenize_chinese_chars": true,
13
- "tokenizer_class": "BertTokenizer",
14
- "unk_token": "[UNK]"
15
  }
 
1
  {
2
  "add_prefix_space": true,
3
+ "bos_token": "[CLS]",
4
  "cls_token": "[CLS]",
5
+ "do_lower_case": false,
6
+ "eos_token": "[SEP]",
7
  "mask_token": "[MASK]",
8
+ "name_or_path": "microsoft/deberta-v3-base",
 
9
  "pad_token": "[PAD]",
10
  "sep_token": "[SEP]",
11
+ "sp_model_kwargs": {},
12
  "special_tokens_map_file": null,
13
+ "split_by_punct": false,
14
+ "tokenizer_class": "DebertaV2Tokenizer",
15
+ "unk_token": "[UNK]",
16
+ "vocab_type": "spm"
17
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f86581a44d3378a536fe67952e3ecc0fe10db95841c20a2d1fbf1bdbfe5a8589
3
  size 3503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a330159fe90edd986bfad9c324118455920beceb6cc1eaed04e9fde6ec47bba6
3
  size 3503