adityaprakhar commited on
Commit
0d0cb1a
·
verified ·
1 Parent(s): b3b76f5

Training in progress, step 500

Browse files
config.json CHANGED
@@ -1,9 +1,12 @@
1
  {
2
- "_name_or_path": "microsoft/layoutlm-base-uncased",
3
  "architectures": [
4
  "LayoutLMForTokenClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
 
 
 
7
  "hidden_act": "gelu",
8
  "hidden_dropout_prob": 0.1,
9
  "hidden_size": 768,
@@ -45,18 +48,18 @@
45
  "parties": 10,
46
  "telephone": 14
47
  },
48
- "layer_norm_eps": 1e-12,
49
  "max_2d_position_embeddings": 1024,
50
- "max_position_embeddings": 512,
51
  "model_type": "layoutlm",
52
  "num_attention_heads": 12,
53
  "num_hidden_layers": 12,
54
- "output_past": true,
55
- "pad_token_id": 0,
56
  "position_embedding_type": "absolute",
 
57
  "torch_dtype": "float32",
58
  "transformers_version": "4.39.0.dev0",
59
- "type_vocab_size": 2,
60
  "use_cache": true,
61
- "vocab_size": 30522
62
  }
 
1
  {
2
+ "_name_or_path": "microsoft/layoutlm-base-cased",
3
  "architectures": [
4
  "LayoutLMForTokenClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "eos_token_id": 2,
9
+ "gradient_checkpointing": false,
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
  "hidden_size": 768,
 
48
  "parties": 10,
49
  "telephone": 14
50
  },
51
+ "layer_norm_eps": 1e-05,
52
  "max_2d_position_embeddings": 1024,
53
+ "max_position_embeddings": 514,
54
  "model_type": "layoutlm",
55
  "num_attention_heads": 12,
56
  "num_hidden_layers": 12,
57
+ "pad_token_id": 1,
 
58
  "position_embedding_type": "absolute",
59
+ "tokenizer_class": "RobertaTokenizer",
60
  "torch_dtype": "float32",
61
  "transformers_version": "4.39.0.dev0",
62
+ "type_vocab_size": 1,
63
  "use_cache": true,
64
+ "vocab_size": 50265
65
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:48408ca2aaa52846d41ea09c9b4189b1ddcae6012f773e03fac23eea82ec0402
3
- size 450585896
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:534dcd70fb6540be4fb48ccd829ca7365b8ead54156d836849d7902695d730a4
3
+ size 511239464
runs/Mar14_06-07-08_1bc8b9c17103/events.out.tfevents.1710396429.1bc8b9c17103.709.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9b2aac7e4406e4059320c478904a74b7830ad2afab9eb031acbc0bd89174c4e
3
+ size 10795
tokenizer.json CHANGED
@@ -1,21 +1,7 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 512,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
- "padding": {
10
- "strategy": {
11
- "Fixed": 512
12
- },
13
- "direction": "Right",
14
- "pad_to_multiple_of": null,
15
- "pad_id": 0,
16
- "pad_type_id": 0,
17
- "pad_token": "[PAD]"
18
- },
19
  "added_tokens": [
20
  {
21
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7066bcf11f4f969e1ee2cadda78e00d46157ec009c614bdcb604a54b96c6fd8d
3
  size 4856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e08db785f258b5f2b05a232a5085485b90a93d895a080da16cac327d8e08f2e1
3
  size 4856