mp-02 commited on
Commit
a5c0eb9
1 Parent(s): 5397dd9

Training in progress, step 120

Browse files
README.md ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-sa-4.0
3
+ base_model: microsoft/layoutlmv3-base
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - precision
8
+ - recall
9
+ - f1
10
+ - accuracy
11
+ model-index:
12
+ - name: layoutlmv3-finetuned-funsd2
13
+ results: []
14
+ ---
15
+
16
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
+ should probably proofread and complete it, then remove this comment. -->
18
+
19
+ # layoutlmv3-finetuned-funsd2
20
+
21
+ This model is a fine-tuned version of [microsoft/layoutlmv3-base](https://huggingface.co/microsoft/layoutlmv3-base) on an unknown dataset.
22
+ It achieves the following results on the evaluation set:
23
+ - Loss: 0.2991
24
+ - Precision: 0.7279
25
+ - Recall: 0.7091
26
+ - F1: 0.7183
27
+ - Accuracy: 0.9111
28
+
29
+ ## Model description
30
+
31
+ More information needed
32
+
33
+ ## Intended uses & limitations
34
+
35
+ More information needed
36
+
37
+ ## Training and evaluation data
38
+
39
+ More information needed
40
+
41
+ ## Training procedure
42
+
43
+ ### Training hyperparameters
44
+
45
+ The following hyperparameters were used during training:
46
+ - learning_rate: 5e-05
47
+ - train_batch_size: 14
48
+ - eval_batch_size: 14
49
+ - seed: 42
50
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
+ - lr_scheduler_type: linear
52
+ - training_steps: 120
53
+
54
+ ### Training results
55
+
56
+ | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
57
+ |:-------------:|:-------:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
58
+ | No log | 0.9091 | 10 | 0.5192 | 0.5148 | 0.4871 | 0.5006 | 0.8322 |
59
+ | No log | 1.8182 | 20 | 0.3608 | 0.6485 | 0.6164 | 0.6320 | 0.8912 |
60
+ | No log | 2.7273 | 30 | 0.2978 | 0.6762 | 0.6616 | 0.6688 | 0.9027 |
61
+ | No log | 3.6364 | 40 | 0.2697 | 0.6768 | 0.6724 | 0.6746 | 0.8996 |
62
+ | No log | 4.5455 | 50 | 0.2737 | 0.6726 | 0.6509 | 0.6616 | 0.8989 |
63
+ | No log | 5.4545 | 60 | 0.2784 | 0.6667 | 0.6724 | 0.6695 | 0.8973 |
64
+ | No log | 6.3636 | 70 | 0.2536 | 0.7054 | 0.6810 | 0.6930 | 0.9096 |
65
+ | No log | 7.2727 | 80 | 0.2803 | 0.7100 | 0.7069 | 0.7084 | 0.9103 |
66
+ | No log | 8.1818 | 90 | 0.2924 | 0.7165 | 0.7026 | 0.7095 | 0.9057 |
67
+ | No log | 9.0909 | 100 | 0.2993 | 0.6801 | 0.6918 | 0.6859 | 0.9004 |
68
+ | No log | 10.0 | 110 | 0.3056 | 0.7013 | 0.6983 | 0.6998 | 0.9057 |
69
+ | No log | 10.9091 | 120 | 0.2991 | 0.7279 | 0.7091 | 0.7183 | 0.9111 |
70
+
71
+
72
+ ### Framework versions
73
+
74
+ - Transformers 4.44.0
75
+ - Pytorch 2.4.0+cu118
76
+ - Datasets 2.20.0
77
+ - Tokenizers 0.19.1
all_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "predict_accuracy": 0.8892307692307693,
3
+ "predict_f1": 0.6833114323258869,
4
+ "predict_loss": 0.3853023946285248,
5
+ "predict_precision": 0.6718346253229974,
6
+ "predict_recall": 0.6951871657754011,
7
+ "predict_runtime": 1.5167,
8
+ "predict_samples_per_second": 11.868,
9
+ "predict_steps_per_second": 1.319
10
+ }
config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/layoutlmv3-base",
3
+ "architectures": [
4
+ "LayoutLMv3ForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "coordinate_size": 128,
10
+ "eos_token_id": 2,
11
+ "has_relative_attention_bias": true,
12
+ "has_spatial_attention_bias": true,
13
+ "hidden_act": "gelu",
14
+ "hidden_dropout_prob": 0.1,
15
+ "hidden_size": 768,
16
+ "id2label": {
17
+ "0": "O",
18
+ "1": "B-HEADER",
19
+ "2": "I-HEADER",
20
+ "3": "B-QUESTION",
21
+ "4": "I-QUESTION",
22
+ "5": "B-ANSWER",
23
+ "6": "I-ANSWER"
24
+ },
25
+ "initializer_range": 0.02,
26
+ "input_size": 224,
27
+ "intermediate_size": 3072,
28
+ "label2id": {
29
+ "B-ANSWER": 5,
30
+ "B-HEADER": 1,
31
+ "B-QUESTION": 3,
32
+ "I-ANSWER": 6,
33
+ "I-HEADER": 2,
34
+ "I-QUESTION": 4,
35
+ "O": 0
36
+ },
37
+ "layer_norm_eps": 1e-05,
38
+ "max_2d_position_embeddings": 1024,
39
+ "max_position_embeddings": 514,
40
+ "max_rel_2d_pos": 256,
41
+ "max_rel_pos": 128,
42
+ "model_type": "layoutlmv3",
43
+ "num_attention_heads": 12,
44
+ "num_channels": 3,
45
+ "num_hidden_layers": 12,
46
+ "pad_token_id": 1,
47
+ "patch_size": 16,
48
+ "rel_2d_pos_bins": 64,
49
+ "rel_pos_bins": 32,
50
+ "second_input_size": 112,
51
+ "shape_size": 128,
52
+ "text_embed": true,
53
+ "torch_dtype": "float32",
54
+ "transformers_version": "4.44.0",
55
+ "type_vocab_size": 1,
56
+ "visual_embed": true,
57
+ "vocab_size": 50265
58
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7821498968442bbbb290d9f95ab619ed81c4c8b17ab488095b4c498c4ce9467a
3
+ size 501355532
predict_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "predict_accuracy": 0.8892307692307693,
3
+ "predict_f1": 0.6833114323258869,
4
+ "predict_loss": 0.3853023946285248,
5
+ "predict_precision": 0.6718346253229974,
6
+ "predict_recall": 0.6951871657754011,
7
+ "predict_runtime": 1.5167,
8
+ "predict_samples_per_second": 11.868,
9
+ "predict_steps_per_second": 1.319
10
+ }
predictions.txt ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ header header other question question question question header header header answer answer answer header question header question header question header question answer header question
2
+ header header header header header header answer header header other other header answer header header header question question question question question question other other
3
+ header header header header header header question question question question question other answer header question header header question header question question header header question question header header question question
4
+ header header header header header header header header header header header header header header question header header question question header header header header other other other other other answer header question header header header header header header header header header header header header header header header header header header header question question question question question question question question question question question question question question question question question question question question question question question question question question question question question question question question question question question
5
+ header header header other other other question other other question question question other answer question question header other header header question header header other other
6
+ header header header header header header answer answer header header header header header header header header header question header question question question question question question question question question question question header header header answer answer header header header header header header answer answer other other question question question header question question header question header question header question question question question question question question question header header header header header header header header question header header question question question question question question question question question question header header header header header header header header header header question header question other question question answer header
7
+ other answer header header header header answer header question header question header question header question header question header question header question question question header
8
+ header header header header header question question question question question other answer header header header header header header question question header header header question header question question header header question header question header question header question
9
+ header header header header header header header header question header header header header question question other header other header header header header answer other other question question header header question question question header header question question question header header header header header header header
10
+ header header header header header header header header header header header header header header header header header header other header header header answer other header header header question question question question header question header header question question question header question other header header other header question question question
11
+ header header header question header header header header other header answer header header header question question header question question question question header question header question answer header header header header header header header header other other other
12
+ answer header header header header header header header header header header header header header header header header header header header header header header question question header header header header header question question question question question question question question other question header header header header header header header header header header header header header header question question question answer question header question header question question question header header header question header header header question header question question question header question header header header header other
13
+ other answer header header header header header header header header header header header header header header question header header header header header header question question question question question question question question question question question question header header header header header question question question question other header question other question answer question question question answer other question question question question question question question header header question question question question header header question header header header header header header header header header header header header header header header header header question question question question question question question question header header header header header header header header header header header header header answer other other other
14
+ header question header header question header header question question question header question question question question question question question other answer answer answer question question question question question question header question header question header question question question header header header header header header question question question header answer
15
+ other header header question header header header header header header header question question question header header other header header header question answer other header question question header question other header question header header header question question question header header header header question question question question header header header header question question
16
+ question question question question header answer question question header header header other question question header answer other other other header question header header header header header header other other header header header question question question header question header header header header header header header header question header question other header other question header header header answer header question question question question question question other other question question other other header other header question
17
+ header header header header header question question header question header header header question header header header header header question other other other other question header header header question other header header other answer answer other other question question question question header header header question other answer header question header question header header header header header other header header header question header question question header header question question header header header header question question answer header question header question header question other other other
18
+ header header header header header question question header header header header question other other header other other answer other question question other other question question other header question other other other header question header question header other other other other other answer header header other
preprocessor_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "apply_ocr": false,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.5,
8
+ 0.5,
9
+ 0.5
10
+ ],
11
+ "image_processor_type": "LayoutLMv3ImageProcessor",
12
+ "image_std": [
13
+ 0.5,
14
+ 0.5,
15
+ 0.5
16
+ ],
17
+ "ocr_lang": null,
18
+ "processor_class": "LayoutLMv3Processor",
19
+ "resample": 2,
20
+ "rescale_factor": 0.00392156862745098,
21
+ "size": {
22
+ "height": 224,
23
+ "width": 224
24
+ },
25
+ "tesseract_config": ""
26
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": true,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": true,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": true,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "<unk>",
46
+ "lstrip": false,
47
+ "normalized": true,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<pad>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "50264": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
+ "normalized": true,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ }
44
+ },
45
+ "apply_ocr": false,
46
+ "bos_token": "<s>",
47
+ "clean_up_tokenization_spaces": true,
48
+ "cls_token": "<s>",
49
+ "cls_token_box": [
50
+ 0,
51
+ 0,
52
+ 0,
53
+ 0
54
+ ],
55
+ "eos_token": "</s>",
56
+ "errors": "replace",
57
+ "mask_token": "<mask>",
58
+ "model_max_length": 512,
59
+ "only_label_first_subword": true,
60
+ "pad_token": "<pad>",
61
+ "pad_token_box": [
62
+ 0,
63
+ 0,
64
+ 0,
65
+ 0
66
+ ],
67
+ "pad_token_label": -100,
68
+ "processor_class": "LayoutLMv3Processor",
69
+ "sep_token": "</s>",
70
+ "sep_token_box": [
71
+ 0,
72
+ 0,
73
+ 0,
74
+ 0
75
+ ],
76
+ "tokenizer_class": "LayoutLMv3Tokenizer",
77
+ "trim_offsets": true,
78
+ "unk_token": "<unk>",
79
+ "use_fast": true
80
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79a00e9e2eb4d526ef08352565518f227fb0a5281651fc78cd4ba3cb25fb2f28
3
+ size 5240
vocab.json ADDED
The diff for this file is too large to render. See raw diff