Training in progress, step 120
Browse files- README.md +77 -0
- all_results.json +10 -0
- config.json +58 -0
- merges.txt +0 -0
- model.safetensors +3 -0
- predict_results.json +10 -0
- predictions.txt +18 -0
- preprocessor_config.json +26 -0
- special_tokens_map.json +51 -0
- tokenizer.json +0 -0
- tokenizer_config.json +80 -0
- training_args.bin +3 -0
- vocab.json +0 -0
README.md
ADDED
@@ -0,0 +1,77 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: cc-by-nc-sa-4.0
|
3 |
+
base_model: microsoft/layoutlmv3-base
|
4 |
+
tags:
|
5 |
+
- generated_from_trainer
|
6 |
+
metrics:
|
7 |
+
- precision
|
8 |
+
- recall
|
9 |
+
- f1
|
10 |
+
- accuracy
|
11 |
+
model-index:
|
12 |
+
- name: layoutlmv3-finetuned-funsd2
|
13 |
+
results: []
|
14 |
+
---
|
15 |
+
|
16 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
17 |
+
should probably proofread and complete it, then remove this comment. -->
|
18 |
+
|
19 |
+
# layoutlmv3-finetuned-funsd2
|
20 |
+
|
21 |
+
This model is a fine-tuned version of [microsoft/layoutlmv3-base](https://huggingface.co/microsoft/layoutlmv3-base) on an unknown dataset.
|
22 |
+
It achieves the following results on the evaluation set:
|
23 |
+
- Loss: 0.2991
|
24 |
+
- Precision: 0.7279
|
25 |
+
- Recall: 0.7091
|
26 |
+
- F1: 0.7183
|
27 |
+
- Accuracy: 0.9111
|
28 |
+
|
29 |
+
## Model description
|
30 |
+
|
31 |
+
More information needed
|
32 |
+
|
33 |
+
## Intended uses & limitations
|
34 |
+
|
35 |
+
More information needed
|
36 |
+
|
37 |
+
## Training and evaluation data
|
38 |
+
|
39 |
+
More information needed
|
40 |
+
|
41 |
+
## Training procedure
|
42 |
+
|
43 |
+
### Training hyperparameters
|
44 |
+
|
45 |
+
The following hyperparameters were used during training:
|
46 |
+
- learning_rate: 5e-05
|
47 |
+
- train_batch_size: 14
|
48 |
+
- eval_batch_size: 14
|
49 |
+
- seed: 42
|
50 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
51 |
+
- lr_scheduler_type: linear
|
52 |
+
- training_steps: 120
|
53 |
+
|
54 |
+
### Training results
|
55 |
+
|
56 |
+
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|
57 |
+
|:-------------:|:-------:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
|
58 |
+
| No log | 0.9091 | 10 | 0.5192 | 0.5148 | 0.4871 | 0.5006 | 0.8322 |
|
59 |
+
| No log | 1.8182 | 20 | 0.3608 | 0.6485 | 0.6164 | 0.6320 | 0.8912 |
|
60 |
+
| No log | 2.7273 | 30 | 0.2978 | 0.6762 | 0.6616 | 0.6688 | 0.9027 |
|
61 |
+
| No log | 3.6364 | 40 | 0.2697 | 0.6768 | 0.6724 | 0.6746 | 0.8996 |
|
62 |
+
| No log | 4.5455 | 50 | 0.2737 | 0.6726 | 0.6509 | 0.6616 | 0.8989 |
|
63 |
+
| No log | 5.4545 | 60 | 0.2784 | 0.6667 | 0.6724 | 0.6695 | 0.8973 |
|
64 |
+
| No log | 6.3636 | 70 | 0.2536 | 0.7054 | 0.6810 | 0.6930 | 0.9096 |
|
65 |
+
| No log | 7.2727 | 80 | 0.2803 | 0.7100 | 0.7069 | 0.7084 | 0.9103 |
|
66 |
+
| No log | 8.1818 | 90 | 0.2924 | 0.7165 | 0.7026 | 0.7095 | 0.9057 |
|
67 |
+
| No log | 9.0909 | 100 | 0.2993 | 0.6801 | 0.6918 | 0.6859 | 0.9004 |
|
68 |
+
| No log | 10.0 | 110 | 0.3056 | 0.7013 | 0.6983 | 0.6998 | 0.9057 |
|
69 |
+
| No log | 10.9091 | 120 | 0.2991 | 0.7279 | 0.7091 | 0.7183 | 0.9111 |
|
70 |
+
|
71 |
+
|
72 |
+
### Framework versions
|
73 |
+
|
74 |
+
- Transformers 4.44.0
|
75 |
+
- Pytorch 2.4.0+cu118
|
76 |
+
- Datasets 2.20.0
|
77 |
+
- Tokenizers 0.19.1
|
all_results.json
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"predict_accuracy": 0.8892307692307693,
|
3 |
+
"predict_f1": 0.6833114323258869,
|
4 |
+
"predict_loss": 0.3853023946285248,
|
5 |
+
"predict_precision": 0.6718346253229974,
|
6 |
+
"predict_recall": 0.6951871657754011,
|
7 |
+
"predict_runtime": 1.5167,
|
8 |
+
"predict_samples_per_second": 11.868,
|
9 |
+
"predict_steps_per_second": 1.319
|
10 |
+
}
|
config.json
ADDED
@@ -0,0 +1,58 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "microsoft/layoutlmv3-base",
|
3 |
+
"architectures": [
|
4 |
+
"LayoutLMv3ForTokenClassification"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0.1,
|
7 |
+
"bos_token_id": 0,
|
8 |
+
"classifier_dropout": null,
|
9 |
+
"coordinate_size": 128,
|
10 |
+
"eos_token_id": 2,
|
11 |
+
"has_relative_attention_bias": true,
|
12 |
+
"has_spatial_attention_bias": true,
|
13 |
+
"hidden_act": "gelu",
|
14 |
+
"hidden_dropout_prob": 0.1,
|
15 |
+
"hidden_size": 768,
|
16 |
+
"id2label": {
|
17 |
+
"0": "O",
|
18 |
+
"1": "B-HEADER",
|
19 |
+
"2": "I-HEADER",
|
20 |
+
"3": "B-QUESTION",
|
21 |
+
"4": "I-QUESTION",
|
22 |
+
"5": "B-ANSWER",
|
23 |
+
"6": "I-ANSWER"
|
24 |
+
},
|
25 |
+
"initializer_range": 0.02,
|
26 |
+
"input_size": 224,
|
27 |
+
"intermediate_size": 3072,
|
28 |
+
"label2id": {
|
29 |
+
"B-ANSWER": 5,
|
30 |
+
"B-HEADER": 1,
|
31 |
+
"B-QUESTION": 3,
|
32 |
+
"I-ANSWER": 6,
|
33 |
+
"I-HEADER": 2,
|
34 |
+
"I-QUESTION": 4,
|
35 |
+
"O": 0
|
36 |
+
},
|
37 |
+
"layer_norm_eps": 1e-05,
|
38 |
+
"max_2d_position_embeddings": 1024,
|
39 |
+
"max_position_embeddings": 514,
|
40 |
+
"max_rel_2d_pos": 256,
|
41 |
+
"max_rel_pos": 128,
|
42 |
+
"model_type": "layoutlmv3",
|
43 |
+
"num_attention_heads": 12,
|
44 |
+
"num_channels": 3,
|
45 |
+
"num_hidden_layers": 12,
|
46 |
+
"pad_token_id": 1,
|
47 |
+
"patch_size": 16,
|
48 |
+
"rel_2d_pos_bins": 64,
|
49 |
+
"rel_pos_bins": 32,
|
50 |
+
"second_input_size": 112,
|
51 |
+
"shape_size": 128,
|
52 |
+
"text_embed": true,
|
53 |
+
"torch_dtype": "float32",
|
54 |
+
"transformers_version": "4.44.0",
|
55 |
+
"type_vocab_size": 1,
|
56 |
+
"visual_embed": true,
|
57 |
+
"vocab_size": 50265
|
58 |
+
}
|
merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7821498968442bbbb290d9f95ab619ed81c4c8b17ab488095b4c498c4ce9467a
|
3 |
+
size 501355532
|
predict_results.json
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"predict_accuracy": 0.8892307692307693,
|
3 |
+
"predict_f1": 0.6833114323258869,
|
4 |
+
"predict_loss": 0.3853023946285248,
|
5 |
+
"predict_precision": 0.6718346253229974,
|
6 |
+
"predict_recall": 0.6951871657754011,
|
7 |
+
"predict_runtime": 1.5167,
|
8 |
+
"predict_samples_per_second": 11.868,
|
9 |
+
"predict_steps_per_second": 1.319
|
10 |
+
}
|
predictions.txt
ADDED
@@ -0,0 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
header header other question question question question header header header answer answer answer header question header question header question header question answer header question
|
2 |
+
header header header header header header answer header header other other header answer header header header question question question question question question other other
|
3 |
+
header header header header header header question question question question question other answer header question header header question header question question header header question question header header question question
|
4 |
+
header header header header header header header header header header header header header header question header header question question header header header header other other other other other answer header question header header header header header header header header header header header header header header header header header header header question question question question question question question question question question question question question question question question question question question question question question question question question question question question question question question question question question question
|
5 |
+
header header header other other other question other other question question question other answer question question header other header header question header header other other
|
6 |
+
header header header header header header answer answer header header header header header header header header header question header question question question question question question question question question question question header header header answer answer header header header header header header answer answer other other question question question header question question header question header question header question question question question question question question question header header header header header header header header question header header question question question question question question question question question question header header header header header header header header header header question header question other question question answer header
|
7 |
+
other answer header header header header answer header question header question header question header question header question header question header question question question header
|
8 |
+
header header header header header question question question question question other answer header header header header header header question question header header header question header question question header header question header question header question header question
|
9 |
+
header header header header header header header header question header header header header question question other header other header header header header answer other other question question header header question question question header header question question question header header header header header header header
|
10 |
+
header header header header header header header header header header header header header header header header header header other header header header answer other header header header question question question question header question header header question question question header question other header header other header question question question
|
11 |
+
header header header question header header header header other header answer header header header question question header question question question question header question header question answer header header header header header header header header other other other
|
12 |
+
answer header header header header header header header header header header header header header header header header header header header header header header question question header header header header header question question question question question question question question other question header header header header header header header header header header header header header header question question question answer question header question header question question question header header header question header header header question header question question question header question header header header header other
|
13 |
+
other answer header header header header header header header header header header header header header header question header header header header header header question question question question question question question question question question question question header header header header header question question question question other header question other question answer question question question answer other question question question question question question question header header question question question question header header question header header header header header header header header header header header header header header header header header question question question question question question question question header header header header header header header header header header header header header answer other other other
|
14 |
+
header question header header question header header question question question header question question question question question question question other answer answer answer question question question question question question header question header question header question question question header header header header header header question question question header answer
|
15 |
+
other header header question header header header header header header header question question question header header other header header header question answer other header question question header question other header question header header header question question question header header header header question question question question header header header header question question
|
16 |
+
question question question question header answer question question header header header other question question header answer other other other header question header header header header header header other other header header header question question question header question header header header header header header header header question header question other header other question header header header answer header question question question question question question other other question question other other header other header question
|
17 |
+
header header header header header question question header question header header header question header header header header header question other other other other question header header header question other header header other answer answer other other question question question question header header header question other answer header question header question header header header header header other header header header question header question question header header question question header header header header question question answer header question header question header question other other other
|
18 |
+
header header header header header question question header header header header question other other header other other answer other question question other other question question other header question other other other header question header question header other other other other other answer header header other
|
preprocessor_config.json
ADDED
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"apply_ocr": false,
|
3 |
+
"do_normalize": true,
|
4 |
+
"do_rescale": true,
|
5 |
+
"do_resize": true,
|
6 |
+
"image_mean": [
|
7 |
+
0.5,
|
8 |
+
0.5,
|
9 |
+
0.5
|
10 |
+
],
|
11 |
+
"image_processor_type": "LayoutLMv3ImageProcessor",
|
12 |
+
"image_std": [
|
13 |
+
0.5,
|
14 |
+
0.5,
|
15 |
+
0.5
|
16 |
+
],
|
17 |
+
"ocr_lang": null,
|
18 |
+
"processor_class": "LayoutLMv3Processor",
|
19 |
+
"resample": 2,
|
20 |
+
"rescale_factor": 0.00392156862745098,
|
21 |
+
"size": {
|
22 |
+
"height": 224,
|
23 |
+
"width": 224
|
24 |
+
},
|
25 |
+
"tesseract_config": ""
|
26 |
+
}
|
special_tokens_map.json
ADDED
@@ -0,0 +1,51 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": true,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"cls_token": {
|
10 |
+
"content": "<s>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": true,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"eos_token": {
|
17 |
+
"content": "</s>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": true,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
},
|
23 |
+
"mask_token": {
|
24 |
+
"content": "<mask>",
|
25 |
+
"lstrip": true,
|
26 |
+
"normalized": true,
|
27 |
+
"rstrip": false,
|
28 |
+
"single_word": false
|
29 |
+
},
|
30 |
+
"pad_token": {
|
31 |
+
"content": "<pad>",
|
32 |
+
"lstrip": false,
|
33 |
+
"normalized": true,
|
34 |
+
"rstrip": false,
|
35 |
+
"single_word": false
|
36 |
+
},
|
37 |
+
"sep_token": {
|
38 |
+
"content": "</s>",
|
39 |
+
"lstrip": false,
|
40 |
+
"normalized": true,
|
41 |
+
"rstrip": false,
|
42 |
+
"single_word": false
|
43 |
+
},
|
44 |
+
"unk_token": {
|
45 |
+
"content": "<unk>",
|
46 |
+
"lstrip": false,
|
47 |
+
"normalized": true,
|
48 |
+
"rstrip": false,
|
49 |
+
"single_word": false
|
50 |
+
}
|
51 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,80 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_prefix_space": true,
|
3 |
+
"added_tokens_decoder": {
|
4 |
+
"0": {
|
5 |
+
"content": "<s>",
|
6 |
+
"lstrip": false,
|
7 |
+
"normalized": true,
|
8 |
+
"rstrip": false,
|
9 |
+
"single_word": false,
|
10 |
+
"special": true
|
11 |
+
},
|
12 |
+
"1": {
|
13 |
+
"content": "<pad>",
|
14 |
+
"lstrip": false,
|
15 |
+
"normalized": true,
|
16 |
+
"rstrip": false,
|
17 |
+
"single_word": false,
|
18 |
+
"special": true
|
19 |
+
},
|
20 |
+
"2": {
|
21 |
+
"content": "</s>",
|
22 |
+
"lstrip": false,
|
23 |
+
"normalized": true,
|
24 |
+
"rstrip": false,
|
25 |
+
"single_word": false,
|
26 |
+
"special": true
|
27 |
+
},
|
28 |
+
"3": {
|
29 |
+
"content": "<unk>",
|
30 |
+
"lstrip": false,
|
31 |
+
"normalized": true,
|
32 |
+
"rstrip": false,
|
33 |
+
"single_word": false,
|
34 |
+
"special": true
|
35 |
+
},
|
36 |
+
"50264": {
|
37 |
+
"content": "<mask>",
|
38 |
+
"lstrip": true,
|
39 |
+
"normalized": true,
|
40 |
+
"rstrip": false,
|
41 |
+
"single_word": false,
|
42 |
+
"special": true
|
43 |
+
}
|
44 |
+
},
|
45 |
+
"apply_ocr": false,
|
46 |
+
"bos_token": "<s>",
|
47 |
+
"clean_up_tokenization_spaces": true,
|
48 |
+
"cls_token": "<s>",
|
49 |
+
"cls_token_box": [
|
50 |
+
0,
|
51 |
+
0,
|
52 |
+
0,
|
53 |
+
0
|
54 |
+
],
|
55 |
+
"eos_token": "</s>",
|
56 |
+
"errors": "replace",
|
57 |
+
"mask_token": "<mask>",
|
58 |
+
"model_max_length": 512,
|
59 |
+
"only_label_first_subword": true,
|
60 |
+
"pad_token": "<pad>",
|
61 |
+
"pad_token_box": [
|
62 |
+
0,
|
63 |
+
0,
|
64 |
+
0,
|
65 |
+
0
|
66 |
+
],
|
67 |
+
"pad_token_label": -100,
|
68 |
+
"processor_class": "LayoutLMv3Processor",
|
69 |
+
"sep_token": "</s>",
|
70 |
+
"sep_token_box": [
|
71 |
+
0,
|
72 |
+
0,
|
73 |
+
0,
|
74 |
+
0
|
75 |
+
],
|
76 |
+
"tokenizer_class": "LayoutLMv3Tokenizer",
|
77 |
+
"trim_offsets": true,
|
78 |
+
"unk_token": "<unk>",
|
79 |
+
"use_fast": true
|
80 |
+
}
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:79a00e9e2eb4d526ef08352565518f227fb0a5281651fc78cd4ba3cb25fb2f28
|
3 |
+
size 5240
|
vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|