MikkoLipsanen commited on
Commit
9ef1847
·
1 Parent(s): 2ca67a2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -10
README.md CHANGED
@@ -86,12 +86,13 @@ Test|2414|5577|179|2445|1097|183|2838|272|374|356
86
  This model was trained using a NVIDIA RTX A6000 GPU with the following hyperparameters:
87
 
88
  - learning rate: 2e-05
89
- - train batch size: 16
90
  - epochs: 10
91
  - optimizer: AdamW with betas=(0.9,0.999) and epsilon=1e-08
92
  - scheduler: linear scheduler with num_warmup_steps=round(len(train_dataloader)/5) and num_training_steps=len(train_dataloader)*epochs
93
  - maximum length of data sequence: 512
94
  - patience: 2 epochs
 
95
 
96
  In the preprocessing stage, the input texts were split into chunks with a maximum length of 300 tokens,
97
  in order to avoid the tokenized chunks exceeding the maximum length of 512. Tokenization was performed
@@ -106,15 +107,15 @@ Evaluation results using the test dataset are listed below:
106
 
107
  ||Precision|Recall|F1-score
108
  -|-|-|-
109
- PERSON|0.91|0.91|0.91
110
- ORG|0.88|0.89|0.89
111
- LOC|0.87|0.89|0.88
112
- GPE|0.93|0.94|0.93
113
- PRODUCT|0.77|0.82|0.80
114
- EVENT|0.66|0.71|0.69
115
- DATE|0.89|0.92|0.91
116
- JON|0.78|0.83|0.80
117
- FIBC|0.88|0.94|0.69
118
  NORP|0.91|0.95|0.93
119
 
120
  The metrics were calculated using the [seqeval](https://github.com/chakki-works/seqeval) library.
 
86
  This model was trained using a NVIDIA RTX A6000 GPU with the following hyperparameters:
87
 
88
  - learning rate: 2e-05
89
+ - train batch size: 24
90
  - epochs: 10
91
  - optimizer: AdamW with betas=(0.9,0.999) and epsilon=1e-08
92
  - scheduler: linear scheduler with num_warmup_steps=round(len(train_dataloader)/5) and num_training_steps=len(train_dataloader)*epochs
93
  - maximum length of data sequence: 512
94
  - patience: 2 epochs
95
+ - classifier dropout: 0.3
96
 
97
  In the preprocessing stage, the input texts were split into chunks with a maximum length of 300 tokens,
98
  in order to avoid the tokenized chunks exceeding the maximum length of 512. Tokenization was performed
 
107
 
108
  ||Precision|Recall|F1-score
109
  -|-|-|-
110
+ PERSON|0.90|0.91|0.90
111
+ ORG|0.84|0.87|0.86
112
+ LOC|0.84|0.86|0.85
113
+ GPE|0.91|0.91|0.91
114
+ PRODUCT|0.73|0.77|0.75
115
+ EVENT|0.69|0.73|0.71
116
+ DATE|0.90|0.92|0.91
117
+ JON|0.83|0.95|0.89
118
+ FIBC|0.95|0.99|0.97
119
  NORP|0.91|0.95|0.93
120
 
121
  The metrics were calculated using the [seqeval](https://github.com/chakki-works/seqeval) library.