David Kagramanyan
commited on
Commit
•
1772cef
1
Parent(s):
df23d26
initial
Browse files- README.md +134 -0
- deploy_endpoint_dix.ipynb +168 -0
- handler.py +46 -0
- loss.tsv +151 -0
- pytorch_model.bin +3 -0
- test.tsv +0 -0
- training.log +0 -0
README.md
ADDED
@@ -0,0 +1,134 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
tags:
|
3 |
+
- flair
|
4 |
+
- token-classification
|
5 |
+
- sequence-tagger-model
|
6 |
+
language: nl
|
7 |
+
datasets:
|
8 |
+
- conll2003
|
9 |
+
widget:
|
10 |
+
- text: "George Washington ging naar Washington."
|
11 |
+
---
|
12 |
+
|
13 |
+
# Dutch NER in Flair (default model)
|
14 |
+
|
15 |
+
This is the standard 4-class NER model for Dutch that ships with [Flair](https://github.com/flairNLP/flair/).
|
16 |
+
|
17 |
+
F1-Score: **92,58** (CoNLL-03)
|
18 |
+
|
19 |
+
Predicts 4 tags:
|
20 |
+
|
21 |
+
| **tag** | **meaning** |
|
22 |
+
|---------------------------------|-----------|
|
23 |
+
| PER | person name |
|
24 |
+
| LOC | location name |
|
25 |
+
| ORG | organization name |
|
26 |
+
| MISC | other name |
|
27 |
+
|
28 |
+
Based on Transformer embeddings and LSTM-CRF.
|
29 |
+
|
30 |
+
---
|
31 |
+
# Demo: How to use in Flair
|
32 |
+
|
33 |
+
Requires: **[Flair](https://github.com/flairNLP/flair/)** (`pip install flair`)
|
34 |
+
|
35 |
+
```python
|
36 |
+
from flair.data import Sentence
|
37 |
+
from flair.models import SequenceTagger
|
38 |
+
|
39 |
+
# load tagger
|
40 |
+
tagger = SequenceTagger.load("flair/ner-dutch")
|
41 |
+
|
42 |
+
# make example sentence
|
43 |
+
sentence = Sentence("George Washington ging naar Washington")
|
44 |
+
|
45 |
+
# predict NER tags
|
46 |
+
tagger.predict(sentence)
|
47 |
+
|
48 |
+
# print sentence
|
49 |
+
print(sentence)
|
50 |
+
|
51 |
+
# print predicted NER spans
|
52 |
+
print('The following NER tags are found:')
|
53 |
+
# iterate over entities and print
|
54 |
+
for entity in sentence.get_spans('ner'):
|
55 |
+
print(entity)
|
56 |
+
|
57 |
+
```
|
58 |
+
|
59 |
+
This yields the following output:
|
60 |
+
```
|
61 |
+
Span [1,2]: "George Washington" [− Labels: PER (0.997)]
|
62 |
+
Span [5]: "Washington" [− Labels: LOC (0.9996)]
|
63 |
+
```
|
64 |
+
|
65 |
+
So, the entities "*George Washington*" (labeled as a **person**) and "*Washington*" (labeled as a **location**) are found in the sentence "*George Washington ging naar Washington*".
|
66 |
+
|
67 |
+
|
68 |
+
---
|
69 |
+
|
70 |
+
### Training: Script to train this model
|
71 |
+
|
72 |
+
The following Flair script was used to train this model:
|
73 |
+
|
74 |
+
```python
|
75 |
+
from flair.data import Corpus
|
76 |
+
from flair.datasets import CONLL_03_DUTCH
|
77 |
+
from flair.embeddings import WordEmbeddings, StackedEmbeddings, FlairEmbeddings
|
78 |
+
|
79 |
+
|
80 |
+
# 1. get the corpus
|
81 |
+
corpus: Corpus = CONLL_03_DUTCH()
|
82 |
+
|
83 |
+
# 2. what tag do we want to predict?
|
84 |
+
tag_type = 'ner'
|
85 |
+
|
86 |
+
# 3. make the tag dictionary from the corpus
|
87 |
+
tag_dictionary = corpus.make_tag_dictionary(tag_type=tag_type)
|
88 |
+
|
89 |
+
# 4. initialize embeddings
|
90 |
+
embeddings = TransformerWordEmbeddings('wietsedv/bert-base-dutch-cased')
|
91 |
+
|
92 |
+
# 5. initialize sequence tagger
|
93 |
+
tagger: SequenceTagger = SequenceTagger(hidden_size=256,
|
94 |
+
embeddings=embeddings,
|
95 |
+
tag_dictionary=tag_dictionary,
|
96 |
+
tag_type=tag_type)
|
97 |
+
|
98 |
+
# 6. initialize trainer
|
99 |
+
trainer: ModelTrainer = ModelTrainer(tagger, corpus)
|
100 |
+
|
101 |
+
# 7. run training
|
102 |
+
trainer.train('resources/taggers/ner-dutch',
|
103 |
+
train_with_dev=True,
|
104 |
+
max_epochs=150)
|
105 |
+
```
|
106 |
+
|
107 |
+
|
108 |
+
---
|
109 |
+
|
110 |
+
### Cite
|
111 |
+
|
112 |
+
Please cite the following paper when using this model.
|
113 |
+
|
114 |
+
```
|
115 |
+
@inproceedings{akbik-etal-2019-flair,
|
116 |
+
title = "{FLAIR}: An Easy-to-Use Framework for State-of-the-Art {NLP}",
|
117 |
+
author = "Akbik, Alan and
|
118 |
+
Bergmann, Tanja and
|
119 |
+
Blythe, Duncan and
|
120 |
+
Rasul, Kashif and
|
121 |
+
Schweter, Stefan and
|
122 |
+
Vollgraf, Roland",
|
123 |
+
booktitle = "Proceedings of the 2019 Conference of the North {A}merican Chapter of the Association for Computational Linguistics (Demonstrations)",
|
124 |
+
year = "2019",
|
125 |
+
url = "https://www.aclweb.org/anthology/N19-4010",
|
126 |
+
pages = "54--59",
|
127 |
+
}
|
128 |
+
```
|
129 |
+
|
130 |
+
---
|
131 |
+
|
132 |
+
### Issues?
|
133 |
+
|
134 |
+
The Flair issue tracker is available [here](https://github.com/flairNLP/flair/issues/).
|
deploy_endpoint_dix.ipynb
ADDED
@@ -0,0 +1,168 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"cells": [
|
3 |
+
{
|
4 |
+
"cell_type": "code",
|
5 |
+
"execution_count": 2,
|
6 |
+
"id": "initial_id",
|
7 |
+
"metadata": {
|
8 |
+
"collapsed": true,
|
9 |
+
"ExecuteTime": {
|
10 |
+
"end_time": "2023-10-05T07:20:29.202015200Z",
|
11 |
+
"start_time": "2023-10-05T07:20:29.190080700Z"
|
12 |
+
}
|
13 |
+
},
|
14 |
+
"outputs": [
|
15 |
+
{
|
16 |
+
"name": "stdout",
|
17 |
+
"output_type": "stream",
|
18 |
+
"text": [
|
19 |
+
"True\n",
|
20 |
+
"0\n",
|
21 |
+
"<torch.cuda.device object at 0x0000028DAB1DB580>\n",
|
22 |
+
"1\n",
|
23 |
+
"NVIDIA GeForce RTX 3090\n"
|
24 |
+
]
|
25 |
+
}
|
26 |
+
],
|
27 |
+
"source": [
|
28 |
+
"import torch\n",
|
29 |
+
"\n",
|
30 |
+
"print(torch.cuda.is_available()) # Returns a bool indicating if CUDA is currently available.\n",
|
31 |
+
"print(torch.cuda.current_device()) # Returns the index of a currently selected device.\n",
|
32 |
+
"print(torch.cuda.device(0)) # Context-manager that changes the selected device.\n",
|
33 |
+
"print(torch.cuda.device_count()) # Returns the number of GPUs available.\n",
|
34 |
+
"print(torch.cuda.get_device_name(0)) # Gets the name of a device."
|
35 |
+
]
|
36 |
+
},
|
37 |
+
{
|
38 |
+
"cell_type": "code",
|
39 |
+
"execution_count": 5,
|
40 |
+
"outputs": [
|
41 |
+
{
|
42 |
+
"name": "stdout",
|
43 |
+
"output_type": "stream",
|
44 |
+
"text": [
|
45 |
+
"2023-10-05 10:37:05,350 SequenceTagger predicts: Dictionary with 20 tags: <unk>, O, S-ORG, S-MISC, B-PER, E-PER, S-PER, S-LOC, B-MISC, E-MISC, B-ORG, E-ORG, I-ORG, I-PER, B-LOC, I-LOC, E-LOC, I-MISC, <START>, <STOP>\n",
|
46 |
+
"non_holiday_pred [{'entity_group': 'PER', 'word': 'George Washington', 'start': 0, 'end': 17, 'score': 0.9970293045043945}, {'entity_group': 'LOC', 'word': 'Washington', 'start': 28, 'end': 38, 'score': 0.9996309280395508}]\n"
|
47 |
+
]
|
48 |
+
}
|
49 |
+
],
|
50 |
+
"source": [
|
51 |
+
"from handler import EndpointHandler\n",
|
52 |
+
"\n",
|
53 |
+
"# init handler\n",
|
54 |
+
"my_handler = EndpointHandler(path=\".\")\n",
|
55 |
+
"\n",
|
56 |
+
"# prepare sample payload\n",
|
57 |
+
"non_holiday_payload = {\"inputs\": \"George Washington ging naar Washington\"}\n",
|
58 |
+
"\n",
|
59 |
+
"\n",
|
60 |
+
"# test the handler\n",
|
61 |
+
"non_holiday_pred=my_handler(non_holiday_payload)\n",
|
62 |
+
"\n",
|
63 |
+
"\n",
|
64 |
+
"# show results\n",
|
65 |
+
"print(\"non_holiday_pred\", non_holiday_pred)\n",
|
66 |
+
"\n"
|
67 |
+
],
|
68 |
+
"metadata": {
|
69 |
+
"collapsed": false,
|
70 |
+
"ExecuteTime": {
|
71 |
+
"end_time": "2023-10-05T07:37:05.789680900Z",
|
72 |
+
"start_time": "2023-10-05T07:37:03.091564500Z"
|
73 |
+
}
|
74 |
+
},
|
75 |
+
"id": "a12c4a4792afc707"
|
76 |
+
},
|
77 |
+
{
|
78 |
+
"cell_type": "code",
|
79 |
+
"execution_count": 3,
|
80 |
+
"outputs": [],
|
81 |
+
"source": [
|
82 |
+
"from typing import Any, Dict, List\n",
|
83 |
+
"import os\n",
|
84 |
+
"from flair.data import Sentence\n",
|
85 |
+
"from flair.models import SequenceTagger"
|
86 |
+
],
|
87 |
+
"metadata": {
|
88 |
+
"collapsed": false,
|
89 |
+
"ExecuteTime": {
|
90 |
+
"end_time": "2023-10-05T07:36:53.389033800Z",
|
91 |
+
"start_time": "2023-10-05T07:36:53.382053200Z"
|
92 |
+
}
|
93 |
+
},
|
94 |
+
"id": "f411919d7d047065"
|
95 |
+
},
|
96 |
+
{
|
97 |
+
"cell_type": "code",
|
98 |
+
"execution_count": 4,
|
99 |
+
"outputs": [
|
100 |
+
{
|
101 |
+
"name": "stdout",
|
102 |
+
"output_type": "stream",
|
103 |
+
"text": [
|
104 |
+
"2023-10-05 10:36:58,072 SequenceTagger predicts: Dictionary with 20 tags: <unk>, O, S-ORG, S-MISC, B-PER, E-PER, S-PER, S-LOC, B-MISC, E-MISC, B-ORG, E-ORG, I-ORG, I-PER, B-LOC, I-LOC, E-LOC, I-MISC, <START>, <STOP>\n"
|
105 |
+
]
|
106 |
+
}
|
107 |
+
],
|
108 |
+
"source": [
|
109 |
+
"tagger = SequenceTagger.load('pytorch_model.bin')"
|
110 |
+
],
|
111 |
+
"metadata": {
|
112 |
+
"collapsed": false,
|
113 |
+
"ExecuteTime": {
|
114 |
+
"end_time": "2023-10-05T07:36:59.846440300Z",
|
115 |
+
"start_time": "2023-10-05T07:36:54.140093700Z"
|
116 |
+
}
|
117 |
+
},
|
118 |
+
"id": "8f497b3807de2e1"
|
119 |
+
},
|
120 |
+
{
|
121 |
+
"cell_type": "code",
|
122 |
+
"execution_count": 12,
|
123 |
+
"outputs": [
|
124 |
+
{
|
125 |
+
"data": {
|
126 |
+
"text/plain": "'0.12.2'"
|
127 |
+
},
|
128 |
+
"execution_count": 12,
|
129 |
+
"metadata": {},
|
130 |
+
"output_type": "execute_result"
|
131 |
+
}
|
132 |
+
],
|
133 |
+
"source": [
|
134 |
+
"import flair\n",
|
135 |
+
"flair.__version__"
|
136 |
+
],
|
137 |
+
"metadata": {
|
138 |
+
"collapsed": false,
|
139 |
+
"ExecuteTime": {
|
140 |
+
"end_time": "2023-10-05T07:36:37.788428800Z",
|
141 |
+
"start_time": "2023-10-05T07:36:37.754490Z"
|
142 |
+
}
|
143 |
+
},
|
144 |
+
"id": "df243c485fd370b"
|
145 |
+
}
|
146 |
+
],
|
147 |
+
"metadata": {
|
148 |
+
"kernelspec": {
|
149 |
+
"name": "torch",
|
150 |
+
"language": "python",
|
151 |
+
"display_name": "torch"
|
152 |
+
},
|
153 |
+
"language_info": {
|
154 |
+
"codemirror_mode": {
|
155 |
+
"name": "ipython",
|
156 |
+
"version": 2
|
157 |
+
},
|
158 |
+
"file_extension": ".py",
|
159 |
+
"mimetype": "text/x-python",
|
160 |
+
"name": "python",
|
161 |
+
"nbconvert_exporter": "python",
|
162 |
+
"pygments_lexer": "ipython2",
|
163 |
+
"version": "2.7.6"
|
164 |
+
}
|
165 |
+
},
|
166 |
+
"nbformat": 4,
|
167 |
+
"nbformat_minor": 5
|
168 |
+
}
|
handler.py
ADDED
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
from typing import Any, Dict, List
|
2 |
+
import os
|
3 |
+
from flair.data import Sentence
|
4 |
+
from flair.models import SequenceTagger
|
5 |
+
|
6 |
+
class EndpointHandler():
|
7 |
+
def __init__(
|
8 |
+
self,
|
9 |
+
path: str,
|
10 |
+
):
|
11 |
+
self.tagger = SequenceTagger.load(os.path.join(path,"pytorch_model.bin"))
|
12 |
+
|
13 |
+
def __call__(self, data: Dict[str, Any]) -> List[Dict[str, Any]]:
|
14 |
+
"""
|
15 |
+
Args:
|
16 |
+
inputs (:obj:`str`):
|
17 |
+
a string containing some text
|
18 |
+
Return:
|
19 |
+
A :obj:`list`:. The object returned should be like [{"entity_group": "XXX", "word": "some word", "start": 3, "end": 6, "score": 0.82}] containing :
|
20 |
+
- "entity_group": A string representing what the entity is.
|
21 |
+
- "word": A substring of the original string that was detected as an entity.
|
22 |
+
- "start": the offset within `input` leading to `answer`. context[start:stop] == word
|
23 |
+
- "end": the ending offset within `input` leading to `answer`. context[start:stop] === word
|
24 |
+
- "score": A score between 0 and 1 describing how confident the model is for this entity.
|
25 |
+
"""
|
26 |
+
inputs = data.pop("inputs", data)
|
27 |
+
sentence: Sentence = Sentence(inputs)
|
28 |
+
|
29 |
+
# Also show scores for recognized NEs
|
30 |
+
self.tagger.predict(sentence, label_name="predicted")
|
31 |
+
|
32 |
+
entities = []
|
33 |
+
for span in sentence.get_spans("predicted"):
|
34 |
+
if len(span.tokens) == 0:
|
35 |
+
continue
|
36 |
+
current_entity = {
|
37 |
+
"entity_group": span.tag,
|
38 |
+
"word": span.text,
|
39 |
+
"start": span.tokens[0].start_position,
|
40 |
+
"end": span.tokens[-1].end_position,
|
41 |
+
"score": span.score,
|
42 |
+
}
|
43 |
+
|
44 |
+
entities.append(current_entity)
|
45 |
+
|
46 |
+
return entities
|
loss.tsv
ADDED
@@ -0,0 +1,151 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
EPOCH TIMESTAMP BAD_EPOCHS LEARNING_RATE TRAIN_LOSS
|
2 |
+
1 14:01:30 0 0.1000 2.779733479175812
|
3 |
+
2 14:03:00 0 0.1000 1.3395338043188438
|
4 |
+
3 14:04:30 0 0.1000 1.0816994036874201
|
5 |
+
4 14:06:01 0 0.1000 0.9621152551255674
|
6 |
+
5 14:07:31 0 0.1000 0.8333948618835874
|
7 |
+
6 14:09:01 0 0.1000 0.7716772701495733
|
8 |
+
7 14:10:31 0 0.1000 0.7374578891911059
|
9 |
+
8 14:12:01 0 0.1000 0.6992316849211342
|
10 |
+
9 14:13:31 0 0.1000 0.6452770539328583
|
11 |
+
10 14:15:01 1 0.1000 0.6467076227960423
|
12 |
+
11 14:16:31 0 0.1000 0.620160369358511
|
13 |
+
12 14:18:01 0 0.1000 0.5951391830658301
|
14 |
+
13 14:19:31 0 0.1000 0.5727337152020544
|
15 |
+
14 14:21:01 0 0.1000 0.5662796683163724
|
16 |
+
15 14:22:32 0 0.1000 0.5356569104979181
|
17 |
+
16 14:24:01 0 0.1000 0.5295078084764318
|
18 |
+
17 14:25:31 0 0.1000 0.5199492649644868
|
19 |
+
18 14:27:01 0 0.1000 0.5174320767832618
|
20 |
+
19 14:28:31 0 0.1000 0.4972612695816236
|
21 |
+
20 14:30:01 0 0.1000 0.4858957131194253
|
22 |
+
21 14:31:30 1 0.1000 0.4954718519734521
|
23 |
+
22 14:33:00 0 0.1000 0.475096299760362
|
24 |
+
23 14:34:29 0 0.1000 0.45335076690739035
|
25 |
+
24 14:35:58 0 0.1000 0.45071878243460617
|
26 |
+
25 14:37:28 0 0.1000 0.4470571254053686
|
27 |
+
26 14:38:57 0 0.1000 0.4351186317511094
|
28 |
+
27 14:40:26 1 0.1000 0.435719607044489
|
29 |
+
28 14:41:55 0 0.1000 0.43325941297743054
|
30 |
+
29 14:43:24 0 0.1000 0.4213554557444703
|
31 |
+
30 14:44:54 0 0.1000 0.41895739234920243
|
32 |
+
31 14:46:23 1 0.1000 0.424293908655134
|
33 |
+
32 14:47:53 0 0.1000 0.41112876903806994
|
34 |
+
33 14:49:22 0 0.1000 0.4081363928368968
|
35 |
+
34 14:50:52 0 0.1000 0.3930869156988258
|
36 |
+
35 14:52:21 1 0.1000 0.4032204799927198
|
37 |
+
36 14:53:50 0 0.1000 0.3917345738818503
|
38 |
+
37 14:55:19 1 0.1000 0.4003689407920226
|
39 |
+
38 14:56:48 0 0.1000 0.385329091319671
|
40 |
+
39 14:58:17 0 0.1000 0.38347142062380785
|
41 |
+
40 14:59:46 0 0.1000 0.3804049695515607
|
42 |
+
41 15:01:14 0 0.1000 0.3764326793770505
|
43 |
+
42 15:02:43 0 0.1000 0.3761323136142176
|
44 |
+
43 15:04:12 1 0.1000 0.3851716780764425
|
45 |
+
44 15:05:41 2 0.1000 0.3771476393326735
|
46 |
+
45 15:07:11 0 0.1000 0.3616421103222757
|
47 |
+
46 15:08:40 1 0.1000 0.3686442003800319
|
48 |
+
47 15:10:09 0 0.1000 0.35834919491894224
|
49 |
+
48 15:11:38 1 0.1000 0.3613178371618956
|
50 |
+
49 15:13:07 0 0.1000 0.3519833675561807
|
51 |
+
50 15:14:36 1 0.1000 0.35567070319611804
|
52 |
+
51 15:16:05 0 0.1000 0.34542505874847756
|
53 |
+
52 15:17:34 1 0.1000 0.34995765023761327
|
54 |
+
53 15:19:03 0 0.1000 0.3352116090110224
|
55 |
+
54 15:20:32 0 0.1000 0.3264291577868991
|
56 |
+
55 15:22:01 1 0.1000 0.3271228834222525
|
57 |
+
56 15:23:30 2 0.1000 0.33473273961462524
|
58 |
+
57 15:25:00 3 0.1000 0.3365086276052345
|
59 |
+
58 15:26:26 4 0.1000 0.33411403429559156
|
60 |
+
59 15:27:48 0 0.0500 0.30620485559487953
|
61 |
+
60 15:29:11 0 0.0500 0.28616658293793346
|
62 |
+
61 15:30:37 0 0.0500 0.2821873968674077
|
63 |
+
62 15:32:02 0 0.0500 0.26961317198653506
|
64 |
+
63 15:33:25 0 0.0500 0.2660407587249055
|
65 |
+
64 15:34:48 0 0.0500 0.2553254587782754
|
66 |
+
65 15:36:12 1 0.0500 0.25559193193912505
|
67 |
+
66 15:37:36 0 0.0500 0.24891968863642114
|
68 |
+
67 15:38:59 1 0.0500 0.2530737343761656
|
69 |
+
68 15:40:22 0 0.0500 0.2413989709992694
|
70 |
+
69 15:41:45 1 0.0500 0.24797574456176188
|
71 |
+
70 15:43:08 2 0.0500 0.242591819476782
|
72 |
+
71 15:44:30 0 0.0500 0.23441884385851713
|
73 |
+
72 15:45:54 1 0.0500 0.23821192036072414
|
74 |
+
73 15:47:16 0 0.0500 0.23189536440066802
|
75 |
+
74 15:48:39 0 0.0500 0.23109626505109998
|
76 |
+
75 15:50:02 0 0.0500 0.21557108603545233
|
77 |
+
76 15:51:24 1 0.0500 0.21703473255675063
|
78 |
+
77 15:52:47 2 0.0500 0.21980291849527603
|
79 |
+
78 15:54:09 3 0.0500 0.21841833248225032
|
80 |
+
79 15:55:32 0 0.0500 0.21489991823322752
|
81 |
+
80 15:56:55 0 0.0500 0.20688325222740825
|
82 |
+
81 15:58:17 1 0.0500 0.22208200507184378
|
83 |
+
82 15:59:40 2 0.0500 0.21044851431989262
|
84 |
+
83 16:01:03 0 0.0500 0.2044361765568073
|
85 |
+
84 16:02:25 0 0.0500 0.20330230971941582
|
86 |
+
85 16:03:48 1 0.0500 0.2057137251027629
|
87 |
+
86 16:05:11 0 0.0500 0.1991372063373908
|
88 |
+
87 16:06:34 1 0.0500 0.2024117647136888
|
89 |
+
88 16:07:57 2 0.0500 0.20274029105392277
|
90 |
+
89 16:09:20 0 0.0500 0.19314180322819285
|
91 |
+
90 16:10:43 0 0.0500 0.18950988753483847
|
92 |
+
91 16:12:06 1 0.0500 0.19186331494432737
|
93 |
+
92 16:13:29 2 0.0500 0.19439757827462423
|
94 |
+
93 16:14:52 0 0.0500 0.18818266501284053
|
95 |
+
94 16:16:15 0 0.0500 0.17917947142552107
|
96 |
+
95 16:17:38 1 0.0500 0.18487144104945355
|
97 |
+
96 16:19:01 2 0.0500 0.18518399219227652
|
98 |
+
97 16:20:25 3 0.0500 0.18472332790111884
|
99 |
+
98 16:21:48 0 0.0500 0.1754296123706855
|
100 |
+
99 16:23:11 1 0.0500 0.18395276073621125
|
101 |
+
100 16:24:33 2 0.0500 0.18039520557364846
|
102 |
+
101 16:25:57 3 0.0500 0.18199163737077997
|
103 |
+
102 16:27:20 0 0.0500 0.17518249248337542
|
104 |
+
103 16:28:42 0 0.0500 0.17339555085596875
|
105 |
+
104 16:30:05 1 0.0500 0.17634159061643812
|
106 |
+
105 16:31:29 2 0.0500 0.17359274017473317
|
107 |
+
106 16:32:52 0 0.0500 0.17105355417817578
|
108 |
+
107 16:34:15 0 0.0500 0.1707773297260969
|
109 |
+
108 16:35:38 0 0.0500 0.1698240860277771
|
110 |
+
109 16:37:01 0 0.0500 0.16659483982202333
|
111 |
+
110 16:38:24 0 0.0500 0.16050158162784373
|
112 |
+
111 16:39:47 1 0.0500 0.1668573046596641
|
113 |
+
112 16:41:09 0 0.0500 0.1591071498572317
|
114 |
+
113 16:42:32 1 0.0500 0.16081542165106177
|
115 |
+
114 16:43:55 2 0.0500 0.1627739645445194
|
116 |
+
115 16:45:18 3 0.0500 0.16070217188352193
|
117 |
+
116 16:46:41 4 0.0500 0.1620385132284246
|
118 |
+
117 16:48:04 0 0.0250 0.14860385387626468
|
119 |
+
118 16:49:27 0 0.0250 0.13816803481079573
|
120 |
+
119 16:50:49 0 0.0250 0.13735185179135037
|
121 |
+
120 16:52:13 0 0.0250 0.13719079037252654
|
122 |
+
121 16:53:35 0 0.0250 0.13567532561401016
|
123 |
+
122 16:54:57 1 0.0250 0.1358587311501177
|
124 |
+
123 16:56:20 0 0.0250 0.1323100252618265
|
125 |
+
124 16:57:43 1 0.0250 0.1347226249356555
|
126 |
+
125 16:59:05 0 0.0250 0.12610901894732418
|
127 |
+
126 17:00:28 0 0.0250 0.12457475662231446
|
128 |
+
127 17:01:51 0 0.0250 0.12143644379754352
|
129 |
+
128 17:03:14 1 0.0250 0.12776582111150792
|
130 |
+
129 17:04:37 2 0.0250 0.12849602556636192
|
131 |
+
130 17:06:00 3 0.0250 0.1244494532290687
|
132 |
+
131 17:07:22 4 0.0250 0.12194784156277648
|
133 |
+
132 17:08:45 0 0.0125 0.12072199874708796
|
134 |
+
133 17:10:07 0 0.0125 0.11447634765735039
|
135 |
+
134 17:11:29 1 0.0125 0.11593401964403625
|
136 |
+
135 17:12:52 2 0.0125 0.11731710933809543
|
137 |
+
136 17:14:14 0 0.0125 0.11113526996137559
|
138 |
+
137 17:15:37 1 0.0125 0.11634638169382372
|
139 |
+
138 17:17:00 2 0.0125 0.11969016508286835
|
140 |
+
139 17:18:23 0 0.0125 0.11103729783820036
|
141 |
+
140 17:19:45 1 0.0125 0.11450310367294866
|
142 |
+
141 17:21:08 2 0.0125 0.11406106056056471
|
143 |
+
142 17:22:31 3 0.0125 0.1144172125010409
|
144 |
+
143 17:23:54 0 0.0125 0.10564635207510402
|
145 |
+
144 17:25:16 0 0.0125 0.1036626111684192
|
146 |
+
145 17:26:39 1 0.0125 0.10585466057826311
|
147 |
+
146 17:28:02 2 0.0125 0.10889074587159686
|
148 |
+
147 17:29:24 3 0.0125 0.11062939536089125
|
149 |
+
148 17:30:47 4 0.0125 0.10591043005896429
|
150 |
+
149 17:32:10 1 0.0063 0.1120014336748192
|
151 |
+
150 17:33:33 0 0.0063 0.1027476362310923
|
pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d1daa8c197b464896aa5d1aafb5abd332f821d3ab80ab83a4d6b0c4ae819e4a2
|
3 |
+
size 501524558
|
test.tsv
ADDED
The diff for this file is too large to render.
See raw diff
|
|
training.log
ADDED
The diff for this file is too large to render.
See raw diff
|
|