update generation examples and params
Browse files
README.md
CHANGED
@@ -1,9 +1,10 @@
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
tags:
|
4 |
-
-
|
5 |
-
|
6 |
-
-
|
|
|
7 |
|
8 |
widget:
|
9 |
- text: "Anna and Mike is going skiing"
|
@@ -26,24 +27,27 @@ ta ohow to remove trents in these nalitives from time series"
|
|
26 |
example_title: "dangling modifier"
|
27 |
- text: "There car broke down so their hitching a ride to they're class."
|
28 |
example_title: "compound-1"
|
|
|
|
|
29 |
|
30 |
inference:
|
31 |
parameters:
|
32 |
-
no_repeat_ngram_size:
|
33 |
max_length: 64
|
34 |
min_length: 4
|
35 |
num_beams: 4
|
36 |
-
repetition_penalty:
|
37 |
-
length_penalty:
|
38 |
early_stopping: True
|
39 |
---
|
40 |
|
41 |
|
42 |
# t5-v1_1-base-ft-jflAUG
|
43 |
|
44 |
-
|
45 |
-
|
46 |
-
-
|
|
|
47 |
|
48 |
## Model description
|
49 |
|
@@ -58,9 +62,8 @@ inference:
|
|
58 |
|
59 |
## Training and evaluation data
|
60 |
|
61 |
-
|
62 |
-
|
63 |
-
## Training procedure
|
64 |
|
65 |
### Training hyperparameters
|
66 |
|
@@ -77,9 +80,6 @@ The following hyperparameters were used during training:
|
|
77 |
- lr_scheduler_warmup_ratio: 0.05
|
78 |
- num_epochs: 5
|
79 |
|
80 |
-
### Training results
|
81 |
-
|
82 |
-
|
83 |
|
84 |
### Framework versions
|
85 |
|
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
tags:
|
4 |
+
- grammar
|
5 |
+
- spelling
|
6 |
+
- punctuation
|
7 |
+
- error-correction
|
8 |
|
9 |
widget:
|
10 |
- text: "Anna and Mike is going skiing"
|
|
|
27 |
example_title: "dangling modifier"
|
28 |
- text: "There car broke down so their hitching a ride to they're class."
|
29 |
example_title: "compound-1"
|
30 |
+
- text: "Which part of Zurich was you going to go hiking in when we were there for the first time together? ! ?"
|
31 |
+
example_title: "chatbot on Zurich"
|
32 |
|
33 |
inference:
|
34 |
parameters:
|
35 |
+
no_repeat_ngram_size: 4
|
36 |
max_length: 64
|
37 |
min_length: 4
|
38 |
num_beams: 4
|
39 |
+
repetition_penalty: 1.51
|
40 |
+
length_penalty: 1
|
41 |
early_stopping: True
|
42 |
---
|
43 |
|
44 |
|
45 |
# t5-v1_1-base-ft-jflAUG
|
46 |
|
47 |
+
> **GOAL:** a more robust and generalized grammar and spelling correction model with minimal impact on the semantics of correct sentences (i.e. it does not change things that do not need to be changed).
|
48 |
+
|
49 |
+
- this model _(at least from preliminary testing)_ can handle large amounts of errors in the source text (i.e. from audio transcription) and still produce cohesive results.
|
50 |
+
- a fine-tuned version of [google/t5-v1_1-base](https://huggingface.co/google/t5-v1_1-base) on an expanded version of the [JFLEG dataset](https://aclanthology.org/E17-2037/).
|
51 |
|
52 |
## Model description
|
53 |
|
|
|
62 |
|
63 |
## Training and evaluation data
|
64 |
|
65 |
+
- trained as text-to-text
|
66 |
+
- JFLEG dataset + additional selected and/or generated grammar corrections
|
|
|
67 |
|
68 |
### Training hyperparameters
|
69 |
|
|
|
80 |
- lr_scheduler_warmup_ratio: 0.05
|
81 |
- num_epochs: 5
|
82 |
|
|
|
|
|
|
|
83 |
|
84 |
### Framework versions
|
85 |
|