Update README.md
Browse files
README.md
CHANGED
@@ -14,6 +14,9 @@ pipeline_tag: text2text-generation
|
|
14 |
This is a basic text-to-text "instruct" model, similar to Google's original [flan-t5](https://huggingface.co/collections/google/flan-t5-release-65005c39e3201fff885e22fb) model series (but not trained for as long).
|
15 |
|
16 |
|
|
|
|
|
|
|
17 |
Fine-tuned from [the base model](https://hf.co/pszemraj/tFINE-900m-e16-d32-1024ctx) on the `pszemraj/flan-subsets-deduped` dataset, subset `flan-v2` for 1 epoch. It achieves the following results on the evaluation set:
|
18 |
- Loss: 1.4134
|
19 |
- Rouge1: 62.9142
|
@@ -23,6 +26,13 @@ Fine-tuned from [the base model](https://hf.co/pszemraj/tFINE-900m-e16-d32-1024c
|
|
23 |
- Gen Len: 12.0586
|
24 |
- Num Input Tokens Seen: 1931815668
|
25 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
26 |
## Usage Example
|
27 |
|
28 |
```py
|
|
|
14 |
This is a basic text-to-text "instruct" model, similar to Google's original [flan-t5](https://huggingface.co/collections/google/flan-t5-release-65005c39e3201fff885e22fb) model series (but not trained for as long).
|
15 |
|
16 |
|
17 |
+
<details>
|
18 |
+
<summary>Details: Click here to expand</summary>
|
19 |
+
|
20 |
Fine-tuned from [the base model](https://hf.co/pszemraj/tFINE-900m-e16-d32-1024ctx) on the `pszemraj/flan-subsets-deduped` dataset, subset `flan-v2` for 1 epoch. It achieves the following results on the evaluation set:
|
21 |
- Loss: 1.4134
|
22 |
- Rouge1: 62.9142
|
|
|
26 |
- Gen Len: 12.0586
|
27 |
- Num Input Tokens Seen: 1931815668
|
28 |
|
29 |
+
### Model features
|
30 |
+
|
31 |
+
- pretrained & fine-tuned at 1024 context length (input)
|
32 |
+
- tokenizer with byte-pair fallback to support understanding and generating text beyond what the original T5 tokenizer does
|
33 |
+
|
34 |
+
</details>
|
35 |
+
|
36 |
## Usage Example
|
37 |
|
38 |
```py
|