pszemraj commited on
Commit
d9ffec9
1 Parent(s): 9282a3e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -0
README.md CHANGED
@@ -14,6 +14,9 @@ pipeline_tag: text2text-generation
14
  This is a basic text-to-text "instruct" model, similar to Google's original [flan-t5](https://huggingface.co/collections/google/flan-t5-release-65005c39e3201fff885e22fb) model series (but not trained for as long).
15
 
16
 
 
 
 
17
  Fine-tuned from [the base model](https://hf.co/pszemraj/tFINE-900m-e16-d32-1024ctx) on the `pszemraj/flan-subsets-deduped` dataset, subset `flan-v2` for 1 epoch. It achieves the following results on the evaluation set:
18
  - Loss: 1.4134
19
  - Rouge1: 62.9142
@@ -23,6 +26,13 @@ Fine-tuned from [the base model](https://hf.co/pszemraj/tFINE-900m-e16-d32-1024c
23
  - Gen Len: 12.0586
24
  - Num Input Tokens Seen: 1931815668
25
 
 
 
 
 
 
 
 
26
  ## Usage Example
27
 
28
  ```py
 
14
  This is a basic text-to-text "instruct" model, similar to Google's original [flan-t5](https://huggingface.co/collections/google/flan-t5-release-65005c39e3201fff885e22fb) model series (but not trained for as long).
15
 
16
 
17
+ <details>
18
+ <summary>Details: Click here to expand</summary>
19
+
20
  Fine-tuned from [the base model](https://hf.co/pszemraj/tFINE-900m-e16-d32-1024ctx) on the `pszemraj/flan-subsets-deduped` dataset, subset `flan-v2` for 1 epoch. It achieves the following results on the evaluation set:
21
  - Loss: 1.4134
22
  - Rouge1: 62.9142
 
26
  - Gen Len: 12.0586
27
  - Num Input Tokens Seen: 1931815668
28
 
29
+ ### Model features
30
+
31
+ - pretrained & fine-tuned at 1024 context length (input)
32
+ - tokenizer with byte-pair fallback to support understanding and generating text beyond what the original T5 tokenizer does
33
+
34
+ </details>
35
+
36
  ## Usage Example
37
 
38
  ```py