Teja-Gollapudi
commited on
Commit
·
15b763c
1
Parent(s):
b093c27
Update README.md
Browse files
README.md
CHANGED
@@ -16,13 +16,23 @@ Instruction-tuned version of SalesForce/Xgen-7b-8k-base. The model is open for <
|
|
16 |
|
17 |
We expanded Open-instruct with additional commercially viable zero-shot COT datasets from Flan v2 to total of 140k instruct-prompt responses. <br>
|
18 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
19 |
|
20 |
-
|
21 |
|
22 |
-
Open-instruct
|
|
|
|
|
23 |
- Mosaic/Dolly-HHRLHF + filtered OASST1 - cc by 3.0
|
24 |
|
25 |
-
Subset of COT SUBMIX (FROM FLAN V2) Zeroshot examples
|
26 |
- ESNLI - MIT
|
27 |
- ECQA - CDLA 1.0 - Sharing
|
28 |
- Strategy - MIT
|
@@ -31,18 +41,6 @@ Subset of COT SUBMIX (FROM FLAN V2) Zeroshot examples
|
|
31 |
- aqua - MIT
|
32 |
- qasc - Apache 2.0
|
33 |
|
34 |
-
<br>
|
35 |
-
|
36 |
-
The model supports up to <b>8192 tokens </b>
|
37 |
-
|
38 |
-
|
39 |
-
|
40 |
-
## License
|
41 |
-
- <b>Commercially Viable </b>
|
42 |
-
- The instruction datasets used for instruction tuning are open for commercial usage.
|
43 |
-
- Language Model, ([Salesforce/xgen-7b-8k-base](https://huggingface.co/Salesforce/xgen-7b-8k-base)) is under apache-2.0
|
44 |
-
- Dataset ([VMware/open-instruct](https://huggingface.co/datasets/VMware/open-instruct)) is under cc-by-sa-3.0
|
45 |
-
|
46 |
|
47 |
|
48 |
## Use in Transformers
|
|
|
16 |
|
17 |
We expanded Open-instruct with additional commercially viable zero-shot COT datasets from Flan v2 to total of 140k instruct-prompt responses. <br>
|
18 |
|
19 |
+
The model supports up to <b>8192 tokens </b>
|
20 |
+
|
21 |
+
## License
|
22 |
+
- <b>Commercially Viable </b>
|
23 |
+
- The instruction datasets used for instruction tuning are open for commercial usage.
|
24 |
+
- Language Model, ([Salesforce/xgen-7b-8k-base](https://huggingface.co/Salesforce/xgen-7b-8k-base)) is under apache-2.0
|
25 |
+
- Dataset ([VMware/open-instruct](https://huggingface.co/datasets/VMware/open-instruct)) is under cc-by-sa-3.0
|
26 |
+
|
27 |
|
28 |
+
## Datasets used for Fine-Tuning
|
29 |
|
30 |
+
### Open-instruct
|
31 |
+
|
32 |
+
**Open-instruct-v1**
|
33 |
- Mosaic/Dolly-HHRLHF + filtered OASST1 - cc by 3.0
|
34 |
|
35 |
+
**Subset of COT SUBMIX (FROM FLAN V2) Zeroshot examples**
|
36 |
- ESNLI - MIT
|
37 |
- ECQA - CDLA 1.0 - Sharing
|
38 |
- Strategy - MIT
|
|
|
41 |
- aqua - MIT
|
42 |
- qasc - Apache 2.0
|
43 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
44 |
|
45 |
|
46 |
## Use in Transformers
|