Update README.md
Browse files
README.md
CHANGED
@@ -8,13 +8,12 @@ language:
|
|
8 |
|
9 |
WangChanLion is a Multilingual, instruction-finetuned on Southeast Asian Languages SEA-LION 7B using open-source, commercially permissible datasets sample from LAION OIG chip2 and infill_dbpedia, DataBricks Dolly v2, OpenAI TL;DR, Hello-SimpleAI HC3, dolphin, iapp_wiki_qa_squad, thaisum, xlsum, scb_mt_enth_2020, han dataset, xp3x and Open-Platypus, a total of ~500k samples. Non-commercial datasets were filtered out. Released under apache 2.0 license. The models are trained to perform a subset of instruction-following tasks we found most relevant: reading comprehension, brainstorming, and creative writing. In this model, we focus on Thai and English datasets. We perform Vicuna-style evaluation using human evaluation. In a similar manner to Dolly v2, we only use open-source, commercially permissive pretrained models and datasets. Our models are neither restricted by non-commercial clauses like LLaMA-based models nor non-compete clauses like models that use self-instruct datasets from ChatGPT.
|
10 |
|
11 |
-
Developers: PyThaiNLP and VISTEC-depa AI Research Institute of Thailand
|
12 |
-
|
13 |
-
Model type: SEA-LION 7B (MPT architecture)
|
14 |
|
15 |
## Model Sources
|
16 |
-
Repository: https://github.com/vistec-AI/WangchanLion
|
17 |
-
Demo: [demo_WangchanLion.ipynb - Colaboratory](https://colab.research.google.com/drive/1y_7oOU3ZJI0h4chUrXFL3K4kelW_OI2G?usp=sharing#scrollTo=4yN3Bo6iAH2L)
|
18 |
|
19 |
# Direct Use
|
20 |
Intended to be used as an instruction-following model for reading comprehension, brainstorming, and creative writing.
|
|
|
8 |
|
9 |
WangChanLion is a Multilingual, instruction-finetuned on Southeast Asian Languages SEA-LION 7B using open-source, commercially permissible datasets sample from LAION OIG chip2 and infill_dbpedia, DataBricks Dolly v2, OpenAI TL;DR, Hello-SimpleAI HC3, dolphin, iapp_wiki_qa_squad, thaisum, xlsum, scb_mt_enth_2020, han dataset, xp3x and Open-Platypus, a total of ~500k samples. Non-commercial datasets were filtered out. Released under apache 2.0 license. The models are trained to perform a subset of instruction-following tasks we found most relevant: reading comprehension, brainstorming, and creative writing. In this model, we focus on Thai and English datasets. We perform Vicuna-style evaluation using human evaluation. In a similar manner to Dolly v2, we only use open-source, commercially permissive pretrained models and datasets. Our models are neither restricted by non-commercial clauses like LLaMA-based models nor non-compete clauses like models that use self-instruct datasets from ChatGPT.
|
10 |
|
11 |
+
- Developers: PyThaiNLP and VISTEC-depa AI Research Institute of Thailand
|
12 |
+
- Model type: SEA-LION 7B (MPT architecture)
|
|
|
13 |
|
14 |
## Model Sources
|
15 |
+
- Repository: https://github.com/vistec-AI/WangchanLion
|
16 |
+
- Demo: [demo_WangchanLion.ipynb - Colaboratory](https://colab.research.google.com/drive/1y_7oOU3ZJI0h4chUrXFL3K4kelW_OI2G?usp=sharing#scrollTo=4yN3Bo6iAH2L)
|
17 |
|
18 |
# Direct Use
|
19 |
Intended to be used as an instruction-following model for reading comprehension, brainstorming, and creative writing.
|