Text Generation
Transformers
PyTorch
Thai
English
llama
openthaigpt
Inference Endpoints
text-generation-inference
Edit model card

πŸ‡ΉπŸ‡­ OpenThaiGPT 1.0.0-alpha

OpenThaiGPT Version 1.0.0-alpha is the first Thai implementation of a 7B-parameter LLaMA v2 Chat model finetuned to follow Thai translated instructions below and makes use of the Huggingface LLaMA implementation.

---- Full Huggingface Checkpoint Model ----

Upgrade from OpenThaiGPT 0.1.0-beta

  • Using Facebook LLama v2 model 7b chat as a base model which is pretrained on over 2 trillion token.
  • Context Length is upgrade from 2048 token to 4096 token
  • Allow research and commerical use.a

Pretrain Model

Support

License

Source Code: License Apache Software License 2.0.
Weight: Research and Commercial uses.

Code and Weight

Colab Demo: https://colab.research.google.com/drive/1kDQidCtY9lDpk49i7P3JjLAcJM04lawu?usp=sharing
Finetune Code: https://github.com/OpenThaiGPT/openthaigpt-finetune-010beta
Inference Code: https://github.com/OpenThaiGPT/openthaigpt
Weight (Lora Adapter): https://huggingface.co/openthaigpt/openthaigpt-1.0.0-alpha-7b-chat
Weight (Huggingface Checkpoint): https://huggingface.co/openthaigpt/openthaigpt-1.0.0-alpha-7b-chat-ckpt-hf
Weight (GGML): https://huggingface.co/openthaigpt/openthaigpt-1.0.0-alpha-7b-chat-ggml
Weight (Quantized 4bit GGML): https://huggingface.co/openthaigpt/openthaigpt-1.0.0-alpha-7b-chat-ggml-q4

Sponsors

Pantip.com, ThaiSC

Powered by

OpenThaiGPT Volunteers, Artificial Intelligence Entrepreneur Association of Thailand (AIEAT), and Artificial Intelligence Association of Thailand (AIAT)

Authors

Disclaimer: Provided responses are not guaranteed.

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 42.05
ARC (25-shot) 50.85
HellaSwag (10-shot) 74.89
MMLU (5-shot) 40.02
TruthfulQA (0-shot) 47.23
Winogrande (5-shot) 69.06
GSM8K (5-shot) 3.87
DROP (3-shot) 8.42
Downloads last month
1,801
Inference API
Model is too large to load in Inference API (serverless). To try the model, launch it on Inference Endpoints (dedicated) instead.

Datasets used to train openthaigpt/openthaigpt-1.0.0-alpha-7b-chat-ckpt-hf

Spaces using openthaigpt/openthaigpt-1.0.0-alpha-7b-chat-ckpt-hf 21