SeaLLM-7B-Chat quantized to run locally with modest GPU

Model Description

  • This is a quantized model of SeaLLM-7B-Chat.
  • SeaLLMs is a family of LLMs pre-trained from Meta's LLaMA 2 and optimized for numerous Southeast Asian languages, including Vietnamese ๐Ÿ‡ป๐Ÿ‡ณ, Indonesian ๐Ÿ‡ฎ๐Ÿ‡ฉ, Thai ๐Ÿ‡น๐Ÿ‡ญ, Malay ๐Ÿ‡ฒ๐Ÿ‡พ, Khmer ๐Ÿ‡ฐ๐Ÿ‡ญ, Lao ๐Ÿ‡ฑ๐Ÿ‡ฆ, Tagalog ๐Ÿ‡ต๐Ÿ‡ญ and Burmese ๐Ÿ‡ฒ๐Ÿ‡ฒ.
  • The quantization has been done with ExLlamaV2, a fast LLM inference library.

Citation

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.