--- language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - mistral - trl base_model: SeaLLMs/SeaLLM-7B-v2 datasets: - facebook/belebele --- # AKA Tagamistral-7b-v1: * Yet another archived test/toy model, fine-tuned on portion of Tagalog part of belebele dataset * Base: [SeaLLM](https://huggingface.co/SeaLLMs/SeaLLM-7B-v2) * [GGUF](https://huggingface.co/922-Narra/tagalog-seallm-7b-v1) ### USAGE This is meant to be mainly a chat model. Best results with "Human" and "Assistant" and prompt with Tagalog. Example: "Ito ay isang chat log sa pagitan ng AI Assistant na nagta-Tagalog at isang Pilipino. Magsimula ng chat:\nHuman: Hello po?\nAssistant:" ### HYPERPARAMS * Trained for 2 epochs * rank: 32 * lora alpha: 32 * lora dropout: 0 * lr: 2e-4 * batch size: 2 * warmup ratio: 0.075 * grad steps: 4 This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [](https://github.com/unslothai/unsloth) ### WARNINGS AND DISCLAIMERS There is still a chance that the model may switch to English or Taglish. Results may improve from a cleaner and bigger portion of the belebele dataset fine-tuned on. Additionally, it is possible that the Tagalog capability still comes mostly from the fine-tuned base more than the dataset. Finally, this model is not guaranteed to output aligned or safe outputs nor is it meant for production use - use at your own risk!