--- language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - llama - gguf base_model: unsloth/llama-3-8b-bnb-4bit --- # Uploaded model - **Developed by:** vutuka - **License:** apache-2.0 - **Finetuned from model :** unsloth/llama-3-8b-bnb-4bit ## LLama CPP ```py from llama_cpp import Llama llm = Llama.from_pretrained( repo_id="vutuka/llama-3-8b-african-aya-f16", filename="*F16.gguf", verbose=False, ) output = llm( "Instruction: Ni nchi zipi zilizo tajiri zaidi duniani ?\n Response: ", # Prompt max_tokens=None, echo=True ) print(output) ``` This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [](https://github.com/unslothai/unsloth)