CSGO Coach Mia, Finetuned on mistralai/Mistral-7B-Instruct-v0.2
Sample usage :
from huggingface_hub import hf_hub_download from llama_cpp import Llama import torch
Specify the path to your .gguf file
model_path = '/content/finetuned8b/finetuned8b.Q5_K_M.gguf'
Instantiate the Llama model
llm = Llama(model_path=model_path)
prompt = "Coach Mia, help me with aiming "
Generation kwargs
generation_kwargs = { "max_tokens":200, "stop":'[INST]', "echo":False, # Echo the prompt in the output "top_k":1 # This is essentially greedy decoding, since the model will always return the highest-probability token. Set this value > 1 for sampling decoding }
res = llm(prompt, **generation_kwargs)
Unpack and the generated text from the LLM response dictionary and print it
print(res["choices"][0]["text"])
res is short for result
#output
100% accuracy. [/INST] Aiming is a crucial aspect of CS:GO. Let's start by analyzing your sensitivity settings and crosshair placement. We can also run some aim training drills to improve your precision.
- Downloads last month
- 2