--- library_name: transformers tags: - unsloth language: - tr --- # Model Card for Model ID Fine-tuned Llama3-8b model with Lora (trained 1 epoch on colap A100 for experimental purposes) Base Model: unsloth/llama-3-8b-bnb-4bit Fine-tuning process video: https://www.youtube.com/watch?v=pK8u4QfdLx0&ab_channel=DavidOndrej Turkish Fine-tune notebook: https://github.com/yudumpacin/LLM/blob/main/Alpaca_%2B_Llama_3_8b_full_Turkish.ipynb Original unsloth notebook: https://colab.research.google.com/drive/135ced7oHytdxu3N2DNe1Z0kqjyYIkDXp?usp=sharing Fine-tuning data : - Yudum/turkish-instruct-dataset which includes; * open question category of atasoglu/databricks-dolly-15k-tr * parsak/alpaca-tr-1k-longest * TFLai/Turkish-Alpaca * umarigan/GPTeacher-General-Instruct-tr # Usage ```python from unsloth import FastLanguageModel model, tokenizer = FastLanguageModel.from_pretrained( model_name = "Yudum/llama3-lora-turkish", max_seq_length = 2048, dtype = None, load_in_4bit = True, ) FastLanguageModel.for_inference(model) # Enable native 2x faster inference alpaca_prompt = """Altta bir görevi tanımlayan bir talimat ile daha fazla bilgi sağlayan bir girdi bulunmaktadır. İsteği uygun şekilde tamamlayan bir yanıt yazın. ### Talimat: {} ### Girdi: {} ### Yanıt: {} """ inputs = tokenizer( [ alpaca_prompt.format( "Paris'teki meşhur kulenin ismi nedir?", # instruction "", # input "", # output - leave this blank for generation! ) ], return_tensors = "pt").to("cuda") outputs = model.generate(**inputs, max_new_tokens = 64, use_cache = True) tokenizer.batch_decode(outputs) ```