Here are the LoRA adapters produced by using my text data with Unsloth's training code, trained for 1000 steps. The base model is the Llama-3-8b-bnb-4bit. If you would like to see the modifications I made to Unsloth's script to make it more concise and adaptable to my own data format, you can find the modified script here
Uploaded model
- Developed by: chrismontes
- License: apache-2.0
- Finetuned from model : unsloth/llama-3-8b-bnb-4bit
This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API:
The model has no pipeline_tag.
Model tree for chrismontes/Dog-LoRA
Base model
meta-llama/Meta-Llama-3-8B
Quantized
unsloth/llama-3-8b-bnb-4bit