--- license: other language: - en tags: - llama - fine tune - light novel - eminence in shadow - konosuba --- This repo is my fine tuned lora of Llama on the first 4 volumes of Eminence in shadow and konosuba to test its ability to record new information. The training used alpaca-lora on a 3090 for 10 hours with : - Micro Batch Size 2, - batch size 64, - 35 epochs, - 3e-4 learning rate, - lora rank 256, - 512 lora alpha, - 0.05 lora dropout, - 352 cutoff