--- license: mit --- This repo contains a low-rank adapter for LLaMA-13b finetuned on Ntropy proprietary dataset (consumer financial transactions). This version of the weights was trained with the following hyperparameters: - Base Model: decapoda-research/llama-13b-hf - Epochs: 10 (load from best epoch) - Batch size: 16 - Cutoff length: 1024 - Learning rate: 3e-4 - Lora r: 16 - Lora target modules: q_proj, k_proj, v_proj, o_proj Instructions for running the adapter can be found at https://github.com/ntropy-network/enrichment_models/blob/main/notebooks/llama.ipynb