base_model: mistralai/Mistral-7B-Instruct-v0.2 model_type: MistralForCausalLM is_mistral_derived_model: true datasets: - path: burkelibbey/colors type: field_instruction: color field_output: description conversation: chatml chat_template: chatml val_set_size: 0.05 output_dir: ./out sequence_len: 512 gradient_accumulation_steps: 4 micro_batch_size: 2 learning_rate: 0.0002 load_in_4bit: true adapter: qlora lora_r: 16 lora_alpha: 8 lora_dropout: 0.05 lora_target_linear: true gradient_checkpointing: true warmup_steps: 10 tokenizer_type: LlamaTokenizer num_epochs: 1 optimizer: paged_adamw_32bit special_tokens: bos_token: "" eos_token: "" unk_token: ""