Caslow's picture
git add lora_model/config.json
268156c
raw
history blame contribute delete
679 Bytes
{
"architectures": ["LlamaForCausalLM"],
"model_type": "llama",
"hidden_size": 4096,
"intermediate_size": 11008,
"num_attention_heads": 32,
"num_hidden_layers": 32,
"vocab_size": 128000,
"max_position_embeddings": 2048,
"bos_token_id": 1,
"eos_token_id": 2,
"pad_token_id": 0,
"initializer_range": 0.02,
"lora_config": {
"lora_alpha": 16,
"lora_dropout": 0,
"r": 16,
"target_modules": [
"k_proj",
"v_proj",
"down_proj",
"gate_proj",
"o_proj",
"up_proj",
"q_proj"
]
},
"peft_type": "LORA",
"task_type": "CAUSAL_LM"
}