license: apache-2.0 | |
# etri-xainlp/llama2-12.8b_lora-dpo_v1 | |
## Model Details | |
**Model Developers** ETRI xainlp team | |
**Input** text only. | |
**Output** text only. | |
**Model Architecture** | |
**Base Model** [meta-llama/Llama-13b-hf](https://huggingface.co/meta-llama/Llama-2-13b-hf) | |
**Training Dataset** | |
- sft+lora: 710k instruction-following set | |
- dpo+lora: 90k user preference set | |
- We use A100 GPU 80GB * 8, when training. |