metadata
license: other
language:
- en
- fr
- es
- hi
- zh
- code
base_model: microsoft/Orca-2-13b
datasets:
- HuggingFaceH4/no_robots
- mlabonne/guanaco-llama2-1k
- OpenAssistant/oasst_top1_2023-08-25
- totally-not-an-llm/EverythingLM-data-V3
- LDJnr/Pure-Dove
- LDJnr/Capybara
- LDJnr/LessWrong-Amplify-Instruct
- LDJnr/Verified-Camel
The "microsoft/Orca-2-13b" model fully fine-tuned on HuggingFaceH4/no_robots, totally-not-an-llm/EverythingLM-data-V3, LDJnr/Capybara, LDJnr/Pure-Dove, LDJnr/LessWrong-Amplify-Instruct, LDJnr/Verified-Camel, mlabonne/guanaco-llama2-1k, and OpenAssistant/oasst_top1_2023-08-25. This model achieved a test loss of 0.39 on LDJnr/Verified-Camel.
Make sure to comply with the microsoft research license. Please read it before using this model.
This model was trained on the ChatML prompt template.