--- license: mit datasets: - unicamp-dl/mmarco - shunk031/jsnli language: - ja --- Model trained on 800,000 Japanese sentences after reducing [oshizo/japanese-e5-mistral-7b_slerp](https://huggingface.co/oshizo/japanese-e5-mistral-7b_slerp) to 8 layers. See this article for details(Japanese) https://note.com/oshizo/n/n9140df790315 See [intfloat/e5-mistral-7b-instruct page](https://huggingface.co/intfloat/e5-mistral-7b-instruct#usage) for model usage.