NeuralKunoichi-EroSumika-4x7B
Collection
OUTDATED
•
3 items
•
Updated
Test merge. Attempt to get good at RP, ERP, general tasks model with 128k context. Every model here has Epiculous/Fett-uccine-Long-Noodle-7B-120k-Context in merge instead of regular MistralYarn 128k. The reason is because i belive Epiculous merged it with Mistral Instruct v0.2 to make first 32k context experience as perfect as possible until we reach YaRN from 32 to 128k, if not - it's sad D:, or, i get something wrong.
Here is the "family tree" of this model, im not writing full model names cause they long af
* NeuralKunoichi-EroSumika 4x7B
*(1) Kunocchini-7b-128k
|
*(2) Mistral-Instruct-v0.2-128k
* Mistral-7B-Instruct-v0.2
|
* Fett-128k
|
*(3) Erosumika-128k
* Erosumika 7B
|
* FFett-128k
|
*(4) Mistral-NeuralHuman-128k
* Fett-128k
|
* Mistral-NeuralHuman
* Mistral_MoreHuman
|
* Mistral-Neural-Story