stories15M_MOE
This model is ModelCloud/tinyllama-15M-stories repeated 4 times to make 4 experts.
The model is used for testing, not intended to be used in production (unless your product is some kind of bedtime story teller)
Weight of router is initialized randomly
shakespeare LoRA adapter
A LoRA adapter trained on first 100 paragraphs of shakespeare can be found inside moe_shakespeare15M
With input: Look in thy glass
- Original model generates:
Look in thy glass was a little girl. She was only three years old and she was three years old. She was
- LoRA adapter generates:
Look in thy glass in love of the eye: That's when when the eye see thy on the sun'
- Downloads last month
- 214,199
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the model is not deployed on the HF Inference API.