stories15M_MOE
This model is ModelCloud/tinyllama-15M-stories repeated 4 times to make 4 experts.
The model is used for testing, not intended to be used in production (unless your product is some kind of bedtime story teller)
Weight of router is initialized randomly
shakespeare LoRA adapter
A LoRA adapter trained on first 100 paragraphs of shakespeare can be found inside moe_shakespeare15M
With input: Look in thy glass
- Original model generates:
Look in thy glass was a little girl. She was only three years old and she was three years old. She was
- LoRA adapter generates:
Look in thy glass in love of the eye: That's when when the eye see thy on the sun'
- Downloads last month
- 17,721
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.