- openai/whisper-small for awesome Speech-to-Text with low latency
- mistralai/Mixtral-8x7B-Instruct-v0.1 for an awesome super powerful LLM Brain
- coqui/XTTS-v2 for a nice and clean voice
Which stack will you personally choose?
Join the community of Machine Learners and AI enthusiasts.
Sign UpI would also use a https://huggingface.co/microsoft/phi-2 model, we need a smaller model for quick inference for easy queries
I would use ph-2 on device for daily conversation and detecting users intention, then pass to cloud hosted much larger LLM to do more complicated stuff (such as web browsing and more).
How would you aim for the cheapest latency using existing tooling?
Cannot access the demo 👀
Same models but with SOLAR 10.7B Instruct, as it has similar performance with less parameters
We are starting like those at pi5, we called 'piBrain'
whisper-tiny - tinyllama - vits
everyone is focusing on the model side, but what about the hardware, id be interested in seeing some Open source versions of this by the community.
On the Rabbit, all the heavy compute is done remotely I think 👀