I set myself the goal to build the fastest AI for the command line!
Of course by first using the fastest language and also the fastest LLM right now:
Mixtral powered by Groq at up to 500 tokens per second!
And the first results are already quite impressive!
You can find the installation instructions here:
Fast, however, means to get the final result fast and sometimes you need more than one model. So today I added support for explicitly specifying a provider:
Which additional features would you like to see?