minstral ai just released their latest generation of open source models. Ministral 8b has great performance even with very small quantizations. So for now the ministral 8b q2 will be used as the new default. This significantly dereases the size of the container while improving performance |
||
|---|---|---|
| .. | ||
| Dockerfile | ||