Introducing Mistral Small 3, a latency-optimized 24B-parameter model that rivals larger models like Llama 3.3 70B and Qwen 32B. It’s a great open alternative to proprietary models, offering impressive speed and accuracy for generative AI tasks. Mistral Small 3 excels in fast-response conversational assistance, low-latency function calling, and fine-tuning for specific domains like legal or medical fields. It can run locally on a single RTX 4090, offering privacy and convenience for different industries. Mistral Small 3 is available for use on various platforms and is set to evolve with enhanced reasoning capabilities in the future.
https://mistral.ai/news/mistral-small-3/