Mistral AI Releases Mistral Small 3.1: A New Benchmark in Open-Source AI Models
4 minutes
Mistral AI has unveiled Mistral Small 3.1, a powerful new language model that combines state-of-the-art performance with an unusually permissive Apache 2.0 license, making it freely available for both research and commercial applications.
The Technical Achievement
Mistral Small 3.1 builds upon its predecessor, Mistral Small 3, with notable improvements across multiple dimensions. The model offers enhanced text processing capabilities while adding robust multimodal understanding—the ability to process and reason about images alongside text. Perhaps most impressively, it features an expanded context window of up to 128,000 tokens, allowing it to process and maintain awareness of substantially longer pieces of text than most comparable models.
According to benchmark tests conducted by Mistral AI, the new model outperforms similar-sized competitors including Google's Gemma 3, Anthropic's Claude 3.5 Haiku, and OpenAI's GPT-4o Mini on numerous standard evaluation metrics. This achievement is particularly noteworthy given that Mistral Small 3.1 maintains competitive inference speeds of 150 tokens per second, making it practical for real-time applications.
Industry Context
The release comes at a critical juncture in the AI industry's evolution. As large language models have grown increasingly capable, concerns about accessibility and concentrated power have intensified. The vast majority of cutting-edge AI models remain proprietary and accessible only through paid APIs controlled by a handful of major technology companies.
Mistral AI, founded in 2023 by former DeepMind and Meta AI researchers, has positioned itself as a European challenger to U.S. AI giants, with a stated philosophy emphasizing open access to advanced AI technology. The company has raised over €385 million in funding, indicating significant investor confidence in its approach.
This release represents the first time an open-source model has demonstrably surpassed the performance of leading proprietary models in the same weight class across text processing, multimodal capabilities, multilingual support, and long-context handling.
Technical Specifications and Capabilities
Mistral Small 3.1's architecture enables it to run on relatively modest hardware—a single NVIDIA RTX 4090 GPU or even a Mac with 32GB of RAM—making it accessible to individual developers and smaller organizations without access to extensive computing resources.
The model excels in:
-
Instruction following and conversational assistance: It demonstrates strong performance on benchmarks measuring language understanding, reasoning, and knowledge.
-
Multimodal comprehension: Unlike many open-source models, it can understand and reason about images alongside text, performing well on visual question-answering tasks.
-
Multilingual support: The model shows robust performance across European, East Asian, and Middle Eastern languages.
-
Long context processing: With its 128K token context window, it can maintain awareness of information across lengthy documents.
-
Low-latency function calling: This enables integration with other software systems, essential for building agentic AI applications.
Practical Applications
The versatility of Mistral Small 3.1 makes it suitable for a wide range of applications:
- Document processing and verification systems
- Medical diagnostics and healthcare assistance
- On-device image processing
- Visual inspection for manufacturing quality control
- Security systems with object detection
- Image-based customer support
- General-purpose virtual assistants
The model's fine-tuning capabilities are particularly promising for specialized domains like legal advice, medical diagnostics, and technical support, where domain-specific expertise is critical.
Building on Open Foundations
Mistral AI has noted the community's rapid adoption of its previous models, highlighting how researchers have built upon these open foundations. For example, Nous Research recently developed DeepHermes 24B based on Mistral Small 3, enhancing its reasoning capabilities. By releasing both the base and instruction-tuned versions of Mistral Small 3.1, the company aims to enable similar innovation with its latest model.
Availability and Deployment
Mistral Small 3.1 is immediately available for download from Hugging Face in both base and instruction-tuned variants. For enterprise users requiring private and optimized inference infrastructure, Mistral AI offers specialized deployment options.
The model can also be accessed through Mistral AI's developer platform "La Plateforme" and is available on Google Cloud Vertex AI. Integration with NVIDIA NIM and Microsoft Azure AI Foundry is expected in the coming weeks, expanding its accessibility across major cloud platforms.
The model is available for download from Hugging Face at Mistral Small 3.1 Base and Mistral Small 3.1 Instruct.
Conclusion
This release represents a significant advancement in the democratization of AI capabilities. By making a high-performance, multimodal model freely available under a permissive license, Mistral AI is challenging the prevailing business model of AI access through controlled APIs and subscription services.
Valeriia Kuka
Valeriia Kuka, Head of Content at Learn Prompting, is passionate about making AI and ML accessible. Valeriia previously grew a 60K+ follower AI-focused social media account, earning reposts from Stanford NLP, Amazon Research, Hugging Face, and AI researchers. She has also worked with AI/ML newsletters and global communities with 100K+ members and authored clear and concise explainers and historical articles.