Mistral AI has stepped up its niche focus on open-source technology and its development mottos through the introduction of an open-source software engineering (SWE) agent model, Devstral. With 24 billion parameters, Devstral is small enough to run on laptops but powerful enough to outperform much larger proprietary models in real-world benchmarks. Developed in collaboration with All Hands AI, the creators of OpenDevin, Devstral marks a significant leap toward autonomous software development. It represents a new frontier where AI-powered agents can deeply understand, navigate, and modify codebases.
This article explores the key innovations behind Devstral, its real-world applications, and what it signals for the future of software engineering.
A Response to Community Backlash
Mistral AI has quickly become one of the most formidable players in the open-source AI arena since its debut in late 2023. Its previous releases, like Codestral and Medium 3, have shown both technical prowess and a willingness to challenge established norms. However, its recent proprietary model, Medium 3, drew criticism from developers who saw it as a departure from the company’s open ethos.
Devstral is Mistral’s answer to that criticism. Released under the permissive Apache 2.0 license, Devstral allows anyone to use, modify, and commercialize the model without restrictions. It’s a reaffirmation that the company still prioritizes openness, developer freedom, and community collaboration.
“We wanted to release something open for developers and other enthusiasts–something they can run locally in an environment of privacy, modify as they wish,” said Baptiste Rozière, a research scientist at Mistral AI.

From Codestral to Devstral: An Evolution
Devstral builds on the success of the Codestral model family. Initially launched in May 2024, Codestral featured 22 billion parameters and supported about 80 programming languages. Its technical merits quickly led to rapid iterations like Codestral-Mamba and Codestral 25.01, gaining traction among IDE plugin developers and enterprise users.
Devstral takes these foundations further, transforming from a fast code completion tool into a full software engineering agent. Unlike traditional LLMs that only suggest code snippets, Devstral is designed to understand context across files, debug issues, and execute multi-step development tasks autonomously.
Benchmark-Breaking Performance
Devstral achieves a 46.8% score on SWE-Bench Verified, a benchmark comprising 500 manually verified real-world GitHub issues. This score surpasses not only all previous open-source models but also proprietary offerings like GPT-4.1-mini, outperforming it by over 20 percentage points.
“At this moment, by significant margin, it is the best model available in the open for SWE-bench verified and code agents,” Rozière said.
The model was fine-tuned using reinforcement learning and safety alignment, ensuring both high performance and robustness. Importantly, the training data excluded repositories cloned from SWE-Bench, reducing overfitting and ensuring generalization.
Engineered for the Agentic Era
What sets Devstral apart is its optimization for agentic AI development. It integrates seamlessly with agentic scaffolding frameworks like OpenHands, SWE-Agent, and OpenDevin.
These platforms allow Devstral to:
- Navigate large codebases
- Execute test cases
- Perform multi-step workflows
For example, within the OpenDevin scaffold, Devstral can act like a backend developer—resolving bugs, updating packages, or modifying code structures autonomously.
Ask it to perform small tasks, like updating the version of a package or changing a tokenization script. It finds just the right spot in your code and makes the changes, which is nice to use,” Rozière continues.
Lightweight, Local, and Enterprise-Ready
At 24 billion parameters, Devstral strikes a sweet spot between performance and efficiency. It can run on:
- RTX 4090 GPUs
- MacBooks with 32GB RAM
- Even air-gapped environments with no internet access
This local deployment capability makes Devstral ideal for privacy-sensitive industries such as finance, healthcare, and government. Moreover, the Apache 2.0 license removes barriers for enterprise adoption, enabling companies to integrate, adapt, and commercialize the model with minimal legal overhead.
Real-World Usage and Deployment
Devstral is already being used internally at Mistral for real development tasks. The company has “dogfooded” the model, testing it across diverse repositories and real-world software engineering workflows.
Developers can deploy Devstral using a wide range of platforms and tools:
- Hugging Face
- Ollama
- Kaggle
- LM Studio
- Unsloth
It supports popular libraries like vLLM, Transformers, and Mistral Inference, and features a generous 128,000 token context window. The tokenizer used is Tekken with a vocabulary of 131,000.
For cloud deployments, Devstral is available via Mistral’s Le Platforme API, priced at:
- $0.10 per million input tokens
- $0.30 per million output tokens
What Lies Ahead?
Devstral is the initial starting point given that it was released to the public in research preview. Mistral and All Hands AI are currently working on bigger brother versions with enhanced capabilities.
“These gaps between small and big will then of course always exist. But we are bridging it.”
With Devstral, the open-source community now has access to a high-performance, low-footprint model capable of powering real software engineering agents. It is not merely a coding assistant but a cornerstone for autonomous AI development systems.
Final Thoughts
Mistral AI’s Devstral is a watershed moment in the evolution of AI-powered development. With unmatched performance in SWE benchmarks, agentic integration, and a highly permissive license, it reestablishes Mistral as a leading force in open-source AI.
By enabling developers to run powerful models on local machines, Devstral not only democratizes AI development but also paves the way for a new era where autonomous coding agents become a standard part of the software lifecycle.
Whether you’re an individual developer, an open-source enthusiast, or an enterprise looking to scale AI-driven development, Devstral offers a rare combination of power, portability, and freedom.