Mistral AI on Tuesday introduced Mistral Medium 3.5, a dense 128-billion-parameter model that consolidates instruction-following, reasoning, and coding into a single system. As a result, the company aligns its flagship capabilities into one unified architecture. At the same time, it expanded its tooling with cloud-based coding agents in Vibe CLI and a new Work mode in Le Chat, both powered by the new model.
Unified Model Architecture
Mistral Medium 3.5 replaces three earlier systems, namely Mistral Medium 3.1, Magistral, and Devstral 2. Consequently, developers now interact with a single model across chat, reasoning, and coding workflows. The model supports a 256,000-token context window, while it also processes both text and image inputs. In addition, it allows adjustable reasoning effort per request, which enables use cases ranging from simple queries to extended agent-driven tasks.
Moreover, benchmark performance shows measurable gains, with 77.6% on SWE-Bench Verified and 91.4% on τ³-Telecom. These scores indicate stronger coding reliability compared to earlier iterations. Meanwhile, the company released open weights on Hugging Face under a Modified MIT License, which permits broad use with limits for large enterprises. Notably, NVIDIA confirmed that deployment can run on as few as four GPUs, which lowers infrastructure barriers.
Cloud Execution and Agent Workflow
Alongside the model, Mistral shifted toward cloud execution. Coding agents in Vibe CLI now run asynchronously, so tasks execute in parallel rather than on local machines. As a result, developers receive notifications once processes complete, which improves workflow efficiency. Additionally, users can launch sessions directly from either Vibe CLI or Le Chat, creating a more integrated environment.
Work mode in Le Chat further extends this approach. It places Medium 3.5 inside an agent framework that selects tools such as email, Slack, and GitHub. Then it executes multi-step operations until completion. However, the system requires explicit user approval before performing sensitive actions, which maintains control over outputs. Currently, this feature is rolling out in preview across Free, Pro, and Team plans.
Competitive Positioning
This release arrives during a period of increased activity in open-weight AI models. Therefore, Mistral positions Medium 3.5 as a cost-efficient alternative for enterprise adoption. The model offers a relatively compact footprint, particularly when quantized to around 70 GB, which brings it closer to consumer-grade hardware environments.
At the same time, the broader market continues to compare open-weight systems with closed-source frontier models. While benchmark results suggest progress, real-world parity remains under evaluation. Nevertheless, Mistral’s approach emphasizes efficiency and accessibility, which could appeal to organizations balancing performance with infrastructure cost.








