Skip to main content
Mistral AI

Mistral AI launches Small 3.1, a multimodal model that outperforms Gemma 3 and GPT‑4o‑mini

Written by
Maren Kessler
AI News
Mistral AI launches Small 3.1, a multimodal model that outperforms Gemma 3 and GPT‑4o‑mini

Photo by Kevin Ku on Unsplash

While earlier models like Gemma 3 struggled with multimodal tasks, Mistral AI’s new Small 3.1—an Apache 2.0‑licensed, multimodal model—already outperforms both Gemma 3 and GPT‑4o‑mini, Mistral AI Twitter reports.

Quick Summary

  • While earlier models like Gemma 3 struggled with multimodal tasks, Mistral AI’s new Small 3.1—an Apache 2.0‑licensed, multimodal model—already outperforms both Gemma 3 and GPT‑4o‑mini, Mistral AI Twitter reports.
  • Key company: Mistral AI

Mistral AI’s Small 3.1 arrives as the company’s most ambitious open‑source release to date, pairing multimodal capability with an Apache 2.0 licence that invites unrestricted commercial use. The model, announced on Mistral’s official Twitter feed, is positioned as a direct challenger to both the recently unveiled Gemma 3 from Google DeepMind and OpenAI’s GPT‑4o‑mini, which have become reference points for lightweight multimodal performance. In a terse tweet, Mistral AI claimed that Small 3.1 “outperforms Gemma 3 and GPT 4o‑mini,” a statement echoed by VentureBeat’s Michael Nuñez, who noted the model’s superiority “with a fraction of parameters” (VentureBeat). The company did not disclose exact benchmark scores, but the emphasis on “multimodal” suggests that Small 3.1 can ingest text, images, and PDFs in a single inference pass—a capability that Gemma 3 reportedly lacked, according to the same VentureBeat piece.

The release is tightly coupled with Mistral’s broader product ecosystem, notably the “le Chat” interface that bundles search, PDF upload, code generation, and image creation under a single UI. A second Mistral tweet highlighted that le Chat runs on an updated Mistral Large model capable of 1,100 tokens per second for “flash queries,” and the service is already downloadable for Android and iOS (Mistral AI Twitter). By integrating Small 3.1 into le Chat, Mistral offers developers a turnkey stack that can be deployed on‑premise or in the cloud without licensing fees, a contrast to the proprietary APIs that power GPT‑4o‑mini and Gemma 3. This strategy aligns with the company’s recent push into cloud services, exemplified by the launch of Mistral Compute, a Europe‑focused AI‑optimized platform that competes with the likes of AWS and Azure (VentureBeat).

From a market‑share perspective, Small 3.1 could shift the calculus for enterprises that have been weighing open‑source flexibility against the raw performance of closed models. The model’s Apache 2.0 licence removes the legal friction that often deters large‑scale adoption of community‑driven AI, while its multimodal edge addresses a growing demand for unified document processing pipelines. Mistral’s own OCR 3 research, released in December 2025, demonstrated “new frontier for both accuracy and efficiency in document processing” (Mistral AI News). By bundling OCR advances with a multimodal language model, Mistral creates a vertically integrated solution that may appeal to sectors such as legal, finance, and healthcare, where secure, on‑premise processing of mixed media is a regulatory imperative.

Analysts have noted that the open‑source arena is becoming increasingly competitive, with Google’s Gemini line and Meta’s Llama series both courting the same developer base. However, the claim that Small 3.1 outperforms GPT‑4o‑mini—a model that OpenAI markets as the most efficient version of its flagship GPT‑4 architecture—places Mistral in a rare position: it can argue performance parity while offering unrestricted commercial rights. VentureBeat’s Carl Franzen reported that Mistral is already iterating on the model, moving from 3.1 to a 3.2 version within weeks, suggesting a rapid development cadence that could keep the model ahead of its closed‑source rivals (VentureBeat). If the performance gap holds up under independent testing, Small 3.1 may become the de‑facto baseline for cost‑sensitive multimodal deployments.

The broader implication for the AI industry is a potential recalibration of how value is extracted from foundational models. Historically, firms have monetized access through API fees, but an open‑weight model that can be self‑hosted erodes that revenue stream. Mistral’s approach—pairing a free, high‑performing core with premium services like Mistral Compute and the le Chat UI—mirrors a “freemium” playbook that could prove sustainable if enterprise customers gravitate toward the bundled ecosystem. As the race for multimodal dominance intensifies, Small 3.1 exemplifies a strategic bet that openness, speed, and integrated tooling can rival the sheer scale of proprietary offerings, reshaping the competitive landscape for the next generation of AI applications.

Sources

Primary source
Independent coverage

This article was created using AI technology and reviewed by the SectorHQ editorial team for accuracy and quality.

More from SectorHQ:📊Intelligence📝Blog
About the author
Maren Kessler
AI News

🏢Companies in This Story

Related Stories