Mistral logo

Mistral Small 3

24B

by Mistral

Mistral Small 3 is a 24B-parameter language model optimized for low-latency performance across common AI tasks. Released under the Apache 2.0 license, it features both pre-trained and instruction-tuned versions designed for efficient local deployment. The model achieves 81% accuracy on the MMLU benchmark and performs competitively with larger models like Llama 3.3 70B and Qwen 32B, while operating at three times the speed on equivalent hardware. [Read the blog post about the model here.](https://mistral.ai/news/mistral-small-3/)

Chat with Mistral Small 3

Pricing

Input Tokens
Per 1M tokens
Free
Output Tokens
Per 1M tokens
Free
Image Processing
Per 1M tokens
$0.00/1M tokens

Supported Modalities

Input

text

Output

text

Performance Benchmarks

Intelligence Index
Overall intelligence score
12.7
Math Index
Mathematical reasoning
4.3
GPQA
Graduate-level questions
46.2%
MMLU Pro
Multitask language understanding
65.2%
HLE
Human-like evaluation
4.1%
LiveCodeBench
Real-world coding tasks
25.2%
AIME 2025
Advanced mathematics
4.3%
MATH 500
Mathematical problem solving
71.5%

Specifications

Context Length
33K tokens
Provider
Mistral
Throughput
226.34 tokens/s
Released
Jan 30, 2025
Model ID
mistralai/mistral-small-24b-instruct-2501

Ready to try it?

Start chatting with Mistral Small 3 right now. No credit card required.

Start Chatting

More from Mistral

View all models