Mistral Small 3
24Bby Mistral
Mistral Small 3 is a 24B-parameter language model optimized for low-latency performance across common AI tasks. Released under the Apache 2.0 license, it features both pre-trained and instruction-tuned versions designed for efficient local deployment. The model achieves 81% accuracy on the MMLU benchmark and performs competitively with larger models like Llama 3.3 70B and Qwen 32B, while operating at three times the speed on equivalent hardware. [Read the blog post about the model here.](https://mistral.ai/news/mistral-small-3/)
Pricing
Input Tokens
Per 1M tokens
Free
Output Tokens
Per 1M tokens
Free
Image Processing
Per 1M tokens
$0.00/1M tokens
Supported Modalities
Input
text
Output
text
Performance Benchmarks
Intelligence Index
Overall intelligence score
12.7
Math Index
Mathematical reasoning
4.3
GPQA
Graduate-level questions
46.2%
MMLU Pro
Multitask language understanding
65.2%
HLE
Human-like evaluation
4.1%
LiveCodeBench
Real-world coding tasks
25.2%
AIME 2025
Advanced mathematics
4.3%
MATH 500
Mathematical problem solving
71.5%
Specifications
- Context Length
- 33K tokens
- Provider
- Mistral
- Throughput
- 226.34 tokens/s
- Released
- Jan 30, 2025
- Model ID
- mistralai/mistral-small-24b-instruct-2501
Ready to try it?
Start chatting with Mistral Small 3 right now. No credit card required.
Start ChattingMore from Mistral
View all modelsCompare Models
Select a model to compare with Mistral Small 3