OpenAI

o4-mini

Multimodal
Zero-eval
#2AIME 2024
#2MathVista
#2BrowseComp
+2 more

by OpenAI

+
+
+
+
About

o4-mini is a multimodal language model developed by OpenAI. It achieves strong performance with an average score of 66.5% across 14 benchmarks. It excels particularly in AIME 2024 (93.4%), AIME 2025 (92.7%), MathVista (84.3%). It supports a 300K token context window for handling large documents. The model is available through 1 API provider. As a multimodal model, it can process and understand text, images, and other input formats seamlessly. Released in 2025, it represents OpenAI's latest advancement in AI technology.

+
+
+
+
Pricing Range
Input (per 1M)$1.10 -$1.10
Output (per 1M)$4.40 -$4.40
Providers1
+
+
+
+
Timeline
AnnouncedApr 16, 2025
ReleasedApr 16, 2025
Knowledge CutoffMay 31, 2024
+
+
+
+
Specifications
Capabilities
Multimodal
+
+
+
+
License & Family
License
Proprietary
Performance Overview
Performance metrics and category breakdown

Overall Performance

14 benchmarks
Average Score
66.5%
Best Score
93.4%
High Performers (80%+)
5

Performance Metrics

Max Context Window
300.0K
Avg Throughput
115.0 tok/s
Avg Latency
5ms
+
+
+
+
All Benchmark Results for o4-mini
Complete list of benchmark scores with detailed information
AIME 2024
text
0.93
93.4%
Self-reported
AIME 2025
text
0.93
92.7%
Self-reported
MathVista
multimodal
0.84
84.3%
Self-reported
MMMU
multimodal
0.82
81.6%
Self-reported
GPQA
text
0.81
81.4%
Self-reported
CharXiv-R
multimodal
0.72
72.0%
Self-reported
TAU-bench Retail
text
0.72
71.8%
Self-reported
Aider-Polyglot
text
0.69
68.9%
Self-reported
SWE-Bench Verified
text
0.68
68.1%
Self-reported
Aider-Polyglot Edit
text
0.58
58.2%
Self-reported
Showing 1 to 10 of 14 benchmarks