MMMU

Multimodal
+
+
+
+
About

MMMU evaluates multimodal models on 11.5K college-level questions across 30 subjects requiring visual perception and domain-specific knowledge in art, science, business, health, and engineering.

+
+
+
+
Evaluation Stats
Total Models16
Organizations9
Verified Results0
Self-Reported0
+
+
+
+
Benchmark Details
Max Score100
+
+
+
+
Performance Overview
Score distribution and top performers

Score Distribution

16 models
Top Score
78.4%
Average Score
30.1%
High Performers (80%+)
0

Top Organizations

#1Alibaba / Qwen
1 model
68.1%
#2ByteDance
1 model
67.6%
#3Kunlun Tech
1 model
55.4%
#4OpenAI
5 models
45.8%
#5Google DeepMind
2 models
33.5%
+
+
+
+
Leaderboard
16 models ranked by performance on MMMU
LicenseLinks
Aug 7, 2025
Proprietary
78.4%
Apr 16, 2025
Proprietary
76.4%
Nov 12, 2025
Proprietary
76.0%
Sep 23, 2025
Apache 2.0
68.1%
Mar 25, 2025
Proprietary
68.0%
May 15, 2025
Proprietary
67.6%
Jul 9, 2025
Apache 2.0
55.4%
Nov 24, 2025
Proprietary
-1.0%
Apr 16, 2025
Proprietary
-1.0%
Apr 17, 2025
Proprietary
-1.0%
Showing 1 to 10 of 16 models
+
+
+
+
Additional Metrics
Extended metrics for top models on MMMU
ModelScoreMMMU-Pro
GPT-578.484.2%
o376.482.9%
GPT-5.176.085.4%
Qwen3-VL-235B-A22B68.178.7%
Gemini 2.5 Pro68.079.6%
Seed 1.5-VL67.677.9%
Skywork-R1V3-38B55.476%
Claude Opus 4.5-1.080.7%
o4 mini-1.081.6%
Gemini 2.5 Flash-1.079.7%
o1-1.078.2%
Grok 3-1.078%
Claude Sonnet 4.5-1.077.8%
InternS1-1.077.7%
Llama 4 Behemoth-1.076.1%
Claude Opus 4.1-1.076.5%
+
+
+
+
Resources