MMMLU
Multilingual
text
+
+
+
+
About
MMMLU (Multilingual Massive Multitask Language Understanding) extends the MMLU benchmark to multiple languages, evaluating language models' knowledge and reasoning capabilities across diverse linguistic contexts. This multilingual adaptation tests models' ability to understand and apply academic knowledge in various languages, assessing both linguistic competency and cross-cultural knowledge transfer.
+
+
+
+
Evaluation Stats
Total Models15
Organizations4
Verified Results0
Self-Reported15
+
+
+
+
Benchmark Details
Max Score1
Language
en
+
+
+
+
Performance Overview
Score distribution and top performers
Score Distribution
15 models
Top Score
89.5%
Average Score
81.5%
High Performers (80%+)
11Top Organizations
#1Anthropic
6 models
87.2%
#2Alibaba Cloud / Qwen Team
1 model
86.7%
#3OpenAI
6 models
81.2%
#4Microsoft
2 models
62.7%
+
+
+
+
Leaderboard
15 models ranked by performance on MMMLU
| License | Links | ||||
|---|---|---|---|---|---|
| Aug 5, 2025 | Proprietary | 89.5% | |||
| Sep 29, 2025 | Proprietary | 89.1% | |||
| May 22, 2025 | Proprietary | 88.8% | |||
| Dec 17, 2024 | Proprietary | 87.7% | |||
| Apr 14, 2025 | Proprietary | 87.3% | |||
| Apr 29, 2025 | Apache 2.0 | 86.7% | |||
| May 22, 2025 | Proprietary | 86.5% | |||
| Feb 24, 2025 | Proprietary | 86.1% | |||
| Feb 27, 2025 | Proprietary | 85.1% | |||
| Oct 15, 2025 | Proprietary | 83.0% |
Showing 1 to 10 of 15 models