
Gemini Diffusion
Zero-eval
#1LBPP (v2)
#1BigCodeBench
#3BIG-Bench Extra Hard
by Google
+
+
+
+
About
Gemini Diffusion is a language model developed by Google. The model shows competitive results across 10 benchmarks. It excels particularly in HumanEval (89.6%), MBPP (76.0%), Global-MMLU-Lite (69.1%). Released in 2025, it represents Google's latest advancement in AI technology.
+
+
+
+
Timeline
AnnouncedMay 20, 2025
ReleasedMay 20, 2025
+
+
+
+
License & Family
License
Proprietary
Performance Overview
Performance metrics and category breakdown
Overall Performance
10 benchmarks
Average Score
46.9%
Best Score
89.6%
High Performers (80%+)
1+
+
+
+
All Benchmark Results for Gemini Diffusion
Complete list of benchmark scores with detailed information
HumanEval | text | 0.90 | 89.6% | Self-reported | |
MBPP | text | 0.76 | 76.0% | Self-reported | |
Global-MMLU-Lite | text | 0.69 | 69.1% | Self-reported | |
LBPP (v2) | text | 0.57 | 56.8% | Self-reported | |
BigCodeBench | text | 0.45 | 45.4% | Self-reported | |
GPQA | text | 0.40 | 40.4% | Self-reported | |
LiveCodeBench | text | 0.31 | 30.9% | Self-reported | |
AIME 2025 | text | 0.23 | 23.3% | Self-reported | |
SWE-Bench Verified | text | 0.23 | 22.9% | Self-reported | |
BIG-Bench Extra Hard | text | 0.15 | 15.0% | Self-reported |