Google

Gemini Diffusion

Zero-eval
#1LBPP (v2)
#1BigCodeBench
#3BIG-Bench Extra Hard

by Google

+
+
+
+
About

Gemini Diffusion is a language model developed by Google. The model shows competitive results across 10 benchmarks. It excels particularly in HumanEval (89.6%), MBPP (76.0%), Global-MMLU-Lite (69.1%). Released in 2025, it represents Google's latest advancement in AI technology.

+
+
+
+
Timeline
AnnouncedMay 20, 2025
ReleasedMay 20, 2025
+
+
+
+
License & Family
License
Proprietary
Performance Overview
Performance metrics and category breakdown

Overall Performance

10 benchmarks
Average Score
46.9%
Best Score
89.6%
High Performers (80%+)
1
+
+
+
+
All Benchmark Results for Gemini Diffusion
Complete list of benchmark scores with detailed information
HumanEval
text
0.90
89.6%
Self-reported
MBPP
text
0.76
76.0%
Self-reported
Global-MMLU-Lite
text
0.69
69.1%
Self-reported
LBPP (v2)
text
0.57
56.8%
Self-reported
BigCodeBench
text
0.45
45.4%
Self-reported
GPQA
text
0.40
40.4%
Self-reported
LiveCodeBench
text
0.31
30.9%
Self-reported
AIME 2025
text
0.23
23.3%
Self-reported
SWE-Bench Verified
text
0.23
22.9%
Self-reported
BIG-Bench Extra Hard
text
0.15
15.0%
Self-reported
+
+
+
+
Resources