AlpacaEval 2.0
text
+
+
+
+
About
AlpacaEval 2.0 is an automated evaluation benchmark for instruction-following language models that achieves 0.98 correlation with ChatBot Arena while being fast, cheap, and reliable. It evaluates models' ability to follow general user instructions using GPT-4 as an auto-annotator, comparing responses to reference outputs. The benchmark features length-controlled win rates, costs less than $10 in OpenAI credits, and runs in under 3 minutes.
+
+
+
+
Evaluation Stats
Total Models4
Organizations2
Verified Results0
Self-Reported4
+
+
+
+
Benchmark Details
Max Score1
Language
en
+
+
+
+
Performance Overview
Score distribution and top performers
Score Distribution
4 models
Top Score
62.7%
Average Score
52.8%
High Performers (80%+)
0Top Organizations
#1IBM
3 models
53.5%
#2DeepSeek
1 model
50.5%
+
+
+
+
Leaderboard
4 models ranked by performance on AlpacaEval 2.0
License | Links | ||||
---|---|---|---|---|---|
Apr 16, 2025 | Apache 2.0 | 62.7% | |||
Apr 16, 2025 | Apache 2.0 | 62.7% | |||
May 8, 2024 | deepseek | 50.5% | |||
May 2, 2025 | Apache 2.0 | 35.2% |