Multipl-E HumanEval

Multilingual
text
+
+
+
+
About

MultiPL-E HumanEval is a variant of the MultiPL-E benchmark specifically based on the HumanEval dataset, extending it to multiple programming languages. It evaluates code generation models' ability to solve the same programming problems across different languages, testing cross-language code generation consistency and programming language understanding.

+
+
+
+
Evaluation Stats
Total Models3
Organizations1
Verified Results0
Self-Reported3
+
+
+
+
Benchmark Details
Max Score1
Language
en
+
+
+
+
Performance Overview
Score distribution and top performers

Score Distribution

3 models
Top Score
75.2%
Average Score
63.8%
High Performers (80%+)
0

Top Organizations

#1Meta
3 models
63.8%
+
+
+
+
Leaderboard
3 models ranked by performance on Multipl-E HumanEval
LicenseLinks
Jul 23, 2024
Llama 3.1 Community License
75.2%
Jul 23, 2024
Llama 3.1 Community License
65.5%
Jul 23, 2024
Llama 3.1 Community License
50.8%
+
+
+
+
Resources