Anthropic

Claude 3.5 Haiku

Zero-eval

by Anthropic

+
+
+
+
About

Claude 3.5 Haiku is a language model developed by Anthropic. It achieves strong performance with an average score of 60.8% across 9 benchmarks. It excels particularly in HumanEval (88.1%), MGSM (85.6%), DROP (83.1%). It supports a 400K token context window for handling large documents. The model is available through 3 API providers. Released in 2024, it represents Anthropic's latest advancement in AI technology.

+
+
+
+
Pricing Range
Input (per 1M)$0.80 -$1.00
Output (per 1M)$4.00 -$5.00
Providers3
+
+
+
+
Timeline
AnnouncedOct 22, 2024
ReleasedOct 22, 2024
+
+
+
+
License & Family
License
Proprietary
Performance Overview
Performance metrics and category breakdown

Overall Performance

9 benchmarks
Average Score
60.8%
Best Score
88.1%
High Performers (80%+)
3

Performance Metrics

Max Context Window
400.0K
Avg Throughput
82.0 tok/s
Avg Latency
0ms
+
+
+
+
All Benchmark Results for Claude 3.5 Haiku
Complete list of benchmark scores with detailed information
HumanEval
text
0.88
88.1%
Self-reported
MGSM
text
0.86
85.6%
Self-reported
DROP
text
0.83
83.1%
Self-reported
MATH
text
0.69
69.4%
Self-reported
MMLU-Pro
text
0.65
65.0%
Self-reported
TAU-bench Retail
text
0.51
51.0%
Self-reported
GPQA
text
0.42
41.6%
Self-reported
SWE-Bench Verified
text
0.41
40.6%
Self-reported
TAU-bench Airline
text
0.23
22.8%
Self-reported