
Llama 3.2 3B Instruct
Zero-eval
#1NIH/Multi-needle
#1InfiniteBench/En.MC
#1Open-rewrite
+2 more
by Meta
+
+
+
+
About
Llama 3.2 3B Instruct is a language model developed by Meta. The model shows competitive results across 15 benchmarks. It excels particularly in NIH/Multi-needle (84.7%), ARC-C (78.6%), GSM8k (77.7%). It supports a 256K token context window for handling large documents. The model is available through 1 API provider. Released in 2024, it represents Meta's latest advancement in AI technology.
+
+
+
+
Pricing Range
Input (per 1M)$0.01 -$0.01
Output (per 1M)$0.02 -$0.02
Providers1
+
+
+
+
Timeline
AnnouncedSep 25, 2024
ReleasedSep 25, 2024
+
+
+
+
Specifications
Training Tokens9.0T
+
+
+
+
License & Family
License
Llama 3.2 Community License
Performance Overview
Performance metrics and category breakdown
Overall Performance
15 benchmarks
Average Score
55.6%
Best Score
84.7%
High Performers (80%+)
1Performance Metrics
Max Context Window
256.0KAvg Throughput
171.5 tok/sAvg Latency
0ms+
+
+
+
All Benchmark Results for Llama 3.2 3B Instruct
Complete list of benchmark scores with detailed information
NIH/Multi-needle | text | 0.85 | 84.7% | Self-reported | |
ARC-C | text | 0.79 | 78.6% | Self-reported | |
GSM8k | text | 0.78 | 77.7% | Self-reported | |
IFEval | text | 0.77 | 77.4% | Self-reported | |
HellaSwag | text | 0.70 | 69.8% | Self-reported | |
BFCL v2 | text | 0.67 | 67.0% | Self-reported | |
MMLU | text | 0.63 | 63.4% | Self-reported | |
InfiniteBench/En.MC | text | 0.63 | 63.3% | Self-reported | |
MGSM | text | 0.58 | 58.2% | Self-reported | |
MATH | text | 0.48 | 48.0% | Self-reported |
Showing 1 to 10 of 15 benchmarks