
Phi-3.5-mini-instruct
Zero-eval
#1Qasper
#1SQuALITY
#1QMSum
+11 more
by Microsoft
+
+
+
+
About
Phi-3.5-mini-instruct is a language model developed by Microsoft. The model shows competitive results across 31 benchmarks. It excels particularly in GSM8k (86.2%), ARC-C (84.6%), RULER (84.1%). It supports a 256K token context window for handling large documents. The model is available through 1 API provider. It's licensed for commercial use, making it suitable for enterprise applications. Released in 2024, it represents Microsoft's latest advancement in AI technology.
+
+
+
+
Pricing Range
Input (per 1M)$0.10 -$0.10
Output (per 1M)$0.10 -$0.10
Providers1
+
+
+
+
Timeline
AnnouncedAug 23, 2024
ReleasedAug 23, 2024
+
+
+
+
Specifications
Training Tokens3.4T
+
+
+
+
License & Family
License
MIT
Performance Overview
Performance metrics and category breakdown
Overall Performance
31 benchmarks
Average Score
58.7%
Best Score
86.2%
High Performers (80%+)
4Performance Metrics
Max Context Window
256.0KAvg Throughput
23.0 tok/sAvg Latency
1ms+
+
+
+
All Benchmark Results for Phi-3.5-mini-instruct
Complete list of benchmark scores with detailed information
GSM8k | text | 0.86 | 86.2% | Self-reported | |
ARC-C | text | 0.85 | 84.6% | Self-reported | |
RULER | text | 0.84 | 84.1% | Self-reported | |
PIQA | text | 0.81 | 81.0% | Self-reported | |
OpenBookQA | text | 0.79 | 79.2% | Self-reported | |
BoolQ | text | 0.78 | 78.0% | Self-reported | |
RepoQA | text | 0.77 | 77.0% | Self-reported | |
Social IQa | text | 0.75 | 74.7% | Self-reported | |
MEGA XStoryCloze | text | 0.73 | 73.5% | Self-reported | |
MBPP | text | 0.70 | 69.6% | Self-reported |
Showing 1 to 10 of 31 benchmarks