Comprehensive side-by-side LLM comparison
DeepSeek R1 Distill Qwen 32B leads with 16.8% higher average benchmark score. DeepSeek R1 Distill Qwen 32B offers 192.0K more tokens in context window than Mistral Small 3 24B Instruct. Both models have similar pricing. Overall, DeepSeek R1 Distill Qwen 32B is the stronger choice for coding tasks.
DeepSeek
DeepSeek-R1-Distill-Qwen-32B was created as a larger distilled variant, designed to transfer more of DeepSeek-R1's reasoning capabilities into a Qwen-based foundation. Built to serve applications requiring enhanced analytical depth, it represents a powerful option in the distilled reasoning model family.
Mistral AI
Mistral Small 24B Instruct was created as the instruction-tuned version of the 24B base model, designed to follow user instructions reliably. Built to serve general-purpose applications requiring moderate capability, it balances performance with deployment practicality.
10 days newer

DeepSeek R1 Distill Qwen 32B
DeepSeek
2025-01-20

Mistral Small 3 24B Instruct
Mistral AI
2025-01-30
Cost per million tokens (USD)

DeepSeek R1 Distill Qwen 32B

Mistral Small 3 24B Instruct
Context window and performance specifications
Average performance across 1 common benchmarks

DeepSeek R1 Distill Qwen 32B

Mistral Small 3 24B Instruct
Mistral Small 3 24B Instruct
2023-10-01
Available providers and their performance metrics

DeepSeek R1 Distill Qwen 32B
DeepInfra

Mistral Small 3 24B Instruct

DeepSeek R1 Distill Qwen 32B

Mistral Small 3 24B Instruct

DeepSeek R1 Distill Qwen 32B

Mistral Small 3 24B Instruct
DeepInfra
Mistral AI