Comprehensive side-by-side LLM comparison
UI-TARS-72B-DPO supports multimodal inputs. Both models have their strengths depending on your specific coding needs.
Mistral AI
Devstral 2, released by Mistral AI on December 9, 2025, is a 123 billion parameter dense transformer model specifically designed for software engineering tasks. It features a 256K token context window and achieved 72.2% on SWE-bench Verified at release, making it a competitive open-weight option for automated coding and agentic development. Devstral 2 targets code generation, multi-file software engineering, and agentic development workflows under a modified MIT license.
ByteDance
UI-TARS-72B-DPO, released by ByteDance in early 2025, is a 72 billion parameter multimodal large language model from the UI-TARS family, built on Qwen-2-VL and fine-tuned for automated GUI interaction and computer control. It features native understanding of screenshots, UI elements, and web interfaces, achieving strong results across GUI benchmarks for perception, grounding, and agentic control. UI-TARS-72B-DPO targets computer-use agents, web automation, and applications requiring robust visual UI reasoning.
11 months newer
UI-TARS-72B-DPO
ByteDance
2025-01

Devstral-2-123B
Mistral AI
2025-12-09
Context window and performance specifications
Available providers and their performance metrics
Devstral-2-123B
OpenRouter
UI-TARS-72B-DPO
Devstral-2-123B
UI-TARS-72B-DPO
Devstral-2-123B
UI-TARS-72B-DPO