MRCR

text
+
+
+
+
About

MRCR (Multi-hop Reasoning for Conversational Recommendation) is a benchmark designed to evaluate models' ability to perform multi-step reasoning in conversational recommendation scenarios. It tests systems' capability to understand user preferences, conduct multi-hop reasoning across knowledge graphs, and provide accurate recommendations through natural dialogue interactions.

+
+
+
+
Evaluation Stats
Total Models6
Organizations1
Verified Results0
Self-Reported6
+
+
+
+
Benchmark Details
Max Score1
Language
en
+
+
+
+
Performance Overview
Score distribution and top performers

Score Distribution

6 models
Top Score
93.0%
Average Score
67.2%
High Performers (80%+)
2

Top Organizations

#1Google
6 models
67.2%
+
+
+
+
Leaderboard
6 models ranked by performance on MRCR
LicenseLinks
May 20, 2025
Proprietary
93.0%
May 1, 2024
Proprietary
82.6%
May 1, 2024
Proprietary
71.9%
Dec 1, 2024
Proprietary
69.2%
Mar 15, 2024
Proprietary
54.7%
May 20, 2025
Proprietary
32.0%
+
+
+
+
Resources