🌱VitaBench: Benchmarking LLM Agents
with Versatile Interactive Tasks in Real-world Applications

Wei He†,*, Yueqing Sun*, Hongyan Hao*, Xueyuan Hao*, Zhikang Xia, Qi Gu,
Chengcheng Han, Dengchang Zhao, Hui Su, Kefeng Zhang, Man Gao, Xi Su,
Xiaodong Cai, Xunliang Cai, Yu Yang, Yunke Zhao
Meituan LongCat Team
†Correspondence to: whe23@m.fudan.edu.cn, guqi03@meituan.com
*Indicates equal contribution.

Abstract

As LLM-based agents are increasingly deployed in real-life scenarios, existing benchmarks fail to capture their inherent complexity of handling extensive information, leveraging diverse resources, and managing dynamic user interactions. To address this gap, we introduce VitaBench, a challenging benchmark that evaluates agents on versatile interactive tasks grounded in real-world settings. Drawing from daily applications in food delivery, in-store consumption, and online travel services, VitaBench presents agents with the most complex life-serving simulation environment to date, comprising 66 tools. Through a framework that eliminates domain-specific policies, we enable flexible composition of these scenarios and tools, yielding 100 cross-scenario tasks (main results) and 300 single-scenario tasks. Each task is derived from multiple real user requests and requires agents to reason across temporal and spatial dimensions, utilize complex tool sets, proactively clarify ambiguous instructions, and track shifting user intent throughout multi-turn conversations. Moreover, we propose a rubric-based sliding window evaluator, enabling robust assessment of diverse solution pathways in complex environments and stochastic interactions. Our comprehensive evaluation reveals that even the most advanced models achieve only 30% success rate on cross-scenario tasks, and less than 50% success rate on others. Overall, we believe VitaBench will serve as a valuable resource for advancing the development of AI agents in practical real-world applications.

Directional Weight Score

Overview

This plot summarizes per-model success rates under our two evaluation regimes. The most striking observation is the consistent and large performance gap: even the best model reaches only 48.3% success on the 300 single-scenario tasks, while performance falls to 30.0% on the 100 cross-scenario tasks. This sharp drop highlights challenges in domain switching, expanded tool selection, and long-horizon coordination. Relative model rankings are broadly preserved, but absolute success rates remain low—indicating substantial headroom for improvement.

Sizes of model trees

Method

VitaBench is built through a two-stage pipeline. Stage I (Framework Design) abstracts real life-serving scenarios — Delivery, In-store Consumption, and OTA — into a directed graph of simplified API tools with explicit pre-/post-conditions and inter-tool dependencies, thereby encoding domain rules directly into tool structures and enabling cross-domain composition. Stage II (Task Creation) constructs tasks from anonymized real user profiles, composite instructions, and realistic environments augmented with curated distractors and transaction histories. Each task is iteratively validated with human checks to ensure clarity while preserving multiple valid solutions, resulting in 400 tasks with comprehensive databases.

Overview of the VitaBench construction pipeline

🏆Leaderboard🏆

Here we present comprehensive evaluation results on VitaBench. Last updated on 2025-09-30. For more detailed analysis, please refer to our paper on arXiv.

Rank Models Cross-Scenarios Delivery In-store OTA
Avg
@4
Pass
@4
Pass
^4
Avg
@4
Pass
@4
Pass
^4
Avg
@4
Pass
@4
Pass
^4
Avg
@4
Pass
@4
Pass
^4
Thinking Models
1o3 (high)30.061.06.053.583.024.053.586.019.037.866.010.0
2Claude-4.1-Opus (w/ thinking)29.056.06.047.580.017.052.578.020.032.357.09.0
3LongCat-Flash-Thinking24.354.03.042.371.013.056.885.025.028.359.06.0
4Gemini-2.5-Pro23.553.05.049.081.016.043.878.012.026.554.06.0
5Claude-4-Sonnet (w/ thinking)23.051.06.046.078.015.051.580.021.029.055.09.0
6GLM-4.5 (w/ thinking)22.848.02.044.577.014.052.880.022.028.855.07.0
7o4-mini (high)19.549.01.044.580.015.046.581.015.023.550.05.0
8Qwen3-235B-A22B-Thinking-250718.845.02.044.078.09.046.080.09.017.541.02.0
9Doubao-Seed-1.6-Thinking17.042.01.030.359.010.043.378.010.018.045.02.0
10DeepSeek-R1-052814.539.00.040.372.011.041.379.07.013.032.02.0
11Gemini-2.5-Flash (think on)5.314.00.032.062.09.023.057.03.018.339.01.0
12Qwen3-32B (w/ thinking)5.024.00.022.853.04.026.560.03.07.318.01.0
Non-thinking Models
1Claude-4.1-Opus (w/o thinking)21.847.03.046.078.013.053.885.021.030.860.09.0
2Claude-4-Sonnet (w/o thinking)21.349.04.039.069.017.046.378.010.025.049.07.0
3LongCat-Flash-Chat20.345.02.039.571.015.050.584.015.022.849.02.0
4GLM-4.5 (w/o thinking)20.047.01.045.872.020.048.382.013.020.345.02.0
5DeepSeek-V3.1 (w/o thinking)16.340.01.034.067.06.042.576.07.018.347.01.0
6Kimi-K2-090515.539.02.035.368.09.042.578.010.022.046.04.0
7Qwen3-235B-A22B-Instruct-250714.338.00.034.366.06.044.887.013.020.045.01.0
8GPT-4.113.835.00.037.867.011.042.571.017.019.842.01.0
9Doubao-Seed-1.610.529.00.037.865.012.039.573.09.018.839.03.0
10Gemini-2.5-Flash (think off)5.817.01.031.065.06.022.846.03.018.544.01.0
11Qwen3-32B (w/o thinking)4.012.00.016.537.03.021.347.02.03.011.00.0
12GPT-5 (minimal)4.09.00.030.064.06.027.060.02.07.822.00.0
13DeepSeek-V3-03243.812.00.025.353.05.034.371.05.010.326.01.0

1We will periodically refresh the dataset by correcting errors, replacing outdated samples, and adding new challenging tasks. All leaderboard metrics are updated concurrently to reflect these changes.
2Due to API stability concerns, we are currently unable to evaluate some models for this benchmark. We are actively working to address these issues and include the latest models.
3While the tasks are grounded in real-world life-serving platforms where the majority of data is originally in Chinese, we are also preparing an English version of the dataset to facilitate broader research use.

BibTeX

@article{he2025vitabench,
    title={VitaBench: Benchmarking LLM Agents with Versatile Interactive Tasks in Real-world Applications}, 
    author={He, Wei and Sun, Yueqing and Hao, Hongyan and Hao, Xueyuan and Xia, Zhikang and Gu, Qi and Han, Chengcheng and Zhao, Dengchang and Su, Hui and Zhang, Kefeng and Gao, Man and Su, Xi and Cai, Xiaodong and Cai, Xunliang and Yang, Yu and Zhao, Yunke},
    journal={arXiv preprint arXiv:2509.26490},
    year={2025}
}