EC-Bench: Enumeration and Counting Benchmark for Ultra-Long Videos
AI 摘要
EC-Bench:长视频计数基准,挑战现有MLLM在长时间序列推理上的能力。
主要贡献
- 提出了EC-Bench,一个长视频枚举和计数基准。
- EC-Bench包含超过30分钟的长视频和相应的枚举证据。
- 评估了22个MLLM在EC-Bench上的性能,发现与人类性能差距较大。
方法论
构建包含长视频和查询对的基准数据集,并使用枚举、计数和时间证据定位作为评估指标,评估MLLM性能。
原文摘要
Counting in long videos remains a fundamental yet underexplored challenge in computer vision. Real-world recordings often span tens of minutes or longer and contain sparse, diverse events, making long-range temporal reasoning particularly difficult. However, most existing video counting benchmarks focus on short clips and evaluate only the final numerical answer, providing little insight into what should be counted or whether models consistently identify relevant instances across time. We introduce EC-Bench, a benchmark that jointly evaluates enumeration, counting, and temporal evidence grounding in long-form videos. EC-Bench contains 152 videos longer than 30 minutes and 1,699 queries paired with explicit evidence spans. Across 22 multimodal large language models (MLLMs), the best model achieves only 29.98% accuracy on Enumeration and 23.74% on Counting, while human performance reaches 78.57% and 82.97%, respectively. Our analysis reveals strong relationships between enumeration accuracy, temporal grounding, and counting performance. These results highlight fundamental limitations of current MLLMs and establish EC-Bench as a challenging benchmark for long-form quantitative video reasoning.