Multimodal Learning 相关度: 8/10

IncompeBench: A Permissively Licensed, Fine-Grained Benchmark for Music Information Retrieval

Benjamin Clavié, Atoof Shakir, Jonah Turner, Sean Lee, Aamir Shakir, Makoto P. Kato
arXiv: 2602.11941v1 发布: 2026-02-12 更新: 2026-02-12

AI 摘要

提出了IncompeBench,一个用于音乐信息检索的高质量、开放许可基准测试集。

主要贡献

  • 构建了包含1574个音乐片段、500个查询和超过125000个相关性判断的基准测试集
  • 使用了多阶段流程,确保了高质量的人工标注数据
  • 数据集以开放许可发布,便于研究人员使用

方法论

采用多阶段标注流程,并进行人工标注,构建高质量的音乐检索数据集,并公开。

原文摘要

Multimodal Information Retrieval has made significant progress in recent years, leveraging the increasingly strong multimodal abilities of deep pre-trained models to represent information across modalities. Music Information Retrieval (MIR), in particular, has considerably increased in quality, with neural representations of music even making its way into everyday life products. However, there is a lack of high-quality benchmarks for evaluating music retrieval performance. To address this issue, we introduce \textbf{IncompeBench}, a carefully annotated benchmark comprising $1,574$ permissively licensed, high-quality music snippets, $500$ diverse queries, and over $125,000$ individual relevance judgements. These annotations were created through the use of a multi-stage pipeline, resulting in high agreement between human annotators and the generated data. The resulting datasets are publicly available at https://huggingface.co/datasets/mixedbread-ai/incompebench-strict and https://huggingface.co/datasets/mixedbread-ai/incompebench-lenient with the prompts available at https://github.com/mixedbread-ai/incompebench-programs.

标签

Music Information Retrieval Benchmark Multimodal Learning

arXiv 分类

cs.IR cs.AI