LMMRec: LLM-driven Motivation-aware Multimodal Recommendation
AI 摘要
LMMRec利用LLM提取动机,融合多模态信息,提升推荐系统性能。
主要贡献
- 提出LMMRec框架,利用LLM理解用户和物品动机
- 采用双编码器结构和对比学习,实现跨模态对齐
- 设计动机协调策略和交互-文本对应方法,降低噪声和语义漂移
方法论
使用LLM提取文本动机,通过双编码器建模文本和交互动机,利用对比学习进行跨模态对齐和噪声抑制。
原文摘要
Motivation-based recommendation systems uncover user behavior drivers. Motivation modeling, crucial for decision-making and content preference, explains recommendation generation. Existing methods often treat motivation as latent variables from interaction data, neglecting heterogeneous information like review text. In multimodal motivation fusion, two challenges arise: 1) achieving stable cross-modal alignment amid noise, and 2) identifying features reflecting the same underlying motivation across modalities. To address these, we propose LLM-driven Motivation-aware Multimodal Recommendation (LMMRec), a model-agnostic framework leveraging large language models for deep semantic priors and motivation understanding. LMMRec uses chain-of-thought prompting to extract fine-grained user and item motivations from text. A dual-encoder architecture models textual and interaction-based motivations for cross-modal alignment, while Motivation Coordination Strategy and Interaction-Text Correspondence Method mitigate noise and semantic drift through contrastive learning and momentum updates. Experiments on three datasets show LMMRec achieves up to a 4.98\% performance improvement.