菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-06
📄 Abstract - Multimodal Enhancement of Sequential Recommendation

We propose a novel recommender framework, MuSTRec (Multimodal and Sequential Transformer-based Recommendation), that unifies multimodal and sequential recommendation paradigms. MuSTRec captures cross-item similarities and collaborative filtering signals, by building item-item graphs from extracted text and visual features. A frequency-based self-attention module additionally captures the short- and long-term user preferences. Across multiple Amazon datasets, MuSTRec demonstrates superior performance (up to 33.5% improvement) over multimodal and sequential state-of-the-art baselines. Finally, we detail some interesting facets of this new recommendation paradigm. These include the need for a new data partitioning regime, and a demonstration of how integrating user embeddings into sequential recommendation leads to drastically increased short-term metrics (up to 200% improvement) on smaller datasets. Our code is availabe at this https URL and will be made publicly available.

顶级标签: multi-modal model training systems
详细标签: sequential recommendation multimodal transformer item-item graphs self-attention collaborative filtering 或 搜索:

多模态序列推荐的增强 / Multimodal Enhancement of Sequential Recommendation


1️⃣ 一句话总结

这篇论文提出了一个名为MuSTRec的新推荐系统框架,它通过结合物品的文本和图像信息来构建物品关系图,并利用注意力机制捕捉用户的长短期偏好,从而在多个数据集上显著超越了现有的先进推荐方法。

源自 arXiv: 2602.07207