菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-07
📄 Abstract - The Unreasonable Effectiveness of Data for Recommender Systems

In recommender systems, collecting, storing, and processing large-scale interaction data is increasingly costly in terms of time, energy, and computation, yet it remains unclear when additional data stops providing meaningful gains. This paper investigates how offline recommendation performance evolves as the size of the training dataset increases and whether a saturation point can be observed. We implemented a reproducible Python evaluation workflow with two established toolkits, LensKit and RecBole, included 11 large public datasets with at least 7 million interactions, and evaluated 10 tool-algorithm combinations. Using absolute stratified user sampling, we trained models on nine sample sizes from 100,000 to 100,000,000 interactions and measured NDCG@10. Overall, raw NDCG usually increased with sample size, with no observable saturation point. To make result groups comparable, we applied min-max normalization within each group, revealing a clear positive trend in which around 75% of the points at the largest completed sample size also achieved the group's best observed performance. A late-stage slope analysis over the final 10-30% of each group further supported this upward trend: the interquartile range remained entirely non-negative with a median near 1.0. In summary, for traditional recommender systems on typical user-item interaction data, incorporating more training data remains primarily beneficial, while weaker scaling behavior is concentrated in atypical dataset cases and in the algorithmic outlier RecBole BPR under our setup.

顶级标签: systems model evaluation data
详细标签: recommender systems scaling laws data efficiency evaluation ndcg 或 搜索:

数据在推荐系统中难以置信的有效性 / The Unreasonable Effectiveness of Data for Recommender Systems


1️⃣ 一句话总结

这项研究发现,对于传统的推荐系统,持续增加训练数据量通常能持续提升推荐效果,在实验规模内并未观察到性能饱和点,表明数据规模仍是提升性能的关键因素。

源自 arXiv: 2604.06420