用于平衡个性化的长期嵌入 / Long-Term Embeddings for Balanced Personalization
1️⃣ 一句话总结
这篇论文提出了一种名为‘长期嵌入’的新方法,通过将稳定的用户长期偏好作为固定语义锚点,有效解决了现有推荐系统过于关注近期行为而忽略长期兴趣的问题,并在实际应用中显著提升了用户参与度和商业指标。
Modern transformer-based sequential recommenders excel at capturing short-term intent but often suffer from recency bias, overlooking stable long-term preferences. While extending sequence lengths is an intuitive fix, it is computationally inefficient, and recent interactions tend to dominate the model's attention. We propose Long-Term Embeddings (LTE) as a high-inertia contextual anchor to bridge this gap. We address a critical production challenge: the point-in-time consistency problem caused by infrastructure constraints, as feature stores typically host only a single "live" version of features. This leads to an offline-online mismatch during model deployments and rollbacks, as models are forced to process evolved representations they never saw during training. To resolve this, we introduce an LTE framework that constrains embeddings to a fixed semantic basis of content-based item representations, ensuring cross-version compatibility. Furthermore, we investigate integration strategies for causal language modeling, considering the data leakage issue that occurs when the LTE and the transformer's short-term sequence share a temporal horizon. We evaluate two representations: a heuristic average and an asymmetric autoencoder with a fixed decoder grounded in the semantic basis to enable behavioral fine-tuning while maintaining stability. Online A/B tests on Zalando demonstrate that integrating LTE as a contextual prefix token using a lagged window yields significant uplifts in both user engagement and financial metrics.
用于平衡个性化的长期嵌入 / Long-Term Embeddings for Balanced Personalization
这篇论文提出了一种名为‘长期嵌入’的新方法,通过将稳定的用户长期偏好作为固定语义锚点,有效解决了现有推荐系统过于关注近期行为而忽略长期兴趣的问题,并在实际应用中显著提升了用户参与度和商业指标。
源自 arXiv: 2604.08181