理解特征交互模型中的深度神经网络:维度坍缩的视角 / Understanding DNNs in Feature Interaction Models: A Dimensional Collapse Perspective
1️⃣ 一句话总结
本文从“维度坍缩”这一新视角出发,通过实验和理论分析证明,在特征交互推荐模型中,深度神经网络(DNN)的主要作用并非高效捕捉高阶特征交互,而是通过缓解嵌入表示的维度坍缩来提升模型的鲁棒性。
DNNs have gained widespread adoption in feature interaction recommendation models. However, there has been a longstanding debate on their roles. On one hand, some works claim that DNNs possess the ability to implicitly capture high-order feature interactions. Conversely, recent studies have highlighted the limitations of DNNs in effectively learning dot products, specifically second-order interactions, let alone higher-order interactions. In this paper, we present a novel perspective to understand the effectiveness of DNNs: their impact on the dimensional robustness of the representations. In particular, we conduct extensive experiments involving both parallel DNNs and stacked DNNs. Our evaluation encompasses an overall study of complete DNN on two feature interaction models, alongside a fine-grained ablation analysis of components within DNNs. Experimental results demonstrate that both parallel and stacked DNNs can effectively mitigate the dimensional collapse of embeddings. Furthermore, a gradient-based theoretical analysis, supported by empirical evidence, uncovers the underlying mechanisms of dimensional collapse.
理解特征交互模型中的深度神经网络:维度坍缩的视角 / Understanding DNNs in Feature Interaction Models: A Dimensional Collapse Perspective
本文从“维度坍缩”这一新视角出发,通过实验和理论分析证明,在特征交互推荐模型中,深度神经网络(DNN)的主要作用并非高效捕捉高阶特征交互,而是通过缓解嵌入表示的维度坍缩来提升模型的鲁棒性。
源自 arXiv: 2604.26489