菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-03
📄 Abstract - Implicit Bias in Deep Linear Discriminant Analysis

While the Implicit Bias(or Implicit Regularization) of standard loss functions has been studied, the optimization geometry induced by discriminative metric-learning objectives remains largely this http URL the best of our knowledge, this paper presents an initial theoretical analysis of the implicit regularization induced by the Deep LDA,a scale invariant objective designed to minimize intraclass variance and maximize interclass distance. By analyzing the gradient flow of the loss on a L-layer diagonal linear network, we prove that under balanced initialization, the network architecture transforms standard additive gradient updates into multiplicative weight updates, which demonstrates an automatic conservation of the (2/L) quasi-norm.

顶级标签: machine learning theory model training
详细标签: implicit bias deep lda gradient flow linear networks optimization geometry 或 搜索:

深度线性判别分析中的隐式偏置 / Implicit Bias in Deep Linear Discriminant Analysis


1️⃣ 一句话总结

这篇论文首次从理论上分析了深度线性判别分析(Deep LDA)目标函数在优化过程中产生的隐式正则化效应,发现在平衡初始化条件下,多层对角线性网络的梯度流会自动保持一种特定的准范数,从而揭示了其内在的几何约束特性。

源自 arXiv: 2603.02622