菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-16
📄 Abstract - Covariance-Aware Transformers for Quadratic Programming and Decision Making

We explore the use of transformers for solving quadratic programs and how this capability benefits decision-making problems that involve covariance matrices. We first show that the linear attention mechanism can provably solve unconstrained QPs by tokenizing the matrix variables (e.g.~$A$ of the objective $\frac{1}{2}x^\top Ax+b^\top x$) row-by-row and emulating gradient descent iterations. Furthermore, by incorporating MLPs, a transformer block can solve (i) $\ell_1$-penalized QPs by emulating iterative soft-thresholding and (ii) $\ell_1$-constrained QPs when equipped with an additional feedback loop. Our theory motivates us to introduce Time2Decide: a generic method that enhances a time series foundation model (TSFM) by explicitly feeding the covariance matrix between the variates. We empirically find that Time2Decide uniformly outperforms the base TSFM model for the classical portfolio optimization problem that admits an $\ell_1$-constrained QP formulation. Remarkably, Time2Decide also outperforms the classical "Predict-then-Optimize (PtO)" procedure, where we first forecast the returns and then explicitly solve a constrained QP, in suitable settings. Our results demonstrate that transformers benefit from explicit use of second-order statistics, and this can enable them to effectively solve complex decision-making problems, like portfolio construction, in one forward pass.

顶级标签: machine learning theory financial
详细标签: transformers quadratic programming portfolio optimization covariance decision making 或 搜索:

用于二次规划和决策的协方差感知Transformer模型 / Covariance-Aware Transformers for Quadratic Programming and Decision Making


1️⃣ 一句话总结

这篇论文提出了一种名为Time2Decide的新方法,通过让Transformer模型直接利用数据之间的协方差矩阵,使其能够有效解决二次规划问题,并在投资组合优化等复杂决策任务中,其单次前向传播的表现优于传统的“先预测后优化”两步法。

源自 arXiv: 2602.14506