菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-19
📄 Abstract - Transformers Learn Robust In-Context Regression under Distributional Uncertainty

Recent work has shown that Transformers can perform in-context learning for linear regression under restrictive assumptions, including i.i.d. data, Gaussian noise, and Gaussian regression coefficients. However, real-world data often violate these assumptions: the distributions of inputs, noise, and coefficients are typically unknown, non-Gaussian, and may exhibit dependency across the prompt. This raises a fundamental question: can Transformers learn effectively in-context under realistic distributional uncertainty? We study in-context learning for noisy linear regression under a broad range of distributional shifts, including non-Gaussian coefficients, heavy-tailed noise, and non-i.i.d. prompts. We compare Transformers against classical baselines that are optimal or suboptimal under the corresponding maximum-likelihood criteria. Across all settings, Transformers consistently match or outperform these baselines, demonstrating robust in-context adaptation beyond classical estimators.

顶级标签: llm theory model evaluation
详细标签: in-context learning distributional robustness linear regression transformers uncertainty 或 搜索:

Transformer在分布不确定性下学习稳健的上下文回归 / Transformers Learn Robust In-Context Regression under Distributional Uncertainty


1️⃣ 一句话总结

这篇论文发现,即使在输入、噪声和系数分布未知、非高斯且存在依赖性的现实复杂数据中,Transformer模型也能通过上下文学习,稳健地进行线性回归,其表现达到或超过了传统最优估计器的水平。

源自 arXiv: 2603.18564