菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-17
📄 Abstract - Can Generative Artificial Intelligence Survive Data Contamination? Theoretical Guarantees under Contaminated Recursive Training

Generative Artificial Intelligence (AI), such as large language models (LLMs), has become a transformative force across science, industry, and society. As these systems grow in popularity, web data becomes increasingly interwoven with this AI-generated material and it is increasingly difficult to separate them from naturally generated content. As generative models are updated regularly, later models will inevitably be trained on mixtures of human-generated data and AI-generated data from earlier versions, creating a recursive training process with data contamination. Existing theoretical work has examined only highly simplified settings, where both the real data and the generative model are discrete or Gaussian, where it has been shown that such recursive training leads to model collapse. However, real data distributions are far more complex, and modern generative models are far more flexible than Gaussian and linear mechanisms. To fill this gap, we study recursive training in a general framework with minimal assumptions on the real data distribution and allow the underlying generative model to be a general universal approximator. In this framework, we show that contaminated recursive training still converges, with a convergence rate equal to the minimum of the baseline model's convergence rate and the fraction of real data used in each iteration. To the best of our knowledge, this is the first (positive) theoretical result on recursive training without distributional assumptions on the data. We further extend the analysis to settings where sampling bias is present in data collection and support all theoretical results with empirical studies.

顶级标签: theory model training machine learning
详细标签: data contamination recursive training generative models convergence analysis theoretical guarantees 或 搜索:

生成式人工智能能在数据污染中存活吗?污染递归训练下的理论保证 / Can Generative Artificial Intelligence Survive Data Contamination? Theoretical Guarantees under Contaminated Recursive Training


1️⃣ 一句话总结

这篇论文首次在理论上证明,即使生成式AI模型在训练中混入了自己早期版本产生的数据(即数据污染),只要每次迭代都包含一定比例的真实人类数据,整个递归训练过程最终仍会收敛,而不会完全崩溃。

源自 arXiv: 2602.16065