菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-11
📄 Abstract - Less is Enough: Synthesizing Diverse Data in Feature Space of LLMs

The diversity of post-training data is critical for effective downstream performance in large language models (LLMs). Many existing approaches to constructing post-training data quantify diversity using text-based metrics that capture linguistic variation, but such metrics provide only weak signals for the task-relevant features that determine downstream performance. In this work, we introduce Feature Activation Coverage (FAC) which measures data diversity in an interpretable feature space. Building upon this metric, we further propose a diversity-driven data synthesis framework, named FAC Synthesis, that first uses a sparse autoencoder to identify missing features from a seed dataset, and then generates synthetic samples that explicitly reflect these features. Experiments show that our approach consistently improves both data diversity and downstream performance on various tasks, including instruction following, toxicity detection, reward modeling, and behavior steering. Interestingly, we identify a shared, interpretable feature space across model families (i.e., LLaMA, Mistral, and Qwen), enabling cross-model knowledge transfer. Our work provides a solid and practical methodology for exploring data-centric optimization of LLMs.

顶级标签: llm model training data
详细标签: data synthesis feature activation sparse autoencoder post-training knowledge transfer 或 搜索:

少即是够:在大型语言模型特征空间中合成多样化数据 / Less is Enough: Synthesizing Diverse Data in Feature Space of LLMs


1️⃣ 一句话总结

这篇论文提出了一种通过分析模型内部特征来合成多样化训练数据的新方法,能有效提升大语言模型在多种任务上的性能,并且发现不同模型家族之间存在可共享的通用特征空间。

源自 arXiv: 2602.10388