菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-01-06
📄 Abstract - Enhancing Linguistic Competence of Language Models through Pre-training with Language Learning Tasks

Language models (LMs) are pre-trained on raw text datasets to generate text sequences token-by-token. While this approach facilitates the learning of world knowledge and reasoning, it does not explicitly optimize for linguistic competence. To bridge this gap, we propose L2T, a pre-training framework integrating Language Learning Tasks alongside standard next-token prediction. Inspired by human language acquisition, L2T transforms raw text into structured input-output pairs to provide explicit linguistic stimulation. Pre-training LMs on a mixture of raw text and L2T data not only improves overall performance on linguistic competence benchmarks but accelerates its acquisition, while maintaining competitive performance on general reasoning tasks.

顶级标签: llm model training natural language processing
详细标签: pre-training linguistic competence language acquisition next-token prediction structured data 或 搜索:

通过语言学习任务预训练增强语言模型的语言能力 / Enhancing Linguistic Competence of Language Models through Pre-training with Language Learning Tasks


1️⃣ 一句话总结

这篇论文提出了一个名为L2T的新预训练框架,它通过模仿人类语言学习的方式,在常规文本预测任务之外,额外加入专门的语言学习任务来训练AI模型,从而让AI更快、更好地掌握语言规则,同时不影响其通用的推理能力。

源自 arXiv: 2601.03448