菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-03
📄 Abstract - LaTeX Compilation: Challenges in the Era of LLMs

As large language models (LLMs) increasingly assist scientific writing, limitations and the significant token cost of TeX become more and more visible. This paper analyzes TeX's fundamental defects in compilation and user experience design to illustrate its limitations on compilation efficiency, generated semantics, error localization, and tool ecosystem in the era of LLMs. As an alternative, Mogan STEM, a WYSIWYG structured editor, is introduced. Mogan outperforms TeX in the above aspects by its efficient data structure, fast rendering, and on-demand plugin loading. Extensive experiments are conducted to verify the benefits on compilation/rendering time and performance in LLM tasks. What's more, we show that due to Mogan's lower information entropy, it is more efficient to use .tmu (the document format of Mogan) to fine-tune LLMs than TeX. Therefore, we launch an appeal for larger experiments on LLM training using the .tmu format.

顶级标签: llm natural language processing systems
详细标签: scientific writing document compilation structured editor information entropy fine-tuning 或 搜索:

LaTeX编译:大语言模型时代的挑战 / LaTeX Compilation: Challenges in the Era of LLMs


1️⃣ 一句话总结

这篇论文指出,在大语言模型辅助科学写作的时代,传统的LaTeX格式在编译效率、错误定位和AI训练成本上存在明显缺陷,并提出一种名为Mogan STEM的新型所见即所得编辑器,其文档格式能显著提升编译速度和降低AI模型训练成本。

源自 arXiv: 2603.02873