大语言模型的格表示假说 / The Lattice Representation Hypothesis of Large Language Models
1️⃣ 一句话总结
这篇论文提出了一个假说,认为大语言模型内部通过几何结构(如向量空间中的半空间交集)自然地编码了概念之间的层次与逻辑关系,从而在连续的神经网络表示和离散的符号推理之间建立了桥梁。
We propose the Lattice Representation Hypothesis of large language models: a symbolic backbone that grounds conceptual hierarchies and logical operations in embedding geometry. Our framework unifies the Linear Representation Hypothesis with Formal Concept Analysis (FCA), showing that linear attribute directions with separating thresholds induce a concept lattice via half-space intersections. This geometry enables symbolic reasoning through geometric meet (intersection) and join (union) operations, and admits a canonical form when attribute directions are linearly independent. Experiments on WordNet sub-hierarchies provide empirical evidence that LLM embeddings encode concept lattices and their logical structure, revealing a principled bridge between continuous geometry and symbolic abstraction.
大语言模型的格表示假说 / The Lattice Representation Hypothesis of Large Language Models
这篇论文提出了一个假说,认为大语言模型内部通过几何结构(如向量空间中的半空间交集)自然地编码了概念之间的层次与逻辑关系,从而在连续的神经网络表示和离散的符号推理之间建立了桥梁。
源自 arXiv: 2603.01227