发现大语言模型中的解耦功能模块 / Discovering Decoupled Functional Modules in Large Language Models
1️⃣ 一句话总结
这项研究提出了一种无监督框架,能够自动将大语言模型中复杂的神经元网络分解成多个独立、可解释的功能模块,从而帮助我们更好地理解模型内部的工作机制。
Understanding the internal functional organization of Large Language Models (LLMs) is crucial for improving their trustworthiness and performance. However, how LLMs organize different functions into modules remains highly unexplored. To bridge this gap, we formulate a functional module discovery problem and propose an Unsupervised LLM Cross-layer MOdule Discovery (ULCMOD) framework that simultaneously disentangles the large set of neurons in the entire LLM into modules while discovering the topics of input samples related to these modules. Our framework introduces a novel objective function and an efficient Iterative Decoupling (IterD) algorithm. Extensive experiments show that our method discovers high-quality, disentangled modules that capture more meaningful semantic information and achieve superior performance in various downstream tasks. Moreover, our qualitative analysis reveals that the discovered modules show semantic coherence, correspond to interpretable specializations, and a clear spatial and hierarchical organization within the LLM. Our work provides a novel tool for interpreting the functional modules of LLMs, filling a critical blank in LLM's interpretability research.
发现大语言模型中的解耦功能模块 / Discovering Decoupled Functional Modules in Large Language Models
这项研究提出了一种无监督框架,能够自动将大语言模型中复杂的神经元网络分解成多个独立、可解释的功能模块,从而帮助我们更好地理解模型内部的工作机制。
源自 arXiv: 2603.17823