菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-06
📄 Abstract - Just Pass Twice: Efficient Token Classification with LLMs for Zero-Shot NER

Large language models encode extensive world knowledge valuable for zero-shot named entity recognition. However, their causal attention mechanism, where tokens attend only to preceding context, prevents effective token classification when disambiguation requires future context. Existing approaches use LLMs generatively, prompting them to list entities or produce structured outputs, but suffer from slow autoregressive decoding, hallucinated entities, and formatting errors. We propose Just Pass Twice (JPT), a simple yet effective method that enables causal LLMs to perform discriminative token classification with full bidirectional context. Our key insight is that concatenating the input to itself lets each token in the second pass attend to the complete sentence, requiring no architectural modifications. We combine these representations with definition-guided entity embeddings for flexible zero-shot generalization. Our approach achieves state-of-the-art results on zero-shot NER benchmarks, surpassing the previous best method by +7.9 F1 on average across CrossNER and MIT benchmarks, being over 20x faster than comparable generative methods.

顶级标签: llm natural language processing model evaluation
详细标签: zero-shot ner token classification efficiency causal attention bidirectional context 或 搜索:

只需两次前向传播:利用大语言模型进行高效零样本命名实体识别的标记分类方法 / Just Pass Twice: Efficient Token Classification with LLMs for Zero-Shot NER


1️⃣ 一句话总结

这篇论文提出了一种名为‘只需两次前向传播’的简单高效方法,通过将输入句子自我拼接,使原本只能单向关注上下文的大语言模型能够利用完整的双向信息进行零样本命名实体识别,从而在显著提升准确率的同时,速度比现有生成式方法快20倍以上。

源自 arXiv: 2604.05158