Zero2Text:针对文本嵌入的零训练跨域逆向攻击 / Zero2Text: Zero-Training Cross-Domain Inversion Attacks on Textual Embeddings
1️⃣ 一句话总结
这篇论文提出了一种名为Zero2Text的新型攻击方法,它无需任何训练数据,就能从AI模型存储的文本向量中高效地还原出原始文本,即使在跨领域和严格的黑盒场景下也有效,并且现有的主流防御手段难以抵挡这种攻击。
The proliferation of retrieval-augmented generation (RAG) has established vector databases as critical infrastructure, yet they introduce severe privacy risks via embedding inversion attacks. Existing paradigms face a fundamental trade-off: optimization-based methods require computationally prohibitive queries, while alignment-based approaches hinge on the unrealistic assumption of accessible in-domain training data. These constraints render them ineffective in strict black-box and cross-domain settings. To dismantle these barriers, we introduce Zero2Text, a novel training-free framework based on recursive online alignment. Unlike methods relying on static datasets, Zero2Text synergizes LLM priors with a dynamic ridge regression mechanism to iteratively align generation to the target embedding on-the-fly. We further demonstrate that standard defenses, such as differential privacy, fail to effectively mitigate this adaptive threat. Extensive experiments across diverse benchmarks validate Zero2Text; notably, on MS MARCO against the OpenAI victim model, it achieves 1.8x higher ROUGE-L and 6.4x higher BLEU-2 scores compared to baselines, recovering sentences from unknown domains without a single leaked data pair.
Zero2Text:针对文本嵌入的零训练跨域逆向攻击 / Zero2Text: Zero-Training Cross-Domain Inversion Attacks on Textual Embeddings
这篇论文提出了一种名为Zero2Text的新型攻击方法,它无需任何训练数据,就能从AI模型存储的文本向量中高效地还原出原始文本,即使在跨领域和严格的黑盒场景下也有效,并且现有的主流防御手段难以抵挡这种攻击。
源自 arXiv: 2602.01757