菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-11
📄 Abstract - Protecting Context and Prompts: Deterministic Security for Non-Deterministic AI

Large Language Model (LLM) applications are vulnerable to prompt injection and context manipulation attacks that traditional security models cannot prevent. We introduce two novel primitives--authenticated prompts and authenticated context--that provide cryptographically verifiable provenance across LLM workflows. Authenticated prompts enable self-contained lineage verification, while authenticated context uses tamper-evident hash chains to ensure integrity of dynamic inputs. Building on these primitives, we formalize a policy algebra with four proven theorems providing protocol-level Byzantine resistance--even adversarial agents cannot violate organizational policies. Five complementary defenses--from lightweight resource controls to LLM-based semantic validation--deliver layered, preventative security with formal guarantees. Evaluation against representative attacks spanning 6 exhaustive categories achieves 100% detection with zero false positives and nominal overhead. We demonstrate the first approach combining cryptographically enforced prompt lineage, tamper-evident context, and provable policy reasoning--shifting LLM security from reactive detection to preventative guarantees.

顶级标签: llm systems security
详细标签: prompt injection cryptographic provenance byzantine resistance formal guarantees preventative security 或 搜索:

保护上下文与提示:为非确定性AI提供确定性安全 / Protecting Context and Prompts: Deterministic Security for Non-Deterministic AI


1️⃣ 一句话总结

这篇论文提出了一种结合密码学验证和形式化证明的新方法,通过确保提示的来源可信和上下文不被篡改,为大语言模型应用提供了主动预防攻击的安全保障,而非事后检测。

源自 arXiv: 2602.10481