多模态检索增强生成(mRAG)隐私的系统性评估 / A Systemic Evaluation of Multimodal RAG Privacy
1️⃣ 一句话总结
这篇论文通过实证研究发现,用于视觉任务的多模态检索增强生成(mRAG)系统在推理过程中存在泄露其背后私有数据集信息的风险,强调了为其开发隐私保护机制的必要性。
The growing adoption of multimodal Retrieval-Augmented Generation (mRAG) pipelines for vision-centric tasks (e.g. visual QA) introduces important privacy challenges. In particular, while mRAG provides a practical capability to connect private datasets to improve model performance, it risks the leakage of private information from these datasets during inference. In this paper, we perform an empirical study to analyze the privacy risks inherent in the mRAG pipeline observed through standard model prompting. Specifically, we implement a case study that attempts to infer the inclusion of a visual asset, e.g. image, in the mRAG, and if present leak the metadata, e.g. caption, related to it. Our findings highlight the need for privacy-preserving mechanisms and motivate future research on mRAG privacy.
多模态检索增强生成(mRAG)隐私的系统性评估 / A Systemic Evaluation of Multimodal RAG Privacy
这篇论文通过实证研究发现,用于视觉任务的多模态检索增强生成(mRAG)系统在推理过程中存在泄露其背后私有数据集信息的风险,强调了为其开发隐私保护机制的必要性。
源自 arXiv: 2601.17644