通过多样性感知红队测试揭示视觉-语言-动作模型的语言脆弱性 / Uncovering Linguistic Fragility in Vision-Language-Action Models via Diversity-Aware Red Teaming
1️⃣ 一句话总结
这篇论文提出了一种新的多样性感知红队测试方法,能够系统性地发现并生成多种多样的、能导致机器人执行失败的自然语言指令,从而有效暴露当前先进的视觉-语言-动作模型在理解语言细微差别时存在的安全隐患。
Vision-Language-Action (VLA) models have achieved remarkable success in robotic manipulation. However, their robustness to linguistic nuances remains a critical, under-explored safety concern, posing a significant safety risk to real-world deployment. Red teaming, or identifying environmental scenarios that elicit catastrophic behaviors, is an important step in ensuring the safe deployment of embodied AI agents. Reinforcement learning (RL) has emerged as a promising approach in automated red teaming that aims to uncover these vulnerabilities. However, standard RL-based adversaries often suffer from severe mode collapse due to their reward-maximizing nature, which tends to converge to a narrow set of trivial or repetitive failure patterns, failing to reveal the comprehensive landscape of meaningful risks. To bridge this gap, we propose a novel \textbf{D}iversity-\textbf{A}ware \textbf{E}mbodied \textbf{R}ed \textbf{T}eaming (\textbf{DAERT}) framework, to expose the vulnerabilities of VLAs against linguistic variations. Our design is based on evaluating a uniform policy, which is able to generate a diverse set of challenging instructions while ensuring its attack effectiveness, measured by execution failures in a physical simulator. We conduct extensive experiments across different robotic benchmarks against two state-of-the-art VLAs, including $\pi_0$ and OpenVLA. Our method consistently discovers a wider range of more effective adversarial instructions that reduce the average task success rate from 93.33\% to 5.85\%, demonstrating a scalable approach to stress-testing VLA agents and exposing critical safety blind spots before real-world deployment.
通过多样性感知红队测试揭示视觉-语言-动作模型的语言脆弱性 / Uncovering Linguistic Fragility in Vision-Language-Action Models via Diversity-Aware Red Teaming
这篇论文提出了一种新的多样性感知红队测试方法,能够系统性地发现并生成多种多样的、能导致机器人执行失败的自然语言指令,从而有效暴露当前先进的视觉-语言-动作模型在理解语言细微差别时存在的安全隐患。
源自 arXiv: 2604.05595