规模化智能体可解释性:在企业担忧与XAI需求之间 / Agentic Explainability at Scale: Between Corporate Fears and XAI Needs
1️⃣ 一句话总结
本文针对企业在规模化部署自主AI智能体时因管理滞后而产生的‘智能体泛滥’风险,提出了结合设计时与运行时解释性技术的解决方案,并设计了一种‘智能体AI卡片’原型,以增强透明度和可控性,从而缓解企业对AI自主性的担忧。
As companies enter the race for agentic AI adoption, fears surface around agentic autonomy and its subsequent risks. These fears compound as companies scale their agentic AI adoption with low-code applications, without a comparable scaling in their governance processes and expertise resulting in a phenomenon known as "Agent Sprawl". While shadow AI tools can help with agentic discovery and identification, few observability tools offer insights into the agents' configuration and settings or the decision-making process during agent-to-agent communication and orchestration. This paper explores AI governance professionals' concerns in enterprise settings, while offering design-time and runtime explainability techniques as suggested by AI governance experts for addressing those fears. Finally, we provide a preliminary prototype of an Agentic AI Card that can help companies feel at ease deploying agents at scale.
规模化智能体可解释性:在企业担忧与XAI需求之间 / Agentic Explainability at Scale: Between Corporate Fears and XAI Needs
本文针对企业在规模化部署自主AI智能体时因管理滞后而产生的‘智能体泛滥’风险,提出了结合设计时与运行时解释性技术的解决方案,并设计了一种‘智能体AI卡片’原型,以增强透明度和可控性,从而缓解企业对AI自主性的担忧。
源自 arXiv: 2604.14984