菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-27
📄 Abstract - LAVA: Layered Audio-Visual Anti-tampering Watermarking for Robust Deepfake Detection and Localization

Proactive watermarking offers a promising approach for deepfake tamper detection and localization in short-form videos. However, existing methods often decouple audio and visual evidence and assume that watermark signals remain reliable under real-world degradations, making tamper localization vulnerable to multimodal misalignment and compression distortions. Moreover, existing semi-fragile visual watermarking methods often degrade significantly under codec compression because their embedding bands overlap with compression-sensitive frequency regions. To address these limitations, we propose Layered Audio-Visual Anti-tampering Watermarking (LAVA), a calibration-aware audio-visual watermark fusion framework for deepfake tamper detection and localization. LAVA leverages cross-modal watermark fusion and calibration-aware alignment to preserve consistent and reliable tamper evidence under compression and audio-visual asynchrony, enabling robust tamper localization. Extensive experiments demonstrate that LAVA achieves near-perfect detection performance (AP = 0.999), remains robust to compression and multimodal misalignment, and significantly improves tamper localization reliability over existing audio-visual fusion baselines.

顶级标签: multi-modal video
详细标签: deepfake detection watermarking tamper localization audio-visual fusion compression robustness 或 搜索:

LAVA:分层视听抗篡改水印,用于稳健的深度伪造检测与定位 / LAVA: Layered Audio-Visual Anti-tampering Watermarking for Robust Deepfake Detection and Localization


1️⃣ 一句话总结

本文提出了一种名为LAVA的新型视听融合水印方法,通过跨模态水印融合与校准对齐技术,有效抵抗视频压缩和音画不同步等现实干扰,实现了对深度伪造视频近乎完美的检测(准确率99.9%)和更可靠的篡改区域定位。

源自 arXiv: 2604.23957