基于预测嵌入的多模态潜在推理 / Multimodal Latent Reasoning via Predictive Embeddings
1️⃣ 一句话总结
这篇论文提出了一种名为Pearl的新方法,它让视觉语言模型在‘潜在空间’里学习使用外部工具(如裁剪、深度估算)的经验,从而在推理时无需实际调用这些工具就能提升图像理解能力,既高效又减少了错误。
Tool-augmented multimodal reasoning enables visual language models (VLMs) to improve perception by interacting with external tools (e.g., cropping, depth estimation). However, such approaches incur substantial inference overhead, require specialized supervision, and are prone to erroneous tool calls. We propose Pearl (Predictive Embedding Alignment for Reasoning in Latent space), a JEPA-inspired framework that learns from expert tool-use trajectories entirely in the latent space, eliminating the need for explicit tool invocation at inference time. Unlike reconstruction-based latent reasoning methods, which autoregressively generate latent tokens and suffer from training-inference mismatch and limited support for multi-step tool use, Pearl directly learns predictive embeddings from multimodal trajectories while preserving the standard vision-language generation pipeline: it is model-agnostic, simple to train, and naturally supports trajectories with multiple tool calls. Experiments across multiple perception benchmarks show that Pearl matches or outperforms standard supervised fine-tuning and reconstruction-based latent reasoning approaches. Furthermore, we provide empirical evidence that reconstruction-based methods primarily learn embeddings rather than image edits in latent space, motivating predictive embedding learning as a more principled alternative.
基于预测嵌入的多模态潜在推理 / Multimodal Latent Reasoning via Predictive Embeddings
这篇论文提出了一种名为Pearl的新方法,它让视觉语言模型在‘潜在空间’里学习使用外部工具(如裁剪、深度估算)的经验,从而在推理时无需实际调用这些工具就能提升图像理解能力,既高效又减少了错误。
源自 arXiv: 2604.08065