菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-09
📄 Abstract - MEGC2026: Micro-Expression Grand Challenge on Visual Question Answering

Facial micro-expressions (MEs) are involuntary movements of the face that occur spontaneously when a person experiences an emotion but attempts to suppress or repress the facial expression, typically found in a high-stakes environment. In recent years, substantial advancements have been made in the areas of ME recognition, spotting, and generation. The emergence of multimodal large language models (MLLMs) and large vision-language models (LVLMs) offers promising new avenues for enhancing ME analysis through their powerful multimodal reasoning capabilities. The ME grand challenge (MEGC) 2026 introduces two tasks that reflect these evolving research directions: (1) ME video question answering (ME-VQA), which explores ME understanding through visual question answering on relatively short video sequences, leveraging MLLMs or LVLMs to address diverse question types related to MEs; and (2) ME long-video question answering (ME-LVQA), which extends VQA to long-duration video sequences in realistic settings, requiring models to handle temporal reasoning and subtle micro-expression detection across extended time periods. All participating algorithms are required to submit their results on a public leaderboard. More details are available at this https URL.

顶级标签: multi-modal computer vision benchmark
详细标签: micro-expression visual question answering multimodal llm video analysis temporal reasoning 或 搜索:

MEGC2026:关于视觉问答的微表情大挑战 / MEGC2026: Micro-Expression Grand Challenge on Visual Question Answering


1️⃣ 一句话总结

这篇论文介绍了2026年微表情大挑战赛,它利用新兴的多模态大模型技术,设置了短时和长时微表情视频问答两项任务,旨在推动微表情自动分析领域的发展。

源自 arXiv: 2603.08927