菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-13
📄 Abstract - rPPG-VQA: A Video Quality Assessment Framework for Unsupervised rPPG Training

Unsupervised remote photoplethysmography (rPPG) promises to leverage unlabeled video data, but its potential is hindered by a critical challenge: training on low-quality "in-the-wild" videos severely degrades model performance. An essential step missing here is to assess the suitability of the videos for rPPG model learning before using them for the task. Existing video quality assessment (VQA) methods are mainly designed for human perception and not directly applicable to the above purpose. In this work, we propose rPPG-VQA, a novel framework for assessing video suitability for rPPG. We integrate signal-level and scene-level analyses and design a dual-branch assessment architecture. The signal-level branch evaluates the physiological signal quality of the videos via robust signal-to-noise ratio (SNR) estimation with a multi-method consensus mechanism, and the scene-level branch uses a multimodal large language model (MLLM) to identify interferences like motion and unstable lighting. Furthermore, we propose a two-stage adaptive sampling (TAS) strategy that utilizes the quality score to curate optimal training datasets. Experiments show that by training on large-scale, "in-the-wild" videos filtered by our framework, we can develop unsupervised rPPG models that achieve a substantial improvement in accuracy on standard benchmarks. Our code is available at this https URL.

顶级标签: medical computer vision model training
详细标签: remote photoplethysmography video quality assessment unsupervised learning signal-to-noise ratio multimodal llm 或 搜索:

rPPG-VQA:一种用于无监督rPPG训练的视频质量评估框架 / rPPG-VQA: A Video Quality Assessment Framework for Unsupervised rPPG Training


1️⃣ 一句话总结

这篇论文提出了一个名为rPPG-VQA的智能框架,它能自动评估视频是否适合用来训练无监督的远程心率监测模型,通过筛选高质量视频来显著提升模型的准确率。

源自 arXiv: 2604.11156