菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-12
📄 Abstract - Learning Perceptual Representations for Gaming NR-VQA with Multi-Task FR Signals

No-reference video quality assessment (NR-VQA) for gaming videos is challenging due to limited human-rated datasets and unique content characteristics including fast motion, stylized graphics, and compression artifacts. We present MTL-VQA, a multi-task learning framework that uses full-reference metrics as supervisory signals to learn perceptually meaningful features without human labels for pretraining. By jointly optimizing multiple full-reference (FR) objectives with adaptive task weighting, our approach learns shared representations that transfer effectively to NR-VQA. Experiments on gaming video datasets show MTL-VQA achieves performance competitive with state-of-the-art NR-VQA methods across both MOS-supervised and label-efficient/self-supervised settings.

顶级标签: computer vision model training video
详细标签: video quality assessment multi-task learning no-reference vqa gaming videos perceptual representation 或 搜索:

利用多任务全参考信号学习感知表征用于游戏视频无参考质量评估 / Learning Perceptual Representations for Gaming NR-VQA with Multi-Task FR Signals


1️⃣ 一句话总结

这篇论文提出了一种名为MTL-VQA的多任务学习框架,它巧妙地利用无需人工标注的全参考视频质量指标作为训练信号,来学习感知特征,从而有效解决了游戏视频因数据稀缺和内容独特而导致的无参考质量评估难题。

源自 arXiv: 2602.11903