EditHF-1M:一个百万规模、包含丰富人类偏好反馈的图像编辑数据集 / EditHF-1M: A Million-Scale Rich Human Preference Feedback for Image Editing
1️⃣ 一句话总结
这篇论文创建了一个包含百万图像和大量人类偏好评分的大规模数据集EditHF-1M,并基于此训练了一个能评估图像编辑质量的AI模型,该模型不仅能准确判断编辑效果,还能作为奖励信号来帮助其他图像编辑AI模型通过强化学习进行自我优化和提升。
Recent text-guided image editing (TIE) models have achieved remarkable progress, while many edited images still suffer from issues such as artifacts, unexpected editings, unaesthetic contents. Although some benchmarks and methods have been proposed for evaluating edited images, scalable evaluation models are still lacking, which limits the development of human feedback reward models for image editing. To address the challenges, we first introduce \textbf{EditHF-1M}, a million-scale image editing dataset with over 29M human preference pairs and 148K human mean opinion ratings, both evaluated from three dimensions, \textit{i.e.}, visual quality, instruction alignment, and attribute preservation. Based on EditHF-1M, we propose \textbf{EditHF}, a multimodal large language model (MLLM) based evaluation model, to provide human-aligned feedback from image editing. Finally, we introduce \textbf{EditHF-Reward}, which utilizes EditHF as the reward signal to optimize the text-guided image editing models through reinforcement learning. Extensive experiments show that EditHF achieves superior alignment with human preferences and demonstrates strong generalization on other datasets. Furthermore, we fine-tune the Qwen-Image-Edit using EditHF-Reward, achieving significant performance improvements, which demonstrates the ability of EditHF to serve as a reward model to scale-up the image editing. Both the dataset and code will be released in our GitHub repository: this https URL.
EditHF-1M:一个百万规模、包含丰富人类偏好反馈的图像编辑数据集 / EditHF-1M: A Million-Scale Rich Human Preference Feedback for Image Editing
这篇论文创建了一个包含百万图像和大量人类偏好评分的大规模数据集EditHF-1M,并基于此训练了一个能评估图像编辑质量的AI模型,该模型不仅能准确判断编辑效果,还能作为奖励信号来帮助其他图像编辑AI模型通过强化学习进行自我优化和提升。
源自 arXiv: 2603.14916