隐式神经表征促进统一的通用视觉编码 / Implicit Neural Representation Facilitates Unified Universal Vision Encoding
1️⃣ 一句话总结
这篇论文提出了一种新型的视觉编码模型,它通过隐式神经表征技术,首次将图像识别和图像生成两大任务统一起来,学习到一个既可用于分类、检测等识别任务,又能用于高质量图像生成的紧凑嵌入空间。
Models for image representation learning are typically designed for either recognition or generation. Various forms of contrastive learning help models learn to convert images to embeddings that are useful for classification, detection, and segmentation. On the other hand, models can be trained to reconstruct images with pixel-wise, perceptual, and adversarial losses in order to learn a latent space that is useful for image generation. We seek to unify these two directions with a first-of-its-kind model that learns representations which are simultaneously useful for recognition and generation. We train our model as a hyper-network for implicit neural representation, which learns to map images to model weights for fast, accurate reconstruction. We further integrate our INR hyper-network with knowledge distillation to improve its generalization and performance. Beyond the novel training design, the model also learns an unprecedented compressed embedding space with outstanding performance for various visual tasks. The complete model competes with state-of-the-art results for image representation learning, while also enabling generative capabilities with its high-quality tiny embeddings. The code is available at this https URL.
隐式神经表征促进统一的通用视觉编码 / Implicit Neural Representation Facilitates Unified Universal Vision Encoding
这篇论文提出了一种新型的视觉编码模型,它通过隐式神经表征技术,首次将图像识别和图像生成两大任务统一起来,学习到一个既可用于分类、检测等识别任务,又能用于高质量图像生成的紧凑嵌入空间。
源自 arXiv: 2601.14256