HyperAIHyperAI

Command Palette

Search for a command to run...

4 months ago

RFSR: Improving ISR Diffusion Models via Reward Feedback Learning

Sun Xiaopeng ; Lin Qinwei ; Gao Yu ; Zhong Yujie ; Feng Chengjian ; Li Dengjie ; Zhao Zheng ; Hu Jie ; Ma Lin

RFSR: Improving ISR Diffusion Models via Reward Feedback Learning

Abstract

Generative diffusion models (DM) have been extensively utilized in imagesuper-resolution (ISR). Most of the existing methods adopt the denoising lossfrom DDPMs for model optimization. We posit that introducing reward feedbacklearning to finetune the existing models can further improve the quality of thegenerated images. In this paper, we propose a timestep-aware training strategywith reward feedback learning. Specifically, in the initial denoising stages ofISR diffusion, we apply low-frequency constraints to super-resolution (SR)images to maintain structural stability. In the later denoising stages, we usereward feedback learning to improve the perceptual and aesthetic quality of theSR images. In addition, we incorporate Gram-KL regularization to alleviatestylization caused by reward hacking. Our method can be integrated into anydiffusion-based ISR model in a plug-and-play manner. Experiments show that ISRdiffusion models, when fine-tuned with our method, significantly improve theperceptual and aesthetic quality of SR images, achieving excellent subjectiveresults. Code: https://github.com/sxpro/RFSR

Code Repositories

sxpro/rfsr
Official
pytorch

Benchmarks

BenchmarkMethodologyMetrics
blind-super-resolution-on-drealsrSeeSR+RFSR
clipiqa: 0.7596
maniqa: 0.5922
musiq: 67.48

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
RFSR: Improving ISR Diffusion Models via Reward Feedback Learning | Papers | HyperAI