Command Palette
Search for a command to run...
Xintian Mao Jiansheng Wang Xingran Xie Qingli Li Yan Wang

Abstract
Due to the computational complexity of self-attention (SA), prevalent techniques for image deblurring often resort to either adopting localized SA or employing coarse-grained global SA methods, both of which exhibit drawbacks such as compromising global modeling or lacking fine-grained correlation. In order to address this issue by effectively modeling long-range dependencies without sacrificing fine-grained details, we introduce a novel approach termed Local Frequency Transformer (LoFormer). Within each unit of LoFormer, we incorporate a Local Channel-wise SA in the frequency domain (Freq-LC) to simultaneously capture cross-covariance within low- and high-frequency local windows. These operations offer the advantage of (1) ensuring equitable learning opportunities for both coarse-grained structures and fine-grained details, and (2) exploring a broader range of representational properties compared to coarse-grained global SA methods. Additionally, we introduce an MLP Gating mechanism complementary to Freq-LC, which serves to filter out irrelevant features while enhancing global learning capabilities. Our experiments demonstrate that LoFormer significantly improves performance in the image deblurring task, achieving a PSNR of 34.09 dB on the GoPro dataset with 126G FLOPs. https://github.com/DeepMed-Lab-ECNU/Single-Image-Deblur
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| deblurring-on-hide-trained-on-gopro | LoFormer | PSNR (sRGB): 31.86 SSIM (sRGB): 0.949 |
| deblurring-on-realblur-j-1 | LoFormer | PSNR (sRGB): 32.90 SSIM (sRGB): 0.933 |
| deblurring-on-realblur-r | LoFormer | PSNR (sRGB): 40.23 SSIM (sRGB): 0.974 |
| image-deblurring-on-gopro | LoFormer | PSNR: 34.09 SSIM: 0.969 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.