Alert button
Picture for Chun-Le Guo

Chun-Le Guo

Alert button

PUGAN: Physical Model-Guided Underwater Image Enhancement Using GAN with Dual-Discriminators

Jun 15, 2023
Runmin Cong, Wenyu Yang, Wei Zhang, Chongyi Li, Chun-Le Guo, Qingming Huang, Sam Kwong

Figure 1 for PUGAN: Physical Model-Guided Underwater Image Enhancement Using GAN with Dual-Discriminators
Figure 2 for PUGAN: Physical Model-Guided Underwater Image Enhancement Using GAN with Dual-Discriminators
Figure 3 for PUGAN: Physical Model-Guided Underwater Image Enhancement Using GAN with Dual-Discriminators
Figure 4 for PUGAN: Physical Model-Guided Underwater Image Enhancement Using GAN with Dual-Discriminators

Due to the light absorption and scattering induced by the water medium, underwater images usually suffer from some degradation problems, such as low contrast, color distortion, and blurring details, which aggravate the difficulty of downstream underwater understanding tasks. Therefore, how to obtain clear and visually pleasant images has become a common concern of people, and the task of underwater image enhancement (UIE) has also emerged as the times require. Among existing UIE methods, Generative Adversarial Networks (GANs) based methods perform well in visual aesthetics, while the physical model-based methods have better scene adaptability. Inheriting the advantages of the above two types of models, we propose a physical model-guided GAN model for UIE in this paper, referred to as PUGAN. The entire network is under the GAN architecture. On the one hand, we design a Parameters Estimation subnetwork (Par-subnet) to learn the parameters for physical model inversion, and use the generated color enhancement image as auxiliary information for the Two-Stream Interaction Enhancement sub-network (TSIE-subnet). Meanwhile, we design a Degradation Quantization (DQ) module in TSIE-subnet to quantize scene degradation, thereby achieving reinforcing enhancement of key regions. On the other hand, we design the Dual-Discriminators for the style-content adversarial constraint, promoting the authenticity and visual aesthetics of the results. Extensive experiments on three benchmark datasets demonstrate that our PUGAN outperforms state-of-the-art methods in both qualitative and quantitative metrics.

* 8 pages, 4 figures, Accepted by IEEE Transactions on Image Processing 2023 
Viaarxiv icon

Unlocking Low-Light-Rainy Image Restoration by Pairwise Degradation Feature Vector Guidance

May 06, 2023
Xin Lin, Jingtong Yue, Chao Ren, Chun-Le Guo, Chongyi Li

Figure 1 for Unlocking Low-Light-Rainy Image Restoration by Pairwise Degradation Feature Vector Guidance
Figure 2 for Unlocking Low-Light-Rainy Image Restoration by Pairwise Degradation Feature Vector Guidance
Figure 3 for Unlocking Low-Light-Rainy Image Restoration by Pairwise Degradation Feature Vector Guidance
Figure 4 for Unlocking Low-Light-Rainy Image Restoration by Pairwise Degradation Feature Vector Guidance

Rain in the dark is a common natural phenomenon. Photos captured in such a condition significantly impact the performance of various nighttime activities, such as autonomous driving, surveillance systems, and night photography. While existing methods designed for low-light enhancement or deraining show promising performance, they have limitations in simultaneously addressing the task of brightening low light and removing rain. Furthermore, using a cascade approach, such as ``deraining followed by low-light enhancement'' or vice versa, may lead to difficult-to-handle rain patterns or excessively blurred and overexposed images. To overcome these limitations, we propose an end-to-end network called $L^{2}RIRNet$ which can jointly handle low-light enhancement and deraining. Our network mainly includes a Pairwise Degradation Feature Vector Extraction Network (P-Net) and a Restoration Network (R-Net). P-Net can learn degradation feature vectors on the dark and light areas separately, using contrastive learning to guide the image restoration process. The R-Net is responsible for restoring the image. We also introduce an effective Fast Fourier - ResNet Detail Guidance Module (FFR-DG) that initially guides image restoration using detail image that do not contain degradation information but focus on texture detail information. Additionally, we contribute a dataset containing synthetic and real-world low-light-rainy images. Extensive experiments demonstrate that our $L^{2}RIRNet$ outperforms existing methods in both synthetic and complex real-world scenarios.

Viaarxiv icon

AMT: All-Pairs Multi-Field Transforms for Efficient Frame Interpolation

Apr 19, 2023
Zhen Li, Zuo-Liang Zhu, Ling-Hao Han, Qibin Hou, Chun-Le Guo, Ming-Ming Cheng

Figure 1 for AMT: All-Pairs Multi-Field Transforms for Efficient Frame Interpolation
Figure 2 for AMT: All-Pairs Multi-Field Transforms for Efficient Frame Interpolation
Figure 3 for AMT: All-Pairs Multi-Field Transforms for Efficient Frame Interpolation
Figure 4 for AMT: All-Pairs Multi-Field Transforms for Efficient Frame Interpolation

We present All-Pairs Multi-Field Transforms (AMT), a new network architecture for video frame interpolation. It is based on two essential designs. First, we build bidirectional correlation volumes for all pairs of pixels, and use the predicted bilateral flows to retrieve correlations for updating both flows and the interpolated content feature. Second, we derive multiple groups of fine-grained flow fields from one pair of updated coarse flows for performing backward warping on the input frames separately. Combining these two designs enables us to generate promising task-oriented flows and reduce the difficulties in modeling large motions and handling occluded areas during frame interpolation. These qualities promote our model to achieve state-of-the-art performance on various benchmarks with high efficiency. Moreover, our convolution-based model competes favorably compared to Transformer-based models in terms of accuracy and efficiency. Our code is available at https://github.com/MCG-NKU/AMT.

* Accepted to CVPR2023 
Viaarxiv icon

RIDCP: Revitalizing Real Image Dehazing via High-Quality Codebook Priors

Apr 08, 2023
Rui-Qi Wu, Zheng-Peng Duan, Chun-Le Guo, Zhi Chai, Chong-Yi Li

Figure 1 for RIDCP: Revitalizing Real Image Dehazing via High-Quality Codebook Priors
Figure 2 for RIDCP: Revitalizing Real Image Dehazing via High-Quality Codebook Priors
Figure 3 for RIDCP: Revitalizing Real Image Dehazing via High-Quality Codebook Priors
Figure 4 for RIDCP: Revitalizing Real Image Dehazing via High-Quality Codebook Priors

Existing dehazing approaches struggle to process real-world hazy images owing to the lack of paired real data and robust priors. In this work, we present a new paradigm for real image dehazing from the perspectives of synthesizing more realistic hazy data and introducing more robust priors into the network. Specifically, (1) instead of adopting the de facto physical scattering model, we rethink the degradation of real hazy images and propose a phenomenological pipeline considering diverse degradation types. (2) We propose a Real Image Dehazing network via high-quality Codebook Priors (RIDCP). Firstly, a VQGAN is pre-trained on a large-scale high-quality dataset to obtain the discrete codebook, encapsulating high-quality priors (HQPs). After replacing the negative effects brought by haze with HQPs, the decoder equipped with a novel normalized feature alignment module can effectively utilize high-quality features and produce clean results. However, although our degradation pipeline drastically mitigates the domain gap between synthetic and real data, it is still intractable to avoid it, which challenges HQPs matching in the wild. Thus, we re-calculate the distance when matching the features to the HQPs by a controllable matching operation, which facilitates finding better counterparts. We provide a recommendation to control the matching based on an explainable solution. Users can also flexibly adjust the enhancement degree as per their preference. Extensive experiments verify the effectiveness of our data synthesis pipeline and the superior performance of RIDCP in real image dehazing.

* Acceptted by CVPR 2023 
Viaarxiv icon

SRFormer: Permuted Self-Attention for Single Image Super-Resolution

Mar 17, 2023
Yupeng Zhou, Zhen Li, Chun-Le Guo, Song Bai, Ming-Ming Cheng, Qibin Hou

Figure 1 for SRFormer: Permuted Self-Attention for Single Image Super-Resolution
Figure 2 for SRFormer: Permuted Self-Attention for Single Image Super-Resolution
Figure 3 for SRFormer: Permuted Self-Attention for Single Image Super-Resolution
Figure 4 for SRFormer: Permuted Self-Attention for Single Image Super-Resolution

Previous works have shown that increasing the window size for Transformer-based image super-resolution models (e.g., SwinIR) can significantly improve the model performance but the computation overhead is also considerable. In this paper, we present SRFormer, a simple but novel method that can enjoy the benefit of large window self-attention but introduces even less computational burden. The core of our SRFormer is the permuted self-attention (PSA), which strikes an appropriate balance between the channel and spatial information for self-attention. Our PSA is simple and can be easily applied to existing super-resolution networks based on window self-attention. Without any bells and whistles, we show that our SRFormer achieves a 33.86dB PSNR score on the Urban100 dataset, which is 0.46dB higher than that of SwinIR but uses fewer parameters and computations. We hope our simple and effective approach can serve as a useful tool for future research in super-resolution model design.

Viaarxiv icon

Embedding Fourier for Ultra-High-Definition Low-Light Image Enhancement

Feb 23, 2023
Chongyi Li, Chun-Le Guo, Man Zhou, Zhexin Liang, Shangchen Zhou, Ruicheng Feng, Chen Change Loy

Figure 1 for Embedding Fourier for Ultra-High-Definition Low-Light Image Enhancement
Figure 2 for Embedding Fourier for Ultra-High-Definition Low-Light Image Enhancement
Figure 3 for Embedding Fourier for Ultra-High-Definition Low-Light Image Enhancement
Figure 4 for Embedding Fourier for Ultra-High-Definition Low-Light Image Enhancement

Ultra-High-Definition (UHD) photo has gradually become the standard configuration in advanced imaging devices. The new standard unveils many issues in existing approaches for low-light image enhancement (LLIE), especially in dealing with the intricate issue of joint luminance enhancement and noise removal while remaining efficient. Unlike existing methods that address the problem in the spatial domain, we propose a new solution, UHDFour, that embeds Fourier transform into a cascaded network. Our approach is motivated by a few unique characteristics in the Fourier domain: 1) most luminance information concentrates on amplitudes while noise is closely related to phases, and 2) a high-resolution image and its low-resolution version share similar amplitude patterns.Through embedding Fourier into our network, the amplitude and phase of a low-light image are separately processed to avoid amplifying noise when enhancing luminance. Besides, UHDFour is scalable to UHD images by implementing amplitude and phase enhancement under the low-resolution regime and then adjusting the high-resolution scale with few computations. We also contribute the first real UHD LLIE dataset, \textbf{UHD-LL}, that contains 2,150 low-noise/normal-clear 4K image pairs with diverse darkness and noise levels captured in different scenarios. With this dataset, we systematically analyze the performance of existing LLIE methods for processing UHD images and demonstrate the advantage of our solution. We believe our new framework, coupled with the dataset, would push the frontier of LLIE towards UHD. The code and dataset are available at https://li-chongyi.github.io/UHDFour.

* ICLR 2023 (Oral)  
* Porject page: https://li-chongyi.github.io/UHDFour 
Viaarxiv icon

Designing An Illumination-Aware Network for Deep Image Relighting

Jul 21, 2022
Zuo-Liang Zhu, Zhen Li, Rui-Xun Zhang, Chun-Le Guo, Ming-Ming Cheng

Figure 1 for Designing An Illumination-Aware Network for Deep Image Relighting
Figure 2 for Designing An Illumination-Aware Network for Deep Image Relighting
Figure 3 for Designing An Illumination-Aware Network for Deep Image Relighting
Figure 4 for Designing An Illumination-Aware Network for Deep Image Relighting

Lighting is a determining factor in photography that affects the style, expression of emotion, and even quality of images. Creating or finding satisfying lighting conditions, in reality, is laborious and time-consuming, so it is of great value to develop a technology to manipulate illumination in an image as post-processing. Although previous works have explored techniques based on the physical viewpoint for relighting images, extensive supervisions and prior knowledge are necessary to generate reasonable images, restricting the generalization ability of these works. In contrast, we take the viewpoint of image-to-image translation and implicitly merge ideas of the conventional physical viewpoint. In this paper, we present an Illumination-Aware Network (IAN) which follows the guidance from hierarchical sampling to progressively relight a scene from a single image with high efficiency. In addition, an Illumination-Aware Residual Block (IARB) is designed to approximate the physical rendering process and to extract precise descriptors of light sources for further manipulations. We also introduce a depth-guided geometry encoder for acquiring valuable geometry- and structure-related representations once the depth information is available. Experimental results show that our proposed method produces better quantitative and qualitative relighting results than previous state-of-the-art methods. The code and models are publicly available on https://github.com/NK-CS-ZZL/IAN.

* Accepted for publication as a Regular paper in the IEEE Transactions on Image Processing (T-IP) 
Viaarxiv icon

Image Harmonization by Matching Regional References

Apr 10, 2022
Ziyue Zhu, Zhao Zhang, Zheng Lin, Ruiqi Wu, Zhi Chai, Chun-Le Guo

Figure 1 for Image Harmonization by Matching Regional References
Figure 2 for Image Harmonization by Matching Regional References
Figure 3 for Image Harmonization by Matching Regional References
Figure 4 for Image Harmonization by Matching Regional References

To achieve visual consistency in composite images, recent image harmonization methods typically summarize the appearance pattern of global background and apply it to the global foreground without location discrepancy. However, for a real image, the appearances (illumination, color temperature, saturation, hue, texture, etc) of different regions can vary significantly. So previous methods, which transfer the appearance globally, are not optimal. Trying to solve this issue, we firstly match the contents between the foreground and background and then adaptively adjust every foreground location according to the appearance of its content-related background regions. Further, we design a residual reconstruction strategy, that uses the predicted residual to adjust the appearance, and the composite foreground to reserve the image details. Extensive experiments demonstrate the effectiveness of our method. The source code will be available publicly.

Viaarxiv icon

Towards An End-to-End Framework for Flow-Guided Video Inpainting

Apr 07, 2022
Zhen Li, Cheng-Ze Lu, Jianhua Qin, Chun-Le Guo, Ming-Ming Cheng

Figure 1 for Towards An End-to-End Framework for Flow-Guided Video Inpainting
Figure 2 for Towards An End-to-End Framework for Flow-Guided Video Inpainting
Figure 3 for Towards An End-to-End Framework for Flow-Guided Video Inpainting
Figure 4 for Towards An End-to-End Framework for Flow-Guided Video Inpainting

Optical flow, which captures motion information across frames, is exploited in recent video inpainting methods through propagating pixels along its trajectories. However, the hand-crafted flow-based processes in these methods are applied separately to form the whole inpainting pipeline. Thus, these methods are less efficient and rely heavily on the intermediate results from earlier stages. In this paper, we propose an End-to-End framework for Flow-Guided Video Inpainting (E$^2$FGVI) through elaborately designed three trainable modules, namely, flow completion, feature propagation, and content hallucination modules. The three modules correspond with the three stages of previous flow-based methods but can be jointly optimized, leading to a more efficient and effective inpainting process. Experimental results demonstrate that the proposed method outperforms state-of-the-art methods both qualitatively and quantitatively and shows promising efficiency. The code is available at https://github.com/MCG-NKU/E2FGVI.

* Accepted to CVPR 2022 
Viaarxiv icon