Alert button
Picture for Chunyi Sun

Chunyi Sun

Alert button

NeRFEditor: Differentiable Style Decomposition for Full 3D Scene Editing

Dec 08, 2022
Chunyi Sun, Yanbin Liu, Junlin Han, Stephen Gould

Figure 1 for NeRFEditor: Differentiable Style Decomposition for Full 3D Scene Editing
Figure 2 for NeRFEditor: Differentiable Style Decomposition for Full 3D Scene Editing
Figure 3 for NeRFEditor: Differentiable Style Decomposition for Full 3D Scene Editing
Figure 4 for NeRFEditor: Differentiable Style Decomposition for Full 3D Scene Editing

We present NeRFEditor, an efficient learning framework for 3D scene editing, which takes a video captured over 360{\deg} as input and outputs a high-quality, identity-preserving stylized 3D scene. Our method supports diverse types of editing such as guided by reference images, text prompts, and user interactions. We achieve this by encouraging a pre-trained StyleGAN model and a NeRF model to learn from each other mutually. Specifically, we use a NeRF model to generate numerous image-angle pairs to train an adjustor, which can adjust the StyleGAN latent code to generate high-fidelity stylized images for any given angle. To extrapolate editing to GAN out-of-domain views, we devise another module that is trained in a self-supervised learning manner. This module maps novel-view images to the hidden space of StyleGAN that allows StyleGAN to generate stylized images on novel views. These two modules together produce guided images in 360{\deg}views to finetune a NeRF to make stylization effects, where a stable fine-tuning strategy is proposed to achieve this. Experiments show that NeRFEditor outperforms prior work on benchmark and real-world scenes with better editability, fidelity, and identity preservation.

* Project page: https://chuny1.github.io/NeRFEditor/nerfeditor.html 
Viaarxiv icon

Harmonizer: Learning to Perform White-Box Image and Video Harmonization

Jul 04, 2022
Zhanghan Ke, Chunyi Sun, Lei Zhu, Ke Xu, Rynson W. H. Lau

Figure 1 for Harmonizer: Learning to Perform White-Box Image and Video Harmonization
Figure 2 for Harmonizer: Learning to Perform White-Box Image and Video Harmonization
Figure 3 for Harmonizer: Learning to Perform White-Box Image and Video Harmonization
Figure 4 for Harmonizer: Learning to Perform White-Box Image and Video Harmonization

Recent works on image harmonization solve the problem as a pixel-wise image translation task via large autoencoders. They have unsatisfactory performances and slow inference speeds when dealing with high-resolution images. In this work, we observe that adjusting the input arguments of basic image filters, e.g., brightness and contrast, is sufficient for humans to produce realistic images from the composite ones. Hence, we frame image harmonization as an image-level regression problem to learn the arguments of the filters that humans use for the task. We present a Harmonizer framework for image harmonization. Unlike prior methods that are based on black-box autoencoders, Harmonizer contains a neural network for filter argument prediction and several white-box filters (based on the predicted arguments) for image harmonization. We also introduce a cascade regressor and a dynamic loss strategy for Harmonizer to learn filter arguments more stably and precisely. Since our network only outputs image-level arguments and the filters we used are efficient, Harmonizer is much lighter and faster than existing methods. Comprehensive experiments demonstrate that Harmonizer surpasses existing methods notably, especially with high-resolution inputs. Finally, we apply Harmonizer to video harmonization, which achieves consistent results across frames and 56 fps at 1080P resolution. Code and models are available at: https://github.com/ZHKKKe/Harmonizer.

Viaarxiv icon

Blind Image Decomposition

Aug 28, 2021
Junlin Han, Weihao Li, Pengfei Fang, Chunyi Sun, Jie Hong, Mohammad Ali Armin, Lars Petersson, Hongdong Li

Figure 1 for Blind Image Decomposition
Figure 2 for Blind Image Decomposition
Figure 3 for Blind Image Decomposition
Figure 4 for Blind Image Decomposition

We present and study a novel task named Blind Image Decomposition (BID), which requires separating a superimposed image into constituent underlying images in a blind setting, that is, both the source components involved in mixing as well as the mixing mechanism are unknown. For example, rain may consist of multiple components, such as rain streaks, raindrops, snow, and haze. Rainy images can be treated as an arbitrary combination of these components, some of them or all of them. How to decompose superimposed images, like rainy images, into distinct source components is a crucial step towards real-world vision systems. To facilitate research on this new task, we construct three benchmark datasets, including mixed image decomposition across multiple domains, real-scenario deraining, and joint shadow/reflection/watermark removal. Moreover, we propose a simple yet general Blind Image Decomposition Network (BIDeN) to serve as a strong baseline for future work. Experimental results demonstrate the tenability of our benchmarks and the effectiveness of BIDeN. Code and project page are available.

* Project page: https://junlinhan.github.io/projects/BID.html. Code: https://github.com/JunlinHan/BID 
Viaarxiv icon

Defocus Blur Detection via Salient Region Detection Prior

Nov 19, 2020
Ming Qian, Min Xia, Chunyi Sun, Zhiwei Wang, Liguo Weng

Figure 1 for Defocus Blur Detection via Salient Region Detection Prior
Figure 2 for Defocus Blur Detection via Salient Region Detection Prior
Figure 3 for Defocus Blur Detection via Salient Region Detection Prior
Figure 4 for Defocus Blur Detection via Salient Region Detection Prior

Defocus blur always occurred in photos when people take photos by Digital Single Lens Reflex Camera(DSLR), giving salient region and aesthetic pleasure. Defocus blur Detection aims to separate the out-of-focus and depth-of-field areas in photos, which is an important work in computer vision. Current works for defocus blur detection mainly focus on the designing of networks, the optimizing of the loss function, and the application of multi-stream strategy, meanwhile, these works do not pay attention to the shortage of training data. In this work, to address the above data-shortage problem, we turn to rethink the relationship between two tasks: defocus blur detection and salient region detection. In an image with bokeh effect, it is obvious that the salient region and the depth-of-field area overlap in most cases. So we first train our network on the salient region detection tasks, then transfer the pre-trained model to the defocus blur detection tasks. Besides, we propose a novel network for defocus blur detection. Experiments show that our transfer strategy works well on many current models, and demonstrate the superiority of our network.

Viaarxiv icon