Alert button
Picture for Siqi Hui

Siqi Hui

Alert button

Understanding the Overfitting of the Episodic Meta-training

Jul 07, 2023
Siqi Hui, Sanping Zhou, Ye deng, Jinjun Wang

Figure 1 for Understanding the Overfitting of the Episodic Meta-training

Despite the success of two-stage few-shot classification methods, in the episodic meta-training stage, the model suffers severe overfitting. We hypothesize that it is caused by over-discrimination, i.e., the model learns to over-rely on the superficial features that fit for base class discrimination while suppressing the novel class generalization. To penalize over-discrimination, we introduce knowledge distillation techniques to keep novel generalization knowledge from the teacher model during training. Specifically, we select the teacher model as the one with the best validation accuracy during meta-training and restrict the symmetric Kullback-Leibler (SKL) divergence between the output distribution of the linear classifier of the teacher model and that of the student model. This simple approach outperforms the standard meta-training process. We further propose the Nearest Neighbor Symmetric Kullback-Leibler (NNSKL) divergence for meta-training to push the limits of knowledge distillation techniques. NNSKL takes few-shot tasks as input and penalizes the output of the nearest neighbor classifier, which possesses an impact on the relationships between query embedding and support centers. By combining SKL and NNSKL in meta-training, the model achieves even better performance and surpasses state-of-the-art results on several benchmarks.

Viaarxiv icon

T-former: An Efficient Transformer for Image Inpainting

May 19, 2023
Ye Deng, Siqi Hui, Sanping Zhou, Deyu Meng, Jinjun Wang

Figure 1 for T-former: An Efficient Transformer for Image Inpainting
Figure 2 for T-former: An Efficient Transformer for Image Inpainting
Figure 3 for T-former: An Efficient Transformer for Image Inpainting
Figure 4 for T-former: An Efficient Transformer for Image Inpainting

Benefiting from powerful convolutional neural networks (CNNs), learning-based image inpainting methods have made significant breakthroughs over the years. However, some nature of CNNs (e.g. local prior, spatially shared parameters) limit the performance in the face of broken images with diverse and complex forms. Recently, a class of attention-based network architectures, called transformer, has shown significant performance on natural language processing fields and high-level vision tasks. Compared with CNNs, attention operators are better at long-range modeling and have dynamic weights, but their computational complexity is quadratic in spatial resolution, and thus less suitable for applications involving higher resolution images, such as image inpainting. In this paper, we design a novel attention linearly related to the resolution according to Taylor expansion. And based on this attention, a network called $T$-former is designed for image inpainting. Experiments on several benchmark datasets demonstrate that our proposed method achieves state-of-the-art accuracy while maintaining a relatively low number of parameters and computational complexity. The code can be found at \href{https://github.com/dengyecode/T-former_image_inpainting}{github.com/dengyecode/T-former\_image\_inpainting}

* ACM Multimedia 2022  
Viaarxiv icon

Auxiliary Loss Adaptation for Image Inpainting

Nov 22, 2021
Siqi Hui, Sanping Zhou, Xingyu Wan, Jixin Wang, Ye Deng, Yang Wu, Zhenghao Gong, Jinjun Wang

Figure 1 for Auxiliary Loss Adaptation for Image Inpainting
Figure 2 for Auxiliary Loss Adaptation for Image Inpainting
Figure 3 for Auxiliary Loss Adaptation for Image Inpainting
Figure 4 for Auxiliary Loss Adaptation for Image Inpainting

Auxiliary losses commonly used in image inpainting lead to better reconstruction performance by incorporating prior knowledge of missing regions. However, it usually requires a lot of effort to fully exploit the potential of auxiliary losses, or otherwise, improperly weighted auxiliary losses would distract the model from the inpainting task, and the effectiveness of an auxiliary loss might vary during the training process. Hence the design of auxiliary losses takes strong domain expertise. To mitigate the problem, in this work, we introduce the Auxiliary Loss Adaptation for Image Inpainting (ALA) algorithm to dynamically adjust the parameters of the auxiliary loss. Our method is based on the principle that the best auxiliary loss is the one that helps increase the performance of the main loss most through several steps of gradient descent. We then examined two commonly used auxiliary losses in inpainting and used ALA to adapt their parameters. Experimental results show that ALA induces more competitive inpainting results than fixed auxiliary losses. In particular, simply combining auxiliary loss with ALA, existing inpainting methods can achieve increased performances without explicitly incorporating delicate network design or structure knowledge prior.

Viaarxiv icon

Auxiliary Loss Adaption for Image Inpainting

Nov 14, 2021
Siqi Hui, Sanping Zhou, Xingyu Wan, Jixin Wang, Ye Deng, Yang Wu, Zhenghao Gong, Jinjun Wang

Figure 1 for Auxiliary Loss Adaption for Image Inpainting
Figure 2 for Auxiliary Loss Adaption for Image Inpainting
Figure 3 for Auxiliary Loss Adaption for Image Inpainting
Figure 4 for Auxiliary Loss Adaption for Image Inpainting

Auxiliary losses commonly used in image inpainting lead to better reconstruction performance by incorporating prior knowledge of missing regions. However, it usually takes a lot of effort to fully exploit the potential of auxiliary losses, since improperly weighted auxiliary losses would distract the model from the inpainting task, and the effectiveness of an auxiliary loss might vary during the training process. Furthermore, the design of auxiliary losses takes domain expertise. In this work, we introduce the Auxiliary Loss Adaption (Adaption) algorithm to dynamically adjust the parameters of the auxiliary loss, to better assist the primary task. Our algorithm is based on the principle that better auxiliary loss is the one that helps increase the performance of the main loss through several steps of gradient descent. We then examined two commonly used auxiliary losses in inpainting and use \ac{ALA} to adapt their parameters. Experimental results show that ALA induces more competitive inpainting results than fixed auxiliary losses. In particular, simply combining auxiliary loss with \ac{ALA}, existing inpainting methods can achieve increased performances without explicitly incorporating delicate network design or structure knowledge prior.

Viaarxiv icon