Alert button
Picture for Guanqi Chen

Guanqi Chen

Alert button

ModLaNets: Learning Generalisable Dynamics via Modularity and Physical Inductive Bias

Jul 04, 2022
Yupu Lu, Shijie Lin, Guanqi Chen, Jia Pan

Figure 1 for ModLaNets: Learning Generalisable Dynamics via Modularity and Physical Inductive Bias
Figure 2 for ModLaNets: Learning Generalisable Dynamics via Modularity and Physical Inductive Bias
Figure 3 for ModLaNets: Learning Generalisable Dynamics via Modularity and Physical Inductive Bias
Figure 4 for ModLaNets: Learning Generalisable Dynamics via Modularity and Physical Inductive Bias

Deep learning models are able to approximate one specific dynamical system but struggle at learning generalisable dynamics, where dynamical systems obey the same laws of physics but contain different numbers of elements (e.g., double- and triple-pendulum systems). To relieve this issue, we proposed the Modular Lagrangian Network (ModLaNet), a structural neural network framework with modularity and physical inductive bias. This framework models the energy of each element using modularity and then construct the target dynamical system via Lagrangian mechanics. Modularity is beneficial for reusing trained networks and reducing the scale of networks and datasets. As a result, our framework can learn from the dynamics of simpler systems and extend to more complex ones, which is not feasible using other relevant physics-informed neural networks. We examine our framework for modelling double-pendulum or three-body systems with small training datasets, where our models achieve the best data efficiency and accuracy performance compared with counterparts. We also reorganise our models as extensions to model multi-pendulum and multi-body systems, demonstrating the intriguing reusable feature of our framework.

* Accepted by ICML 2022 
Viaarxiv icon

Less is More: Adaptive Curriculum Learning for Thyroid Nodule Diagnosis

Jul 02, 2022
Haifan Gong, Hui Cheng, Yifan Xie, Shuangyi Tan, Guanqi Chen, Fei Chen, Guanbin Li

Figure 1 for Less is More: Adaptive Curriculum Learning for Thyroid Nodule Diagnosis
Figure 2 for Less is More: Adaptive Curriculum Learning for Thyroid Nodule Diagnosis
Figure 3 for Less is More: Adaptive Curriculum Learning for Thyroid Nodule Diagnosis
Figure 4 for Less is More: Adaptive Curriculum Learning for Thyroid Nodule Diagnosis

Thyroid nodule classification aims at determining whether the nodule is benign or malignant based on a given ultrasound image. However, the label obtained by the cytological biopsy which is the golden standard in clinical medicine is not always consistent with the ultrasound imaging TI-RADS criteria. The information difference between the two causes the existing deep learning-based classification methods to be indecisive. To solve the Inconsistent Label problem, we propose an Adaptive Curriculum Learning (ACL) framework, which adaptively discovers and discards the samples with inconsistent labels. Specifically, ACL takes both hard sample and model certainty into account, and could accurately determine the threshold to distinguish the samples with Inconsistent Label. Moreover, we contribute TNCD: a Thyroid Nodule Classification Dataset to facilitate future related research on the thyroid nodules. Extensive experimental results on TNCD based on three different backbone networks not only demonstrate the superiority of our method but also prove that the less-is-more principle which strategically discards the samples with Inconsistent Label could yield performance gains. Source code and data are available at https://github.com/chenghui-666/ACL/.

* Accepted to MICCAI 2022 with Student Travel Award 
Viaarxiv icon

Depthwise Non-local Module for Fast Salient Object Detection Using a Single Thread

Jan 22, 2020
Haofeng Li, Guanbin Li, Binbin Yang, Guanqi Chen, Liang Lin, Yizhou Yu

Figure 1 for Depthwise Non-local Module for Fast Salient Object Detection Using a Single Thread
Figure 2 for Depthwise Non-local Module for Fast Salient Object Detection Using a Single Thread
Figure 3 for Depthwise Non-local Module for Fast Salient Object Detection Using a Single Thread
Figure 4 for Depthwise Non-local Module for Fast Salient Object Detection Using a Single Thread

Recently deep convolutional neural networks have achieved significant success in salient object detection. However, existing state-of-the-art methods require high-end GPUs to achieve real-time performance, which makes them hard to adapt to low-cost or portable devices. Although generic network architectures have been proposed to speed up inference on mobile devices, they are tailored to the task of image classification or semantic segmentation, and struggle to capture intra-channel and inter-channel correlations that are essential for contrast modeling in salient object detection. Motivated by the above observations, we design a new deep learning algorithm for fast salient object detection. The proposed algorithm for the first time achieves competitive accuracy and high inference efficiency simultaneously with a single CPU thread. Specifically, we propose a novel depthwise non-local moudule (DNL), which implicitly models contrast via harvesting intra-channel and inter-channel correlations in a self-attention manner. In addition, we introduce a depthwise non-local network architecture that incorporates both depthwise non-local modules and inverted residual blocks. Experimental results show that our proposed network attains very competitive accuracy on a wide range of salient object detection datasets while achieving state-of-the-art efficiency among all existing deep learning based algorithms.

* Accepted as a regular paper in the IEEE Transactions on Cybernetics 
Viaarxiv icon

Motion Guided Attention for Video Salient Object Detection

Oct 03, 2019
Haofeng Li, Guanqi Chen, Guanbin Li, Yizhou Yu

Figure 1 for Motion Guided Attention for Video Salient Object Detection
Figure 2 for Motion Guided Attention for Video Salient Object Detection
Figure 3 for Motion Guided Attention for Video Salient Object Detection
Figure 4 for Motion Guided Attention for Video Salient Object Detection

Video salient object detection aims at discovering the most visually distinctive objects in a video. How to effectively take object motion into consideration during video salient object detection is a critical issue. Existing state-of-the-art methods either do not explicitly model and harvest motion cues or ignore spatial contexts within optical flow images. In this paper, we develop a multi-task motion guided video salient object detection network, which learns to accomplish two sub-tasks using two sub-networks, one sub-network for salient object detection in still images and the other for motion saliency detection in optical flow images. We further introduce a series of novel motion guided attention modules, which utilize the motion saliency sub-network to attend and enhance the sub-network for still images. These two sub-networks learn to adapt to each other by end-to-end training. Experimental results demonstrate that the proposed method significantly outperforms existing state-of-the-art algorithms on a wide range of benchmarks. We hope our simple and effective approach will serve as a solid baseline and help ease future research in video salient object detection. Code and models will be made available.

* 10 pages, 4 figures, ICCV 2019, code: https://github.com/lhaof/Motion-Guided-Attention 
Viaarxiv icon