Alert button
Picture for Kangcheng Liu

Kangcheng Liu

Alert button

Outram: One-shot Global Localization via Triangulated Scene Graph and Global Outlier Pruning

Sep 16, 2023
Pengyu Yin, Haozhi Cao, Thien-Minh Nguyen, Shenghai Yuan, Shuyang Zhang, Kangcheng Liu, Lihua Xie

One-shot LiDAR localization refers to the ability to estimate the robot pose from one single point cloud, which yields significant advantages in initialization and relocalization processes. In the point cloud domain, the topic has been extensively studied as a global descriptor retrieval (i.e., loop closure detection) and pose refinement (i.e., point cloud registration) problem both in isolation or combined. However, few have explicitly considered the relationship between candidate retrieval and correspondence generation in pose estimation, leaving them brittle to substructure ambiguities. To this end, we propose a hierarchical one-shot localization algorithm called Outram that leverages substructures of 3D scene graphs for locally consistent correspondence searching and global substructure-wise outlier pruning. Such a hierarchical process couples the feature retrieval and the correspondence extraction to resolve the substructure ambiguities by conducting a local-to-global consistency refinement. We demonstrate the capability of Outram in a variety of scenarios in multiple large-scale outdoor datasets. Our implementation is open-sourced: https://github.com/Pamphlett/Outram.

* 8 pages, 5 figures 
Viaarxiv icon

Path Planning for Multiple Tethered Robots Using Topological Braids

Apr 29, 2023
Muqing Cao, Kun Cao, Shenghai Yuan, Kangcheng Liu, Yan Loi Wong, Lihua Xie

Figure 1 for Path Planning for Multiple Tethered Robots Using Topological Braids
Figure 2 for Path Planning for Multiple Tethered Robots Using Topological Braids
Figure 3 for Path Planning for Multiple Tethered Robots Using Topological Braids
Figure 4 for Path Planning for Multiple Tethered Robots Using Topological Braids

Path planning for multiple tethered robots is a challenging problem due to the complex interactions among the cables and the possibility of severe entanglements. Previous works on this problem either consider idealistic cable models or provide no guarantee for entanglement-free paths. In this work, we present a new approach to address this problem using the theory of braids. By establishing a topological equivalence between the physical cables and the space-time trajectories of the robots, and identifying particular braid patterns that emerge from the entangled trajectories, we obtain the key finding that all complex entanglements stem from a finite number of interaction patterns between 2 or 3 robots. Hence, non-entanglement can be guaranteed by avoiding these interaction patterns in the trajectories of the robots. Based on this finding, we present a graph search algorithm using the permutation grid to efficiently search for a feasible topology of paths and reject braid patterns that result in an entanglement. We demonstrate that the proposed algorithm can achieve 100% goal-reaching capability without entanglement for up to 10 drones with a slack cable model in a high-fidelity simulation platform. The practicality of the proposed approach is verified using three small tethered UAVs in indoor flight experiments.

* Accepted for presentation in Robotics: Science and Systems 2023 
Viaarxiv icon

3D Semantic Segmentation in the Wild: Learning Generalized Models for Adverse-Condition Point Clouds

Apr 03, 2023
Aoran Xiao, Jiaxing Huang, Weihao Xuan, Ruijie Ren, Kangcheng Liu, Dayan Guan, Abdulmotaleb El Saddik, Shijian Lu, Eric Xing

Figure 1 for 3D Semantic Segmentation in the Wild: Learning Generalized Models for Adverse-Condition Point Clouds
Figure 2 for 3D Semantic Segmentation in the Wild: Learning Generalized Models for Adverse-Condition Point Clouds
Figure 3 for 3D Semantic Segmentation in the Wild: Learning Generalized Models for Adverse-Condition Point Clouds
Figure 4 for 3D Semantic Segmentation in the Wild: Learning Generalized Models for Adverse-Condition Point Clouds

Robust point cloud parsing under all-weather conditions is crucial to level-5 autonomy in autonomous driving. However, how to learn a universal 3D semantic segmentation (3DSS) model is largely neglected as most existing benchmarks are dominated by point clouds captured under normal weather. We introduce SemanticSTF, an adverse-weather point cloud dataset that provides dense point-level annotations and allows to study 3DSS under various adverse weather conditions. We study all-weather 3DSS modeling under two setups: 1) domain adaptive 3DSS that adapts from normal-weather data to adverse-weather data; 2) domain generalizable 3DSS that learns all-weather 3DSS models from normal-weather data. Our studies reveal the challenge while existing 3DSS methods encounter adverse-weather data, showing the great value of SemanticSTF in steering the future endeavor along this very meaningful research direction. In addition, we design a domain randomization technique that alternatively randomizes the geometry styles of point clouds and aggregates their embeddings, ultimately leading to a generalizable model that can improve 3DSS under various adverse weather effectively. The SemanticSTF and related codes are available at \url{https://github.com/xiaoaoran/SemanticSTF}.

* CVPR2023 
Viaarxiv icon

SVCNet: Scribble-based Video Colorization Network with Temporal Aggregation

Mar 21, 2023
Yuzhi Zhao, Lai-Man Po, Kangcheng Liu, Xuehui Wang, Wing-Yin Yu, Pengfei Xian, Yujia Zhang, Mengyang Liu

Figure 1 for SVCNet: Scribble-based Video Colorization Network with Temporal Aggregation
Figure 2 for SVCNet: Scribble-based Video Colorization Network with Temporal Aggregation
Figure 3 for SVCNet: Scribble-based Video Colorization Network with Temporal Aggregation
Figure 4 for SVCNet: Scribble-based Video Colorization Network with Temporal Aggregation

In this paper, we propose a scribble-based video colorization network with temporal aggregation called SVCNet. It can colorize monochrome videos based on different user-given color scribbles. It addresses three common issues in the scribble-based video colorization area: colorization vividness, temporal consistency, and color bleeding. To improve the colorization quality and strengthen the temporal consistency, we adopt two sequential sub-networks in SVCNet for precise colorization and temporal smoothing, respectively. The first stage includes a pyramid feature encoder to incorporate color scribbles with a grayscale frame, and a semantic feature encoder to extract semantics. The second stage finetunes the output from the first stage by aggregating the information of neighboring colorized frames (as short-range connections) and the first colorized frame (as a long-range connection). To alleviate the color bleeding artifacts, we learn video colorization and segmentation simultaneously. Furthermore, we set the majority of operations on a fixed small image resolution and use a Super-resolution Module at the tail of SVCNet to recover original sizes. It allows the SVCNet to fit different image resolutions at the inference. Finally, we evaluate the proposed SVCNet on DAVIS and Videvo benchmarks. The experimental results demonstrate that SVCNet produces both higher-quality and more temporally consistent videos than other well-known video colorization approaches. The codes and models can be found at https://github.com/zhaoyuzhi/SVCNet.

* under revision of IEEE Transactions on Image Processing 
Viaarxiv icon

DoubleBee: A Hybrid Aerial-Ground Robot with Two Active Wheels

Mar 20, 2023
Muqing Cao, Xinhang Xu, Shenghai Yuan, Kun Cao, Kangcheng Liu, Lihua Xie

Figure 1 for DoubleBee: A Hybrid Aerial-Ground Robot with Two Active Wheels
Figure 2 for DoubleBee: A Hybrid Aerial-Ground Robot with Two Active Wheels
Figure 3 for DoubleBee: A Hybrid Aerial-Ground Robot with Two Active Wheels
Figure 4 for DoubleBee: A Hybrid Aerial-Ground Robot with Two Active Wheels

We present the dynamic model and control of DoubleBee, a novel hybrid aerial-ground vehicle consisting of two propellers mounted on tilting servo motors and two motor-driven wheels. DoubleBee exploits the high energy efficiency of a bicopter configuration in aerial mode, and enjoys the low power consumption of a two-wheel self-balancing robot on the ground. Furthermore, the propeller thrusts act as additional control inputs on the ground, enabling a novel decoupled control scheme where the attitude of the robot is controlled using thrusts and the translational motion is realized using wheels. A prototype of DoubleBee is constructed using commercially available components. The power efficiency and the control performance of the robot are verified through comprehensive experiments. Challenging tasks in indoor and outdoor environments demonstrate the capability of DoubleBee to traverse unstructured environments, fly over and move under barriers, and climb steep and rough terrains.

Viaarxiv icon

FAC: 3D Representation Learning via Foreground Aware Feature Contrast

Mar 11, 2023
Kangcheng Liu, Aoran Xiao, Xiaoqin Zhang, Shijian Lu, Ling Shao

Figure 1 for FAC: 3D Representation Learning via Foreground Aware Feature Contrast
Figure 2 for FAC: 3D Representation Learning via Foreground Aware Feature Contrast
Figure 3 for FAC: 3D Representation Learning via Foreground Aware Feature Contrast
Figure 4 for FAC: 3D Representation Learning via Foreground Aware Feature Contrast

Contrastive learning has recently demonstrated great potential for unsupervised pre-training in 3D scene understanding tasks. However, most existing work randomly selects point features as anchors while building contrast, leading to a clear bias toward background points that often dominate in 3D scenes. Also, object awareness and foreground-to-background discrimination are neglected, making contrastive learning less effective. To tackle these issues, we propose a general foreground-aware feature contrast (FAC) framework to learn more effective point cloud representations in pre-training. FAC consists of two novel contrast designs to construct more effective and informative contrast pairs. The first is building positive pairs within the same foreground segment where points tend to have the same semantics. The second is that we prevent over-discrimination between 3D segments/objects and encourage foreground-to-background distinctions at the segment level with adaptive feature learning in a Siamese correspondence network, which adaptively learns feature correlations within and across point cloud views effectively. Visualization with point activation maps shows that our contrast pairs capture clear correspondences among foreground regions during pre-training. Quantitative experiments also show that FAC achieves superior knowledge transfer and data efficiency in various downstream 3D semantic segmentation and object detection tasks.

* 11 pages, IEEE/CVF Conference on Computer Vision and Pattern Recognition 
Viaarxiv icon

Learning-Based Defect Recognitions for Autonomous UAV Inspections

Feb 13, 2023
Kangcheng Liu

Figure 1 for Learning-Based Defect Recognitions for Autonomous UAV Inspections
Figure 2 for Learning-Based Defect Recognitions for Autonomous UAV Inspections
Figure 3 for Learning-Based Defect Recognitions for Autonomous UAV Inspections
Figure 4 for Learning-Based Defect Recognitions for Autonomous UAV Inspections

Automatic crack detection and segmentation play a significant role in the whole system of unmanned aerial vehicle inspections. In this paper, we have implemented a deep learning framework for crack detection based on classical network architectures including Alexnet, VGG, and Resnet. Moreover, inspired by the feature pyramid network architecture, a hierarchical convolutional neural network (CNN) deep learning framework which is efficient in crack segmentation is also proposed, and its performance of it is compared with other state-of-the-art network architecture. We have summarized the existing crack detection and segmentation datasets and established the largest existing benchmark dataset on the internet for crack detection and segmentation, which is open-sourced for the research community. Our feature pyramid crack segmentation network is tested on the benchmark dataset and gives satisfactory segmentation results. A framework for automatic unmanned aerial vehicle inspections is also proposed and will be established for the crack inspection tasks of various concrete structures. All our self-established datasets and codes are open-sourced at: https://github.com/KangchengLiu/Crack-Detection-and-Segmentation-Dataset-for-UAV-Inspection

* ROBIO 2019 Oral 
Viaarxiv icon

A LiDAR-Inertial-Visual SLAM System with Loop Detection

Jan 13, 2023
Kangcheng Liu

Figure 1 for A LiDAR-Inertial-Visual SLAM System with Loop Detection
Figure 2 for A LiDAR-Inertial-Visual SLAM System with Loop Detection
Figure 3 for A LiDAR-Inertial-Visual SLAM System with Loop Detection
Figure 4 for A LiDAR-Inertial-Visual SLAM System with Loop Detection

We have proposed, to the best of our knowledge, the first-of-its-kind LiDAR-Inertial-Visual-Fused simultaneous localization and mapping (SLAM) system with a strong place recognition capacity. Our proposed SLAM system is consist of visual-inertial odometry (VIO) and LiDAR inertial odometry (LIO) subsystems. We propose the LIO subsystem utilizing the measurement from the LiDAR and the inertial sensors to build the local odometry map, and propose the VIO subsystem which takes in the visual information to construct the 2D-3D associated map. Then, we propose an iterative Kalman Filter-based optimization function to optimize the local project-based 2D-to-3D photo-metric error between the projected image pixels and the local 3D points to make the robust 2D-3D alignment. Finally, we have also proposed the back-end pose graph global optimization and the elaborately designed loop closure detection network to improve the accuracy of the whole SLAM system. Extensive experiments deployed on the UGV in complicated real-world circumstances demonstrate that our proposed LiDAR-Visual-Inertial localization system outperforms the current state-of-the-art in terms of accuracy, efficiency, and robustness.

* 2022 12th International Conference on CYBER Technology in Automation, Control, and Intelligent Systems (IEEE Cyber Oral) 
Viaarxiv icon

An Integrated Visual System for Unmanned Aerial Vehicles Tracking and Landing on the Ground Vehicles

Dec 31, 2022
Kangcheng Liu

Figure 1 for An Integrated Visual System for Unmanned Aerial Vehicles Tracking and Landing on the Ground Vehicles
Figure 2 for An Integrated Visual System for Unmanned Aerial Vehicles Tracking and Landing on the Ground Vehicles
Figure 3 for An Integrated Visual System for Unmanned Aerial Vehicles Tracking and Landing on the Ground Vehicles
Figure 4 for An Integrated Visual System for Unmanned Aerial Vehicles Tracking and Landing on the Ground Vehicles

The vision of unmanned aerial vehicles is very significant for UAV-related applications such as search and rescue, landing on a moving platform, etc. In this work, we have developed an integrated system for the UAV landing on the moving platform, and the UAV object detection with tracking in the complicated environment. Firstly, we have proposed a robust LoG-based deep neural network for object detection and tracking, which has great advantages in robustness to object scale and illuminations compared with typical deep network-based approaches. Then, we have also improved based on the original Kalman filter and designed an iterative multi-model-based filter to tackle the problem of unknown dynamics in real circumstances of motion estimations. Next, we implemented the whole system and do ROS Gazebo-based testing in two complicated circumstances to verify the effectiveness of our design. Finally, we have deployed the proposed detection, tracking, and motion estimation strategies into real applications to do UAV tracking of a pillar and obstacle avoidance. It is demonstrated that our system shows great accuracy and robustness in real applications.

* ICCA 2022 (Oral) 
Viaarxiv icon