Alert button
Picture for Zhichao Ye

Zhichao Ye

Alert button

EC-SfM: Efficient Covisibility-based Structure-from-Motion for Both Sequential and Unordered Images

Feb 21, 2023
Zhichao Ye, Chong Bao, Xin Zhou, Haomin Liu, Hujun Bao, Guofeng Zhang

Figure 1 for EC-SfM: Efficient Covisibility-based Structure-from-Motion for Both Sequential and Unordered Images
Figure 2 for EC-SfM: Efficient Covisibility-based Structure-from-Motion for Both Sequential and Unordered Images
Figure 3 for EC-SfM: Efficient Covisibility-based Structure-from-Motion for Both Sequential and Unordered Images
Figure 4 for EC-SfM: Efficient Covisibility-based Structure-from-Motion for Both Sequential and Unordered Images

Structure-from-Motion is a technology used to obtain scene structure through image collection, which is a fundamental problem in computer vision. For unordered Internet images, SfM is very slow due to the lack of prior knowledge about image overlap. For sequential images, knowing the large overlap between adjacent frames, SfM can adopt a variety of acceleration strategies, which are only applicable to sequential data. To further improve the reconstruction efficiency and break the gap of strategies between these two kinds of data, this paper presents an efficient covisibility-based incremental SfM. Different from previous methods, we exploit covisibility and registration dependency to describe the image connection which is suitable to any kind of data. Based on this general image connection, we propose a unified framework to efficiently reconstruct sequential images, unordered images, and the mixture of these two. Experiments on the unordered images and mixed data verify the effectiveness of the proposed method, which is three times faster than the state of the art on feature matching, and an order of magnitude faster on reconstruction without sacrificing the accuracy. The source code is publicly available at https://github.com/openxrlab/xrsfm

Viaarxiv icon

Generative Category-Level Shape and Pose Estimation with Semantic Primitives

Oct 03, 2022
Guanglin Li, Yifeng Li, Zhichao Ye, Qihang Zhang, Tao Kong, Zhaopeng Cui, Guofeng Zhang

Figure 1 for Generative Category-Level Shape and Pose Estimation with Semantic Primitives
Figure 2 for Generative Category-Level Shape and Pose Estimation with Semantic Primitives
Figure 3 for Generative Category-Level Shape and Pose Estimation with Semantic Primitives
Figure 4 for Generative Category-Level Shape and Pose Estimation with Semantic Primitives

Empowering autonomous agents with 3D understanding for daily objects is a grand challenge in robotics applications. When exploring in an unknown environment, existing methods for object pose estimation are still not satisfactory due to the diversity of object shapes. In this paper, we propose a novel framework for category-level object shape and pose estimation from a single RGB-D image. To handle the intra-category variation, we adopt a semantic primitive representation that encodes diverse shapes into a unified latent space, which is the key to establish reliable correspondences between observed point clouds and estimated shapes. Then, by using a SIM(3)-invariant shape descriptor, we gracefully decouple the shape and pose of an object, thus supporting latent shape optimization of target objects in arbitrary poses. Extensive experiments show that the proposed method achieves SOTA pose estimation performance and better generalization in the real-world dataset. Code and video are available at https://zju3dv.github.io/gCasp

* CoRL 2022, 17 pages, 13 figures 
Viaarxiv icon