Alert button
Picture for Sihyeon Kim

Sihyeon Kim

Alert button

Semantic-Aware Implicit Template Learning via Part Deformation Consistency

Aug 23, 2023
Sihyeon Kim, Minseok Joo, Jaewon Lee, Juyeon Ko, Juhan Cha, Hyunwoo J. Kim

Figure 1 for Semantic-Aware Implicit Template Learning via Part Deformation Consistency
Figure 2 for Semantic-Aware Implicit Template Learning via Part Deformation Consistency
Figure 3 for Semantic-Aware Implicit Template Learning via Part Deformation Consistency
Figure 4 for Semantic-Aware Implicit Template Learning via Part Deformation Consistency

Learning implicit templates as neural fields has recently shown impressive performance in unsupervised shape correspondence. Despite the success, we observe current approaches, which solely rely on geometric information, often learn suboptimal deformation across generic object shapes, which have high structural variability. In this paper, we highlight the importance of part deformation consistency and propose a semantic-aware implicit template learning framework to enable semantically plausible deformation. By leveraging semantic prior from a self-supervised feature extractor, we suggest local conditioning with novel semantic-aware deformation code and deformation consistency regularizations regarding part deformation, global deformation, and global scaling. Our extensive experiments demonstrate the superiority of the proposed method over baselines in various tasks: keypoint transfer, part label transfer, and texture transfer. More interestingly, our framework shows a larger performance gain under more challenging settings. We also provide qualitative analyses to validate the effectiveness of semantic-aware deformation. The code is available at https://github.com/mlvlab/PDC.

* ICCV camera-ready version 
Viaarxiv icon

Self-positioning Point-based Transformer for Point Cloud Understanding

Mar 29, 2023
Jinyoung Park, Sanghyeok Lee, Sihyeon Kim, Yunyang Xiong, Hyunwoo J. Kim

Figure 1 for Self-positioning Point-based Transformer for Point Cloud Understanding
Figure 2 for Self-positioning Point-based Transformer for Point Cloud Understanding
Figure 3 for Self-positioning Point-based Transformer for Point Cloud Understanding
Figure 4 for Self-positioning Point-based Transformer for Point Cloud Understanding

Transformers have shown superior performance on various computer vision tasks with their capabilities to capture long-range dependencies. Despite the success, it is challenging to directly apply Transformers on point clouds due to their quadratic cost in the number of points. In this paper, we present a Self-Positioning point-based Transformer (SPoTr), which is designed to capture both local and global shape contexts with reduced complexity. Specifically, this architecture consists of local self-attention and self-positioning point-based global cross-attention. The self-positioning points, adaptively located based on the input shape, consider both spatial and semantic information with disentangled attention to improve expressive power. With the self-positioning points, we propose a novel global cross-attention mechanism for point clouds, which improves the scalability of global self-attention by allowing the attention module to compute attention weights with only a small set of self-positioning points. Experiments show the effectiveness of SPoTr on three point cloud tasks such as shape classification, part segmentation, and scene segmentation. In particular, our proposed model achieves an accuracy gain of 2.6% over the previous best models on shape classification with ScanObjectNN. We also provide qualitative analyses to demonstrate the interpretability of self-positioning points. The code of SPoTr is available at https://github.com/mlvlab/SPoTr.

* Accepted paper at CVPR 2023 
Viaarxiv icon

Metropolis-Hastings Data Augmentation for Graph Neural Networks

Mar 26, 2022
Hyeonjin Park, Seunghun Lee, Sihyeon Kim, Jinyoung Park, Jisu Jeong, Kyung-Min Kim, Jung-Woo Ha, Hyunwoo J. Kim

Figure 1 for Metropolis-Hastings Data Augmentation for Graph Neural Networks
Figure 2 for Metropolis-Hastings Data Augmentation for Graph Neural Networks
Figure 3 for Metropolis-Hastings Data Augmentation for Graph Neural Networks
Figure 4 for Metropolis-Hastings Data Augmentation for Graph Neural Networks

Graph Neural Networks (GNNs) often suffer from weak-generalization due to sparsely labeled data despite their promising results on various graph-based tasks. Data augmentation is a prevalent remedy to improve the generalization ability of models in many domains. However, due to the non-Euclidean nature of data space and the dependencies between samples, designing effective augmentation on graphs is challenging. In this paper, we propose a novel framework Metropolis-Hastings Data Augmentation (MH-Aug) that draws augmented graphs from an explicit target distribution for semi-supervised learning. MH-Aug produces a sequence of augmented graphs from the target distribution enables flexible control of the strength and diversity of augmentation. Since the direct sampling from the complex target distribution is challenging, we adopt the Metropolis-Hastings algorithm to obtain the augmented samples. We also propose a simple and effective semi-supervised learning strategy with generated samples from MH-Aug. Our extensive experiments demonstrate that MH-Aug can generate a sequence of samples according to the target distribution to significantly improve the performance of GNNs.

* 10 pages, 5 figures 
Viaarxiv icon

Point Cloud Augmentation with Weighted Local Transformations

Oct 11, 2021
Sihyeon Kim, Sanghyeok Lee, Dasol Hwang, Jaewon Lee, Seong Jae Hwang, Hyunwoo J. Kim

Figure 1 for Point Cloud Augmentation with Weighted Local Transformations
Figure 2 for Point Cloud Augmentation with Weighted Local Transformations
Figure 3 for Point Cloud Augmentation with Weighted Local Transformations
Figure 4 for Point Cloud Augmentation with Weighted Local Transformations

Despite the extensive usage of point clouds in 3D vision, relatively limited data are available for training deep neural networks. Although data augmentation is a standard approach to compensate for the scarcity of data, it has been less explored in the point cloud literature. In this paper, we propose a simple and effective augmentation method called PointWOLF for point cloud augmentation. The proposed method produces smoothly varying non-rigid deformations by locally weighted transformations centered at multiple anchor points. The smooth deformations allow diverse and realistic augmentations. Furthermore, in order to minimize the manual efforts to search the optimal hyperparameters for augmentation, we present AugTune, which generates augmented samples of desired difficulties producing targeted confidence scores. Our experiments show our framework consistently improves the performance for both shape classification and part segmentation tasks. Particularly, with PointNet++, PointWOLF achieves the state-of-the-art 89.7 accuracy on shape classification with the real-world ScanObjectNN dataset.

* 9 pages, Accepted to ICCV 2021 
Viaarxiv icon