Alert button
Picture for Taeksoo Kim

Taeksoo Kim

Alert button

NCHO: Unsupervised Learning for Neural 3D Composition of Humans and Objects

May 23, 2023
Taeksoo Kim, Shunsuke Saito, Hanbyul Joo

Figure 1 for NCHO: Unsupervised Learning for Neural 3D Composition of Humans and Objects
Figure 2 for NCHO: Unsupervised Learning for Neural 3D Composition of Humans and Objects
Figure 3 for NCHO: Unsupervised Learning for Neural 3D Composition of Humans and Objects
Figure 4 for NCHO: Unsupervised Learning for Neural 3D Composition of Humans and Objects

Deep generative models have been recently extended to synthesizing 3D digital humans. However, previous approaches treat clothed humans as a single chunk of geometry without considering the compositionality of clothing and accessories. As a result, individual items cannot be naturally composed into novel identities, leading to limited expressiveness and controllability of generative 3D avatars. While several methods attempt to address this by leveraging synthetic data, the interaction between humans and objects is not authentic due to the domain gap, and manual asset creation is difficult to scale for a wide variety of objects. In this work, we present a novel framework for learning a compositional generative model of humans and objects (backpacks, coats, scarves, and more) from real-world 3D scans. Our compositional model is interaction-aware, meaning the spatial relationship between humans and objects, and the mutual shape change by physical contact is fully incorporated. The key challenge is that, since humans and objects are in contact, their 3D scans are merged into a single piece. To decompose them without manual annotations, we propose to leverage two sets of 3D scans of a single person with and without objects. Our approach learns to decompose objects and naturally compose them back into a generative human model in an unsupervised manner. Despite our simple setup requiring only the capture of a single subject with objects, our experiments demonstrate the strong generalization of our model by enabling the natural composition of objects to diverse identities in various poses and the composition of multiple objects, which is unseen in training data.

Viaarxiv icon

Semi-supervised Feature-Level Attribute Manipulation for Fashion Image Retrieval

Jul 11, 2019
Minchul Shin, Sanghyuk Park, Taeksoo Kim

Figure 1 for Semi-supervised Feature-Level Attribute Manipulation for Fashion Image Retrieval
Figure 2 for Semi-supervised Feature-Level Attribute Manipulation for Fashion Image Retrieval
Figure 3 for Semi-supervised Feature-Level Attribute Manipulation for Fashion Image Retrieval
Figure 4 for Semi-supervised Feature-Level Attribute Manipulation for Fashion Image Retrieval

With a growing demand for the search by image, many works have studied the task of fashion instance-level image retrieval (FIR). Furthermore, the recent works introduce a concept of fashion attribute manipulation (FAM) which manipulates a specific attribute (e.g color) of a fashion item while maintaining the rest of the attributes (e.g shape, and pattern). In this way, users can search not only "the same" items but also "similar" items with the desired attributes. FAM is a challenging task in that the attributes are hard to define, and the unique characteristics of a query are hard to be preserved. Although both FIR and FAM are important in real-life applications, most of the previous studies have focused on only one of these problem. In this study, we aim to achieve competitive performance on both FIR and FAM. To do so, we propose a novel method that converts a query into a representation with the desired attributes. We introduce a new idea of attribute manipulation at the feature level, by matching the distribution of manipulated features with real features. In this fashion, the attribute manipulation can be done independently from learning a representation from the image. By introducing the feature-level attribute manipulation, the previous methods for FIR can perform attribute manipulation without sacrificing their retrieval performance.

* Accepted to BMVC 2019 
Viaarxiv icon

Unsupervised Visual Attribute Transfer with Reconfigurable Generative Adversarial Networks

Jul 31, 2017
Taeksoo Kim, Byoungjip Kim, Moonsu Cha, Jiwon Kim

Figure 1 for Unsupervised Visual Attribute Transfer with Reconfigurable Generative Adversarial Networks
Figure 2 for Unsupervised Visual Attribute Transfer with Reconfigurable Generative Adversarial Networks
Figure 3 for Unsupervised Visual Attribute Transfer with Reconfigurable Generative Adversarial Networks
Figure 4 for Unsupervised Visual Attribute Transfer with Reconfigurable Generative Adversarial Networks

Learning to transfer visual attributes requires supervision dataset. Corresponding images with varying attribute values with the same identity are required for learning the transfer function. This largely limits their applications, because capturing them is often a difficult task. To address the issue, we propose an unsupervised method to learn to transfer visual attribute. The proposed method can learn the transfer function without any corresponding images. Inspecting visualization results from various unsupervised attribute transfer tasks, we verify the effectiveness of the proposed method.

Viaarxiv icon

Learning to Discover Cross-Domain Relations with Generative Adversarial Networks

May 15, 2017
Taeksoo Kim, Moonsu Cha, Hyunsoo Kim, Jung Kwon Lee, Jiwon Kim

Figure 1 for Learning to Discover Cross-Domain Relations with Generative Adversarial Networks
Figure 2 for Learning to Discover Cross-Domain Relations with Generative Adversarial Networks
Figure 3 for Learning to Discover Cross-Domain Relations with Generative Adversarial Networks
Figure 4 for Learning to Discover Cross-Domain Relations with Generative Adversarial Networks

While humans easily recognize relations between data from different domains without any supervision, learning to automatically discover them is in general very challenging and needs many ground-truth pairs that illustrate the relations. To avoid costly pairing, we address the task of discovering cross-domain relations given unpaired data. We propose a method based on generative adversarial networks that learns to discover relations between different domains (DiscoGAN). Using the discovered relations, our proposed network successfully transfers style from one domain to another while preserving key attributes such as orientation and face identity. Source code for official implementation is publicly available https://github.com/SKTBrain/DiscoGAN

* Accepted to International Conference on Machine Learning (ICML) 2017 
Viaarxiv icon