Alert button
Picture for Mo Han

Mo Han

Alert button

Segmentation and Classification of EMG Time-Series During Reach-to-Grasp Motion

Add code
Bookmark button
Alert button
Apr 19, 2021
Mo Han, Mehrshad Zandigohar, Mariusz P. Furmanek, Mathew Yarossi, Gunar Schirner, Deniz Erdogmus

Figure 1 for Segmentation and Classification of EMG Time-Series During Reach-to-Grasp Motion
Figure 2 for Segmentation and Classification of EMG Time-Series During Reach-to-Grasp Motion
Figure 3 for Segmentation and Classification of EMG Time-Series During Reach-to-Grasp Motion
Figure 4 for Segmentation and Classification of EMG Time-Series During Reach-to-Grasp Motion
Viaarxiv icon

Multimodal Fusion of EMG and Vision for Human Grasp Intent Inference in Prosthetic Hand Control

Add code
Bookmark button
Alert button
Apr 08, 2021
Mehrshad Zandigohar, Mo Han, Mohammadreza Sharif, Sezen Yagmur Gunay, Mariusz P. Furmanek, Mathew Yarossi, Paolo Bonato, Cagdas Onal, Taskin Padir, Deniz Erdogmus, Gunar Schirner

Figure 1 for Multimodal Fusion of EMG and Vision for Human Grasp Intent Inference in Prosthetic Hand Control
Figure 2 for Multimodal Fusion of EMG and Vision for Human Grasp Intent Inference in Prosthetic Hand Control
Figure 3 for Multimodal Fusion of EMG and Vision for Human Grasp Intent Inference in Prosthetic Hand Control
Figure 4 for Multimodal Fusion of EMG and Vision for Human Grasp Intent Inference in Prosthetic Hand Control
Viaarxiv icon

From Hand-Perspective Visual Information to Grasp Type Probabilities: Deep Learning via Ranking Labels

Add code
Bookmark button
Alert button
Mar 08, 2021
Mo Han, Sezen Ya{ğ}mur Günay, İlkay Yıldız, Paolo Bonato, Cagdas D. Onal, Taşkın Padır, Gunar Schirner, Deniz Erdo{ğ}muş

Figure 1 for From Hand-Perspective Visual Information to Grasp Type Probabilities: Deep Learning via Ranking Labels
Figure 2 for From Hand-Perspective Visual Information to Grasp Type Probabilities: Deep Learning via Ranking Labels
Figure 3 for From Hand-Perspective Visual Information to Grasp Type Probabilities: Deep Learning via Ranking Labels
Figure 4 for From Hand-Perspective Visual Information to Grasp Type Probabilities: Deep Learning via Ranking Labels
Viaarxiv icon

HANDS: A Multimodal Dataset for Modeling Towards Human Grasp Intent Inference in Prosthetic Hands

Add code
Bookmark button
Alert button
Mar 08, 2021
Mo Han, Sezen Ya{ğ}mur Günay, Gunar Schirner, Taşkın Padır, Deniz Erdo{ğ}muş

Figure 1 for HANDS: A Multimodal Dataset for Modeling Towards Human Grasp Intent Inference in Prosthetic Hands
Figure 2 for HANDS: A Multimodal Dataset for Modeling Towards Human Grasp Intent Inference in Prosthetic Hands
Figure 3 for HANDS: A Multimodal Dataset for Modeling Towards Human Grasp Intent Inference in Prosthetic Hands
Figure 4 for HANDS: A Multimodal Dataset for Modeling Towards Human Grasp Intent Inference in Prosthetic Hands
Viaarxiv icon

Towards Creating a Deployable Grasp Type Probability Estimator for a Prosthetic Hand

Add code
Bookmark button
Alert button
Jan 13, 2021
Mehrshad Zandigohar, Mo Han, Deniz Erdogmus, Gunar Schirner

Figure 1 for Towards Creating a Deployable Grasp Type Probability Estimator for a Prosthetic Hand
Figure 2 for Towards Creating a Deployable Grasp Type Probability Estimator for a Prosthetic Hand
Figure 3 for Towards Creating a Deployable Grasp Type Probability Estimator for a Prosthetic Hand
Figure 4 for Towards Creating a Deployable Grasp Type Probability Estimator for a Prosthetic Hand
Viaarxiv icon

Universal Physiological Representation Learning with Soft-Disentangled Rateless Autoencoders

Add code
Bookmark button
Alert button
Sep 28, 2020
Mo Han, Ozan Ozdenizci, Toshiaki Koike-Akino, Ye Wang, Deniz Erdogmus

Figure 1 for Universal Physiological Representation Learning with Soft-Disentangled Rateless Autoencoders
Figure 2 for Universal Physiological Representation Learning with Soft-Disentangled Rateless Autoencoders
Figure 3 for Universal Physiological Representation Learning with Soft-Disentangled Rateless Autoencoders
Figure 4 for Universal Physiological Representation Learning with Soft-Disentangled Rateless Autoencoders
Viaarxiv icon

Disentangled Adversarial Autoencoder for Subject-Invariant Physiological Feature Extraction

Add code
Bookmark button
Alert button
Aug 26, 2020
Mo Han, Ozan Ozdenizci, Ye Wang, Toshiaki Koike-Akino, Deniz Erdogmus

Figure 1 for Disentangled Adversarial Autoencoder for Subject-Invariant Physiological Feature Extraction
Figure 2 for Disentangled Adversarial Autoencoder for Subject-Invariant Physiological Feature Extraction
Figure 3 for Disentangled Adversarial Autoencoder for Subject-Invariant Physiological Feature Extraction
Figure 4 for Disentangled Adversarial Autoencoder for Subject-Invariant Physiological Feature Extraction
Viaarxiv icon

Disentangled Adversarial Transfer Learning for Physiological Biosignals

Add code
Bookmark button
Alert button
Apr 15, 2020
Mo Han, Ozan Ozdenizci, Ye Wang, Toshiaki Koike-Akino, Deniz Erdogmus

Figure 1 for Disentangled Adversarial Transfer Learning for Physiological Biosignals
Figure 2 for Disentangled Adversarial Transfer Learning for Physiological Biosignals
Figure 3 for Disentangled Adversarial Transfer Learning for Physiological Biosignals
Viaarxiv icon