Picture for Wesley P. Chan

Wesley P. Chan

University of British Columbia

Robot Gaze During Autonomous Navigation and its Effect on Social Presence

Add code
May 10, 2023
Figure 1 for Robot Gaze During Autonomous Navigation and its Effect on Social Presence
Figure 2 for Robot Gaze During Autonomous Navigation and its Effect on Social Presence
Figure 3 for Robot Gaze During Autonomous Navigation and its Effect on Social Presence
Figure 4 for Robot Gaze During Autonomous Navigation and its Effect on Social Presence
Viaarxiv icon

Autonomous social robot navigation in unknown urban environments using semantic segmentation

Add code
Aug 25, 2022
Figure 1 for Autonomous social robot navigation in unknown urban environments using semantic segmentation
Figure 2 for Autonomous social robot navigation in unknown urban environments using semantic segmentation
Figure 3 for Autonomous social robot navigation in unknown urban environments using semantic segmentation
Figure 4 for Autonomous social robot navigation in unknown urban environments using semantic segmentation
Viaarxiv icon

Design and Implementation of a Human-Robot Joint Action Framework using Augmented Reality and Eye Gaze

Add code
Aug 25, 2022
Figure 1 for Design and Implementation of a Human-Robot Joint Action Framework using Augmented Reality and Eye Gaze
Figure 2 for Design and Implementation of a Human-Robot Joint Action Framework using Augmented Reality and Eye Gaze
Figure 3 for Design and Implementation of a Human-Robot Joint Action Framework using Augmented Reality and Eye Gaze
Figure 4 for Design and Implementation of a Human-Robot Joint Action Framework using Augmented Reality and Eye Gaze
Viaarxiv icon

Metrics for Evaluating Social Conformity of Crowd Navigation Algorithms

Add code
Feb 02, 2022
Figure 1 for Metrics for Evaluating Social Conformity of Crowd Navigation Algorithms
Figure 2 for Metrics for Evaluating Social Conformity of Crowd Navigation Algorithms
Figure 3 for Metrics for Evaluating Social Conformity of Crowd Navigation Algorithms
Figure 4 for Metrics for Evaluating Social Conformity of Crowd Navigation Algorithms
Viaarxiv icon

ARviz -- An Augmented Reality-enabled Visualization Platform for ROS Applications

Add code
Oct 29, 2021
Figure 1 for ARviz -- An Augmented Reality-enabled Visualization Platform for ROS Applications
Figure 2 for ARviz -- An Augmented Reality-enabled Visualization Platform for ROS Applications
Figure 3 for ARviz -- An Augmented Reality-enabled Visualization Platform for ROS Applications
Figure 4 for ARviz -- An Augmented Reality-enabled Visualization Platform for ROS Applications
Viaarxiv icon

A Proposed Set of Communicative Gestures for Human Robot Interaction and an RGB Image-based Gesture Recognizer Implemented in ROS

Add code
Sep 21, 2021
Figure 1 for A Proposed Set of Communicative Gestures for Human Robot Interaction and an RGB Image-based Gesture Recognizer Implemented in ROS
Figure 2 for A Proposed Set of Communicative Gestures for Human Robot Interaction and an RGB Image-based Gesture Recognizer Implemented in ROS
Figure 3 for A Proposed Set of Communicative Gestures for Human Robot Interaction and an RGB Image-based Gesture Recognizer Implemented in ROS
Figure 4 for A Proposed Set of Communicative Gestures for Human Robot Interaction and an RGB Image-based Gesture Recognizer Implemented in ROS
Viaarxiv icon

An Experimental Validation and Comparison of Reaching Motion Models for Unconstrained Handovers: Towards Generating Humanlike Motions for Human-Robot Handovers

Add code
Aug 29, 2021
Figure 1 for An Experimental Validation and Comparison of Reaching Motion Models for Unconstrained Handovers: Towards Generating Humanlike Motions for Human-Robot Handovers
Figure 2 for An Experimental Validation and Comparison of Reaching Motion Models for Unconstrained Handovers: Towards Generating Humanlike Motions for Human-Robot Handovers
Figure 3 for An Experimental Validation and Comparison of Reaching Motion Models for Unconstrained Handovers: Towards Generating Humanlike Motions for Human-Robot Handovers
Figure 4 for An Experimental Validation and Comparison of Reaching Motion Models for Unconstrained Handovers: Towards Generating Humanlike Motions for Human-Robot Handovers
Viaarxiv icon

Group Surfing: A Pedestrian-Based Approach to Sidewalk Robot Navigation

Add code
Apr 13, 2021
Figure 1 for Group Surfing: A Pedestrian-Based Approach to Sidewalk Robot Navigation
Figure 2 for Group Surfing: A Pedestrian-Based Approach to Sidewalk Robot Navigation
Figure 3 for Group Surfing: A Pedestrian-Based Approach to Sidewalk Robot Navigation
Figure 4 for Group Surfing: A Pedestrian-Based Approach to Sidewalk Robot Navigation
Viaarxiv icon

Virtual Barriers in Augmented Reality for Safe and Effective Human-Robot Cooperation in Manufacturing

Add code
Apr 12, 2021
Figure 1 for Virtual Barriers in Augmented Reality for Safe and Effective Human-Robot Cooperation in Manufacturing
Figure 2 for Virtual Barriers in Augmented Reality for Safe and Effective Human-Robot Cooperation in Manufacturing
Figure 3 for Virtual Barriers in Augmented Reality for Safe and Effective Human-Robot Cooperation in Manufacturing
Figure 4 for Virtual Barriers in Augmented Reality for Safe and Effective Human-Robot Cooperation in Manufacturing
Viaarxiv icon

Seeing Thru Walls: Visualizing Mobile Robots in Augmented Reality

Add code
Apr 08, 2021
Figure 1 for Seeing Thru Walls: Visualizing Mobile Robots in Augmented Reality
Figure 2 for Seeing Thru Walls: Visualizing Mobile Robots in Augmented Reality
Figure 3 for Seeing Thru Walls: Visualizing Mobile Robots in Augmented Reality
Figure 4 for Seeing Thru Walls: Visualizing Mobile Robots in Augmented Reality
Viaarxiv icon