Picture for Wesley P. Chan

Wesley P. Chan

University of British Columbia

Robot Gaze During Autonomous Navigation and its Effect on Social Presence

May 10, 2023
Figure 1 for Robot Gaze During Autonomous Navigation and its Effect on Social Presence
Figure 2 for Robot Gaze During Autonomous Navigation and its Effect on Social Presence
Figure 3 for Robot Gaze During Autonomous Navigation and its Effect on Social Presence
Figure 4 for Robot Gaze During Autonomous Navigation and its Effect on Social Presence
Viaarxiv icon

Autonomous social robot navigation in unknown urban environments using semantic segmentation

Aug 25, 2022
Figure 1 for Autonomous social robot navigation in unknown urban environments using semantic segmentation
Figure 2 for Autonomous social robot navigation in unknown urban environments using semantic segmentation
Figure 3 for Autonomous social robot navigation in unknown urban environments using semantic segmentation
Figure 4 for Autonomous social robot navigation in unknown urban environments using semantic segmentation
Viaarxiv icon

Design and Implementation of a Human-Robot Joint Action Framework using Augmented Reality and Eye Gaze

Aug 25, 2022
Figure 1 for Design and Implementation of a Human-Robot Joint Action Framework using Augmented Reality and Eye Gaze
Figure 2 for Design and Implementation of a Human-Robot Joint Action Framework using Augmented Reality and Eye Gaze
Figure 3 for Design and Implementation of a Human-Robot Joint Action Framework using Augmented Reality and Eye Gaze
Figure 4 for Design and Implementation of a Human-Robot Joint Action Framework using Augmented Reality and Eye Gaze
Viaarxiv icon

Metrics for Evaluating Social Conformity of Crowd Navigation Algorithms

Feb 02, 2022
Figure 1 for Metrics for Evaluating Social Conformity of Crowd Navigation Algorithms
Figure 2 for Metrics for Evaluating Social Conformity of Crowd Navigation Algorithms
Figure 3 for Metrics for Evaluating Social Conformity of Crowd Navigation Algorithms
Figure 4 for Metrics for Evaluating Social Conformity of Crowd Navigation Algorithms
Viaarxiv icon

ARviz -- An Augmented Reality-enabled Visualization Platform for ROS Applications

Add code
Oct 29, 2021
Figure 1 for ARviz -- An Augmented Reality-enabled Visualization Platform for ROS Applications
Figure 2 for ARviz -- An Augmented Reality-enabled Visualization Platform for ROS Applications
Figure 3 for ARviz -- An Augmented Reality-enabled Visualization Platform for ROS Applications
Figure 4 for ARviz -- An Augmented Reality-enabled Visualization Platform for ROS Applications
Viaarxiv icon

A Proposed Set of Communicative Gestures for Human Robot Interaction and an RGB Image-based Gesture Recognizer Implemented in ROS

Add code
Sep 21, 2021
Figure 1 for A Proposed Set of Communicative Gestures for Human Robot Interaction and an RGB Image-based Gesture Recognizer Implemented in ROS
Figure 2 for A Proposed Set of Communicative Gestures for Human Robot Interaction and an RGB Image-based Gesture Recognizer Implemented in ROS
Figure 3 for A Proposed Set of Communicative Gestures for Human Robot Interaction and an RGB Image-based Gesture Recognizer Implemented in ROS
Figure 4 for A Proposed Set of Communicative Gestures for Human Robot Interaction and an RGB Image-based Gesture Recognizer Implemented in ROS
Viaarxiv icon

An Experimental Validation and Comparison of Reaching Motion Models for Unconstrained Handovers: Towards Generating Humanlike Motions for Human-Robot Handovers

Aug 29, 2021
Figure 1 for An Experimental Validation and Comparison of Reaching Motion Models for Unconstrained Handovers: Towards Generating Humanlike Motions for Human-Robot Handovers
Figure 2 for An Experimental Validation and Comparison of Reaching Motion Models for Unconstrained Handovers: Towards Generating Humanlike Motions for Human-Robot Handovers
Figure 3 for An Experimental Validation and Comparison of Reaching Motion Models for Unconstrained Handovers: Towards Generating Humanlike Motions for Human-Robot Handovers
Figure 4 for An Experimental Validation and Comparison of Reaching Motion Models for Unconstrained Handovers: Towards Generating Humanlike Motions for Human-Robot Handovers
Viaarxiv icon

Group Surfing: A Pedestrian-Based Approach to Sidewalk Robot Navigation

Apr 13, 2021
Figure 1 for Group Surfing: A Pedestrian-Based Approach to Sidewalk Robot Navigation
Figure 2 for Group Surfing: A Pedestrian-Based Approach to Sidewalk Robot Navigation
Figure 3 for Group Surfing: A Pedestrian-Based Approach to Sidewalk Robot Navigation
Figure 4 for Group Surfing: A Pedestrian-Based Approach to Sidewalk Robot Navigation
Viaarxiv icon

Virtual Barriers in Augmented Reality for Safe and Effective Human-Robot Cooperation in Manufacturing

Apr 12, 2021
Figure 1 for Virtual Barriers in Augmented Reality for Safe and Effective Human-Robot Cooperation in Manufacturing
Figure 2 for Virtual Barriers in Augmented Reality for Safe and Effective Human-Robot Cooperation in Manufacturing
Figure 3 for Virtual Barriers in Augmented Reality for Safe and Effective Human-Robot Cooperation in Manufacturing
Figure 4 for Virtual Barriers in Augmented Reality for Safe and Effective Human-Robot Cooperation in Manufacturing
Viaarxiv icon

Seeing Thru Walls: Visualizing Mobile Robots in Augmented Reality

Apr 08, 2021
Figure 1 for Seeing Thru Walls: Visualizing Mobile Robots in Augmented Reality
Figure 2 for Seeing Thru Walls: Visualizing Mobile Robots in Augmented Reality
Figure 3 for Seeing Thru Walls: Visualizing Mobile Robots in Augmented Reality
Figure 4 for Seeing Thru Walls: Visualizing Mobile Robots in Augmented Reality
Viaarxiv icon