Alert button
Picture for Masood Dehghan

Masood Dehghan

Alert button

Managing Temporal Resolution in Continuous Value Estimation: A Fundamental Trade-off

Dec 17, 2022
Zichen Zhang, Johannes Kirschner, Junxi Zhang, Francesco Zanini, Alex Ayoub, Masood Dehghan, Dale Schuurmans

Figure 1 for Managing Temporal Resolution in Continuous Value Estimation: A Fundamental Trade-off
Figure 2 for Managing Temporal Resolution in Continuous Value Estimation: A Fundamental Trade-off
Figure 3 for Managing Temporal Resolution in Continuous Value Estimation: A Fundamental Trade-off
Figure 4 for Managing Temporal Resolution in Continuous Value Estimation: A Fundamental Trade-off
Viaarxiv icon

Analyzing Neural Jacobian Methods in Applications of Visual Servoing and Kinematic Control

Jun 10, 2021
Michael Przystupa, Masood Dehghan, Martin Jagersand, A. Rupam Mahmood

Figure 1 for Analyzing Neural Jacobian Methods in Applications of Visual Servoing and Kinematic Control
Figure 2 for Analyzing Neural Jacobian Methods in Applications of Visual Servoing and Kinematic Control
Figure 3 for Analyzing Neural Jacobian Methods in Applications of Visual Servoing and Kinematic Control
Figure 4 for Analyzing Neural Jacobian Methods in Applications of Visual Servoing and Kinematic Control
Viaarxiv icon

A Quantitative Analysis of Activities of Daily Living: Insights into Improving Functional Independence with Assistive Robotics

Apr 08, 2021
Laura Petrich, Jun Jin, Masood Dehghan, Martin Jagersand

Figure 1 for A Quantitative Analysis of Activities of Daily Living: Insights into Improving Functional Independence with Assistive Robotics
Figure 2 for A Quantitative Analysis of Activities of Daily Living: Insights into Improving Functional Independence with Assistive Robotics
Figure 3 for A Quantitative Analysis of Activities of Daily Living: Insights into Improving Functional Independence with Assistive Robotics
Figure 4 for A Quantitative Analysis of Activities of Daily Living: Insights into Improving Functional Independence with Assistive Robotics
Viaarxiv icon

Assistive arm and hand manipulation: How does current research intersect with actual healthcare needs?

Jan 07, 2021
Laura Petrich, Jun Jin, Masood Dehghan, Martin Jagersand

Figure 1 for Assistive arm and hand manipulation: How does current research intersect with actual healthcare needs?
Figure 2 for Assistive arm and hand manipulation: How does current research intersect with actual healthcare needs?
Figure 3 for Assistive arm and hand manipulation: How does current research intersect with actual healthcare needs?
Figure 4 for Assistive arm and hand manipulation: How does current research intersect with actual healthcare needs?
Viaarxiv icon

U$^2$-Net: Going Deeper with Nested U-Structure for Salient Object Detection

May 18, 2020
Xuebin Qin, Zichen Zhang, Chenyang Huang, Masood Dehghan, Osmar R. Zaiane, Martin Jagersand

Figure 1 for U$^2$-Net: Going Deeper with Nested U-Structure for Salient Object Detection
Figure 2 for U$^2$-Net: Going Deeper with Nested U-Structure for Salient Object Detection
Figure 3 for U$^2$-Net: Going Deeper with Nested U-Structure for Salient Object Detection
Figure 4 for U$^2$-Net: Going Deeper with Nested U-Structure for Salient Object Detection
Viaarxiv icon

A Geometric Perspective on Visual Imitation Learning

Mar 05, 2020
Jun Jin, Laura Petrich, Masood Dehghan, Martin Jagersand

Figure 1 for A Geometric Perspective on Visual Imitation Learning
Figure 2 for A Geometric Perspective on Visual Imitation Learning
Figure 3 for A Geometric Perspective on Visual Imitation Learning
Figure 4 for A Geometric Perspective on Visual Imitation Learning
Viaarxiv icon

Understanding Contexts Inside Robot and Human Manipulation Tasks through a Vision-Language Model and Ontology System in a Video Stream

Mar 02, 2020
Chen Jiang, Masood Dehghan, Martin Jagersand

Figure 1 for Understanding Contexts Inside Robot and Human Manipulation Tasks through a Vision-Language Model and Ontology System in a Video Stream
Figure 2 for Understanding Contexts Inside Robot and Human Manipulation Tasks through a Vision-Language Model and Ontology System in a Video Stream
Figure 3 for Understanding Contexts Inside Robot and Human Manipulation Tasks through a Vision-Language Model and Ontology System in a Video Stream
Figure 4 for Understanding Contexts Inside Robot and Human Manipulation Tasks through a Vision-Language Model and Ontology System in a Video Stream
Viaarxiv icon

Visual Geometric Skill Inference by Watching Human Demonstration

Nov 08, 2019
Jun Jin, Laura Petrich, Zichen Zhang, Masood Dehghan, Martin Jagersand

Figure 1 for Visual Geometric Skill Inference by Watching Human Demonstration
Figure 2 for Visual Geometric Skill Inference by Watching Human Demonstration
Figure 3 for Visual Geometric Skill Inference by Watching Human Demonstration
Figure 4 for Visual Geometric Skill Inference by Watching Human Demonstration
Viaarxiv icon

Long range teleoperation for fine manipulation tasks under time-delay network conditions

Mar 21, 2019
Jun Jin, Laura Petrich, Shida He, Masood Dehghan, Martin Jagersand

Figure 1 for Long range teleoperation for fine manipulation tasks under time-delay network conditions
Figure 2 for Long range teleoperation for fine manipulation tasks under time-delay network conditions
Figure 3 for Long range teleoperation for fine manipulation tasks under time-delay network conditions
Figure 4 for Long range teleoperation for fine manipulation tasks under time-delay network conditions
Viaarxiv icon

Evaluation of state representation methods in robot hand-eye coordination learning from demonstration

Mar 02, 2019
Jun Jin, Masood Dehghan, Laura Petrich, Steven Weikai Lu, Martin Jagersand

Figure 1 for Evaluation of state representation methods in robot hand-eye coordination learning from demonstration
Figure 2 for Evaluation of state representation methods in robot hand-eye coordination learning from demonstration
Figure 3 for Evaluation of state representation methods in robot hand-eye coordination learning from demonstration
Figure 4 for Evaluation of state representation methods in robot hand-eye coordination learning from demonstration
Viaarxiv icon