Alert button
Picture for Ahmed H. Qureshi

Ahmed H. Qureshi

Alert button

DeRi-IGP: Manipulating Rigid Objects Using Deformable Objects via Iterative Grasp-Pull

Sep 09, 2023
Zixing Wang, Ahmed H. Qureshi

Figure 1 for DeRi-IGP: Manipulating Rigid Objects Using Deformable Objects via Iterative Grasp-Pull
Figure 2 for DeRi-IGP: Manipulating Rigid Objects Using Deformable Objects via Iterative Grasp-Pull
Figure 3 for DeRi-IGP: Manipulating Rigid Objects Using Deformable Objects via Iterative Grasp-Pull
Figure 4 for DeRi-IGP: Manipulating Rigid Objects Using Deformable Objects via Iterative Grasp-Pull

Heterogeneous systems manipulation, i.e., manipulating rigid objects via deformable (soft) objects, is an emerging field that remains in its early stages of research. Existing works in this field suffer from limited action and operational space, poor generalization ability, and expensive development. To address these challenges, we propose a universally applicable and effective moving primitive, Iterative Grasp-Pull (IGP), and a sample-based framework, DeRi-IGP, to solve the heterogeneous system manipulation task. The DeRi-IGP framework uses local onboard robots' RGBD sensors to observe the environment, comprising a soft-rigid body system. It then uses this information to iteratively grasp and pull a soft body (e.g., rope) to move the attached rigid body to a desired location. We evaluate the effectiveness of our framework in solving various heterogeneous manipulation tasks and compare its performance with several state-of-the-art baselines. The result shows that DeRi-IGP outperforms other methods by a significant margin. In addition, we also demonstrate the advantage of the large operational space of IGP in the long-distance object acquisition task within both simulated and real environments.

Viaarxiv icon

AnyPose: Anytime 3D Human Pose Forecasting via Neural Ordinary Differential Equations

Sep 09, 2023
Zixing Wang, Ahmed H. Qureshi

Figure 1 for AnyPose: Anytime 3D Human Pose Forecasting via Neural Ordinary Differential Equations
Figure 2 for AnyPose: Anytime 3D Human Pose Forecasting via Neural Ordinary Differential Equations
Figure 3 for AnyPose: Anytime 3D Human Pose Forecasting via Neural Ordinary Differential Equations
Figure 4 for AnyPose: Anytime 3D Human Pose Forecasting via Neural Ordinary Differential Equations

Anytime 3D human pose forecasting is crucial to synchronous real-world human-machine interaction, where the term ``anytime" corresponds to predicting human pose at any real-valued time step. However, to the best of our knowledge, all the existing methods in human pose forecasting perform predictions at preset, discrete time intervals. Therefore, we introduce AnyPose, a lightweight continuous-time neural architecture that models human behavior dynamics with neural ordinary differential equations. We validate our framework on the Human3.6M, AMASS, and 3DPW dataset and conduct a series of comprehensive analyses towards comparison with existing methods and the intersection of human pose and neural ordinary differential equations. Our results demonstrate that AnyPose exhibits high-performance accuracy in predicting future poses and takes significantly lower computational time than traditional methods in solving anytime prediction tasks.

Viaarxiv icon

Structural Concept Learning via Graph Attention for Multi-Level Rearrangement Planning

Sep 05, 2023
Manav Kulshrestha, Ahmed H. Qureshi

Robotic manipulation tasks, such as object rearrangement, play a crucial role in enabling robots to interact with complex and arbitrary environments. Existing work focuses primarily on single-level rearrangement planning and, even if multiple levels exist, dependency relations among substructures are geometrically simpler, like tower stacking. We propose Structural Concept Learning (SCL), a deep learning approach that leverages graph attention networks to perform multi-level object rearrangement planning for scenes with structural dependency hierarchies. It is trained on a self-generated simulation data set with intuitive structures, works for unseen scenes with an arbitrary number of objects and higher complexity of structures, infers independent substructures to allow for task parallelization over multiple manipulators, and generalizes to the real world. We compare our method with a range of classical and model-based baselines to show that our method leverages its scene understanding to achieve better performance, flexibility, and efficiency. The dataset, supplementary details, videos, and code implementation are available at: https://manavkulshrestha.github.io/scl

* Accepted to Conference on Robot Learning (CoRL) 2023 
Viaarxiv icon

Efficient Q-Learning over Visit Frequency Maps for Multi-agent Exploration of Unknown Environments

Jul 30, 2023
Xuyang Chen, Ashvin N. Iyer, Zixing Wang, Ahmed H. Qureshi

The robot exploration task has been widely studied with applications spanning from novel environment mapping to item delivery. For some time-critical tasks, such as rescue catastrophes, the agent is required to explore as efficiently as possible. Recently, Visit Frequency-based map representation achieved great success in such scenarios by discouraging repetitive visits with a frequency-based penalty. However, its relatively large size and single-agent settings hinder its further development. In this context, we propose Integrated Visit Frequency Map, which encodes identical information as Visit Frequency Map with a more compact size, and a visit frequency-based multi-agent information exchange and control scheme that is able to accommodate both representations. Through tests in diverse settings, the results indicate our proposed methods can achieve a comparable level of performance of VFM with lower bandwidth requirements and generalize well to different multi-agent setups including real-world environments.

* Accepted by IROS 2023. 8 pages 
Viaarxiv icon

SIMF: Semantics-aware Interactive Motion Forecasting for Autonomous Driving

Jun 26, 2023
Vidyaa Krishnan Nivash, Ahmed H. Qureshi

Figure 1 for SIMF: Semantics-aware Interactive Motion Forecasting for Autonomous Driving
Figure 2 for SIMF: Semantics-aware Interactive Motion Forecasting for Autonomous Driving
Figure 3 for SIMF: Semantics-aware Interactive Motion Forecasting for Autonomous Driving
Figure 4 for SIMF: Semantics-aware Interactive Motion Forecasting for Autonomous Driving

Autonomous vehicles require motion forecasting of their surrounding multi-agents (pedestrians and vehicles) to make optimal decisions for navigation. The existing methods focus on techniques to utilize the positions and velocities of these agents and fail to capture semantic information from the scene. Moreover, to mitigate the increase in computational complexity associated with the number of agents in the scene, some works leverage Euclidean distance to prune far-away agents. However, distance-based metric alone is insufficient to select relevant agents and accurately perform their predictions. To resolve these issues, we propose Semantics-aware Interactive Motion Forecasting (SIMF) method to capture semantics along with spatial information, and optimally select relevant agents for motion prediction. Specifically, we achieve this by implementing a semantic-aware selection of relevant agents from the scene and passing them through an attention mechanism to extract global encodings. These encodings along with agents' local information are passed through an encoder to obtain time-dependent latent variables for a motion policy predicting the future trajectories. Our results show that the proposed approach outperforms state-of-the-art baselines and provides more accurate predictions in a scene-consistent manner.

Viaarxiv icon

MANER: Multi-Agent Neural Rearrangement Planning of Objects in Cluttered Environments

Jun 10, 2023
Vivek Gupta, Praphpreet Dhir, Jeegn Dani, Ahmed H. Qureshi

Figure 1 for MANER: Multi-Agent Neural Rearrangement Planning of Objects in Cluttered Environments
Figure 2 for MANER: Multi-Agent Neural Rearrangement Planning of Objects in Cluttered Environments
Figure 3 for MANER: Multi-Agent Neural Rearrangement Planning of Objects in Cluttered Environments
Figure 4 for MANER: Multi-Agent Neural Rearrangement Planning of Objects in Cluttered Environments

Object rearrangement is a fundamental problem in robotics with various practical applications ranging from managing warehouses to cleaning and organizing home kitchens. While existing research has primarily focused on single-agent solutions, real-world scenarios often require multiple robots to work together on rearrangement tasks. This paper proposes a comprehensive learning-based framework for multi-agent object rearrangement planning, addressing the challenges of task sequencing and path planning in complex environments. The proposed method iteratively selects objects, determines their relocation regions, and pairs them with available robots under kinematic feasibility and task reachability for execution to achieve the target arrangement. Our experiments on a diverse range of environments demonstrate the effectiveness and robustness of the proposed framework. Furthermore, results indicate improved performance in terms of traversal time and success rate compared to baseline approaches.

* The videos and supplementary material are available at https://sites.google.com/view/maner-supplementary 
Viaarxiv icon

Progressive Learning for Physics-informed Neural Motion Planning

Jun 01, 2023
Ruiqi Ni, Ahmed H. Qureshi

Figure 1 for Progressive Learning for Physics-informed Neural Motion Planning
Figure 2 for Progressive Learning for Physics-informed Neural Motion Planning
Figure 3 for Progressive Learning for Physics-informed Neural Motion Planning
Figure 4 for Progressive Learning for Physics-informed Neural Motion Planning

Motion planning (MP) is one of the core robotics problems requiring fast methods for finding a collision-free robot motion path connecting the given start and goal states. Neural motion planners (NMPs) demonstrate fast computational speed in finding path solutions but require a huge amount of expert trajectories for learning, thus adding a significant training computational load. In contrast, recent advancements have also led to a physics-informed NMP approach that directly solves the Eikonal equation for motion planning and does not require expert demonstrations for learning. However, experiments show that the physics-informed NMP approach performs poorly in complex environments and lacks scalability in multiple scenarios and high-dimensional real robot settings. To overcome these limitations, this paper presents a novel and tractable Eikonal equation formulation and introduces a new progressive learning strategy to train neural networks without expert data in complex, cluttered, multiple high-dimensional robot motion planning scenarios. The results demonstrate that our method outperforms state-of-the-art traditional MP, data-driven NMP, and physics-informed NMP methods by a significant margin in terms of computational planning speed, path quality, and success rates. We also show that our approach scales to multiple complex, cluttered scenarios and the real robot set up in a narrow passage environment. The proposed method's videos and code implementations are available at https://github.com/ruiqini/P-NTFields.

* Accepted to Robotics: Science and Systems (RSS) 2023 
Viaarxiv icon

Multi-Stage Monte Carlo Tree Search for Non-Monotone Object Rearrangement Planning in Narrow Confined Environments

May 26, 2023
Hanwen Ren, Ahmed H. Qureshi

Figure 1 for Multi-Stage Monte Carlo Tree Search for Non-Monotone Object Rearrangement Planning in Narrow Confined Environments
Figure 2 for Multi-Stage Monte Carlo Tree Search for Non-Monotone Object Rearrangement Planning in Narrow Confined Environments
Figure 3 for Multi-Stage Monte Carlo Tree Search for Non-Monotone Object Rearrangement Planning in Narrow Confined Environments
Figure 4 for Multi-Stage Monte Carlo Tree Search for Non-Monotone Object Rearrangement Planning in Narrow Confined Environments

Non-monotone object rearrangement planning in confined spaces such as cabinets and shelves is a widely occurring but challenging problem in robotics. Both the robot motion and the available regions for object relocation are highly constrained because of the limited space. This work proposes a Multi-Stage Monte Carlo Tree Search (MS-MCTS) method to solve non-monotone object rearrangement planning problems in confined spaces. Our approach decouples the complex problem into simpler subproblems using an object stage topology. A subgoal-focused tree expansion algorithm that jointly considers the high-level planning and the low-level robot motion is designed to reduce the search space and better guide the search process. By fitting the task into the MCTS paradigm, our method produces optimistic solutions by balancing exploration and exploitation. The experiments demonstrate that our method outperforms the existing methods regarding the planning time, the number of steps, and the total move distance. Moreover, we deploy our MS-MCTS to a real-world robot system and verify its performance in different confined environments.

Viaarxiv icon

DeRi-Bot: Learning to Collaboratively Manipulate Rigid Objects via Deformable Objects

May 22, 2023
Zixing Wang, Ahmed H. Qureshi

Figure 1 for DeRi-Bot: Learning to Collaboratively Manipulate Rigid Objects via Deformable Objects
Figure 2 for DeRi-Bot: Learning to Collaboratively Manipulate Rigid Objects via Deformable Objects
Figure 3 for DeRi-Bot: Learning to Collaboratively Manipulate Rigid Objects via Deformable Objects
Figure 4 for DeRi-Bot: Learning to Collaboratively Manipulate Rigid Objects via Deformable Objects

Recent research efforts have yielded significant advancements in manipulating objects under homogeneous settings where the robot is required to either manipulate rigid or deformable (soft) objects. However, the manipulation under heterogeneous setups that involve both deformable and rigid objects remains an unexplored area of research. Such setups are common in various scenarios that involve the transportation of heavy objects via ropes, e.g., on factory floors, at disaster sites, and in forestry. To address this challenge, we introduce DeRi-Bot, the first framework that enables the collaborative manipulation of rigid objects with deformable objects. Our framework comprises an Action Prediction Network (APN) and a Configuration Prediction Network (CPN) to model the complex pattern and stochasticity of soft-rigid body systems. We demonstrate the effectiveness of DeRi-Bot in moving rigid objects to a target position with ropes connected to robotic arms. Furthermore, DeRi-Bot is a distributive method that can accommodate an arbitrary number of robots or human partners without reconfiguration or retraining. We evaluate our framework in both simulated and real-world environments and show that it achieves promising results with strong generalization across different types of objects and multi-agent settings, including human-robot collaboration.

* 8 pages, 7 figures 
Viaarxiv icon

Merging Decision Transformers: Weight Averaging for Forming Multi-Task Policies

Mar 14, 2023
Daniel Lawson, Ahmed H. Qureshi

Figure 1 for Merging Decision Transformers: Weight Averaging for Forming Multi-Task Policies
Figure 2 for Merging Decision Transformers: Weight Averaging for Forming Multi-Task Policies
Figure 3 for Merging Decision Transformers: Weight Averaging for Forming Multi-Task Policies
Figure 4 for Merging Decision Transformers: Weight Averaging for Forming Multi-Task Policies

Recent work has shown the promise of creating generalist, transformer-based, policies for language, vision, and sequential decision-making problems. To create such models, we generally require centralized training objectives, data, and compute. It is of interest if we can more flexibly create generalist policies, by merging together multiple, task-specific, individually trained policies. In this work, we take a preliminary step in this direction through merging, or averaging, subsets of Decision Transformers in weight space trained on different MuJoCo locomotion problems, forming multi-task models without centralized training. We also propose that when merging policies, we can obtain better results if all policies start from common, pre-trained initializations, while also co-training on shared auxiliary tasks during problem-specific finetuning. In general, we believe research in this direction can help democratize and distribute the process of which forms generally capable agents.

Viaarxiv icon