Picture for Roozbeh Mottaghi

Roozbeh Mottaghi

ADAPT: Actively Discovering and Adapting to Preferences for any Task

Add code
Apr 05, 2025
Viaarxiv icon

Grounding Multimodal LLMs to Embodied Agents that Ask for Help with Reinforcement Learning

Add code
Apr 02, 2025
Viaarxiv icon

From an Image to a Scene: Learning to Imagine the World from a Million 360 Videos

Add code
Dec 10, 2024
Viaarxiv icon

PARTNR: A Benchmark for Planning and Reasoning in Embodied Multi-agent Tasks

Add code
Oct 31, 2024
Figure 1 for PARTNR: A Benchmark for Planning and Reasoning in Embodied Multi-agent Tasks
Figure 2 for PARTNR: A Benchmark for Planning and Reasoning in Embodied Multi-agent Tasks
Figure 3 for PARTNR: A Benchmark for Planning and Reasoning in Embodied Multi-agent Tasks
Figure 4 for PARTNR: A Benchmark for Planning and Reasoning in Embodied Multi-agent Tasks
Viaarxiv icon

Situated Instruction Following

Add code
Jul 15, 2024
Figure 1 for Situated Instruction Following
Figure 2 for Situated Instruction Following
Figure 3 for Situated Instruction Following
Figure 4 for Situated Instruction Following
Viaarxiv icon

Towards Open-World Mobile Manipulation in Homes: Lessons from the Neurips 2023 HomeRobot Open Vocabulary Mobile Manipulation Challenge

Add code
Jul 09, 2024
Figure 1 for Towards Open-World Mobile Manipulation in Homes: Lessons from the Neurips 2023 HomeRobot Open Vocabulary Mobile Manipulation Challenge
Figure 2 for Towards Open-World Mobile Manipulation in Homes: Lessons from the Neurips 2023 HomeRobot Open Vocabulary Mobile Manipulation Challenge
Figure 3 for Towards Open-World Mobile Manipulation in Homes: Lessons from the Neurips 2023 HomeRobot Open Vocabulary Mobile Manipulation Challenge
Figure 4 for Towards Open-World Mobile Manipulation in Homes: Lessons from the Neurips 2023 HomeRobot Open Vocabulary Mobile Manipulation Challenge
Viaarxiv icon

Track2Act: Predicting Point Tracks from Internet Videos enables Diverse Zero-shot Robot Manipulation

Add code
May 02, 2024
Figure 1 for Track2Act: Predicting Point Tracks from Internet Videos enables Diverse Zero-shot Robot Manipulation
Figure 2 for Track2Act: Predicting Point Tracks from Internet Videos enables Diverse Zero-shot Robot Manipulation
Figure 3 for Track2Act: Predicting Point Tracks from Internet Videos enables Diverse Zero-shot Robot Manipulation
Figure 4 for Track2Act: Predicting Point Tracks from Internet Videos enables Diverse Zero-shot Robot Manipulation
Viaarxiv icon

GOAT-Bench: A Benchmark for Multi-Modal Lifelong Navigation

Add code
Apr 09, 2024
Figure 1 for GOAT-Bench: A Benchmark for Multi-Modal Lifelong Navigation
Figure 2 for GOAT-Bench: A Benchmark for Multi-Modal Lifelong Navigation
Figure 3 for GOAT-Bench: A Benchmark for Multi-Modal Lifelong Navigation
Figure 4 for GOAT-Bench: A Benchmark for Multi-Modal Lifelong Navigation
Viaarxiv icon

Controllable Human-Object Interaction Synthesis

Add code
Dec 06, 2023
Figure 1 for Controllable Human-Object Interaction Synthesis
Figure 2 for Controllable Human-Object Interaction Synthesis
Figure 3 for Controllable Human-Object Interaction Synthesis
Figure 4 for Controllable Human-Object Interaction Synthesis
Viaarxiv icon

GOAT: GO to Any Thing

Add code
Nov 10, 2023
Figure 1 for GOAT: GO to Any Thing
Figure 2 for GOAT: GO to Any Thing
Figure 3 for GOAT: GO to Any Thing
Figure 4 for GOAT: GO to Any Thing
Viaarxiv icon