Picture for Alexander Toshev

Alexander Toshev

Apple

STAIR: Learning Sparse Text and Image Representation in Grounded Tokens

Add code
Feb 08, 2023
Viaarxiv icon

Perceptual Grouping in Vision-Language Models

Add code
Oct 18, 2022
Figure 1 for Perceptual Grouping in Vision-Language Models
Figure 2 for Perceptual Grouping in Vision-Language Models
Figure 3 for Perceptual Grouping in Vision-Language Models
Figure 4 for Perceptual Grouping in Vision-Language Models
Viaarxiv icon

Retrospectives on the Embodied AI Workshop

Add code
Oct 17, 2022
Figure 1 for Retrospectives on the Embodied AI Workshop
Figure 2 for Retrospectives on the Embodied AI Workshop
Figure 3 for Retrospectives on the Embodied AI Workshop
Figure 4 for Retrospectives on the Embodied AI Workshop
Viaarxiv icon

Gesture2Path: Imitation Learning for Gesture-aware Navigation

Add code
Sep 19, 2022
Figure 1 for Gesture2Path: Imitation Learning for Gesture-aware Navigation
Figure 2 for Gesture2Path: Imitation Learning for Gesture-aware Navigation
Figure 3 for Gesture2Path: Imitation Learning for Gesture-aware Navigation
Figure 4 for Gesture2Path: Imitation Learning for Gesture-aware Navigation
Viaarxiv icon

GAUDI: A Neural Architect for Immersive 3D Scene Generation

Add code
Jul 27, 2022
Figure 1 for GAUDI: A Neural Architect for Immersive 3D Scene Generation
Figure 2 for GAUDI: A Neural Architect for Immersive 3D Scene Generation
Figure 3 for GAUDI: A Neural Architect for Immersive 3D Scene Generation
Figure 4 for GAUDI: A Neural Architect for Immersive 3D Scene Generation
Viaarxiv icon

A Protocol for Validating Social Navigation Policies

Add code
Apr 11, 2022
Figure 1 for A Protocol for Validating Social Navigation Policies
Figure 2 for A Protocol for Validating Social Navigation Policies
Figure 3 for A Protocol for Validating Social Navigation Policies
Figure 4 for A Protocol for Validating Social Navigation Policies
Viaarxiv icon

Do As I Can, Not As I Say: Grounding Language in Robotic Affordances

Add code
Apr 04, 2022
Figure 1 for Do As I Can, Not As I Say: Grounding Language in Robotic Affordances
Figure 2 for Do As I Can, Not As I Say: Grounding Language in Robotic Affordances
Figure 3 for Do As I Can, Not As I Say: Grounding Language in Robotic Affordances
Figure 4 for Do As I Can, Not As I Say: Grounding Language in Robotic Affordances
Viaarxiv icon

Socially Compliant Navigation Dataset (SCAND): A Large-Scale Dataset of Demonstrations for Social Navigation

Add code
Mar 28, 2022
Figure 1 for Socially Compliant Navigation Dataset (SCAND): A Large-Scale Dataset of Demonstrations for Social Navigation
Figure 2 for Socially Compliant Navigation Dataset (SCAND): A Large-Scale Dataset of Demonstrations for Social Navigation
Figure 3 for Socially Compliant Navigation Dataset (SCAND): A Large-Scale Dataset of Demonstrations for Social Navigation
Figure 4 for Socially Compliant Navigation Dataset (SCAND): A Large-Scale Dataset of Demonstrations for Social Navigation
Viaarxiv icon

Value Function Spaces: Skill-Centric State Abstractions for Long-Horizon Reasoning

Add code
Nov 04, 2021
Figure 1 for Value Function Spaces: Skill-Centric State Abstractions for Long-Horizon Reasoning
Figure 2 for Value Function Spaces: Skill-Centric State Abstractions for Long-Horizon Reasoning
Figure 3 for Value Function Spaces: Skill-Centric State Abstractions for Long-Horizon Reasoning
Figure 4 for Value Function Spaces: Skill-Centric State Abstractions for Long-Horizon Reasoning
Viaarxiv icon

ReLMoGen: Leveraging Motion Generation in Reinforcement Learning for Mobile Manipulation

Add code
Aug 18, 2020
Figure 1 for ReLMoGen: Leveraging Motion Generation in Reinforcement Learning for Mobile Manipulation
Figure 2 for ReLMoGen: Leveraging Motion Generation in Reinforcement Learning for Mobile Manipulation
Figure 3 for ReLMoGen: Leveraging Motion Generation in Reinforcement Learning for Mobile Manipulation
Figure 4 for ReLMoGen: Leveraging Motion Generation in Reinforcement Learning for Mobile Manipulation
Viaarxiv icon