Look and Tell: A Dataset for Multimodal Grounding Across Egocentric and Exocentric Views

Add code
Oct 26, 2025
Figure 1 for Look and Tell: A Dataset for Multimodal Grounding Across Egocentric and Exocentric Views
Figure 2 for Look and Tell: A Dataset for Multimodal Grounding Across Egocentric and Exocentric Views
Figure 3 for Look and Tell: A Dataset for Multimodal Grounding Across Egocentric and Exocentric Views
Figure 4 for Look and Tell: A Dataset for Multimodal Grounding Across Egocentric and Exocentric Views

Share this with someone who'll enjoy it:

View paper onarxiv icon

Share this with someone who'll enjoy it: