Alert button
Picture for Anna Rohrbach

Anna Rohrbach

Alert button

CLIP-It! Language-Guided Video Summarization

Jul 01, 2021
Medhini Narasimhan, Anna Rohrbach, Trevor Darrell

Figure 1 for CLIP-It! Language-Guided Video Summarization
Figure 2 for CLIP-It! Language-Guided Video Summarization
Figure 3 for CLIP-It! Language-Guided Video Summarization
Figure 4 for CLIP-It! Language-Guided Video Summarization
Viaarxiv icon

DETReg: Unsupervised Pretraining with Region Priors for Object Detection

Jun 08, 2021
Amir Bar, Xin Wang, Vadim Kantorov, Colorado J Reed, Roei Herzig, Gal Chechik, Anna Rohrbach, Trevor Darrell, Amir Globerson

Figure 1 for DETReg: Unsupervised Pretraining with Region Priors for Object Detection
Figure 2 for DETReg: Unsupervised Pretraining with Region Priors for Object Detection
Figure 3 for DETReg: Unsupervised Pretraining with Region Priors for Object Detection
Figure 4 for DETReg: Unsupervised Pretraining with Region Priors for Object Detection
Viaarxiv icon

NewsCLIPpings: Automatic Generation of Out-of-Context Multimodal Media

Apr 13, 2021
Grace Luo, Trevor Darrell, Anna Rohrbach

Figure 1 for NewsCLIPpings: Automatic Generation of Out-of-Context Multimodal Media
Figure 2 for NewsCLIPpings: Automatic Generation of Out-of-Context Multimodal Media
Figure 3 for NewsCLIPpings: Automatic Generation of Out-of-Context Multimodal Media
Figure 4 for NewsCLIPpings: Automatic Generation of Out-of-Context Multimodal Media
Viaarxiv icon

Identity-Aware Multi-Sentence Video Description

Aug 22, 2020
Jae Sung Park, Trevor Darrell, Anna Rohrbach

Figure 1 for Identity-Aware Multi-Sentence Video Description
Figure 2 for Identity-Aware Multi-Sentence Video Description
Figure 3 for Identity-Aware Multi-Sentence Video Description
Figure 4 for Identity-Aware Multi-Sentence Video Description
Viaarxiv icon

Are You Looking? Grounding to Multiple Modalities in Vision-and-Language Navigation

Jun 10, 2019
Ronghang Hu, Daniel Fried, Anna Rohrbach, Dan Klein, Trevor Darrell, Kate Saenko

Figure 1 for Are You Looking? Grounding to Multiple Modalities in Vision-and-Language Navigation
Figure 2 for Are You Looking? Grounding to Multiple Modalities in Vision-and-Language Navigation
Figure 3 for Are You Looking? Grounding to Multiple Modalities in Vision-and-Language Navigation
Figure 4 for Are You Looking? Grounding to Multiple Modalities in Vision-and-Language Navigation
Viaarxiv icon

Language-Conditioned Graph Networks for Relational Reasoning

May 10, 2019
Ronghang Hu, Anna Rohrbach, Trevor Darrell, Kate Saenko

Figure 1 for Language-Conditioned Graph Networks for Relational Reasoning
Figure 2 for Language-Conditioned Graph Networks for Relational Reasoning
Figure 3 for Language-Conditioned Graph Networks for Relational Reasoning
Figure 4 for Language-Conditioned Graph Networks for Relational Reasoning
Viaarxiv icon

Viewpoint Invariant Change Captioning

Jan 08, 2019
Dong Huk Park, Trevor Darrell, Anna Rohrbach

Figure 1 for Viewpoint Invariant Change Captioning
Figure 2 for Viewpoint Invariant Change Captioning
Figure 3 for Viewpoint Invariant Change Captioning
Figure 4 for Viewpoint Invariant Change Captioning
Viaarxiv icon

Adversarial Inference for Multi-Sentence Video Description

Dec 13, 2018
Jae Sung Park, Marcus Rohrbach, Trevor Darrell, Anna Rohrbach

Figure 1 for Adversarial Inference for Multi-Sentence Video Description
Figure 2 for Adversarial Inference for Multi-Sentence Video Description
Figure 3 for Adversarial Inference for Multi-Sentence Video Description
Figure 4 for Adversarial Inference for Multi-Sentence Video Description
Viaarxiv icon

Speaker-Follower Models for Vision-and-Language Navigation

Oct 27, 2018
Daniel Fried, Ronghang Hu, Volkan Cirik, Anna Rohrbach, Jacob Andreas, Louis-Philippe Morency, Taylor Berg-Kirkpatrick, Kate Saenko, Dan Klein, Trevor Darrell

Figure 1 for Speaker-Follower Models for Vision-and-Language Navigation
Figure 2 for Speaker-Follower Models for Vision-and-Language Navigation
Figure 3 for Speaker-Follower Models for Vision-and-Language Navigation
Figure 4 for Speaker-Follower Models for Vision-and-Language Navigation
Viaarxiv icon