Get our free extension to see links to code for papers anywhere online!

 Add to Chrome

 Add to Firefox

CatalyzeX Code Finder - Browser extension linking code for ML papers across the web! | Product Hunt Embed
What is More Likely to Happen Next? Video-and-Language Future Event Prediction

Oct 15, 2020
Jie Lei, Licheng Yu, Tamara L. Berg, Mohit Bansal

* EMNLP 2020 (17 pages) 

  Access Paper or Ask Questions

MART: Memory-Augmented Recurrent Transformer for Coherent Video Paragraph Captioning

May 11, 2020
Jie Lei, Liwei Wang, Yelong Shen, Dong Yu, Tamara L. Berg, Mohit Bansal

* ACL 2020 (12 pages) 

  Access Paper or Ask Questions

TVR: A Large-Scale Dataset for Video-Subtitle Moment Retrieval

Jan 24, 2020
Jie Lei, Licheng Yu, Tamara L. Berg, Mohit Bansal

* 18 pages 

  Access Paper or Ask Questions

IMP: Instance Mask Projection for High Accuracy Semantic Segmentation of Things

Jun 15, 2019
Cheng-Yang Fu, Tamara L. Berg, Alexander C. Berg


  Access Paper or Ask Questions

TVQA+: Spatio-Temporal Grounding for Video Question Answering

Apr 25, 2019
Jie Lei, Licheng Yu, Tamara L. Berg, Mohit Bansal

* 13 pages 

  Access Paper or Ask Questions

Multi-Target Embodied Question Answering

Apr 09, 2019
Licheng Yu, Xinlei Chen, Georgia Gkioxari, Mohit Bansal, Tamara L. Berg, Dhruv Batra

* 10 pages, 6 figures 

  Access Paper or Ask Questions

Dance Dance Generation: Motion Transfer for Internet Videos

Mar 30, 2019
Yipin Zhou, Zhaowen Wang, Chen Fang, Trung Bui, Tamara L. Berg


  Access Paper or Ask Questions

TVQA: Localized, Compositional Video Question Answering

Sep 05, 2018
Jie Lei, Licheng Yu, Mohit Bansal, Tamara L. Berg

* EMNLP 2018 (13 pages; Data and Leaderboard at: http://tvqa.cs.unc.edu

  Access Paper or Ask Questions

Visual to Sound: Generating Natural Sound for Videos in the Wild

Jun 01, 2018
Yipin Zhou, Zhaowen Wang, Chen Fang, Trung Bui, Tamara L. Berg

* Project page: http://bvision11.cs.unc.edu/bigpen/yipin/visual2sound_webpage/visual2sound.html 

  Access Paper or Ask Questions

MAttNet: Modular Attention Network for Referring Expression Comprehension

Mar 27, 2018
Licheng Yu, Zhe Lin, Xiaohui Shen, Jimei Yang, Xin Lu, Mohit Bansal, Tamara L. Berg

* Equation of word attention fixed; MAttNet+Grabcut results added 

  Access Paper or Ask Questions

Combining Multiple Cues for Visual Madlibs Question Answering

Feb 07, 2018
Tatiana Tommasi, Arun Mallya, Bryan Plummer, Svetlana Lazebnik, Alexander C. Berg, Tamara L. Berg

* submitted to IJCV -- under review 

  Access Paper or Ask Questions

Image2GIF: Generating Cinemagraphs using Recurrent Deep Q-Networks

Jan 27, 2018
Yipin Zhou, Yale Song, Tamara L. Berg

* WACV2018 

  Access Paper or Ask Questions

Hierarchically-Attentive RNN for Album Summarization and Storytelling

Aug 09, 2017
Licheng Yu, Mohit Bansal, Tamara L. Berg

* To appear at EMNLP-2017 (7 pages) 

  Access Paper or Ask Questions

A Joint Speaker-Listener-Reinforcer Model for Referring Expressions

Apr 17, 2017
Licheng Yu, Hao Tan, Mohit Bansal, Tamara L. Berg

* Some typo fixed; comprehension results on refcocog updated; more human evaluation results added 

  Access Paper or Ask Questions

Learning Temporal Transformations From Time-Lapse Videos

Aug 27, 2016
Yipin Zhou, Tamara L. Berg

* ECCV2016 

  Access Paper or Ask Questions

When was that made?

Aug 12, 2016
Sirion Vittayakorn, Alexander C. Berg, Tamara L. Berg


  Access Paper or Ask Questions

Solving Visual Madlibs with Multiple Cues

Aug 11, 2016
Tatiana Tommasi, Arun Mallya, Bryan Plummer, Svetlana Lazebnik, Alexander C. Berg, Tamara L. Berg

* accepted at BMVC 2016 

  Access Paper or Ask Questions

Modeling Context in Referring Expressions

Aug 10, 2016
Licheng Yu, Patrick Poirson, Shan Yang, Alexander C. Berg, Tamara L. Berg

* 19 pages, 6 figures, in ECCV 2016; authors, references and acknowledgement updated 

  Access Paper or Ask Questions

Visual Madlibs: Fill in the blank Image Generation and Question Answering

May 31, 2015
Licheng Yu, Eunbyung Park, Alexander C. Berg, Tamara L. Berg

* 10 pages; 8 figures; 4 tables 

  Access Paper or Ask Questions