Get our free extension to see links to code for papers anywhere online!

Chrome logo Add to Chrome

Firefox logo Add to Firefox

Picture for Lijuan Wang

SwinBERT: End-to-End Transformers with Sparse Attention for Video Captioning


Nov 25, 2021
Kevin Lin, Linjie Li, Chung-Ching Lin, Faisal Ahmed, Zhe Gan, Zicheng Liu, Yumao Lu, Lijuan Wang


  Access Paper or Ask Questions

An Empirical Study of Training End-to-End Vision-and-Language Transformers


Nov 25, 2021
Zi-Yi Dou, Yichong Xu, Zhe Gan, Jianfeng Wang, Shuohang Wang, Lijuan Wang, Chenguang Zhu, Pengchuan Zhang, Lu Yuan, Nanyun Peng, Zicheng Liu, Michael Zeng


  Access Paper or Ask Questions

VIOLET : End-to-End Video-Language Transformers with Masked Visual-token Modeling


Nov 24, 2021
Tsu-Jui Fu, Linjie Li, Zhe Gan, Kevin Lin, William Yang Wang, Lijuan Wang, Zicheng Liu

* Code is available at https://github.com/tsujuifu/pytorch_violet 

  Access Paper or Ask Questions

Scaling Up Vision-Language Pre-training for Image Captioning


Nov 24, 2021
Xiaowei Hu, Zhe Gan, Jianfeng Wang, Zhengyuan Yang, Zicheng Liu, Yumao Lu, Lijuan Wang


  Access Paper or Ask Questions

Crossing the Format Boundary of Text and Boxes: Towards Unified Vision-Language Modeling


Nov 23, 2021
Zhengyuan Yang, Zhe Gan, Jianfeng Wang, Xiaowei Hu, Faisal Ahmed, Zicheng Liu, Yumao Lu, Lijuan Wang


  Access Paper or Ask Questions

Florence: A New Foundation Model for Computer Vision


Nov 22, 2021
Lu Yuan, Dongdong Chen, Yi-Ling Chen, Noel Codella, Xiyang Dai, Jianfeng Gao, Houdong Hu, Xuedong Huang, Boxin Li, Chunyuan Li, Ce Liu, Mengchen Liu, Zicheng Liu, Yumao Lu, Yu Shi, Lijuan Wang, Jianfeng Wang, Bin Xiao, Zhen Xiao, Jianwei Yang, Michael Zeng, Luowei Zhou, Pengchuan Zhang


  Access Paper or Ask Questions

UFO: A UniFied TransfOrmer for Vision-Language Representation Learning


Nov 19, 2021
Jianfeng Wang, Xiaowei Hu, Zhe Gan, Zhengyuan Yang, Xiyang Dai, Zicheng Liu, Yumao Lu, Lijuan Wang


  Access Paper or Ask Questions

An Empirical Study of GPT-3 for Few-Shot Knowledge-Based VQA


Sep 10, 2021
Zhengyuan Yang, Zhe Gan, Jianfeng Wang, Xiaowei Hu, Yumao Lu, Zicheng Liu, Lijuan Wang


  Access Paper or Ask Questions

OVIS: Open-Vocabulary Visual Instance Search via Visual-Semantic Aligned Representation Learning


Aug 08, 2021
Sheng Liu, Kevin Lin, Lijuan Wang, Junsong Yuan, Zicheng Liu

* 10 pages 

  Access Paper or Ask Questions

Is Object Detection Necessary for Human-Object Interaction Recognition?


Jul 27, 2021
Ying Jin, Yinpeng Chen, Lijuan Wang, Jianfeng Wang, Pei Yu, Zicheng Liu, Jenq-Neng Hwang


  Access Paper or Ask Questions

End-to-End Semi-Supervised Object Detection with Soft Teacher


Jun 17, 2021
Mengde Xu, Zheng Zhang, Han Hu, Jianfeng Wang, Lijuan Wang, Fangyun Wei, Xiang Bai, Zicheng Liu


  Access Paper or Ask Questions

VALUE: A Multi-Task Benchmark for Video-and-Language Understanding Evaluation


Jun 08, 2021
Linjie Li, Jie Lei, Zhe Gan, Licheng Yu, Yen-Chun Chen, Rohit Pillai, Yu Cheng, Luowei Zhou, Xin Eric Wang, William Yang Wang, Tamara Lee Berg, Mohit Bansal, Jingjing Liu, Lijuan Wang, Zicheng Liu

* VALUE is available at https://value-leaderboard.github.io/ 

  Access Paper or Ask Questions

Compressing Visual-linguistic Model via Knowledge Distillation


Apr 05, 2021
Zhiyuan Fang, Jianfeng Wang, Xiaowei Hu, Lijuan Wang, Yezhou Yang, Zicheng Liu


  Access Paper or Ask Questions

Mesh Graphormer


Apr 01, 2021
Kevin Lin, Lijuan Wang, Zicheng Liu


  Access Paper or Ask Questions

DAP: Detection-Aware Pre-training with Weak Supervision


Mar 30, 2021
Yuanyi Zhong, Jianfeng Wang, Lijuan Wang, Jian Peng, Yu-Xiong Wang, Lei Zhang

* To appear in CVPR 2021 

  Access Paper or Ask Questions

Adversarial Feature Augmentation and Normalization for Visual Recognition


Mar 22, 2021
Tianlong Chen, Yu Cheng, Zhe Gan, Jianfeng Wang, Lijuan Wang, Zhangyang Wang, Jingjing Liu


  Access Paper or Ask Questions

SEED: Self-supervised Distillation For Visual Representation


Jan 12, 2021
Zhiyuan Fang, Jianfeng Wang, Lijuan Wang, Lei Zhang, Yezhou Yang, Zicheng Liu

* Accepted as a conference paper at ICLR 2021 

  Access Paper or Ask Questions

VinVL: Making Visual Representations Matter in Vision-Language Models


Jan 02, 2021
Pengchuan Zhang, Xiujun Li, Xiaowei Hu, Jianwei Yang, Lei Zhang, Lijuan Wang, Yejin Choi, Jianfeng Gao


  Access Paper or Ask Questions

End-to-End Human Pose and Mesh Reconstruction with Transformers


Dec 17, 2020
Kevin Lin, Lijuan Wang, Zicheng Liu


  Access Paper or Ask Questions

MiniVLM: A Smaller and Faster Vision-Language Model


Dec 13, 2020
Jianfeng Wang, Xiaowei Hu, Pengchuan Zhang, Xiujun Li, Lijuan Wang, Lei Zhang, Jianfeng Gao, Zicheng Liu


  Access Paper or Ask Questions

TAP: Text-Aware Pre-training for Text-VQA and Text-Caption


Dec 08, 2020
Zhengyuan Yang, Yijuan Lu, Jianfeng Wang, Xi Yin, Dinei Florencio, Lijuan Wang, Cha Zhang, Lei Zhang, Jiebo Luo


  Access Paper or Ask Questions

VIVO: Surpassing Human Performance in Novel Object Captioning with Visual Vocabulary Pre-Training


Sep 28, 2020
Xiaowei Hu, Xi Yin, Kevin Lin, Lijuan Wang, Lei Zhang, Jianfeng Gao, Zicheng Liu


  Access Paper or Ask Questions

A Study on Effects of Implicit and Explicit Language Model Information for DBLSTM-CTC Based Handwriting Recognition


Jul 31, 2020
Qi Liu, Lijuan Wang, Qiang Huo

* Accepted by ICDAR-2015 

  Access Paper or Ask Questions

M3P: Learning Universal Representations via Multitask Multilingual Multimodal Pre-training


Jun 04, 2020
Haoyang Huang, Lin Su, Di Qi, Nan Duan, Edward Cui, Taroon Bharti, Lei Zhang, Lijuan Wang, Jianfeng Gao, Bei Liu, Jianlong Fu, Dongdong Zhang, Xin Liu, Ming Zhou

* 10 pages,2 figures 

  Access Paper or Ask Questions

Hashing-based Non-Maximum Suppression for Crowded Object Detection


May 22, 2020
Jianfeng Wang, Xi Yin, Lijuan Wang, Lei Zhang


  Access Paper or Ask Questions

Oscar: Object-Semantics Aligned Pre-training for Vision-Language Tasks


May 18, 2020
Xiujun Li, Xi Yin, Chunyuan Li, Pengchuan Zhang, Xiaowei Hu, Lei Zhang, Lijuan Wang, Houdong Hu, Li Dong, Furu Wei, Yejin Choi, Jianfeng Gao

* Code and pre-trained models are released: https://github.com/microsoft/Oscar 

  Access Paper or Ask Questions