Alert button
Picture for Zhe Gan

Zhe Gan

Alert button

Violet

An Empirical Study of Training End-to-End Vision-and-Language Transformers

Add code
Bookmark button
Alert button
Nov 03, 2021
Zi-Yi Dou, Yichong Xu, Zhe Gan, Jianfeng Wang, Shuohang Wang, Lijuan Wang, Chenguang Zhu, Nanyun, Peng, Zicheng Liu, Michael Zeng

Figure 1 for An Empirical Study of Training End-to-End Vision-and-Language Transformers
Figure 2 for An Empirical Study of Training End-to-End Vision-and-Language Transformers
Figure 3 for An Empirical Study of Training End-to-End Vision-and-Language Transformers
Figure 4 for An Empirical Study of Training End-to-End Vision-and-Language Transformers
Viaarxiv icon

An Empirical Study of GPT-3 for Few-Shot Knowledge-Based VQA

Add code
Bookmark button
Alert button
Sep 10, 2021
Zhengyuan Yang, Zhe Gan, Jianfeng Wang, Xiaowei Hu, Yumao Lu, Zicheng Liu, Lijuan Wang

Figure 1 for An Empirical Study of GPT-3 for Few-Shot Knowledge-Based VQA
Figure 2 for An Empirical Study of GPT-3 for Few-Shot Knowledge-Based VQA
Figure 3 for An Empirical Study of GPT-3 for Few-Shot Knowledge-Based VQA
Figure 4 for An Empirical Study of GPT-3 for Few-Shot Knowledge-Based VQA
Viaarxiv icon

Simpler, Faster, Stronger: Breaking The log-K Curse On Contrastive Learners With FlatNCE

Add code
Bookmark button
Alert button
Jul 02, 2021
Junya Chen, Zhe Gan, Xuan Li, Qing Guo, Liqun Chen, Shuyang Gao, Tagyoung Chung, Yi Xu, Belinda Zeng, Wenlian Lu, Fan Li, Lawrence Carin, Chenyang Tao

Figure 1 for Simpler, Faster, Stronger: Breaking The log-K Curse On Contrastive Learners With FlatNCE
Figure 2 for Simpler, Faster, Stronger: Breaking The log-K Curse On Contrastive Learners With FlatNCE
Figure 3 for Simpler, Faster, Stronger: Breaking The log-K Curse On Contrastive Learners With FlatNCE
Figure 4 for Simpler, Faster, Stronger: Breaking The log-K Curse On Contrastive Learners With FlatNCE
Viaarxiv icon

Chasing Sparsity in Vision Transformers: An End-to-End Exploration

Add code
Bookmark button
Alert button
Jun 09, 2021
Tianlong Chen, Yu Cheng, Zhe Gan, Lu Yuan, Lei Zhang, Zhangyang Wang

Figure 1 for Chasing Sparsity in Vision Transformers: An End-to-End Exploration
Figure 2 for Chasing Sparsity in Vision Transformers: An End-to-End Exploration
Figure 3 for Chasing Sparsity in Vision Transformers: An End-to-End Exploration
Figure 4 for Chasing Sparsity in Vision Transformers: An End-to-End Exploration
Viaarxiv icon

VALUE: A Multi-Task Benchmark for Video-and-Language Understanding Evaluation

Add code
Bookmark button
Alert button
Jun 08, 2021
Linjie Li, Jie Lei, Zhe Gan, Licheng Yu, Yen-Chun Chen, Rohit Pillai, Yu Cheng, Luowei Zhou, Xin Eric Wang, William Yang Wang, Tamara Lee Berg, Mohit Bansal, Jingjing Liu, Lijuan Wang, Zicheng Liu

Figure 1 for VALUE: A Multi-Task Benchmark for Video-and-Language Understanding Evaluation
Figure 2 for VALUE: A Multi-Task Benchmark for Video-and-Language Understanding Evaluation
Figure 3 for VALUE: A Multi-Task Benchmark for Video-and-Language Understanding Evaluation
Figure 4 for VALUE: A Multi-Task Benchmark for Video-and-Language Understanding Evaluation
Viaarxiv icon

Chasing Sparsity in Vision Transformers:An End-to-End Exploration

Add code
Bookmark button
Alert button
Jun 08, 2021
Tianlong Chen, Yu Cheng, Zhe Gan, Lu Yuan, Lei Zhang, Zhangyang Wang

Figure 1 for Chasing Sparsity in Vision Transformers:An End-to-End Exploration
Figure 2 for Chasing Sparsity in Vision Transformers:An End-to-End Exploration
Figure 3 for Chasing Sparsity in Vision Transformers:An End-to-End Exploration
Figure 4 for Chasing Sparsity in Vision Transformers:An End-to-End Exploration
Viaarxiv icon

Adversarial VQA: A New Benchmark for Evaluating the Robustness of VQA Models

Add code
Bookmark button
Alert button
Jun 01, 2021
Linjie Li, Jie Lei, Zhe Gan, Jingjing Liu

Figure 1 for Adversarial VQA: A New Benchmark for Evaluating the Robustness of VQA Models
Figure 2 for Adversarial VQA: A New Benchmark for Evaluating the Robustness of VQA Models
Figure 3 for Adversarial VQA: A New Benchmark for Evaluating the Robustness of VQA Models
Figure 4 for Adversarial VQA: A New Benchmark for Evaluating the Robustness of VQA Models
Viaarxiv icon

Playing Lottery Tickets with Vision and Language

Add code
Bookmark button
Alert button
Apr 23, 2021
Zhe Gan, Yen-Chun Chen, Linjie Li, Tianlong Chen, Yu Cheng, Shuohang Wang, Jingjing Liu

Figure 1 for Playing Lottery Tickets with Vision and Language
Figure 2 for Playing Lottery Tickets with Vision and Language
Figure 3 for Playing Lottery Tickets with Vision and Language
Figure 4 for Playing Lottery Tickets with Vision and Language
Viaarxiv icon

CUPID: Adaptive Curation of Pre-training Data for Video-and-Language Representation Learning

Add code
Bookmark button
Alert button
Apr 13, 2021
Luowei Zhou, Jingjing Liu, Yu Cheng, Zhe Gan, Lei Zhang

Figure 1 for CUPID: Adaptive Curation of Pre-training Data for Video-and-Language Representation Learning
Figure 2 for CUPID: Adaptive Curation of Pre-training Data for Video-and-Language Representation Learning
Figure 3 for CUPID: Adaptive Curation of Pre-training Data for Video-and-Language Representation Learning
Figure 4 for CUPID: Adaptive Curation of Pre-training Data for Video-and-Language Representation Learning
Viaarxiv icon