Alert button
Picture for Wenhui Wang

Wenhui Wang

Alert button

Foundation Transformers

Add code
Bookmark button
Alert button
Oct 12, 2022
Hongyu Wang, Shuming Ma, Shaohan Huang, Li Dong, Wenhui Wang, Zhiliang Peng, Yu Wu, Payal Bajaj, Saksham Singhal, Alon Benhaim, Barun Patra, Zhun Liu, Vishrav Chaudhary, Xia Song, Furu Wei

Figure 1 for Foundation Transformers
Figure 2 for Foundation Transformers
Figure 3 for Foundation Transformers
Figure 4 for Foundation Transformers
Viaarxiv icon

Image as a Foreign Language: BEiT Pretraining for All Vision and Vision-Language Tasks

Add code
Bookmark button
Alert button
Aug 31, 2022
Wenhui Wang, Hangbo Bao, Li Dong, Johan Bjorck, Zhiliang Peng, Qiang Liu, Kriti Aggarwal, Owais Khan Mohammed, Saksham Singhal, Subhojit Som, Furu Wei

Figure 1 for Image as a Foreign Language: BEiT Pretraining for All Vision and Vision-Language Tasks
Figure 2 for Image as a Foreign Language: BEiT Pretraining for All Vision and Vision-Language Tasks
Figure 3 for Image as a Foreign Language: BEiT Pretraining for All Vision and Vision-Language Tasks
Figure 4 for Image as a Foreign Language: BEiT Pretraining for All Vision and Vision-Language Tasks
Viaarxiv icon

Language Models are General-Purpose Interfaces

Add code
Bookmark button
Alert button
Jun 13, 2022
Yaru Hao, Haoyu Song, Li Dong, Shaohan Huang, Zewen Chi, Wenhui Wang, Shuming Ma, Furu Wei

Figure 1 for Language Models are General-Purpose Interfaces
Figure 2 for Language Models are General-Purpose Interfaces
Figure 3 for Language Models are General-Purpose Interfaces
Figure 4 for Language Models are General-Purpose Interfaces
Viaarxiv icon

VL-BEiT: Generative Vision-Language Pretraining

Add code
Bookmark button
Alert button
Jun 02, 2022
Hangbo Bao, Wenhui Wang, Li Dong, Furu Wei

Figure 1 for VL-BEiT: Generative Vision-Language Pretraining
Figure 2 for VL-BEiT: Generative Vision-Language Pretraining
Figure 3 for VL-BEiT: Generative Vision-Language Pretraining
Figure 4 for VL-BEiT: Generative Vision-Language Pretraining
Viaarxiv icon

TimeReplayer: Unlocking the Potential of Event Cameras for Video Interpolation

Add code
Bookmark button
Alert button
Mar 25, 2022
Weihua He, Kaichao You, Zhendong Qiao, Xu Jia, Ziyang Zhang, Wenhui Wang, Huchuan Lu, Yaoyuan Wang, Jianxing Liao

Figure 1 for TimeReplayer: Unlocking the Potential of Event Cameras for Video Interpolation
Figure 2 for TimeReplayer: Unlocking the Potential of Event Cameras for Video Interpolation
Figure 3 for TimeReplayer: Unlocking the Potential of Event Cameras for Video Interpolation
Figure 4 for TimeReplayer: Unlocking the Potential of Event Cameras for Video Interpolation
Viaarxiv icon

AutoDistil: Few-shot Task-agnostic Neural Architecture Search for Distilling Large Language Models

Add code
Bookmark button
Alert button
Jan 29, 2022
Dongkuan Xu, Subhabrata Mukherjee, Xiaodong Liu, Debadeepta Dey, Wenhui Wang, Xiang Zhang, Ahmed Hassan Awadallah, Jianfeng Gao

Figure 1 for AutoDistil: Few-shot Task-agnostic Neural Architecture Search for Distilling Large Language Models
Figure 2 for AutoDistil: Few-shot Task-agnostic Neural Architecture Search for Distilling Large Language Models
Figure 3 for AutoDistil: Few-shot Task-agnostic Neural Architecture Search for Distilling Large Language Models
Figure 4 for AutoDistil: Few-shot Task-agnostic Neural Architecture Search for Distilling Large Language Models
Viaarxiv icon

Distilled Dual-Encoder Model for Vision-Language Understanding

Add code
Bookmark button
Alert button
Dec 16, 2021
Zekun Wang, Wenhui Wang, Haichao Zhu, Ming Liu, Bing Qin, Furu Wei

Figure 1 for Distilled Dual-Encoder Model for Vision-Language Understanding
Figure 2 for Distilled Dual-Encoder Model for Vision-Language Understanding
Figure 3 for Distilled Dual-Encoder Model for Vision-Language Understanding
Figure 4 for Distilled Dual-Encoder Model for Vision-Language Understanding
Viaarxiv icon

VLMo: Unified Vision-Language Pre-Training with Mixture-of-Modality-Experts

Add code
Bookmark button
Alert button
Nov 03, 2021
Wenhui Wang, Hangbo Bao, Li Dong, Furu Wei

Figure 1 for VLMo: Unified Vision-Language Pre-Training with Mixture-of-Modality-Experts
Figure 2 for VLMo: Unified Vision-Language Pre-Training with Mixture-of-Modality-Experts
Figure 3 for VLMo: Unified Vision-Language Pre-Training with Mixture-of-Modality-Experts
Figure 4 for VLMo: Unified Vision-Language Pre-Training with Mixture-of-Modality-Experts
Viaarxiv icon

s2s-ft: Fine-Tuning Pretrained Transformer Encoders for Sequence-to-Sequence Learning

Add code
Bookmark button
Alert button
Oct 26, 2021
Hangbo Bao, Li Dong, Wenhui Wang, Nan Yang, Furu Wei

Figure 1 for s2s-ft: Fine-Tuning Pretrained Transformer Encoders for Sequence-to-Sequence Learning
Figure 2 for s2s-ft: Fine-Tuning Pretrained Transformer Encoders for Sequence-to-Sequence Learning
Figure 3 for s2s-ft: Fine-Tuning Pretrained Transformer Encoders for Sequence-to-Sequence Learning
Figure 4 for s2s-ft: Fine-Tuning Pretrained Transformer Encoders for Sequence-to-Sequence Learning
Viaarxiv icon