Picture for Edward Cui

Edward Cui

GEM: A General Evaluation Benchmark for Multimodal Tasks

Add code
Jun 18, 2021
Figure 1 for GEM: A General Evaluation Benchmark for Multimodal Tasks
Figure 2 for GEM: A General Evaluation Benchmark for Multimodal Tasks
Figure 3 for GEM: A General Evaluation Benchmark for Multimodal Tasks
Figure 4 for GEM: A General Evaluation Benchmark for Multimodal Tasks
Viaarxiv icon

M3P: Learning Universal Representations via Multitask Multilingual Multimodal Pre-training

Add code
Jun 04, 2020
Figure 1 for M3P: Learning Universal Representations via Multitask Multilingual Multimodal Pre-training
Figure 2 for M3P: Learning Universal Representations via Multitask Multilingual Multimodal Pre-training
Figure 3 for M3P: Learning Universal Representations via Multitask Multilingual Multimodal Pre-training
Figure 4 for M3P: Learning Universal Representations via Multitask Multilingual Multimodal Pre-training
Viaarxiv icon

XGLUE: A New Benchmark Dataset for Cross-lingual Pre-training, Understanding and Generation

Add code
Apr 19, 2020
Figure 1 for XGLUE: A New Benchmark Dataset for Cross-lingual Pre-training, Understanding and Generation
Figure 2 for XGLUE: A New Benchmark Dataset for Cross-lingual Pre-training, Understanding and Generation
Figure 3 for XGLUE: A New Benchmark Dataset for Cross-lingual Pre-training, Understanding and Generation
Figure 4 for XGLUE: A New Benchmark Dataset for Cross-lingual Pre-training, Understanding and Generation
Viaarxiv icon

XGPT: Cross-modal Generative Pre-Training for Image Captioning

Add code
Mar 04, 2020
Figure 1 for XGPT: Cross-modal Generative Pre-Training for Image Captioning
Figure 2 for XGPT: Cross-modal Generative Pre-Training for Image Captioning
Figure 3 for XGPT: Cross-modal Generative Pre-Training for Image Captioning
Figure 4 for XGPT: Cross-modal Generative Pre-Training for Image Captioning
Viaarxiv icon

ImageBERT: Cross-modal Pre-training with Large-scale Weak-supervised Image-Text Data

Add code
Jan 23, 2020
Figure 1 for ImageBERT: Cross-modal Pre-training with Large-scale Weak-supervised Image-Text Data
Figure 2 for ImageBERT: Cross-modal Pre-training with Large-scale Weak-supervised Image-Text Data
Figure 3 for ImageBERT: Cross-modal Pre-training with Large-scale Weak-supervised Image-Text Data
Figure 4 for ImageBERT: Cross-modal Pre-training with Large-scale Weak-supervised Image-Text Data
Viaarxiv icon