Alert button
Picture for Li Dong

Li Dong

Alert button

Adapt-and-Distill: Developing Small, Fast and Effective Pretrained Language Models for Domains

Add code
Bookmark button
Alert button
Jun 29, 2021
Yunzhi Yao, Shaohan Huang, Wenhui Wang, Li Dong, Furu Wei

Figure 1 for Adapt-and-Distill: Developing Small, Fast and Effective Pretrained Language Models for Domains
Figure 2 for Adapt-and-Distill: Developing Small, Fast and Effective Pretrained Language Models for Domains
Figure 3 for Adapt-and-Distill: Developing Small, Fast and Effective Pretrained Language Models for Domains
Figure 4 for Adapt-and-Distill: Developing Small, Fast and Effective Pretrained Language Models for Domains
Viaarxiv icon

DeltaLM: Encoder-Decoder Pre-training for Language Generation and Translation by Augmenting Pretrained Multilingual Encoders

Add code
Bookmark button
Alert button
Jun 25, 2021
Shuming Ma, Li Dong, Shaohan Huang, Dongdong Zhang, Alexandre Muzio, Saksham Singhal, Hany Hassan Awadalla, Xia Song, Furu Wei

Figure 1 for DeltaLM: Encoder-Decoder Pre-training for Language Generation and Translation by Augmenting Pretrained Multilingual Encoders
Figure 2 for DeltaLM: Encoder-Decoder Pre-training for Language Generation and Translation by Augmenting Pretrained Multilingual Encoders
Figure 3 for DeltaLM: Encoder-Decoder Pre-training for Language Generation and Translation by Augmenting Pretrained Multilingual Encoders
Figure 4 for DeltaLM: Encoder-Decoder Pre-training for Language Generation and Translation by Augmenting Pretrained Multilingual Encoders
Viaarxiv icon

Learning to Sample Replacements for ELECTRA Pre-Training

Add code
Bookmark button
Alert button
Jun 25, 2021
Yaru Hao, Li Dong, Hangbo Bao, Ke Xu, Furu Wei

Figure 1 for Learning to Sample Replacements for ELECTRA Pre-Training
Figure 2 for Learning to Sample Replacements for ELECTRA Pre-Training
Figure 3 for Learning to Sample Replacements for ELECTRA Pre-Training
Figure 4 for Learning to Sample Replacements for ELECTRA Pre-Training
Viaarxiv icon

BEiT: BERT Pre-Training of Image Transformers

Add code
Bookmark button
Alert button
Jun 15, 2021
Hangbo Bao, Li Dong, Furu Wei

Figure 1 for BEiT: BERT Pre-Training of Image Transformers
Figure 2 for BEiT: BERT Pre-Training of Image Transformers
Figure 3 for BEiT: BERT Pre-Training of Image Transformers
Figure 4 for BEiT: BERT Pre-Training of Image Transformers
Viaarxiv icon

Consistency Regularization for Cross-Lingual Fine-Tuning

Add code
Bookmark button
Alert button
Jun 15, 2021
Bo Zheng, Li Dong, Shaohan Huang, Wenhui Wang, Zewen Chi, Saksham Singhal, Wanxiang Che, Ting Liu, Xia Song, Furu Wei

Figure 1 for Consistency Regularization for Cross-Lingual Fine-Tuning
Figure 2 for Consistency Regularization for Cross-Lingual Fine-Tuning
Figure 3 for Consistency Regularization for Cross-Lingual Fine-Tuning
Figure 4 for Consistency Regularization for Cross-Lingual Fine-Tuning
Viaarxiv icon

Improving Pretrained Cross-Lingual Language Models via Self-Labeled Word Alignment

Add code
Bookmark button
Alert button
Jun 11, 2021
Zewen Chi, Li Dong, Bo Zheng, Shaohan Huang, Xian-Ling Mao, Heyan Huang, Furu Wei

Figure 1 for Improving Pretrained Cross-Lingual Language Models via Self-Labeled Word Alignment
Figure 2 for Improving Pretrained Cross-Lingual Language Models via Self-Labeled Word Alignment
Figure 3 for Improving Pretrained Cross-Lingual Language Models via Self-Labeled Word Alignment
Figure 4 for Improving Pretrained Cross-Lingual Language Models via Self-Labeled Word Alignment
Viaarxiv icon

A Semi-supervised Multi-task Learning Approach to Classify Customer Contact Intents

Add code
Bookmark button
Alert button
Jun 10, 2021
Li Dong, Matthew C. Spencer, Amir Biagi

Figure 1 for A Semi-supervised Multi-task Learning Approach to Classify Customer Contact Intents
Figure 2 for A Semi-supervised Multi-task Learning Approach to Classify Customer Contact Intents
Figure 3 for A Semi-supervised Multi-task Learning Approach to Classify Customer Contact Intents
Figure 4 for A Semi-supervised Multi-task Learning Approach to Classify Customer Contact Intents
Viaarxiv icon

Memory-Efficient Differentiable Transformer Architecture Search

Add code
Bookmark button
Alert button
May 31, 2021
Yuekai Zhao, Li Dong, Yelong Shen, Zhihua Zhang, Furu Wei, Weizhu Chen

Figure 1 for Memory-Efficient Differentiable Transformer Architecture Search
Figure 2 for Memory-Efficient Differentiable Transformer Architecture Search
Figure 3 for Memory-Efficient Differentiable Transformer Architecture Search
Figure 4 for Memory-Efficient Differentiable Transformer Architecture Search
Viaarxiv icon

Zero-shot Cross-lingual Transfer of Neural Machine Translation with Multilingual Pretrained Encoders

Add code
Bookmark button
Alert button
Apr 18, 2021
Guanhua Chen, Shuming Ma, Yun Chen, Li Dong, Dongdong Zhang, Jia Pan, Wenping Wang, Furu Wei

Figure 1 for Zero-shot Cross-lingual Transfer of Neural Machine Translation with Multilingual Pretrained Encoders
Figure 2 for Zero-shot Cross-lingual Transfer of Neural Machine Translation with Multilingual Pretrained Encoders
Figure 3 for Zero-shot Cross-lingual Transfer of Neural Machine Translation with Multilingual Pretrained Encoders
Figure 4 for Zero-shot Cross-lingual Transfer of Neural Machine Translation with Multilingual Pretrained Encoders
Viaarxiv icon