Picture for Melvin Johnson

Melvin Johnson

SLAM: A Unified Encoder for Speech and Language Modeling via Speech-Text Joint Pre-Training

Add code
Oct 20, 2021
Figure 1 for SLAM: A Unified Encoder for Speech and Language Modeling via Speech-Text Joint Pre-Training
Figure 2 for SLAM: A Unified Encoder for Speech and Language Modeling via Speech-Text Joint Pre-Training
Figure 3 for SLAM: A Unified Encoder for Speech and Language Modeling via Speech-Text Joint Pre-Training
Figure 4 for SLAM: A Unified Encoder for Speech and Language Modeling via Speech-Text Joint Pre-Training
Viaarxiv icon

Multilingual Document-Level Translation Enables Zero-Shot Transfer From Sentences to Documents

Add code
Sep 21, 2021
Figure 1 for Multilingual Document-Level Translation Enables Zero-Shot Transfer From Sentences to Documents
Figure 2 for Multilingual Document-Level Translation Enables Zero-Shot Transfer From Sentences to Documents
Figure 3 for Multilingual Document-Level Translation Enables Zero-Shot Transfer From Sentences to Documents
Figure 4 for Multilingual Document-Level Translation Enables Zero-Shot Transfer From Sentences to Documents
Viaarxiv icon

HintedBT: Augmenting Back-Translation with Quality and Transliteration Hints

Add code
Sep 09, 2021
Figure 1 for HintedBT: Augmenting Back-Translation with Quality and Transliteration Hints
Figure 2 for HintedBT: Augmenting Back-Translation with Quality and Transliteration Hints
Figure 3 for HintedBT: Augmenting Back-Translation with Quality and Transliteration Hints
Figure 4 for HintedBT: Augmenting Back-Translation with Quality and Transliteration Hints
Viaarxiv icon

MergeDistill: Merging Pre-trained Language Models using Distillation

Add code
Jun 05, 2021
Figure 1 for MergeDistill: Merging Pre-trained Language Models using Distillation
Figure 2 for MergeDistill: Merging Pre-trained Language Models using Distillation
Figure 3 for MergeDistill: Merging Pre-trained Language Models using Distillation
Figure 4 for MergeDistill: Merging Pre-trained Language Models using Distillation
Viaarxiv icon

nmT5 -- Is parallel data still relevant for pre-training massively multilingual language models?

Add code
Jun 03, 2021
Figure 1 for nmT5 -- Is parallel data still relevant for pre-training massively multilingual language models?
Figure 2 for nmT5 -- Is parallel data still relevant for pre-training massively multilingual language models?
Figure 3 for nmT5 -- Is parallel data still relevant for pre-training massively multilingual language models?
Figure 4 for nmT5 -- Is parallel data still relevant for pre-training massively multilingual language models?
Viaarxiv icon

XTREME-R: Towards More Challenging and Nuanced Multilingual Evaluation

Add code
Apr 15, 2021
Figure 1 for XTREME-R: Towards More Challenging and Nuanced Multilingual Evaluation
Figure 2 for XTREME-R: Towards More Challenging and Nuanced Multilingual Evaluation
Figure 3 for XTREME-R: Towards More Challenging and Nuanced Multilingual Evaluation
Figure 4 for XTREME-R: Towards More Challenging and Nuanced Multilingual Evaluation
Viaarxiv icon

Gradient-guided Loss Masking for Neural Machine Translation

Add code
Feb 26, 2021
Figure 1 for Gradient-guided Loss Masking for Neural Machine Translation
Figure 2 for Gradient-guided Loss Masking for Neural Machine Translation
Figure 3 for Gradient-guided Loss Masking for Neural Machine Translation
Figure 4 for Gradient-guided Loss Masking for Neural Machine Translation
Viaarxiv icon

They, Them, Theirs: Rewriting with Gender-Neutral English

Add code
Feb 12, 2021
Figure 1 for They, Them, Theirs: Rewriting with Gender-Neutral English
Figure 2 for They, Them, Theirs: Rewriting with Gender-Neutral English
Figure 3 for They, Them, Theirs: Rewriting with Gender-Neutral English
Figure 4 for They, Them, Theirs: Rewriting with Gender-Neutral English
Viaarxiv icon

Distilling Large Language Models into Tiny and Effective Students using pQRNN

Add code
Jan 21, 2021
Figure 1 for Distilling Large Language Models into Tiny and Effective Students using pQRNN
Figure 2 for Distilling Large Language Models into Tiny and Effective Students using pQRNN
Figure 3 for Distilling Large Language Models into Tiny and Effective Students using pQRNN
Figure 4 for Distilling Large Language Models into Tiny and Effective Students using pQRNN
Viaarxiv icon

Rethinking embedding coupling in pre-trained language models

Add code
Oct 24, 2020
Figure 1 for Rethinking embedding coupling in pre-trained language models
Figure 2 for Rethinking embedding coupling in pre-trained language models
Figure 3 for Rethinking embedding coupling in pre-trained language models
Figure 4 for Rethinking embedding coupling in pre-trained language models
Viaarxiv icon