Alert button
Picture for Melvin Johnson

Melvin Johnson

Alert button

Multilingual Document-Level Translation Enables Zero-Shot Transfer From Sentences to Documents

Add code
Bookmark button
Alert button
Sep 21, 2021
Biao Zhang, Ankur Bapna, Melvin Johnson, Ali Dabirmoghaddam, Naveen Arivazhagan, Orhan Firat

Figure 1 for Multilingual Document-Level Translation Enables Zero-Shot Transfer From Sentences to Documents
Figure 2 for Multilingual Document-Level Translation Enables Zero-Shot Transfer From Sentences to Documents
Figure 3 for Multilingual Document-Level Translation Enables Zero-Shot Transfer From Sentences to Documents
Figure 4 for Multilingual Document-Level Translation Enables Zero-Shot Transfer From Sentences to Documents
Viaarxiv icon

HintedBT: Augmenting Back-Translation with Quality and Transliteration Hints

Add code
Bookmark button
Alert button
Sep 09, 2021
Sahana Ramnath, Melvin Johnson, Abhirut Gupta, Aravindan Raghuveer

Figure 1 for HintedBT: Augmenting Back-Translation with Quality and Transliteration Hints
Figure 2 for HintedBT: Augmenting Back-Translation with Quality and Transliteration Hints
Figure 3 for HintedBT: Augmenting Back-Translation with Quality and Transliteration Hints
Figure 4 for HintedBT: Augmenting Back-Translation with Quality and Transliteration Hints
Viaarxiv icon

MergeDistill: Merging Pre-trained Language Models using Distillation

Add code
Bookmark button
Alert button
Jun 05, 2021
Simran Khanuja, Melvin Johnson, Partha Talukdar

Figure 1 for MergeDistill: Merging Pre-trained Language Models using Distillation
Figure 2 for MergeDistill: Merging Pre-trained Language Models using Distillation
Figure 3 for MergeDistill: Merging Pre-trained Language Models using Distillation
Figure 4 for MergeDistill: Merging Pre-trained Language Models using Distillation
Viaarxiv icon

nmT5 -- Is parallel data still relevant for pre-training massively multilingual language models?

Add code
Bookmark button
Alert button
Jun 03, 2021
Mihir Kale, Aditya Siddhant, Noah Constant, Melvin Johnson, Rami Al-Rfou, Linting Xue

Figure 1 for nmT5 -- Is parallel data still relevant for pre-training massively multilingual language models?
Figure 2 for nmT5 -- Is parallel data still relevant for pre-training massively multilingual language models?
Figure 3 for nmT5 -- Is parallel data still relevant for pre-training massively multilingual language models?
Figure 4 for nmT5 -- Is parallel data still relevant for pre-training massively multilingual language models?
Viaarxiv icon

XTREME-R: Towards More Challenging and Nuanced Multilingual Evaluation

Add code
Bookmark button
Alert button
Apr 15, 2021
Sebastian Ruder, Noah Constant, Jan Botha, Aditya Siddhant, Orhan Firat, Jinlan Fu, Pengfei Liu, Junjie Hu, Graham Neubig, Melvin Johnson

Figure 1 for XTREME-R: Towards More Challenging and Nuanced Multilingual Evaluation
Figure 2 for XTREME-R: Towards More Challenging and Nuanced Multilingual Evaluation
Figure 3 for XTREME-R: Towards More Challenging and Nuanced Multilingual Evaluation
Figure 4 for XTREME-R: Towards More Challenging and Nuanced Multilingual Evaluation
Viaarxiv icon

Gradient-guided Loss Masking for Neural Machine Translation

Add code
Bookmark button
Alert button
Feb 26, 2021
Xinyi Wang, Ankur Bapna, Melvin Johnson, Orhan Firat

Figure 1 for Gradient-guided Loss Masking for Neural Machine Translation
Figure 2 for Gradient-guided Loss Masking for Neural Machine Translation
Figure 3 for Gradient-guided Loss Masking for Neural Machine Translation
Figure 4 for Gradient-guided Loss Masking for Neural Machine Translation
Viaarxiv icon

They, Them, Theirs: Rewriting with Gender-Neutral English

Add code
Bookmark button
Alert button
Feb 12, 2021
Tony Sun, Kellie Webster, Apu Shah, William Yang Wang, Melvin Johnson

Figure 1 for They, Them, Theirs: Rewriting with Gender-Neutral English
Figure 2 for They, Them, Theirs: Rewriting with Gender-Neutral English
Figure 3 for They, Them, Theirs: Rewriting with Gender-Neutral English
Figure 4 for They, Them, Theirs: Rewriting with Gender-Neutral English
Viaarxiv icon

Distilling Large Language Models into Tiny and Effective Students using pQRNN

Add code
Bookmark button
Alert button
Jan 21, 2021
Prabhu Kaliamoorthi, Aditya Siddhant, Edward Li, Melvin Johnson

Figure 1 for Distilling Large Language Models into Tiny and Effective Students using pQRNN
Figure 2 for Distilling Large Language Models into Tiny and Effective Students using pQRNN
Figure 3 for Distilling Large Language Models into Tiny and Effective Students using pQRNN
Figure 4 for Distilling Large Language Models into Tiny and Effective Students using pQRNN
Viaarxiv icon

Rethinking embedding coupling in pre-trained language models

Add code
Bookmark button
Alert button
Oct 24, 2020
Hyung Won Chung, Thibault Févry, Henry Tsai, Melvin Johnson, Sebastian Ruder

Figure 1 for Rethinking embedding coupling in pre-trained language models
Figure 2 for Rethinking embedding coupling in pre-trained language models
Figure 3 for Rethinking embedding coupling in pre-trained language models
Figure 4 for Rethinking embedding coupling in pre-trained language models
Viaarxiv icon

Explicit Alignment Objectives for Multilingual Bidirectional Encoders

Add code
Bookmark button
Alert button
Oct 15, 2020
Junjie Hu, Melvin Johnson, Orhan Firat, Aditya Siddhant, Graham Neubig

Figure 1 for Explicit Alignment Objectives for Multilingual Bidirectional Encoders
Figure 2 for Explicit Alignment Objectives for Multilingual Bidirectional Encoders
Figure 3 for Explicit Alignment Objectives for Multilingual Bidirectional Encoders
Figure 4 for Explicit Alignment Objectives for Multilingual Bidirectional Encoders
Viaarxiv icon