Get our free extension to see links to code for papers anywhere online!

Chrome logo  Add to Chrome

Firefox logo Add to Firefox

On the Role of Parallel Data in Cross-lingual Transfer Learning


Dec 20, 2022
Machel Reid, Mikel Artetxe

Add code

* Preprint 

   Access Paper or Ask Questions

  • Share via Twitter
  • Share via Facebook
  • Share via LinkedIn
  • Share via Whatsapp
  • Share via Messenger
  • Share via Email

DiffusER: Discrete Diffusion via Edit-based Reconstruction


Oct 30, 2022
Machel Reid, Vincent J. Hellendoorn, Graham Neubig

Add code

* Preprint. Work in progress 

   Access Paper or Ask Questions

  • Share via Twitter
  • Share via Facebook
  • Share via LinkedIn
  • Share via Whatsapp
  • Share via Messenger
  • Share via Email

M2D2: A Massively Multi-domain Language Modeling Dataset


Oct 13, 2022
Machel Reid, Victor Zhong, Suchin Gururangan, Luke Zettlemoyer

Add code

* EMNLP 2022 

   Access Paper or Ask Questions

  • Share via Twitter
  • Share via Facebook
  • Share via LinkedIn
  • Share via Whatsapp
  • Share via Messenger
  • Share via Email

Learning to Model Editing Processes


May 24, 2022
Machel Reid, Graham Neubig

Add code


   Access Paper or Ask Questions

  • Share via Twitter
  • Share via Facebook
  • Share via LinkedIn
  • Share via Whatsapp
  • Share via Messenger
  • Share via Email

Large Language Models are Zero-Shot Reasoners


May 24, 2022
Takeshi Kojima, Shixiang Shane Gu, Machel Reid, Yutaka Matsuo, Yusuke Iwasawa

Add code


   Access Paper or Ask Questions

  • Share via Twitter
  • Share via Facebook
  • Share via LinkedIn
  • Share via Whatsapp
  • Share via Messenger
  • Share via Email

A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for African News Translation


May 04, 2022
David Ifeoluwa Adelani, Jesujoba Oluwadara Alabi, Angela Fan, Julia Kreutzer, Xiaoyu Shen, Machel Reid, Dana Ruiter, Dietrich Klakow, Peter Nabende, Ernie Chang, Tajuddeen Gwadabe, Freshia Sackey, Bonaventure F. P. Dossou, Chris Chinenye Emezue, Colin Leong, Michael Beukman, Shamsuddeen Hassan Muhammad, Guyo Dub Jarso, Oreen Yousuf, Andre Niyongabo Rubungo, Gilles Hacheme, Eric Peter Wairagala, Muhammad Umair Nasir, Benjamin Ayoade Ajibade, Tunde Oluwaseyi Ajayi, Yvonne Wambui Gitau, Jade Abbott, Mohamed Ahmed, Millicent Ochieng, Anuoluwapo Aremu, Perez Ogayo, Jonathan Mukiibi, Fatoumata Ouoba Kabore, Godson Koffi Kalipe, Derguene Mbaye, Allahsera Auguste Tapo, Victoire Memdjokam Koagne, Edwin Munkoh-Buabeng, Valencia Wagner, Idris Abdulmumin, Ayodele Awokoya, Happy Buzaaba, Blessing Sibanda, Andiswa Bukula, Sam Manthalu

Add code

* Accepted to NAACL 2022 

   Access Paper or Ask Questions

  • Share via Twitter
  • Share via Facebook
  • Share via LinkedIn
  • Share via Whatsapp
  • Share via Messenger
  • Share via Email

Can Wikipedia Help Offline Reinforcement Learning?


Jan 28, 2022
Machel Reid, Yutaro Yamada, Shixiang Shane Gu

Add code


   Access Paper or Ask Questions

  • Share via Twitter
  • Share via Facebook
  • Share via LinkedIn
  • Share via Whatsapp
  • Share via Messenger
  • Share via Email

AfroMT: Pretraining Strategies and Reproducible Benchmarks for Translation of 8 African Languages


Sep 10, 2021
Machel Reid, Junjie Hu, Graham Neubig, Yutaka Matsuo

Add code

* EMNLP 2021 

   Access Paper or Ask Questions

  • Share via Twitter
  • Share via Facebook
  • Share via LinkedIn
  • Share via Whatsapp
  • Share via Messenger
  • Share via Email

PARADISE: Exploiting Parallel Data for Multilingual Sequence-to-Sequence Pretraining


Aug 04, 2021
Machel Reid, Mikel Artetxe

Add code

* Preprint 

   Access Paper or Ask Questions

  • Share via Twitter
  • Share via Facebook
  • Share via LinkedIn
  • Share via Whatsapp
  • Share via Messenger
  • Share via Email

LEWIS: Levenshtein Editing for Unsupervised Text Style Transfer


May 18, 2021
Machel Reid, Victor Zhong

Add code

* ACL-IJCNLP 2021 (Findings) 

   Access Paper or Ask Questions

  • Share via Twitter
  • Share via Facebook
  • Share via LinkedIn
  • Share via Whatsapp
  • Share via Messenger
  • Share via Email
1
2
>>