no code implementations • RepL4NLP (ACL) 2022 • Machel Reid, Mikel Artetxe
Despite the success of multilingual sequence-to-sequence pretraining, most existing approaches rely on monolingual corpora and do not make use of the strong cross-lingual signal contained in parallel data.
no code implementations • insights (ACL) 2022 • Itsuki Okimura, Machel Reid, Makoto Kawano, Yutaka Matsuo
The reason for this is that within NLP, the impact of proposed data augmentation methods on performance has not been evaluated in a unified manner, and effective data augmentation methods are unclear.
no code implementations • NAACL (AmericasNLP) 2021 • Francis Zheng, Machel Reid, Edison Marrese-Taylor, Yutaka Matsuo
This paper describes UTokyo’s submission to the AmericasNLP 2021 Shared Task on machine translation systems for indigenous languages of the Americas.
1 code implementation • 24 May 2022 • Machel Reid, Graham Neubig
We introduce baseline results and metrics on this task, finding that modeling editing processes improves performance on a variety of axes on both our proposed task and related downstream tasks compared to previous single-step models of edits.
1 code implementation • 24 May 2022 • Takeshi Kojima, Shixiang Shane Gu, Machel Reid, Yutaka Matsuo, Yusuke Iwasawa
Pretrained large language models (LLMs) are widely used in many sub-fields of natural language processing (NLP) and generally known as excellent few-shot learners with task-specific exemplars.
Ranked #1 on
Arithmetic Reasoning
on MultiArith
1 code implementation • 4 May 2022 • David Ifeoluwa Adelani, Jesujoba Oluwadara Alabi, Angela Fan, Julia Kreutzer, Xiaoyu Shen, Machel Reid, Dana Ruiter, Dietrich Klakow, Peter Nabende, Ernie Chang, Tajuddeen Gwadabe, Freshia Sackey, Bonaventure F. P. Dossou, Chris Chinenye Emezue, Colin Leong, Michael Beukman, Shamsuddeen Hassan Muhammad, Guyo Dub Jarso, Oreen Yousuf, Andre Niyongabo Rubungo, Gilles Hacheme, Eric Peter Wairagala, Muhammad Umair Nasir, Benjamin Ayoade Ajibade, Tunde Oluwaseyi Ajayi, Yvonne Wambui Gitau, Jade Abbott, Mohamed Ahmed, Millicent Ochieng, Anuoluwapo Aremu, Perez Ogayo, Jonathan Mukiibi, Fatoumata Ouoba Kabore, Godson Koffi Kalipe, Derguene Mbaye, Allahsera Auguste Tapo, Victoire Memdjokam Koagne, Edwin Munkoh-Buabeng, Valencia Wagner, Idris Abdulmumin, Ayodele Awokoya, Happy Buzaaba, Blessing Sibanda, Andiswa Bukula, Sam Manthalu
We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pre-training?
1 code implementation • 28 Jan 2022 • Machel Reid, Yutaro Yamada, Shixiang Shane Gu
In this paper, we look to take advantage of this formulation of reinforcement learning as sequence modeling and investigate the transferability of pre-trained sequence models on other domains (vision, language) when finetuned on offline RL tasks (control, games).
1 code implementation • EMNLP 2021 • Machel Reid, Junjie Hu, Graham Neubig, Yutaka Matsuo
Reproducible benchmarks are crucial in driving progress of machine translation research.
1 code implementation • 4 Aug 2021 • Machel Reid, Mikel Artetxe
Despite the success of multilingual sequence-to-sequence pretraining, most existing approaches rely on monolingual corpora, and do not make use of the strong cross-lingual signal contained in parallel data.
1 code implementation • Findings (ACL) 2021 • Machel Reid, Victor Zhong
Moreover, compared to previous methods on unsupervised data synthesis, our method results in higher quality parallel style pairs and improves model performance.
1 code implementation • Findings (EMNLP) 2021 • Machel Reid, Edison Marrese-Taylor, Yutaka Matsuo
In light of this, we explore parameter-sharing methods in Transformers with a specific focus on generative models.
no code implementations • 1 Jan 2021 • Machel Reid, Edison Marrese-Taylor, Yutaka Matsuo
We also perform equally well as Transformer-big with 40% less parameters and outperform the model by 0. 7 BLEU with 12M less parameters.
Ranked #20 on
Machine Translation
on WMT2014 English-German
1 code implementation • EMNLP 2020 • Machel Reid, Edison Marrese-Taylor, Yutaka Matsuo
In this paper, we tackle the task of definition modeling, where the goal is to learn to generate definitions of words and phrases.
1 code implementation • 20 Apr 2020 • Edison Marrese-Taylor, Machel Reid, Yutaka Matsuo
Document editing has become a pervasive component of the production of information, with version control systems enabling edits to be efficiently stored and applied.
no code implementations • 9 Mar 2020 • Machel Reid, Edison Marrese-Taylor, Yutaka Matsuo
The contrast between the need for large amounts of data for current Natural Language Processing (NLP) techniques, and the lack thereof, is accentuated in the case of African languages, most of which are considered low-resource.