1 code implementation • WMT (EMNLP) 2021 • Pinzhen Chen, Jindřich Helcl, Ulrich Germann, Laurie Burchell, Nikolay Bogoychev, Antonio Valerio Miceli Barone, Jonas Waldendorf, Alexandra Birch, Kenneth Heafield
This paper presents the University of Edinburgh’s constrained submissions of English-German and English-Hausa systems to the WMT 2021 shared task on news translation.
no code implementations • EAMT 2022 • Peggy van der Kreeft, Alexandra Birch, Sevi Sariisik, Felipe Sánchez-Martínez, Wilker Aziz
The GoURMET project, funded by the European Commission’s H2020 program (under grant agreement 825299), develops models for machine translation, in particular for low-resourced languages.
no code implementations • BioNLP (ACL) 2022 • Matúš Falis, Hang Dong, Alexandra Birch, Beatrice Alex
We propose data augmentation and synthesis techniques in order to address these scenarios.
no code implementations • MTSummit 2021 • Alexandra Birch, Barry Haddow, Antonio Valerio Miceli Barone, Jindrich Helcl, Jonas Waldendorf, Felipe Sánchez Martínez, Mikel Forcada, Víctor Sánchez Cartagena, Juan Antonio Pérez-Ortiz, Miquel Esplà-Gomis, Wilker Aziz, Lina Murady, Sevi Sariisik, Peggy van der Kreeft, Kay Macquarrie
We find that starting from an existing large model pre-trained on 50languages leads to far better BLEU scores than pretraining on one high-resource language pair with a smaller model.
no code implementations • WMT (EMNLP) 2020 • Rachel Bawden, Alexandra Birch, Radina Dobreva, Arturo Oncevay, Antonio Valerio Miceli Barone, Philip Williams
We describe the University of Edinburgh’s submissions to the WMT20 news translation shared task for the low resource language pair English-Tamil and the mid-resource language pair English-Inuktitut.
no code implementations • IWSLT 2017 • Pawel Przybysz, Marcin Chochowski, Rico Sennrich, Barry Haddow, Alexandra Birch
This paper describes the joint submission of Samsung Research and Development, Warsaw, Poland and the University of Edinburgh team to the IWSLT MT task for TED talks.
no code implementations • IWSLT 2016 • Marcin Junczys-Dowmunt, Alexandra Birch
This paper describes the submission of the University of Edinburgh team to the IWSLT MT task for TED talks.
no code implementations • IWSLT 2016 • Maria Nădejde, Alexandra Birch, Philipp Koehn
String-to-tree MT systems translate verbs without lexical or syntactic context on the source side and with limited target-side context.
no code implementations • IWSLT (EMNLP) 2018 • Philip Williams, Marcin Chochowski, Pawel Przybysz, Rico Sennrich, Barry Haddow, Alexandra Birch
This paper describes the joint submission to the IWSLT 2018 Low Resource MT task by Samsung R&D Institute, Poland, and the University of Edinburgh.
no code implementations • EMNLP (IWSLT) 2019 • Joanna Wetesko, Marcin Chochowski, Pawel Przybysz, Philip Williams, Roman Grundkiewicz, Rico Sennrich, Barry Haddow, None Barone, Valerio Miceli, Alexandra Birch
This paper describes the joint submission to the IWSLT 2019 English to Czech task by Samsung RD Institute, Poland, and the University of Edinburgh.
1 code implementation • 1 Jun 2022 • Laurie Burchell, Alexandra Birch, Kenneth Heafield
We also find evidence that lexical diversity is more important than syntactic for back translation performance.
no code implementations • 6 May 2022 • Arturo Oncevay, Duygu Ataman, Niels van Berkel, Barry Haddow, Alexandra Birch, Johannes Bjerva
In this work, we propose to reduce the rigidity of such claims, by quantifying morphological typology at the word and segment level.
no code implementations • 4 May 2022 • Jindřich Helcl, Barry Haddow, Alexandra Birch
In this paper, we point out flaws in the evaluation methodology present in the literature on NAR models and we provide a fair comparison between a state-of-the-art NAR model and the autoregressive submissions to the shared task.
no code implementations • RepL4NLP (ACL) 2022 • Antonio Valerio Miceli-Barone, Alexandra Birch, Rico Sennrich
Neural machine learning models can successfully model language that is similar to their training distribution, but they are highly susceptible to degradation under distribution shift, which occurs in many practical applications when processing out-of-domain (OOD) text.
1 code implementation • EMNLP 2021 • Nikita Moghe, Mark Steedman, Alexandra Birch
In this work, we enhance the transfer learning process by intermediate fine-tuning of pretrained multilingual models, where the multilingual models are fine-tuned with different but related data and/or tasks.
1 code implementation • EMNLP 2021 • Matúš Falis, Hang Dong, Alexandra Birch, Beatrice Alex
We propose a set of metrics for hierarchical evaluation using the depth-based representation.
Multi Label Text Classification
Multi-Label Text Classification
no code implementations • 1 Sep 2021 • Barry Haddow, Rachel Bawden, Antonio Valerio Miceli Barone, Jindřich Helcl, Alexandra Birch
We present a survey covering the state of the art in low-resource machine translation research.
no code implementations • Findings (ACL) 2021 • Christos Baziotis, Ivan Titov, Alexandra Birch, Barry Haddow
Unsupervised cross-lingual pretraining has achieved strong results in neural machine translation (NMT), by drastically reducing the need for large parallel data.
1 code implementation • EACL 2021 • Farid Arthaud, Rachel Bawden, Alexandra Birch
Machine translation (MT) models used in industries with constantly changing topics, such as translation or news agencies, need to adapt to new data to maintain their performance over time.
1 code implementation • EMNLP 2020 • Arturo Oncevay, Barry Haddow, Alexandra Birch
Sparse language vectors from linguistic typology databases and learned embeddings from tasks like multilingual machine translation have been investigated in isolation, without analysing how they could benefit from each other's language characterisation.
1 code implementation • EMNLP 2020 • Christos Baziotis, Barry Haddow, Alexandra Birch
A common solution is to exploit the knowledge of language models (LM) trained on abundant monolingual data.
1 code implementation • 19 Feb 2020 • Zaixiang Zheng, Xiang Yue, Shu-Jian Huang, Jia-Jun Chen, Alexandra Birch
Document-level machine translation manages to outperform sentence level models by a small margin, but have failed to be widely adopted.
1 code implementation • ICLR 2020 • Duygu Ataman, Wilker Aziz, Alexandra Birch
Translation into morphologically-rich languages challenges neural machine translation (NMT) models with extremely sparse vocabularies where atomic treatment of surface forms is unrealistic.
no code implementations • WS 2019 • Hiroaki Hayashi, Yusuke Oda, Alexandra Birch, Ioannis Konstas, Andrew Finch, Minh-Thang Luong, Graham Neubig, Katsuhito Sudoh
This document describes the findings of the Third Workshop on Neural Generation and Translation, held in concert with the annual conference of the Empirical Methods in Natural Language Processing (EMNLP 2019).
1 code implementation • WS 2019 • Duygu Ataman, Orhan Firat, Mattia A. Di Gangi, Marcello Federico, Alexandra Birch
Neural Machine Translation (NMT) models generally perform translation using a fixed-size lexical vocabulary, which is an important bottleneck on their generalization capability and overall translation quality.
no code implementations • WS 2019 • Rachel Bawden, Nikolay Bogoychev, Ulrich Germann, Roman Grundkiewicz, Faheem Kirefu, Antonio Valerio Miceli Barone, Alexandra Birch
For all translation directions, we created or used back-translations of monolingual data in the target language as additional synthetic training data.
no code implementations • WS 2018 • Mikel L. Forcada, Carolina Scarton, Lucia Specia, Barry Haddow, Alexandra Birch
A popular application of machine translation (MT) is gisting: MT is consumed as is to make sense of text in a foreign language.
no code implementations • WS 2018 • Alexandra Birch, Andrew Finch, Minh-Thang Luong, Graham Neubig, Yusuke Oda
This document describes the findings of the Second Workshop on Neural Machine Translation and Generation, held in concert with the annual conference of the Association for Computational Linguistics (ACL 2018).
2 code implementations • ACL 2018 • Marcin Junczys-Dowmunt, Roman Grundkiewicz, Tomasz Dwojak, Hieu Hoang, Kenneth Heafield, Tom Neckermann, Frank Seide, Ulrich Germann, Alham Fikri Aji, Nikolay Bogoychev, André F. T. Martins, Alexandra Birch
We present Marian, an efficient and self-contained Neural Machine Translation framework with an integrated automatic differentiation engine based on dynamic computation graphs.
no code implementations • NAACL 2018 • Rachel Bawden, Rico Sennrich, Alexandra Birch, Barry Haddow
Despite gains using BLEU, multi-encoder models give limited improvement in the handling of discourse phenomena: 50% accuracy on our coreference test set and 53. 5% for coherence/cohesion (compared to a non-contextual baseline of 50%).
no code implementations • WS 2017 • Rico Sennrich, Alexandra Birch, Anna Currey, Ulrich Germann, Barry Haddow, Kenneth Heafield, Antonio Valerio Miceli Barone, Philip Williams
This paper describes the University of Edinburgh's submissions to the WMT17 shared news translation and biomedical translation tasks.
3 code implementations • WS 2017 • Antonio Valerio Miceli Barone, Jindřich Helcl, Rico Sennrich, Barry Haddow, Alexandra Birch
It has been shown that increasing model depth improves the quality of neural machine translation.
4 code implementations • EACL 2017 • Rico Sennrich, Orhan Firat, Kyunghyun Cho, Alexandra Birch, Barry Haddow, Julian Hitschler, Marcin Junczys-Dowmunt, Samuel Läubli, Antonio Valerio Miceli Barone, Jozef Mokry, Maria Nădejde
We present Nematus, a toolkit for Neural Machine Translation.
no code implementations • WS 2017 • Maria Nadejde, Siva Reddy, Rico Sennrich, Tomasz Dwojak, Marcin Junczys-Dowmunt, Philipp Koehn, Alexandra Birch
Our results on WMT data show that explicitly modeling target-syntax improves machine translation quality for German->English, a high-resource pair, and for Romanian->English, a low-resource pair and also several syntactic phenomena including prepositional phrase attachment.
1 code implementation • EMNLP 2016 • Alexandra Birch, Omri Abend, Ondrej Bojar, Barry Haddow
Human evaluation of machine translation normally uses sentence-level measures such as relative ranking or adequacy scales.
1 code implementation • WS 2016 • Rico Sennrich, Barry Haddow, Alexandra Birch
We participated in the WMT 2016 shared news translation task by building neural translation systems for four language pairs, each trained in both directions: English<->Czech, English<->German, English<->Romanian and English<->Russian.
Ranked #1 on
Machine Translation
on WMT2016 English-Russian
2 code implementations • ACL 2016 • Rico Sennrich, Barry Haddow, Alexandra Birch
Neural Machine Translation (NMT) has obtained state-of-the art performance for several language pairs, while only using parallel data for training.
24 code implementations • ACL 2016 • Rico Sennrich, Barry Haddow, Alexandra Birch
Neural machine translation (NMT) models typically operate with a fixed vocabulary, but translation is an open-vocabulary problem.
Ranked #1 on
Machine Translation
on WMT2015 English-Russian