no code implementations • IWSLT 2017 • Pawel Przybysz, Marcin Chochowski, Rico Sennrich, Barry Haddow, Alexandra Birch
This paper describes the joint submission of Samsung Research and Development, Warsaw, Poland and the University of Edinburgh team to the IWSLT MT task for TED talks.
no code implementations • EAMT 2022 • Peggy van der Kreeft, Alexandra Birch, Sevi Sariisik, Felipe Sánchez-Martínez, Wilker Aziz
The GoURMET project, funded by the European Commission’s H2020 program (under grant agreement 825299), develops models for machine translation, in particular for low-resourced languages.
no code implementations • EMNLP (IWSLT) 2019 • Joanna Wetesko, Marcin Chochowski, Pawel Przybysz, Philip Williams, Roman Grundkiewicz, Rico Sennrich, Barry Haddow, None Barone, Valerio Miceli, Alexandra Birch
This paper describes the joint submission to the IWSLT 2019 English to Czech task by Samsung RD Institute, Poland, and the University of Edinburgh.
no code implementations • WMT (EMNLP) 2020 • Rachel Bawden, Alexandra Birch, Radina Dobreva, Arturo Oncevay, Antonio Valerio Miceli Barone, Philip Williams
We describe the University of Edinburgh’s submissions to the WMT20 news translation shared task for the low resource language pair English-Tamil and the mid-resource language pair English-Inuktitut.
no code implementations • NAACL 2022 • Jindřich Helcl, Barry Haddow, Alexandra Birch
In this paper, we point out flaws in the evaluation methodology present in the literature on NAR models and we provide a fair comparison between a state-of-the-art NAR model and the autoregressive submissions to the shared task.
no code implementations • IWSLT 2016 • Marcin Junczys-Dowmunt, Alexandra Birch
This paper describes the submission of the University of Edinburgh team to the IWSLT MT task for TED talks.
no code implementations • IWSLT 2016 • Maria Nădejde, Alexandra Birch, Philipp Koehn
String-to-tree MT systems translate verbs without lexical or syntactic context on the source side and with limited target-side context.
no code implementations • IWSLT (EMNLP) 2018 • Philip Williams, Marcin Chochowski, Pawel Przybysz, Rico Sennrich, Barry Haddow, Alexandra Birch
This paper describes the joint submission to the IWSLT 2018 Low Resource MT task by Samsung R&D Institute, Poland, and the University of Edinburgh.
1 code implementation • WMT (EMNLP) 2021 • Pinzhen Chen, Jindřich Helcl, Ulrich Germann, Laurie Burchell, Nikolay Bogoychev, Antonio Valerio Miceli Barone, Jonas Waldendorf, Alexandra Birch, Kenneth Heafield
This paper presents the University of Edinburgh’s constrained submissions of English-German and English-Hausa systems to the WMT 2021 shared task on news translation.
no code implementations • MTSummit 2021 • Alexandra Birch, Barry Haddow, Antonio Valerio Miceli Barone, Jindrich Helcl, Jonas Waldendorf, Felipe Sánchez Martínez, Mikel Forcada, Víctor Sánchez Cartagena, Juan Antonio Pérez-Ortiz, Miquel Esplà-Gomis, Wilker Aziz, Lina Murady, Sevi Sariisik, Peggy van der Kreeft, Kay Macquarrie
We find that starting from an existing large model pre-trained on 50languages leads to far better BLEU scores than pretraining on one high-resource language pair with a smaller model.
no code implementations • AMTA 2022 • Jonas Waldendorf, Alexandra Birch, Barry Hadow, Antonio Valerio Micele Barone
Dictionary-based data augmentation techniques have been used in the field of domain adaptation to learn words that do not appear in the parallel training data of a machine translation model.
no code implementations • BioNLP (ACL) 2022 • Matúš Falis, Hang Dong, Alexandra Birch, Beatrice Alex
We propose data augmentation and synthesis techniques in order to address these scenarios.
no code implementations • 24 Aug 2024 • Zhonghe Zhang, Xiaoyu He, Vivek Iyer, Alexandra Birch
Machine Translation of Culture-Specific Items (CSIs) poses significant challenges.
no code implementations • 23 Aug 2024 • Vivek Iyer, Bhavitvya Malik, Pavel Stepachev, Pinzhen Chen, Barry Haddow, Alexandra Birch
Similarly, diversity during SFT has been shown to promote significant transfer in LLMs across languages and tasks.
1 code implementation • 13 Jun 2024 • Weixuan Wang, Barry Haddow, Wei Peng, Alexandra Birch
In our work, we investigate how neuron activation is shared across languages by categorizing neurons into four distinct groups according to their responses across different languages for a particular input: all-shared, partial-shared, specific, and non-activated.
no code implementations • 3 May 2024 • Guillem Ramírez, Alexandra Birch, Ivan Titov
Either a cascading strategy is used, where a smaller LLM or both are called sequentially, or a routing strategy is used, where only one model is ever called.
1 code implementation • 2 May 2024 • Wenhao Zhu, ShuJian Huang, Fei Yuan, Cheng Chen, Jiajun Chen, Alexandra Birch
In this paper, we explore how broadly this method can be applied by examining its effects in reasoning with executable code and reasoning with common sense.
no code implementations • 24 Apr 2024 • Mateusz Klimaszewski, Piotr Andruszkiewicz, Alexandra Birch
Modular deep learning is the state-of-the-art solution for lifting the curse of multilinguality, preventing the impact of negative interference and enabling cross-lingual performance in Multilingual Pre-trained Language Models.
1 code implementation • 27 Mar 2024 • Mateusz Klimaszewski, Piotr Andruszkiewicz, Alexandra Birch
Moreover, we propose a method that allows the transfer of modules between incompatible PLMs without any change in the inference complexity.
no code implementations • 29 Feb 2024 • Tsz Kin Lam, Alexandra Birch, Barry Haddow
It also avoids speech discretization in inference and is more robust to the DSU tokenization.
1 code implementation • 2 Feb 2024 • Laurie Burchell, Alexandra Birch, Robert P. Thompson, Kenneth Heafield
Code switching (CS) is a very common phenomenon in written and spoken communication but one that is handled poorly by many natural language processing applications.
no code implementations • 1 Feb 2024 • Giulio Zhou, Tsz Kin Lam, Alexandra Birch, Barry Haddow
While there has been a growing interest in developing direct speech translation systems to avoid propagating errors and losing non-verbal content, prior work in direct S2TT has struggled to conclusively establish the advantages of integrating the acoustic signal directly into the translation process.
1 code implementation • 29 Jan 2024 • Nikita Moghe, Arnisa Fazla, Chantal Amrhein, Tom Kocmi, Mark Steedman, Alexandra Birch, Rico Sennrich, Liane Guillou
We benchmark metric performance, assess their incremental performance over successive campaigns, and measure their sensitivity to a range of linguistic phenomena.
1 code implementation • 24 Jan 2024 • Matúš Falis, Aryo Pradipta Gema, Hang Dong, Luke Daines, Siddharth Basetti, Michael Holder, Rose S Penfold, Alexandra Birch, Beatrice Alex
Neural coding models were trained on baseline and augmented data and evaluated on a MIMIC-IV test set.
1 code implementation • 15 Jan 2024 • Wenhao Zhu, ShuJian Huang, Fei Yuan, Shuaijie She, Jiajun Chen, Alexandra Birch
A typical solution is to translate instruction data into all languages of interest, and then train on the resulting multilingual data, which is called translate-training.
1 code implementation • 20 Dec 2023 • Weixuan Wang, Barry Haddow, Alexandra Birch
Knowledge represented in Large Language Models (LLMs) is quite often incorrect and can also become obsolete over time.
no code implementations • 16 Nov 2023 • Nikolay Bogoychev, Pinzhen Chen, Barry Haddow, Alexandra Birch
Deploying large language models (LLMs) encounters challenges due to intensive computational and memory requirements.
no code implementations • 21 Oct 2023 • Vivek Iyer, Edoardo Barba, Alexandra Birch, Jeff Z. Pan, Roberto Navigli
Lexical ambiguity is a significant and pervasive challenge in Neural Machine Translation (NMT), with many state-of-the-art (SOTA) NMT systems struggling to handle polysemous words (Campolungo et al., 2022).
no code implementations • 20 Oct 2023 • Guillem Ramírez, Matthias Lindemann, Alexandra Birch, Ivan Titov
To curtail the frequency of these calls, one can employ a smaller language model -- a student -- which is continuously trained on the responses of the LLM.
1 code implementation • 15 Oct 2023 • Weixuan Wang, Barry Haddow, Alexandra Birch, Wei Peng
Large language models (LLMs) have been treated as knowledge bases due to their strong performance in knowledge probing tasks.
no code implementations • 20 Sep 2023 • Vivek Iyer, Pinzhen Chen, Alexandra Birch
Resolving semantic ambiguity has long been recognised as a central challenge in the field of Machine Translation.
1 code implementation • 23 May 2023 • Laurie Burchell, Alexandra Birch, Nikolay Bogoychev, Kenneth Heafield
We achieve this by training on a curated dataset of monolingual data, the reliability of which we ensure by auditing a sample from each source and each language manually.
no code implementations • 23 May 2023 • Christos Baziotis, Biao Zhang, Alexandra Birch, Barry Haddow
Next, we analyze the impact of scale (from 90M to 1. 6B parameters) and find it is important for both methods, particularly DAE.
1 code implementation • 28 Mar 2023 • Nuno M. Guerreiro, Duarte Alves, Jonas Waldendorf, Barry Haddow, Alexandra Birch, Pierre Colombo, André F. T. Martins
Large-scale multilingual machine translation systems have demonstrated remarkable ability to translate directly between numerous languages, making them increasingly appealing for real-world applications.
no code implementations • 17 Jan 2023 • Biao Zhang, Barry Haddow, Alexandra Birch
Research on prompting has shown excellent performance with little or even no supervised training across many tasks.
no code implementations • 20 Dec 2022 • Nikita Moghe, Tom Sherborne, Mark Steedman, Alexandra Birch
We calculate the correlation between the metric's ability to predict a good/bad translation with the success/failure on the final task for the Translate-Test setup.
no code implementations • 20 Dec 2022 • Nikita Moghe, Evgeniia Razumovskaia, Liane Guillou, Ivan Vulić, Anna Korhonen, Alexandra Birch
We use MULTI3NLU++ to benchmark state-of-the-art multilingual models for the NLU tasks of intent detection and slot labelling for TOD systems in the multilingual setting.
1 code implementation • DeepLo 2022 • Laurie Burchell, Alexandra Birch, Kenneth Heafield
We also find evidence that lexical diversity is more important than syntactic for back translation performance.
no code implementations • NAACL 2022 • Arturo Oncevay, Duygu Ataman, Niels van Berkel, Barry Haddow, Alexandra Birch, Johannes Bjerva
In this work, we propose to reduce the rigidity of such claims, by quantifying morphological typology at the word and segment level.
no code implementations • 4 May 2022 • Jindřich Helcl, Barry Haddow, Alexandra Birch
In this paper, we point out flaws in the evaluation methodology present in the literature on NAR models and we provide a fair comparison between a state-of-the-art NAR model and the autoregressive submissions to the shared task.
no code implementations • RepL4NLP (ACL) 2022 • Antonio Valerio Miceli-Barone, Alexandra Birch, Rico Sennrich
Neural machine learning models can successfully model language that is similar to their training distribution, but they are highly susceptible to degradation under distribution shift, which occurs in many practical applications when processing out-of-domain (OOD) text.
1 code implementation • EMNLP 2021 • Nikita Moghe, Mark Steedman, Alexandra Birch
In this work, we enhance the transfer learning process by intermediate fine-tuning of pretrained multilingual models, where the multilingual models are fine-tuned with different but related data and/or tasks.
1 code implementation • EMNLP 2021 • Matúš Falis, Hang Dong, Alexandra Birch, Beatrice Alex
We propose a set of metrics for hierarchical evaluation using the depth-based representation.
Multi Label Text Classification Multi-Label Text Classification +1
no code implementations • CL (ACL) 2022 • Barry Haddow, Rachel Bawden, Antonio Valerio Miceli Barone, Jindřich Helcl, Alexandra Birch
We present a survey covering the state of the art in low-resource machine translation research.
1 code implementation • Findings (ACL) 2021 • Christos Baziotis, Ivan Titov, Alexandra Birch, Barry Haddow
Unsupervised cross-lingual pretraining has achieved strong results in neural machine translation (NMT), by drastically reducing the need for large parallel data.
1 code implementation • EACL 2021 • Farid Arthaud, Rachel Bawden, Alexandra Birch
Machine translation (MT) models used in industries with constantly changing topics, such as translation or news agencies, need to adapt to new data to maintain their performance over time.
1 code implementation • EMNLP 2020 • Christos Baziotis, Barry Haddow, Alexandra Birch
A common solution is to exploit the knowledge of language models (LM) trained on abundant monolingual data.
1 code implementation • EMNLP 2020 • Arturo Oncevay, Barry Haddow, Alexandra Birch
Sparse language vectors from linguistic typology databases and learned embeddings from tasks like multilingual machine translation have been investigated in isolation, without analysing how they could benefit from each other's language characterisation.
1 code implementation • 19 Feb 2020 • Zaixiang Zheng, Xiang Yue, Shu-Jian Huang, Jia-Jun Chen, Alexandra Birch
Document-level machine translation manages to outperform sentence level models by a small margin, but have failed to be widely adopted.
1 code implementation • ICLR 2020 • Duygu Ataman, Wilker Aziz, Alexandra Birch
Translation into morphologically-rich languages challenges neural machine translation (NMT) models with extremely sparse vocabularies where atomic treatment of surface forms is unrealistic.
no code implementations • WS 2019 • Hiroaki Hayashi, Yusuke Oda, Alexandra Birch, Ioannis Konstas, Andrew Finch, Minh-Thang Luong, Graham Neubig, Katsuhito Sudoh
This document describes the findings of the Third Workshop on Neural Generation and Translation, held in concert with the annual conference of the Empirical Methods in Natural Language Processing (EMNLP 2019).
1 code implementation • WS 2019 • Duygu Ataman, Orhan Firat, Mattia A. Di Gangi, Marcello Federico, Alexandra Birch
Neural Machine Translation (NMT) models generally perform translation using a fixed-size lexical vocabulary, which is an important bottleneck on their generalization capability and overall translation quality.
no code implementations • WS 2019 • Rachel Bawden, Nikolay Bogoychev, Ulrich Germann, Roman Grundkiewicz, Faheem Kirefu, Antonio Valerio Miceli Barone, Alexandra Birch
For all translation directions, we created or used back-translations of monolingual data in the target language as additional synthetic training data.
no code implementations • WS 2018 • Mikel L. Forcada, Carolina Scarton, Lucia Specia, Barry Haddow, Alexandra Birch
A popular application of machine translation (MT) is gisting: MT is consumed as is to make sense of text in a foreign language.
no code implementations • WS 2018 • Alexandra Birch, Andrew Finch, Minh-Thang Luong, Graham Neubig, Yusuke Oda
This document describes the findings of the Second Workshop on Neural Machine Translation and Generation, held in concert with the annual conference of the Association for Computational Linguistics (ACL 2018).
2 code implementations • ACL 2018 • Marcin Junczys-Dowmunt, Roman Grundkiewicz, Tomasz Dwojak, Hieu Hoang, Kenneth Heafield, Tom Neckermann, Frank Seide, Ulrich Germann, Alham Fikri Aji, Nikolay Bogoychev, André F. T. Martins, Alexandra Birch
We present Marian, an efficient and self-contained Neural Machine Translation framework with an integrated automatic differentiation engine based on dynamic computation graphs.
no code implementations • NAACL 2018 • Rachel Bawden, Rico Sennrich, Alexandra Birch, Barry Haddow
Despite gains using BLEU, multi-encoder models give limited improvement in the handling of discourse phenomena: 50% accuracy on our coreference test set and 53. 5% for coherence/cohesion (compared to a non-contextual baseline of 50%).
no code implementations • WS 2017 • Rico Sennrich, Alexandra Birch, Anna Currey, Ulrich Germann, Barry Haddow, Kenneth Heafield, Antonio Valerio Miceli Barone, Philip Williams
This paper describes the University of Edinburgh's submissions to the WMT17 shared news translation and biomedical translation tasks.
3 code implementations • WS 2017 • Antonio Valerio Miceli Barone, Jindřich Helcl, Rico Sennrich, Barry Haddow, Alexandra Birch
It has been shown that increasing model depth improves the quality of neural machine translation.
4 code implementations • EACL 2017 • Rico Sennrich, Orhan Firat, Kyunghyun Cho, Alexandra Birch, Barry Haddow, Julian Hitschler, Marcin Junczys-Dowmunt, Samuel Läubli, Antonio Valerio Miceli Barone, Jozef Mokry, Maria Nădejde
We present Nematus, a toolkit for Neural Machine Translation.
no code implementations • WS 2017 • Maria Nadejde, Siva Reddy, Rico Sennrich, Tomasz Dwojak, Marcin Junczys-Dowmunt, Philipp Koehn, Alexandra Birch
Our results on WMT data show that explicitly modeling target-syntax improves machine translation quality for German->English, a high-resource pair, and for Romanian->English, a low-resource pair and also several syntactic phenomena including prepositional phrase attachment.
1 code implementation • EMNLP 2016 • Alexandra Birch, Omri Abend, Ondrej Bojar, Barry Haddow
Human evaluation of machine translation normally uses sentence-level measures such as relative ranking or adequacy scales.
1 code implementation • WS 2016 • Rico Sennrich, Barry Haddow, Alexandra Birch
We participated in the WMT 2016 shared news translation task by building neural translation systems for four language pairs, each trained in both directions: English<->Czech, English<->German, English<->Romanian and English<->Russian.
Ranked #1 on Machine Translation on WMT2016 Czech-English
2 code implementations • ACL 2016 • Rico Sennrich, Barry Haddow, Alexandra Birch
Neural Machine Translation (NMT) has obtained state-of-the art performance for several language pairs, while only using parallel data for training.
26 code implementations • ACL 2016 • Rico Sennrich, Barry Haddow, Alexandra Birch
Neural machine translation (NMT) models typically operate with a fixed vocabulary, but translation is an open-vocabulary problem.
Ranked #1 on Machine Translation on WMT2015 English-Russian