no code implementations • WMT (EMNLP) 2020 • Nikita Moghe, Christian Hardmeier, Rachel Bawden
Our baseline systems are transformer-big models that are pre-trained on the WMT’19 News Translation task and fine-tuned on pseudo-in-domain web crawled data and in-domain task data.
1 code implementation • 27 Aug 2024 • Vilém Zouhar, Pinzhen Chen, Tsz Kin Lam, Nikita Moghe, Barry Haddow
The COMET metric has blazed a trail in the machine translation community, given its strong correlation with human judgements of translation quality.
1 code implementation • 29 Jan 2024 • Nikita Moghe, Arnisa Fazla, Chantal Amrhein, Tom Kocmi, Mark Steedman, Alexandra Birch, Rico Sennrich, Liane Guillou
We benchmark metric performance, assess their incremental performance over successive campaigns, and measure their sensitivity to a range of linguistic phenomena.
1 code implementation • 16 Nov 2023 • Nikita Moghe, Patrick Xia, Jacob Andreas, Jason Eisner, Benjamin Van Durme, Harsh Jhamtani
Users of natural language interfaces, generally powered by Large Language Models (LLMs), often must repeat their preferences each time they make a similar request.
no code implementations • 2 Nov 2023 • Chantal Amrhein, Nikita Moghe, Liane Guillou
We benchmark the performance of segmentlevel metrics submitted to WMT 2023 using the ACES Challenge Set (Amrhein et al., 2022).
no code implementations • 20 Dec 2022 • Nikita Moghe, Evgeniia Razumovskaia, Liane Guillou, Ivan Vulić, Anna Korhonen, Alexandra Birch
We use MULTI3NLU++ to benchmark state-of-the-art multilingual models for the NLU tasks of intent detection and slot labelling for TOD systems in the multilingual setting.
no code implementations • 20 Dec 2022 • Nikita Moghe, Tom Sherborne, Mark Steedman, Alexandra Birch
We calculate the correlation between the metric's ability to predict a good/bad translation with the success/failure on the final task for the Translate-Test setup.
1 code implementation • 27 Oct 2022 • Chantal Amrhein, Nikita Moghe, Liane Guillou
As machine translation (MT) metrics improve their correlation with human judgement every year, it is crucial to understand the limitations of such metrics at the segment level.
Ranked #1 on Machine Translation on ACES
1 code implementation • EMNLP 2021 • Nikita Moghe, Mark Steedman, Alexandra Birch
In this work, we enhance the transfer learning process by intermediate fine-tuning of pretrained multilingual models, where the multilingual models are fine-tuned with different but related data and/or tasks.
1 code implementation • WS 2020 • Nikita Moghe, Priyesh Vijayan, Balaraman Ravindran, Mitesh M. Khapra
This requires capturing structural, sequential and semantic information from the conversation context and the background resources.
1 code implementation • EMNLP 2018 • Nikita Moghe, Siddhartha Arora, Suman Banerjee, Mitesh M. Khapra
Existing dialog datasets contain a sequence of utterances and responses without any explicit background knowledge associated with them.
no code implementations • COLING 2018 • Suman Banerjee, Nikita Moghe, Siddhartha Arora, Mitesh M. Khapra
("Can you help me in booking a table at this restaurant?").