1 code implementation • EANCS 2021 • Alexandru Coca, Bo-Hsiang Tseng, Bill Byrne
The evaluation of dialogue systems in interaction with simulated users has been proposed to improve turn-level, corpus-based metrics which can only evaluate test cases encountered in a corpus and cannot measure system’s ability to sustain multi-turn interactions.
no code implementations • ECNLP (ACL) 2022 • Xiaoyu Shen, Gianni Barlacchi, Marco del Tredici, Weiwei Cheng, Bill Byrne, Adrià Gispert
In this paper, we build a benchmark with annotations for both evidence selection and answer generation covering 6 information sources.
1 code implementation • GeBNLP (COLING) 2020 • Danielle Saunders, Rosie Sallis, Bill Byrne
Neural Machine Translation (NMT) has been shown to struggle with grammatical gender that is dependent on the gender of human referents, which can cause gender bias effects.
no code implementations • EMNLP 2021 • Eva Hasler, Tobias Domhan, Jonay Trenous, Ke Tran, Bill Byrne, Felix Hieber
Building neural machine translation systems to perform well on a specific target domain is a well-studied problem.
no code implementations • Findings (ACL) 2022 • Tisha Anders, Alexandru Coca, Bill Byrne
Our approach is to augment the training set of a given target corpus with alien corpora which have different semantic representations.
1 code implementation • 30 May 2024 • David Stap, Eva Hasler, Bill Byrne, Christof Monz, Ke Tran
In particular, we observe a decline in the ability to perform formality steering, to produce technical translations through few-shot examples, and to perform document-level translation.
no code implementations • 17 Apr 2024 • Dawei Zhu, Sony Trenous, Xiaoyu Shen, Dietrich Klakow, Bill Byrne, Eva Hasler
Recent research has shown that large language models (LLMs) can achieve remarkable translation performance through supervised fine-tuning (SFT) using only a small amount of parallel data.
1 code implementation • 10 Apr 2024 • Jinghong Chen, Weizhe Lin, Jingbiao Mei, Bill Byrne
The Directed Acyclic Transformer is a fast non-autoregressive (NAR) model that performs well in Neural Machine Translation.
no code implementations • 17 Mar 2024 • Igor Sterner, Weizhe Lin, Jinghong Chen, Bill Byrne
Two approaches have emerged to input images into large language models (LLMs).
no code implementations • 15 Mar 2024 • Hakim Sidahmed, Samrat Phatale, Alex Hutcheson, Zhuonan Lin, Zhang Chen, Zac Yu, Jarvis Jin, Simral Chaudhary, Roman Komarytsia, Christiane Ahlheim, Yonghao Zhu, Bowen Li, Saravanan Ganesh, Bill Byrne, Jessica Hoffmann, Hassan Mansoor, Wei Li, Abhinav Rastogi, Lucas Dixon
In this work, we empirically evaluate the setup of Parameter Efficient Reinforcement Learning from Human Feedback (PE-RLHF) that leverages LoRA fine-tuning for Reward Modeling, and Reinforcement Learning.
1 code implementation • 13 Feb 2024 • Weizhe Lin, Jingbiao Mei, Jinghong Chen, Bill Byrne
Large Multimodal Models (LMMs) excel in natural language and visual understanding but are challenged by exacting tasks such as Knowledge-based Visual Question Answering (KB-VQA) which involve the retrieval of relevant information from document collections to use in shaping answers to questions.
Ranked #1 on Retrieval on InfoSeek (using extra training data)
1 code implementation • 14 Nov 2023 • Guangyu Yang, Jinghong Chen, Weizhe Lin, Bill Byrne
Minimum Bayes Risk (MBR) decoding can significantly improve translation performance of Multilingual Large Language Models (MLLMs).
no code implementations • 14 Nov 2023 • Jingbiao Mei, Jinghong Chen, Weizhe Lin, Bill Byrne, Marcus Tomalin
Hateful memes have emerged as a significant concern on the Internet.
Ranked #2 on Meme Classification on Hateful Memes
no code implementations • 12 Oct 2023 • Junxiao Shen, John J. Dudley, Jingyao Zheng, Bill Byrne, Per Ola Kristensson
However, the task of prompting large language models to specialize in specific text prediction tasks can be challenging, particularly for designers without expertise in prompt engineering.
1 code implementation • NeurIPS 2023 • Weizhe Lin, Jinghong Chen, Jingbiao Mei, Alexandru Coca, Bill Byrne
FLMR addresses two major limitations in RA-VQA's retriever: (1) the image representations obtained via image-to-text transforms can be incomplete and inaccurate and (2) relevance scores between queries and documents are computed with one-dimensional embeddings, which can be insensitive to finer-grained relevance.
Ranked #1 on Retrieval on OK-VQA
no code implementations • 23 Sep 2023 • Alexandru Coca, Bo-Hsiang Tseng, Jinghong Chen, Weizhe Lin, Weixuan Zhang, Tisha Anders, Bill Byrne
Schema-guided dialogue state trackers can generalise to new domains without further training, yet they are sensitive to the writing style of the schemata.
1 code implementation • 16 May 2023 • Xiaoyu Shen, Akari Asai, Bill Byrne, Adrià De Gispert
To study this practical industrial task, we present xPQA, a large-scale annotated cross-lingual PQA dataset in 12 languages across 9 branches, and report results in (1) candidate ranking, to select the best English candidate containing the information to answer a non-English question; and (2) answer generation, to generate a natural-sounding non-English answer based on the selected English candidate.
no code implementations • 19 Mar 2023 • Weizhe Lin, Zhilin Wang, Bill Byrne
The widely used Fact-based Visual Question Answering (FVQA) dataset contains visually-grounded questions that require information retrieval using common sense knowledge graphs to answer.
1 code implementation • 29 Jan 2023 • Jinghong Chen, Weizhe Lin, Bill Byrne
We show that SGSAcc can be applied to evaluate utterances generated from a wide range of dialogue actions in the Schema Guided Dialogue (SGD) dataset with good agreement with human judgment.
1 code implementation • 7 Oct 2022 • Weizhe Lin, Bill Byrne
The strong retrieval ability of our model significantly reduces the number of retrieved documents needed in training, yielding significant benefits in answer quality and computation required for training.
Ranked #2 on Retrieval on OK-VQA
no code implementations • 5 Aug 2022 • Xiaoyu Shen, Svitlana Vakulenko, Marco del Tredici, Gianni Barlacchi, Bill Byrne, Adrià De Gispert
Dense retrieval (DR) approaches based on powerful pre-trained language models (PLMs) achieved significant advances and have become a key component for modern open-domain question-answering systems.
1 code implementation • NAACL 2022 • Tobias Domhan, Eva Hasler, Ke Tran, Sony Trenous, Bill Byrne, Felix Hieber
Vocabulary selection, or lexical shortlisting, is a well-known technique to improve latency of Neural Machine Translation models by constraining the set of allowed output words during inference.
no code implementations • NLP4ConvAI (ACL) 2022 • Marco del Tredici, Xiaoyu Shen, Gianni Barlacchi, Bill Byrne, Adrià De Gispert
In conversational QA, models have to leverage information in previous turns to answer upcoming questions.
no code implementations • 2 Apr 2022 • Weizhe Lin, Linjun Shou, Ming Gong, Pei Jian, Zhilin Wang, Bill Byrne, Daxin Jiang
Knowledge graph (KG) based Collaborative Filtering is an effective approach to personalizing recommendation systems for relatively static domains such as movies and books, by leveraging structured information from KG to enrich both item and user representations.
1 code implementation • ACL 2021 • Bo-Hsiang Tseng, Yinpei Dai, Florian Kreyssig, Bill Byrne
Our goal is to develop a modelling framework that can incorporate new dialogue scenarios through self-play between the two agents.
1 code implementation • Findings (ACL) 2022 • Danielle Saunders, Rosie Sallis, Bill Byrne
Neural machine translation inference procedures like beam search generate the most likely output under the model.
1 code implementation • EMNLP 2021 • Weizhe Lin, Bo-Hsiang Tseng, Bill Byrne
Dialogue State Tracking is central to multi-domain task-oriented dialogue systems, responsible for extracting information from user utterances.
Ranked #1 on Multi-domain Dialogue State Tracking on MULTIWOZ 2.0
no code implementations • ACL 2021 • Bill Byrne, Karthik Krishnamoorthi, Saravanan Ganesh, Mihir Sanjay Kale
In terms of data, we introduce TicketTalk, a movie ticketing dialog dataset with 23, 789 annotated conversations.
no code implementations • NLP4CALL (COLING) 2020 • Andrew Caines, Helen Yannakoudakis, Helena Edmondson, Helen Allen, Pascual Pérez-Paredes, Bill Byrne, Paula Buttery
The Teacher-Student Chatroom Corpus (TSCC) is a collection of written conversations captured during one-to-one lessons between teachers and learners of English.
no code implementations • AACL (WAT) 2020 • Danielle Saunders, Weston Feely, Bill Byrne
One possible approach to this problem uses sub-character decomposition for training and test sentences.
1 code implementation • 11 Oct 2020 • Danielle Saunders, Rosie Sallis, Bill Byrne
Neural Machine Translation (NMT) has been shown to struggle with grammatical gender that is dependent on the gender of human referents, which can cause gender bias effects.
no code implementations • WMT (EMNLP) 2020 • Danielle Saunders, Bill Byrne
The 2020 WMT Biomedical translation task evaluated Medline abstract translations.
no code implementations • ACL 2020 • Danielle Saunders, Felix Stahlberg, Bill Byrne
We find that each of these lines of research has a clear space in it for the other, and propose merging them with a scheme that allows a document-level evaluation metric to be used in the NMT training objective.
2 code implementations • ACL 2020 • Danielle Saunders, Bill Byrne
During inference we propose a lattice-rescoring scheme which outperforms all systems evaluated in Stanovsky et al (2019) on WinoMT with no degradation of general test set BLEU, and we show this scheme can be applied to remove gender bias in the output of `black box` online commercial MT systems.
no code implementations • 26 Nov 2019 • Bo-Hsiang Tseng, Marek Rei, Paweł Budzianowski, Richard E. Turner, Bill Byrne, Anna Korhonen
Dialogue systems benefit greatly from optimizing on detailed annotations, such as transcribed utterances, internal dialogue state representations and dialogue act labels.
no code implementations • IJCNLP 2019 • Bo-Hsiang Tseng, Marek Rei, Pawe{\l} Budzianowski, Richard Turner, Bill Byrne, Anna Korhonen
Dialogue systems benefit greatly from optimizing on detailed annotations, such as transcribed utterances, internal dialogue state representations and dialogue act labels.
no code implementations • WS 2019 • Filip Radlinski, Krisztian Balog, Bill Byrne, Karthik Krishnamoorthi
Studying the dialogues in one domain, we present a brief quantitative analysis of how people describe movie preferences at scale.
1 code implementation • IJCNLP 2019 • Bill Byrne, Karthik Krishnamoorthi, Chinnadhurai Sankar, Arvind Neelakantan, Daniel Duckworth, Semih Yavuz, Ben Goodrich, Amit Dubey, Andy Cedilnik, Kyu-Young Kim
A significant barrier to progress in data-driven approaches to building dialog systems is the lack of high quality, goal-oriented conversational data.
no code implementations • IJCNLP 2019 • Felix Stahlberg, Bill Byrne
We report on search errors and model errors in neural machine translation (NMT).
no code implementations • WS 2019 • Felix Stahlberg, Danielle Saunders, Adri{\`a} de Gispert, Bill Byrne
Two techniques provide the fabric of the Cambridge University Engineering Department{'}s (CUED) entry to the WMT19 evaluation campaign: elastic weight consolidation (EWC) and different forms of language modelling (LMs).
no code implementations • WS 2019 • Zheng Yuan, Felix Stahlberg, Marek Rei, Bill Byrne, Helen Yannakoudakis
In this paper, we describe our submission to the BEA 2019 shared task on grammatical error correction.
no code implementations • WS 2019 • Felix Stahlberg, Bill Byrne
We describe two entries from the Cambridge University Engineering Department to the BEA 2019 Shared Task on grammatical error correction.
no code implementations • WS 2019 • Danielle Saunders, Felix Stahlberg, Bill Byrne
The 2019 WMT Biomedical translation task involved translating Medline abstracts.
no code implementations • 11 Jun 2019 • Felix Stahlberg, Danielle Saunders, Adria de Gispert, Bill Byrne
Two techniques provide the fabric of the Cambridge University Engineering Department's (CUED) entry to the WMT19 evaluation campaign: elastic weight consolidation (EWC) and different forms of language modelling (LMs).
no code implementations • ACL 2019 • Danielle Saunders, Felix Stahlberg, Adria de Gispert, Bill Byrne
We investigate adaptive ensemble weighting for Neural Machine Translation, addressing the case of improving performance on a new and potentially unknown domain without sacrificing performance on the original domain.
no code implementations • NAACL 2019 • Felix Stahlberg, Christopher Bryant, Bill Byrne
Language model based GEC (LM-GEC) is a promising alternative which does not rely on annotated training data.
1 code implementation • WS 2018 • Felix Stahlberg, Danielle Saunders, Bill Byrne
We propose to achieve explainable neural machine translation (NMT) by changing the output representation to explain itself.
no code implementations • WS 2018 • Felix Stahlberg, Adria de Gispert, Bill Byrne
The University of Cambridge submission to the WMT18 news translation task focuses on the combination of diverse models of translation.
no code implementations • NAACL 2018 • Eva Hasler, Adrià De Gispert, Gonzalo Iglesias, Bill Byrne
Despite the impressive quality improvements yielded by neural machine translation (NMT) systems, controlling their translation output to adhere to user-provided terminology constraints remains an open problem.
no code implementations • ACL 2018 • Danielle Saunders, Felix Stahlberg, Adria de Gispert, Bill Byrne
We explore strategies for incorporating target syntax into Neural Machine Translation.
no code implementations • NAACL 2018 • Gonzalo Iglesias, William Tambellini, Adrià De Gispert, Eva Hasler, Bill Byrne
We describe a batched beam decoding algorithm for NMT with LMBR n-gram posteriors, showing that LMBR techniques still yield gains on top of the best recently reported results with Transformers.
no code implementations • WS 2018 • Felix Stahlberg, Danielle Saunders, Gonzalo Iglesias, Bill Byrne
SGNMT is a decoding platform for machine translation which allows paring various modern neural models of translation with different kinds of constraints and symbolic models.
no code implementations • EMNLP 2017 • Lucas Sterckx, Jason Naradowsky, Bill Byrne, Thomas Demeester, Chris Develder
Comprehending lyrics, as found in songs and poems, can pose a challenge to human and machine readers alike.
1 code implementation • WS 2017 • Eva Hasler, Felix Stahlberg, Marcus Tomalin, Adri`a de Gispert, Bill Byrne
We compare several language models for the word-ordering task and propose a new bag-to-sequence neural model based on attention-based sequence-to-sequence models.
1 code implementation • EMNLP 2017 • Felix Stahlberg, Eva Hasler, Danielle Saunders, Bill Byrne
This paper introduces SGNMT, our experimental platform for machine translation research.
no code implementations • EMNLP 2017 • Felix Stahlberg, Bill Byrne
Ensembling is a well-known technique in neural machine translation (NMT) to improve system performance.
no code implementations • EACL 2017 • Felix Stahlberg, Adrià De Gispert, Eva Hasler, Bill Byrne
This makes our approach much more flexible than $n$-best list or lattice rescoring as the neural decoder is not restricted to the SMT search space.
no code implementations • WS 2016 • Felix Stahlberg, Eva Hasler, Bill Byrne
This paper presents the University of Cambridge submission to WMT16.
no code implementations • ACL 2016 • Felix Stahlberg, Eva Hasler, Aurelien Waite, Bill Byrne
We investigate the use of hierarchical phrase-based SMT lattices in end-to-end neural machine translation (NMT).
no code implementations • NAACL 2016 • Daniel Beck, Adrià De Gispert, Gonzalo Iglesias, Aurelien Waite, Bill Byrne
We address the problem of automatically finding the parameters of a statistical machine translation system that maximize BLEU scores while ensuring that decoding speed exceeds a minimum value.