no code implementations • EAMT 2022 • Peggy van der Kreeft, Alexandra Birch, Sevi Sariisik, Felipe Sánchez-Martínez, Wilker Aziz
The GoURMET project, funded by the European Commission’s H2020 program (under grant agreement 825299), develops models for machine translation, in particular for low-resourced languages.
no code implementations • MTSummit 2021 • Alexandra Birch, Barry Haddow, Antonio Valerio Miceli Barone, Jindrich Helcl, Jonas Waldendorf, Felipe Sánchez Martínez, Mikel Forcada, Víctor Sánchez Cartagena, Juan Antonio Pérez-Ortiz, Miquel Esplà-Gomis, Wilker Aziz, Lina Murady, Sevi Sariisik, Peggy van der Kreeft, Kay Macquarrie
We find that starting from an existing large model pre-trained on 50languages leads to far better BLEU scores than pretraining on one high-resource language pair with a smaller model.
1 code implementation • 23 Jul 2024 • Pedro Ferreira, Wilker Aziz, Ivan Titov
Explanation regularisation (ER) has been introduced as a way to guide models to make their predictions in a manner more akin to humans, i. e., making their attributions "plausible".
1 code implementation • 27 Feb 2024 • Evgenia Ilia, Wilker Aziz
This can be seen as assessing a form of calibration, which, in the context of text classification, Baan et al. (2022) termed calibration to human uncertainty.
no code implementations • 25 Feb 2024 • Joris Baan, Raquel Fernández, Barbara Plank, Wilker Aziz
With the rise of increasingly powerful and user-facing NLP systems, there is growing interest in assessing whether they have a good representation of uncertainty by evaluating the quality of their predictive distribution over outcomes.
no code implementations • 28 Jul 2023 • Joris Baan, Nico Daheim, Evgenia Ilia, Dennis Ulmer, Haau-Sing Li, Raquel Fernández, Barbara Plank, Rico Sennrich, Chrysoula Zerva, Wilker Aziz
Recent advances of powerful Language Models have allowed Natural Language Generation (NLG) to emerge as an important technology that can not only perform traditional tasks like summarisation or translation, but also serve as a natural language interface to a variety of applications.
1 code implementation • 19 May 2023 • Mario Giulianelli, Joris Baan, Wilker Aziz, Raquel Fernández, Barbara Plank
In Natural Language Generation (NLG) tasks, for any input, multiple communicative goals are plausible, and any goal can be put into words, or produced, in multiple ways.
1 code implementation • 13 Apr 2023 • Angelos Nalmpantis, Apostolos Panagiotopoulos, John Gkountouras, Konstantinos Papakostas, Wilker Aziz
The lack of interpretability of the Vision Transformer may hinder its use in critical real-world applications despite its effectiveness.
1 code implementation • 28 Oct 2022 • Joris Baan, Wilker Aziz, Barbara Plank, Raquel Fernández
Calibration is a popular framework to evaluate whether a classifier knows when it does not know - i. e., its predictive probabilities are a good indication of how likely a prediction is to be correct.
no code implementations • 6 Apr 2022 • Claartje Barkhof, Wilker Aziz
We propose a framework for the statistical evaluation of variational auto-encoders (VAEs) and test two instances of this framework in the context of modelling images of handwritten digits and a corpus of English text.
1 code implementation • EMNLP 2021 • Nicola De Cao, Wilker Aziz, Ivan Titov
Generative approaches have been recently shown to be effective for both Entity Disambiguation and Entity Linking (i. e., joint mention detection and disambiguation).
Ranked #5 on Entity Linking on AIDA-CoNLL
1 code implementation • 10 Aug 2021 • Bryan Eikema, Wilker Aziz
The mode and other high-probability translations found by beam search have been shown to often be inadequate in a number of ways.
1 code implementation • ICLR 2022 • António Farinhas, Wilker Aziz, Vlad Niculae, André F. T. Martins
Neural networks and other machine learning models compute continuous representations, while humans communicate mostly through discrete symbols.
3 code implementations • EMNLP 2021 • Nicola De Cao, Wilker Aziz, Ivan Titov
We present KnowledgeEditor, a method which can be used to edit this knowledge and, thus, fix 'bugs' or unexpected predictions without the need for expensive re-training or fine-tuning.
1 code implementation • 24 Oct 2020 • Dhruba Pujary, Camilo Thorne, Wilker Aziz
The detection and normalization of diseases in biomedical texts are key biomedical natural language processing tasks.
1 code implementation • NeurIPS 2020 • Gonçalo M. Correia, Vlad Niculae, Wilker Aziz, André F. T. Martins
In this paper, we propose a new training strategy which replaces these estimators by an exact yet efficient marginalization.
no code implementations • 11 Jun 2020 • Rob Hesselink, Wilker Aziz
Normalising flows (NFs) for discrete data are challenging because parameterising bijective transformations of discrete variables requires predicting discrete/integer parameters.
2 code implementations • 8 Jun 2020 • Nicola De Cao, Wilker Aziz
We propose a novel distribution, the Power Spherical distribution, which retains some of the important aspects of the vMF (e. g., support on the hyper-sphere, symmetry about its mean direction parameter, known KL from other vMF distributions) while addressing its main drawbacks (i. e., scalability and numerical stability).
no code implementations • COLING 2020 • Bryan Eikema, Wilker Aziz
We argue that the evidence corroborates the inadequacy of MAP decoding more than casts doubt on the model and its training algorithm.
2 code implementations • EMNLP 2020 • Nicola De Cao, Michael Schlichtkrull, Wilker Aziz, Ivan Titov
Attribution methods assess the contribution of inputs to the model prediction.
1 code implementation • ICLR 2020 • Duygu Ataman, Wilker Aziz, Alexandra Birch
Translation into morphologically-rich languages challenges neural machine translation (NMT) models with extremely sparse vocabularies where atomic treatment of surface forms is unrealistic.
no code implementations • WS 2019 • Alex Birch, ra, Barry Haddow, Ivan Tito, Antonio Valerio Miceli Barone, Rachel Bawden, Felipe S{\'a}nchez-Mart{\'\i}nez, Mikel L. Forcada, Miquel Espl{\`a}-Gomis, V{\'\i}ctor S{\'a}nchez-Cartagena, Juan Antonio P{\'e}rez-Ortiz, Wilker Aziz, Andrew Secker, Peggy van der Kreeft
1 code implementation • ACL 2019 • Jasmijn Bastings, Wilker Aziz, Ivan Titov
The success of neural networks comes hand in hand with a desire for more interpretability.
1 code implementation • ACL 2020 • Tom Pelsmaeker, Wilker Aziz
We concentrate on one such model, the variational auto-encoder, which we argue is an important building block in hierarchical probabilistic models of language.
4 code implementations • 9 Apr 2019 • Nicola De Cao, Ivan Titov, Wilker Aziz
Recently, as an alternative to hand-crafted bijections, Huang et al. (2018) proposed neural autoregressive flow (NAF) which is a universal approximator for density functions.
Ranked #1 on Density Estimation on UCI MINIBOONE
no code implementations • 18 Jan 2019 • Jasmijn Bastings, Wilker Aziz, Ivan Titov, Khalil Sima'an
Recently it was shown that linguistic structure predicted by a supervised parser can be beneficial for neural machine translation (NMT).
1 code implementation • ACL 2019 • Iacer Calixto, Miguel Rios, Wilker Aziz
In this work, we propose to model the interaction between visual and textual features for multi-modal neural machine translation (MMT) through a latent variable model.
Ranked #9 on Multimodal Machine Translation on Multi30K
1 code implementation • NAACL 2019 • Nicola De Cao, Wilker Aziz, Ivan Titov
Most research in reading comprehension has focused on answering questions based on individual documents or even single paragraphs.
1 code implementation • WS 2019 • Bryan Eikema, Wilker Aziz
We present a deep generative model of bilingual sentence pairs for machine translation.
no code implementations • ACL 2018 • Wilker Aziz, Philip Schulz
Using DGMs one can easily design latent variable models that account for missing observations and thereby enable unsupervised and semi-supervised learning with neural networks.
1 code implementation • ACL 2018 • Philip Schulz, Wilker Aziz, Trevor Cohn
The process of translation is ambiguous, in that there are typically many valid trans- lations for a given sentence.
1 code implementation • NAACL 2018 • Miguel Rios, Wilker Aziz, Khalil Sima'an
This work exploits translation data as a source of semantically relevant learning signal for models of word representation.
no code implementations • WS 2017 • Jan-Thorsten Peter, Hermann Ney, Ond{\v{r}}ej Bojar, Ngoc-Quan Pham, Jan Niehues, Alex Waibel, Franck Burlot, Fran{\c{c}}ois Yvon, M{\=a}rcis Pinnis, Valters {\v{S}}ics, Jasmijn Bastings, Miguel Rios, Wilker Aziz, Philip Williams, Fr{\'e}d{\'e}ric Blain, Lucia Specia
no code implementations • EMNLP 2017 • Jasmijn Bastings, Ivan Titov, Wilker Aziz, Diego Marcheggiani, Khalil Sima'an
We present a simple and effective approach to incorporating syntactic structure into neural attention-based encoder-decoder models for machine translation.
no code implementations • COLING 2016 • Philip Schulz, Wilker Aziz
In order to make our model useful in practice, we devise an auxiliary variable Gibbs sampler that allows us to resample alignment links in constant time independently of the target sentence length.
1 code implementation • LREC 2016 • Karin Sim Smith, Wilker Aziz, Lucia Specia
We describe COHERE, our coherence toolkit which incorporates various complementary models for capturing and measuring different aspects of text coherence.
no code implementations • 13 Sep 2015 • Raymond W. M. Ng, Mortaza Doulaty, Rama Doddipatla, Wilker Aziz, Kashif Shah, Oscar Saz, Madina Hasan, Ghada Alharbi, Lucia Specia, Thomas Hain
The USFD primary system incorporates state-of-the-art ASR and MT techniques and gives a BLEU score of 23. 45 and 14. 75 on the English-to-French and English-to-German speech-to-text translation task with the IWSLT 2014 data.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +4
no code implementations • LREC 2012 • Wilker Aziz, Sheila Castilho, Lucia Specia
Given the significant improvements in Machine Translation (MT) quality and the increasing demand for translations, post-editing of automatic translations is becoming a popular practice in the translation industry.