Search Results for author: Alberto Poncelas

Found 25 papers, 1 papers with code

On Machine Translation of User Reviews

no code implementations RANLP 2021 Maja Popović, Alberto Poncelas, Marija Brkic, Andy Way

This work investigates neural machine translation (NMT) systems for translating English user reviews into Croatian and Serbian, two similar morphologically complex languages.

Machine Translation NMT +1

Neural Machine Translation between similar South-Slavic languages

no code implementations WMT (EMNLP) 2020 Maja Popović, Alberto Poncelas

This paper describes the ADAPT-DCU machine translation systems built for the WMT 2020 shared task on Similar Language Translation.

Machine Translation NMT +1

Neural Machine Translation for translating into Croatian and Serbian

no code implementations VarDial (COLING) 2020 Maja Popović, Alberto Poncelas, Marija Brkic, Andy Way

Furthermore, translation performance from English is much better than from German, partly because German is morphologically more complex and partly because the corpus consists mostly of parallel human translations instead of original text and its human translation.

Machine Translation NMT +1

The Impact of Indirect Machine Translation on Sentiment Classification

no code implementations AMTA 2020 Alberto Poncelas, Pintu Lohar, Andy Way, James Hadley

Furthermore, as performing a direct translation is not always possible, we explore the performance of automatic classifiers on sentences that have been translated using a pivot MT system.

Classification General Classification +4

Selecting Backtranslated Data from Multiple Sources for Improved Neural Machine Translation

no code implementations ACL 2020 Xabier Soto, Dimitar Shterionov, Alberto Poncelas, Andy Way

Machine translation (MT) has benefited from using synthetic training data originating from translating monolingual corpora, a technique known as backtranslation.

Machine Translation Translation

Multiple Segmentations of Thai Sentences for Neural Machine Translation

no code implementations LREC 2020 Alberto Poncelas, Wichaya Pidchamook, Chao-Hong Liu, James Hadley, Andy Way

Thai is a low-resource language, so it is often the case that data is not available in sufficient quantities to train an Neural Machine Translation (NMT) model which perform to a high level of quality.

Machine Translation NMT +2

A Tool for Facilitating OCR Postediting in Historical Documents

1 code implementation LREC 2020 Alberto Poncelas, Mohammad Aboomar, Jan Buts, James Hadley, Andy Way

This paper reports on a tool built for postediting the output of Tesseract, more specifically for correcting common errors in digitized historical documents.

Language Modelling Optical Character Recognition +1

Selecting Artificially-Generated Sentences for Fine-Tuning Neural Machine Translation

no code implementations WS 2019 Alberto Poncelas, Andy Way

Neural Machine Translation (NMT) models tend to achieve best performance when larger sets of parallel sentences are provided for training.

Machine Translation NMT +2

Combining SMT and NMT Back-Translated Data for Efficient NMT

no code implementations9 Sep 2019 Alberto Poncelas, Maja Popovic, Dimitar Shterionov, Gideon Maillette de Buy Wenniger, Andy Way

Neural Machine Translation (NMT) models achieve their best performance when large sets of parallel data are used for training.

Machine Translation NMT +1

ABI Neural Ensemble Model for Gender Prediction Adapt Bar-Ilan Submission for the CLIN29 Shared Task on Gender Prediction

no code implementations23 Feb 2019 Eva Vanmassenhove, Amit Moryossef, Alberto Poncelas, Andy Way, Dimitar Shterionov

In contradiction with the results described in previous comparable shared tasks, our neural models performed better than our best traditional approaches with our best feature set-up.

Gender Prediction

Data Selection with Feature Decay Algorithms Using an Approximated Target Side

no code implementations IWSLT (EMNLP) 2018 Alberto Poncelas, Gideon Maillette de Buy Wenniger, Andy Way

A limitation of these methods to date is that using the source-side test set does not by itself guarantee that sentences are selected with correct translations, or translations that are suitable given the test-set domain.

Machine Translation NMT +2

Extracting In-domain Training Corpora for Neural Machine Translation Using Data Selection Methods

no code implementations WS 2018 Catarina Cruz Silva, Chao-Hong Liu, Alberto Poncelas, Andy Way

Data selection is a process used in selecting a subset of parallel data for the training of machine translation (MT) systems, so that 1) resources for training might be reduced, 2) trained models could perform better than those trained with the whole corpus, and/or 3) trained models are more tailored to specific domains.

Machine Translation NMT +1

Understanding Meanings in Multilingual Customer Feedback

no code implementations5 Jun 2018 Chao-Hong Liu, Declan Groves, Akira Hayakawa, Alberto Poncelas, Qun Liu

Understanding and being able to react to customer feedback is the most fundamental task in providing good customer service.

General Classification

Investigating Backtranslation in Neural Machine Translation

no code implementations17 Apr 2018 Alberto Poncelas, Dimitar Shterionov, Andy Way, Gideon Maillette de Buy Wenniger, Peyman Passban

A prerequisite for training corpus-based machine translation (MT) systems -- either Statistical MT (SMT) or Neural MT (NMT) -- is the availability of high-quality parallel data.

Machine Translation NMT +1

Cannot find the paper you are looking for? You can Submit a new open access paper.