no code implementations • EAMT 2022 • Mattia Di Gangi, Nick Rossenbach, Alejandro Pérez, Parnia Bahar, Eugen Beck, Patrick Wilken, Evgeny Matusov
The revoicing usually comes with a changed script, mostly in a different language, and the revoicing should reproduce the original emotions, coherent with the body language, and lip synchronized.
no code implementations • IWSLT (EMNLP) 2018 • Evgeny Matusov, Patrick Wilken, Parnia Bahar, Julian Schamper, Pavel Golik, Albert Zeyer, Joan Albert Silvestre-Cerda, Adrià Martínez-Villaronga, Hendrik Pesch, Jan-Thorsten Peter
This work describes AppTek’s speech translation pipeline that includes strong state-of-the-art automatic speech recognition (ASR) and neural machine translation (NMT) components.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+4
no code implementations • ACL (IWSLT) 2021 • Parnia Bahar, Patrick Wilken, Mattia A. Di Gangi, Evgeny Matusov
This paper describes the offline and simultaneous speech translation systems developed at AppTek for IWSLT 2021.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+4
no code implementations • IWSLT (ACL) 2022 • Patrick Wilken, Panayota Georgakopoulou, Evgeny Matusov
This paper addresses the problem of evaluating the quality of automatically generated subtitles, which includes not only the quality of the machine-transcribed or translated speech, but also the quality of line segmentation and subtitle timing.
no code implementations • IWSLT (ACL) 2022 • Patrick Wilken, Evgeny Matusov
To participate in the Isometric Spoken Language Translation Task of the IWSLT 2022 evaluation, constrained condition, AppTek developed neural Transformer-based systems for English-to-German with various mechanisms of length control, ranging from source-side and target-side pseudo-tokens to encoding of remaining length in characters that replaces positional encoding.
no code implementations • 12 May 2022 • Patrick Wilken, Evgeny Matusov
To participate in the Isometric Spoken Language Translation Task of the IWSLT 2022 evaluation, constrained condition, AppTek developed neural Transformer-based systems for English-to-German with various mechanisms of length control, ranging from source-side and target-side pseudo-tokens to encoding of remaining length in characters that replaces positional encoding.
no code implementations • 11 May 2022 • Patrick Wilken, Panayota Georgakopoulou, Evgeny Matusov
This paper addresses the problem of evaluating the quality of automatically generated subtitles, which includes not only the quality of the machine-transcribed or translated speech, but also the quality of line segmentation and subtitle timing.
no code implementations • WS 2020 • Parnia Bahar, Patrick Wilken, Tamer Alkhouli, Andreas Guta, Pavel Golik, Evgeny Matusov, Christian Herold
AppTek and RWTH Aachen University team together to participate in the offline and simultaneous speech translation tracks of IWSLT 2020.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+4
no code implementations • WS 2020 • Patrick Wilken, Tamer Alkhouli, Evgeny Matusov, Pavel Golik
In simultaneous machine translation, the objective is to determine when to produce a partial translation given a continuous stream of source words, with a trade-off between latency and quality.
no code implementations • 9 Oct 2019 • Patrick Wilken, Evgeny Matusov
In this work, we explore the usefulness of target factors in neural machine translation (NMT) beyond their original purpose of predicting word lemmas and their inflections, as proposed by Garc\`ia-Mart\`inez et al., 2016.
no code implementations • WS 2019 • Evgeny Matusov, Patrick Wilken, Yota Georgakopoulou
In this work, we customized a neural machine translation system for translation of subtitles in the domain of entertainment.
no code implementations • WS 2018 • Jos{\'e} G. Camargo de Souza, Michael Kozielski, Prashant Mathur, Ernie Chang, Marco Guerini, Matteo Negri, Marco Turchi, Evgeny Matusov
The setting requires the generation process to be fast and the generated title to be both human-readable and concise.
no code implementations • ACL 2018 • Pavel Petrushkov, Shahram Khadivi, Evgeny Matusov
We empirically investigate learning from partial feedback in neural machine translation (NMT), when partial feedback is collected by asking users to highlight a correct chunk of a translation.
no code implementations • NAACL 2018 • Julia Kreutzer, Shahram Khadivi, Evgeny Matusov, Stefan Riezler
We present the first real-world application of methods for improving neural machine translation (NMT) with human reinforcement, based on explicit and implicit user feedback collected on the eBay e-commerce platform.
no code implementations • EMNLP 2017 • Leonard Dahlmann, Evgeny Matusov, Pavel Petrushkov, Shahram Khadivi
In this paper, we introduce a hybrid search for attention-based neural machine translation (NMT).
no code implementations • MTSummit 2017 • Shahram Khadivi, Patrick Wilken, Leonard Dahlmann, Evgeny Matusov
In this paper, we discuss different methods which use meta information and richer context that may accompany source language input to improve machine translation quality.
no code implementations • EACL 2017 • Iacer Calixto, Daniel Stein, Evgeny Matusov, Pintu Lohar, Sheila Castilho, Andy Way
We evaluate our models quantitatively using BLEU and TER and find that (i) additional synthetic data has a general positive impact on text-only and multi-modal NMT models, and that (ii) using a multi-modal NMT model for re-ranking n-best lists improves TER significantly across different n-best list sizes.
no code implementations • WS 2017 • Iacer Calixto, Daniel Stein, Evgeny Matusov, Sheila Castilho, Andy Way
Nonetheless, human evaluators ranked translations from a multi-modal NMT model as better than those of a text-only NMT over 88{\%} of the time, which suggests that images do help NMT in this use-case.
1 code implementation • AMTA 2016 • Wenhu Chen, Evgeny Matusov, Shahram Khadivi, Jan-Thorsten Peter
In this paper, we propose an effective way for biasing the attention mechanism of a sequence-to-sequence neural machine translation (NMT) model towards the well-studied statistical word alignment models.