no code implementations • SemEval (NAACL) 2022 • Nidhir Bhavsar, Rishikesh Devanathan, Aakash Bhatnagar, Muskaan Singh, Petr Motlicek, Tirthankar Ghosal
This work represents the system proposed by team Innovators for SemEval 2022 Task 8: Multilingual News Article Similarity.
no code implementations • ACL (WAT) 2021 • Shantipriya Parida, Subhadarshi Panda, Ketan Kotwal, Amulya Ratna Dash, Satya Ranjan Dash, Yashvardhan Sharma, Petr Motlicek, Ondřej Bojar
Our submission tops in English→Malayalam Multimodal translation task (text-only translation, and Malayalam caption), and ranks second-best in English→Hindi Multimodal translation task (text-only translation, and Hindi caption).
1 code implementation • LTEDI (ACL) 2022 • Muskaan Singh, Petr Motlicek
The increased expansion of abusive content on social media platforms negatively affects online users.
no code implementations • LTEDI (ACL) 2022 • Muskaan Singh, Petr Motlicek
Given social media postings in English, the submitted system classify the signs of depression into three labels, namely “not depressed,” “moderately depressed,” and “severely depressed.” Our best model is ranked 3^{rd} position with 0. 54% accuracy .
1 code implementation • IWSLT (ACL) 2022 • Aakash Bhatnagar, Nidhir Bhavsar, Muskaan Singh, Petr Motlicek
In this paper, we propose a hierarchical approach to generate isometric translation on MUST-C dataset, we achieve a BERTscore of 0. 85, a length ratio of 1. 087, a BLEU score of 42. 3, and a length range of 51. 03%.
no code implementations • IWSLT 2016 • Alexandros Lazaridis, Ivan Himawan, Petr Motlicek, Iosif Mporas, Philip N. Garner
We experiment with three different scenarios using, i) French, as a source language uncorrelated to the target language, ii) Ukrainian, as a source language correlated to the target one and finally iii) English as a source language uncorrelated to the target language using a relatively large amount of data in respect to the other two scenarios.
no code implementations • AACL (WAT) 2020 • Shantipriya Parida, Petr Motlicek, Amulya Ratna Dash, Satya Ranjan Dash, Debasish Kumar Mallick, Satya Prakash Biswal, Priyanka Pattnaik, Biranchi Narayan Nayak, Ondřej Bojar
We have participated in the English-Hindi Multimodal task and Indic task.
1 code implementation • LTEDI (ACL) 2022 • Deepanshu Khanna, Muskaan Singh, Petr Motlicek
With the increase of users on social media platforms, manipulating or provoking masses of people has become a piece of cake.
no code implementations • LEGAL (LREC) 2022 • Mickaël Rigault, Claudia Cevenini, Khalid Choukri, Martin Kocour, Karel Veselý, Igor Szoke, Petr Motlicek, Juan Pablo Zuluaga-Gomez, Alexander Blatt, Dietrich Klakow, Allan Tart, Pavel Kolčárek, Jan Černocký
In this paper the authors detail the various legal and ethical issues faced during the ATCO2 project.
no code implementations • COLING (CreativeSumm) 2022 • Aditya Upadhyay, Nidhir Bhavsar, Aakash Bhatnagar, Muskaan Singh, Petr Motlicek
This paper documents our approach for the Creative-Summ 2022 shared task for Automatic Summarization of Creative Writing.
no code implementations • ICON 2020 • Maël Fabien, Esau Villatoro-Tello, Petr Motlicek, Shantipriya Parida
Identifying the author of a given text can be useful in historical literature, plagiarism detection, or police investigations.
no code implementations • ICON 2020 • Shantipriya Parida, Esau Villatoro-Tello, Sajit Kumar, Maël Fabien, Petr Motlicek
Language detection is considered a difficult task especially for similar languages, varieties, and dialects.
no code implementations • NAACL (AmericasNLP) 2021 • Shantipriya Parida, Subhadarshi Panda, Amulya Dash, Esau Villatoro-Tello, A. Seza Doğruöz, Rosa M. Ortega-Mendoza, Amadeo Hernández, Yashvardhan Sharma, Petr Motlicek
This paper describes the team (“Tamalli”)’s submission to AmericasNLP2021 shared task on Open Machine Translation for low resource South American languages.
no code implementations • MMTLRL (RANLP) 2021 • Shantipriya Parida, Subhadarshi Panda, Satya Prakash Biswal, Ketan Kotwal, Arghyadeep Sen, Satya Ranjan Dash, Petr Motlicek
Multimodal Machine Translation (MMT) systems utilize additional information from other modalities beyond text to improve the quality of machine translation (MT).
1 code implementation • LTEDI (ACL) 2022 • Muskaan Singh, Petr Motlicek
It reflects the belief to achieve an objective, discovers a new path, or become motivated to formulate pathways. In this paper we classify given a social media post, hope speech or not hope speech, using ensembled voting of BERT, ERNIE 2. 0 and RoBERTa for English language with 0. 54 macro F1-score (2^{st} rank).
no code implementations • 3 Feb 2025 • Yacouba Kaloga, Shashi Kumar, Petr Motlicek, Ina Kodrasi
Based on the SOTD, we propose Optimal Temporal Transport Classification (OTTC) loss for ASR and contrast its behavior with CTC.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+1
no code implementations • 6 Nov 2024 • Shashi Kumar, Iuliia Thorbecke, Sergio Burdisso, Esaú Villatoro-Tello, Manjunath K E, Kadri Hacioğlu, Pradeep Rangappa, Petr Motlicek, Aravind Ganapathiraju, Andreas Stolcke
Recent research has demonstrated that training a linear connector between speech foundation encoders and large language models (LLMs) enables this architecture to achieve strong ASR capabilities.
1 code implementation • 24 Oct 2024 • Sergio Burdisso, Srikanth Madikeri, Petr Motlicek
Efficiently deriving structured workflows from unannotated dialogs remains an underexplored and formidable challenge in computational linguistics.
1 code implementation • 23 Oct 2024 • Dairazalia Sánchez-Cortés, Sergio Burdisso, Esaú Villatoro-Tello, Petr Motlicek
Our experiments, targeting two challenging bias descriptors, factual reporting and political bias, showed a significant performance improvement at the source media level.
no code implementations • 20 Sep 2024 • Iuliia Thorbecke, Juan Zuluaga-Gomez, Esaú Villatoro-Tello, Shashi Kumar, Pradeep Rangappa, Sergio Burdisso, Petr Motlicek, Karthik Pandia, Aravind Ganapathiraju
The training of automatic speech recognition (ASR) with little to no supervised data remains an open question.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+2
no code implementations • 20 Sep 2024 • Iuliia Thorbecke, Juan Zuluaga-Gomez, Esaú Villatoro-Tello, Andres Carofilis, Shashi Kumar, Petr Motlicek, Karthik Pandia, Aravind Ganapathiraju
The language model is used as an additional support to keyword biasing when the language model is combined with bias entities in a single context graph to take care of the overall performance.
no code implementations • 5 Jul 2024 • Shashi Kumar, Srikanth Madikeri, Juan Zuluaga-Gomez, Iuliia Thorbecke, Esaú Villatoro-Tello, Sergio Burdisso, Petr Motlicek, Karthik Pandia, Aravind Ganapathiraju
In addition to ASR, we conduct experiments on 3 different tasks: speaker change detection, endpointing, and NER.
no code implementations • 5 Jul 2024 • Shashi Kumar, Srikanth Madikeri, Juan Zuluaga-Gomez, Esaú Villatoro-Tello, Iuliia Thorbecke, Petr Motlicek, Manjunath K E, Aravind Ganapathiraju
Our experiments on the AMI dataset reveal that the XLSR-Transducer achieves 4% absolute WER improvement over Whisper large-v2 and 8% over a Zipformer transducer model trained from scratch.
1 code implementation • 22 Apr 2024 • Sergio Burdisso, Ernesto Reyes-Ramírez, Esaú Villatoro-Tello, Fernando Sánchez-Vega, Pastor López-Monroy, Petr Motlicek
Finally, to highlight the magnitude of this bias, we achieve a 0. 90 F1 score by intentionally exploiting it, the highest result reported to date on this dataset using only textual information.
1 code implementation • 15 Apr 2024 • Sergio Burdisso, Dairazalia Sánchez-Cortés, Esaú Villatoro-Tello, Petr Motlicek
Contrary to previous research, our proposed approach models the problem as the estimation of a reliability degree, and not a reliability label, based on how all the news media sources interact with each other on the Web.
1 code implementation • 3 Jul 2023 • Sergio Burdisso, Esaú Villatoro-Tello, Srikanth Madikeri, Petr Motlicek
We propose a simple approach for weighting self-connecting edges in a Graph Convolutional Network (GCN) and show its impact on depression detection from transcribed clinical interviews.
2 code implementations • 29 May 2023 • Florian Mai, Juan Zuluaga-Gomez, Titouan Parcollet, Petr Motlicek
In particular, multi-head HyperConformer achieves comparable or higher recognition performance while being more efficient than Conformer in terms of inference speed, memory, parameter count, and available training data.
no code implementations • 2 May 2023 • Juan Zuluaga-Gomez, Iuliia Nigmatulina, Amrutha Prasad, Petr Motlicek, Driss Khalil, Srikanth Madikeri, Allan Tart, Igor Szoke, Vincent Lenders, Mickael Rigault, Khalid Choukri
This paper explores the lessons learned from the ATCO2 project, a project that aimed to develop a unique platform to collect and preprocess large amounts of ATC data from airspace in real time.
no code implementations • 16 Apr 2023 • Juan Zuluaga-Gomez, Amrutha Prasad, Iuliia Nigmatulina, Petr Motlicek, Matthias Kleinert
The overall pipeline is composed of the following submodules: (i) automatic speech recognition (ASR) system that transforms audio into a sequence of words; (ii) high-level air traffic control (ATC) related entity parser that understands the transcribed voice communication; and (iii) a text-to-speech submodule that generates a spoken utterance that resembles a pilot based on the situation of the dialogue.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+2
1 code implementation • 16 Dec 2022 • Esaú Villatoro-Tello, Srikanth Madikeri, Juan Zuluaga-Gomez, Bidisha Sharma, Seyyed Saeed Sarfjoo, Iuliia Nigmatulina, Petr Motlicek, Alexei V. Ivanov, Aravind Ganapathiraju
In this paper, we perform an exhaustive evaluation of different representations to address the intent classification problem in a Spoken Language Understanding (SLU) setup.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+5
no code implementations • 14 Dec 2022 • Amrutha Prasad, Juan Zuluaga-Gomez, Petr Motlicek, Saeed Sarfjoo, Iuliia Nigmatulina, Karel Vesely
The system understands the voice communications issued by the ATCo, and, in turn, it generates a spoken prompt that follows the pilot's phraseology to the initial communication.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+1
3 code implementations • 8 Nov 2022 • Juan Zuluaga-Gomez, Karel Veselý, Igor Szöke, Alexander Blatt, Petr Motlicek, Martin Kocour, Mickael Rigault, Khalid Choukri, Amrutha Prasad, Seyyed Saeed Sarfjoo, Iuliia Nigmatulina, Claudia Cevenini, Pavel Kolčárek, Allan Tart, Jan Černocký, Dietrich Klakow
In this paper, we introduce the ATCO2 corpus, a dataset that aims at fostering research on the challenging ATC field, which has lagged behind due to lack of annotated data.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+6
1 code implementation • 8 Sep 2022 • Martin Fajcik, Muskaan Singh, Juan Zuluaga-Gomez, Esaú Villatoro-Tello, Sergio Burdisso, Petr Motlicek, Pavel Smrz
In this paper, we describe our shared task submissions for Subtask 2 in CASE-2022, Event Causality Identification with Casual News Corpus.
1 code implementation • 8 Sep 2022 • Sergio Burdisso, Juan Zuluaga-Gomez, Esau Villatoro-Tello, Martin Fajcik, Muskaan Singh, Pavel Smrz, Petr Motlicek
In this paper, we describe our participation in the subtask 1 of CASE-2022, Event Causality Identification with Casual News Corpus.
1 code implementation • 28 Jul 2022 • Martin Fajcik, Petr Motlicek, Pavel Smrz
We propose to disentangle the per-evidence relevance probability and its contribution to the final veracity probability in an interpretable way -- the final veracity probability is proportional to a linear ensemble of per-evidence relevance probabilities.
2 code implementations • 31 Mar 2022 • Juan Zuluaga-Gomez, Amrutha Prasad, Iuliia Nigmatulina, Saeed Sarfjoo, Petr Motlicek, Matthias Kleinert, Hartmut Helmke, Oliver Ohneiser, Qingran Zhan
Recent work on self-supervised pre-training focus on leveraging large-scale unlabeled speech data to build robust end-to-end (E2E) acoustic models (AM) that can be later fine-tuned on downstream tasks e. g., automatic speech recognition (ASR).
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+1
no code implementations • 8 Feb 2022 • Iuliia Nigmatulina, Juan Zuluaga-Gomez, Amrutha Prasad, Seyyed Saeed Sarfjoo, Petr Motlicek
Automatic Speech Recognition (ASR), as the assistance of speech communication between pilots and air-traffic controllers, can significantly reduce the complexity of the task and increase the reliability of transmitted information.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+5
2 code implementations • 12 Oct 2021 • Juan Zuluaga-Gomez, Seyyed Saeed Sarfjoo, Amrutha Prasad, Iuliia Nigmatulina, Petr Motlicek, Karel Ondrej, Oliver Ohneiser, Hartmut Helmke
We propose a system that combines SAD and a BERT model to perform speaker change detection and speaker role detection (SRD) by chunking ASR transcripts, i. e., SD with a defined number of speakers together with SRD.
no code implementations • 27 Aug 2021 • Iuliia Nigmatulina, Rudolf Braun, Juan Zuluaga-Gomez, Petr Motlicek
Automatic Speech Recognition (ASR) can be used as the assistance of speech communication between pilots and air-traffic controllers.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+4
no code implementations • 27 Aug 2021 • Amrutha Prasad, Juan Zuluaga-Gomez, Petr Motlicek, Saeed Sarfjoo, Iuliia Nigmatulina, Oliver Ohneiser, Hartmut Helmke
In this work, we propose to (1) automatically segment the ATCO and pilot data based on an intuitive approach exploiting ASR transcripts and (2) subsequently consider an automatic recognition of ATCOs' and pilots' voice as two separate tasks.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+1
2 code implementations • 16 Jul 2021 • Rudolf A. Braun, Srikanth Madikeri, Petr Motlicek
A common problem for automatic speech recognition systems is how to recognize words that they did not see during training.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+3
no code implementations • 8 Apr 2021 • Juan Zuluaga-Gomez, Iuliia Nigmatulina, Amrutha Prasad, Petr Motlicek, Karel Veselý, Martin Kocour, Igor Szöke
Results show that `unseen domains' (e. g. data from airports not present in the supervised training data) are further aided by contextual SSL when compared to standalone SSL.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+2
1 code implementation • 4 Nov 2020 • Yihui Fu, Zhuoyuan Yao, Weipeng He, Jian Wu, Xiong Wang, Zhanheng Yang, Shimin Zhang, Lei Xie, DongYan Huang, Hui Bu, Petr Motlicek, Jean-Marc Odobez
In this challenge, we open source a sizable speech, keyword, echo and noise corpus for promoting data-driven methods, particularly deep-learning approaches on KWS and SSL.
Sound Audio and Speech Processing
1 code implementation • 7 Oct 2020 • Srikanth Madikeri, Sibo Tong, Juan Zuluaga-Gomez, Apoorv Vyas, Petr Motlicek, Hervé Bourlard
We present a simple wrapper that is useful to train acoustic models in PyTorch using Kaldi's LF-MMI training framework.
Audio and Speech Processing Sound
3 code implementations • 18 Jun 2020 • Juan Zuluaga-Gomez, Petr Motlicek, Qingran Zhan, Karel Vesely, Rudolf Braun
We demonstrate that the cross-accent flaws due to speakers' accents are minimized due to the amount of data, making the system feasible for ATC environments.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+1
no code implementations • LREC 2020 • Shantipriya Parida, Satya Ranjan Dash, Ond{\v{r}}ej Bojar, Petr Motlicek, Priyanka Pattnaik, Debasish Kumar Mallick
The preparation of parallel corpora is a challenging task, particularly for languages that suffer from under-representation in the digital world.
no code implementations • IJCNLP 2019 • Shantipriya Parida, Petr Motlicek
We propose an iterative data augmentation approach which uses synthetic data along with the real summarization data for the German language.
no code implementations • WS 2019 • Shantipriya Parida, Ond{\v{r}}ej Bojar, Petr Motlicek
This paper describes the Idiap submission to WAT 2019 for the English-Hindi Multi-Modal Translation Task.
no code implementations • 15 Sep 2019 • Mary Ellen Foster, Bart Craenen, Amol Deshmukh, Oliver Lemon, Emanuele Bastianelli, Christian Dondrup, Ioannis Papaioannou, Andrea Vanzo, Jean-Marc Odobez, Olivier Canévet, Yuanzhouhan Cao, Weipeng He, Angel Martínez-González, Petr Motlicek, Rémy Siegfried, Rachid Alami, Kathleen Belhassein, Guilhem Buisan, Aurélie Clodic, Amandine Mayima, Yoan Sallami, Guillaume Sarthou, Phani-Teja Singamaneni, Jules Waldhart, Alexandre Mazel, Maxime Caniot, Marketta Niemelä, Päivi Heikkilä, Hanna Lammi, Antti Tammela
In the EU-funded MuMMER project, we have developed a social robot designed to interact naturally and flexibly with users in public spaces such as a shopping mall.
no code implementations • 8 Aug 2019 • Subhadeep Dey, Petr Motlicek, Trung Bui, Franck Dernoncourt
In this paper, we explore various approaches for semi supervised learning in an end to end automatic speech recognition (ASR) framework.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+2
1 code implementation • 30 Nov 2017 • Weipeng He, Petr Motlicek, Jean-Marc Odobez
We propose to use neural networks for simultaneous detection and localization of multiple sound sources in human-robot interaction.
no code implementations • LREC 2014 • Volha Petukhova, Martin Gropp, Dietrich Klakow, Gregor Eigner, Mario Topf, Stefan Srb, Petr Motlicek, Blaise Potard, John Dines, Olivier Deroo, Ronny Egeler, Uwe Meinz, Steffen Liersch, Anna Schmidt
We first start with human-human Wizard of Oz experiments to collect human-human data in order to model natural human dialogue behaviour, for better understanding of phenomena of human interactions and predicting interlocutors actions, and then replace the human Wizard by an increasingly advanced dialogue system, using evaluation data for system improvement.
no code implementations • JEPTALNRECITAL 2012 • Gw{\'e}nol{\'e} Lecorv{\'e}, John Dines, Thomas Hain, Petr Motlicek