no code implementations • EMNLP 2021 • Nathaniel Berger, Stefan Riezler, Sebastian Ebert, Artem Sokolov
Recently more attention has been given to adversarial attacks on neural networks for natural language processing (NLP).
no code implementations • EMNLP 2021 • Andrea Schioppa, David Vilar, Artem Sokolov, Katja Filippova
Fine-grained control of machine translation (MT) outputs along multiple attributes is critical for many modern MT applications and is a requirement for gaining users’ trust.
no code implementations • Findings (EMNLP) 2021 • Julia Kreutzer, David Vilar, Artem Sokolov
Training data for machine translation (MT) is often sourced from a multitude of large corpora that are multi-faceted in nature, e. g. containing contents from multiple domains or different levels of quality or complexity.
no code implementations • 27 Feb 2023 • Irina Bejan, Artem Sokolov, Katja Filippova
Increasingly larger datasets have become a standard ingredient to advancing the state of the art in NLP.
1 code implementation • 23 Dec 2022 • Vladimir Kondratenko, Artem Sokolov, Nikolay Karpov, Oleg Kutuzov, Nikita Savushkin, Fyodor Minkin
We present a new data set for speech emotion recognition (SER) tasks called Dusha.
Speech Emotion Recognition
Speech Emotion Recognition in Russian
1 code implementation • 6 Dec 2021 • Andrea Schioppa, Polina Zablotskaia, David Vilar, Artem Sokolov
We address efficient calculation of influence functions for tracking predictions back to the training data.
no code implementations • 13 Oct 2021 • Julia Kreutzer, David Vilar, Artem Sokolov
Training data for machine translation (MT) is often sourced from a multitude of large corpora that are multi-faceted in nature, e. g. containing contents from multiple domains or different levels of quality or complexity.
no code implementations • 16 Sep 2021 • Nathaniel Berger, Stefan Riezler, Artem Sokolov, Sebastian Ebert
Recently more attention has been given to adversarial attacks on neural networks for natural language processing (NLP).
no code implementations • 9 Sep 2021 • Luca Hormann, Artem Sokolov
We apply imitation learning (IL) to tackle the NMT exposure bias problem with error-correcting oracles, and evaluate an SMT lattice-based oracle which, despite its excellent performance in an unconstrained oracle translation task, turned out to be too pruned and idiosyncratic to serve as the oracle for IL.
2 code implementations • 21 Aug 2021 • Denis Schapiro, Clarence Yapp, Artem Sokolov, Sheila M. Reynolds, Yu-An Chen, Damir Sudar, Yubin Xie, Jeremy L. Muhlich, Raquel Arias-Camison, Sarah Arena, Adam J. Taylor, Milen Nikolov, Madison Tyler, Jia-Ren Lin, Erik A. Burlingame, Human Tumor Atlas Network, Young H. Chang, Samouil L Farhi, Vésteinn Thorsson, Nithya Venkatamohan, Julia L. Drewes, Dana Pe'er, David A. Gutman, Markus D. Herrmann, Nils Gehlenborg, Peter Bankhead, Joseph T. Roland, John M. Herndon, Michael P. Snyder, Michael Angelo, Garry Nolan, Jason R. Swedlow, Nikolaus Schultz, Daniel T. Merrick, Sarah A. Mazzilli, Ethan Cerami, Scott J. Rodig, Sandro Santagata, Peter K. Sorger
The imminent release of tissue atlases combining multi-channel microscopy with single cell sequencing and other omics data from normal and diseased specimens creates an urgent need for data and metadata standards that guide data deposition, curation and release.
no code implementations • 20 Apr 2021 • Shahin Amiriparian, Artem Sokolov, Ilhan Aslan, Lukas Christ, Maurice Gerczuk, Tobias Hübner, Dmitry Lamanov, Manuel Milling, Sandra Ottl, Ilya Poduremennykh, Evgeniy Shuranov, Björn W. Schuller
Text encodings from automatic speech recognition (ASR) transcripts and audio representations have shown promise in speech emotion recognition (SER) ever since.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+3
no code implementations • 5 Apr 2021 • Vasiliy Kuzmin, Fyodor Kravchenko, Artem Sokolov, Jie Geng
In this paper, we describe the work that we have done to participate in Task1 of the ConferencingSpeech2021 challenge.
no code implementations • 22 Mar 2021 • Julia Kreutzer, Isaac Caswell, Lisa Wang, Ahsan Wahab, Daan van Esch, Nasanbayar Ulzii-Orshikh, Allahsera Tapo, Nishant Subramani, Artem Sokolov, Claytone Sikasote, Monang Setyawan, Supheakmungkol Sarin, Sokhar Samb, Benoît Sagot, Clara Rivera, Annette Rios, Isabel Papadimitriou, Salomey Osei, Pedro Ortiz Suarez, Iroro Orife, Kelechi Ogueji, Andre Niyongabo Rubungo, Toan Q. Nguyen, Mathias Müller, André Müller, Shamsuddeen Hassan Muhammad, Nanda Muhammad, Ayanda Mnyakeni, Jamshidbek Mirzakhalov, Tapiwanashe Matangira, Colin Leong, Nze Lawson, Sneha Kudugunta, Yacine Jernite, Mathias Jenny, Orhan Firat, Bonaventure F. P. Dossou, Sakhile Dlamini, Nisansa de Silva, Sakine Çabuk Ballı, Stella Biderman, Alessia Battisti, Ahmed Baruwa, Ankur Bapna, Pallavi Baljekar, Israel Abebe Azime, Ayodele Awokoya, Duygu Ataman, Orevaoghene Ahia, Oghenefego Ahia, Sweta Agrawal, Mofetoluwa Adeyemi
With the success of large-scale pre-training and multilingual modeling in Natural Language Processing (NLP), recent years have seen a proliferation of large, web-mined text datasets covering hundreds of languages.
1 code implementation • 2 Jun 2020 • Mayumi Ohta, Nathaniel Berger, Artem Sokolov, Stefan Riezler
Interest in stochastic zeroth-order (SZO) methods has recently been revived in black-box optimization scenarios such as adversarial black-box attacks to deep neural networks.
no code implementations • IWSLT (EMNLP) 2018 • Julia Kreutzer, Artem Sokolov
Most modern neural machine translation (NMT) systems rely on presegmented inputs.
no code implementations • 12 Jun 2018 • Artem Sokolov, Julian Hitschler, Mayumi Ohta, Stefan Riezler
Stochastic zeroth-order (SZO), or gradient-free, optimization allows to optimize arbitrary functions by relying only on function evaluations under parameter perturbations, however, the iteration complexity of SZO methods suffers a factor proportional to the dimensionality of the perturbed function.
16 code implementations • 15 Dec 2017 • Felix Hieber, Tobias Domhan, Michael Denkowski, David Vilar, Artem Sokolov, Ann Clifton, Matt Post
Written in Python and built on MXNet, the toolkit offers scalable training and inference for the three most prominent encoder-decoder architectures: attentional recurrent neural networks, self-attentional transformers, and fully convolutional networks.
no code implementations • EMNLP 2017 • Carolin Lawrence, Artem Sokolov, Stefan Riezler
The goal of counterfactual learning for statistical machine translation (SMT) is to optimize a target SMT system from logged data that consist of user feedback to translations that were predicted by another, historic SMT system.
no code implementations • 28 Jul 2017 • Carolin Lawrence, Artem Sokolov, Stefan Riezler
The goal of counterfactual learning for statistical machine translation (SMT) is to optimize a target SMT system from logged data that consist of user feedback to translations that were predicted by another, historic SMT system.
no code implementations • WS 2017 • Artem Sokolov, Julia Kreutzer, Kellen Sunderland, Pavel Danchenko, Witold Szymaniak, Hagen Fürstenau, Stefan Riezler
We introduce and describe the results of a novel shared task on bandit learning for machine translation.
1 code implementation • ACL 2017 • Julia Kreutzer, Artem Sokolov, Stefan Riezler
Bandit structured prediction describes a stochastic optimization framework where learning is performed from partial feedback.
1 code implementation • NeurIPS 2016 • Artem Sokolov, Julia Kreutzer, Christopher Lo, Stefan Riezler
Stochastic structured prediction under bandit feedback follows a learning protocol where on each of a sequence of iterations, the learner receives an input, predicts an output structure, and receives partial feedback in form of a task loss evaluation of the predicted structure.
no code implementations • 18 Jan 2016 • Artem Sokolov, Stefan Riezler, Tanguy Urvoy
We present an application to discriminative reranking in Statistical Machine Translation (SMT) where the learning algorithm only has access to a 1-BLEU loss evaluation of a predicted translation instead of obtaining a gold standard reference translation.