no code implementations • Findings (ACL) 2022 • Markus Freitag, David Vilar, David Grangier, Colin Cherry, George Foster
In this work we propose a method for training MT systems to achieve a more natural style, i. e. mirroring the style of text originally written in the target language.
no code implementations • 14 Nov 2022 • Maartje ter Hoeve, David Grangier, Natalie Schluter
The central bottleneck for low-resource NLP is typically regarded to be the quantity of accessible data, overlooking the contribution of data quality.
2 code implementations • 7 Sep 2022 • Zalán Borsos, Raphaël Marinier, Damien Vincent, Eugene Kharitonov, Olivier Pietquin, Matt Sharifi, Olivier Teboul, David Grangier, Marco Tagliasacchi, Neil Zeghidour
We introduce AudioLM, a framework for high-quality audio generation with long-term consistency.
1 code implementation • ICLR 2022 • Rachid Riad, Olivier Teboul, David Grangier, Neil Zeghidour
In particular, we show that introducing our layer into a ResNet-18 architecture allows keeping consistent high performance on CIFAR10, CIFAR100 and ImageNet even when training starts from poor random stride configurations.
no code implementations • 17 Nov 2021 • Markus Freitag, David Grangier, Qijun Tan, Bowen Liang
In Neural Machine Translation, it is typically assumed that the sentence with the highest estimated probability should also be the translation with the highest quality as measured by humans.
no code implementations • ACL 2022 • David Grangier, Dan Iter
This work connects language model adaptation with concepts of machine learning theory.
no code implementations • 15 Sep 2021 • Dan Iter, David Grangier
Domain adaptation of neural networks commonly relies on three training phases: pretraining, selected data training and then fine tuning.
1 code implementation • ICLR 2021 • Lucio M. Dery, Yann Dauphin, David Grangier
In this case, careful consideration is needed to select tasks and model parameterizations such that updates from the auxiliary tasks actually help the primary task.
no code implementations • NAACL 2022 • Kelly Marchisio, Markus Freitag, David Grangier
Modern unsupervised machine translation (MT) systems reach reasonable translation quality under clean and controlled data conditions.
no code implementations • 28 May 2021 • Neil Zeghidour, Olivier Teboul, David Grangier
Our neural algorithm presents the diarization task as an iterative process: it repeatedly builds a representation for each speaker before predicting the voice activity of each speaker conditioned on the extracted representations.
3 code implementations • 29 Apr 2021 • Markus Freitag, George Foster, David Grangier, Viresh Ratnakar, Qijun Tan, Wolfgang Macherey
Human evaluation of modern high-quality machine translation systems is a difficult problem, and there is increasing evidence that inadequate evaluation procedures can lead to erroneous conclusions.
no code implementations • 21 Oct 2020 • Aaqib Saeed, David Grangier, Olivier Pietquin, Neil Zeghidour
We propose CHARM, a method for training a single neural network across inconsistent input channels.
2 code implementations • 21 Oct 2020 • Aaqib Saeed, David Grangier, Neil Zeghidour
We introduce COLA, a self-supervised pre-training approach for learning a general-purpose representation of audio.
Ranked #4 on
Spoken Command Recognition
on Speech Command v2
1 code implementation • WMT (EMNLP) 2020 • Markus Freitag, George Foster, David Grangier, Colin Cherry
When used in place of original references, the paraphrased versions produce metric scores that correlate better with human judgment.
no code implementations • ACL 2020 • Parker Riley, Isaac Caswell, Markus Freitag, David Grangier
Machine translation has an undesirable propensity to produce {``}translationese{''} artifacts, which can lead to higher BLEU scores while being liked less by human raters.
1 code implementation • ACL 2020 • Daphne Ippolito, David Grangier, Douglas Eck, Chris Callison-Burch
We propose a sentence-level language model which selects the next sentence in a story from a finite set of fluent alternatives.
2 code implementations • EMNLP 2020 • Markus Freitag, David Grangier, Isaac Caswell
The quality of automatic metrics for machine translation has been increasingly called into question, especially for high-quality systems.
2 code implementations • 12 Mar 2020 • Aurko Roy, Mohammad Saffar, Ashish Vaswani, David Grangier
This work builds upon two lines of research: it combines the modeling flexibility of prior work on content-based sparse attention with the efficiency gains from approaches based on local, temporal sparse attention.
Ranked #5 on
Image Generation
on ImageNet 64x64
(Bits per dim metric)
no code implementations • 20 Feb 2020 • Neil Zeghidour, David Grangier
Wavesplit infers a set of source representations via clustering, which addresses the fundamental permutation problem of separation.
Ranked #3 on
Speech Separation
on WHAMR!
no code implementations • 10 Nov 2019 • Parker Riley, Isaac Caswell, Markus Freitag, David Grangier
Machine translation has an undesirable propensity to produce "translationese" artifacts, which can lead to higher BLEU scores while being liked less by human raters.
3 code implementations • ACL 2019 • Angela Fan, Yacine Jernite, Ethan Perez, David Grangier, Jason Weston, Michael Auli
We introduce the first large-scale corpus for long-form question answering, a task requiring elaborate and in-depth answers to open-ended questions.
no code implementations • WS 2019 • Isaac Caswell, Ciprian Chelba, David Grangier
Recent work in Neural Machine Translation (NMT) has shown significant quality gains from noised-beam decoding during back-translation, a method to generate synthetic parallel data.
no code implementations • WS 2019 • Daphne Ippolito, David Grangier, Chris Callison-Burch, Douglas Eck
Story infilling involves predicting words to go into a missing span from a story.
no code implementations • ACL 2019 • Aurko Roy, David Grangier
We compare with MT-based approaches on paraphrase identification, generation, and training augmentation.
6 code implementations • NAACL 2019 • Myle Ott, Sergey Edunov, Alexei Baevski, Angela Fan, Sam Gross, Nathan Ng, David Grangier, Michael Auli
fairseq is an open-source sequence modeling toolkit that allows researchers and developers to train custom models for translation, summarization, language modeling, and other text generation tasks.
1 code implementation • 21 Jan 2019 • Dario Pavllo, Christoph Feichtenhofer, Michael Auli, David Grangier
Previous work on predicting or generating 3D human pose sequences regresses either joint rotations or joint positions.
9 code implementations • CVPR 2019 • Dario Pavllo, Christoph Feichtenhofer, David Grangier, Michael Auli
We start with predicted 2D keypoints for unlabeled video, then estimate 3D poses and finally back-project to the input 2D keypoints.
Ranked #11 on
Weakly-supervised 3D Human Pose Estimation
on Human3.6M
(Number of Frames Per View metric)
Monocular 3D Human Pose Estimation
Weakly-supervised 3D Human Pose Estimation
3 code implementations • EMNLP 2018 • Sergey Edunov, Myle Ott, Michael Auli, David Grangier
An effective method to improve neural machine translation with monolingual data is to augment the parallel training corpus with back-translations of target language sentences.
Ranked #2 on
Machine Translation
on WMT2014 English-German
(using extra training data)
no code implementations • NAACL 2018 • David Grangier, Michael Auli
We also evaluate our model for paraphrasing through a user study.
5 code implementations • WS 2018 • Myle Ott, Sergey Edunov, David Grangier, Michael Auli
Sequence to sequence learning models still require several days to reach state of the art performance on large benchmark datasets using a single machine.
Ranked #12 on
Machine Translation
on WMT2014 English-French
1 code implementation • 16 May 2018 • Dario Pavllo, David Grangier, Michael Auli
Deep learning for predicting or generating 3D human pose sequences is an active research area.
1 code implementation • ICML 2018 • Myle Ott, Michael Auli, David Grangier, Marc'Aurelio Ranzato
We propose tools and metrics to assess how uncertainty in the data is captured by the model distribution and how it affects search strategies that generate translations.
1 code implementation • NAACL 2018 • Sergey Edunov, Myle Ott, Michael Auli, David Grangier, Marc'Aurelio Ranzato
There has been much recent work on training neural attention models at the sequence-level using either reinforcement learning-style methods or by optimizing the beam.
Ranked #4 on
Machine Translation
on IWSLT2015 German-English
no code implementations • WS 2018 • Angela Fan, David Grangier, Michael Auli
Current models for document summarization disregard user preferences such as the desired length, style, the entities that the user might be interested in, or how much of the document the user has already read.
no code implementations • HLT 2018 • David Grangier, Michael Auli
We also evaluate our model for paraphrasing through a user study.
36 code implementations • ICML 2017 • Jonas Gehring, Michael Auli, David Grangier, Denis Yarats, Yann N. Dauphin
The prevalent approach to sequence to sequence learning maps an input sequence to a variable length output sequence via recurrent neural networks.
10 code implementations • ICML 2017 • Yann N. Dauphin, Angela Fan, Michael Auli, David Grangier
The pre-dominant approach to language modeling to date is based on recurrent neural networks.
Ranked #18 on
Language Modelling
on One Billion Word
2 code implementations • ACL 2017 • Jonas Gehring, Michael Auli, David Grangier, Yann N. Dauphin
The prevalent approach to neural machine translation relies on bi-directional LSTMs to encode the source sentence.
Ranked #7 on
Machine Translation
on IWSLT2015 German-English
no code implementations • 20 Oct 2016 • Roman Novak, Michael Auli, David Grangier
Existing machine translation decoding algorithms generate translations in a strictly monotonic fashion and never revisit previous decisions.
no code implementations • 1 Oct 2016 • Gurvan L'Hostis, David Grangier, Michael Auli
Classical translation models constrain the space of possible outputs by selecting a subset of translation rules based on the input sentence.
12 code implementations • ICML 2017 • Edouard Grave, Armand Joulin, Moustapha Cissé, David Grangier, Hervé Jégou
We propose an approximate strategy to efficiently train neural network based language models over very large vocabularies.
no code implementations • 24 Jun 2016 • Camille Jandot, Patrice Simard, Max Chickering, David Grangier, Jina Suh
In text classification, dictionaries can be used to define human-comprehensible features.
3 code implementations • EMNLP 2016 • Remi Lebret, David Grangier, Michael Auli
This paper introduces a neural model for concept-to-text generation that scales to large, rich domains.
Ranked #4 on
Table-to-Text Generation
on WikiBio
2 code implementations • ACL 2016 • Welin Chen, David Grangier, Michael Auli
Training neural network language models over large vocabularies is still computationally very costly compared to count-based models such as Kneser-Ney.
no code implementations • 17 Nov 2015 • Yann N. Dauphin, David Grangier
Contrary to a classical neural network, a belief network can predict more than the expected value of the output $Y$ given the input $X$.
no code implementations • 16 Sep 2014 • Patrice Simard, David Chickering, Aparna Lakshmiratan, Denis Charles, Leon Bottou, Carlos Garcia Jurado Suarez, David Grangier, Saleema Amershi, Johan Verwey, Jina Suh
Based on the machine's output, the teacher can revise the definition of the task or make it more precise.
no code implementations • NeurIPS 2010 • Samy Bengio, Jason Weston, David Grangier
Multi-class classification becomes challenging at test time when the number of classes is very large and testing against every possible class can become computationally infeasible.
no code implementations • NeurIPS 2010 • David Grangier, Iain Melvin
Our proposal maps (feature, value) pairs into an embedding space and then non-linearly combines the set of embedded vectors.
no code implementations • NeurIPS 2009 • Bing Bai, Jason Weston, David Grangier, Ronan Collobert, Kunihiko Sadamasa, Yanjun Qi, Corinna Cortes, Mehryar Mohri
We present a class of nonlinear (polynomial) models that are discriminatively trained to directly map from the word content in a query-document or document-document pair to a ranking score.