Search Results for author: Julien Chaumond

Found 9 papers, 6 papers with code

AfroDigits: A Community-Driven Spoken Digit Dataset for African Languages

no code implementations22 Mar 2023 Chris Chinenye Emezue, Sanchit Gandhi, Lewis Tunstall, Abubakar Abid, Josh Meyer, Quentin Lhoest, Pete Allen, Patrick von Platen, Douwe Kiela, Yacine Jernite, Julien Chaumond, Merve Noyan, Omar Sanseviero

The advancement of speech technologies has been remarkable, yet its integration with African languages remains limited due to the scarcity of African speech corpora.

DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter

31 code implementations NeurIPS 2019 Victor Sanh, Lysandre Debut, Julien Chaumond, Thomas Wolf

As Transfer Learning from large-scale pre-trained models becomes more prevalent in Natural Language Processing (NLP), operating these large models in on-the-edge and/or under constrained computational training or inference budgets remains challenging.

Hate Speech Detection Knowledge Distillation +8

TransferTransfo: A Transfer Learning Approach for Neural Network Based Conversational Agents

21 code implementations23 Jan 2019 Thomas Wolf, Victor Sanh, Julien Chaumond, Clement Delangue

We introduce a new approach to generative data-driven dialogue systems (e. g. chatbots) called TransferTransfo which is a combination of a Transfer learning based training scheme and a high-capacity Transformer model.

Ranked #3 on Dialogue Generation on Persona-Chat (using extra training data)

Dialogue Generation Information Retrieval +2

Continuous Learning in a Hierarchical Multiscale Neural Network

no code implementations ACL 2018 Thomas Wolf, Julien Chaumond, Clement Delangue

We reformulate the problem of encoding a multi-scale representation of a sequence in a language model by casting it in a continuous learning framework.

Language Modelling Meta-Learning

Meta-Learning a Dynamical Language Model

no code implementations28 Mar 2018 Thomas Wolf, Julien Chaumond, Clement Delangue

We consider the task of word-level language modeling and study the possibility of combining hidden-states-based short-term representations with medium-term representations encoded in dynamical weights of a language model.

Language Modelling Meta-Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.