Search Results for author: Matthieu Labeau

Found 28 papers, 6 papers with code

Code-switched inspired losses for spoken dialog representations

no code implementations EMNLP 2021 Pierre Colombo, Emile Chapuis, Matthieu Labeau, Chloé Clavel

Spoken dialogue systems need to be able to handle both multiple languages and multilinguality inside a conversation (e. g in case of code-switching).

Retrieval Spoken Dialogue Systems

EZCAT: an Easy Conversation Annotation Tool

no code implementations LREC 2022 Gaël Guibon, Luce Lefeuvre, Matthieu Labeau, Chloé Clavel

We also present our first usage of EZCAT along with our annotation schema we used to annotate confidential customer service conversations.

Management

Polysemy in Spoken Conversations and Written Texts

1 code implementation LREC 2022 Aina Garí Soler, Matthieu Labeau, Chloé Clavel

Our discourses are full of potential lexical ambiguities, due in part to the pervasive use of words having multiple senses.

The Impact of Word Splitting on the Semantic Content of Contextualized Word Representations

1 code implementation22 Feb 2024 Aina Garí Soler, Matthieu Labeau, Chloé Clavel

When deriving contextualized word representations from language models, a decision needs to be made on how to obtain one for out-of-vocabulary (OOV) words that are segmented into subwords.

Semantic Similarity Semantic Textual Similarity

Exploiting Edge Features in Graphs with Fused Network Gromov-Wasserstein Distance

no code implementations28 Sep 2023 Junjie Yang, Matthieu Labeau, Florence d'Alché-Buc

Pairwise comparison of graphs is key to many applications in Machine learning ranging from clustering, kernel-based classification/regression and more recently supervised graph prediction.

Improving Multimodal fusion via Mutual Dependency Maximisation

no code implementations EMNLP 2021 Pierre Colombo, Emile Chapuis, Matthieu Labeau, Chloe Clavel

We demonstrate that our new penalties lead to a consistent improvement (up to $4. 3$ on accuracy) across a large variety of state-of-the-art models on two well-known sentiment analysis datasets: \texttt{CMU-MOSI} and \texttt{CMU-MOSEI}.

Multimodal Sentiment Analysis

Code-switched inspired losses for generic spoken dialog representations

no code implementations27 Aug 2021 Emile Chapuis, Pierre Colombo, Matthieu Labeau, Chloe Clavel

Spoken dialog systems need to be able to handle both multiple languages and multilinguality inside a conversation (\textit{e. g} in case of code-switching).

Retrieval

The importance of fillers for text representations of speech transcripts

no code implementations EMNLP 2020 Tanvi Dinkar, Pierre Colombo, Matthieu Labeau, Chloé Clavel

While being an essential component of spoken language, fillers (e. g."um" or "uh") often remain overlooked in Spoken Language Understanding (SLU) tasks.

Spoken Language Understanding

Compositional Languages Emerge in a Neural Iterated Learning Model

1 code implementation ICLR 2020 Yi Ren, Shangmin Guo, Matthieu Labeau, Shay B. Cohen, Simon Kirby

The principle of compositionality, which enables natural language to represent complex concepts via a structured combination of simpler ones, allows us to convey an open-ended set of messages using a limited vocabulary.

Experimenting with Power Divergences for Language Modeling

no code implementations IJCNLP 2019 Matthieu Labeau, Shay B. Cohen

In this paper, we experiment with several families (alpha, beta and gamma) of power divergences, generalized from the KL divergence, for learning language models with an objective different than standard MLE.

Language Modelling

Learning with Noise-Contrastive Estimation: Easing training by learning to scale

no code implementations COLING 2018 Matthieu Labeau, Alex Allauzen, re

Noise-Contrastive Estimation (NCE) is a learning criterion that is regularly used to train neural language models in place of Maximum Likelihood Estimation, since it avoids the computational bottleneck caused by the output softmax.

Language Modelling Machine Translation +1

Algorithmes \`a base d'\'echantillonage pour l'entra\^\inement de mod\`eles de langue neuronaux (Here the title in English)

no code implementations JEPTALNRECITAL 2018 Matthieu Labeau, Alex Allauzen, re

L{'}estimation contrastive bruit{\'e}e (NCE) et l{'}{\'e}chantillonage par importance (IS) sont des proc{\'e}dures d{'}entra{\^\i}nement bas{\'e}es sur l{'}{\'e}chantillonage, que l{'}on utilise habituellement {\`a} la place de l{'}estimation du maximum de vraisemblance (MLE) pour {\'e}viter le calcul du softmax lorsque l{'}on entra{\^\i}ne des mod{\`e}les de langue neuronaux.

Adaptation au domaine pour l'analyse morpho-syntaxique (Domain Adaptation for PoS tagging)

no code implementations JEPTALNRECITAL 2017 {\'E}l{\'e}onor Bartenlian, Margot Lacour, Matthieu Labeau, Alex Allauzen, re, Guillaume Wisniewski, Fran{\c{c}}ois Yvon

Ce travail cherche {\`a} comprendre pourquoi les performances d{'}un analyseur morpho-syntaxiques chutent fortement lorsque celui-ci est utilis{\'e} sur des donn{\'e}es hors domaine.

Domain Adaptation POS +1

Repr\'esentations continues d\'eriv\'ees des caract\`eres pour un mod\`ele de langue neuronal \`a vocabulaire ouvert (Opening the vocabulary of neural language models with character-level word representations)

no code implementations JEPTALNRECITAL 2017 Matthieu Labeau, Alex Allauzen, re

Les repr{\'e}sentations continues des mots sont calcul{\'e}es {\`a} la vol{\'e}e {\`a} partir des caract{\`e}res les composant, gr{\`a}ce {\`a} une couche convolutionnelle suivie d{'}une couche de regroupement (pooling).

SENTER

An experimental analysis of Noise-Contrastive Estimation: the noise distribution matters

no code implementations EACL 2017 Matthieu Labeau, Alex Allauzen, re

Noise Contrastive Estimation (NCE) is a learning procedure that is regularly used to train neural language models, since it avoids the computational bottleneck caused by the output softmax.

Language Modelling Machine Translation +1

Cannot find the paper you are looking for? You can Submit a new open access paper.