Search Results for author: Stéphane Dupont

Found 15 papers, 6 papers with code

Improved Soccer Action Spotting using both Audio and Video Streams

no code implementations9 Nov 2020 Bastien Vanderplaetse, Stéphane Dupont

Action spotting and classification are the tasks that consist in finding the temporal anchors of events in a video and determine which event they are.

Action Classification Action Spotting +2

Modulated Fusion using Transformer for Linguistic-Acoustic Emotion Recognition

1 code implementation5 Oct 2020 Jean-Benoit Delbrouck, Noé Tits, Stéphane Dupont

This paper aims to bring a new lightweight yet powerful solution for the task of Emotion Recognition and Sentiment Analysis.

Emotion Recognition Sentiment Analysis

AVECL-UMONS database for audio-visual event classification and localization

no code implementations2 Oct 2020 Mathilde Brousmiche, Stéphane Dupont, Jean Rouat

We introduce the AVECL-UMons dataset for audio-visual event classification and localization in the context of office environments.

Classification General Classification

Modulated Self-attention Convolutional Network for VQA

no code implementations8 Oct 2019 Jean-Benoit Delbrouck, Antoine Maiorca, Nathan Hubens, Stéphane Dupont

As new data-sets for real-world visual reasoning and compositional question answering are emerging, it might be needed to use the visual feature extraction as a end-to-end process during training.

Question Answering Visual Question Answering +1

Adversarial reconstruction for Multi-modal Machine Translation

no code implementations7 Oct 2019 Jean-Benoit Delbrouck, Stéphane Dupont

Even with the growing interest in problems at the intersection of Computer Vision and Natural Language, grounding (i. e. identifying) the components of a structured description in an image still remains a challenging task.

Machine Translation

Object-oriented Targets for Visual Navigation using Rich Semantic Representations

no code implementations22 Nov 2018 Jean-Benoit Delbrouck, Stéphane Dupont

When searching for an object humans navigate through a scene using semantic information and spatial relationships.

Visual Navigation

Bringing back simplicity and lightliness into neural image captioning

no code implementations15 Oct 2018 Jean-Benoit Delbrouck, Stéphane Dupont

So far, the goal has been to maximize scores on automated metric and to do so, one has to come up with a plurality of new modules and techniques.

Image Captioning Machine Translation

UMONS Submission for WMT18 Multimodal Translation Task

1 code implementation15 Oct 2018 Jean-Benoit Delbrouck, Stéphane Dupont

This paper describes the UMONS solution for the Multimodal Machine Translation Task presented at the third conference on machine translation (WMT18).

Image Captioning Multimodal Machine Translation

Proceedings of eNTERFACE 2015 Workshop on Intelligent Interfaces

no code implementations19 Jan 2018 Matei Mancas, Christian Frisson, Joëlle Tilmanne, Nicolas D'Alessandro, Petr Barborka, Furkan Bayansar, Francisco Bernard, Rebecca Fiebrink, Alexis Heloir, Edgar Hemery, Sohaib Laraba, Alexis Moinet, Fabrizio Nunnari, Thierry Ravet, Loïc Reboursière, Alvaro Sarasua, Mickaël Tits, Noé Tits, François Zajéga, Paolo Alborno, Ksenia Kolykhalova, Emma Frid, Damiano Malafronte, Lisanne Huis in't Veld, Hüseyin Cakmak, Kevin El Haddad, Nicolas Riche, Julien Leroy, Pierre Marighetto, Bekir Berker Türker, Hossein Khaki, Roberto Pulisci, Emer Gilmartin, Fasih Haider, Kübra Cengiz, Martin Sulir, Ilaria Torre, Shabbir Marzban, Ramazan Yazıcı, Furkan Burak Bâgcı, Vedat Gazi Kılı, Hilal Sezer, Sena Büsra Yenge, Charles-Alexandre Delestage, Sylvie Leleu-Merviel, Muriel Meyer-Chemenska, Daniel Schmitt, Willy Yvart, Stéphane Dupont, Ozan Can Altiok, Aysegül Bumin, Ceren Dikmen, Ivan Giangreco, Silvan Heller, Emre Külah, Gueorgui Pironkov, Luca Rossetto, Yusuf Sahillioglu, Heiko Schuldt, Omar Seddati, Yusuf Setinkaya, Metin Sezgin, Claudiu Tanase, Emre Toyan, Sean Wood, Doguhan Yeke, Françcois Rocca, Pierre-Henri De Deken, Alessandra Bandrabur, Fabien Grisard, Axel Jean-Caurant, Vincent Courboulay, Radhwan Ben Madhkour, Ambroise Moreau

The 11th Summer Workshop on Multimodal Interfaces eNTERFACE 2015 was hosted by the Numediart Institute of Creative Technologies of the University of Mons from August 10th to September 2015.

Modulating and attending the source image during encoding improves Multimodal Translation

1 code implementation9 Dec 2017 Jean-Benoit Delbrouck, Stéphane Dupont

We propose a new and fully end-to-end approach for multimodal translation where the source text encoder modulates the entire visual input processing using conditional batch normalization, in order to compute the most informative image features for our task.

Visually Grounded Word Embeddings and Richer Visual Features for Improving Multimodal Neural Machine Translation

no code implementations4 Jul 2017 Jean-Benoit Delbrouck, Stéphane Dupont, Omar Seddati

In Multimodal Neural Machine Translation (MNMT), a neural model generates a translated sentence that describes an image, given the image itself and one source descriptions in English.

Machine Translation Object Detection +1

An empirical study on the effectiveness of images in Multimodal Neural Machine Translation

no code implementations EMNLP 2017 Jean-Benoit Delbrouck, Stéphane Dupont

In state-of-the-art Neural Machine Translation (NMT), an attention mechanism is used during decoding to enhance the translation.

Machine Translation

Cannot find the paper you are looking for? You can Submit a new open access paper.