Search Results for author: Jean-Benoit Delbrouck

Found 15 papers, 7 papers with code

QIAI at MEDIQA 2021: Multimodal Radiology Report Summarization

1 code implementation NAACL (BioNLP) 2021 Jean-Benoit Delbrouck, Cassie Zhang, Daniel Rubin

This paper describes the solution of the QIAI lab sent to the Radiology Report Summarization (RRS) challenge at MEDIQA 2021.

Modulated Self-attention Convolutional Network for VQA

no code implementations8 Oct 2019 Jean-Benoit Delbrouck, Antoine Maiorca, Nathan Hubens, Stéphane Dupont

As new data-sets for real-world visual reasoning and compositional question answering are emerging, it might be needed to use the visual feature extraction as a end-to-end process during training.

Question Answering Visual Question Answering +1

Adversarial reconstruction for Multi-modal Machine Translation

no code implementations7 Oct 2019 Jean-Benoit Delbrouck, Stéphane Dupont

Even with the growing interest in problems at the intersection of Computer Vision and Natural Language, grounding (i. e. identifying) the components of a structured description in an image still remains a challenging task.

Machine Translation Translation

Object-oriented Targets for Visual Navigation using Rich Semantic Representations

no code implementations22 Nov 2018 Jean-Benoit Delbrouck, Stéphane Dupont

When searching for an object humans navigate through a scene using semantic information and spatial relationships.

Visual Navigation

Bringing back simplicity and lightliness into neural image captioning

no code implementations15 Oct 2018 Jean-Benoit Delbrouck, Stéphane Dupont

So far, the goal has been to maximize scores on automated metric and to do so, one has to come up with a plurality of new modules and techniques.

Image Captioning Machine Translation +1

UMONS Submission for WMT18 Multimodal Translation Task

1 code implementation15 Oct 2018 Jean-Benoit Delbrouck, Stéphane Dupont

This paper describes the UMONS solution for the Multimodal Machine Translation Task presented at the third conference on machine translation (WMT18).

Image Captioning Multimodal Machine Translation +1

Transformer for Emotion Recognition

1 code implementation3 May 2018 Jean-Benoit Delbrouck

This paper describes the UMONS solution for the OMG-Emotion Challenge.

Emotion Recognition

Modulating and attending the source image during encoding improves Multimodal Translation

1 code implementation9 Dec 2017 Jean-Benoit Delbrouck, Stéphane Dupont

We propose a new and fully end-to-end approach for multimodal translation where the source text encoder modulates the entire visual input processing using conditional batch normalization, in order to compute the most informative image features for our task.

Translation

Visually Grounded Word Embeddings and Richer Visual Features for Improving Multimodal Neural Machine Translation

no code implementations4 Jul 2017 Jean-Benoit Delbrouck, Stéphane Dupont, Omar Seddati

In Multimodal Neural Machine Translation (MNMT), a neural model generates a translated sentence that describes an image, given the image itself and one source descriptions in English.

Machine Translation Object Detection +2

Multimodal Compact Bilinear Pooling for Multimodal Neural Machine Translation

no code implementations23 Mar 2017 Jean-Benoit Delbrouck, Stephane Dupont

Recently, the effectiveness of the attention mechanism has also been explored for multimodal tasks, where it becomes possible to focus both on sentence parts and image regions.

Machine Translation Question Answering +2

Cannot find the paper you are looking for? You can Submit a new open access paper.