1 code implementation • ACL 2022 • Jean-Benoit Delbrouck, Khaled Saab, Maya Varma, Sabri Eyuboglu, Pierre Chambon, Jared Dunnmon, Juan Zambrano, Akshay Chaudhari, Curtis Langlotz
There is a growing need to model interactions between data modalities (e. g., vision, language) — both to improve AI predictions on existing tasks and to enable new applications.
no code implementations • NAACL (TeachingNLP) 2021 • Jean-Benoit Delbrouck
MiniVQA is a Jupyter notebook to build a tailored VQA competition for your students.
1 code implementation • NAACL (BioNLP) 2021 • Jean-Benoit Delbrouck, Cassie Zhang, Daniel Rubin
This paper describes the solution of the QIAI lab sent to the Radiology Report Summarization (RRS) challenge at MEDIQA 2021.
1 code implementation • 14 Sep 2023 • Dave Van Veen, Cara Van Uden, Louis Blankemeier, Jean-Benoit Delbrouck, Asad Aali, Christian Bluethgen, Anuj Pareek, Malgorzata Polacin, William Collins, Neera Ahuja, Curtis P. Langlotz, Jason Hom, Sergios Gatidis, John Pauly, Akshay S. Chaudhari
Sifting through vast textual data and summarizing key information imposes a substantial burden on how clinicians allocate their time.
1 code implementation • 22 Aug 2023 • Maya Varma, Jean-Benoit Delbrouck, Sarah Hooper, Akshay Chaudhari, Curtis Langlotz
The first key contribution of this work is to demonstrate through systematic evaluations that as the pairwise complexity of the training dataset increases, standard VLMs struggle to learn region-attribute relationships, exhibiting performance degradations of up to 37% on retrieval tasks.
1 code implementation • 2 May 2023 • Dave Van Veen, Cara Van Uden, Maayane Attias, Anuj Pareek, Christian Bluethgen, Malgorzata Polacin, Wah Chiu, Jean-Benoit Delbrouck, Juan Manuel Zambrano Chaves, Curtis P. Langlotz, Akshay S. Chaudhari, John Pauly
We systematically investigate lightweight strategies to adapt large language models (LLMs) for the task of radiology report summarization (RRS).
no code implementations • 23 Nov 2022 • Pierre Chambon, Christian Bluethgen, Jean-Benoit Delbrouck, Rogier van der Sluijs, Małgorzata Połacin, Juan Manuel Zambrano Chaves, Tanishq Mathew Abraham, Shivanshu Purohit, Curtis P. Langlotz, Akshay Chaudhari
We present evidence that the resulting model (RoentGen) is able to create visually convincing, diverse synthetic CXR images, and that the output can be controlled to a new extent by using free-form text prompts including radiology-specific language.
1 code implementation • 15 Nov 2022 • Zhihong Chen, Maya Varma, Xiang Wan, Curtis Langlotz, Jean-Benoit Delbrouck
We then conduct extensive experiments to evaluate the performance of models both within and across modality-anatomy pairs in MIMIC-RRS.
1 code implementation • 21 Oct 2022 • Jean-Benoit Delbrouck, Pierre Chambon, Christian Bluethgen, Emily Tsai, Omar Almusa, Curtis P. Langlotz
To overcome this limitation, we propose a new method, the RadGraph reward, to further improve the factual completeness and correctness of generated radiology reports.
2 code implementations • ICLR 2022 • Sabri Eyuboglu, Maya Varma, Khaled Saab, Jean-Benoit Delbrouck, Christopher Lee-Messer, Jared Dunnmon, James Zou, Christopher Ré
In this work, we address these challenges by first designing a principled evaluation framework that enables a quantitative comparison of SDMs across 1, 235 slice discovery settings in three input domains (natural images, medical images, and time-series data).
1 code implementation • EMNLP (nlpbt) 2020 • Jean-Benoit Delbrouck, Noé Tits, Stéphane Dupont
This paper aims to bring a new lightweight yet powerful solution for the task of Emotion Recognition and Sentiment Analysis.
1 code implementation • WS 2020 • Jean-Benoit Delbrouck, Noé Tits, Mathilde Brousmiche, Stéphane Dupont
Understanding expressed sentiment and emotions are two crucial factors in human multimodal language.
Ranked #4 on
Multimodal Sentiment Analysis
on CMU-MOSEI
(using extra training data)
1 code implementation • 31 Oct 2019 • Jean-Benoit Delbrouck, Bastien Vanderplaetse, Stéphane Dupont
Recently, generative adversarial networks (GAN) have gathered a lot of interest.
no code implementations • 8 Oct 2019 • Jean-Benoit Delbrouck, Antoine Maiorca, Nathan Hubens, Stéphane Dupont
As new data-sets for real-world visual reasoning and compositional question answering are emerging, it might be needed to use the visual feature extraction as a end-to-end process during training.
no code implementations • 7 Oct 2019 • Jean-Benoit Delbrouck, Stéphane Dupont
Even with the growing interest in problems at the intersection of Computer Vision and Natural Language, grounding (i. e. identifying) the components of a structured description in an image still remains a challenging task.
no code implementations • 22 Nov 2018 • Jean-Benoit Delbrouck, Stéphane Dupont
When searching for an object humans navigate through a scene using semantic information and spatial relationships.
1 code implementation • 15 Oct 2018 • Jean-Benoit Delbrouck, Stéphane Dupont
This paper describes the UMONS solution for the Multimodal Machine Translation Task presented at the third conference on machine translation (WMT18).
no code implementations • 15 Oct 2018 • Jean-Benoit Delbrouck, Stéphane Dupont
So far, the goal has been to maximize scores on automated metric and to do so, one has to come up with a plurality of new modules and techniques.
1 code implementation • 3 May 2018 • Jean-Benoit Delbrouck
This paper describes the UMONS solution for the OMG-Emotion Challenge.
1 code implementation • 9 Dec 2017 • Jean-Benoit Delbrouck, Stéphane Dupont
We propose a new and fully end-to-end approach for multimodal translation where the source text encoder modulates the entire visual input processing using conditional batch normalization, in order to compute the most informative image features for our task.
no code implementations • 4 Jul 2017 • Jean-Benoit Delbrouck, Stéphane Dupont, Omar Seddati
In Multimodal Neural Machine Translation (MNMT), a neural model generates a translated sentence that describes an image, given the image itself and one source descriptions in English.
no code implementations • EMNLP 2017 • Jean-Benoit Delbrouck, Stéphane Dupont
In state-of-the-art Neural Machine Translation (NMT), an attention mechanism is used during decoding to enhance the translation.
no code implementations • 23 Mar 2017 • Jean-Benoit Delbrouck, Stephane Dupont
Recently, the effectiveness of the attention mechanism has also been explored for multimodal tasks, where it becomes possible to focus both on sentence parts and image regions.