no code implementations • RepL4NLP (ACL) 2022 • Adnen Abdessaied, Ekta Sood, Andreas Bulling
We propose the Video Language Co-Attention Network (VLCN) – a novel memory-enhanced model for Video Question Answering (VideoQA).
no code implementations • 3 Mar 2025 • Adnen Abdessaied, Anna Rohrbach, Marcus Rohrbach, Andreas Bulling
We present V$^2$Dial - a novel expert-based model specifically geared towards simultaneously handling image and video input data for multimodal conversational tasks.
no code implementations • 2 Jul 2024 • Adnen Abdessaied, Lei Shi, Andreas Bulling
Then, it predicts the missing underlying structure of the selected constituents of each modality by learning local latent graphs using a novel multi-modal graph structure learning method.
no code implementations • 21 May 2024 • Matteo Bortoletto, Constantin Ruhdorfer, Adnen Abdessaied, Lei Shi, Andreas Bulling
This finding calls for a deeper understanding of the role of ToM in CPA and beyond, as well as new methods for modelling and evaluating mental states in computational collaborative agents.
no code implementations • 20 Feb 2024 • Adnen Abdessaied, Manuel von Hochmeister, Andreas Bulling
OLViT addresses these challenges by maintaining a global dialog state based on the output of an Object State Tracker (OST) and a Language State Tracker (LST): while the OST attends to the most important objects within the video, the LST keeps track of the most important linguistic co-references to previous dialog turns.
no code implementations • 25 Oct 2023 • Adnen Abdessaied, Lei Shi, Andreas Bulling
We propose $\mathbb{VD}$-$\mathbb{GR}$ - a novel visual dialog model that combines pre-trained language models (LMs) with graph neural networks (GNNs).
1 code implementation • COLING 2022 • Adnen Abdessaied, Mihai Bâce, Andreas Bulling
We propose Neuro-Symbolic Visual Dialog (NSVD) -the first method to combine deep learning and symbolic program execution for multi-round visually-grounded reasoning.