Search Results for author: Laura Hanu

Found 4 papers, 1 papers with code

Language as the Medium: Multimodal Video Classification through text only

no code implementations19 Sep 2023 Laura Hanu, Anita L. Verő, James Thewlis

Despite an exciting new wave of multimodal machine learning models, current approaches still struggle to interpret the complex contextual relationships between the different modalities present in videos.

Action Recognition Video Classification +1

VTC: Improving Video-Text Retrieval with User Comments

1 code implementation19 Oct 2022 Laura Hanu, James Thewlis, Yuki M. Asano, Christian Rupprecht

In this paper, we a) introduce a new dataset of videos, titles and comments; b) present an attention-based mechanism that allows the model to learn from sometimes irrelevant data such as comments; c) show that by using comments, our method is able to learn better, more contextualised, representations for image, video and audio representations.

Representation Learning Retrieval +3

Learning Context-Adapted Video-Text Retrieval by Attending to User Comments

no code implementations29 Sep 2021 Laura Hanu, Yuki M Asano, James Thewlis, Christian Rupprecht

Learning strong representations for multi-modal retrieval is an important problem for many applications, such as recommendation and search.

Retrieval Text Retrieval +1

Evaluating Robustness to Context-Sensitive Feature Perturbations of Different Granularities

no code implementations29 Jan 2020 Isaac Dunn, Laura Hanu, Hadrien Pouget, Daniel Kroening, Tom Melham

We cannot guarantee that training datasets are representative of the distribution of inputs that will be encountered during deployment.

Autonomous Driving

Cannot find the paper you are looking for? You can Submit a new open access paper.