Search Results for author: Kayode Olaleye

Found 4 papers, 1 papers with code

YFACC: A Yorùbá speech-image dataset for cross-lingual keyword localisation through visual grounding

no code implementations10 Oct 2022 Kayode Olaleye, Dan Oneata, Herman Kamper

We collect and release a new single-speaker dataset of audio captions for 6k Flickr images in Yor\`ub\'a -- a real low-resource language spoken in Nigeria.

Visual Grounding

Keyword localisation in untranscribed speech using visually grounded speech models

1 code implementation2 Feb 2022 Kayode Olaleye, Dan Oneata, Herman Kamper

Masked-based localisation gives some of the best reported localisation scores from a VGS model, with an accuracy of 57% when the system knows that a keyword occurs in an utterance and need to predict its location.

Keyword Spotting TAG

Attention-Based Keyword Localisation in Speech using Visual Grounding

no code implementations16 Jun 2021 Kayode Olaleye, Herman Kamper

Visually grounded speech models learn from images paired with spoken captions.

Visual Grounding

Towards localisation of keywords in speech using weak supervision

no code implementations14 Dec 2020 Kayode Olaleye, Benjamin van Niekerk, Herman Kamper

Of the two forms of supervision, the visually trained model performs worse than the BoW-trained model.

Cannot find the paper you are looking for? You can Submit a new open access paper.