Search Results for author: Olga Slizovskaia

Found 8 papers, 4 papers with code

Locate This, Not That: Class-Conditioned Sound Event DOA Estimation

no code implementations8 Mar 2022 Olga Slizovskaia, Gordon Wichern, Zhong-Qiu Wang, Jonathan Le Roux

Existing systems for sound event localization and detection (SELD) typically operate by estimating a source location for all classes at every time instant.

Sound Event Localization and Detection

Solos: A Dataset for Audio-Visual Music Analysis

1 code implementation14 Jun 2020 Juan F. Montesinos, Olga Slizovskaia, Gloria Haro

In this paper, we present a new dataset of music performance videos which can be used for training machine learning methods for multiple tasks such as audio-visual blind source separation and localization, cross-modal correspondences, cross-modal generation and, in general, any audio-visual selfsupervised task.

Audio Source Separation Audio-Visual Synchronization +1 Audio and Speech Processing Databases Sound

Conditioned Source Separation for Music Instrument Performances

1 code implementation8 Apr 2020 Olga Slizovskaia, Gloria Haro, Emilia Gómez

In music source separation, the number of sources may vary for each piece and some of the sources may belong to the same family of instruments, thus sharing timbral characteristics and making the sources more correlated.

Sound Audio and Speech Processing

A Case Study of Deep-Learned Activations via Hand-Crafted Audio Features

no code implementations3 Jul 2019 Olga Slizovskaia, Emilia Gómez, Gloria Haro

We also propose a technique for measuring the similarity between activation maps and audio features which typically presented in the form of a matrix, such as chromagrams or spectrograms.

End-to-End Sound Source Separation Conditioned On Instrument Labels

no code implementations5 Nov 2018 Olga Slizovskaia, Leo Kim, Gloria Haro, Emilia Gomez

Can we perform an end-to-end music source separation with a variable number of sources using a deep learning model?

Music Source Separation

Timbre Analysis of Music Audio Signals with Convolutional Neural Networks

3 code implementations20 Mar 2017 Jordi Pons, Olga Slizovskaia, Rong Gong, Emilia Gómez, Xavier Serra

The focus of this work is to study how to efficiently tailor Convolutional Neural Networks (CNNs) towards learning timbre representations from log-mel magnitude spectrograms.

Sound

Cannot find the paper you are looking for? You can Submit a new open access paper.