Search Results for author: Jan Østergaard

Found 11 papers, 2 papers with code

How to train your ears: Auditory-model emulation for large-dynamic-range inputs and mild-to-severe hearing losses

1 code implementation15 Mar 2024 Peter Leer, Jesper Jensen, Zheng-Hua Tan, Jan Østergaard, Lars Bramsløw

Our results show that this new optimization objective significantly improves the emulation performance of deep neural networks across relevant input sound levels and auditory-model frequency channels, without increasing the computational load during inference.

Speech Enhancement

Neural Networks Hear You Loud And Clear: Hearing Loss Compensation Using Deep Neural Networks

no code implementations15 Mar 2024 Peter Leer, Jesper Jensen, Laurel Carney, Zheng-Hua Tan, Jan Østergaard, Lars Bramsløw

In this study, we propose a DNN-based approach for hearing-loss compensation, which is trained on the outputs of hearing-impaired and normal-hearing DNN-based auditory models in response to speech signals.

Music Classification Speaker Identification +2

Self-supervised Pretraining for Robust Personalized Voice Activity Detection in Adverse Conditions

no code implementations27 Dec 2023 Holger Severin Bovbjerg, Jesper Jensen, Jan Østergaard, Zheng-Hua Tan

Our experiments show that self-supervised pretraining not only improves performance in clean conditions, but also yields models which are more robust to adverse conditions compared to purely supervised learning.

Action Detection Activity Detection +1

Investigating the Design Space of Diffusion Models for Speech Enhancement

no code implementations7 Dec 2023 Philippe Gonzalez, Zheng-Hua Tan, Jan Østergaard, Jesper Jensen, Tommy Sonne Alstrøm, Tobias May

To address this, we extend this framework to account for the progressive transformation between the clean and noisy speech signals.

Image Generation Speech Enhancement

Diffusion-Based Speech Enhancement in Matched and Mismatched Conditions Using a Heun-Based Sampler

no code implementations5 Dec 2023 Philippe Gonzalez, Zheng-Hua Tan, Jan Østergaard, Jesper Jensen, Tommy Sonne Alstrøm, Tobias May

We show that the proposed system substantially benefits from using multiple databases for training, and achieves superior performance compared to state-of-the-art discriminative models in both matched and mismatched conditions.

Image Generation Speech Enhancement

Head Orientation Estimation with Distributed Microphones Using Speech Radiation Patterns

no code implementations4 Dec 2023 Kaspar Müller, Bilgesu Çakmak, Paul Didier, Simon Doclo, Jan Østergaard, Tobias Wolff

Determining the head orientation of a talker is not only beneficial for various speech signal processing applications, such as source localization or speech enhancement, but also facilitates intuitive voice control and interaction with smart environments or modern car assistants.

Speech Enhancement

Distributed Adaptive Norm Estimation for Blind System Identification in Wireless Sensor Networks

1 code implementation1 Mar 2023 Matthias Blochberger, Filip Elvander, Randall Ali, Jan Østergaard, Jesper Jensen, Marc Moonen, Toon van Waterschoot

Distributed signal-processing algorithms in (wireless) sensor networks often aim to decentralize processing tasks to reduce communication cost and computational complexity or avoid reliance on a single device (i. e., fusion center) for processing.

Minimum Processing Near-end Listening Enhancement

no code implementations31 Oct 2022 Andreas Jonas Fuglsig, Jesper Jensen, Zheng-Hua Tan, Lars Søndergaard Bertelsen, Jens Christian Lindof, Jan Østergaard

The intelligibility and quality of speech from a mobile phone or public announcement system are often affected by background noise in the listening environment.

Joint Far- and Near-End Speech Intelligibility Enhancement based on the Approximated Speech Intelligibility Index

no code implementations15 Nov 2021 Andreas Jonas Fuglsig, Jan Østergaard, Jesper Jensen, Lars Søndergaard Bertelsen, Peter Mariager, Zheng-Hua Tan

However, the existing optimal mutual information based method requires a complicated system model that includes natural speech variations, and relies on approximations and assumptions of the underlying signal distributions.

Speech Enhancement

Deep Sound Field Reconstruction in Real Rooms: Introducing the ISOBEL Sound Field Dataset

no code implementations12 Feb 2021 Miklas Strøm Kristoffersen, Martin Bo Møller, Pablo Martínez-Nuevo, Jan Østergaard

Moreover, the paper advances on a recent deep learning-based method for sound field reconstruction using a very low number of microphones, and proposes an approach for modeling both magnitude and phase response in a U-Net-like neural network architecture.

Cannot find the paper you are looking for? You can Submit a new open access paper.