Search Results for author: Byeonggeun Kim

Found 12 papers, 1 papers with code

Improving Small Footprint Few-shot Keyword Spotting with Supervision on Auxiliary Data

no code implementations31 Aug 2023 Seunghan Yang, Byeonggeun Kim, Kyuhong Shim, Simyung Chang

Few-shot keyword spotting (FS-KWS) models usually require large-scale annotated datasets to generalize to unseen target keywords.

Keyword Spotting Multi-Task Learning +1

Scalable Weight Reparametrization for Efficient Transfer Learning

no code implementations26 Feb 2023 Byeonggeun Kim, Jun-Tae Lee, Seunghan Yang, Simyung Chang

Efficient transfer learning involves utilizing a pre-trained model trained on a larger dataset and repurposing it for downstream tasks with the aim of maximizing the reuse of the pre-trained model.

Keyword Spotting Transfer Learning

TTN: A Domain-Shift Aware Batch Normalization in Test-Time Adaptation

no code implementations10 Feb 2023 Hyesu Lim, Byeonggeun Kim, Jaegul Choo, Sungha Choi

In this paper, we identify that CBN and TBN are in a trade-off relationship and present a new test-time normalization (TTN) method that interpolates the statistics by adjusting the importance between CBN and TBN according to the domain-shift sensitivity of each BN layer.

Test-time Adaptation

Personalized Keyword Spotting through Multi-task Learning

no code implementations28 Jun 2022 Seunghan Yang, Byeonggeun Kim, Inseop Chung, Simyung Chang

We design two personalized KWS tasks; (1) Target user Biased KWS (TB-KWS) and (2) Target user Only KWS (TO-KWS).

Keyword Spotting Multi-Task Learning +1

Dummy Prototypical Networks for Few-Shot Open-Set Keyword Spotting

no code implementations28 Jun 2022 Byeonggeun Kim, Seunghan Yang, Inseop Chung, Simyung Chang

We also verify our method on a standard benchmark, miniImageNet, and D-ProtoNets shows the state-of-the-art open-set detection rate in FSOSR.

Keyword Spotting Metric Learning +1

Domain Generalization with Relaxed Instance Frequency-wise Normalization for Multi-device Acoustic Scene Classification

no code implementations24 Jun 2022 Byeonggeun Kim, Seunghan Yang, Jangho Kim, Hyunsin Park, JunTae Lee, Simyung Chang

While using two-dimensional convolutional neural networks (2D-CNNs) in image processing, it is possible to manipulate domain information using channel statistics, and instance normalization has been a promising way to get domain-invariant features.

Acoustic Scene Classification Domain Generalization +1

Domain Generalization on Efficient Acoustic Scene Classification using Residual Normalization

no code implementations12 Nov 2021 Byeonggeun Kim, Seunghan Yang, Jangho Kim, Simyung Chang

Moreover, we introduce an efficient architecture, BC-ResNet-ASC, a modified version of the baseline architecture with a limited receptive field.

Acoustic Scene Classification Classification +5

Towards Robust Domain Generalization in 2D Neural Audio Processing

no code implementations29 Sep 2021 Byeonggeun Kim, Seunghan Yang, Jangho Kim, Hyunsin Park, Jun-Tae Lee, Simyung Chang

While using two-dimensional convolutional neural networks (2D-CNNs) in image processing, it is possible to manipulate domain information using channel statistics, and instance normalization has been a promising way to get domain-invariant features.

Acoustic Scene Classification Domain Generalization +3

Broadcasted Residual Learning for Efficient Keyword Spotting

4 code implementations8 Jun 2021 Byeonggeun Kim, Simyung Chang, Jinkyu Lee, Dooyong Sung

We present a broadcasted residual learning method to achieve high accuracy with small model size and computational load.

Keyword Spotting

Query-by-example on-device keyword spotting

no code implementations11 Oct 2019 Byeonggeun Kim, Mingu Lee, Jinkyu Lee, Yeonseok Kim, Kyuwoong Hwang

A keyword spotting (KWS) system determines the existence of, usually predefined, keyword in a continuous speech stream.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +2

Orthogonality Constrained Multi-Head Attention For Keyword Spotting

no code implementations10 Oct 2019 Mingu Lee, Jinkyu Lee, Hye Jin Jang, Byeonggeun Kim, Wonil Chang, Kyuwoong Hwang

Augmenting regularization terms which penalize positional and contextual non-orthogonality between the attention heads encourages to output different representations from separate subsequences, which in turn enables leveraging structured information without explicit sequence models such as hidden Markov models.

Keyword Spotting

Cannot find the paper you are looking for? You can Submit a new open access paper.