1 code implementation • 6 Dec 2023 • Jang-Hyun Kim, Junyoung Yeom, Sangdoo Yun, Hyun Oh Song
This paper presents a context key/value compression method for Transformer language models in online scenarios, where the context continually expands.
1 code implementation • NeurIPS 2023 • Jang-Hyun Kim, Sangdoo Yun, Hyun Oh Song
To this end, we present scalable and effective algorithms for detecting label errors and outlier data based on the relational graph structure of data.
2 code implementations • 30 May 2022 • Jang-Hyun Kim, Jinuk Kim, Seong Joon Oh, Sangdoo Yun, Hwanjun Song, JoonHyun Jeong, Jung-Woo Ha, Hyun Oh Song
The great success of machine learning with massive amounts of data comes at a price of huge computation costs and storage for training and tuning.
4 code implementations • NeurIPS 2021 • Gaon An, Seungyong Moon, Jang-Hyun Kim, Hyun Oh Song
However, prior methods typically require accurate estimation of the behavior policy or sampling from OOD data points, which themselves can be a non-trivial problem.
Ranked #1 on Gym halfcheetah-random on D4RL
1 code implementation • ICLR 2021 • Jang-Hyun Kim, Wonho Choo, Hosan Jeong, Hyun Oh Song
While deep neural networks show great performance on fitting to the training distribution, improving the networks' generalization performance to the test distribution and robustness to the sensitivity to input perturbations still remain as a challenge.
1 code implementation • ICML 2020 • Jang-Hyun Kim, Wonho Choo, Hyun Oh Song
While deep neural networks achieve great performance on fitting the training distribution, the learned networks are prone to overfitting and are susceptible to adversarial attacks.
Ranked #2 on Image Classification on Tiny-ImageNet
no code implementations • 5 Mar 2020 • Jang-Hyun Kim, Jongmin Lee, Hee-Seok Oh
In this study, we propose a new approach to construct principal curves on a sphere by a projection of the data onto a continuous curve.
7 code implementations • ICLR 2019 • Hyeong-Seok Choi, Jang-Hyun Kim, Jaesung Huh, Adrian Kim, Jung-Woo Ha, Kyogu Lee
Most deep learning-based models for speech enhancement have mainly focused on estimating the magnitude of spectrogram while reusing the phase from noisy speech for reconstruction.
1 code implementation • 21 Dec 2018 • Jang-Hyun Kim, Jaejun Yoo, Sanghyuk Chun, Adrian Kim, Jung-Woo Ha
We present a hybrid framework that leverages the trade-off between temporal and frequency precision in audio representations to improve the performance of speech enhancement task.
Audio and Speech Processing Sound