no code implementations • 18 Jun 2024 • Dongwon Jo, Taesu Kim, Yulhwa Kim, Jae-Joon Kim
Binarization, which converts weight parameters to binary values, has emerged as an effective strategy to reduce the size of large language models (LLMs).
1 code implementation • 15 Feb 2024 • Taesu Kim, Jongho Lee, Daehyun Ahn, Sarang Kim, Jiwoong Choi, Minkyu Kim, HyungJun Kim
We introduce QUICK, a group of novel optimized CUDA kernels for the efficient inference of quantized Large Language Models (LLMs).
1 code implementation • 14 Feb 2024 • Jiwon Song, Kyungseok Oh, Taesu Kim, HyungJun Kim, Yulhwa Kim, Jae-Joon Kim
In this paper, we introduce SLEB, a novel approach designed to streamline LLMs by eliminating redundant transformer blocks.
no code implementations • 3 Jul 2023 • Jiwoong Choi, Minkyu Kim, Daehyun Ahn, Taesu Kim, Yulhwa Kim, Dongwon Jo, Hyesung Jeon, Jae-Joon Kim, HyungJun Kim
The emergence of diffusion models has greatly broadened the scope of high-fidelity image synthesis, resulting in notable advancements in both practical implementation and academic research.
2 code implementations • 4 Jun 2023 • Changhun Lee, Jungyu Jin, Taesu Kim, HyungJun Kim, Eunhyeok Park
Large language models (LLMs) with hundreds of billions of parameters require powerful server-grade GPUs for inference, limiting their practical deployment.
no code implementations • 15 Mar 2023 • Suhee Jo, Younggun Lee, Yookyung Shin, Yeongtae Hwang, Taesu Kim
In recent years, emotional text-to-speech has shown considerable progress.
no code implementations • 13 Jul 2022 • Yookyung Shin, Younggun Lee, Suhee Jo, Yeongtae Hwang, Taesu Kim
Expressive text-to-speech has shown improved performance in recent years.
no code implementations • 5 Jul 2022 • Gyunpyo Lee, Taesu Kim, Hyeon-Jeong Suk
Therefore, we release GP22, composed of car styling features defined by automotive designers.
1 code implementation • 6 Oct 2021 • Jaesung Tae, Hyeongju Kim, Taesu Kim
We present EdiTTS, an off-the-shelf speech editing methodology based on score-based generative modeling for text-to-speech synthesis.
no code implementations • ICLR 2019 • Daehyun Ahn, Dongsoo Lee, Taesu Kim, Jae-Joon Kim
In this paper, we propose a new sparse matrix format in order to enable a highly parallel decoding process of the entire sparse matrix.
no code implementations • 27 Nov 2018 • Suwon Shon, Young-Gun Lee, Taesu Kim
In this paper, we proposed Random Speaker-variability Subspace (RSS) projection to map a data into LSH based hash tables.
2 code implementations • 23 Nov 2018 • Young-Gun Lee, Suwon Shon, Taesu Kim
First, we train the speech synthesis network bilingually in English and Korean and analyze how the network learns the relations of phoneme pronunciation between the languages.
1 code implementation • 6 Nov 2018 • Young-Gun Lee, Taesu Kim
We propose prosody embeddings for emotional and expressive speech synthesis networks.
no code implementations • journal 2018 • Young-Gun Lee, Taesu Kim, Soo-Young Lee
We propose a neural text-to-speech (TTS) model that can imitate a new speaker's voice using only a small amount of speech sample.
no code implementations • ICLR 2018 • Dongsoo Lee, Daehyun Ahn, Taesu Kim, Pierce I. Chuang, Jae-Joon Kim
Hence, pruning is usually restricted to inference with a batch size of one, for which an efficient parallel matrix-vector multiplication method exists.
no code implementations • 30 Mar 2017 • Hyungjun Kim, Taesu Kim, Jinseok Kim, Jae-Joon Kim
Artificial Neural Network computation relies on intensive vector-matrix multiplications.