1 code implementation • 13 Nov 2022 • TaeHoon Kim, Mark Marsden, Pyunghwan Ahn, Sangyun Kim, Sihaeng Lee, Alessandra Sala, Seung Hwan Kim
However, we find that large-scale bidirectional training between image and text enables zero-shot image captioning.
1 code implementation • 15 Mar 2022 • Jinsu Yoo, TaeHoon Kim, Sihaeng Lee, Seung Hwan Kim, Honglak Lee, Tae Hyun Kim
Recent transformer-based super-resolution (SR) methods have achieved promising results against conventional CNN-based methods.
1 code implementation • CVPR 2022 • TaeHoon Kim, Gwangmo Song, Sihaeng Lee, Sangyun Kim, Yewon Seo, Soonyoung Lee, Seung Hwan Kim, Honglak Lee, Kyunghoon Bae
Unlike other models, BiART can distinguish between image (or text) as a conditional reference and a generation target.
Ranked #1 on Image Reconstruction on ImageNet 256x256
no code implementations • 19 Nov 2021 • Alexandra Vioni, Myrsini Christidou, Nikolaos Ellinas, Georgios Vamvoukakis, Panos Kakoulidis, TaeHoon Kim, June Sig Sung, Hyoungmin Park, Aimilios Chalamandaris, Pirros Tsiakoulis
This paper presents a method for controlling the prosody at the phoneme level in an autoregressive attention-based text-to-speech system.
1 code implementation • 17 Jun 2020 • Taehoon Kim, Youngjoon Yoo, Jihoon Yang
In this paper, we present a new network architecture search (NAS) procedure to find a network that guarantees both full-precision (FLOAT32) and quantized (INT8) performances.
1 code implementation • 20 May 2018 • Taehoon Kim, Jihoon Yang
Seq2CNN is trained end-to-end to classify various-length texts without preprocessing inputs into fixed length.
Ranked #10 on Text Classification on Yahoo! Answers