no code implementations • 6 Dec 2023 • Junhyuk So, Jungwon Lee, Eunhyeok Park
The substantial computational costs of diffusion models, especially due to the repeated denoising steps necessary for high-quality image generation, present a major obstacle to their widespread adoption.
no code implementations • NeurIPS 2023 • Junhyuk So, Jungwon Lee, Daehyun Ahn, HyungJun Kim, Eunhyeok Park
The diffusion model has gained popularity in vision applications due to its remarkable generative performance and versatility.
no code implementations • 26 Oct 2022 • Yoojin Choi, Mostafa El-Khamy, Jungwon Lee
We propose a novel method for training a conditional generative adversarial network (CGAN) without the use of training data, called zero-shot learning of a CGAN (ZS-CGAN).
no code implementations • 17 Jun 2021 • Yoojin Choi, Mostafa El-Khamy, Jungwon Lee
In the conventional generative replay, the generative model is pre-trained for old data and shared in extra memory for later incremental learning.
no code implementations • 27 Apr 2021 • Weituo Hao, Mostafa El-Khamy, Jungwon Lee, Jianyi Zhang, Kevin J Liang, Changyou Chen, Lawrence Carin
Federated learning has emerged as an important distributed learning paradigm, where a server aggregates a global model from many client-trained models while having no access to the client data.
1 code implementation • 8 May 2020 • Yoojin Choi, Jihwan Choi, Mostafa El-Khamy, Jungwon Lee
The synthetic data are generated from a generator, while no data are used in training the generator and in quantization.
1 code implementation • 8 May 2020 • Abdelrahman Abdelhamed, Mahmoud Afifi, Radu Timofte, Michael S. Brown, Yue Cao, Zhilu Zhang, WangMeng Zuo, Xiaoling Zhang, Jiye Liu, Wendong Chen, Changyuan Wen, Meng Liu, Shuailin Lv, Yunchao Zhang, Zhihong Pan, Baopu Li, Teng Xi, Yanwen Fan, Xiyu Yu, Gang Zhang, Jingtuo Liu, Junyu Han, Errui Ding, Songhyun Yu, Bumjun Park, Jechang Jeong, Shuai Liu, Ziyao Zong, Nan Nan, Chenghua Li, Zengli Yang, Long Bao, Shuangquan Wang, Dongwoon Bai, Jungwon Lee, Youngjung Kim, Kyeongha Rho, Changyeop Shin, Sungho Kim, Pengliang Tang, Yiyun Zhao, Yuqian Zhou, Yuchen Fan, Thomas Huang, Zhihao LI, Nisarg A. Shah, Wei Liu, Qiong Yan, Yuzhi Zhao, Marcin Możejko, Tomasz Latkowski, Lukasz Treszczotko, Michał Szafraniuk, Krzysztof Trojanowski, Yanhong Wu, Pablo Navarrete Michelini, Fengshuo Hu, Yunhua Lu, Sujin Kim, Wonjin Kim, Jaayeon Lee, Jang-Hwan Choi, Magauiya Zhussip, Azamat Khassenov, Jong Hyun Kim, Hwechul Cho, Priya Kansal, Sabari Nathan, Zhangyu Ye, Xiwen Lu, Yaqi Wu, Jiangxin Yang, Yanlong Cao, Siliang Tang, Yanpeng Cao, Matteo Maggioni, Ioannis Marras, Thomas Tanay, Gregory Slabaugh, Youliang Yan, Myungjoo Kang, Han-Soo Choi, Kyungmin Song, Shusong Xu, Xiaomu Lu, Tingniao Wang, Chunxia Lei, Bin Liu, Rajat Gupta, Vineet Kumar
This challenge is based on a newly collected validation and testing image datasets, and hence, named SIDD+.
5 code implementations • 5 May 2020 • Andreas Lugmayr, Martin Danelljan, Radu Timofte, Namhyuk Ahn, Dongwoon Bai, Jie Cai, Yun Cao, Junyang Chen, Kaihua Cheng, SeYoung Chun, Wei Deng, Mostafa El-Khamy, Chiu Man Ho, Xiaozhong Ji, Amin Kheradmand, Gwantae Kim, Hanseok Ko, Kanghyu Lee, Jungwon Lee, Hao Li, Ziluan Liu, Zhi-Song Liu, Shuai Liu, Yunhua Lu, Zibo Meng, Pablo Navarrete Michelini, Christian Micheloni, Kalpesh Prajapati, Haoyu Ren, Yong Hyeok Seo, Wan-Chi Siu, Kyung-Ah Sohn, Ying Tai, Rao Muhammad Umer, Shuangquan Wang, Huibing Wang, Timothy Haoning Wu, Hao-Ning Wu, Biao Yang, Fuzhi Yang, Jaejun Yoo, Tongtong Zhao, Yuanbo Zhou, Haijie Zhuo, Ziyao Zong, Xueyi Zou
This paper reviews the NTIRE 2020 challenge on real world super-resolution.
no code implementations • 14 Feb 2020 • Qing-Feng Liu, Mostafa El-Khamy, Dongwoon Bai, Jungwon Lee
The proposed Global and Selective Attention Network (GSANet) features Atrous Spatial Pyramid Pooling (ASPP) with a novel sparsemax global attention and a novel selective attention that deploys a condensation and diffusion mechanism to aggregate the multi-scale contextual information from the extracted deep features.
no code implementations • 10 Dec 2019 • Ryan Szeto, Mostafa El-Khamy, Jungwon Lee, Jason J. Corso
To combine the benefits of image and video models, we propose an image-to-video model transfer method called Hyperconsistency (HyperCon) that transforms any well-trained image model into a temporally consistent video model without fine-tuning.
no code implementations • Interspeech 2019 • Jaeyoung Kim, Mostafa El-Khamy, Jungwon Lee
Second, three loss functions based on SDR, PESQ and STOI are proposed to minimize the metric mismatch.
Sound Audio and Speech Processing
no code implementations • 13 Oct 2019 • Jaeyoung Kim, Mostafa El-Khamy, Jungwon Lee
Transformer neural networks (TNN) demonstrated state-of-art performance on many natural language processing (NLP) tasks, replacing recurrent neural networks (RNNs), such as LSTMs or GRUs.
Audio and Speech Processing Sound
no code implementations • 25 Sep 2019 • J. Jon Ryu, Yoojin Choi, Young-Han Kim, Mostafa El-Khamy, Jungwon Lee
A new variational autoencoder (VAE) model is proposed that learns a succinct common representation of two correlated data variables for conditional and joint generation tasks.
no code implementations • ICCV 2019 • Yoojin Choi, Mostafa El-Khamy, Jungwon Lee
Our model also shows comparable and sometimes better performance than the state-of-the-art learned image compression models that deploy multiple networks trained for varying rates.
no code implementations • 11 Jun 2019 • Mostafa El-Khamy, Haoyu Ren, Xianzhi Du, Jungwon Lee
In this paper, we introduce the problem of estimating the real world depth of elements in a scene captured by two cameras with different field of views, where the first field of view (FOV) is a Wide FOV (WFOV) captured by a wide angle lens, and the second FOV is contained in the first FOV and is captured by a tele zoom lens.
no code implementations • 7 Jun 2019 • Haoyu Ren, Mostafa El-Khamy, Jungwon Lee
We introduce two different scene understanding modules based on scene classification and coarse depth estimation respectively.
no code implementations • 27 May 2019 • J. Jon Ryu, Yoojin Choi, Young-Han Kim, Mostafa El-Khamy, Jungwon Lee
A new bimodal generative model is proposed for generating conditional and joint samples, accompanied with a training method with learning a succinct bottleneck representation.
no code implementations • 19 Apr 2019 • Xianzhi Du, Mostafa El-Khamy, Jungwon Lee
A stacked atrous multiscale network is proposed to aggregate rich multiscale contextual information from the cost volume which allows for estimating the disparity with high accuracy at multiple scales.
no code implementations • 21 Feb 2019 • Yoojin Choi, Mostafa El-Khamy, Jungwon Lee
We consider the optimization of deep convolutional neural networks (CNNs) such that they provide good performance while having reduced complexity if deployed on either conventional systems with spatial-domain convolution or lower-complexity systems designed for Winograd convolution.
1 code implementation • 26 Jan 2019 • Jaeyoung Kim, Mostafa El-Khamy, Jungwon Lee
First, the network optimization is performed on the time-domain signals after ISTFT to avoid spectrum mismatch.
no code implementations • 16 Oct 2018 • Haoyu Ren, Mostafa El-Khamy, Jungwon Lee
The results show that DN-ResNets are more efficient, robust, and perform better denoising than current state of art deep learning methods, as well as the popular variants of the BM3D algorithm, in cases of blind and non-blind denoising of images corrupted with Poisson, Gaussian or Poisson-Gaussian noise.
no code implementations • 1 Sep 2018 • Yoojin Choi, Mostafa El-Khamy, Jungwon Lee
In training low-precision networks, gradient descent in the backward pass is performed with high-precision weights while quantized low-precision weights and activations are used in the forward pass to calculate the loss function for training.
no code implementations • 21 May 2018 • Yoojin Choi, Mostafa El-Khamy, Jungwon Lee
In particular, the proposed framework produces one compressed model whose convolutional filters can be made sparse either in the spatial domain or in the Winograd domain.
no code implementations • 2 May 2018 • Xianzhi Du, Mostafa El-Khamy, Vlad I. Morariu, Jungwon Lee, Larry Davis
The classification system further classifies the generated candidates based on opinions of multiple deep verification networks and a fusion network which utilizes a novel soft-rejection fusion method to adjust the confidence in the detection results.
no code implementations • NIPS Workshop CDNNRIA 2018 • Yoojin Choi, Mostafa El-Khamy, Jungwon Lee
In this paper, we investigate lossy compression of deep neural networks (DNNs) by weight quantization and lossless source coding for memory-efficient deployment.
no code implementations • 11 Nov 2017 • Haoyu Ren, Mostafa El-Khamy, Jungwon Lee
We propose methodologies to train highly accurate and efficient deep convolutional neural networks (CNNs) for image super resolution (SR).
no code implementations • 27 Oct 2017 • Jaeyoung Kim, Mostafa El-Khamy, Jungwon Lee
Despite the remarkable progress achieved on automatic speech recognition, recognizing far-field speeches mixed with various noise sources is still a challenging task.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +4
3 code implementations • 10 Jan 2017 • Jaeyoung Kim, Mostafa El-Khamy, Jungwon Lee
The residual LSTM provides an additional spatial shortcut path from lower layers for efficient training of deep networks with multiple LSTM layers.
no code implementations • 5 Dec 2016 • Yoojin Choi, Mostafa El-Khamy, Jungwon Lee
Network quantization is one of network compression techniques to reduce the redundancy of deep neural networks.
no code implementations • 11 Oct 2016 • Xianzhi Du, Mostafa El-Khamy, Jungwon Lee, Larry S. Davis
A single shot deep convolutional network is trained as a object detector to generate all possible pedestrian candidates of different sizes and occlusions.
Ranked #21 on Pedestrian Detection on Caltech