1 code implementation • Findings (ACL) 2022 • KiYoon Yoo, Jangho Kim, Jiho Jang, Nojun Kwak
Word-level adversarial attacks have shown success in NLP models, drastically decreasing the performance of transformer-based models in recent years.
no code implementations • 28 Jun 2022 • Byeonggeun Kim, Seunghan Yang, Jangho Kim, Simyung Chang
The goal of the task is to design an audio scene classification system for device-imbalanced datasets under the constraints of model complexity.
no code implementations • 24 Jun 2022 • Byeonggeun Kim, Seunghan Yang, Jangho Kim, Hyunsin Park, JunTae Lee, Simyung Chang
While using two-dimensional convolutional neural networks (2D-CNNs) in image processing, it is possible to manipulate domain information using channel statistics, and instance normalization has been a promising way to get domain-invariant features.
1 code implementation • 3 Mar 2022 • KiYoon Yoo, Jangho Kim, Jiho Jang, Nojun Kwak
Word-level adversarial attacks have shown success in NLP models, drastically decreasing the performance of transformer-based models in recent years.
1 code implementation • 25 Nov 2021 • Jiho Jang, Seonhoon Kim, KiYoon Yoo, Chaerin Kong, Jangho Kim, Nojun Kwak
Through self-distillation, the intermediate layers are better suited for instance discrimination, making the performance of an early-exited sub-network not much degraded from that of the full network.
no code implementations • 12 Nov 2021 • Byeonggeun Kim, Seunghan Yang, Jangho Kim, Simyung Chang
Moreover, we introduce an efficient architecture, BC-ResNet-ASC, a modified version of the baseline architecture with a limited receptive field.
no code implementations • 29 Sep 2021 • Byeonggeun Kim, Seunghan Yang, Jangho Kim, Hyunsin Park, Jun-Tae Lee, Simyung Chang
While using two-dimensional convolutional neural networks (2D-CNNs) in image processing, it is possible to manipulate domain information using channel statistics, and instance normalization has been a promising way to get domain-invariant features.
no code implementations • 10 Sep 2021 • Jangho Kim, Jayeon Yoo, Yeji Song, KiYoon Yoo, Nojun Kwak
To alleviate this problem, dynamic pruning methods have emerged, which try to find diverse sparsity patterns during training by utilizing Straight-Through-Estimator (STE) to approximate gradients of pruned weights.
no code implementations • 25 Jun 2021 • Jangho Kim, Simyung Chang, Nojun Kwak
Unlike traditional pruning and KD, PQK makes use of unimportant weights pruned in the pruning process to make a teacher network for training a better student network without pre-training the teacher model.
no code implementations • 25 Mar 2021 • Jangho Kim, Simyung Chang, Sungrack Yun, Nojun Kwak
We verify the usefulness of PPP on a couple of tasks in computer vision and Keyword spotting.
1 code implementation • NeurIPS 2020 • Jangho Kim, KiYoon Yoo, Nojun Kwak
Second, we empirically show that PSG acting as a regularizer to a weight vector is favorable for model compression domains such as quantization and pruning.
no code implementations • ICML 2020 • Inseop Chung, SeongUk Park, Jangho Kim, Nojun Kwak
By training a network to fool the corresponding discriminator, it can learn the other network's feature map distribution.
no code implementations • 28 Nov 2019 • Jangho Kim, Yash Bhalgat, Jinwon Lee, Chirag Patel, Nojun Kwak
First, Self-studying (SS) phase fine-tunes a quantized low-precision student network without KD to obtain a good initialization.
1 code implementation • 19 Apr 2019 • Jangho Kim, Minsung Hyun, Inseop Chung, Nojun Kwak
We propose a learning framework named Feature Fusion Learning (FFL) that efficiently trains a powerful classifier through a fusion module which combines the feature maps generated from parallel neural networks.
no code implementations • 27 Sep 2018 • Jangho Kim, Jeesoo Kim, Nojun Kwak
The C-Net guarantees no degradation in the performance of the previously learned tasks and the H-Net shows high confidence in finding the origin of an input sample.
no code implementations • 7 Sep 2018 • Jangho Kim, Jeesoo Kim, Nojun Kwak
The StackNet guarantees no degradation in the performance of the previously learned tasks and the index module shows high confidence in finding the origin of an input sample.
no code implementations • 9 Jul 2018 • Jeesoo Kim, Jangho Kim, Jaeyoung Yoo, Daesik Kim, Nojun Kwak
Using a subnetwork based on a precedent work of image completion, our model makes the shape of an object.
2 code implementations • NeurIPS 2018 • Jangho Kim, SeongUk Park, Nojun Kwak
Among the model compression methods, a method called knowledge transfer is to train a student network with a stronger teacher network.