Search Results for author: Jangho Kim

Found 19 papers, 6 papers with code

Paraphrasing Complex Network: Network Compression via Factor Transfer

2 code implementations NeurIPS 2018 Jangho Kim, SeongUk Park, Nojun Kwak

Among the model compression methods, a method called knowledge transfer is to train a student network with a stronger teacher network.

Model Compression Transfer Learning

Vehicle Image Generation Going Well with The Surroundings

no code implementations9 Jul 2018 Jeesoo Kim, Jangho Kim, Jaeyoung Yoo, Daesik Kim, Nojun Kwak

Using a subnetwork based on a precedent work of image completion, our model makes the shape of an object.

Colorization Image Generation +7

StackNet: Stacking Parameters for Continual learning

no code implementations7 Sep 2018 Jangho Kim, Jeesoo Kim, Nojun Kwak

The StackNet guarantees no degradation in the performance of the previously learned tasks and the index module shows high confidence in finding the origin of an input sample.

Continual Learning

HC-Net: Memory-based Incremental Dual-Network System for Continual learning

no code implementations27 Sep 2018 Jangho Kim, Jeesoo Kim, Nojun Kwak

The C-Net guarantees no degradation in the performance of the previously learned tasks and the H-Net shows high confidence in finding the origin of an input sample.

Continual Learning Hippocampus

Feature Fusion for Online Mutual Knowledge Distillation

1 code implementation19 Apr 2019 Jangho Kim, Minsung Hyun, Inseop Chung, Nojun Kwak

We propose a learning framework named Feature Fusion Learning (FFL) that efficiently trains a powerful classifier through a fusion module which combines the feature maps generated from parallel neural networks.

Knowledge Distillation

QKD: Quantization-aware Knowledge Distillation

no code implementations28 Nov 2019 Jangho Kim, Yash Bhalgat, Jinwon Lee, Chirag Patel, Nojun Kwak

First, Self-studying (SS) phase fine-tunes a quantized low-precision student network without KD to obtain a good initialization.

Knowledge Distillation Quantization

Feature-map-level Online Adversarial Knowledge Distillation

no code implementations ICML 2020 Inseop Chung, SeongUk Park, Jangho Kim, Nojun Kwak

By training a network to fool the corresponding discriminator, it can learn the other network's feature map distribution.

Knowledge Distillation

Position-based Scaled Gradient for Model Quantization and Pruning

1 code implementation NeurIPS 2020 Jangho Kim, KiYoon Yoo, Nojun Kwak

Second, we empirically show that PSG acting as a regularizer to a weight vector is favorable for model compression domains such as quantization and pruning.

Model Compression Position +1

Prototype-based Personalized Pruning

no code implementations25 Mar 2021 Jangho Kim, Simyung Chang, Sungrack Yun, Nojun Kwak

We verify the usefulness of PPP on a couple of tasks in computer vision and Keyword spotting.

Keyword Spotting Model Compression

PQK: Model Compression via Pruning, Quantization, and Knowledge Distillation

no code implementations25 Jun 2021 Jangho Kim, Simyung Chang, Nojun Kwak

Unlike traditional pruning and KD, PQK makes use of unimportant weights pruned in the pruning process to make a teacher network for training a better student network without pre-training the teacher model.

Keyword Spotting Knowledge Distillation +2

Dynamic Collective Intelligence Learning: Finding Efficient Sparse Model via Refined Gradients for Pruned Weights

1 code implementation10 Sep 2021 Jangho Kim, Jayeon Yoo, Yeji Song, KiYoon Yoo, Nojun Kwak

To alleviate this problem, dynamic pruning methods have emerged, which try to find diverse sparsity patterns during training by utilizing Straight-Through-Estimator (STE) to approximate gradients of pruned weights.

Towards Robust Domain Generalization in 2D Neural Audio Processing

no code implementations29 Sep 2021 Byeonggeun Kim, Seunghan Yang, Jangho Kim, Hyunsin Park, Jun-Tae Lee, Simyung Chang

While using two-dimensional convolutional neural networks (2D-CNNs) in image processing, it is possible to manipulate domain information using channel statistics, and instance normalization has been a promising way to get domain-invariant features.

Acoustic Scene Classification Domain Generalization +3

Domain Generalization on Efficient Acoustic Scene Classification using Residual Normalization

no code implementations12 Nov 2021 Byeonggeun Kim, Seunghan Yang, Jangho Kim, Simyung Chang

Moreover, we introduce an efficient architecture, BC-ResNet-ASC, a modified version of the baseline architecture with a limited receptive field.

Acoustic Scene Classification Classification +5

Self-Distilled Self-Supervised Representation Learning

1 code implementation25 Nov 2021 Jiho Jang, Seonhoon Kim, KiYoon Yoo, Chaerin Kong, Jangho Kim, Nojun Kwak

Through self-distillation, the intermediate layers are better suited for instance discrimination, making the performance of an early-exited sub-network not much degraded from that of the full network.

Representation Learning Self-Supervised Learning

Detection of Word Adversarial Examples in Text Classification: Benchmark and Baseline via Robust Density Estimation

no code implementations3 Mar 2022 KiYoon Yoo, Jangho Kim, Jiho Jang, Nojun Kwak

Word-level adversarial attacks have shown success in NLP models, drastically decreasing the performance of transformer-based models in recent years.

Adversarial Defense Density Estimation +3

Domain Generalization with Relaxed Instance Frequency-wise Normalization for Multi-device Acoustic Scene Classification

no code implementations24 Jun 2022 Byeonggeun Kim, Seunghan Yang, Jangho Kim, Hyunsin Park, JunTae Lee, Simyung Chang

While using two-dimensional convolutional neural networks (2D-CNNs) in image processing, it is possible to manipulate domain information using channel statistics, and instance normalization has been a promising way to get domain-invariant features.

Acoustic Scene Classification Domain Generalization +1

Magnitude Attention-based Dynamic Pruning

no code implementations8 Jun 2023 Jihye Back, Namhyuk Ahn, Jangho Kim

Existing pruning methods utilize the importance of each weight based on specified criteria only when searching for a sparse structure but do not utilize it during training.

Efficient Exploration

Cannot find the paper you are looking for? You can Submit a new open access paper.