Search Results for author: Jongwoo Ko

Found 12 papers, 10 papers with code

DistiLLM: Towards Streamlined Distillation for Large Language Models

2 code implementations6 Feb 2024 Jongwoo Ko, Sungnyun Kim, Tianyi Chen, Se-Young Yun

Knowledge distillation (KD) is widely used for compressing a teacher model to a smaller student model, reducing its inference cost and memory footprint while preserving model capabilities.

Instruction Following Knowledge Distillation

Improving Adaptability and Generalizability of Efficient Transfer Learning for Vision-Language Models

1 code implementation27 Nov 2023 Yongjin Yang, Jongwoo Ko, Se-Young Yun

Vision-Language Models (VLMs) like CLIP have demonstrated remarkable applicability across a variety of downstream tasks, including zero-shot image classification.

General Knowledge Image Classification +2

Fine tuning Pre trained Models for Robustness Under Noisy Labels

no code implementations24 Oct 2023 Sumyeong Ahn, Sihyeon Kim, Jongwoo Ko, Se-Young Yun

To tackle this issue, researchers have explored methods for Learning with Noisy Labels to identify clean samples and reduce the influence of noisy labels.

Denoising Learning with noisy labels

NASH: A Simple Unified Framework of Structured Pruning for Accelerating Encoder-Decoder Language Models

1 code implementation16 Oct 2023 Jongwoo Ko, Seungjoon Park, Yujin Kim, Sumyeong Ahn, Du-Seong Chang, Euijai Ahn, Se-Young Yun

Structured pruning methods have proven effective in reducing the model size and accelerating inference speed in various network architectures such as Transformers.

Fast and Robust Early-Exiting Framework for Autoregressive Language Models with Synchronized Parallel Decoding

1 code implementation9 Oct 2023 Sangmin Bae, Jongwoo Ko, Hwanjun Song, Se-Young Yun

To tackle the high inference latency exhibited by autoregressive language models, previous studies have proposed an early-exiting framework that allocates adaptive computation paths for each token based on the complexity of generating the subsequent token.

CUDA: Curriculum of Data Augmentation for Long-Tailed Recognition

1 code implementation10 Feb 2023 Sumyeong Ahn, Jongwoo Ko, Se-Young Yun

To handle this restriction, several methods have been developed that increase the representations of minority samples by leveraging the features of the majority samples.

Data Augmentation Long-tail Learning

Revisiting Intermediate Layer Distillation for Compressing Language Models: An Overfitting Perspective

1 code implementation3 Feb 2023 Jongwoo Ko, Seungjoon Park, Minchan Jeong, Sukjin Hong, Euijai Ahn, Du-Seong Chang, Se-Young Yun

Knowledge distillation (KD) is a highly promising method for mitigating the computational problems of pre-trained language models (PLMs).

Knowledge Distillation

Synergy with Translation Artifacts for Training and Inference in Multilingual Tasks

1 code implementation18 Oct 2022 Jaehoon Oh, Jongwoo Ko, Se-Young Yun

Translation has played a crucial role in improving the performance on multilingual tasks: (1) to generate the target language data from the source language data for training and (2) to generate the source language data from the target language data for inference.

Sentence Sentence Classification +1

A Gift from Label Smoothing: Robust Training with Adaptive Label Smoothing via Auxiliary Classifier under Label Noise

1 code implementation15 Jun 2022 Jongwoo Ko, Bongsoo Yi, Se-Young Yun

While existing methods address this problem in various directions, they still produce unpredictable sub-optimal results since they rely on the posterior information estimated by the feature extractor corrupted by noisy labels.

Self-Contrastive Learning

no code implementations29 Sep 2021 Sangmin Bae, Sungnyun Kim, Jongwoo Ko, Gihun Lee, Seungjong Noh, Se-Young Yun

This paper proposes a novel contrastive learning framework, called Self-Contrastive (SelfCon) Learning, that self-contrasts within multiple outputs from the different levels of a multi-exit network.

Contrastive Learning

Self-Contrastive Learning: Single-viewed Supervised Contrastive Framework using Sub-network

1 code implementation29 Jun 2021 Sangmin Bae, Sungnyun Kim, Jongwoo Ko, Gihun Lee, Seungjong Noh, Se-Young Yun

To this end, we propose Self-Contrastive (SelfCon) learning, which self-contrasts within multiple outputs from the different levels of a single network.

Contrastive Learning

FINE Samples for Learning with Noisy Labels

1 code implementation NeurIPS 2021 Taehyeon Kim, Jongwoo Ko, Sangwook Cho, Jinhwan Choi, Se-Young Yun

Our framework, coined as filtering noisy instances via their eigenvectors (FINE), provides a robust detector with derivative-free simple methods having theoretical guarantees.

General Classification Learning with noisy labels

Cannot find the paper you are looking for? You can Submit a new open access paper.