Search Results for author: Hwanjun Song

Found 37 papers, 17 papers with code

Can Your Model Tell a Negation from an Implicature? Unravelling Challenges With Intent Encoders

no code implementations7 Mar 2024 Yuwei Zhang, Siffi Singh, Sailik Sengupta, Igor Shalyminov, Hang Su, Hwanjun Song, Saab Mansour

The triplet task gauges the model's understanding of two semantic concepts paramount in real-world conversational systems-- negation and implicature.

Clustering intent-classification +2

Semi-Supervised Dialogue Abstractive Summarization via High-Quality Pseudolabel Selection

1 code implementation6 Mar 2024 Jianfeng He, Hang Su, Jason Cai, Igor Shalyminov, Hwanjun Song, Saab Mansour

Semi-supervised dialogue summarization (SSDS) leverages model-generated summaries to reduce reliance on human-labeled data and improve the performance of summarization models.

Abstractive Text Summarization Natural Language Understanding

MAGID: An Automated Pipeline for Generating Synthetic Multi-modal Datasets

no code implementations5 Mar 2024 Hossein Aboutalebi, Hwanjun Song, Yusheng Xie, Arshit Gupta, Justin Sun, Hang Su, Igor Shalyminov, Nikolaos Pappas, Siffi Singh, Saab Mansour

Development of multimodal interactive systems is hindered by the lack of rich, multimodal (text, images) conversational data, which is needed in large quantities for LLMs.

Image-text matching Retrieval +1

TofuEval: Evaluating Hallucinations of LLMs on Topic-Focused Dialogue Summarization

1 code implementation20 Feb 2024 Liyan Tang, Igor Shalyminov, Amy Wing-mei Wong, Jon Burnsky, Jake W. Vincent, Yu'an Yang, Siffi Singh, Song Feng, Hwanjun Song, Hang Su, Lijia Sun, Yi Zhang, Saab Mansour, Kathleen McKeown

We find that there are diverse errors and error distributions in model-generated summaries and that non-LLM based metrics can capture all error types better than LLM-based evaluators.

Hallucination News Summarization +2

Adaptive Shortcut Debiasing for Online Continual Learning

no code implementations14 Dec 2023 Doyoung Kim, Dongmin Park, Yooju Shin, Jihwan Bang, Hwanjun Song, Jae-Gil Lee

We propose a novel framework DropTop that suppresses the shortcut bias in online continual learning (OCL) while being adaptive to the varying degree of the shortcut bias incurred by continuously changing environment.

Continual Learning

Enhancing Abstractiveness of Summarization Models through Calibrated Distillation

no code implementations20 Oct 2023 Hwanjun Song, Igor Shalyminov, Hang Su, Siffi Singh, Kaisheng Yao, Saab Mansour

Our experiments show that DisCal outperforms prior methods in abstractive summarization distillation, producing highly abstractive and informative summaries.

Abstractive Text Summarization Informativeness +1

Fast and Robust Early-Exiting Framework for Autoregressive Language Models with Synchronized Parallel Decoding

1 code implementation9 Oct 2023 Sangmin Bae, Jongwoo Ko, Hwanjun Song, Se-Young Yun

To tackle the high inference latency exhibited by autoregressive language models, previous studies have proposed an early-exiting framework that allocates adaptive computation paths for each token based on the complexity of generating the subsequent token.

Prompt-Guided Transformers for End-to-End Open-Vocabulary Object Detection

no code implementations25 Mar 2023 Hwanjun Song, Jihwan Bang

Prompt-OVD is an efficient and effective framework for open-vocabulary object detection that utilizes class embeddings from CLIP as prompts, guiding the Transformer decoder to detect objects in both base and novel classes.

object-detection Open Vocabulary Object Detection +1

Re-thinking Federated Active Learning based on Inter-class Diversity

1 code implementation CVPR 2023 Sangmook Kim, Sangmin Bae, Hwanjun Song, Se-Young Yun

In this work, we first demonstrate that the superiority of two selector models depends on the global and local inter-class diversity.

Active Learning Federated Learning

Q-HyViT: Post-Training Quantization for Hybrid Vision Transformer with Bridge Block Reconstruction

no code implementations22 Mar 2023 Jemin Lee, Yongin Kwon, Jeman Park, Misun Yu, Sihyeong Park, Hwanjun Song

To overcome these challenges, we propose a new post-training quantization method, which is the first to quantize efficient hybrid ViTs (MobileViTv1, MobileViTv2, Mobile-Former, EfficientFormerV1, EfficientFormerV2) with a significant margin (an average improvement of 8. 32\% for 8-bit and 26. 02\% for 6-bit) compared to existing PTQ methods (EasyQuant, FQ-ViT, and PTQ4ViT).

Quantization

Generating Instance-level Prompts for Rehearsal-free Continual Learning

no code implementations ICCV 2023 Dahuin Jung, Dongyoon Han, Jihwan Bang, Hwanjun Song

However, we observe that the use of a prompt pool creates a domain scalability problem between pre-training and continual learning.

Continual Learning

e-CLIP: Large-Scale Vision-Language Representation Learning in E-commerce

no code implementations1 Jul 2022 Wonyoung Shin, Jonghun Park, Taekang Woo, Yongwoo Cho, Kwangjin Oh, Hwanjun Song

Understanding vision and language representations of product content is vital for search and recommendation applications in e-commerce.

Attribute Attribute Extraction +3

Dataset Condensation via Efficient Synthetic-Data Parameterization

2 code implementations30 May 2022 Jang-Hyun Kim, Jinuk Kim, Seong Joon Oh, Sangdoo Yun, Hwanjun Song, JoonHyun Jeong, Jung-Woo Ha, Hyun Oh Song

The great success of machine learning with massive amounts of data comes at a price of huge computation costs and storage for training and tuning.

Dataset Condensation

ReFine: Re-randomization before Fine-tuning for Cross-domain Few-shot Learning

no code implementations11 May 2022 Jaehoon Oh, Sungnyun Kim, Namgyu Ho, Jin-Hwa Kim, Hwanjun Song, Se-Young Yun

Cross-domain few-shot learning (CD-FSL), where there are few target samples under extreme differences between source and target domains, has recently attracted huge attention.

cross-domain few-shot learning Transfer Learning

FedRN: Exploiting k-Reliable Neighbors Towards Robust Federated Learning

1 code implementation3 May 2022 Sangmook Kim, Wonyoung Shin, Soohyuk Jang, Hwanjun Song, Se-Young Yun

Robustness is becoming another important challenge of federated learning in that the data collection process in each client is naturally accompanied by noisy labels.

Federated Learning

Online Continual Learning on a Contaminated Data Stream with Blurry Task Boundaries

2 code implementations CVPR 2022 Jihwan Bang, Hyunseo Koh, Seulki Park, Hwanjun Song, Jung-Woo Ha, Jonghyun Choi

A large body of continual learning (CL) methods, however, assumes data streams with clean labels, and online learning scenarios under noisy data streams are yet underexplored.

Continual Learning

Meta-Learning for Online Update of Recommender Systems

1 code implementation19 Mar 2022 Minseok Kim, Hwanjun Song, Yooju Shin, Dongmin Park, Kijung Shin, Jae-Gil Lee

It is featured with an adaptive learning rate for each parameter-interaction pair for inducing a recommender to quickly learn users' up-to-date interest.

Meta-Learning Recommendation Systems

Understanding Cross-Domain Few-Shot Learning Based on Domain Similarity and Few-Shot Difficulty

2 code implementations1 Feb 2022 Jaehoon Oh, Sungnyun Kim, Namgyu Ho, Jin-Hwa Kim, Hwanjun Song, Se-Young Yun

This data enables self-supervised pre-training on the target domain, in addition to supervised pre-training on the source domain.

cross-domain few-shot learning

Data Collection and Quality Challenges in Deep Learning: A Data-Centric AI Perspective

no code implementations13 Dec 2021 Steven Euijong Whang, Yuji Roh, Hwanjun Song, Jae-Gil Lee

In this survey, we study the research landscape for data collection and data quality primarily for deep learning applications.

BIG-bench Machine Learning Fairness +2

Task-Agnostic Undesirable Feature Deactivation Using Out-of-Distribution Data

1 code implementation NeurIPS 2021 Dongmin Park, Hwanjun Song, Minseok Kim, Jae-Gil Lee

A deep neural network (DNN) has achieved great success in many machine learning tasks by virtue of its high expressive power.

Coherence-based Label Propagation over Time Series for Accelerated Active Learning

no code implementations ICLR 2022 Yooju Shin, Susik Yoon, Sundong Kim, Hwanjun Song, Jae-Gil Lee, Byung Suk Lee

Time-series data are ubiquitous these days, but lack of the labels in time-series data is regarded as a hurdle for its broad applicability.

Active Learning Time Series +1

Influential Rank: A New Perspective of Post-training for Robust Model against Noisy Labels

no code implementations14 Jun 2021 Seulki Park, Hwanjun Song, Daeho Um, Dae Ung Jo, Sangdoo Yun, Jin Young Choi

Deep neural network can easily overfit to even noisy labels due to its high capacity, which degrades the generalization performance of a model.

Learning with noisy labels

Robust Learning by Self-Transition for Handling Noisy Labels

no code implementations8 Dec 2020 Hwanjun Song, Minseok Kim, Dongmin Park, Yooju Shin, Jae-Gil Lee

In the seeding phase, the network is updated using all the samples to collect a seed of clean samples.

MORPH

Learning from Noisy Labels with Deep Neural Networks: A Survey

1 code implementation16 Jul 2020 Hwanjun Song, Minseok Kim, Dongmin Park, Yooju Shin, Jae-Gil Lee

Deep learning has achieved remarkable success in numerous domains with help from large amounts of big data.

Carpe Diem, Seize the Samples Uncertain "At the Moment" for Adaptive Batch Selection

no code implementations19 Nov 2019 Hwanjun Song, Minseok Kim, Sundong Kim, Jae-Gil Lee

Compared with existing batch selection methods, the results showed that Recency Bias reduced the test error by up to 20. 97% in a fixed wall-clock training time.

How does Early Stopping Help Generalization against Label Noise?

no code implementations19 Nov 2019 Hwanjun Song, Minseok Kim, Dongmin Park, Jae-Gil Lee

In this paper, we claim that such overfitting can be avoided by "early stopping" training a deep neural network before the noisy labels are severely memorized.

MLAT: Metric Learning for kNN in Streaming Time Series

no code implementations23 Oct 2019 Dongmin Park, Susik Yoon, Hwanjun Song, Jae-Gil Lee

Learning a good distance measure for distance-based classification in time series leads to significant performance improvement in many tasks.

Metric Learning Time Series +1

Prestopping: How Does Early Stopping Help Generalization Against Label Noise?

no code implementations25 Sep 2019 Hwanjun Song, Minseok Kim, Dongmin Park, Jae-Gil Lee

In this paper, we claim that such overfitting can be avoided by "early stopping" training a deep neural network before the noisy labels are severely memorized.

SELFIE: Refurbishing Unclean Samples for Robust Deep Learning

1 code implementation15 Jun 2019 Hwanjun Song, Minseok Kim, Jae-Gil Lee

Owing to the extremely high expressive power of deep neural networks, their side effect is to totally memorize training data even when the labels are extremely noisy.

Learning with noisy labels

Cannot find the paper you are looking for? You can Submit a new open access paper.