Search Results for author: Youngwan Lee

Found 17 papers, 7 papers with code

Training-Free Exponential Extension of Sliding Window Context with Cascading KV Cache

no code implementations24 Jun 2024 Jeffrey Willette, Heejun Lee, Youngwan Lee, Myeongjae Jeon, Sung Ju Hwang

The context window within a transformer provides a form of active memory for the current task, which can be useful for few-shot learning and conditional generation, both which depend heavily on previous context tokens.

Few-Shot Learning MMLU

HiP Attention: Sparse Sub-Quadratic Attention with Hierarchical Attention Pruning

no code implementations14 Jun 2024 Heejun Lee, Geon Park, Youngwan Lee, Jina Kim, Wonyoung Jeong, Myeongjae Jeon, Sung Ju Hwang

In modern large language models (LLMs), increasing sequence lengths is a crucial challenge for enhancing their comprehension and coherence in handling complex tasks such as multi-modal question answering.

Question Answering

PC-LoRA: Low-Rank Adaptation for Progressive Model Compression with Knowledge Distillation

no code implementations13 Jun 2024 Injoon Hwang, Haewon Park, Youngwan Lee, Jooyoung Yang, SunJae Maeng

Low-rank adaption (LoRA) is a prominent method that adds a small number of learnable parameters to the frozen pre-trained weights for parameter-efficient fine-tuning.

Knowledge Distillation Model Compression +1

Visualizing the loss landscape of Self-supervised Vision Transformer

no code implementations28 May 2024 Youngwan Lee, Jeffrey Ryan Willette, Jonghee Kim, Sung Ju Hwang

To further investigate the reason for better generalization of the self-supervised ViT when trained by MAE (MAE-ViT) and the effect of the gradient correction of RC-MAE from the perspective of optimization, we visualize the loss landscapes of the self-supervised vision transformer by both MAE and RC-MAE and compare them with the supervised ViT (Sup-ViT).

Dynamic and Super-Personalized Media Ecosystem Driven by Generative AI: Unpredictable Plays Never Repeating The Same

no code implementations19 Feb 2024 Sungjun Ahn, Hyun-Jeong Yim, Youngwan Lee, Sung-Ik Park

This paper introduces a media service model that exploits artificial intelligence (AI) video generators at the receive end.

KOALA: Empirical Lessons Toward Memory-Efficient and Fast Diffusion Models for Text-to-Image Synthesis

no code implementations7 Dec 2023 Youngwan Lee, KwanYong Park, Yoorhim Cho, Yong-Ju Lee, Sung Ju Hwang

As text-to-image (T2I) synthesis models increase in size, they demand higher inference costs due to the need for more expensive GPUs with larger memory, which makes it challenging to reproduce these models in addition to the restricted access to training datasets.

Denoising Image Generation +1

Exploring The Role of Mean Teachers in Self-supervised Masked Auto-Encoders

1 code implementation5 Oct 2022 Youngwan Lee, Jeffrey Willette, Jonghee Kim, Juho Lee, Sung Ju Hwang

Masked image modeling (MIM) has become a popular strategy for self-supervised learning~(SSL) of visual representations with Vision Transformers.

Classification Instance Segmentation +4

MPViT: Multi-Path Vision Transformer for Dense Prediction

3 code implementations CVPR 2022 Youngwan Lee, Jonghee Kim, Jeff Willette, Sung Ju Hwang

While Convolutional Neural Networks (CNNs) have been the dominant architectures for such tasks, recently introduced Vision Transformers (ViTs) aim to replace them as a backbone.

Instance Segmentation object-detection +3

Diverse Temporal Aggregation and Depthwise Spatiotemporal Factorization for Efficient Video Classification

1 code implementation1 Dec 2020 Youngwan Lee, Hyung-Il Kim, Kimin Yun, Jinyoung Moon

By using the proposed temporal modeling method (T-OSA), and the efficient factorized component (D(2+1)D), we construct two types of VoV3D networks, VoV3D-M and VoV3D-L.

Ranked #30 on Action Recognition on Something-Something V1 (using extra training data)

3D Architecture Action Recognition +2

Adversarial Training with Stochastic Weight Average

no code implementations21 Sep 2020 Joong-won Hwang, Youngwan Lee, Sungchan Oh, Yuseok Bae

Moreover, we further improved SWA to be adequate to adversarial training.

CenterMask: Real-Time Anchor-Free Instance Segmentation

1 code implementation CVPR 2020 Youngwan Lee, Jongyoul Park

We propose a simple yet efficient anchor-free instance segmentation, called CenterMask, that adds a novel spatial attention-guided mask (SAG-Mask) branch to anchor-free one stage object detector (FCOS) in the same vein with Mask R-CNN.

Real-time Instance Segmentation Segmentation +1

CenterMask : Real-Time Anchor-Free Instance Segmentation

8 code implementations arXiv 2019 Youngwan Lee, Jongyoul Park

We hope that CenterMask and VoVNetV2 can serve as a solid baseline of real-time instance segmentation and backbone network for various vision tasks, respectively.

 Ranked #1 on Object Detection on COCO test-dev (Hardware Burden metric)

Panoptic Segmentation Real-time Instance Segmentation +3

An Energy and GPU-Computation Efficient Backbone Network for Real-Time Object Detection

14 code implementations22 Apr 2019 Youngwan Lee, Joong-won Hwang, Sangrok Lee, Yuseok Bae, Jongyoul Park

As DenseNet conserves intermediate features with diverse receptive fields by aggregating them with dense connection, it shows good performance on the object detection task.

object-detection Real-Time Object Detection +2

Rank of Experts: Detection Network Ensemble

no code implementations1 Dec 2017 Seung-Hwan Bae, Youngwan Lee, Youngjoo Jo, Yuseok Bae, Joong-won Hwang

The recent advances of convolutional detectors show impressive performance improvement for large scale object detection.

Object object-detection +1

Wide-Residual-Inception Networks for Real-time Object Detection

no code implementations4 Feb 2017 Youngwan Lee, Byeonghak Yim, Huien Kim, Eunsoo Park, Xuenan Cui, Taekang Woo, Hakil Kim

Since convolutional neural network(CNN)models emerged, several tasks in computer vision have actively deployed CNN models for feature extraction.

Object object-detection +1

Cannot find the paper you are looking for? You can Submit a new open access paper.