Search Results for author: Song Park

Found 9 papers, 7 papers with code

Rotary Position Embedding for Vision Transformer

1 code implementation20 Mar 2024 Byeongho Heo, Song Park, Dongyoon Han, Sangdoo Yun

This study provides a comprehensive analysis of RoPE when applied to ViTs, utilizing practical implementations of RoPE for 2D vision data.

Position

SeiT++: Masked Token Modeling Improves Storage-efficient Training

1 code implementation15 Dec 2023 Minhyun Lee, Song Park, Byeongho Heo, Dongyoon Han, Hyunjung Shim

A recent breakthrough by SeiT proposed the use of Vector-Quantized (VQ) feature vectors (i. e., tokens) as network inputs for vision classification.

Classification Data Augmentation +2

SeiT: Storage-Efficient Vision Training with Tokens Using 1% of Pixel Storage

1 code implementation ICCV 2023 Song Park, Sanghyuk Chun, Byeongho Heo, Wonjae Kim, Sangdoo Yun

We need billion-scale images to achieve more generalizable and ground-breaking vision models, as well as massive dataset storage to ship the images (e. g., the LAION-4B dataset needs 240TB storage space).

Continual Learning

Few-shot Font Generation with Weakly Supervised Localized Representations

2 code implementations22 Dec 2021 Song Park, Sanghyuk Chun, Junbum Cha, Bado Lee, Hyunjung Shim

Existing methods learn to disentangle style and content elements by developing a universal style representation for each font style.

Font Generation

StyleAugment: Learning Texture De-biased Representations by Style Augmentation without Pre-defined Textures

no code implementations24 Aug 2021 Sanghyuk Chun, Song Park

Hence, StyleAugment let the model observe abundant confounding cues for each image by on-the-fly the augmentation strategy, while the augmented images are more realistic than artistic style transferred images.

Data Augmentation Style Transfer

Multiple Heads are Better than One: Few-shot Font Generation with Multiple Localized Experts

4 code implementations ICCV 2021 Song Park, Sanghyuk Chun, Junbum Cha, Bado Lee, Hyunjung Shim

MX-Font extracts multiple style features not explicitly conditioned on component labels, but automatically by multiple experts to represent different local concepts, e. g., left-side sub-glyph.

Disentanglement Font Generation +1

Few-shot Font Generation with Localized Style Representations and Factorization

3 code implementations23 Sep 2020 Song Park, Sanghyuk Chun, Junbum Cha, Bado Lee, Hyunjung Shim

However, learning component-wise styles solely from reference glyphs is infeasible in the few-shot font generation scenario, when a target script has a large number of components, e. g., over 200 for Chinese.

Font Generation

Cannot find the paper you are looking for? You can Submit a new open access paper.