Search Results for author: Susung Hong

Found 11 papers, 7 papers with code

MusicInfuser: Making Video Diffusion Listen and Dance

no code implementations18 Mar 2025 Susung Hong, Ira Kemelmacher-Shlizerman, Brian Curless, Steven M. Seitz

We introduce MusicInfuser, an approach for generating high-quality dance videos that are synchronized to a specified music track.

Video Generation

Spatiotemporal Skip Guidance for Enhanced Video Diffusion Sampling

no code implementations27 Nov 2024 Junha Hyung, Kinam Kim, Susung Hong, Min-Jung Kim, Jaegul Choo

Our contributions include: (1) introducing STG as an efficient, high-performing guidance technique for video diffusion models, (2) eliminating the need for auxiliary models by simulating a weak model through layer skipping, and (3) ensuring quality-enhanced guidance without compromising sample diversity or dynamics unlike CFG.

Diversity

Smoothed Energy Guidance: Guiding Diffusion Models with Reduced Energy Curvature of Attention

1 code implementation1 Aug 2024 Susung Hong

Recent attempts to extend guidance to unconditional models have relied on heuristic techniques, resulting in suboptimal generation quality and unintended effects.

Image Generation

Effective Rank Analysis and Regularization for Enhanced 3D Gaussian Splatting

no code implementations17 Jun 2024 Junha Hyung, Susung Hong, Sungwon Hwang, Jaeseong Lee, Jaegul Choo, Jin-Hwa Kim

3D reconstruction from multi-view images is one of the fundamental challenges in computer vision and graphics.

3DGS 3D Reconstruction

Retrieval-Augmented Score Distillation for Text-to-3D Generation

1 code implementation5 Feb 2024 Junyoung Seo, Susung Hong, Wooseok Jang, Inès Hyeonsu Kim, Minseop Kwak, Doyup Lee, Seungryong Kim

We leverage the retrieved asset to incorporate its geometric prior in the variational objective and adapt the diffusion model's 2D prior toward view consistency, achieving drastic improvements in both geometry and fidelity of generated scenes.

3D Generation 3D geometry +3

DirecT2V: Large Language Models are Frame-Level Directors for Zero-Shot Text-to-Video Generation

1 code implementation23 May 2023 Susung Hong, Junyoung Seo, Heeseong Shin, Sunghwan Hong, Seungryong Kim

In the paradigm of AI-generated content (AIGC), there has been increasing attention to transferring knowledge from pre-trained text-to-image (T2I) models to text-to-video (T2V) generation.

Text-to-Video Generation Video Generation +1

Debiasing Scores and Prompts of 2D Diffusion for View-consistent Text-to-3D Generation

1 code implementation NeurIPS 2023 Susung Hong, Donghoon Ahn, Seungryong Kim

In this work, we explore existing frameworks for score-distilling text-to-3D generation and identify the main causes of the view inconsistency problem -- the embedded bias of 2D diffusion models.

3D Generation Language Modelling +1

Neural Matching Fields: Implicit Representation of Matching Fields for Visual Correspondence

1 code implementation6 Oct 2022 Sunghwan Hong, Jisu Nam, Seokju Cho, Susung Hong, Sangryul Jeon, Dongbo Min, Seungryong Kim

Existing pipelines of semantic correspondence commonly include extracting high-level semantic features for the invariance against intra-class variations and background clutters.

Semantic correspondence

Improving Sample Quality of Diffusion Models Using Self-Attention Guidance

5 code implementations ICCV 2023 Susung Hong, Gyuseong Lee, Wooseok Jang, Seungryong Kim

Denoising diffusion models (DDMs) have attracted attention for their exceptional generation quality and diversity.

Denoising Diversity +1

Cannot find the paper you are looking for? You can Submit a new open access paper.