Search Results for author: JiHwan Kim

Found 14 papers, 3 papers with code

Generating Animated Layouts as Structured Text Representations

no code implementations2 May 2025 Yeonsang Shin, JiHwan Kim, Yumin Song, Kyungseung Lee, Hyunhee Chung, Taeyoung Na

Despite the remarkable progress in text-to-video models, achieving precise control over text elements and animated graphics remains a significant challenge, especially in applications such as video advertisements.

Layout Generation

CrashFixer: A crash resolution agent for the Linux kernel

no code implementations29 Apr 2025 Alex Mathai, Chenxi Huang, Suwei Ma, JiHwan Kim, Hailie Mitchell, Aleksandr Nogikh, Petros Maniatis, Franjo Ivančić, Junfeng Yang, Baishakhi Ray

In this work, we build upon kGym, which shares a benchmark for system-level Linux kernel bugs and a platform to run experiments on the Linux kernel.

Code Repair

Activating Self-Attention for Multi-Scene Absolute Pose Regression

1 code implementation3 Nov 2024 Miso Lee, JiHwan Kim, Jae-Pil Heo

Based on the statistical analysis, we reveal that queries and keys are mapped in completely different spaces while only a few keys are blended into the query region.

Camera Pose Estimation Pose Estimation +1

Single-shot reconstruction of three-dimensional morphology of biological cells in digital holographic microscopy using a physics-driven neural network

no code implementations30 Sep 2024 JiHwan Kim, Youngdo Kim, Hyo Seung Lee, Eunseok Seo, Sang Joon Lee

However, existing deep learning-based phase retrieval methods have technical limitations in generalization performance and three-dimensional (3D) morphology reconstruction from a single-shot hologram of biological cells.

Deep Learning Image Reconstruction +1

Prediction-Feedback DETR for Temporal Action Detection

no code implementations29 Aug 2024 JiHwan Kim, Miso Lee, Cheol-Ho Cho, Jihyun Lee, Jae-Pil Heo

Temporal Action Detection (TAD) is fundamental yet challenging for real-world video applications.

Action Detection Prediction

Long-term Pre-training for Temporal Action Detection with Transformers

no code implementations23 Aug 2024 JiHwan Kim, Miso Lee, Jae-Pil Heo

Temporal action detection (TAD) is challenging, yet fundamental for real-world video applications.

Action Detection

Mutually-Aware Feature Learning for Few-Shot Object Counting

no code implementations19 Aug 2024 Yerim Jeon, SuBeen Lee, JiHwan Kim, Jae-Pil Heo

Few-shot object counting has garnered significant attention for its practicality as it aims to count target objects in a query image based on given exemplars without the need for additional training.

Object Counting

Boundary-Recovering Network for Temporal Action Detection

no code implementations18 Aug 2024 JiHwan Kim, Jaehyun Choi, Yerim Jeon, Jae-Pil Heo

To this end, we propose Boundary-Recovering Network (BRN) to address the vanishing boundary problem.

Action Detection object-detection +1

FIFO-Diffusion: Generating Infinite Videos from Text without Training

1 code implementation19 May 2024 JiHwan Kim, Junoh Kang, Jinyoung Choi, Bohyung Han

We propose a novel inference technique based on a pretrained diffusion model for text-conditional video generation.

 Ranked #1 on Video Generation on UCF-101 (FVD128 metric)

Text-to-Video Generation Video Generation

Self-Feedback DETR for Temporal Action Detection

no code implementations ICCV 2023 JiHwan Kim, Miso Lee, Jae-Pil Heo

In this paper, we point out the problem in the self-attention of DETR for TAD; the attention modules focus on a few key elements, called temporal collapse problem.

Action Detection Decoder +1

When Crowd Meets Persona: Creating a Large-Scale Open-Domain Persona Dialogue Corpus

no code implementations1 Apr 2023 Won Ik Cho, Yoon Kyung Lee, Seoyeon Bae, JiHwan Kim, Sangah Park, Moosung Kim, Sowon Hahn, Nam Soo Kim

Building a natural language dataset requires caution since word semantics is vulnerable to subtle text change or the definition of the annotated concept.

Dialogue Generation Question Answering +2

Bootstrap Equilibrium and Probabilistic Speaker Representation Learning for Self-supervised Speaker Verification

no code implementations16 Dec 2021 Sung Hwan Mun, Min Hyun Han, Dongjune Lee, JiHwan Kim, Nam Soo Kim

In this paper, we propose self-supervised speaker representation learning strategies, which comprise of a bootstrap equilibrium speaker representation learning in the front-end and an uncertainty-aware probabilistic speaker embedding training in the back-end.

Contrastive Learning Representation Learning +1

Self-Supervised Video GANs: Learning for Appearance Consistency and Motion Coherency

no code implementations CVPR 2021 Sangeek Hyun, JiHwan Kim, Jae-Pil Heo

The proposed tasks enable the discriminators to learn representations of appearance and temporal context, and force the generator to synthesize videos with consistent appearance and natural flow of motions.

Contrastive Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.