1 code implementation • 14 Mar 2024 • Byeongjun Park, Hyojun Go, Jin-Young Kim, Sangmin Woo, Seokil Ham, Changick Kim
To achieve this, we employ a sparse mixture-of-experts within each transformer block to utilize semantic information and facilitate handling conflicts in tasks through parameter isolation.
1 code implementation • 26 Dec 2023 • Sangmin Woo, Byeongjun Park, Hyojun Go, Jin-Young Kim, Changick Kim
This work introduces HarmonyView, a simple yet effective diffusion sampling technique adept at decomposing two intricate aspects in single-image 3D generation: consistency and diversity.
no code implementations • 25 Oct 2023 • Se-Ho Kim, Inyong Koo, Inyoung Lee, Byeongjun Park, Changick Kim
During training, DiffRef3D gradually adds noise to the residuals between proposals and target objects, then applies the noisy residuals to proposals to generate hypotheses.
no code implementations • 14 Oct 2023 • Byeongjun Park, Changick Kim
Dynamic radiance fields have emerged as a promising approach for generating novel views from a monocular video.
2 code implementations • 11 Oct 2023 • Byeongjun Park, Sangmin Woo, Hyojun Go, Jin-Young Kim, Changick Kim
Diffusion models generate highly realistic images by learning a multi-step denoising process, naturally embodying the principles of multi-task learning (MTL).
no code implementations • 15 Sep 2022 • Byeongjun Park, Hyojun Go, Changick Kim
Although recent methods generate high-quality novel views, synthesizing with only one explicit or implicit 3D geometry has a trade-off between two objectives that we call the "seesaw" problem: 1) preserving reprojected contents and 2) completing realistic out-of-view regions.
no code implementations • 31 Aug 2022 • JeongSoo Kim, Sangmin Woo, Byeongjun Park, Changick Kim
Camera traps, unmanned observation devices, and deep learning-based image recognition systems have greatly reduced human effort in collecting and analyzing wildlife images.
no code implementations • 15 Feb 2022 • Byeongjun Park, JeongSoo Kim, Seungju Cho, Heeseon Kim, Changick Kim
Here, we propose a unified framework and introduce two datasets for long-tailed camera-trap recognition.
no code implementations • 8 Nov 2021 • Byeongjun Park, Taekyung Kim, Hyojun Go, Changick Kim
In this paper, we propose residual guidance loss that enables the depth estimation network to embed the discriminative feature by transferring the discriminability of auto-encoded features.