no code implementations • 25 Oct 2023 • Se-Ho Kim, Inyong Koo, Inyoung Lee, Byeongjun Park, Changick Kim
During training, DiffRef3D gradually adds noise to the residuals between proposals and target objects, then applies the noisy residuals to proposals to generate hypotheses.
1 code implementation • ICCV 2023 • Inyong Koo, Inyoung Lee, Se-Ho Kim, Hee-Seon Kim, Woo-jin Jeon, Changick Kim
Motivated by this, we propose Point Generation R-CNN (PG-RCNN), a novel end-to-end detector that generates semantic surface points of foreground objects for accurate detection.
1 code implementation • 25 Jan 2022 • Sangmin Woo, Jinyoung Park, Inyong Koo, Sumin Lee, Minki Jeong, Changick Kim
To our surprise, we found that training schedule shows divide-and-conquer-like pattern: time segments are first diversified regardless of the target, then coupled with each target, and fine-tuned to the target again.
no code implementations • 25 May 2021 • Inyong Koo, Minki Jeong, Changick Kim
In this work, we propose a novel framework that generates class representations by extracting features from class-relevant regions of the images.