no code implementations • 7 Aug 2024 • Seok Hwan Lee, Taein Son, Soo Won Seo, Jisong Kim, Jun Won Choi
Video action detection (VAD) is a formidable vision task that involves the localization and classification of actions within the spatial and temporal dimensions of a video clip.
1 code implementation • 23 Jul 2024 • Jae Soon Baik, In Young Yoon, Kun Hoon Kim, Jun Won Choi
The performance of these methods is limited because they use only the training samples within each class for class centroid estimation, making the quality of centroids susceptible to long-tailed distributions and noisy labels.
1 code implementation • 18 Jul 2024 • Sehwan Choi, Jungho Kim, Hongjae Shin, Jun Won Choi
PQG extracts instance-level positional queries by embedding BEV positional information into Mask-Aware Queries, while GFE utilizes BEV Segmentation Masks to generate point-level geometric features.
1 code implementation • 17 Jun 2024 • Yecheol Kim, Junho Lee, Changsoo Park, Hyoung won Kim, Inho Lim, Christopher Chang, Jun Won Choi
TODA efficiently utilizes all available data, including labeled data in the source domain, and both labeled data and unlabeled data in the target domain to enhance domain adaptation performance.
1 code implementation • 11 Mar 2024 • Konyul Park, Yecheol Kim, Junho Koh, Byungwoo Park, Jun Won Choi
Through STV grids, points within each pillar are individually encoded using Vertical PFE (V-PFE), Temporal PFE (T-PFE), and Horizontal PFE (H-PFE).
no code implementations • CVPR 2024 • Geonho Bang, Kwangjin Choi, Jisong Kim, Dongsuk Kum, Jun Won Choi
The inherent noisy and sparse characteristics of radar data pose challenges in finding effective representations for 3D object detection.
no code implementations • 4 Mar 2024 • Jisong Kim, Geonho Bang, Kwangjin Choi, Minjae Seong, Jaechang Yoo, Eunjong Pyo, Jun Won Choi
The PillarGen model performs the following three steps: 1) pillar encoding, 2) Occupied Pillar Prediction (OPP), and 3) Pillar to Point Generation (PPG).
no code implementations • 10 Sep 2023 • Kyoung Ok Yang, Junho Koh, Jun Won Choi
Various types of sensors have been considered to develop human action recognition (HAR) models.
no code implementations • 17 Jul 2023 • Jisong Kim, Minjae Seong, Geonho Bang, Dongsuk Kum, Jun Won Choi
While LiDAR sensors have been successfully applied to 3D object detection, the affordability of radar and camera sensors has led to a growing interest in fusing radars and cameras for 3D object detection.
Ranked #5 on 3D Object Detection on nuscenes Camera-Radar
no code implementations • 12 May 2023 • Minjae Lee, Seongmin Park, Hyungmin Kim, Minyong Yoon, Janghwan Lee, Jun Won Choi, Nam Sung Kim, Mingu Kang, Jungwook Choi
3D object detection using point cloud (PC) data is essential for perception pipelines of autonomous driving, where efficient encoding is key to meeting stringent resource and latency requirements.
1 code implementation • ICCV 2023 • Youngseok Kim, Juyeb Shin, Sanmin Kim, In-Jae Lee, Jun Won Choi, Dongsuk Kum
Autonomous driving requires an accurate and fast 3D perception system that includes 3D object detection, tracking, and segmentation.
Ranked #2 on 3D Multi-Object Tracking on nuscenes Camera-Radar
1 code implementation • 1 Dec 2022 • Junho Koh, Junhyung Lee, Youngwoo Lee, Jaekyum Kim, Jun Won Choi
While conventional 3D object detectors use a set of unordered LiDAR points acquired over a fixed time interval, recent studies have revealed that substantial performance improvement can be achieved by exploiting the spatio-temporal context present in a sequence of LiDAR point sets.
Ranked #26 on 3D Object Detection on nuScenes
1 code implementation • 24 Nov 2022 • Yecheol Kim, Konyul Park, Minwook Kim, Dongsuk Kum, Jun Won Choi
Fusing data from cameras and LiDAR sensors is an essential technique to achieve robust 3D object detection.
Ranked #1 on 3D Object Detection on KITTI Cars Hard
no code implementations • ICCV 2023 • Sehwan Choi, Jungho Kim, Junyong Yun, Jun Won Choi
The trajectory refinement network enhances each of the M proposals using 1) tube-query scene attention (TQSA) and 2) proposal-level interaction attention (PIA) mechanisms.
Ranked #13 on Motion Forecasting on Argoverse CVPR 2020
no code implementations • 29 Oct 2022 • Youngseok Kim, Sanmin Kim, Sangmin Sim, Jun Won Choi, Dongsuk Kum
In this way, our 3D detection network can be supervised by more depth supervision from raw LiDAR points, which does not require any human annotation cost, to estimate accurate depth without explicitly predicting the depth map.
1 code implementation • 30 Sep 2022 • Junhyung Lee, Junho Koh, Youngwoo Lee, Jun Won Choi
LiDAR sensors are widely used for 3D object detection in various mobile robotics applications.
no code implementations • 14 Sep 2022 • Youngseok Kim, Sanmin Kim, Jun Won Choi, Dongsuk Kum
Camera and radar sensors have significant advantages in cost, reliability, and maintenance compared to LiDAR.
Ranked #7 on 3D Object Detection on nuscenes Camera-Radar
no code implementations • 21 Jul 2022 • Jun Ho Lee, Jae Soon Baik, Tae Hwan Hwang, Jun Won Choi
By combining the aforementioned metrics, we present the proposed {\it self-ensemble-based robust training} (SRT) method, which can filter the samples with noisy labels to reduce their influence on training.
no code implementations • 5 Jul 2022 • Jae Soon Baik, In Young Yoon, Jun Won Choi
There is growing interest in the challenging visual perception task of learning from long-tailed class distributions.
no code implementations • 5 Jul 2022 • Jae Soon Baik, In Young Yoon, Jun Won Choi
The student models are given a fixed number of temporal self-ensemble models, and the teacher model is constructed by averaging the weights of the student models.
no code implementations • 14 Dec 2021 • Junho Koh, Jaekyum Kim, Jinhyuk Yoo, Yecheol Kim, Jun Won Choi
The detector constructs the spatio-temporal features via the weighted temporal aggregation of the spatial features obtained by the camera and LiDAR fusion.
1 code implementation • CVPR 2021 • ByeoungDo Kim, Seong Hyeon Park, Seokhwan Lee, Elbek Khoshimjonov, Dongsuk Kum, Junsoo Kim, Jeong Soo Kim, Jun Won Choi
In this paper, we address the problem of predicting the future motion of a dynamic agent (called a target agent) given its current and past states as well as the information on its environment.
no code implementations • 19 Feb 2021 • Sun Hong Lim, Sunwoo Kim, Byonghyo Shim, Jun Won Choi
In this paper, we propose a deep learning-based beam tracking method for millimeter-wave (mmWave)communications.
1 code implementation • 20 Nov 2020 • Junho Koh, Jaekyum Kim, Younji Shin, Byeongwon Lee, Seungji Yang, Jun Won Choi
In this paper, we propose a new video object detector (VoD) method referred to as temporal feature aggregation and motion-aware VoD (TM-VoD), which produces a joint representation of temporal image sequences and object motion.
1 code implementation • ECCV 2020 • Jin Hyeok Yoo, Yecheol Kim, Jisong Kim, Jun Won Choi
First, the method employs auto-calibrated projection, to transform the 2D camera features to a smooth spatial feature map with the highest correspondence to the LiDAR features in the bird's eye view (BEV) domain.
1 code implementation • International Conference on Computer Vision Workshops 2019 • Dawei Du, Pengfei Zhu, Longyin Wen, Xiao Bian, Haibin Lin, QinGhua Hu, Tao Peng, Jiayu Zheng, Xinyao Wang, Yue Zhang, Liefeng Bo, Hailin Shi, Rui Zhu, Aashish Kumar, Aijin Li, Almaz Zinollayev, Anuar Askergaliyev, Arne Schumann, Binjie Mao, Byeongwon Lee, Chang Liu, Changrui Chen, Chunhong Pan, Chunlei Huo, Da Yu, Dechun Cong, Dening Zeng, Dheeraj Reddy Pailla, Di Li, Dong Wang, Donghyeon Cho, Dongyu Zhang, Furui Bai, George Jose, Guangyu Gao, Guizhong Liu, Haitao Xiong, Hao Qi, Haoran Wang, Heqian Qiu, Hongliang Li, Huchuan Lu, Ildoo Kim, Jaekyum Kim, Jane Shen, Jihoon Lee, Jing Ge, Jingjing Xu, Jingkai Zhou, Jonas Meier, Jun Won Choi, Junhao Hu, Junyi Zhang, Junying Huang, Kaiqi Huang, Keyang Wang, Lars Sommer, Lei Jin, Lei Zhang
Results of 33 object detection algorithms are presented.
no code implementations • 1 Aug 2019 • Jin Hyeok Yoo, Dongsuk Kum, Jun Won Choi
Convolutional neural network (CNN) has led to significant progress in object detection.
no code implementations • 17 Jul 2018 • Jaekyum Kim, Junho Koh, Yecheol Kim, Jaehyung Choi, Youngbae Hwang, Jun Won Choi
The goal of multi-modal learning is to use complimentary information on the relevant task provided by the multiple modalities to achieve reliable and robust performance.
no code implementations • 18 Feb 2018 • Seong Hyeon Park, ByeongDo Kim, Chang Mook Kang, Chung Choo Chung, Jun Won Choi
We employ the encoder-decoder architecture which analyzes the pattern underlying in the past trajectory using the long short-term memory (LSTM) based encoder and generates the future trajectory sequence using the LSTM based decoder.
no code implementations • 24 Apr 2017 • ByeoungDo Kim, Chang Mook Kang, Seung Hi Lee, Hyunmin Chae, Jaekyum Kim, Chung Choo Chung, Jun Won Choi
Our approach is data-driven and simple to use in that it learns complex behavior of the vehicles from the massive amount of trajectory data through deep neural network model.
2 code implementations • 8 Feb 2017 • Hyunmin Chae, Chang Mook Kang, ByeoungDo Kim, Jaekyum Kim, Chung Choo Chung, Jun Won Choi
In this paper, we propose a new autonomous braking system based on deep reinforcement learning.