Search Results for author: Jongmin Lee

Found 26 papers, 10 papers with code

Batch Reinforcement Learning with Hyperparameter Gradients

no code implementations ICML 2020 Byung-Jun Lee, Jongmin Lee, Peter Vrancx, Dongho Kim, Kee-Eung Kim

We consider the batch reinforcement learning problem where the agent needs to learn only from a fixed batch of data, without further interaction with the environment.

Continuous Control reinforcement-learning +1

MFOS: Model-Free & One-Shot Object Pose Estimation

no code implementations3 Oct 2023 Jongmin Lee, Yohann Cabon, Romain Brégier, Sungjoo Yoo, Jerome Revaud

Existing learning-based methods for object pose estimation in RGB images are mostly model-specific or category based.

Object Pose Estimation

SACReg: Scene-Agnostic Coordinate Regression for Visual Localization

no code implementations21 Jul 2023 Jerome Revaud, Yohann Cabon, Romain Brégier, Jongmin Lee, Philippe Weinzaepfel

Instead of encoding the scene coordinates into the network weights, our model takes as input a database image with some sparse 2D pixel to 3D coordinate annotations, extracted from e. g. off-the-shelf Structure-from-Motion or RGB-D data, and a query image for which are predicted a dense 3D coordinate map and its confidence, based on cross-attention.

Image Retrieval regression +2

Learning Rotation-Equivariant Features for Visual Correspondence

no code implementations CVPR 2023 Jongmin Lee, Byungjin Kim, SeungWook Kim, Minsu Cho

The resultant features and their orientations are further processed by group aligning, a novel invariant mapping technique that shifts the group-equivariant features by their orientations along the group dimension.

Pose Estimation Self-Supervised Learning

Local Metric Learning for Off-Policy Evaluation in Contextual Bandits with Continuous Actions

1 code implementation24 Oct 2022 Haanvid Lee, Jongmin Lee, Yunseon Choi, Wonseok Jeon, Byung-Jun Lee, Yung-Kyun Noh, Kee-Eung Kim

We consider local kernel metric learning for off-policy evaluation (OPE) of deterministic policies in contextual bandits with continuous action spaces.

Metric Learning Multi-Armed Bandits +1

Self-Supervised Learning of Image Scale and Orientation

1 code implementation15 Jun 2022 Jongmin Lee, Yoonwoo Jeong, Minsu Cho

We study the problem of learning to assign a characteristic pose, i. e., scale and orientation, for an image region of interest.

Pose Estimation Self-Supervised Learning

COptiDICE: Offline Constrained Reinforcement Learning via Stationary Distribution Correction Estimation

1 code implementation ICLR 2022 Jongmin Lee, Cosmin Paduraru, Daniel J. Mankowitz, Nicolas Heess, Doina Precup, Kee-Eung Kim, Arthur Guez

We consider the offline constrained reinforcement learning (RL) problem, in which the agent aims to compute a policy that maximizes expected return while satisfying given cost constraints, learning only from a pre-collected dataset.

Offline RL Off-policy evaluation +1

Self-Supervised Equivariant Learning for Oriented Keypoint Detection

1 code implementation CVPR 2022 Jongmin Lee, Byungjin Kim, Minsu Cho

Detecting robust keypoints from an image is an integral part of many computer vision problems, and the characteristic orientation and scale of keypoints play an important role for keypoint description and matching.

Keypoint Detection Self-Supervised Learning +1

LobsDICE: Offline Learning from Observation via Stationary Distribution Correction Estimation

2 code implementations28 Feb 2022 Geon-Hyeong Kim, Jongmin Lee, Youngsoo Jang, Hongseok Yang, Kee-Eung Kim

We consider the problem of learning from observation (LfO), in which the agent aims to mimic the expert's behavior from the state-only demonstrations by experts.

Imitation Learning

Neural Tangent Kernel Analysis of Deep Narrow Neural Networks

1 code implementation7 Feb 2022 Jongmin Lee, Joo Young Choi, Ernest K. Ryu, Albert No

The tremendous recent progress in analyzing the training dynamics of overparameterized neural networks has primarily focused on wide networks and therefore does not sufficiently address the role of depth in deep learning.

DemoDICE: Offline Imitation Learning with Supplementary Imperfect Demonstrations

no code implementations ICLR 2022 Geon-Hyeong Kim, Seokin Seo, Jongmin Lee, Wonseok Jeon, HyeongJoo Hwang, Hongseok Yang, Kee-Eung Kim

We consider offline imitation learning (IL), which aims to mimic the expert's behavior from its demonstration without further interaction with the environment.

Imitation Learning

Rotation-Equivariant Keypoint Detection

no code implementations29 Sep 2021 Jongmin Lee, Byungjin Kim, Minsu Cho

Therefore, we propose a rotation-invariant keypoint detection method using rotation-equivariant CNNs.

Keypoint Detection Translation

Offline Reinforcement Learning for Large Scale Language Action Spaces

no code implementations ICLR 2022 Youngsoo Jang, Jongmin Lee, Kee-Eung Kim

GPT-Critic is essentially free from the issue of diverging from human language since it learns from the sentences sampled from the pre-trained language model.

Language Modelling Offline RL +2

OptiDICE: Offline Policy Optimization via Stationary Distribution Correction Estimation

1 code implementation21 Jun 2021 Jongmin Lee, Wonseok Jeon, Byung-Jun Lee, Joelle Pineau, Kee-Eung Kim

We consider the offline reinforcement learning (RL) setting where the agent aims to optimize the policy solely from the data without further environment interactions.

Offline RL Reinforcement Learning (RL)

Monte-Carlo Planning and Learning with Language Action Value Estimates

no code implementations ICLR 2021 Youngsoo Jang, Seokin Seo, Jongmin Lee, Kee-Eung Kim

Interactive Fiction (IF) games provide a useful testbed for language-based reinforcement learning agents, posing significant challenges of natural language understanding, commonsense reasoning, and non-myopic planning in the combinatorial search space.

Natural Language Understanding reinforcement-learning +1

Representation Balancing Offline Model-based Reinforcement Learning

no code implementations ICLR 2021 Byung-Jun Lee, Jongmin Lee, Kee-Eung Kim

We present a new objective for model learning motivated by recent advances in the estimation of stationary distribution corrections.

Model-based Reinforcement Learning Offline RL +2

Reinforcement Learning for Control with Multiple Frequencies

no code implementations NeurIPS 2020 Jongmin Lee, ByungJun Lee, Kee-Eung Kim

Many real-world sequential decision problems involve multiple action variables whose control frequencies are different, such that actions take their effects at different periods.

Continuous Control reinforcement-learning +1

Learning to Compose Hypercolumns for Visual Correspondence

1 code implementation ECCV 2020 Juhong Min, Jongmin Lee, Jean Ponce, Minsu Cho

Feature representation plays a crucial role in visual correspondence, and recent methods for image matching resort to deeply stacked convolutional layers.

object-detection Semantic correspondence

Spherical Principal Curves

no code implementations5 Mar 2020 Jang-Hyun Kim, Jongmin Lee, Hee-Seok Oh

In this study, we propose a new approach to construct principal curves on a sphere by a projection of the data onto a continuous curve.

Dimensionality Reduction

SPair-71k: A Large-scale Benchmark for Semantic Correspondence

no code implementations28 Aug 2019 Juhong Min, Jongmin Lee, Jean Ponce, Minsu Cho

In this paper, we present a new large-scale benchmark dataset of semantically paired images, SPair-71k, which contains 70, 958 image pairs with diverse variations in viewpoint and scale.

Semantic correspondence

Hyperpixel Flow: Semantic Correspondence with Multi-layer Neural Features

1 code implementation ICCV 2019 Juhong Min, Jongmin Lee, Jean Ponce, Minsu Cho

Establishing visual correspondences under large intra-class variations requires analyzing images at different levels, from features linked to semantics and context to local patterns, while being invariant to instance-specific details.

Semantic correspondence

Monte-Carlo Tree Search for Constrained POMDPs

no code implementations NeurIPS 2018 Jongmin Lee, Geon-Hyeong Kim, Pascal Poupart, Kee-Eung Kim

In this paper, we present CC-POMCP (Cost-Constrained POMCP), an online MCTS algorithm for large CPOMDPs that leverages the optimization of LP-induced parameters and only requires a black-box simulator of the environment.

Decision Making

Attentive Semantic Alignment with Offset-Aware Correlation Kernels

no code implementations ECCV 2018 Paul Hongsuck Seo, Jongmin Lee, Deunsol Jung, Bohyung Han, Minsu Cho

Semantic correspondence is the problem of establishing correspondences across images depicting different instances of the same object or scene class.

Semantic correspondence Translation

Cannot find the paper you are looking for? You can Submit a new open access paper.