Search Results for author: Insoon Yang

Found 13 papers, 4 papers with code

On Task-Relevant Loss Functions in Meta-Reinforcement Learning and Online LQR

no code implementations9 Dec 2023 Jaeuk Shin, Giho Kim, Howon Lee, Joonho Han, Insoon Yang

Designing a competent meta-reinforcement learning (meta-RL) algorithm in terms of data usage remains a central challenge to be tackled for its successful real-world applications.

Meta Reinforcement Learning

Distributionally Robust Differential Dynamic Programming with Wasserstein Distance

1 code implementation16 May 2023 Astghik Hakobyan, Insoon Yang

To achieve this, we use the Kantrovich duality principle to decompose the value function in a novel way and derive closed-form expressions of the distributionally robust control and worst-case distribution policies to be used in each iteration of our DDP algorithm.

Wasserstein Distributionally Robust Control of Partially Observable Linear Stochastic Systems

no code implementations9 Dec 2022 Astghik Hakobyan, Insoon Yang

Distributionally robust control (DRC) aims to effectively manage distributional ambiguity in stochastic systems.

Anderson Acceleration for Partially Observable Markov Decision Processes: A Maximum Entropy Approach

no code implementations28 Nov 2022 MinGyu Park, Jaeuk Shin, Insoon Yang

Inspired by the quasi-Newton interpretation of AA, we propose a maximum entropy variant of QMDP, which we call soft QMDP, to fully benefit from AA.

Decision Making

Wasserstein Distributionally Robust Control of Partially Observable Linear Systems: Tractable Approximation and Performance Guarantee

no code implementations31 Mar 2022 Astghik Hakobyan, Insoon Yang

The key idea is to reformulate the WDRC problem as a novel minimax control problem with an approximate Wasserstein penalty.

On Affine Policies for Wasserstein Distributionally Robust Unit Commitment

no code implementations29 Mar 2022 Youngchae Cho, Insoon Yang

The proposed model is formulated as a WDRO problem relying on an affine policy, which nests an infinite-dimensional worst-case expectation problem and satisfies the non-anticipativity constraint.

Computational Efficiency

Improved Regret Analysis for Variance-Adaptive Linear Bandits and Horizon-Free Linear Mixture MDPs

no code implementations5 Nov 2021 Yeoneung Kim, Insoon Yang, Kwang-Sung Jun

For linear bandits, we achieve $\tilde O(\min\{d\sqrt{K}, d^{1. 5}\sqrt{\sum_{k=1}^K \sigma_k^2}\} + d^2)$ where $d$ is the dimension of the features, $K$ is the time horizon, and $\sigma_k^2$ is the noise variance at time step $k$, and $\tilde O$ ignores polylogarithmic dependence, which is a factor of $d^3$ improvement.

LEMMA

Training Wasserstein GANs without gradient penalties

no code implementations27 Oct 2021 Dohyun Kwon, Yeoneung Kim, Guido Montúfar, Insoon Yang

We propose a stable method to train Wasserstein generative adversarial networks.

On Anderson acceleration for partially observable Markov decision processes

no code implementations29 Mar 2021 Melike Ermis, MinGyu Park, Insoon Yang

This paper proposes an accelerated method for approximately solving partially observable Markov decision process (POMDP) problems offline.

Risk-sensitive safety analysis using Conditional Value-at-Risk

1 code implementation28 Jan 2021 Margaret P. Chapman, Riccardo Bonalli, Kevin M. Smith, Insoon Yang, Marco Pavone, Claire J. Tomlin

In addition, we propose a second definition for risk-sensitive safe sets and provide a tractable method for their estimation without using a parameter-dependent upper bound.

Hamilton-Jacobi Deep Q-Learning for Deterministic Continuous-Time Systems with Lipschitz Continuous Controls

1 code implementation27 Oct 2020 Jeongho Kim, Jaeuk Shin, Insoon Yang

In this paper, we propose Q-learning algorithms for continuous-time deterministic optimal control problems with Lipschitz continuous controls.

Continuous Control Q-Learning

Hamilton-Jacobi-Bellman Equations for Q-Learning in Continuous Time

no code implementations L4DC 2020 Jeongho Kim, Insoon Yang

The performance of the proposed Q-learning algorithm is demonstrated using 1-, 10- and 20-dimensional dynamical systems.

Q-Learning reinforcement-learning +1

Cannot find the paper you are looking for? You can Submit a new open access paper.