Search Results for author: Li-Cheng Lan

Found 8 papers, 3 papers with code

Expert Proximity as Surrogate Rewards for Single Demonstration Imitation Learning

1 code implementation1 Feb 2024 Chia-Cheng Chiang, Li-Cheng Lan, Wei-Fang Sun, Chien Feng, Cho-Jui Hsieh, Chun-Yi Lee

In this paper, we focus on single-demonstration imitation learning (IL), a practical approach for real-world applications where obtaining numerous expert demonstrations is costly or infeasible.

Imitation Learning Reinforcement Learning (RL)

Can Agents Run Relay Race with Strangers? Generalization of RL to Out-of-Distribution Trajectories

no code implementations26 Apr 2023 Li-Cheng Lan, huan zhang, Cho-Jui Hsieh

With extensive experimental evaluation, we show the prevalence of \emph{generalization failure} on controllable states from stranger agents.

Reinforcement Learning (RL)

Are AlphaZero-like Agents Robust to Adversarial Perturbations?

1 code implementation7 Nov 2022 Li-Cheng Lan, huan zhang, Ti-Rong Wu, Meng-Yu Tsai, I-Chen Wu, Cho-Jui Hsieh

Given that the state space of Go is extremely large and a human player can play the game from any legal state, we ask whether adversarial states exist for Go AIs that may lead them to play surprisingly wrong actions.

Adversarial Attack Game of Go

Learning to Schedule Learning rate with Graph Neural Networks

no code implementations ICLR 2022 Yuanhao Xiong, Li-Cheng Lan, Xiangning Chen, Ruochen Wang, Cho-Jui Hsieh

By constructing a directed graph for the underlying neural network of the target problem, GNS encodes current dynamics with a graph message passing network and trains an agent to control the learning rate accordingly via reinforcement learning.

Benchmarking Image Classification +2

Learning to Stop: Dynamic Simulation Monte-Carlo Tree Search

no code implementations14 Dec 2020 Li-Cheng Lan, Meng-Yu Tsai, Ti-Rong Wu, I-Chen Wu, Cho-Jui Hsieh

This implies that a significant amount of resources can be saved if we are able to stop the searching earlier when we are confident with the current searching result.

Atari Games

How much progress have we made in neural network training? A New Evaluation Protocol for Benchmarking Optimizers

no code implementations19 Oct 2020 Yuanhao Xiong, Xuanqing Liu, Li-Cheng Lan, Yang You, Si Si, Cho-Jui Hsieh

For end-to-end efficiency, unlike previous work that assumes random hyperparameter tuning, which over-emphasizes the tuning time, we propose to evaluate with a bandit hyperparameter tuning strategy.

Benchmarking Graph Mining

Multiple Policy Value Monte Carlo Tree Search

2 code implementations31 May 2019 Li-Cheng Lan, Wei Li, Ting-Han Wei, I-Chen Wu

Many of the strongest game playing programs use a combination of Monte Carlo tree search (MCTS) and deep neural networks (DNN), where the DNNs are used as policy or value evaluators.

Multi-Labelled Value Networks for Computer Go

no code implementations30 May 2017 Ti-Rong Wu, I-Chen Wu, Guan-Wun Chen, Ting-Han Wei, Tung-Yi Lai, Hung-Chun Wu, Li-Cheng Lan

First, the MSE of the ML value network is generally lower than the value network alone.

Cannot find the paper you are looking for? You can Submit a new open access paper.