no code implementations • 25 Mar 2024 • Saad Abdul Ghani, Zizhao Wang, Peter Stone, Xuesu Xiao
In our new Dynamic Learning from Learned Hallucination (Dyna-LfLH), we design and learn a novel latent distribution and sample dynamic obstacles from it, so the generated training data can be used to learn a motion planner to navigate in dynamic environments.
no code implementations • 23 Jan 2024 • Zizhao Wang, Caroline Wang, Xuesu Xiao, Yuke Zhu, Peter Stone
Two desiderata of reinforcement learning (RL) algorithms are the ability to learn from relatively little experience and the ability to learn policies that generalize to a range of problem specifications.
1 code implementation • 27 Jun 2022 • Zizhao Wang, Xuesu Xiao, Zifan Xu, Yuke Zhu, Peter Stone
Learning dynamics models accurately is an important goal for Model-Based Reinforcement Learning (MBRL), but most MBRL methods learn a dense dynamics model which is vulnerable to spurious correlations and therefore generalizes poorly to unseen states.
no code implementations • 24 Mar 2021 • Iretiayo Akinola, Zizhao Wang, Peter Allen
We propose a vision-based reinforcement learning (RL) approach for closed-loop trajectory generation in an arm reaching problem.
1 code implementation • 20 Sep 2019 • Antonio Khalil Moretti, Zizhao Wang, Luhuan Wu, Iddo Drori, Itsik Pe'er
We apply SVO to three nonlinear latent dynamics tasks and provide statistics to rigorously quantify the predictions of filtered and smoothed objectives.
no code implementations • ICLR Workshop DeepGenStruct 2019 • Antonio Moretti, Zizhao Wang, Luhuan Wu, Itsik Pe'er
The task of recovering nonlinear dynamics and latent structure from a population recording is a challenging problem in statistical neuroscience motivating the development of novel techniques in time series analysis.