no code implementations • 18 Dec 2023 • Samuel Yang-Zhao, Kee Siong Ng, Marcus Hutter
Prior approximations of AIXI, a Bayesian optimality notion for general reinforcement learning, can only approximate AIXI's Bayesian environment model using an a-priori defined set of models.
no code implementations • 13 Oct 2022 • Samuel Yang-Zhao, Tianyu Wang, Kee Siong Ng
We propose a practical integration of logical state abstraction with AIXI, a Bayesian optimality notion for reinforcement learning agents, to significantly expand the model class that AIXI agents can be approximated over to complex history-dependent and structured environments.
no code implementations • 5 Jun 2022 • Dawei Chen, Samuel Yang-Zhao, John Lloyd, Kee Siong Ng
This paper introduces the factored conditional filter, a new filtering algorithm for simultaneously tracking states and estimating parameters in high-dimensional state spaces.
no code implementations • 28 May 2019 • Marcus Hutter, Samuel Yang-Zhao, Sultan J. Majeed
The convergence of many reinforcement learning (RL) algorithms with linear function approximation has been investigated extensively but most proofs assume that these methods converge to a unique solution.