Amortized Posterior on Latent Variables in Gaussian Process

29 Sep 2021  ·  Qing Sun ·

Deep neural networks have achieved impressive performance on a variety of domains. However, performing tasks in partially observed, dynamic environments is still an open problem. Gaussian Process (GPs) is well-known for capturing uncertainty in model parameters. However, it simply assumes a fixed Gaussian prior on latent variables. Thus, agents are not able to update their beliefs about latent variables as observing data points. Instead, in this paper, we propose to replace the prior with an amortized posterior, which enables quick adaptation, especially to abrupt changes. Experiments show that our proposed method can adjust behaviors on the fly (e.g., blind “Predator” take 56% more chance to approach “Prey”), correct mistakes to escape bad situations (e.g., 25% ↑ on avoiding repeating to hit objects with negative rewards), and update beliefs quickly (e.g., 9% faster convergence rate on learning new concepts).

PDF Abstract
No code implementations yet. Submit your code now

Tasks


Datasets


  Add Datasets introduced or used in this paper

Results from the Paper


  Submit results from this paper to get state-of-the-art GitHub badges and help the community compare results to other papers.

Methods