Vision-language navigation (VLN) is the task of navigating an embodied agent to carry out natural language instructions inside real 3D environments.
#2 best model for Vision-Language Navigation on Room2Room
We illustrate the promise of ALE by developing and benchmarking domain-independent agents designed using well-established AI techniques for both reinforcement learning and planning.
#8 best model for Atari Games on Atari 2600 Centipede
This paper focuses on building a model that reasons about the long-term future and demonstrates how to use this for efficient planning and exploration.
But is there a way to remove the need for real world human demonstrations during training?