Search Results for author: Vincent Micheli

Found 6 papers, 2 papers with code

Transformers are Sample-Efficient World Models

1 code implementation1 Sep 2022 Vincent Micheli, Eloi Alonso, François Fleuret

Deep reinforcement learning agents are notoriously sample inefficient, which considerably limits their application to real-world problems.

Atari Games 100k reinforcement-learning +1

MineRL Diamond 2021 Competition: Overview, Results, and Lessons Learned

no code implementations17 Feb 2022 Anssi Kanervisto, Stephanie Milani, Karolis Ramanauskas, Nicholay Topin, Zichuan Lin, Junyou Li, Jianing Shi, Deheng Ye, Qiang Fu, Wei Yang, Weijun Hong, Zhongyue Huang, Haicheng Chen, Guangjun Zeng, Yue Lin, Vincent Micheli, Eloi Alonso, François Fleuret, Alexander Nikulin, Yury Belousov, Oleg Svidchenko, Aleksei Shpilman

With this in mind, we hosted the third edition of the MineRL ObtainDiamond competition, MineRL Diamond 2021, with a separate track in which we permitted any solution to promote the participation of newcomers.

Language Models are Few-Shot Butlers

1 code implementation EMNLP 2021 Vincent Micheli, François Fleuret

Pretrained language models demonstrate strong performance in most NLP tasks when fine-tuned on small task-specific datasets.

reinforcement-learning Reinforcement Learning (RL)

Multi-task Reinforcement Learning with a Planning Quasi-Metric

no code implementations8 Feb 2020 Vincent Micheli, Karthigan Sinnathamby, François Fleuret

We introduce a new reinforcement learning approach combining a planning quasi-metric (PQM) that estimates the number of steps required to go from any state to another, with task-specific "aimers" that compute a target state to reach a given goal.

reinforcement-learning Reinforcement Learning (RL)

Cannot find the paper you are looking for? You can Submit a new open access paper.