Search Results for author: Thibault Lahire

Found 3 papers, 2 papers with code

Importance Sampling for Stochastic Gradient Descent in Deep Neural Networks

1 code implementation29 Mar 2023 Thibault Lahire

Stochastic gradient descent samples uniformly the training set to build an unbiased gradient estimate with a limited number of samples.

Actor Loss of Soft Actor Critic Explained

no code implementations31 Dec 2021 Thibault Lahire

This technical report is devoted to explaining how the actor loss of soft actor critic is obtained, as well as the associated gradient estimate.

Large Batch Experience Replay

1 code implementation4 Oct 2021 Thibault Lahire, Matthieu Geist, Emmanuel Rachelson

The optimal sampling distribution being intractable, we make several approximations providing good results in practice and introduce, among others, LaBER (Large Batch Experience Replay), an easy-to-code and efficient method for sampling the replay buffer.

Atari Games Reinforcement Learning (RL)

Cannot find the paper you are looking for? You can Submit a new open access paper.