Random Ensemble Mixture (REM) is an easy to implement extension of DQN inspired by Dropout. The key intuition behind REM is that if one has access to multiple estimates of Q-values, then a weighted combination of the Q-value estimates is also an estimate for Q-values. Accordingly, in each training step, REM randomly combines multiple Q-value estimates and uses this random combination for robust training.
Source: An Optimistic Perspective on Offline Reinforcement LearningPaper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
Reinforcement Learning (RL) | 6 | 15.38% |
DQN Replay Dataset | 3 | 7.69% |
Offline RL | 3 | 7.69% |
Super-Resolution | 2 | 5.13% |
Atari Games | 2 | 5.13% |
Disentanglement | 1 | 2.56% |
Electroencephalogram (EEG) | 1 | 2.56% |
Management | 1 | 2.56% |
Continual Learning | 1 | 2.56% |