Random Ensemble Mixture (REM) is an easy to implement extension of DQN inspired by Dropout. The key intuition behind REM is that if one has access to multiple estimates of Q-values, then a weighted combination of the Q-value estimates is also an estimate for Q-values. Accordingly, in each training step, REM randomly combines multiple Q-value estimates and uses this random combination for robust training.
Source: An Optimistic Perspective on Offline Reinforcement LearningPaper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
Reinforcement Learning (RL) | 6 | 9.84% |
EEG | 4 | 6.56% |
DQN Replay Dataset | 3 | 4.92% |
Offline RL | 3 | 4.92% |
Computational Efficiency | 2 | 3.28% |
Management | 2 | 3.28% |
Continual Learning | 2 | 3.28% |
Diversity | 2 | 3.28% |
Super-Resolution | 2 | 3.28% |