Monte-Carlo Tree Search is a planning algorithm that accumulates value estimates obtained from Monte Carlo simulations in order to successively direct simulations towards more highly-rewarded trajectories. We execute MCTS after encountering each new state to select an agent's action for that state: it is executed again to select the action for the next state. Each execution is an iterative process that simulates many trajectories starting from the current state to the terminal state. The core idea is to successively focus multiple simulations starting at the current state by extending the initial portions of trajectories that have received high evaluations from earlier simulations.
Source: Sutton and Barto, Reinforcement Learning (2nd Edition)
Image Credit: Chaslot et al
Paper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
Reinforcement Learning (RL) | 47 | 16.15% |
Reinforcement Learning | 44 | 15.12% |
Decision Making | 20 | 6.87% |
Model-based Reinforcement Learning | 20 | 6.87% |
Board Games | 16 | 5.50% |
Deep Reinforcement Learning | 14 | 4.81% |
Atari Games | 12 | 4.12% |
Game of Go | 6 | 2.06% |
Continuous Control | 6 | 2.06% |
Component | Type |
|
---|---|---|
🤖 No Components Found | You can add them if they exist; e.g. Mask R-CNN uses RoIAlign |