About cookies on this site Our websites require some cookies to function properly (required). In addition, other cookies may be used with your consent to analyze site usage, improve the user experience and for advertising. For more information, please review your options. By visiting our website, you agree to our processing of information as described in IBM’sprivacy statement. To provide a smooth navigation, your cookie preferences will be shared across the IBM web domains listed here.
Publication
AAAI 2020
Workshop paper
Efficient Exploration with Failure Ratio for Deep Reinforcement Learning
Abstract
A combination of Monte Carlo tree search (MCTS) and deep reinforcement learning has demonstrated incredibly high performance and has been attracting much attention lately. However, the convergence of learning is very time-consuming. On the other hand, when we want to acquire skills efficiently, it is important to learn from failure, locating its cause, and modifying the strategy accordingly. Using the analogy of this context, we propose an efficient tree search method by introducing a failure ratio that has high values in important phases. We applied our method to Othello board game. We conducted experiments and showed that our method has a higher winning ratio than the state-of-the-art method, especially in the early stage of learning.