Article citationsMore>>
Silver, D., Huang, A., Maddison, C.J., Guez, A., Sifre, L., Van Den Driessche, G., Schrittwieser, J., Antonoglou, I., Panneershelvam, V., Lanctot, M., et al. (2016) Mastering the Game of Go with Deep Neural Networks and Tree Search. Nature, 529, 484-489.
http://dx.doi.org/10.1038/nature16961
has been cited by the following article:
-
TITLE:
Exploring Deep Reinforcement Learning with Multi Q-Learning
AUTHORS:
Ethan Duryea, Michael Ganger, Wei Hu
KEYWORDS:
Reinforcement Learning, Deep Learning, Multi Q-Learning
JOURNAL NAME:
Intelligent Control and Automation,
Vol.7 No.4,
November
15,
2016
ABSTRACT: Q-learning is a popular temporal-difference reinforcement learning algorithm which often explicitly stores state values using lookup tables. This implementation has been proven to converge to the optimal solution, but it is often beneficial to use a function-approximation system, such as deep neural networks, to estimate state values. It has been previously observed that Q-learning can be unstable when using value function approximation or when operating in a stochastic environment. This instability can adversely affect the algorithm’s ability to maximize its returns. In this paper, we present a new algorithm called Multi Q-learning to attempt to overcome the instability seen in Q-learning. We test our algorithm on a 4 × 4 grid-world with different stochastic reward functions using various deep neural networks and convolutional networks. Our results show that in most cases, Multi Q-learning outperforms Q-learning, achieving average returns up to 2.5 times higher than Q-learning and having a standard deviation of state values as low as 0.58.
Related Articles:
-
Wei Hu, James Hu
-
Rabeb Zarai, Mnaouer Kachout, Mohamed A. G. Hazber, Mohammed A. Mahdi
-
Fahed Awad, Rufaida Shamroukh
-
Hossein Haroonabadi, Hassan Barati
-
Wen Li, Fucang Jia, Qingmao Hu