Article citationsMore>>
Mnih, V., Kavukcuoglu, K., Silver, D., Rusu, A.A., Veness, J., Bellemare, M.G., Graves, A., Riedmiller, M., Fidjeland, A.K., Ostrovski, G., et al. (2015) Human-Level Control through Deep Reinforcement Learning. Nature, 518, 529-533.
http://dx.doi.org/10.1038/nature14236
has been cited by the following article:
-
TITLE:
Exploring Deep Reinforcement Learning with Multi Q-Learning
AUTHORS:
Ethan Duryea, Michael Ganger, Wei Hu
KEYWORDS:
Reinforcement Learning, Deep Learning, Multi Q-Learning
JOURNAL NAME:
Intelligent Control and Automation,
Vol.7 No.4,
November
15,
2016
ABSTRACT: Q-learning is a popular temporal-difference reinforcement learning algorithm which often explicitly stores state values using lookup tables. This implementation has been proven to converge to the optimal solution, but it is often beneficial to use a function-approximation system, such as deep neural networks, to estimate state values. It has been previously observed that Q-learning can be unstable when using value function approximation or when operating in a stochastic environment. This instability can adversely affect the algorithm’s ability to maximize its returns. In this paper, we present a new algorithm called Multi Q-learning to attempt to overcome the instability seen in Q-learning. We test our algorithm on a 4 × 4 grid-world with different stochastic reward functions using various deep neural networks and convolutional networks. Our results show that in most cases, Multi Q-learning outperforms Q-learning, achieving average returns up to 2.5 times higher than Q-learning and having a standard deviation of state values as low as 0.58.
Related Articles:
-
Wei Hu, James Hu
-
Ethan Duryea, Michael Ganger, Wei Hu
-
Jai Dev Chandel, Nand Lal Singh
-
Md Mahmudul Hasan, Ishtiaque Zaman, Miao He, Michael Giesselmann
-
Daniel Schilling Weiss Nguyen, Richard Odigie