Journal of Data Analysis and Information Processing

Volume 4, Issue 4 (November 2016)

ISSN Print: 2327-7211   ISSN Online: 2327-7203

Google-based Impact Factor: 1.59  Citations  

Double Sarsa and Double Expected Sarsa with Shallow and Deep Learning

HTML  XML Download Download as PDF (Size: 5404KB)  PP. 159-176  
DOI: 10.4236/jdaip.2016.44014    2,891 Downloads   7,696 Views  Citations

ABSTRACT

Double Q-learning has been shown to be effective in reinforcement learning scenarios when the reward system is stochastic. We apply the idea of double learning that this algorithm uses to Sarsa and Expected Sarsa, producing two new algorithms called Double Sarsa and Double Expected Sarsa that are shown to be more robust than their single counterparts when rewards are stochastic. We find that these algorithms add a significant amount of stability in the learning process at only a minor computational cost, which leads to higher returns when using an on-policy algorithm. We then use shallow and deep neural networks to approximate the actionvalue, and show that Double Sarsa and Double Expected Sarsa are much more stable after convergence and can collect larger rewards than the single versions.

Share and Cite:

Ganger, M. , Duryea, E. and Hu, W. (2016) Double Sarsa and Double Expected Sarsa with Shallow and Deep Learning. Journal of Data Analysis and Information Processing, 4, 159-176. doi: 10.4236/jdaip.2016.44014.

Cited by

[1] Meta-Active Learning in Probabilistically Safe Optimization
IEEE Robotics and …, 2022
[2] Investigation of Maximization Bias in Sarsa Variants
2021 IEEE Symposium Series on …, 2021
[3] Market Making in a Limit Order Book: Classical Optimal Control and Reinforcement Learning Approaches
2021
[4] 基于 FM 与 DQN 结合的视频推荐算法
计算机与数字工程, 2021
[5] An Efficient Impersonation Attack Detection Method in Fog Computing
2021
[6] Servo Health Monitoring Based on Feature Learning via Deep Neural Network
IEEE …, 2021
[7] Handbook of automated scoring: Theory into practice
2020
[8] Robust control under worst‐case uncertainty for unknown nonlinear systems using modified reinforcement learning
2020
[9] Meta-active Learning in Probabilistically-Safe Optimization
2020
[10] A New Reinforcement Learning Algorithm Based on Counterfactual Experience Replay
2020
[11] Functional test generation from UI test scenarios using reinforcement learning for android applications
2020
[12] UAV Landmark Detection Based on Convolutional Neural Network
2020
[13] Reinforcement Learning Approach for Inspect/Correct Tasks
2020
[14] Reinforcement Learning-Driven Test Generation for Android GUI Applications using Formal Specifications
2019
[15] An information network security policy learning algorithm based on Sarsa with optimistic initial values
2019
[16] Training a Quantum Neural Network to Solve the Contextual Multi-Armed Bandit Problem
2019
[17] Q Learning with Quantum Neural Networks
2019
[18] Reinforcement Learning with Deep Quantum Neural Networks
2019
[19] Distributional Reinforcement Learning with Quantum Neural Networks
2019
[20] Deep Reinforcement Learning with Adaptive Update Target Combination
2019
[21] Deep Reinforcement Learning with Applications in Transportation
2019
[22] Empirical Analysis of Decision Making of an AI Agent on IBM's 5Q Quantum Computer
2018
[23] Causal Analytics and Risk Analytics
Causal Analytics for Applied Risk Analysis, 2018
[24] Ensemble Network Architecture for Deep Reinforcement Learning
Mathematical Problems in Engineering, 2018
[25] Deep Reinforcement Learning with Sarsa and Q-Learning: A Hybrid Approach
2018
[26] 基于重抽样优选缓存经验回放机制的深度强化学习方法
2018
[27] User Feedback-Based Reinforcement Learning for Vehicle Comfort Control
2018
[28] Double Sarsa Based Machine Learning to Improve Quality of Video Streaming over HTTP through Wireless Networks
2018
[29] Reinforcement Learning in R
2017
[30] Hluboké učení v počítačových hrách

Copyright © 2024 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.