Averaged-DQN: Variance reduction and stabilization for Deep Reinforcement Learning

Oron Ansehel, Nir Baram, Nahum Shimkin

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Instability and variability of Deep Reinforcement Learning (DRL) algorithms tend to adversely affect their performance. Averaged-DQN is a simple extension to the DQN algorithm, based on averaging previously learned Q-values estimates, which leads to a more stable training procedure and improved performance by reducing approximation error variance in the target values. To understand the effect of the algorithm, we examine the source of value function estimation errors and provide an analytical comparison within a simplified model. We further present experiments on the Arcade Learning Environment benchmark that demonstrate significantly improved stability and performance due to the proposed extension.

Original languageEnglish
Title of host publication34th International Conference on Machine Learning, ICML 2017
Pages240-253
Number of pages14
ISBN (Electronic)9781510855144
StatePublished - 2017
Event34th International Conference on Machine Learning, ICML 2017 - Sydney, Australia
Duration: 6 Aug 201711 Aug 2017

Publication series

Name34th International Conference on Machine Learning, ICML 2017
Volume1

Conference

Conference34th International Conference on Machine Learning, ICML 2017
Country/TerritoryAustralia
CitySydney
Period6/08/1711/08/17

All Science Journal Classification (ASJC) codes

  • Software
  • Human-Computer Interaction
  • Computational Theory and Mathematics

Fingerprint

Dive into the research topics of 'Averaged-DQN: Variance reduction and stabilization for Deep Reinforcement Learning'. Together they form a unique fingerprint.

Cite this