Opposition-based Q(λ) algorithm

Maryam Shokri, Hamid R. Tizhoosh, Mohamed Kamel

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

The problem of delayed reward in reinforcement learning is usually tackled by implementing the mechanism of eligibility traces. In this paper we introduce an extension of eligibility traces to solve one of the challenging problems in reinforcement learning. The concept of opposition traces is proposed in this work to deal with large state space problems in reinforcement learning applications. We combine the idea of opposition and eligibility traces to construct the oppositionbased Q(λ). The results are compared with the conventional Watkins' Q(λ) and reflect a remarkable performance increase.

Original languageEnglish (US)
Title of host publicationInternational Joint Conference on Neural Networks 2006, IJCNN '06
Pages254-261
Number of pages8
StatePublished - 2006
EventInternational Joint Conference on Neural Networks 2006, IJCNN '06 - Vancouver, BC, Canada
Duration: Jul 16 2006Jul 21 2006

Publication series

NameIEEE International Conference on Neural Networks - Conference Proceedings
ISSN (Print)1098-7576

Conference

ConferenceInternational Joint Conference on Neural Networks 2006, IJCNN '06
Country/TerritoryCanada
CityVancouver, BC
Period7/16/067/21/06

ASJC Scopus subject areas

  • Software

Fingerprint

Dive into the research topics of 'Opposition-based Q(λ) algorithm'. Together they form a unique fingerprint.

Cite this