Using Continuous Action Spaces to Solve Discrete Problems

  • Hado van Hasselt
  • , Marco Wiering

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

55 Citations (Scopus)

Abstract

Real-world control problems are often modeled as Markov Decision Processes (MDPs) with discrete action spaces to facilitate the use of the many reinforcement learning algorithms that exist to find solutions for such MDPs. For many of these problems an underlying continuous action space can be assumed. We investigate the performance of the Cacla algorithm, which uses a continuous actor, on two such MDPs: the mountain car and the cart pole. We show that Cacla has clear advantages over discrete algorithms such as Q-learning and Sarsa, even though its continuous actions get rounded to actions in the same finite action space that may contain only a small number of actions. In particular, we show that Cacla retains much better performance when the action space is changed by removing some actions after some time of learning.
Original languageEnglish
Title of host publicationProceedings of the International Joint Conference on Neural Networks
Subtitle of host publicationIJCNN
Number of pages8
Publication statusPublished - 2009

Keywords

  • Reinforcement learning
  • Machine learning

Fingerprint

Dive into the research topics of 'Using Continuous Action Spaces to Solve Discrete Problems'. Together they form a unique fingerprint.

Cite this