Skip to main navigation Skip to search Skip to main content

Human feedback in continuous actor-critic reinforcement learning

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

16 Scopus citations

Abstract

Reinforcement learning is utilized in contexts where an agent tries to learn from the environment. Using continuous actions, the performance may be improved in comparison to using discrete actions, however, this leads to excessive time to find a proper policy. In this work, we focus on including human feedback in reinforcement learning for a continuous action space. We unify the policy and the feedback to favor actions of low probability density. Furthermore, we compare the performance of the feedback for the continuous actor-critic algorithm and test our experiments in the cart-pole balancing task. The obtained results show that the proposed approach increases the accumulated reward in comparison to the autonomous learning method.

Original languageEnglish
Title of host publicationESANN 2019 - Proceedings, 27th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning
PublisherESANN (i6doc.com)
Pages661-666
Number of pages6
ISBN (Electronic)9782875870650
StatePublished - 2019
Event27th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning, ESANN 2019 - Bruges, Belgium
Duration: 24 Apr 201926 Apr 2019

Publication series

NameESANN 2019 - Proceedings, 27th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning

Conference

Conference27th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning, ESANN 2019
Country/TerritoryBelgium
CityBruges
Period24/04/1926/04/19

Fingerprint

Dive into the research topics of 'Human feedback in continuous actor-critic reinforcement learning'. Together they form a unique fingerprint.

Cite this