Ir directamente a la navegación principal Ir directamente a la búsqueda Ir directamente al contenido principal

Elastic step DDPG: Multi-step reinforcement learning for improved sample efficiency

  • Adrian Ly
  • , Richard Dazeley
  • , Peter Vamplew
  • , Francisco Cruz
  • , Sunil Aryal
  • Deakin University
  • Federation University Australia
  • UNSW Sydney

Producción científica: Capítulo del libro/informe/acta de congresoContribución a la conferenciarevisión exhaustiva

4 Citas (Scopus)

Resumen

A major challenge in deep reinforcement learning is that it requires more data to converge to an policy for complex problems. One way to improve sample efficiency is to use n-step updates to reduce the number of samples required to converge to a good policy. However n-step updates are known to be brittle and difficult to tune. Elastic Step DQN has shown that it is possible to automate the value of n in DQN to solve problems involving discrete action spaces, however the efficacy of the technique when applied on more complex problems and against problems with continuous action spaces is yet to be shown. In this paper we adapt the innovations proposed by Elastic Step DQN onto the DDPG algorithm and show empirically that Elastic Step DDPG is able to achieve a much stronger final training policy and is more sample efficient than DDPG.

Idioma originalInglés
Título de la publicación alojadaIJCNN 2023 - International Joint Conference on Neural Networks, Proceedings
EditorialInstitute of Electrical and Electronics Engineers Inc.
ISBN (versión digital)9781665488679
DOI
EstadoPublicada - 2023
Publicado de forma externa

Serie de la publicación

NombreProceedings of the International Joint Conference on Neural Networks
Volumen2023-June

Huella

Profundice en los temas de investigación de 'Elastic step DDPG: Multi-step reinforcement learning for improved sample efficiency'. En conjunto forman una huella única.

Citar esto