Exploring the limits of hierarchical world models in reinforcement learning

  • Hierarchical model-based reinforcement learning (HMBRL) aims to combine the sample efficiency of model-based reinforcement learning with the abstraction capability of hierarchical reinforcement learning. While HMBRL has great potential, the structural and conceptual complexities of current approaches make it challenging to extract general principles, hindering understanding and adaptation to new use cases, and thereby impeding the overall progress of the field. In this work we describe a novel HMBRL framework and evaluate it thoroughly. We construct hierarchical world models that simulate the environment at various levels of temporal abstraction. These models are used to train a stack of agents that communicate top-down by proposing goals to their subordinate agents. A significant focus of this study is the exploration of a static and environment agnostic temporal abstraction, which allows concurrent training of models and agents throughout the hierarchy. Unlike most goal-conditioned H(MB)RL approaches, it also leads to comparatively low dimensional abstract actions. Although our HMBRL approach did not outperform traditional methods in terms of final episode returns, it successfully facilitated decision-making across two levels of abstraction. A central challenge in enhancing our method's performance, as uncovered through comprehensive experimentation, is model exploitation on the abstract level of our world model stack. We provide an in depth examination of this issue, discussing its implications and suggesting directions for future research to overcome this challenge. By sharing these findings, we aim to contribute to the broader discourse on refining HMBRL methodologies.

Download full text files

Export metadata

Additional Services

Share in Twitter Search Google Scholar
Metadaten
Author:Robin SchiewerGND, Anand SubramoneyGND, Laurenz WiskottORCiDGND
URN:urn:nbn:de:hbz:294-118252
DOI:https://doi.org/10.1038/s41598-024-76719-w
Parent Title (English):Scientific reports
Publisher:Springer Nature
Place of publication:London
Document Type:Article
Language:English
Date of Publication (online):2025/03/13
Date of first Publication:2024/11/06
Publishing Institution:Ruhr-Universität Bochum, Universitätsbibliothek
Tag:Open Access Fonds
Goal-conditioned reinforcement learning; Hierarchical reinforcement learning; Model exploitation; Model-based reinforcement learning; Temporal abstraction
Volume:14
Issue:Article 26856
First Page:26856-1
Last Page:26856-24
Note:
Dieser Beitrag ist auf Grund des DEAL-Springer-Vertrages frei zugänglich.
Note:
Article Processing Charge funded by the Deutsche Forschungsgemeinschaft (DFG) and the Open Access Publication Fund of Ruhr-Universität Bochum.
Institutes/Facilities:Institut für Neuroinformatik
Dewey Decimal Classification:Allgemeines, Informatik, Informationswissenschaft / Informatik
open_access (DINI-Set):open_access
faculties:Fakultät für Informatik
Licence (English):License LogoCreative Commons - CC BY 4.0 - Attribution 4.0 International