Please use this identifier to cite or link to this item: http://bura.brunel.ac.uk/handle/2438/30166
Full metadata record
DC FieldValueLanguage
dc.contributor.authorWang, H-
dc.contributor.authorTang, Z-
dc.contributor.authorSun, Y-
dc.contributor.authorWang, F-
dc.contributor.authorZhang, S-
dc.contributor.authorChen, Y-
dc.date.accessioned2024-11-18T11:58:42Z-
dc.date.available2024-11-18T11:58:42Z-
dc.date.issued2024-08-12-
dc.identifierORCiD: Haoran Wang https://orcid.org/0000-0002-4622-0119-
dc.identifierORCiD: Zeshen Tang https://orcid.org/0000-0001-8765-6464-
dc.identifierORCiD: Fang Wang https://orcid.org/0000-0003-1987-9150-
dc.identifierORCiD: Siyu Zhang https://orcid.org/0000-0002-0001-0204-
dc.identifierORCiD: Yeming Chen https://orcid.org/0009-0005-5515-1943-
dc.identifier.citationWang, H. et al. (2024) 'Guided Cooperation in Hierarchical Reinforcement Learning via Model-Based Rollout', IEEE Transactions on Neural Networks and Learning Systems, 36 (5), pp. 8455 - 8469. doi: 10.1109/tnnls.2024.3425809.en_US
dc.identifier.issn2162-237X-
dc.identifier.urihttps://bura.brunel.ac.uk/handle/2438/30166-
dc.descriptionA preprint version of the article is available at arXiv:2309.13508v2 [cs.LG], https://arxiv.org/abs/2309.13508 ([v2] Sat, 6 Apr 2024 17:07:13 UTC (4,747 KB)) . It is archived on this institutional repository but it has not been certified by peer review. Comments: Resubmitted a revised version, in which we provided more illustrative examples, corrected the writing errors, and added references.en_US
dc.descriptionThis article has supplementary downloadable material available at https://doi.org/10.1109/TNNLS.2024.3425809, provided by the authors.-
dc.description.abstractGoal-conditioned hierarchical reinforcement learning (HRL) presents a promising approach for enabling effective exploration in complex, long-horizon reinforcement learning (RL) tasks through temporal abstraction. Empirically, heightened interlevel communication and coordination can induce more stable and robust policy improvement in hierarchical systems. Yet, most existing goal-conditioned HRL algorithms have primarily focused on the subgoal discovery, neglecting interlevel cooperation. Here, we propose a novel goal-conditioned HRL framework named Guided Cooperation via Model-Based Rollout (GCMR; code is available at https://github.com/HaoranWang-TJ/GCMR_ACLG_official), aiming to bridge interlayer information synchronization and cooperation by exploiting forward dynamics. First, the GCMR mitigates the state-transition error within off-policy correction via model-based rollout, thereby enhancing sample efficiency. Second, to prevent disruption by the unseen subgoals and states, lower level Q -function gradients are constrained using a gradient penalty with a model-inferred upper bound, leading to a more stable behavioral policy conducive to effective exploration. Third, we propose a one-step rollout-based planning, using higher level critics to guide the lower level policy. Specifically, we estimate the value of future states of the lower level policy using the higher level critic function, thereby transmitting global task information downward to avoid local pitfalls. These three critical components in GCMR are expected to facilitate interlevel cooperation significantly. Experimental results demonstrate that incorporating the proposed GCMR framework with a disentangled variant of hierarchical reinforcement learning guided by landmarks (HIGL), namely, adjacency constraint and landmark-guided planning (ACLG), yields more stable and robust policy improvement compared with various baselines and significantly outperforms previous state-of-the-art (SOTA) algorithms.en_US
dc.format.extent8455 - 8469-
dc.format.mediumPrint-Electronic-
dc.languageEnglish-
dc.language.isoen_USen_US
dc.publisherInstitute of Electrical and Electronics Engineers (IEEE)en_US
dc.relation.urihttps://arxiv.org/abs/2309.13508-
dc.rightsCopyright © 2024 Institute of Electrical and Electronics Engineers (IEEE). Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. See: https://journals.ieeeauthorcenter.ieee.org/become-an-ieee-journal-author/publishing-ethics/guidelines-and-policies/post-publication-policies/-
dc.rights.urihttps://journals.ieeeauthorcenter.ieee.org/become-an-ieee-journal-author/publishing-ethics/guidelines-and-policies/post-publication-policies/-
dc.subjectdeep reinforcement learning (DRL)en_US
dc.subjectgoal conditioningen_US
dc.subjecthierarchical reinforcement learning (HRL)en_US
dc.subjectinterlevel cooperationen_US
dc.subjectmodel-based rollouten_US
dc.titleGuided Cooperation in Hierarchical Reinforcement Learning via Model-Based Rollouten_US
dc.typeArticleen_US
dc.date.dateAccepted2024-06-30-
dc.identifier.doihttps://doi.org/10.1109/tnnls.2024.3425809-
dc.relation.isPartOfIEEE Transactions on Neural Networks and Learning Systems-
pubs.issue5-
pubs.publication-statusPublished-
pubs.volume36-
dc.identifier.eissn2162-2388-
dc.rights.holderInstitute of Electrical and Electronics Engineers (IEEE)-
Appears in Collections:Dept of Computer Science Research Papers

Files in This Item:
File Description SizeFormat 
Preprint.pdfCopyright © 2024 Institute of Electrical and Electronics Engineers (IEEE). Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. See: https://journals.ieeeauthorcenter.ieee.org/become-an-ieee-journal-author/publishing-ethics/guidelines-and-policies/post-publication-policies/ .5.95 MBAdobe PDFView/Open


Items in BURA are protected by copyright, with all rights reserved, unless otherwise indicated.