Please use this identifier to cite or link to this item: http://bura.brunel.ac.uk/handle/2438/18020
Full metadata record
DC FieldValueLanguage
dc.contributor.authorShi, H-
dc.contributor.authorSun, Y-
dc.contributor.authorLi, G-
dc.contributor.authorWang, F-
dc.contributor.authorWang, D-
dc.contributor.authorLi, J-
dc.date.accessioned2019-05-08T16:52:39Z-
dc.date.available2019-03-19-
dc.date.available2019-05-08T16:52:39Z-
dc.date.issued2019-03-19-
dc.identifier.citationShi, H., Sun, Y., Li, G., Wang,F., Wang, D. and Li, J. (2019) 'Hierarchical Intermittent Motor Control With Deterministic Policy Gradient,' IEEE Access, 7, pp. 41799-41810. doi: 10.1109/ACCESS.2019.2904910.en_US
dc.identifier.urihttps://bura.brunel.ac.uk/handle/2438/18020-
dc.description© 2019 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.en_US
dc.description.abstractIt has been evidenced that the neural motor control exploits the hierarchical and intermittent representation. In this paper, we propose a hierarchical deep reinforcement learning (DRL) method to learn the continuous control policy across multiple levels, by unifying the neuroscience principle of the minimum transition hypothesis. The control policies in the two levels of the hierarchy operate at different time scales. The high-level controller produces the intermittent actions to set a sequence of goals for the low-level controller, which in turn conducts the basic skills with the modulation of goals. The goal planning and the basic motor skills are trained jointly with the proposed algorithm: hierarchical intermittent deep deterministic policy gradient (HI-DDPG). The performance of the method is validated in two continuous control problems. The results show that the method successfully learns to temporally decompose compound tasks into sequences of basic motions with sparse transitions and outperforms the previous DRL methods that lack a hierarchical continuous representation.en_US
dc.description.sponsorship10.13039/501100001809-National Natural Science Foundation of China; 10.13039/501100003399-Science and Technology Commission of Shanghai Municipality; 10.13039/100007219-Natural Science Foundation of Shanghai;-
dc.description.sponsorship10.13039/501100001809-National Natural Science Foundation of China; 10.13039/501100003399-Science and Technology Commission of Shanghai Municipality; 10.13039/100007219-Natural Science Foundation of Shanghai;-
dc.format.extent41799 - 41810-
dc.language.isoenen_US
dc.publisherIEEEen_US
dc.rightsThis journal is 100% open access, which means that all content is freely available without charge to users or their institutions. All articles are currently published under Creative Commons licenses (either CCBY or CCBY-NC-ND), and the author retains copyright. Users are allowed to read, download, copy, distribute, print, search, or link to the full texts of the articles published under CCBY, or use them for any other lawful purpose, as long as proper attribution is given. Articles published under CCBY-NC-ND are also available to users under the same conditions as CCBY, but the reuse cannot be for commercial purposes or change the work in any way.-
dc.subjecthierarchical reinforcement learningen_US
dc.subjectintermittent controlen_US
dc.subjectdeterministic policy gradienten_US
dc.subjectcontinuous action control, motor controlen_US
dc.titleHierarchical Intermittent Motor Control with Deterministic Policy Gradienten_US
dc.typeArticleen_US
dc.identifier.doihttps://doi.org/10.1109/ACCESS.2019.2904910-
dc.relation.isPartOfIEEE Access-
pubs.publication-statusPublished-
pubs.volume7-
dc.identifier.eissn2169-3536-
Appears in Collections:Dept of Computer Science Research Papers

Files in This Item:
File Description SizeFormat 
FullText.pdf4.91 MBAdobe PDFView/Open


Items in BURA are protected by copyright, with all rights reserved, unless otherwise indicated.