Publication:
Deep Reinforcement Learning for Continuous Action Control

dc.contributor.advisor Kasmarik, Kathryn en_US
dc.contributor.advisor Abbass, Hussein en_US
dc.contributor.author Yang, Zhaoyang en_US
dc.date.accessioned 2022-03-22T16:30:21Z
dc.date.available 2022-03-22T16:30:21Z
dc.date.issued 2017 en_US
dc.description.abstract Deep reinforcement learning has greatly improved the performance of learning agent by combining the strong generalization and extraction ability of deep learning models with the bootstrapping nature of reinforcement learning. Many works have achieved unprecedented results, especially on discrete action control tasks. However, much less work has been done to deal with robotic control in a continuous action space. All single-thread based algorithms in this domain can only control the robot to solve basic tasks, and only one task at a time. This thesis aims to make up for these limitations. We first proposed a novel deep reinforcement learning network architecture that can reduce the number of parameters needed for learning a single basic skill in a continuous action space by more than 70%. We then proposed a novel multi-task deep reinforcement learning algorithm to learn multiple basic tasks simultaneously. It makes use of the proposed network architecture to reduce the number of parameters needed for learning multiple tasks by more than 80%. Finally, we proposed a novel hierarchical deep reinforcement learning algorithm which consists of two levels of hierarchy. It adapted the proposed multi-task learning algorithm in its first level of hierarchy to learn multiple basic skills and then learn to reuse these skills in its second level of hierarchy to solve compound tasks. We conducted several sets of experiments to test both the proposed network architecture and the algorithms with a simulated Pioneer 3AT robot in Gazebo 2 in a ROS Indigo environment. Results show that agents built with the proposed network architecture can learn skills that are as good as the ones learned by agents built with traditional convolutional neural networks. Also, all basic skills learned by the proposed multi-task learning algorithm achieve comparable performance to the skills learned independently by single-task learning algorithm. Results also show that the proposed hierarchical learning algorithm can learn both high performance basic skills and compound skills within the same learning process. The performance of the proposed algorithm on solving compound tasks has outperforms both a state-of-the-art single-thread based continuous action control algorithm and a well-known discrete action control algorithm. en_US
dc.identifier.uri http://hdl.handle.net/1959.4/59035
dc.language English
dc.language.iso EN en_US
dc.publisher UNSW, Sydney en_US
dc.rights CC BY-NC-ND 3.0 en_US
dc.rights.uri https://creativecommons.org/licenses/by-nc-nd/3.0/au/ en_US
dc.subject.other Continuous Action Control en_US
dc.subject.other Deep Learning en_US
dc.subject.other Reinforcement Learning en_US
dc.title Deep Reinforcement Learning for Continuous Action Control en_US
dc.type Thesis en_US
dcterms.accessRights open access
dcterms.rightsHolder Yang, Zhaoyang
dspace.entity.type Publication en_US
unsw.accessRights.uri https://purl.org/coar/access_right/c_abf2
unsw.identifier.doi https://doi.org/10.26190/unsworks/20144
unsw.relation.faculty UNSW Canberra
unsw.relation.originalPublicationAffiliation Yang, Zhaoyang, Engineering & Information Technology, UNSW Canberra, UNSW en_US
unsw.relation.originalPublicationAffiliation Kasmarik, Kathryn, Engineering & Information Technology, UNSW Canberra, UNSW en_US
unsw.relation.originalPublicationAffiliation Abbass, Hussein, Engineering & Information Technology, UNSW Canberra, UNSW en_US
unsw.relation.school School of Engineering and Information Technology *
unsw.thesis.degreetype Masters Thesis en_US
Files
Original bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
public version.pdf
Size:
5.49 MB
Format:
application/pdf
Description:
Resource type