Publication:
Program distribution estimation with grammar models

dc.contributor.author Shan, Yin en_US
dc.date.accessioned 2022-03-22T09:10:20Z
dc.date.available 2022-03-22T09:10:20Z
dc.date.issued 2005 en_US
dc.description.abstract This thesis studies grammar-based approaches in the application of Estimation of Distribution Algorithms (EDA) to the tree representation widely used in Genetic Programming (GP). Although EDA is becoming one of the most active fields in Evolutionary computation (EC), the solution representation in most EDA is a Genetic Algorithms (GA) style linear representation. The more complex tree representations, resembling GP, have received only limited exploration. This is unfortunate, because tree representations provide a natural and expressive way of representing solutions for many problems. This thesis aims to help fill this gap, exploring grammar-based approaches to extending EDA to GP-style tree representations. This thesis firstly provides a comprehensive survey of current research on EDA with emphasis on EDA with GP-style tree representation. The thesis attempts to clarify the relationship between EDA with conventional linear representations and those with a GP-style tree representation, and to reveal the unique difficulties which face this research. Secondly, the thesis identifies desirable properties of probabilistic models for EDA with GP-style tree representation, and derives the PRODIGY framework as a consequence. Thirdly, following the PRODIGY framework, three methods are proposed. The first method is Program Evolution with Explicit Learning (PEEL). Its incremental general-to-specific grammar learning method balances the effectiveness and efficiency of the grammar learning. The second method is Grammar Model-based Program Evolution (GMPE). GMPE realises the PRODIGY framework by introducing elegant inference methods from the formal grammar field. GMPE provides good performance on some problems, but also provides a means to better understand some aspects of conventional GP, especially the building block hypothesis. The third method is Swift GMPE (sGMPE), which is an extension of GMPE, aiming at reducing the computational cost. Fourthly, a more accurate Minimum Message Length metric for grammar learning in PRODIGY is derived in this thesis. This metric leads to improved performance in the GMPE system, but may also be useful in grammar learning in general. It is also relevant to the learning of other probabilistic graphical models. en_US
dc.identifier.uri http://hdl.handle.net/1959.4/38737
dc.language English
dc.language.iso EN en_US
dc.publisher UNSW, Sydney en_US
dc.rights CC BY-NC-ND 3.0 en_US
dc.rights.uri https://creativecommons.org/licenses/by-nc-nd/3.0/au/ en_US
dc.subject.other Distribution algorithms en_US
dc.subject.other estimation of distribution en_US
dc.subject.other explicit learning en_US
dc.subject.other genetic programming en_US
dc.subject.other grammar models en_US
dc.subject.other linear representation en_US
dc.subject.other probabilistic models en_US
dc.subject.other Program Evolution with Explicit learning (PEEL) en_US
dc.subject.other PRODIGY en_US
dc.subject.other tree representation en_US
dc.title Program distribution estimation with grammar models en_US
dc.type Thesis en_US
dcterms.accessRights open access
dcterms.rightsHolder Shan, Yin
dspace.entity.type Publication en_US
unsw.accessRights.uri https://purl.org/coar/access_right/c_abf2
unsw.identifier.doi https://doi.org/10.26190/unsworks/18079
unsw.relation.faculty UNSW Canberra
unsw.relation.originalPublicationAffiliation Shan, Yin, Information Technology & Electrical Engineering, Australian Defence Force Academy, UNSW en_US
unsw.relation.school School of Engineering and Information Technology *
unsw.thesis.degreetype PhD Doctorate en_US
Files
Original bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
whole.pdf
Size:
1.19 MB
Format:
application/pdf
Description:
Resource type