{"@context":"http://iiif.io/api/presentation/2/context.json","@id":"https://repo.library.stonybrook.edu/cantaloupe/iiif/2/manifest.json","@type":"sc:Manifest","label":"Machine Learning, Evolutionary Algorithms, and the Inference of Mathematical Truths","metadata":[{"label":"dc.description.sponsorship","value":"This work is sponsored by the Stony Brook University Graduate School in compliance with the requirements for completion of degree."},{"label":"dc.format","value":"Monograph"},{"label":"dc.format.medium","value":"Electronic Resource"},{"label":"dc.identifier.uri","value":"http://hdl.handle.net/11401/77467"},{"label":"dc.language.iso","value":"en_US"},{"label":"dc.publisher","value":"The Graduate School, Stony Brook University: Stony Brook, NY."},{"label":"dcterms.abstract","value":"In this thesis we set out to find whether the true data generating formula behind a set of data points can be automatically inferred from the data points alone. We start with the topic of machine learning and quickly realize that black box models can only approximate the real world which creates the motivation to move on to evolutionary algorithms as a vehicle to implement symbolic regression. Through a series of experiments we discover that the mean-squared error cost function is easily fooled by decoy solutions and is unable to make use of all the information presented in the training examples. Based on this result we develop the concept of feature signatures which uniquely define a set of training examples and possess several desirable properties, the most important being invariance to linear transformations. Armed with this concept we conduct several more numerical experiments based on common analytical functions and real world data sets which ultimately lead to the experimental evidence we need to support the thesis."},{"label":"dcterms.available","value":"2017-09-20T16:52:45Z"},{"label":"dcterms.contributor","value":"Doboli, Alex"},{"label":"dcterms.creator","value":"Hensley, Asher"},{"label":"dcterms.dateAccepted","value":"2017-09-20T16:52:45Z"},{"label":"dcterms.dateSubmitted","value":"2017-09-20T16:52:45Z"},{"label":"dcterms.description","value":"Department of Electrical Engineering."},{"label":"dcterms.extent","value":"219 pg."},{"label":"dcterms.format","value":"Monograph"},{"label":"dcterms.identifier","value":"http://hdl.handle.net/11401/77467"},{"label":"dcterms.issued","value":"2013-12-01"},{"label":"dcterms.language","value":"en_US"},{"label":"dcterms.provenance","value":"Made available in DSpace on 2017-09-20T16:52:45Z (GMT). No. of bitstreams: 1\nHensley_grad.sunysb_0771M_11593.pdf: 1321872 bytes, checksum: cca383630cc1377000e536bd87e5dfe9 (MD5)\n Previous issue date: 1"},{"label":"dcterms.publisher","value":"The Graduate School, Stony Brook University: Stony Brook, NY."},{"label":"dcterms.subject","value":"symbolic regression"},{"label":"dcterms.title","value":"Machine Learning, Evolutionary Algorithms, and the Inference of Mathematical Truths"},{"label":"dcterms.type","value":"Thesis"},{"label":"dc.type","value":"Thesis"}],"description":"This manifest was generated dynamically","viewingDirection":"left-to-right","sequences":[{"@type":"sc:Sequence","canvases":[{"@id":"https://repo.library.stonybrook.edu/cantaloupe/iiif/2/canvas/page-1.json","@type":"sc:Canvas","label":"Page 1","height":1650,"width":1275,"images":[{"@type":"oa:Annotation","motivation":"sc:painting","resource":{"@id":"https://repo.library.stonybrook.edu/cantaloupe/iiif/2/47%2F26%2F41%2F47264108429569569884362442570927609896/full/full/0/default.jpg","@type":"dctypes:Image","format":"image/jpeg","height":1650,"width":1275,"service":{"@context":"http://iiif.io/api/image/2/context.json","@id":"https://repo.library.stonybrook.edu/cantaloupe/iiif/2/47%2F26%2F41%2F47264108429569569884362442570927609896","profile":"http://iiif.io/api/image/2/level2.json"}},"on":"https://repo.library.stonybrook.edu/cantaloupe/iiif/2/canvas/page-1.json"}]}]}]}