Show simple item record

dc.identifier.urihttp://hdl.handle.net/1951/59730
dc.identifier.urihttp://hdl.handle.net/11401/71296
dc.description.sponsorshipThis work is sponsored by the Stony Brook University Graduate School in compliance with the requirements for completion of degree.en_US
dc.formatMonograph
dc.format.mediumElectronic Resourceen_US
dc.language.isoen_US
dc.publisherThe Graduate School, Stony Brook University: Stony Brook, NY.
dc.typeDissertation
dcterms.abstractThis dissertation focuses on training autonomous agents to plan and act under uncertainty, specifically for cases where the underlying state spaces are continuous in nature. Partially Observable Markov Decision Processes (POMDPs) are a class of models aimed at training agents to seek high rewards or low costs while navigating a state space without knowing their true location. Information regarding an agent's location is gathered in the form of possibly nonlinear and noisy measurements as a function of the true location. An exactly solved POMDP allows an agent to optimally balance seeking rewards and seeking information regarding its position in state space. It is computationally intractable to solve POMDPs for state domains that are continuous, motivating the need for efficient approximate solutions. The algorithm considered in this thesis is the Parametric POMDP (PPOMDP) method. PPOMDP represents an agent's knowledge as a parameterised probability distribution and is able to infer the impact of future actions and observations. The contribution of this thesis is in enhancing the PPOMDP algorithm making significant improvements in training and plan execution times. Several aspects of the original algorithm are generalized and the impact on training time, execution time, and performance are measured on a variety of classic robot navigation models in the literature today. In addition, a mathematically principled threefold adaptive sampling scheme is implemented. With an adaptive sampling scheme the algorithm automatically varies sampling according to the complexity of posterior distributions. Finally, a forward search algorithm is proposed to improve execution performance for sparse belief sets by searching several ply deeper than allowed by previous implementations.
dcterms.available2013-05-22T17:34:56Z
dcterms.available2015-04-24T14:46:55Z
dcterms.contributorXing, Haipengen_US
dcterms.contributorZhu, Weien_US
dcterms.contributorHu, Jiaqiaoen_US
dcterms.contributorZhang, Minghua.en_US
dcterms.creatorKnapik, Timothy Ryan
dcterms.dateAccepted2013-05-22T17:34:56Z
dcterms.dateAccepted2015-04-24T14:46:55Z
dcterms.dateSubmitted2013-05-22T17:34:56Z
dcterms.dateSubmitted2015-04-24T14:46:55Z
dcterms.descriptionDepartment of Applied Mathematics and Statisticsen_US
dcterms.extent132 pg.en_US
dcterms.formatApplication/PDFen_US
dcterms.formatMonograph
dcterms.identifierKnapik_grad.sunysb_0771E_11143en_US
dcterms.identifierhttp://hdl.handle.net/1951/59730
dcterms.identifierhttp://hdl.handle.net/11401/71296
dcterms.issued2012-12-01
dcterms.languageen_US
dcterms.provenanceMade available in DSpace on 2013-05-22T17:34:56Z (GMT). No. of bitstreams: 1 Knapik_grad.sunysb_0771E_11143.pdf: 1433740 bytes, checksum: 4ae62e04bfae35ab2298f5030910c931 (MD5) Previous issue date: 1en
dcterms.provenanceMade available in DSpace on 2015-04-24T14:46:55Z (GMT). No. of bitstreams: 3 Knapik_grad.sunysb_0771E_11143.pdf.jpg: 1894 bytes, checksum: a6009c46e6ec8251b348085684cba80d (MD5) Knapik_grad.sunysb_0771E_11143.pdf.txt: 183114 bytes, checksum: 142d64b0fdc9646abd11ab7db6cd9f81 (MD5) Knapik_grad.sunysb_0771E_11143.pdf: 1433740 bytes, checksum: 4ae62e04bfae35ab2298f5030910c931 (MD5) Previous issue date: 1en
dcterms.publisherThe Graduate School, Stony Brook University: Stony Brook, NY.
dcterms.subjectMathematics
dcterms.subjectContinuous State Space, Partially Observable Markov Decision Processes
dcterms.titleApproximating Partially Observable Markov Decision Processes with Parametric Belief Distributions for Continuous State Spaces
dcterms.typeDissertation


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record