Show simple item record

dc.identifier.urihttp://hdl.handle.net/11401/76062
dc.description.sponsorshipThis work is sponsored by the Stony Brook University Graduate School in compliance with the requirements for completion of degree.en_US
dc.formatMonograph
dc.format.mediumElectronic Resourceen_US
dc.language.isoen_US
dc.publisherThe Graduate School, Stony Brook University: Stony Brook, NY.
dc.typeDissertation
dcterms.abstractThe focus of this work is on practical applications of stochastic multi-armed bandits (MABs) in two distinctive settings. First, we develop and present REGA, a novel adaptive sampling-based algorithm for control of finite-horizon Markov decision processes (MDPs) with very large state spaces and small action spaces. We apply a variant of the epsilon-greedy multi-armed bandit algorithm to each stage of the MDP in a recursive manner, thus computing an estimation of the " reward-to-go" value at each stage of the MDP. We provide a finite-time analysis of REGA. In particular, we provide a bound on the probability that the approximation error exceeds a given threshold, where the bound is given in terms of the number of samples collected at each stage of the MDP. We empirically compare REGA against other sampling-based algorithms and find that our algorithm is competitive. We discuss measures to aid against the curse of dimensionality due to the backwards induction nature of REGA, necessary when the MDP horizon is large. Second, we introduce e-Discovery, a topic of extreme significance to the legal industry, which pertains to the ability of sifting through large volumes of data in order to identify the " needle in the haystack" documents relevant to a lawsuit or investigation. Surprisingly, the topic has not been explicitly investigated in academia. Looking at the problem from a scheduling perspective, we highlight the main properties and challenges pertaining to this topic and outline a formal model for the problem. We examine an approach based on related work from the field of scheduling theory and provide simulation results that demonstrate the performance of our approach against a very large data set. We also provide an approach based on list-scheduling that incorporates a side multi-armed bandit in lieu of standard heuristics. Necessarily, we propose the first MAB algorithm that accounts for both sleeping bandits and bandits with history. The empirical results are encouraging. Surveys of multi-armed bandits as well as scheduling theory are included. Many new and known open problems are proposed and/or documented.
dcterms.available2017-09-18T23:49:57Z
dcterms.contributorArkin, Estheren_US
dcterms.contributorHu, Jiaqiaoen_US
dcterms.contributorDeng, Yuefanen_US
dcterms.contributorOrtiz, Luis.en_US
dcterms.creatorMuqattash, Isa Mithqal
dcterms.dateAccepted2017-09-18T23:49:57Z
dcterms.dateSubmitted2017-09-18T23:49:57Z
dcterms.descriptionDepartment of Applied Mathematics and Statistics.en_US
dcterms.extent118 pg.en_US
dcterms.formatMonograph
dcterms.formatApplication/PDFen_US
dcterms.identifierhttp://hdl.handle.net/11401/76062
dcterms.issued2014-12-01
dcterms.languageen_US
dcterms.provenanceMade available in DSpace on 2017-09-18T23:49:57Z (GMT). No. of bitstreams: 1 Muqattash_grad.sunysb_0771E_12112.pdf: 1705588 bytes, checksum: 06cdc910ffc781300ce04fb3012c18bd (MD5) Previous issue date: 1en
dcterms.publisherThe Graduate School, Stony Brook University: Stony Brook, NY.
dcterms.subjectElectronic Discovery, Markov Decision Process (MDP), Multi-Armed Bandit (MAB), Optimization Under Uncertainties, Sampling, Stochastic Scheduling
dcterms.subjectApplied mathematics
dcterms.titleMulti-Armed Bandits with Applications to Markov Decision Processes and Scheduling Problems
dcterms.typeDissertation


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record