Show simple item record

dc.contributor.authorHu, Weici
dc.identifier.otherbibid: 10361629
dc.description.abstractWe consider a class of stochastic sequential allocation problems - restless multi-armed bandits (RMAB) with a finite horizon and multiple pulls per period. Leveraging the Lagrangian relaxation of the problem, we propose an index-based policy that uses the optimal Lagrange multipliers to index individual arms, and prove that the policy is asymptotically optimal as the number of arms tends to infinity. We also demonstrate numerically that this index-based policy outperforms state-of-the-art heuristics in several instances of RMAB. In addition, we study two other applications of sequential resource allocation problems which are extensions of the RMAB problem, and demonstrate how our index policy can be adapted to these settings.
dc.subjectIndex-based Policy
dc.subjectRestless Bandit
dc.subjectSequential Resource Allocation
dc.subjectStochastic Dynamic Program
dc.subjectOperations research
dc.titleSequential Resource Allocation Under Uncertainty: An Index Policy Approach
dc.typedissertation or thesis Research University of Philosophy D., Operations Research
dc.contributor.chairFrazier, Peter
dc.contributor.committeeMemberTopaloglu, Huseyin
dc.contributor.committeeMemberJoachims, Thorsten

Files in this item


This item appears in the following Collection(s)

Show simple item record