Near-optimality for multi-action multi-resource restless bandits with many arms /

We consider multi-action restless bandits with multiple resource constraints, also referred to as weakly coupled Markov decision processes. This problem is important in recommender systems, active learning, revenue management, and many other areas. An optimal policy can be theoretically found by sol...

Full description

Bibliographic Details
Main Author: Zhang, Xiangyu (Author)
Other Authors: Frazier, Peter (Thesis advisor)
Format: Thesis Book
Language:English
Published: Ann Arbor, Michigan : ProQuest Information and Learning, 2022
Subjects:

Internet

Cornell University

Holdings details from Cornell University
Call Number: Thesis T57 2022 Z438