WiOpt 2021 Workshops
Workshop on Reinforcement Learning and Stochastic Control in Queues and Networks
Session 1
Approximate planning and learning for partially observed systems
Aditya Mahajan (McGill University, Canada)
that this approach works well in practice. Joint work with Jayakumar Subramanian, Amit Sinha, and Raihan Seraj.
Approximation Benefits of Policy Gradient Methods with Aggregated States
Dan Russo (Columbia University, USA)
Towards an approximate information state for multi-agent RL problems
Vijay Subramanian (University of Michigan, USA)
Compressive state representation learning towards small-data RL applications
Mengdi Wang (Princeton University, USA)
Online Reinforcement Learning for MDPs and POMDPs via Posterior Sampling
Rahul Jain (University of Southern California, USA)
Session Chair
Vijay Subramanian (University of Michigan, USA)
Virtual Lunch Break
Session 2
Learning algorithm for optimal network control
Eytan Modiano (MIT, USA)
Data-Driven Stochastic Network Control via Reinforcement Learning
Qiaomin Xie (University of Wisconsin–Madison, USA)
Given a stable policy, we further develop a model-based RL method and prove that it converges to the optimal policy. Our method demonstrates promising performance in a variety of network control problems including routing, dynamic server allocation and switch scheduling.
Job Dispatching Policies for Queueing Systems with Unknown Service Rates
Weina Wang (Carnegie Mellon University, USA)
Learning based meta-scheduling in wireless networks
Sanjay Shakkottai (The University of Texas at Austin, USA)
A Provably-Efficient Model-Free Algorithm for Constrained Markov Decision Processes
Lei Ying (University of Michigan, USA)
Session Chair
Gauri Joshi and Weina Wang (Carnegie Mellon University, USA)
Made with in Toronto · Privacy Policy · © 2022 Duetone Corp.