Policy gradient in robust mdps with global convergence guarantee

Q Wang, CP Ho, M Petrik - International Conference on …, 2023 - proceedings.mlr.press
Abstract Robust Markov decision processes (RMDPs) provide a promising framework for
computing reliable policies in the face of model errors. Many successful reinforcement …

A survey of decision making and optimization under uncertainty

AJ Keith, DK Ahner - Annals of Operations Research, 2021 - Springer
Recent advances in decision making have incorporated both risk and ambiguity in decision
theory and optimization methods. These methods implement a variety of uncertainty …

Fast bellman updates for wasserstein distributionally robust mdps

Z Yu, L Dai, S Xu, S Gao, CP Ho - Advances in Neural …, 2023 - proceedings.neurips.cc
Markov decision processes (MDPs) often suffer from the sensitivity issue under model
ambiguity. In recent years, robust MDPs have emerged as an effective framework to …

Partial policy iteration for l1-robust markov decision processes

CP Ho, M Petrik, W Wiesemann - Journal of Machine Learning Research, 2021 - jmlr.org
Robust Markov decision processes (MDPs) compute reliable solutions for dynamic decision
problems with partially-known transition probabilities. Unfortunately, accounting for …

Beyond confidence regions: Tight bayesian ambiguity sets for robust mdps

M Petrik, RH Russel - Advances in neural information …, 2019 - proceedings.neurips.cc
Abstract Robust MDPs (RMDPs) can be used to compute policies with provable worst-case
guarantees in reinforcement learning. The quality and robustness of an RMDP solution are …

Fast Bellman updates for robust MDPs

CP Ho, M Petrik, W Wiesemann - … Conference on Machine …, 2018 - proceedings.mlr.press
We describe two efficient, and exact, algorithms for computing Bellman updates in robust
Markov decision processes (MDPs). The first algorithm uses a homotopy continuation …

Fast Algorithms for -constrained S-rectangular Robust MDPs

B Behzadian, M Petrik, CP Ho - Advances in Neural …, 2021 - proceedings.neurips.cc
Abstract Robust Markov decision processes (RMDPs) are a useful building block of robust
reinforcement learning algorithms but can be hard to solve. This paper proposes a fast …

Robust satisficing mdps

H Ruan, S Zhou, Z Chen, CP Ho - … Conference on Machine …, 2023 - proceedings.mlr.press
Despite being a fundamental building block for reinforcement learning, Markov decision
processes (MDPs) often suffer from ambiguity in model parameters. Robust MDPs are …

A sufficient statistic for influence in structured multiagent environments

F Oliehoek, S Witwicki, L Kaelbling - Journal of Artificial Intelligence …, 2021 - jair.org
Making decisions in complex environments is a key challenge in artificial intelligence (AI).
Situations involving multiple decision makers are particularly complex, leading to …

Sequential convex programming for the efficient verification of parametric MDPs

M Cubuktepe, N Jansen, S Junges, JP Katoen… - … Conference on Tools …, 2017 - Springer
Multi-objective verification problems of parametric Markov decision processes under
optimality criteria can be naturally expressed as nonlinear programs. We observe that many …