Stochastic networked control systems
Our goal in writing this book has been to provide a comprehensive, mathematically rigorous,
but still accessible treatment of the interaction between information and control in multi …
but still accessible treatment of the interaction between information and control in multi …
[LIBRO][B] Markov decision processes with their applications
Q Hu, W Yue - 2007 - books.google.com
Markov decision processes (MDPs), also called stochastic dynamic programming, were first
studied in the 1960s. MDPs can be used to model and solve dynamic decision-making …
studied in the 1960s. MDPs can be used to model and solve dynamic decision-making …
Further results on exponential estimates of Markovian jump systems with mode-dependent time-varying delays
H Gao, Z Fei, J Lam, B Du - IEEE Transactions on Automatic …, 2010 - ieeexplore.ieee.org
This technical note studies the problem of exponential estimates for Markovian jump
systems with mode-dependent interval time-varying delays. A novel Lyapunov-Krasovskii …
systems with mode-dependent interval time-varying delays. A novel Lyapunov-Krasovskii …
Learning in Markov decision processes under constraints
We consider reinforcement learning (RL) in Markov Decision Processes in which an agent
repeatedly interacts with an environment that is modeled by a controlled Markov process. At …
repeatedly interacts with an environment that is modeled by a controlled Markov process. At …
Constrained-cost adaptive dynamic programming for optimal control of discrete-time nonlinear systems
Q Wei, T Li - IEEE Transactions on Neural Networks and …, 2023 - ieeexplore.ieee.org
For discrete-time nonlinear systems, this research is concerned with optimal control
problems (OCPs) with constrained cost, and a novel value iteration with constrained cost …
problems (OCPs) with constrained cost, and a novel value iteration with constrained cost …
Continuous-time Markov decision processes
The study of continuous-time Markov decision processes dates back at least to the 1950s,
shortly after that of its discrete-time analogue. Since then, the theory has rapidly developed …
shortly after that of its discrete-time analogue. Since then, the theory has rapidly developed …
A convex analytic approach to risk-aware Markov decision processes
In classical Markov decision process (MDP) theory, we search for a policy that, say,
minimizes the expected infinite horizon discounted cost. Expectation is, of course, a risk …
minimizes the expected infinite horizon discounted cost. Expectation is, of course, a risk …
From infinite to finite programs: Explicit error bounds with applications to approximate dynamic programming
We consider linear programming (LP) problems in infinite dimensional spaces that are in
general computationally intractable. Under suitable assumptions, we develop an …
general computationally intractable. Under suitable assumptions, we develop an …
Stochastic nestedness and the belief sharing information pattern
S Yuksel - IEEE Transactions on Automatic Control, 2009 - ieeexplore.ieee.org
Solutions to decentralized stochastic optimization problems lead to recursions in which the
state space enlarges with the time-horizon, thus leading to non-tractability of classical …
state space enlarges with the time-horizon, thus leading to non-tractability of classical …
[LIBRO][B] Finite Approximations in discrete-time stochastic control
Control and optimization of dynamical systems in the presence of stochastic uncertainty is a
mature field with a large range of applications. A comprehensive treatment of such problems …
mature field with a large range of applications. A comprehensive treatment of such problems …