Stochastic networked control systems

S Yüksel, T Basar - AMC, 2013 - Springer
Our goal in writing this book has been to provide a comprehensive, mathematically rigorous,
but still accessible treatment of the interaction between information and control in multi …

[图书][B] Markov decision processes with their applications

Q Hu, W Yue - 2007 - books.google.com
Markov decision processes (MDPs), also called stochastic dynamic programming, were first
studied in the 1960s. MDPs can be used to model and solve dynamic decision-making …

Learning in Markov decision processes under constraints

R Singh, A Gupta, NB Shroff - arXiv preprint arXiv:2002.12435, 2020 - arxiv.org
We consider reinforcement learning (RL) in Markov Decision Processes in which an agent
repeatedly interacts with an environment that is modeled by a controlled Markov process. At …

Further results on exponential estimates of Markovian jump systems with mode-dependent time-varying delays

H Gao, Z Fei, J Lam, B Du - IEEE Transactions on Automatic …, 2010 - ieeexplore.ieee.org
This technical note studies the problem of exponential estimates for Markovian jump
systems with mode-dependent interval time-varying delays. A novel Lyapunov-Krasovskii …

Constrained-cost adaptive dynamic programming for optimal control of discrete-time nonlinear systems

Q Wei, T Li - IEEE Transactions on Neural Networks and …, 2023 - ieeexplore.ieee.org
For discrete-time nonlinear systems, this research is concerned with optimal control
problems (OCPs) with constrained cost, and a novel value iteration with constrained cost …

Continuous-time Markov decision processes

A Piunovskiy, Y Zhang - Probability Theory and Stochastic Modelling, 2020 - Springer
The study of continuous-time Markov decision processes dates back at least to the 1950s,
shortly after that of its discrete-time analogue. Since then, the theory has rapidly developed …

A convex analytic approach to risk-aware Markov decision processes

WB Haskell, R Jain - SIAM Journal on Control and Optimization, 2015 - SIAM
In classical Markov decision process (MDP) theory, we search for a policy that, say,
minimizes the expected infinite horizon discounted cost. Expectation is, of course, a risk …

From infinite to finite programs: Explicit error bounds with applications to approximate dynamic programming

P Mohajerin Esfahani, T Sutter, D Kuhn… - SIAM journal on …, 2018 - SIAM
We consider linear programming (LP) problems in infinite dimensional spaces that are in
general computationally intractable. Under suitable assumptions, we develop an …

Stochastic nestedness and the belief sharing information pattern

S Yuksel - IEEE Transactions on Automatic Control, 2009 - ieeexplore.ieee.org
Solutions to decentralized stochastic optimization problems lead to recursions in which the
state space enlarges with the time-horizon, thus leading to non-tractability of classical …

[图书][B] Finite Approximations in discrete-time stochastic control

N Saldi, T Linder, S Yüksel - 2018 - Springer
Control and optimization of dynamical systems in the presence of stochastic uncertainty is a
mature field with a large range of applications. A comprehensive treatment of such problems …