Policy iteration for continuous-time average reward Markov decision processes in Polish spaces
From MaRDI portal
Publication:963139
DOI10.1155/2009/103723zbMath1192.90243OpenAlexW2046510288WikidataQ58646522 ScholiaQ58646522MaRDI QIDQ963139
Quanxin Zhu, Xinsong Yang, Chuang Xia Huang
Publication date: 8 April 2010
Published in: Abstract and Applied Analysis (Search for Journal in Brave)
Full work available at URL: https://eudml.org/doc/226313
Continuous-time Markov processes on general state spaces (60J25) Markov and semi-Markov decision processes (90C40)
Related Items (3)
Average sample-path optimality for continuous-time Markov decision processes in Polish spaces ⋮ Policy iteration algorithms for zero-sum stochastic differential games with long-run average payoff criteria ⋮ Variance minimization for continuous-time Markov decision processes: two approaches
Cites Work
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Average optimality for continuous-time Markov decision processes in Polish spaces
- Policy iteration for average cost Markov control processes on Borel spaces
- Computable exponential convergence rates for stochastically ordered Markov processes
- Average optimality for continuous-time Markov decision processes with a policy iteration approach
- Average optimality inequality for continuous-time Markov decision processes in Polish spaces
- Another set of conditions for Markov decision processes with average sample-path costs
- Markov Decision Processes with Variance Minimization: A New Condition and Approach
- Bias and Overtaking Optimality for Continuous-Time Jump Markov Decision Processes in Polish Spaces
- On undiscounted markovian decision processes with compact action spaces
- On the Convergence of Policy Iteration in Finite State Undiscounted Markov Decision Processes: The Unichain Case
- Counter examples for compact action markov decision chains with average reward criteria
- Conditions for optimality in dynamic programming and for the limit of n-stage optimal policies to be optimal
- A new policy iteration scheme for Markov decision processes using Schweitzer's formula
- The policy iteration algorithm for average reward Markov decision processes with general state space
- Convergence Properties of Policy Iteration
- Drift and monotonicity conditions for continuous-time controlled markov chains with an average criterion
- Optimal Control of Ergodic Continuous-Time Markov Chains with Average Sample-Path Rewards
- Multichain Markov Renewal Programs
- Another Set of Conditions for Strongn(n = −1, 0) Discount Optimality in Markov Decision Processes
- A Note on the Convergence of Policy Iteration in Markov Decision Processes with Compact Action Spaces
This page was built for publication: Policy iteration for continuous-time average reward Markov decision processes in Polish spaces