Policy iteration type algorithms for recurrent state Markov decision processes
From MaRDI portal
Publication:1886500
DOI10.1016/S0305-0548(03)00190-4zbMath1067.90159OpenAlexW2021512803WikidataQ115023230 ScholiaQ115023230MaRDI QIDQ1886500
Publication date: 18 November 2004
Published in: Computers \& Operations Research (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1016/s0305-0548(03)00190-4
Cites Work
- Unnamed Item
- Unnamed Item
- On the solvability of Bellman's functional equations for Markov renewal programming
- A Brouwer fixed-point mapping approach to communicating Markov decision processes
- Denumerable Undiscounted Semi-Markov Decision Processes with Unbounded Rewards
- On undiscounted markovian decision processes with compact action spaces
- On the Convergence of Policy Iteration in Finite State Undiscounted Markov Decision Processes: The Unichain Case
- Counter examples for compact action markov decision chains with average reward criteria
- The Existence of a Stationary $\varepsilon $-Optimal Policy for a Finite Markov Chain
- An Analysis of Stochastic Shortest Path Problems
- On Controlled Finite State Markov Processes with Compact Control Sets
- The optimality equation in average cost denumerable state semi-Markov decision problems, recurrency conditions and algorithms
- A New Value Iteration method for the Average Cost Dynamic Programming Problem
- Optimal decision procedures for finite Markov chains. Part II: Communicating systems
- Discrete-Time Controlled Markov Processes with Average Cost Criterion: A Survey
- Existence of a Stationary Control for a Markov Chain Maximizing the Average Reward
This page was built for publication: Policy iteration type algorithms for recurrent state Markov decision processes