Finding optimal memoryless policies of POMDPs under the expected average reward criterion
From MaRDI portal
Publication:418072
DOI10.1016/j.ejor.2010.12.014zbMath1237.90250OpenAlexW2055418958MaRDI QIDQ418072
Baoqun Yin, Hong-Sheng Xi, Yan-Jie Li
Publication date: 14 May 2012
Published in: European Journal of Operational Research (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1016/j.ejor.2010.12.014
Related Items (2)
Uses Software
Cites Work
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Basic ideas for event-based optimization of Markov systems
- Optimization of a special case of continuous-time Markov decision processes with compact action set
- A survey of algorithmic methods for partially observed Markov decision processes
- Perturbation realization, potentials, and sensitivity analysis of Markov processes
- The Optimal Control of Partially Observable Markov Processes over a Finite Horizon
- CONVERGENCE OF SIMULATION-BASED POLICY ITERATION
- Simulation-based optimization of Markov reward processes
- The $n$th-Order Bias Optimality for Multichain Markov Decision Processes
- Event-Based Optimization of Markov Systems
- Potential-Based Online Policy Iteration Algorithms for Markov Decision Processes
- Performance optimization algorithms based on potentials for semi-Markov control processes
This page was built for publication: Finding optimal memoryless policies of POMDPs under the expected average reward criterion