Strong Uniform Value in Gambling Houses and Partially Observable Markov Decision Processes
DOI10.1137/15M1043340zbMath1346.90796arXiv1505.07495OpenAlexW365305493MaRDI QIDQ2818183
Publication date: 6 September 2016
Published in: SIAM Journal on Control and Optimization (Search for Journal in Brave)
Full work available at URL: https://arxiv.org/abs/1505.07495
Dynamic programming (90C39) Applications of Markov chains and discrete-time Markov processes on general state spaces (social mobility, learning theory, industrial processes, etc.) (60J20) Markov and semi-Markov decision processes (90C40) Dynamical systems and their relations with probability theory and stochastic processes (37A50)
Related Items (4)
Cites Work
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Uniform value in dynamic programming
- A convex analytic approach to Markov decision processes
- Stochastic optimal control. The discrete time case
- Incomplete information in Markovian decision models
- Markov chains and invariant probabilities
- Blackwell optimality in Markov decision processes with partial observation.
- Partially Observable Total-Cost Markov Decision Processes with Weakly Continuous Transition Probabilities
- Reduction of a Controlled Markov Model with Incomplete Data to a Problem with Complete Information in the Case of Borel State and Control Space
- Denumerable Constrained Markov Decision Processes and Finite Approximations
- Average Cost Dynamic Programming Equations For Controlled Markov Chains With Partial Observations
- Discrete-Time Controlled Markov Processes with Average Cost Criterion: A Survey
- Discrete Dynamic Programming
- Discrete-Time Markovian Decision Processes with Incomplete State Observation
- Long-Term Values in Markov Decision Processes and Repeated Games, and a New Distance for Probability Spaces
- Stochastic games
This page was built for publication: Strong Uniform Value in Gambling Houses and Partially Observable Markov Decision Processes