Splitting Randomized Stationary Policies in Total-Reward Markov Decision Processes

From MaRDI portal
Publication:2884309

DOI10.1287/moor.1110.0525zbMath1243.90233OpenAlexW1991591460MaRDI QIDQ2884309

Uriel G. Rothblum, Eugene A. Feinberg

Publication date: 24 May 2012

Published in: Mathematics of Operations Research (Search for Journal in Brave)

Full work available at URL: https://doi.org/10.1287/moor.1110.0525




Related Items (19)

Conditions for the solvability of the linear programming formulation for constrained discounted Markov decision processesStationary Markov Nash Equilibria for Nonzero-Sum Constrained ARAT Markov GamesConstrained continuous-time Markov decision processes on the finite horizonConstrained optimality for finite horizon semi-Markov decision processes in Polish spacesThe multi-armed bandit, with constraintsExtreme Occupation Measures in Markov Decision Processes with an Absorbing StateNash equilibria for total expected reward absorbing Markov games: the constrained and unconstrained casesAbsorbing Markov decision processesA Convex Programming Approach for Discrete-Time Markov Decision Processes under the Expected Total Reward CriterionAn exact iterative search algorithm for constrained Markov decision processesSufficiency of Deterministic Policies for Atomless Discounted and Uniformly Absorbing MDPs with Multiple CriteriaNote on discounted continuous-time Markov decision processes with a lower bounding functionExtreme point characterization of constrained nonstationary infinite-horizon Markov decision processes with finite state spaceConstrained discounted Markov decision processes with Borel state spacesRisk-sensitive semi-Markov decision processes with general utilities and multiple criteriaConstrained Markov Decision Processes with Expected Total Reward CriteriaOptimality of Mixed Policies for Average Continuous-Time Markov Decision Processes with ConstraintsAbsorbing Continuous-Time Markov Decision Processes with Total Cost CriteriaOn the reduction of total‐cost and average‐cost MDPs to discounted MDPs




This page was built for publication: Splitting Randomized Stationary Policies in Total-Reward Markov Decision Processes