Deprecated: $wgMWOAuthSharedUserIDs=false is deprecated, set $wgMWOAuthSharedUserIDs=true, $wgMWOAuthSharedUserSource='local' instead [Called from MediaWiki\HookContainer\HookContainer::run in /var/www/html/w/includes/HookContainer/HookContainer.php at line 135] in /var/www/html/w/includes/Debug/MWDebug.php on line 372
Discrete-Time Controlled Markov Processes with Average Cost Criterion: A Survey - MaRDI portal

Discrete-Time Controlled Markov Processes with Average Cost Criterion: A Survey

From MaRDI portal
Publication:4695407

DOI10.1137/0331018zbMath0770.93064OpenAlexW1967037758WikidataQ60167559 ScholiaQ60167559MaRDI QIDQ4695407

Mrinal K. Ghosh, Emmanuel Fernández-Gaucherand, Steven I. Marcus, Vivek S. Borkar, Aristotle Arapostathis

Publication date: 13 September 1993

Published in: SIAM Journal on Control and Optimization (Search for Journal in Brave)

Full work available at URL: http://hdl.handle.net/1903/5155



Related Items

On the optimality equation for average cost Markov control processes with Feller transition probabilities, LP based upper and lower bounds for Cesàro and Abel limits of the optimal values in problems of control of stochastic discrete time systems, Unnamed Item, A note on the Ross-Taylor theorem, Markov Decision Processes with Variance Minimization: A New Condition and Approach, Policy iteration type algorithms for recurrent state Markov decision processes, Computational aspects in applied stochastic control, Attention allocation for decision making queues, Approximate dynamic programming for stochastic linear control problems on compact state spaces, Controlled semi-Markov chains with risk-sensitive average cost criterion, Unnamed Item, On Some Ergodic Impulse Control Problems with Constraint, Optimality equations and inequalities in a class of risk-sensitive average cost Markov decision chains, Risk sensitive control of Markov processes in countable state space, Average cost optimality in inventory models with Markovian demands, A Mixed Value and Policy Iteration Method for Stochastic Control with Universally Measurable Policies, On computing average cost optimal policies with application to routing to parallel queues, Note on stability estimation in average Markov control processes, On some algorithms for limiting average Markov decision processes, A heuristic to solve a sea cargo revenue management problem, Approximation of average cost optimal policies for general Markov decision processes with unbounded costs, Redundant data transmission in control/estimation over lossy networks, Value iteration in average cost Markov control processes on Borel spaces, Dynamic Limit Growth Indices in Discrete Time, Optimal life histories for structured populations in fluctuating environments, Controlled Markov processes on the infinite planning horizon: Weighted and overtaking cost criteria, A semimartingale characterization of average optimal stationary policies for Markov decision processes, From Infinite to Finite Programs: Explicit Error Bounds with Applications to Approximate Dynamic Programming, Constrained Semi-Markov decision processes with average rewards, Stationary policies for lower bounds on the minimum average cost of discrete-time nonlinear control systems, Discounted approximations to the risk-sensitive average cost in finite Markov chains, Exact decomposition approaches for Markov decision processes: a survey, Passage-detector-based traffic queue estimation in intelligent transportation systems: A computational study of competing algorithms, The policy iteration algorithm for average continuous control of piecewise deterministic Markov processes, A pause control approach to the value iteration scheme in average Markov decision processes, A note on the convergence rate of the value iteration scheme in controlled Markov chains, Optimal eviction policies for stochastic address traces, Another set of verifiable conditions for average Markov decision processes with Borel spaces, ``Super-overtaking optimal policies for Markov control processes, Sample-Path Optimal Stationary Policies in Stable Markov Decision Chains with the Average Reward Criterion, Strong \(n\)-discount and finite-horizon optimality for continuous-time Markov decision processes, Discounted approximations in risk-sensitive average Markov cost chains with finite state space, Optimality Gap of Constant-Order Policies Decays Exponentially in the Lead Time for Lost Sales Models, Average sample-path optimality for continuous-time Markov decision processes in Polish spaces, Randomization and simplification in dynamic decision-making., Bias and Overtaking Optimality for Continuous-Time Jump Markov Decision Processes in Polish Spaces, On optimal control theory in marine oil spill management: a Markovian decision approach, Approximate receding horizon approach for Markov decision processes: average reward case, An axiomatic approach to Markov decision processes, Average control of Markov decision processes with Feller transition probabilities and general action spaces, Growth Optimal Investment with Transaction Costs, STRONG AVERAGE OPTIMALITY FOR CONTROLLED NONHOMOGENEOUS MARKOV CHAINS*, Maximizing the set of recurrent states of an MDP subject to convex constraints, The LP approach in average reward MDPs with multiple cost constraints: The countable state case, Partially observed semi-Markov zero-sum games with average payoff, The average cost of Markov chains subject to total variation distance uncertainty, Quadratic costs and second moments of jump linear systems with general Markov chain, Approximation of Markov decision processes with general state space, Infinite horizon optimality criteria for equipment replacement under technological change, Sample-path optimality and variance-maximization for Markov decision processes, Convergence of value functions for finite horizon Markov decision processes with constraints, A note on the vanishing interest rate approach in average Markov decision chains with continuous and bounded costs, Another set of conditions for average optimality in Markov control processes, Sparse and constrained stochastic predictive control for networked systems, Solutions of the average cost optimality equation for Markov decision processes with weakly continuous kernel: the fixed-point approach revisited, Quantitative model-checking of controlled discrete-time Markov processes, Cooperation dynamics in repeated games of adverse selection, Planning for the long run: programming with patient, Pareto responsive preferences, Continuous-time Markov decision processes with \(n\)th-bias optimality criteria, On the adaptive control of a class of partially observed Markov decision processes, The discounted method and equivalence of average criteria for risk-sensitive Markov decision processes on Borel spaces, On the vanishing discount factor approach for Markov decision processes with weakly continuous transition probabilities, Stochastic Differential Games with Multiple Modes and Applications to Portfolio Optimization, Exit time risk-sensitive control for systems of cooperative agents, A discounted approach in communicating average Markov decision chains under risk-aversion, On the existence of stationary optimal policies for partially observed MDPs under the long-run average cost criterion, Another set of conditions for Markov decision processes with average sample-path costs, Phase Transitions for Controlled Markov Chains on Infinite Graphs, Successive approximations in partially observable controlled Markov chains with risk-sensitive average criterion, Unnamed Item, Inventory models with Markovian demands and cost functions of polynomial growth, Strong Uniform Value in Gambling Houses and Partially Observable Markov Decision Processes, On the relation between discounted and average optimal value functions, Ergodic control of partially observed Markov chains, On structural properties of optimal average cost functions in Markov decision processes with Borel spaces and universally measurable policies, Dispatching to parallel servers. Solutions of Poisson's equation for first-policy improvement, Contractive approximations in risk-sensitive average semi-Markov decision chains on a finite state space, Average optimality for Markov decision processes in borel spaces: a new condition and approach, Single sample path-based optimization of Markov chains, Convex analytic method revisited: further optimality results and performance of deterministic policies in average cost stochastic control, The convergence of value iteration in average cost Markov decision chains, Solutions of the average cost optimality equation for finite Markov decision chains: Risk-sensitive and risk-neutral criteria, Application of average dynamic programming to inventory systems, Blackwell optimality in Markov decision processes with partial observation., Stochastic approximations of constrained discounted Markov decision processes, Automated verification and synthesis of stochastic hybrid systems: a survey, Reversible Markov decision processes and the Gaussian free field, Constrained markov decision processes with compact state and action spaces: the average case, Continuity of the optimal average cost in Markov decision chains with small risk-sensitivity, Weak conditions for average optimality in Markov control processes, Robustness to Incorrect Priors and Controlled Filter Stability in Partially Observed Stochastic Control, Optimal Ergodic Harvesting under Ambiguity, Finite-Memory Strategies in POMDPs with Long-Run Average Objectives, Blackwell optimal policies in a Markov decision process with a Borel state space, Contractive approximations in average Markov decision chains driven by a risk-seeking controller, Risk-Sensitive Reinforcement Learning via Policy Gradient Search, Average criteria in denumerable semi-Markov decision chains under risk-aversion, Unnamed Item, Dual sourcing models with stock-out dependent substitution, A dynamic analytic method for risk-aware controlled martingale problems, Markov decision processes under risk sensitivity: a discount vanishing approach, Risk-sensitivity vanishing limit for controlled Markov processes, On Iteration Improvement for Averaged Expected Cost Control for One-Dimensional Ergodic Diffusions, The Vanishing Discount Approach in a class of Zero-Sum Finite Games with Risk-Sensitive Average Criterion, Average Cost Optimality Inequality for Markov Decision Processes with Borel Spaces and Universally Measurable Policies, Another Set of Conditions for Strongn(n = −1, 0) Discount Optimality in Markov Decision Processes, Algorithms for optimization and stabilization of controlled Markov chains., Denumerable controlled Markov chains with average reward criterion: Sample path optimality, Vanishing discount approximations in controlled Markov chains with risk-sensitive average criterion, Protocol invariance and the timing of decisions in dynamic games, Average Reward Markov Decision Processes with Multiple Cost Constraints, The value function in ergodic control of diffusion processes with partial observations, A perturbation approach to approximate value iteration for average cost Markov decision processes with Borel spaces and bounded costs, Infinite Horizon Average Cost Dynamic Programming Subject to Total Variation Distance Ambiguity, History-dependent Evaluations in Partially Observable Markov Decision Process, Ergodic risk-sensitive control of Markov processes on countable state space revisited, Approximation of average cost Markov decision processes using empirical distributions and concentration inequalities, Zero and non-zero sum risk-sensitive Semi-Markov games