Sensitivity analysis of stationary performance measures for Markov chains (Q1922200)
From MaRDI portal
| This is the item page for this Wikibase entity, intended for internal use and editing purposes. Please use this page instead for the normal view: Sensitivity analysis of stationary performance measures for Markov chains |
scientific article; zbMATH DE number 927172
| Language | Label | Description | Also known as |
|---|---|---|---|
| English | Sensitivity analysis of stationary performance measures for Markov chains |
scientific article; zbMATH DE number 927172 |
Statements
Sensitivity analysis of stationary performance measures for Markov chains (English)
0 references
12 March 1997
0 references
Many practical systems, such as communication systems, manufacturing systems for example, can be well described by Markov chains. In this very nice paper, the author is interested in derivative estimation of performance measures of Markov chains. He considers \(X\) a finite state space, irreducible, and a periodic Markov chain with transition matrix \(P(\theta)\) for a discrete-time system, or an infinitesimal generator \(Q(\theta)\) for a continuous-time system, and he assumes that \(\theta\in (a,b)\subset \mathbb{R}\). The state space and the cost function do not depend on \(\theta\) explicitly, since \(J(\theta)=E[f(\theta)]\), which represents the mean cost of making a state transition when the chain is stationary, is a function of \(\theta\) through \(P(\theta)\) or \(Q(\theta)\) only. This paper proposes an algorithm which is a special implementation of the structural infinitesimal perturbation analysis (SIPA) algorithm; it is designed for estimation via simulation but can be easily tailored to address the needs of derivative estimation from sample path observation. The basic procedure is to simulate the same Markov chain but with a new cost function. At each step, the value of this new cost function is estimated from some extra simulations. At one step, the extra simulations are guaranteed to require bounded computational effort. The author theoretically proves that the algorithm converges to \(J(\theta)\) at rate \(T^{-1/2}\), as the simulation time \(T\) goes to infinity. The rate \(T^{-1/2}\) is the best possible rate for derivative estimation methods such as perturbation analysis and likelihood ratio, the two best-known methods for derivative estimation. This work is different from previous ones in that the author not only imposes minimum assumptions but also guarantees a consistent estimate that attains the best possible rate of convergence.
0 references
derivative estimation
0 references
Markov chains
0 references
structural infinitesimal perturbation analysis
0 references
simulation
0 references
consistent estimate
0 references
rate of convergence
0 references
0 references
0 references