Policy structure for discrete time Markov chain disorder problems (Q1077336)
From MaRDI portal
| This is the item page for this Wikibase entity, intended for internal use and editing purposes. Please use this page instead for the normal view: Policy structure for discrete time Markov chain disorder problems |
scientific article; zbMATH DE number 3956843
| Language | Label | Description | Also known as |
|---|---|---|---|
| English | Policy structure for discrete time Markov chain disorder problems |
scientific article; zbMATH DE number 3956843 |
Statements
Policy structure for discrete time Markov chain disorder problems (English)
0 references
1986
0 references
A Markov chain disorder problem (MCDP) is a partially observed Markov decision problem where a policy of response must be made to an unobservable transition to an undesirable state. Models for discrete time MCDP and standard functional equations for the minimal expected discounted cost are given. The functional equations yield stationary Markov policies which achieve the desired minima. Sufficient conditions are given for the policies to have a ''likelihood consistency property'', and such policies can be used when an ''optimal'' policy is unattainable or undesirable.
0 references
quality control
0 references
Markov chain disorder problem
0 references
partially observed Markov decision problem
0 references
discrete time
0 references
minimal expected discounted cost
0 references
likelihood consistency property
0 references
0 references