Abstract
This paper deals with denumerable-state continuous-time controlled Markov chains with possibly unbounded transition and reward rates. It concerns optimality criteria that improve the usual expected average reward criterion. First, we show the existence of average reward optimal policies with minimal average variance. Then we compare the variance minimization criterion with overtaking optimality. We present an example showing that they are opposite criteria, and therefore we cannot optimize them simultaneously. This leads to a multiobjective problem for which we identify the set of Pareto optimal policies (also known as nondominated policies).
Similar content being viewed by others
References
Bäuerle N (2005) Benchmark and mean-variance problems for insurers. Math Methods Oper Res 62: 159–165
Bhatt AG, Borkar VS (1996) Occupation measures for controlled Markov processes: characterization and optimality. Ann Probab 24: 1531–1562
Bhattacharya RN (1982) On the functional central limit theorem and the law of the iterated logarithm for Markov processes. Z Wahrsch Verw Geb 60: 185–201
Down D, Meyn SP, Tweedie RL (1995) Exponential and uniform ergodicity of Markov processes. Ann Probab 23: 1671–1691
Guo XP, Hernández-Lerma O (2003a) Continuous-time controlled Markov chains with discounted rewards. Acta Appl Math 79: 195–216
Guo XP, Hernández-Lerma O (2003b) Drift and monotonicity conditions for continuous-time controlled Markov chains. IEEE Trans Automat Control 48: 236–244
Guo XP, Hernández-Lerma O, Prieto-Rumeau T (2006) A survey of recent results on continuous-time Markov decision processes. Top 14: 177–261
Hernández-Lerma O, Romera R (2004) The scalarization approach to multiobjective Markov control problems: why does it work? Appl Math Optim 50: 279–293
Hernández-Lerma O, Vega-Amaya O, Carrasco G (1999) Sample-path optimality and variance- minimization of average cost Markov control processes. SIAM J Control Optim 38: 79–93
Kurtz TG, Stockbridge RH (1998) Existence of Markov controls and characterization of optimal Markov controls. SIAM J Control Optim 36: 609–653 (Erratum, ibid. 37:1310–1311)
Mandl P, Romera MR (1987) On controlled Markov processes with average cost criterion. Kybernetica (Prague) 23: 433–442
Prieto-Rumeau T (2006) Blackwell optimality in the class of Markov policies for continuous-time controlled Markov chains. Acta Appl Math 92: 77–96
Prieto-Rumeau T, Hernández-Lerma O (2005) The Laurent series, sensitive discount and Blackwell optimality for continuous-time controlled Markov chains. Math Methods Oper Res 61: 123–145
Prieto-Rumeau T, Hernández-Lerma O (2006) Bias optimality for continuous-time controlled Markov chains. SIAM J Control Optim 45: 51–73
Prieto-Rumeau T, Hernández-Lerma O (2008) Ergodic control of continuous-time Markov chains with pathwise constraints. SIAM J Control Optim 47: 1888–1908
Ye L, Guo XP, Hernández-Lerma O (2008) Existence and regularity of a nonhomogeneous transition matrix under measurability conditions. J Theoret Probab 21: 604–627
Author information
Authors and Affiliations
Corresponding author
Additional information
This research was partially supported by CONACyT Grant 45693-F.
Rights and permissions
About this article
Cite this article
Prieto-Rumeau, T., Hernández-Lerma, O. Variance minimization and the overtaking optimality approach to continuous-time controlled Markov chains. Math Meth Oper Res 70, 527–540 (2009). https://doi.org/10.1007/s00186-008-0276-z
Received:
Revised:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00186-008-0276-z
Keywords
- Continuous-time controlled Markov chains
- Markov decision processes
- Average reward optimality
- Overtaking optimality
- Average variance