Skip to main content

Total Reward Variance in Discrete and Continuous Time Markov Chains

  • Conference paper
Operations Research Proceedings 2004

Part of the book series: Operations Research Proceedings ((ORP,volume 2004))

Abstract

This note studies the variance of total cumulative rewards for Markov reward chains in both discrete and continuous time. It is shown that parallel results can be obtained for both cases.

First, explicit formulae are presented for the variance within finite time. Next, the infinite time horizon is considered. Most notably, it is concluded that the variance has a linear growth rate. Explicit expressions are provided, related to the standard average reward case, to compute this growth rate.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Benito, F. (1982): Calculating the variance in Markov processes with random reward. Trabajos de Estadistica y de Investigacion Operativa, 33, 73–85

    MATH  MathSciNet  Google Scholar 

  2. Mandl, P. (1971): On the variance in controlled Markov chains. Kybernetika, 7, 1–12

    MATH  MathSciNet  Google Scholar 

  3. Puterman, M. L. (1994): Markov Decision Processes — Discrete Stochastic Dynamic Programming. Wiley, New York

    Google Scholar 

  4. Ross, S.M. (1970): Applied Probability Models with Optimization Applications. Holden-Day, San Francisco, CA

    Google Scholar 

  5. Sladký, K., Sitař, M. (2004): Optimal solutions for undiscounted variance penalized Markov decision chains. In: Dynamic Stochastic Optimization (Marti, K., Ermoliev, Y., Pflug, G., Eds.), LNEMS, Vol. 532, Springer, Berlin, pp. 43–66

    Google Scholar 

  6. Sobel, M. J. (1982): The variance of discounted Markov decision processes. J. Appl. Probab., 19, 794–802

    Article  MATH  MathSciNet  Google Scholar 

  7. van Dijk, N. M., Sladky, K. (2004): On total reward variance in continuous-time Markov reward chains. Manuscript

    Google Scholar 

  8. White, D. J. (1988): Mean, variance and probability criteria in finite Markov decision processes: A review. J. Optim. Theory Appl., 56, 1–29

    Article  MATH  MathSciNet  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2005 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Sladký, K., van Dijk, N.M. (2005). Total Reward Variance in Discrete and Continuous Time Markov Chains. In: Fleuren, H., den Hertog, D., Kort, P. (eds) Operations Research Proceedings 2004. Operations Research Proceedings, vol 2004. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-27679-3_40

Download citation

Publish with us

Policies and ethics