Skip to main content

Dynamic Programming or Direct Comparison?,

  • Chapter
Three Decades of Progress in Control Sciences
  • 711 Accesses

Summary

The standard approach to stochastic control is dynamic programming. In our recent research, we proposed an alternative approach based on direct comparison of the performance of any two policies. This approach has a number of advantages: the results may be derived in a simple and intuitive way; the approach applies to different optimization problems, including finite and infinite horizon, discounting and average performance, discrete time discrete states and continuous time and continuous stats, etc., in the same way; and it may be generalized to some non-standard problems where dynamic programming fails. This approach also links stochastic control to perturbation analysis, reinforcement learning and other research subjects in optimization, which may stimulate new research directions.

Supported in part by a grant from Hong Kong UGC.

Tribute to Chris Byrnes and Anders Lindquist.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 169.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Akian, M., Sulem, A., Taksar, M.: Dynamic Optimization of Long Term Growth Rate for a Portfolio with Transaction Costs and Logarithmic Utility. Mathematical Finance 11, 153–188 (2001)

    Article  MathSciNet  MATH  Google Scholar 

  2. Bertsekas, D.P.: Dynamic Programming and Optimal Control, vol. I, II. Athena Scientific, Belmont (2007)

    Google Scholar 

  3. Brockett, R.: Stochastic Control. Preprint (2009)

    Google Scholar 

  4. Cao, X.R.: Realization Probabilities - The Dynamics of Queueing Systems. Springer, New York (1994)

    Book  Google Scholar 

  5. Cao, X.R.: Stochastic Learning and Optimization - a Sensitivity-Based Approach. Springer, Heidelberg (2007)

    Book  MATH  Google Scholar 

  6. Cao, X.R., Zhang, J.: The nth-Order Bias Optimality for Multi-chain Markov Decision Processes. IEEE Transactions on Automatic Control 53, 496–508 (2008)

    Article  MathSciNet  Google Scholar 

  7. Cao, X.R.: Stochastic Control via Direct Comparison. Submitted to IEEE Transaction on Automatic Control (2009)

    Google Scholar 

  8. Cao, X.R.: Singular Stochastic Control and Composite Markov Processes. Manuscript to be submitted (2009)

    Google Scholar 

  9. Fleming, W.H., Soner, H.M.: Controlled Markov Processes and Viscosity Solutions, 2nd edn. Springer, Heidelberg (2006)

    MATH  Google Scholar 

  10. Ho, Y.C., Cao, X.R.: Perturbation Analysis of Discrete-Event Dynamic Systems. Kluwer Academic Publisher, Boston (1991)

    Book  MATH  Google Scholar 

  11. Meyn, S.P.: The Policy Iteration Algorithm for Average Reward Markov Decision Processes with General State Space. IEEE Transactions on Automatic Control 42, 1663–1680 (1997)

    Article  MathSciNet  MATH  Google Scholar 

  12. Muthuraman, K., Zha, H.: Simulation-based portfolio optimization for large portfolios with transaction costs. Mathematical Finance 18, 115–134 (2008)

    Article  MathSciNet  MATH  Google Scholar 

  13. Puterman, M.L.: Markov Decision Processes: Discrete Stochastic Dynamic Programming. Wiley, Chichester (1994)

    Book  MATH  Google Scholar 

  14. Oksendal, B., Sulem, A.: Applied Stochastic Control of Jump Diffusions. Springer, Heidelberg (2007)

    Book  Google Scholar 

  15. Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction. MIT Press, Cambridge (1998)

    Google Scholar 

  16. Veinott, A.F.: Discrete Dynamic Programming with Sensitive Discount Optimality Criteria. The Annals of Mathematical Statistics 40(5), 1635–1660 (1969)

    Article  MathSciNet  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2010 Springer Berlin Heidelberg

About this chapter

Cite this chapter

Cao, XR. (2010). Dynamic Programming or Direct Comparison?, . In: Hu, X., Jonsson, U., Wahlberg, B., Ghosh, B. (eds) Three Decades of Progress in Control Sciences. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-11278-2_5

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-11278-2_5

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-11277-5

  • Online ISBN: 978-3-642-11278-2

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics