Skip to main content

Stochastic Optimal Control with Imperfect State Information over a Finite Horizon

  • Chapter
  • First Online:
Neural Approximations for Optimal Control and Decision

Abstract

Discrete-time stochastic optimal control problems are considered. These problems are stated over a finite number of decision stages. The state vector is assumed to be observed through a noisy measurement channel. Because of the very general assumptions under which the problems are stated, obtaining analytically optimal solutions is practically impossible. Note that the controller has to retain the vector of all the measures and of all the controls in memory, up to the most recent decision stage. Such measures and controls constitute the “information vector” that the control function depends on. The increasing dimension of the information vector makes it practically impossible to use dynamic programming. Then, we resort to the “extended Ritz method” (ERIM). The ERIM consists in substituting the admissible functions with fixed-structure parametrized functions containing vectors of “free” parameters. Of course, if the number of decision stages is large, the application of the ERIM is also impossible. Therefore, an approximate approach is followed by truncating the information vector and retaining in the memory only a suitable “limited-memory information vector.”

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 149.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Hardcover Book
USD 199.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    We recall three basic operations on the conditional probability density functions that are used in this section.

    • The so-called chain rule:

      $$\begin{aligned} p(\varvec{a},\varvec{b}\, \vert \, \varvec{c}) = p(\varvec{a}\, \vert \, \varvec{b}, \varvec{c}) \, p(\varvec{b}\, \vert \, \varvec{c}) \, . \end{aligned}$$
      (8.14)
    • The integrated version of (8.14):

      $$\begin{aligned} p(\varvec{a}\, \vert \, \varvec{c}) = \int \, p(\varvec{a}\, \vert \, \varvec{b}, \varvec{c}) \, p(\varvec{b}\, \vert \, \varvec{c}) \, d \varvec{b}\, . \end{aligned}$$
      (8.15)
    • The Bayes formula:

      $$\begin{aligned} p(\varvec{a}\, \vert \, \varvec{b}, \varvec{c}) = \frac{p(\varvec{b}\, \vert \, \varvec{a}, \varvec{c}) \, p(\varvec{a}\, \vert \, \varvec{c})}{\displaystyle \int \, (\mathrm{numerator}) \, d \varvec{a}} \, . \end{aligned}$$
      (8.16)

References

  1. Alessandri A, Baglietto M, Battistelli G (2008) Moving-horizon state estimation for nonlinear discrete-time systems: new stability results and approximation schemes. Automatica 44:1753–1765

    Article  MathSciNet  Google Scholar 

  2. Alessandri A, Baglietto M, Battistelli G, Gaggero M (2011) Moving-horizon state estimation for nonlinear systems using neural networks. IEEE Trans. Neural Netw 22:768–780

    Article  Google Scholar 

  3. Alessandri A, Baglietto M, Parisini T, Zoppoli R (1999) A neural state estimator with bounded errors for nonlinear systems. IEEE Trans Autom Control 44:2028–2042

    MathSciNet  MATH  Google Scholar 

  4. Alessandri A, Parisini T, Zoppoli R (1997) Neural approximations for nonlinear finite-memory state estimation. Int J Control 67:275–302

    Article  Google Scholar 

  5. Alessandri A, Parisini T, Zoppoli R (2001) Sliding-window neural state estimation in a power plant heater line. Int J Adapt Control Signal Process 15:815–836

    Article  Google Scholar 

  6. Aoki M (1967) Optimization of stochastic systems. Academic Press

    Google Scholar 

  7. Aoki M (1989) Optimization of stochastic systems. Academic Press, 2nd edn

    Google Scholar 

  8. Bertsekas DP (2005) Dynamic programming and optimal control, vol 1. Athena Scientific

    Google Scholar 

  9. Bertsekas DP (2005) Dynamic programming and suboptimal control: a survey from ADP to MPC. Eur J Control 11:310–334

    Article  MathSciNet  Google Scholar 

  10. Chui CK, Chen G (2017) Kalman filtering. Springer, 5th edn

    Google Scholar 

  11. Garcia CE, Prett DM, Morari M (1989) Model predictive control: theory and practice—a survey. Automatica 25:335–348

    Article  Google Scholar 

  12. Jazwinski AH (1968) Limited memory optimal filters. IEEE Trans Autom Control 13:558–563

    Article  Google Scholar 

  13. Jazwinski AH (1970) Stochastic processes and filtering theory. Academic Press

    Google Scholar 

  14. Lemon A, Lall S (2016) Approximate sufficient statistics for team decision problems. In: 2016 AAAI spring symposium series. Challenges and opportunities in multiagent learning for the real world

    Google Scholar 

  15. Mayne DQ, Rawlings JB, Rao CV, Scokaert POM (2000) Constrained model predictive control: stability and optimality. Automatica 36:789–814

    Article  MathSciNet  Google Scholar 

  16. Messner A, Papageorgiou M (1992) Motorway network control via nonlinear optimization. In: Proceedings of the 1st meeting of the EURO working group on urban traffic and transportation, pp 1–24

    Google Scholar 

  17. Papageorgiou M (1983) Applications of automatic control concepts to traffic flow modeling and control. Lecture Notes in Control and Information Sciences. Springer

    Google Scholar 

  18. Parisini T, Zoppoli R (1994) Neural networks for nonlinear state estimation. Int J Robust and Nonlinear Control 4:231–248

    Article  MathSciNet  Google Scholar 

  19. Parisini T, Zoppoli R (1996) Neural approximations for multistage optimal control of nonlinear stochastic systems. IEEE Trans Autom Control 41:889–895

    Article  MathSciNet  Google Scholar 

  20. Parisini T, Zoppoli R (1998) Neural approximations for infinite-horizon optimal control of nonlinear stochastic systems. IEEE Trans Neural Netw 9:1388–1408

    Article  Google Scholar 

  21. Payne HJ (1971) Models of freeway traffic and control. Simul Council Proc 1:51–61

    Google Scholar 

  22. Rawlings JB, Mayne DQ, Diehl MM (2018) Model predictive control: theory, computation, and design. Nob Hill Publishing

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Riccardo Zoppoli .

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Zoppoli, R., Sanguineti, M., Gnecco, G., Parisini, T. (2020). Stochastic Optimal Control with Imperfect State Information over a Finite Horizon. In: Neural Approximations for Optimal Control and Decision. Communications and Control Engineering. Springer, Cham. https://doi.org/10.1007/978-3-030-29693-3_8

Download citation

Publish with us

Policies and ethics