Solutions of any optimal control problem are described by trajectories of a Hamiltonian system. The system is intrinsically associated to the problem by a procedure that is a geometric elaboration of the Lagrange multipliers rule. The intimate relation of the optimal control and Hamiltonian dynamics is fruitful for both domains; among other things, it leads to a clarification and a far going generalization of important classical results about Riemannian geodesic flows.
Preview
Unable to display preview. Download preview PDF.
References
Agrachev, A.A.: Geometry of optimal control problems and Hamiltonian systems. In: C.I.M.E. Lecture Notes in Mathematics, Springer, Berlin (to appear)
Agrachev, A.A., Sachkov, Yu.L. : Control Theory from the Geometric Viewpoint. Springer, Berlin (2004)
Gamkrelidze, R.V.: Principles of Optimal Control Theory. Plenum Publishing, New York (1978)
Jurdjevic, V.: Geometric Control Theory. Cambridge University Press, Cambridge (1997)
Pontryagin, L.S., Boltyanskij, V.G., Gamkrelidze, R.V., Mishchenko, E.F.: The Mathematical Theory of Optimal Processes. Pergamon, Oxfrod (1964)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2008 Springer Science + Business Media B.V
About this paper
Cite this paper
Agrachev, A. (2008). Hamiltonian systems and optimal control. In: Craig, W. (eds) Hamiltonian Dynamical Systems and Applications. NATO Science for Peace and Security Series. Springer, Dordrecht. https://doi.org/10.1007/978-1-4020-6964-2_8
Download citation
DOI: https://doi.org/10.1007/978-1-4020-6964-2_8
Publisher Name: Springer, Dordrecht
Print ISBN: 978-1-4020-6962-8
Online ISBN: 978-1-4020-6964-2
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)