An Introduction to Mathematical Optimal Control Theory
by Lawrence C. Evans
Publisher: University of California, Berkeley 2010
Number of pages: 126
Contents: Introduction; Controllability, bang-bang principle; Linear time-optimal control; The Pontryagin Maximum Principle; Dynamic programming; Game theory; Introduction to stochastic control theory; Proofs of the Pontryagin Maximum Principle.
Home page url
Download or read it online for free here:
by Richard Weber - University of Cambridge
Topics: Dynamic Programming; Dynamic Programming Examples; Dynamic Programming over the Infinite Horizon; Positive Programming; Negative Programming; Bandit Processes and Gittins Index; Average-cost Programming; LQ Regulation; Controllability; etc.
by Shkelzen Cakaj - InTech
Topics covered: parametric representation of shapes, modeling of dynamic continuous fluid flow process, plant layout optimal plot plan, atmospheric modeling, cellular automata simulations, thyristor switching characteristics simulation, etc.
by Hans Zwart, Birgit Jacob - CIMPA
Topics from the table of contents: Introduction; Homogeneous differential equation; Boundary Control Systems; Transfer Functions; Well-posedness; Stability and Stabilizability; Systems with Dissipation; Mathematical Background.
by B.D.O. Anderson, J.B. Moore - Prentice-Hall
Numerous examples highlight this treatment of the use of linear quadratic Gaussian methods for control system design. It explores linear optimal control theory from an engineering viewpoint, with illustrations of practical applications.