**An Introduction to Mathematical Optimal Control Theory**

by Lawrence C. Evans

**Publisher**: University of California, Berkeley 2010**Number of pages**: 126

**Description**:

Contents: Introduction; Controllability, bang-bang principle; Linear time-optimal control; The Pontryagin Maximum Principle; Dynamic programming; Game theory; Introduction to stochastic control theory; Proofs of the Pontryagin Maximum Principle.

Download or read it online for free here:

**Download link**

(690KB, PDF)

## Similar books

**Modeling, Simulation and Optimization: Tolerance and Optimal Control**

by

**Shkelzen Cakaj**-

**InTech**

Topics covered: parametric representation of shapes, modeling of dynamic continuous fluid flow process, plant layout optimal plot plan, atmospheric modeling, cellular automata simulations, thyristor switching characteristics simulation, etc.

(

**10452**views)

**Optimal Control: Linear Quadratic Methods**

by

**B.D.O. Anderson, J.B. Moore**-

**Prentice-Hall**

Numerous examples highlight this treatment of the use of linear quadratic Gaussian methods for control system design. It explores linear optimal control theory from an engineering viewpoint, with illustrations of practical applications.

(

**13238**views)

**Optimization and Control**

by

**Richard Weber**-

**University of Cambridge**

Topics: Dynamic Programming; Dynamic Programming Examples; Dynamic Programming over the Infinite Horizon; Positive Programming; Negative Programming; Bandit Processes and Gittins Index; Average-cost Programming; LQ Regulation; Controllability; etc.

(

**7507**views)

**Stochastic Optimal Control: The Discrete-Time Case**

by

**Dimitri P. Bertsekas, Steven E. Shreve**-

**Athena Scientific**

This research monograph is the authoritative and comprehensive treatment of the mathematical foundations of stochastic optimal control of discrete-time systems, including the treatment of the intricate measure-theoretic issues.

(

**8570**views)