Optimal Control: Linear Quadratic Methods
by B.D.O. Anderson, J.B. Moore
Publisher: Prentice-Hall 1989
ISBN/ASIN: 0486457664
Number of pages: 394
Description:
Numerous examples highlight this treatment of the use of linear quadratic Gaussian methods for control system design. It explores linear optimal control theory from an engineering viewpoint, with illustrations of practical applications. Key topics include loop-recovery techniques, frequency shaping, and controller reduction. Numerous examples and complete solutions.
Download or read it online for free here:
Download link
(18MB, PDF)
Similar books
Distributed-Parameter Port-Hamiltonian Systemsby Hans Zwart, Birgit Jacob - CIMPA
Topics from the table of contents: Introduction; Homogeneous differential equation; Boundary Control Systems; Transfer Functions; Well-posedness; Stability and Stabilizability; Systems with Dissipation; Mathematical Background.
(11918 views)
Optimization and Controlby Richard Weber - University of Cambridge
Topics: Dynamic Programming; Dynamic Programming Examples; Dynamic Programming over the Infinite Horizon; Positive Programming; Negative Programming; Bandit Processes and Gittins Index; Average-cost Programming; LQ Regulation; Controllability; etc.
(14522 views)
Stochastic Optimal Control: The Discrete-Time Caseby Dimitri P. Bertsekas, Steven E. Shreve - Athena Scientific
This research monograph is the authoritative and comprehensive treatment of the mathematical foundations of stochastic optimal control of discrete-time systems, including the treatment of the intricate measure-theoretic issues.
(15933 views)
An Introduction to Mathematical Optimal Control Theoryby Lawrence C. Evans - University of California, Berkeley
Contents: Introduction; Controllability, bang-bang principle; Linear time-optimal control; The Pontryagin Maximum Principle; Dynamic programming; Game theory; Introduction to stochastic control theory; Proofs of the Pontryagin Maximum Principle.
(16795 views)