Contents
Preface
1. Background and Preview
2. Highlights of Classical Control Theory
3. State Variables and the State Space Description of Dynamic Systems
4. Fundamentals of Matrix Algebra
5. Vectors and Linear Vector Spaces
6. Simulatneous Linear Equations
7. Eigenvalues and Eigenvectors
8. Functions of Square Matrices and the Cayley-Hamilton Theorem
9. Analysis of Continuous-and Discrete-Time Linear State Equations
10. Stability
11. Controllability and Observability for Linear Systems
12. The Relationship Between State Variable and Transfer Function Description of Systems
13. Design of Linear Feedback Control Systems
14. An Introduction to Optimal Control Theory
15. An Introduction to Nonlinear Control Systems
Answers to Problems
Index