Finite-time partial stability, stabilization, semistabilization, and optimal feedback control

Thumbnail Image
L'afflitto, Andrea
Haddad, Wassim M.
Associated Organizations
Supplementary to
Asymptotic stability is a key notion of system stability for controlled dynamical systems as it guarantees that the system trajectories are bounded in a neighborhood of a given isolated equilibrium point and converge to this equilibrium over the infinite horizon. In some applications, however, asymptotic stability is not the appropriate notion of stability. For example, for systems with a continuum of equilibria, every neighborhood of an equilibrium contains another equilibrium and a nonisolated equilibrium cannot be asymptotically stable. Alternatively, in stabilization of spacecraft dynamics via gimballed gyroscopes, it is desirable to find state- and output-feedback control laws that guarantee partial-state stability of the closed-loop system, that is, stability with respect to part of the system state. Furthermore, we may additionally require finite-time stability of the closed-loop system, that is, convergence of the system's trajectories to a Lyapunov stable equilibrium in finite time. The Hamilton-Jacobi-Bellman optimal control framework provides necessary and sufficient conditions for the existence of state-feedback controllers that minimize a given performance measure and guarantee asymptotic stability of the closed-loop system. In this research, we provide extensions of the Hamilton-Jacobi-Bellman optimal control theory to develop state-feedback control laws that minimize nonlinear-nonquadratic performance criteria and guarantee semistability, partial-state stability, finite-time stability, and finite-time partial state stability of the closed-loop system.
Date Issued
Resource Type
Resource Subtype
Rights Statement
Rights URI