Lyapunov Stability

Get Lyapunov Stability essential facts below. View Videos or join the Lyapunov Stability discussion. Add Lyapunov Stability to your PopFlock.com topic list for future reference or share this resource on social media.
## History

## Definition for continuous-time systems

### System of deviations

### Lyapunov's second method for stability

## Definition for discrete-time systems

## Stability for linear state space models

## Stability for systems with inputs

## Example

## Barbalat's lemma and stability of time-varying systems

## See also

## References

## Further reading

This article uses material from the Wikipedia page available here. It is released under the Creative Commons Attribution-Share-Alike License 3.0.

Lyapunov Stability

Various types of stability may be discussed for the solutions of differential equations or difference equations describing dynamical systems. The most important type is that concerning the stability of solutions near to a point of equilibrium. This may be discussed by the theory of **Aleksandr Lyapunov**. In simple terms, if the solutions that start out near an equilibrium point stay near forever, then is **Lyapunov stable**. More strongly, if is Lyapunov stable and all solutions that start out near converge to , then is asymptotically stable. The notion of **exponential stability** guarantees a minimal rate of decay, i.e., an estimate of how quickly the solutions converge. The idea of Lyapunov stability can be extended to infinite-dimensional manifolds, where it is known as structural stability, which concerns the behavior of different but "nearby" solutions to differential equations. Input-to-state stability (ISS) applies Lyapunov notions to systems with inputs.

In the restricted three-body problem, Lyapunov orbits are curved paths around a Lagrangian point that lie entirely in the plane of the two primary bodies, in contrast with halo orbits and Lissajous orbits, which move above and below the plane also.

Lyapunov stability is named after Aleksandr Mikhailovich Lyapunov, a Russian mathematician who defended the thesis *The General Problem of Stability of Motion* at Kharkov University in 1892.^{[1]} A. M. Lyapunov was a pioneer in successful endeavoring to develop the global approach to the analysis of the stability of nonlinear dynamical systems by comparison with the widely spread local method of linearizing them about points of equilibrium. His work, initially published in Russian and then translated to French, received little attention for many years. The mathematical theory of stability of motion, founded by A. M. Lyapunov, considerably anticipated the time for its implementation in science and technology. Moreover Lyapunov did not himself make application in this field, his own interest being in the stability of rotating fluid masses with astronomical application. He did not have doctoral students who followed the research in the field of stability and his own destiny was terribly tragic because of the Russian revolution of 1917^{[]}. For several decades the theory of stability sank into complete oblivion. The Russian-Soviet mathematician and mechanician Nikolay Gur'yevich Chetaev working at the Kazan Aviation Institute in the 1930s was the first who realized the incredible magnitude of the discovery made by A. M. Lyapunov. Actually, his figure as a great scientist is comparable to the one of A. M. Lyapunov. The contribution to the theory made by N. G. Chetaev^{[2]} was so significant that many mathematicians, physicists and engineers consider him Lyapunov's direct successor and the next-in-line scientific descendant in the creation and development of the mathematical theory of stability.

The interest in it suddenly skyrocketed during the Cold War period when the so-called "Second Method of Lyapunov" (see below) was found to be applicable to the stability of aerospace guidance systems which typically contain strong nonlinearities not treatable by other methods. A large number of publications appeared then and since in the control and systems literature.^{[3]}^{[4]}^{[5]}^{[6]}^{[7]}
More recently the concept of the Lyapunov exponent (related to Lyapunov's First Method of discussing stability) has received wide interest in connection with chaos theory. Lyapunov stability methods have also been applied to finding equilibrium solutions in traffic assignment problems.^{[8]}

Consider an autonomous nonlinear dynamical system

- ,

where denotes the system state vector, an open set containing the origin, and is a continuous vector field on . Suppose has an equilibrium at so that then

- This equilibrium is said to be
**Lyapunov stable**, if, for every , there exists a such that, if , then for every we have . - The equilibrium of the above system is said to be
**asymptotically stable**if it is Lyapunov stable and there exists such that if , then . - The equilibrium of the above system is said to be
**exponentially stable**if it is asymptotically stable and there exist such that if , then , for all .

Conceptually, the meanings of the above terms are the following:

- Lyapunov stability of an equilibrium means that solutions starting "close enough" to the equilibrium (within a distance from it) remain "close enough" forever (within a distance from it). Note that this must be true for
*any*that one may want to choose. - Asymptotic stability means that solutions that start close enough not only remain close enough but also eventually converge to the equilibrium.
- Exponential stability means that solutions not only converge, but in fact converge faster than or at least as fast as a particular known rate .

The trajectory * is (locally) **attractive* if

- as

for all trajectories that start close enough to , and *globally attractive* if this property holds for all trajectories.

That is, if *x* belongs to the interior of its stable manifold, it is *asymptotically stable* if it is both attractive and stable. (There are examples showing that attractivity does not imply asymptotic stability. Such examples are easy to create using homoclinic connections.)

If the Jacobian of the dynamical system at an equilibrium happens to be a stability matrix (i.e., if the real part of each eigenvalue is strictly negative), then the equilibrium is asymptotically stable.

Instead of considering stability only near an equilibrium point (a constant solution ), one can formulate similar definitions of stability near an arbitrary solution . However, one can reduce the more general case to that of an equilibrium by a change of variables called a "system of deviations". Define , obeying the differential equation:

- .

This is no longer an autonomous system, but it has a guaranteed equilibrium point at whose stability is equivalent to the stability of the original solution .

Lyapunov, in his original 1892 work, proposed two methods for demonstrating stability.^{[1]} The first method developed the solution in a series which was then proved convergent within limits. The second method, which is now referred to as the Lyapunov stability criterion or the Direct Method, makes use of a *Lyapunov function V(x)* which has an analogy to the potential function of classical dynamics. It is introduced as follows for a system having a point of equilibrium at . Consider a function such that

- if and only if
- if and only if
- for all values of . Note: for asymptotic stability, for is required.

Then *V(x)* is called a Lyapunov function and the system is stable in the sense of Lyapunov (Note that is required; otherwise for example would "prove" that is locally stable). An additional condition called "properness" or "radial unboundedness" is required in order to conclude global stability. Global asymptotic stability (GAS) follows similarly.

It is easier to visualize this method of analysis by thinking of a physical system (e.g. vibrating spring and mass) and considering the energy of such a system. If the system loses energy over time and the energy is never restored then eventually the system must grind to a stop and reach some final resting state. This final state is called the attractor. However, finding a function that gives the precise energy of a physical system can be difficult, and for abstract mathematical systems, economic systems or biological systems, the concept of energy may not be applicable.

Lyapunov's realization was that stability can be proven without requiring knowledge of the true physical energy, provided a Lyapunov function can be found to satisfy the above constraints.

The definition for discrete-time systems is almost identical to that for continuous-time systems. The definition below provides this, using an alternate language commonly used in more mathematical texts.

Let (*X*, *d*) be a metric space and *f* : *X* -> *X* a continuous function. A point *x* in *X* is said to be **Lyapunov stable**, if,

We say that *x* is **asymptotically stable** if it belongs to the interior of its stable set, *i.e.* if,

A linear state space model

- ,

where is a finite matrix, is asymptotically stable (in fact, exponentially stable) if all real parts of the eigenvalues of are negative. This condition is equivalent to the following one:^{[9]}

is negative definite for some positive definite matrix . (The relevant Lyapunov function is .)

Correspondingly, a time-discrete linear state space model

is asymptotically stable (in fact, exponentially stable) if all the eigenvalues of have a modulus smaller than one.

This latter condition has been generalized to switched systems: a linear switched discrete time system (ruled by a set of matrices )

is asymptotically stable (in fact, exponentially stable) if the joint spectral radius of the set is smaller than one.

A system with inputs (or controls) has the form

where the (generally time-dependent) input u(t) may be viewed as a *control*, *external input*,
*stimulus*, *disturbance*, or *forcing function*. It has been shown ^{[10]} that near to a point of equilibrium which is Lyapunov stable the system remains stable under small disturbances. For larger input disturbances the study of such systems is the subject of control theory and applied in control engineering. For systems with inputs, one must quantify the effect of inputs on the stability of the system. The main two approaches to this analysis are BIBO stability (for linear systems) and input-to-state stability (ISS) (for nonlinear systems)

Consider an equation, where compared to the Van der Pol oscillator equation the friction term is changed:

Here is a good example of an unsuccessful try to find a Lyapunov function that proves stability.

Let

so that the corresponding system is

- The equilibrium is

Let us choose as a Lyapunov function

which is clearly positive definite. Its derivative is

It seems that if the parameter is positive, stability is asymptotic for But this is wrong, since does not depend on , and will be 0 everywhere on the axis. The equilibrium is Lyapunov stable.

Assume that f is a function of time only.

- Having does not imply that has a limit at . For example, .
- Having approaching a limit as does not imply that . For example, .
- Having lower bounded and decreasing () implies it converges to a limit. But it does not say whether or not as .

Barbalat's Lemma says:

- If has a finite limit as and if is uniformly continuous (or is bounded), then as .
^{[11]}

An alternative version is as follows:

- Let and . If and , then as
^{[12]}

In the following form the Lemma is true also in the vector valued case:

- Let be a uniformly continuous function with values in a Banach space and assume that has a finite limit as . Then as .
^{[13]}

The following example is taken from page 125 of Slotine and Li's book *Applied Nonlinear Control*.

Consider a non-autonomous system

This is non-autonomous because the input is a function of time. Assume that the input is bounded.

Taking gives

This says that by first two conditions and hence and are bounded. But it does not say anything about the convergence of to zero. Moreover, the invariant set theorem cannot be applied, because the dynamics is non-autonomous.

Using Barbalat's lemma:

- .

This is bounded because , and are bounded. This implies as and hence . This proves that the error converges.

- ^
^{a}^{b}Lyapunov, A. M.*The General Problem of the Stability of Motion*(In Russian), Doctoral dissertation, Univ. Kharkov 1892 English translations: (1)*Stability of Motion*, Academic Press, New-York & London, 1966 (2)*The General Problem of the Stability of Motion*, (A. T. Fuller trans.) Taylor & Francis, London 1992. Included is a biography by Smirnov and an extensive bibliography of Lyapunov's work. **^**Chetaev, N. G. On stable trajectories of dynamics, Kazan Univ Sci Notes, vol.4 no.1 1936; The Stability of Motion, Originally published in Russian in 1946 by ?. . - ?-. ., -.Translated by Morton Nadler, Oxford, 1961, 200 pages.**^**Letov, A. M. (1955).*?*[*Stability of Nonlinear Control Systems*] (in Russian). Moscow: Gostekhizdat. English tr. Princeton 1961**^**Kalman, R. E.; Bertram, J. F (1960). "Control System Analysis and Design Via the "Second Method" of Lyapunov: I--Continuous-Time Systems".*Journal of Basic Engineering*.**82**(2): 371-393. doi:10.1115/1.3662604.**^**LaSalle, J. P.; Lefschetz, S. (1961).*Stability by Lyapunov's Second Method with Applications*. New York: Academic Press.**^**Parks, P. C. (1962). "Liapunov's method in automatic control theory".*Control*. I Nov 1962 II Dec 1962.**^**Kalman, R. E. (1963). "Lyapunov functions for the problem of Lur'e in automatic control".*Proc Natl Acad Sci USA*.**49**(2): 201-205. Bibcode:1963PNAS...49..201K. doi:10.1073/pnas.49.2.201. PMC 299777. PMID 16591048.**^**Smith, M. J.; Wisten, M. B. (1995). "A continuous day-to-day traffic assignment model and the existence of a continuous dynamic user equilibrium".*Annals of Operations Research*.**60**(1): 59-79. doi:10.1007/BF02031940. S2CID 14034490.**^**Goh, B. S. (1977). "Global stability in many-species systems".*The American Naturalist*.**111**(977): 135-143. doi:10.1086/283144. S2CID 84826590.**^**Malkin I.G. Theory of Stability of Motion, Moscow 1952 (Gostekhizdat) Chap II para 4 (Russian) Engl. transl, Language Service Bureau, Washingotn AEC -tr-3352; originally On stability under constantly acting disturbances Prikl Mat 1944, vol. 8 no.3 241-245 (Russian); Amer. Math. Soc. transl. no. 8**^**I. Barb?lat, Systèmes d'équations différentielles d'oscillations non Linéaires, Rev. Math. Pures Appl. 4 (1959) 267-270, p. 269.**^**B. Farkas et al., Variations on Barb?lat's Lemma, Amer. Math. Monthly (2016) 128, no. 8, 825-830, DOI: 10.4169/amer.math.monthly.123.8.825, p. 827.**^**B. Farkas et al., Variations on Barb?lat's Lemma, Amer. Math. Monthly (2016) 128, no. 8, 825-830, DOI: 10.4169/amer.math.monthly.123.8.825, p. 826.

- Bhatia, Nam Parshad; Szeg?, Giorgio P. (2002).
*Stability theory of dynamical systems*. Springer. ISBN 978-3-540-42748-3. - Chervin, Robert (1971).
*Lyapunov Stability and Feedback Control of Two-Stream Plasma Systems*(PhD). Columbia University. - Gandolfo, Giancarlo (1996).
*Economic Dynamics*(Third ed.). Berlin: Springer. pp. 407-428. ISBN 978-3-540-60988-9. - Parks, P. C. (1992). "A. M. Lyapunov's stability theory--100 years on".
*IMA Journal of Mathematical Control & Information*.**9**(4): 275-303. doi:10.1093/imamci/9.4.275. - Slotine, Jean-Jacques E.; Weiping Li (1991).
*Applied Nonlinear Control*. NJ: Prentice Hall. - Teschl, G. (2012).
*Ordinary Differential Equations and Dynamical Systems*. Providence: American Mathematical Society. ISBN 978-0-8218-8328-0. - Wiggins, S. (2003).
*Introduction to Applied Nonlinear Dynamical Systems and Chaos*(2nd ed.). New York: Springer Verlag. ISBN 978-0-387-00177-7.

*This article incorporates material from asymptotically stable on PlanetMath, which is licensed under the Creative Commons Attribution/Share-Alike License.*

This article uses material from the Wikipedia page available here. It is released under the Creative Commons Attribution-Share-Alike License 3.0.

Popular Products

Music Scenes

Popular Artists