Вы находитесь на странице: 1из 100

Electronic Journal of Dierential Equations, Monograph 08, 2007, (101 pages).

ISSN: 1072-6691. URL: http://ejde.math.txstate.edu or http://ejde.math.unt.edu


ftp ejde.math.txstate.edu (login: ftp)
AN ALGORITHM FOR CONSTRUCTING
LYAPUNOV FUNCTIONS
SIGURDUR FREYR HAFSTEIN
Abstract. In this monograph we develop an algorithm for constructing Lya-
punov functions for arbitrary switched dynamical systems x = f(t, x), pos-
sessing a uniformly asymptotically stable equilibrium. Let x = fp(t, x), p P,
be the collection of the ODEs, to which the switched system corresponds. The
number of the vector elds fp on the right-hand side of the dierential equa-
tion is assumed to be nite and we assume that their components f
p,i
are C
2
functions and that we can give some bounds, not necessarily close, on their
second-order partial derivatives. The inputs of the algorithm are solely a nite
number of the function values of the vector elds fp and these bounds. The
domain of the Lyapunov function constructed by the algorithm is only limited
by the size of the equilibriums region of attraction. Note, that the concept of a
Lyapunov function for the arbitrary switched system x = f(t, x) is equivalent
to the concept of a common Lyapunov function for the systems x = fp(t, x),
p P, and that if P contains exactly one element, then the switched system
is just a usual ODE x = f (t, x). We give numerous examples of Lyapunov
functions constructed by our method at the end of this monograph.
Contents
1. Introduction 2
2. Outline and Categorization 4
3. Preliminaries 8
3.1. Continuous dynamical systems 8
3.2. Arbitrary switched systems 9
3.3. Dini derivatives 12
3.4. Stability of arbitrary switched systems 14
3.5. Three useful lemmas 17
3.6. Linear programming 19
4. Lyapunovs Direct Method for Switched Systems 20
5. Converse Theorem for Switched Systems 24
5.1. Converse theorems 24
5.2. A converse theorem for arbitrary switched systems 25
6. Construction of Lyapunov Functions 33
2000 Mathematics Subject Classication. 35J20, 35J25.
Key words and phrases. Lyapunov functions; switched systems; converse theorem;
piecewise ane functions.
c 2007 Texas State University - San Marcos.
Submitted August 29, 2006. Published August 15, 2007.
1
2 S. F. HAFSTEIN EJDE-2007/MON. 08
6.1. Continuous piecewise ane functions 34
6.2. The linear programming problem 39
6.3. Denition of the functions , , and V
Lya
44
6.4. Implications of the constraints (LC1) 44
6.5. Implications of the constraints (LC2) 45
6.6. Implications of the constraints (LC3) 46
6.7. Implications of the constraints (LC4) 46
6.8. Summary of the results and their consequences 48
6.9. The autonomous case 52
7. Constructive Converse Theorems 64
7.1. The assumptions 64
7.2. The assignments 65
7.3. The constraints (LC1) are fullled 66
7.4. The constraints (LC2) are fullled 67
7.5. The constraints (LC3) are fullled 68
7.6. The constraints (LC4) are fullled 68
7.7. Summary of the results 70
7.8. The autonomous case 71
8. An Algorithm for Constructing Lyapunov Functions 74
8.1. The algorithm in the nonautonomous case 74
8.2. The algorithm in the autonomous case 76
9. Examples of Lyapunov functions generated by linear programming 78
9.1. An autonomous system 78
9.2. An arbitrary switched autonomous system 80
9.3. A variable structure system 84
9.4. A variable structure system with sliding modes 88
9.5. A one-dimensional nonautonomous switched system 91
9.6. A two-dimensional nonautonomous switched system 93
10. Conclusion 96
List of Symbols 98
References 99
1. Introduction
Let T be a nonempty set and equip it with the discrete metric, let | R
n
be
a domain containing the origin, and let | | be a norm on R
n
. For every p T
assume that f
p
: R
0
| R
n
satises the local Lipschitz condition: for every
compact ( R
0
| there is a constant L
p,C
such that (t, x), (t, y) ( implies
|f
p
(t, x) f
p
(t, y)| L
p,C
|x y|. Dene B
,R
:= x R
n
: |x| < R for every
R > 0. We consider the switched system x = f

(t, x), where is an arbitrary right-


continuous mapping R
0
T of which the discontinuity-points form a discrete set.
In this monograph we establish the claims made in the abstract in the following
three steps:
First, we show that the origin is a uniformly asymptotically stable equilibrium
of the arbitrary switched system x = f

(t, x), whenever there exists a common


Lyapunov function for the systems x = f
p
(t, x), p T, and we show how to derive
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 3
a lower bound on the equilibriums region of attraction from such a Lyapunov
function.
Second, we show that if B
,R
| is a subset of the region of attraction of the
arbitrary switched system x = f

(t, x) and the vector elds f


p
, p T, satisfy the
Lipschitz condition: there exists a constant L such that for every p T and every
(s, x), (t, y) R
0
B
,R
the inequality |f
p
(s, x) f
p
(t, y)| L([s t[ +|xy|)
holds; then for every 0 < R

< R, there exists a common Lyapunov function


V : R
0
B
,R
R for the systems x = f
p
(t, x), p T.
Third, assuming that the set T is nite and that the second-order partial deriva-
tives of the components of the vector elds f
p
are bounded, we write down a linear
programming problem using the function values of the vector elds f
p
on a dis-
crete set and bounds on the second-order partial derivatives of their components.
Then we show how to parameterize a common Lyapunov function for the systems
x = f
p
(t, x), p T, from a feasible solution to this linear programming problem.
We then use these results to give an algorithm for constructing such a common
Lyapunov function for the systems x = f
p
(t, x), p T, and we prove that it always
succeeds in a nite number of steps if there exists a common Lyapunov function
for the systems at all.
Let us be more specic on this last point. Consider the systems x = f
p
(t, x), p
T, and assume that they possess a common Lyapunov function W : R
0
1 R,
where 1 | is a domain containing the origin. That is, there exist functions , ,
and of class / such that
(|x|) W(t, x) (|x|)
and

x
W(t, x) f
p
(t, x) +
W
t
(t, x) (|x|)
for every p T, every t R
0
and every x 1. The second inequality can
equivalently be formulated as
limsup
h0
W(t +h, x +hf
p
(t, x)) W(t, x)
h
(|x|).
Now, let t
1
, t
2
R be arbitrary constants such that 0 t
1
< t
2
< + and let (
and T be arbitrary compact subsets of R
n
of positive measure such that the origin
is in the interior of T and T ( 1. We will prove that the algorithm will always
succeed in generating a continuous function V : [t
1
, t
2
] ( R with the property,
that there exist functions

, and

of class /, such that

(|x|) V (t, x)

(|x|)
for all t [t
1
, t
2
] and all x ( and
limsup
h0
V (t +h, x +hf
p
(t, x)) V (t, x)
h

(|x|)
for every p T, every t [t
1
, t
2
], and every x (T. Note, that the last inequality
is not necessarily valid for x T, but because one can take T as small as one wishes,
this is nonessential.
It is reasonable to consider the autonomous case separately, because then there
exist common autonomous Lyapunov functions whenever the origin is an asymptot-
ically stable equilibrium, and the algorithm is then also able to construct common
autonomous Lyapunov functions for the systems x = f
p
(x), p T.
4 S. F. HAFSTEIN EJDE-2007/MON. 08
2. Outline and Categorization
This monograph is thematically divided into three parts. In the rst part,
which consists of the sections 3, 4, and 5, we develop a stability theory for ar-
bitrary switched systems. In Section 3 we introduce switched dynamical systems
x = f

(t, x) and discuss some elementary properties of their solutions. We further


consider the stability of isolated equilibria of arbitrary switched systems and we
prove in Section 4 that if such a system possesses a Lyapunov function, then the
equilibrium is uniformly asymptotically stable. These results are quite straightfor-
ward if one is familiar with the Lyapunov stability theory for ordinary dierential
equations (ODEs), but, because we consider Lyapunov functions that are merely
continuous and not necessarily dierentiable in this work, we establish the most
important results. Switched systems have gained much interest recently. For an
overview see, for example, [33] and [31].
In Section 5 we prove a converse theorem on uniform asymptotic stability for
arbitrary switched nonlinear, nonautonomous, continuous systems. In the litera-
ture there are numerous results regarding the existence of Lyapunov functions for
switched systems. A short non-exhaustive overview follows: In [48] Narendra and
Balakrishnan consider the problem of common quadratic Lyapunov functions for
a set of autonomous linear systems, in [15], in [46], in [32], and in [2] the results
were considerably improved by Gurvits; Mori, Mori, and Kuroe; Liberzon, Hes-
panha, and Morse; and Agrachev and Liberzon respectively. Shim, Noh, and Seo
in [57] and Vu and Liberzon in [61] generalized the approach to commuting au-
tonomous nonlinear systems. The resulting Lyapunov function is not necessarily
quadratic. Dayawansa and Martin proved in [41] that a set of linear autonomous
systems possesses a common Lyapunov function, whenever the corresponding arbi-
trary switched system is asymptotically stable, and they proved that even in this
simple case there might not exist any quadratic Lyapunov function. The same
authors generalized their approach to exponentially stable nonlinear, autonomous
systems in [7]. Mancilla-Aguilar and Garca used results from Lin, Sontag, and
Wang in [35] to prove a converse Lyapunov theorem on asymptotically stable non-
linear, autonomous switched systems in [38].
In this work we prove a converse Lyapunov theorem for uniformly asymptotically
stable nonlinear switched systems and we allow the systems to depend explicitly
on the time t, that is, we work the nonautonomous case out. We proceed as
follows: Assume that the functions f
p
, of the arbitrary switched system x = f

(t, x),
: R
0
T, satisfy the Lipschitz condition: there exists a constant L such
that |f
p
(t, x) f
p
(t, y)|
2
L|x y|
2
for all p T, all t 0, and all x, y
in some compact neighborhood of the origin. Then, by combining a Lyapunov
function construction method by Massera for ODEs, see, for example, [42] or Section
5.7 in [60], with the construction method presented by Dayawansa and Martin in
[7], it is possible to construct a Lyapunov function V for the system. However,
we need the Lyapunov function to be smooth, so we prove that if the functions
f
p
, p T, satisfy the Lipschitz condition: there exists a constant L such that
|f
p
(t, x) f
p
(s, y)|
2
L([t s[ + |x y|
2
) for all p T, all s, t 0, and all
x, y in some compact neighborhood of the origin in the state-space, then we can
smooth out the Lyapunov function to be innitely dierentiable except possibly at
the origin.
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 5
By Lin, Sontag, and Wang, in [35, Lemma 4.3], this implies for autonomous
systems that there exists a (

Lyapunov function for the system. This has been


stated in the literature before by Wilson [66], but the claim was incorrect in that
context as shown below. In this monograph, however, we do not need this to hold
true, neither in the autonomous nor in the nonautonomous case.
Wilson states in [66] that if ^ R
n
is a neighborhood of the origin
and W ((^) (

(^ 0) is Lipschitz at the origin, that is


[W(x)[ L|x| for some constant L, then
x W(x) exp(1/W(x))
is a (

(^) function. In [47] Nadzieja repairs some other parts of


Wilsons proof and notices also that this is by no means canoni-
cal. He, however, argues that this must hold true because x
W(x) exp(1/W(x)) converges faster to zero than any rational
function in |x| grows at the origin. Unfortunately, this argument is
not satisfactory because some arbitrary derivative of W might still
diverge to fast at the origin. As a counterexample one can take a
function that oscillates heavily at the origin, for example
W : R R, W(x) = xsin(exp(
1
x
2
)).
It is not dicult to see that
d
dx
_
W(x) exp(1/W(x))

= W

(x) exp(1/W(x))
W(x) + 1
W(x)
does not have a limit when x approaches zero.
In the second part of this monograph, which consists of the sections 6, 7 and
8, we give an algorithmic construction scheme of a linear programming problem
for the switched system x = f

(t, x), : R
0
T, T ,= nite, where the f
p
are assumed to be (
2
functions. Further, we prove that if this linear programming
problem possesses a feasible solution, then such a solution can be used to param-
eterize a function that is a Lyapunov function for all of the individual systems
and then a Lyapunov function for the arbitrary switched system. We then use
this fact to derive an algorithm for constructing Lyapunov functions for nonlinear,
nonautonomous, arbitrary switched systems that possess a uniformly asymptoti-
cally stable equilibrium.
In Section 6 we introduce the function space CPWA, a set of continuous functions
R
n
R that are piecewise ane (often called piecewise linear in the literature)
with a certain simplicial boundary conguration. The spaces CPWA are essentially
the function spaces PWL, presented by Julian, Desages, and Agamennoni in [26],
Julian, Guivant, and Desages in [27], and by Julian in [25], with variable grid
sizes. A function space CPWA dened on a compact domain is a nite dimensional
vector space over R, which makes it particularly well suited as the foundation
for the search of a parameterized Lyapunov function. Another property which
renders them appropriate as a search space, is that a function in (
2
can be neatly
approximated by a function in CPWA as shown in Lemma 6.6.
In Section 6 we further dene our linear programming problem in Denition 6.8
and then we show how to use a feasible solution to it to parameterize a Lyapunov
function for the corresponding switched system. We discuss the autonomous case
6 S. F. HAFSTEIN EJDE-2007/MON. 08
separately, because in this case it is possible to parameterize an autonomous Lya-
punov function with a more simple linear programming problem, which is dened
in Denition 6.12. These results are generalizations of former results by the author,
presented in [39, 40, 17, 16, 18].
In Section 7 we prove, that if we construct a linear programming problem as
in Denition 6.8 for a switched system that possesses a uniformly asymptotically
stable equilibrium, then, if the boundary conguration of the function space CPWA
is suciently closely meshed, there exist feasible solutions to the linear program-
ming problem. There are algorithms, for example the simplex algorithm, that
always nd a feasible solution to a linear programming problem, provided there
exists at least one. This implies that we have reduced the problem of constructing
a Lyapunov function for the arbitrary switched system to a more simple prob-
lem of choosing an appropriate boundary conguration for the CPWA space. If
the systems x = f
p
(t, x), p T, are autonomous and exponentially stable, then
it was proved in [17] that it is even possible to calculate the mesh-sizes directly
from the original data, that is, the functions f
p
. This, however, is much more re-
strictive than necessary, because a systematic scan of boundary congurations is
considerably more eective, will lead to success for merely uniformly asymptotically
stable nonautonomous systems, and delivers a boundary conguration that is more
coarsely meshed. Just as in Section 5 we consider the more simple autonomous
case separately.
In Section 8 we use the results from Section 7 to dene our algorithm in Procedure
8.1 to construct Lyapunov functions and we prove in Theorem 8.2 that it always
delivers a Lyapunov function, if the arbitrary switched system possesses a uniformly
asymptotically stable equilibrium. For autonomous systems we do the same in
Procedure 8.3 and Theorem 8.4. These procedures and theorems are generalizations
of results presented by the author in [17, 16, 18].
In the last decades there have been several proposals of how to numerically con-
struct Lyapunov functions. For comparison to the construction method presented
in this work some of these are listed below. This list is by no means exhaustive.
In [59] Vandenberghe and Boyd present an interior-point algorithm for construct-
ing a common quadratic Lyapunov function for a nite set of autonomous linear
systems and in [34] Liberzon and Tempo took a somewhat dierent approach to
do the same and introduced a gradient decreasing algorithm. Booth methods are
numerically ecient, but unfortunately, limited by the fact that there might exist
Lyapunov functions for the system, non of which is quadratic. In [23], [24], and
[22] Johansson and Rantzer proposed construction methods for piecewise quadratic
Lyapunov functions for piecewise ane autonomous systems. Their construction
scheme is based on continuity matrices for the partition of the respective state-
space. The generation of these continuity matrices remains, to the best knowledge
of the author, an open problem. Further, piecewise quadratic Lyapunov functions
seem improper for the following reason:
Let V be a Lyapunov function for some autonomous system. Ex-
panding in power series about some y in the state-space gives
V (x) V (y) +V (y) (x y) +
1
2
(x y)
T
H
V
(y)(x y),
where H
V
(y) is the Hessian matrix of V at y, as a second-order
approximation. Now, if y is an equilibrium of the system, then
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 7
necessarily V (y) = 0 and V (y) = 0 and the second-order approx-
imation simplies to
V (x)
1
2
(x y)
T
H
V
(y)(x y),
which renders it very reasonable to make the Lyapunov function
ansatz
W(x) = (x y)
T
A(x y)
for some square matrix A and then try to determine a suitable
matrix A. This will, if the equilibrium is exponentially stable, de-
liver a function that is a Lyapunov function for the system in some
(possibly small) vicinity of the equilibrium.
However, if y is not an equilibrium of the system, then V (y) ,= 0
and V (y) ,= 0 and using several second-order power series expan-
sions about dierent points in the state-space that are not equilib-
ria, each one supposed to be valid on some subset of the state-space,
becomes problematic. Either one has to try to glue the local approx-
imations together in a continuous fashion, which causes problems
because the result is in general not a Lyapunov function for the
system, or one has to consider non-continuous Lyapunov functions,
which causes even more problems.
Brayton and Tong in [4, 5], Ohta, Imanishi, Gong, and Haneda in [49], Michel,
Sarabudla, and Miller in [45], and Michel, Nam, and Vittal in [44] reduced the Lya-
punov function construction for a set of autonomous linear systems to the design
of a balanced polytope fullling certain invariance properties. Polanski in [52] and
Koutsoukos and Antsaklis in [30] consider the construction of a Lyapunov function
of the form V (x) := |Wx|

, where W is a matrix, for autonomous linear systems


by linear programming. Julian, Guivant, and Desages in [27] and Julian in [25]
presented a linear programming problem to construct piecewise ane Lyapunov
functions for autonomous piecewise ane systems. This method can be used for
autonomous, nonlinear systems if some a posteriori analysis of the generated Lya-
punov function is done. The dierence between this method and our (non-switched
and autonomous) method is described in Section 6.2 in [40]. In [21] Johansen uses
linear programming to parameterize Lyapunov functions for autonomous nonlin-
ear systems. His results are, however, only valid within an approximation error,
which is dicult to determine. P. Parrilo in [51] and Papachristodoulou and Prajna
in [50] consider the numerical construction of Lyapunov functions that are pre-
sentable as sums of squares for autonomous polynomial systems under polynomial
constraints. Recently, Giesl proposed in [11, 12, 13, 14] a method to construct
Lyapunov functions for autonomous systems with exponentially stable equilibrium
by solving numerically a generalized Zubov equation (see, for example, [68] and for
an extension to perturbed systems [6])
V (x) f (x) = p(x), (2.1)
where usually p(x) = |x|
2
. A solution to the partial dierential equation (2.1)
is a Lyapunov function for the system. He uses radial basis functions to nd a
numerical approximation to the solution of (2.1).
In the third part of this monograph in Section 9 we give several examples of Lya-
punov functions generated by the linear programming problems from Denition 6.8
8 S. F. HAFSTEIN EJDE-2007/MON. 08
(nonautonomous) and Denition 6.12 (autonomous). Further, in Section 10, we
give some nal words. The examples are as follows: In Section 9.1 we generate a
Lyapunov function for a two-dimensional, autonomous, nonlinear ODE, of which
the equilibrium is asymptotically stable but not exponentially stable. In Section 9.2
we consider three dierent two-dimensional, autonomous, nonlinear ODEs and we
generate a Lyapunov function for each of them. Then we generate a Lyapunov func-
tion for the corresponding arbitrary switched system. In Section 9.3 we generate
a Lyapunov function for a two-dimensional, autonomous, piecewise linear variable
structure system without sliding modes. Variable structure systems are switched
systems, where the switching is not arbitrary but is performed in dependence of
the current state-space position of the system. Such systems are not discussed the
theoretical part of this monograph, but as explained in the example such an ex-
tension is straight forward. For variable structure systems, however, one cannot
use the theorems that guarantee the success of the linear programming problem in
parameterizing a Lyapunov function. In Section 9.4 we generate a Lyapunov func-
tion for a two-dimensional, autonomous, piecewise ane variable structure system
with sliding modes. In Section 9.5 we generate Lyapunov functions for two dif-
ferent one-dimensional, nonautonomous, nonlinear systems. We then parameterize
a Lyapunov function for the corresponding arbitrary switched system. Finally, in
Section 9.6, we parameterize Lyapunov functions for two dierent two-dimensional,
nonautonomous, nonlinear systems. Then we generate a Lyapunov function for the
corresponding arbitrary switched system.
3. Preliminaries
In this section we list some results, most of them well known or straightforward
extensions of well known results, that we will need later on in this monograph and
we give some references for further reading.
3.1. Continuous dynamical systems. A continuous dynamical system is a sys-
tem, of which the dynamics can be modelled by an ODE of the form
x = f (t, x). (3.1)
This equation is called the state equation of the dynamical system. We refer to x
as the state of the system and to the set of all possible states as the state-space of
the system. Further, we refer to t as the time of the system. For our purposes it is
advantageous to assume that f is a vector eld R
0
| R
n
, where | R
n
is a
domain in R
n
.
One possibility to secure the existence and uniqueness of a solution for any initial
time t
0
and any initial state in the state-space of a system, is given by a Lipschitz
condition (for more general conditions see, for example, [63, III.12.VII]).
Let | | be a norm on R
n
and assume that f in (3.1) satises the local Lipschitz
condition: for every compact ( R
0
| there exists a constant L
C
R such that
|f (t, x) f (t, y)| L
C
|x y| for every (t, x), (t, y) (. (3.2)
Then there exists a unique global solution to the initial value problem
x = f (t, x), x(t
0
) = ,
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 9
for every t
0
R
0
and every | (see, for example, Theorems VI and IX in
III.10 in [63]) and we denote this solution by t (t, t
0
, ) and we say that is
the solution to the state equation (3.1).
For this reason we will, in this monograph, only consider continuous dynamical
systems, of which the dynamics are modelled by an ODE
x = f (t, x), (3.3)
where f : R
0
| R
n
satises a local Lipschitz condition as in (3.2).
Two well known facts about ODEs that we will need in this monograph are given
in the next two theorems:
Theorem 3.1. Let | R
n
be a domain, f : R
0
| R
n
satisfy a local Lipschitz
condition as in (3.2), and be the solution to the state equation x = f (t, x). Let
m N
0
and assume that f [(
m
(R
0
|)]
n
, that is, every component f
i
of f is
in (
m
(R
0
|), then ,

[(
m
(dom())]
n
.
Proof. See, for example, the Corollary at the end of III.13 in [63].
Theorem 3.2. Let 1 R be a nonempty interval, | | be a norm on R
n
, and let
| be a domain in R
n
. Let f , g : 1 | R
n
be continuous mappings and assume
that there exists a constant L R such that f satises the Lipschitz condition: there
exists a constant L such that
|f (t, x) f (t, y)| L|x y| for all t 1 and all x, y |.
Let t
0
1 and let , | and denote the (unique) global solution to the initial
value problem
x = f (t, x), x(t
0
) := ,
by y : 1
y
R
n
and let z : 1
z
R
n
be any solution to the initial value problem
x = g(t, x), x(t
0
) = .
Set := 1
y
1
z
and let and be constants, such that
| | and |f (t, z(t)) g(t, z(t))|
for all t . Then the inequality
|y(t) z(t)| e
L|tt0|
+

L
(e
L|tt0|
1)
holds for all t .
Proof. See, for example, Theorem III.12.V in [63].
3.2. Arbitrary switched systems. A switched system is basically a family of
dynamical systems and a switching signal, where the switching signal determines
which system in the family describes the dynamics at what times or states. As we
are concerned with the stability of switched systems under arbitrary switchings,
the following denition of a switching signal is sucient for our needs.
Denition 3.3 (Switching signal). Let T be a nonempty set and equip it with
the discrete metric d(p, q) := 1 if p ,= q. A switching signal : R
0
T is a
right-continuous function, of which the discontinuity-points are a discrete subset of
R
0
. The discontinuity-points are called switching times. For technical reasons it
is convenient to count zero with the switching times, so we agree upon that zero is
a switching time as well. We denote the set of all switching signals R
0
T by
o
P
.
10 S. F. HAFSTEIN EJDE-2007/MON. 08
With the help of the switching signal in the last denition we can dene the
concept of a switched system and its solution.
Denition 3.4 (Solution to a switched system). Let | R
n
be a domain, let T
be a nonempty set, and let f
p
: R
0
| R
n
, p T, be a family of mappings,
of which each f
p
, p T, satises a local Lipschitz condition as in (3.2). For every
switching signal o
P
we dene the solution t

(t, s, ) to the initial value


problem
x = f

(t, x), x(s) = , (3.4)


in the following way:
Denote by t
0
, t
1
, t
2
, . . . the switching times of in an increasing order. If there
is a largest switching time t
k
we set t
k+1
:= + and if there is no switching
time besides zero we set t
1
:= +. Let s R
0
and let k N
0
be such that
t
k
s < t
k+1
. Then

is dened by gluing together the trajectories of the systems


x = f
p
(t, x), p T,
using p := (s) between s and t
k+1
, p := (t
k+1
) between t
k+1
and t
k+2
, and in
general p := (t
i
) between t
i
and t
i+1
, i k + 1. Mathematically this can be
expressed inductively as follows:
Forward solution:
(i)

(s, s, ) = for all s R


0
and all |.
(ii) Denote by y the solution to the initial value problem
x = f
(s)
(t, x), x(s) = ,
on the interval [s, t
k+1
[, where k N
0
is such that t
k
s < t
k+1
. Then
we dene

(t, s, ) on the domain of t y(t) by

(t, s, ) := y(t).
If the closure of graph(y) is a compact subset of [s, t
k+1
] |, then the
limit lim
tt
k+1

y(t) exists and is in | and we dene

(t
k+1
, s, ) :=
lim
tt
k+1

y(t).
(iii) Assume

(t
i
, s, ) | is dened for some integer i k +1 and denote by
y the solution to the initial value problem
x = f
(ti)
(t, x), x(t
i
) =

(t
i
, s, ),
on the interval [t
i
, t
i+1
[ . Then we dene

(t, s, ) on the domain of t


y(t) by

(t, s, ) := y(t). If the closure of graph(y) is a compact subset of


[t
i
, t
i+1
] |, then the limit lim
tti+1
y(t) exists and is in | and we dene

(t
i+1
, s, ) := lim
tti+1
y(t).
Backward solution:
(i)

(s, s, ) = for all s R


0
and all |.
(ii) Denote by y the solution to the initial value problem
x = f
(t
k
)
(t, x), x(s) = ,
on the interval ]t
k
, s], where k N
0
is such that t
k
s < t
k+1
. Then
we dene

(t, s, ) on the domain of t y(t) by

(t, s, ) := y(t). If
the closure of graph(y) is not empty and a compact subset of [t
k
, s] |,
then the limit lim
tt
k
+
y(t) exists and is in | and we dene

(t
k
, s, ) :=
lim
tt
k
+
y(t).
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 11
(iii) Assume

(t
i
, s, ) | is dened for some integer i, 0 < i k and denote
by y the solution to the initial value problem
x = f
(ti)
(t, x), x(t
i
) =

(t
i
, s, ),
on the interval ]t
i1
, t
i
]. Then we dene

(t, s, ) on the domain of t


y(t) by

(t, s, ) := y(t). If the closure of graph(y) is a compact subset of


[t
i1
, t
i
] |, then the limit lim
tti1+
y(t) exists and is in | and we dene

(t
i1
, s, ) := lim
tti1+
y(t).
Thus, for every o
P
we have dened the solution

to the dierential equation


x = f

(t, x).
Note, just as one usually suppresses the time dependency of x in (3.1), that
is, writes x = f (t, x) instead of x(t) = f (t, x(t)), one usually suppresses the time
dependency of the switching signal too, that is, writes x = f

(t, x) instead of
x(t) = f
(t)
(t, x(t)).
Now, that we have dened the solution to (3.4) for every o
P
, we can dene
the switched system and its solution.
Switched System 3.5. Let | R
n
be a domain, let T be a nonempty set, and
let f
p
: R
0
| R
n
, p T, be a family of mappings, of which each f
p
, p T,
satises a local Lipschitz condition as in (3.2).
The arbitrary switched system
x = f

(t, x), o
P
,
is simply the collection of all the dierential equations
x = f

(t, x) : o
P
,
whose solutions we dened in Denition 3.4. The solution to the arbitrary
switched system is the collection of all the solutions

to the individual switched


systems.
Because the trajectories of the Switched System 3.5 are dened by gluing together
trajectory-pieces of the corresponding continuous systems, they inherit the following
important property: For every o
P
, every s R
0
, and every | the closure
of the graph of t

(t, s, ), t s, is not a compact subset of R


0
| and the
closure of the graph of t

(t, s, ), t s, is not a compact subset of R


>0
|.
Further, note that if , o
P
, ,= , then in general

(t, t
0
, ) is not equal to

(t, t
0
, ) and that if the Switched System 3.5 is autonomous, that is, none of the
vector elds f
p
, p T, does depend on the time t, then

(t, t

, x) =

(t t

, 0, ), where (s) := (s +t

) for all s 0,
for all t t

0 and all |. Therefore, we often suppress the middle argument


of the solution to an autonomous system and simply write

(t, ).
We later need the following generalization of Theorem 3.2 to switched systems.
Theorem 3.6. Consider the Switched System 3.5, let | | be a norm on R
n
, and
assume that the functions f
p
satisfy the Lipschitz condition: there exists a constant
L such that
|f
p
(t, x) f
p
(t, y)| L|x y|
12 S. F. HAFSTEIN EJDE-2007/MON. 08
for all t 0, all x, y |, and all p T. Let t
0
0, let , |, let , o
P
,
and assume there is a constant 0 such that
|f
(t)
(t, x) f
(t)
(t, x)|
for all t 0 and all x |.
Denote the solution to the initial value problem
x = f

(t, x), x(s


0
) = ,
by y : 1
y
R
n
and denote the solution to the initial value problem
x = f

(t, x), x(s


0
) = ,
by z : 1
z
R
n
. Set := 1
y
1
z
and set := | |. Then the inequality
|y(t) z(t)| e
L|ts0|
+

L
(e
L|ts0|
1) (3.5)
holds for all t .
Proof. We prove only inequality (3.5) for t s
0
, the case t < s
0
follows similarly.
Let s
1
be the smallest real number larger than s
0
that is a switching time of or
a switching time of . If there is no such a number, then set s
1
:= sup
xJ
x. By
Theorem 3.2 inequality (3.5) holds for all t, s
0
t < s
1
. If s
1
= sup
xJ
x we
are nished with the proof, otherwise s
1
and inequality (3.5) holds for t = s
1
too. In the second case, let s
2
be the smallest real number larger than s
1
that is a
switching time of or a switching time of . If there is no such a number, then set
s
2
:= sup
xJ
x. Then, by Theorem 3.2,
|y(t) z(t)|
_
e
L(s1s0)
+

L
(e
L(s1s0)
1)
_
e
L(ts1)
+

L
(e
L(ts1)
1)
= e
L(ts0)
+

L
e
L(ts0)


L
e
L(ts1)
+

L
e
L(ts1)


L
= e
L(ts0)
+

L
(e
L(ts0)
1)
for all t such that s
1
t < s
2
. As this argumentation can, if necessary, be repeated
ad innitum, inequality (3.5) holds for all t s
0
such that t .
3.3. Dini derivatives. The Italian mathematician Ulisse Dini introduced in 1878
in his textbook [8] on analysis the so-called Dini derivatives. They are a gener-
alization of the classical derivative and inherit some important properties from it.
Because the Dini derivatives are point-wise dened, they are more suited for our
needs than some more modern approaches to generalize the concept of a derivative
like Sobolev Spaces (see, for example, [1]) or distributions (see, for example, [64]).
The Dini derivatives are dened as follows:
Denition 3.7 (Dini derivatives). Let 1 R, g : 1 R be a function, and y 1.
(i) Assume y is a limit point of 1 ]y, +[. Then the right-hand upper Dini
derivative D
+
of g at the point y is dened by
D
+
g(y) := limsup
xy+
g(x) g(y)
x y
:= lim
0+
_
sup
xI]y,+[
0<xy
g(x) g(y)
x y
_
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 13
and the right-hand lower Dini derivative D
+
of g at the point y is dened
by
D
+
g(y) := liminf
xy+
g(x) g(y)
x y
:= lim
0+
_
inf
xI]y,+[
0<xy
g(x) g(y)
x y
_
.
(ii) Assume y is a limit point of 1 ] , y[. Then the left-hand upper Dini
derivative D

of g at the point y is dened by


D

g(y) := limsup
xy
:= lim
0
_
sup
xI],y[
xy<0
g(x) g(y)
x y
_
and the left-hand lower Dini derivative D

of g at the point y is dened by


D

g(y) := liminf
xy
g(x) g(y)
x y
:= lim
0
_
inf
xI],y[
xy<0
g(x) g(y)
x y
_
.
The four Dini derivatives dened in Denition 3.7 are sometimes called the de-
rived numbers of g at y, or more exactly the right-hand upper derived number,
the right-hand lower derived number, the left-hand upper derived number, and the
left-hand lower derived number respectively.
It is clear from elementary calculus, that if g : 1 R is a function from a
nonempty open subset 1 R into R and y 1, then all four Dini derivatives
D
+
g(y), D
+
g(y), D

g(y), and D

g(y) of g at the point y exist. This means that


if 1 is a nonempty open interval, then the functions D
+
g, D
+
g, D

g, D

g : 1 R
dened in the canonical way, are all properly dened. It is not dicult to see that
if this is the case, then the classical derivative g

: 1 R of g exists, if and only if


the Dini derivatives are all real-valued and D
+
g = D
+
g = D

g = D

g.
Using limsup and liminf instead of the usual limit in the denition of a derivative
has the advantage, that they are always properly dened. The disadvantage is, that
because of the elementary
limsup
xy+
[g(x) +h(x)] limsup
xy+
g(x) + limsup
xy+
h(x),
a derivative dened in this way is not a linear operation anymore. However, when
the right-hand limit of the function h exists, then it is easy to see that
limsup
xy+
[g(x) +h(x)] = limsup
xy+
g(x) + lim
xy+
h(x).
This leads to the following lemma, which we will need later.
Lemma 3.8. Let g and h be real-valued functions, the domains of which are subsets
of R, and let D

D
+
, D
+
, D

, D

be a Dini derivative. Let y R be such,


that the Dini derivative D

g(y) is properly dened and h is dierentiable at y in


the classical sense. Then
D

[g +h](y) = D

g(y) +h

(y).
The reason why Dini derivatives are so useful for the applications in this mono-
graph, is the following generalization of the Mean-value theorem of dierential
calculus and its corollary.
Theorem 3.9 (Mean-value theorem for Dini derivatives). Let 1 be an interval of
strictly positive measure in R, let ( be a countable subset of 1, and let g : 1 R be
14 S. F. HAFSTEIN EJDE-2007/MON. 08
a continuous function. Let D

D
+
, D
+
, D

, D

be a Dini derivative and let


be an interval, such that D

g(x) for all x 1 (. Then


g(x) g(y)
x y

for all x, y 1, x ,= y.
Proof. See, for example, Theorem 12.24 in [62].
The previous theorem has an obvious corollary.
Corollary 3.10. Let 1 be an interval of strictly positive measure in R, let ( be
a countable subset of 1, let g : 1 R be a continuous function, and let D


D
+
, D
+
, D

, D

be a Dini derivative. Then:


D

g(x) 0 for all x 1 (, implies that g is a monotonically


increasing function on 1.
D

g(x) > 0 for all x 1 (, implies that g is a strictly monotonically


increasing function on 1.
D

g(x) 0 for all x 1 (, implies that g is a monotonically


decreasing function on 1.
D

g(x) < 0 for all x 1 (, implies that g is a strictly monotonically


decreasing function on 1.
3.4. Stability of arbitrary switched systems. The concepts equilibrium point
and stability are motivated by the desire to keep a dynamical system in, or at
least close to, some desirable state. The term equilibrium or equilibrium point of
a dynamical system, is used for a state of the system that does not change in the
course of time, that is, if the system is at an equilibrium at time t = 0, then it will
stay there for all times t > 0.
Denition 3.11 (Equilibrium point). A state y in the state-space of the Switched
System 3.5 is called an equilibrium or an equilibrium point of the system, if and
only if f
p
(t, y) = 0 for all p T and all t 0.
If y is an equilibrium point of Switched System 3.5, then obviously the initial
value problem
x = f

(t, x), x(0) = y


has the solution x(t) = y for all t 0 regardless of the switching signal o. The
solution with y as an initial value in the state-space is thus a constant vector and
the state does not change in the course of time. By a translation in the state-space
one can always reach that y = 0 without aecting the dynamics. Hence, there is no
loss of generality in assuming that a particular equilibrium point is at the origin.
A real-world system is always subject to some uctuations in the state. There
are some external eects that are unpredictable and cannot be modelled, some
dynamics that have (hopefully) very little impact on the behavior of the system
are neglected in the modelling, etc. Even if the mathematical model of a physical
system would be perfect, which hardly seems possible, the system state would
still be subject to quantum mechanical uctuations. The concept of local stability
in the theory of dynamical systems is motivated by the desire, that the system
state stays at least close to an equilibrium point after small uctuations in the
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 15
state. Any system that is expected to do something useful must have a predictable
behavior to some degree. This excludes all equilibria that are not locally stable as
usable working points for a dynamical system. Local stability is thus a minimum
requirement for an equilibrium. It is, however, not a very strong property. It merely
states, that there are disturbances that are so small, that they do not have a great
eect on the system in the long run. In this monograph we will concentrate on
uniform asymptotic stability on a set containing the equilibrium. This means that
we are demanding that the uniform asymptotic stability property of the equilibrium
is not merely valid for some, possibly arbitrary small, neighborhood of the origin,
but this property must hold on a a priori dened neighborhood of the origin. This
is a much more robust and powerful concept. It denotes, that all disturbances
up to a certain known degree are ironed out by the dynamics of the system, and,
because the domain of the Lyapunov functions is only limited by the size of the
equilibriums region of attraction, that we can get a reasonable lower bound on the
region of attraction.
The common stability concepts are most practically characterized by the use of
so-called /, L, and /L functions.
Denition 3.12 (Comparison functions /, L, and /L). The function classes /,
L, and /L of comparison functions are dened as follows:
(i) A continuous function : R
0
R
0
is said to be of class /, if and only if
(0) = 0, it is strictly monotonically increasing, and lim
r+
(r) = +.
(ii) A continuous function : R
0
R
0
is said to be of class L, if and only
if it is strictly monotonically decreasing and lim
s+
(s) = 0.
(iii) A continuous function : R
0
R
0
R
0
is said to be of class /L, if
and only if for every xed s R
0
the mapping r (r, s) is of class /
and for every xed r R
0
the mapping s (r, s) is of class L.
Note that some authors make a dierence between strictly monotonically in-
creasing functions that vanish at the origin and strictly monotonically increasing
functions that vanish at the origin and additionally asymptotically approach in-
nity at innity. They usually denote the functions of the former type as class /
functions and the functions of the latter type as class /

functions. We are not


interested in functions of the former type and in this work / always implies
lim
r+
(r) = +.
We now dene various stability concepts for equilibrium points of switched dy-
namical systems with help of the comparison functions.
Denition 3.13 (Stability concepts for equilibria). Assume that the origin is an
equilibrium point of the Switched System 3.5, denote by the solution to the
system, and let | | be an arbitrary norm on R
n
.
(i) The origin is said to be a uniformly stable equilibrium point of the Switched
System 3.5 on a neighborhood ^ | of the origin, if and only if there exists
an / such that for every o
P
, every t t
0
0, and every ^
the following inequality holds
|

(t, t
0
, )| (||).
(ii) The origin is said to be a uniformly asymptotically stable equilibrium point
of the Switched System 3.5 on the neighborhood ^ | of the origin, if and
16 S. F. HAFSTEIN EJDE-2007/MON. 08
only if there exists a /L such that for every o
P
, every t t
0
0,
and every ^ the following inequality holds
|

(t, t
0
, )| (||, t t
0
). (3.6)
(iii) The origin is said to be a uniformly exponentially stable equilibrium point
of the Switched System 3.5 on the neighborhood ^ | of the origin, if and
only if there exist constants k > 0 and > 0, such that for every o
P
,
every t t
0
0, and every ^ the following inequality holds
|

(t, t
0
, )| ke
(tt0)
||.
The stability denitions above imply, that if the origin is a uniformly exponen-
tially stable equilibrium of the Switched System 3.5 on the neighborhood ^, then
the origin is a uniformly asymptotically stable equilibrium on ^ as well, and, if the
origin is a uniformly asymptotically stable equilibrium of the Switched System 3.5
on the neighborhood ^, then the origin is a uniformly stable equilibrium on ^.
If the Switched System 3.5 is autonomous, then the stability concepts presented
above for the systems equilibria are uniform in a canonical way, that is, independent
of t
0
, and the denitions are somewhat more simple.
Denition 3.14. (Stability concepts for equilibria of autonomous systems) As-
sume that the origin is an equilibrium point of the Switched System 3.5, denote by
the solution to the system, let | | be an arbitrary norm on R
n
, and assume that
the system is autonomous.
(i) The origin is said to be a stable equilibrium point of the autonomous
Switched System 3.5 on a neighborhood ^ | of the origin, if and only
if there exists an / such that for every o
P
, every t 0, and every
^ the following inequality holds
|

(t, )| (||).
(ii) The origin is said to be an asymptotically stable equilibrium point of the
autonomous Switched System 3.5 on the neighborhood ^ | of the origin,
if and only if there exists a /L such that for every o
P
, every t 0,
and every ^ the following inequality holds
|

(t, )| (||, t).


(iii) The origin is said to be an exponentially stable equilibrium point of the
Switched System 3.5 on the neighborhood ^ | of the origin, if and only
if there exist constants k > 0 and > 0, such that for every o
P
, every
t 0, and every ^ the following inequality holds
|

(t, )| ke
t
||.
The set of those points in the state-space of a dynamical system, that are at-
tracted to an equilibrium point by the dynamics of the system, is of great impor-
tance. It is called the region of attraction of the equilibrium. Some authors prefer
domain, basin, or even bassin instead of region. For nonautonomous systems it
might depend on the initial time.
Denition 3.15 (Region of attraction). Assume that y = 0 is an equilibrium
point of the Switched System 3.5 and let be the solution to the system. For
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 17
every t
0
R
0
the set

t0
Att
:= | : limsup
t+

(t, t
0
, ) = 0 for all o
P

is called the region of attraction with respect to t


0
of the equilibrium at the origin.
The region of attraction
Att
of the equilibrium at the origin is dened by

Att
:=

t00

t0
Att
.
Thus, for the Switched System 3.5,
Att
implies lim
t+

(t, t
0
, ) = 0
for all o
P
and all t
0
0.
3.5. Three useful lemmas. It is often more convenient to work with smooth
rather that merely continuous functions and later on we need estimates by convex
(

/ functions. The two next lemmas state some useful facts in this regard.
Lemma 3.16. Let f : R
>0
R
0
be a monotonically decreasing function. Then
there exists a function g : R
>0
R
>0
with the following properties:
(i) g (

(R
>0
).
(ii) g(x) > f(x) for all x R
>0
.
(iii) g is strictly monotonically decreasing.
(iv) lim
x0+
g(x) = + and lim
x+
g(x) = lim
x+
f(x).
(v) g is invertible and g
1
(

(g(R
>0
)).
Proof. We dene the function

h : R
>0
R
>0
by,

h(x) :=
_
f
_
1
n+1
_
+
1
x
, if x [
1
n+1
,
1
n
[ for some n N
>0
,
f(n) +
1
x
, if x [n, n + 1[ for some n N
>0
,
and the function h : R
>0
R
>0
by
h(x) :=

h(x tanh(x)).
Then h is a strictly monotonically decreasing measurable function and because

h
is, by its denition, strictly monotonically decreasing and larger than f, we have
h(x + tanh(x)) =

h(x + tanh(x) tanh(x + tanh(x))) >

h(x) > f(x)


for all x R
>0
.
Let (

(R) such that (x) 0 for all x R, supp() ] 1, 1[, and


_
R
(x)dx = 1. We claim that the function g : R
>0
R
>0
,
g(x) :=
_
x+tanh(x)
xtanh(x)

_
x y
tanh(x)
_
h(y)
tanh(x)
dy =
_
1
1

1
(y)h(x y tanh(x))dy,
fullls the properties (i)(v).
Proposition (i) follows from elementary Lebesgue integration theory. Proposition
(ii) follows from
g(x) =
_
1
1
(y)h(x y tanh(x))dy
>
_
1
1
(y)h(x + tanh(x))dy
18 S. F. HAFSTEIN EJDE-2007/MON. 08
>
_
1
1
(y)f(x)dy = f(x).
To see that g is strictly monotonically decreasing let t > s > 0 and consider that
t y tanh(t) > s y tanh(s) (3.7)
for all y in the interval [1, 1]. Inequality (3.7) follows from
t y tanh(t) [s y tanh(s)] = t s y[tanh(t) tanh(s)]
= t s y(t s)(1 tanh
2
(s +
t,s
(t s))) > 0,
for some
t,s
[0, 1], where we used the Mean-value theorem. But then
h(t y tanh(t)) < h(s y tanh(s))
for all y [1, 1] and the denition of g implies that g(t) < g(s). Thus, proposition
(iii) is fullled.
Proposition iv) is obvious from the denition of g. Clearly g is invertible and by
the chain rule
[g
1
]

(x) =
1
g

(g
1
(x))
,
so it follows by mathematical induction that g
1
(

(g(R
>0
)), that is, proposition
(v).
Lemma 3.17. Let /. Then, for every R > 0, there is a function
R
/,
such that:
(i)
R
is a convex function.
(ii)
R
restricted to R
>0
is innitely dierentiable.
(iii) For all 0 x R we have
R
(x) (x).
Proof. By Lemma 3.16 there is a function g, such that g (

(R
>0
), g(x) > 1/(x)
for all x > 0, lim
t0+
g(x) = +, and g is strictly monotonically decreasing. Then
the function
R
: R
0
R
0
, dened through

R
(x) :=
1
R
_
x
0
d
g()
,
has the desired properties. First,
R
(0) = 0 and for every 0 < x R we have

R
(x) =
1
R
_
x
0
d
g()

1
g(x)
< (x).
Second, to prove that
R
is a convex / function is suces to prove that the second
derivative of
R
is strictly positive. But this follows immediately because for every
x > 0 we have g

(x) < 0, which implies


d
2

R
dx
2
(x) =
g

(x)
R[g(x)]
2
> 0.

The third existence lemma is the well known and very useful Masseras lemma
[42].
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 19
Lemma 3.18 (Masseras lemma). Let f L and R
>0
. Then there is a function
g (
1
(R
0
), such that g, g

/, g restricted to R
>0
is a (

(R
>0
) function,
_
+
0
g(f(t))dt < +, and
_
+
0
g

(f(t))e
t
dt < +.
Note, that because g, g

/ in Masseras lemma above, we have for every mea-


surable function u : R
0
R
0
, such that u(t) f(t) for all t R
0
, that
_
+
0
g(u(t))dt
_
+
0
g(f(t))dt and
_
+
0
g

(u(t))e
t
dt
_
+
0
g

(f(t))e
t
dt.
It is further worth noting that Masseras lemma can be proved quite simply by
using Lemma 3.16, which implies that there is a strictly monotonically decreasing
(

(R
>0
) bijective function h : R
>0
R
>0
such that h(x) > f(x) for all x > 0 and
h
1
(

(R
>0
). The function g : R
0
R
0
,
g(t) :=
_
t
0
e
(1+)h
1
()
d,
then fullls the claimed properties.
3.6. Linear programming. For completeness we spend a few words on linear
programming problems. A linear programming problem is a set of linear constraints,
under which a linear function is to be minimized. There are several equivalent
possibilities to state a linear programming problem, one of them is
minimize g(x) := c x,
given Cx b, x 0,
(3.8)
where r, s > 0 are integers, C R
sr
is a matrix, b R
s
and c R
r
are vectors,
and x y denotes x
i
y
i
for all i. The function g is called the objective of
the linear programming problem and the conditions Cx b and x 0 together
are called the constraints. A feasible solution to the linear programming problem
is a vector x

R
r
that satises the constraints, that is, x

0 and Cx

b.
There are numerous algorithms known to solve linear programming problems, the
most commonly used being the simplex method (see, for example, [56]) or interior
point algorithms, for example, the primal-dual logarithmic barrier method (see, for
example, [53]). Both need a starting feasible solution for initialization. A feasible
solution to (3.8) can be found by introducing slack variables y R
s
and solving
the linear programming problem:
minimize g(
_
x
y
_
) :=
s

i=1
y
i
,
given
_
C I
s

_
x
y
_
b,
_
x
y
_
0,
(3.9)
which has the feasible solution x = 0 and y = ([b
1
[, [b
2
[, . . . , [b
s
[). If the linear
programming problem (3.9) has the solution g([x

]) = 0, then x

is a feasible
solution to (3.8), if the minimum of g is strictly larger than zero, then (3.8) does
not have any feasible solution.
20 S. F. HAFSTEIN EJDE-2007/MON. 08
4. Lyapunovs Direct Method for Switched Systems
The Russian mathematician and engineer Alexandr Mikhailovich Lyapunov pub-
lished a revolutionary work in 1892 on the stability of motion, where he introduced
two methods to study the stability of general continuous dynamical systems. An
English translation of this work can be found in [36].
The more important of these two methods, known as Lyapunovs second method
or Lyapunovs direct method, enables one to prove the stability of an equilibrium
of (3.3) without integrating the dierential equation. It states, that if y = 0 is an
equilibrium point of the system, V (
1
(R
0
|) is a positive denite function,
that is, there exist functions
1
,
2
/ such that

1
(|x|
2
) V (t, x)
2
(|x|
2
)
for all x | and all t R
0
, and is the solution to the ODE (3.3). Then the
equilibrium is uniformly asymptotically stable, if there is an / such that the
inequality
d
dt
V (t, (t, t
0
, )) = [
x
V ](t, (t, t
0
, )) f (t, (t, t
0
, )) +
V
t
(t, (t, t
0
, ))
(|(t, t
0
, )|
2
)
(4.1)
holds for all (t, t
0
, ) in an open neighborhood ^ | of the equilibrium y. In
this case the equilibrium is uniformly asymptotically stable on a neighborhood,
which depends on V , of the origin. The function V satisfying (4.1) is said to
be a Lyapunov function for (3.3). The direct method of Lyapunov is covered in
practically all modern textbooks on nonlinear systems and control theory. Some
good examples are [19, 20, 28, 54, 60, 3, 65].
We will prove, that if the time-derivative in the inequalities above is replaced
with a Dini derivative with respect to t, then the assumption V (
1
(R
0
|) can
be replaced with the less restrictive assumption, that V is merely continuous. The
same is done in Theorem 42.5 in [19], but a lot of details are left out. Further, we
generalize the results to arbitrary switched systems.
Before we state and prove the direct method of Lyapunov for switched systems,
we prove a lemma that we use in its proof.
Lemma 4.1. Assume that the origin is an equilibrium of the Switched System 3.5
and let | | be a norm on R
n
. Further, assume that there is a function /, such
that for all o
P
and all t t
0
0 the inequality
|

(t, t
0
, )| (||) (4.2)
holds for all in some bounded neighborhood ^ | of the origin.
Under these assumptions the following two propositions are equivalent:
(i) There exists a function L, such that
|

(t, t
0
, )|
_
(||)(t t
0
)
for all o
P
, all t t
0
0, and all ^.
(ii) For every > 0 there exists a T > 0, such that for every t
0
0, every
o
P
, and every ^ the inequality
|

(t, t
0
, )|
holds for all t T +t
0
.
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 21
Proof. Let R > 0 be so large that ^ B
,R
and set C := max1, (R).
Note that Proposition (i) implies proposition (ii): For every > 0 we set T :=

1
(/
_
(R)) and proposition (ii) follows immediately.
Proposition (ii) implies proposition (i): For every > 0 dene

T() as the
inmum of all T > 0 with the property, that for every t
0
0, every o
P
, and
every ^ the inequality
|

(t, t
0
, )|
holds for all t T +t
0
.
Then

T is a monotonically decreasing function R
>0
R
0
and, because of (4.2),

T() = 0 for all > (R). By Lemma 3.16 there exists a strictly monotonically
decreasing (

(R
>0
) bijective function g : R
>0
R
>0
, such that g() >

T() for
all > 0. Now, for every pair t > t
0
0 set

:= g
1
(t t
0
) and note that because
t = g(

) +t
0


T(

) +t
0
we have
g
1
(t t
0
) =

(t, t
0
, )|.
But then
(s) :=
_
_
2C C/g(1) s, if s [0, g(1)],
_
Cg
1
(s), if s > g(1),
is an L function such that
_
|

(t, t
0
, )| (t t
0
),
for all t t
0
0 and all ^, and therefore
|

(t, t
0
, )|
_
(||)(t t
0
).

We come to the main theorem of this section: The Lyapunovs direct method
for arbitrary switched systems.
Theorem 4.2. Assume that the Switched System 3.5 has an equilibrium at the
origin. Let | | be a norm on R
n
and let R > 0 be a constant such that the closure
of the ball B
,R
is a subset of |. Let V : R
0
B
,R
R be a continuous
function and assume that there exist functions
1
,
2
/ such that

1
(||) V (t, )
2
(||)
for all t 0 and all B
,R
. Denote the solution to the Switched System 3.5
by and set d :=
1
2
(
1
(R)). Finally, let D

D
+
, D
+
, D

, D

be a Dini
derivative with respect to the time t, which means, for example with D

= D
+
, that
D
+
[V (t,

(t, t
0
, ))] := limsup
h0+
V (t +h,

(t +h, t
0
, )) V (t,

(t, t
0
, ))
h
.
Then the following propositions are true:
(i) If for every o
P
, every |, and every t t
0
0, such that

(t, t
0
, ) B
,R
, the inequality
D

[V (t,

(t, t
0
, ))] 0 (4.3)
holds, then the origin is a uniformly stable equilibrium of the Switched Sys-
tem 3.5 on B
,d
.
22 S. F. HAFSTEIN EJDE-2007/MON. 08
(ii) If there exists a function /, with the property that for every o
P
,
every |, and every t t
0
0, such that

(t, t
0
, ) B
,R
, the
inequality
D

[V (t,

(t, t
0
, ))] (|

(t, t
0
, )|) (4.4)
holds, then the origin is a uniformly asymptotically stable equilibrium of the
Switched System 3.5 on B
,d
.
Proof. Proposition (i): Let t
0
0, B
,d
, and o
P
all be arbitrary but xed.
By the note after the denition of Switched System 3.5 either

(t, t
0
, ) B
,R
for all t t
0
or there is a t

> t
0
such that

(s, t
0
, ) B
,R
for all s [t
0
, t

[
and

(t

, t
0
, ) B
,R
. Assume that the second possibility applies. Then, by
inequality (4.3) and Corollary 3.10

1
(R) V (t

(t

, t
0
, )) V (t
0
, )
2
(||) <
2
(d),
which is contradictory to d =
1
2
(
1
(R)). Therefore

(t, t
0
, ) B
,R
for all
t t
0
.
But then it follows by inequality (4.3) and Corollary 3.10 that

1
(|

(t, t
0
, )|) V (t,

(t, t
0
, )) V (t
0
, )
2
(||),
for all t t
0
, so
|

(t, t
0
, )|
1
1
(
2
(||))
for all t t
0
. Because
1
1

2
is a class / function, it follows, because t
0
0,
B
,d
, and o
P
were arbitrary, that the equilibrium at the origin is a
uniformly stable equilibrium point of the Switched System 3.5 on B
,d
.
Proposition (ii): Inequality (4.4) implies inequality (4.3) so Lemma 4.1 applies
and it suces to show that for every > 0 there is a nite T > 0, such that
t T +t
0
implies |

(t, t
0
, )| (4.5)
for all t
0
0, all B
,d
, and all o
P
. To prove this choose an arbitrary
> 0 and set

:= mind,
1
2
(
1
()) and T :=

2
(d)
(

)
.
We rst prove that for every o
P
the following proposition:
B
,d
and t
0
0 implies |

(t

, t
0
, )| <

(4.6)
for some t

[t
0
, T +t
0
]. We prove (4.6) by contradiction. Assume that
|

(t, t
0
, )|

(4.7)
for all t [t
0
, T +t
0
]. Then
0 <
1
(

)
1
(|

(T +t
0
, t
0
, )|) V (T +t
0
,

(T +t
0
, t
0
, )). (4.8)
By Theorem 3.9 and the assumption (4.7), there is an s [t
0
, T +t
0
], such that
V (T +t
0
,

(T +t
0
, t
0
, )) V (t
0
, )
T
[D

V ](s, (s, t
0
, ))]
(|

(s, t
0
, )|)
(

),
that is
V (T +t
0
,

(T +t
0
, t
0
, )) V (t
0
, ) T(

)
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 23

2
(||) T(

)
<
2
(d) T(

)
=
2
(d)

2
(d)
(

)
(

) = 0,
which is contradictory to (4.8). Therefore proposition (4.6) is true.
Now, let t

be as in (4.6) and let t > T +t


0
be arbitrary. Then, because
s V (s,

(s, t
0
, )), s t
0
,
is strictly monotonically decreasing by inequality (4.4) and Corollary 3.10, we get
by (4.6), that

1
(|

(t, t
0
, )|) V (t,

(t, t
0
, ))
V (t

(t

, t
0
, ))

2
(|

(t

, t
0
, )|)
<
2
(

)
= min
2
(d),
1
()

1
(),
and we have proved (4.5). The proposition (ii) follows.

The function V in the last theorem is called a Lyapunov function for the Switched
System 3.5.
Denition 4.3 (Lyapunov function). Assume that the Switched System 3.5 has
an equilibrium at the origin. Denote the solution to the Switched System 3.5 by
and let | | be a norm on R
n
. Let R > 0 be a constant such that the closure
of the ball B
,R
is a subset of |. A continuous function V : R
0
B
,R
R
is called a Lyapunov function for the Switched System 3.5 on B
,R
, if and only if
there exists a Dini derivative D

D
+
, D
+
, D

, D

with respect to the time t


and functions
1
,
2
, / with the properties that:
(L1)

1
(||) V (t, )
2
(||)
for all t 0 and all B
,R
.
(L2)
D

[V (t,

(t, t
0
, ))] (|

(t, t
0
, )|)
for every o
P
, every |, and every t t
0
0,
such that

(t, t
0
, ) B
,R
.
The Direct Method of Lyapunov (Theorem 4.2) can thus, by Denition 4.3, be
rephrased as follows:
Assume that the Switched System 3.5 has an equilibrium point at
the origin and that there exists a Lyapunov function dened on the
ball B
,R
, of which the closure is a subset of |, for the system.
Then there is a d, 0 < d < R, such that the origin is a uniformly
asymptotically stable equilibrium point of the system on the ball
B
,d
(which implies that B
,d
is a subset of the equilibriums
region of attraction). If the comparison functions
1
and
2
in the
24 S. F. HAFSTEIN EJDE-2007/MON. 08
condition (L1) for a Lyapunov function are known, then we can
take d =
1
2
(
1
(R)).
5. Converse Theorem for Switched Systems
In the last section we proved, that the existence of a Lyapunov function V
for the Switched System 3.5 is a sucient condition for the uniform asymptotic
stability of an equilibrium at the origin. In this section we prove the converse of
this theorem. That is, if the origin is a uniformly asymptotically stable equilibrium
of the Switched System 3.5, then there exists a Lyapunov function for the system.
Later, in Section 8, we prove that our algorithm always succeeds in construct-
ing a Lyapunov function for a switched system if the system possesses a Lyapunov
function, whose second-order derivatives are bounded on every compact subset of
the state-space that do not contain the origin. Thus, it is not sucient for our pur-
poses to prove the existence of a merely continuous Lyapunov function. Therefore,
we prove that if the origin is a uniformly asymptotically stable equilibrium of the
Switched System 3.5, then there exists a Lyapunov function for the system that is
innitely dierentiable with a possible exception at the origin.
5.1. Converse theorems. There are several theorems known, similar to Theo-
rem 4.2, where one either uses more or less restrictive assumptions regarding the
Lyapunov function than in Theorem 4.2. Such theorems are often called Lyapunov-
like theorems. An example for less restrictive assumptions is Theorem 46.5 in [19]
or equivalently Theorem 4.10 in [28], where the solution to a continuous system is
shown to be uniformly bounded, and an example for more restrictive assumptions is
Theorem 5.17 in [54], where an equilibrium is proved to be uniformly exponentially
stable. The Lyapunov-like theorems all have the form:
If one can nd a function V for a dynamical system, such that
V satises the properties X, then the system has the stability
property Y .
A natural question awakened by any Lyapunov-like theorem is whether its converse
is true or not, that is, if there is a corresponding theorem of the form:
If a dynamical system has the stability property Y , then there
exists a function V for the dynamical system, such that V satises
the properties X.
Such theorems are called converse theorems in the Lyapunov stability theory. For
nonlinear systems they are more involved than Lyapunovs direct method and the
results came rather late and did not stem from Lyapunov himself. The converse
theorems are covered quite thoroughly in Chapter VI in [19]. Some further general
references are Section 5.7 in [60] and Section 4.3 in [28]. More specic references
were given here in Section 2.
About the techniques to prove such theorems W. Hahn writes on page 225 in his
book Stability of Motion [19]:
In the converse theorems the stability behavior of a family of mo-
tions p(t, a, t
0
)
1
is assumed to be known. For example, it might
be assumed that the expression |p(t, a, t
0
)| is estimated by known
1
In our notation p(t, a, t
0
) = (t, t
0
, a).
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 25
comparison functions (secs. 35 and 36). Then one attempts to con-
struct by means of a nite or transnite procedure, a Lyapunov
function which satises the conditions of the stability theorem un-
der consideration.
In this section we prove a converse theorem on uniform asymptotic stability
of an arbitrary switched systems equilibrium, where the functions f
p
, p T, of
the systems x = f
p
(t, x), satisfy the common Lipschitz condition: there exists a
constant L such that
|f
p
(s, x) f
p
(t, y)| L([s t[ +|x y|)
for all p T, all s, t R
0
, and all x, y B
,R
. To construct a Lyapunov function
that is merely Lipschitz in its state-space argument, it suces that the functions
f
p
, p T, satisfy the common Lipschitz condition:
|f
p
(t, x) f
p
(t, y)| L
x
|x y| (5.1)
for all p T, all t R
0
, and all x, y B
,R
, as shown in Theorem 5.2. However,
our procedure to smooth it to a (

function everywhere except at the origin, as done


in Theorem 5.4, does not necessarily work if (s, t) |f
p
(s, x) f
p
(t, x)|/[s t[,
s ,= t, is unbounded. Note, that this additional assumption does not aect the
growth of the functions f
p
, p T, but merely excludes innitely fast oscillations
in the temporal domain. The Lipschitz condition (5.1) already takes care of that
|f
p
(t, x)| L
x
|x| L
x
R for all t 0 and all x B
,R
because f
p
(t, 0) = 0.
5.2. A converse theorem for arbitrary switched systems. The construction
here of a smooth Lyapunov function for the Switched System 3.5 is quite long
and technical. We therefore arrange the proof in a few denitions, lemmas, and
theorems. First, we use Masseras lemma (Lemma 3.18) to dene the functions
W

, o
P
, and them in turn to dene the function W, and after that we prove
that the function W is a Lyapunov function for the Switched System 3.5 used in
its construction.
Denition 5.1 (The functions W

and W). Assume that the origin is a uniformly


asymptotically stable equilibrium point of the Switched System 3.5 on the ball
B
,R
|, where | | is a norm on R
n
and R > 0, and let /L be such that
|

(t, t
0
, )| (||, t t
0
) for all o
P
, all B
,R
, and all t t
0
0.
Assume further, that there exists a constant L for the functions f
p
, such that
|f
p
(t, x) f
p
(t, y)| L|x y|
for all t 0, all x, y B
,R
, and all p T. By Masseras lemma (Lemma 3.18)
there exists a function g (
1
(R
0
), such that g, g

/, g is innitely dierentiable
on R
>0
,
_
+
0
g((R, ))d < +, and
_
+
0
g

((R, ))e
L
d < +.
(i) For every o
P
we dene the function W

for all t 0 and all B


,R
by
W

(t, ) :=
_
+
t
g(|

(, t, )|)d.
26 S. F. HAFSTEIN EJDE-2007/MON. 08
(ii) We dene the function W for all t 0 and all B
,R
by
W(t, ) := sup
S
P
W

(t, ).
Note, that if the Switched System 3.5 is autonomous, then W does not
depend on t, that is, it is time-invariant.
The function W from the denition above (Denition 5.1) is a Lyapunov function
for the Switched System 3.5 used in its construction. This is proved in the next
theorem.
Theorem 5.2 (Converse theorem for switched systems). The function W in Deni-
tion 5.1 is a Lyapunov function for the Switched System 3.5 used in its construction.
Further, there exists a constant L
W
> 0 such that
[W(t, ) W(t, )[ L
W
| | (5.2)
for all t 0 and all , B
,R
, where the norm | | and the constant R are the
same as in Denition 5.1.
Proof. We have to show that the function W complies to the conditions (L1) and
(L2) of Denition 4.3. Because

(u, t, ) = +
_
u
t
f
()
(,

(, t, ))d,
and |f
(s)
(s, y)| LR for all s 0 and all y B
,R
, we conclude |

(u, t, )|
|| (u t)LR for all u t 0, B
,R
, and all o
P
. Therefore,
|

(u, t, )|
||
2
whenever t u t +
||
2LR
,
which implies
W

(t, ) :=
_
+
t
g(|

(, t, )|)d
||
2LR
g(||/2)
and then
1
(||) W(t, ) for all t 0 and all B
,R
, where
1
(x) :=
x/(2LR)g(x/2) is a / function.
By the denition of W,
W(t, )
_
t+h
t
g(|

(, t, )|)d +W(t +h,

(t +h, t, ))
(reads: supremum over all trajectories emerging from
at time t is not less than over any particular trajectory
emerging from at time t)
for all B
,R
, all t 0, all small enough h > 0, and all o
P
, from which
limsup
h0+
W(t +h,

(t +h, t, )) W(t, )
h
g(||)
follows. Because g / this implies that the condition (L2) from Denition 4.3
holds for the function W.
Now, assume that there is an L
W
> 0 such that inequality (5.2) holds. Then
W(t, )
2
(||) for all t 0 and all B
,R
, where
2
(x) := L
W
x is a
class / function. Thus, it only remains to prove inequality (5.2). However, as this
inequality is a byproduct of the next lemma, we spare us the proof here.
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 27
The results of the next lemma are needed in the proof of our converse theorem on
uniform asymptotic stability of a switched systems equilibrium and as a convenient
side eect it completes the proof of Theorem 5.2.
Lemma 5.3. The function W in Denition 5.1 satises for all t s 0, all
, B
,R
, and all o
P
the inequality
W(t, ) W(s, ) C|

(t, s, )|
_
t
s
g(|

(, s, )|)d, (5.3)
where
C :=
_
+
0
g

((R, ))e
L
d < +.
Especially,
[W(t, ) W(t, )[ C| | (5.4)
for all t 0 and all , B
,R
.
The norm | |, the constants R, L, and the functions and g are, of course, the
same as in Denition 5.1.
Proof. By the Mean-value theorem and Theorem 3.6 we have
W

(t, ) W

(s, ) (5.5)
=
_
+
t
g(|

(, t, )|)d
_
+
s
g(|

(, s, )|)d

_
+
t

g(|

(, t, )|) g(|

(, s, )|)

d
_
t
s
g(|

(, s, )|)d
=
_
+
t

g(|

(, t, )|) g(|

(, t,

(t, s, ))|)

d
_
t
s
g(|

(, s, )|)d

_
+
t
g

((R, t))|

(, t, )

(, t,

(t, s, ))|d
_
t
s
g(|

(, s, )|)d

_
+
t
g

((R, t))e
L(t)
|

(t, s, )|d
_
t
s
g(|

(, s, )|)d
= C|

(t, s, )|
_
t
s
g(|

(, s, )|)d.
We now show that we can replace W

(t, ) W

(s, ) by W(t, ) W(s, ) on


the leftmost side of inequality (5.5) without violating the relations. That this is
possible might seem a little surprising at rst sight. However, a closer look reveals
that this is not surprising at all because the rightmost side of inequality (5.5) only
depends on the values of (z) for s z t and because W

(t, ) W(s, )
W

(t, ) W

(s, ), where the left-hand side only depends on the values of (z)
for z t, .
To rigidly prove the validity of this replacement let > 0 be an arbitrary constant
and choose a o
P
, such that
W(t, ) W

(t, ) <

2
, (5.6)
and a u > 0 so small that
ug((||, 0)) + 2CR(e
u
1) <

2
. (5.7)
28 S. F. HAFSTEIN EJDE-2007/MON. 08
We dene o
P
by
() :=
_
(), if 0 < t +u,
(), if t +u.
Then
W(t, ) W(s, ) W(t, ) W

(s, )
[W(t, ) W

(t, )] + [W

(t, ) W

(t, )] + [W

(t, ) W

(s, )].
(5.8)
By the Mean-value theorem, Theorem 3.6, and inequality (5.7),
W

(t, ) W

(t, ) (5.9)
=
_
t+u
t
[g(|

(, t, )|) g(|

(, t, )|)]d
+
_
+
t+u
[g(|

(, t +u,

(t +u, t, ))|) g(|

(, t +u,

(t +u, t, ))|)]d
ug((||, 0)) +
_
+
t+u
g

((R, t u))
|

(, t +u,

(t +u, t, ))

(, t +u,

(t +u, t, ))|d
ug((||, 0))
+
_
+
t+u
g

((R, t u))e
L(tu)
|

(t +u, t, )

(t +u, t, )|d
ug((||, 0)) +
_
+
t+u
g

((R, t u))e
L(tu)
2RL
e
Lu
1
L
d
= ug((||, 0)) + 2R(e
u
1)
_
+
0
g

((R, ))e
L
d.
<

2
.
Because and coincide on [s, t], we get by (5.5), that
W

(t, ) W

(s, ) C|

(t, s, )|
_
t
s
g(|

(, s, )|)d
= C|

(t, s, )|
_
t
s
g(|

(, s, )|)d.
(5.10)
Hence, by (5.8), (5.6), (5.9), and (5.10) we conclude that
W(t, ) W(s, ) < +C|

(t, s, )|
_
t
s
g(|

(, s, )|)d
and because > 0 was arbitrary we have proved inequality (5.3).
Inequality (5.4) is a trivial consequence of inequality (5.3), just set s = t and
note that and can be reversed.
Finally, we come to the central theorem of this section. It is the promised
converse Lyapunov theorem for a uniformly asymptotically stable equilibrium of
the Switched System 3.5.
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 29
Theorem 5.4 (Smooth converse theorem for switched systems). Assume that the
origin is a uniformly asymptotically stable equilibrium point of the Switched System
3.5 on the ball B
,R
|, R > 0, where | | is a norm on R
n
. Assume further,
that the functions f
p
, p T, satisfy the common Lipschitz condition: there exists a
constant L > 0 such that
|f
p
(s, x) f
p
(t, y)| L([s t[ +|x y|) (5.11)
for all s, t 0, all x, y B
,R
, and all p T.
Then, for every 0 < R

< R, there exists a Lyapunov function V : R


0

B
,R
R for the switched system, that is innitely dierentiable at every point
(t, x) R
0
B
,R
, x ,= 0.
Further, if the Switched System 3.5 is autonomous, then there exists a time-
invariant Lyapunov function V : B
,R
R for the system, that is innitely
dierentiable at every point x B
,R
, x ,= 0.
Proof. The proof is long and technical, even after all the preparation we have
done, so we split it into two parts. In part I we introduce some constants and
functions that we will use in the rest of the proof and in part II we dene a function
V (

(R
0

_
B
,R
0

) and prove that it is a Lyapunov function for the


system.
Part I: Because the assumptions of the theorem imply the assumptions made in
Denition 5.1, we can dene the functions W

: R
0
B
,R
R
0
, o
P
, and
W : R
0
B
,R
R
0
just as in the denition. As in Denition 5.1, denote by
g be the function from Masseras lemma 3.18 in the denition of the functions W

,
and set
C :=
_
+
0
g

((R, ))e
L
d,
where, once again, is the same function as in Denition 5.1.
Let m, M > 0 be constants such that
|x|
2
m|x| and |x| M|x|
2
for all x R
n
and let a be a constant such that
a > 2m and set y

:=
mR
a
.
Dene
K :=
g(y

)
a
L
_
C
_
m(1 +M)R +mR
_
4
3
LR +M
_
+g(4R/3)mR
_
,
and set
:= min
_
a
3g(y

)
,
a(R R

)
R

g(y

)
,
a
2mRLg(y

)
,
1
K
_
. (5.12)
Note that is a real-valued constant that is strictly larger than zero. We dene the
function : R
0
R
0
by
(x) :=
_ x
a
0
g(z)dz. (5.13)
The denition of implies
(x) g(x/a)
x
a

a
3g(y

)
g(x/a)
x
a

x
3
(5.14)
30 S. F. HAFSTEIN EJDE-2007/MON. 08
for all 0 x mR and

(x) =

a
g(x/a) (5.15)
for all x 0.
Dene the function by (x) := g(2x/3) g(x/2) for all x 0. Then (0) = 0
and for every x > 0 we have

(x) =
2
3
g

(2x/3)
1
2
g

(x/2) > 0
because g

/, that is /.
Part II: Let (

(R) be a nonnegative function with supp() ] 1, 0[ and


_
R
(x) = 1 and let (

(R
n
) be a nonnegative function with supp() B
2,1
and
_
R
n
(x)d
n
x = 1. Extend W on R R
n
by setting it equal to zero outside of
R
0
B
,R
. We claim that the function V : R
0
B
,R
R
0
, V (t, 0) := 0
for all t 0, and
V (t, ) :=
_
R
_
R
n

_
t
(||
2
)
_

_
y
(||
2
)
_
W[, y]

n+1
(||
2
)
d
n
yd
=
_
R
_
R
n
()(y)W[t (||
2
), (||
2
)y]d
n
yd
for all t 0 and all B
,R
0, is a (

(R
0

_
B
,R
0

) Lyapunov
function for the switched system. Note, that if the Switched System 3.5 in question
is autonomous, then W is time-invariant, which implies that V is time-invariant
too.
Because, for every |y|
2
1 and every || < R

, we have by (5.14) and (5.12),


that
| (||
2
)y|
_
1 +
(||
2
)
||
2
_
||

_
1 +
g(||
2
/a)
||
2

||
2
a
_
||
<
_
1 +
a(R R

)g(y

)
R

g(y

)a
_
R

= R,
so V is properly dened on R
0
B
,R
. But then, by construction, V (

(R
0

_
B
,R
0

). It remains to be shown that V fullls the conditions (L1) and (L2)


in Denition 4.3 of a Lyapunov function.
By Theorem 5.2 and Lemma 5.3 there is a function
1
/ and a constant
L
W
> 0, such that

1
(||) W(t, ) L
W
||
for all t 0 and all B
,R
. By inequality (5.14) we have for all B
,R
and
all |y|
2
1, that
| (||
2
)y| |
||
2
3

||
2
| =
2
3
||, (5.16)
| (||
2
)y| | +
||
2
3

||
2
| =
4
3
||. (5.17)
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 31
Hence

1
(2||/3) =
_
R
_
R
n
()(y)
1
(2||/3)d
n
yd

_
R
_
R
n
()(y)
1
(| (||
2
)y|)d
n
yd

_
R
_
R
n
()(y)W[t (||
2
), (||
2
)y]d
n
yd
= V (t, )

_
R
_
R
n
()(y)L
W
| (||
2
)y|d
n
yd

4L
W
3
||,
(5.18)
and the function V fullls the condition (L1).
We now prove that V fullls the condition (L2). To do this let t 0, B
,R
,
and o
P
be arbitrary, but xed throughout the rest of the proof. Denote by 1
the maximum interval in R
0
on which s

(s, t, ) is dened and set


q(s, ) := s (|

(s, t, )|
2
)
for all s 1 and all 1 0 and dene
D(h, y, ) := W[q(t +h, ),

(t +h, t, ) (|

(t +h, t, )|
2
)y]
W[q(t, ), (||
2
)y]
for all h such that t +h 1, all |y|
2
1, and all 1 0. Then
V (t +h,

(t +h, t, )) V (t, ) =
_
R
_
R
n
()(y)D(h, y, )d
n
yd
for all h such that t +h 1, especially this equality holds for all h in an interval of
the form [0, h

[, where 0 < h

+.
We are going to show that
limsup
h0+
D(h, y, )
h
(||). (5.19)
If we can prove that (5.19) holds, then, by Fatous lemma,
limsup
h0+
V (t +h,

(t +h, t, )) V (t, )
h

_
R
_
R
n
()
n
(y) limsup
h0+
D(h, y, )
h
d
n
yd
(||),
and we would have proved that the condition (L2) is fullled.
To prove inequality (5.19) observe that q(t, ) 0 for all 1 0 and, for
every s > t that is smaller than any switching-time (discontinuity-point) of larger
than t, and because of (5.12) and (5.11), we have
dq
ds
(s, ) = 1
g(|

(s, t, )|
2
/a)
a

(s, t, )
|

(s, t, )|
2
f
(s)
(s,

(s, t, ))
1
g(y

)LmR
a
32 S. F. HAFSTEIN EJDE-2007/MON. 08

1
2
,
so q(t +h, ) q(t, ) 0 for all small enough h 0.
Now, denote by the constant switching signal (t) in o
P
, that is (s) := (t)
for all s 0, and consider that by Lemma 5.3
D(h, y, )
h

C
h
_
_
_

(t +h, t, ) (|

(t +h, t, )|
2
)y

(q(t +h, ), q(t, ), (||


2
)y)
_
_
_

1
h
_
q(t+h,)
q(t,)
g(|

(s, q(t, ), (||


2
)y)|)ds
= C
_
_
_

(t +h, t, )
h

(|

(t +h, t, )|
2
) (||
2
)
h
y

(q(t +h, ), q(t, ), (||


2
)y) [ (||
2
)y]
h
_
_
_

1
h
_
q(t+h,)
q(t,)
g(|

(s, q(t, ), (||


2
)y)|)ds.
For the next calculations we need s q(s, ) to be dierentiable at t. If it is not,
which might be the case if t is a switching time of , we replace with

o
P
where

(s) :=
_
(t), if 0 s t,
(s), if s t.
Note that this does not aect the numerical value
limsup
h0+
D(h, y, )
h
because

(t + h) = (t + h) for all h 0. Hence, with p := (t), and by (5.11),


the chain rule, (5.16), and (5.17),
limsup
h0+
D(h, y, )
h
C
_
_
_f
p
(t, ) f
p
(q(t, ), (||
2
)y)
dq
dt

(t

, )

=t

(||
2
)
d
dt

(t

, t, )|
2

=t
y
_
_
_
g(|

(q(t, ), q(t, ), (||


2
)y)|)
dq
dt

(t

, )

=t
= C
_
_
_f
p
(t, ) f
p
(q(t, ), (||
2
)y)
_
1

(||
2
)

||
2
f
p
(t, )
_

(||
2
)[

||
2
f
p
(t, )]y
_
_
_
g(| (||
2
)y|)
_
1

(||
2
)

||
2
f
p
(t, )
_
C
_
_
f
p
(t, ) f
p
(q(t, ), (||
2
)y)
_
_
+C

(||
2
)|f
p
(t, )|
2
_
|f
p
(q(t, ), (||
2
)y)| +|y|

EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 33


g(2||/3) +g(4||/3)

(||
2
)|f
p
(t, )|
2
CL
_
[t q(t, )[ +(||
2
)|y|)
_
]
+C

(||
2
)mLR
_
L| (||
2
)y)| +M|y|
2

g(2||/3) +g(4||/3)

(||
2
)mLR
C
_
L(1 +M)(||
2
) +

(||
2
)mLR
_
L
4
3
|| +M
_
g(2||/3) +g(4||/3)

(||
2
)mLR.
Therefore, by (5.14), (5.15), and (5.12), and with x := ||, we can further simplify,
limsup
h0+
D(h, y, )
h
g(2x/3) +

a
g(mx/a)L
_
C
_
m(1 +M)x +mR
_
4
3
Lx +M
_
+g(4x/3)mR
_
g(2x/3) +Kg(x/2)
(x),
and because t 0, B
,R
, and o
P
were arbitrary, we have proved that V
is a Lyapunov function for the system.
Now, we have proved the main theorem of this section, our much wanted converse
theorem for the arbitrary Switched System 3.5.
6. Construction of Lyapunov Functions
In this section we present a procedure to construct Lyapunov functions for the
Switched System 3.5. After a few preliminaries on piecewise ane functions we give
an algorithmic description of how to derive a linear programming problem from the
Switched System 3.5 (Denition 6.8), and we prove that if the linear programming
problem possesses a feasible solution, then it can be used to parameterize a Lya-
punov function for the system. Then, in Section 8 and after some preparation in
Section 7, we present an algorithm that systematically generates linear program-
ming problems for the Switched System 3.5 and we prove, that if the switched
system possesses a Lyapunov function at all, then the algorithm generates, in a
nite number of steps, a linear programming problem that has a feasible solution.
Because there are algorithms that always nd a feasible solution to a linear pro-
gramming problem if one exists, this implies that we have derived an algorithm for
constructing Lyapunov functions, whenever one exists. Further, we consider the
case when the Switched System 3.5 is autonomous separately, because in this case
it is possible to parameterize a time-independent Lyapunov function for the system.
Let us be a little more specic on these points before we start to derive the results:
To construct a Lyapunov function with a linear programming problem, one needs
a class of continuous functions that are easily parameterized. That is, we need a
class of functions that is general enough to be used as a search-space for Lyapunov
functions, but it has to be a nite-dimensional vector space so that its functions
are uniquely characterized by a nite number of real numbers. The class of the
continuous piecewise ane functions CPWA is a well suited candidate.
The algorithm for parameterizing a Lyapunov function for the Switched System
3.5 consists roughly of the following steps:
34 S. F. HAFSTEIN EJDE-2007/MON. 08
(i) Partition a neighborhood of the equilibrium under consideration in a family
S of simplices.
(ii) Limit the search for a Lyapunov function V for the system to the class of
continuous functions that are ane on any S S.
(iii) State linear inequalities for the values of V at the vertices of the simplices
in S, so that if they can be fullled, then the function V , which is uniquely
determined by its values at the vertices, is a Lyapunov function for the
system in the whole area.
We rst partition R
n
into n-simplices and use this partition to dene the function
spaces CPWA of continuous piecewise ane functions R
n
R. A function in
CPWA is uniquely determined by its values at the vertices of the simplices in S.
Then we present a linear programming problem, algorithmically derived from the
Switched System 3.5, and prove that a CPWA Lyapunov function for the system can
be parameterized from any feasible solution to this linear programming problem.
Finally, in Section 7, we prove that if the equilibrium of the Switched System
3.5 is uniformly asymptotically stable, then any simplicial partition with small
enough simplices leads to a linear programming problem that does have a feasible
solution. Because, by Theorem 4.2 and Theorem 5.4, a Lyapunov function exists for
the Switched System 3.5 exactly when the equilibrium is uniformly asymptotically
stable, and because it is always possible to algorithmically nd a feasible solution
if at least one exists, this proves that the algorithm we present in Section 8 can
parameterize a Lyapunov function for the Switched System 3.5 if the system does
possess a Lyapunov functions at all.
6.1. Continuous piecewise ane functions. To construct a Lyapunov function
by linear programming, one needs a class of continuous functions that are easily
parameterized. Our approach is a simplicial partition of R
n
, on which we dene the
nite dimensional R-vector space CPWA of continuous functions, that are ane on
every of the simplices. We rst discuss an appropriate simplicial partition of R
n
and then dene the function space CPWA. The same is done in considerable more
detail in Chapter 4 in [40].
The simplices S

, where Perm[1, 2, . . . , n], will serve as the atoms of our


partition of R
n
. They are dened in the following way:
Denition 6.1 (The simplices S

). For every Perm[1, 2, . . . , n] we dene


the n-simplex
S

:= y R
n
: 0 y
(1)
y
(2)
y
(n)
1,
where y
(i)
is the (i)-th component of the vector y. An equivalent denition of
the n-simplex S

is
S

= con
_
n

j=1
e
(j)
,
n

j=2
e
(j)
, . . . ,
n

j=n+1
e
(j)
_
=
_
n+1

i=1

i
n+1

j=i
e
(j)
: 0
i
1 for i = 1, 2, . . . , n + 1 and
n+1

i=1

i
= 1
_
,
where e
(i)
is the (i)-th unit vector in R
n
.
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 35
For every Perm[1, 2, . . . , n] the set S

is an n-simplex with the volume


1/n! and, more importantly, if , Perm[1, 2, . . . , n], then
S

= conx R
n
: x is a vertex of S

and x is a vertex of S

. (6.1)
Thus, we can dene a continuous function p : [0, 1]
n
R that is ane on every S

,
Perm[1, 2, . . . , n], by just specifying it values at the vertices of the hypercube
[0, 1]
n
. That is, if x S

, then
x =
n+1

i=1

i
n+1

j=i
e
(j)
where 0
i
1 for i = 1, 2, . . . , n + 1 and

n+1
i=1

i
= 1, Then we set
p(x) = p
_
n+1

i=1

i
n+1

j=i
e
(j)
_
=
n+1

i=1

i
p
_
n+1

j=i
e
(j)
_
.
The function p is now well dened and continuous because of (6.1). We could now
proceed by partitioning R
n
into the simplices (z + S

)
zZ
n
,Perm[{1,2,...,n}]
, but
we prefer a simplicial partition of R
n
that is invariable with respect to reections
through the hyperplanes e
i
x = 0, i = 1, 2, . . . , n, as a domain for the function
space CPWA. We construct such a partition by rst partitioning R
n
0
into the
family (z +S

)
zZ
n
0
, Perm[{1,2,...,n}]
and then we extend this partition on R
n
by
use of the reection functions R
J
, where P(1, 2, . . . , n).
Denition 6.2 (Reection functions R
J
). For every P(1, 2, . . . , n), we
dene the reection function R
J
: R
n
R
n
,
R
J
(x) :=
n

i=1
(1)

J
(i)
x
i
e
i
for all x R
n
, where
J
: 1, 2, . . . , n 0, 1 is the characteristic function of
the set .
Clearly R
J
, where := j
1
, j
2
, . . . , j
k
, represents reections through the hy-
perplanes e
j1
x = 0, e
j2
x = 0, . . . , e
j
k
x = 0 in succession.
The simplicial partition of R
n
that we use for the denition of the function spaces
CPWA of continuous piecewise ane functions is
(R
J
(z +S

))
zZ
n
0
, JP({1,2,...,n}), Perm[{1,2,...,n}]
.
Similar to (6.1), this partition has the advantageous property, that from
S, S


_
R
J
(z +S

) : z Z
n
0
, P(1, 2, . . . , n), Perm[1, 2, . . . , n]
_
follows, that S S

is the convex hull of the vertices that are common to S and


S

. This leads to the following theorem:


Theorem 6.3. Let (q
z
)
zZ
n be a collection of real numbers. Then there is exactly
one continuous function p : R
n
R with the following properties:
(i) p(z) = q
z
for every z Z
n
.
(ii) For every P(1, 2, . . . , n), every Perm[1, 2, . . . , n], and every
z Z
n
0
, the restriction of the function p to the simplex R
J
(z + S

) is
ane.
Proof. See, for example, Corollary 4.12 in [40].
36 S. F. HAFSTEIN EJDE-2007/MON. 08
A CPWA space is a set of continuous ane functions from a subset of R
n
into
R with a given boundary conguration. If the subset is compact, then the bound-
ary conguration makes it possible to parameterize the functions in the respective
CPWA space with a nite number of real-valued parameters. Further, the CPWA
spaces are vector spaces over R in a canonical way. They are thus well suited
as a foundation, in the search of a Lyapunov function with a linear programming
problem.
We rst dene the function spaces CPWA for subsets of R
n
that are the unions
of n-dimensional cubes.
Denition 6.4 (CPWA function on a simple grid). Let Z Z
n
, Z , = , be such
that the interior of the set
^ :=
_
zZ
(z + [0, 1]
n
),
is connected. The function space CPWA[^] is then dened as follows.
A function p : ^ R is in CPWA[^], if and only if:
(i) p is continuous.
(ii) For every simplex R
J
(z +S

) ^, where z Z
n
0
, P(1, 2, . . . , n),
and Perm[1, 2, . . . ], the restriction p[
R
J
(z+S)
is ane.
We will need continuous piecewise ane functions, dened by their values on
grids with smaller grid steps than one, and we want to use grids with variable grid
steps. We achieve this by using images of Z
n
under mappings R
n
R
n
, of which
the components are continuous and strictly increasing functions R R, ane on
the intervals [m, m + 1] for all integers m, and map the origin on itself. We call
such R
n
R
n
mappings piecewise scaling functions.
Note that if y
i,j
, i = 1, 2, . . . , n and j Z, are real numbers such that y
i,j
< y
i,j+1
and y
i,0
= 0 for all i = 1, 2, . . . , n and all j Z, then we can dene a piecewise
scaling function PS : R
n
R
n
by

PS
i
(j) := y
i,j
for all i = 1, 2, . . . , n and all j Z.
Moreover, the piecewise scaling functions R
n
R
n
are exactly the functions, that
can be constructed in this way.
In the next denition we use piecewise scaling functions to dene general CPWA
spaces.
Denition 6.5 (CPWA function, general). Let PS : R
n
R
n
be a piecewise
scaling function and let Z Z
n
, Z , = , be such that the interior of the set
^ :=
_
zZ
(z + [0, 1]
n
)
is connected. The function space CPWA[PS, ^] is dened as
CPWA[PS, ^] := p PS
1
: p CPWA[^]
and we denote by S[PS, ^] the set of the simplices in the family
(PS(R
J
(z +S

)))
zZ
n
0
, JP({1,2,...,n}), Perm[{1,2,...,n}]
that are contained in the image PS(^) of ^ under PS.
Clearly
x R
n
: x is a vertex of a simplex in S[PS, ^] = PS(^ Z
n
)
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 37
and every function in CPWA[PS, ^] is continuous and is uniquely determined by
its values on the grid PS(^ Z
n
).
We use functions from CPWA[PS, ^] to approximate functions in (
2
(PS(^)),
that have bounded second-order derivatives. The next lemma gives an upper bound
of the approximation error of such a linearization.
Lemma 6.6. Let Perm[1, 2, . . . , n], let P(1, 2, . . . , n), let z Z
n
0
,
let R
J
be a reection function, and let PS be a piecewise scaling function. Denote
by S the n-simplex that is the convex combination of the vertices
y
i
:= PS(R
J
(z +
n+1

j=i
e
(j)
)), i = 1, 2, . . . , n + 1,
and let f (
2
(|) be a function dened on a domain S | R
n
. For every
i = 1, 2, . . . , n + 1 and every k = 1, 2, . . . , n dene the constant
A
k,i
:= [e
k
(y
i
y
n+1
)[
and for every r, s = 1, 2, . . . , n let B
rs
be a constant, such that
B
rs
max
xS


2
f
x
r
x
s
(x)

.
Dene for every i = 1, 2, . . . , n + 1 the constant
E
i
:=
1
2
n

r,s=1
B
rs
A
r,i
(A
s,1
+A
s,i
).
Then for every convex combination
y :=
n+1

i=1

i
y
i
, (6.2)
of the vertices of the simplex S we have

f(y)
n+1

i=1

i
f(y
i
)


n+1

i=1

i
E
i
.
Proof. Let y be as in equation (6.2). Then, by Taylors theorem, there is a vector
y
x
on the line-segment between y
n+1
and y, such that
f(y) = f(y
n+1
) +f(y
n+1
) (y y
n+1
)
+
1
2
n

r,s=1
[e
r
(y y
n+1
)][e
s
(y y
n+1
)]

2
f
x
r
x
s
(y
x
)
=
n+1

i=1

i
_
f(y
n+1
) +f(y
n+1
) (y
i
y
n+1
)
+
1
2
n

r,s=1
[e
r
(y
i
y
n+1
)][e
s
(y y
n+1
)]

2
f
x
r
x
s
(y
x
)
_
and for every i = 1, 2, . . . , n there is a vector y
i,x
on the line-segment between y
i
and y
n+1
such that
f(y
i
) = f(y
n+1
) +f(y
n+1
) (y
i
y
n+1
)
38 S. F. HAFSTEIN EJDE-2007/MON. 08
+
1
2
n

r,s=1
[e
r
(y
i
y
n+1
)][e
s
(y
i
y
n+1
)]

2
f
x
r
x
s
(y
i,x
) .
Further, because a simplex is a convex set, the vectors y
x
and y
1,x
, y
2,x
, . . . , y
n,x
are all in S. But then

f(y)
n+1

i=1

i
f(y
i
)

1
2
n+1

i=1

i
n

r,s=1
[e
r
(y
i
y
n+1
)[ ([e
s
(y y
n+1
)[ +[e
s
(y
i
y
n+1
)[) B
rs
=
1
2
n+1

i=1

i
n

r,s=1
B
rs
A
r,i
([e
s
(y y
n+1
)[ +A
s,i
)
and because
[e
s
(y y
n+1
)[
n+1

i=1

i
[e
s
(y
i
y
n+1
)[ [e
s
(y
1
y
n+1
)[ = A
s,1
it follows that

f(y)
n+1

i=1

i
f(y
i
)


1
2
n+1

i=1

i
n

r,s=1
B
rs
A
r,i
(A
s,1
+A
s,i
) =
n+1

i=1

i
E
i
.

An ane function p, dened on a simplex S R


n
and with values in R, has
the algebraic form p(x) = w x + q, where w is a constant vector in R
n
and q is
constant in R. Another characterization of p is given by specifying its values at
the vertices as stated. The next lemma gives a formula for the components of the
vector w when the values of p at the vertices of S are known and S is a simplex in
S[PS, ^].
Lemma 6.7. Let PS : R
n
R
n
be a piecewise scaling function, let z Z
n
0
, let
P(1, 2, . . . , n), let Perm[1, 2, . . . , n], and let p(x) := w x + q be an
ane function dened on the n-simplex with the vertices
y
i
:= PS(R
J
(z +
n

j=i
e
(j)
)), i = 1, 2, . . . , n.
Then
w =
n

i=1
p(y
i
) p(y
i+1
)
e
(i)
(y
i
y
i+1
)
e
(i)
.
Proof. For any i 1, 2, . . . , n we have
p(y
i
) p(y
i+1
) = w (y
i
y
i+1
)
=
n

k=1
w
(k)
[e
(k)
(y
i
y
i+1
)]
= w
(i)
[e
(i)
(y
i
y
i+1
)]
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 39
because the components of the vectors y
i
and y
i+1
are all equal with except of the
(i)-th one. But then
w
(i)
=
p(y
i
) p(y
i+1
)
e
(i)
(y
i
y
i+1
)
and we have nished the proof.
Now, that we have dened the function spaces CPWA we are ready to state
our linear programming problem, of which every feasible solution parameterizes a
CPWA Lyapunov function for the Switched System 3.5 used in the derivation of
its linear constraints.
6.2. The linear programming problem. We come to the linear programming
problem, of which every feasible solution parameterizes a Lyapunov function for
the Switched System 3.5. The Lyapunov function is of class CPWA. We rst
dene the linear programming problem in Denition 6.8. In the denition the
linear constraints are grouped into four classes, (LC1), (LC2), (LC3), and (LC4),
for linear constraints 1, 2, 3, and 4 respectively. Then we show how the variables
of the linear programming problem that fulll these constraints can be used to
parameterize functions that meet the conditions (L1) and (L2) of Denition 4.3,
the denition of a Lyapunov function. Then we state and discuss the results in
Section 6.8. Finally, we consider a more simple linear programming, dened in
Denition 6.12, for autonomous systems and we show that it is equivalent to the
linear programming problem in Denition 6.8 with additional constraints that force
the parameterized CPWA Lyapunov function to be time-invariant.
The next denition plays a central role in this work. It is generalization of the
linear programming problems presented in [40], [39], [17], and [16] to serve the
nonautonomous Switched System 3.5.
Denition 6.8. (Linear programming problemLP (f
p
: p T, ^, PS, t, T, ||))
Consider the Switched System 3.5 where the set T has a nite number of elements.
Let T

and T

be constants such that 0 T

< T

and let PS : R
n
R
n
be a
piecewise scaling function and ^ | be such that the interior of the set
/ :=
_
zZ
n
, PS(z+[0,1]
n
)N
PS(z + [0, 1]
n
)
is a connected set that contains the origin. Let
T := PS( ]d

1
, d
+
1
[ ]d

2
, d
+
2
[ . . . ]d

n
, d
+
n
[ )
be a set, of which the closure is contained in the interior of /, and either T =
or d

i
and d
+
i
are integers such that d

i
1 and 1 d
+
i
for every i = 1, 2, . . . , n.
Finally, let | | be an arbitrary norm on R
n
and t := (t
0
, t
1
, . . . , t
M
) R
M+1
,
M N
>0
, be a vector such that T

=: t
0
< t
1
< < t
M
:= T

.
We assume that the components of the f
p
, p T, have bounded
second-order partial derivatives on [T

, T

] (/ T).
Before we go on, it is very practical to introduce an alternate notation for the
vectors (t, x) RR
n
, because it considerably shortens the formulae in the linear
programming problem. We identify the time t with the zeroth component x
0
of the
vector
x := ( x
0
, x
1
, . . . , x
n
)
40 S. F. HAFSTEIN EJDE-2007/MON. 08
and x with the components 1 to n, that is t := x
0
and x
i
:= x
i
for all i = 1, 2, . . . , n.
Then, the systems
x = f
p
(t, x), p T,
can be written in the equivalent form
d
d x
0
x =

f
p
( x), p T,
where

f
p
( x) :=
_

f
p,0
( x),

f
p,1
( x),

f
p,2
( x), . . . ,

f
p,n
( x)

:=
_
1, f
p,1
(t, x), f
p,2
(t, x), . . . , f
p,n
(t, x)

,
(Recall that f
p,i
denotes the i-th component of the function f
p
.) that is,

f
p,0
:= 1
and

f
p,i
( x) := f
p,i
(t, x), where x = (t, x), for all p T and all i = 1, 2, . . . , n.
Further, let PS
0
: R R be a piecewise scaling function such that PS
0
(i) := t
i
for all i = 0, 1, . . . , M and dene the piecewise scaling function

PS : R R
n
R R
n
through

PS( x) :=
_
PS
0
( x
0
), PS
1
( x
1
), . . . , PS
n
( x
n
))

,
that is,

PS( x) =
_
PS
0
(t), PS(x)

,
where x = (t, x).
We will use the standard orthonormal basis in R
n+1
= R R
n
, but start the
indexing at zero (use e
0
, e
1
, . . . , e
n
), that is,
x :=
n

i=0
x
i
e
i
= te
0
+
n

i=1
x
i
e
i
.
Because we do not have to consider negative time-values t = x
0
< 0, it is more
convenient to use reection functions that do always leave the zeroth-component of
x = (t, x) unchanged. Therefore, we dene for every reection function R
J
: R
n

R
n
, where 1, 2, . . . , n, the function

R
J
: R R
n
R R
n
through

R
J
( x) :=
_
x
0
, R
J
(x)

:= te
0
+
n

i=1
(1)

J
(i)
x
i
e
i
.
We dene the seminorm | |

: R R
n
R
0
through
|( x
0
, x
1
, . . . , x
n
)|

:= |( x
1
, x
2
, . . . , x
n
)|.
Then, obviously, | x|

= |x| for all x = (t, x) R R


n
. The linear programming
problem LP(f
p
: p T, ^, PS, t, T, | |) is now constructed in the following way:
(i) Dene the sets
( := x R R
n
: x

PS(Z Z
n
)
_
[T

, T

] (/ T)
_

and
A

:= |x| : x PS(Z
n
) /.
The set ( is the grid, on which we will derive constraints on the values
of the CPWA Lyapunov function, and A

is the set of distances of all


relevant points in the state-space to the origin with respect to the norm
| |.
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 41
(ii) Dene for every Perm[0, 1, . . . , n] and every i = 0, 1, . . . , n + 1 the
vector
x

i
:=
n

j=i
e
(j)
,
where, of course, the empty sum is interpreted as 0 R R
n
.
(iii) Dene the set Z through: The tuple (z, ), where z := (z
0
, z
1
, . . . , z
n
)
Z
0
Z
n
0
and P(1, 2, . . . , n, is an element of Z, if and only if

PS(

R
J
(z + [0, 1]
n+1
)) [T

, T

]
_
/ T
_
.
Note that this denition implies that
_
(z,J)Z

PS(

R
J
(z + [0, 1]
n+1
)) = [T

, T

]
_
/ T
_
.
iv) For every (z, ) Z, every Perm[0, 1, . . . , n], and every
i = 0, 1, . . . , n + 1 we set
y
(z,J)
,i
:=

PS(

R
J
(z +x

i
)).
The vectors y
(z,J)
,i
are the vertices of the simplices in our simplicial partition
of the set [T

, T

]
_
/ T
_
. The position of the simplex is given by
(z, ), where z
0
species the position in time and (z
1
, z
2
, . . . , z
n
) species
the position in the state-space. Further, species the simplex and i
species the vertex of the simplex.
v) Dene the set
} :=
_
y
(z,J)
,k
, y
(z,J)
,k+1

Perm[0, 1, . . . , n], (z, ) Z,


and k 0, 1, . . . , n
_
.
The set } is the set of all pairs of neighboring grid points in the grid (.
(vi) For every p T, every (z, ) Z, and every r, s = 0, 1, . . . , n let B
(z,J)
p,rs
be a real-valued constant, such that
B
(z,J)
p,rs
max
i=1,2,...,n
sup
x
f
PS(
e
R
J
(z+[0,1]
n+1
))


2
f
p,i
x
r
x
s
( x)

.
The constants B
(z,J)
p,rs
are local bounds on the second-order partial deriva-
tives of the components of the functions

f
p
, p T, with regard to the in-
nity norm | |

, similar to the constants B


rs
in Lemma 6.6. Note, that
because

f
p,0
:= 1, the zeroth-components can be left out in the denition
of the B
(z,J)
p,rs
because they are identically zero anyways. Further, for every
r, s = 0, 1, . . . , n and every x = (t, x),

2
f
p,i
x
r
x
s
( x) =

2
f
p,i
x
r
x
s
(t, x)
if we read x
0
as t on the right-hand side of the equation. Finally, note
that if B is a constant such that
B sup
x[T

,T

](M\D)
|

2
f
p
x
r
x
s
( x)|

,
42 S. F. HAFSTEIN EJDE-2007/MON. 08
for all p T and all r, s = 0, 1, . . . , n, then, of course, we can set B
(z,J)
p,rs
= B
for all p T, all (z, ) Z, and all r, s = 0, 1, . . . , n. Tighter bounds,
however, might save a lot of computational eorts in a search for a feasible
solution to the linear programming problem.
(vii) For every (z, ) Z, every i, k = 0, 1, . . . , n, and every
Perm[0, 1, . . . , n], dene
A
(z,J)
,k,i
:=

e
k
(y
(z,J)
,i
y
(z,J)
,n+1
)

.
The A
(z,J)
,k,i
are constants similar to the constants A
k,i
in Lemma 6.6.
(viii) Dene the constant
x
min,M
:= min|x| : x PS(Z
n
) /,
where / is the boundary of the set /.
(ix) For every p T, every (z, ) Z, every Perm[0, 1, . . . , n], and every
i = 0, 1, . . . , n + 1 set
E
(z,J)
p,,i
:=
1
2
n

r,s=0
B
(z,J)
p,rs
A
(z,J)
,r,i
(A
(z,J)
,s,i
+A
(z,J)
,s,0
). (6.3)
(ix) Let > 0 and > 0 be arbitrary constants.
The variables of the linear programming problem are:
,
[y], for all y A

,
[y], for all y A

,
V [ x], for all x (,
C[ x, y], for all x, y }.
Considering Denition 4.3, the denition of a Lyapunov function, the variables
[y] correspond to the function
1
, the variables [y] to the function , and the
variables V [ x] to the Lyapunov function V , the x
0
component representing the
time t. The variables C[ x, y] are local bounds on the gradient
x
V
Lya
of the
Lyapunov function V
Lya
to be constructed and is a corresponding global bound.
The linear constraints of the linear programming problem are:
(LC1) Let y
0
, y
1
, . . . , y
K
be the elements of A

in an increasing order. Then


[y
0
] = [y
0
] = 0,
y
1
[y
1
],
y
1
[y
1
],
and for every i = 1, 2, . . . , K 1:
[y
i
] [y
i1
]
y
i
y
i1

[y
i+1
] [y
i
]
y
i+1
y
i
,
[y
i
] [y
i1
]
y
i
y
i1

[y
i+1
] [y
i
]
y
i+1
y
i
.
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 43
(LC2) For every x (:
[| x|

] V [ x].
If T = , then, whenever | x|

= 0:
V [ x] = 0.
If T , = , then, whenever ( x
1
, x
2
, . . . , x
2
) PS(Z
n
) T:
V [ x] [x
min,M
] .
Further, if T , = , then for every i = 1, 2, . . . , n and every j = 0, 1, . . . , M:
V [PS
0
(j)e
0
+ PS
i
(d

i
)e
i
] PS
i
(d

i
),
V [PS
0
(j)e
0
+ PS
i
(d
+
i
)e
i
] PS
i
(d
+
i
).
(LC3) For every x, y }:
C[ x, y] | x y|

V [ x] V [ y] C[ x, y] | x y|

| x y|

.
(LC4) For every p T, every (z, ) Z, every Perm[0, 1, . . . , n], and every
i = 0, 1, . . . , n + 1:

_
|y
(z,J)
,i
|

j=0
_
V [y
(z,J)
,j
] V [y
(z,J)
,j+1
]
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)

f
p,(j)
(y
(z,J)
,i
) +E
(z,J)
p,,i
C[y
(z,J)
,j
, y
(z,J)
,j+1
]
_
.
As the objective of the linear programming problem is not needed to parameterize
a CPWA Lyapunov function we do not dene it here.
Note that the values of the constants > 0 and > 0 do not aect whether there
is a feasible solution to the linear program or not. If there is a feasible solution for
:=

> 0 and :=

> 0, then there is a feasible solution for all :=

> 0
and :=

> 0. Just multiply the numerical values of all variables of the feasible
solution with
max

.
Further note that if |y
(z,J)
,i
|

= 0, then

f
p,(j)
(y
(z,J)
,i
) = 0 for all j 0, 1, . . . , n
such that (j) ,= 0 and if (j) = 0, then V [y
(z,J)
,j
] V [y
(z,J)
,j+1
] = 0. Thus, the con-
straints (LC4) reduce to
0
n

j=0
E
(z,J)
p,,i
C[y
(z,J)
,j
, y
(z,J)
,j+1
], (6.4)
which looks contradictory at rst glance. However, if |y
(z,J)
,i
|

= 0 then necessarily
i = n + 1 and
E
(z,J)
p,,n+1
:=
1
2
n

r,s=0
B
(z,J)
p,rs
A
(z,J)
,r,n+1
(A
(z,J)
,s,n+1
+A
(z,J)
,s,0
) = 0
because A
(z,J)
,r,n+1
= 0 for all r = 0, 1, . . . , n, so (6.4) is not contradictory and even
trivially fullled.
Finally, if the Switched System 3.5 is autonomous, then we know by Theorem
5.4 that there exists a time-invariant Lyapunov function for the system. To reect
this fact one is tempted to additionally include the constraints V [ x] = V [ y] for
44 S. F. HAFSTEIN EJDE-2007/MON. 08
every pair x, y ( such that | x y|

= 0 in the linear programming problem to


limit the search to time-invariant Lyapunov functions. However, as we will show in
Theorem 6.14, this is equivalent to a more simple linear programming problem if
the Switched System 3.5 is autonomous, namely, the linear programming problem
dened in Denition 6.12.
In the next sections we prove that a feasible solution to the linear programming
problem dened in Denition 6.8 parameterizes a CPWA Lyapunov function for
the Switched System 3.5 used for its construction. For this proof the variable is
not needed. However, it will be needed for the analysis in Section 6.8.
6.3. Denition of the functions , , and V
Lya
. Let y
0
, y
1
, . . . , y
K
be the el-
ements of A

in an increasing order. We dene the piecewise ane functions


, : R
0
R,
(y) := [y
i
] +
[y
i+1
] [y
i
]
y
i+1
y
i
(y y
i
),
(y) := [y
i
] +
[y
i+1
] [y
i
]
y
i+1
y
i
(y y
i
),
for all y [y
i
, y
i+1
] and all i = 0, 1, . . . , K 1. The values of and on ]y
K
, +[
do not really matter, but to have everything properly dened, we set
(y) := [y
K1
] +
[y
K
] [y
K1
]
y
K
y
K1
(y y
K1
),
(y) := [y
K1
] +
[y
K
] [y
K1
]
y
K
y
K1
(y y
K1
)
for all y > y
K
. Clearly the functions and are continuous.
The function V
Lya
CPWA[

PS,

PS
1_
[T

, T

]
_
/ T
__
] is dened by as-
signing
V
Lya
( x) := V [ x]
for all x (. We will sometimes write V
Lya
(t, x) for V
Lya
( x) and V [t, x] for V [ x].
It is then to be understood that t := x
0
and x := ( x
1
, x
2
, . . . , x
n
).
In the next four sections we will successively derive the implications the linear
constraints (LC1), (LC2), (LC3), and (LC4) have on the functions , , and V
Lya
.
6.4. Implications of the constraints (LC1). Let y
0
, y
1
, . . . , y
K
be the elements
of A

in an increasing order. We are going to show that the constraints (LC1)


imply, that the functions and are convex and strictly increasing on [0, +[ .
Because y
0
= 0, (y
0
) = [y
0
] = 0, and (y
0
) = [y
0
] = 0, this means that they
are convex / functions. The constraints are the same for and , so it suces to
show this for the function .
From the denition of , it is clear that it is continuous and that
(x) (y)
x y
=
[y
i+1
] [y
i
]
y
i+1
y
i
(6.5)
for all x, y [y
i
, y
i+1
] and all i = 0, 1, . . . , K 1. From y
0
= 0, [y
0
] = 0, and
y
1
[y
1
] we get

[y
1
] [y
0
]
y
1
y
0

[y
2
] [y
1
]
y
2
y
1

[y
K
] [y
K1
]
y
K
y
K1
.
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 45
But then D
+
is a positive and increasing function on R
0
and it follows from
Corollary 3.10, that is a strictly increasing function.
The function is convex, if and only if for every y R
>0
there are constants
a
y
, b
y
R, such that
a
y
y +b
y
= (y) and a
y
x +b
y
(x)
for all x R
0
(see, for example, Section 17 in Chapter 11 in [62]). Let y R
>0
.
Because the function D
+
is increasing, it follows by Theorem 3.9, that for every
x R
0
, there is a c
x,y
R, such that
(x) = (y) +c
x,y
(x y)
and c
x,y
D
+
(y) if x < y and c
x,y
D
+
(y) if x > y. This means that
(x) = (y) +c
x,y
(x y) D
+
(y)x +(y) D
+
(y)y
for all x R
0
. Because y was arbitrary, the function is convex.
6.5. Implications of the constraints (LC2). Dene the constant
V
Lya
M,min
:= min
xM
t[T

,T

]
V
Lya
(t, x)
and if T , = the constant
V
Lya
D,max
:= max
xD
t[T

,T

]
V
Lya
(t, x).
We are going to show that the constraints (LC2) imply, that
(|x|) V
Lya
(t, x) (6.6)
for all t [T

, T

] and all x / T and that


V
Lya
D,max
V
Lya
M,min

if T , = .
We rst show that they imply, that
(| x|

) V
Lya
( x)
for all x (, which obviously implies (6.6). Let x (. Then there is a (z, ) Z,
a Perm[0, 1, . . . , n], and constants
0
,
1
, . . . ,
n+1
[0, 1], such that
x =
n+1

i=0

i
y
(z,J)
,i
and
n+1

i=0

i
= 1.
Then
(| x|

) = (|
n+1

i=0

i
y
(z,J)
,i
|

) (
n+1

i=0

i
|y
(z,J)
,i
|

n+1

i=0

i
(|y
(z,J)
,i
|

) =
n+1

i=0

i
[|y
(z,J)
,i
|

]
n+1

i=0

i
V [y
(z,J)
,i
]
=
n+1

i=0

i
V
Lya
(y
(z,J)
,i
) = V
Lya
(
n+1

i=0

i
y
(z,J)
,i
) = V
Lya
( x).
46 S. F. HAFSTEIN EJDE-2007/MON. 08
Now consider the case T , = . From the denition of V
Lya
and the constants
V
Lya
D,max
and V
Lya
M,min
it is clear, that
V
Lya
D,max
= max
xDPS(Z
n
)
u=0,1,...,M
V [t
u
, x],
V
Lya
M,min
= min
xMPS(Z
n
)
u=0,1,...,M
V [t
u
, x].
Let x /PS(Z
n
) and u 0, 1, . . . , M be such that V [t
u
, x] = V
Lya
M,min
, then
V
Lya
D,max
[x
min,M
] = (x
min,M
)
(|x|) V [t
u
, x]
= V
Lya
M,min
.
6.6. Implications of the constraints (LC3). The constraints (LC3) imply that

V [ x] V [ y]
| x y|

C[ x, y]
for every x, y } and these local bounds C[ x, y] on the gradient
x
V
Lya
will
be used in the next section.
6.7. Implications of the constraints (LC4). We are going to show that the
constraints (LC4) and (LC3) together imply that
(|

(t, t

, )|) limsup
h0+
V
Lya
(t +h,

(t +h, t

, )) V
Lya
(t,

(t, t

, ))
h
(6.7)
for all o
P
and all (t,

(t, t

, )) in the interior of [T

, T

] (/ T).
Let o
P
and x := (t,

(t, t

, )) in the interior of [T

, T

] (/ T) be
arbitrary, but xed throughout this section, and set x :=

(t, t

, ) and p := (t).
We claim that there is a (z, ) Z, a Perm[0, 1, . . . , n], and constants

0
,
1
, . . . ,
n+1
[0, 1], such that
x =
n+1

i=0

i
y
(z,J)
,i
,
n+1

i=0

i
= 1,
x +h

f
p
( x) cony
(z,J)
,0
, y
(z,J)
,1
, . . . , y
(z,J)
,n+1

(6.8)
for all h [0, a], where a > 0 is some constant.
We prove this claim by a contradiction. Assume that it does not hold true.
The vector x is contained in some of the simplices in the simplicial partition of
[T

, T

]
_
/ T
_
, say S
1
, S
2
, . . . , S
k
. Simplices are convex sets so we necessarily
have
_
x +
1
j

f
p
( x) : j N
>0
_
S
i
=
for every i = 1, 2, . . . , k. But then there must be a simplex S in the simplicial
partition, dierent to the simplices S
1
, S
2
, . . . , S
k
, such that the intersection
_
x +
1
j

f
p
( x) : j N
>0
_
S
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 47
contains an innite number of elements. This implies that there is a sequence in S
that converges to x, which is a contradiction, because S is a closed set and x / S.
Therefore (6.8) holds.
Because is a convex function, we have
(| x|

)
n+1

i=0

_
|y
(z,J)
,i
|

(6.9)
as was shown in Section 6.5. From the denition of V
Lya
it follows, that there is a
vector w R R
n
, such that
V
Lya
( y) = w ( y y
(z,J)
,n+1
) +V
Lya
(y
(z,J)
,n+1
) (6.10)
for all y cony
(z,J)
,0
, y
(z,J)
,1
, . . . , y
(z,J)
,n+1
.
It follows by Holders inequality, that
w

f
p
( x) = w
n+1

i=0

f
p
(y
(z,J)
,i
) +w
_

f
p
( x)
n+1

i=0

f
p
(y
(z,J)
,i
)
_

n+1

i=0

i
w

f
p
(y
(z,J)
,i
) +|w|
1
|

f
p
( x)
n+1

i=0

f
p
(y
(z,J)
,i
)|

.
(6.11)
By Lemma 6.6 and the assignment in (6.3),
|

f
p
( x)
n+1

i=0

f
p
(y
(z,J)
,i
)|

= max
j=0,1,...,n


f
p,j
( x)
n+1

i=0

i

f
p,j
(y
(z,J)
,i
)

1
2
n+1

i=0

i
n

r,s=0
B
(z,J)
p,rs
A
(z,J)
,r,i
(A
(z,J)
,s,0
+A
(z,J)
,s,i
)

n+1

i=0

i
E
(z,J)
p,,i
,
which implies that we have derived the inequality
w

f
p
( x)
n+1

i=0

i
_
w

f
p
(y
(z,J)
,i
) +|w|
1
E
(z,J)
p,,i
_
. (6.12)
We come to the vector w. By Lemma 6.7, the constraints (LC3), and because

e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)

= |y
(z,J)
,j
y
(z,J)
,j+1
|

for all j = 0, 1, . . . , n, we obtain the inequality


|w|
1
=
n

j=0

V [y
(z,J)
,j
] V [y
(z,J)
,j+1
]
|y
(z,J)
,j
y
(z,J)
,j+1
|

j=0
C[y
(z,J)
,j
, y
(z,J)
,j+1
].
This inequality combined with (6.12) gives
w

f
p
( x)
n+1

i=0

i
n

j=0
_
V [y
(z,J)
,j
] V [y
(z,J)
,j+1
]
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)

f
p,(j)
(y
(z,J)
,i
)
+E
(z,J)
p,,i
C[y
(z,J)
,j
, y
(z,J)
,j+1
]
_
.
(6.13)
48 S. F. HAFSTEIN EJDE-2007/MON. 08
We are going to show that inequality (6.13) together with the constraints (LC4)
imply that inequality (6.7) holds. First, note that because V
Lya
satises a Lipschitz
condition with Lipschitz constant, say L
V
> 0 with respect to the norm | |, we
have
limsup
h0+

V
Lya
(t +h,

(t +h, t

, )) V
Lya
(t +h, x +hf
p
(t, x))
h

limsup
h0+
L
V
_
_

(t +h, t

, ) x
h
f
p
(t, x)
_
_
= L
V
|f
p
(t, x) f
p
(t, x)| = 0.
Hence, by Lemma 3.8 and the representation (6.10) of V
Lya
,
limsup
h0+
V
Lya
(t +h,

(t +h, t

, )) V
Lya
(t,

(t, t

, ))
h
= limsup
h0+
V
Lya
( x +h

f
p
( x)) V
Lya
( x)
h
= limsup
h0+
hw

f
p
( x)
h
= w

f
p
( x),
and we obtain by (6.13), (LC4), and (6.9) that
limsup
h0+
V
Lya
(t +h,

(t +h, t

, )) V
Lya
(t,

(t, t

, ))
h
= w

f
p
( x)

n+1

i=0

i
n

j=0
_
V [y
(z,J)
,j
] V [y
(z,J)
,j+1
]
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)

f
p,(j)
(y
(z,J)
,i
) +E
(z,J)
p,,i
C[y
(z,J)
,j
, y
(z,J)
,j+1
]
_

n+1

i=1

_
|y
(z,J)
,i
|

(|

(t, t

, )|).
Hence, inequality (6.7) holds for all o
P
and all (t,

(t, t

, )) in the interior of
[T

, T

] (/ T).
6.8. Summary of the results and their consequences. We start by summing
up the results we have proved after the denition of the linear programming problem
in a theorem.
Theorem 6.9 (CPWA Lyapunov functions by linear programming).
Consider the linear programming problem LP(f
p
: p T, ^, PS, t, T, | |) in
Denition 6.8 and assume that it possesses a feasible solution. Let the functions
, , and V
Lya
be dened as in Section 6.3 from the numerical values of the vari-
ables [y], [y], and V [ x] from a feasible solution. Then the inequality
(|x|) V
Lya
(t, x)
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 49
holds for all x /T and all t [T

, T

]. If T = we have (0) = V
Lya
(t, 0) = 0
for all t [T

, T

]. If T , = we have, with
V
Lya
M,min
:= min
xM
t[T

,T

]
V
Lya
(t, x),
V
Lya
D,max
:= max
xD
t[T

,T

]
V
Lya
(t, x),
that
V
Lya
D,max
V
Lya
M,min
.
Further, with as the solution to the Switched System 3.5 that we used in the
construction of the linear programming problem, the inequality
(|

(t, t

, )|) limsup
h0+
V
Lya
(t +h,

(t +h, t

, )) V
Lya
(t,

(t, t

, ))
h
(6.14)
hold true for all o
P
and all (t,

(t, t

, )) in the interior of [T

, T

] (/ T).
We now come to the important question:
Which information on the stability behavior of the Switched System
3.5 can we extract from the Lyapunov-like function V
Lya
dened
in Section 6.3?
Before we answer this question we discuss the implications secured by a continuously
dierentiable Lyapunov function on the stability behavior of a non-switched system
to get an idea what we can expect. To do this consider the system
x = f (t, x),
where f [(
1
(R
0
1)]
n
and 1 is a bounded domain in R
n
containing the origin,
and assume that there is a function W (
1
(R
0
1) and functions a, b, c /,
such that
a(||) W(t, ) b(||)
for all 1 and all t 0 and
d
dt
W(t, (t, t

, )) = [
x
W](t, (t, t

, )) f (t, (t, t

, )) +
W
t
(t, (t, t

, ))
c(|(t, t

, )|)
for all (t, (t, t

, )) R
0
1, where is the solution to the dierential equation
x = f (t, x).
For our analysis we let (t

, ) R
0
1 be arbitrary but constant and set y(t) :=
W(t, (t, t

, )). Then y(t

) = W(t

, ) and y satises the dierential inequality


y(t) c(b
1
(y(t)))
for all t such that (t, t

, ) 1. Now, assume that there are constants b

> 0 and
c

> 0, such that b(|x|) b

|x| and c

|| c(|x|) for all x 1. In this simple


case it is quite simple to derive the inequality
y(t) y(t

) exp
_

(t t

)
_
,
which is valid for all t t

if
W(t

, ) < inf
st

, yV
W(s, y).
50 S. F. HAFSTEIN EJDE-2007/MON. 08
We are going to show that a very similar analysis can be done for a switched
system and the corresponding Lyapunov-like function V
Lya
if the arbitrary norm
| | used in Denition 6.8 of the linear programming problem is a p-norm | |
p
,
1 p +, but rst we prove a technical lemma that will be used in the proof of
the theorem.
Lemma 6.10. Let [a, b[ be an interval in R, < a < b +, and let y, z :
[a, b[ R be functions such that y(a) z(a), y is continuous, and z is dierentiable.
Assume that there is a function s : R R that satises the local Lipschitz condition,
for every compact ( R there is a constant L
C
such that
[y() y()[ L
C
[ [, for all , (,
and assume further that
D
+
y(t) s(y(t)) and z(t) = s(z(t))
for all t [a, b[ . Then y(t) z(t) for all t [a, b[ .
Proof. Assume that the proposition of the lemma does not hold. Then there is a
t
0
[a, b[ such that y(t) z(t) for all t [a, t
0
] and an > 0 such that y(t) > z(t)
for all t ]t
0
, t
0
+ ]. Let L > 0 be a local Lipschitz constant for s on the interval
[y(t
0
), y(t
0
+)]. Then, by Lemma 3.8,
D
+
(y z)(t) = D
+
y(t) z(t) s(y(t)) +s(z(t)) L(y(t) z(t))
for every t [t
0
, t
0
+]. But then, with w(t) := y(t) z(t) for all t [a, b[ , we have
limsup
h0+
w(t +h)e
L(t+h)
w(t)e
Lt
h
e
Lt
limsup
h0+
w(t +h)(e
Lh
1)
h
+e
Lt
limsup
h0+
w(t +h) w(t)
h
= Le
Lt
w(t) +e
Lt
D
+
w(t)
Le
Lt
w(t) +Le
Lt
w(t) = 0,
for all t [t
0
, t
0
+ ], which implies, by Corollary 3.10, that the function t
e
Lt
w(t) is monotonically decreasing on the same interval. Because w(t
0
) = 0 this
is contradictory to y(t) > z(t) for all t ]t
0
, t
0
+] and therefore the proposition of
the lemma must hold true.
We come to the promised theorem, where the implications of the function V
Lya
on the stability behavior of the Switched System 3.5 are specied. Here with k-
norm we mean the norm |x|
k
:=
_
n
i=1
[x
i
[
k
_
1/k
if 1 k < + and |x|

:=
max
i=1,2,...,n
[x
i
[. Unfortunately, these norms are usually called p-norms, which is
inappropriate in this context because the alphabet p is used to index the functions
f
p
, p T.
Theorem 6.11 (Implications of the Lyapunov function V
Lya
).
Make the same assumptions and denitions as in Theorem 6.9 and assume ad-
ditionally that the norm | | in the linear programming problem LP(f
p
: p
T, ^, PS, t, T, | |) is a k-norm, 1 k +. Dene the set T through T := 0
if T = and
T := T
_
x / T : max
t[T

,T

]
V
Lya
(t, x) V
Lya
D,max
_
, if T , = ,
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 51
and the set / through
/ :=
_
x / T : max
t[T

,T

]
V
Lya
(t, x) < V
Lya
M,min
_
.
Set q := k (k 1)
1
if 1 < k < +, q := 1 if k = +, and q := + if k = 1,
and dene the constant
E
q
:= |
n

i=1
e
i
|
q
.
Then the following propositions hold true:
(i) If

(t, t

, ) T for some particular o


P
, T

t T

, t

0, and
|, then

(s, t

, ) T for all s [t, T

].
(ii) If

(t, t

, ) / T for some particular o


P
, T

t T

, t

0, and
|, then the inequality
V
Lya
(s,

(s, t

, )) V
Lya
(t,

(t, t

, )) exp
_


E
q
(s t)
_
(6.15)
holds for all s such that

(s

, t

, ) / T for all t s

s T

.
(iii) If

(t, t

, ) / for some particular o


P
, T

t T

, t

0,
and |, then the solution

either fullls inequality (6.15) for all


t s T

, or there is a T

]t, T

], such that the solution

fullls
inequality (6.15) for all t s T

(T

, t

, ) T, and

(s, t

, ) T
for all T

s T

.
Proof. Proposition (i) is trivial if T = . To prove proposition (i) when T , =
dene for every > 0 the set
T

:= x R
n
: |x y|
2
< for some y T .
Because V
Lya
D,max
V
Lya
M,min
by Theorem 6.9, it follows that T

/ for all
small enough > 0. For every such small > 0 notice, that inequality (6.14) and
Corollary 3.10 together imply, that if

(t, t

, ) T

for some particular o


P
,
T

t T

, t

0, and |, then

(s, t

, ) T

for all s [t, T

]. Then the
proposition (i) follows, because if

(t, t

, ) T , then

(t, t

, )

>0
T

, and
therefore

(s, t

, )

>0
T

= T for all s [t, T

].
To prove proposition (ii) rst note that the linear constraints (LC2) and (LC3)
imply that V
Lya
(t, x) |x|
1
for all t [T

, T

] and all x /. To see this just


notice that at least for one i 1, 2, . . . , n we must have either
x
i
PS
i
(d
+
i
) or x
i
PS
i
(d

i
)
because x / T. Then either
V
Lya
(t, x
i
e
i
) PS
i
(d
+
i
) + [x
1
PS
i
(d
+
i
)[ = [x
i
[
or
V
Lya
(t, x
i
e
i
) PS
i
(d

i
) + [x
1
PS
i
(d

i
)[ = [x
i
[,
so
V
Lya
(t, x
i
e
i
) [x
i
[,
which in turn implies, for any j 1, 2, . . . , n, j ,= i, that
V
Lya
(t, x
i
e
i
+x
j
e
j
) V
Lya
(t, x
i
e
i
) + [x
j
[ ([x
i
[ +[x
j
[)
52 S. F. HAFSTEIN EJDE-2007/MON. 08
and by mathematical induction V
Lya
(t, x) |x|
1
. Then, by Holders inequality,
V
Lya
(t, x) |x|
1
=
_
n

i=1
e
i
_

_
n

i=1
[x
i
[e
i
_
E
q
|x|
k
,
so by the linear constraints (LC1) and inequality (6.14), we have for every o
P
,
T

t T

, t

0, and |, such that

(t, t

, ) / T, that


E
q
V
Lya
(t,

(t, t

, ))
|

(t, t

, )|
k
(|

(t, t

, )|
k
)
limsup
h0+
V
Lya
(t +h,

(t +h, t

, )) V
Lya
(t,

(t, t

, ))
h
.
The dierential equation y(s) = E
q
/ y(s) has solution y(s) = exp[E
q
(s
t

)/] y(t

). Hence, by Lemma 6.10,


V (s,

(s, t

, )) V (t,

(t, t

, )) exp
_


E
q
(s t)
_
and proposition (ii) holds.
Proposition (iii) is a direct consequence of the propositions (i) and (ii) and the
denition of the set /. It merely states that if it is impossible for a solution to exit
the set / T at the boundary /, then it either exits at the boundary T or it
does not exit at all.
6.9. The autonomous case. As was discussed after Denition 6.8, one is tempted
to try to parameterize a time-invariant Lyapunov function for the Switched System
3.5 if it is autonomous. The reason for this is that we proved in Theorem 5.4 that if
it is autonomous, then there exists a time-invariant Lyapunov function. In the next
denition we present a linear programming problem that does exactly this. It is a
generalization of the linear programming problem presented in [40], [39], [17], and
[16] to serve the Switched System 3.5 in the particular case that it is autonomous.
Denition 6.12. (Linear programming problem LP(f
p
: p T, ^, PS, T, | |))
Consider the Switched System 3.5, where the set T has a nite number of elements
and the functions f
p
: | R
n
, p T are time-invariant. Let PS : R
n
R
n
be a
piecewise scaling function and ^ | be such that the interior of the set
/ :=
_
zZ
n
, PS(z+[0,1]
n
)N
PS(z + [0, 1]
n
)
is a connected set that contains the origin. Let | | be an arbitrary norm on R
n
and let
T := PS( ]d

1
, d
+
1
[ ]d

2
, d
+
2
[ . . . ]d

n
, d
+
n
[ )
be a set, of which the closure is contained in the interior of /, and either T =
or d

i
and d
+
i
are integers such that d

i
1 and 1 d
+
i
for all i = 1, 2, . . . , n.
We assume that the components of the f
p
, p T, have bounded
second-order partial derivatives on / T.
The linear programming problem LP(f
p
: p T, ^, PS, T, | |) is now con-
structed in the following way:
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 53
(i) Dene the sets
(
a
:= x R
n
: x PS(Z
n
)
_
/ T
_
,
A

:= |x| : x PS(Z
n
) /.
(ii) Dene for every Perm[1, 2, . . . , n] and every i = 1, . . . , n + 1 the
vector
x

i
:=
n

j=i
e
(j)
.
(iii) Dene the set Z
a
through:
Z
a
:=
_
(z, ) Z
n
0
P(1, 2, . . . , n) : PS(R
J
(z + [0, 1]
n
)) / T
_
.
iv) For every (z, ) Z
a
, every Perm[1, 2, . . . , n], and every i =
1, 2, . . . , n + 1 we set
y
(z,J)
,i
:= PS(R
J
(z +x

i
)).
(v) Dene the set
}
a
:=
_
y
(z,J)
,k
, y
(z,J)
,k+1
: Perm[1, 2, . . . , n], (z, ) Z
a
, k 1, 2, . . . , n
_
.
(vi) For every p T, every (z, ) Z
a
, and every r, s = 1, 2, . . . , n let B
(z,J)
p,rs
be a real-valued constant, such that
B
(z,J)
p,rs
max
i=1,2,...,n
sup
xPS(R
J
(z+[0,1]
n
))

2
f
p,i
x
r
x
s
(x)

.
(vii) For every (z, ) Z
a
, every i, k = 1, 2, . . . , n, and every
Perm[1, 2, . . . , n], dene
A
(z,J)
,k,i
:=

e
k
(y
(z,J)
,i
y
(z,J)
,n+1
)

.
(viii) Dene the constant
x
min,M
:= min|x| : x PS(Z
n
) /,
where / is the boundary of the set /.
(ix) For every p T, every (z, ) Z
a
, every Perm[1, 2, . . . , n], and
every i = 1, 2, . . . , n + 1 set
E
(z,J)
p,,i
:=
1
2
n

r,s=1
B
(z,J)
p,rs
A
(z,J)
,r,i
(A
(z,J)
,s,i
+A
(z,J)
,s,1
). (6.16)
(ix) Let > 0 and > 0 be arbitrary constants.
The variables of the linear programming problem are:

a
,

a
[y], for all y A

a
[y], for all y A

,
V
a
[x], for all x (
a
,
C
a
[x, y], for all x, y }
a
.
The linear constraints of the linear programming problem are:
54 S. F. HAFSTEIN EJDE-2007/MON. 08
(LC1a) Let y
0
, y
1
, . . . , y
K
be the elements of A

in an increasing order. Then

a
[y
0
] =
a
[y
0
] = 0,
y
1

a
[y
1
],
y
1

a
[y
1
],
and for every i = 1, 2, . . . , K 1:

a
[y
i
]
a
[y
i1
]
y
i
y
i1


a
[y
i+1
]
a
[y
i
]
y
i+1
y
i
,

a
[y
i
]
a
[y
i1
]
y
i
y
i1


a
[y
i+1
]
a
[y
i
]
y
i+1
y
i
.
(LC2a) For every x (
a
:

a
[|x|] V
a
[x].
If T = , then:
V
a
[0] = 0.
If T , = , then, for every x PS(Z
n
) T:
V
a
[x]
a
[x
min,M
] .
Further, if T , = , then for every i = 1, 2, . . . , n:
V
a
[PS
i
(d

i
)e
i
]
a
PS
i
(d

i
) and V
a
[PS
i
(d
+
i
)e
i
]
a
PS
i
(d
+
i
).
(LC3a) For every x, y }
a
:
C
a
[x, y] |x y|

V
a
[x] V
a
[y] C
a
[x, y] |x y|


a
|x y|

.
(LC4a) For every p T, every (z, ) Z
a
, every Perm[1, 2, . . . , n], and
every i = 1, 2, . . . , n + 1:

a
_
|y
(z,J)
,i
|

j=1
_
V
a
[y
(z,J)
,j
] V
a
[y
(z,J)
,j+1
]
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)
f
p,(j)
(y
(z,J)
,i
+E
(z,J)
p,,i
C
a
[y
(z,J)
,j
, y
(z,J)
,j+1
]
_
.
As the objective of the linear programming problem is not needed to parameterize
a CPWA Lyapunov function we do not dene it here.
Obviously, the two rst comments after Denition 6.8 apply equally to the linear
programming problem from this denition. Further, if the functions f
p
, p T, in
Denition 6.12 are linear, then obviously we can set B
(z,J)
p,rs
:= 0 for all p T,
all (z, ) Z
a
, and all r, s = 1, 2, . . . , n, and then the error terms E
(z,J)
p,,i
are
all identically zero. Linear problems are thus the most easy to solve with the
linear programming problem because we can drop the variables C[x, y] and the
constraints (LC3) out of the linear programming problem altogether.
If the linear programming problem LP(f
p
: p T, ^, PS, T, | |) from Deni-
tion 6.12 possesses a feasible solution, then we can use this solution to parameterize
a time-invariant CPWA Lyapunov function for the autonomous Switched System
3.5 used in the construction of the linear programming problem. The denition of
the parameterized CPWA Lyapunov function in the autonomous case is in essence
identical to the denition in the nonautonomous case.
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 55
Denition 6.13. Assume that

a
,

a
[y], for all y A

a
[y], for all y A

,
V
a
[x], for all x (
a
,
C
a
[x, y], for all x, y }
a
.
is a feasible solution to LP(f
p
: p T, ^, PS, T, | |) from Denition 6.12. Then
we dene the function V
Lya
a
trough V
Lya
a
CPWA[PS, PS
1
_
/ T
_
] and
V
Lya
a
(x) := V
a
[x] for all x (
a
.
Further, we dene the function
a
from the numerical values of the variables
a
[y]
and
a
from the numerical values of the variables
a
[y], just as the functions and
were dened in Section 6.3 from the numerical values of the variables [y] and
[y] respectively.
That V
Lya
a
in Denition 6.13 is a Lyapunov function for the autonomous Switched
System 3.5, that is equivalent to a time-invariant Lyapunov function parameter-
ized by the linear programming problem LP(f
p
: p T, ^, PS, t, T, | |) from
Denition 6.8, is proved in the next theorem.
Theorem 6.14. Consider the Switched System 3.5 where the set T is nite. Let
T

and T

be constants such that 0 T

< T

and let PS : R
n
R
n
be a piecewise
scaling function and ^ | be such that the interior of the set
/ :=
_
zZ
n
, PS(z+[0,1]
n
)N
PS(z + [0, 1]
n
)
is a connected set that contains the origin. Let | | be an arbitrary norm on R
n
and let
T := PS( ]d

1
, d
+
1
[ ]d

2
, d
+
2
[ . . . ]d

n
, d
+
n
[ )
be a set, of which the closure is contained in the interior of /, and either T =
or d

i
and d
+
i
are integers such that d

i
1 and 1 d
+
i
for all i = 1, 2, . . . , n.
Finally, let t := (t
0
, t
1
, . . . , t
M
) R
M+1
, M N
>0
be a vector such that T

=: t
0
<
t
1
< < t
M
:= T

.
Assume the the Switched System 3.5 is autonomous, that is, that the f
p
, p T
are time-independent, and assume that the second-order partial derivatives of their
components are bounded on /T. Then, the linear programming problem LP(f
p
:
p T, ^, PS, t, T, | |) from Denition 6.8 with the additional linear constraints:
(LC-A) For every x, y ( such that | x y|

= 0:
V [ x] = V [ y].
For every x, y } such that | x y|

= 0:
C[ x, y] = 0.
Is equivalent to the linear programming problem LP(f
p
: p T, ^, PS, T, | |)
from Denition 6.12, in the following sense:
56 S. F. HAFSTEIN EJDE-2007/MON. 08
(i) If V
Lya
is a Lyapunov function, dened as in 6.3 from a feasible solution
to the linear programming problem LP(f
p
: p T, ^, PS, t, T, | |) from
Denition 6.8 that additionally satises the constraints (LC-A), then V
Lya
does not depend on t and we can parameterize a Lyapunov function W
Lya
with the linear programming problem LP(f
p
: p T, ^, PS, T, | |) from
Denition 6.12, such that
W
Lya
(x) = V
Lya
(T

, x) for all x / T.
(ii) If W
Lya
is a Lyapunov function, dened as the function V
Lya
a
in Denition
6.13, from a feasible solution to the linear programming problem LP(f
p
:
p T, ^, PS, T, | |) from Denition 6.12, then we can parameterize a
Lyapunov function V
Lya
by use of the linear programming problem LP(f
p
:
p T, ^, PS, t, T, | |) from Denition 6.8 with (LC-A) as additional
constraints, such that
V
Lya
(t, x) = W
Lya
(x) for all t [T

, T

] and all x / T.
In both cases one should use the same numerical values for the bounds B
(z,J)
p,rs
on
the second-order partial derivatives of the components of the functions f
p
, p T,
and for the constants and .
Proof. We start by proving proposition (i):
Assume that

a
,

a
[y], for all y A

a
[y], for all y A

,
V
a
[x], for all x (
a
,
C
a
[x, y], for all x, y }
a
.
is a feasible solution to the (autonomous) linear programming problem
LP(f
p
: p T, ^, PS, T, | |) from Denition 6.12, dene C
a
[x, x] := 0 for all
x (
a
, and set
:=
a
,
[y] :=
a
[y], for all y A

,
[y] :=
a
[y], for all y A

,
V [ x] := V
a
[( x
1
, x
2
, . . . , x
n
)], for all x (,
C[ x, y] := C
a
[( x
1
, x
2
, . . . , x
n
), ( y
1
, y
2
, . . . , y
n
)], for all x, y }.
We claim that , [y], [y], V [ x], and C[ x, y] is a feasible solution to the linear
programming problem LP(f
p
: p T, ^, PS, t, T, | |) from Denition 6.8 that
additionally satises the constraints (LC-A). If this is the case, then clearly
V
Lya
CPWA[

PS,

PS
1_
[T

, T

]
_
/ T
__
],
dened through V
Lya
( x) := V [ x] for all x (,
is the promised Lyapunov function.
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 57
It is a simple task so conrm that they satisfy the constraints (LC1), (LC2),
(LC3), and (LC-A), so we only prove that they fulll the constraints (LC4), which
is not as obvious.
Let p T, (z, ) Z, Perm[0, 1, . . . , n], and i 0, 1, . . . , n + 1 be
arbitrary, but xed throughout this part of the proof. We have to show that

_
|y
(z,J)
,i
|

j=0
_
V [y
(z,J)
,j
] V [y
(z,J)
,j+1
]
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)

f
p,(j)
(y
(z,J)
,i
) +E
(z,J)
p,,i
C[y
(z,J)
,j
, y
(z,J)
,j+1
]
_
.
(6.17)
We dene the mapping l

: 0, 1, . . . , n + 1 1, . . . , n + 1 through
l

(k) :=
_
k + 1, if 0 k
1
(0),
k, otherwise,
and Perm[1, 2, . . . , n] through
(k) :=
_
(k 1), if 1 k
1
(0),
(k), if
1
(0) < k n.
Further, we set
z

:= (z
1
, z
2
, . . . , z
n
), where z = (z
0
, z
1
, . . . , z
n
).
Note that by these denitions and the denitions of y
(z,J)
,j
and y
(z

,J)
,j
we have for
all j = 0, 1, . . . , n + 1 and all r = 1, 2, . . . , n, that
e
r
y
(z,J)
,j
= e
r


PS(

R
J
(z +
n

k=j
e
(k)
))
= e
r
PS(R
J
(z

+
n

k=j, (k)=0
e
(k)
))
= e
r
PS(R
J
(z

+
n

k=l(j)
e
(k)
))
= e
r
y
(z

,J)
,l(j)
.
(6.18)
Especially, because l

(0) = 1, we have
e
r
y
(z,J)
,0
= e
r
y
(z

,J)
,1
for all r = 1, 2, . . . , n.
But then |y
(z,J)
,i
|

= |y
(z

,J)
,l(i)
| and for every r = 1, 2, . . . , n and every j =
0, 1, . . . , n we have
A
(z,J)
,r,j
=

e
r
(y
(z,J)
,j
y
(z,J)
,n+1
)

e
r
(y
(z

,J)
,l(j)
y
(z

,J)
,n+1
)

= A
(z

,J)
,r,l(j)
. (6.19)
Further, because

f
p
does not depend on the rst argument,

f
p,(j)
(y
(z,J)
,i
) =
_
1, if (j) = 0,
f
p,(l(j))
(y
(z

,J)
,l(i)
), if j 1, 2, . . . , n
1
(0),
(6.20)
58 S. F. HAFSTEIN EJDE-2007/MON. 08
and similarly
V [y
(z,J)
,j
] = V
a
[y
(z

,J)
,l(j)
] for all j = 0, 1, . . . , n + 1, (6.21)
and
C[y
(z,J)
,j
, y
(z,J)
,j+1
] = C
a
[y
(z

,J)
,l(j)
, y
(z

,J)
,l(j+1)
] for all j = 0, 1, . . . , n. (6.22)
Especially,
V [y
(z,J)
,j
] V [y
(z,J)
,j+1
] = C[y
(z,J)
,j
, y
(z,J)
,j+1
] = 0 if (j) = 0. (6.23)
For the bounds B
(z,J)
p,rs
on the second-order partial derivatives of the components
of the f
p
we demand,
B
(z,J)
p,rs
max
i=1,2,...,n
sup
x
f
PS(
e
R
J
(z+[0,1]
n+1
))


2
f
p,i
x
r
x
s
( x)

,
which is compatible with
B
(z,J)
p,rs
:= 0 if r = 0 or s = 0
and
B
(z,J)
p,rs
:= B
(z

,J)
p,rs
for r, s = 1, 2, . . . , n.
This together with (6.19) implies that
E
(z,J)
p,,i
:=
1
2
n

r,s=0
B
(z,J)
p,rs
A
(z,J)
,r,i
(A
(z,J)
,s,i
+A
(z,J)
,s,0
) (6.24)
=
1
2
n

r,s=1
B
(z

,J)
p,rs
A
(z,J)
,r,l(i)
(A
(z

,J)
,s,l(i)
+A
(z

,J)
,s,1
)
= E
(z

,J)
p,,l(i)
.
Now, by assumption,

a
_
|y
(z

,J)
,l(i)
|

j=1
_
V
a
[y
(z

,J)
,j
] V
a
[y
(z

,J)
,j+1
]
e
(j)
(y
(z

,J)
,j
y
(z

,J)
,j+1
)
f
p,(j)
(y
(z

,J)
,l(i)
) +E
(z

,J)
p,,l(i)
C
a
[y
(z

,J)
,j
, y
(z

,J)
,j+1
]
_
,
so by (6.18), the denition of the function l

, (6.24), (6.20), (6.21), (6.22), and


(6.23), we have

_
|y
(z,J)
,i
|

=
a
_
|y
(z

,J)
,l(i)
|

j=1
_
V
a
[y
(z

,J)
,j
] V
a
[y
(z

,J)
,j+1
]
e
(j)
(y
(z

,J)
,j
y
(z

,J)
,j+1
)
f
p,(j)
(y
(z

,J)
,l(i)
) +E
(z

,J)
p,,l(i)
C
a
[y
(z

,J)
,j
, y
(z

,J)
,j+1
]
_
=
n

j=0
(j)=0
_
V
a
[y
(z

,J)
,l(j)
] V
a
[y
(z

,J)
,l(j+1)
]
e
(l(j))
(y
(z

,J)
,l(j)
y
(z

,J)
,l(j+1)
)
f
p,(l(j))
(y
(z

,J)
,l(i)
)
+E
(z

,J)
p,,l(i)
C
a
[y
(z

,J)
,l(j)
, y
(z

,J)
,l(j+1)
]
_
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 59
=
n

j=0
(j)=0
_
V [y
(z,J)
,j
] V [y
(z,J)
,j+1
]
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)

f
p,(j)
(y
(z,J)
,i
) +E
(z,J)
p,,i
C[y
(z,J)
,j
, y
(z,J)
,j+1
]
_
=
n

j=0
_
V [y
(z,J)
,j
] V [y
(z,J)
,j+1
]
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)

f
p,(j)
(y
(z,J)
,i
) +E
(z,J)
p,,i
C[y
(z,J)
,j
, y
(z,J)
,j+1
]
_
and we have proved (6.17).
We now prove proposition (ii): Assume that
,
[y], for all y A

,
[y], for all y A

,
V [ x], for all x (,
C[ x, y], for all x, y }.
is a solution to the linear programming problem LP(f
p
: p T, ^, PS, t, T, | |)
from Denition 6.8 and set

a
:= ,

a
[y] := [y], for all y A

a
[y] := [y], for all y A

,
V
a
[x] := V [(T

, x)], for all x (


a
,
C
a
[x, y] := C[(T

, x), (T

, y)], for all x, y }


a
.
We claim that
a
,
a
[y],
a
[y], V
a
[x], and C
a
[x, y] is a feasible solution to the
linear programming problem LP(f
p
: p T, ^, PS, T, | |) from Denition 6.12,
that
V
Lya
CPWA[

PS,

PS
1_
[T

, T

]
_
/ T
__
],
dened through V
Lya
( x) := V [ x] for all x (,
does not depend on the rst argument (the time), and that
W
Lya
CPWA[PS, PS
1
, / T],
dened through W
Lya
(x) := V
a
[x] for all x (
a
,
is the promised Lyapunov function.
First we prove that the function values of V
Lya
do not depend on the rst
argument. To do this it is obviously enough to show that this holds for every
simplex in the simplicial partition of [T

, T

]
_
/ T
_
. Let (z, ) Z and
Perm[0, 1, . . . , n] be arbitrary and let
x, y cony
(z,J)
,0
, y
(z,J)
,1
, . . . , y
(z,J)
,n+1

be such that
e
r
x = e
r
y for all r = 1, 2, . . . , n.
We are going to show that V
Lya
( x) = V
Lya
( y).
60 S. F. HAFSTEIN EJDE-2007/MON. 08
Set k :=
1
(0) and let
0
,
1
, . . . ,
n+1
[0, 1] and
0
,
1
, . . . ,
n+1
[0, 1] be
such that
x =
n+1

i=0

i
y
(z,J)
,i
, y =
n+1

i=0

i
y
(z,J)
,i
,
n+1

i=0

i
=
n+1

i=0

i
= 1.
From the denition of the y
(z,J)
,i
it follows that
e
(0)

n+1

i=0

i
y
(z,J)
,i
= e
(0)

n+1

i=0

i
y
(z,J)
,i
, hence
0
=
0
,
which implies
e
(1)

n+1

i=1

i
y
(z,J)
,i
= e
(1)

n+1

i=1

i
y
(z,J)
,i
, hence
1
=
1
,
.
.
.
which implies
e
(k1)

n+1

i=k1

i
y
(z,J)
,i
= e
(k1)

n+1

i=k1

i
y
(z,J)
,i
, hence
k1
=
k1
,
which implies
e
(k+1)

n+1

i=k

i
y
(z,J)
,i
= e
(k+1)

n+1

i=k

i
y
(z,J)
,i
, hence
k
+
k+1
=
k
+
k+1
,
which implies
e
(k+2)

n+1

i=k+2

i
y
(z,J)
,i
= e
(k+2)

n+1

i=k+2

i
y
(z,J)
,i
, hence
k+2
=
k+2
,
.
.
.
which implies
e
(n)

n+1

i=n

i
y
(z,J)
,i
= e
(n)

n+1

i=n

i
y
(z,J)
,i
, hence
n
=
n
.
Then
n+1
=
n+1
and because by (LC-A) we have V [y
(z,J)
,k
] = V [y
(z,J)
,k+1
] we get
V
Lya
( x) = V
Lya
(
n+1

i=0

i
y
(z,J)
,i
) =
n+1

i=0

i
V [y
(z,J)
,i
]
=
n+1

i=0
i=k,k+1

i
V [y
(z,J)
,i
] + (
k
+
k+1
)V [y
(z,J)
,k
]
=
n+1

i=0
i=k,k+1

i
V [y
(z,J)
,i
] + (
k
+
k+1
)V [y
(z,J)
,k
]
=
n+1

i=0

i
V [y
(z,J)
,i
] = V
Lya
(
n+1

i=0

i
y
(z,J)
,i
)
= V
Lya
( y)
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 61
and we have proved that V
Lya
does not depend on the rst argument.
Now, let p T, (z, ) Z
a
, Perm[1, 2, . . . , n], and i 1, 2, . . . , n+1 be
arbitrary, but xed throughout the rest of the proof. To nish the proof we have
to show that

a
_
|y
(z,J)
,i
|

j=1
_
V
a
[y
(z,J)
,j
] V
a
[y
(z,J)
,j+1
]
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)
f
p,(j)
(y
(z,J)
,i
) +E
(z,J)
p,,i
C
a
[y
(z,J)
,j
, y
(z,J)
,j+1
]
_
.
(6.25)
We dene Perm[0, 1, 2, . . . , n] trough
(k) :=
_
0, if k = 0,
(k), if k 1, 2, . . . , n,
and z

( through
z

:= (T

, z
1
, z
2
, . . . , z
n
), where z = (z
1
, z
2
, . . . , z
n
).
Then, for every r = 1, 2, . . . , n, one easily veries that
e
r
y
(z

,J)
,0
= e
r
y
(z

,J)
,1
and, for all r = 1, 2, . . . , n and all j = 1, 2, . . . , n + 1, that
e
r
y
(z

,J)
,j
= e
r


PS(

R
J
(z

+
n

k=j
e
(k)
))
= e
r
PS(R
J
(z +
n

k=j
e
(k)
)) = e
r
y
(z,J)
,j
.
(6.26)
Then |y
(z

,J)
,i
|

= |y
(z,J)
,i
| and for every r = 1, 2, . . . , n
A
(z

,J)
,r,0
=

e
r
(y
(z

,J)
,0
y
(z

,J)
,n+1
)

e
r
(y
(z,J)
,1
y
(z,J)
,n+1
)

= A
(z,J)
,r,1
. (6.27)
and for every every j, r = 1, 2, . . . , n we have
A
(z

,J)
,r,j
=

e
r
(y
(z

,J)
,j
y
(z

,J)
,n+1
)

e
r
(y
(z,J)
,j
y
(z,J)
,n+1
)

= A
(z,J)
,r,j
. (6.28)
For every k = 0, 1, . . . , M dene
z
k
:= (t
k
, z
1
, z
2
, . . . , z
n
), where z := (z
1
, z
2
, . . . , z
n
),
and dene for every r, s = 1, 2, . . . , n
B
(z,J)
p,rs
:= min
k=0,1,...,n
B
(z
k
,J)
p,rs
.
Now set
B
(z
k
,J)
p,00
:= 0 for all k = 0, 1, . . . , M
and
B
(z
k
,J)
p,rs
:= B
(z,J)
p,rs
for all k = 0, 1, . . . , M and all r, s = 1, 2, . . . , n,
and consider, that with these possibly tighter bounds B
(z
k
,J)
p,rs
in the linear pro-
gramming problem LP(f
p
: p T, ^, PS, t, T, | |), the , [y], [y], V [ x], and
62 S. F. HAFSTEIN EJDE-2007/MON. 08
C[ x, y] are, of course, still a solution. Therefore we can just as well assume that
these were the bounds ab initio. It follows by (6.28) that
E
(z

,J)
p,,i
=
1
2
n

r,s=0
B
(z

,J)
p,rs
A
(z

,J)
,r,i
(A
(z

,J)
,s,i
+A
(z,J)
,s,0
) (6.29)
=
1
2
n

r,s=1
B
(z,J)
p,rs
A
(z,J)
,r,i
(A
(z,J)
,s,i
+A
(z,J)
,s,1
)
= E
(z,J)
p,,i
.
By assumption

_
|y
(z

,J)
,i
|

j=0
_
V [y
(z

,J)
,j
] V [y
(z

,J)
,j+1
]
e
(j)
(y
(z

,J)
,j
y
(z

,J)
,j+1
)

f
p,(j)
(y
(z

,J)
,i
) +E
(z

,J)
p,,i
C[y
(z

,J)
,j
, y
(z

,J)
,j+1
]
_
,
so, by (6.26), (LC-A), because f
p
does not depend on the rst argument, and (6.29),
we get

_
|y
(z,J)
,i
|

=
_
|y
(z

,J)
,i
|

j=0
_
V [y
(z

,J)
,j
] V [y
(z

,J)
,j+1
]
e
(j)
(y
(z

,J)
,j
y
(z

,J)
,j+1
)

f
p,(j)
(y
(z

,J)
,i
) +E
(z

,J)
p,,i
C[y
(z

,J)
,j
, y
(z

,J)
,j+1
]
_
=
n

j=1
_
V
a
[y
(z,J)
,j
] V
a
[y
(z,J)
,j+1
]
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)
f
p,(j)
(y
(z,J)
,i
) +E
(z,J)
p,,i
C
a
[y
(z,J)
,j
, y
(z,J)
,j+1
]
_
and we have proved (6.25) and the proof is complete.
An immediate consequence of Theorem 6.14 is a theorem, similar to Theorem 6.9,
but for autonomous systems possessing a time-invariant CPWA Lyapunov function.
Theorem 6.15. (Autonomous CPWA Lyapunov functions by linear programming)
Consider the linear programming problem LP(f
p
: p T, ^, PS, T, | |) from
Denition 6.12 and assume that it possesses a feasible solution. Let the functions

a
,
a
, and V
Lya
a
be dened as in Denition 6.13 from the numerical values of the
variables
a
[y],
a
[y], and V
a
[ x] from a feasible solution. Then the inequality

a
(|x|) V
Lya
a
(x)
holds for all x / T. If T = we have
a
(0) = V
Lya
a
(t, 0) = 0. If T ,= we
have, with
V
Lya
M,min
:= min
xM
V
Lya
a
(x),
V
Lya
D,max
:= max
xD
V
Lya
a
(x),
that
V
Lya
D,max
V
Lya
M,min
.
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 63
Further, with as the solution to the Switched System 3.5 that we used in the
construction of the linear programming problem, the inequality

a
(|

(t, )|) limsup


h0+
V
Lya
(

(t +h, )) V
Lya
(

(t, ))
h
hold true for all o
P
and all in the interior of / T.
Proof. Follows directly by Theorem 6.9 and Theorem 6.14.
We conclude this discussion with a theorem, that is the equivalent of Theorem
6.11 for autonomous systems possessing a time-invariant CPWA Lyapunov function,
parameterized by the linear programming problem LP(f
p
: p T, ^, PS, T, | |)
from Denition 6.12. It states some stability properties such a system must have.
Theorem 6.16 (Implications of the Lyapunov function V
Lya
a
).
Make the same assumptions and denitions as in Theorem 6.15 and assume ad-
ditionally that the norm | | in the linear programming problem LP(f
p
: p
T, ^, PS, T, | |) is a k-norm
2
, 1 k +. Dene the set T through T := 0
if T = and
T := T
_
x / T : V
Lya
a
(x) V
Lya
D,max
_
, if T , = ,
and the set / through
/ :=
_
x / T : V
Lya
(x) < V
Lya
M,min
_
.
Set q := k (k 1)
1
if 1 < k < +, q := 1 if k = +, and q := + if k = 1,
and dene the constant
E
q
:= |
n

i=1
e
i
|
q
.
Then the following propositions hold true:
(i) If T , then

(t, ) T for all o


P
and all t 0.
(ii) If / T, the inequality
V
Lya
a
(

(t, )) V
Lya
a
() exp
_


a
E
q
t
_
(6.30)
holds for all t such that

(t

, ) / T for all 0 t

t.
(iii) If / T and T = , then inequality (6.30) holds for all t 0 and all
o
P
. If / T and T , = , then, for every o
P
there is a t

0,
such that inequality (6.30) holds for all 0 t t

(t

, ) T , and

(t, ) T for all t t

.
The proof follows directly by Theorem 6.11.
2
With k-norm we mean the norm x
k
:=
`P
n
i=1
|x
i
|
k

1/k
if 1 k < + and x :=
max
i=1,2,...,n
|x
i
|. Unfortunately, these norms are usually called p-norms, which is inappropriate
in this context because the alphabet p is used to index the functions fp, p P.
64 S. F. HAFSTEIN EJDE-2007/MON. 08
7. Constructive Converse Theorems
We show how to combine the results from Theorem 5.4, a non-constructive con-
verse Lyapunov theorem, and the linear programming problem from Denition 6.8,
to prove a constructive converse Lyapunov theorem. We will do this for the general,
not necessarily autonomous, case. Thereafter, we will do the same for autonomous
switched systems and we will prove that in this case, we can parameterize a time-
invariant Lyapunov function by the use of the linear programming problem from
Denition 6.12.
The structure of this section is somewhat unconventional because we start with
a proof of the yet to be stated Theorem 7.1. In the proof we assign values to
the constants and to the variables of the linear programming problem such that a
feasible solution results. By these assignments we use the numerical values of the
Lyapunov function from Theorem 5.4. Note, that because Theorem 5.4 is a pure
existence theorem, the numerical values of this Lyapunov function are not known.
However, our knowledge about these numerical values and their relations is sub-
stantial. Indeed, we have enough information to prove that the linear constraints
(LC1), (LC2), (LC3), and (LC4), of the linear programming problem in Denition
6.8 are fullled by the numerical values we assign to the variables and to the con-
stants. Because there are well-known algorithms to nd a feasible solution to a
linear programming problem if the set of feasible solutions is not empty, this im-
plies that we can always parameterize a Lyapunov function by the use of the linear
programming problem in Denition 6.8, whenever the underlying system possesses
a Lyapunov function at all.
7.1. The assumptions. Consider the Switched System 3.5 and assume that the set
T is nite and that f
p
is a [(
2
(R
0
|)]
n
function for every p T. Further, assume
that there is an a > 0 such that [a, a]
n
| and W (
2
(R
0
([a, a]
n
0)) is
a Lyapunov function for the switched system. By Theorem 5.4 this is, for example,
the case if the origin is a uniformly asymptotically stable equilibrium of the Switched
System 3.5, [a, a]
n
is a subset of its region of attraction, and the functions f
p
all
satisfy the Lipschitz condition: for every p T there exists a constant L
p
such that
|f
p
(t, x)f
p
(s, y)| L
p
([st[|xy|), for all s, t R
0
and all x, y [a, a]
n
.
By Denition 4.3 there exist, for an arbitrary norm | | on R
n
, class / functions
, , and , such that
(|x|) W(t, x) (|x|)
and
[
x
W](t, x) f
p
(t, x) +
W
t
(t, x) (|x|) (7.1)
for all (t, x) R
>0
(] a, a[
n
0) and all p T. Further, by Lemma 3.17, we
can assume without loss of generality that and are convex functions. Now, let
0 T

< T

< +be arbitrary and let T

[a, a]
n
be an arbitrary neighborhood
of the origin. Especially, the set T

,= can be taken as small as one wishes. We


are going to prove that we can parameterize a CPWA Lyapunov function on the set
[T

, T

]
_
[a, a]
n
T

_
. We will start by assigning values to the constants and the
variables of the linear programming problem LP(f
p
: p T, ] a, a[
n
, PS, t, T, |
|) in Denition 6.8. This includes that we dene the piecewise scaling function
PS, the vector t, and the set T T

. Thereafter, we will prove that the linear


constraints of the linear programming problem are all fullled by these values.
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 65
7.2. The assignments. First, we determine a constant B that is an upper bound
on all second-order partial derivatives of the components of the functions f
p
, p T.
That is, with x = ( x
0
, x
1
, . . . , x
n
) := (t, x) and

f
p
( x) = (

f
p,0
( x),

f
p,1
( x), . . . ,

f
p,n
( x)) = (1, f
p,1
(t, x), f
p,2
(t, x), . . . , f
p,n
(t, x)),
we need a constant B < + such that
B max
pP, i,r,s=0,1,...,n, x[T

,T

][a,a]
n


2
f
p,i
x
r
x
s
( x)

.
We must, at least in principle, be able to assign a numerical value to the constant
B. This is in contrast to the rest of the constants and variables, where the mere
knowledge of the existence of the appropriate values suces. However, because
B is an arbitrary upper bound (no assumptions are needed about its quality) on
the second-order partial derivatives of the components of the functions f
p
on the
compact set [T

, T

] [a, a]
n
, this should not cause any diculties if the algebraic
form of the components is known. It might sound strange that the mere existence
of the appropriate values to be assigned to the other variables suces in a construc-
tive theorem. However, as we will prove later on, if they exist then the simplex
algorithm, for example, will successfully determine valid values for them.
With
x

min
:= min
x=a
|x|
we set
:=
(x

min
)
2
and let m

be a strictly positive integer, such that


[
a
2
m

,
a
2
m

]
n
x R
n
: (|x|) T

(7.2)
and set
T :=]
a
2
m

,
a
2
m

[
n
.
Note that we do not know the numerical values of the constants and m

because
and are unknown. However, their mere existence allows us to properly dene
and m

. We will keep on introducing constants in this way. Their existence is


secured in the sense that there exists a constant with the following property.
Set
x

:= 2
m

min
,

:=
1
2
(x

),
A

:= sup
pP, x[T

,T

][a,a]
n
|

f
p
( x)|
2
.
We dene

W( x) := W(t, x), where x := (t, x), and assign
C := max
r=0,1,...,n
x[T

,T

]([a,a]
n
\D)

W
x
r
( x)

,
B

:= (n + 1)
3
2
max
r,s=0,1,...,n
x[T

,T

]([a,a]
n
\D)

W
x
r
x
s
( x)

,
C

:= (n + 1)
3
CB.
66 S. F. HAFSTEIN EJDE-2007/MON. 08
We set a

:= maxT

, a and let m m

be an integer, such that


a

2
m

_
(A

)
2
+ 4x

2C

,
and set d := 2
mm

.
We dene the piecewise scaling function PS : R
n
R
n
through
PS(j
1
, j
2
, . . . , j
n
) := a2
m
(j
1
, j
2
, . . . , j
n
)
for all (j
1
, j
2
, . . . , j
n
) Z
n
and the vector t := (t
0
, t
1
, . . . , t
2
m), where
t
j
:= T

+ 2
m
j(T

)
for all j = 0, 1, . . . , 2
m
.
We assign the following values to the variables and the remaining constants of
the linear programming problem LP(f
p
: p T, ^, PS, t, T, | |):
B
(z,J)
p,rs
:= B, for all p T, all (z, ) Z, and all r, s = 0, 1, . . . , n,
[y] := (y), for all y A

,
[y] :=

y, for all y A

,
V [ x] :=

W( x) for all x (,
C[ x, y] := C, for all x, y },
:= max
_
C, a
1
2
m

max
i=1,2,...,n
(a2
m

|e
i
|)
_
,
:= min

, (y
1
)/y
1
, where y
1
:= miny : y A

and y ,= 0.
We now show that the linear constraints (LC1), (LC2), (LC3), and (LC4) of the
linear programming problem LP(f
p
: p T, ^, PS, t, T, | |) are satised by
these values.
7.3. The constraints (LC1) are fullled. Let y
0
, y
1
, . . . , y
K
be the elements of
A

in an increasing order. We have to show that [y


0
] = [y
0
] = 0, y
1
[y
1
],
y
1
[y
1
], and that for every i = 1, 2, . . . , K 1:
[y
i
] [y
i1
]
y
i
y
i1

[y
i+1
] [y
i
]
y
i+1
y
i
,
[y
i
] [y
i1
]
y
i
y
i1

[y
i+1
] [y
i
]
y
i+1
y
i
.
Proof. Clearly [y
0
] = [y
0
] = 0 because y
0
= 0 and
y
1

y
1
= [y
1
] and y
1

(y
1
)
y
1
y
1
= [y
1
].
Because is convex we have for all i = 1, 2, . . . , K 1 that
y
i
y
i1
y
i+1
y
i1
(y
i+1
) +
y
i+1
y
i
y
i+1
y
i1
(y
i1
) (y
i
),
that is
(y
i
) (y
i1
)
y
i
y
i1
=
[y
i
] [y
i1
]
y
i
y
i1

[y
i+1
] [y
i
]
y
i+1
y
i
=
(y
i+1
) (y
i
)
y
i+1
y
i
.
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 67
Finally, we clearly have for every i = 1, 2, . . . , K 1 that

2
=
[y
i
] [y
i1
]
y
i
y
i1

[y
i+1
] [y
i
]
y
i+1
y
i
=

2
.

7.4. The constraints (LC2) are fullled. We have to show that for every x (
we have
[| x|

] V [ x],
that for every x = ( x
0
, x
1
, . . . , x
n
), such that ( x
1
, x
2
, . . . , x
n
) PS(Z
n
) T we
have
V [ x] [x
min,M
] ,
and that for every i = 1, 2, . . . , n and every j = 0, 1, . . . , 2
m
we have
V [PS
0
(j)e
0
+ PS
i
(d

i
)e
i
] PS(d

i
)
and
V [PS
0
(j)e
0
+ PS
i
(d

i
)e
i
] PS
i
(d
+
i
).
Proof. Clearly,
[| x|

] = (| x|

)

W( x) = V [ x]
for all x (. For every x = ( x
0
, x
1
, . . . , x
n
), such that ( x
1
, x
2
, . . . , x
n
) PS(Z
n
)
T, we have by (7.2) that
V [ x] =

W( x) (| x|

) = (x

min
) (x
min,M
) = [x
min,M
] .
Finally, note that d
+
i
= d

i
= d = 2
mm

for all i = 1, 2, . . . , n, which implies


that for every i = 1, 2, . . . , n and j = 0, 1, . . . , 2
m
we have
V [PS
0
(j)e
0
+ PS(d
+
i
)e
i
] = V [PS
0
(j)e
0
+a2
m

e
i
]
= W(t
j
, a2
m

e
i
)
(a2
m

|e
i
|)
a2
m

= PS
i
(d
+
i
)
and
V [PS
0
(j)e
0
+ PS(d

i
)e
i
] = V [PS
0
(j)e
0
a2
m

e
i
]
= W(t
j
, a2
m

e
i
)
(a2
m

|e
i
|)
a2
m

= PS(d

i
).

68 S. F. HAFSTEIN EJDE-2007/MON. 08
7.5. The constraints (LC3) are fullled. We have to show that x, y }
implies the inequalities:
C[ x, y] | x y|

V [ x] V [ y] C[ x, y] | x y|

| x y|

.
Proof. Let x, y }. Then there is an i 0, 1, . . . , n such that x y =
e
i
| x y|

. By the Mean-value theorem there is a ]0, 1[ such that

V [ x] V [ y]
| x y|

W( x)

W( y)
| x y|

W
x
i
( y +( x y))

.
Hence, by the denition of the constants C and ,

V [ x] V [ y]
| x y|

C ,
which implies that the constraints (LC3) are fullled.
7.6. The constraints (LC4) are fullled. We have to show that for arbitrary
p T, (z, ) Z, Perm[0, 1, . . . , n], and i 0, 1, . . . , n + 1 we have

_
|y
(z,J)
,i
|

j=0
_
V [y
(z,J)
,j
] V [y
(z,J)
,j+1
]
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)

f
p,(j)
(y
(z,J)
,i
) +E
(z,J)
p,,i
C[y
(z,J)
,j
, y
(z,J)
,j+1
]
_
.
(7.3)
Proof. With the values we have assigned to the variables and the constants of the
linear programming problem we have for every p T, every (z, ) Z, every
Perm[0, 1, . . . , n], every i, j = 0, 1, . . . , n, and with h := a

2
m
, that
A
(z,J)
,i,j
h,
E
(z,J)
p,,i
:=
1
2
n

r,s=0
B
(z,J)
p,rs
A
(z,J)
,r,i
(A
(z,J)
,s,i
+A
(z,J)
,s,1
) (n + 1)
2
Bh
2
,
n

j=0
C[y
(z,J)
,j
, y
(z,J)
,j+1
] (n + 1)C.
Hence, inequality (7.3) follows if we can prove that

_
|y
(z,J)
,i
|

|y
(z,J)
,i
|

j=0

W(y
(z,J)
,j
)

W(y
(z,J)
,j+1
)
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)

f
p,(j)
(y
(z,J)
,i
) +C

h
2
.
(7.4)
Now, by the Cauchy-Schwarz inequality and inequality (7.1),
n

j=0

W(y
(z,J)
,j
)

W(y
(z,J)
,j+1
)
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)

f
p,(j)
(y
(z,J)
,i
)
=
n

j=0
_
W(y
(z,J)
,j
)

W(y
(z,J)
,j+1
)
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)

W
x
(j)
(y
(z,J)
,i
)
_

f
p,(j)
(y
(z,J)
,i
)
+
x

W(y
(z,J)
,i
)

f
p
(y
(z,J)
,i
)
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 69

_
_
_
n

j=0
_
W(y
(z,J)
,j
)

W(y
(z,J)
,j+1
)
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)

W
x
(j)
(y
(z,J)
,i
)
_
e
j
_
_
_
2
|

f
p
(y
(z,J)
,i
)|
2
(|y
(z,J)
,i
|

)
By the Mean-value theorem there is an y on the line-segment between the vectors
y
(z,J)
,j
and y
(z,J)
,j+1
, such that

W(y
(z,J)
,j
)

W(y
(z,J)
,j+1
)
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)
=

W
x
(j)
(y)
and an y

on the line-segment between the vectors y and y


(z,J)
,i
such that

W
x
(j)
(y)

W
x
(j)
(y
(z,J)
,i
) =
_

W
x
(j)

(y

) (y y
(z,J)
,i
).
Because y and y
(z,J)
,i
are both elements of the simplex PS(R
J
(z +S

)), we have
|y y
(z,J)
,i
|
2
h

n + 1
and because
_
_
_

W
x
(j)

(y

)
_
_
2

n + 1 max
r,s=0,1,...,n
x[T

,T

]([a,a]
n
\D)

W
x
r
x
s
( x)

,
we obtain
_
_
_
n

j=0
_
W(y
(z,J)
,j
)

W(y
(z,J)
,j+1
)
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)

W
x
(j)
(y
(z,J)
,i
)
_
e
j
_
_
_
2
hB

.
Finally, by the denition of A

, |

f
p
(y
(z,J)
,i
)|
2
A

. Putting the pieces together


delivers the inequality
n

j=0

W(y
(z,J)
,j
)

W(y
(z,J)
,j+1
)
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)

f
p,(j)
(y
(z,J)
,i
) hB

(|y
(z,J)
,i
|

).
From this inequality and because (x) 2

x for all x x

and because of the


fact that |y
(z,J)
,i
|

, inequality (7.4) holds if

|y
(z,J)
,i
|

hA

|y
(z,J)
,i
|

+h
2
C

.
This last inequality follows from
h :=
a

2
m

_
(A

)
2
+ 4x

2C

,
which implies
0 hA

+h
2
C

hA

|y
(z,J)
,i
|

+h
2
C

.
Because p T, (z, ) Z, Perm[0, 1, . . . , n], and i 0, 1, . . . , n + 1 were
arbitrary, the proof is complete
70 S. F. HAFSTEIN EJDE-2007/MON. 08
In the last proof we took care of that the second-order polynomial
P(z) := z
2
C

+zA

has two distinct real-valued roots, one smaller than zero and one larger then zero.
Further, because h := a

2
m
> 0 is not larger than the positive root, we have
P(h) 0, which is exactly what we need in the proof so that everything adds up.
7.7. Summary of the results. In this Section 7, we have delivered a proof of the
following theorem:
Theorem 7.1 (Constructive converse theorem for arbitrary switched systems).
Consider the Switched System 3.5 where T is a nite set, let a > 0 be a real-valued
constant such that [a, a]
n
|, and assume that at least one of the following two
assumptions holds:
(i) There exists a Lyapunov function W (
2
(R
0
([a, a]
n
0)) for the
Switched System 3.5.
(ii) The origin is a uniformly asymptotically stable equilibrium point of the
Switched System 3.5, the set [a, a]
n
is contained in its region of attraction,
and the functions f
p
satisfy the Lipschitz condition: for every p T there
exists a constant L
P
such that
|f
p
(s, x) f
p
(t, y)| L
p
([s t[ +|x y|)
for all s, t R
0
and all x, y [a, a]
n
.
Then, for every constants 0 T

< T

< + and every neighborhood ^


[a, a]
n
of the origin, no matter how small, it is possible to parameterize a Lyapunov
function V
Lya
of class CPWA,
V
Lya
: [T

, T

]
_
[a, a]
n
^
_
R,
for the Switched System 3.5 by using the linear programming problem dened in
Denition 6.8.
More concretely: Let m be a positive integer and dene the piecewise scaling
function PS : R
n
R
n
, the set T, and the vector t of the linear programming
problem through
PS(j
1
, j
2
, . . . , j
n
) := a2
m
(j
1
, j
2
, . . . , j
n
),
T :=] 2
k
a
2
m
, 2
k
a
2
m
[
n
^,
for some integer 1 k < m, and
t := (t
0
, t
1
, . . . , t
M
), where t
i
:= T

+j2
m
(T

) for all j = 0, 1, . . . , 2
m
.
Then the linear programming problem LP(f
p
: p T, [a, a]
n
, PS, t, T, | |) in
Denition 6.8 possesses a feasible solution, whenever m is large enough.
Proof. Note that by Theorem 5.4 assumption (ii) implies assumption (i). But then,
by the arguments already delivered in this Section 7, the propositions of the theorem
follow.
Note that we have, in this Section 7, actually proved substantially more than
stated in Theorem 7.1. Namely, we did derive formulae for the values of the pa-
rameters that are needed to initialize the linear programming problem in Denition
6.8. These formulae do depend on the unknown Lyapunov function W, so we can-
not extract the numerical values. However, these formulae are concrete enough to
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 71
derive the promised algorithm for generating a CPWA Lyapunov function. This
will be done in Section 8.
7.8. The autonomous case. The circumstances are close to identical when the
Switched System 3.5 is autonomous.
Theorem 7.2 (Converse theorem for autonomous switched systems).
Consider the Switched System 3.5 where T is a nite set and assume that it is
autonomous. Let a > 0 be a real-valued constant such that [a, a]
n
|, and
assume that at least one of the following two assumptions holds:
(i) There exists a Lyapunov function W (
2
([a, a]
n
0) for the Switched
System 3.5.
(ii) The origin is an asymptotically stable equilibrium point of the Switched
System 3.5, the set [a, a]
n
is contained in its region of attraction, and the
functions f
p
satisfy the Lipschitz condition: for every p T there exists a
constant L
P
such that
|f
p
(x) f
p
(y)| L
p
|x y|, for all x, y [a, a]
n
.
Then, for every neighborhood ^ [a, a]
n
of the origin, no matter how small, it is
possible to parameterize a time-invariant Lyapunov function V
Lya
of class CPWA,
V
Lya
: [a, a]
n
^ R,
for the Switched System 3.5 by using the linear programming problem from Deni-
tion 6.12.
More concretely: Let m be a positive integer and dene the piecewise scaling
function PS : R
n
R
n
and the set T of the linear programming problem through
PS(j
1
, j
2
, . . . , j
n
) := a2
m
(j
1
, j
2
, . . . , j
n
)
and
T :=] 2
k
a
2
m
, 2
k
a
2
m
[
n
^,
for some integer 1 k < m. Then, the linear programming problem LP(f
p
: p
T, [a, a]
n
, PS, T, | |) in Denition 6.12 possesses a feasible solution, whenever
m is large enough.
Proof. The proof is essentially a slimmed down version of the proof of Theorem
7.1, so we will not go very thoroughly into details.
First, note that by Theorem 5.4 assumption (ii) implies assumption (i), so in both
cases there are functions , , / and a function W (
2
([a, a]
n
0) R,
such that
(|x|) W(x) (|x|),
W(x) f
p
(x) (|x|)
for all x ] a, a[
n
0 and all p T. Further, by Lemma 3.17, we can assume
without loss of generality that and are convex functions. With
x

min
:= min
x=a
|x|
we set
:=
(x

min
)
2
72 S. F. HAFSTEIN EJDE-2007/MON. 08
and let m

be a strictly positive integer, such that


[
a
2
m

,
a
2
m

]
n
x R
n
: (|x|) ^
and set
T :=]
a
2
m

,
a
2
m

[
n
Set
x

:= min
x=a2
m

|x|,

:=
1
2
(x

),
C := max
i=1,2,...,n
x[a,a]
n
\D

W
x
i
(x)

,
and determine a constant B such that
B max
pP
i,r,s=1,2,...,n
x[a,a]
n


2
f
p,i
x
r
x
s
(x)

.
Assign
A

:= sup
pP
x[a,a]
n
|f
p
(x)|
2
,
B

:= n
3
2
max
r,s=1,2,...,n
x[a,a]
n
\D

2
W
x
r
x
s
(x)

,
C

:= n
3
BC,
and let m m

be an integer such that


a
2
m

_
(A

)
2
+ 4x

2C

and set d := 2
mm

. We dene the piecewise scaling function PS : R


n
R
n
through
PS(j
1
, j
2
, . . . , j
n
) := a2
m
(j
1
, j
2
, . . . , j
n
) for all (j
1
, j
2
, . . . , j
n
) Z
n
.
We assign the following values to the variables and the remaining constants of the
linear programming problem LP(f
p
: p T, [a, a]
n
, PS, T, | |):
B
(z,J)
rs
:= B, for all (z, ) Z
a
and all r, s = 1, 2, . . . , n,

a
[y] := (y), for all y A

a
[y] :=

x, for all y A

,
V
a
[x] := W(x) for all x (
a
,
C
a
[x, y] := C, for all x, y }
a
,
:= min

, (y
1
)/y
1
, where y
1
:= miny : y A

and y ,= 0.
That the linear constraints (LC1a), (LC2a), and (LC3a) are all satised follows
very similarly to how the linear constraints (LC1), (LC2), and (LC3) follow in the
nonautonomous case, so we only show that the constraints (LC4a) are fullled.
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 73
To do this let (z, ) Z
a
, Perm[1, 2, . . . , n], and i 1, 2, . . . , n + 1 be
arbitrary, but xed throughout the rest of the proof. We have to show that

a
[|y
(z,J)
,i
|]

j=1
V
a
[y
(z,J)
,j
] V
a
[y
(z,J)
,j+1
]
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)
f
p,(j)
(y
(z,J)
,i
) +E
(z,J)
p,,i
n

j=1
C
a
[y
(z,J)
,j
, y
(z,J)
,j+1
].
(7.5)
With the values we have assigned to the variables and the constants of the linear
programming problem, inequality (7.5) holds if

|y
(z,J)
,i
|
n

j=1
W[y
(z,J)
,j
] W[y
(z,J)
,j+1
]
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)
f
(j)
(y
(z,J)
,i
) +h
2
C

with h := a2
m
. Now, by the Mean-value theorem and because (x) 2

x for
all x x

,
n

j=1
W[y
(z,J)
,j
] W[y
(z,J)
,j+1
]
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)
f
(j)
(y
(z,J)
,i
) +h
2
C

=
n

j=1
_
W[y
(z,J)
,j
] W[y
(z,J)
,j+1
]
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)

W

(j)
(y
(z,J)
,i
)
_
f
(j)
(y
(z,J)
,i
)
+W(y
(z,J)
,i
) f (y
(z,J)
,i
) +h
2
C

_
_
_
n

j=1
_
W[y
(z,J)
,j
] W[y
(z,J)
,j+1
]
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)

W

(j)
(y
(z,J)
,i
)
_
e
j
_
_
_
2
|f
(j)
(y
(z,J)
,i
)|
2
(|y
(z,J)
,i
|) +h
2
C

hA

|y
(z,J)
,i
| +h
2
C

.
Hence, if

|y
(z,J)
,i
| hA

|y
(z,J)
,i
| +h
2
C

,
inequality (7.5) follows. But, this last inequality follows from
h :=
a
2
m

_
(A

)
2
+ 4x

2C

,
which implies
0 hA

+h
2
C

hA

|y
(z,J)
,i
| +h
2
C

In Section 8 we will use Theorem 7.1 to derive an algorithm for parameterizing


a CPWA Lyapunov function for the Switched System 3.5 and, if the Switched
System 3.5 is autonomous, we will use Theorem 7.2 to derive an algorithm for
parameterizing a time-invariant CPWA Lyapunov function for the system.
74 S. F. HAFSTEIN EJDE-2007/MON. 08
8. An Algorithm for Constructing Lyapunov Functions
In this section use the results from Theorem 7.1 and Theorem 7.2 to prove that
the systematic scan of the initiating parameters of the linear programming problem
from Denition 6.8 in Procedure 8.1 is an algorithm for constructing Lyapunov
functions for the Switched System 3.5, whenever one exists, and that Procedure 8.3
is an algorithm for constructing time-invariant Lyapunov functions for the Switched
System 3.5 if it is autonomous, again, whenever one exists. However, we rst give
a short discussion on algorithms, because we intend to prove that our procedure
to generate Lyapunov functions is concordant with the concept of an algorithm,
whenever the system in question possesses a Lyapunov function.
Donald Knuth writes in his classic work The Art of Computer Programming on
algorithms [29]:
The modern meaning for algorithm is quite similar to that of recipe,
process, method, technique, procedure, routine, rigmarole, except
that the word algorithm connotes something just a little dierent.
Besides merely being a nite set of rules that gives a sequence of
operations for solving a specic type of problem, an algorithm has
ve important features:
(1) Finiteness. An algorithm must always terminate in a nite
number of steps. [. . . ]
(2) Deniteness. Each step of an algorithm must be precisely
dened; the actions to be carried out must be be rigorously
and unambiguously specied for each case. [. . . ]
(3) Input. An algorithm has zero or more inputs: quantities that
are given to it initially before the algorithm begins, or dy-
namically as the algorithm runs. These inputs are taken from
specied sets of objects. [. . . ]
(4) Output. An algorithm has one or more outputs: quantities
that have a specied relation to the inputs. [. . . ]
(5) Eectiveness. An algorithm is also generally expected to be
eective, in the sense that its operations must all be suciently
basic that they can in principle be done exactly and in a nite
length of time by someone using pencil and paper.
The construction scheme for a Lyapunov function we are going to derive here
does comply to all of these features whenever the equilibrium at the origin is an
uniformly asymptotically stable equilibrium of the Switched System 3.5, and is
therefore an algorithm for constructing Lyapunov functions for arbitrary switched
systems possessing a uniformly asymptotically stable equilibrium.
8.1. The algorithm in the nonautonomous case. We begin by dening a pro-
cedure to construct Lyapunov functions and then we prove that it is an algorithm
for constructing Lyapunov functions for arbitrary switched systems possessing a
uniformly asymptotically stable equilibrium.
Procedure 8.1. Consider the Switched System 3.5 where T is a nite set, let a > 0
be a constant such that [a, a]
n
|, and let ^ | be an arbitrary neighborhood
of the origin. Further, let T

and T

be arbitrary real-valued constants such that


EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 75
0 T

< T

and let | | be an arbitrary norm on R


n
.
We assume that the components of the f
p
, p T, have bounded
second-order partial derivatives on [T

, T

] [a, a]
n
.
First, we have to determine a constant B such that
B max
pP, i,r,s=0,1,...,n
x[T

,T

][a,a]
n


2
f
p,i
x
r
x
s
( x)

.
The process has two integer variables that have to be initialized, namely m and N.
They should be initialized as follows: Set N := 0 and assign the smallest possible
positive integer to m such that
] a2
m
, a2
m
[
n
^.
The process consists of the following steps:
(i) Dene the piecewise scaling function PS : R
n
R
n
and the vector t :=
(t
0
, t
1
, . . . , t
2
m), through
PS(j
1
, j
2
, . . . , j
n
) := a2
m
(j
1
, j
2
, . . . , j
n
), for all (j
1
, j
2
, . . . , j
n
) Z
n
and
t
i
:= T

+i
T

2
m
, for i = 0, 1, . . . , 2
m
.
(ii) For every N

= 0, 1, . . . , N we do the following:
Generate the linear programming problem
LP(f
p
: p T, [a, a]
n
, PS, t, ] a2
N

m
, a2
N

m
[
n
, | |)
as dened in Denition 6.8 and check whether it possesses a feasible solu-
tion or not. If one of the linear programming problems possesses a feasible
solution, then go to step (iii). If none of them possesses a feasible solution,
then assign m := m+ 1 and N := N + 1 and go back to step i).
(iii) Use the feasible solution to parameterize a CPWA Lyapunov function for
the Switched System 3.5 as described in Section 6.3.
After all the preparation we have done, the proof that Procedure 8.1 is an algo-
rithm for constructing Lyapunov functions for arbitrary switched systems possess-
ing a uniformly asymptotically stable equilibrium is remarkably short.
Theorem 8.2 (Procedure 8.1 is an algorithm). Consider the Switched System 3.5
where T is a nite set and let a > 0 be a constant such that [a, a]
n
|.
We assume that the components of the f
p
, p T, have bounded
second-order partial derivatives on [T

, T

] [a, a]
n
for
every 0 T

< T

< +.
Assume further, that at least one of the following two assumptions holds:
(i) There exists a Lyapunov function W (
2
(R
0
([a, a]
n
0)) for the
Switched System 3.5.
(ii) The origin is a uniformly asymptotically stable equilibrium point of the
Switched System 3.5, the set [a, a]
n
is contained in its region of attraction,
and the functions f
p
satisfy the Lipschitz condition: for every p T there
exists a constant L
P
such that
|f
p
(s, x) f
p
(t, y)| L
p
([s t[ +|x y|),
76 S. F. HAFSTEIN EJDE-2007/MON. 08
for all s, t R
0
and all x, y [a, a]
n
.
Then, for every constants 0 T

< T

< + and every neighborhood ^ [a, a]


n
of the origin, no matter how small, the Procedure 8.1 delivers, in a nite number
of steps, a CPWA Lyapunov function V
Lya
,
V
Lya
: [T

, T

]
_
[a, a]
n
^
_
R,
for the Switched System 3.5.
Proof. Follows directly from what we have shown in Section 7. With the same
notations as there, the linear programming problem
LP(f
p
: p T, [a, a]
n
, PS, t, ] a2
N

m
, a2
N

m
[
n
, | |)
possesses a feasible solution, when m is so large that
maxT

, a
2
m

_
(A

)
2
+ 4x

2C

and N

, 0 N

N, is such that
[
a2
N

2
m
,
a2
N

2
m
]
n
x R
n
: (|x|) ^.

Because we have already proved in Theorem 4.2 and Theorem 5.4 that the
Switched System 3.5 possesses a Lyapunov function, if and only if an equilibrium of
the system is uniformly asymptotically stable, Theorem 8.2 implies the statement:
It is always possible, in a nite number of steps, to construct a
Lyapunov function for the Switched System 3.5 with the methods
presented in this monograph, whenever one exists.
8.2. The algorithm in the autonomous case. The procedure to construct Lya-
punov functions for autonomous systems mimics Procedure 6.8.
Procedure 8.3. Consider the Switched System 3.5 where T is a nite set, let a > 0
be a constant such that [a, a]
n
|, and let ^ | be an arbitrary neighborhood
of the origin. Further, assume that the system is autonomous and let | | be an
arbitrary norm on R
n
.
We assume that the components of the f
p
, p T, have bounded
second-order partial derivatives on [a, a]
n
.
First, we have to determine a constant B such that
B max
pP
i,r,s=1,2,...,n
x[a,a]
n


2
f
p,i
x
r
x
s
(x)

.
The procedure has two integer variables that have to be initialized, namely m and N.
They should be initialized as follows: Set N := 0 and assign the smallest possible
positive integer to m such that
] a2
m
, a2
m
[
n
^.
The procedure consists of the following steps:
(i) Dene the piecewise scaling function PS : R
n
R
n
through
PS(j
1
, j
2
, . . . , j
n
) := a2
m
(j
1
, j
2
, . . . , j
n
), for all (j
1
, j
2
, . . . , j
n
) Z
n
.
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 77
(ii) For every N

= 0, 1, . . . , N we do the following:
Generate the linear programming problem
LP(f
p
: p T, [a, a]
n
, PS, ] a2
N

m
, a2
N

m
[
n
, | |)
as dened in Denition 6.12 and check whether it possesses a feasible solu-
tion or not. If one of the linear programming problems possesses a feasible
solution, then go to step (iii). If none of them possesses a feasible solution,
then assign m := m+ 1 and N := N + 1 and go back to step i).
(iii) Use the feasible solution to parameterize a CPWA Lyapunov function for
the Switched System 3.5 as described in Denition 6.13.
The proof that Procedure 8.3 is an algorithm for constructing time-invariant
Lyapunov functions for arbitrary switched systems possessing an asymptotically
stable equilibrium is essentially identical to the proof of Theorem 8.2, where the
nonautonomous case is treated.
Theorem 8.4 (Procedure 8.3 is an algorithm). Consider the Switched System 3.5
where T is a nite set and assume that it is autonomous. Let a > 0 be a constant
such that [a, a]
n
|.
We assume that the components of the f
p
, p T, have bounded
second-order partial derivatives on [a, a]
n
.
Assume further, that at least one of the following two assumptions holds:
(i) There exists a time-invariant Lyapunov function W (
2
([a, a]
n
0)
for the Switched System 3.5.
(ii) The origin is an asymptotically stable equilibrium point of the Switched
System 3.5, the set [a, a]
n
is contained in its region of attraction, and the
functions f
p
satisfy the Lipschitz condition: for every p T there exists a
constant L
P
such that
|f
p
(x) f
p
(y)| L
p
|x y|, for all x, y [a, a]
n
.
Then, for every neighborhood ^ [a, a]
n
of the origin, no matter how small,
the Procedure 8.3 delivers, in a nite number of steps, a time-invariant Lyapunov
function V
Lya
of class CPWA,
V
Lya
: [a, a]
n
^ R,
for the autonomous Switched System 3.5.
Proof. Almost identical to the proof of Theorem 8.2. With the same notation as in
Section 7, the linear programming problem
LP(f
p
: p T, [a, a]
n
, PS, ] a2
N

m
, a2
N

m
[
n
, | |)
possesses a feasible solution, when m is so large that
a
2
m

_
(A

)
2
+ 4x

2C

and N

, 0 N

N, is such that
[
a2
N

2
m
,
a2
N

2
m
]
n
x R
n
: (|x|) ^.

78 S. F. HAFSTEIN EJDE-2007/MON. 08
Because we have already proved in Theorem 4.2 and Theorem 5.4 that the au-
tonomous Switched System 3.5 possesses a time-invariant Lyapunov function, if and
only if an equilibrium of the system is asymptotically stable, Theorem 8.4 implies
the statement:
It is always possible, in a nite number of steps, to construct a time-
invariant Lyapunov function for the autonomous Switched System
3.5 with the methods presented in this monograph, whenever one
exists.
9. Examples of Lyapunov functions generated by linear programming
In this section we give some examples of the construction of Lyapunov functions
by the linear programming problem LP(f
p
: p T, ^, PS, t, T, | |) from Def-
inition 6.8 and the linear programming problem LP(f
p
: p T, ^, PS, T, | |)
from Denition 6.12. In all the examples we will use the innity norm, that is
| | := | |

, in the linear programming problems. Further, we will use piece-


wise scaling functions PS, whose components PS
i
are all odd functions, that is
(recall that the i-th component PS
i
of PS does only depend on the i-th variable
x
i
of the argument x) PS
i
(x
i
) = PS
i
(x
i
). Because are only be interested in
the values of a piecewise scaling functions on compact subsets [m, m]
n
R
n
,
m N
>0
, this implies that we can dene such a function by specifying n vectors
ps
i
:= (ps
i,1
, ps
i,2
, . . . , ps
i,m
), i = 1, 2, . . . , n. If we say that the piecewise scaling
function PS is dened through the ordered vector tuple (ps
1
, ps
2
, . . . , ps
n
), we
mean that PS(0) := 0 and that for every i = 1, 2, . . . , n and every j = 1, 2, . . . , m,
we have
PS
i
(j) := ps
i,j
and PS
i
(j) := ps
i,j
.
If we say that the piecewise scaling function PS is dened through the vector ps, we
mean that it is dened trough the vector tuple (ps
1
, ps
2
, . . . , ps
n
), where ps
i
:= ps
for all i = 1, 2, . . . , n.
The linear programming problems were all solved by use of the GNU Linear
programming kit (GLPK), version 4.8, developed by Andrew Makhorin. It is a
free software that is available for download on the internet. The parameterized
Lyapunov functions were drawn with gnuplot, version 3.7, developed by Thomas
Williams and Colin Kelley. Just as GLPK, gnuplot is a free software that is available
for download on the internet. The author is indebted to these developers.
9.1. An autonomous system. As a st example of the use of the linear program-
ming problem from Denition 6.12 and Procedure 8.3 we consider the continuous
system
x = f (x), where f (x, y) :=
_
x
3
(y 1)

x
4
(1+x
2
)
2

y
1+y
2
_
. (9.1)
This system is taken from Example 65 in Section 5.3 in [60]. The Jacobian of f at
the origin has the eigenvalues 0 and 1. Hence, the origin is not an exponentially
stable equilibrium point (see, for example, Theorem 4.4 in [28] or Theorem 15 in
Section 5.5 in [60]). We initialize Procedure 8.3 with
a :=
8
15
and ^ :=]
2
15
,
2
15
[
2
.
Further, with
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 79
Figure 1. A Lyapunov function for the system (9.1) generated by
Algorithm 8.3.
Figure 2. A Lyapunov function for the system (9.1), parameter-
ized with the linear programming problem from Denition 6.12,
with a larger domain than the Lyapunov function on Figure 1.
x
(z,J)
:=

e
1
PS(R
J
(z +e
1
))

and y
(z,J)
:=

e
2
PS(R
J
(z +e
2
))

, (9.2)
80 S. F. HAFSTEIN EJDE-2007/MON. 08
we set (note that for the constants B
(z,J)
p,rs
the index p is redundant because the
system is non-switched)
B
(z,J)
11
:= 6x
(z,J)
(1 +y
(z,J)
),
B
(z,J)
12
:= 3x
2
(z,J)
,
B
(z,J)
22
:=
_
_
_
6y
(z,J)
(1+y
2
(z,J)
)
2

8y
3
(z,J)
(1+y
2
(z,J)
)
3
, if y
(z,J)

2 1,
1.46, else,
for all (z, ) Z in the linear programming problems. This is more eective
than using one constant B larger than all B
(z,J)
p,rs
for all (z, ) Z and all r, s =
1, 2, . . . , n, as done to shorten the proof of Theorem 7.1.
Procedure 8.1 succeeds in nding a feasible solution to the linear programming
problem with m = 4 and D = 2. The corresponding Lyapunov function of class
CPWA is drawn in Figure 1. We used this Lyapunov function as a starting point
to parameterize a CPWA Lyapunov function with a larger domain and succeeded
with ^ := [1, 1]
2
, T :=] 0.133, 0.133[
2
, and PS dened through the vector
ps := (0.033, 0.067, 0.1, 0.133, 0.18, 0.25, 0.3, 0.38, 0.45, 0.55, 0.7, 0.85, 0.93, 1)
as described at the beginning of Section 9. It is drawn on Figure 2.
Note that the domain of the Lyapunov function on Figure 1, where we used
the Procedure 8.3 to scan the parameters of the linear programming problem from
Denition 6.12, is much smaller than that of the Lyapunov function on Figure 2,
where we used another trial-and-error procedure to scan the parameters. This is
typical! The power of Procedure 8.3 and Theorem 8.4 is that they tell us that a
systematic scan will lead to a success if there exists a Lyapunov function for the
system. However, as Procedure 8.3 will not try to increase the distance between
the points in the grid ( of the linear programming problem far away from the
equilibrium, it is not particularly well suited to parameterize Lyapunov functions
with large domains. To actually parameterize Lyapunov functions a trial-and-
error procedure that rst tries to parameterize a Lyapunov function in a small
neighborhood of the equilibrium, and if it succeeds it tries to extend the grid with
larger grid-steps farther away from the equilibrium, is more suited.
In Figure 3 the sets T, T , and / from Lemma 6.16 are drawn for this particular
Lyapunov function. The innermost square is the boundary of T, the outmost gure
is the boundary of the set /, and in between the boundary of T is plotted. Every
solution to the system (9.1) with an initial value in / will reach the square
[0.133, 0.133]
2
in a nite time t

and will stay in the set T for all t t

.
9.2. An arbitrary switched autonomous system. Consider the autonomous
systems
x = f
1
(x), where f
1
(x, y) :=
_
y
x y(1 x
2
+ 0.1x
4
)
_
, (9.3)
x = f
2
(x), where f
2
(x, y) :=
_
y +x(x
2
+y
2
1)
x +y(x
2
+y
2
1)
_
, (9.4)
and
x = f
3
(x), (9.5)
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 81
Figure 3. The sets T, T , and / from Lemma 6.11 for the Lya-
punov function on Figure 2 for the system (9.1).
where f
3
(x, y) :=
_
1.5y
x
1.5
+y
_
_
x
1.5
_
2
+y
2
1
_
_
.
The systems (9.3) and (9.4) are taken from Exercise 1.16 in [28] and from page 194
in [54] respectively.
First, we used the linear programming problem from Denition 6.12 to param-
eterize a Lyapunov function for each of the systems (9.3), (9.4), and (9.5) individ-
ually. We dene x
(z,J)
and y
(z,J)
as in formula (9.2) and for the system (9.3) we
set
B
(z,J)
1,11
:= 2y
(z,J)
+ 1.2y
(z,J)
x
2
(z,J)
,
B
(z,J)
1,12
:= 2x
(z,J)
+ 0.4x
3
(z,J)
,
B
(z,J)
1,22
:= 0,
for the system (9.4) we set
B
(z,J)
2,11
:= max6x
(z,J)
, 2y
(z,J)
,
B
(z,J)
2,12
:= max2x
(z,J)
, 2y
(z,J)
,
B
(z,J)
2,22
:= max2x
(z,J)
, 6y
(z,J)
,
and for the system (9.5) we set
B
(z,J)
3,11
:=
8
9
y
(z,J)
,
B
(z,J)
3,12
:=
8
9
x
(z,J)
,
82 S. F. HAFSTEIN EJDE-2007/MON. 08
Figure 4. A Lyapunov function for the system (9.3) generated by
the linear programming problem from Denition 6.12.
B
(z,J)
3,22
:= 6y
(z,J)
.
Figure 5. A Lyapunov function for the system (9.4) generated by
the linear programming problem from Denition 6.12.
We parameterized a CPWA Lyapunov function for the system (9.3) by use of
the linear programming problem from Denition 6.12 with ^ := [1.337, 1.337]
2
,
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 83
T := , and PS dened trough the vector
ps := (0.0906, 0.316, 0.569, 0.695, 0.909, 1.016, 1.163, 1.236, 1.337)
as described at the beginning of Section 9. The Lyapunov function is depicted on
Figure 4.
We parameterized a CPWA Lyapunov function for the system (9.4) by use of
the linear programming problem from Denition 6.12 with ^ := [0.818, 0.818]
2
,
T := , and PS dened trough the vector
ps := (0.188, 0.394, 0.497, 0.639, 0.8, 0.745, 0.794, 0.806, 0.818)
as described at the beginning of Section 9. The Lyapunov function is depicted on
Figure 5.
We parameterized a CPWA Lyapunov function for the system (9.5) by use of
the linear programming problem from Denition 6.12 with ^ := [0.506, 0.506]
2
,
T :=] 0.01, 0.01[
2
, and PS dened trough the vector
ps := (0.01, 0.0325, 0.0831, 0.197, 0.432, 0.461, 0.506)
as described at the beginning of Section 9. The Lyapunov function is depicted on
Figure 6.
Figure 6. A Lyapunov function for the system (9.5) generated by
the linear programming problem from Denition 6.12.
Finally, we parameterized a CPWA Lyapunov function for the switched system
x = f
p
(x), p 1, 2, 3, (9.6)
where the functions f
1
, f
2
, and f
3
are, of course, the functions from (9.3), (9.4),
and (9.5), by use of the linear programming problem from Denition 6.12 with
^ := [0.612, 0.612]
2
, T :=] 0.01, 0.01[
2
, and PS dened trough the vector
ps := (0.01, 0.0325, 0.0831, 0.197, 0.354, 0.432, 0.535, 0.586, 0.612)
84 S. F. HAFSTEIN EJDE-2007/MON. 08
as described at the beginning of Section 9. The Lyapunov function is depicted on
Figure 7. Note, that this Lyapunov function is a Lyapunov function for all of the
systems (9.3), (9.4), and (9.5) individually.
The equilibriums region of attraction, secured by this Lyapunov function, and
the set T are drawn on Figure 8. Every solution to the system (9.6) that starts in
the larger set will reach the smaller set in a nite time.
9.3. A variable structure system. Consider the linear systems
x = A
1
x, where A
1
:=
_
0.1 1
2 0.1
_
and
x = A
2
x, where A
2
:=
_
0.1 2
1 0.1
_
.
These systems are taken from [33]. It is easy to verify that the matrices A
1
and A
2
both have the eigenvalues

= 0.1i

2. Therefore, by elementary linear stability


theory, the systems (9.3) and (9.3) are both unstable.
On Figure 9 the trajectories of the systems (9.3) and (9.3) with the initial value
(1, 0) are depicted. That the norm of the solutions is growing with t in the long
run is clear. However, it is equally clear, that the solution to (9.3) is decreasing on
the sets
Q
2
:= (x
1
, x
2
) : x
1
0 and x
2
> 0 and Q
4
:= (x
1
, x
2
) : x
1
0 and x
2
> 0
and that the solution to (9.3) is decreasing on the sets
Q
1
:= (x
1
, x
2
) : x
1
> 0 and x
2
0 and Q
3
:= (x
1
, x
2
) : x
1
< 0 and x
2
0.
Figure 7. A Lyapunov function for the arbitrary switched system
(9.6) generated by the linear programming problem from Denition
6.12.
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 85
Figure 8. The region of attraction secured by the Lyapunov func-
tion on Figure 7 for the switched system. All solution that start
in the larger set are asymptotically attracted to the smaller set at
the origin.
Figure 9. Trajectory of the system x = A
1
x (left) and of x = A
2
x
(right) starting at (1, 0).
Now, consider the switched system
x = A
p
x, p 1, 2, (9.7)
where the matrices A
1
and A
2
are the same as in (9.3) and (9.3). Obviously, this
system is not stable under arbitrary switching, but, if we only consider solution
trajectories (t, )

(t, ), such that

(t, ) Q
2
Q
4
, implies (t) = 1, and (9.8)

(t, ) Q
1
Q
3
, implies (t) = 2, (9.9)
86 S. F. HAFSTEIN EJDE-2007/MON. 08
then we would expect all trajectories under consideration to be asymptotically at-
tracted to the equilibrium. The switched system (9.7), together with the constraints
(9.8) and (9.9), is said to be a variable structure system. The reason is quite obvi-
ous, the structure of the right-hand side of the system (9.7) depends on the current
position in the state-space.
It is a simple task to modify the linear programming problem from Denition 6.12
to parameterize a Lyapunov function for the variable structure system. Usually,
one would include the constraint (LC4a), that is,

_
|y
(z,J)
,i
|

j=1
_
V [y
(z,J)
,j
] V [y
(z,J)
,j+1
]
e
(j)
(y
(z,J)
,j
y
(z,J)
,j+1
)

f
p,(j)
(y
(z,J)
,i
) +E
(z,J)
p,,i
C[y
(z,J)
,j
, y
(z,J)
,j+1
]
_
.
for every p T, every (z, ) Z, every Perm[1, 2, . . . , n], and every i =
1, 2, . . . , n + 1. In the modied linear programming problem however, we exclude
the constraints for some values of p, (z, ), , and i. It goes as follows:
(i) Whenever p = 2 and either = 1 or = 2, we do not include the
constraint (LC4a), for these particular values of p, (z, ), , and i, in the
linear programming problem.
(ii) Whenever p = 1 and either = or = 1, 2, we do not include the
constraints (LC4a), for these particular values of p, (z, ), , and i, in the
linear programming problem.
We parameterized a Lyapunov function for the variable structure system by use
of this modied linear programming problem with ^ := [1.152, 1.152]
2
, T :=
] 0.01, 0.01[
2
, and PS dened trough the vector
ps := (0.00333, 0.00667, 0.01, 0.0133, 0.0166, 0.0242,
Figure 10. A Lyapunov function for the variable structure system
(9.7) generated by an altered version of the linear programming
problem from Denition 6.12.
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 87
0.0410, 0.0790, 0.157, 0.319, 0.652, 1.152)
as described at the beginning of Section 9. The Lyapunov function V
Lya
is depicted
on Figure 10.
Now, one might wonder, what information we can extract from this function
V
Lya
, which is parameterized by our modied linear programming problem. Denote
by
a
the function that is constructed from the variables
a
[y
i
] as in Denition 6.13.
Then it is easy to see that for every
x
_
^ T
_

_
Q
1
Q
3
_
we have
limsup
h0+
V
Lya
(x +hA
2
x) V
Lya
(x)
h

a
(|x|

)
and for every x
_
^ T
_

_
Q
2
Q
4
_
, we have
limsup
h0+
V
Lya
(x +hA
1
x) V
Lya
(x)
h

a
(|x|

).
But this includes all trajectories of the system (9.7) that comply with the constraints
(9.8) and (9.9) so
limsup
h0+
V
Lya
(

(t +h, )) V
Lya
(

(t, ))
h

a
(|

(t, )|

)
for all

(t, ) in the interior of ^ T and all trajectories under consideration


and therefore V
Lya
is a Lyapunov function for the variable structure system. The
equilibriums region of attraction, secured by this Lyapunov function, is drawn on
Figure 11.
Figure 11. The region of attraction secured by the Lyapunov
function in Figure 10 for the variable structure system. All solution
that start in the larger set are asymptotically attracted to the
smaller set at the origin.
88 S. F. HAFSTEIN EJDE-2007/MON. 08
9.4. A variable structure system with sliding modes. Dene the matrix A
and the vector p through
A :=
_
0 1
1 0
_
and p :=
_
1
1
_
and consider the systems
x = f
1
(x), where f
1
(x) := Ax, (9.10)
x = f
2
(x), where f
2
(x) := p, (9.11)
x = f
3
(x), where f
3
(x) := p. (9.12)
The eigenvalues of the matrix A in (9.10) are

= 1 and the equilibrium at


the origin is therefore a saddle point of the system and is not stable. The systems
(9.11) and (9.12) do not even possess an equilibrium. Let the sets Q
1
, Q
2
, Q
3
, and
Figure 12. A Lyapunov function for the variable structure system
(9.13) generated by an altered version of the linear programming
problem from Denition 6.12.
Q
4
be dened as in the last example and consider the variable structure system
where we use the system (9.10) in Q
2
and Q
4
, the system (9.11) in Q
1
, and the
system (9.12) in Q
3
. A look the direction eld of the system (9.10) suggests that
this variable structure system might be stable, but the problem is that the system
does not possess a properly dened solution compatible with our solution concept
in Denition 3.4. The reason is that a trajectory, for example leaving Q
4
to Q
1
, is
sent straight back by the dynamics in Q
1
to Q
4
, where it will, of course, be sent
straight back to Q
1
. This phenomena is often called chattering and the sets x
R
2
: x
1
= 0 and x R
2
: x
2
= 0 are called the sliding modes of the dynamics. A
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 89
solution concept for such variable structure systems has been developed by Filippov
and others, see, for example [9], [10], and [55], or, for a brief review, [67].
Figure 13. The region of attraction secured by the Lyapunov
function in Figure 12 for the variable structure system. All solution
that start in the larger set are asymptotically attracted to the
smaller set at the origin.
Figure 14. A Lyapunov function for the variable structure system
(9.13) generated by an altered version of the linear programming
problem from Denition 6.12.
90 S. F. HAFSTEIN EJDE-2007/MON. 08
Even though Filippovs solution trajectories are supposed to be close to the true
trajectories if the switching is fast, we will use a more simple and more robust
technique here to prove the stability of the system. Our approach is very simple,
set h := 0.005 and dene the sets
o
1,2
:= x R
n
: [x
1
[ < h and x
2
> 0
o
2,3
:= x R
n
: x
1
< 0 and [x
2
[ < h
o
3,4
:= x R
n
: [x
1
[ < h and x
2
< 0
o
4,1
:= x R
n
: x
1
> 0 and [x
2
[ < h
T

:=] 2h, 2h[


2
.
We will generate a Lyapunov function with [0.957, 0.957]
2
as domain for the vari-
able structure system so we can consider T to be a small neighborhood of the origin
and the o
i,j
to be thin stripes between Q
i
and Q
j
.
We parameterized a Lyapunov function V
Lya
for the variable structure system
by use of a modied linear programming problem with ^ := [0, 957, 0, 957]
2
,
T := T

, and PS dened trough the vector


ps := (0.005, 0.01, 0.015, 0.0263, 0.052, 0.109, 0.237, 0.525, 0.957)
as described at the beginning of Section 9.
The modication we used to the linear programming problem in Denition 6.12,
similar to the modication in the last example, was to only include the constraints
(LC4a) in the linear programming problem for some sets of parameters
p T, (z, ) Z, Perm[1, 2, . . . , n], i 1, 2, . . . , n + 1.
Exactly, for every simplex S in the simplicial partition, S T = , we included
the constraints (LC4a) for every vertex of the simplex, if and only if:
S Q
1

_
o
1,2
o
4,1
_
and p = 2,
S Q
2

_
o
1,2
o
2,3
_
_
Q
4

_
o
4,1
o
3,4
_
and p = 1,
S Q
3

_
o
2,3
o
3,4
_
and p = 3,
S o
1,2
o
4,1
and
_
p = 1 or p = 2
_
,
S o
2,3
o
3,4
and
_
p = 1 or p = 3
_
.
This implies for the Lyapunov function V
Lya
, where the function
a
/ is con-
structed from the variables
a
[y
i
] as in Denition 6.13, that:
V1) For every x in the interior of the set
_
Q
1
o
1,2
o
4,1
_
T we have
limsup
h0+
V
Lya
(x +hf
2
(x)) V
Lya
(x)
h

a
(|x|

).
V2) For every x in the interior of the set
_
Q
2
Q
4
o
1,2
o
2,3
o
3,4
o
4,1
_
T,
we have
limsup
h0+
V
Lya
(x +hf
1
(x)) V
Lya
(x)
h

a
(|x|

).
V3) For every x in the interior of the set
_
Q
3
o
2,3
o
3,4
_
T, we have
limsup
h0+
V
Lya
(x +hf
3
(x)) V
Lya
(x)
h

a
(|x|

).
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 91
Now, let f
1
, f
2
, and f
3
be the functions from the systems (9.10), (9.11), and
(9.12) and consider the variable structure system
x = f
p
(x), p 1, 2, 3, (9.13)
under the following constraints:
(i)

(t, ) ^ T and

(t, ) in the interior of Q


1

_
o
1,2
o
4,1
_
implies (t) = 2.
(ii)

(t, ) ^ T and

(t, ) in the interior of


Q
2

_
o
1,2
o
2,3
_
Q
4

_
o
4,1
o
3,4
_
implies (t) = 1.
(iii)

(t, ) ^ T and

(t, ) in the interior of Q


3

_
o
2,3
o
3,4
_
implies (t) = 3.
(iii)

(t, ) ^ T and

(t, ) in the interior of o


1,2
o
4,1
implies
(t) 1, 2.
(iii)

(t, ) ^ T and

(t, ) in the interior of o


2,3
o
3,4
implies
(t) 1, 3.
One should make one self clear what these constraints imply. For example, if
Q
2

_
o
1,2
o
2,3
_
, then we must use the dynamics x = Ax until t

(t, )
leaves Q
2

_
o
1,2
o
2,3
_
. If then, for example,

(t

, ) o
1,2
for some t

> 0, then
every switching between the systems x = Ax and x = p is allowed as long as
t

(t, ) stays in o
1,2
. However, if, for example,

(t

, ) Q
1

_
o
1,2
o
4,1
_
for some t

> t

, then we must use the dynamics x = p until t

(t, x) leaves
Q
1

_
o
1,2
o
4,1
_
.
By V1, V2, and V3 we have for every trajectory t

(t, ) under consideration


that
limsup
h0+
V
Lya
(

(t +h, )) V
Lya
(

(t, ))
h

a
(|

(t, )|

),
so the function V
Lya
is a Lyapunov function for this system.
The parameterized Lyapunov function V
Lya
for the system (9.13) is depicted
on Figure 12 and its region of attraction on Figure 13. Because it is dicult to
recognize the structure of the Lyapunov function close to the origin, a Lyapunov
function for the same system, but with a much smaller domain, is depicted on
Figure 14.
9.5. A one-dimensional nonautonomous switched system. Consider the one-
dimensional systems
x = f
1
(t, x), where f
1
(t, x) :=
x
1 +t
(9.14)
and
x = f
2
(t, x), where f
2
(t, x) :=
tx
1 +t
. (9.15)
The system (9.14) has the closed-form solution
(t, t
0
, ) =
1 +t
0
1 +t
and the system (9.15) has the closed-form solution
(t, t
0
, ) = e
(tt0)
1 +t
1 +t
0
.
92 S. F. HAFSTEIN EJDE-2007/MON. 08
Figure 15. A Lyapunov function for the nonautonomous system
(9.14) generated by the linear programming problem from Deni-
tion 6.8.
The origin in the state-space is therefore, for every xed t
0
, an asymptotically
stable equilibrium point of the system (9.14) and, because
[[e
(tt0)
1 +t
1 +t
0
2[[e

tt
0
2
,
a uniformly exponentially stable equilibrium point of the system (9.15). However,
as can easily be veried, it is not a uniformly asymptotically stable equilibrium
point of the system (9.14). This implies that the system (9.14) cannot possess a
Lyapunov function that is dened for all t 0. Note however, that this does not
imply that we cannot parameterize a Lyapunov-like function on a compact time
interval for the system (9.14).
We set
t
(z,J)
:= e
0


PS(z) and x
(z,J)
:= [e
1


PS(R
J
(z +e
1
))[
and dene the constants B
(z,J)
p,rs
from the linear programming problem from Deni-
tion 6.8 by
B
(z,J)
p,00
:=
2x
(z,J)
(1 +t
(z,J)
)
3
,
B
(z,J)
p,01
:=
1
(1 +t
(z,J)
)
2
,
B
(z,J)
p,11
:= 0
for p 1, 2. We parameterized a CPWA Lyapunov function for the system (9.14),
the system (9.15), and the switched system
x = f
p
(t, x), p 1, 2 (9.16)
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 93
Figure 16. A Lyapunov function for the nonautonomous system
(9.15) generated by the linear programming problem from Deni-
tion 6.8.
by use of the linear programming problem from Denition 6.8 with ^ :=]1.1, 1.1[,
T :=] 0.11, 0.11[, PS dened through the vector
ps := (0.11, 0.22, 0.33, 0.44, 0.55, 0.66, 0.77, 0.88, 0.99, 1.1)
as described at the beginning of Section 9, and the vector
t := (0, 0.194, 0.444, 0.75, 1.111, 1.528, 2, 2.528, 3.111, 3.75, 4.444, 5.194, 6,
6.861, 7.778, 8.75, 9.778, 10.861, 12, 13.194, 14.444, 15.75, 17.111, 18.528, 20).
The Lyapunov function for the system (9.14) is depicted on Figure 15, the Lyapunov
function for the system (9.15) on Figure 15, and the Lyapunov function for the
arbitrary switched system (9.16) on Figure 9.16.
9.6. A two-dimensional nonautonomous switched system.
Consider the two-dimensional systems
x = f
1
(t, x), where f
1
(t, x, y) :=
_
2x +y cos(t)
xcos(t) 2y
_
(9.17)
and
x = f
2
(t, x), where f
2
(t, x, y) :=
_
2x +y sin(t)
xsin(t) 2y
_
. (9.18)
We set
x
(z,J)
:= [e
1


PS(R
J
(z +e
1
))[ and y
(z,J)
:= [e
2


PS(R
J
(z +e
2
))[
and assign values to the constants B
(z,J)
p,rs
from the linear programming problem in
Denition 6.8 as follows:
B
(z,J)
p,00
:= maxx
(z,J)
, y
(z,J)
,
94 S. F. HAFSTEIN EJDE-2007/MON. 08
Figure 17. A Lyapunov function for the switched nonautono-
mous system (9.16) generated by the linear programming problem
from Denition 6.8.
Figure 18. The function (x, y) V (2, x, y), where V (t, x, y) is
the parameterized Lyapunov function for the nonautonomous sys-
tem (9.17).
B
(z,J)
p,11
:= 0,
B
(z,J)
p,22
:= 0,
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 95
Figure 19. The function (x, y) V (2, x, y), where V (t, x, y) is
the parameterized Lyapunov function for the nonautonomous sys-
tem (9.18).
B
(z,J)
p,01
:= 1,
B
(z,J)
p,02
:= 1,
B
(z,J)
p,12
:= 0
for p 1, 2. We parameterized a Lyapunov function for the system (9.17), the
system (9.18), and the switched system
x = f
p
(t, x), p 1, 2 (9.19)
by use of the linear programming problem from Denition 6.8 with
^ :=] 0.55, 0.55[
2
, T :=] 0.11, 0.11[
2
, PS dened through the vector
ps := (0.11, 0.22, 0.33, 0.44, 0.55)
as described at the beginning of Section 9, and the vector
t := (0, 0.3125, 0.75, 1.3125, 2).
Because the Lyapunov functions are functions from R R
2
into R it is hardly
possible to draw them in any sensible way on a two-dimensional sheet. Therefore,
we only draw them exemplary for the xed time-value t := 2. On gures 18, 19,
and 20, the state-space dependency of the parameterized Lyapunov functions for
the systems (9.17), (9.18), and (9.19) respectively are depicted.
10. Conclusion
In this monograph we developed an algorithm for constructing Lyapunov func-
tions for nonlinear, nonautonomous, arbitrary switched continuous systems possess-
ing a uniformly asymptotically stable equilibrium. The necessary stability theory
96 S. F. HAFSTEIN EJDE-2007/MON. 08
5
NNS3-5.jpg
Figure 20. The function (x, y) V (2, x, y), where V (t, x, y)
is the parameterized Lyapunov function for the nonautonomous
switched system (9.19).
of switched systems, including a converse Lyapunov theorem for arbitrary switched
nonlinear, nonautonomous systems possessing a uniformly asymptotically stable
equilibrium (Theorem 5.4), was developed in the sections 3, 4, and 5. In the sec-
tions 6, 7, and 8 we presented a linear programming problem in Denition 6.8 that
can be constructed from a nite set of nonlinear and nonautonomous dierential
equations x = f
p
(t, x), p T, where the components of the f
p
are (
2
, and we proved
that every feasible solution to the linear programming problem can be used to pa-
rameterize a common Lyapunov function for the systems. Further, we proved that
if the origin in the state-space is a uniformly asymptotically stable equilibrium of
the switched system x = f

(t, x), : R
0
T, then Procedure 8.1, which uses the
linear programming problem from Denition 6.8, is an algorithm for constructing
a Lyapunov function for the switched system. Finally, in Section 9, we gave several
examples of Lyapunov functions that we generated by use of the linear program-
ming problem. Especially, we generated Lyapunov functions for variable structure
systems with sliding modes.
It is the belief of the author that this work is a considerable advance in the Lya-
punov stability theory of dynamical systems and he hopes to have convinced the
reader that the numerical construction of Lyapunov functions, even for arbitrary
switched, nonlinear, nonautonomous, continuous systems, is not only a theoretical
possibility, but is capable of being developed to a standard tool in system analysis
software in the near future. Thus, the new algorithm presented in this monograph
should give system engineers a considerable advantage in comparison to the tradi-
tional approach of linearization and pure local analysis.
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 97
List of Symbols
R the set of real numbers
R
0
the real-numbers larger than or equal to zero
R
>0
the real-numbers larger than zero
Z the integers
Z
0
the integers larger than or equal to zero
Z
>0
the integers larger than zero
/
n
set of n-tuples of elements belonging to a set /
R
n
the n-dimensional Euclidean space, n N
>0
/ the topological closure of a set / R
n
R R := R +
/ the boundary of a set /
dom(f) the domain of a function f
f(|) the image of a set | under a mapping f
f
1
(|) the preimage of a set | with respect to a mapping f
((|) continuous real-valued functions with domain |
(
k
(|) k-times continuously dierentiable real-valued functions
with domain |
[(
k
(|)]
n
vector elds f = (f
1
, f
2
, . . . , f
n
) of which f
i
(
k
(|)
for i = 1, 2, . . . , n
/ strictly monotonically increasing functions on [0, +[
vanishing at the origin
L strictly monotonically decreasing functions on [0, +[,
approaching zero at innity
P(/) the power set of a set /
Perm[/] the permutation group of /, i.e., the set of all bijective functions
/ /
con/ the convex hull of a set /
graph(f) the graph of a function f
e
i
the i-th unit vector
x y the inner product of vectors x and y
|x|
p
p-norm of a vector x, |x|
p
:=
_
i
[x
i
[
p
_1
p
if 1 p < +
and |x|

:= max
i
[x
i
[
f

the derivative of a function f


x the time-derivative of a vector-valued function x
f the gradient of a scalar eld f : R
n
R
f the Jacobian of a vector eld f : R
m
R
n

A
the characteristic function of a set /

ij
the delta Kronecker, equal to 1 if i = j and equal to 0 if i ,= j
[a, b] [a, b] := x R[a x b
]a, b] ]a, b] := x R[a < x b
supp(f) supp(f) := x R
n
: f(x) ,= 0
R
J
reection function with respect to the set 1, 2, . . . , n,
see Denition 6.2
PS piecewise scaling function, see page 36
S[PS, ^] a simplicial partition of a the set PS(^) R
n
,
see Denition 6.5
f
p,i
the i-th component of the vector eld f
p
B
,R
B
,R
:= x R
n
: |x| < R
o
P
the set of all switching signals R
0
T, see Denition 3.3
x = f

(t, x) arbitrary switched system, see Switched System 3.5


98 S. F. HAFSTEIN EJDE-2007/MON. 08
References
[1] R. Adams. Sobolev Spaces. Acad. Press, New York, 1975.
[2] A. Agrachev and D. Liberzon. Stability of switched systems: a Lie-algebraic condition. Siam
Journal on Control and Optimization, 40:253269, 2001.
[3] N. Bhatia and G. Szego. Stability Theory of Dynamical Systems. Springer, 1970.
[4] R. Brayton and C. Tong. Stability of dynamical systems: a constructive approach. IEEE
Transactions on Circuits and Systems, 26:224234, 1979.
[5] R. Brayton and C. Tong. Constructive stability and asymptotic stability of dynamical sys-
tems. IEEE Transactions on Circuits and Systems, 27:11211130, 1980.
[6] F. Camilli, L. Gr une, and F. Wirth. A generalization of Zubovs method to perturbed systems.
SIAM Journal of Control and Optimization, 40(2):496515, 2002.
[7] W. Dayawansa and C. Martin. A Converse Lyapunov Theorem for a Class of Dynamical
Systems which Undergo Switching. IEEE Transactions on Automatic Control, 44:751760,
1999.
[8] U. Dini. Fondamenti per la teoria delle funzioni di variabili reali. Pisa, 1878, (in Italian).
[9] A. Filippov. Dierential equations with discontinuous right-hand side. American Mathemat-
ical Society Translations, 42:191231, 1964.
[10] A. Filippov. Dierential equations with second member discontinuous on intersecting surfaces.
Dierential Equations, 415:12921299, 1980.
[11] P. Giesl. Construction of global Lyapunov functions using radial basis functions. Habilitation
Thesis: TU M unchen, Germany, 2004.
[12] P. Giesl. Construction of Global Lyapunov Functions Using Radial Basis Functions. Springer,
2007.
[13] P. Giesl. Construction of global Lyapunov functions using radial basis functions with a single
operator. Discrete and Continuous Dynamical Systems - Series B, 7:101124, 2007.
[14] P. Giesl. On the determination of the basin of attraction of a periodic orbit in two-dimensional
systems. Discrete and Continuous Dynamical Systems - Series B, 7:101124, 2007.
[15] I. Gurvits. Stability of discrete linear inclusion. Linear Algebra and its Applications, 231:47
85, 1995.
[16] S. Hafstein. A constructive converse Lyapunov theorem on asymptotically stable dierential
equations stability. University Duisburg-Essen, Schriftenreihe des Instituts f ur Mathematik,
SM-DU-576, 2004.
[17] S. Hafstein. A constructive converse Lyapunov theorem on exponential stability. Discrete and
Continuous Dynamical Systems - Series A, 10-3:657678, 2004.
[18] S. Hafstein. A constructive converse Lyapunov theorem on asymptotic stability for nonlinear
autonomous ordinary dierential equations. Dynamical Systems: An International Journal,
20:281299, 2005.
[19] W. Hahn. Stability of Motion. Springer, New York, 1967.
[20] M. Hirsch, S. Smale, and R. Devaney. Dierential Equations, Dynamical Systems & An
Introduction to Chaos. Elsevier, 2004.
[21] T. Johansen. Computation of Lyapunov Functions for Smooth Nonlinear Systems using Con-
vex Optimization. Automatica, 36:16171626, 2000.
[22] M. Johansson. Piecewise Linear Control Systems. Ph.D. Thesis: Lund Institute of Technol-
ogy, Lund, Sweden, 1999.
[23] M. Johansson and A. Rantzer. Computation of piecewise quadratic Lyapunov functions for
hybrid systems. In Proceedings of the European Control Conference ECC97, 1997.
[24] M. Johansson and A. Rantzer. On the computation of piecewise quadratic Lyapunov func-
tions. In Proceedings of the 36th IEEE Conference on Decision and Control, 1997.
[25] P. Julian. A High Level Canonical Piecewise Linear Representation: Theory and Applica-
tions. Ph.D. Thesis: Universidad Nacional del Sur, Bahia Blanca, Argentina, 1999.
[26] P. Julian, A. Desages, and O. Agamennoni. High-level canonical piecewise linear representa-
tion using a simplicial partition. IEEE Transactions on Circuits and Systems, 46:463480,
1999.
[27] P. Julian, J. Guivant, and A. Desages. A parametrization of piecewise linear Lyapunov func-
tion via linear programming. International Journal of Control, 72:702715, 1999.
[28] H. Khalil. Nonlinear Systems. Macmillan, New York, 1992.
EJDE-2007/MON. 08 LYAPUNOV FUNCTIONS 99
[29] D. Knuth. The Art of Computer Programming, Volume 1, Fundamental Algorithms. Addison
Wesley, 3. edition, 1997.
[30] X. Koutsoukos and P. Antsaklis. Design of Stabilizing Switching Control Laws for Discrete and
Continuous-Time Linear Systems Using Piecewise-Linear Lyapunov Functions. International
Journal of Control, 75:932945, 2002.
[31] D. Liberzon. Switching in Systems and Control. Birkhaeuser, 2003.
[32] D. Liberzon, J. Hespanha, and A. Morse. Stability of switched systems: a Lie-algebraic
condition. Systems and Control Letters, 37:117122, 1999.
[33] D. Liberzon and A. Morse. Basic problems in stability and design of switched systems. IEEE
Control Systems Magazine, 19-5:5970, 1999.
[34] D. Liberzon and R. Tempo. Common Lyapunov Functions and Gradient Algorithms. IEEE
Transactions on Automatic Control, 49:990994, 2004.
[35] Y. Lin, E. Sontag, and Y. Wang. A Smooth Converse Lyapunov Theorem for Robust Stability.
SIAM Journal on Control and Optimization, 34:124160, 1996.
[36] A. Lyapunov. The general problem of the stability of motion. International Journal of Con-
trol, 55:531773, 1992, (English translation of the original Russian text).
[37] I. Malkin. On a question of reversability of Liapunovs theorem on asymptotic stability. In
J. Aggarwal and M. Vidyasagar, editors, Nonlinear Systems: Stability Analysis, pages 161
170. Stroudsburg:Dowden, Hutchinson & Ross, 1977.
[38] J. Mancilla-Aguilar and R. Garcia. A converse Lyapunov theorem for nonlinear switched
systems. System & Control Letters, 41:6771, 2000.
[39] S. Marinosson. Lyapunov function construction for ordinary dierential equations with linear
programming. Dynamical Systems: An International Journal, 17:137150, 2002.
[40] S. Marinosson. Stability Analysis of Nonlinear Systems with Linear Programming A Lya-
punov Functions Based Approach. Ph.D. Thesis: Gerhard-Mercator-University, Duisburg,
Germany, 2002.
[41] C. Martin and W. Dayawansa. On the existence of a Lyapunov function for a family of
switching systems. In Proceedings of the 35th IEEE Conference on Decision and Control,
1996.
[42] J. Massera. On Liapunos conditions of stability. Annals of Mathematics, 50:705721, 1949.
[43] J. Massera. Contributions to stability theory. Annals of Mathematics, 64-1:182206, 1956.
[44] A. Michel, B. Nam, and V. Vittal. Computer generated Lyapunov functions for intercon-
nected systems: Improved Results with Applications to Power Systems. IEEE Transactions
on Circuits and Systems, CAS-31:189198, 1984.
[45] A. Michel, N. Sarabudla, and R. Miller. Stability analysis of complex dynamical systems:
some computational methods. Circuits and Systems Signal Processing, 1:171202, 1982.
[46] Y. Mori, T. Mori, and Y. Kuroe. A Solution to the Common Lyapunov Function Problem
for Continuous-Time System. In Proceedings of the 36th IEEE Conference on Decision and
Control, 1997.
[47] T. Nadzieja. Construction of a smooth Lyapunov function for an asymptotically stable set.
Czechoslovak Mathematical Journal, 40 (115):195199, 1990.
[48] K. Narendra and J. Balakrishnan. A common Lyapunov functions for stable LTI systems with
commuting A-matrices. IEEE Transactions on Automatic Control, 39:24692471, 1994.
[49] Y. Ohta, H. Imanishi, L. Gong, and H. Haneda. Computer generated Lyapunov functions
for a class of non-linear systems. IEEE Transactions on Circuits and Systems, 40:343354,
1993.
[50] A. Papachristodoulou and S. Prajna. The Construction of Lyapunov Functions using the
Sum of Squares Decomposition. In Proceedings of the 41st IEEE Conference on Decision
and Control, pages 34823487, 2002.
[51] P. Parrilo. Structured Semidenite Programs and Semialgebraic Geometry Methods in Ro-
bustness and Optimization. Ph.D. Thesis: Caltech, Pasadena, USA, 2000.
[52] A. Polanski. Lyapunov function construction by linear programming. IEEE Transactions on
Automatic Control, 42:10131016, 1997.
[53] C. Roos, T. Terlaky, and J. Vial. Theory and Algorithms for Linear Optimization. Wiley,
1997.
[54] S. Sastry. Nonlinear Systems: Analysis, Stability, and Control. Springer, New York, 1999.
[55] D. Schevitz and B. Paden. Lyapunov stability theory on non-smooth systems. IEEE Trans-
actions on Automatic Control, 39:19101914, 1994.
100 S. F. HAFSTEIN EJDE-2007/MON. 08
[56] A. Schrijver. Theory of Linear and Integer Programming. Wiley, 1983.
[57] H. Shim, D. Noh, and Seo J. Common Lyapunov function for exponentially stable nonlinear
systems. In 4th SIAM Conference on Control & its Applications, pages 12181223, 1998.
[58] J.-J. Slotline and W. Li. Applied Nonlinear Control. Prentice-Hall, New Jersey, 1991.
[59] L. Vandenberghe and S. Boyd. A polynominal-time algorithm for determing quadratic Lya-
punov functions for nonlinear systems. In Proceedings of the European Conference on Circuit
Theory and Design, pages 10651068, 1993.
[60] M. Vidyasagar. Nonlinear System Analysis. Prentice Hall, Englewood Clis, 2. edition, 1993.
[61] L. Vu and D. Liberzon. Common Lyapunov functions for families of commuting nonlinear
systems. System & Control Letters, 54:405416, 2005.
[62] W. Walter. Analysis I. Springer, Berlin, 2. edition, 1990.
[63] W. Walter. Ordinary Dierential Equations. Springer, New York, 1. edition, 1991.
[64] W. Walter. Einf uhrung in die Theorie der Distributionen. BI-Wiss.-Verl., Mannheim, 3.
edition, 1994, (in German).
[65] J. Willems. Stability Theory of Dynamical Systems. Nelson, 1970.
[66] F. Wilson. Smoothing derivatives of functions and applications. Translations of the American
Mathematical Society, 139:413428, 1969.
[67] Q. Wu, S. Onyshko, N. Sepehri, and A. Thornton-Trump. On construction of smooth Lya-
punov functions for non-smooth systems. International Journal of Control, 69-3:443457,
1998.
[68] V. Zubov. Methods of A. M. Lyapunov and their Application. P. Noordho, Groningen,
Netherlands, 1964.
Sigurdur Freyr Hafstein
School of Science and Engineering, Reykjavik University, Reykjavik, Iceland
E-mail address: sigurdurh@ru.is

Вам также может понравиться