IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. XX, NO. Y, MONTH 1999
1
Asymptotic Controllability Implies Feedback Stabilization F.H. Clarke Yu.S. Ledyaev E.D. Sontag A.I. Subbotin
Abstract |
It is shown that every asymptotically controllable system can be globally stabilized by means of some (discontinuous) feedback law. The stabilizing strategy is based on pointwise optimization of a smoothed version of a control-Lyapunov function, iteratively sending trajectories into smaller and smaller neighborgoods of a desired equilibrium. A major technical problem, and one of contributions of the present paper, concerns the precise meaning of \solution" when using a discontinuous controller. I. Introduction
A longstanding open question in nonlinear control theory concerns the relationship between asymptotic controllability to the origin in Rn of a nonlinear system
x_ = f (x; u)
(1)
by an \open loop" control u : [0; +1) ! U and the existence of a feedback control k : Rn ! U which stabilizes trajectories of the system
x_ = f (x; k(x))
(2)
with respect to the origin. For the special case of linear control systems x_ = Ax + Bu, this relationship is well understood: asymptotic controllability is equivalent to the existence of a continuous (even linear) stabilizing feedback law. But it is well-known that continuous feedback laws may fail to exist even for simple asymptotically controllable nonlinear systems. This is especially easy to see, as discussed in [27], for onedimensional (U=R, n=1) systems (1): in that case asymptotic controllability is equivalent to the property \for each F.H. Clarke is with Universite' de Lyon I (BAT 101), 69622 Villeurbanne, France, and CRM, Universite de Montreal, Montreal (Qc) Canada H3C 3J7; email:
[email protected]. His work was supported in part by the Natural Sciences and Engineering Research Council of Canada and Le Fonds FCAR du Quebec Yu.S. Ledyaev was visiting Rutgers University when this work was done. His permanent address is Steklov Institute of Mathematics, Moscow 117966, Russia; email:
[email protected]. His work was supported in part by Russian Fund for Fundamental Research Grant 96-01-00219, by NSERC and La Fonde FCAR du Quebec, and by the Rutgers Center for Systems and Control (SYCON) E.D. Sontag is with Dept. of Mathematics, Rutgers University, New Brunswick, NJ 08903, USA; email:
[email protected]. His work was supported in part by US Air Force Grant AFOSR-94-0293 A.I. Subbotin is with Inst. of Mathematics & Mechanics, Ekaterinburg 620219, Russia; email:
[email protected]. His work was supported in part by Russian Fund for Fundamental Research Grant 96-01-00219
x 6= 0 there is some value u so that xf (x; u) < 0", but it is easy to construct examples of functions f , even ana-
lytic, for which this property is satis ed but for which no possible continuous section k : R n f0g ! R exists so that xf (x; k(x)) < 0 for all nonzero x. General results regarding the nonexistence of continuous feedback were presented in the paper [3], where techniques from topological degree theory were used (an exposition is given in the textbook [25]). These negative results led to the search for feedback laws which are not necessarily of the form u = k(x), k a continuous function. One possible approach consists of looking for dynamical feedback laws, where additional \memory" variables are introduced into a controller, and as a very special case, time-varying (even periodic) continuous feedback u = k(t; x). Such time-varying laws were shown in [27] to be always possible in the case of one-dimensional systems, and in the major work [9] (see also [10]) it was shown that they are also always possible when the original system is completely controllable and has \no drift", meaning essentially that f (x; 0) = 0 for all states (see also [26] for numerical algorithms and an alternative proof of the time-varying result for analytic systems). However, for the general case of asymptotically controllable systems with drift, no dynamic or time-varying solutions are known. Thus it is natural to ask about the existence of discontinuous feedback laws u = k(x). Such feedbacks are often obtained when solving optimal-control problems, for example, so it is interesting to search for general theorems insuring their existence. Unfortunately, allowing nonregular feedback leads to an immediate diculty: how should one de ne the meaning of solution x() of the dierential equation (2) with discontinuous right-hand side? One of the best-known candidates for the concept of solution of (2) is that of a Filippov solution (cf. [13]), which is de ned as the solution of a certain dierential inclusion with multivalued right-hand side which is built from f (x; k(x)). However, it follows from the results in [21], [11] that the existence of a discontinuous stabilizing feedback in the Filippov sense implies the same Brockett necessary conditions [3] as the existence of a continuous stabilizing feedback does. Moreover, it is shown in [11] that the existence of a stabilizing feedback in the Filippov sense is equivalent to the existence of a continuous stabilizing one, in the case of systems ane in controls. In conclusion, there is no hope of obtaining general results if one insists on the use of Filippov
2
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. XX, NO. Y, MONTH 1999
solutions. In this paper, we develop a concept of solution of (2) for arbitrary feedback k(x) which has (a) a clear and reasonable physical meaning (perhaps even more so than the de nitions derived from dierential inclusions), and (b) allows proving the desired general theorem. Our notion is borrowed from the theory of positional dierential games, and it was systematically studied in that context by Krasovskii and Subbotin in [19]. There have been several papers dealing with rather general theorems on discontinuous stabilization. One of the best known is [30], which provided piecewise analytic feedback laws for analytic systems which satisfy controllability conditions. The de nition of \feedback" given in that paper involves a speci cation of \exit rules" for certain lower-dimensional submanifolds, and these cannot be expressed in terms of a true feedback law (even in the sense of this paper). Sampling is a strategy commonly used in digital control (see e.g. [25] for a discussion of sampling in a general nonlinear systems context), and an approach to sampled control of nonlinear systems was given in [22]. Sampling is not true feedback, in that one typically uses a xed sampling rate, or perhaps a predetermined sampling schedule, and intersample behavior is not accounted for. A stabilization approach introduced in [16] is based on a sampling-like method under strong controllability Lie algebraic conditions on the system being controlled. One may interpret our solutions as representing the behavior of sampling, with a xed feedback law being used, as the sampling periods tend to zero { indeed, such a speedup of sampling is essential as we approach the target state, to avoid an overshoot during the sampling interval, as well as far from the target, due to possible explosion times in the dynamics. A. De nition of Feedback Solution From now on, we assume that U is a locally compact metric space and that the mapping f : Rn U ! Rn : (x; u) 7! f (x; u) is continuous, and locally Lipschitz on x uniformly on compact subsets of Rn U. We use jxj to denote the usual Euclidean norm of x 2 Rn , and hx; yi for the inner product of two such vectors. A locally bounded (for each compact, image is relatively compact) function k : Rn ! U will be called a feedback . Any in nite sequence = fti gi0 consisting of numbers
0 = t 0 < t1 < t2 < : : : with limi!1 ti = 1 is called a partition (of [0; +1)), and the number d() := sup(ti+1 ? ti ) i0
when a new measurement is taken. This notion of solution is an accurate model of the process used in computer control (\sampling"). De nition I.1: Assume given a feedback k, a partition , and an x0 2 Rn . For each i, i = 0; 1; 2; : : :, recursively solve x_ (t) = f (x(t); k(x(ti ))) ; t 2 [ti ; ti+1 ] (3) using as initial value x(ti ) the endpoint of the solution on the preceding interval (and starting with x(t0 ) = x0 ). The -trajectory of (2) starting from x0 is the function x() thus obtained. Observe that this solution may fail to be de ned on all of [0; +1), because of possible nite escape times in one of the intervals, in which case we only have a trajectory de ned on some maximal interval. In our results, however the construction will provide a feedback for which solutions are globally de ned; we say in that case that the trajectory is well-de ned . Remark I.2: It is worth pointing out that the concept of solution introduced above is quite dierent from the \Euler solution" which would be obtained when attempting to solve the dierential equation (2) using Euler's method: in that case, on each interval [ti ; ti+1 ] one would have the formula x(t) = x(ti ) + (t ? ti )f (x(ti ); k(x(ti ))), corresponding to the solution of the dierent equation x_ = f (x(ti ); k(x(ti ))). This alternative de nition does not have any physical meaning in terms of the original system. (It is amusing to note, however, that this could be interpreted as the -trajectory of the system x_ = u under the feedback u = f (x; k(x)), so in that sense, Euler approximations are a particular case of our setup.) 2 B. Statement of Main Result The main objective of this paper is to explore the relationship between the existence of stabilizing (discontinuous) feedback and asymptotic controllability of the open loop system (1). We rst de ne the meaning of (globally) stabilizing feedback. This is a feedback law which, for fast enough sampling, drives all states asymptotically to the origin and with small overshoot. Of course, since sampling is involved, when near the origin it is impossible to guarantee arbitrarily small displacements unless a faster sampling rate is used, and, for technical reasons (for instance, due to the existence of possible explosion times), one might also need to sample faster for large states. Thus the sampling rate needed may depend on the accuracy desired when controlling to zero as well as on the rough size of the initial states, and this fact is captured in the following de nition. (The \s" in \s-stabilizing" is for \sampling".) De nition I.3: The feedback k : Rn ! U is said to sstabilize the system (1) if for each pair
0 0, and T = means of the following procedure: on each interval [ti ; ti+1 ], T (r; R) > 0 such that, for every partition with d() < the initial state is measured, ui = k(x(ti )) is computed, and A more cumbersome but descriptive notation would be \clssthen the constant control u ui is applied until time ti+1 , stabilizing", for stabilization under \closed-loop system sampling"
CLARKE, LEDYAEV, SONTAG, AND SUBBOTIN: ASYMPTOTIC CONTROLLABILITY IMPLIES FEEDBACK STABILIZATION
3
and for any initial state x0 such that jx0 j R, the - function (called just \Lyapunov function" for a control systrajectory x() of (2) starting from x0 is well-de ned and it tem in [25]; see also [17], [18]), (b) methods of nonsmooth holds that: 1. (uniform attractiveness) jx(t)j r 8 t T ; 2. (overshoot boundedness) jx(t)j M (R) 8 t 0; 3. (Lyapunov stability) limR#0 M (R) = 0. Remark I.4: As mentioned in Section VI, if a continuous feedback k stabilizes the system (1) in the usual sense (namely, it makes the origin of (2) globally asymptotically stable), then it also s-stabilizes. In this sense, the present notion generalizes the classical notion of stabilization. 2 We next recall the de nition of (global, null-) asymptotic controllability. By a control we mean a measurable function u : [0; +1) ! U which is locally essentially bounded (meaning that, for each T > 0 there is some compact subset UT U so that u(t) 2 UT for a.a. t 2 [0; T ]). In general, we use the notation x(t; x0 ; u) to denote the solution of (1) at time t 0, with initial condition x0 and control u. The expression x(t; x0 ; u) is de ned on some maximal interval [0; tmax (x0 ; u)). De nition I.5: The system (1) is asymptotically controllable if: 1. (attractiveness) for each x0 2 Rn there exists some control u such that the trajectory x(t) = x(t; x0 ; u) is de ned for all t 0 and x(t) ! 0 as t ! +1; 2. (Lyapunov stability) for each " > 0 there exists > 0 such that for each x0 2 Rn with jx0 j < there is a control u as in 1. such that jx(t)j < " for all t 0; 3. (bounded controls) there are a neighborhood X0 of 0 in Rn , and a compact subset U0 of U such that, if the initial state x0 in 2. satis es also x0 2 X0 , then the control in 2. can be chosen with u(t) 2 U0 for almost all t. This is a natural generalization to control systems of the concept of uniform asymptotic stability of solutions of differential equations. The last property { which is not part of the standard de nition of asymptotic controllability given in textbooks, e.g. [25] { is introduced here for technical reasons, and it has the eect of ruling out the case in which the only way to control to zero is by using controls that must \go to in nity" as the state approaches the origin. Our main result is as follows. Theorem 1: The system (1) is asymptotically controllable if and only if it admits an s-stabilizing feedback. One implication is trivial: existence of an s-stabilizing feedback is easily seen to imply asymptotic controllability. Note that the bounded overshoot property, together with the fact that k is locally bounded, insures that the control applied (namely, a piecewise constant control which switches at the \sampling times" in the partition) is bounded. The attractiveness property holds by iteratively controlling to balls of small radius and using the overshoot and stability estimate to insure convergence to the origin. Finally, the Lyapunov stability property holds by construction. The interesting implication is the converse, namely the construction of the feedback law. The main ingredients in this construction are: (a) the notion of control-Lyapunov
analysis, and (c) techniques from positional dierential games. We review these ingredients in the next section, and then develop further technical results; latter sections contain the proof as well as a robustness result. II. Some Preliminaries
We start with a known characterization of asymptotic controllability in terms of control-Lyapunov functions. Given a function V : Rn ! R and a vector v 2 Rn , the lower directional derivative of V in the direction of v is DV (x; v) := lim inf 1t (V (x + tv0 ) ? V (x)) : t#0
v0 ! v The function v 7! DV (x; v) is lower semicontinuous. For a set F Rn , coF denotes its convex hull.
De nition II.1: A control-Lyapunov pair for the system (1) consists of two continuous functions V; W : Rn ! R0 such that the following properties hold: 1. (positive de niteness) V (x) > 0 and W (x) > 0 for all x 6= 0, and V (0) = 0; 2. (properness) the set fx j V (x) g is bounded for each ; 3. (in nitesimal decrease) for each bounded subset G Rn there is some compact subset U0 U such that min DV (x; v) ?W (x) (4) v2cof (x;U ) 0
for every x 2 G. If V is part of a control-Lyapunov pair (V; W ), it is a control-Lyapunov function (clf). It was shown in [23] that asymptotic controllability is equivalent to the existence of a pair of functions (V; W ) which satisfy the properties given above, except that property 3 is expressed in an apparently weaker fashion, namely, by means of derivatives along trajectories (corresponding to relaxed controls). In [28] it was observed that in fact one can reformulate the de nition in the above terms, so we obtain as follows. Theorem 2: The system (1) is asymptotically controllable if and only if it admits a control-Lyapunov function. Observe that when the function V is smooth, condition (4) can be written in the more familiar form found in the literature, namely: min hrV (x); f (x; u)i ?W (x) :
u2U0
(5)
In contrast to the situation with stability of (non-controlled) dierential equations, a system may be asymptotically controllable system and yet there may not exist any possible smooth clf V . In other words, there is no analogue of the classical theorems due to Massera and Kurzweil. This issue is intimately related to that of existence of continuous feedback, via what is known as Artstein's Theorem (cf. [2], [17], [18], [24]), which asserts that existence of a dierentiable V
4
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. XX, NO. Y, MONTH 1999
is denoted by @p V (x) and is called the proximal subdierential of V at x. If the function V is dierentiable at x, then we have from the de nition that the only possible subgradient of V at x is = rV (x), but, we note that, unless further regularity is imposed (for instance, if V is of class C 2 ) then @p V (x) may well be empty. There is a simple relation between proximal subgradients of V at x and the lower directional derivatives DV (x; v), which follows immediately from the de nitions: for any 2 0 n where and @p V (x) are the proximal subgradients and the @p V (x) and any v 2 R , subdierential, respectively, of the function V at the point h; vi DV (x; v) : (8) x. The use of proximal subgradients as substitutes for the gradient for a nondierentiable function plays a central role We now x a parameter 2 (0; 1] and de ne the following in our construction of feedback. The concept was originally function: developed in nonsmooth analysis for the study of optimiz 1 jy ? xj2 : ation problems, see [4]. (9) V ( y ) + V (x) := yinf 2Rn 22 The relation (6) says that V is a \proximal supersolution" of the corresponding Hamilton-Jacobi equation, which via This is called the inf-convolution of the function V (with the results in [6] is known to be equivalent to the statement multiple of the squared-norm). The function V is well that locally V is a viscosity supersolution (cf. [12], [14]) of aknown in classical convex analysis as the \Iosida-Moreau the same equation. On the other hand, the relation (4) says regularization" of (convex) V . In the case of a lower semithat V is an upper minimax solution of the same equation V which is bounded below (for instance, for a ([29]). The coincidence of these two solution concepts re- continuous clf V , which is continuous and nonnegative), the function
ects the deep and intrinsic connection between invariance V is locally Lipschitz is an approximation of V in the properties of function V with respect to trajectories of the sense that lim V (xand ) = V (x); cf. [5]. In this case, the # 0 control system (1) and the characterization of these proper- set of minimizing points y in We choose ties in terms of proximal subgradients of V . For more dis- one of them and denote it by y(9)(xis). nonempty. (We are not asserting cussion on these aspects of viscosity and other generalized the existence of a regular choice; any assignment will do for solution concepts of rst-order PDEs, and their connection our purposes.) to nonsmooth analysis, the reader is referred to [6] and the By de nition, then, book [29]. Finally, we rely on methods developed in the theory of V (x) = V (y (x)) + 21 2 jy (x) ? xj2 positional dierential games in [19]. These techniques were used together with nonsmooth analysis tools in the con V (y) + 21 2 jy ? xj2 (10) struction of discontinuous feedback for dierential games of pursuit in [7] and games of xed duration in [8], and n these results are relevant to the construction of stabilizing for all y 2 R . The vector feedback in our main result. (11) (x) := x ?y2 (x) This paper is organized as follows. In Section III, we provide a self-contained exposition of proximal subgradients and their basic properties. These methods are used will be of special interest; it is simultaneously a proximal in Section IV for construction of \semiglobal" (that is to subgradient of V at y (x) and a proximal super gradient of say, on each compact set) stabilizing feedback, and the res- V at x, as we discuss next. n Lemma III.1: For any x 2 R , ults are extended to the global case in Section V. Finally, section VI discusses robustness issues. (x) 2 @p V (y (x)) : (12) Proof: We rearrange terms in (10) and obtain, for any III. Proximal Subgradients and y 2 Rn , Inf-Convolutions We recall the concept of proximal subgradient, one of the V (y) V (y (x)) + h (x); y ? y (x)i? 21 2 jy ? y (x)j2 ; basic building blocks of nonsmooth analysis. A vector "Rn is a proximal subgradient (respectively, supergradient ) of the function V : Rn ! (?1; +1] at which implies (12), by de nitionn of subgradients. x if there exists some > 0 such that, for all y in some Lemma III.2: For any x 2 R , (x) is a proximal supergradient of V at x. neighborhood of x, Proof: The assertion will be shown if we prove that, V (y) V (x) + h; y ? xi ? jy ? xj2 (7) for every x; y 2 Rn , (respectively, V (y) V (x)+ h; y ? xi + jy ? xj2 ). The set V (y) V (x) + h (x); y ? xi + 21 2 jy ? xj2 : (13) of proximal subgradients of V at x (which may be empty)
is equivalent, for systems ane in controls, to there being a stabilizing regular feedback. Nevertheless, it is possible to reinterpret the condition (5) in such a manner that relation (5) does hold in general, namely by using a suitable generalization of the gradient. Speci cally, we will remark later that we may use the proximal subgradients of V at x instead of rV (x), replacing (5) by: min h; f (x; u)i ?W (x) for every 2 @p V (x) ; (6) u2U
CLARKE, LEDYAEV, SONTAG, AND SUBBOTIN: ASYMPTOTIC CONTROLLABILITY IMPLIES FEEDBACK STABILIZATION
By de nition, we have
V (y) V (y (x)) + 21 2 jy (x) ? yj2
V (x) = V (y (x)) + 21 2 jy (x) ? xj2 :
5
which follows from the de nitions of V and (R). Recall that 2 (0; 1], so it follows from here that also p
jxj R ) jy (x)j R + 2 (R) :
(20)
We next show that V is uniformly approximated by V on B . Subtracting and rearranging quadratic terms, we ob- RLemma III.4: For any x 2 BR , tain (13). p 1 We deduce from Equation (13) that, for any 2 R and 2 ( R ) : (21) V ( x ) V ( x ) V ( x ) + ! n R any v 2 R , Proof: The rst inequality is just (18). To prove the 2 jv j2 second one, we use the obvious relation V (x + v) V (x) + h (x); vi + 22 : (14) V (x) V (y (x)) : This plays a role analogous to that of Taylor expansions for evaluating increments of the function V . By the estimate in Lemma III.3, We introduce some additional notations and obtain sevp eral estimates which are used later. In general, we use BR V (y (x)) V (x) ? !R 2 (R) : to denote the closed ball of radius R. We assume that the function V is a clf as in De nition II.1. For each R > 0 we Equation (21) follows from these two inequalities. consider the numbers We use later the inclusions (R) := max V (x) and (R) := min V (x) B(R) GR GR ; jxjR
jxjR
(22)
which are valid for all R > 0 and > 0 and which follow directly from the de nition of and (18). Lemma III.5: For any R > 0 and satisfying p !R 2 (R) < 21 (R) (23) we have and GR int BR : (24) (R) := max f j B GR g : Proof: Let x 2 GR . By de nition of GR , V (x) 1 (R), so the fact that V (x) equals the expression in (10) 2 It is easy to verify that the following relations hold: implies that (R) (R) > 0 ; (R) > 0 8 R > 0 ; (15) V (y (x)) 21 (R) and 21 2 jy (x) ? xj2 21 (R) : lim (R) = Rlim (R) = 0 ; (16) R#0 #0 It follows that y (x) 2 BR by de nition of (R) and lim (R) = Rlim (R) = 1 : (17) jy (x) ? xj p2 (R) because of (15). Then R!1 !1 p Finally, the function V (x) V (y (x)) + !R 2 (R) < (R) ; !R () := o and this implies that jxj < R by de nition of (R). n max V (x) ? V (x0 ) jx ? x0 j ; x; x0 2 BR+p2 (R) as well as the sets
1 GR := x V (x) 2 (R) ; GR := x V (x) 12 (R) ;
is the modulus of continuity of the function V on the ball BR+p2 (R) . The next Lemma provides an upper estimate of the distance between x and y (x). p Lemma III.3: For any x 2 BR , jy (x) ? xj 2 (R). Proof: The conclusion follows from the obvious inequality V (x) V (x) (18) as well as the inequality 1 2 22 jy (x) ? xj V (x) ? V (y (x)) V (x) (R) ; (19)
IV. Semi-Global Practical Stabilization
From now on, we assume that an asymptotically controllable system (1) has been given. Applying Theorem 2, we pick a Lyapunov pair (V; W ). Choose any 0 < r < R. We will construct a feedback that sends all states in the ball of radius R (\semiglobal" stabilization) into the ball of radius r (\practical" stability, as we do not yet ask that states converge to zero). By the de nition of Lyapunov pair, there is some compact subset U0 U so that property (4) holds for every x 2 BR+p2 (R) . Because of the relation (8) between proximal subgradients and lower directional derivatives, we have that also condition (6) holds for the Lyapunov pair, for all x 2 BR+p2 (R) .
6
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. XX, NO. Y, MONTH 1999
Proof: It follows from (25) and (29) that Pick any 2 (0; 1]. In terms of the vectors (x) introduced in Equation (11), we de ne a function k : BR ! U0 by letting k (x) be a pointwise minimizer of h (x); f (x; k (x))i h (x); f (x; u)i: umin h (x); f (y (x); u)i + ` j (x)j jy (x) ? xj : 2U0 h (x); f (x; k (x))i = umin h (x); f (x; u)i : (25) But in view of Equation (12), and condition (6), applied at 2U0 the point y (x) 2 BR+p2 (R) (recall (20)), we have that The choice x 7! k (x) is not required to have any particular regularity properties. We use the subscript = (; r; R) to the rst term in the right-hand side is upper bounded by emphasize the dependence of the function k on the particu- ?W (y (x)). Regarding the second term, note that the eslar parameters (which represent respectively the \degree of timate smoothing" of V that is used in its construction, the radius j (x)j jy (x) ? xj 2(V (x) ? V (y (x))) of the ball to which we are controlling, and the radius of p 2 !R 2 (R) the ball on which the feedback will be eective). The next theorem says that for any xed r; R, we can choose arbitrarily small > 0 and then > 0 such that the set GR is holds, by the de nition of (x), the rst inequality in Equation (19), the conclusion of Lemma III.3, (20), and invariant with respect to any -trajectory of the de nition of !R . So we obtain x_ = f (x; k (x)) ; (26) p 2 ( R ) : h ( x ) ; f ( x; k ( x )) i ? W ( y ( x )) + 2 ` ! R and that x(t) enters and stays in Br for all large t, provided (32) that the diameter of the partition satis es d() . Theorem 3: Let V be a clf. Then, for any 0 < r < R Since (r) jxj R, we obtain from Lemma III.3, (20), there are 0 = 0 (r; R) and T = T (r; R) such that, for any and (30) that 2 (0; 0 ) there exists > 0 such that for any x0 2 GR 1 (r) jy (x)j R + p2 (R) : and any partition with d() , the -trajectory x() 2 of (26) starting at x0 must satisfy: This implies that W (y (x)) 3 and from (32) and (30) x(t) 2 GR ; 8t 0 (27) we obtain (31), as claimed. Now we consider any -trajectory of (26) corresponding and to a partition = fti gi0 with d() , where satis es x(t) 2 Br ; 8t T : (28) the inequality 0 Observe that, because of (17) and (22), for every R > ! p 0 there is some R > 0 such that BR0 GR . Thus the 2 `m 2 ( R ) m Theorem says that there is a feedback that steers every (33) + 22 : state of BR0 into the neighborhood Br ; in this sense the result is semiglobal. The proof of the Theorem will take Lemma IV.2: Let ; satisfy (23), (30), and (33), and the rest of this section and will be based on a sequence assume that for some index i it is the case that x(ti ) 2 of lemmas. The main idea behind the proof is to use the G n B ( r) . Then new function V (for suciently small ) as a Lyapunov R function and to use the \Taylor expansion" formula (14) to V (x(t)) ? V (x(ti )) ?(t ? ti ) (34) estimate the variations of V along -trajectories. By continuity of f (x; u) and the local Lipschitz property for all t 2 [ti ; ti+1 ]. In particular, x(t) 2 GR for all such t. assumed, we know that there are some constants `; m such Proof: Denote xi := x(ti ) and consider the largest that t 2 (ti ; ti+1 ] such that x( ) 2 BR for all 2 [ti ; t]. Note such a t exists, since xi 2 int BR due to (24). Pick any jf (x; u) ? f (x0 ; u)j ` jx ? x0 j ; jf (x; u)j m (29) that t 2 [ti ; t]. We have from (29) that for all x; x0 2 BR and all u 2 U0 . Let jx( ) ? xi j m( ? ti ) : (35) p 1 1 := 3 min W (y) 2 (r) jyj R + 2 (R) : In general, x(t) = xi + (t ? ti )fi ; (36) Note that > 0, due to the positivity of W for x 6= 0. where Lemma IV.1: Let 2 (0; 1] satisfy Zt 1 p p 1 2 (R) < 2 (r) ; 2` !R 2 (R) < : (30) fi = t ? ti ti f (x( ); k (xi )) d = f (xi ; k (xi )) + i : From (29) and (35) we obtain estimates Then, for any x 2 BR n B(r) it holds that
h (x); f (x; k (x))i ?2 :
(31)
jfi j m ; ji j `m(t ? ti ) :
(37)
CLARKE, LEDYAEV, SONTAG, AND SUBBOTIN: ASYMPTOTIC CONTROLLABILITY IMPLIES FEEDBACK STABILIZATION
7
Now, using the \Taylor expansion" formula (14) and (36), Proof: It follows from Lemma IV.2 and minimality of we conclude that N that x(ti ) 2 GR n Gr ; i = 0; 1; : : :; N ? 1 ; V (x(t)) ? V (x(ti )) = V (xi + (t ? ti )fi ) ? V (xi ) (t ? ti )h (xi ); fi i + 21 2 (t ? ti )2 jfi j2 : (38) (applying Lemma IV.2 recursively and using that B(r) Gr ) and On the other hand, using (36) and (37) as well as 0 V (x(tN )) V (x(0)) ? tN 21 (R) ? tN ; Lemma IV.1, we have: so (39) is valid. h (xi ); fi i h (xi ); p f (xi ; k (xi ))i + j (xi )j ji j Lemma IV.4: Assume that, in addition to the previously 2 (R) imposed constraints, and also satisfy the following two ?2 + `m conditions: p 1 2 ( R ) ! (40) < 4 (r) ; R which implies that (which actually implies (23)) and V (x(t)) ? V (x(ti )) ! p 1 ! (41) 2 (R) 1 R (m ) < (r) : 2 4 ?2 + `m + 22 m (t ? ti ) : Then, x(t) 2 Br for all t tN . Proof: If x(ti ) 2 Gr , then This implies (34) for all t 2 (ti ; t]. In particular, (34) im plies that x(t) 2 int BR , which contradicts the maximality 3 of t unless t = ti+1 . Therefore (34) holds for all t 2 [ti ; ti+1 ]. V (x(t)) V (x(t)) V (x(ti )) + !R (m) < 4 (r) (42)
Next we establish that every trajectory enters the ball for all t 2 [ti ; ti+1 ] (using (29) and (41)). Then, Br at time tN , where N is the least integer such that V (x(ti+1 )) < 34 (r) ; x(tN ) 2 Gr , and stays inside thereafter. To do this, we rst show that there is a uniform upper bound on such times tN . and if x(t ) 62 G we can apply Lemma IV.2 to conclude i+1 r Figure 1 may help in understanding the constructions. that (43) V (x(t)) < 43 (r) for all t 2 [ti ; tj ], where j is the least integer such that tj > ti and x(tj ) 2 Gr (if there is any). Starting from tj , we may repeat the argument. We conclude that (43) holds for all t 0. Due to Lemma III.4 and (40), B
R+
2β (R)
B
y (x(t)) α
R
x(t)
xo
α GR
x(t N) B ρ (r)
V < γ( r)
Gr
GR
V (x(t)) V (x(t)) + !R
p
2 (R) < (r) ;
which means that x(t) 2 Br for all t tN , as claimed. To conclude the proof of Theorem 3, let 0 be the supremum of the set of all > 0 which satisfy conditions (30) and (40). Then, for any 2 (0; 0 ) we can choose satisfying (33) and (41), so that, for each partition with d() and every -trajectory starting from a state in GR , the inclusions (27) and (28) hold. V. Proof of Global Result
We now prove Theorem 1. The idea is to partition the state space Rn into a number of \spherical shells" (more precisely, sets built out of sublevel sets of the funcFig. 1. The various sets appearing in the constructions tions V obtained when smoothing-out the original controlLyapunov function) and to use a suitable feedback, conLemma IV.3: Let satisfy (23) and (30), and pick any structed as in the semiglobal Theorem 3, in each shell. so that (33) is valid. Then, for any -trajectory x() with We assume given an asymptotically controllable system, d() and every x(0) 2 GR , it holds that and a clf V for it, and we pick an arbitrary R0 > 0. There is then a sequence fRj g+j=1?1 satisfying
( R ) (39) tN T = 2 : 2Rj (Rj+1 ) ; j = 0; 1; 2; : : :
8
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. XX, NO. Y, MONTH 1999
(just de ne the Rj 's inductively for j = 1; 2; : : : and for j = ?1; ?2; : : :; this is possible because of (17)). We also denote rj := 21 (Rj?1 ) for all j . We have that, for each integer j , (Rj ) < Rj < 2Rj < (Rj+1 ) (44) and lim R = 0 ; j!lim R = 1: (45) j !?1 j +1 j
Since the sets Hj plus the origin constitute a partition of the state space, we may de ne a map k : R ! U by means of the rule k(x) := kj (x) ; 8 x 2 Hj?1 (50) for each integer j , and k(0) = u0 , where u0 is any xed element of U0 . To complete the proof of the Theorem, we need to show that this feedback is s-stabilizing. Lemma V.2: The set GRjj is invariant with respect to Consider any integer j . For the pair (rj ; Rj ), Theorem 3 -trajectories of (2), when (50) is used and provides the existence of numbers j > 0, j > 0, and Tj > 0, and a map kj : BRj ! Uj , kj := k(j ;rj ;Rj ) , d() < min j ; Rmj?1 : (51) j j such that GRj is invariant with respect to all -trajectories Proof: Consider any -trajectory starting at a state of (26) when k = kj and d() j , and for each such in GRj , where satis es (51). It is enough to show that j trajectory it holds that x(t) 2 GRjj for all t 2 [ti ; ti+1 ] if x(ti ) 2 GRjj . Pick any i. There are three possibilities which must be treated jx(t)j rj ; 8t Tj : (46) such separately: (a) x(ti ) 2 Hj?1 , (b) x(ti ) 2 H`?1 for some Recall that in the construction of kj , we used the fact that ` j ?1, and (c) x(jti ) = 0. In the rst case, k(x) = kj (x) is there is some compact subset U0 U, to be called here Uj known to leave GRj invariant, so there is nothing to prove. Assume now that we are in case (b). Observe that the to distinguish the sets used for the dierent indices j , so partition may not be ne enough to guarantee that GR`` that condition (6) holds for the Lyapunov pair, for all x 2 BRj +p2 (Rj ) . Since the Rj form an increasing sequence, is invariant under the feedback k` , so we need to argue a dierent way, namely that the trajectory cannot go and since if the min in condition (6) also holds if we enlarge in too the sampling time is small. Since x(ti ) 2 U0 , we may, and will, assume that the Uj = U0 for all j < 0 H farbecause ` B , jx(t )j < R . Pick any t 2 [t ; t ] G 1 R` i ` i i+1 R` and that Uj Uj+1 for all j 0. In Equation (29) we so`?that ]. It follows from the j x ( t ) j R for all t 2 [ t ; t j i picked a bound m on the values of jf (x; u)j for x 2 BRj choice of m and the fact that k(x(t )) 2 U U that ` j and all u 2 Uj ; we call this bound mj to emphasize the jx(t) ? x(t )jj m (t ? t ) for all t 2 i[t ; t] and from here i j i i dependence on j , and observe that mj mj+1 for all j , that because of the monotonicity of the sets Uj and BRj . Finally, we introduce the sets on which we will use the jx(t)j mj (t ? ti ) + R` Rj?1 + R` 2Rj?1 ; (52) dierent feedbacks kj : for all t 2 [ti ; t]. We claim that j : n G Hj := GRjj+1 Rj +1 jx(t)j < Rj (53) T Lemma V.1: For each i 6= j , Hi Hj = ;, and for all t 2 [ti ; ti+1 ]. Indeed, if this were not the case, then there is some t 2 [ti ; ti+1 ] so that jx(t)j Rj for all t 2 [ti ; t] +[ 1 Rn n f0g = Hj : (47) and jx(t)j = Rj . Since also jx(t)j Rj for all t 2 [ti ; t], Equation (52) holds, which gives Rj 2Rj?1 , contradictj =?1 ing (44). We conclude that (53) holds, so the argument Proof: We know from (44), (22), and (24) that leading to (52) can be applied with t = ti+1 and we conj+1 j (48) clude that the trajectory stays in B2Rj?1 , which is included GRj BRj B2Rj GRj+1 in GRjj by (48). T j for all j . This implies that Hi GRj and Hi Hj = ; There remains the special case x(ti ) = 0. Recall that whenever i < j . For any two integers M < N , we obtain k(0) 2 U0 Uj . Then the argument given for case (b) from (48) that can be repeated, where in (52) we have 0 instead of R` ; we conclude that the trajectory stays in BRj?1 , so it is in GRjj . N [ M fx j RM jxj RN g Hj = GRNN+1 We now prove that the feedback k is s-stabilizing. We +1 n GRM : j =M claim that for any 0 < r < R, there exist = (r; R) > 0, = T (r; R) > 0, and M = M (R), such that, when (50) The rst inclusion together with (45) implies (47) when Tis used and d() < , every -trajectory x() of (2) with taking M ! ?1 and N ! +1. j x (0) j R satis es: We use later the inclusion
Brj int GRjj??11 which follows from (22) and the de nition of rj .
(49)
jx(t)j r ; 8t T
and
jx(t)j M ; 8t 0 :
(54)
CLARKE, LEDYAEV, SONTAG, AND SUBBOTIN: ASYMPTOTIC CONTROLLABILITY IMPLIES FEEDBACK STABILIZATION
We start by choosing an integer K = K (r) and the least integer N = N (R) such that
GRKK Br and BR GRNN ; and de ne
:= min
T := and
min ; RK ?1 ; K j N j m N
N X j =K +1
Tj ;
M := RN (R) :
9
VI. Robustness Properties
One of the main justi cations for the use of feedback control as opposed to open-loop control lies in the robustness (55) properties of the former. Certainly, any feedback law automatically accounts for sudden changes in states. However, it is not necessarily the case that a given feedback law, especially a discontinuous one, will stabilize a system which is not identical to the model assumed for its design. We show in this section that the s-stabilizing feedback control k de ned in (50) is indeed insensitive to small perturbations in the system dynamics. In particular, we establish the existence of a continuous function : Rn ! R0 , (x) > 0 for x 6= 0 (which depends on the system (1) being controlled), such that for any continuous function g : Rn U ! Rn satisfying
Consider the -trajectory starting from some x0 2 BR . We claim that at some instant ti 2 such that ti T it holds that x(ti ) 2 GRKK . If x0 62 GRKK , then x0 2 HN 0 for some integer K N 0 N ? 1. The set GRNN00 is invariant, by Lemma V.2 (notice the choice of , consistent with what is needed there). Therefore the trajectory remains in HN 0 until it enters GR??11 in some interval (ti?1 ; ti ], for some < N 0 . We note that ti TN 0 , since due to Theorem 3 if x() remains in HN 0 until moment TN 0 then
x(TN 0 ) 2 BrN 0 : Because of (49), this contradicts our assumption that x() remains in HN 0 . If K + 1, we are done. Otherwise we repeat the argument. In conclusion, x(ti ) 2 GRKK for some ti T , as claimed. But then Lemma V.2 insures that x() stays in this set for all future t ti , which in accordance with (55) implies (54). It follows from these considerations that x(t) remains in GRNN BRN for all t 0. Thus jx(t)j M (R) for all t, and the claim is established. We conclude that the feedback k has the attractiveness and bounded overshoot properties required by De nition I.3. To complete the proof of the Theorem, we must still verify that k has the Lyapunov stability property postulated in that de nition, namely, that the function M (R) just de ned satis es also lim M (R) = 0 : (56) R#0
jg(x; u) ? f (x; u)j (x) ; for all x 2 Rn ; u 2 U ; (57) the same feedback k is s-stabilizing also for the control system x_ = g(x; u) : (58) (Note that this result also demonstrates, in view of Theorem 1, that the asymptotic controllability property is structurally stable with respect to perturbations g of the system (1) which satisfy (57).) Actually, we prove an even more general result than the s-stabilization of (58), expressed in terms of dierential inclusions. Let us consider -trajectories of the dierential inclusion
x_ 2 f (x; k(x)) + B(x)
(59)
where k : Rn ! U is an arbitrary (not necessarily continuous) function and : Rn ! R0 is continuous. For a given partition of the interval [0; +1), a (complete) trajectory x() of (59) is any absolutely continuous function de ned on [0; 1) and satisfying recursively the dierential inclusion
x_ (t) 2 f (x(t); k(x(ti ))) + B(x(t)) ; a.a. t 2 [ti ; ti+1 ] (60) for all i. De nition VI.1: The feedback k : Rn ! U is said to
robustly s-stabilize the system (1) if there exists some continuous function : Rn ! R0 , (x) > 0 for x 6= 0, such that for every pair 0 < r < R there exist = (r; R) > 0, T = T (r; R) > 0, and M = M (R), such that, for every partition with d() < , and every -trajectory x() of (59) with jx(0)j R, all conditions 1., 2., 3. in De nition I.3 It follows fromthe de nition of N = N (R) that BR is not hold. contained in GRNN??11 . Therefore The concept of robust s-stabilizing feedback k formulated in terms of solutions of dierential inclusions is closely con(RN ?1 ) < R nected with robustness properties of the feedback k in the sense of stabilization of the perturbed system and we get from this inequality that x_ = f (x; k(x)) + w(t) (61) lim N (R) = ?1 : where w : [0; +1) ! Rn is an integrable function (which R#0 we call a \disturbance"). For the partition of [0; +1) Then (56) follows from the fact that limj!?1 Rj = 0. the -trajectory of (61) starting from x0 is the absolutely
10
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. XX, NO. Y, MONTH 1999
continuous function recursively de ned on intervals [ti ; ti+1 ] Finally, we can repeat this argument to obtain that jx(t)j as solutions of the following dierential equation: r for all t > T (r0 ; R), which proves the Proposition. (Observe that since x() is a solution of an ordinary dierential x_ (t) = f (x(t); k(x(ti ))) + w(t) ; t 2 [ti ; ti+1 ] : (62) equation, boundedness implies global existence.) The main result of this Section is as follows: Remark VI.2: There is an obvious relation between Theorem 4: Let the system (1) be asymptotically con-trajectories of the dierential inclusion (59) and - trollable. Then the feedback k de ned in (50) is robustly trajectories of the perturbed system (61): x() is a - s-stabilizing. trajectory of (59) if and only if there is a measurable disWe base the proof of this Theorem on the proof of the turbance w() such that following semiglobal practical robust stabilization result, in the same manner that the proof of Theorem 1 is based on jw(t)j (x(t)) ; a.a. t 0 (63) Theorem 3. Theorem 5: Let V be a clf. Then, for every 0 < r < R and x() is a -trajectories of (61). 2 there exist 0 = 0 (r; R) and T = T (r; R) such that for We show that a robustly stabilizing feedback provides any 2 (0; ) there exists > 0, > 0 such that for stabilization of the perturbed system (62) in the following any x0 2 GR , 0any feedback k 0de ned as in (25), and any sense. disturbance w() satisfying (64), and any partition with De nition VI.3: The feedback k : Rn ! U is said to d() < , the -trajectory x() of the perturbed system be robustly practically stabilizing for (61) if for every pair x_ = f (x; k (x)) + w(t) (65) 0 < r < R there exist > 0, T > 0, and 0 > 0, such that, for any disturbance w() satisfying starting from x0 must satisfy (27) and (28). The proof of this Theorem follows from the proof of Thejw(t)j 0 ; a.a. t 0 ; (64) orem 3 with some minor changes. We explain just the and any partition such that d() < , every -trajectory needed changes, not repeating all details, and using the of (62) with jx(0)j R is well-de ned on [0; 1) and condi- same notations as in Section IV. tions 1., 2., 3. in De nition I.3 hold. 2 We start by de ning We have the following relation between the two concepts m0 := max fjf (x; u)j j x 2 BR ; u 2 U0 g of robustness. Proposition VI.4: If the feedback k is robustly s-stabil- and rede ne the constant m now as izing, then it is robustly practically stabilizing for (61). m := 2m0 : Proof: Let us x any pair 0 < r < R and choose r0 such that Then the upper bound 0 in (64) on the magnitude of the 2r0 < r; M (2r0 ) < r : admissible disturbance is given as ) ( We de ne constants 0 : (66) 0 = min m ; p 2 2 (R) 0 := min 21 (x) j r0 jxj M (R) For this choice of 0 we have that the inequalities in (29) and can be replaced by 0 r 0 0 := min (r ; R); 2m ; jf (x; u) ? f (x0 ; u)j ` jx ? x0 j ; jf (x; u) + wj m ; (67) where m := maxfjf (x; u)j + (x) j jxj rg, and where the which hold for all x; x0 2 BR , all u 2 U0 , and all w 2 B0 . functions and are as in De nition VI.1. Then for every Because of (66) we have from Lemma III.3 that for all x 2 partition with d() < , any -trajectory x() of the per- BR and w 2 B0 : turbed system(65) with the disturbance w() satisfying (64) and jx(0)j R is a -trajectory of the dierential incluh (x); wi j (x)j 0 21 : (68) sion (59) as long as r0 jx(t)j M (R). Since k is robustly s-stabilizing, jx(t)j stays bounded by M (R) until the rst Now we analyze the behavior of any -trajectory of the time ti T (r0 ; R), ti 2 , such that jx(ti )j r0 . Assume perturbed system (65) assuming that the disturbance satthat after this moment, x(tl ), l = i + 1; : : : ; j ? 1 belongs is es (64). In place of Lemma IV.2, we substitute the folto the ball Br0 , but x(tj ) 62 Br0 . Due to the choice of 0 we lowing: have that x(t) 2 B2r0 for t 2 [ti ; tj ]. Then r0 jx(tj )j 2r0 Lemma VI.5: Let ; satisfy (23), (30), and (33),, and and x() again is a -trajectory of the dierential inclu- assume for some index i it is the case that x(ti ) 2 sion (59) driven by the feedback k until the next moment G n B that. Then ( r ) R tk > tj such that jx(tk )j < r0 . It follows that (69) V (x(t)) ? V (x(ti )) ? 21 (t ? ti ) jx(t)j M (2r0 ) < r for t 2 [tj ; tk ] :
CLARKE, LEDYAEV, SONTAG, AND SUBBOTIN: ASYMPTOTIC CONTROLLABILITY IMPLIES FEEDBACK STABILIZATION
for all t 2 [ti ; ti+1 ]. In particular, x(t) 2 GR for all such t. The proof of this Lemma follows along the lines of the proof of Lemma IV.2, but with one important modi cation due to the de nition (62) of x() as a -trajectory of (65). Namely, fi in (36) is de ned now instead as follows:
fi = t ?1 t i
Zt ti
[f (x( ); k (xi )) + w( )] d
Zt w( ) d + i : = f (xi ; k (xi )) + t ?1 t i
ti
Note that due to (67) the estimate (37) holds again in this case. This implies, because of (68):
h (xi ); fi i h (xi ); f (xi ; k (xi ))i + j (xi )j j0 j + j (xi )j ji j and therefore, due to (31), (37), and (68): p
11
To conclude the proof of this theorem, we let 0 be the supremum of all the > 0 satisfying (23) and (30). Then, for every 2 (0; 0 ) we can choose 0 satisfying (66) and satisfying (33) and (41) so that for any disturbance w() bounded by 0 and any -trajectory x() of the perturbed system (65) with d() starting from GR the inclusions (27) and (28) hold. Proof of Theorem 4. We use the same notations as in Section V and follow closely the scheme of the proof of Theorem 1, pointing out the necessary modi cations. Let
m0j := max jf (x; u)j j x 2 BRj ; u 2 Uj ; (
)
; j := min m0j ; pj 2 2 (Rj ) mj = 2m0j : Without loss of generality, we can assume that mj is monotone increasing. We de ne the function
h (xi ); fi i ? 23 + 2 (R) `m : (x) := inf fe(y) + jy ? xj j y 2 Rn g ; Next, applying the \Taylor expansion" formula (38), we where e is a function de ned as follows: e(0) = 0, obtain: e(x) = j for x 2 Hj?1 : (71) V (x(t)) ? V (x(ti )) It is obvious that the function so de ned is Lipschitz with ! p 2 (R) 1 3 constant 1 in all of Rn and it satis es ? 2 + `m + 22 m2 (t ? ti ) : 0 < (x) j for x 2 Hj?1 : (72) In view of the choice of and , we obtain that (69) holds We now consider the -trajectory x() of the dierential for all t 2 (ti ; t]. In particular, (69) implies that x(t) 2 inclusion In view of Remark VI.2, x() is also a soluint BR , which contradicts the maximality of t unless t = tion of the(59). perturbed system (61) corresponding to some ti+1 . Therefore (69) holds for all t 2 [ti ; ti+1 ]. disturbance w ( ) which satis es (63). The decreasing property for V along every -trajectory Due to the inequality (72), we obtain that if x(ti ) 2 Hj?1 of the perturbed system (65) lets us establish that every for some t , j , and i -trajectory enters the ball Br at time tN , where N is the least integer such that x(tN ) 2 Gr . et := sup ft ti j x(t) 2 Hj ?1 g ; Lemma VI.6: Let satisfy (23) and (30), and 0 satisfy (66). Pick > 0 so that (33) holds. Then for every dis- then on [ti ; et], x() is a -trajectory of the perturbed systurbance w() satisfying (64), and any -trajectory of (65) tem (65) with = j and by j . Note that if w() bounded with d() and every x(0) 2 GR , it holds that x() stays in GRjj then x(t) ? x(et) mj (t ? et) for any t; et. This means in accordance with Theorem 5, Lemmas VI.5
( R ) and VI.6 that we have a result analogous to Lemma V.2: (70) tN T = 2 : j Lemma VI.7: The set G is invariant with respect to Rj Proof: It follows from Lemma VI.5 and minimality of -trajectories of the dierential inclusion (59), when (50) N that is used and (51) holds. 2 Moreover, due to Lemma VI.6 we have that for every j x(ti ) 2 GR n Gr ; i = 0; 1; : : :; N ? 1 ; there is a Tj such that for any -trajectory of (59) starting applying Lemma IV.2 recursively and using that B(r) Gr from Hj?1 there is a moment t0 Tj such that x(t0 ) 2 we obtain GRjj??11 if d() j . Then, to complete the proof of the fact that k is robustly s-stabilizing we need to repeat word by 0 V (x(tN )) V (x(0)) ? 12 tN 12 (( (R) ? tN ) : word the end of the proof in Section V, starting from (54). Corollary VI.8: If the system(1) is asymptotically conThis shows (70). Finally, note that Lemma IV.4 is valid for any - trollable then there is a continuous function : Rn ! R0 , trajectory x() of the perturbed system (65) and any dis- (x) > 0 for x 6= 0 such that any continuous function turbance w() satisfying (64), because of the choice of 0 g(x; u) satisfying (57), any s-stabilizing feedback (50) is in (66) and the estimates (67). s-stabilizing for the system (58). 2
12
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. XX, NO. Y, MONTH 1999
Without loss of generality, we can assume that tk ! et 2 [0; T ] and xk () ! x() uniformly on [0; T ] (by the Arzelax_ = g(x; k(x)) (73) Ascoli theorem, since the functions are equibounded and equicontinuous because their derivatives are estimated by is a -trajectory of the dierential inclusion (59); thus the an upper bound of f (x; k(x)) on BR ). Because of continuCorollary follows from Theorem 4. ity of k and the fact that k ! 0, we conclude that x() is a trajectory of (2) with jx(0)j R. But by (76) we A. Concluding Remarks have that x(et) = 2M (R), which contradicts condition 2. It easily follows from Theorem 4 and Proposition VI.4 in De nition I.3. The proof that (75) holds is analogous. that the feedback k in (50) is robust with respect to actuator We now prove Proposition VI.9. We 0 < r < R, and errors. By this we mean that, for arbitrary 0 < r < R, for all suciently small perturbations e() and w(), all - choose r0 > 0 such that trajectories of 2M (2r0 ) < r ; 2r0 < r ; x_ (t) = f (x(t); k(x(t)) + e(t)) + w(t) and 0 1 r 1 0 0 0 0 with suciently small sampling periods d() satisfy condi := min ( 2 r ; R); ( 2 r ; 2r ); 2m ; tions 1. and 2. in De nition I.3. We leave to reader to ll the details. where m = maxff (x; k(x)) j jxj rg. On the other hand, we have not made any statements reThen we obtain from Lemma VI.10 that any -trajectory garding measurement errors. Indeed, our technique is not of (2) with d() 0 satis es (74) and jx(T )j 21 r0 robust with respect to such errors. It is possible to pro- for T = T ( 21 r0 ; R). This means that for some ti 2 duce examples of systems for which every s-stabilizing state we have jx(ti )j < r0 and x(t) stays in Br0 for those mofeedback will fail to stabilize if the input k(x(t) + e(t)) is ments ti ; ti+1 ; : : : ; tj?1 while jx(tj )j r0 . Due to the choice used instead of k(x(t)). This question is related to the fact of 0 , we have that jx(tj )j 2r0 < r. We again apply that feedback robust to measurement error results essen- Lemma VI.10, now for T 0 = T ( 21 r0 ; 2r0 ) and M 0 = M (2r0 ), tially in solutions of certain dierential inclusions (see [15]). to obtain that However, it is possible to provide a robust dynamic (rather jx(t)j < 2M 0 < r for all t 2 [tj ; tj + T 0] : than purely state) stabilizing feedback; the paper [20] deals with that issue. the rst time tk 2 such that tk > tj It was mentioned in Section II that a continuous feed- After that, we nd 0 , and repeat the arguments above. This and j x ( t ) j < r k back which stabilizes in the usual sense is in particular s-stabilizing. We now make this precise. For a continuous implies that x(t) stays in Br for all t T , which proves the feedback k (not necessarily Lipschitz, so there is no unique- Proposition. ness of solutions), stabilizability is taken here to mean that, References for each 0 < r < R there exist T = T (r; R) and M = M (R) [1] Aubin, J.-P., and A. Cellina, Dierential Inclusions: Set-Valued such that for every trajectory of the closed loop system (2) Maps and Viability Theory , Springer-Verlag, Berlin, 1984. with jx(0)j R we have conditions 1., 2., 3. in De ni- [2] Artstein, Z., \Stabilization with relaxed controls," Nonl. Anal., TMA 7(1983): 1163-1173. tion I.3. R.W., \Asymptotic stability and feedback stabilizaProposition VI.9: If k is a continuous stabilizing feed- [3] Brockett, tion," in Dierential Geometric Control theory (R.W. Brockett, back then it is s-stabilizing. R.S. Millman, and H.J. Sussmann, eds.), Birkhauser, Boston, 1983, pp. 181-191. We rst establish: [4] Clarke, F.H., Methods of Dynamic and Nonsmooth OptimizaLemma VI.10: Let T (r; R) and M (R) be as in the de ntion. Volume 57 of CBMS-NSF Regional Conference Series in Applied Mathematics, S.I.A.M., Philadelphia, 1989. ition of continuous stabilizing feedback. Then there exists F.H., Yu.S. Ledyaev, and P. Wolenski, \Proximal ana = (r; R) such that, for any partition with d() and [5] Clarke, lysis and minimization principles," J. Math. Anal. Appl. 196 any -trajectory of (2) with jx(0)j R, (1995): 722-735 Indeed, it is obvious that any -trajectory of
jx(t)j < 2M (R) for all t 2 [0; T ] and
(74)
jx(T )j < 2r :
(75) Proof: Assume on the contrary that there is a sequence k # 0 and partitions k with d(k ) k , and k trajectories xk () starting from BR such that (74) does not hold. This means that there is some sequence tk 2 [0; T ] such that
jxk (t)j < 2M (Ri ) 8 t 2 [0; tk ) ; jxk (tk )j = 2M (R) : (76)
[6] Clarke, F.H., Yu.S. Ledyaev, R.J. Stern, and P. Wolenski, \Qualitative properties of trajectories of control systems: A survey," J. Dynamical and Control Systems 1(1995): 1-48 . [7] Clarke, F.H., Yu.S. Ledyaev, and A.I. Subbotin, \Universal feedback strategies for dierential games of pursuit," SIAM J. Control , 1997, to appear. [8] Clarke, F.H., Yu.S. Ledyaev, and A.I. Subbotin, \Universal feedback via proximal aiming in problems of control and differential games," Rapport CRM-2386, Centre de Recherches Mathematiques, U. de Montreal, 1994. [9] Coron, J-M., \Global asymptotic stabilization for controllable systems without drift," Math of Control, Signals, and Systems 5(1992): 295-312. [10] Coron, J-M., \Stabilization in nite time of locally controllable systems by means of continuous time-varying feedback laws," SIAM J. Control and Opt. 33(1995): 804-833.
CLARKE, LEDYAEV, SONTAG, AND SUBBOTIN: ASYMPTOTIC CONTROLLABILITY IMPLIES FEEDBACK STABILIZATION
[11] Coron, J.-M., and L. Rosier, \A relation between continuous time-varying and discontinuous feedback stabilization," J.Math. Systems, Estimation, and Control 4(1994): 67-84. [12] Crandall, M., and P.-L. Lions (1983) \Viscosity solutions of Hamilton-Jacobi equations," Trans. Amer. Math. Soc. 277(1983): 1-42. [13] Filippov, A.F., \Dierential equations with discontinuous righthand side," Matem. Sbornik 5(1960): 99-127. English trans. in Amer. Math. Translations 42(1964): 199-231. [14] Fleming, W.H., and H.M. Soner, Controlled Markov Processes and Viscosity Solutions , Springer-Verlag, New York, 1993. [15] Hermes, H., \Discontinuous vector elds and feedback control," in Dierential Equations and Dynamic Systems (J.K. Hale and J.P. La Salle, eds.), Academic Press, New York, 1967. [16] Hermes, H., \On the synthesis of stabilizing feedback control via Lie algebraic methods," SIAM J. Control and Opt. 18(1980): 352-361. [17] Isidori, A., Nonlinear Control Systems, Third Edition , SpringerVerlag, London, 1989. [18] Krstic, M., I. Kanellakopoulos, and P. Kokotovic, Nonlinear and adaptive control design , John Wiley & Sons, New York, 1995. [19] Krasovskii, N.N., and A.I. Subbotin, Positional dierential games , Nauka, Moscow, 1974 [in Russian]. French translation Jeux dierentiels, Editions Mir, Moscou, 1979. Revised English translation Game-Theoretical Control Problems, SpringerVerlag, New York, 1988. [20] Ledyaev, Yu.S., and E.D. Sontag, \A remark on robust stabilization of general asymptotically controllable systems," submitted. [21] Ryan, E.P., \On Brockett's condition for smooth stabilizability and its necessity in a context of nonsmooth feedback," SIAM J. Control Optim. 32(1994): 1597-1604. [22] Sontag, E.D., \Nonlinear regulation: The piecewise linear approach," IEEE Trans. Autom. Control AC-26(1981): 346-358. [23] Sontag E.D., \A Lyapunov-like characterization of asymptotic controllability," SIAM J. Control and Opt. 21(1983): 462-471. [24] Sontag, E.D., \A `universal' construction of Artstein's theorem on nonlinear stabilization," Systems and Control Letters, 13(1989): 117-123. [25] Sontag E.D., Mathematical Control Theory, Deterministic Finite Dimensional Systems, Springer-Verlag, New York, 1990. [26] Sontag E.D., \Control of systems without drift via generic loops," IEEE Trans. Autom. Control 40(1995): 1210-1219. [27] Sontag, E.D., and H.J. Sussmann, \Remarks on continuous feedback," in Proc. IEEE Conf. Decision and Control, Albuquerque, Dec. 1980, IEEE Publications, Piscataway, pp. 916-921. [28] Sontag, E.D., and H.J. Sussmann, \Nonsmooth controlLyapunov functions," Proc. IEEE Conf. Decision and Control, New Orleans, Dec. 1995, IEEE Publications, 1995, pp. 27992805. [29] Subbotin, A.I., Generalized Solutions of First-Order PDEs: The Dynamical Optimization Perspective , Birkhauser, Boston, 1995. [30] Sussmann, H.J., \Subanalytic sets and feedback control," J. Differential Equations 31(1979: 31-52.
13