CONTROL OF NON-MINIMUM PHASE SYSTEMS USING EXTENDED HIGH-GAIN OBSERVERS By Mohammed “Shahid” Nazrulla A DISSERTATION Submitted to Michigan State University in partial fulfillment of the requirements for the degree of DOCTOR OF PHILOSOPHY Electrical Engineering 2010 ABSTRACT CONTROL OF NON-MINIMUM PHASE SYSTEMS USING EXTENDED HIGH-GAIN OBSERVERS By Mohammed “Shahid” Nazrulla Control schemes that achieve stabilization and output regulation in the case of nonminimum phase nonlinear systems are presented. These techniques utilize continuously implemented sliding mode control, and an extended high-gain observer to generate estimates of the output and its first ρ − 1 derivatives (ρ is the relative degree of the system), in addition to a signal that renders the zero dynamics of the given plant observable. The results include exponential stability of the origin of the output feedback system in the stabilization problem, exponential stability of the origin of the error system in the regulation problem, and asymptotic recovery of the performance of the state feedback designs in both instances. The stabilization and regulation problems pose unique challenges in the case of non-minimum phase nonlinear systems. This work provides a design methodology that can be incorporated into any robust control scheme. The design process conforms to the nonlinear separation principle, and comprises the following three steps. 1) Augment a dynamic stabilizing compensator to the system and moreover, design it such that the resultant system is minimum phase and relative degree one with respect to a virtual output to be constructed using the state estimates. 2) Use any robust control technique such as sliding mode control, saturated highgain feedback, Lyapunov redesign, etc. to design a static state feedback control law that stabilizes the relative degree one minimum phase system obtained in the previous step. 3) Design an extended high-gain observer (EHGO) to estimate the system output and its first ρ − 1 derivatives, and additionally, the coupling signal that is required to solve the auxiliary design problem stemming from step 1). Replace all the states and the coupling signal in the state feedback control design with their estimates, and if necessary, saturate the resulting control law outside a compact set of interest, or saturate the estimates obtained from the EHGO. Simulation examples are provided—in particular, the stabilization algorithm is applied to the Translational Oscillator with a Rotating Actuator (TORA) nonlinear benchmark system. DEDICATION To my mother Zeenath, who is Inspiration personified. iv ACKNOWLEDGMENT I would first and foremost like to acknowledge the invaluable help and guidance of my advisor, Professor Hassan Khalil. I am extremely fortunate to have had the opportunity to work so closely over the past several years with such a pre-eminent scholar as Dr. Khalil. His wisdom and his vast, seemingly boundless knowledge of our field were the most precious and useful resources of all. He has always been available to answer my numerous questions, and he often exhibited great patience with me. For his unwavering support and encouragement, I shall be everlastingly grateful. I would also like to thank my other committee members, Dr. Tan, Dr. Mukherjee and Dr. Choi, for their support and tremendously helpful feedback during my research activities. I would not be at this momentous juncture if not for my parents, Sanaulla M. Nazrulla and Zeenath Nazrulla, to whom I owe everything. They both unfailingly led by example when it came to teaching and encouraging me to always think for myself, and this more than any other factor has brought me to the place where I find myself today. My brother Maajid, although younger than me by a few years, has nonetheless had much to teach me, and has inspired me to strive to better myself. It would be terribly remiss of me not to recognize the role played by each and every teacher I have had in my life, who along with my family and friends taught me everything I know. I owe a tremendous debt to every one of these individuals. I could not possibly list all of my friends, but I would particularly like to thank my cohorts, colleagues and friends in Attaullah, Alexis, Jeff, Luma, Jim, Chen, Ranran, Muataz, Mohamed Edardar and Joonho; our many discussions and conversations over the years helped me in more ways than I could describe in a few words. Finally, I would like to thank Prof. Weil for helping me typeset this document. Once again, thank you, one and all, for helping me get to where I am today. v TABLE OF CONTENTS List of Figures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . viii List of Tables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . x 1 Introduction . . . . . . . . . . . . . . . . . . . . . 1.1 Background and Preliminaries . . . . . . . . . . . . . . . 1.2 Some Definitions . . . . . . . . . . . . . . . . . . . . . . 1.3 Stabilization of Non-Minimum Phase Systems . . . . . . 1.4 Regulation of Non-Minimum Phase Systems . . . . . . . 1.5 Extended High-Gain Observers in Nonlinear Control . . 1.6 Sliding Mode Control . . . . . . . . . . . . . . . . . . . . 1.7 Organization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 1 6 8 15 22 24 26 2 Stabilization of Non-Minimum Phase Systems . . . . . . . . 2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2 Problem Statement . . . . . . . . . . . . . . . . . . . . . . . . . . 2.3 Controller Design for the State Feedback System . . . . . . . . . . 2.4 Output Feedback Control Using an Extended High-Gain Observer 2.4.1 Output Feedback Controller Design . . . . . . . . . . . . . 2.4.2 Exponential Stability and Trajectory Convergence . . . . . 2.5 Example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.5.1 Design of the Stabilizing Controller . . . . . . . . . . . . . 2.5.2 Numerical Simulations . . . . . . . . . . . . . . . . . . . . 2.6 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27 27 29 31 39 39 44 45 46 49 52 . . . . . . . 3 Output Regulation of Non-Minimum Phase Systems . . . . . . . 3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.2 Problem Description and Assumptions . . . . . . . . . . . . . . . . . 3.2.1 Background and Preliminaries . . . . . . . . . . . . . . . . . . 3.2.2 State Feedback Regulator Design . . . . . . . . . . . . . . . . 3.2.3 Output Feedback Control Using Extended High-Gain Observers 3.2.3.1 Output Feedback Regulator Design . . . . . . . . . . 3.2.3.2 Exponential Stability of the Output Feedback System and Trajectory Convergence . . . . . . . . . . . . . . 3.3 Example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.3.1 The Auxiliary Problem . . . . . . . . . . . . . . . . . . . . . . 3.3.2 The Output Feedback Regulator . . . . . . . . . . . . . . . . . 3.3.3 Numerical Simulations . . . . . . . . . . . . . . . . . . . . . . vi 56 56 59 59 63 72 72 77 78 80 81 81 3.4 3.3.3.1 Discussion . . . . . . . . . . . . . . . . . . . . . . . . Concluding Remarks . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 Stabilization and Regulation of the TORA System . . . . . . 4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.2 TORA Dynamics: A Simplified Model of a Dual Spin Spacecraft . 4.3 Stabilization of the TORA System . . . . . . . . . . . . . . . . . 4.3.1 Design of the Stabilizing Controller . . . . . . . . . . . . . 4.3.2 Output Feedback Controller for the TORA . . . . . . . . . 4.3.3 Output Feedback Design Based on Isidori [21] . . . . . . . 4.3.4 Numerical Simulations . . . . . . . . . . . . . . . . . . . . 4.3.5 Discussion of the Results . . . . . . . . . . . . . . . . . . . 4.4 Concluding Remarks . . . . . . . . . . . . . . . . . . . . . . . . . 5 Conclusions and Future Work . 5.1 Concluding Remarks . . . . . . 5.1.1 Main Contributions . . . 5.2 Future Directions . . . . . . . . . . . . . . . . Appendix A.1 Locally Lipschitz property of ∆0 /ε A.2 Solution of the PDE (2.14) . . . . . A.2.1 Properties of the solution . A.2.2 Details of the solution . . . A.3 Proof of Proposition 2.1 . . . . . . A.4 Proof of Proposition 3.1 . . . . . . Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81 83 . 88 . 88 . 89 . 92 . 94 . 97 . 98 . 98 . 100 . 107 . . . . 112 112 113 115 . . . . . . 117 117 118 118 119 122 124 . . . . . . . . . . . . . . . . . . . . . . . . . . 129 vii LIST OF FIGURES 2.1 Recovery of the Auxiliary System performance. . . . . . . . . . . . . 48 2.2 Recovery of the Partial State Feedback System performance . . . . . 49 2.3 A comparison between Isidori’s and the EHGO-based design . . . . . 50 2.4 A comparison of the control efforts . . . . . . . . . . . . . . . . . . . 51 2.5 A comparison between Isidori’s and the EHGO design for an = 0.86 . 52 2.6 A comparison between Isidori’s and the EHGO design for an = 1.04 . 53 2.7 Response of the EHGO design when an = 0.76 . . . . . . . . . . . . . 54 2.8 Response of the EHGO-based design when an = 1.05 . . . . . . . . . 55 3.1 Trajectories of z1 for the nominal value of mc . . . . . . . . . . . . . 82 3.2 Trajectories of z2 for the nominal value of mc . . . . . . . . . . . . . 83 3.3 Trajectories of y for the nominal value of mc . . . . . . . . . . . . . . 84 3.4 Trajectories of x2 for the nominal value of mc . . . . . . . . . . . . . 85 3.5 Comparison of the control efforts . . . . . . . . . . . . . . . . . . . . 86 3.6 The response of the EHGO design when mc = 0.777 kg . . . . . . . . 87 4.1 TORA schematic . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89 4.2 Recovery of the Reduced System performance by decreasing εa . . . . 100 viii 4.3 Recovery of the Auxiliary System performance by decreasing µ (εa = 0.01) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101 4.4 Recovery of the Partial State Feedback System’s performance by decreasing ε (µ = 0.1) . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102 4.5 Recovery of the Partial State Feedback System’s performance by decreasing ε (µ = 0.1) . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103 4.6 A comparison of the responses of the EHGO and Isidori designs . . . 104 4.7 A comparison of the control efforts . . . . . . . . . . . . . . . . . . . 105 4.8 A comparison of transient performance when y(0) = 0.086 . . . . . . 106 4.9 A comparison of the control effort when y(0) = 0.086 . . . . . . . . . 107 4.10 Response of the EHGO + SMC design when y(0) = 0.541 . . . . . . . 108 4.11 Control effort of the EHGO + SMC design when y(0) = 0.541 . . . . 109 4.12 Response under +100% perturbation in mc and mr . . . . . . . . . . 110 4.13 Response under −50% perturbation in mc and mr ix . . . . . . . . . . 111 LIST OF TABLES 4.1 TORA system parameters. . . . . . . . . . . . . . . . . . . . . . . . . 99 4.2 TORA stabilization: simulation parameters and initial conditions. . . 99 x Chapter 1 Introduction Engineers have always concerned themselves with applying humanity’s collective knowledge toward harnessing natural resources and phenomena in order to solve or to address commonly encountered technological challenges, and for the betterment of human society at large. Control engineering, with its interdisciplinary roots, is uniquely positioned among all of the engineering disciplines in that it facilitates the manipulation and control of large and complex physical systems, provided the relevant system models and associated control solutions are sufficiently accurate in some sense, not to mention tractable and amenable to economically and technologically feasible implementation methods. 1.1 Background and Preliminaries Control theory primarily addresses the problem of designing control algorithms for physical systems that can be modeled using techniques derived largely—although not exclusively—from classical physics and the theory of ordinary differential equations. We can encompass a very broad range of deterministic (as opposed to stochastic) 1 control problems by considering systems modeled by the equations x = f (t, x, u, w, θ), ˙ (1.1a) y = h(t, x, u, w, θ), (1.1b) where x ∈ D ⊂ Rn is the system state, u ∈ Rp is a control input to the system that we wish to design, w ∈ Rd is a vector containing disturbance and reference signals, θ ∈ Θ ⊂ Rp is a vector of constant unknown system parameters, y ∈ Rm given by the mapping h : [0, ∞) × D × Rp × Rd × Θ → Rm is the system output, and t ∈ [0, ∞) is a special scalar parameter—usually interpreted as time—that characterizes the trajectories that comprise the solution set of (1.1), provided one exists. A system that lacks an explicit dependence on time in the functions f (·) and h(·) is said to be autonomous or time-invariant, while a system like (1.1) above is said to be nonautonomous or time-varying. Finally, the mapping f : [0, ∞)×D×Rp ×Rd ×Θ → Rn is a vector-valued function that describes the dynamic behavior of the states x of the physical system (1.1) or plant, in engineering parlance. Generally speaking, the control task is to design the input u of the system (1.1) to, on the one hand, stabilize the latter in the absence of disturbances and/or reference signals w, or optionally, to also reject the effect of these disturbances, and/or to track any reference signals that happen to be contained in w. The control input could be a static function of the output measurements y and possibly also the reference signals, but it is often the case that it is necessary to design an additional dynamic system— broadly referred to as a compensator —to generate the desired signal u that is capable of achieving the control task; mathematically, this implies that the augmented system formed by the interconnection of this aforementioned dynamic compensator and the plant (1.1) has to be stable in some sense. A significant class of control problems admits solutions that require the control input u—also known as a control law —to 2 be a function of the plant states x, possibly in addition to the states of a dynamic compensator (if one is required). As can be seen from (1.1), the only signal that is available to the control designer is the output measurement y, so in order to implement a control law that needs to be a function of the state vector x, it is often necessary to construct a dynamical system called a state estimator or observer that is capable of reconstructing the states from the output measurements. Thus, observer-based output feedback control designs can be considered to fall under the broader class of control schemes in which the input u is generated by combining output measurements y with the states of a dynamic compensator. The advantage in the case of a certain class of systems that allow for observerbased control schemes is the idea of the separation principle—in essence, for this class of systems, it may be possible to separate the observer and control designs into two distinct sub-problems. In the first step, the control law u is designed as a function of the states x and if necessary, the states of a dynamic compensator. Such a law is referred to as a state feedback control law. In the next step, an observer is designed using the output y and our knowledge of the plant model in order to generate estimates of the states x. In the final stage, the results of the previous steps are combined by substituting the estimates of x into the control law, thus yielding an output feedback control law that is capable of controlling the plant. This type of solution is often simpler or easier to develop and to implement than the alternative of attempting to design a control scheme using only the output y, and without the benefit of the state estimates—indeed, in the absence of the latter, the control problem could very well be intractable. In the sequel, we shall revisit the idea of the separation principle for a certain type of observer-based nonlinear control system, and discuss some of the differing interpretations of this concept. If both f (t, x, 0, 0, θ) and h(t, x, 0, 0, θ) happen to be linear in x, then (1.1) is said to be a linear system. Similarly, if u is linear in x, and the compensator subsystem 3 that generates u—if one is present—is also linear in its state variables, then the interconnection of the compensator with the plant (1.1), called the closed-loop system, will likewise be linear in the resulting augmented state vector, which basically comprises the states of the plant as well as the compensator. The theory of linear systems, which is essentially an amalgamation of the theory of linear differential equations and results from complex analysis and frequency domain functional analysis, has been thoroughly researched for a period spanning most of the past two centuries. In practice, a significant majority of control designs tend to rely upon results from linear system theory. Many systems can be modeled adequately using linear differential equations. Even in the event that nonlinearities or model uncertainties are present, the theory of robust linear control has successfully addressed many types of problems in which systems can be modeled as being essentially linear, but incorporating certain types of structural model uncertainties. Thus, linear system theory has historically been very successful in addressing many issues pertaining to linear and quasi-linear systems, and robust control problems. Linear systems and control will be revisited periodically in the discussion to follow. Despite the ubiquity of linear control methods in engineering applications, it is often the case that nonlinearities are inherent in certain physical systems such as robots or physical systems encountered in the study of the mechanics of rigid bodies, flexible structures, etc., while other systems contain components that exhibit severe nonlinearities, either by happenstance or design, that diverge very quickly from linear approximations. One of the most common examples of the latter type of nonlinearity is saturation, which simulates physical limitations on the so-called control effort (effectively, bounds on the magnitude and/or time rate of change of the control signal u). For instance, the input voltage to a motor—or practically any electrical, electromechanical or electromagnetic device, for that matter—if increased beyond some critical value, will often cause actuator saturation. Other examples 4 include, but aren’t limited to, hysteresis, dead bands, backlash nonlinearities, switching or relay action, etc. Thus, control schemes that take into account these kinds of nonlinear phenomena might be able to produce superior results compared to linear control methods—examples where this has proven true include electromechanical systems, robots, aerospace applications, digital and analog electronic devices, smart material-based or other types of systems with significant model nonlinearities, etc. Nonlinear control has been a very active area of research for the past four decades. Prior to the 1970s, most results in nonlinear control addressed challenges pertaining to rigid body mechanics, nonlinear circuit and electronic elements, aeronautics and aerospace applications, and these problems were largely analyzed using techniques such as the describing functions method and phase plane analysis, or were posed and solved in an optimal control framework [4]. Research into this field gathered pace during the seventies, and the year 1978 saw the publication of the first major textbook on nonlinear system theory by Vidyasagar [42]. Since then, many other books on nonlinear systems and control have been published, serving to collect all the major results and design techniques up to that point—of particular note are the books by Isidori [19], first published in 1985, Khalil [29] (first edition: 1992) and Sastry [38] (1999). In addition to the aforementioned books that provide a broad coverage of Lyapunov stability theory (which forms the backbone of stability analysis for nonlinear systems), standard nonlinear control design techniques, and elements of differential geometry needed for the exposition of the control theoretic results contained therein, many other books and monographs have been published on specific topics within the purview of nonlinear systems, such as robust nonlinear control, adaptive control, regulation, etc. Some of these books will be mentioned while discussing these particular topics. In contrast, linear system theory and control had been thoroughly researched, and many major results had been published by the 1970s. The globally linear nature 5 of linear spaces naturally led to global results, unlike the case of nonlinear control, where local or semiglobal results are much more common. A semiglobal result (such as stabilization) is one that can be achieved on a compact set of finite size containing the desired equilibrium set, but the former can be made arbitrary large in certain cases by appropriate choice of some design parameter such as controller gain, for instance. A watershed moment in modern control theory occurred circa 1960, with the publication of a pair of seminal papers by Kalman, the first of which introduced the concepts of controllability and observability [24]. Indeed, the juncture at which these papers were published is often regarded as the dividing line between classical and modern control theory. Over the next couple of decades, considerable advances were made in the development of robust control schemes for linear systems, culminating in the publication of a paper on H∞ control by Zames in 1981 [44]. 1.2 Some Definitions We take this opportunity to define a few terms before we discuss the types of problems that are tackled in this work. To this end, let us consider the following autonomous system x = f (x, u), ˙ (1.2) where f : D × R → Rn is a locally Lipschitz map from a domain (D × R) ⊂ (Rn × R) into Rn . Suppose there exists a state feedback control law u = Γ(x), (1.3) where Γ : D → R. Also, suppose that x∗ is an equilibrium point of the closed-loop system (1.2)–(1.3); consequently, we have f (x∗ , Γ(x∗ )) = 0. We state the definitions of the stability of the equilibrium point x∗ (or the lack thereof) and some related 6 concepts below. The following definition is taken from [29]. Definition 1.1. The equilibrium point x = x∗ of (1.2)–(1.3) is • stable if, for every ε > 0, there exists a δ = δ(ε) > 0 such that x(0) − x∗ < δ ⇒ x(t) − x∗ < ε, ∀t ≥ 0 • unstable, if it is not stable • asymptotically stable, if it is stable and δ can be chosen such that x(0) − x∗ < δ ⇒ lim x(t) = x∗ t→∞ Let φ(t; x) be a solution of (1.2)–(1.3) that starts at initial state x at time t = 0. When the equilibrium point x = x∗ of (1.2)–(1.3) is asymptotically stable, its region of attraction is defined as the set of all points x such that φ(t; x) is defined for all t ≥ 0 and limt→∞ φ(t; x) = x∗ . If it is the case that for any initial state x, the trajectory φ(t; x) approaches x∗ as t → ∞ no matter how large x − x∗ gets, then the region of attraction is the whole space Rn . If an asymptotically stable equilibrium point x∗ possesses this aforementioned property, it is said to be globally asymptotically stable. For the purposes of the next definition, we suppose that the unforced version of system (1.2), i.e. x = f (x, 0), ˙ (1.4) has an asymptotically stable equilibrium point at the origin x = 0. The following definition is adapted from [29]. Definition 1.2. The system (1.2) is said to be regionally input-to-state stable if there exist a class KL function β and a class K function γ such that, for any initial state 7 x(0) ∈ X , where X ⊂ Rn is a compact set, and for any bounded input u(t), the solution x(t) exists for all t ≥ 0 and satisfies x(t) ≤ β( x(0) , t) + γ sup u(τ ) 0≤τ ≤t . (1.5) Inequality (1.5) guarantees that for any bounded input u(t), the state x(t) will be bounded. Furthermore, as t increases, the state x(t) will be ultimately bounded by a class K function of supt≥0 u(t) . It can also be shown that if u(t) → 0 as t → ∞, then limt→∞ x(t) = 0. Since, with u(t) ≡ 0, (1.5) reduces to x(t) ≤ β( x(0) , t), regional input-to-state stability implies that the origin of the unforced system (1.4) is asymptotically stable. Moreover, if inequality (1.5) holds for any initial state x(0) ∈ Rn , then we drop the modifier “regionally” from the above definition and state that the system (1.2) is simply input-to-state-stable, which in turn implies that the origin of (1.4) is globally asymptotically stable. 1.3 Stabilization of Non-Minimum Phase Systems The stabilization problem for non-minimum phase systems (to be defined shortly) forms the crux of this work. In order to introduce the concepts and terminology associated with this problem, we shall consider single-input, single-output (SISO— i.e. u ∈ R and y ∈ R) autonomous systems that have the following structure. xp = f (xp ) + g(xp )u, ˙ y = h(xp ). 8 (1.6a) (1.6b) All the other variables apart from u and y belong to the domains defined above for the system (1.1), and the mappings f : D → Rn , g : D → Rn , and h : D → R are assumed to be sufficiently smooth for all xp ∈ D. Moreover, f (0) = 0. In the special case of linear time-invariant systems, we can take f (xp ) = Axp , g = B, and h(xp ) = Cxp , which gives xp = Axp + Bu, ˙ y = Cxp , (1.7a) (1.7b) where A ∈ Rn×n , B ∈ Rn and C ∈ R1×n . The system equations (1.7) are referred to as the state-space realization for a linear system with input u and output y. The input-to-output map for (1.7) can also be represented in transfer function form as H(s) = bm sm + bm−1 sm−1 + . . . + b0 Y (s) = , U (s) sn + an−1 sn−1 + . . . + a0 (1.8) where m < n and bm = 0, and the polynomials Y (s) and U (s) are the Laplace transforms of the time domain signals y(t) and u(t), respectively. The difference between the degrees of the numerator and denominator polynomials above, n − m, is called the relative degree of the linear system described by H(s), or equivalently, by (1.7). Given an input-output map for a linear system such as (1.8), several non-unique, albeit equivalent state-space realizations of Y (s)/U (s) exist in the time domain. Two of the most common state-space realizations of transfer functions are the controllable canonical form and the observable canonical form. The controllable canonical representation of (1.8) is given by ˙ xp = Ac xp + Bc u, ˜ ˜ y = C c xp , ˜ 9 (1.9a) (1.9b) where the triple {Ac , Bc , Cc } is given by         Ac =         0 1 0 ··· 0 0 0 0 1 ··· 0 0 0 . . . 0 . . . 0 . . . ··· ... 0 . . . 0 . . . 0 0 0 ··· 0  1 −a0 −a1 −a2 · · · Cc = −an−2 −an−1 b0 b1 · · · bm 0 · · ·        ,          0   0       0 Bc =  .  , . .     0     1 0 , where the coefficients ai and bj , 0 ≤ i ≤ n − 1 and 0 ≤ j ≤ m, are the same as the ones that appear in (1.8). We note that   C Ai x ,  c c ˜p 1 ≤ i ≤ n − m − 1, (i) = y  C An−m x + C An−m−1 B u, i = n − m,  c c ˜p c c c and consequently, the relative degree ρ = n − m can be interpreted as the minimum number of times the output y has to be differentiated in order for the control input u to show up in the resulting expression. If all the roots of the numerator polynomial of H(s) are in the open left half plane, then (1.7) is said to be minimum phase, or non-minimum phase otherwise. The above interpretation of relative degree in terms of successive derivatives of the output is also applicable to a certain class of nonlinear systems in the form of (1.6). Indeed, the derivative of the output is y = Lf h(xp ) + Lg h(xp )u, where ˙ Lf h(xp ) ∂h(xp ) f (xp ) ∂xp 10 is called the Lie derivative of h with respect to f , and can be interpreted as the total time derivative of h along the trajectories of f . Now, if Lg h(xp ) = 0, then y will be ˙ independent of u, in which case we could continue calculating successive derivatives of y until u finally appears, and the result would be as follows.   i  L h(xp ), 1 ≤ i ≤ ρ − 1, f (i) = y  Lρ h(x ) + L Lρ−1 h(x )u, i = ρ,  p g f p f for some positive integer ρ, in analogous fashion to the linear case. It is apparent from the above equation that the state feedback control ρ v − L h(xp ) f , u= ρ−1 Lg L h(xp ) f which can be implemented if f , g and h are known precisely, will reduce the inputoutput map to y (ρ) = v, which is a chain of ρ integrators, and hence this number is referred to as the relative degree of the system (1.1). We note that the system will have a well defined relative ρ−1 h(xp ) = 0, for all xp ∈ D0 . degree in a region D0 ⊂ D, only if Lg L f The above procedure of taking successive derivatives of the output until the input appears suggests that we could define a transformation T : D0 → D0 as [29]   z    x xt = T (xp ) =   ϕ(xp )   π(xp ) ϕ1 (xp ) · · · ϕn−ρ (xp ) 11 h(xp ) · · · L ρ−1 h(xp ) f ⊤ , (1.10) where ϕ1 to ϕn−ρ satisfy the partial differential equations ∂ϕi g(xp ) = 0, for 1 ≤ i ≤ n − ρ, ∀xp ∈ D0 . ∂xp (1.11) The functions ϕ1 to ϕn−ρ can be shown to exist at least locally, and moreover, can be chosen so as to make (1.10) a diffeomorphism on D0 [29, Theorem 13.1]. A diffeomorphism is a continuously differentiable map that also has a continuously differentiable inverse. The change of variables (1.10) thus transforms (1.6) into z = f0 (z, x), ˙ (1.12a) x = Ax + B[b0 (z, x) + a0 (z, x)u], ˙ (1.12b) y = Cx, (1.12c) where x ∈ Rρ , z ∈ Rn−ρ , (A, B, C) is a chain of ρ integrators in canonical form, and ∂ϕ f (xp ) , ∂xp xp =T −1 (xt ) ρ , b0 (z, x) = L h(xp ) f xp =T −1 (xt ) f0 (z, x) = a0 (z, x) = Lg L ρ−1 . h(xp ) f xp =T −1 (xt ) Equation (1.12) is called the normal form representation of the system (1.6). In this form, the system comprises an external part given by the x-equation in (1.12), ˙ and an internal part given by the z-equation. If the functions a0 (·) and b0 (·) are ˙ known, then the external part of (1.12) can linearized by the state feedback control u = [v − b0 (z, x)]/a0 (z, x), but we should also note that the internal part will be made unobservable from the output y by this same control law. Setting x = 0 in the 12 internal dynamics of (1.12) gives z = f0 (z, 0), ˙ (1.13) which is referred to as the zero dynamics of the system (1.6). The system is said to be minimum phase if the equilibrium point of (1.13) is asymptotically stable, and non-minimum phase otherwise. In this work, we are interested in designing output feedback stabilizers and regulators for systems that are potentially non-minimum phase. There are several applications that give rise to system models that can be expressed in the form of (1.12), and these have been documented and analyzed extensively in the literature. Common sources of such models are mechanical and electromechanical systems in which the outputs comprise position or displacement variables, but not their derivatives (velocities, accelerations, and so forth)—typical examples of such models include induction motors [30], an inverted-pendulum-on-acart system [29], and the translational oscillator with a rotating actuator (TORA) benchmark system [6]. The latter two examples happen to be non-minimum phase, as well—the open loop TORA system has a pair of conjugate poles on the imaginary axis, while the inverted pendulum system has a pole in the right half of the complex plane. Concerning non-minimum phase systems, it should be noted that historically, no real distinction has been necessary between the aforementioned type of plant or a minimum phase plant in developing control design techniques for linear systems, as long as the model is assumed to be completely known. The only difference of note between the two types of systems is the well known fact that right half plane zeros tend to reduce the gain and phase margins of the closed loop system, thus imposing constraints upon the performance measures that can be achieved by the design. For 13 instance, [17] sets forth trade-offs that are inherent in the controller designs for plants with right half plane zeros and poles, and the limitations imposed on the sensitivity and complementary sensitivity functions of the closed loop system. Reference [45] extends the so-called loop transfer recovery (LTR) procedure to non-minimum phase linear plants, and shows that when applying LTR to a non-minimum phase system, there is a trade-off between the feedback properties of the state feedback loop and the ability to recover these properties in the output feedback design. Non-minimum phase systems pose a much greater design challenge when the plant dynamics are nonlinear. Indeed, it is fairly common in the literature pertaining to nonlinear systems to assume that the plant under consideration is minimum phase— specifically, that the zero dynamics are either input-to-state stable or globally asymptotically stable. A few papers on the topic of output feedback control of non-minimum phase systems have been published in the past decade, however. Specifically, the papers [25, 34, 21, 36] deal with the case of stabilization of an equilibrium point. Reference [21] in particular introduced a high-gain feedback based design tool for a very large class of systems, possibly non-minimum phase, in the normal form. The extension of the aforementioned technique to render it independent of any particular control scheme and to show that the resulting design is robust to uncertainties in the control coefficient (the signal a0 (·) in (1.12)) and in other system parameters forms the crux of this dissertation. This extension is possible by utilizing a so-called extended high-gain observer (EHGO) to estimate the signal b0 (·) in (1.12), which renders the zero dynamics observable if this same signal is viewed as the output of a reduced order auxiliary system that incorporates the zero dynamics of (1.12). In accordance with Isidori’s design tool [21], if a stabilizing controller is known for the auxiliary system, then it is possible to construct an output feedback controller that stabilizes (1.12). While Isidori used a high-gain feedback design, the EHGO-based output feedback controller in our work may incorporate any robust control scheme 14 such as sliding mode control (SMC), saturated high-gain feedback, etc. The stabilization problem is tackled in Chapter 2. 1.4 Regulation of Non-Minimum Phase Systems The next chapter of this dissertation focuses on the problem of output feedback regulation of non-minimum phase systems. The regulation problem is a special case of the output tracking and disturbance rejection problems. In the latter case, the control design is required to make the output of the system match an external refernce signal and to reject unknown disturbance signals. The objective is the same in the regulation problem, but the external signals are assumed to be sinusoidal with known frequencies or constant—this allows for the incorporation of an internal model into the system, which is capable of asymptotically generating the reference signal to be tracked and the disturbances to be rejected. As is the case with unstable zero dynamics, there are some unique challenges when designing regulators for nonlinear systems as opposed to the linear case. The regulation problem for linear systems was solved in the 1970s. As mentioned earlier, no distinction is necessary between minimum and non-minimum phase plants in the design of output feedback regulators for linear plants; moreover, it is possible for the internal model to have poles in the right half complex plane, unlike in the case of nonlinear systems, whose internal model poles are confined to the imaginary axis. The key papers on output feedback regulation of linear systems are by Francis and Wonham [15], who introduced the internal model principle, and by Davison [12]. Generally speaking, the presence of nonlinearities in a system model results in the generation of higher order harmonics of any sinusoidal signals that enter into the system. Consequently, the internal model must reproduce these higher order harmonics if we are to successfully solve the regulation problem—this issue is explained in 15 further detail in [26]. Nevertheless, regulation theory is fairly well developed even in the case of nonlinear systems, and the major results have been collected in standard reference works and textbooks such as [20, 8, 29, 18], to name a few. On the other hand, just as with the stabilization problem, there are fewer results in the specific case of non-minimum phase systems. This is not surprising since the inclusion of a marginally stable internal model contributes additional marginally stable dynamics to the overall system, which specifically manifest themselves as critically stable zero dynamics (in addition to others that may come from the plant) that are not observable from the output tracking error. Some recent papers dealing with unstable zero dynamics are [32, 10], and they represent extensions of the design tool of [21] to the regulation problem. All these papers utilize a high gain feedback controller design and do not consider uncertainty in the control coefficient. In Chapter 3 of this dissertation, we seek to extend the idea of [21] in analogous fashion to Chapter 2 by utilizing an extended high-gain observer, which allows for control designs other than high gain feedback, and we also consider uncertain control coefficients. In this work, the regulation problem is tackled in Chapter 3 for single-input, single-output systems of the form [27] xp = f (xp , θ) + g(xp , θ)u + ϕ(xp , d, θ), ˙ y = h(xp , θ) + γ(d, θ) − yr , (1.14a) (1.14b) where xp ∈ Rn is the system state, u ∈ R is the input, y ∈ R is the measured output, d ∈ Rp is a time varying disturbance input and yr ∈ R is a time varying reference signal, with both yr and d having known models, and θ is a vector of uncertain constant parameters that belongs to a compact set Θ ⊂ Rl . The functions f , g, h, ϕ and γ all depend continuously on θ, and for all θ ∈ Θ, they are sufficiently smooth in xp and d for all xp ∈ Dθ , an open connected subset of Rn that could depend on θ, 16 and all d in a compact set of interest. Moreover, ϕ(x, 0, θ) = 0 and γ(0, θ) = 0, for all θ ∈ Θ and xp ∈ Dθ . The control input u is to be designed so as to make the output y go to zero. The disturbance-free system (1.14), with d = 0, is assumed to have a uniform relative degree ρ ≤ n for all θ ∈ Θ and xp ∈ Dθ , that is, Lg h(xp , θ) = Lg Lf h(xp , θ) = · · · = Lg L ρ−2 h(xp , θ) = 0, f and Lg L ρ−1 h(xp , θ) ≥ κ > 0, f where κ is independent of θ. Furthermore, we assume that there exists a diffeomorphism T : D θ × Θ → Dθ ,   z  T : xp →   x (1.15a) (1.15b) that transforms (1.14), with d = 0, into the normal form z = f0 (z, x, θ), ˙ xi = xi+1 , ˙ 1 ≤ i ≤ ρ − 1, xρ = b0 (z, x, θ) + a0 (z, x, θ)u, ˙ y = x1 − yr . (1.16a) (1.16b) (1.16c) (1.16d) We note that (1.16) will be identical to the normal form equations (1.12) above in the absence of uncertainties θ. Systems transformable into the normal form (1.16) are studied in Chapter 2, and therefore the results therein are applicable to any 17 disturbance-free system of the form (1.14) (i.e. d = 0 and yr = 0 in the stabilization problem). When d = 0, we assume that (1.15) transforms the disturbance-driven system (1.14) into the form z = fd (z, x1 , . . . , xm+1 , d, θ), ˙ (1.17a) xi = xi+1 + ψi (x1 , . . . , xi , d, θ), ˙ 1 ≤ i ≤ m − 1, m > 1, (1.17b) xi = xi+1 + ψi (x1 , . . . , xi , d, θ), ˙ 1 < m ≤ i ≤ ρ − 1, (1.17c) xρ = b0 (z, x, θ) + a0 (z, x, θ)u + ψρ (z, x, d, θ), ˙ (1.17d) y = x1 + γ(d, θ) − yr , (1.17e) where 1 ≤ m ≤ ρ − 1, and the functions ψi vanish at d = 0. Let ρ0 be the disturbance relative degree, that is, the disturbance appears for the first time in y (ρ0 ) , and set ρ∗ = ρ − ρ0 . Also, let D(t)    d(t)    .  ∈D , . .  d   ∗  d(ρ ) (t) Y    yr (t)   .   .  ∈ Y,  .    (ρ) yr (t) ∗ where Dd ∈ R(ρ +1)p and Y ∈ Rρ+1 are compact sets. We assume that D(t) and Y (t) are generated by the exosystem  w = S0 w, ˙  D    = Γ0 w, Y (1.18a) (1.18b) where S0 ∈ Rnw ×nw has distinct eigenvalues on the imaginary axis, and w(t) belongs 18 to a compact set W ⊂ Rnw . We now define the functions x1 (w, θ) through xm (w, θ) ¯ ¯ as x1 = yr − γ(d, θ), ¯ ∂ xi ¯ S w − ψi (¯1 , . . . , xi , d, θ), x ¯ xi+1 = ¯ ∂w 0 1 ≤ i ≤ m − 1. Next, we assume there exists a unique mapping z (w, θ) that solves the partial differ¯ ential equation ∂z ¯ S w = fd (¯, x1 , . . . , xm , (∂ xm /∂w)S0 w − ψm (¯, x1 , . . . , xm , d, θ), d, θ), (1.19) z ¯ ¯ ¯ z ¯ ¯ ∂w 0 for all w ∈ W . Using z (w, θ), we now define xm+1 (w, θ) to xρ (w, θ) as ¯ ¯ ¯ xi+1 = ¯ ∂ xi ¯ S w − ψi (¯, x1 , . . . , xi , d, θ), z ¯ ¯ ∂w 0 m ≤ i ≤ ρ − 1. The steady-state zero-error manifold is given by {z = z (w, θ), x = x(w, θ)}. The ¯ ¯ steady-state value of the control input on this manifold is given by χ(w, θ) = 1 [(∂ xρ /∂w)S0 w − b0 (¯, x, θ) − ψρ (¯, x, d, θ)]. ¯ z ¯ z ¯ a0 (¯, x, θ) z ¯ (1.20) Suppose there exists a set of real numbers c0 , . . . , cq−1 , independent of θ, such that χ(w, θ) satisfies the identity q q−1 Ls χ = c0 χ + c1 Ls χ + · · · + cq−1 Ls χ, for all (w, θ) ∈ W × Θ, where Ls χ = (∂χ/∂w)S0 w, and the characteristic polynomial equation λq − cq−1 λq−1 − · · · − c1 λ − c0 = 0 19 has distinct roots on the imaginary axis. By defining  0 1 0  0 0 1    0 0 0 S=. . . . . . . . .   0 0 0   c0 c1 c2 Γ= ··· 1 0 0 ···  0 0   ··· 0 0     ··· 0 0   . . , ... . .  . .   ··· 0 1    · · · cq−2 cq−1 0 0 1×q ,    χ     L χ   s     2   Ls χ  τ =  . ,  .   .     q−2  L   s χ   q−1 Ls χ it can be shown that χ(w, θ) is generated by the internal model ∂τ (w, θ) S0 w = Sτ (w, θ), ∂w χ(w, θ) = Γτ (w, θ). In order to convert the regulation problem to one of stabilizing an equilibrium point, we apply the change of variables η = z − z (w, θ), ¯ ei = y (i−1) , (1.21a) 1 ≤ i ≤ ρ. (1.21b) By a series of calculations, it can be verified that e1 = x1 − x1 (w, θ), ¯ ei = xi − xi (w, θ) − ϑi (e1 , . . . , ei−1 , w, θ), ¯ ei = xi − xi (w, θ) − ϑi (η, e1 , . . . , ei−1 , w, θ), ¯ 20 2 ≤ i ≤ m, m + 1 ≤ i ≤ ρ, where the functions ϑi vanish at (η, e) = (0, 0). With the change of variables (1.21), we obtain the following representation for system (1.17) in the new variables. η = f (η, e, w, θ), ˙ ei = ei+1 , ˙ 1 ≤ i ≤ ρ − 1, eρ = b(η, e, w, θ) + a(η, e, w, θ)u, ˙ y = e1 . (1.22a) (1.22b) (1.22c) (1.22d) The functions f (·), a(·) and b(·) satisfy f (0, 0, w, θ) = 0, a(0, 0, w, θ) = a0 (¯(w, θ), x(w, θ), θ) z ¯ b(0, 0, w, θ) = −χ(w, θ)a0 (¯(w, θ), x(w, θ), θ). z ¯ In these new variables, the zero-error manifold is given by {η = 0, e = 0}. The above procedure shows how a system of the form (1.14) can be converted to error coordinates, with a view toward obtaining a solution to the regulation problem. In Chapter 3, we consider a special case of the system (1.14), and we begin the analysis and the design of the regulator for that system with normal form equations that are a special case of (1.17). Specifically, the normal form equations assumed for the system studied in Chapter 3 do not have any perturbation terms in the chain of integrators, unlike equations (1.17) above. 21 1.5 Extended High-Gain Observers in Nonlinear Control A common feature of all the designs presented in this work is the utilization of the extended high-gain observer (EHGO) to estimate the output and its derivatives up to the ρth order, which is the relative degree of the system, plus one additional signal, namely b0 (·) in (1.12) or a perturbed version of this signal, which renders the zero dynamics observable if this function is viewed as a virtual output of the system. The EHGO was also utilized by Freidovich and Khalil [16] as a disturbance observer, and the analyses in this work share some similarities with this paper. During the analysis of both the stabilization and regulation designs, it is shown that the output feedback system has exactly the same structure as the more general high-gain observer based closed loop systems studied by Atassi and Khalil [2, 3], and consequently, the EHGO-based designs in the subsequent chapters conform to the separation principle of Atassi and Khalil, in which the output feedback design exhibits the following three properties: a) recovery of the stability of the state feedback design, b) asymptotic recovery of the region of attraction of the state feedback case, and c) closeness of trajectories of the output and state feedback designs. An important consideration to take into account when designing control schemes that utilize high-gain observers is the so-called peaking phenomenon that occurs when the high-gain parameter ε > 0, which is required to be small, is made sufficiently small [14, 29]. On the one hand, it can be shown by means of singular perturbation analysis [29] that reducing ε has the effect of reducing or even eliminating the effect of uncertainties or disturbances in the system, but on the other hand, the peaking phenomenon will inevitably manifest itself as ε is made progressively smaller. This phenomenon is essentially caused by an impulse-like response in the observer states that typically gets transmitted to the plant via the control input, which is often a 22 function of the state estimates obtained from the observer. Consequently, this results in the somewhat counter-intuitive behavior of the output feedback system relative to the state feedback system, whereby the trajectories of the former start to deviate further away from those of the latter as ε is reduced and the peaking phenomenon becomes more pronounced. It is possible to mitigate the above-mentioned effect of the peaking phenomenon by saturating the control, and possibly also the state estimates themselves, if necessary, outside a compact set of interest. This has the effect of creating a buffer that protects the plant from peaking. Notwithstanding the saturation of the output feedback control, the performance of the state feedback control can still be recovered by judicious choices of the high-gain observer parameter ε and the saturation level for the control. The reason this can be achieved is as follows: a small enough ε ensures that there is a separation of time scales between the states of the observer, which will exhibit fast responses, and those of the plant, which comprise the slow dynamics of the overall system. Once saturation has been incorporated into the output feedback design, the control will saturate during the peaking period. However, since the peaking period shrinks to zero as ε tends to zero, for sufficiently small ε, the peaking period becomes so small that the states of the plant remain close to their initial values, even as the observer states peak rapidly and then quickly approach the state trajectories of the plant. Hence, subsequent to the peaking period, the state estimation errors become O(ε) and the output feedback control becomes close to the state feedback control. Therefore, when the output feedback control is either globally bounded or made to saturate outside a compact set, the trajectories of the resulting closed-loop system asymptotically approach those of the state feedback system as ε tends to zero. Indeed, it was proved by Atassi and Khalil [2, 3] that this technique of combining high-gain observers with globally bounded state feedback designs, which was originally proposed by Esfandiari and Khalil [14], 1) leads to a separation prin23 ciple that is independent of the state feedback design, 2) facilitates the recovery of the region of attraction and the trajectories achieved under state feedback, and 3) is robust to certain model uncertainties. 1.6 Sliding Mode Control Another common thread in this work is the use of continuously-implemented sliding mode control. Firstly, sliding mode control is a scheme in which a virtual signal, typically denoted s, is defined as a function of the system states, and possibly compensator states, if the latter exists. The control law basically comprises the product of an envelope function and a switching term sgn(s). Switching occurs about s = 0, and one objective of sliding mode control is to drive the system such that s = 0 within a finite amount of time and for all time thereafter; consequently, the set of all system states satisfying s = 0 is referred to as the sliding manifold. Continuouslyimplemented sliding mode control is essentially an approximation of sliding mode control, whereby the switching term sgn(s) is approximated by sat(s/µ), where µ > 0 is a design parameter that is typically chosen to be small. When s > µ, the signum and saturation functions both evaluate to ±1, and hence the difference between the two arises when s ≤ µ—the set of all system states that satisfy this inequality is referred to as the boundary layer. Using this terminology, inside the boundary layer, the continuous sliding mode control acts as high-gain feedback. The saturation function approximates the switching achieved by the signum function about the origin if µ, which controls the slope of the linear region of the saturation function, is chosen to be small enough. Indeed, as µ → 0, the saturation function approaches the signum function. While sliding mode control is a popular control scheme for nonlinear systems due to its robustness properties, it tends to suffer from a phenomenon known as chattering, 24 especially in most practical applications. In an ideal system, the trajectories under sliding mode control are expected to start sliding along the manifold s = 0 soon as this manifold is reached. However, in reality, there will be a delay between the time the sign of s changes and the time the control reacts to this change and effects the switch. During this delay period, the trajectory passes from a region, say s > 0, through the manifold s = 0 and beyond, so that now, s < 0. When the control eventually switches, the trajectory reverses direction and heads back toward the manifold, but delays in the control switching will ensure that it will overshoot and pass through to the other side of the manifold yet again. This process repeats several times over a rather short period of time and results in very high frequency oscillations in the immediate vicinity of the sliding manifold, and this is what is referred to as chattering. Chattering results in persistent oscillations and low control accuracy, high heat losses in electrical circuits, and considerable wear and tear of mechanical components such as relays, motors, etc. It may excite unmodeled high-frequency dynamics, which can not only degrade the performance of the system, but may also lead to instability [29]. Continuously-implemented sliding mode control can eliminate chattering because no switching occurs inside the boundary layer—instead, the system will operate under linear feedback within this region. The aforementioned fact necessitates the design of this type of control scheme in two parts, one outside the boundary layer, where we typically seek to make this control identical to sliding mode control, and the other component is the design of the linear feedback control inside the boundary layer. This scheme generally does not recover the equilibrium point of sliding mode control, but instead achieves ultimate boundedness of the trajectories with an ultimate bound proportional to the size of the boundary layer thickness parameter µ. Consequently, this design can guarantee that the trajectories will reach a small set containing the desired equilibrium point, but what happens inside that set is problem dependent. As µ is made smaller, the ultimate bound decreases, as well, but pushing µ too small 25 will result in chattering, which this techique is often chosen to avoid in the first place. There are certain instances in which the origin of the closed-loop system can be made asymptotically stable under continuously-implemented sliding mode control, provided the parameter µ is chosen small enough. This is the case for the stabilization and regulation problems studied in Chapters 2 and 3, respectively. 1.7 Organization This dissertation is organized as follows. Chapter 2 presents an output feedback controller design to stabilize the origin of a non-minimum phase system. Chapter 3 deals with the regulation problem for non-minimum phase systems, which is solved by introducing a servocompensator driven by the tracking error and then designing a stabilizing compensator to stabilize the resultant system, in a manner that is analogous to the technique of Davison [12] for linear systems. While each chapter ends with a simulation example, it is always helpful to demonstrate the designs on a non-trivial application example. Consequently, in Chapter 4, the stabilization design of Chapter 2 is implemented on the translational oscillator with a rotating actuator (TORA) benchmark system, and simulation results are provided. Finally, Chapter 5 contains concluding remarks and a discussion of possible future directions of this work. 26 Chapter 2 Stabilization of Non-Minimum Phase Systems 2.1 Introduction The problem of stabilizing nonlinear systems via output feedback has drawn much attention in recent years. On the basis of certain assumptions made about the system structure, various control schemes have achieved global and semi-global results. A common assumption in many of these results, however, calls for the zero dynamics of the system under consideration to be either input-to-state stable, or globally asymptotically stable. While stabilization results abound for various classes of minimum phase nonlinear systems, there have been a few notable developments in the case of non-minimum phase systems in recent times, as well. Some examples include the work by Karagiannis et al.[25], which uses a reduced order observer in conjunction with backstepping and the small-gain theorem; Marino and Tomei [34], which provides a stability result on a class of systems that is required to be minimum phase with respect to some linear combination of its states, though it may be non-minimum phase with respect to its output; and also a result by Isidori [21] that guarantees 27 robust, semiglobal practical stabilization of nonlinear systems in the normal form, which could be non-minimum phase. The work [21] is particularly noteworthy because it provides a simple and very useful design tool for a broad class of nonlinear systems. The basic idea of an auxiliary system derived from the original system to help solve the stabilization problem for the latter forms the basis of this chapter, and was set forth in [21]. In this chapter we arrive at the auxiliary system of [21] from a different perspective. We employ an extended high gain observer (EHGO) as a virtual sensor to estimate a key signal that renders the zero dynamics of a non-minimum phase system observable, thus facilitating the design of a stabilizing compensator. This signal is used to drive a dynamic compensator, which is designed with the objective of making the resulting augmented system, comprising the plant and the compensator, a relative-degree-one minimum-phase system with respect to a virtual output. The stabilization of the augmented system is then pursued by a separation approach that builds on the work of Atassi and Khalil [2]. The design of the compensator is equivalent to the stabilization of the auxiliary system of [21]. The relationship between the stabilization of the auxiliary system and the creation of a relative-degree-one minimum-phase system was exploited in the proof of [21], but the perspective adopted in this chapter opens up new avenues that can be explored further in problems beyond stabilization, such as regulation, for example. It also opens the door for different approaches to the design of the stabilizing controller. While in this chapter we present a continuouslyimplemented sliding mode controller to stabilize the augmented system, other robust control techniques could also be used. In previous work, the EHGO was used as a disturbance observer [16]. The analysis of the output feedback system in the current paper has some similarities with [16]. While the techniques of both this chapter and [21] rely on the stabilization of the same auxiliary system, the controller designs are different. The paper [21] uses 28 high-gain feedback to implement the output feedback control in such a way that the closed-loop system can be represented as the feedback connection of an asymptotically stable system with a memoryless high-gain system. In this work, we do not use highgain feedback. Instead, an EHGO is used to estimate the signals that are needed to implement the state feedback controller. This difference between the two controllers could be significant in some applications because high-gain feedback, by its nature, would cause a large spike in the control signal in the initial transient, which will not be observed with the proposed controller. It is worth mentioning that both controllers use high-gain observers to estimate derivatives of the output from the output measurement. The dimension of the extended observer used in the current paper is higher by one, however. There are two main technical differences between the results presented in this chapter and those of [21]. First, our design allows for uncertainty in the control coefficient, which was not addressed in [21]. Handling this uncertainty dominates our analysis in Section 2.3. Second, we prove exponential stability of the closed-loop system, under the assumption of exponential stability of the closed-loop auxiliary system. The paper [21] proves only practical stabilization, but it requires the less stringent condition of asymptotic stability of the closed-loop auxiliary system. 2.2 Problem Statement A single-input, single-output system with relative degree ρ, under a suitable diffeomorphism, can be expressed in the following normal form [19]. η = f (η, ξ, θ), ˙ (2.1) ˙ ξi = ξi+1 , 1 ≤ i ≤ ρ − 1, (2.2) ˙ ξρ = b(η, ξ, θ) + a(η, ξ, θ)u, (2.3) 29 y = ξ1 , (2.4) where η ∈ Dη ⊂ Rn−ρ , ξ ∈ Dξ ⊂ Rρ , with Dη and Dξ being domains containing η = 0 and ξ = 0 respectively, y is a measured output, θ ∈ Θ ⊂ Rp is a vector of unknown constant parameters and Θ is a compact set. Assumption 2.1. The functions a(·) and b(·) are continuous in θ and continuously differentiable with respect to (η, ξ), with locally Lipschitz derivatives. The function f (·) is continuous in θ and locally Lipschitz in (η, ξ). In addition, a(η, ξ, θ) = 0, b(0, 0, θ) = 0 and f (0, 0, θ) = 0, ∀θ ∈ Θ. In this paper, the objective is to find a robust, stabilizing output feedback controller for systems of the form (2.1)–(2.3), which could potentially include systems with unstable zero dynamics. To this end, we begin by first considering the state feedback case. We shall eventually utilize an extended high gain observer in the output feedback design, and hence for the purposes of the state feedback design, we can assume that the signals ξ1 , . . . , ξρ−1 and b(η, ξ, θ) + ∆a (η, ξ, θ)u, (2.5) where ∆a (η, ξ, θ) = a(η, ξ, θ)−an (ξ) and an (ξ) is a known nominal model of a(η, ξ, θ), are all available for feedback. During the analysis of the output feedback system, it will be shown that the extended high gain observer recovers the signal given by (2.5) in addition to the output and its first ρ − 1 derivatives. The expression in (2.5) is based on the anticipation that with the extended high-gain observer, we will be able to estimate this quantity arbitrarily closely. Roughly speaking, this can be seen ˙ from (2.3) as follows. If the derivative ξρ = y (ρ) and a(·) were available, we could ˙ calculate b from the expression b = ξρ − au. If only an estimate an (·) were available ˙ for the uncertain term a(·), we would use bn = ξρ − an u, which results in the error 30 bn − b = (a − an )u = ∆a u. By requiring an to depend only on ξ, we have set up the problem such that the variables that need to be estimated are the derivatives of the output y up to the ρth order, which will be the role of the extended high-gain observer. Note that if a(·) were a known function of ξ, we could take an = a. We require the following assumption about a(·) and an (·). Assumption 2.2. Suppose that a(η, ξ, θ)/an (ξ) ≥ k0 > 0, for all (η, ξ) ∈ Dη × Dξ and θ ∈ Θ. Moreover, an (ξ) is locally Lipschitz in ξ over Dξ and globally bounded in ξ. 2.3 Controller Design for the State Feedback System In the case of minimum phase systems, the use of high-gain observers allows us to reduce the problem of stabilizing a system with relative degree higher than one to one of stabilizing a relative degree one system. This idea is explicated in [33]. For ρ−1 instance, we could define a virtual output s = ξρ + i=1 ki ξi , where the constants ki are chosen such that the system with output s is minimum phase. Then, a static stabilizing control law utilizing the virtual output s could be designed using standard techniques, and this design could then be implemented by replacing the virtual output s with its estimate from a high-gain observer. This definition of the virtual output will not work if the system (2.1)–(2.4) is non-minimum phase. However, with an extended high-gain observer, we will have access to the states ξ1 to ξρ in addition to the signal b(η, ξ, θ) + ∆a (η, ξ, θ)u, which can all be used in the state feedback design. We augment the plant (2.1)–(2.4) with the dynamic compensator ˙ φ = L(φ, ξ1 , . . . , ξρ−1 ) + M (φ, ξ1 , . . . , ξρ−1 )[b(η, ξ, θ) + ∆a (η, ξ, θ)u], 31 (2.6) s = ξρ − N (φ, ξ1 , . . . , ξρ−1 ), (2.7) where φ ∈ Rr . The augmented system will have relative degree one with respect to the virtual output s if 1− ∂N ∆a M ≥ k † > 0, ∂φ a (2.8) where k † is some real constant. Our task now is to design L(·), M (·), and N (·) such that the augmented system is minimum phase. It turnes out that this task is equivalent to the stabilization of the auxiliary system of [21], as we shall see in the development to follow. With the change of variables ξρ → s, the augmented system is given by η = f (η, ξ1 , . . . , ξρ−1 , s + N (φ, ξ1 , . . . , ξρ−1 ), θ), ˙ ˙ ξ1 = ξ2 , . . . (2.9a) (2.9b) (2.9c) ˙ ξρ−1 = s + N (φ, ξ1 , . . . , ξρ−1 ), (2.9d) ˙ φ = M (φ, ξ1 , . . . , ξρ−1 ) b(η, ξ1 , . . . , ξρ−1 , s + N (·), θ) + ∆a (η, ξ1 , . . . , ξρ−1 , s + N (·), θ)u + L(φ, ξ1 , . . . , ξρ−1 ), ∂N [L(·) + M (·)(b(·) + ∆a (·)u)] ∂φ ∂N ∂N ∂N + a(·)u − ξ2 − . . . − ξρ−1 − [s + N (·)]. ∂ξ1 ∂ξρ−2 ∂ξρ−1 s = b(η, ξ1 , . . . , ξρ−1 , s + N (·), θ) − ˙ (2.9e) (2.9f) (2.9g) (2.9h) We now set xa = η ⊤ ξ1 · · · ⊤ ξρ−2 ξρ−1 , ∂N ∂N ∂N ∂N (s + N (·)) − ξρ−1 − . . . − ξ , (L + M b) − ∂φ ∂ξρ−1 ∂ξρ−2 ∂ξ1 2 ∂N M ∆a , gb (xa , φ, s, θ) = a − ∂φ fb (xa , φ, s, θ) = b − 32       fa =       f (η, ξ1 , . . . , ξρ−1 , s + N (·), θ) ξ2 . . . ξρ−1 s + N (·)       .      Due to Assumption 2.2 and (2.8), we have gb /an ≥ k † k0 > 0. (2.10) The system equations can be rewritten in a compact form as follows. xa = fa (xa , φ, s, θ), ˙ (2.11) ˙ φ = L(xa , φ) + M (xa , φ)[b(xa , φ, s, θ) + ∆a (xa , φ, s, θ)u], (2.12) s = fb (xa , φ, s, θ) + gb (xa , φ, s, θ)u. ˙ (2.13) With s viewed as the output, this system has relative degree one. The internal dynamics are given by (2.11), (2.12). However, the system is not in the normal form due to the term M ∆a u in (2.12). We thus apply a change of variables that eliminates the control input from (2.12). Let T : φ → q such that ∂T ∂T M ∆a + g =0 ∂φ ∂s b and x⊤ (T (xa , φ, s, θ))⊤ s a ⊤ is a diffeomorphism of (2.14) x⊤ φ ⊤ s a ⊤ . When ∆a = 0, we have T = φ. Hence, for sufficiently small |∆a |, it can be argued that a function T exists [11]. Let φ = P (xa , q, s, θ) be the inverse transformation. Due to 33 the continuous dependence of the solution of (2.14) on ∆a [11], we can write T and P as1 T = φ + Φ(xa , φ, s, θ), P = q + Q(xa , q, s, θ), where Φ and Q are of the order of |∆a |. The transformed system equations due to (2.14) are given by xa = fa (xa , q + Q(xa , q, s, θ), s, θ), ˙ (2.15) q = L(xa , q + Q(xa , q, s, θ)) + M (xa , q + Q(xa , q, s, θ))b(xa , q + Q(xa , q, s, θ), s, θ) ˙ ∂Φ ∂Φ [L(·) + M (·)b(·)] fa (xa , q + Q(xa , q, s, θ), s, θ) + ∂xa ∂φ ∂Φ + f (xa , q + Q(xa , q, s, θ), s, θ), ∂s b + (2.16) s = gb (xa , q + Q(xa , q, s, θ), s, θ)u + fb (xa , q + Q(xa , q, s, θ), s, θ). ˙ (2.17) When ∆a = 0, (2.11)–(2.13) simplify to xa = fa (xa , φ, s, θ), ˙ ˙ φ = L(xa , φ) + M (xa , φ)b(xa , φ, s, θ), s = fb (xa , φ, s, θ) + a(xa , φ, s, θ)u. ˙ Let xs and x⊤ q ⊤ a ⊤ , (2.18)   fa (xa , q + Q(xa , q, s, θ), s, θ)   F (xs , s, θ) =  . ∂Φ f + ∂Φ (L + M b) + ∂Φ f L + Mb + ∂xa a ∂φ ∂s b (2.19) Since xs includes the states (ξ1 , . . . , ξρ−1 , q), we write L(·), M (·) and N (·) as L(xs ), 1 See Section A.2 for details regarding the existence of solutions to equation (2.14), and continuous dependence of the former on the parameters of the latter. 34 M (xs ) and N (xs ). The internal dynamics (2.15)–(2.16) can now be written as xs = F (xs , s, θ). ˙ (2.20) Our objective is to design the compensator {L, M, N } such that the internal dynamics (2.20) are input to state stable with s viewed as an input. The control law based on continuously-implemented sliding mode control is taken to be u=− β(ξ, φ) sat an (ξ) s µ , (2.21) where β(·) is to be chosen. The closed-loop state feedback system is given by (2.13), (2.20), (2.21). We now make the following assumptions. Assumption 2.3. Let D ⊂ Rn+r be a domain containing the origin of the system (2.13), (2.20), (2.21) such that (η, ξ, φ) ∈ Dη × Dξ × Dφ ⇒ (xs , s) ∈ D. Moreover, suppose there exists a continuously differentiable Lyapunov function V (xs , θ) such that α1 ( xs ) ≤ V (xs , θ) ≤ α2 ( xs ), ∂V F (xs , s, θ) ≤ −α3 ( xs ), ∀ xs ≥ γ(|s|), ∂xs (2.22) (2.23) for all (xs , s) ∈ D ⊂ Rn+r , where α1 (·), α2 (·), α3 (·) and γ(·) are class K functions. Assumption 2.4. 1) L(xs ), M (xs ) and β(ξ, φ) are locally Lipschitz functions in their arguments over the domain of interest, and L(0) = 0. 2) L(xs ), M (xs ) and β(ξ, φ) are globally bounded functions of ξ. 3) With s = 0, the origin xs = 0 of (2.20) is locally exponentially stable. 35 Remark 2.1. 1) The set D is well-defined for any sufficiently smooth N (·) in the definition of the variable s = ξρ − N (·). 2) Inequality (2.23) is equivalent to regional input-to-state stability of the system (2.20) with s viewed as an input. Let us now consider the set Ω {V (xs , θ) ≤ c0 } × {|s| ≤ c}, (2.24) where c > µ, c0 ≥ α2 (γ(c)), and c, c0 are chosen such that Ω is a compact subset of D. As in the development preceding [29, Theorem 14.1 (§14.1.2)], we show that Ω is a positively invariant set for the system (2.20), and can thus serve as an estimate of the region of attraction. We now have g (xa , φ, s, θ) β(ξ, φ) sat ss = s fb (xa , φ, s, θ) − b ˙ an (ξ) s µ . β(·) is chosen such that β(ξ, φ) ≥ β0 + an (ξ)fb (xa , φ, s, θ) , gb (xa , φ, s, θ) β0 > 0. (2.25) Hence, for s ≥ µ, we have ss = sfb − (gb /an )β |s| ≤ (gb /an )( an fb /gb − β) |s| ˙ ≤ −(gb /an )β0 |s| ≤ −k † k0 β0 |s| . Hence, whenever |s(0)| > µ, |s(t)| will decrease until it reaches the set {|s| ≤ µ} in finite time and will remain inside thereafter. 36 We would now like to study the behavior of the remaining states given by (2.20). In order to proceed with the analysis, we utilize Assumption 2.3, which calls for the existence of a Lyapunov function for the internal dynamics. The derivative of V along the trajectories of (2.20) is given by ∂V ˙ F (xs , s, θ). V = ∂xs Thus, on the boundary V = c0 , we have ∂V −1 ˙ V = F (xs , s, θ) ≤ −α3 ( xs ) ≤ −α3 (α2 (c0 )). ∂xs Hence, Ω is a positively invariant set for the system (2.20). As a consequence of the preceding analysis, all trajectories starting in Ω, stay in Ω and reach the set Ω ∩ {|s| ≤ µ} in finite time. Inside this set, we have ˙ V ≤ −α3 ( xs ), ∀ xs ≥ γ(µ). The above analysis shows that the trajectories enter the positively invariant set Ωµ = {V (xs , θ) ≤ α2 (γ(µ))} × {|s| ≤ µ} in finite time. Now, inside the set Ωµ , the closed-loop system is given by the following equations in the singularly perturbed form. xa = f a , ˙ q = L + Mb + ˙ (2.26a) ∂Φ ∂Φ ∂Φ (L + M b) + f , fa + ∂xa ∂φ ∂s b g µs = µfb − b βs. ˙ an (2.26b) (2.26c) 37 We can use standard singular perturbation arguments [29, Chapter 11] to show that the origin of (2.26) is exponentially stable for sufficiently small µ, with a region of attraction independent of µ—the details are provided in the Appendix. Hence, for sufficiently small µ, all trajectories in Ωµ converge to the origin. We summarize this result in the following proposition. Proposition 2.1. Consider the closed-loop system given by (2.13), (2.20), (2.21). Let Assumptions 2.1 through 2.4 hold. Then, there exists a positive constant µ∗ such that the origin of (2.13), (2.20), (2.21) is exponentially stable for all 0 < µ < µ∗ . Moreover, Ω, given by (2.24), is an estimate of the region of attraction. Proof. See Section A.3 for details of the proof. Remark 2.2. 1) In [21], the closed-loop auxiliary system (which corresponds to the internal dynamics (2.20) in our paper, with s = 0 and ∆a = 0) was assumed to be asymptotically stable, and consequently, Theorem 1 of that paper establishes practical stabilization for the state feedback system. This is generally the result in the event that the state feedback system fails to satisfy certain interconnection conditions. If assumptions are made such that these additional conditions are satisfied, then the state feedback system could be shown to be asymptotically stable as well [29]. 2) Assumption 2.3 requires the compensator {L, M, N } to be designed such that (2.20) is input to state stable with s viewed as the input. This problem is identical to the auxiliary design problem of [21] when ∆a = 0. In [21], the interconnection of the auxiliary system with its dynamic compensator was assumed to be globally asymptotically stable, with a Lyapunov function independent of the unknown system parameters. The fulfillment of this assumption is still an open problem, but as shown in [21, Proposition 2], in the case of linear systems, the stabilizability 38 and detectability of the original system is a sufficient condition for the existence of a stabilizing controller for the auxiliary system. 3) We assumed local exponential stability of the origin of the internal dynamics (2.20) in Assumption 2.4, and this in conjunction with the regularity conditions given by Assumptions 2.1 and 2.4 allowed for us to prove exponential stability of the origin of the closed-loop state feedback system in Theorem 2.1. 4) If all the assumptions hold globally, then we could take D = Rn+r , and with the appropriate choice of controller parameters, the positively invariant set Ω ⊂ D that depends on these parameters can also be made arbitrarily large, and thus our design allows for semi-global stabilization of the state feedback system. 2.4 Output Feedback Control Using an Extended High-Gain Observer 2.4.1 Output Feedback Controller Design The design of the output feedback controller in this paper closely follows the work by Freidovich and Khalil [16]. It relies upon the estimates of the states and of b(η, ξ, θ) that are obtained by using an extended high-gain observer for the system (2.1)–(2.3), which is taken as ˙ ˆ ˆ ˆ ξi = ξi+1 + (αi /εi )(ξ1 − ξ1 ), 1 ≤ i ≤ ρ − 1, ˙ ˆ ˆ ˆ ξρ = σ + an (ξ)u + (αρ /ερ )(ξ1 − ξ1 ), ˆ ˙ ˆ σ = (αρ+1 /ερ+1 )(ξ1 − ξ1 ), ˆ 39 (2.27) (2.28) (2.29) where ε is a positive constant to be specified, and the positive constants αi are chosen such that the roots of λρ+1 + α1 λρ + . . . + αρ λ + αρ+1 = 0 are in the open left-half ˆ ˆ plane. It is apparent from (2.27)–(2.29) that the ξ1 , . . . , ξρ are used to estimate the output and its first ρ − 1 derivatives, while σ is intended to provide an estimate for ˆ b(·). With the aid of these estimates, the output feedback controller for the original system can be taken as ˙ ˆ ˆ ˆ ˆ φ = L(φ, ξ1 , . . . , ξρ−1 ) + M (φ, ξ1 , . . . , ξρ−1 )ˆ , σ ˆ β(ξ, φ) ˆ l(ξ, φ) = − sat ˆ an (ξ) ˆ ˆ ˆ ξρ − N (φ, ξ1 , . . . , ξρ−1 ) µ , ˆ u = K sat(l(ξ, φ)/K), (2.30) (2.31) (2.32) where K> β(ξ, φ) . max (xs ,s)∈Ω an (ξ) (2.33) The control is saturated at ±K outside the compact set of interest in order to protect the system from peaking during the observer’s transient response.2 The closed loop system under output feedback can now be expressed as η = f (η, ξ, θ), ˙ ˙ ˆ ξ = Aξ + B[b(η, ξ, θ) + a(η, ξ, θ)K sat(l(ξ, φ)/K)], (2.35) ˙ ˆ ˆ ˆ ˆ φ = L(φ, ξ1 , . . . , ξρ−1 ) + M (φ, ξ1 , . . . , ξρ−1 )ˆ , σ (2.36) ˙ ˆ ˆ ˆ ˆ ˆ ξ = Aξ + B σ + an (ξ)K sat(l(ξ, φ)/K) + H(ε)(y − C ξ), ˆ (2.37) ˆ ˙ σ = αρ+1 /ερ+1 (y − C ξ), ˆ (2.38) y = Cξ, 2 See (2.34) (2.39) [14] for the role played by saturation in overcoming the effect of peaking. 40 where  0   0  . A = . .   0  0 1 0 ··· 0 . . . 0 0 C= H(ε) = 0    0 . . . B =  , 0     1   1 · · · 0  . .. . . . . , . .   0 · · · 1  0 ··· 0 1 0 ··· 0 , (α1 /ε) (α2 /ε2 ) · · · (αρ /ερ ) ⊤ . We now introduce a change of variables [16] νi νρ+1 ˆ (ξi − ξi )/ερ+1−i , 1 ≤ i ≤ ρ, ∆a (ε, η, ξ, ν, θ)Kgε (l(ξ, φ)/K) + b(η, ξ, θ) − σ , ˆ (2.40) (2.41) ˆ where ∆a (ε, η, ξ, ν, θ) = a(η, ξ, θ) − an (ξ), and gε (·) is an odd function defined by [16]    y, 0 ≤ y ≤ 1,    2 gε (y) =  y + (y − 1)/ε − 0.5(y − 1)/ε, 1 < y < 1 + ε,     1 + 0.5ε, y ≥ 1 + ε. The function gε (·) is nondecreasing, continuously differentiable with a locally Lipschitz derivative, bounded uniformly in ε on any bounded interval of ε, and satisfies ′ gε ≤ 1 and |gε (y) − sat(y)| ≤ ε/2 for all y ∈ R. The closed loop system can now be expressed in terms of (2.34)–(2.36), 41 (2.39) and ¯ ¯ εν = ε[B1 ∆1 (ξ, η, ν, φ, θ, ε) + B2 ∆2 (ξ, η, ν, φ, θ, ε)] + Λν, ˙ (2.42) where  −α1    −α 2   .  . Λ= .    −αρ  −αρ+1 1 0 ··· 0 . . . 0 0 0        1 · · · 0  0 ¯ B  . .. . ¯ . . .  , B1 =   , B2 =   , ∆2 (ξ, η, ν, φ, θ, ε) = ∆0 /ε, . .  B 0  0 · · · 1  0 ··· 0 ∆0 (ξ, η, ν, φ, θ, ε) = ˆ a(η, ξ, θ)K[gε (l(ξ, φ)/K) − gε (l(ξ, φ)/K)] ˆ ˆ +an (ξ)Kgε (l(ξ, φ)/K) − an (ξ)Kgε (l(ξ, φ)/K) ˆ ˆ ˆ +[a(η, ξ, θ) − an (ξ)]K[sat(l(ξ, φ)/K) − gε (l(ξ, φ)/K)], ∆1 (ξ, η, ν, φ, θ, ε) = ′ ˙ ˙ b + ∆a Kgε (l(ξ, φ)/K) + ∆a gε (l(ξ, φ)/K) + ∂l(ξ, φ) ˆ ˆ L(φ, ξ1 , . . . , ξρ−1 ) + ∆a Kgε ∂φ ˆ ˆ + b − νρ+1 M (φ, ξ1 , . . . , ξρ−1 ) ∂l(ξ, φ) ˙ ξ ∂ξ l(ξ, φ) K . We note that the functions ∆1 and ∆2 are locally Lipschitz in their arguments and bounded from above by affine-in- η functions, uniformly in ε. This is because an (·) and gε (·) are continuously differentiable with locally Lipschitz derivatives and globally bounded. Moreover, we can use (2.40) and the definition of gε to show that ∆0 /ε is 42 locally Lipschitz [16].3 The slow dynamics can be written as   ˙ xs  xo =   = fr (xo , ν, ε, θ), ˙ s ˙ where xo x⊤ s s (2.43) ⊤ ,   φ(η, ξ, θ)       ξ2     .   . .    , fr (xo , ν, ε, θ) =     ξρ       f1 (ξ, η, ν, φ, θ, ε)     ˆ b(·) − ∂N F (xs , s, θ) + a(·)K sat(l(ξ, φ)/K) ∂xs where f1 (·) = M (·)[b + ∆a (ε, η, ξ, ν, θ)Kgε (l(ξ, φ)/K)] + L(·) − M (·)νρ+1 , xs and ⊤ ˆ ˆ . Then, the F are defined as in (2.18), (2.19), and xs = η ⊤ ξ1 , . . . , ξρ−1 q ⊤ ˆ reduced system is obtained by setting ε = 0 in (2.43), which yields    F (xs , s, θ)  xo = fr (xo , 0, 0, θ) =  ˙ , f2 (ξ, η, ν, φ, θ, ε) (2.44) where f2 (ξ, η, ν, φ, θ, ε) = b(η, ξ, θ) − ∂N F (xs , s, θ) + a(η, ξ, θ)K sat(l(ξ, φ)/K). We ∂xs note that in the set Ω given by (2.24), the reduced system (2.44) is identical to the closed-loop state feedback system (2.13), (2.20), (2.21), which we know to have an exponentially stable equilibrium point at its origin. 3 See Section A.1 for details. 43 The dynamics of the fast variable ν are given by the boundary layer system ∂ν = Λν, ∂τ τ t , ε (2.45) which we note is exponentially stable. In the subsection below, we state some results pertaining to the recovery of the performance of the system (2.13), (2.20) in the case of output feedback with sufficiently small ε. These results follow directly from the work of Atassi and Khalil [2] because equations (2.34)–(2.36) and (2.42) are precisely of the same form as equations (10)–(13) in [2]. It should also be noted that while the structure of system (2.34)–(2.36), (2.42) is very similar to that of the closed-loop system studied by Freidovich and Khalil [16], a key difference between this particular work and the former is the fact that both the nonlinear terms in the fast dynamics (2.42) are O(ε) in this work, while there is one term in [16] that is not so. The reason for such a term to appear in [16] was the fact that the estimate σ was used in the control law, whereas ˆ in this work, this quantity is not utilized directly in the control. We can see this by examining (2.32) and (2.36)—ˆ appears in the compensator dynamics, but u is not σ explicitly dependent upon it. 2.4.2 Exponential Stability and Trajectory Convergence Let (xo (t, ε), ν(t, ε)) denote the trajectory of the system (2.42), (2.43) starting from (xo (0), ν(0)). Also, let xr (t) be the solution of (2.44) starting from xo (0). We know that the system (2.13), (2.20), (2.21) has an exponentially stable equilibrium point, so let us suppose that R is its region of attraction. Let S be a compact set in the interior of R and Q be a compact subset of Rρ+1 . The recovery of exponential stability and the fact that xo (t, ε) converges to xr (t) as ε → 0, uniformly in t, for all t ≥ 0, is established by the following theorem [2, Theorems 1, 2, 3 and 5], provided xo (0) ∈ S 44 and ˆ ξ(0)⊤ σ (0) ˆ ⊤ ∈ Q. Theorem 2.1. Let Assumptions 2.1 through 2.4 hold, and suppose the vector field fr (xo , 0, 0, θ) is continuously differentiable around the origin. Moreover, assume xo (0) ∈ S and ⊤ ˆ ⊤ σ (0) ∈ Q. Then, ξ(0) ˆ 1) there exists ε∗ > 0 such that, for every 0 < ε ≤ ε∗ , the origin of system (2.43), 1 1 (2.42) is exponentially stable, and S × Q is included in the region of attraction. 2) given any δ > 0, there exists ε∗ > 0 such that, for every 0 < ε ≤ ε∗ , we have 2 2 xo (t, ε) − xr (t) ≤ δ, ∀t ≥ 0. (2.46) Remark 2.3. Theorem 1 of [2] establishes boundedness of (xo , ν) for all trajectories starting in S ×Q. Theorem 2 of [2] shows that the trajectories are ultimately bounded, and the ultimate bound can be made arbitrarily small by choosing ε small enough. Finally, Theorem 5 of [2] establishes that the origin is exponentially stable with a region of attraction that is independent of ε. The combination of the three results yields the first bullet of the foregoing theorem. The second bullet follows from Theorem 3 of [2]. 2.5 Example We illustrate the design procedure presented in this chapter by using a second order system as an example. This system has the following structure. x1 = tan x1 + x2 , ˙ 45 (2.47) x2 = x1 + u, ˙ y = x2 . (2.48) (2.49) We note that this system has a relative degree of one, and is already in the normal form. We can thus write the auxiliary system as x1 = tan x1 + ua , ˙ (2.50) ya = x1 . (2.51) The design of the stabilizing controller is now carried out by first considering the auxiliary system and proceeding in a step-by-step fashion, in accordance with the procedure detailed in Sections 2.3 and 2.4. 2.5.1 Design of the Stabilizing Controller We begin by noting that we could linearize the closed-loop auxiliary system equation with the choice of the input ua = − tan ya −ya , resulting in a stable system. However, the design procedure requires that ya serve as an input to a dynamic controller, and that the control input ua not be driven directly by ya —a static output feedback law is therefore unsuitable. Hence, a dynamic compensator for the auxiliary system (2.50), (2.51) is designed using a low-pass filter, and feedback linearization is in effect achieved by utilizing the filter’s output. This gives the controller ˙ φ = (ya − φ)/εa , (2.52) ua = − tan φ − φ. (2.53) A choice of controller parameter εa = 0.1 yields closed-loop auxiliary system with an exponentially stable equilibrium point at the origin. Now, based on our knowledge 46 of the stabilizing dynamic controller (2.52), (2.53) for the auxiliary system (2.50), (2.51), we choose our sliding manifold for the partial state feedback system, and the resulting controller is given by s = x2 + φ + tan φ, (2.54) ˙ φ = (x1 − φ)/εa , (2.55) u=− β sat an s µ , (2.56) where the constant β satisfies the inequality (2.25), and an is a nominal value of the control coefficient a = 1 in the original plant. To stabilize the system using only output feedback, we utilize the following extended high-gain observer. ˙ x2 = σ + an u + (α1 /ε)(y − x2 ) ˆ ˆ ˆ ˙ σ = (α2 /ε2 )(y − x2 ). ˆ ˆ (2.57) (2.58) Hence, the extended-high-gain-observer-based output feedback controller is given by σ−φ ˆ , 0.6εa β y + φ + tan φ u=− sat an µ ˙ φ = 0.6 sat (2.59) . (2.60) The expression on the right-hand side of (2.59) was saturated at ±0.6, and these values were determined by means of simulation. This was done because Theorem 2.1 (Assumption 2.4) requires that the functions L(·) and M (·) be globally bounded. In the following subsection, the above design is compared via simulation with a control design performed according to the high-gain feedback method of Isidori [21]. It is based on the same controller (2.52)–(2.53) for the auxiliary system as the one 47 0.5 x1 0.4 0.3 0.2 0.1 0 −0.1 0 2 t (s) 4 6 Figure 2.1: Recovery of the Auxiliary System performance (auxiliary sys: solid, state feedback with µ = 1: dashed, and µ = 0.1: dash-dot). designed above. Thus, the output feedback controller for the full system is given by ˙ φ = [k(y + tan φ + φ) − φ]/εa , (2.61) u = {(2 + tan2 φ)[φ − k(y + tan φ + φ)]/εa − k(y + tan φ + φ)}/an . (2.62) We note that since the controller is driven only by the output, an observer is not required in this method. 48 x1 0.5 0 0 1 2 3 4 5 6 1 2 3 4 5 6 1 2 3 4 5 6 y 0.5 −0.5 −1.5 0 u 10 0 −10 0 t (s) Figure 2.2: Recovery of the Partial State Feedback System performance by decreasing ε (solid: state feedback response, output feedback response with ε = 0.01: dashed, ε = 0.001: dash-dot). 2.5.2 Numerical Simulations The step-by-step tuning procedure leading to performance recovery is illustrated by Figures 2.1 and 2.2. The constant β was chosen to be 8. The compensator parameter εa was fixed at 0.1 and the width of the boundary layer was reduced in the partial state feedback system from µ = 1 down to 0.1, and Figure 2.1 shows the recovery of the auxiliary system performance. Next, µ was fixed at 0.1 and the extended highgain observer parameter ε was reduced from ε = 10−2 to 10−3 , and Figure 2.2 shows the recovery of the performance of the state feedback system. Next, in order to compare the performance of the above design with that of Isidori’s, the design parameters for each method were chosen such that the initial 49 y 0 −0.5 −1 0 1 2 3 4 5 6 1 2 3 4 5 6 x1 0.6 0.4 0.2 0 0 t (s) Figure 2.3: A comparison of the perfomance of the Isidori (dash-dot) and the EHGO-based (solid) designs. condition satisfies (x1 (0), y(0), φ(0)) ∈ I = [−0.5, 0.5] × [−0.1, 0.1] × {0}. To this end, the high-gain control parameter that appears in Isidori’s design was chosen to be k = 12. Figure 2.3 and Figure 2.4 show how the performance compares between Isidori’s method and the output feedback controller. Upon comparing the control efforts in Figure 2.4, it is apparent that the initial transient response of Isidori’s design causes u to start out at a value of about 25, while in the EHGO-based SMC design, it starts out at 8. Subsequently, however, the overshoot and maximum rate of change are about the same in each case. This fact is used as the basis for performance 50 20 u 10 0 −10 0 1 2 3 4 5 6 1 2 3 4 5 6 0.6 z 0.4 0.2 0 −0.2 0 Figure 2.4: dash-dot). t (s) A comparison of the control efforts (EHGO: solid, Isidori: comparison between the two designs. It is apparent from Figures 2.3 through 2.6 that the design procedure of this paper was able to provide better transient response than the high-gain feedback approach. Figure 2.5 shows a comparison between the two methods when the nominal value of the control coefficient is an = 0.86, while in Figure 2.6, it is an = 1.04. The EHGO-based SMC design is able to tolerate a slightly wider range of uncertainties, and Figures 2.7 and 2.8 show the simulation results for an = 0.76 and an = 1.05, respectively. We note that the settling time in Figure 2.7 is much longer than in all the other cases. 51 y 0 −0.5 −1 −1.5 0 2 4 6 8 10 2 4 6 8 10 30 u 20 10 0 0 t (s) Figure 2.5: A comparison between Isidori’s (dash-dot) and the EHGO-based (solid) design for an = 0.86. 2.6 Conclusion A robust, stabilizing output feedback controller for systems in the normal form, which could potentially include unstable zero dynamics, was presented. The control scheme adopted herein incorporated continuously-implemented sliding mode control—chosen for its robustness properties as well as its ability to prescribe or constrain the motion of trajectories in the sliding phase—and an extended high gain observer to estimate one of the unknown functions. Stabilization in the case of an unknown control coefficient and uncertain constant parameters was shown for the state as well as output feedback cases. The control design presented in this chapter is an alternative to the design of [21]. 52 0.5 y 0 −0.5 −1 0 2 4 6 8 10 2 4 6 8 10 20 u 10 0 −10 0 t (s) Figure 2.6: A comparison between Isidori’s (dash-dot) and the EHGO-based (solid) design for an = 1.04. A main contribution of this chapter is that it shows how a non-minimum phase plant can be stabilized despite uncertainty in the control coefficient a(·), and it demonstrates the potential of the EHGO acting as a virtual sensor that can be used to estimate signals whose knowledge is necessary for the successful realization of the aforementioned control task. Although sliding mode control was used in this work, the controller design is introduced in Section 2.3 in such a way that the technique can be extended to other control schemes quite easily, and the EHGO-as-a-virtual-sensor concept can be applied to a broader class of problems, as well. The paper by Isidori [21] assumes asymptotic stability of the origin of the auxiliary system, and it proves practical stabilization of the output feedback system by means of a dynamic high-gain feedback based controller. In this work, the auxiliary system is assumed to be locally 53 0.5 y 0 −0.5 −1 −1.5 0 5 10 5 10 15 20 25 15 20 25 8 u 6 4 2 0 −2 0 t (s) Figure 2.7: Response of the EHGO-based design when an = 0.76. exponentially stable. With the aid of some regularity assumptions, it was shown that an extended high-gain observer based dynamic, continuously-implemented sliding mode controller is able to exponentially stabilize the output feedback system in the presence of uncertainty in the control coefficient a(·). The tuning of the design parameters µ and ε to achieve a desired performance merits comment. The separation result of Atassi and Khalil [2] was invoked so as to perform the state feedback design first, which employed a continuously implemented sliding mode control instead of employing it “as-is” to address the issue of chattering. Consequently, a key component in the state feedback design is the tuning of the boundary layer parameter µ to recover the properties of “true” sliding mode control—however, if µ is made too small, the issue of chattering arises yet again, 54 y 0 −0.5 −1 0 1 2 1 2 3 4 5 6 3 4 5 6 u 5 0 −5 0 t (s) Figure 2.8: Response of the EHGO-based design when an = 1.05. so we are limited in terms of being unable to make this parameter arbitrarily small [29]. Similarly, in the case of the output feedback design, we would like to make the extended high-gain observer parameter ε as small as possible in order to recover the performance of the state feedback system. However, in most practical applications, we are limited by bandwith constraints, and so ε cannot be made arbitrarily small. Moreover, the presence of measurement noise results in the need for a trade-off between state feedback performance recovery and robustness to model uncertainties on the one hand, and the potential for the exacerbation of the effect of measurement noise on the other. Specifically, measurement noise imposes a lower bound on ε that is O(ϕ1/(ρ+1) ), where ϕ is an upper bound on the amplitude of the measurement noise, in accordance with the results of Ahrens and Khalil [1, 5]. 55 Chapter 3 Output Regulation of Non-Minimum Phase Systems 3.1 Introduction A lot of research has been conducted in the past couple of decades on the problem of output regulation of nonlinear systems; references [9, 40, 26, 39] are but a tiny cross-section of the standard literature on this topic. Most of the available results have in fact been collected in well known books on nonlinear control, such as [18, 19]. Despite this wealth of literature, one area of nonlinear control is still very much an open avenue of research in many respects, due mainly to the challenging nature of the problem—and this is the problem of robust output feedback regulation of nonminimum phase systems. A major breakthrough was achieved on the stabilization problem by the pioneering work of [21]. This work was later followed up by [32, 10], in which the the same technique was extended and applied to the problem of nonlinear output regulation of non-minimum phase systems. Both of the aforementioned approaches utilize high-gain feedback techniques. The results contained therein do not consider uncertain control coefficients, and this par56 ticular issue is addressed in this chapter. This work is an extension, in analogous fashion, of the stabilization problem investigated in Chapter 2. The approach utilizes an extended high-gain observer (EHGO) to estimate a “virtual output” in addition to the system output (or tracking error in the case of regulation) and its first ρ derivatives (this number being the relative degree of the system), in conjunction with a continuous implementation of sliding mode control. The key idea of [21, 32] is retained in both these approaches, and consequently, the results take advantage of the knowledge of a dynamic controller for the auxiliary subsystem associated with the original problem, and the definition of the latter in [36] and in this work is identical to the original work of [21, 32, 10], respectively. Despite the aforementioned similarities between this work and prior work by [21, 32, 10], we should note that our approach to the output feedback design is different. The papers [21, 32, 10] use high-gain feedback to implement the output feedback control in such a way that the closed-loop system can be represented as the feedback connection of an asymptotically stable system with a memoryless high-gain system. In this work, we do not use high-gain feedback. Instead, an EHGO is used to estimate all of the signals that are needed to implement the state feedback controller, thus allowing for us to obtain results in keeping with the spirit of the Separation Principle. Consequently, our approach does not confine us to any specific control design method such as high-gain feedback, sliding mode control (SMC), Lyapunov redesign, etc., although continuously implemented SMC was chosen for this work. Moreover, highgain feedback, by its nature, would cause a large spike in the control signal in the initial transient, which will not be observed with the proposed controller. Another technical difference between the results of the current work and those of [32, 10] is that our design allows for uncertainty in the control coefficient, which was not addressed in the former works. Handling this uncertainty dominates our analysis in Section 3.2.2. A preliminary version of this chapter appears in [35]. The main difference between 57 the aforementioned paper and this work is that we use the standard servocompensator driven by the tracking error herein, whereas in [35], the servo is driven by the virtual output obtained from the extended high-gain observer. The approach adopted in this chapter simplifies some of the technical details in the analysis of the state feedback system, as well as the design of the stabilizing controller for the auxiliary system. The example used in this chapter is a linearized pendulum model, which is also different from the second order system utilized in [35]. This chapter is organized as follows—in Section 3.2, the problem statement and some preliminary assumptions are introduced. In the main Section 3.2.2 that follows, the control design is carried out initially for the state feedback case, in which the output and its first ρ derivatives are assumed to be available, along with a virtual output that renders the zero dynamics of the system observable. The state feedback regulation result is summarized, and this is followed by the output feedback design that utilizes an extended high-gain observer to estimate the states corresponding to the chain of integrators in the normal form, as well as an estimate of the aforementioned virtual output, albeit potentially perturbed due to any uncertainty in the control coefficient. The section is concluded by stating the main result of this chapter. Appropriate assumptions are made in this section as required. Section 3.3 illustrates the result using a third order system as an example, and some numerical results are presented, along with a comparison of the technique of this chapter and that of [32] via simulation of this example using both approaches. Section 3.4 contains some concluding remarks summarizing the results. 58 3.2 Problem Description and Assumptions 3.2.1 Background and Preliminaries In this chapter, we consider a single-input, single-output system that is in the normal form and has a well-defined relative degree, and that could potentially include unstable zero dynamics. The system model takes the form z = f0 (z, x, w, θ), ˙ (3.1a) x = Ax + B[b0 (z, x, w, θ) + a0 (z, x, w, θ)u], ˙ (3.1b) y = Cx − q(w, θ), (3.1c) where (z, x) ∈ Dz × Dx ⊂ Rn is the system state, u ∈ R is the input and y ∈ R the output, θ ∈ Θ ⊂ Rp is a vector of constant parameters, where Θ is a compact set, q(w, θ) is a reference signal, the control coefficient a0 (z, x, w, θ) = 0, and the disturbance w : [0, ∞) → Rd is generated by the following neutrally stable exosystem. w = S0 w. ˙ (3.2) Moreover, the matrices A ∈ Rρ×ρ , B ∈ Rρ×1 and C ∈ R1×ρ have the following structure.  0   0  . A = . .   0  0 and C = 1 0 ··· 1 0 ··· 0 . . . 0 0 0    1 · · · 0  . .. . . . . , . .   0 · · · 1  0 ··· 0   0 . . . B =  , 0     1 0 . The goal in this chapter is to design a control algorithm that regulates the output y to zero. 59 Let Ωs ⊂ Rn and W ⊂ Rd be given compact sets. Then, bearing in mind the aforementioned goal of output regulation, we shall require the following assumptions. Assumption 3.1. Let x(w, θ) ¯ where Ls q q(w, θ) Ls q(w, θ) · · · ρ−1 Ls q(w, θ) ⊤ , (∂q/∂w)S0 w. There exists a unique mapping z (w, θ) that solves the ¯ partial differential equation ∂ z (w, θ) ¯ S0 w = f0 (¯(w, θ), x(w, θ), w, θ), z ¯ ∂w for all w ∈ W . Under this assumption, the system (3.1) has a steady-state zero-error manifold given by {(z, x) | z = z (w, θ), x = x(w, θ)}. The steady-state value of the control ¯ ¯ signal u on this manifold is given by ρ Ls q(w, θ) − b0 (¯(w, θ), x(w, θ), w, θ) z ¯ c(w, θ) = . a0 (¯(w, θ), x(w, θ), w, θ) z ¯ Assumption 3.2. There exist real constants d0 , . . . , dm−1 , all independent of θ, such that c(w, θ) satisfies the identity m−1 c, Lm c = d0 c + d1 Ls c + · · · + dm−1 Ls s for all (w, θ) ∈ W × Θ, and the characteristic polynomial λm − dm−1 λm−1 − · · · − d0 has distinct zeros on the imaginary axis. 60 (3.3) Assumption 3.2 is motivated by the nonlinear counterpart to the well known internal model principle from regulation theory [12]. When the system is nonlinear, its controller must reproduce, in addition to the trajectories generated by the exosystem, any higher order nonlinear deformations of the aforementioned trajectories that may be present (see, for example, [26]). Let us now define  0 1 0 ··· 0       0 0 1 ··· 0    . . . .  ... . . . . , S= . . . .       0 0 0 ··· 1    d0 d1 d2 · · · dm−1 and Γ = 1 0 ··· 0  c       Ls c     .   . , τ = .     m−2  Ls c   m−1 c Ls ∈ R1×m . It can be shown that c(w, θ) is generated by the internal model ∂τ (w, θ) S0 w = Sτ (w, θ), ∂w (3.4a) c(w, θ) = Γτ (w, θ), (3.4b) where the pair (S, Γ) is observable. Suppose we choose matrices J and K1 such that (S, J) is controllable and F = S − JK1 is Hurwitz. It can be shown that there exists a unique matrix X such that [8] SX = XS, and − K1 X = Γ. Hence, it is easy to verify that c(w, θ) can also be generated by Ls (Xτ (w, θ)) = SXτ (w, θ), 61 (3.5a) c(w, θ) = −K1 Xτ (w, θ). (3.5b) We will have occasion to use (3.5) when designing the servocompensator. We solve the output regulation problem by converting it to a stabilization problem. To this end, we define and apply the change of variables η z − z (w, θ) and e ¯ ⊤ x − x(w, θ), and also define ξ ¯ . The system equations (3.1) are e1 · · · eρ−1 then transformed into η = f (η, ξ, eρ , w, θ), ˙ (3.6a) ˙ ξ = P ξ + Qeρ , (3.6b) eρ = b(η, ξ, eρ , w, θ) + a(η, ξ, eρ , w, θ)[u − c(w, θ)], ˙ y = Rξ = ξ1 = e1 , (3.6c) (3.6d) where the control coefficient a(η, ξ, eρ , w, θ) = a0 (z, x, w, θ) and the functions f (·) and b(·) are such that, for all (w, θ) ∈ W × Θ, f (0, 0, 0, w, θ) = 0 and b(0, 0, 0, w, θ) = 0. The triple (P, Q, R) has the same structure as (A, B, C) in (3.1), except that P ∈ R(ρ−1)×(ρ−1) , Q ∈ R(ρ−1)×1 and R ∈ R1×(ρ−1) , respectively. The zero-error manifold in terms of the new variables now becomes {(η, e) | η = 0, e = 0}. Next, we augment the above system with the servocompensator ˙ ζ = Sζ + Jξ1 , (3.7) where ζ ∈ Rm . When we examine the error system (3.6), we notice that in order for regulation to 62 occur, i.e., for y(t) → 0, the control u(t) would have to asymptotically converge to a steady state given by c(w, θ). This is obviously not implementable due to w, θ being unknown. Nevertheless, this problem has been solved in the case of minimum phase systems: [18, 28, 39, 37]. However, the techniques presented in the aforementioned papers assume that the zero dynamics of the system are either input to state stable or globally asymptotically stable, and would hence fail in the case of non-minimum phase systems. The challenge of stabilizing non-minimum phase systems was tackled by Isidori, leading to a design tool that can be applied to a large class of systems in which the coupling term b(·) between the internal and external dynamics of a system in the normal form is exploited in the design of the dynamic output feedback stabilizer [21]. This tool was extended to the regulation problem by Marconi et al.[32]. Regulation is more complicated when compared to the stabilization problem, since it not only involves the stabilization of the error system, but also the asymptotic reconstruction of the steady-state control c(w, θ). As a matter of fact, the steady state control has to be reconstructed by adding a marginally stable internal model unit to the compensator. Thus, the regulation problem clearly poses a greater challenge than the stabilization case. In the next subsection we present the state feedback control design approach that addresses the above-mentioned challenges. 3.2.2 State Feedback Regulator Design As in the stabilization case, the state feedback design for the system (3.6), (3.7) is carried out by utilizing virtual signals ξ1 , . . . , ξρ−1 and σ = b(·) − an (·)K1 ζ − a(·)c(·) + ∆a (η, ξ, eρ , w, θ)u, 63 (3.8) where ∆a (η, ξ, eρ , w, θ) = a(η, ξ, eρ , w, θ) − an (ξ, eρ ). These signals can be estimated by introducing the following extended-high gain observer (EHGO). ˙ ˆ ˆ ˆ ξ = P ξ + Qˆρ + H(ε)(y − ξ1 ), e ˙ ˆˆ ˆ eρ = σ + an (ξ, eρ )(u + K1 ζ) + (ˆ ρ /ερ )(y − ξ1 ), ˆ ˆ α ˙ ˆ σ = (ˆ ρ+1 /ερ+1 )(y − ξ1 ), ˆ α where H(ε) = α1 /ε, α2 /ε2 , · · · ˆ ˆ αρ−1 /ερ−1 ˆ ⊤ , and an (ξ, eρ ) is a nominal function used in the subsequent state feedback control design in place of a(η, ξ, eρ , w, θ). In the above observer equations, we have followed the standard notational convention of labeling the estimates of the virtual signals with the hatted versions of the corresponding symbols. It can be shown [16] using high gain observer theory [3] in conjunction with singular perturbation analysis [31] ˆ ˆ that the signals ξ1 , . . . , ξρ−1 , eρ , σ become arbitrarily close to ξ1 , . . . , ξρ−1 , eρ and ˆ ˆ σ, respectively. In the case of minimum phase systems, the regulation problem for a plant with well-defined relative degree can always be reduced to one of regulating a relative degree one system by utilizing a high-gain observer. For example, a virtual ρ−1 output could be defined as s = eρ + i=1 ki ξi , with the constants ki chosen so as to render the system minimum phase with respect to the output s. Then, a static control law utilizing the virtual output s could be designed in accordance with standard techniques, and this design could be implemented by replacing s with its estimate from a high-gain observer. This technique will not work if the system is non-minimum phase, however. Since we are considering non-minimum phase systems in this work, we introduce, in analogous fashion to the stabilization case of [36], the dynamic compensator given 64 by ˙ φ = L(ξ, φ, ζ) + M (ξ, φ, ζ)σ. (3.9) Next, we define a new virtual output s = eρ − N (ξ, φ, ζ). (3.10) Our objective is to design the triple {L, M, N } such that the augmented system (3.6), (3.7), (3.9), (3.10) with s taken as its output is minimum phase. In order to continue with the analysis and bearing in mind the goal of reducing the regulation problem to one of stabilization, we introduce the change of variables eρ → s, ζ → ψ ζ − Xτ (w, θ) and u → v u + K1 ζ = u + K1 ψ − c(w, θ) (due to (3.5)) to obtain the following equations describing the augmented system. η = f (η, ξ, s + N (ξ, φ, ψ + Xτ ), w, θ), ˙ (3.11a) ˙ ψ = Sψ + Jξ1 , (3.11b) ˙ ξ = P ξ + Q[s + N (ξ, φ, ψ + Xτ )], (3.11c) ∂N ∂N (Sψ + Jξ1 ) − SXτ ∂ζ ∂ζ ∂N ∂N [P ξ + Q(s + N (·))] − [L + M (b − aK1 ψ + ∆a v)], − ∂ξ ∂φ s = b(·) − a(·)K1 ψ + a(·)v − ˙ ˙ φ = L(·) + M (·)[b(·) − a(·)K1 ψ + ∆a (·)v]. (3.11d) (3.11e) (3.11f) ˙ We note that if ρ = 1, then the ξ-equation in (3.11) will not exist, and ξ1 would be replaced by s + N (·). Now, let fs (η, ξ, s, φ, ψ, w, θ) = ∂N ∂N [P ξ + Q(s + N (·))] − SXτ ∂ξ ∂ζ ∂N ∂N [L(·) + M (·)(b(·) − a(·)K1 ψ)] − (Sψ + Jξ1 ), − ∂φ ∂ζ b(·) − a(·)K1 ψ − 65 gs (η, ξ, s, φ, ψ, w, θ) = a(·) − ∂N M (·)∆a (·), ∂φ fφ (η, ξ, s, φ, ψ, w, θ) = L(·) + M (·)[b(·) − a(·)K1 ψ]. The system (3.11) will have a relative degree of one with respect to s if gs (η, ξ, s, φ, ψ, w, θ) ≥ k † > 0, a(η, ξ, s, φ, ψ, w, θ) (3.12) for some positive real constant k † . When ∆a = 0, we transform (3.11) into the normal form by the change of variable T : φ → q that satisfies the partial differential equation ∂T ∂T gs + M ∆a = 0. ∂s ∂φ Due to (3.13), we note that the vector η ⊤ ξ ⊤ s q ⊤ ψ ⊤ of η ⊤ ξ ⊤ s φ⊤ ψ ⊤ (3.13) ⊤ is a diffeomorphism ⊤ . When ∆a = 0, we can take T = φ, and the system (3.11) will be in the normal form. Since a solution exists for the ∆a = 0 case, we can argue that for sufficiently small |∆a |, (3.13) still has a solution T [11]. If we let φ = T −1 (·) be the inverse transformation, then due to the continuous dependence of the solution of (3.13) on ∆a [11], we can write T and T −1 as T = φ + Φ(η, ξ, s, φ, ψ, w, θ), T −1 = q + ϕ(η, ξ, s, q, ψ, w, θ), where Φ and ϕ are of the order of O (|∆a |). The transformed system equations are now given by η = f (η, ξ, s + N (ξ, q + ϕ(·), ψ + Xτ, w, θ), w, θ), ˙ (3.14a) ˙ ψ = Sψ + Jξ1 , (3.14b) 66 ˙ ξ = P ξ + Q[s + N (·)], (3.14c) s = fs (η, ξ, s, q + ϕ(·), ψ, w, θ) + gs (·)v, ˙ (3.14d) q= ˙ ∂Φ ∂Φ fφ + [P ξ + Q(s + N )] ∂φ ∂ξ ∂Φ ∂Φ ∂Φ ∂Φ f+ fs + S0 w + (Sψ + Jξ1 ). + ∂η ∂s ∂w ∂ψ 1+ (3.14e) (3.14f) With s viewed as the output, and due to (3.12), this is a relative degree one system in the normal form. The zero dynamics shall have trajectories confined to the set {(η, ξ, s, q, ψ)|s = 0}, and are given by η = f (η, ξ, N (ξ, q + ϕ, ψ + Xτ, w, θ), w, θ), ˙ (3.15a) ˙ ψ = Sψ + Jξ1 , (3.15b) ˙ ξ = P ξ + QN (·), (3.15c) q= ˙ 1+ ∂Φ ∂φ fφ + ∂Φ ∂Φ ∂Φ ∂Φ f+ (P ξ + QN ) + (Sψ + Jξ1 ) + S w. (3.15d) ∂η ∂ξ ∂ψ ∂w 0 When ∆a = 0, (3.15) reduces to η = f (η, ξ, N (ξ, q, ψ + Xτ, w, θ), w, θ), ˙ (3.16a) ˙ ψ = Sψ + Jξ1 , (3.16b) ˙ ξ = P ξ + QN (·), (3.16c) q = L(·) + M (·)[b(·) − a(·)K1 ψ]. ˙ (3.16d) This can be viewed as a feedback interconnection between the extended auxiliary plant of [32]: η = f (η, ξ, ua , w, θ), ˙ (3.17a) ˙ ξ = P ξ + Qua , (3.17b) 67 ˙ ψ = Sψ + Jξ1 , ya = b(·) − a(·)K1 ψ, (3.17c) (3.17d) and the auxiliary controller ˙ φ = L(ξ, φ, ζ) + M (ξ, φ, ζ)ya , ua = N (ξ, φ, ζ). (3.18a) (3.18b) We note that the system (3.17)–(3.18) is very similar to the extended auxiliary system of [32]—the only difference is in the structure of the servocompensator given by the ˙ ˙ ψ-equation. In [32], both the ψ and the q equations are driven by b(·) − a(·)K1 ψ, ˙ ˙ whereas here, the ψ equation is driven by ξ1 . This fact makes the analysis of the closed-loop system tractable in the case of unknown control coefficient a(·), whereas in the method of [32], the lack of knowledge of this function would complicate the analysis because the counterpart of the transformation that maps ζ to ψ would depend on a(·) using this method; moreover, the control coefficient is assumed to be 1 in this latter work. Next, let xs = η⊤ ψ⊤ ξ ⊤ q⊤ ⊤ and   f (η, ξ, N (ξ, q + ϕ, ψ + Xτ, w, θ), w, θ)      Sψ + Jξ1   F (xs , s, w, θ) =  ,   P ξ + Q[s + N (xs )]     f1 (xs , s + N (xs ), w, θ) where f1 (xs , s + N (xs ), w, θ) = 1 + ∂Φ fφ + ∂Φ f + ∂Φ (Sψ + Jξ1 ) + ∂Φ [P ξ + ∂φ ∂η ∂ψ ∂ξ ∂Φ S w. With these newly defined variables, (3.14) can be rewritten Q(s + N (xs ))] + ∂w 0 68 as xs = F (xs , s, w, θ), ˙ (3.19) s = fs (xs , s + N, w, θ) + gs (xs , s + N, w, θ)v. ˙ (3.20) We note that the system (3.19)–(3.20) is relative degree one with s viewed as the output. Our goal now is to design the dynamic compensator {L, M, N } such that the resulting augmented system is minimum phase, and the internal dynamics (3.19) are input-to-state stable with s viewed as an input to the latter. Finally, the stabilizing control input v for the augmented system can be designed using any viable robust control technique for relative degree one minimum phase systems such as high gain feedback, Lyapunov redesign, etc. We choose continuously implemented sliding mode control (SMC) in this chapter, but it is important to note that we are not confined to any particular technique. SMC offers several advantages such as control saturation at a predetermined level—which mitigates the effects of the peaking phenomenon caused by the high gain observer—trajectory convergence to a region inside the boundary layer in finite time, and high gain feedback inside the aforementioned region. The control input v is thus designed as v=− β(ξ, eρ , φ, ζ) sat an (ξ, eρ ) s µ , (3.21) where µ is a constant design parameter that specifies the thickness of the boundary layer, and β(·) is chosen in accordance with Assumption 2.2 and inequality (3.22) below. Assumption 3.3. It is assumed that a(xs , s + N (xs ), w, θ)/an (xs , s + N (xs )) ≥ k0 > 0, 69 for all (xs , s) ∈ D and (w, θ) ∈ W × Θ. Moreover, an (xs , s + N (xs )) is locally Lipschitz in (xs , s) over the domain of interest, and globally bounded in ξ. The function β(·) is also chosen to satisfy β(ξ, eρ , φ, ζ) ≥ β0 + an (ξ, eρ )fs (xs , s + N, w, θ) , gs (xs , s + N, w, θ) (3.22) where β0 > 0. Assumption 2.2 and (3.22) ensure that when s ≥ µ, we have ss ≤ ˙ −k † k0 β0 |s|, and so whenever |s(0)| > µ, |s(t)| will decrease until it reaches the set {|s| ≤ µ} in finite time and remain inside thereafter. The closed-loop state feedback system is given by (3.19), (3.20), (3.21). We require this system to satisfy the following assumptions. Assumption 3.4. 1) Let D ⊂ Rm+n+r be a domain containing the origin, and let (xs , s) ∈ D. Moreover, suppose there exists a continuously differentiable Lyapunov function V (xs , w, θ) such that α2 ( xs ) ≥ V (xs , w, θ) ≥ α1 ( xs ), −α3 ( xs ) ≥ ∂V ∂V F (xs , s, w, θ) + S w, ∂xs ∂w 0 (3.23a) ∀ xs ≥ γ(|s|), (3.23b) for all (xs , s) ∈ D ⊂ Rm+n+r , and (w, θ) ∈ W × Θ, where α1 (·), α2 (·), α3 (·) and γ(·) are class K functions. 2) With s = 0, the origin xs = 0 of (3.19) is locally exponentially stable. Remark 3.1. 1. The set D is well-defined for any sufficiently smooth N (·) in the definition of the variable s = eρ − N (·). 70 2. Inequality (3.23b) is equivalent to regional input-to-state stability of (3.19) with s viewed as an input. Assumption 3.5. 1) L(xs ), M (xs ) and β(xs ) are locally Lipschitz functions in their arguments over the domain of interest, and L(0) = 0. 2) L(xs ) and M (xs ) are globally bounded functions of ξ. Remark 3.2. We require the global boundedness assumption on L(xs ) and M (xs ) to protect the system from the peaking phenomenon that may occur in the output feedback implementation when the extended high-gain observer is introduced [3, 29]. The peaking will be transmitted to the plant states in the closed-loop system, and in the absence of the above assumption, there is a possibility of the state trajectories having a finite escape time. Under the above assumptions, we obtain the following result pertaining to the stability of this state feedback error system. Proposition 3.1. Consider the closed loop state feedback error system (3.19), (3.20). Let Assumptions 3.1 through 3.5 hold. Then, there exists a positive constant µ∗ such that (3.19), (3.20) has an exponentially stable equilibrium at (xs = 0, s = 0) for all µ ∈ (0, µ∗ ]. Moreover, the set Ω defined below is an estimate of the region of attraction. Ω {V (xs , w, θ) ≤ c0 } × {|s| ≤ c}, (3.24) where c > µ, c0 ≥ α2 (γ(c)), and c, c0 are chosen such that Ω is a compact subset of D. Proof. See Section A.4 for the proof. Remark 3.3. 71 1) Assumption 2.3 requires the compensator {L, M, N } to be designed such that (3.19) is input to state stable with s viewed as the input. This problem is similar to the auxiliary design problem of [32] when ∆a = 0, with the difference being in the structure of the servocompensator employed in each work. In [32], the interconnection of the extended auxiliary system with its dynamic compensator was assumed to be globally asymptotically stable and locally exponentially stable, with a Lyapunov function independent of the unknown system parameters. 2) If all the assumptions hold globally, then we could take D = Rm+n+r , and with the appropriate choice of controller parameter β(·) (see (3.22)), the positively invariant set Ω ⊂ D that depends on this parameter can be made arbitrarily large, and thus our design allows for semi-global stabilization of the state feedback error system. 3.2.3 Output Feedback Control Using Extended High-Gain Observers 3.2.3.1 Output Feedback Regulator Design In order to estimate the states e of the error system (3.6) and the virtual output σ, an extended high-gain observer with the following structure is introduced. ˙ ˆ ˆ ˆ ξ = P ξ + Qˆρ + H(ε)(y − ξ1 ), e ˙ ˆˆ ˆ eρ = σ + an (ξ, eρ )(u + K1 ζ) + (ˆ ρ /ερ )(y − ξ1 ), ˆ ˆ α ˙ ˆ σ = (ˆ ρ+1 /ερ+1 )(y − ξ1 ), ˆ α (3.25a) (3.25b) (3.25c) ⊤ , and the constants α are chosen ˆ α1 /ε, α2 /ε2 , · · · αρ−1 /ερ−1 ˆ ˆ ˆ so as to make the polynomial λρ+1 + α1 λρ + . . . + αρ λ + αρ+1 Hurwitz. The ˆ ˆ ˆ where H(ε) = structure of this observer is very similar to the one used in Chapter 2, with the only 72 difference being the fact that this observer is driven by the signal v = u + K1 ζ instead of by u. It shall be established during the analysis of the output feedback ˆ system to follow that as ε → 0, we shall have ξ → ξ, eρ → eρ , and moreover, ˆ σ → b − an K1 ζ − ac + (a − an )u = σ. ˆ The state and virtual output estimates obtained from this observer are used to replace the unavailable states and unmeasured signals in the state feedback design, and the extended high-gain observer-based output feedback regulator is given by (3.25) and ˙ ζ = Sζ + Jy, (3.26a) ˙ ˆ ˆ φ = L(ξ, φ, ζ) + M (ξ, φ, ζ)ˆ , σ (3.26b) ˆ u = K sat(l(ξ, φ)/K) − K1 ζ, (3.26c) l(·) = ˆˆ −β(ξ, eρ , φ, ζ) sat ˆˆ an (ξ, eρ ) ˆ eρ − N (ξ, φ, ζ) ˆ µ , (3.26d) where K> max (xs ,s)∈Ω ˆˆ β(ξ, eρ , φ, ζ) . ˆˆ an (ξ, eρ ) (3.27) The sliding mode control component v is saturated at ±K outside Ω in order to protect the system from peaking during the observer’s transient response. We now introduce a change of variables as follows. ˆ νi = (ξi − ξi )/ερ+1−i , 1 ≤ i ≤ ρ − 1, νρ = (eρ − eρ )/ε, ˆ (3.28a) (3.28b) νρ+1 = b(η, ξ, eρ , w, θ) − a(η, ξ, eρ , w, θ)K1 ψ − σ ˆ + ∆a (η, ξ, eρ , w, θ)Kgε (l(xs , s)/K), 73 (3.28c) (3.28d) where gε is an odd function defined by gε (z) =   z,          0 ≤ z ≤ 1, z + (z − 1)/ε − 0.5(z 2 − 1)/ε, 1 < z < 1 + ε, 1 + 0.5ε, z ≥ 1 + ε, ′ and we note that gε ≤ 1 and |gε (z) − sat(z)| ≤ ε/2. Moreover, (3.28) is very similar to the change of variables utilized in Chapter 2, except for the additional term involving ψ in the νρ+1 -equation. With the change of variables (3.28), and after applying the transformations ζ → ψ, eρ → s and φ → q, the closed-loop output feedback system (3.6), (3.25), (3.26), can be written as xs = F (xs , s, w, θ) + G(xs , s, ν, ε, w, θ), ˙ (3.29a) s = fs (·) + gs (·)K sat(l(xs , s)/K) + εK1 Jνρ a(xs , s + N (xs ), w, θ) ˙ (3.29b) − a(xs , s + N (xs ), w, θ)K[sat(l(xs , s)/K) − sat(ln (xs , s, ν, ε)/K)], (3.29c) and ¯ ¯ ¯ εν = Aν + ε[B1 ∆1 (η, ξ, eρ , ν, w, θ, ε) + B2 ∆2 (η, ξ, eρ , ν, w, θ, ε)], ˙ where  0        0   G(xs , s, ν, ε, w, θ) =  ,   0     M (xs , ν, ε)∆a Kgε (l(xs , s)/K) − νρ+1 74 (3.30) l(xs , s) = ln (·) = −β(ξ, eρ , φ, ζ) sat an (ξ, eρ ) ˆˆ −β(ξ, eρ , φ, ζ) sat ˆˆ a(ξ, eρ ) eρ − N (ξ, φ, ψ + Xτ ) µ , eρ − ενρ − N (xs , ν, ε) µ ˆˆ = l(ξ, eρ , q, ζ),  −ˆ 1 α    −ˆ  α2  . ¯ A= .  .    −ˆ ρ α  −ˆ ρ+1 α 1 0 ··· 0 . . . 1 . . . 0 0 0 0   0     0     ¯ = . B1  .  , .     0   1 (ρ+1)×1 0    · · · 0  ... . . , .   · · · 1  ··· 0 (ρ+1)×(ρ+1)   0      0     ¯ = . B2  .  , .      1   0 (ρ+1)×1 ∆2 (η, ξ, eρ , ν, w, θ, ε) = ∆0 (η, ξ, eρ , ν, w, θ, ε)/ε, ∆0 (η, ξ, eρ , ν, w, θ, ε) = −K∆a (η, ξ, eρ , w, θ)[gε (l(·)/K) − gε (ln (·)/k)] +[a(η, ξ, eρ , w, θ) − an (ξ, eρ , ν, ε)]K[sat(ln (xs , s, ν, ε)/K) −gε (ln (xs , s, ν, ε)/K)] + K[an (ξ, eρ ) − an (ξ, eρ , ν, ε)]gε (ln (xs , s, ν, ε)/K) ∆1 (η, ξ, eρ , ν, w, θ, ε) = ′ ˙ ˙ b(η, ξ, eρ , w, θ) − a(η, ξ, eρ , w, θ)K1 ψ + ∆a (η, ξ, eρ , w, θ)gε (l(xs , s)/K)l(ξ, eρ , φ) ˙ ˙ ˙ −a(η, ξ, eρ , w, θ)K1 ψ + ∆a (η, ξ, eρ , w, θ)Kgε (l(xs , s)/K). 75 We note that the functions ∆0 and ∆1 are locally Lipschitz in their arguments. This is because an (·) and gε (·) are continuously differentiable with locally Lipschitz derivatives and globally bounded. Moreover, we can use (3.28) and the definition of gε to show that ∆0 /ε is locally Lipschitz [16]. We can now write the slow dynamics in a compact form as xo = fr (xo , ν, ε, w, θ), ˙ where xo = x⊤ s s (3.31) ⊤ ,   F (xs , s, w, θ) + G(xs , s, nu, ε, w, θ) fr (xo , ν, ε, w, θ) =  , f2 (xs , s, ν, ε, w, θ) f2 (xs , s, ν, ε, w, θ) = b1 (xs , s, w, θ) + a(·)[K sat(ln (·)/K) + εK1 Jνρ ]   ∂N  0  ∂N − F (xs , s, w, θ),   ∆a K sat(ln (·)/K) − ∂xs M (x ) ∂xs a s where b1 (xo , w, θ) = b(η, ξ, eρ , w, θ) − K1 ψa(η, ξ, eρ , w, θ). Then, the reduced system is given by xo = fr (xo , 0, 0, w, θ), ˙ (3.32) and it is easily verified that in the compact set Ω, given by (3.24), this reduced system is identical to the state feedback system (3.19), (3.20), which, due to Proposition 3.1, is exponentially stable with respect to its zero error manifold {η = 0, e = 0}. The boundary layer system is given by ∂ν ¯ = Aν, ∂τ ˜ 76 t τ= . ˜ ε (3.33) ¯ Since A is Hurwitz, the boundary layer system is exponentially stable with respect to {ν = 0}. At this stage in our analysis, we are ready to state some results pertaining to the recovery of the performance of the system (3.19), (3.20) in the case of output feedback as ε → 0. These results follow directly from the work of [3] because the singularly perturbed system (3.29)–(3.30) is of the same form as the system described by equations (10)–(13) in [3]. Consequently, the analysis of (3.29)–(3.30) can be carried out in the same vein, and so only the results of that analysis by [3] are reproduced below. Moreover, while the structure of system (3.29)–(3.30) is also very similar to that of the closed-loop system analyzed by [16], a key difference between our work and the latter is that both the perturbation terms in the fast dynamics (3.30) are O(ε) in this work, while there is one term in [16] that is not. Such a term appeared in [16] because the estimate σ was used in the control u, whereas in our ˆ work, this quantity is not utilized directly in the static control. 3.2.3.2 Exponential Stability of the Output Feedback System and Trajectory Convergence Let (xo (t, ε), ν(t, ε)) denote the trajectory of the system (3.30), (3.31) starting from ˆ (xo (0), ξ(0)). Also, let xr (t) be the solution of (3.19), (3.20) starting at xo (0). We know that the trajectory (xs (t), s(t)) of the system (3.19), (3.20) is exponentially stable with respect to {(xs , s) = (0, 0)}, so let us suppose that R is its region of attraction. Let S be a compact set in the interior of R, and let Q be a compact subset of Rρ+1 . The recovery of exponential stability of the zero error manifold, and the fact that xo (t) converges to xr (t) as ε → 0, uniformly in t, for all t ≥ 0, is established by the following theorem, [3, Theorems 1, 2, 3 and 5], provided that ˆ xo (0) ∈ S and ξ(0) ∈ Q. 77 Theorem 3.1. Let Assumptions 3.1 through 3.5 hold. Moreover, assume xo (0) ∈ S and ν(0) ∈ Q. Then, 1. there exists ε∗ > 0 such that, for every 0 < ε ≤ ε∗ , the system (3.30), (3.31) is 1 1 exponentially stable with respect to the manifold {(xo , ν) = (0, 0)}, and S × Q is included in the region of attraction. 2. given any δ > 0, there exists ε∗ > 0 such that, for every 0 < ε ≤ ε∗ , we have 2 2 xo (t, ε) − xr (t) ≤ δ, ∀t ≥ 0. (3.34) Remark 3.4. Theorem 1 of [3] establishes boundedness of (xo , ν) for all trajectories starting in S ×Q. Theorem 2 of [3] shows that the trajectories are ultimately bounded, and that the ultimate bound can be made arbitrarily small by choosing ε small enough. Finally, Theorem 5 of [3] establishes that {(xo , ν) = (0, 0)} is exponentially stable with a region of attraction that is independent of ε. The combination of the three results yields the first item of the foregoing theorem. The second item follows from Theorem 3 of [3]. 3.3 Example We illustrate the design procedure for an extended high-gain observer based sliding mode output feedback regulator using a linearized inverted pendulum on a cart model as an example. The system is given by [10] x z1 = z2 − 2 , ˙ h g z2 = z1 , ˙ h x1 = x2 , ˙ (3.35a) (3.35b) (3.35c) 78 x2 = (u + w1 − mp gz1 ) ˙ 1 , mc y = x1 , (3.35d) (3.35e) where z1 is the angle of the pendulum with respect to the vertical, x1 is the position of the cart, u is the control input, w1 is a sinusoidal disturbance of known frequency ω, mc is the mass of the cart, mp is the mass of the pendulum (assumed to be concentrated in the bob), h is its length, and g is the acceleration due to gravity. mp and h are assumed to be uncertain parameters in the ranges 0.19 kg ≤ mp ≤ 0.23 kg and 0.55 m ≤ h ≤ 0.67 m. An exosystem generates the sinusoidal disturbance signal w1 , and this system is modeled by w = S0 w, where ˙    0 ω S0 =  . −ω 0 The goal is to utilize the output measurement y to drive a feedback controller designed to make this system output converge to zero despite the presence of the input disturbance w1 and uncertain parameters mp and h. We conclude from the steady-state analysis of (3.35) that the steady-state control is c(w) = −w1 , and this allows us to determine the S matrix of the servocompensator that satisfies Assumption 3.2 and (3.5), and an appropriate gain matrix K1 to make S − JK1 Hurwitz. We now proceed with the first step in our design by augmenting (3.35) with the servocompensator ˙ ζ = Sζ + Jy, where   1  0 S=  −ω 2 0   0 and J =   . 1 79 (3.36) Next, we choose a matrix K1 such that the eigenvalues of (S − JK1 ) are −1 and −2. The matrix is thus given by K1 = 3.3.1 3 − ω2, 2 . The Auxiliary Problem After applying the change of variable ζ → ψ = ζ −Xτ , the auxiliary system associated with the augmented system (3.35), (3.36) is given by     ˙  ψ  S 0 0    z   0 0 1  ˙1    =  z   0 g/h 0  ˙2      x1 ˙ 0 0 0 −1 mc   J  ψ   0      0   z1  −1/h      ua ,   +  0   z2   0          x1 1 0 A ya =  (3.37a) B −K1 , mp g, 0, 0 ψ ⊤ z 1 z 2 x1 ⊤ . (3.37b) C We now design an observer based controller for (3.37), and it is given by ˙ φ = (A − HCn − BK)φ + Hya ), (3.38a) ua = −Kφ, (3.38b) ˆ where Cn = −1 −K1 , mp g, 0, 0 is a nominal version of the matrix C, mc is mc ˆ a nominal parameter used in lieu of mc , and the gain matrices K and H are designed using the LQR technique to make A − BK and A − HCn Hurwitz. These gain matrices were designed to be K= −0.022, 0.100, −17.099, −4.265, −10.000 , H= 0.082, 0.568, −4.763, −19.097, −0.100 ⊤ 80 . 3.3.2 The Output Feedback Regulator The EHGO-based output feedback regulator for (3.35) is designed in accordance with Section 3.2.3, and is thus given by ˙ x1 = x2 + (ˆ 1 /ε)(y − x1 ), ˆ α ˆ (3.39a) ˙ x2 = σ + v/mc + (ˆ 2 /ε2 )(y − x1 ), ˆ ˆ ˆ α ˆ (3.39b) ˙ σ = (ˆ 3 /ε3 )(y − x1 ), ˆ α ˆ (3.39c) ˙ ζ = Sζ + Jy, (3.39d) ˙ φ = (A − HCn − BK)φ + H σ ), ˆ (3.39e) v = −β mc sat ˆ x2 + Kφ ˆ µ u = v − K1 ζ. 3.3.3 , (3.39f) (3.39g) Numerical Simulations Simulations were performed for the output feedback design given in Section 3.3.2, and the results were compared with that of the design provided in [10]. The results are depicted in the figures that follow. 3.3.3.1 Discussion A comparison was made with the design presented in this chapter and the one based on [10] by means of simulations. The results showed that both the high-gain feedback design method of [10] and the EHGO-based design can achieve regulation for the targeted range of uncertainties in the plant parameters mp and h. Figure 3.1 to Figure 3.4 show the state and output trajectories for the case when mc = mc = 0.82 ˆ kg. It can be seen from these figures that the EHGO design provides better transient performance and shorter settling times. Another key difference is the performance 81 0.5 0.4 0.3 0.2 z1 0.1 0 −0.1 −0.2 −0.3 −0.4 −0.5 0 50 t (s) 100 Figure 3.1: Trajectories of z1 for the nominal value of mc : design (solid), and design from [10] (dash-dot). 150 EHGO + SMC in the presence of uncertainty in the nominal plant parameter mc , which also affects the control coefficient, which is basically 1/mc , and the range of uncertainty allowed for by each design. The high-gain feedback design of [10] achieves regulation when mc ∈ [0.819, 0.8206] kg, which is a range of uncertainty from −0.12% to +0.0732%, while the EHGO design allows for mc ∈ [0.777, 0.84] kg, which is a range of −5.24% to +2.44%. Moreover, the control effort corresponding to the EHGO method is consistently lower than the method of [10] by a factor of about 5 during the transient phase—this can be seen in Figure 3.5. Figure 3.6 shows the output trajectory of the EHGO design when mc = 0.777 kg—the design of [10] is unstable for this particular parameter value. 82 0.8 0.6 0.4 z2 0.2 0 −0.2 −0.4 −0.6 −0.8 0 50 t (s) 100 150 Figure 3.2: Trajectories of the state z2 for the nominal value of mc . 3.4 Concluding Remarks In this chapter, an output feedback regulator design was presented for non-minimum phase systems with uncertainties in the control coefficient, and in the constant system parameters. This design utilizes an extended high-gain observer and sliding mode control. The fundamental idea is to use the knowledge of a stabilizing controller for an associated auxiliary problem in designing the controller for the full system—this approach was first introduced by [21] and later used by [32] to solve the regulation problem. The contribution of this chapter is that it provides a design procedure that reduces the problem of controlling a non-minimum phase nonlinear system to one of designing a controller for a relative degree one minimum phase augmented system under certain assumptions, and uncertainty in the control coefficient is considered, as well. It is shown that this design is able to solve the regulation problem in the 83 0.5 0.4 0.3 y 0.2 0.1 0 −0.1 −0.2 −0.3 0 50 t (s) 100 150 Figure 3.3: Trajectories of the output for the nominal value of mc . presence of the aforementioned uncertainty, provided it is small enough. Some numerical simulations of a linearized inverted pendulum on a cart model are included to illustrate the procedure and the resulting performance. 84 5 4 3 2 x 2 1 0 −1 −2 −3 −4 −5 0 50 t (s) 100 150 Figure 3.4: Trajectories of the derivative of the output for the nominal value of mc . 85 u 100 0 −100 0 50 100 150 100 150 20 u 10 0 −10 −20 0 50 t (s) Figure 3.5: The control effort of the algorithm from [10] (top) and the EHGO based design (bottom). 86 8 7 6 y 5 4 3 2 1 0 0 50 100 150 t (s) 200 250 300 Figure 3.6: The response of the EHGO design when mc = 0.777 kg. 87 Chapter 4 Stabilization and Regulation of the TORA System 4.1 Introduction The translational oscillator with a rotating actuator (TORA) system, or rotationaltranslational actuator (RTAC), as it is also known, has been widely used in the literature for the past several years as a benchmark problem for testing various nonlinear control schemes. The problem was introduced in [43, 6]. In [22], two control laws are presented, the first a cascade controller and the second one a feedback passivating controller. A control law based on an L2 disturbance attenuation approach was proposed in [41]. Output feedback controllers dependent on measurements of the rotational and translational positions were proposed in [23], while controllers requiring only the measurement of the rotational position were presented in [25] and [13]. In [7], four different controllers were experimentally evaluated on an RTAC testbed. The TORA problem was originally conceived as a simplified version of the dynamics of dual-spin spacecraft [43]. The interaction between rotation and translation in the oscillating eccentric rotor is analogous to the interaction between spin and 88 Ñ Ñ                                                         Å                   ¢ ¢ ¢              ¢ ¢ ¢                                                                                            Ñ Ñ × Ü Ù ¹ Ü  Ä  Ý Ñ × Figure 4.1: Translating Oscillator with Rotating Actuator (TORA). nutation in a dual-spin spacecraft. 4.2 TORA Dynamics: A Simplified Model of a Dual Spin Spacecraft Figure 4.1 depicts a schematic diagram that is based on experimental setups of the TORA benchmark system.1 This system consists of a platform or cart of mass mc connected to a fixed reference frame via a linear spring with spring constant k. The cart is constrained to only move horizontally. Attached to the cart is an unbalanced mass mr with moment of inertia I about its center of mass, located at a distance lr from its rotational axis. Let θ be the rotational angle of the unbalanced mass, and xc denote the translational position of the center of mass of this proof mass. The control torque applied to the proof mass at its axis of rotation is denoted by u, and we also assume that there is a disturbance torque d acting at the same point— this disturbance signal could be a model of the effect of friction, or a periodic and 1 Figure taken from [29]. 89 persistent disturbance torque generated by a collision between the proof mass and an obstacle, for example. The rotating proof mass can be used to transfer a linear force to the cart, which in turn can be used to dampen the translational oscillations of the platform by a suitable design of the input torque u to the proof mass. We now derive the dynamic model for the TORA system. Figure 4.1 shows a free body diagram of the proof mass next to the schematic, and as indicated by this diagram, the pivot where the proof mass is attached to the platform is subject to the external input torque u and the forces Fx and Fy applied by the cart in reaction to u and d. Hence, Newton’s Second Law yields the following equations at the center of mass of the proof mass [29] d2 (xc + lr sin θ) = Fx , dt2 d2 (lr cos θ) = Fy , mr dt2 ¨ u + Fy lr sin θ − Fx cos θ = I θ. mr By Newton’s Third Law, the platform is subject to the forces −Fx and −Fy , but the force equation along the vertical is trivial and yields no information, since the frame of the testbed is constructed to absorb forces in the vertical direction and constrain the motion of the platform to the horizontal direction only. Newton’s Second Law gives the following dynamic equation governing the motion of the platform. mc xc = −Fx − kxc . ¨ Upon eliminating the internal forces Fx and Fy and carrying out the differentiation to the fullest extent possible, the dynamics of the TORA system can be expressed as 90 follows.    ¨ u  θ  D(θ)   =  , ˙ mr lr θ2 sin θ − kxc xc ¨  where   2 m l cos θ rr  I + m r lr  D(θ) =  . mr lr cos θ mr + mc The determinant of D(θ) is 2 2 2 δ(θ) = (I + mr lr )(mr + mc ) − m2 lr cos2 θ ≥ (I + mr lr )mc + mr I > 0, r ¨ and so we can solve the above equation for θ and xc to obtain ¨      ¨ u 1  mr + mc −mr lr cos θ   θ   .  = δ(θ) −m l cos θ I + m l2 ˙ mr lr θ2 sin θ − kxc xc ¨ rr rr (4.1) ¨ We note that since both the θ and xc equations in (4.1) above depend on u, the ¨ ⊤ ˙ obvious choice of state variable, to wit q = xc xc θ θ , will not yield a set ˙ of equations for the plant dynamics in the normal form. Consequently, we apply the following change of variables [29] z 1 = xc + mr lr sin θ , mr + mc z 2 = xc + ˙ ˙ mr lr θ cos θ , mr + mc x1 = θ, ˙ x2 = θ, (4.2) and it can be verified that (4.2) transforms (4.1) into the following normal form. z1 = z2 , ˙ z2 = ˙ k mr + mc (4.3a) mr lr sin x1 − z1 , mr + mc x1 = x2 , ˙ (4.3b) (4.3c) 91 x2 = ˙ 1 δ(x1 ) (mr + mc )u − mr lr cos x1 mr lr sin x1 × mr lr x2 sin x1 − k z1 − 2 mr + mc (4.3d) , y = x1 . 4.3 (4.3e) (4.3f) Stabilization of the TORA System In this section, we consider the problem of stabilizing the TORA system about its equilibrium point, which is the origin. The model equations in the normal form are given by (4.3), and we make the substitutions η1 = z1 , η2 = z2 , ξ1 = x1 and ξ2 = x2 for convenience, so that the equations in this section match the notation used in Chapter 2. Moreover, in this section, we do not consider disturbances or reference signals, and so (4.3) now becomes. η1 = η2 , ˙ η2 = ˙ (4.4) k mr + mc ˙ ξ1 = ξ2 , 1 ˙ ξ2 = δ(ξ1 ) mr lr sin ξ1 − η1 , mr + mc (4.5) (4.6) (mr + mc )u − mr lr cos ξ1 mr lr sin ξ1 2 × mr lr ξ2 sin ξ1 − k η1 − mr + mc y = ξ1 . , (4.7) (4.8) Uncertainty in the cart mass mc and rotor mass mr results in a system with uncertain functions a(·) and b(·) corresponding to (2.1)–(2.4). The design is carried out for nominal values of mr and mc , and then the effect of perturbing these parameters is studied via simulation. We assume that the output y is the only measured variable, and the goal is to stabilize this system about its origin by feeding back the output. 92 In accordance with the technique presented in Chapter 2, we assume that the signals ξ and σ = b(η, ξ, θ) are available, since they can be estimated by the extended high-gain observer to be introduced later, and so we augment (4.4)–(4.8) with a compensator ˙ φ = L(φ, ξ1 ) + M (φ, ξ1 )σ, (4.9) and we define a new variable s = ξ2 − N (φ, ξ1 ), where L(·), M (·) and N (·) are to be designed with the goal of rendering the resulting system minimum phase when s is viewed as its output. We now perform the change of variable ξρ → s. We thus obtain the augmented system η1 = η2 , ˙ η2 = ˙ (4.10a) k mr + mc mr lr sin ξ1 − η1 , mr + mc ˙ ξ1 = ξ2 , 1 s= ˙ (mr + mc )u − mr lr cos ξ1 δ(ξ1 ) mr lr sin ξ1 2 × mr lr ξ2 sin ξ1 − k η1 − mr + mc ∂N [s + N (φ, ξ1 )], − ∂ξ1 mr lr cos ξ1 ˙ φ = L(φ, ξ1 ) + M (φ, ξ1 ) k η1 δ(ξ1 ) mr lr sin ξ1 − mr lr (N (φ, ξ1 ))2 sin ξ1 , − mr + mc (4.10b) (4.10c) (4.10d) (4.10e) (4.10f) (4.10g) (4.10h) which is clearly relative degree one with respect to s, and with this signal viewed as the output, we see that the zero dynamics {(η, ξ1 ) | s = 0} obey η1 = η2 , ˙ η2 = ˙ k mr + mc (4.11) mr lr sin ξ1 − η1 , mr + mc 93 (4.12) ˙ ξ 1 = ua , (4.13) ua = N (φ, ξ1 ), (4.14) ˙ φ = L(φ, ξ1 ) + M (φ, ξ1 )σ, mr lr sin ξ1 σ = k η1 − mr + mc (4.15) mr lr cos ξ1 − mr lr u2 sin ξ1 . a δ(ξ1 ) (4.16) We note that this system is identical to the closed-loop auxiliary system as introduced by Isidori [21]. The problem is now reduced to one of stabilizing (4.11)–(4.16), and we do this in the subsections to follow. 4.3.1 Design of the Stabilizing Controller In order to design the stabilizing compensator (4.14)–(4.16), we shall adopt the strategy of first linearizing the input-output map of (4.11)–(4.13), then designing a linear compensator for the transformed system, and then implementing the inverse transformations to obtain the nonlinear control {L, M, N } in the original coordinates. Hence, we define a new output for the auxiliary system as ya = ˜ δ(ξ1 )σ mr lr sin ξ1 m r lr 2 ua sin ξ1 + + mr lr k cos ξ1 k mr + mc ˜ ha (ya , ξ1 , ua ). (4.17) With the newly defined output ya , the output equation is given by ˜ ya = η1 . ˜ (4.18) The design of a stabilizing feedback controller for the system (4.11)–(4.13) and (4.18) can be simplified by viewing sin ξ1 as a virtual control input to the system (4.11)– (4.12), and then using either backstepping or high-gain feedback to determine the 94 control ua . We shall use the high-gain feedback approach. To this end, let ua = (va − ξ1 )/εa , where va is a control input to be designed and εa is a positive number. The open loop auxiliary system can now be expressed in terms of (4.11), (4.12), (4.18) and ˙ εa ξ 1 = va − ξ 1 . (4.19) For εa small enough, (4.11), (4.12) and (4.19) will exhibit a two time-scale behavior and can thus be regarded as a singularly perturbed system. The reduced system, obtained by setting εa = 0 ⇒ va = ξ1 , is given by (4.11), (4.12) and (4.18), and for this reduced system, the quantity sin ξ1 = sin va wa (4.20) is regarded as the control input, and can now be designed as a linear control. To summarize, the control problem for the auxiliary system reduces to a problem of finding a compensator with a strictly proper transfer function for the linear system η1 = η2 , ˙ η2 = ˙ k mr + mc (4.21) m r lr w a − η1 , mr + mc ya = η1 . ˜ (4.22) (4.23) It can be shown that the above system is observable and controllable, and a state feedback controller wa = k1 η1 + k2 η2 was found using LQR theory. The values of k1 and k2 providing near-optimal trade-off between the transient performance of the states and the control effort were computed in terms of a parameter ρ1 used to 95 minimize the cost function J= ∞ 0 η1 − 2 m r lr 2 + ρ1 wa dt, wa mr + mc subject to the constraint |wa | ≤ 1. This parameter was tuned to ensure that wa just meets the constraint. This yielded the parameter values of k1 = −1 and k2 = −8. Next, a full-order observer was designed for (4.21)–(4.23) as ˙ φ1 = φ2 + α1 (˜a − φ1 ), ˜ y (4.24) kφ1 kmr lr ˙ + φ2 = α2 (˜a − φ1 ) − ˜ y wa , mr + mc (mr + mc )2 (4.25) where α1 = 2 and α2 = 1. The observer-based control law can now be taken as ˜ ˜ wa = k1 φ1 + k2 φ2 , (4.26) and (4.24)–(4.26) provide the strictly proper dynamic compensator alluded to earlier. We now substitute the transformations into the compensator equations above and find the stabilizing controller for the auxiliary system (4.11)–(4.16) to be ˙ φ1 = φ2 + α1 (ha (σ, ξ1 , N (φ, ξ1 )) − φ1 ), ˜ ˜ kφ1 ˙ φ2 = α2 (ha (σ, ξ1 , N (φ, ξ1 )) − φ1 ) − ˜ ˜ mr + mc kmr lr sat(k1 φ1 + k2 φ2 ), + (mr + mc )2 ua = N (φ, ξ1 ), where N (φ, ξ1 ) = (4.27) (4.28) (4.29) sin−1 [sat(k1 φ1 + k2 φ2 )] − ξ1 /εa . We note that we used the saturated control sat(wa ) = sat(k1 φ1 + k2 φ2 ) in order to ensure that its contribution to the control effort does not exceed unity magnitude in the presence of peaking. 96 Thus, the partial state feedback controller for the system (4.4)–(4.7) is given by (4.27)–(4.28) and u=− 4.3.2 βδ(ξ1 ) sat mr + mc ξ2 − N (φ, ξ1 ) µ . (4.30) Output Feedback Controller for the TORA The output feedback controller for the system (4.4)–(4.8) can now be designed using an extended high-gain observer. ˙ ˆ ˆ ˆ ξ1 = ξ2 + (α1 /ε)(y − ξ1 ), ˙ ˆ ˆ ξ2 = σ + an (y)u + (α2 /ε2 )(y − ξ1 ), ˆ ˙ ˆ σ = (α3 /ε3 )(y − ξ1 ), ˆ ˙ ˜ φ1 = φ2 + α1 ha (ˆ , y, N (φ, y)) − φ1 , ˜ ˆ σ kφ1 ˙ ˜ φ2 = α2 ha (ˆ , y, N (φ, y)) − φ1 − ˜ ˆ σ mr + mc ˆ ˆ k m r lr ˆ sat(k1 φ1 + k2 φ2 ), + (mr + mc ) 2 ˆ ˆ ˆ ξ − N (φ, y) β sat 2 , u=− an (y) µ (4.31) (4.32) (4.33) (4.34) (4.35) (4.36) ˆ where δ(ξ1 ) = (I + mr lr )(mr + mc ) − m2 lr cos2 ξ1 , the nominal function an (y) = ˆ 2 ˆ ˆ ˆr 2 ˆ (mr + mc )/δ(y), and ˆ ˆ ˆ σ ˜ ha (ˆ , y, N (φ, y)) = ˆ σ δ(y)ˆ m r lr ˆ mr lr sin y ˆ + (N (φ, y))2 sin y + , mr lr k cos y ˆ k mr + mc ˆ ˆ and mr and mc are nominal values for mr and mc , respectively. ˆ ˆ 97 4.3.3 Output Feedback Design Based on Isidori [21] We now follow the procedure of [21] to design an output feedback stabilizer for (4.4)– (4.8) using a high gain approach, and we shall compare the design from the previous subsection with this one by means of numerical simulations. As per [21], the design is again based on the stabilizer for auxiliary system, and the high gain observer based output feedback stabilizer was found to be ˙ ˆ ˆ ˆ ξ1 = ξ2 + (ˆ 1 /ε1 )(y − ξ1 ), α ˙ ˆ ˆ ξ2 = (ˆ 2 /ε2 )(y − ξ1 ), α 2 ˙ ˜ ˆ φ1 = α1 ha (kI (ξ2 − N (φ, y)), y, N (φ, y)) − φ1 + φ2 , ˜ ˆ ˆ ˜ ˙ φ2 = α2 ha (kI (ξ2 − N (φ, y)), y, N (φ, y)) − φ1 ˜ ˆ kφ1 k m r lr ˆ + − sat(k1 φ1 + k2 φ2 ), mr + mc ( mr + mc ) 2 ˆ ˆ ˆ ˆ ˆ δ(y) ˆ ˆ ˙ u= N − (ξ2 − N (φ, y))kI , mr + mc ˆ ˆ (4.37) (4.38) (4.39) (4.40) (4.41) ˆ ˙ ˙ ˆ ˙ where N = k1 p(φ)φ1 + k2 p(φ)φ2 − ξ2 /εa , the constant kI is a high gain parameter, p(φ) = us (k1 φ1 + k2 φ2 + 1) − us (−k1 φ1 − k2 φ2 − 1) εa 1 − [sat(k1 φ1 + k2 φ2 )]2 , and us (·) is the unit step function. 4.3.4 Numerical Simulations Numerical simulations were performed using the system parameters shown in Table 4.12 and controller parameters given in Table 4.2. Figures 4.2 through 4.13 show the effects of tuning certain design parameters and of perturbations in the masses of the cart and the proof mass. 2 These values are taken from [43] 98 Parameter Symbol Cart mass Rotor mass Dist. from rotor pivot to its CM Rotor moment of inertia Spring constant Value mc mr lr I k 1.3608 kg 0.096 kg 0.0592 m 0.0002175 kg m2 186.3 N/m Table 4.1: TORA system parameters. Module Parameter Value Auxiliary controller α1 ˜ α2 ˜ K εa 2 1 −1, −8 {0.1, 0.01} State feedback controller β µ 85 {1, 0.1} Extended high-gain observer α1 ˆ α2 ˆ α3 ˆ ε 3 3 1 {0.01, 0.001} Plant (all other ICs set to 0) θ(0) xc (0) 0 rad 0.025 m Table 4.2: TORA stabilization: simulation parameters and initial conditions. 99 η1 0.02 0 −0.02 0 5 10 5 t (s) 10 error 0.01 0 −0.01 0 Figure 4.2: Recovery of the Reduced System performance by decreasing εa (η1 curves: solid is the reduced system, dotted is εa = 0.1, dash-dot is εa = 0.01; error curves: solid is εa = 0.1 and dash-dot is εa = 0.01). 4.3.5 Discussion of the Results Figures 4.2 through 4.5 show the results of a step-by-step tuning of the design parameters, starting with matching the performance of the fourth-order reduced system in the first step and ending with the tuning of the extended high-gain observer parameter to match the partial state feedback case. In the process, the full output feedback system is made to match the performance of the reduced system, whose state equation is linear and fourth order—however, since the compensator’s eigenvalues are located quite far into the left half-plane relative to those of the reduced system’s plant, the state equation for the latter is essentially second order and linear. Therefore, this 100 error ξ1 0.5 0 −0.5 0 −3 x 10 1 0 −1 0 5 10 5 t (s) 10 Figure 4.3: Recovery of the Auxiliary System performance by decreasing µ (bottom: solid is µ = 1, dash-dot is µ = 0.1). design methodology enabled us to enforce linear second order-like response characteristics on the controlled TORA system. We explain the tuning procedure to recover the reduced system’s performance below. The errors in each step are computed using the system from the previous step as the new target. 1) The top panel in Figure 4.2 shows the target response trajectory we seek to recover. The solid line is the response of the reduced system (4.11)–(4.12), (4.18), (4.24)– (4.26). The dashed lines are the responses of the auxiliary system (4.11)–(4.16), (4.27)–(4.29), and in these cases, high-gain feedback parameter values of εa = 0.1 and 0.01 were utilized. The bottom panel shows how reducing εa from 0.1 to 0.01 allows for the auxiliary system to better approximate the response of the reduced system—indeed, the settling time was reduced drastically. The εa parameter is 101 y 0.5 0 −0.5 0 2 4 2 4 6 8 10 6 8 10 error 0.2 0 −0.2 0 t (s) Figure 4.4: Recovery of the Partial State Feedback System’s performance by decreasing ε. (Top: solid is state feedback, dashed is output feedback with ε = 0.01, dash-dot is ε = 0.001. Bottom: solid is ε = 0.001, dash-dot is ε = 0.01.) now fixed at 0.01 and we move to the next step. 2) Now, we try and match the performance of the auxiliary system with the partial state feedback system (4.4)–(4.7), (4.27)–(4.28), (4.30). Figure 4.3 shows the results. The reference trajectory used to compute the error in this step is the auxiliary system state ξ1 . The switch was made from η1 to ξ1 because the former is not an output of the full system. A µ value of 0.1 was deemed to provide an acceptably small error, so it was fixed at this value. 3) We now tune the extended high-gain observer parameter ε to recover the perfor102 x2 0.02 0 −0.02 0 2 4 6 8 10 2 4 6 8 10 −3 error 5 x 10 0 −5 0 t (s) Figure 4.5: Recovery of the Partial State Feedback System’s performance by decreasing ε (µ = 0.1). mance of the partial state feedback system from the previous step. Figures 4.4 and 4.5 show the recovery of the performance of the state feedback system by reducing ε from 0.01 to 0.001, in the output feedback system (4.4)–(4.8), (4.31)–(4.36). We note from the bottom panels of these two figures that reducing ε down to 10−3 results in a significant reduction in the error. We therefore choose the final design parameter as ε = 10−3 . The above three-step tuning procedure suggests that we pick the design parameters εa = 0.01, µ = 0.1 and ε = 10−3 so that the performance of the reduced system is recovered by the output feedback system when the latter is initialized at y(0) = 0 (i.e. the trajectory starts out on the sliding manifold, which is also the zero dynamics manifold of the closed loop system when s is viewed as the output). 103 0.5 y 0 −0.5 0 2 4 6 8 10 6 8 10 x2 0.02 0 −0.02 0 2 4 t (s) Figure 4.6: A comparison of the Extended High-Gain Observer (solid) and Isidori (dash-dot) designs. Next, the EHGO-and-SMC design presented in this paper was compared with the design based on [21]. Figure 4.6 shows the best responses that were achieved after tuning the parameters of both designs. Figure 4.7 shows the control effort required under each method, and it can be seen that they are virtually identical when y(0) = 0. Starting with Figure 4.6, the responses of the EHGO + SMC design are shown as solid curves, while those of the high gain feedback design [21] are shown as dash-dotted lines. It should be noted that in the design based on [21], the high gain feedback parameter was tuned to kI = 500, which is comparable to β = 500 in the EHGO + SMC design. Another point to note is that the high gain observer parameter for the design from [21] needed to be set to ε = 10−6 to ensure stability of the output feedback system, whereas in the EHGO-and-SMC case, a value of ε = 10−3 was 104 Control Effort 0.05 0 −0.05 0 2 4 6 t (s) 8 10 Figure 4.7: A comparison of the responses of the control efforts---EHGO (solid) and Isidori (dash-dot). sufficient to guarantee stability and matching of the state feedback performance— this was the case for all subsequent simulations discussed below, as well. Next, y(0) was chosen to be non-zero, and this immediately revealed differences between the transient performance of the high gain feedback approach [21] and that of the SMC design from our paper. Figure 4.8 shows a comparison of the transient performance when y(0) = 0.086. Figure 4.9 shows that the control effort expended by the high gain feedback design is quite large during the transient phase, while the sliding mode control saturates at a magnitude of under 0.3 during this period. Figures 4.10 and 4.11 show the response of the EHGO + SMC design when y(0) = 0.541—this initial condition brings the system to the verge of instability. The high gain feedback 105 y 1 0 −1 0 2 4 6 8 10 8 10 x2 0.02 0 −0.02 0 2 4 6 t (s) Figure 4.8: A comparison of transient performance when y(0) = 0.086. design is unstable when initiated at this rotor angle, and so its response is not shown in these two figures. An important point to note is that due to (4.17), the closed loop systems under both designs will have singularities at 2jπ ± π/2, j ∈ Z, and Figures 4.8, 4.10 show the transient response of y approaching π/2—a saturated control input is preferable in such instances in order to obtain a larger region of attraction. Finally, in order to investigate the robustness of the two designs, the masses of the cart and the rotor were perturbed and some simulations were run. Figure 4.12 shows the result when the cart and eccentric rotor masses were both perturbed by +100%. Figure 4.13 shows the result when the cart and eccentric rotor masses were perturbed −50%. The −50% perturbation brought both designs to the verge of instability. The conclusion of this particular investigation is that both designs are able to tolerate 106 Control 0.5 0 −0.5 −1 0 2 4 6 8 10 0.15 0.2 Control 0 −5 −10 0 0.05 0.1 t (s) Figure 4.9: A comparison of the control effort when y(0) = 0.086. similar amounts of perturbations in the cart and rotor mass parameters. 4.4 Concluding Remarks In this chapter, the stabilization algorithm from Chapter 2 was employed to stabilize the translational oscillator with rotating actuator (TORA) benchmark system. We were able to simplify the original stabilization problem to one of finding a stabilizing controller for an associated auxiliary system—in order to find this controller, a highgain feedback approach was employed, which in turn reduced the problem further to a constrained linear control problem for a simple second order linear system. A linear dynamic controller satisfying the control constraint was obtained by optimizing 107 y 1 0 −1 0 2 4 6 8 10 8 10 x2 0.02 0 −0.02 0 2 4 6 t (s) Figure 4.10: Response of the EHGO + SMC based design when y(0) = 0.541. a quadratic cost function. This immediately led to a set of equations for the nonlinear controller that stabilizes the auxiliary system, and the remainder of the output feedback stabilization problem was solved on the basis of the knowledge of the controller for the auxiliary system, in accordance with the design procedure presented in in Chapter 2. The simulations also show that the performance of the closed-loop reduced system, which is only fourth order, can be asymptotically recovered by appropriate tuning of a single high-gain feedback parameter εa —this is possible so long as the initial conditions place the trajectories of the system on the sliding manifold, thus eliminating the reaching phase. In the full output feedback system, the performance can be fine-tuned further by reducing the width of the boundary layer µ in the sliding mode control, and 108 0.3 Control Effort 0.2 0.1 0 −0.1 −0.2 0 Figure 4.11: 0.541. 2 4 6 t (s) 8 10 Control effort of the EHGO + SMC based design when y(0) = also by reducing the extended high-gain observer parameter ε. Through simulations, it was observed that the stabilizing controller design allows for uncertainties between −50% and +100% for both the cart mass and the rotor mass. An important point to note is that decreasing the ε parameter in the extended high-gain observer too much could potentially pose practical difficulties during implementation due to the effects of measurement noise as discussed in Chapter 2. That being said, the two designs presented in this chapter fully comply with the performance requirements and constraints set forth in [6], to wit: |xc | ≤ 0.025 m, and |u| ≤ 0.1 Nm. In conclusion, the stabilizing controller designed in this chapter for the TORA 109 0.5 y 0 −0.5 0 2 4 6 8 10 6 8 10 x2 0.02 0 −0.02 0 2 4 t (s) Figure 4.12: Response under +100% perturbation in mc and mr . system provides good transient performance and exhibits robustness to perturbations of the cart and rotor masses. The design procedure is fairly simple and systematic in that it provides a framework for converting the problem to one of stabilizing a relative degree one system, followed by order reduction, provided a stabilizing controller can be found for the auxiliary system of order one less than that of the original problem—this proved to be true for the TORA system, thus demonstrating that this technique may be profitably employed to solve other application problems in the future. Furthermore, this design approach was ultimately able to recover the performance characteristics of a full-order observer-based second order linear system, so long as the initial conditions are located on the sliding manifold of the full system. 110 0.5 y 0 −0.5 0 5 10 5 t (s) 10 x2 0.02 0 −0.02 0 Figure 4.13: Response under −50% perturbation in mc and mr . 111 Chapter 5 Conclusions and Future Work 5.1 Concluding Remarks This dissertation is concerned with the problems of robust stabilization and regulation, by means of output measurements alone, of non-minimum phase systems that are transformable into the normal form. This work is an extension of the design tool of Isidori [21], and the chapter on regulation is similarly an extension of the technique presented in the paper by Marconi et al.[32]. Nevertheless, this work represents a significant contribution to the control literature because it presents a design methodology that is more versatile than [21, 32, 10], in that any robust control technique can be applied in the design of the control law, unlike the papers by Isidori and coworkers, in which high-gain feedback control was an intrinsic component of the design. This extension to other robust control schemes is possible using our approach due to the fact that an extended high-gain observer is used to estimate an unknown function that renders the zero dynamics of the plant observable, and thus allows us to stabilize the system despite its being non-minimum phase. The contributions of this work are discussed in further detail in Section 5.1.1. In order to derive the stability results presented in this work, continuously112 implemented sliding mode control was chosen—this was also a natural choice because sliding mode control is known to have good robustness properties, and since the control effort saturates outside the boundary layer, this design precludes the possibility of large spikes during the transient phase of the system response, thus providing protection against the peaking phenomenon, which is a common feature of output feedback systems that employ high-gain observers. 5.1.1 Main Contributions The primary contribution of this work is the following three-step design procedure for stabilizing the origin of a non-minimum phase system in the normal form. 1) Augment a dynamic compensator {L, M, N } to the system and moreover, design this triple such that the resultant system is minimum phase and relative degree one with respect to the virtual output s = eρ −N (·). This task is equivalent to the stabilization of the auxiliary system of Isidori [21], since this closed-loop auxiliary system coincides with the zero dynamics of the aforementioned augmented system with s viewed as the output. 2) Now use any of several standard robust control techniques such as sliding mode control, saturated high-gain feedback, Lyapunov redesign, etc. to design a static state feedback control law that stabilizes the relative degree one minimum phase system obtained from the previous step. 3) Finally, design an extended high-gain observer to estimate the system output and its first ρ − 1 derivatives, and additionally, the signal b that is required to solve the auxiliary design problem in step 1). Replace all the states and the signal b in the state feedback control design with their estimates, and if necessary, saturate the resulting control law outside a compact set of interest, or saturate the estimates obtained from the EHGO. 113 It is important to note that we were able to separate steps 1) and 2) in the design process in anticipation of the fact that the extended high-gain observer would be used in the final step to estimate b along with y and its first ρ − 1 derivatives. The approach pursued by [21, 32, 10] is different; the recovery of the signal b is intrinsic to the design of the control scheme employed in these papers, meaning that steps 1) and 2) go hand-in-hand if this technique is utilized, and hence it becomes necessary to implement their particular high-gain feedback based control law. Another contribution of this work is the fact that our design allows for uncertainty in the control coefficient a(·), which was not addressed by Isidori and coworkers. This constitutes the main technical difference between the results presented in this work and those of [21]. Handling this uncertainty dominates our analysis in Sections 2.3 and 3.2.2. In both Chapters 2 and 3, we prove exponential stability of the closed-loop system, under the assumption of exponential stability of the closed-loop auxiliary system. The paper [21] proves only practical stabilization, but it requires the less stringent condition of asymptotic stability of the closed-loop auxiliary system. Simulation examples afforded opportunities to compare our design with that of [21]. The main difference was that in all the examples, the high-gain feedback design of Isidori exhibits a sharp transient response and much larger control effort during this phase, when compared to our designs. Moreover, in the example involving a tangent nonlinearity in Chapter 2, the size of the estimate of the region of attraction was much larger in our case than in the case of the high-gain feedback design of [21]. The stabilization design of Chapter 2 was applied to the nonlinear TORA benchmark system in Chapter 4, and this served to demonstrate that our design procedure can be successfully applied to practical problems and non-trivial physical systems. Once again, comparisons were made between our design method and that of Isidori [21], and the results were consistent with all the other observations made by simulating the aformentioned academic examples. In particular, our design was able to stay 114 within the performance limitations of |u| ≤ 0.1 and |xc | ≤ 0.025 as specified in [6] over a much wider range of initial conditions than the high-gain feedback design of Isidori [21]. 5.2 Future Directions The stabilization and regulation designs that we presented for non-minimum phase systems could be extended in several directions, and could potentially also be unified with other control schemes that employ extended high-gain observers. The following are some ideas for immediate extensions of this work. • Apply the conditional integrator in conjunction with the extended high-gain observer to solve the problem of causing a non-minimum phase system to track a constant reference and reject constant disturbances. • Extend the above problem to output regulation of non-minimum phase systems using a conditional servocompensator. This could potentially allow for the design of the stabilizing compensator for the plant alone, without sonsidering the dynamics of the servocompensator, thus paving the way for reusing stabilization designs in regulation problems. Moreover, the conditional servocompensator may be able to asymptotically recover the performance of “ideal” sliding mode control. • Extend our techniques to incorporate control schemes such as saturated highgain feedback, Lyapunov redesign, etc. • Combine the extensions of the results presented in this work with that of Freidovich and Khalil [16] to obtain a general framework that provides for the application of extended high-gain observers in solving tracking and disturbance rejection problems for systems that are possibly non-minimum phase. 115 APPENDIX 116 Proofs and Technical Details A.1 Locally Lipschitz property of ∆0/ε In this section, we illustrate the locally Lipschitz property of the quantity ∆0 /ε of Chapter 2 by first looking at the sub-expression 1 ˆ [gε (l(ξ, φ)/K) − gε (l(ξ, φ)/K)]. ε (A.1) To simplify the analysis, we define G(ξ, φ) gε (l(ξ, φ)/K). Hence, by the Fundamental Theorem of Calculus, we have 1 1 ˆ ˆ ˆ − G(ξ + λ(ξ − ξ), φ) [G(ξ, φ) − G(ξ, φ)] = G(ξ + λ(ξ − ξ), φ) ε ε λ=0 λ=1 1 ∂ 1 ˆ ˆ = G(ξ + λ(ξ − ξ), φ) dλ(ξ − ξ) ε 0 ∂ξ 1 ∂ ˆ =− G(ξ + λ(ξ − ξ), φ) dλ 0 ∂ξ ⊤ . (A.2) × ερ−1 ν1 , . . . , ενρ−1 , νρ We note that both gε (·) and l(·) are differentiable and have locally Lipschitz derivatives with respect to their arguments, and hence (A.2) is locally Lipschitz uniformly 117 in ε. We can use similar arguments for the remaining terms of ∆0 /ε to show that they are also locally Lipschitz uniformly in ε. A.2 A.2.1 Solution of the PDE (2.14) Properties of the solution In this subsection, we establish the existence of the transformation q = T (xa , φ, s, θ) as a solution of (2.14), and determine the conditions under which it is a diffeomorphism. Since a(·) = 0, we can write (2.14) as ∂T ∂N ∂T − ∂φ ∂s ∂φ Let ϕ = φ⊤ s M ∆a ∂T + = 0. a ∂s (A.3) ⊤ and   ∆a M a   g (ϕ, θ) =  ¯ . 1 − ∂N M ∆a a ∂φ We note that for sufficiently small |∆a /a|, g has full rank, and so we shall make this ¯ assumption henceforth. The PDE (A.3) can now be written as ∂T g (ϕ, θ) = 0. ¯ ∂ϕ Since g is of full rank, the Frobenius theorem [19] guarantees the existence of a solution ¯ to the above PDE. Now, let ˜ ϕ = T (xa , ϕ, s, θ) = ˜ q⊤ s ⊤ = 118 (T (xa , φ, s, θ))⊤ s ⊤ . ˜ We seek conditions under which T is a diffeomorphism. We note that its Jacobian matrix is  ˜  ∂T ∂T =  ∂φ ∂ϕ 0  ∂T ∂s  ,  1 ˜ so that ∂ T is nonsingular if and only if ∂T is nonsingular. The latter condition will ∂ϕ ∂φ be utilized when finding the solution to (A.3) in the following subsection. ˜ Finally, we conclude by the inverse function theorem [19] that if ∂ T is nonsingular ∂ϕ for some ϕ in the domain of interest, then there exists an open neighborhood of ϕ in ˜ this domain such that the restriction of T to this neighborhood is a diffeomorphism. A.2.2 Details of the solution We begin by noting that when ∆a = 0, a particular solution of (A.3) can be taken to be T = φ. Now, for ∆a = 0, we need to employ the method of characteristics for (quasi-) linear differential equations, which reduces (A.3) to a system of ODE along so-called characteristic curves of the PDE. To this end, we rewrite (A.3) as a system of scalar differential equations as follows. ∂Ti ∂Ti ∂N − ∂φ ∂s ∂φ M ∆a ∂Ti + = 0, a ∂s 1 ≤ i ≤ r, (A.4) where Ti is the ith component of T . A solution Ti = qi (xa , φ, s, θ) defines an integral surface in the φ, s, Ti -space. The direction of the normal to the surface is ⊤ ∂Ti ∂Ti , so that (A.4) can be interpreted as the condition that each , , −1 ∂φ ∂s point on the integral surface has the property that the vector ⊤ ∆a ∂N ∆a ⊤ M , 1− M, 0 a a ∂φ 119 (A.5) at that point is tangent to the surface. Thus (A.4) defines a direction field given by (A.5), called the characteristic directions, having the property that a surface Ti = qi (xa , φ, s, θ) is an integral surface if and only if, at each of its points, the tangent hyperplane contains the characteristic direction. The integral curves of this field, i.e. the family of curves whose tangent lines coincide with the characteristic directions, are called characteristic curves. Now, for a PDE of the form m ai i=1 ∂u = a, ∂xi the characteristic directions tangent to the integral surface are given by du : dxm : · · · : dx2 : dx1 = a : am : · · · : a2 : a1 , and hence, we can now reduce the above PDE to a system of m ODE du = a, dp dxi = ai , dp 1 ≤ i ≤ m, that describes a family of m characteristic curves. In our problem, the characteristic ODE are given by dTi = 0, dp ds ∆a ∂N =1− M, dp a ∂φ dφ ∆a =M , dp a 1 ≤ i ≤ r. (A.6) The integral surface can now be constructed by solving the following initial value problem: let an r-dimensional manifold C be given by φ = ϑ, s = 0, Ti = Ti0 (ϑ), 120 1 ≤ i ≤ r, ⊤ where ϑ = ϑ1 ϑ2 · · · ϑr is a vector of parameters, and suppose that the rank of the matrix ⊤ ∂s ⊤ ∂ϑ ∂φ ⊤ ∂ϑ is r. We assume that the projection C0 of this manifold on the (φ, s)-space is free of double points, that is, different points on C0 correspond to different sets of ϑ. We also note that the boundary conditions Ti = Ti0 on the manifold C must be chosen ˜ such that the condition ∂ T = 0 obtained in the previous subsection is satisfied. ∂φ We obtain the solutions φ(ϑ, p), s(ϑ, p) and Ti (xa , ϑ, p, θ) of the system of characteristic ODE (A.6), which, at p = 0, coincide with the prescribed functions of ϑ describing the initial manifold C. We now consider the quantities ϑ, p to be expressed in terms of φ and s, respectively, and substitute them into Ti (xa , ϑ, p, θ), so that the Ti appear as functions of φ and s. This introduction of φ and s as new independent variables is possible if the following inequality holds. ∆ ∂φ ∂p ∂φ ∂ϑ ∂s ∂p = 0. ∂s ∂ϑ p=0 (A.7) ∂φ We note that when ∆a = 0, we have ∂s = 1, = I and ∂s = 0, so the above ∂p ∂ϑ ∂ϑ inequality (A.7) holds, because ∆ = −1 in this case. Consequently, (A.7) also holds for small enough |∆a /a|. Let us now define ∆a ds, a − ∂N M ∆a ∂φ Ψ (A.8) and ˜ ∆ ∂N M Ψ. ∂φ 121 (A.9) The solutions of (A.6) are therefore given by Ti = Ti0 (ϑ). φ = ϑ + M Ψ, ˜ ⇒ p = s + ∆, ˜ s = p − ∆, ϑ = φ − M Ψ, ˜ Ti (xa , φ − M Ψ, s + ∆, θ) = Ti0 (φ − M Ψ). We now consider the case when ∆a = 0. We then have a solution Ti = Ti0 (φ); moreover, we can choose T 0 = φ, which yields the solution T = φ (where T and T 0 are vectors comprising the elements Ti and Ti0 , respectively). We note that with the same choice of initial condition, the solution for the ∆a = 0 case is T = φ − M Ψ. Due to the well known result that an ODE exhibits continuous dependence on parameters [29], we conclude that the solutions of the second and third ODEs in (A.6) will be close to s and φ, respectively, provided the perturbation |∆a /a| is small. Therefore, the resulting solution T = φ − M Ψ of the original PDE (A.3) will also be close to φ. A.3 Proof of Proposition 2.1 We pick up the analysis of the closed-loop state feedback system of Chapter 2 where we left off at equation (2.26). We first show that the reduced system and the boundary layer system have exponentially stable equilibrium points at xs = 0 and s = 0, respectively, and then we use a composite Lyapunov function to show that the origin of the whole system is also an exponentially stable equilibrium point. Item 3) of Assumption 2.4 states that when s = 0, the origin of (2.20) is locally exponentially stable. Hence, the origin of the reduced system in (2.26) is locally exponentially stable by hypothesis, and by a converse Lyapunov theorem [29, Theorem 4.14], there exists a Lyapunov function V1 (xs , θ) that satisfies the following 122 inequalities c1 xs 2 ≤ V1 (xs , θ) ≤ c2 xs 2 , ∂V1 F (xs , 0, θ) ≤ −c3 xs 2 , ∂xs ∂V1 ≤ c 4 xs , ∂xs (A.10a) (A.10b) (A.10c) in some neighborhood Nxs of xs = 0. The boundary layer system is given by g ds a = − b βs = − dτ an an 1− ∂N ∆a M ∂φ a s. (A.11) Let V2 (s) = s2 /2 be a Lyapunov function candidate for the boundary layer system (A.11). We then obtain ∂V2 ds a · =− ∂s dτ an 1− ∂N ∆a M ∂φ a s2 ≤ −β0 k0 k † s2 . Hence, the origin of the boundary layer system is exponentially stable uniformly in xs . By the smoothness of F and fb and the fact that each function vanishes at the origin, we have F (xs , s, θ) − F (xs , 0, θ) ≤ k1 |s| , (A.12a) fb (xs , s, θ) ≤ k2 xs + k3 |s| (A.12b) in some neighborhood N of (xs , s) = (0, 0). Suppose µ is chosen small enough such that Ωµ ⊂ Nxs and Ωµ ⊂ N . We now consider the composite Lyapunov function candidate W (xs , s, θ) = V1 (xs , θ) + s2 /2. 123 From (A.10) and (A.12), it can be shown that [29] ˙ W ≤ −c3 xs 2 + c4 k1 xs |s| + k2 xs |s| + k3 s2 − β0 k0 k † s2 /µ    ⊤  −c4 k1 −k2 c3   xs   xs   2 (A.13) = −  .   † −c4 k1 −k2 β0 k0 k |s| |s| − k3 µ 2 There exists a µ∗ > 0 such that for all 0 < µ < µ∗ , the right-hand side of (A.13) is negative definite in Ωµ , which then yields ˙ W ≤ −2γW for some γ > 0. Therefore, W (xs (t), s(t), θ) ≤ W (xs (t0 ), s(t0 ), θ)e−2γ(t−t0 ) and, from the properties of V1 and V2 , we obtain xs (t) s(t) ≤ xs (t0 ) K ∗ e−γ(t−t0 ) , s(t0 ) for some K ∗ > 0. This completes the proof of Proposition 2.1. A.4 Proof of Proposition 3.1 It was already shown prior to stating the proposition that the trajectories of the closed-loop state feedback system (3.19), (3.20) reach the set {|s| ≤ µ} in finite time and remain inside thereafter, and so we continue the analysis by considering the set Ω defined in (3.24) below the statement of the proposition. 124 Assumption 3.4 calls for the existence of a Lyapunov function V (xs , w, θ) for the internal dynamics. The derivative of V is given by ∂V ∂V ˙ S w. F (xs , s, w, θ) + V = ∂xs ∂w 0 Thus, on the boundary V = c0 , we have ∂V ∂V −1 ˙ F (xs , s, w, θ) + S w ≤ −α3 ( xs ) ≤ −α3 (α2 (c0 )). V = ∂xs ∂w 0 Hence, Ω is a positively invariant set for the system (3.19). As a consequence of the preceding analysis, all trajectories starting in Ω, stay in Ω and reach the set Ω ∩ {|s| ≤ µ} in finite time. Inside this set, we have ˙ V ≤ −α3 ( xs ), ∀ xs ≥ γ(µ). The above analysis shows that the trajectories enter the positively invariant set Ωµ = {V (xs , w, θ) ≤ α2 (γ(µ))} × {|s| ≤ µ} in finite time. Now, inside the set Ωµ , the closed-loop system is given by the following equations in the singularly perturbed form. xs = F (xs , s, w, θ), ˙ µs = µfs (xs , s + N, w, θ) − ˙ (A.14a) β(xs , s) gs (xs , s + N, w, θ)s. an (xs , s) (A.14b) Item 2) of Assumption 3.4 states that when s = 0, the origin of the xs -equation in ˙ (A.14) is locally exponentially stable. Hence, the origin of the reduced system in (A.14) is locally exponentially stable by hypothesis, and by a converse Lyapunov the125 orem [29, Theorem 4.14], there exists a Lyapunov function V1 (xs , w, θ) that satisfies the following inequalities c1 xs 2 ≤ V1 (xs , w, θ) ≤ c2 xs 2 , ∂V ∂V1 F (xs , 0, w, θ) + 1 S0 w ≤ −c3 xs 2 , ∂xs ∂w ∂V1 ≤ c 4 xs , ∂xs (A.15a) (A.15b) (A.15c) in some neighborhood Nxs of xs = 0. The boundary layer system is given by ds a gs = − βs = − dτ an an 1− ∂N ∆a M ∂φ a s. (A.16) Let V2 (s) = s2 /2 be a Lyapunov function candidate for the boundary layer system (A.16). We then obtain ∂V2 ds a · =− ∂s dτ an 1− ∂N ∆a M ∂φ a s2 ≤ −β0 k0 k † s2 , due to (3.22), (3.12) and Assumption 3.3. Hence, the origin of the boundary layer system is exponentially stable uniformly in xs . By the smoothness of F and fs and the fact that each function vanishes at the origin, we have F (xs , s, w, θ) − F (xs , 0, w, θ) ≤ k1 |s| , (A.17a) fs (xs , s + N, w, θ) ≤ k2 xs + k3 |s| (A.17b) in some neighborhood N of (xs , s) = (0, 0). Suppose µ is chosen small enough such that Ωµ ⊂ Nxs and Ωµ ⊂ N . We now consider the composite Lyapunov function candidate W (xs , s, w, θ) = V1 (xs , w, θ) + s2 /2. 126 From (A.15) and (A.17), it can be shown that [29] ˙ W ≤ −c3 xs 2 + c4 k1 xs |s| + k2 xs |s| + k3 s2 − β0 k0 k † s2 /µ    ⊤  −c4 k1 −k2 c3   xs   xs   2 (A.18) = −  .   † −c4 k1 −k2 β0 k0 k |s| |s| − k3 µ 2 There exists a µ∗ > 0 such that for all 0 < µ < µ∗ , the right-hand side of (A.18) is negative definite in Ωµ , which then yields ˙ W ≤ −2γr W for some γr > 0. Therefore, W (xs (t), s(t), w(t), θ) ≤ W (xs (t0 ), s(t0 ), w(t0 ), θ)e−2γr (t−t0 ) and, from the properties of V1 and V2 , we obtain xs (t) s(t) ≤ xs (t0 ) s(t0 ) Kr e−γr (t−t0 ) , for some Kr > 0. This completes the proof of Proposition 3.1. 127 BIBLIOGRAPHY 128 BIBLIOGRAPHY [1] Jeffrey H. Ahrens and Hassan K. Khalil. High-gain observers in the presence of measurement noise: A switched-gain approach. Automatica, 45(4):936–943, Apr 2009. [2] A. N. Atassi and H. K. Khalil. A separation principle for the stabilization of a class of nonlinear systems. IEEE Transactions on Automatic Control, 44(9):1672–1687, Sep 1999. [3] A. N. Atassi and H. K. Khalil. A separation principle for the control of a class of nonlinear systems. IEEE Transactions on Automatic Control, 46(5):742–746, May 2001. [4] M. Athans and P. L. Falb. Optimal Control: An Introduction to the Theory and Its Applications. McGraw-Hill, Inc, New York, NY, 1966. [5] A. A. Ball and H. K. Khalil. High-gain observers in the presence of measurement noise: A nonlinear gain approach. In 47th IEEE Conference on Decision and Control, 2008, pages 2288–2293, Dec 2008. [6] R. T. Bupp, D. S. Bernstein, and V. T. Coppola. A benchmark problem for nonlinear control design. International Journal of Robust and Nonlinear Control, 8(4-5):307–310, Apr 15 1998. [7] R. T. Bupp, D. S. Bernstein, and V. T. Coppola. Experimental implementation of integrator backstepping and passive nonlinear controllers on the RTAC testbed. International Journal of Robust and Nonlinear Control, 8(4-5):435–457, Apr 15 1998. [8] C. I. Byrnes, F. D. Priscoli, and A. Isidori. Output Regulation of Uncertain Nonlinear Systems. Birkh¨user: Boston, Cambridge, MA, 1997. a [9] C. I. Byrnes, F. D. Priscoli, A. Isidori, and W. Kang. Structurally stable output regulation of nonlinear systems. Automatica, 33(3):369–385, Mar 1997. 129 [10] F. Celani, A. Isidori, and L. Marconi. A reduction paradigm for output regulation. International Journal of Robust and Nonlinear Control, 18(7):756–781, May 10 2008. [11] R. Courant and D. Hilbert. Methods of Mathematical Physics, Volume II. John Wiley & Sons, Hoboken, NJ, 1962. [12] E. Davison. The robust control of a servomechanism problem for linear timeinvariant multivariable systems. Automatic Control, IEEE Transactions on, 21(1):25–34, Feb 1976. [13] G. Escobar, R. Ortega, and H. Sira-Ramirez. Output-feedback global stabilization of a nonlinear benchmark system using a saturated passivity-based controller. IEEE Transactions on Control Systems Technology, 7(2):289–293, Mar 1999. [14] F. Esfandiari and H. K. Khalil. Output-feedback stabilization of fully linearizable systems. International Journal of Control, 56(5):1007–1037, Nov 1992. [15] B. A. Francis and W. M. Wonham. The internal model principle for linear multivariable regulators. Applied Mathematics & Optimization, 2:170–194, 1975. 10.1007/BF01447855. [16] L. B. Freidovich and H. K. Khalil. Performance recovery of feedback linearization based designs. IEEE Transactions on Automatic Control, 53(10):2324–2334, Nov 2008. [17] J. Freudenberg and D. Looze. Right half plane poles and zeros and design tradeoffs in feedback systems. Automatic Control, IEEE Transactions on, 30(6):555– 565, Jun 1985. [18] J. Huang. Nonlinear Output Regulation: Theory and Applications. Philadelphia, PA, 2004. SIAM, [19] A. Isidori. Nonlinear Control Systems. Springer-Verlag, D¨sseldorf, Germany, u 1995. [20] A. Isidori. Nonlinear Control Systems II. Springer-Verlag London Ltd, London, UK, 1999. [21] A. Isidori. A tool for semiglobal stabilization of uncertain non-minimum-phase nonlinear systems via output feedback. IEEE Transactions on Automatic Control, 45(10):1817–1827, Oct 2000. [22] M. Jankovic, D. Fontaine, and P. V. Kokotovi´. TORA example: Cascadec and passivity-based control designs. IEEE Transactions on Control Systems Technology, 4(3):292–297, May 1996. 130 [23] Z. P. Jiang and I. Kanellakopoulos. Global output-feedback tracking for a benchmark nonlinear system. IEEE Transactions on Automatic Control, 45(5):1023– 1027, May 2000. [24] R. E. Kalman. Contributions to the Theory of Optimal Control. Boletin de la Sociedad Matematica Mexicana, 5:102–119, 1981. [25] D. Karagiannis, Z. P. Jiang, R. Ortega, and A. Astolfi. Output-feedback stabilization of a class of uncertain non-minimum-phase nonlinear systems. Automatica, 41(9):1609–1615, Sep 2005. [26] H. K. Khalil. Robust servomechanism output-feedback controllers for feedback linearizable systems. Automatica, 30(10):1587–1599, Oct 1994. [27] H. K. Khalil. On the design of robust servomechanisms for minimum phase nonlinear systems. International Journal of Robust and Nonlinear Control, 10(5):339–361, Apr 30 2000. [28] H. K. Khalil. Universal integral controllers for minimum-phase nonlinear systems. IEEE Transactions on Automatic Control, 45(3):490–494, Mar 2000. [29] H. K. Khalil. Nonlinear Systems. Prentice Hall, Upper Saddle River, NJ, third edition, 2002. [30] H. K. Khalil and E. G. Strangas. Robust speed control of induction motors using position and current measurements. IEEE Transactions on Automatic Control, 41(8):1216–1220, Aug 1996. [31] P. Kokotovi´, H. K. Khalil, and J. O’Reilly. Singular Perturbation Methods in c Control: Analysis and Design. SIAM, Philadelphia, PA, 1999. [32] L. Marconi, A. Isidori, and A. Serrani. Non-resonance conditions for uniform observability in the problem of nonlinear output regulation. Systems & Control Letters, 53(3-4):281–298, Nov 2004. [33] Lorenzo Marconi and Laurent Praly. Uniform practical nonlinear output regulation. IEEE Transactions on Automatic Control, 53(5):1184–1202, Jun 2008. [34] R. Marino and P. Tomei. A class of globally output feedback stabilizable nonlinear nonminimum phase systems. IEEE Transactions on Automatic Control, 50(12):2097–2101, Dec 2005. [35] S. Nazrulla and H. K. Khalil. Output regulation of non-minimum phase nonlinear systems using an extended high-gain observer. In Proceedings of the IEEE International Conference on Control and Automation, pages 732–737, Dec 2009. [36] S. Nazrulla and H. K. Khalil. Robust stabilization of non-minimum phase nonlinear systems using extended high-gain observers. IEEE Transactions on Automatic Control, Preprint, 2010. 131 [37] V. O. Nikiforov. Adaptive nonlinear tracking with complete compensation of unknown disturbances. European Journal of Control, 4(2):132–139, 1998. [38] S. Sastry. Nonlinear Systems: Analysis, Stability and Control. Springer, New York, NY, 1999. [39] A. Serrani, A. Isidori, and L. Marconi. Semiglobal robust output regulation of minimum-phase nonlinear systems. International Journal of Robust and Nonlinear Control, 10(5):379–396, Apr 30 2000. [40] S. Seshagiri and H. K. Khalil. Robust output regulation of minimum phase nonlinear systems using conditional servocompensators. International Journal of Robust and Nonlinear Control, 15(2):83–102, Jan 25 2005. [41] P. Tsiotras, M. Corless, and M. A. Rotea. An L2 disturbance attenuation solution to the nonlinear benchmark problem. International Journal of Robust and Nonlinear Control, 8(4-5):311–330, Apr 15 1998. [42] M. Vidyasagar. Nonlinear Systems Analysis. Prentice-Hall, Englewood Cliffs, NJ, second edition, 1993. [43] Chih-Jian Wan, D. S. Bernstein, and V. T. Coppola. Global stabilization of the oscillating eccentric rotor. In Proceedings of the IEEE Conference on Decision and Control, volume 4, pages 4024–4029, Dec 1994. [44] G. Zames. Feedback and optimal sensitivity: Model reference transformations, multiplicative seminorms, and approximate inverses. IEEE Transactions on Automatic Control, 26(2):301–320, 1981. [45] Z. Zhang and J. S. Freudenberg. Loop transfer recovery for nonminimum phase plants. Automatic Control, IEEE Transactions on, 35(5):547–553, May 1990. 132