This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

State Compensation Linearization and Control

Quan Quan, Jinrui Ren Q. Quan is with the School of Automation Science and Electrical Engineering, Beihang University, Beijing 100191, China. (e-mail: qq_buaa@buaa.edu.cn.J. Ren is with the School of Automation, Xi’an University of Posts & Telecommunications, Shaanxi, Xi’an 710121, China. (e-mail: renjinrui@xupt.edu.cn).
Abstract

The linearization method builds a bridge from mature methods for linear systems to nonlinear systems and has been widely used in various areas. There are currently two main linearization methods: Jacobian linearization and feedback linearization. However, the Jacobian linearization method has approximate and local properties, and the feedback linearization method has a singularity problem and loses the physical meaning of the obtained states. Thus, as a kind of complementation, a new linearization method named state compensation linearization is proposed in the paper. Their differences, advantages, and disadvantages are discussed in detail. Based on the state compensation linearization, a state-compensation-linearization-based control framework is proposed for a class of nonlinear systems. Under the new framework, the original problem can be simplified. The framework also allows different control methods, especially those only applicable to linear systems, to be incorporated. Three illustrative examples are also given to show the process and effectiveness of the proposed linearization method and control framework.

Index Terms:
State compensation linearization; additive state decomposition; stabilization; tracking.

I Introduction

Linearization, aiming to extract a linear relation from a nonlinear dynamical system, is a common process in control designs. This is on one hand caused by the fact that most of the dynamical systems are nonlinear in nature, and is incentivized by the rich results of linear control theories and methodologies. On the other hand, humans are more used to dealing with linear systems especially when deploying and tuning controllers on real systems. The internal model hypothesis suggests that the central nervous system constructs models of the body and the physical world, and that humans use approximate feedforward plant inversion to interact with linear systems [1], whereas experiments in [2] showed that such a linear plant inversion in humans’ brains may be less precise when directly dealing with nonlinear systems. Therefore, linearization is important in both systematic control design [3, 4, 5] and facilitating human-machine interaction. In the control community, linear systems are regarded as a class of standard and fundamental systems, where theories have been nearly fully developed. Although more advanced nonlinear control methods have been developed rapidly in recent years, compared with linear methods, they are more difficult to apply in practice. Linearization thus allows mature results of linear systems, for example, frequency-domain methods [6], to be applied for nonlinear systems. In engineering practice, linearization can be considered as an intermediate step, and the subsequent analysis and design in the linear domain will become much easier to implement and understand [7, 8].

There are two main linearization methods: Jacobian linearization (i.e., linearization by Taylor’s expansion) and feedback linearization [3],[4],[9]. Jacobian linearization is the most popular one because it is simple, intuitive, and easy-to-use. However, the Jacobian linearization method is an approximate linearization method, which neglects higher-order nonlinearity terms in Taylor’s expansion directly. Moreover, Taylor’s expansion holds locally, and thus the Jacobian linearization has a local property. Different from the Jacobian linearization, feedback linearization is an accurate linearization method, which transforms a nonlinear system into an equivalent linear system of the Brunowski form, namely multiple chains of integrators. Nevertheless, feedback linearization is somewhat difficult to apply directly in some situations [4]. Full state measurement is also required for implementing feedback linearization, and robustness may not be guaranteed when there exist uncertain parameters or unmodelled dynamics [3]. Moreover, feedback linearization may lead to a singularity problem, i.e., a vanishing denominator, which makes the system uncontrollable [4, pp.213-216]. In order to avoid the singularity condition, the working state of the system must be restricted [10]. In order to improve the robustness property of classical feedback linearization, robust feedback linearization [11, 12, 13] has been developed. It brings the original nonlinear system into the Jacobian approximation around the origin rather than in the Brunowski form. In this case, the original nonlinear system is only partially transformed, which allows to preserve the good robustness property obtained by a linear control law which it is associated with. However, other disadvantages of feedback linearization still remain in robust feedback linearization. There are also some other linearization methods, for example, the feedback linearization using Gaussian processes to predict the unknown functions [14].

To complement the existing linearization methods, a new linearization method named state compensation linearization, which is based on the additive state decomposition, is proposed in this paper. The idea of compensation has already existed in active disturbance rejection control (ADRC) [15],[16], which can also be used to accomplish linearization, although it mainly aims to compensate for disturbances. The nonlinear part of a dynamical system has been forced to be a part of lumped disturbances, which can be compensated via a feedforward strategy based on the estimates from an extended state observer (ESO) [17],[18]. The final linearized system is a cascade integral plant. The significant feature of ADRC is that it requires minimum information about a dynamic system except for the relative degree of the system. However, for most physical systems, the nonlinear dynamics may be known or at least partially known. The control effect can be improved if the known nonlinear dynamics could be exploited in design. Besides, how to select the relative degree for non-minimum phase (NMP) systems is still an open problem in ADRC [16].

Additive state decomposition [19],[20] is a kind of generalization of the superposition principle in linear systems to nonlinear systems. As a task-oriented problem decomposition, additive state decomposition can reduce the complexity of a task, which is the addition of several subtasks. The essential idea of additive state decomposition can be expressed by the following equation

A=B+(AB)𝐶{A=B+}\underset{C}{\underbrace{\left(A-B\right)}} (1)

where AA is the original system/task/problem, BB is the primary system/task/problem, and C=ABC=A-B represents the secondary system/task/problem. It is obvious that the equality holds. In order to accomplish linearization, the linear system (primary system B{B}) can be obtained by compensating for the nonlinearity of the original system using the secondary system CC. This can be described as

B=AC.{B=A-}C. (2)

The secondary system CC can be designed as a system without any uncertainties. On the other hand, because the sum of the primary system and the secondary system is the original system, all the uncertainties, such as the higher-order unmodeled dynamics of the real system, are left in the primary system (the linear system). Based on the proposed state compensation linearization method, a state-compensation-linearization-based control framework is then established. The basic idea is to decompose an original system into several subsystems taking charge of simpler subtasks. Then one can design controllers for these subtasks and finally combine them to achieve the original control task. Such control methods were also proposed in our previous work in [19],[21]. However, the method was only applied to the control problem for some special systems. In this paper, the state-compensation-linearization-based control framework is proposed for more general systems.

The main contributions of this paper are twofold.

(i) A new linearization method called state compensation linearization is proposed. It is a linearization method complementary to the Jacobian linearization method and the feedback linearization method. Unlike the existing two linearization methods, the state compensation linearization method uses a compensation strategy. The linear system (primary system) can be obtained by compensating for the nonlinearity of the original system using the secondary system and its controller. This makes state compensation linearization flexible and combines the advantages of Jacobian linearization and feedback linearization.

(ii) A state-compensation-linearization-based control framework is established. Because the primary system is linear, the classical frequency-domain and time-domain methods for linear systems can be applied. In this new framework, the nonlinearity of the original system is compensated by the nonlinear secondary system and its stabilizing controller. Then, the original stabilization/tracking problem for the nonlinear system is simplified to a stabilization/tracking problem for the linear primary system, and different control methods are allowed to be applied to the system.

The remaining parts of this paper are organized as follows. In Section II, necessary preliminaries are introduced. Next, Section III details a new linearization method called state compensation linearization, which is followed by state compensation linearization applications to control design in Section IV. Three illustrative examples are provided in Section V to show the effectiveness of the proposed method. Section VI concludes the paper.

II Preliminaries: Additive state decomposition

Additive state decomposition has already appeared in some other domains. The additive state decomposition mentioned here is limited to the meaning in the author’s paper [19]. Consider the following class of differential dynamic systems

A:{𝐱˙=𝐟(𝐱,𝐮)+𝐝,𝐱(0)=𝐱0𝐲=𝐡(𝐱)A:\left\{\begin{array}[]{l}\mathbf{\dot{x}}=\mathbf{f}\left({\mathbf{x,}}\mathbf{u}\right)+\mathbf{d},\mathbf{x}\left(0\right)=\mathbf{x}_{0}\\ \mathbf{y}=\mathbf{h}\left({\mathbf{x}}\right)\end{array}\right. (3)

where 𝐱n\mathbf{x}\in\mathbb{R}^{n} is the state, 𝐮m\mathbf{u}\in\mathbb{R}^{m} is the input, 𝐲p\mathbf{y}\in\mathbb{R}^{p} is the output, 𝐝n\mathbf{d}\in\mathbb{R}^{n} denotes the disturbance, functions 𝐟:n×mn\mathbf{f:\mathbb{R}}^{n}\mathbf{\times\mathbb{R}}^{m}\rightarrow\mathbb{R}^{n} and 𝐡:nm\mathbf{\mathbf{h}:\mathbb{R}}^{n}\rightarrow\mathbb{R}^{m}. There is no restriction on functions 𝐟\mathbf{f} and 𝐡\mathbf{h}. For the system (3), we make

Assumption 1. For a given external input 𝐮\mathbf{u}, the system (3) with the initial value 𝐱0\mathbf{x}_{0} has a unique solution 𝐱\mathbf{x}^{*} on [0,)\left[0,\infty\right).

Noteworthy, most systems in practice satisfy Assumption 1, i.e., the uniqueness of solutions.

First, a primary system is brought in, having the same dimension as the original system:

B:{𝐱˙p=𝐟p(𝐱p,𝐮p)+𝐝,𝐱p(0)=𝐱p,0𝐲p=𝐡p(𝐱p)B:\left\{\begin{array}[]{l}\mathbf{\dot{x}}_{\text{p}}=\mathbf{f}_{\text{p}}\left({\mathbf{x}_{\text{p}},}\mathbf{u}_{\text{p}}\right)+\mathbf{d},\mathbf{x}_{\text{p}}\left(0\right)=\mathbf{x}_{\text{p},0}\\ \mathbf{y}_{\text{p}}=\mathbf{h}_{\text{p}}\left({\mathbf{x}_{\text{p}}}\right)\end{array}\right. (4)

where 𝐱p\mathbf{x}_{\text{p}}n,𝐮pm\in\mathbb{R}^{n},\mathbf{u}_{\text{p}}\in\mathbb{R}^{m}, 𝐲pp\mathbf{y}_{\text{p}}\in\mathbb{R}^{p}, 𝐟p:n×mn\mathbf{\mathbf{f}_{\text{p}}:\mathbb{R}}^{n}\mathbf{\times\mathbb{R}}^{m}\rightarrow\mathbb{R}^{n} and 𝐡p:nm\mathbf{\mathbf{h}_{\text{p}}:\mathbb{R}}^{n}\rightarrow\mathbb{R}^{m}. Functions 𝐟p\mathbf{f}_{\text{p}} and 𝐡p\mathbf{h}_{\text{p}} are determined by the designer. From the original system (3) and the primary system (4), the following secondary system is derived:

C=AB:{𝐱˙𝐱˙p=𝐟(𝐱,𝐮)𝐟p(𝐱p,𝐮p)𝐲𝐲p=𝐡(𝐱)𝐡p(𝐱p).C=A-B:\left\{\begin{array}[]{l}\mathbf{\dot{x}}-\mathbf{\dot{x}}_{\text{p}}=\mathbf{f}\left({\mathbf{x,}}\mathbf{u}\right)-\mathbf{f}_{\text{p}}\left({\mathbf{x}_{\text{p}},}\mathbf{u}_{\text{p}}\right)\\ \mathbf{y}-\mathbf{y}_{\text{p}}=\mathbf{h}\left({\mathbf{x}}\right)-\mathbf{h}_{\text{p}}\left({\mathbf{x}_{\text{p}}}\right).\end{array}\right.

New variables 𝐱s\mathbf{x}_{\text{s}}n,𝐮sm,\in\mathbb{R}^{n},\mathbf{u}_{\text{s}}\in\mathbb{R}^{m}, 𝐲sp\mathbf{y}{{}_{\text{s}}}\in\mathbb{R}^{p} are defined as follows

𝐱s𝐱𝐱p,𝐮s𝐮𝐮,p𝐲s𝐲𝐲p.{\mathbf{x}_{\text{s}}}\triangleq{\mathbf{x}}-{\mathbf{x}_{\text{p}},}\mathbf{u}{{}_{\text{s}}}\triangleq\mathbf{u}-\mathbf{u}{{}_{\text{p}},}\mathbf{y}{{}_{\text{s}}}\triangleq\mathbf{y}-\mathbf{y}_{\text{p}}. (5)

Then, the secondary system can be further written as follows

𝐱˙s\displaystyle\mathbf{\dot{x}}_{\text{s}} =𝐟(𝐱p+𝐱s,𝐮p+𝐮)s𝐟p(𝐱p,𝐮p),𝐱s(0)=𝐱s,0\displaystyle=\mathbf{f}\left({\mathbf{x}_{\text{p}}}+{\mathbf{x}_{\text{s}},}\mathbf{u}_{\text{p}}+\mathbf{u}{{}_{\text{s}}}\right)-\mathbf{f}_{\text{p}}\left({\mathbf{x}_{\text{p}},\mathbf{u}_{\text{p}}}\right),\mathbf{x}_{\text{s}}\left(0\right)=\mathbf{x}_{\text{s,0}}
𝐲s\displaystyle\mathbf{y}_{\text{s}} =𝐡(𝐱p+𝐱s)𝐡p(𝐱p).\displaystyle=\mathbf{h}\left({\mathbf{x}_{\text{p}}}+{\mathbf{x}_{\text{s}}}\right)-\mathbf{h}_{\text{p}}\left({\mathbf{x}_{\text{p}}}\right). (6)

From the definition (5), it follows

𝐱=𝐱p+𝐱s,𝐮=𝐮+p𝐮,s𝐲=𝐲p+𝐲s.{\mathbf{x}}={\mathbf{x}_{\text{p}}}+{\mathbf{x}_{\text{s}},}\mathbf{u}=\mathbf{u}{{}_{\text{p}}}+\mathbf{u}{{}_{\text{s}},\mathbf{y}}={\mathbf{y}_{\text{p}}}+{\mathbf{y}_{\text{s}}.} (7)

Now we can state

Lemma 1. Under Assumption 1, suppose 𝐱p\mathbf{x}_{\text{p}} and 𝐱s\mathbf{x}_{\text{s}} are the solutions of the system (4) and system (6) respectively, and the initial conditions of (3), (4) and (6) satisfy 𝐱0=𝐱p,0+𝐱s,0\mathbf{x}_{0}=\mathbf{x}_{\text{p},0}+\mathbf{x}_{\text{s,0}}. Then 𝐱=𝐱p+𝐱s{\mathbf{x}}={\mathbf{x}_{\text{p}}}+{\mathbf{x}_{\text{s}}}.

Proof. See reference [22].

Additive state decomposition is applicable not only to the above class of differential dynamic systems but also to almost any dynamic or static systems. As for a concrete decomposition process, it requires a specific analysis for a specific problem.

III A new linearization method: State compensation linearization

In this section, two existing linearization methods are recalled firstly. Then, as a new linearization method, state compensation linearization is proposed. Finally, the comparison of the three linearization methods is summarized.

For clarification, the following general nonlinear system is considered

𝐱˙=𝐟(𝐱,𝐮)+𝐝, 𝐱(0)=𝐱0\mathbf{\dot{x}=f}\left(\mathbf{x,u}\right)+\mathbf{d}\mathbf{,}\text{ }\mathbf{x}\left(0\right)=\mathbf{x}_{0} (8)

where 𝐱n\mathbf{x}\in\mathbb{R}^{n} is the state, 𝐮m\mathbf{u}\in\mathbb{R}^{m} is the input, 𝐝n\mathbf{d}\in\mathbb{R}^{n} denotes the disturbance, 𝐱0\mathbf{x}_{0} is the initial state, and 𝐟:n×mn\mathbf{f:\mathbb{R}}^{n}\mathbf{\times\mathbb{R}}^{m}\rightarrow\mathbb{R}^{n}. For simplicity, let 𝐟(𝟎,𝟎)=𝟎,\mathbf{f}\left(\mathbf{0,0}\right)=\mathbf{0,} which implies that the origin is the equilibrium point of the nominal system (8) with 𝐝=𝟎\mathbf{d}=\mathbf{0}.

Assumption 2. The state 𝐱\mathbf{x} is measurable.

Assumption 3. The function 𝐟(𝐱,𝐮)\mathbf{f}\left(\mathbf{x,u}\right) is continuously differentiable relative to 𝐱\mathbf{x} and 𝐮\mathbf{u} around the equilibrium point.

III-A Existing linearization methods

There are currently two main linearization methods: Jacobian linearization (i.e., linearization by Taylor’s expansion) and feedback linearization.

III-A1 Jacobian linearization

Jacobian linearization is based on the equilibrium point of the considered system. Considering that, when 𝐝=𝟎\mathbf{d}=\mathbf{0}, the equilibrium point of system (8) is the origin, the nonlinear system is linearized as

𝐱˙=𝐀1𝐱+𝐁1𝐮\mathbf{\dot{x}=A}_{1}\mathbf{x}+\mathbf{B}_{1}\mathbf{u} (9)

where 𝐀1=𝐟𝐱|𝐱=𝟎,𝐮=𝟎n×n,𝐁1=𝐟𝐮|𝐱=𝟎,𝐮=𝟎n×m\mathbf{A}_{1}=\left.\frac{\partial\mathbf{f}}{\partial\mathbf{x}}\right|_{\mathbf{x}=\mathbf{0,u}=\mathbf{0}}\in\mathbf{\mathbb{R}}^{n\times n},\mathbf{B}_{1}=\left.\frac{\partial\mathbf{f}}{\partial\mathbf{u}}\right|_{\mathbf{x}=\mathbf{0,u}=\mathbf{0}}\in\mathbf{\mathbb{R}}^{n\times m}. The detailed conditions for Jacobian linearization can be seen in [4]. Then, a controller is often designed as 𝐮=𝐋1(𝐱)\mathbf{u=L}_{1}\left(\mathbf{x}\right), where 𝐋1()\mathbf{L}_{1}\left(\mathbf{\cdot}\right) is a linear function, and the feedback state 𝐱\mathbf{x} is that of the real system (8).

Jacobian linearization is the most widely-used method to deal with nonlinear systems in practice. It is simple and intuitive. Moreover, this linearization method is easy-to-use because it can be applied to most nonlinear functions and the linearization process can be implemented by a program automatically. After Jacobian linearization, the physical meaning of states remains the same as before so that it can be easily understood. However, this method has only a local property. This implies that linearization may cause worse performance, or even instability of the resulting closed-loop system when the states are far away from the equilibrium points (trim states). In order to reduce approximation error, the number of equilibrium points required for the entire space has to be increased. It may be particularly large. As a result, this will result in a large number of linear models [23]. Then, the controller gains need to be scheduled according to the linear models [24]. Trajectory linearization is a special kind of Jacobian linearization that linearizes the system at every point on the nominal trajectory [25]. In essence, the obtained linearized systems are time-varying, and time-varying analysis and control design are necessary.

III-A2 Feedback linearization

First, it should be pointed out that the feedback linearization method is only applicable to precisely known systems without disturbances and uncertainties. Thus, it is assumed that 𝐝=𝟎\mathbf{d}=\mathbf{0} in the following discussion. Note that the following affine nonlinear systems are often considered in feedback linearization

𝐱˙=𝐟(𝐱)+𝐠(𝐱)𝐮\mathbf{\dot{x}=}\mathbf{f}\left(\mathbf{x}\right)+\mathbf{g}\left(\mathbf{x}\right)\mathbf{u} (10)

where 𝐟:nn\mathbf{f}:\mathbf{\mathbb{R}}^{n}\rightarrow\mathbb{R}^{n} and 𝐠:nn×m\mathbf{g}:\mathbf{\mathbb{R}}^{n}\rightarrow\mathbb{R}^{n\times m} are smooth vector fields. The detailed conditions for feedback linearization can been seen in [4]. Find a state transformation 𝐳=𝐓(𝐱)n\mathbf{z=T}\left(\mathbf{x}\right)\in\mathbb{R}^{n} and an input transformation 𝐮=𝐮(𝐳,𝐯)m\mathbf{u=u}\left(\mathbf{z,v}\right)\in\mathbb{R}^{m} so that the system (8) can be converted into the following form [4]

𝐳˙=𝐀2𝐳+𝐁2𝐯\mathbf{\dot{z}=A}_{2}\mathbf{z+B}_{2}\mathbf{v} (11)

where 𝐀2n×n,\mathbf{A}_{2}\in\mathbb{R}^{n\times n}, 𝐁2n×m\mathbf{B}_{2}\in\mathbb{R}^{n\times m}. For system (11), the controller is designed as 𝐯=𝐋2(𝐳)\mathbf{v=L}_{2}\left(\mathbf{z}\right). The feedback control is based on the variable 𝐳\mathbf{z} after the coordinate transformation. Finally, the actual control input can be obtained by 𝐮=𝐮(𝐳,𝐯)\mathbf{u=u}\left(\mathbf{z,v}\right).

After feedback linearization, system (11) no longer has multiple equilibrium points and has a linear property in the entire transformed space. However, the coordinate transformation is complicated. There are often many situations where feedback linearization is difficult to apply: (i) the exact model of the nonlinear system is not available, and no robustness is guaranteed in the presence of parameter uncertainties or unmodeled dynamics; (ii) the control law is not defined at singularity points; (iii) it is not easy to use for NMP systems, which may lead to unstable control; (iv) the physical meaning of new states differs from that of the original system, which makes the controller parameter adjustment difficult.

To avoid the disadvantages of the two existing linearization methods, a new linearization method named state compensation linearization, which is based on the additive state decomposition, is proposed as a kind of complementation in the next subsection.

III-B State compensation linearization as a kind of complementation

In this part, the additive state decomposition is applied to linearization, which leads to a new linearization method named state compensation linearization. For the original nonlinear system (8), denoted as AA, the primary system is chosen as

𝐱˙p=𝐀1𝐱p+𝐁1𝐮p+𝐝,𝐱p(0)=𝐱0\mathbf{\dot{x}}_{\text{p}}=\mathbf{A}_{1}\mathbf{x}_{\text{p}}+\mathbf{B}_{1}\mathbf{u}_{\text{p}}+\mathbf{d},\mathbf{x}_{\text{p}}\left(0\right)=\mathbf{x}_{0} (12)

which is a linear system and denoted as BB. Here, 𝐀1=𝐟𝐱|𝐱=𝟎,𝐮=𝟎n×n,𝐁1=𝐟𝐮|𝐱=𝟎,𝐮=𝟎n×m\mathbf{A}_{1}=\left.\frac{\partial\mathbf{f}}{\partial\mathbf{x}}\right|_{\mathbf{x}=\mathbf{0,u}=\mathbf{0}}\in\mathbf{\mathbb{R}}^{n\times n},\mathbf{B}_{1}=\left.\frac{\partial\mathbf{f}}{\partial\mathbf{u}}\right|_{\mathbf{x}=\mathbf{0,u}=\mathbf{0}}\in\mathbf{\mathbb{R}}^{n\times m}, which are the same as in (9). Then, the secondary system is determined by the original system and the primary system using the rule (6), namely C=ABC=A-B, that

𝐱˙s=𝐟(𝐱,𝐮)𝐀1𝐱p𝐁1𝐮p,𝐱s(0)=𝟎,\mathbf{\dot{x}}_{\text{s}}=\mathbf{f}\left(\mathbf{x},\mathbf{u}\right)-\mathbf{A}_{1}\mathbf{x}_{\text{p}}-\mathbf{B}_{1}\mathbf{u}_{\text{p}},\mathbf{x}_{\text{s}}\left(0\right)=\mathbf{0}, (13)

which can be then written as

𝐱˙s=𝐟(𝐱,𝐮)+𝐀1(𝐱s𝐱)+𝐁1(𝐮s𝐮),𝐱s(0)=𝟎\mathbf{\dot{x}}_{\text{s}}=\mathbf{f}\left(\mathbf{x},\mathbf{u}\right)+\mathbf{A}_{1}\left(\mathbf{x}_{\text{s}}-\mathbf{x}\right)+\mathbf{B}_{1}\left(\mathbf{u}_{\text{s}}-\mathbf{u}\right),\mathbf{x}_{\text{s}}\left(0\right)=\mathbf{0} (14)

where

𝐱=𝐱p+𝐱s,𝐮=𝐮+p𝐮.s{\mathbf{x}}={\mathbf{x}_{\text{p}}}+{\mathbf{x}_{\text{s}},}\mathbf{u}=\mathbf{u}{{}_{\text{p}}}+\mathbf{u}{{}_{\text{s}}.} (15)

In order to accomplish linearization, the linear system (12) (B=ACB=A-C) can be obtained by compensating for the nonlinearity of the original system (AA) using the secondary system (CC) and its controller 𝐮s=𝐋4(𝐱,𝐱s)\mathbf{u}_{\text{s}}=\mathbf{L}_{4}\left(\mathbf{x},\mathbf{x}_{\text{s}}\right) as shown in Fig. 1. The system from the input 𝐮p\mathbf{u}_{\text{p}} to the state 𝐱p\mathbf{x}_{\text{p}} is linear. In the subsystems (12) and (13), the secondary system (13) is a virtual exact system, which exists only in design. On the other hand, since the sum of the primary system and the secondary system is the original system, all the uncertainties, such as the higher-order unmodeled dynamics of the real system, are left in the primary system. After state compensation linearization, the controller for the linear primary system (12) can be designed as 𝐮p=𝐋3(𝐱p)\mathbf{u}_{\text{p}}=\mathbf{L}_{3}\left(\mathbf{x}_{\text{p}}\right).

Refer to caption
Figure 1: The construction diagram of the linear primary system from the input 𝐮p\mathbf{u}_{\text{p}} to the state 𝐱p\mathbf{x}_{\text{p}}. The primary system can be obtained by subtracting the complementary secondary system from the real system.

Based on the above content, the following theorem about the conditions for state compensation linearization can be stated.

Theorem 1. System (8) can be linearized by state compensation linearization if

i). in the entire space 𝐱n,𝐮m\mathbf{x}\in\mathbb{R}^{n},\mathbf{u}\in\mathbb{R}^{m}, Assumption 1 is satisfied.

ii). it can be linearizable by Jacobian linearization.

Proof. First, if Condition i) is met, according to Lemma 1, additive state decomposition holds. Next, if system (8) can be linearizable by Jacobian linearization, then the matrices 𝐀1,𝐁1\mathbf{A}_{1},\mathbf{B}_{1} can be obtained, and thus the primary system (12) can be determined. Then, according to additive state decomposition, the secondary system (13) is determined by the original system and the primary system using the rule (6), namely C=ABC=A-B. Next, one can design a controller 𝐮s=𝐋4(𝐱,𝐱s)\mathbf{u}_{\text{s}}=\mathbf{L}_{4}\left(\mathbf{x},\mathbf{x}_{\text{s}}\right) to stabilize the secondary system. Consequently, by compensating for the nonlinearity of the original system using the secondary system and its controller, the linear system (i.e., the primary system (12)) can be obtained. \square

State compensation linearization no longer has multiple equilibrium points. Moreover, in state compensation linearization, 𝐱=𝐱p+𝐱s,{\mathbf{x}}={\mathbf{x}_{\text{p}}}+{\mathbf{x}_{\text{s}},} there are no coordinate transformation or unit transformation. In some sense, 𝐱p\mathbf{x}_{\text{p}} can be considered as a part of 𝐱\mathbf{x}, and they have the same dimension and unit. Thus, the physical meaning of the states remains the same after state compensation linearization. This property of state compensation linearization is similar to that of Jacobian linearization. In contrast, the new states obtained from feedback linearization have different physical meanings with the original states because of coordinate transformation.

Remark 1. In state compensation linearization, one keeps both ‘linear primary’ and ‘nonlinear secondary’ component systems. Furthermore, it is not necessarily assumed that the secondary system is ‘small’ in some sense. The secondary system may be significant when the original system has a strong nonlinearity. However, what we often do is to stabilize the secondary system so that it does become ‘small’ eventually, leaving one free to work with the primary system. AA and BB are said to be equivalent if the nonlinear secondary system ABA-B can be stabilized by a controller. Then, checking the equivalence means checking whether the system ABA-B can be stabilized. In that case a nonlinear system AA is equivalent to the linear system BB for which ABA-B can be stabilized. The difference between state compensation linearization and Jacobian linearization is that state compensation linearization does not throw the ‘error’ system in Jacobian linearization away.

Remark 2. In state compensation linearization, one often needs to construct a stabilizing controller for a nonlinear secondary system. Although designing a stabilizing controller may cost some effort, the benefit is also attractive. The nonlinearity in the secondary system can be better considered, and thus better performance can be expected. Besides, because all the uncertainties and the initial value are considered in the primary system, the stabilizing control problem for the secondary system is a stabilizing control problem of an exact nonlinear system with zero initial states, which is often not complicated.

III-C Comparison of the three linearization methods

Although the linearized systems (12) and (9) have the same 𝐀1,𝐁1\mathbf{A}_{1},\mathbf{B}_{1}, the state and input in (12) are 𝐱p\mathbf{x}_{\text{p}} and 𝐮p\mathbf{u}_{\text{p}} instead of the real state and input 𝐱\mathbf{x} and 𝐮\mathbf{u} in (9).

Under some specific conditions, state compensation linearization can be transformed into Jacobian linearization. Considering the original system (8) with 𝐝=𝟎\mathbf{d}=\mathbf{0}. In state compensation linearization, let 𝐱s=𝟎\mathbf{x}_{\text{s}}=\mathbf{0}, 𝐮s=𝟎\mathbf{u}_{\text{s}}=\mathbf{0}, then the resulting linearized system is the same as that of Jacobian linearization, namely 𝐱˙=𝐀1𝐱+𝐁1𝐮\mathbf{\dot{x}=A}_{1}\mathbf{x}+\mathbf{B}_{1}\mathbf{u}.

State compensation linearization seems similar to Jacobian linearization. The difference is that the latter drops the higher-order small nonlinear terms directly and Taylor’s expansion holds locally, which determines its approximate and local properties. In contrast, state compensation linearization uses a compensation strategy, which makes state compensation linearization extend to the whole space and become a kind of accurate linearization method. This property is similar to that of feedback linearization. Thus, state compensation linearization combines the advantages of the two linearization methods in some sense.

Along with state compensation linearization, an original control task can be assigned to the linearized primary system and nonlinear secondary system by dividing the original control task into two simpler control tasks. From this perspective, additive state decomposition is also a task-oriented problem decomposition that offers a general way to decompose an original system into several subsystems in charge of simpler subtasks. Thanks to task decomposition, one can design controllers for simpler subtasks more easily than designing a controller for the original task. These controllers can finally be combined to accomplish the original control task. In the following, state compensation linearization applications to nonlinear control design are discussed.

IV State compensation linearization applications to nonlinear control design

Based on the state compensation linearization method in the previous section, a state-compensation-linearization-based control framework for a class of nonlinear systems is proposed in this subsection. State-compensation-linearization-based stabilizing control is discussed in the following for illustration.

IV-A Problem formulation

Consider the nonlinear system (8), state compensation linearization has been done in Subsection III-B. Then, the control objective is to make 𝐱(t)𝟎\mathbf{x}\left(t\right)\rightarrow\mathbf{0} or 𝐱(t)(𝟎n×1,σ)\mathbf{x}\left(t\right)\rightarrow\mathcal{B}\left(\mathbf{0}_{n\times 1},\sigma\right)111(𝐨,σ){𝐱n|𝐱𝐨σ},\mathcal{B}\left(\mathbf{o},\sigma\right)\triangleq\left\{\mathbf{x}\in\mathbb{R}^{n}\left|\left\|\mathbf{x}-\mathbf{o}\right\|\leq\sigma\right.\right\}, and 𝝃(t)(𝐨,σ)\boldsymbol{\xi}\left(t\right)\rightarrow\mathcal{B}\left(\mathbf{o},\sigma\right) signifies min𝐲(𝐨,σ)\underset{\mathbf{y}\in\mathcal{B}\left(\mathbf{o},\sigma\right)}{\min} 𝝃(t)𝐲0.\left\|\boldsymbol{\xi}\left(t\right)-\mathbf{y}\right\|\rightarrow 0. as tt\rightarrow\infty.

IV-B Observer design

It should be pointed out that the primary system and the secondary system are virtual systems, which do not exist in practice and just exist in the analysis. Thus, an observer is necessary to provide the variables of the primary system and the secondary system for the later control design. By using the state compensation linearization method, the unknown disturbance and the initial value are assigned in the primary system, and the secondary system is an exact nonlinear system. Thus, the following observer is built based on the secondary system to observe states [19],[21].

Theorem 2. Suppose that 𝐀1\mathbf{A}_{1} is stable, and an observer is designed to estimate 𝐱p\mathbf{x}_{\text{p}}, 𝐱s\mathbf{x}_{\text{s}} in (12) and (13) as

𝐱^˙s\displaystyle\mathbf{\dot{\hat{x}}}_{\text{s}} =𝐟(𝐱,𝐮)+𝐀1(𝐱^s𝐱)+𝐁1(𝐮s𝐮),𝐱^s(0)=𝟎\displaystyle=\mathbf{f}\left(\mathbf{x},\mathbf{u}\right)+\mathbf{A}_{1}\left(\mathbf{\hat{x}}_{\text{s}}-\mathbf{x}\right)+\mathbf{B}_{1}\left(\mathbf{u}_{\text{s}}-\mathbf{u}\right),\mathbf{\hat{x}}_{\text{s}}\left(0\right)=\mathbf{0} (16)
𝐱^p\displaystyle\mathbf{\hat{x}}_{\text{p}} =𝐱𝐱^s\displaystyle=\mathbf{x-\hat{x}}_{\text{s}} (17)

where 𝐱^s\mathbf{\hat{x}}_{\text{s}} and 𝐱^p\mathbf{\hat{x}}_{\text{p}} are the estimation of 𝐱s\mathbf{x}_{\text{s}} and 𝐱p\mathbf{x}_{\text{p}}. Then 𝐱^p=𝐱p\hat{\mathbf{x}}_{\text{p}}=\mathbf{x}_{\text{p}}, 𝐱^s=𝐱s\mathbf{\hat{x}}_{\text{s}}=\mathbf{x}_{\text{s}}.

Proof. Subtracting (16) from (13) results in 𝐱~˙s=𝐀1𝐱~s\dot{\tilde{\mathbf{x}}}_{\text{s}}=\mathbf{A}_{1}\tilde{\mathbf{x}}_{\text{s}} and 𝐱~s(0)=𝟎\tilde{\mathbf{x}}_{\text{s}}\left(0\right)=\mathbf{0}, where 𝐱~s𝐱s𝐱^s\tilde{\mathbf{x}}_{\text{s}}\triangleq\mathbf{x}_{\text{s}}-\hat{\mathbf{x}}_{\text{s}}. Then, considering that 𝐀1\mathbf{A}_{1} is stable, 𝐱~s=𝟎\tilde{\mathbf{x}}_{\text{s}}=\mathbf{0}, which implies that 𝐱^s=𝐱s\mathbf{\hat{x}}_{\text{s}}=\mathbf{x}_{\text{s}}. Consequently, by (15), it can be obtained that 𝐱^p=𝐱𝐱^s=𝐱p\mathbf{\hat{x}}_{\text{p}}=\mathbf{x-\hat{x}}_{\text{s}}=\mathbf{x}_{\text{p}}. \square

Remark 3. The designed observer is an open-loop observer, in which 𝐀1\mathbf{A}_{1} must be stable. Otherwise, state feedback is needed to obtain a stable 𝐀1\mathbf{A}_{1}. If a closed-loop observer is adopted here, it will be rather difficult to analyze the stability of the closed-loop system consisting of a nonlinear controller and the observer because separation principle does not hold in nonlinear systems.

Remark 4. The initial values 𝐱s(0)\mathbf{x}_{\text{s}}\left(0\right) and 𝐱^s(0)\mathbf{\hat{x}}_{\text{s}}\left(0\right) are both assigned as 𝐱s(0)=𝟎\mathbf{x}_{\text{s}}\left(0\right)=\mathbf{0} and 𝐱^s(0)=𝟎\mathbf{\hat{x}}_{\text{s}}\left(0\right)=\mathbf{0}. If there is an initial value measurement error, it will be assigned to and considered in the primary system in the form of disturbances.

IV-C State-compensation-linearization-based control framework

Based on state compensation linearization, the original system (8) is decomposed into two systems: an LTI system including all disturbances as the primary system, together with the secondary system whose equilibrium point is the origin. Since the states of the primary system and the secondary system can be observed, the original stabilizing problem for the system (8) is correspondingly decomposed into two problems: a stabilizing problem for the nonlinear secondary system (for the nonlinearity compensation objective) and a stabilizing problem for an LTI primary system (for the control objective). Because the primary system is linear, the classical frequency-domain and time-domain methods can be applied, such as lead-lag compensation. Therefore, the original problem turns into two simpler subproblems. The final objective is achieved once the objectives of the two subproblems are achieved.

  • Problem 1 Consider the secondary system (13). Design the secondary controller as

    𝐮s=𝐋(𝐱,𝐱s)\mathbf{u}_{\text{s}}=\mathbf{L}(\mathbf{x},\mathbf{x_{\text{s}}}) (18)

    such that 𝐱s(t)𝟎\mathbf{x}_{\text{s}}\left(t\right)\rightarrow\mathbf{0} as tt\rightarrow\infty, where 𝐋(,)\mathbf{L}(\cdot,\cdot) is a nonlinear function.

After the compensation of the secondary system and its controller (18), one can get a linear system (primary system). Next, just a stabilizing problem for a linear system needs to be considered.

  • Problem 2 Consider the primary system (12). Design the primary controller as

    𝐮p=1(𝐇(s)𝐊𝐱p(s))\mathbf{u}_{\text{p}}=\mathcal{L}^{-1}(\mathbf{H}(s)\mathbf{K}\mathbf{x_{\text{p}}}(s)) (19)

    such that 𝐱p(t)𝟎\mathbf{x}_{\text{p}}\left(t\right)\rightarrow\mathbf{0} or 𝐱p(t)(𝟎n×1,σ)\mathbf{x}_{\text{p}}\left(t\right)\rightarrow\mathcal{B}\left(\mathbf{0}_{n\times 1},\sigma\right) as tt\rightarrow\infty, where H(s) is a stable transfer function, 1\mathcal{L}^{-1} denotes the inverse Laplace transformation, and state feedback matrix 𝐊m×n\mathbf{K\in\mathbb{R}^{\text{$m\times n$}}}.

Remark 5. In the controller (18), the state feedback matrix 𝐊\mathbf{K} can be determined by the linear quadratic regulator (LQR) method or the “PID Tuner App” in the Matlab software. Besides, one can use some kind of robust methods to construct a controller for the primary system.

The controller of the original system can be obtained by combing the controllers of the primary system and the secondary system. With the solutions to the two problems in hand, we can state

Theorem 3. For system (8) under Assumptions 1-3, suppose (i) 𝐀1\mathbf{A}_{1} is stable; (ii) Problems 1-2 are solved; (iii) the controller for system (8) is designed as

𝐱^˙s\displaystyle\mathbf{\dot{\hat{x}}}_{\text{s}} =𝐟(𝐱,𝐮)+𝐀1(𝐱^s𝐱)+𝐁1(𝐮s𝐮)\displaystyle=\mathbf{f}\left(\mathbf{x},\mathbf{u}\right)+\mathbf{A}_{1}\left(\mathbf{\hat{x}}_{\text{s}}-\mathbf{x}\right)+\mathbf{B}_{1}\left(\mathbf{u}_{\text{s}}-\mathbf{u}\right)
𝐱^p\displaystyle\hat{\mathbf{x}}_{\text{p}} =𝐱𝐱^s\displaystyle=\mathbf{x}-\mathbf{\hat{x}}_{\text{s}}
𝐮\displaystyle\mathbf{u} =1(𝐇(s)𝐊𝐱^p(s))+𝐋(𝐱,𝐱^s).\displaystyle\mathcal{=L}^{-1}(\mathbf{H}(s)\mathbf{K}\hat{\mathbf{x}}_{\text{p}}(s))+\mathbf{L}(\mathbf{x},\mathbf{\hat{x}}_{\text{s}}). (20)

Then, the state of system (8) satisfies 𝐱(t)𝟎\mathbf{x}\left(t\right)\rightarrow\mathbf{0} or 𝐱(t)(𝟎n×1,σ)\mathbf{x}\left(t\right)\rightarrow\mathcal{B}\left(\mathbf{0}_{n\times 1},\sigma\right) as tt\rightarrow\infty.

Proof. It is easy to obtain that 𝐱^p=𝐱p\mathbf{\hat{x}}_{\text{p}}=\mathbf{x}_{\text{p}} and 𝐱^s=𝐱s\mathbf{\hat{x}}_{\text{s}}=\mathbf{x}_{\text{s}} with the observer (16)-(17). Then, since Problem 1 is solved, 𝐱s(t)𝟎\mathbf{x}_{\text{s}}\left(t\right)\rightarrow\mathbf{0} as tt\rightarrow\infty. Furthermore, 𝐱p(t)𝟎\mathbf{x}_{\text{p}}\left(t\right)\rightarrow\mathbf{0} or 𝐱p(t)(𝟎n×1,σ)\mathbf{x}_{\text{p}}\left(t\right)\rightarrow\mathcal{B}\left(\mathbf{0}_{n\times 1},\sigma\right) as tt\rightarrow\infty with Problem 2 solved. By additive state decomposition, we have 𝐱=𝐱p+𝐱s𝟎\mathbf{x=\mathbf{x}}_{\text{p}}+\mathbf{x}_{\text{s}}\rightarrow\mathbf{0} or 𝐱(t)(𝟎n×1,σ)\mathbf{x}\left(t\right)\rightarrow\mathcal{B}\left(\mathbf{0}_{n\times 1},\sigma\right) as tt\rightarrow\infty. \square

In this new framework, the original problem is simplified, and different control methods are allowed to be applied to the systems. The closed-loop system is shown in Fig. 2. The controller based on state compensation linearization includes three parts: the controller of the primary system, the controller of the secondary system and the observer. In practice, the primary system and the secondary system do not need to have physical status. They are only the models existing in the design.

Refer to caption
Figure 2: The closed-loop control diagram based on the state-compensation-linearization-based stabilizing control for a class of nonlinear systems

Remark 6. The basic idea of the state-compensation-linearization-based tracking control framework is similar to that of the state-compensation-linearization-based stabilizing control framework. The most significant difference is that a tracking control problem rather than a stabilizing control problem needs to be specified to the linear primary system.

V Simulation study

In the following, three examples are given to illustrate and compare different linearization methods for control design.

V-A Example 1: A bilinear system

Consider a bilinear system as follows

x˙\displaystyle\dot{x} =4x+xu+d\displaystyle=-4x+xu+d
y\displaystyle y =x\displaystyle=x (21)

where x,y,u,dx,y,u,d\in\mathbb{R} represent the state, output, input and disturbance, respectively. It is assumed that d0d\neq 0. The tracking control goal is to make yydy\rightarrow y_{\text{d}} as tt\rightarrow\infty.

V-A1 Jacobian-linearization-based-control (JLC) and feedback-linearization-based control (FLC)

The two existing linearization methods are attempted to linearize system (21) so that the given tracking control goal can be achieved by using linear control methods. When the Jacobian linearization method is applied, the first thing is to obtain equilibrium points of system (21). Unfortunately, the stability property of the nonlinear system (21) is affected by the disturbance dd, which is unknown. Additionally, uu can be arbitrary values when x=0x=0. This implies that equilibrium points of system (21) cannot be obtained. Without equilibrium points, it is impossible to linearize system (21) by Taylor’s expansion. When the feedback linearization method is applied, let y˙=v\dot{y}=v. Then, v=4x+xu+dv=-4x+xu+d. There exists a singularity problem that system (21) becomes uncontrollable [26] when x=0x=0. Thus, neither of the two existing linearization methods is suitable for the bilinear system.

V-A2 State-compensation-linearization-based control (SCLC)

Considering the two existing linearization methods are not readily applicable to system (21), the state compensation linearization method is then adopted. First, a linear primary system (BB) is designed as follows

x˙p\displaystyle\dot{x}_{\text{p}} =4xp+ydup+d,xp(0)=x0\displaystyle=-4x_{\text{p}}+y_{\text{d}}u_{\text{p}}+d,x_{\text{p}}\left(0\right)=x_{0}
yp\displaystyle y_{\text{p}} =xp\displaystyle=x_{\text{p}} (22)

where let up=uu_{\text{p}}=u, ydy_{\text{d}} is the desired output. In the linearization process, because the desired input udu_{\text{d}} is unknown, ud=0u_{\text{d}}=0 is chosen for simplicity. Then, the following secondary system (C=ABC=A-B) is determined by subtracting the primary system (22) from the original system (21) with rule (6)

x˙s\displaystyle\dot{x}_{\text{s}} =4xs+xuydu\displaystyle=-4x_{\text{s}}+xu-y_{\text{d}}u
=(u4)xs+(xpyd)u\displaystyle=\left(u-4\right)x_{\text{s}}+\left(x_{\text{p}}-y_{\text{d}}\right)u
ys\displaystyle y_{\text{s}} =xs.\displaystyle=x_{\text{s}}. (23)

According to additive state decomposition, it follows

x=x+px,su=u.px=x{{}_{\text{p}}}+x{{}_{\text{s}},}u=u{{}_{\text{p}}.} (24)

Then, the linear system (22) (B=ACB=A-C) can be obtained by compensating for the nonlinearity of the original system (21) (AA) using the secondary system (CC). The structure is displayed in Fig. 3. If the controller for the primary system (22) is well designed, namely xpydx_{\text{p}}\rightarrow y_{\text{d}}, uudu\rightarrow u_{\text{d}}. Then, the secondary system is approaching the dynamics of x˙s=(ud4)xs,\dot{x}_{\text{s}}=\left(u_{\text{d}}-4\right)x_{\text{s}}, which implies that xs0x_{\text{s}}\rightarrow 0 as long as ud<4u_{\text{d}}<4. The final linear system (22) is independent of system (23). The remaining problem is to design a controller for the linear system (22), for which any linear control methods can be utilized. Noteworthy, the proposed state compensation linearization method can avoid the equilibrium point problem of the Jacobian linearization method and the singularity problem of the feedback linearization method. Let yd=20y_{\text{d}}=20, d=3d=3, x0=1.x_{0}=-1. A proportional-integral-differential (PID) controller is designed for the primary system (22), and the PID parameters are kp=0.66,ki=1.33,kd=0.02k_{\text{p}}=0.66,k_{\text{i}}=1.33,k_{\text{d}}=-0.02, which is tuned by the “PID Tuner App” in the Matlab software. The simulation result is displayed in Fig. 4. The output can track the reference output well. It is noticed that due to continuity, yy goes from 1-1 to 2020, namely x=0x=0 at a certain time. This will make feedback linearization fail.

Refer to caption
Figure 3: The state compensation linearization structure. The primary system (22) from the input upu_{\text{p}} to the state xx_{\text{p }} is linear, which can be obtained by subtracting the complementary secondary system (23) from the original system (21).
Refer to caption
Figure 4: Output response of system (21) based on the state compensation linearization method (Example 1)

V-B Example 2: An NMP system with saturation

Consider an NMP system with input saturation as follows

𝐱˙\displaystyle\mathbf{\dot{x}} =𝐀𝐱+𝐛sat(u),𝐱(0)=𝐱0\displaystyle=\mathbf{Ax}+\mathbf{b}\text{sat}\left(u\right),\mathbf{x}\left(0\right)=\mathbf{x}_{0}
y\displaystyle y =𝐜T𝐱\displaystyle=\mathbf{c}^{\text{T}}\mathbf{x} (25)

with

𝐀=[010001464],𝐛=[001],𝐜=[101].\mathbf{A=}\left[\begin{array}[]{ccc}0&1&0\\ 0&0&1\\ -4&-6&-4\end{array}\right],\mathbf{b}=\left[\begin{array}[]{c}0\\ 0\\ 1\end{array}\right],\mathbf{c}=\left[\begin{array}[]{c}-1\\ 0\\ 1\end{array}\right].

The saturation function sat()\left(\cdot\right) is defined as

sat(u)={2,u>2u,2>u>22,u<2.\text{sat}\left(u\right)=\left\{\begin{array}[]{l}2,u>2\\ u,2>u>-2\\ -2,u<-2.\end{array}\right.

The system state is measurable. The control objective is to design uu such that yy tracks ydy_{\text{d}}. Next, SCLC is adopted. System (25) can be decomposed into a linear primary system

𝐱˙p\displaystyle\mathbf{\dot{x}}_{\text{p}} =𝐀𝐱p+𝐛up,𝐱p(0)=𝐱0\displaystyle=\mathbf{Ax}_{\text{p}}+\mathbf{b}u_{\text{p}},\mathbf{x}_{\text{p}}\left(0\right)=\mathbf{x}_{0}
yp\displaystyle y_{\text{p}} =𝐜T𝐱p\displaystyle=\mathbf{c}^{\text{T}}\mathbf{x}_{\text{p}} (26)

and a nonlinear secondary system

𝐱˙s\displaystyle\mathbf{\dot{x}}_{\text{s}} =𝐀𝐱s+𝐛(sat(u)u),𝐱s(0)=𝟎\displaystyle=\mathbf{Ax}_{\text{s}}+\mathbf{b}\left(\text{sat}\left(u\right)-u\right),\mathbf{x}_{\text{s}}\left(0\right)=\mathbf{0}
ys\displaystyle y_{\text{s}} =𝐜T𝐱s\displaystyle=\mathbf{c}^{\text{T}}\mathbf{x}_{\text{s}} (27)

where let up=uu_{\text{p}}=u. According to additive state decomposition, it follows

𝐱=𝐱p+𝐱s,u=u.p{\mathbf{x}}={\mathbf{x}_{\text{p}}}+{\mathbf{x}_{\text{s}},}u=u{{}_{\text{p}}.} (28)

Then, the linear system (26) (B=ACB=A-C) can be obtained by compensating for the nonlinearity of the original system (25) (AA) using the secondary system (CC). If the controller for the primary system (26) is well designed, sat(up)up\left(u_{\text{p}}\right)\rightarrow u_{\text{p}}, upudu_{\text{p}}\rightarrow u_{\text{d}}, which also implies uudu\rightarrow u_{\text{d}}. Then, sat(up)u0\left(u_{\text{p}}\right)-u\rightarrow 0. In this case, since 𝐀\mathbf{A} is stable, the secondary system (27) can stabilize itself. The remaining problem is to design a tracking controller for (26) without saturation. By using state compensation linearization and the remaining tracking controller for (26), yydy\rightarrow y_{\text{d}} can hold globally. Let 𝐱0=𝟎\mathbf{x}_{0}=\mathbf{0} and

yd(t)={sin(0.25t),0t4π0,4πt.y_{\text{d}}\left(t\right)=\left\{\begin{array}[]{cc}\sin(0.25t),&0\leq t\leq 4\pi\\ 0,&4\pi\leq t.\end{array}\right.

With a PID controller (the PID parameters are kp=0.5,ki=1.3,kd=0k_{\text{p}}=-0.5,k_{\text{i}}=-1.3,k_{\text{d}}=0, which is tuned by the “PID Tuner App” in the Matlab software) designed for the primary system (26), the simulation result is displayed in Fig. 5(a). The simulation result of JLC is shown in Fig. 5(b) for comparison. For the SCLC, the system undergoes input saturation from 3.33.3s to 11.511.5s (7.2s duration); for the JLC, with the same PID controller, the system undergoes input saturation from 3.53.5s to 16.216.2s (12.7s duration). It can be seen that the SCLC can exit saturation earlier than the JLC and can achieve better tracking performance. In terms of FLC, because the saturation function is irreversible, the real input uu cannot be obtained. Moreover, because system (25) is an NMP system, FLC may lead to an unstable controller.

Refer to caption
Figure 5: Output response (Example 2): (a) Compensation linearization method; (b) Jacobian linearization method.

V-C Example 3: A mismatched nonlinear system

Consider a mismatched nonlinear system as follows

x˙1\displaystyle\dot{x}_{1} =x2+sinx2+d1,\displaystyle=x_{2}+\sin x_{2}+d_{1},
x˙2\displaystyle\dot{x}_{2} =2x13x2+2x22+u+d2,𝐱(0)=𝐱0\displaystyle=-2x_{1}-3x_{2}+2x_{2}^{2}+u+d_{2},\mathbf{x}\left(0\right)=\mathbf{x}_{0} (29)

where 𝐱=[x1x2]T\mathbf{x=}\left[x_{1}\ x_{2}\right]^{\text{T}}is the state, uu is the input, 𝐱0\mathbf{x}_{0} is the initial state value, and 𝐝=[d1d2]T\mathbf{d=}\left[d_{1}\ d_{2}\right]^{\text{T}}represents the unknown uncertainties.

V-C1 SCLC

Based on the proposed state compensation linearization, for the original system (29), the linear primary system (12) is chosen as

x˙p,1\displaystyle\dot{x}_{\text{p},1} =2xp,2+d1,\displaystyle=2x_{\text{p},2}+d_{1},
x˙p,2\displaystyle\dot{x}_{\text{p},2} =2xp,13xp,2+up+d2,𝐱p(0)=𝐱0.\displaystyle=-2x_{\text{p},1}-3x_{\text{p},2}+u_{\text{p}}+d_{2},\mathbf{x}_{\text{p}}\left(0\right)=\mathbf{x}_{0}. (30)

Its state-space form is 𝐱˙=𝐀𝐱+𝐁u+𝐝\mathbf{\dot{x}=}\mathbf{A}\mathbf{x}+\mathbf{B}u+\mathbf{d} with

𝐀=[2023],𝐁=[01].\mathbf{A}=\left[\begin{array}[]{cc}2&0\\ -2&-3\end{array}\right],\mathbf{B}=\left[\begin{array}[]{c}0\\ 1\end{array}\right].

Then, the secondary system (13) can be obtained as

x˙s,1\displaystyle\dot{x}_{\text{s},1} =2xs,2x+sin(x),\displaystyle=2x_{\text{s},2}-x+\text{sin}(x),
x˙s,2\displaystyle\dot{x}_{\text{s},2} =2xs,13xs,2+2x+2us,𝐱s(0)=𝟎.\displaystyle=-2x_{\text{s},1}-3x_{\text{s},2}+2x{}^{2}+u_{\text{s}},\mathbf{x}_{\text{s}}\left(0\right)=\mathbf{0}. (31)

Next, controller design is carried out. For (30), the LQR method is employed to calculate a feedback matrix 𝐤2×1\mathbf{k}\in\mathbb{R}^{2\times 1}, and a controller is designed as

up=𝐤T𝐱^p.u_{\text{p}}=\mathbf{k}^{\text{T}}\mathbf{\hat{x}}_{\text{p}}.

For (31), a backstepping controller is designed as

us\displaystyle u_{\text{s}} =2x^s,1+3x^s,22x2\displaystyle=2\hat{x}_{\text{s},1}+3\hat{x}_{\text{s},2}-2x^{2}
a1x^s,12+1(sin(x^s,2)+x^s,2+g)czs,2\displaystyle-a\frac{1}{\hat{x}_{\text{s},1}^{2}+1}\left(\sin\left(\hat{x}_{\text{s},2}\right)+\hat{x}_{\text{s},2}+g\right)-cz_{\text{s},2}

where zs,2=x^s,2+aarctanx^s,1z_{\text{s},2}=\hat{x}_{\text{s},2}+a\arctan\hat{x}_{\text{s},1}, g=sin(x)sin(x^s,2)g=\sin\left(x\right)-\sin\left(\hat{x}_{\text{s},2}\right), a,c>0a,c>0 are parameters to be specified later. By combining the two controllers, the final controller follows

u\displaystyle u =up+us=𝐤T𝐱^p+2x^s,1+3x^s,22x2\displaystyle=u_{\text{p}}+u_{\text{s}}=\mathbf{k}^{\text{T}}\mathbf{\hat{x}}_{\text{p}}+2\hat{x}_{\text{s},1}+3\hat{x}_{\text{s},2}-2x^{2}
a1x^s,12+1(sin(x^s,2)+x^s,2+g)czs,2.\displaystyle-a\frac{1}{\hat{x}_{\text{s},1}^{2}+1}\left(\sin\left(\hat{x}_{\text{s},2}\right)+\hat{x}_{\text{s},2}+g\right)-cz_{\text{s},2}.

V-C2 JLC for comparison

By using Jacobian linearization, the following linearized system is obtained

x˙1\displaystyle\dot{x}_{1} =2x2\displaystyle=2x_{2}
x˙2\displaystyle\dot{x}_{2} =2x13x2+u.\displaystyle=-2x_{1}-3x_{2}+u. (32)

Its state-space form is 𝐱˙=𝐀𝐱+𝐁u\mathbf{\dot{x}=}\mathbf{A}\mathbf{x}+\mathbf{B}u with

𝐀=[2023],𝐁=[01].\mathbf{A}=\left[\begin{array}[]{cc}2&0\\ -2&-3\end{array}\right],\mathbf{B}=\left[\begin{array}[]{c}0\\ 1\end{array}\right].

Based on the linear system (32), a controller is designed as

u=𝐤T𝐱.u=\mathbf{k}^{\text{T}}\mathbf{x}.

V-C3 FLC for comparison

By using feedback linearization, the following linearized system is obtained

z˙1\displaystyle\dot{z}_{1} =z2\displaystyle=z_{2}
z˙2\displaystyle\dot{z}_{2} =v\displaystyle=v (33)

where z1=x1,z2=x2+sinx2z_{1}=x_{1},z_{2}=x_{2}+\sin x_{2}, vv is the virtual input. Its state-space form is 𝐳˙=𝐀𝐳+𝐁v\mathbf{\dot{z}=}\mathbf{A}\mathbf{z}+\mathbf{B}v with

𝐀=[1000],𝐁=[01].\mathbf{A}=\left[\begin{array}[]{cc}1&0\\ 0&0\end{array}\right],\mathbf{B}=\left[\begin{array}[]{c}0\\ 1\end{array}\right].

Based on the linear system (33), a controller is designed as

v=𝐤T𝐳.v=\mathbf{k}^{\text{T}}\mathbf{z}.

Then

u=v1+cosx2+2x1+3x22x22.u=\frac{v}{1+\cos x_{2}}+2x_{1}+3x_{2}-2x_{2}^{2}.

A singularity problem may be encountered that the system becomes uncontrollable and the control input tends to infinity when x2=π±2kπ,k=1,2,3,x_{2}=-\pi\pm 2k\pi,k=1,2,3,···, i.e., cosx2=1\cos x_{2}=-1.

V-C4 Robust-Feedback-linearization-based control (RFLC) for comparison

By using robust feedback linearization, the following linearized system is obtained

z˙1\displaystyle\dot{z}_{1} =2z2\displaystyle=2z_{2}
z˙2\displaystyle\dot{z}_{2} =2z13z2+v\displaystyle=-2z_{1}-3z_{2}+v (34)

where z1=x1,z2=0.5x2+0.5sinx2z_{1}=x_{1},z_{2}=0.5x_{2}+0.5\sin x_{2}. Its state-space form is 𝐱˙=𝐀𝐱+𝐁v\mathbf{\dot{x}=}\mathbf{A}\mathbf{x}+\mathbf{B}v with

𝐀=[2023],𝐁=[01].\mathbf{A}=\left[\begin{array}[]{cc}2&0\\ -2&-3\end{array}\right],\mathbf{B}=\left[\begin{array}[]{c}0\\ 1\end{array}\right].

Based on the linear system (34), a controller is designed as

v=𝐤T𝐳.v=\mathbf{k}^{\text{T}}\mathbf{z}.

Then

u=2v1+cosx2+2x1+3x22x224x1+3x2+3sinx21+cosx2.u=\frac{2v}{1+\cos x_{2}}+2x_{1}+3x_{2}-2x_{2}^{2}-\frac{4x_{1}+3x_{2}+3\sin x_{2}}{1+\cos x_{2}}.

Note that, similarly to FLC, a singularity problem may be encountered.

V-C5 ADRC for comparison [27, 28]

By using ADRC, the following linearized system is obtained

x˙1\displaystyle\dot{x}_{1} =x2\displaystyle=x_{2}
x˙2\displaystyle\dot{x}_{2} =bu\displaystyle=bu (35)

where bb is the estimation of 1+cosx21+\cos x_{2}. Its state-space form is 𝐱˙=𝐀𝐱+𝐁u\mathbf{\dot{x}=}\mathbf{A}\mathbf{x}+\mathbf{B}u with

𝐀=[1000],𝐁=[0b].\mathbf{A}=\left[\begin{array}[]{cc}1&0\\ 0&0\end{array}\right],\mathbf{B}=\left[\begin{array}[]{c}0\\ b\end{array}\right].

The extended state is

x3(1+cosx2)(2x13x2+2x22)+(1+cosx2b)u.x_{3}\triangleq\left(1+\cos x_{2}\right)\left(-2x_{1}-3x_{2}+2x_{2}^{2}\right)+\left(1+\cos x_{2}-b\right)u.

An linear extended state observer (LESO) is designed as

x^˙1\displaystyle\dot{\hat{x}}_{1} =x^23ω0(x^1y)\displaystyle=\hat{x}_{2}-3\omega_{0}\left(\hat{x}_{1}-y\right)
x^˙2\displaystyle\dot{\hat{x}}_{2} =x^33ω02(x^1y)+bu\displaystyle=\hat{x}_{3}-3\omega_{0}^{2}\left(\hat{x}_{1}-y\right)+bu
x^˙3\displaystyle\dot{\hat{x}}_{3} =ω03(x^1y)\displaystyle=-\omega_{0}^{3}\left(\hat{x}_{1}-y\right)

where the parameter ω0\omega_{0} can be specified later. Then, a controller is designed as

u=b1x^3+u0(x^1,x^2)u=-b^{-1}\hat{x}_{3}+u_{0}\left(\hat{x}_{1},\hat{x}_{2}\right) (36)

where u0(x^1,x^2)u_{0}\left(\hat{x}_{1},\hat{x}_{2}\right) is designed based on (35) as

u0=𝐤T𝐱.u_{0}=\mathbf{k}^{\text{T}}\mathbf{x}.

It can be seen from (35) that bb must be reversible. However, in this example, bb is the estimation of 1+cosx21+\cos x_{2}, and thus bb is irreversible when x2=π±2kπ,k=1,2,3,x_{2}=-\pi\pm 2k\pi,k=1,2,3,···, i.e., cosx2=1\cos x_{2}=-1. It can be concluded that, similarly to FLC and RFLC, a singularity problem may also be encountered.

V-C6 Simulation results

The LQR method is utilized to determine the feedback matrix 𝐤\mathbf{k} for all the five control design methods and the parameters are all selected as 𝐐=diag(10,10)\mathbf{Q}=\text{diag}\left(10,10\right), 𝐑=1\mathbf{R}=1. Besides, let a=10a=10, c=10c=10, ω0=2\omega_{0}=2.

The following four scenarios are presented in the simulation to evaluate and compare the performance of our approach with other linearization methods.

(i) 𝐱0=[2 2]T\mathbf{x}_{0}=\left[2\ 2\right]^{\text{T}}, 𝐝=𝟎\mathbf{d=}\mathbf{0}.

(ii) 𝐱0=[5 5]T\mathbf{x}_{0}=\left[5\ 5\right]^{\text{T}}, 𝐝=𝟎\mathbf{d=}\mathbf{0}.

(iii) 𝐱0=[2 2]T\mathbf{x}_{0}=\left[2\ 2\right]^{\text{T}}, 𝐝=[1 1]T\mathbf{d=}\left[1\ 1\right]^{\text{T}}.

(iv) 𝐱0=[2 2]T\mathbf{x}_{0}=\left[2\ 2\right]^{\text{T}}, 𝐝=𝟎\mathbf{d=}\mathbf{0}, a time delay 0.2s0.2\text{s} is added into the input channel of the system (29).

Two commonly-used performance indices: the integral absolute error (IAE) and the integral of time-weighted absolute error (ITAE) are used to evaluate the performance of the above five methods quantitatively. Please refer to [29] to find the computation method of IAE and ITAE.

In Scenario (i), the corresponding state and input response is depicted in Fig. 6(a), which shows that all the five methods can achieve stabilization control. Note that the inputs of FLC and RFLC have large initial values. From Tab. I, it can be seen that RFLC has the smallest IAE and ITAE values, which is followed by SCLC in the second place. In Scenario (ii), when the initial state is increased to 𝐱0=[5,5]T\mathbf{x}_{0}=\left[5,5\right]^{\text{T}}, the corresponding state and input response is displayed in Fig. 6(b). JLC causes instability and is thus not included in the figure. It can be found that a singularity phenomenon is incurred by FLC, while SCLC still works well. RFLC still has large initial input values. In Scenario (iii), a constant disturbance is considered. From Fig. 6(c) and Tab. I, one can observe that ADRC has the best performance. SCLC also performs well. In Scenario (iv), RFLC causes instability because of the singularity problem and thus do not included in Fig. 6(d). Tab. I shows that SCLC has the best performance.

It can be concluded that, for system (29), JLC has a local property, and causes instability under large initial state values. FLC has a singularity problem and does not have enough robustness. In order to avoid the singularity condition, the working state of the system must be restricted. RFLC has better robustness than FLC. ADRC can tackle unknown disturbances better. However, RFLC and ADRC also have a singularity problem (Some other simulations have also be done where instability is incurred by the singularity problem in RFLC and ADRC, especially when the initial values are large). By contrast, SCLC avoids the local property of JLC, has better robustness than FLC, and avoids the singularity problem in FLC, RFLC, and ADRC. Besides, SCLC can accommodate the input time delay better. This is because all the input time delay is allocated into the primary system by using state compensation linearization. The secondary system deals with the known nonlinearity without the effect of the input time delay. The special strategy mitigates the effect of input time delay on the system. Thus, state compensation linearization can be complementary to the existing linearization methods. For some control problems, if the existing linearization methods do not work well, SCLC is worth trying. The simulation results verify the theoretical conclusion well.

TABLE I: Performance indices of the controllers
Sce. Index SCLC JLC FLC RFLC ADRC
(i) IAE 1.915 2.495 3.485 1.762 2.788
ITAE 1.055 1.851 3.842 0.947 4.719
(ii) IAE 5.026 - 12.438 3.695 5.409
ITAE 2.866 - 16.546 1.960 6.840
(iii) IAE 10.908 13.129 20.027 10.485 9.626
ITAE 48.440 57.575 95.280 47.286 36.893
(iv) IAE 2.324 2.906 2.765 - 3.466
ITAE 1.445 2.223 2.577 - 6.018
Refer to caption
Figure 6: State and input response.

VI Conclusions

By virtue of additive state decomposition, this paper proposes a new linearization process for control design, namely state compensation linearization. The nonlinear terms can be removed from the original system through the compensation at the input and state so that a linear primary system can be obtained. For some control problems like flight control design with challenging aircraft dynamics, the state compensation linearization may offer some advantages over traditional Jacobian linearization and feedback linearization by bridging the gap between linear control theories and nonlinear systems. Based on the proposed state compensation linearization, this paper proposes a control framework for a class of nonlinear systems so that their closed-loop behavior can follow a specified linear system. Three examples are provided to show the concrete design procedure. Simulation results have illustrated that state-compensation-linearization-based control outperforms Jacobian-linearization-based control and feedback-linearization-based control in some challenging cases. In future research, state-compensation-linearization-based stability margin analysis for nonlinear systems would be considered.

References

  • [1] D. M. Wolpert, R. C. Miall, and M. Kawato, “Internal models in the cerebellum,” Trends in Cognitive Sciences, vol. 2, no. 9, pp. 338–347, 1998.
  • [2] M. Seigler, S. Wang, S. Koushkbaghi, and J. B. Hoagg, “On the control strategies that humans use to interact with linear systems with output nonlinearities,” in 2018 IEEE International Conference on Systems, Man, and Cybernetics (SMC).   IEEE, 2018, pp. 3732–3737.
  • [3] H. K. Khalil, Nonlinear systems.   Upper Saddle River, NJ: Prentice hall, 2002.
  • [4] J.-J. E. Slotine and W. Li, Applied nonlinear control.   Englewood Cliffs, NJ: Prentice hall, 1991.
  • [5] B. Jakubczyk, W. Respondek et al., “On linearization of control systems,” Bull.acad.pol.sci.ser.sci.math, vol. 28, no. 9-10, pp. 517–522, 1980.
  • [6] C.-T. Chen, Introduction to linear system theory.   Holt, Rinehart and Winston, 1970.
  • [7] R. Chi, Y. Hui, B. Huang, and Z. Hou, “Adjacent-agent dynamic linearization-based iterative learning formation control,” IEEE Transactions on Cybernetics, vol. 50, no. 10, pp. 4358–4369, 2020.
  • [8] Q. Gan and C. J. Harris, “Linearization and state estimation of unknown discrete-time nonlinear dynamic systems using recurrent neurofuzzy networks,” IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), vol. 29, no. 6, pp. 802–817, 1999.
  • [9] X. Peng, K. Guo, X. Li, and Z. Geng, “Cooperative moving-target enclosing control for multiple nonholonomic vehicles using feedback linearization approach,” IEEE Transactions on Systems, Man, and Cybernetics: Systems, vol. 51, no. 8, pp. 4929–4935, 2021.
  • [10] A. Alasty and H. Salarieh, “Nonlinear feedback control of chaotic pendulum in presence of saturation effect,” Chaos, Solitons and Fractals, vol. 31, no. 2, pp. 292–304, 2007.
  • [11] H. Guillard and H. Bourles, “Robust feedback linearization,” in Proceedings of the 14th International Symposium on Mathematical Theory of Networks and Systems, 2000, pp. 1009–1014.
  • [12] J. de Jesús Rubio, “Robust feedback linearization for nonlinear processes control,” ISA transactions, vol. 74, pp. 155–164, 2018.
  • [13] A.-C. Huang and S.-Y. Yu, “Input-output feedback linearization control of uncertain systems using function approximation techniques,” in 2016 12th World Congress on Intelligent Control and Automation (WCICA).   IEEE, 2016, pp. 132–137.
  • [14] J. Umlauft, T. Beckers, M. Kimmel, and S. Hirche, “Feedback linearization using gaussian processes,” in 2017 IEEE 56th Annual Conference on Decision and Control (CDC).   IEEE, 2017, pp. 5249–5255.
  • [15] J. Han, “From PID to active disturbance rejection control,” IEEE Transactions on Industrial Electronics, vol. 56, no. 3, pp. 900–906, 2009.
  • [16] W.-H. Chen, J. Yang, L. Guo, and S. Li, “Disturbance-observer-based control and related methods-An overview,” IEEE Transactions on Industrial Electronics, vol. 63, no. 2, pp. 1083–1095, 2015.
  • [17] G. Qi, X. Li, and Z. Chen, “Problems of extended state observer and proposal of compensation function observer for unknown model and application in UAV,” IEEE Transactions on Systems, Man, and Cybernetics: Systems, vol. 52, no. 5, pp. 2899–2910, 2022.
  • [18] A. Liu, W.-A. Zhang, L. Yu, H. Yan, and R. Zhang, “Formation control of multiple mobile robots incorporating an extended state observer and distributed model predictive approach,” IEEE Transactions on Systems, Man, and Cybernetics: Systems, vol. 50, no. 11, pp. 4587–4597, 2020.
  • [19] Q. Quan, K.-Y. Cai, and H. Lin, “Additive-state-decomposition-based tracking control framework for a class of nonminimum phase systems with measurable nonlinearities and unknown disturbances,” International Journal of Robust and Nonlinear Control, vol. 25, no. 2, pp. 163–178, 2015.
  • [20] Z.-B. Wei, Q. Quan, and K.-Y. Cai, “Output feedback ILC for a class of nonminimum phase nonlinear systems with input saturation: An additive-state-decomposition-based method,” IEEE Transactions on Automatic Control, vol. 62, no. 1, pp. 502–508, 2016.
  • [21] Q. Quan, H. Lin, and K.-Y. Cai, “Output feedback tracking control by additive state decomposition for a class of uncertain systems,” International Journal of Systems Science, vol. 45, no. 9, pp. 1799–1813, 2014.
  • [22] Q. Quan and K.-Y. Cai, “Additive decomposition and its applications to internal-model-based tracking,” in Proceedings of the 48h IEEE Conference on Decision and Control (CDC) held jointly with 2009 28th Chinese Control Conference.   IEEE, 2009, pp. 817–822.
  • [23] R. A. Nichols, R. T. Reichert, and W. J. Rugh, “Gain scheduling for H-infinity controllers: A flight control example,” IEEE Transactions on Control Systems Technology, vol. 1, no. 2, pp. 69–79, 1993.
  • [24] D. J. Leith and W. E. Leithead, “Survey of gain-scheduling analysis and design,” International Journal of Control, vol. 73, no. 11, pp. 1001–1025, 2000.
  • [25] Y. Liu, J. J. Zhu, R. L. Williams II, and J. Wu, “Omni-directional mobile robot controller based on trajectory linearization,” Robotics and Autonomous Systems, vol. 56, no. 5, pp. 461–479, 2008.
  • [26] L. Tie, “On controllability and near-controllability of discrete-time inhomogeneous bilinear systems without drift,” Automatica, vol. 50, no. 7, pp. 1898–1908, 2014.
  • [27] Y. Huang and W. Xue, “Active disturbance rejection control: Methodology and theoretical analysis,” ISA transactions, vol. 53, no. 4, pp. 963–976, 2014.
  • [28] Z. Gao, “Active disturbance rejection control: A paradigm shift in feedback control system design,” in 2006 American Control Conference.   IEEE, 2006, pp. 2399–2405.
  • [29] L. Oliveira, A. Bento, V. J. Leite, and F. Gomide, “Evolving granular feedback linearization: Design, analysis, and applications,” Applied Soft Computing, vol. 86, p. 105927, 2020.