This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Conditions for Convergence of Dynamic Regressor Extension and Mixing Parameter Estimators Using LTI Filters

Bowen Yi and Romeo Ortega \IEEEmembershipLife Fellow, IEEE B. Yi is with Australian Centre for Field Robotics & Sydney Institute for Robotics and Intelligent Systems, The University of Sydney, Sydney, NSW 2006, Australia (email: bowen.yi@sydney.edu.au) R. Ortega is with Departamento Académico de Sistemas Digitales, ITAM, Ciudad de México, México and Department of Control Systems and Informatics, ITMO University, Saint Petersburg 197101, Russia (email: romeo.ortega@itam.mx)
Abstract

In this note we study the conditions for convergence of the recently introduced dynamic regressor extension and mixing (DREM) parameter estimator when the extended regressor is generated using LTI filters. In particular, we are interested in relating these conditions with the ones required for convergence of the classical gradient (or least squares), namely the well-known persistent excitation (PE) requirement on the original regressor vector, ϕ(t)q\phi(t)\in\mathbb{R}^{q}, with qq\in\mathbb{N} the number of unknown parameters. Moreover, we study the case when only interval excitation (IE) is available, under which DREM, concurrent and composite learning schemes ensure global convergence, being the convergence for DREM in finite time. Regarding PE we prove, under some mild technical assumptions, that if ϕ\phi is PE then the scalar regressor of DREM, ΔN\Delta_{N}\in\mathbb{R}, is also PE ensuring exponential convergence. Concerning IE we prove that if ϕ\phi is IE then ΔN\Delta_{N} is also IE. All these results are established in the almost sure sense, namely proving that the set of filter parameters for which the claims do not hold is of zero measure. The main technical tool used in our proof is inspired by a study of Luenberger observers for nonautonomous nonlinear systems recently reported in the literature.

{IEEEkeywords}

parameter estimation, system identification, adaptive control

1 Introduction

We consider in the paper the problem of online estimation of the unknown parameter vector θq\theta\in\mathbb{R}^{q} from the linear regression equation (LRE)

y(t)=ϕ(t)θ,y(t)=\phi^{\top}(t)\theta, (1)

where ϕ(t)q\phi(t)\in\mathbb{R}^{q} is bounded and, for simplicity, we assume y(t)y(t)\in\mathbb{R}. Here, we restrict ourselves to on-line recursive algorithms, which are attractive due to their robustness and direct applicability to adaptive control.

It is well-known [24] that the classical gradient estimator

θ^˙(t)=γϕ(t)[y(t)ϕ(t)θ^(t)],γ>0,\dot{\hat{\theta}}(t)=\gamma\phi(t)[y(t)-\phi^{\top}(t)\hat{\theta}(t)],\;\gamma>0,

ensures global exponential stability (GES) of the zero equilibrium of the associated linear time-varying (LTV) error equation

θ~˙(t)=γϕ(t)ϕ(t)θ~(t),\dot{\tilde{\theta}}(t)=-\gamma\phi(t)\phi^{\top}(t)\tilde{\theta}(t), (2)

with θ~(t):=θ^(t)θ{\tilde{\theta}}(t):=\hat{\theta}(t)-\theta, if and only if the regressor ϕ\phi is (T,δ)(T,\delta)-PE, that is, it satisfies

tt+Tϕ(s)ϕ(s)𝑑sδIq,\int_{t}^{t+T}\phi(s)\phi^{\top}(s)ds\geq\delta I_{q}, (3)

for some T>0,δ>0T>0,\;\delta>0 and all t0t\geq 0—where we underscore the uniformity in time requirement. Another property of the gradient estimator that follows directly from (2) is monotonicity of the norm of the parameter error, that is,

|θ~(tb)||θ~(ta)|,tbta0.|\tilde{\theta}(t_{b})|\leq|\tilde{\theta}(t_{a})|,\quad\forall t_{b}\geq t_{a}\geq 0. (4)

It has recently been shown [5, 12, 23] that global asymptotic stability (GAS)—but not exponential—of the error equation (2) is ensured under the strictly weaker condition of generalized PE. Namely,

τkτk+1ϕ(s)ϕ(s)𝑑sδkIq\int_{\tau_{k}}^{\tau_{k+1}}\phi(s)\phi^{\top}(s)ds\geq\delta_{k}I_{q}

where {δk}k\{\delta_{k}\}_{k\in\mathbb{N}} is a sequence of positive numbers, and {τk}k\{\tau_{k}\}_{k\in\mathbb{N}} is a strictly increasing sequence of positive times such that τk\tau_{k}\to\infty as kk\to\infty, together with a technical assumption of the relation between δk\delta_{k} and the integral of |ϕ(t)|2|\phi(t)|^{2}. It is widely accepted that both, the PE and the generalized PE conditions, are extremely restrictive, a situation that has motivated the development of new estimation algorithms that relax these assumptions. The interested reader is referred to [21] for a recent survey of this literature and [8] for a novel interesting algorithm. The main objective of this paper is to establish the connection with the classical PE requirement and the new condition for convergence of the recently introduced DREM estimator [ARAetaltac17].

Notation. We use \mathbb{C} to represent the complex plane, and >0\mathbb{C}_{>0} for the open right half-plane. For a complex-valued matrix An×mA\in\mathbb{C}^{n\times m}, A𝖧A^{\mathsf{H}} denotes the Hermitian transpose. Given a real-valued symmetric matrix Pn×nP\in\mathbb{R}^{n\times n}, λ𝚖𝚒𝚗(P)\lambda_{\tt min}(P) is its smallest eigenvalue. We use 𝟏n{\bf 1}_{n} to represent an nn-dimensional vector of ones.

2 Extended Regressor Equations and DREM

In this section we give the background material for the development of the DREM estimator.

2.1 Extended LRE

A key component of all the new modified estimators is the construction of an extended LRE (ELRE). This idea was first reported in [16] within the context of system identification and later used in [14] for adaptive observers and in [15] for adaptive controller designs. In both cases, the ELRE is created applying stable, linear filters to the LRE (1). Namely, we introduce a linear, bounded-input-bounded-output (BIBO), single-input \ell-output operator :{\cal H}:{\cal L}_{\infty}\to{\cal L}_{\infty}^{\ell} to define the ELRE

Y(t)=Φ(t)θ,Y(t)=\Phi(t)\theta, (5)

where Y(t):=[y](t)Y(t):={\cal H}[y](t)\in\mathbb{R}^{\ell} and the extended regressor matrix is defined as

Φ(t):=[[ϕ1](t)||[ϕq](t)]×q.\Phi(t):=[{\cal H}[\phi_{1}](t)~{}|~{}\ldots~{}|{\cal H}[\phi_{q}](t)]\in\mathbb{R}^{\ell\times q}. (6)

Applying a gradient-descent estimation to the ELRE (5) yields

θ^˙(t)=γΦ(t)[Y(t)Φ(t)θ^(t)],γ>0,\dot{\hat{\theta}}(t)=\gamma\Phi^{\top}(t)[Y(t)-\Phi(t)\hat{\theta}(t)],\;\gamma>0,

whose corresponding parameter estimation error equation is

θ~˙(t)=γΦ(t)Φ(t)θ~(t).\dot{\tilde{\theta}}(t)=-\gamma\Phi^{\top}(t)\Phi(t)\tilde{\theta}(t). (7)

Notice that, in contrast to (2), the matrix Φ(t)Φ(t)q×q\Phi^{\top}(t)\Phi(t)\in\mathbb{R}^{q\times q} is not necessarily of rank one. This is the central property that motivates the extension of the regressor. However, it is well known [17, Subsection 6.5.3(a.iv)] and [21, Proposition 2], that the provable stability properties of the error equation (7) are the same as the ones of (2), and still require the PE condition for GES.111The only provable advantage of the new estimator is that the convergence speed can be improved increasing γ\gamma. However, as discussed in [21, Remark 6], the interest of increasing the gain in adaptive systems is highly questionable.

Two different ways to generate the ELRE have been studied in the literature. In [16], they used =q\ell=q and the BIBO operator {\cal H} is obtained with the stable, linear time-invariant (LTI) filters

i(p)=λip+λi,iq¯:={1,,q},{\cal H}_{i}(p)={\lambda_{i}\over p+\lambda_{i}},\;i\in\bar{q}:=\{1,\ldots,q\}, (8)

with the differential operator p:=ddtp:={d\over dt}, λi>0\lambda_{i}>0 and λiλj\lambda_{i}\neq\lambda_{j} for all iji\neq j. We refer in the sequel to the ELRE as Lion’s (L-ELRE). Its state space realization can be written as

Φ˙\displaystyle\dot{\Phi} =ΛΦ+Λ[ϕϕ],\displaystyle=-\Lambda\Phi+\Lambda\begin{bmatrix}\phi^{\top}\\ \vdots\\ \phi^{\top}\end{bmatrix}, (9)
Y˙\displaystyle\dot{Y} =ΛY+Λcol(y,,y),\displaystyle=-\Lambda Y+\Lambda\mbox{col}(y,\ldots,y),

with Λ=diag(λ1,,λ)\Lambda=\mbox{diag}(\lambda_{1},\ldots,\lambda_{\ell}).

In [14, 15] they also consider =q\ell=q and the operator {\cal H} is LTV of the form222It is clear that i{\cal H}_{i} is an LTV operator of the form x˙i=Aixi+bi(t)u\dot{x}_{i}=A_{i}x_{i}+b_{i}(t)u with Ai=αA_{i}=-\alpha and bi(t)=ϕib_{i}(t)=\phi_{i}.

i(p,t)=1p+αϕi(t),iq¯,{\cal H}_{i}(p,t)={1\over p+\alpha}\phi_{i}(t),\;i\in\bar{q}, (10)

with α>0\alpha>0. A state space realization of Kreisselmeier’s ELRE—called K-ELRE—is given by

Φ˙(t)\displaystyle\dot{\Phi}(t) =αΦ(t)+ϕ(t)ϕ(t)\displaystyle=-\alpha\Phi(t)+\phi(t)\phi^{\top}(t) (11)
Y˙(t)\displaystyle\dot{Y}(t) =αY(t)+ϕ(t)y(t).\displaystyle=-\alpha Y(t)+\phi(t)y(t).

In recent years, ELREs have been used to ease the PE requirement in novel estimator schemes—see the recent survey in [21] for more details and a complete list of references. Two examples of these estimators are the concurrent [10] and composite learning [22]. Within these methodologies, a dynamic data stack is built to discretely record online historical data, and the convergence of parameter estimation is managed monitoring the excitation over an interval. That is, the PE condition (3) is replaced by the strictly weaker assumption that the regressor is IE, whose definition is given as follows.

Definition 1

A bounded signal ϕq\phi\in\mathbb{R}^{q} is (t0,tc,μ)(t_{0},t_{c},\mu)-IE if there exist t00t_{0}\geq 0 and tc>0t_{c}>0 such that

t0t0+tcϕ(s)ϕ(s)𝑑sμIq\int_{t_{0}}^{t_{0}+t_{c}}\phi(s)\phi^{\top}(s)ds\geq\mu I_{q} (12)

for some μ>0\mu>0.

The IE condition is called “exciting over a finite time interval” in [25, Definition 3.1, pp. 108], where it was used to analyze some convergence properties of the estimation error in the gradient algorithm.

2.2 DREM estimator

A new estimator that has attracted a lot of attention, and has proven to be very successful to solve many theoretical and practical open problems is DREM, first proposed in [ARAetaltac17] and recently reviewed in [19]. The main idea of DREM is to generate, out of the ELRE (5), qq scalar LREs. Towards this end, we also fix =q\ell=q and then introduce the key mixing step of multiplying from the left (5) by the adjugate of the (square) matrix Φ(t)\Phi(t), denoted adj{Φ(t)}\mbox{adj}\{\Phi(t)\}, to get

𝒴i(t)=Δ(t)θi,iq¯,{\cal Y}_{i}(t)=\Delta(t)\theta_{i},\;i\in\bar{q}, (13)

where 𝒴(t):=adj{Φ(t)}Y(t){\cal Y}(t):=\mbox{adj}\{\Phi(t)\}Y(t) and

The DREM design is completed with the qq scalar estimators

θ^˙i(t)=γiΔ(t)[𝒴i(t)Δ(t)θ^i(t)],γi>0,iq¯,\dot{\hat{\theta}}_{i}(t)=\gamma_{i}\Delta(t)[{\cal Y}_{i}(t)-\Delta(t)\hat{\theta}_{i}(t)],\;\gamma_{i}>0,\;i\in\bar{q}, (14)

with associated error equations

θ~˙i(t)=γiΔ2(t)θ~i(t),iq¯,\dot{\tilde{\theta}}_{i}(t)=-\gamma_{i}\Delta^{2}(t)\tilde{\theta}_{i}(t),\;i\in\bar{q}, (15)

for which the following proposition can be easily proved [ARAetaltac17, 19].

Proposition 1

The systems (15) enjoy the following feature.

  1. F1

    The origin is GAS Δ2\;\Longleftrightarrow\;\Delta\notin{\cal L}_{2}.

  2. F2

    The origin is GES Δ\;\Longleftrightarrow\;\Delta is PE.

  3. F3

    For all tbta0t_{b}\geq t_{a}\geq 0 we have |θ~i(tb)||θ~i(ta)|,iq¯|\tilde{\theta}_{i}(t_{b})|\leq|\tilde{\theta}_{i}(t_{a})|,\;i\in\bar{q}.

Moreover, a variation of DREM that converges in finite time under the weaker IE assumption (12) has been recently reported in [20] and, as discussed in [19], it has proven instrumental to solve many practical and theoretical open problems.

Note that in DREM we have replaced the convergence condition on the vector ϕ\phi being PE by a condition—either non-square integrability for GAS or PE for GES—on a scalar quantity Δ(t)\Delta(t), which is the determinant of the extended regressor matrix (2.2). A natural question that arises is the relationship between the excitation properties of the original regressor ϕ(t)\phi(t) and the new scalar regressor Δ(t)\Delta(t). This question has been recently answered in [ARAetaltac20] for the case of K-ELRE where the following results are proven.

Proposition 2

Consider the extended regressor matrix (6) generated via (10), and its determinant (2.2).

  1. C1

    ϕ\phi is PE Δ\;\Longleftrightarrow\;\Delta is PE.

  2. C2

    ϕ\phi is (t0,tc,μ)(t_{0},t_{c},\mu)-IE Δ\;\Longrightarrow\;\Delta is also (t0,tc,μ)(t_{0},t_{c},\mu)-IE.

The results above prove that, in a scenario with suitable excitation, DREM with K-ELRE has the same convergence properties as the standard gradient, with the following additional advantages:

  1. (i)

    GAS under the non-square integrability condition of the scalar regressor Δ\Delta that, as shown in [19, Proposition 3] is strictly weaker than PE of the regressor ϕ\phi;

  2. (ii)

    element-by-element monotonicity of the parameter errors, which is strictly stronger than (4);

  3. (iii)

    ability to tune, via γi\gamma_{i}, the convergence rate of each parameter error, in an independent way;

  4. (iv)

    possibility to ensure finite convergence time under IE [20, Proposition 3] without the injection of high-gain.

The main objective of this paper is to prove a similar result for DREM with L-ELRE.333In [ARAetalijacsp18] this question was studied for the particular case of systems identification, when the regressor is generated via LTI filtering of a sum of sinusoidal signals. In this way we conclusively establish the superiority of DREM—in either one of its forms, K- or L-ELRE—over classical estimators. Instrumental to establish our results is to adopt a Kazantzis-Kravaris-Luenberger (KKL) observer perspective of the DREM estimator, as done in [18]. In this way, we can invoke a fundamental result on injectivity of the key mapping of KKL observers for nonautonomous nonlinear systems recently reported in [6]. This result extends to the nonautonomous case the previous results of [ANDPRA] for autonomous systems, allowing then to include the study at hand.

3 Two Minor Modifications to the DREM Estimator

To establish our results we introduce two slight modifications to the procedure described above. First, we do not select the number of filters equal to the dimension of the parameter vector, instead we set444This choice is made for simplicity, the results being true for any >q\ell>q.

=q+1.\ell=q+1. (16)

Moreover, in order to use the result of [11] in the proof of our main claim, we allow λi\lambda_{i} to be in >0\mathbb{C}_{>0}. This modification is similar to the procedure used in the design of KKL observers, first proposed in [13] and intensively studied in [ANDPRA, 7] where—to ensure injectivity of a key mapping—the number of LTI filters is selected larger than the dimension of the systems state. Such an approach was suggested in [18, Section 5], where the DREM estimator is revisited as a KKL observer for the LTV system hal-03245139)

  • [5] N. Barabanov and R. Ortega, On global asymptotic stability of x˙=ϕ(t)ϕ(t)x\dot{x}=\phi(t)\phi^{\top}(t)x with ϕ(t)\phi(t) bounded and not persistently exciting, Systems & Control Letters, vol. 109, pp. 24-27, 2017.
  • [6] P. Bernard and V. Andrieu, Luenberger observers for nonautonomous nonlinear systems, IEEE Trans. Automatic Control, vol. 64, pp. 270-281, 2019.
  • [7] P. Bernard, Observer Design for Nonlinear Systems, Springer, vol. 479, 2019.
  • [8] N. M. Boffi and J. J. E. Slotine, Implicit regularization and momentum algorithms in nonlinearly parameterized adaptive control and prediction, Neural Computation, vol. 33, no. 3, pp. 590–673, 2021.
  • [9] A. Bobtsov, B. Yi, R. Ortega and A. Astolfi, Generation of new exciting regressors for consistent on-line estimation for a scalar parameter, ArXiv Preprint, 2021. (arXiv:2104.02210)
  • [10] G. Chowdhary, T. Yucelen, M. Mhlegg and E. Johnson, Concurrent learning adaptive control of linear systems with exponentially convergent bounds, Int. J. on Adaptive Control and Signal Processing, vol. 27, no. 4, pp. 280-301, 2013.
  • [11] J.M. Coron, On the stabilization of controllable and observable systems by output feedback law, Mathematics of Control, Signals and Systems, vol. 7, pp. 187-216, 1994.
  • [12] D. Efimov, N. Barabanov and R. Ortega, Robustness of linear time-varying systems with relaxed excitation, Int. J. on Adaptive Control and Signal Processing, vol. 33, no. 12, pp. 1885-1900, 2019.
  • [13] N. Kazantzis and C. Kravaris, Nonlinear observer design using Lyapunov’s auxiliary theorem, Systems & Control Letters, vol. 34, no. 5, pp. 241-247, 1998.
  • [14] G. Kreisselmeier, Adaptive observers with exponential rate of convergence, IEEE Trans. Automatic Control, vol. 22, no. 1, pp. 2-8, 1977.
  • [15] G. Kreisselmeier and D. Joos, Rate of convergence in model reference adaptive control, IEEE Trans. Automatic Control, vol. 27, pp. 710-713, 1982.
  • [16] P.M. Lion, Rapid identification of linear and nonlinear systems, AIAA Journal, vol.5, pp. 1835-1842, 1967.
  • [17] K. Narendra and A. Annaswamy, Stable Adaptive Systems, Prentice Hall, New Jersey, 1989.
  • [18] R. Ortega, L. Praly, S. Aranovskiy, B. Yi and W. Zhang, On dynamic regressor extension and mixing parameter estimators: Two Luenberger observers interpretations, Automatica, vol. 95, pp. 548-551, 2018.
  • [19] R. Ortega, S. Aranovskiy, A. Pyrkin, A Astolfi and A. Bobtsov, New results on parameter estimation via dynamic regressor extension and mixing: Continuous and discrete-time cases, IEEE Trans. Automatic Control, vol. 66, pp. 2265–2272, 2021.
  • [20] R. Ortega, D. Gerasimov, N. Barabanov and V. Nikiforov, Adaptive control of linear multivariable systems using dynamic regressor extension and mixing estimators: Removing the high-frequency gain assumption, Automatica, vol. 110, 108589, pp. 1-8, 2019.
  • [21] R. Ortega, V. Nikiforov and D. Gerasimov, On modified parameter estimators for identification and adaptive control: A unified framework and some new schemes, Annual Reviews in Control, vol. 50, pp. 278–293, 2020.
  • [22] Y. Pan and H. Yu, Composite learning robot control with guaranteed parameter convergence, Automatica, vol. 89, pp. 398-406, 2018.
  • [23] L. Praly, Convergence of the gradient algorithm for linear regression models in the continuous and discrete-time cases, Int. Rep. MINES ParisTech, Centre Automatique et Systèmes, December 26, 2017.
  • [24] S. Sastry and M. Bodson, Adaptive Control: Stability, Convergence and Robustness, Prentice-Hall, New Jersey, 1989.
  • [25] G. Tao, Adaptive Control Design and Analysis, John Wiley & Sons: New Jersey, 2003.