This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Anytime Control under Practical Communication Model

Wanchun Liu Daniel E. Quevedo Yonghui Li and Branka Vucetic
Abstract
W. Liu, Y. Li and B. Vucetic are with School of Electrical and Information Engineering, The University of Sydney, Australia. Emails: {wanchun.liu, yonghui.li, branka.vucetic}@sydney.edu.au. D. E. Quevedo is with the School of Electrical Engineering and Robotics, Queensland University of Technology (QUT), Brisbane, Australia. Email: daniel.quevedo@qut.edu.au.

We investigate a novel anytime control algorithm for wireless networked control with random dropouts. The controller computes sequences of tentative future control commands using time-varying (Markovian) computational resources. The sensor-controller and controller-actuator channel states are spatial- and time-correlated, and are modeled as a multi-state Markov process. To compensate the effect of packet dropouts, a dual-buffer mechanism is proposed. We develop a novel cycle-cost-based approach to obtain the stability conditions on the nonlinear plant, controller, network and computational resources.

Index Terms:
Control over communications, nonlinear systems, stability of nonlinear systems, Markov fading channels.

I Introduction

During past decades, significant attention has focused on embedded or networked control systems that have limited and time-varying controller’s computation capability due to high requirements on multitasking operations. In particular, assuming constant and limited computational resources, bounds on computational time of specific optimization algorithms for achieving stability were derived in [1, 2]. For time-varying computational resources, a dynamic computation task scheduling method was proposed for model predictive controllers [3]. On-demand computation scheduling of control input based on plant states were investigated for periodic, event-triggered and self-triggered policies in [4, 5, 6], respectively.

Another stream of research considers anytime algorithms for robust control and making efficient use of time-varying computational resources. In general, an anytime algorithm can provide a solution even with limited computational resources, and refines the solution when more resources are available. In the pioneering work [7], an anytime control system was proposed, where the number of updated states varies with the available computation time known to the controller a priori. In [8], an anytime control algorithm for a multi-input linear system was proposed for the scenario when the computation availability is unknown a priori. The main idea was to first calculate the most important component of the control vector and then calculate the less important ones as more computational resource becomes available. In [9], a sequence-based anytime control method was proposed, which can calculate a tentative sequence of future control input for as many time steps as allowed by the available computational resources at each time step. The pre-calculated control sequence can compensate for the time steps when no computational resource is available for control calculations. Following this work, sequence-based anytime control systems with Markovian processor availability, event-triggered sensor updates and multiple control laws were investigated in [10, 11, 12], respectively. In [7, 8, 9, 10], the sensor-controller and controller-actuator channels were assumed to be perfect and error-free. In [11] and [12], the sensor-controller channel was assumed to have independent and identically distributed (i.i.d.) packet dropouts222Note that i.i.d. packet dropout channel is very commonly considered in the literature of networked control [13]., i.e., only binary-level (on-off) channel states, were considered, while the controller-sensor channel was assumed to be perfect.

Motivation. The existing work of anytime control [7, 8, 9, 10, 11, 12] cannot effectively handle the fully distributed networked control scenario, where both the sensor-controller and controller-actuator communication links are wireless. Moreover, the existing research only considered simple wireless channel models, which cannot capture the key features of practical wireless channels that are time-varying and correlated [14]. Therefore, anytime control design and analysis for fully distributed networked control system in practical wireless channels are critical in practice, but also present new challenges.

Novelty and contributions. In this work, we consider for the first time the sequence-based anytime control of a wireless networked control system (WNCS) in a generalized dual imperfect channel, where the sensor-controller and controller-actuator channel states are spatial- and time-correlated and are modeled as a multi-state Markov process. Different from most of the existing works [9, 10, 11, 12], where only the controller has a buffer to keep the calculated sequence of control inputs, we propose to use anytime control with buffers at both the controller and the actuator nodes. The latter is used to compensate for dropouts in the controller-actuator channel. Moreover, the available computational resource of the controller is allowed to be time-correlated, modeled as a multi-state Markov process. Such a dual-channel-dual-buffer anytime control system has practical advantages but brings significant challenges to its analysis due to the complex system state updating rule, compared to previous setups. We propose a novel cycle-cost-based approach to derive sufficient conditions for stochastic stability of the overall WNCS. Our stability conditions are stated in terms of plant dynamics, network dynamics, buffer properties and computational resource dynamics. We further show that, under suitable assumptions, the conditions derived guarantee robust stability when plant disturbances are taken into account.

The remainder of the paper is organized as follows: Section II presents the system model of anytime control in the dual-channel-dual-buffer WNCS. Section III develops the stability condition. Section IV provides robust stability analysis. Section V draws conclusions.

Notation: Sets are denoted by calligraphic capital letters, e.g., 𝒜\mathcal{A}. 𝒜\\mathcal{A}\backslash\mathcal{B} denotes set subtraction. Matrices and vectors are denoted by capital and lowercase upright bold letters, e.g., 𝐀\mathbf{A} and 𝐚\mathbf{a}, respectively. 𝖤[A]\mathsf{E}\left[A\right] is the expectation of the random variable AA. The conditional probability 𝖯𝗋𝗈𝖻[A|B]=0\mathsf{Prob}\left[{A}|{B}\right]=0 if 𝖯𝗋𝗈𝖻[B]=0\mathsf{Prob}\left[{B}\right]=0. ()(\cdot)^{\top} is the matrix transpose operator. |𝐯||\mathbf{v}| is the Euclidean norm of vector 𝐯\mathbf{v}. \mathbb{N} and 0\mathbb{N}_{0} denote the sets of positive and non-negative integers, respectively. m\mathbb{R}^{m} denotes the mm-dimensional Euclidean space. [𝐀]j,k[\mathbf{A}]_{j,k} and [𝐯]j[\mathbf{v}]_{j} denote the element at the jjth row and kkth column of a matrix 𝐀\mathbf{A}, and the jjth element of a vector 𝐯\mathbf{v}, respectively. λmax(𝐀)\lambda_{\max}(\mathbf{A}) denotes the spectral radius of 𝐀\mathbf{A}. diag{𝐯}\text{diag}\{\mathbf{v}\} denotes the diagonal matrix generated by the vector 𝐯\mathbf{v}. {v}0\{v\}_{\mathbb{N}_{0}} denotes the semi-infinite sequence {v0,v1,}\{v_{0},v_{1},\cdots\}. A function ϕ:00\phi:\mathbb{R}_{\geq 0}\rightarrow\mathbb{R}_{\geq 0} is of class-𝒦\mathcal{K}_{\infty} (ϕ𝒦\phi\in\mathcal{K}_{\infty}) if it is continuous, strictly increasing, and zero at zero. 𝟎i\mathbf{0}_{i} and 𝟎i×j\mathbf{0}_{i\times j} denotes the all-zero i×ii\times i and i×ji\times j matrices, respectively. 𝐀=𝟎\mathbf{A}=\mathbf{0} indicates that 𝐀\mathbf{A} has all zero elements.

II Anytime Control in a Dual-Channel-Dual-Buffer WNCS

We consider a WNCS consisting of a plant system, a remote controller and a wireless network placed between the plant and the controller. A sensor sends its measurements of the plant to the controller, and the controller computes and sends control commands to a remote actuator via the wireless network as illustrated in Fig. 1. The dual wireless channel (i.e., the sensor-controller and controller-actuator channels) setup is different from [10, 11, 9], which assumed either perfect transmissions in two channels or a single (imperfect) wireless channel from the sensor to the controller.

Refer to caption
Figure 1: The dual-channel-dual-buffer WNCS.

Each sampling period of the plant, denoted by TsT_{s}, is divided into four phases: sensor-controller (S-C) transmission, command computation, controller-actuator (C-A) transmission and implementation of control as illustrated in Fig. 2.

Refer to caption
Figure 2: A communications-computing-control process.

We consider a process-noise-free discrete-time non-linear plant model

𝐱(t+1)=f(𝐱(t),𝐮(t)),\mathbf{x}(t+1)=f(\mathbf{x}(t),\mathbf{u}(t)), (1)

where 𝐱(t)ls\mathbf{x}(t)\in\mathbb{R}^{l_{s}} and 𝐮(t)lu\mathbf{u}(t)\in\mathbb{R}^{l_{u}} are the plant state and the control input at time tt. Note that a more practical model with process noise will be investigated in Section IV.

As a consequence of stochastic computational resources at the controller and packet dropouts in the sensing and control channels (which will be described later in more detail), the plant may have to operate in open loop for arbitrarily long time intervals. This may lead to performance degradation and potential loss of stability. Thus, throughout this work we will analyze the stability conditions of the considered system where the stability is defined as follows.

Definition 1.

The process-noise-free dynamical system (1) is stochastically stable, if for some ψ𝒦\psi\in\mathcal{K}_{\infty}, the expected value k=0𝖤[ψ(|𝐱(k)|)]<\sum_{k=0}^{\infty}\mathsf{E}\left[\psi(|\mathbf{x}(k)|)\right]<\infty.

Our standing assumption is that the plant is globally controllable (in the idealized closed loop case):

Assumption 1 (​​[10, 11, 9]).

There exist functions V:ls0,ψ1,ψ2𝒦V:\mathbb{R}^{l_{s}}\rightarrow\mathbb{R}_{\geq 0},\psi_{1},\psi_{2}\in\mathcal{K}_{\infty}, a constant ρ(0,1)\rho\in(0,1), and a control policy κ:lslu\kappa:\mathbb{R}^{l_{s}}\rightarrow\mathbb{R}^{l_{u}}, such that for all xlsx\in\mathbb{R}^{l_{s}}

ψ1(|𝐱|)V(𝐱)ψ2(|𝐱|)\displaystyle\psi_{1}(|\mathbf{x}|)\leq V(\mathbf{x})\leq\psi_{2}(|\mathbf{x}|)
V(f(𝐱,κ(𝐱)))ρV(𝐱).\displaystyle V(f(\mathbf{x},\kappa(\mathbf{x})))\leq\rho V(\mathbf{x}).
Assumption 2 (​​[10, 11, 9]).

There exists α>0\alpha>0 such that

V(f(𝐱,𝟎))<αV(𝐱),𝐱ls.V(f(\mathbf{x},\mathbf{0}))<\alpha V(\mathbf{x}),\forall\mathbf{x}\in\mathbb{R}^{l_{s}}.

The initial plant state satisfies

𝖤[ψ2(|𝐱(0)|)]<,\mathsf{E}\left[\psi_{2}(|\mathbf{x}(0)|)\right]<\infty,

where ψ2𝒦\psi_{2}\in\mathcal{K}_{\infty} as in Assumption 1.

II-A Dual Markov Fading Channels

We consider wireless fading channels for the S-C and C-A transmissions [15, 16, 17], where wireless channel status varies with time due to multi-path propagation and shadowing caused by obstacles affecting the radio-frequency (RF) wave propagation. The time-varying channel conditions can be modeled as Markov processes [14]. Furthermore, in practice, the two channel conditions can be correlated, named as spatial correlation, which is caused by the same environment obstacles [18].

For wireless packet transmissions, there exists a fundamental tradeoff between reliability, i.e., the packet drop probability, and data rate, which determines the amount of information bits that a packet can carry [19]. For a fixed channel condition, increasing the data rate of a packet can lead to a higher packet drop probability. For a good channel condition, one can reduce the packet drop probability while maintain the fixed data rate. Alternatively, one can increase the data rate while maintaining the packet drop probability [20]. In this work, we will keep the data rate fixed for the S-C transmissions, since there is nothing gained from aggregating past sensor measurements in the state feedback case. However, it is well known that sending control sequences can be beneficial to compensate for packet dropouts. To accommodate this in a fading channel environment, we allow C-A transmissions to contain packets of varying data rate. The rates depend on the channel condition and provide a guaranteed packet-dropout probability. Thus, a longer control sequence can be transmitted to the actuator under a better channel condition with the same reliability.

C-A channel. At time slot tt, the controller can at most transmit B(t){0,1,,B¯}B(t)\in\mathcal{B}\triangleq\{0,1,\cdots,\bar{B}\} commands to the actuator with a guaranteed packet drop probability γ¯\bar{\gamma}. In other words, B(t)B(t) denotes the C-A channel quality. In this sense, B(t)B(t) can be treated as the capacity of the channel under the packet drop probability requirement γ¯\bar{\gamma}. Let γ(t)=1\gamma(t)=1 and γ(t)=0\gamma(t)=0 denote the successful and failed transmissions in time slot tt.

S-C channel. At time slot tt, the S-C channel power gain takes values from {h1,h2,,hB¯}\{h_{1},h_{2},\cdots,h_{\bar{B}^{\prime}}\}. Let B(t){1,2,,B¯}B^{\prime}(t)\in\mathcal{B}^{\prime}\triangleq\{1,2,\cdots,\bar{B}^{\prime}\} denote the index of the channel power gain. Thus, B(t)B^{\prime}(t) denotes the S-C channel quality. Let γ(t)=1\gamma^{\prime}(t)=1 and γ(t)=0\gamma^{\prime}(t)=0 denote the successful and failed transmissions in time slot tt. The packet drop probability at time tt is

γ¯(t)=g(hB(t)){g(h1),g(h2),,g(hB¯)},\bar{\gamma}^{\prime}(t)=g(h_{B^{\prime}(t)})\in\{g(h_{1}),g(h_{2}),\cdots,g(h_{\bar{B}^{\prime}})\}, (2)

where g()g(\cdot) is the packet drop probability function in terms of the channel power gain.

Then, we assume that the joint C-A and S-C channel condition {(B,B)}\{(B,B^{\prime})\}_{\mathbb{N}} is a time-homogeneous Markov process, and the state transition probability is given as

pi,j\displaystyle p_{i,j} 𝖯𝗋𝗈𝖻[(B(t+1),B(t+1))=bj|(B(t),B(t))=bi],\displaystyle\triangleq\mathsf{Prob}\left[(B(t+1),B^{\prime}(t+1))=b_{j}|(B(t),B^{\prime}(t))=b_{i}\right], (3)
bi,bj×.\displaystyle\hskip 165.02606pt\forall b_{i},b_{j}\in\mathcal{B}\times\mathcal{B}^{\prime}.
Remark 1.

Our current channel model jointly considers both the spatial-correlated S-C and C-A channels, the time-correlated fading channel conditions and variable data rate requirements. To the best of our knowledge, this has never been considered in the literature of WNCSs and is more general than existing models. For example, independent dual channels with i.i.d. packet dropouts were considered in [21], which is a special case of our model when the fading channels degrade to static ones and the channels’ spacial correlation is perfectly canceled.

II-B Anytime Control with Dual Buffers

When considering perfect transmission between the controller and the actuator, as in [10, 11, 9], the system only needs one buffer at the controller to store the computed control commands. If imperfect transmissions are taken into account, it is convenient to include a command buffer at the actuator to provide robustness against packet dropouts, see e.g. [22] for a general packetized predictive control method333Markovian communication and computational resources were not considered in [22].. Clearly, the dual-buffer system introduces a more complex state updating process.

Let Λc\Lambda_{c} and Λa\Lambda_{a} denote the length of the controller’s and the actuator’s buffer, respectively. Then, the buffer state at the controller after its transmission phase is denoted as

𝐛c(t)[𝐛c,1(t),𝐛c,2(t),,𝐛c,Λc(t)],\vec{\mathbf{b}}_{c}(t)\triangleq[\mathbf{b}_{c,1}(t)^{\top},\mathbf{b}_{c,2}(t)^{\top},\cdots,\mathbf{b}_{c,\Lambda_{c}}(t)^{\top}]^{\top}, (4)

where 𝐛c,i(t)lu,i{1,2,,Λc}\mathbf{b}_{c,i}(t)\in\mathbb{R}^{l_{u}},\forall i\in\{1,2,\cdots,\Lambda_{c}\}. The buffer state at the actuator right after the C-A transmission but before the implementation of a control command, i.e., the pre-control buffer, is denoted as

𝐛a(t)[𝐛a,1(t),𝐛a,2(t),,𝐛a,Λa(t)],\vec{\mathbf{b}}_{a}(t)\triangleq[\mathbf{b}_{a,1}(t)^{\top},\mathbf{b}_{a,2}(t)^{\top},\cdots,\mathbf{b}_{a,\Lambda_{a}}(t)^{\top}]^{\top}, (5)

where 𝐛a,i(t)lu,i{1,2,,Λa}\mathbf{b}_{a,i}(t)\in\mathbb{R}^{l_{u}},\forall i\in\{1,2,\cdots,\Lambda_{a}\}. In general, the buffers 𝐛a(t)\vec{\mathbf{b}}_{a}(t) and 𝐛c(t)\vec{\mathbf{b}}_{c}(t) keep the calculated sequences of control command, and the buffer updating rules will be given in the following part.

The control input is the first element in the actuator’s buffer, i.e.,

𝐮(t)=𝐛a,1(t),\mathbf{u}(t)=\mathbf{b}_{a,1}(t), (6)

which can be treated as the previously predicted control command for the current time slot. The buffer state at the actuator right after the control implementation, i.e., the post-control state, is

𝐛a(t)𝐒a𝐛a(t),\vec{\mathbf{b}}^{\prime}_{a}(t)\triangleq\mathbf{S}_{a}\vec{\mathbf{b}}_{a}(t), (7)

where the buffer shift matrices are defined as

𝐒i[𝟎lu𝐈lu𝟎lu𝟎lu𝟎lu𝟎lu𝐈lu𝟎lu𝟎lu𝟎lu𝐈lu𝟎lu𝟎lu]Λilu×Λilu,i=c or a.\color[rgb]{0,0,0}\mathbf{S}_{i}\triangleq\begin{bmatrix}\mathbf{0}_{l_{u}}&\mathbf{I}_{l_{u}}&\mathbf{0}_{l_{u}}&\cdots&\mathbf{0}_{l_{u}}\\ \vdots&\ddots&\ddots&\ddots&\vdots\\ \mathbf{0}_{l_{u}}&\cdots&\mathbf{0}_{l_{u}}&\mathbf{I}_{l_{u}}&\mathbf{0}_{l_{u}}\\ \mathbf{0}_{l_{u}}&\cdots&\cdots&\mathbf{0}_{l_{u}}&\mathbf{I}_{l_{u}}\\ \mathbf{0}_{l_{u}}&\cdots&\cdots&\cdots&\mathbf{0}_{l_{u}}\\ \end{bmatrix}\in\mathbb{R}^{\Lambda_{i}l_{u}\times\Lambda_{i}l_{u}},i=c\text{ or }a. (8)

Let N(t)N(t) denote the number of calculated tentative future control commands at time tt. The process {N}\{N\}_{\mathbb{N}} is a time-homogeneous Markov process with the transition probability

qi,j𝖯𝗋𝗈𝖻[N(t+1)=j|N(t)=i],i,j𝒩,q_{i,j}\triangleq\mathsf{Prob}\left[N(t+1)=j|N(t)=i\right],i,j\in\mathcal{N}, (9)

where 𝒩{0,1,,N¯}\mathcal{N}\triangleq\{0,1,\cdots,\bar{N}\}. It is assumed that N¯Λc\bar{N}\leq\Lambda_{c}.

The controller’s operations are described as

  1. 1.

    If γ(t)N(t)>0\gamma^{\prime}(t)N(t)>0, the controller has a new update from the sensor and is available for computation. In this case, it discards all the existing commands in its buffer and generates a sequence of N(t)N(t) control commands to control the plant in time slots tt to (t+N(t)1)(t+N(t)-1). The sequence of tentative controls is

    𝐮(t)=[𝐮1(t),𝐮2(t),,𝐮N(t)(t)].\vec{\mathbf{u}}(t)=[\mathbf{u}_{1}(t)^{\top},\mathbf{u}_{2}(t)^{\top},\cdots,\mathbf{u}_{N(t)}(t)^{\top}]^{\top}. (10)

    The buffer state before transmission is written as

    [𝐮(t),𝟎lu×1,,𝟎lu×1ΛcN(t)].[\vec{\mathbf{u}}(t)^{\top},\underbrace{\mathbf{0}_{l_{u}\times 1}^{\top},\cdots,\mathbf{0}_{l_{u}\times 1}^{\top}}_{\Lambda_{c}-N(t)}]^{\top}.

    Specifically, the controller calculates the control sequence based on the anytime control algorithm proposed in [9], which is rewritten as

    𝐮i(t)=κ(𝐱i(t)),i=1,,N(t)\displaystyle\mathbf{u}_{i}(t)=\kappa(\mathbf{x}_{i}^{\prime}(t)),i=1,\cdots,N(t) (11)
    𝐱i(t)={𝐱(t),i=1f(𝐱i1(t),𝐮i1(t)),i=2,,N(t)\displaystyle\mathbf{x}^{\prime}_{i}(t)=\begin{cases}\mathbf{x}(t),&i=1\\ f(\mathbf{x}^{\prime}_{i-1}(t),\mathbf{u}_{i-1}(t)),&i=2,\cdots,N(t)\end{cases}

    where κ()\kappa(\cdot) is defined in Assumption 1.

    Considering the C-A channel capacity and the actuator’s buffer length, the controller transmits min{B(t),N(t),Λa}\min\{B(t),N(t),\Lambda_{a}\} commands to the actuator. If the transmission is successful, the buffer shifts by min{B(t),N(t),Λa}\min\{B(t),N(t),\Lambda_{a}\} steps. Otherwise, the controller erases its buffer. This is because the first computed control command 𝐮1(t)\mathbf{u}_{1}(t) cannot be implemented in the current time slot tt, and the rest of computed control commands, which are calculated based on the successful implementation of the first control command, become useless. It is clear that at time instances where the actuator has run out of buffer contents, we have 𝐛a(t)=𝟎\vec{\mathbf{b}}_{a}(t)=\mathbf{0} and 𝐮(t)=𝟎𝐮1(t)\mathbf{u}(t)=\mathbf{0}\neq\mathbf{u}_{1}(t). For the case that 𝐛a(t)𝟎\vec{\mathbf{b}}_{a}(t)\neq\mathbf{0}, the predicted control input 𝐮(t)\mathbf{u}(t) is equal to currently calculated control command 𝐮1(t)\mathbf{u}_{1}(t) only in the perfect process-noise-free scenario, and 𝐮(t)𝐮1(t)\mathbf{u}(t)\neq\mathbf{u}_{1}(t) in general.

  2. 2.

    If γ(t)=0\gamma^{\prime}(t)=0, the controller does not have a new update from the sensor. In this case, it does not generate any new control command. If N(t)=0N(t)=0, the controller does not have the computational resource to generate any new control command.
    In these two cases, if the plant is out of control in the previous time slot, i.e., 𝐛a(t1)=𝟎\vec{\mathbf{b}}_{a}(t-1)=\mathbf{0}, the controller erases its buffer due to the same reason in case 1); otherwise, the controller sends the buffered commands to the actuator as much as it can, subject to the constraints of the C-A channel capacity and the actuator’s buffer length.

The actuator’s operations are described as

  1. 1.

    If γ(t)=0\gamma(t)=0, the actuator’s buffer is shifted by one step, i.e., 𝐛a(t)=𝐒a𝐛a(t1)\vec{\mathbf{b}}_{a}(t)=\mathbf{S}_{a}\vec{\mathbf{b}}_{a}(t-1), since the first command in the buffer of the previous time slot was used for control.

  2. 2.

    If γ(t)=1\gamma(t)=1 and γ(t)N(t)>0\gamma^{\prime}(t)N(t)>0, the actuator erases the previous commands and stores the received ones. This is because the newly calculated control commands are expected to perform better than the previous calculated commands, especially when the process noise has a large variance. 444 Note that when the number of tentative future commands in the newly received packet is less than the previous ones, we still need to erase all the previous commands, not just part of them. The reason is that each buffered control command at the actuator was calculated assuming the successful implementation of the previously calculated control command. Thus, if the first few control commands at the actuator’s buffer are removed (i.e., cannot be applied for control), then the rest of the buffer is of little use.

  3. 3.

    If 𝐛a(t1)=𝟎\vec{\mathbf{b}}_{a}(t-1)=\mathbf{0} and γ(t)N(t)=0\gamma^{\prime}(t)N(t)=0, no operation on the actuator’s buffer is required as there is no new commands transmitted.

  4. 4.

    If 𝐛a(t1)𝟎\vec{\mathbf{b}}_{a}(t-1)\neq\mathbf{0}, γ(t)=1\gamma^{\prime}(t)=1 and γ(t)N(t)=0\gamma^{\prime}(t)N(t)=0, the actuator shifts its buffer by one step due to the same reason in 1) and stores the received commands in the buffer right after the existing commands.

Let λc(t)\lambda_{c}(t) and λa(t)\lambda_{a}(t) denote the effective buffer lengths at the controller and the actuator, respectively. Intuitively, λc(t)\lambda_{c}(t) and λa(t)\lambda_{a}(t) jointly determine the closed-loop performance of plant, as a larger λc(t)\lambda_{c}(t) and a larger λa(t)\lambda_{a}(t) indicate that the plant will be delivered tentative control values for a longer time. Based on the controller’s and the actuator’s operations, the actuator is not necessary to have a larger buffer than the controller. Thus, we assume that ΛaΛc\Lambda_{a}\leq\Lambda_{c}.

Let L(t)L(t) denote the number of tentative commands to be transmitted. Based on the controller’s operations, L(t)L(t) can be written as

L(t)={min{B(t),N(t),Λa},​​​​​​if γ(t)N(t)>0min{B(t),λc(t1),Λaλa(t1)},​​​​​if γ(t)N(t)=0,λa(t1)00,​​​​​​otherwise.L(t)\!=\!\begin{cases}\!\min\{B(t),N(t),\Lambda_{a}\},&\text{\!\!\!\!\!\!if }\gamma^{\prime}(t)N(t)>0\\ \!\min\{B(t),\lambda_{c}(t-1),\Lambda_{a}-\lambda_{a}(t-1)\},&\begin{aligned} &\text{\!\!\!\!\!if }\gamma^{\prime}(t)N(t)=0,\\ &\!\!\!\!\!\lambda_{a}(t-1)\neq 0\end{aligned}\\ \!0,&\text{\!\!\!\!\!\!otherwise.}\\ \end{cases} (12)

Then, the buffer-updating rules based on the controller’s and the actuator’s operations are

𝐛c(t)={𝐒cL(t)[𝐮(t)𝟎],​​​​​if γ(t)N(t)>0,γ(t)=1𝟎,​​​​​if γ(t)N(t)>0,γ(t)=0𝟎,​​​​​if γ(t)N(t)=0,λa(t1)=0𝐒cL(t)𝐛c(t1),​​​​​if γ(t)N(t)=0,λa(t1)0,γ(t)=1𝐛c(t1),​​​​​if γ(t)N(t)=0,λa(t1)0,γ(t)=0\vec{\mathbf{b}}_{c}(t)\!=\!\begin{cases}\mathbf{S}_{c}^{L(t)}\begin{bmatrix}\vec{\mathbf{u}}(t)\\ \mathbf{0}\end{bmatrix},&\text{\!\!\!\!\!if }\gamma^{\prime}(t)N(t)>0,\gamma(t)=1\\ \mathbf{0},&\text{\!\!\!\!\!if }\gamma^{\prime}(t)N(t)>0,\gamma(t)=0\\ \mathbf{0},&\text{\!\!\!\!\!if }\gamma^{\prime}(t)N(t)=0,\lambda_{a}(t-1)=0\\ \mathbf{S}_{c}^{L(t)}\vec{\mathbf{b}}_{c}(t-1),&\text{\!\!\!\!\!if }\gamma^{\prime}(t)N(t)=0,\lambda_{a}(t-1)\!\neq\!0,\gamma(t)\!=\!1\\ \vec{\mathbf{b}}_{c}(t-1),&\text{\!\!\!\!\!if }\gamma^{\prime}(t)N(t)=0,\lambda_{a}(t-1)\!\neq\!0,\gamma(t)\!=\!0\\ \end{cases} (13)

and

𝐛a(t)={𝐒a𝐛a(t1), if γ(t)=0[[𝐮1(t),,𝐮L(t)(t)]𝟎], if γ(t)N(t)>0,γ(t)=1𝟎, if γ(t)N(t)=0,𝐛a(t)=𝟎[[𝐛a,2(t1),,𝐛a,λa(t1)(t)][𝐛c,1(t),,𝐛c,L(t)(t)]𝟎],if γ(t)N(t)=0,𝐛a(t)𝟎,γ(t)=1.\vec{\mathbf{b}}_{a}(\!t)\!\!=\!\!\begin{cases}\mathbf{S}_{a}\vec{\mathbf{b}}_{a}(t-1),\text{\hskip 62.59596pt if }\gamma(t)=0\\ \!\!\!\begin{bmatrix}[\mathbf{u}_{1}(t)^{\top},\cdots,\mathbf{u}_{L(t)}(t)^{\top}]^{\top}\\ \mathbf{0}\end{bmatrix},\text{ if }\gamma^{\prime}(t)N(t)>0,\gamma(t)=1\\ \mathbf{0},\begin{aligned} \text{\hskip 102.43008pt if }\gamma^{\prime}(t)N(t)=0,\vec{\mathbf{b}}_{a}(t)=\mathbf{0}\end{aligned}\\ \!\!\!\begin{bmatrix}\![\mathbf{b}_{a,2}(t-1)^{\top},\cdots,\mathbf{b}_{a,\lambda_{a}(t-1)}(t)^{\top}]^{\top}\\ [\mathbf{b}_{c,1}(t)^{\top},\cdots,\mathbf{b}_{c,L(t)}(t)^{\top}]^{\top}\\ \mathbf{0}\end{bmatrix}\!\!\!,\!\begin{aligned} \text{if }\gamma^{\prime}(t)N(t)\!=\!0,\\ \vec{\mathbf{b}}_{a}(t)\!\neq\!\mathbf{0},\gamma(t)\!=\!1.\end{aligned}\\ \end{cases} (14)

From (13) and (14), the updating rules of the effective buffer lengths λc(t)\lambda_{c}(t) and λa(t)\lambda_{a}(t) are

λc(t)={N(t)L(t),if γ(t)N(t)>0,γ(t)=10,if γ(t)N(t)>0,γ(t)=00,if γ(t)N(t)=0,λa(t1)=0λc(t+1)L(t),​​​​if γ(t)N(t)=0,λa(t1)0,γ(t)=1λc(t1),​​​​if γ(t)N(t)=0,λa(t1)0,γ(t)=0\lambda_{c}(t)=\begin{cases}N(t)-L(t),&\text{if }\gamma^{\prime}(t)N(t)>0,\gamma(t)=1\\ 0,&\text{if }\gamma^{\prime}(t)N(t)>0,\gamma(t)=0\\ 0,&\text{if }\gamma^{\prime}(t)N(t)=0,\lambda_{a}(t-1)=0\\ \lambda_{c}(t+1)-L(t),&\text{\!\!\!\!if }\gamma^{\prime}(t)N(t)=0,\lambda_{a}(t-1)\!\neq\!0,\!\gamma(t)\!=\!1\\ \lambda_{c}(t-1),&\text{\!\!\!\!if }\gamma^{\prime}(t)N(t)=0,\lambda_{a}(t-1)\!\neq\!0,\!\gamma(t)\!=\!0\end{cases} (15)

and

λa(t)={max{λa(t1)1,0},​if γ(t)=0L(t),​if γ(t)N(t)>0,γ(t)=10,​if γ(t)N(t)=0,λa(t1)=0λa(t1)+L(t)1,​if γ(t)N(t)=0,λa(t1)0,γ(t)=1.\lambda_{a}(t)\!=\!\begin{cases}\!\max\{\lambda_{a}(t\!-\!1)\!-\!1,\!0\},&\text{\!if }\gamma(t)=0\\ \!L(t),&\text{\!if }\gamma^{\prime}(t)N(t)>0,\gamma(t)=1\\ \!0,&\text{\!if }\gamma^{\prime}(t)N(t)\!=\!0,\lambda_{a}(t-1)\!=\!0\\ \!\lambda_{a}(t-1)+L(t)-1,&\begin{aligned} &\text{\!if }\gamma^{\prime}(t)N(t)\!=\!0,\lambda_{a}(t-1)\!\neq\!0,\\ &\gamma(t)=1.\end{aligned}\end{cases} (16)

III Stability of the Anytime Control System

Based on the anytime control method described in (12), (13) and (14), and following the established stability analysis framework adopting stochastic Lyapunov functions [10, 11, 9], to investigate the stability condition of the system (1), we only need to focus on the plant events that the actuator runs out of control commands, i.e., λa(t)=0,t0\lambda_{a}(t)=0,\forall t\in\mathbb{N}_{0}. However, since the process {λa}0\{\lambda_{a}\}_{\mathbb{N}_{0}} has an infinite memory and is not a Markov process, the methods in [10, 11, 9] are not directly applicable. Instead, we shall analyze the control system through the aggregated Markov process {Z}0\{Z\}_{\mathbb{N}_{0}} defined as

Z(t)\displaystyle Z(t) (λc(t),λa(t),B(t+1),B(t+1),N(t+1))\displaystyle\triangleq\left(\lambda_{c}(t),\lambda_{a}(t),B(t+1),B^{\prime}(t+1),N(t+1)\right) (17)
𝒳c×𝒳a×××𝒩,t0,\displaystyle\hskip 73.97733pt\in\mathcal{X}_{c}\times\mathcal{X}_{a}\times\mathcal{B}\times\mathcal{B}^{\prime}\times\mathcal{N},t\in\mathbb{N}_{0},

where 𝒳c{0,1,,Λc}\mathcal{X}_{c}\triangleq\{0,1,\cdots,\Lambda_{c}\} and 𝒳a{0,1,,min{Λa,N¯}}\mathcal{X}_{a}\triangleq\{0,1,\cdots,\min\{\Lambda_{a},\bar{N}\}\}. Assume that Z(t),t0Z(t),\forall t\in\mathbb{N}_{0}, belongs to the finite set 𝒮{s0,s1,,sS}\mathcal{S}\triangleq\{s_{0},s_{1},\cdots,s_{S}\} with cardinality SS. Different from [10], which only needs to analyze an aggregated process of two processes, we need to investigate the aggregation of five processes, where both {λc}0\{\lambda_{c}\}_{\mathbb{N}_{0}} and {λa}0\{\lambda_{a}\}_{\mathbb{N}_{0}} are correlated with {B,B,N}0\{B,B^{\prime},N\}_{\mathbb{N}_{0}}.

Since the control process is divided by the open-loop events with λa(t)=0\lambda_{a}(t)=0, we define 𝒦={kn}n0\mathcal{K}=\{k_{n}\}_{n\in\mathbb{N}_{0}} as the sequence of time steps with λa(t)=0\lambda_{a}(t)=0. We name the time sequence between knk_{n} and kn+1k_{n+1} as the (n+1)(n+1)th cycle of the process, n0\forall n\in\mathbb{N}_{0}. Then, the number of time steps between consecutive elements of 𝒦\mathcal{K} is

Δn+1=kn+1kn.\Delta_{n+1}=k_{n+1}-k_{n}. (18)

Without loss of generality, let the set 𝒮0{s0,s1,,sS0}𝒮\mathcal{S}_{0}\triangleq\{s_{0},s_{1},\cdots,s_{S_{0}}\}\in\mathcal{S} with cardinality of S0S_{0} denote the subset of 𝒮\mathcal{S} consisting of all the states with λa(t)=0\lambda_{a}(t)=0, and hence Z(kn)𝒮0,n0Z(k_{n})\in\mathcal{S}_{0},\forall n\in\mathbb{N}_{0}. In [10, 11, 9], the set 𝒮0\mathcal{S}_{0} has only one state. In our scenario, S0>1S_{0}>1 introduces more challenges in analyzing the process {kn}n0\{k_{n}\}_{n\in\mathbb{N}_{0}}. The state transition process of {Z}0\{Z\}_{\mathbb{N}_{0}} is illustrated in Fig. 3.

Refer to caption
Figure 3: An illustration of the state transition process of {Z}0\{Z\}_{\mathbb{N}_{0}}.

In what follows, we study the properties of {Z}0\{Z\}_{\mathbb{N}_{0}} and {Z(kn)}n0\{Z(k_{n})\}_{n\in\mathbb{N}_{0}} and then analyze the stability condition.

III-A Properties of {Z}0\{Z\}_{\mathbb{N}_{0}} and {Z(kn)}n0\{Z(k_{n})\}_{n\in\mathbb{N}_{0}}

For ease of analysis, we need the following assumption about the aggregated process {(B,B,N)}0\{(B,B^{\prime},N)\}_{\mathbb{N}_{0}}.

Assumption 3.

{(B,B,N)}0××𝒩\{(B,B^{\prime},N)\}_{\mathbb{N}_{0}}\in\mathcal{B}\times\mathcal{B}^{\prime}\times\mathcal{N} is an irreducible and aperiodic (IA) Markov process.

Note that wireless fading channel conditions are commonly modeled as IA Markov processes [14, 15, 23, 24], thus it is reasonable to consider an IA Markov process {(B,B)}0×\{(B,B^{\prime})\}_{\mathbb{N}_{0}}\in\mathcal{B}\times\mathcal{B}^{\prime}. Also the computation availability {N}0𝒩\{N\}_{\mathbb{N}_{0}}\in\mathcal{N} is commonly modeled as an IA Markov process [10]. Since an aggregation of IA Markov processes is still an IA Markov process [25], it is reasonable to consider the IA Markov process {(B,B,N)}0\{(B,B^{\prime},N)\}_{\mathbb{N}_{0}}.

For the initial state of {Z}0\{Z\}_{\mathbb{N}_{0}}, it is practical to assume that both buffers are empty:

Assumption 4.

Let λc(0)=λa(0)=0\lambda_{c}(0)=\lambda_{a}(0)=0, B(0)=B0B(0)=B_{0}\in\mathcal{B}, B(0)=B0B^{\prime}(0)=B^{\prime}_{0}\in\mathcal{B}^{\prime} and N(0)=N0𝒩N(0)=N_{0}\in\mathcal{N}.

Lemma 1.

Consider Assumptions 3 and 4. Then {Z}0\{Z\}_{\mathbb{N}_{0}} is an IA Markov process.

Proof.

(1) Irreducibility. Considering the initial state Z(0)Z(0), to prove that it is possible to get to any state from any state in {Z}0\{Z\}_{\mathbb{N}_{0}}, we only need to show that any state (λ~c,λ~a,B~,B~,N~)𝒮(\tilde{\lambda}_{c},\tilde{\lambda}_{a},\tilde{B},\tilde{B}^{\prime},\tilde{N})\in\mathcal{S} can return to Z(0)Z(0) [25]. Since {(B,B,N)}0\{(B,B^{\prime},N)\}_{\mathbb{N}_{0}} is an IA Markov process, we assume that (B~,B~,N~)(\tilde{B},\tilde{B}^{\prime},\tilde{N}) can get to (B0,B0,N0)(B_{0},B^{\prime}_{0},N_{0}) in ll steps with probability p>0p>0, where l>Λal>\Lambda_{a}. Since γ¯>0\bar{\gamma}>0, when the packet dropout event consecutively occurs for ll times with probability γ¯l>0\bar{\gamma}^{l}>0, the effective buffer lengths will be zero at both controller and actuator. Thus, the state (λ~c,λ~a,B~,B~,N~)(\tilde{\lambda}_{c},\tilde{\lambda}_{a},\tilde{B},\tilde{B}^{\prime},\tilde{N}) can return to Z(0)Z(0) in ll steps with a non-zero probability.

(2) Aperiodicity. To prove the aperiodicity of {Z}0\{Z\}_{\mathbb{N}_{0}}, we only need to show that a single state state Z(0)Z(0) is aperiodic due to the irreducebility of {Z}0\{Z\}_{\mathbb{N}_{0}} [25]. Due to the aperiodicity of {(B,B,N)}0\{(B,B^{\prime},N)\}_{\mathbb{N}_{0}}, the period of the state (B0,B0,N0)(B_{0},B^{\prime}_{0},N_{0}) is 11 and is written as [25]

gcd{t>0:𝖯𝗋𝗈𝖻[(B(t),B(t),N(t))=(B0,B0,N0)|\displaystyle\gcd\{t>0:\mathsf{Prob}\left[(B(t),B^{\prime}(t),N(t))=(B_{0},B^{\prime}_{0},N_{0})|\right. (19)
(B(0),B(0),N(0))=(B0,B0,N0)]>0}=1.\displaystyle\left.\hskip 62.59596pt(B(0),B^{\prime}(0),N(0))=(B_{0},B^{\prime}_{0},N_{0})\right]>0\}=1.

If state (B0,B0,N0)(B_{0},B^{\prime}_{0},N_{0}) can reach itself in tt\in\mathbb{N} steps, due to the non-zero probability of consecutive packet dropout of tt times, the state (0,0,B0,B0,N0)(0,0,B_{0},B^{\prime}_{0},N_{0}) can reach itself in tt steps as well. Then, from (19), the period of the state (0,0,B0,B0,N0)(0,0,B_{0},B^{\prime}_{0},N_{0}) is 11. ∎

From Lemma 1, {Z}0\{Z\}_{\mathbb{N}_{0}} has a unique stationary distribution. We shall denote the state transition probability of {Z}0\{Z\}_{\mathbb{N}_{0}} as per

vi,j\displaystyle v_{i,j} 𝖯𝗋𝗈𝖻[Z(t+1)=sj|Z(t)=si],si,sj𝒮,t0.\displaystyle\triangleq\mathsf{Prob}\left[Z(t+1)=s_{j}|Z(t)=s_{i}\right],\forall s_{i},s_{j}\in\mathcal{S},t\in\mathbb{N}_{0}. (20)

Note that vi,jv_{i,j} can be numerically calculated based on (3), (9), (15) and (16), though it does not have a closed-from expression due to the complexity introduced by the dual-buffer updating process (15) and (16).

Let 𝐕S×S\mathbf{V}\in\mathbb{R}^{S\times S} denote the state transition probability matrix, i.e., [𝐕]i,j=vi,j[\mathbf{V}]_{i,j}=v_{i,j}, and

𝐕=[𝐕0,0𝐕0,1𝐕1,0𝐕1,1]\mathbf{V}=\begin{bmatrix}\mathbf{V}_{0,0}&\!\!\!\framebox{}\!\!\!&\mathbf{V}_{0,1}\\ \hline\cr\mathbf{V}_{1,0}&\!\!\!\framebox{}\!\!\!&\mathbf{V}_{1,1}\\ \end{bmatrix} (21)

where 𝐕0,0S0×S0\mathbf{V}_{0,0}\in\mathbb{R}^{S_{0}\times S_{0}}, 𝐕0,1S0×(SS0)\mathbf{V}_{0,1}\in\mathbb{R}^{S_{0}\times(S-S_{0})}, 𝐕1,0(SS0)×S0\mathbf{V}_{1,0}\in\mathbb{R}^{(S-S_{0})\times S_{0}} and 𝐕1,1(SS0)×(SS0)\mathbf{V}_{1,1}\in\mathbb{R}^{(S-S_{0})\times(S-S_{0})}.

Lemma 2.

{Z(kn)}n0\{Z(k_{n})\}_{n\in\mathbb{N}_{0}} is an IA Markov process under Assumptions 3 and 4. The state transition probability matrix is

𝐕~l=1𝐃(l)S0×S0,\tilde{\mathbf{V}}\triangleq\sum_{l=1}^{\infty}\mathbf{D}(l)\in\mathbb{R}^{S_{0}\times S_{0}},

where

𝐃(l)={𝐕0,0,l=1𝐕0,1𝐕1,1l2𝐕1,0,l>1\mathbf{D}(l)=\begin{cases}\mathbf{V}_{0,0},&l=1\\ \mathbf{V}_{0,1}\mathbf{V}_{1,1}^{l-2}\mathbf{V}_{1,0},&l>1\end{cases}

The stationary distribution of si𝒮0s_{i}\in\mathcal{S}_{0}, πi\pi_{i}, is the unique solution of

𝝅𝐕~=𝝅,\bm{\pi}^{\top}\tilde{\mathbf{V}}=\bm{\pi}^{\top}, (22)

where 𝝅[π1,π2,,πS0]\bm{\pi}\triangleq[\pi_{1},\pi_{2},\cdots,\pi_{S_{0}}]^{\top}.

Proof.

The irreducibility of {Z(kn)}n0\{Z(k_{n})\}_{n\in\mathbb{N}_{0}} is obvious, because {Z}0\{Z\}_{\mathbb{N}_{0}} is irreducible. For the aperiodicity, we only need to prove that there exist one state of {Z(kn)}n0\{Z(k_{n})\}_{n\in\mathbb{N}_{0}} with period 11 [25]. Due to the aperiodicity of {(B,B,N)}0\{(B,B^{\prime},N)\}_{\mathbb{N}_{0}}, the period of the state (B,B,N)(B,B^{\prime},N) is 11. If the state (B,B,N)(B,B^{\prime},N) can reach itself in nn\in\mathbb{N} steps, due to the non-zero probability of consecutive packet dropout of nn times, the state (0,0,B,B,N)𝒮0(0,0,B,B^{\prime},N)\in\mathcal{S}_{0} can reach itself in nn\in\mathbb{N} steps as well without passing through any state within 𝒮1𝒮\𝒮0\mathcal{S}_{1}\triangleq\mathcal{S}\backslash\mathcal{S}_{0}. Therefore, the period of the state (0,0,B,B,N)(0,0,B,B^{\prime},N) of the process {Z(kn)}n0\{Z(k_{n})\}_{n\in\mathbb{N}_{0}} is 11. In the following, we derive the state transition probability matrix.

We define the conditional probability

di,j(l)\displaystyle d_{i,j}(l) (23)
𝖯𝗋𝗈𝖻[Δn+1=l,Z(kn+1)=sj|Z(kn)=si],si,sj𝒮0,l,\displaystyle\!\triangleq\!\mathsf{Prob}\left[\Delta_{n+1}\!=\!l,Z(k_{n+1})\!=\!s_{j}|Z(k_{n})\!=\!s_{i}\right],s_{i},s_{j}\!\in\!\mathcal{S}_{0},l\in\!\mathbb{N},

and note that

di,j(1)=vi,j,si,sj𝒮0.d_{i,j}(1)=v_{i,j},\forall s_{i},s_{j}\in\mathcal{S}_{0}. (24)

Let Δ~n+1\tilde{\Delta}_{n+1} denote the number of steps to go from Z(kn)=si𝒮0Z(k_{n})=s_{i}\in\mathcal{S}_{0} to sj𝒮1s_{j}\in\mathcal{S}_{1} without passing through any states in 𝒮0\mathcal{S}_{0}. We define the following conditional probability

d~i,j(l)\displaystyle\!\tilde{d}_{i,j}(l) (25)
𝖯𝗋𝗈𝖻[Δ~n+1=l,Z(kn+Δ~n+1)=sj|Z(kn)=si,kn+Δ~n+1<kn+1],\displaystyle\!\!\triangleq\!\!\mathsf{Prob}\left[\!\tilde{\Delta}_{n+1}\!\!=\!l,\!Z(\!k_{n}\!+\!\tilde{\Delta}_{n+1}\!)\!\!=\!s_{j}\!|\!Z(\!k_{n}\!)\!=\!s_{i},k_{n}\!\!+\!\tilde{\Delta}_{n+1}\!<\!k_{n+1}\!\right]\!,
si𝒮0,sj𝒮1,l\displaystyle\hskip 150.79968pt\forall s_{i}\in\mathcal{S}_{0},s_{j}\in\mathcal{S}_{1},l\in\mathbb{N}

and

d~i,j(1)=vi,j,si𝒮0,sj𝒮1.\tilde{d}_{i,j}(1)=v_{i,j},\forall s_{i}\in\mathcal{S}_{0},s_{j}\in\mathcal{S}_{1}. (26)

Then, it can be shown that

d~i,k(l+1)\displaystyle\tilde{d}_{i,k}(l+1) =k𝒮1d~i,k(l)vk,k,\displaystyle=\sum_{k^{\prime}\in\mathcal{S}_{1}}\tilde{d}_{i,k^{\prime}}(l)v_{k^{\prime},k}, (27)
di,j(l+1)\displaystyle d_{i,j}(l+1) =k𝒮1d~i,k(l)vk,j.\displaystyle=\sum_{k\in\mathcal{S}_{1}}\tilde{d}_{i,k}(l)v_{k,j}. (28)

Let’s introduce 𝐃(l)S0×S0\mathbf{D}(l)\in\mathbb{R}^{S_{0}\times S_{0}} and 𝐃~(l)S0×(SS0)\tilde{\mathbf{D}}(l)\in\mathbb{R}^{S_{0}\times(S-S_{0})}, where [𝐃(l)]i,j=di,j(l)[\mathbf{D}(l)]_{i,j}=d_{i,j}(l) and [𝐃~(l)]i,j=d~i,j(l)[\tilde{\mathbf{D}}(l)]_{i,j}=\tilde{d}_{i,j}(l). From (27), we have

𝐃~(l+1)=𝐃~(l)𝐕1,1,\tilde{\mathbf{D}}(l+1)=\tilde{\mathbf{D}}(l)\mathbf{V}_{1,1}, (29)

and hence

𝐃~(l)=𝐃~(1)𝐕1,1l1.\tilde{\mathbf{D}}(l)=\tilde{\mathbf{D}}(1)\mathbf{V}_{1,1}^{l-1}. (30)

Similarly, from (28), we have

𝐃(l+1)=𝐃~(l)𝐕1,0.\mathbf{D}(l+1)=\tilde{\mathbf{D}}(l)\mathbf{V}_{1,0}. (31)

Using (24), (25), (30) and (31), we have

𝐃(l)={𝐃(1)=𝐕0,0,l=1𝐃~(1)𝐕1,1l2𝐕1,0=𝐕0,1𝐕1,1l2𝐕1,0,l>1.\mathbf{D}(l)=\begin{cases}\mathbf{D}(1)=\mathbf{V}_{0,0},&l=1\\ \tilde{\mathbf{D}}(1)\mathbf{V}_{1,1}^{l-2}\mathbf{V}_{1,0}=\mathbf{V}_{0,1}\mathbf{V}_{1,1}^{l-2}\mathbf{V}_{1,0},&l>1.\end{cases} (32)

Therefore, the state transition probability matrix satisfies 𝐕~l=1𝐃(l).\tilde{\mathbf{V}}\triangleq\sum_{l=1}^{\infty}\mathbf{D}(l).

III-B Analysis of the Stability Condition

Similar to [10, 11, 9], the stability of the WNCS depends on the statistics of {Δn}n\{\Delta_{n}\}_{n\in\mathbb{N}} in (18), which denotes the time duration between consecutive open-loop events. Different to [10, 9], in the considered case, the process {Δn}n\{\Delta_{n}\}_{n\in\mathbb{N}} is not i.i.d. For WNCS with a single channel, in [11] an event-triggered setup was considered, leading to {Δn}n\{\Delta_{n}\}_{n\in\mathbb{N}} which is not i.i.d. However, our current setup is different from [11]. In particular, {Δn}n\{\Delta_{n}\}_{n\in\mathbb{N}} is formed by the first return time of a set of states in 𝒮0\mathcal{S}_{0} rather than a single state (as in [11]) . Therefore, the approach in [11] cannot be adopted directly. In the following, we propose a novel cycle-cost-based approach to obtain sufficient stability conditions.

Lemma 3.

The plant is stochastically stable if

n=1𝖤[Ξ(n)]<,\sum_{n=1}^{\infty}\mathsf{E}\left[\Xi(n)\right]<\infty, (33)

where

Ξ(n)=αnρi=1n(Δi1),n,\Xi(n)=\alpha^{n}\rho^{\sum_{i=1}^{n}(\Delta_{i}-1)},\forall n\in\mathbb{N}, (34)

and ρ\rho and α\alpha were defined in Assumptions 1 and 2, respectively.

Proof.

From Assumptions 1 and 2, and Definition 1, the plant is stable if we can prove that

𝖤[t=0V(𝐱(t))]<.\mathsf{E}\left[\sum_{t=0}^{\infty}V(\mathbf{x}(t))\right]<\infty. (35)

By using Assumptions 1 and 2, we have

t=knkn+11V(𝐱(t))\displaystyle\sum_{t=k_{n}}^{k_{n+1}-1}V(\mathbf{x}(t)) (1+αl=0kn+1kn2ρl)V(𝐱(kn))\displaystyle\leq\left(1+\alpha\sum_{l=0}^{k_{n+1}-k_{n}-2}\rho^{l}\right)V(\mathbf{x}(k_{n})) (36)
<(1+α1ρ)V(𝐱(kn)),\displaystyle<\left(1+\frac{\alpha}{1-\rho}\right)V(\mathbf{x}(k_{n})),

and hence

t=0V(𝐱(t))(1+α1ρ)n=0V(𝐱(kn)).\sum_{t=0}^{\infty}V(\mathbf{x}(t))\leq\left(1+\frac{\alpha}{1-\rho}\right)\sum_{n=0}^{\infty}V(\mathbf{x}(k_{n})). (37)

Then, it is easy to obtain that

V(𝐱(kn))αnρi=1n(Δi1)V(𝐱(0)).V(\mathbf{x}(k_{n}))\leq\alpha^{n}\rho^{\sum_{i=1}^{n}(\Delta_{i}-1)}V(\mathbf{x}(0)). (38)

Since 𝖤[V(𝐱(kn))]αn𝖤[ρi=1n(Δi1)]𝖤[V(𝐱(0))]\mathsf{E}\left[V(\mathbf{x}(k_{n}))\right]\leq\alpha^{n}\mathsf{E}\left[\rho^{\sum_{i=1}^{n}(\Delta_{i}-1)}\right]\mathsf{E}\left[V(\mathbf{x}(0))\right] and 𝖤[V(𝐱(0))]<\mathsf{E}\left[V(\mathbf{x}(0))\right]<\infty, (33) is proved. ∎

From Lemma 3, to find the stability condition of the system, we only need to investigate the process {Ξ}\{\Xi\}_{\mathbb{N}}. However, {Ξ}\{\Xi\}_{\mathbb{N}} is not Markovian, whereas the underlying process {Z(kn)}n0\{Z(k_{n})\}_{n\in\mathbb{N}_{0}} is. In the following, we investigate two stability conditions, with and without exploring the state transition properties of the underlying process. The results are stated in Theorems 1 and 2, respectively.

Before proceeding, we need the technical lemma below.

Lemma 4.

The following inequality holds

𝖤[Ξ(n+1)]<(αρmaxi,j{1,,S0}ri,j)𝖤[Ξ(n)],n,\mathsf{E}\left[\Xi(n+1)\right]<\left(\frac{\alpha}{\rho}\max_{i,j\in\{1,\cdots,S_{0}\}}r_{i,j}\right)\mathsf{E}\left[\Xi(n)\right],\forall n\in\mathbb{N},

where

ri,j\displaystyle r_{i,j} 𝖤[ρΔn+1|Z(kn)=si,Z(kn+1)=sj]\displaystyle\triangleq\mathsf{E}\left[\rho^{\Delta_{n+1}}|Z(k_{n})=s_{i},Z(k_{n+1})=s_{j}\right]
=l=1ρl[𝐃(l)]i,j[𝐕~]i,j<,i,j{1,,S0},\displaystyle=\frac{\sum_{l=1}^{\infty}\rho^{l}[\mathbf{D}(l)]_{i,j}}{[\tilde{\mathbf{V}}]_{i,j}}<\infty,i,j\in\{1,\cdots,S_{0}\},

and 𝐃(l)\mathbf{D}(l) and 𝐕~\tilde{\mathbf{V}} are defined in Lemma 2.

Proof.

From (34), it can be shown that

Ξ(n+1)=Ξ(n)αρρΔn+1,\Xi(n+1)=\Xi(n)\frac{\alpha}{\rho}\rho^{\Delta_{n+1}}, (39)

and hence

𝖤[Ξ(n+1)]=𝖤[Ξ(n)αρρΔn+1]\displaystyle\mathsf{E}\left[\Xi(n+1)\right]=\mathsf{E}\left[\Xi(n)\frac{\alpha}{\rho}\rho^{\Delta_{n+1}}\right] (40)
=αρi=1S0j=1S0𝖤[Ξ(n)ρΔn+1|Z(kn)=si,Z(kn+1)=sj]\displaystyle=\frac{\alpha}{\rho}\sum_{i=1}^{S_{0}}\sum_{j=1}^{S_{0}}\mathsf{E}\left[\Xi(n)\rho^{\Delta_{n+1}}|Z(k_{n})=s_{i},Z(k_{n+1})=s_{j}\right]
×𝖯𝗋𝗈𝖻[Z(kn)=si,Z(kn+1)=sj]\displaystyle\hskip 113.81102pt\times\mathsf{Prob}\left[Z(k_{n})=s_{i},Z(k_{n+1})=s_{j}\right]
αρmaxi,j{1,,S0}ri,ji=1S0j=1S0𝖤[Ξ(n)|Z(kn)=si,Z(kn+1)=sj]\displaystyle\leq\frac{\alpha}{\rho}\max_{i,j\in\{1,\cdots,S_{0}\}}r_{i,j}\sum_{i=1}^{S_{0}}\sum_{j=1}^{S_{0}}\mathsf{E}\left[\Xi(n)|Z(k_{n})=s_{i},Z(k_{n+1})=s_{j}\right]
×𝖯𝗋𝗈𝖻[Z(kn)=si,Z(kn+1)=sj]\displaystyle\hskip 113.81102pt\times\mathsf{Prob}\left[Z(k_{n})=s_{i},Z(k_{n+1})=s_{j}\right]
=αρmaxi,j{1,,S0}ri,j𝖤[Ξ(n)],\displaystyle=\frac{\alpha}{\rho}\max_{i,j\in\{1,\cdots,S_{0}\}}r_{i,j}\mathsf{E}\left[\Xi(n)\right],

where

ri,j\displaystyle r_{i,j} 𝖤[ρΔn+1|Z(kn)=si,Z(kn+1)=sj]\displaystyle\triangleq\mathsf{E}\left[\rho^{\Delta_{n+1}}|Z(k_{n})=s_{i},Z(k_{n+1})=s_{j}\right] (41)
=Δn+1=1ρl𝖯𝗋𝗈𝖻[Δn+1=l,Z(kn+1)=sj|Z(kn)=si]𝖯𝗋𝗈𝖻[Z(kn+1)=sj|Z(kn)=si]\displaystyle=\sum_{\Delta_{n+1}=1}^{\infty}\rho^{l}\frac{\mathsf{Prob}\left[\Delta_{n+1}=l,Z(k_{n+1})=s_{j}|Z(k_{n})=s_{i}\right]}{\mathsf{Prob}\left[Z(k_{n+1})=s_{j}|Z(k_{n})=s_{i}\right]}
=l=1ρl[𝐃(l)]i,j[𝐕~]i,j,\displaystyle=\frac{\sum_{l=1}^{\infty}\rho^{l}[\mathbf{D}(l)]_{i,j}}{[\tilde{\mathbf{V}}]_{i,j}},

and ri,jr_{i,j} is bounded due to the fact that l=1[𝐃(l)]i,j<1\sum_{l=1}^{\infty}[\mathbf{D}(l)]_{i,j}<1 and ρ<1\rho<1. ∎

Using Lemmas 3 and 4, it is straightforward to have the following result.

Theorem 1.

Suppose that Assumptions 1-4 hold. The system (1) is stochastically stable if

Ωαρmaxi,j{1,,S0}ri,j<1.\Omega^{\prime}\triangleq\frac{\alpha}{\rho}\max_{i,j\in\{1,\cdots,S_{0}\}}r_{i,j}<1.
Remark 2.

For the special case with S0=1S_{0}=1, i.e., there exists only a single state in 𝒮\mathcal{S} leading to no control being applied, Lemma 2 shows that the processes {Z(kn)}n0\{Z(k_{n})\}_{n\in\mathbb{N}_{0}} is i.i.d. and so is the process {Δn}n\{\Delta_{n}\}_{n\in\mathbb{N}}. Such a special case is identical to that of [10], and we see that the stability condition in Theorem 1 reduces to

Ω=αρl=1ρl𝖯𝗋𝗈𝖻[Δn=l]<1,\Omega^{\prime}=\frac{\alpha}{\rho}\sum_{l=1}^{\infty}\rho^{l}\mathsf{Prob}\left[\Delta_{n}=l\right]<1, (42)

which is identical to that of [10].

Remark 3.

The sufficient condition in Theorem 1 is obtained by considering the worst case scenario of the state transitions, i.e., considering the pair of states sis_{i} and sjs_{j} such that it has the largest conditional expectation 𝖤[ρΔn+1|Z(kn)=si,Z(kn+1)=sj]\mathsf{E}\left[\rho^{\Delta_{n+1}}|Z(k_{n})=s_{i},Z(k_{n+1})=s_{j}\right]. However, such a method does not take into account the state transition probabilities in 𝐕~\mathbf{\tilde{V}} defined in Lemma 2. Thus, the sufficient condition in Theorem 1 is conservative.

The following lemma is needed to obtain a less conservative stability condition in Theorem 2.

Lemma 5.

Assuming that {Z}0\{Z\}_{\mathbb{N}_{0}} evolves in the steady state, for any arbitrarily small ϵ>0\epsilon>0, there exist μ>0\mu>0 and K>0K>0 such that

𝖤[Ξ(n)]<μ(αρ(λmax(𝐔)+ϵ))n1,n>K,\mathsf{E}\left[\Xi(n)\right]<\mu\Big{(}\frac{\alpha}{\rho}\left(\lambda_{\max}(\mathbf{U})+\epsilon\right)\Big{)}^{n-1},\forall n>K, (43)

where 𝐔[𝐔1𝐔2𝐔S0]S02×S02,\mathbf{U}\triangleq\begin{bmatrix}\mathbf{U}_{1}^{\top}&\!\!\!\framebox{}\!\!\!&\mathbf{U}_{2}^{\top}&\!\!\!\framebox{}\!\!\!&\cdots&\!\!\!\framebox{}\!\!\!&\mathbf{U}_{S_{0}}^{\top}\end{bmatrix}^{\top}\in\mathbb{R}^{S^{2}_{0}\times S^{2}_{0}},

𝐔i[r1,i𝐟1𝟎𝟎𝟎r2,i𝐟2𝟎𝟎𝟎rS0,i𝐟S0]S0×S02,\mathbf{U}_{i}\triangleq\begin{bmatrix}r_{1,i}\mathbf{f}^{\top}_{1}&\mathbf{0}&\cdots&\mathbf{0}\\ \mathbf{0}&r_{2,i}\mathbf{f}^{\top}_{2}&\cdots&\mathbf{0}\\ \vdots&\cdots&\ddots&\vdots\\ \mathbf{0}&\cdots&\mathbf{0}&r_{S_{0},i}\mathbf{f}^{\top}_{S_{0}}\end{bmatrix}\in\mathbb{R}^{S_{0}\times S^{2}_{0}},

and [𝐟1𝐟2𝐟S0]𝐙𝐕~𝐙1\begin{bmatrix}\mathbf{f}_{1}&\!\!\!\framebox{}\!\!\!&\mathbf{f}_{2}&\!\!\!\framebox{}\!\!\!&\cdots&\!\!\!\framebox{}\!\!\!&\mathbf{f}_{S_{0}}\end{bmatrix}\triangleq\mathbf{Z}\tilde{\mathbf{V}}\mathbf{Z}^{-1} and 𝐙=diag{𝝅}\mathbf{Z}=\text{diag}\{\bm{\pi}\}.

Proof.

See Appendix A. ∎

From Lemmas 3 and 5, it is straightforward to derive the following result.

Theorem 2.

Suppose that Assumptions 1-4 hold. System (1) is stochastically stable if

Ωαρλmax(𝐔)<1.\Omega\triangleq\frac{\alpha}{\rho}\lambda_{\max}(\mathbf{U})<1.
Remark 4.

This stability condition has considered the effect of all the conditional expectations 𝖤[ρΔn+1|Z(kn)=si,Z(kn+1)=sj]\mathsf{E}\left[\rho^{\Delta_{n+1}}|Z(k_{n})=s_{i},Z(k_{n+1})=s_{j}\right] and the state transition probability matrix of {Z(kn)}n\{Z(k_{n})\}_{n\in\mathbb{N}}, i.e., 𝐕~\tilde{\mathbf{V}}. Using Perron-Frobenius theorem [27], since 𝐔\mathbf{U} is non-negative, we have λmax(𝐔)maxi{1,,S0}j=1S0[𝐔]i,jmaxi,j{1,,S0}ri,j\lambda_{\max}(\mathbf{U})\leq\max_{i\in\{1,\cdots,S_{0}\}}\sum_{j=1}^{S_{0}}[\mathbf{U}]_{i,j}\leq\max_{i,j\in\{1,\cdots,S_{0}\}}r_{i,j}. Thus, the sufficient condition Theorem 2 is less restrictive than that in Theorem 1.

III-C Numerical Example for Stability Condition Calculation

We set the buffer lengths Λc=Λa=2\Lambda_{c}=\Lambda_{a}=2, the maximum number of calculated control commands per time step N¯=2\bar{N}=2, the maximum number of commands that can be transmitted via the C-A channel B¯=2\bar{B}=2, the number of channel states of the S-C channel B¯=2\bar{B}^{\prime}=2, and the packet drop probabilities in the two states as 0.20.2 and 0.010.01. We assume that ρ=0.8\rho=0.8. The Markov state transition probability matrices of the C-A and S-C channel state (B(t),B(t)){0,1,2}×{1,2}(B(t),B^{\prime}(t))\in\{0,1,2\}\times\{1,2\}, and the processor’s computational resource N(t){0,1,2}N(t)\in\{0,1,2\} are

𝐌=[0.240.160.060.040.300.200.040.360.010.090.050.450.120.080.060.040.420.280.020.180.010.090.070.63000.300.200.300.20000.050.450.050.45],\mathbf{M}=\begin{bmatrix}0.24&0.16&0.06&0.04&0.30&0.20\\ 0.04&0.36&0.01&0.09&0.05&0.45\\ 0.12&0.08&0.06&0.04&0.42&0.28\\ 0.02&0.18&0.01&0.09&0.07&0.63\\ 0&0&0.30&0.20&0.30&0.20\\ 0&0&0.05&0.45&0.05&0.45\end{bmatrix}, (44)

and

𝐍=[0.10.20.700.60.40.10.30.6].\mathbf{N}=\begin{bmatrix}0.1&0.2&0.7\\ 0&0.6&0.4\\ 0.1&0.3&0.6\end{bmatrix}. (45)

Thus, the process Z(t)Z(t) defined in (17) has 3×3×3×2×3=1623\times 3\times 3\times 2\times 3=162 states. Using the state transition rules of λc(t)\lambda_{c}(t) and λa(t)\lambda_{a}(t) in (15) and (16), and the channel and computational resource state transition probabilities (3) and (9), the state transition matrix of Z(t)Z(t) can be calculated. Due to the space limitation, the matrix is presented in [28] with 7272 transient states (highlighted in yellow), and thus Z(t)Z(t) has S=16272=90S=162-72=90 recurrent (effective) states including S0=54S_{0}=54 states with λa(t)=0\lambda_{a}(t)=0 (highlighted in red). By removing the transient states, the 90×9090\times 90 state transition matrix 𝐕\mathbf{V} of the effective states can be easily obtained. Due to the space limitation, it is not possible to show the calculation of the stability conditions based on the 90×9090\times 90 matrix 𝐕\mathbf{V}. So we only present the comparison result of the stability conditions in Theorems 1 and 2:

maxi,j{1,,S0}ri,j=0.8λmax(𝐔)=0.0016,\max_{i,j\in\{1,\cdots,S_{0}\}}r_{i,j}=0.8\gg\lambda_{\max}(\mathbf{U})=0.0016, (46)

showing that the sufficient stability condition in Theorem 2 is less restrictive than Theorem 1.

For ease of illustration, we will show the calculation of the stability conditions based on a randomly generated (small) 𝐕\mathbf{V} with S=4S=4 and assume that S0=2S_{0}=2, where

𝐕=[0.100.100.100.700.300.200.100.400.600.200.100.100.900.050.020.03]4×4,\mathbf{V}=\begin{bmatrix}0.10&0.10&\!\!\!\framebox{}\!\!\!&0.10&0.70\\ 0.30&0.20&\!\!\!\framebox{}\!\!\!&0.10&0.40\\ \hline\cr 0.60&0.20&\!\!\!\framebox{}\!\!\!&0.10&0.10\\ 0.90&0.05&\!\!\!\framebox{}\!\!\!&0.02&0.03\end{bmatrix}\in\mathbb{R}^{4\times 4}, (47)

and thus 𝐕0,0\mathbf{V}_{0,0}, 𝐕0,1\mathbf{V}_{0,1}, 𝐕1,0\mathbf{V}_{1,0} and 𝐕1,1\mathbf{V}_{1,1} are obtained directly from (21).

Using Lemma 2, the state transition matrix of {Z(kn)}n0\{Z(k_{n})\}_{n\in\mathbb{N}_{0}} and its stationary distribution are obtained as

𝐕~=[0.83780.16220.75460.2454],\tilde{\mathbf{V}}=\begin{bmatrix}0.8378&0.1622\\ 0.7546&0.2454\end{bmatrix}, (48)

and

𝝅=[0.8231 0.1769].\bm{\pi}=[0.8231\ 0.1769]^{\top}. (49)

Taking (48) and (49) into (41), we have ri,j=[𝐑]i,jr_{i,j}=[\mathbf{R}]_{i,j}, where

𝐑=[0.65110.73230.69710.7673].\mathbf{R}=\begin{bmatrix}0.6511&0.7323\\ 0.6971&0.7673\end{bmatrix}. (50)

Taking (48), (49) and (50) into the definition of 𝐔\mathbf{U} in Lemma 5, we have

𝐔=[0.36950.071600000.01650.00540.41560.080500000.01810.0059].\mathbf{U}=\begin{bmatrix}0.3695&0.0716&0&0\\ 0&0&0.0165&0.0054\\ 0.4156&0.0805&0&0\\ 0&0&0.0181&0.0059\end{bmatrix}. (51)

By applying (50) and (51) onto Theorems 1 and 2, respectively, we have

maxi,j{1,,S0}ri,j=0.7673>λmax(𝐔)=0.3731.\max_{i,j\in\{1,\cdots,S_{0}\}}r_{i,j}=0.7673>\lambda_{\max}(\mathbf{U})=0.3731. (52)

IV Robustness to Process Noise

In this section, we investigate the stability condition of the plant system below with process noise:

𝐱(t+1)=f(𝐱(t),𝐮(t),𝐰(t)),\mathbf{x}(t+1)=f(\mathbf{x}(t),\mathbf{u}(t),\mathbf{w}(t)), (53)

where 𝐰(t)ls\mathbf{w}(t)\in\mathbb{R}^{l_{s}} is a white noise process, which is independent with the other random processes of the system.

For ease of analysis, we consider uniform bounds and continuity as follows.

Assumption 5 (​​[10]).

There exists βx\beta_{x}, βu\beta_{u}, βw\beta_{w}, βV\beta_{V}, βκ\beta_{\kappa}, ρ\rho, α\alpha and η>0\eta\in\mathbb{R}_{>0}, such that, 𝐱,𝐳,𝐰ls\forall\mathbf{x},\mathbf{z},\mathbf{w}\in\mathbb{R}^{l_{s}} and 𝐮,𝐯lu\forall\mathbf{u},\mathbf{v}\in\mathbb{R}^{l_{u}} the following are satisfied

|f(𝐱,𝐮,𝐰)f(𝐳,𝐯,𝟎)|βx|𝐱𝐳|+βu|𝐮𝐯|+βw|𝐰|\displaystyle|f(\mathbf{x,u,w})-f(\mathbf{z,v,0})|\leq\beta_{x}|\mathbf{x-z}|+\beta_{u}|\mathbf{u-v}|+\beta_{w}|\mathbf{w}| (54)
|V(𝐱)V(𝐳)|βV|𝐱𝐳|\displaystyle|V(\mathbf{x})-V(\mathbf{z})|\leq\beta_{V}|\mathbf{x-z}|
|κ(𝐱)κ(𝐳)|βκ|𝐱𝐳|\displaystyle|\kappa(\mathbf{x})-\kappa(\mathbf{z})|\leq\beta_{\kappa}|\mathbf{x-z}|
V(f(𝐱,κ(𝐱),𝐰))ρV(𝐱)+η|𝐰|\displaystyle V(f(\mathbf{x},\kappa(\mathbf{x}),\mathbf{w}))\leq\rho V(\mathbf{x})+\eta|\mathbf{w}|
V(f(𝐱,𝟎,𝐰))αV(𝐱)+η|𝐰|.\displaystyle V(f(\mathbf{x},\mathbf{0},\mathbf{w}))\leq\alpha V(\mathbf{x})+\eta|\mathbf{w}|.

When considering unbounded process noise, the stability condition in Definition 1 cannot be satisfied. Thus, we consider the following stability condition in terms of the average cost [10].

Definition 2.

The dynamical system with process noise (53) is stochastically stable, if for some ψ𝒦\psi\in\mathcal{K}_{\infty}, the average expected value lim supT1Tk=0T𝖤[ψ(|𝐱(k)|)]<\limsup\limits_{T\rightarrow\infty}\frac{1}{T}\sum_{k=0}^{T}\mathsf{E}\left[\psi(|\mathbf{x}(k)|)\right]<\infty.

Theorem 3.

Suppose that Assumptions 1-5 holds. The system (53) is stochastically stable if Ω<1\Omega<1.

Proof.

See Appendix B. ∎

Remark 5.

Theorem 3 shows that the stability condition for the process-noise-free case holds for the process-noise-present one as well under Assumption 5, which is in line with [10]. Note that in the case of Markov jump linear systems, stability conditions for noise-free cases are equivalent to conditions for noisy-cases as well, see Theorem 3.33 of Chapter 3 of [29].

V Simulation Results

We present simulation results of the proposed dual-buffer anytime control algorithm and the baseline single-buffer algorithm [10], where the S-C and C-A channels are modeled as multi-state Markov chains. The system parameters are the same as in the numerical example in Section III. We consider an open-loop unstable constrained plant model of (53) in the presence of noise [11]:

[x1(t)x2(t)]=[x2(t)+u1(t)sat(x1(t)+x2(t))+u2(t)]+[w1(t)w2(t)],\begin{bmatrix}x_{1}(t)\\ x_{2}(t)\end{bmatrix}=\begin{bmatrix}x_{2}(t)+u_{1}(t)\\ -\text{sat}(x_{1}(t)+x_{2}(t))+u_{2}(t)\end{bmatrix}+\begin{bmatrix}w_{1}(t)\\ w_{2}(t)\end{bmatrix}, (55)

where

sat(μ)={10, if μ<10μ, if μ[10,10]10, if μ>10\text{sat}(\mu)=\begin{cases}-10,\text{ if }\mu<-10\\ \mu,\text{ if }\mu\in[-10,10]\\ 10,\text{ if }\mu>10\end{cases} (56)

and the process noise w1(t)w_{1}(t) and w2(t)w_{2}(t) are independent zero-mean i.i.d. Gaussian and 𝖤[w1(t)2]=𝖤[w2(t)2]=0.1\mathsf{E}\left[w_{1}(t)^{2}\right]=\mathsf{E}\left[w_{2}(t)^{2}\right]=0.1. We take the control policy as κ(𝐱)=[x2, 0.505sat(x1+x2)]\kappa(\mathbf{x})=[-x_{2},\ 0.505\text{sat}(x_{1}+x_{2})]^{\top}.

In Fig. 4, we show the simulation of |𝐱(t)||\mathbf{x}(t)| with 800800 time steps for both the dual-buffer anytime method and the baseline method. It is clear that the dual-buffer method significantly outperforms the baseline especially when the initial process state 𝐱(0)\mathbf{x}(0) is far from the origin. Therefore, adding the command buffer at the actuator can effectively overcome the effect of packet dropouts, leading to significant control performance improvement.

Refer to caption
Figure 4: Simulation of |𝐱(t)||\mathbf{x}(t)| with different initiations.

VI Conclusions

We have studied an anytime control algorithm for dual-channel-dual-buffer WNCS with random computational resources over correlated channels. We have proposed a novel approach to derive sufficient conditions for stochastic stability in the case of nonlinear plant models with disturbances. The stability conditions are stated in terms of plant dynamics, network dynamics, buffer properties and computational resource dynamics. Our numerical results have shown that the proposed dual-buffer anytime control system can provide significant control performance improvement when compared to the conventional single-buffer system as it effectively overcomes the effect of packet dropouts. In future work, we will consider optimal control sequence design. In addition, we may include an extension to large-scale WNCSs with multiple plants and controllers and that need to share network and communication resources.

References

  • [1] L. K. McGovern and E. Feron, “Requirements and hard computational bounds for real-time optimization in safety-critical control systems,” in Proc. IEEE CDC, 1998, pp. 3366–3371.
  • [2] ——, “Closed-loop stability of systems driven by real-time, dynamic optimization algorithms,” in Proc. IEEE CDC, 1999, pp. 3690–3696.
  • [3] D. Henriksson, A. Cervin, J. Akesson, and K. . Arzen, “On dynamic real-time scheduling of model predictive controllers,” in Proc. IEEE CDC, 2002, pp. 1325–1330.
  • [4] A. Cervin, M. Velasco, P. Marti, and A. Camacho, “Optimal online sampling period assignment: Theory and experiments,” IEEE Trans. Control Syst. Technol., vol. 19, no. 4, pp. 902–910, 2011.
  • [5] P. Tabuada, “Event-triggered real-time scheduling of stabilizing control tasks,” IEEE Trans. Autom. Control, vol. 52, no. 9, pp. 1680–1685, 2007.
  • [6] X. Wang and M. D. Lemmon, “Self-triggered feedback control systems with finite-gain l2l_{2} stability,” IEEE Trans. Autom. Control, vol. 54, no. 3, pp. 452–467, 2009.
  • [7] R. Bhattacharya and G. J. Balas, “Anytime control algorithm: Model reduction approach,” J. Guidance, Control, Dynamics, vol. 27, no. 5, pp. 767–776, Sep.-Oct. 2004.
  • [8] V. Gupta and F. Luo, “On a control algorithm for time-varying processor availability,” IEEE Trans. Autom. Control, vol. 58, no. 3, pp. 743–748, 2013.
  • [9] D. E. Quevedo and V. Gupta, “Sequence-based anytime control,” IEEE Trans. Autom. Control, vol. 58, no. 2, pp. 377–390, 2013.
  • [10] D. E. Quevedo, W.-J. Ma, and V. Gupta, “Anytime control using input sequences with Markovian processor availability,” IEEE Trans. Autom. Control, vol. 60, no. 2, pp. 515–521, Feb. 2015. [Online]. Available: https://arxiv.org/pdf/1405.0751.pdf
  • [11] D. E. Quevedo, V. Gupta, W. Ma, and S. Yüksel, “Stochastic stability of event-triggered anytime control,” IEEE Trans. Autom. Control, vol. 59, no. 12, pp. 3373–3379, 2014.
  • [12] T. V. Dang, K. Ling, and D. E. Quevedo, “Stability analysis of event-triggered anytime control with multiple control laws,” IEEE Trans. Autom. Control, vol. 64, no. 1, pp. 420–426, 2019.
  • [13] X.-S. Zhan, J. Wu, T. Jiang, and X.-W. Jiang, “Optimal performance of networked control systems under the packet dropouts and channel noise,” ISA Transactions, vol. 58, pp. 214–221, 2015. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S0019057815001299
  • [14] P. Sadeghi, R. A. Kennedy, P. B. Rapajic, and R. Shams, “Finite-state Markov modeling of fading channels - a survey of principles and applications,” IEEE Signal Process. Mag., vol. 25, no. 5, pp. 57–80, Sep. 2008.
  • [15] K. Huang, W. Liu, Y. Li, B. Vucetic, and A. Savkin, “Optimal downlink-uplink scheduling of wireless networked control for Industrial IoT,” IEEE Internet Things J., vol. 7, no. 3, pp. 1756–1772, Mar. 2020.
  • [16] K. Huang, W. Liu, Y. Li, and B. Vucetic, “To retransmit or not: Real-time remote estimation in wireless networked control,” in Proc. IEEE ICC, May 2019, pp. 1–7.
  • [17] K. Huang, W. Liu, M. Shirvanimoghaddam, Y. Li, and B. Vucetic, “Real-time remote estimation with hybrid ARQ in wireless networked control,” IEEE Trans. Wireless Commun., vol. 19, no. 5, pp. 3490–3504, 2020.
  • [18] D. Tse and P. Viswanath, Fundamentals of wireless communication.   Cambridge university press, 2005.
  • [19] Y. Polyanskiy, H. V. Poor, and S. Verdu, “Channel coding rate in the finite blocklength regime,” IEEE Trans. Inf. Theory, vol. 56, no. 5, pp. 2307–2359, May 2010.
  • [20] A. J. Goldsmith and Soon-Ghee Chua, “Variable-rate variable-power mqam for fading channels,” IEEE Trans. Commun., vol. 45, no. 10, pp. 1218–1230, 1997.
  • [21] L. Schenato, B. Sinopoli, M. Franceschetti, K. Poolla, and S. S. Sastry, “Foundations of control and estimation over lossy networks,” Proc. IEEE, vol. 95, no. 1, pp. 163–187, Jan. 2007.
  • [22] D. E. Quevedo and D. Nešić, “Robust stability of packetized predictive control of nonlinear systems with disturbances and markovian packet losses,” Automatica, vol. 48, no. 8, pp. 1803 – 1811, 2012.
  • [23] W. Liu, D. E. Quevedo, Y. Li, K. H. Johansson, and B. Vucetic, “Remote state estimation with smart sensors over markov fading channels,” submitted to IEEE Trans. Autom. Control, 2020. [Online]. Available: https://arxiv.org/abs/2005.07871
  • [24] W. Liu, D. E. Quevedo, K. H. Johansson, Y. Li, and B. Vucetic, “Remote state estimation of multiple systems over multiple markov fading channels,” submitted to IEEE Trans. Autom. Control, 2021. [Online]. Available: https://arxiv.org/abs/2104.04181
  • [25] R. Durrett, Probability: Theory and Examples.   Cambridge university press, 2019, vol. 49.
  • [26] W. Liu, D. E. Quevedo, Y. Li, and B. Vucetic, “Anytime control under practical communication model,” submitted to IEEE Trans. Autom. Control, 2021. [Online]. Available: https://arxiv.org/abs/2012.00962
  • [27] H. Minc, Nonnegative Matrices.   Wiley, 1988.
  • [28] State Transition Matrix of Z(t)Z(t), May 2021. [Online]. Available: {https://drive.google.com/file/d/12hnTfMwKoM9Rx4eXSlq9N0Sak_6M8dtb/view?usp=sharing/}
  • [29] O. L. V. Costa, M. D. Fragoso, and R. P. Marques, Discrete-time Markov jump linear systems.   Springer Science & Business Media, 2006.

A. Proof of Lemma 5

Direct calculations give that

𝖤[Ξ(n+1)|Z(kn)=si,Z(kn+1)=sj]\displaystyle\mathsf{E}\left[\Xi(n+1)|Z(k_{n})=s_{i},Z(k_{n+1})=s_{j}\right] (57)
=𝖤[Ξ(n)αρρΔn+1|Z(kn)=si,Z(kn+1)=sj]\displaystyle=\mathsf{E}\left[\Xi(n)\frac{\alpha}{\rho}\rho^{\Delta_{n+1}}|Z(k_{n})=s_{i},Z(k_{n+1})=s_{j}\right]
=α𝖤[Ξ(n)|Z(kn)=si]𝖤[ρΔn+1|Z(kn)=si,Z(kn+1)=sj]\displaystyle=\alpha^{\prime}\mathsf{E}\left[\Xi(n)|Z(k_{n})=s_{i}\right]\mathsf{E}\left[\rho^{\Delta_{n+1}}|Z(k_{n})=s_{i},Z(k_{n+1})=s_{j}\right]
=α𝖤[Ξ(n)|Z(kn)=si]ri,j\displaystyle=\alpha^{\prime}\mathsf{E}\left[\Xi(n)|Z(k_{n})=s_{i}\right]r_{i,j}
=αri,ji=1S0𝖤[Ξ(n)|Z(kn1)=si,Z(kn)=si]\displaystyle=\alpha^{\prime}r_{i,j}\sum_{i^{\prime}=1}^{S_{0}}\mathsf{E}\left[\Xi(n)|Z(k_{n-1})=s_{i^{\prime}},Z(k_{n})=s_{i}\right]
×𝖯𝗋𝗈𝖻[Z(kn1)=si|Z(kn)=si]\displaystyle\hskip 99.58464pt\times\mathsf{Prob}\left[Z(k_{n-1})=s_{i^{\prime}}|Z(k_{n})=s_{i}\right]
=αri,ji=1S0𝖤[Ξ(n)|Z(kn1)=si,Z(kn)=si]fi,i,\displaystyle=\alpha^{\prime}r_{i,j}\sum_{i^{\prime}=1}^{S_{0}}\mathsf{E}\left[\Xi(n)|Z(k_{n-1})=s_{i^{\prime}},Z(k_{n})=s_{i}\right]{f_{i^{\prime},i}},

where ααρ\alpha^{\prime}\triangleq\frac{\alpha}{\rho} and

fi,i𝖯𝗋𝗈𝖻[Z(kn1)=si|Z(kn)=si]=πiπiv~i,i,i,i=1,,S0.f_{i^{\prime},i}\!\triangleq\!\!\mathsf{Prob}\left[Z(k_{n-1})\!=\!s_{i^{\prime}}|Z(k_{n})\!=\!s_{i}\right]\!=\!\frac{\pi_{i^{\prime}}}{\pi_{i}}\tilde{v}_{i^{\prime},i},\!\forall i^{\prime},i\!=\!1,\cdots,\!S_{0}. (58)

From (58), we define 𝐅S0×S0\mathbf{F}\in\mathbb{R}^{S_{0}\times S_{0}} such that [𝐅]i,j=fi,j[\mathbf{F}]_{i,j}=f_{i,j}, i.e.,

𝐅[𝐟1𝐟2𝐟S0]=𝐙𝐕~𝐙1,\mathbf{F}\triangleq\begin{bmatrix}\mathbf{f}_{1}&\!\!\!\framebox{}\!\!\!&\mathbf{f}_{2}&\!\!\!\framebox{}\!\!\!&\cdots&\!\!\!\framebox{}\!\!\!&\mathbf{f}_{S_{0}}\end{bmatrix}=\mathbf{Z}\tilde{\mathbf{V}}\mathbf{Z}^{-1}, (59)

where 𝐙=diag{𝝅}\mathbf{Z}=\text{diag}\{\bm{\pi}\} and 𝐟i[f1,i,f2,i,,fS0,i]S0\mathbf{f}_{i}\triangleq[f_{1,i},f_{2,i},\cdots,f_{S_{0},i}]^{\top}\in\mathbb{R}^{S_{0}}.

Let’s introduce 𝚵(n)S0×S0,n\mathbf{\Xi}(n)\in\mathbb{R}^{S_{0}\times S_{0}},\forall n\in\mathbb{N}, where

[𝚵(n)]i,j𝖤[Ξ(n)|Z(kn1)=si,Z(kn)=sj]\displaystyle\left[\bm{\Xi}(n)\right]_{i,j}\triangleq\mathsf{E}\left[\Xi(n)|Z(k_{n-1})=s_{i},Z(k_{n})=s_{j}\right] (60)

Let 𝝃i(n)[[𝚵(n)]1,i,[𝚵(n)]2,i,,[𝚵(n)]S0,i]\bm{\xi}_{i}(n)\triangleq\left[[\mathbf{\Xi}(n)]_{1,i},[\mathbf{\Xi}(n)]_{2,i},\cdots,[\mathbf{\Xi}(n)]_{S_{0},i}\right]^{\top}. From (57), we have

𝚵(n+1)=\displaystyle\mathbf{\Xi}(n+1)= (61)
[αr1,1𝐟1𝝃1(n)αr1,2𝐟1𝝃1(n)αr1,S0𝐟1𝝃1(n)αr2,1𝐟2𝝃2(n)αr2,2𝐟2𝝃2(n)αr2,S0𝐟2𝝃2(n)αrS0,1𝐟S0𝝃S0(n)αrS0,2𝐟S0𝝃S0(n)αrS0,S0𝐟S0𝝃S0(n)].\displaystyle\!\begin{bmatrix}\!\alpha^{\prime}{r_{1,1}}\mathbf{f}^{\top}_{1}\bm{\xi}_{1}(n)&\alpha^{\prime}{r_{1,2}}\mathbf{f}^{\top}_{1}\bm{\xi}_{1}(n)&\cdots&\alpha^{\prime}{r_{1,S_{0}}}\mathbf{f}^{\top}_{1}\bm{\xi}_{1}(n)\\ \!\alpha^{\prime}{r_{2,1}}\mathbf{f}^{\top}_{2}\bm{\xi}_{2}(n)&\alpha^{\prime}{r_{2,2}}\mathbf{f}^{\top}_{2}\bm{\xi}_{2}(n)&\cdots&\alpha^{\prime}{r_{2,S_{0}}}\mathbf{f}^{\top}_{2}\bm{\xi}_{2}(n)\\ \!\vdots&\vdots&\ddots&\vdots\\ \!\alpha^{\prime}{r_{S_{0},1}}\mathbf{f}^{\top}_{S_{0}}\bm{\xi}_{S_{0}}(n)&\alpha^{\prime}{r_{{S_{0}},2}}\mathbf{f}^{\top}_{S_{0}}\bm{\xi}_{S_{0}}(n)&\cdots&\alpha^{\prime}{r_{{S_{0}},{S_{0}}}}\mathbf{f}^{\top}_{S_{0}}\bm{\xi}_{S_{0}}(n)\end{bmatrix}\!\!.

After some algebraic transformation, it can be obtained that

𝐯(n+1)=𝐀𝐔𝐯(n),\mathbf{v}(n+1)=\mathbf{AU}\mathbf{v}(n), (62)

where

𝐯(n)[𝝃1(n),𝝃2(n),,𝝃S0(n)]S02,n,\mathbf{v}(n)\triangleq[\bm{\xi}_{1}(n)^{\top},\bm{\xi}_{2}(n)^{\top},\cdots,\bm{\xi}_{S_{0}}(n)^{\top}]^{\top}\in\mathbb{R}^{S^{2}_{0}},\forall n\in\mathbb{N}, (63)
𝐀diag{α,,α}S02,\mathbf{A}\triangleq\text{diag}\{\alpha^{\prime},\cdots,\alpha^{\prime}\}\in\mathbb{R}^{S^{2}_{0}}, (64)
𝐔[𝐔1𝐔2𝐔S0]S02×S02,\mathbf{U}\triangleq\begin{bmatrix}\mathbf{U}_{1}\\ \mathbf{U}_{2}\\ \vdots\\ \mathbf{U}_{S_{0}}\\ \end{bmatrix}\in\mathbb{R}^{S^{2}_{0}\times S^{2}_{0}}, (65)

and 𝐔i,i=1,,S0\mathbf{U}_{i},i=1,\cdots,S_{0}, is defined under (43).

From (62), it follows that

𝐯(n)=𝐀n1𝐔n1𝐯(1).\mathbf{v}(n)=\mathbf{A}^{n-1}\mathbf{U}^{n-1}\mathbf{v}(1). (66)

From [23, Lemma 2], for any arbitrarily small ϵ>0\epsilon>0, there exists μ>0\mu^{\prime}>0 and K>0K>0, such that

maxi,j{1,,S0}[𝐔n]i,j<μ(λmax(𝐔)+ϵ)n,n>K.\max_{i,j\in\{1,\cdots,S_{0}\}}[\mathbf{U}^{n}]_{i,j}<\mu^{\prime}\left(\lambda_{\max}(\mathbf{U})+\epsilon\right)^{n},\forall n>K. (67)

Taking (67) into (66), we have

𝖤[Ξ(n)]maxi{1,,S02}[𝐯(n)]i<μ(α(λmax(𝐔)+ϵ))n1,ϵ>0,n>K,\mathsf{E}\left[\Xi(n)\right]\!\leq\!\!\!\!\!\max_{i\in\{1,\cdots,S^{2}_{0}\}}[\mathbf{v}(n)]_{i}\!<\!\mu\Big{(}\!\alpha^{\prime}\left(\!\lambda_{\max}(\mathbf{U})+\epsilon\!\right)\!\!\Big{)}^{n-1},\forall\epsilon\!>\!0,n\!>\!K, (68)

where μ=μS02maxi{1,,S02}[𝐯(1)]i<\mu=\mu^{\prime}S^{2}_{0}\max_{i\in\{1,\cdots,S^{2}_{0}\}}[\mathbf{v}(1)]_{i}<\infty. This competes the proof.

B. Proof of Theorem 3

The proof follows the same steps of [10, Theorem 7.1] in analyzing the noise effect on the performance, since the analysis only depends on the process {Z(kn)}n0\{Z(k_{n})\}_{n\in{\mathbb{N}_{0}}}. From [10], it can be obtained that

𝖤[t=kn1kn1V(𝐱(t))]1+αρ1ρ𝖤[Ξ(n)]𝖤[V(𝐱(0))]\displaystyle\mathsf{E}\left[\sum_{t=k_{n-1}}^{k_{n}-1}V(\mathbf{x}(t))\right]\leq\frac{1+\alpha-\rho}{1-\rho}\mathsf{E}\left[\Xi(n)\right]\mathsf{E}\left[V(\mathbf{x}(0))\right] (69)
+1+αρ1ρ(𝖤[Ξ(n1)]+𝖤[Ξ(n2)]++1)w¯+w¯,\displaystyle+\frac{1+\alpha-\rho}{1-\rho}\left(\mathsf{E}\left[\Xi(n-1)\right]+\mathsf{E}\left[\Xi(n-2)\right]+\cdots+1\right)\bar{w}+\bar{w}^{\prime},

where Ξ(n)\Xi(n) is defined in (34), w¯=Ψ1𝖤[|𝐰(t)|]\bar{w}=\Psi_{1}\mathsf{E}\left[|\mathbf{w}(t)|\right], w¯=Ψ2𝖤[|𝐰(t)|]\bar{w}^{\prime}=\Psi_{2}\mathsf{E}\left[|\mathbf{w}(t)|\right], and Ψ1\Psi_{1} and Ψ2\Psi_{2} are bounded constant determined by βx\beta_{x}, βu\beta_{u}, βw\beta_{w}, βV\beta_{V}, βκ\beta_{\kappa}, ρ\rho, and η\eta (see [10, Appendix D] for the calculation of Ψ1\Psi_{1} and Ψ2\Psi_{2}).

Based on (69), Theorem 3 holds if we can prove

lim supni=1n𝖤[Ξ(i)]<.\limsup\limits_{n\rightarrow\infty}\sum_{i=1}^{n}\mathsf{E}\left[\Xi(i)\right]<\infty. (70)

Taking (43) into (70), for any arbitrarily small ϵ>0\epsilon>0, there exists μ,K>0\mu^{\prime},K^{\prime}>0 such that, n>K\forall n>K^{\prime}, the following are satisfied

i=1n𝖤[Ξ(i)]\displaystyle\sum_{i=1}^{n}\mathsf{E}\left[\Xi(i)\right] μ((αρ(λmax(𝐔)+ϵ))n2\displaystyle\leq\mu^{\prime}\left(\Big{(}\frac{\alpha}{\rho}\left(\lambda_{\max}(\mathbf{U})+\epsilon\right)\Big{)}^{n-2}\right. (71)
+(αρ(λmax(𝐔)+ϵ))n3++1)+c,\displaystyle\left.\hskip 14.22636pt+\Big{(}\frac{\alpha}{\rho}\left(\lambda_{\max}(\mathbf{U})+\epsilon\right)\Big{)}^{n-3}+\cdots+1\right)+c,

where cc is a constant. Then, it is clear that (70) holds if αρ(λmax(𝐔))<1\frac{\alpha}{\rho}\left(\lambda_{\max}(\mathbf{U})\right)<1.