This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Resilient Average Consensus with Adversaries via Distributed Detection and Recovery

Liwei Yuan    \IEEEmembershipMember, IEEE and Hideaki Ishii    \IEEEmembershipFellow, IEEE This work was supported in the part by JSPS under Grant-in-Aid for Scientific Research Grant No. 22H01508, and in the part by the Yuelushan Center for Industrial Innovation under Grant 2023YCII0102. L. Yuan is with the College of Electrical and Information Engineering, Hunan University, Changsha, 410082, China (e-mail: yuanliwei@hnu.edu.cn). H. Ishii is with the Department of Information Physics and Computing, The University of Tokyo, Tokyo, 113-8656, Japan (e-mail: hideaki_ishii@ipc.i.u-tokyo.ac.jp).
Abstract

We study the problem of resilient average consensus in multi-agent systems where some of the agents are subject to failures or attacks. The objective of resilient average consensus is for non-faulty/normal agents to converge to the average of their initial values despite the erroneous effects from malicious agents. To this end, we propose a successful distributed iterative resilient average consensus algorithm for the multi-agent networks with general directed topologies. The proposed algorithm has two parts at each iteration: detection and averaging. For the detection part, we propose two distributed algorithms and one of them can detect malicious agents with only the information from direct in-neighbors. For the averaging part, we extend the applicability of an existing averaging algorithm where normal agents can remove the effects from malicious agents so far, after they are detected. Another important feature of our method is that it can handle the case where malicious agents are neighboring and collaborating with each other to mislead the normal ones from averaging. This case cannot be solved by existing detection approaches in related literature. Moreover, our algorithm is efficient in storage usage especially for large-scale networks as each agent only requires the values of neighbors within two hops. Lastly, numerical examples are given to verify the efficacy of the proposed algorithms.

{IEEEkeywords}

Average consensus, directed topologies, distributed detection, resilient consensus.

1 Introduction

\IEEEPARstart

Distributed consensus in multi-agent systems is a fundamental and well-studied topic across different research areas including systems control, computer science, and communication [2, 1, 3, 4]. Under this broad topic, a particular problem that has been extensively studied is that of average consensus where agents try to reach consensus on the average of their values through local interactions among nearby agents [9, 10, 6, 7, 5, 11, 8, 12, 13]. Average consensus algorithms are also useful to maintain the total of the resources invariant and have found applications in, e.g., economic dispatch problems for power systems [14], distributed computation of PageRank for the search engine of Google [15, 16]. As concerns for cyber-security sharply rise in our society, consensus protocols that properly function even in the presence of faults and adversarial agents have been actively studied; see, e.g., [20, 21, 19, 17, 18]. The objective is for the non-faulty/normal agents to reach consensus without being affected by the misbehaviors of adversarial agents. In this context, resilient algorithms for performing average consensus have remained somewhat limited except for the recent works [22, 23, 24]. A major challenge is that, normal agents should reach consensus on the exact average of their initial values despite adversarial agents’ misbehaviors, which may include adding erroneous values to the normal agents’ values during the interactions with normal neighbors.

In this paper, we propose an iterative distributed algorithm to tackle the resilient average consensus (RAC) problem in general directed networks under the attacks by the so-called malicious agents. Such an agent is capable to send arbitrary but identical values to its neighbors at each iteration [25, 26, 19]. This is the typical way of communication in broadcast networks [27]. There are basically two types of approaches for handling the resilient consensus problem, where normal agents need to reach a common value but not necessarily the average of the initial values: (i) mean subsequence reduced (MSR) algorithms [28, 29, 19, 31, 32, 30, 33, 34, 35] and (ii) detection and isolation algorithms [37, 36]. In MSR algoirthms, agents utilize only the values in a time-varying safety interval to update their next values, with no capabilities to recognize whether a neighbor is adversarial or not. On the other hand, in detection and isolation algorithms, agents detect the neighbors violating the given consensus protocol and remove the values of such neighbors for updating their next values. This property makes the detection approach a good basis for our RAC algorithm. The reason is that the information of identities of normal agents must be known by the algorithms, which is the key to accumulate the values of normal agents for averaging.

Table 1: Comparisons with Related Resilient Averaging Works
Algorithm 3 [38] [23] [22]
Network
type
Directed Undirected Undirected Directed
Adversary
type
Malicious Malicious Malicious Byzantine
Neighboring
adversaries
Yes No No Yes
Communication
range
Two-hop Two-hop Two-hop Flooding

Our RAC algorithm is based on the detection approach and has two parts: detection and averaging. Existing related works for the RAC problem share this structure [22, 23, 38]. However, our method has certain advantages over them in different aspects as listed in Table I. More specifically, the work [22] proposed a secure broadcast and retrieval algorithm for the RAC problem in directed networks. There, each normal agent uses a certified propagation algorithm to broadcast its initial value to all agents and retrieve the initial values of normal agents for averaging. This approach would cost a huge amount of storage and time for collecting the values of all normal agents in a large-scale network. The work [23] proposed a detection and compensation algorithm for the RAC problem in undirected networks. It utilizes the two-hop neighbors’ information to detect misbehaving neighbors and it requires a doubly stochastic adjacent matrix for averaging. As a result, their algorithm is applied in undirected networks only and also cannot handle the case where malicious agents are neighboring with each other. Recently, the authors of [24] proposed an RAC algorithm for directed networks. It allows normal agents to dynamically remove or add the values received from neighbors, however, with the assumption that each normal agent can have access to a correct detection of neighbors. Then in [39, 38], the same authors proposed a detection and compensation algorithm for RAC problem in undirected networks. However, their detection requires the direct communication with two-hop neighbors and it cannot handle the case of neighboring malicious agents either.

In [36], we proposed a secure detection algorithm for resilient consensus, where each normal agent acts as a detector of its neighbors. An important feature is that it can guarantee the fully distributed detection of malicious neighbors in general directed networks. Besides, it is able to tackle the case of neighboring malicious agents. This is accomplished through the majority voting [41, 40] under a certain topology requirement on the network. In this paper, we exploit these properties and develop a novel RAC algorithm based on the two-hop detection approach.

The contributions of this paper are summarized as follows. We propose a novel RAC algorithm under which normal agents can iteratively detect malicious neighbors and converge to the average of their initial values in general directed networks. The proposed algorithm consists of the detection part and the averaging part. Specifically, for the averaging part, we employ the running-sum based algorithm from [24], where each node has local buffers to store the total effects received from its in-neighbors. It allows the normal nodes to precisely recover from the influence of malicious neighbors once any misbehavior is detected. We also improve the applicability of the averaging algorithm by relaxing the necessary assumptions in [24]. In particular, it is sufficient for normal agents to access the correct detection of only in- and out-neighbors for our RAC algorithm, which can save storage resources. Furthermore, we extend the class of misbehaviors of the malicious nodes and consider scenarios where they may go beyond manipulating their identities and also remain to act normally at all times.

For the detection part, we propose two novel algorithms which allow normal nodes to monitor their neighbors and detect as soon as malicious agents perform any misbehaviors in the messages that they broadcast. The fundamental idea is to exploit the two-hop communication so that the normal agents have access to the inputs of their neighbors. This will enable them to obtain multiple reconstructed versions of the outputs of their neighbors and then to compare them to find the true outputs. The difference between the two algorithms lies in the levels in the capabilities for the normal agents to share the detection information among themselves. The first algorithm assumes the availability of authenticated mobile detectors, which help to reduce the requirement on the network connectivity. It will be referred to as the sharing detection algorithm. Our second algorithm is more significant in that it can be implemented in a fully distributed fashion in our RAC algorithm. Here, each normal node is able to acquire all the inputs of an in-neighbor through the majority voting under a necessary graph structure. Besides, it obtains the detection information of any two-hop in-neighbor (in-neighbor’s in-neighbor) by the same approach. As a result, normal nodes can independently detect all the malicious neighbors violating the given averaging algorithm in general directed networks.

Both detection algorithms can handle the case of neighboring malicious nodes, which cannot be solved by related works for the RAC problem [23, 39, 38]. Moreover, we provide tight graph conditions for our algorithms to achieve the detection and averaging functions, respectively. We also prove that the graph condition for the fully distributed detection algorithm can be simplified for undirected networks, which makes it more convenient to check whether a graph meets the condition or not. We emphasize that although the topology requirement may be dense, we can generate the directed/undirected network topologies that satisfy our conditions in large scale. Lastly, we provide extensive examples to show the efficacy of our RAC algorithm in large-scale networks as well as in an extreme adversarial situation, where over half of the nodes in the network are compromised by malicious attackers.

The rest of this paper is organized as follows. Section II outlines preliminaries on graph notions and the problem settings. Section III presents the novel RAC algorithm with an emphasis on the averaging part. Sections IV and V present the sharing detection algorithm and the fully distributed detection algorithm, respectively. Moreover, tight graph conditions for the proposed algorithms to achieve resilient average consensus are proved. Section VI provides numerical examples to demonstrate the efficacy of the proposed algorithms. Finally, Section VII concludes the paper.

2 Preliminaries and Problem Setting

In this section, we present preliminaries on graph theory, the average consensus algorithm, and the problem settings.

2.1 Graph Notions

Consider the directed graph 𝒢=(𝒱,)\mathcal{G}=(\mathcal{V},\mathcal{E}) consisting of the node set 𝒱={1,,n}\mathcal{V}=\{1,...,n\} and the edge set 𝒱×𝒱\mathcal{E}\subset\mathcal{V}\times\mathcal{V}. Here, the edge (j,i)(j,i)\in\mathcal{E} indicates that node ii can receive information from node jj. Node jj is said to be an in-neighbor of node ii, and node ii is an out-neighbor of node jj. The sets of in-neighbors and out-neighbors of node ii are denoted by 𝒩i={j𝒱:(j,i)}\mathcal{N}_{i}^{-}=\{j\in\mathcal{V}:\,(j,i)\in\mathcal{E}\} and 𝒩i+={j𝒱:(i,j)}\mathcal{N}_{i}^{+}=\{j\in\mathcal{V}:\,(i,j)\in\mathcal{E}\}, respectively. The in-degree and out-degree of node ii are given by di=|𝒩i|d_{i}^{-}=\left|\mathcal{N}_{i}^{-}\right| and di+=|𝒩i+|d_{i}^{+}=\left|\mathcal{N}_{i}^{+}\right|, respectively. Here, |𝒮|\left|\mathcal{S}\right| is the cardinality of a finite set 𝒮\mathcal{S}. If the graph 𝒢\mathcal{G} is undirected, the edge (j,i)(j,i)\in\mathcal{E} indicates (i,j)(i,j)\in\mathcal{E}. A complete graph 𝒦n=(𝒱,)\mathcal{K}_{n}=(\mathcal{V},\mathcal{E}) is defined by ={(i,j)𝒱×𝒱:ij}\mathcal{E}=\{(i,j)\in\mathcal{V}\times\mathcal{V}:i\neq j\}. A path from node i1i_{1} to imi_{m} is a sequence of distinct nodes (i1,i2,,im)(i_{1},i_{2},\dots,i_{m}), where (ij,ij+1)(i_{j},i_{j+1})\in\mathcal{E} for j=1,,m1j=1,\dots,m-1. It is also referred to as an (m1)(m-1)-hop path. We say that node imi_{m} is reachable from node i1i_{1}. Node i1i_{1} is an (m1)(m-1)-hop in-neighbor of node imi_{m}. A directed graph 𝒢\mathcal{G} is said to be strongly connected111An undirected graph is simply said to be connected if every node is reachable from every other node. if every node is reachable from every other node. An undirected graph 𝒢\mathcal{G} is said to be kk^{\prime}-connected if it contains at least k+1k^{\prime}+1 nodes and does not contain a set of k1k^{\prime}-1 nodes whose removal disconnects the graph.

2.2 Average Consensus and the Running-sum Algorithm

The problem of multi-agent average consensus can be described as follows: Consider a system with nn agents interacting over the network modeled by the directed graph 𝒢=(𝒱,)\mathcal{G}=(\mathcal{V},\mathcal{E}). Each agent i𝒱i\in\mathcal{V} has a scalar state xi[k]x_{i}[k]\in\mathbb{R} to be updated over time k0k\in\mathbb{Z}_{\geq 0}. The goal is to design distributed algorithms that allow agents to eventually converge to the average value of their initial states X¯=1ni=1nxi[0]\overline{X}=\frac{1}{n}\sum_{i=1}^{n}x_{i}[0], where each agent utilizes only the local information from their neighboring agents during the consensus forming. The push-sum ratio consensus algorithm [42] was proposed to achieve this goal through two iterative processes on each agent. Here, we describe this algorithm for the time-varying graph 𝒢[k]=(𝒱,[k])\mathcal{G}[k]=(\mathcal{V},\mathcal{E}[k]), where [k]\mathcal{E}[k]\subseteq\mathcal{E}. Denote the set of out-neighbors of agent ii at time kk by 𝒩i+[k]\mathcal{N}_{i}^{+}[k] and the out-degree by di+[k]=|𝒩i+[k]|d_{i}^{+}[k]=|\mathcal{N}_{i}^{+}[k]|; we employ similar notations for the set of in-neighbors 𝒩i[k]\mathcal{N}_{i}^{-}[k] and the in-degree di[k]d_{i}^{-}[k].

We first introduce the push-sum algorithm, which is the basis of the running-sum algorithm. Each node ii has two state variables, yi[k]y_{i}[k] and zi[k]z_{i}[k], and updates them as

yi[k+1]\displaystyle y_{i}[k+1] =j𝒩i[k]{i}yj[k]1+dj+[k],\displaystyle=\sum_{j\in\mathcal{N}_{i}^{-}[k]\cup\{i\}}\frac{y_{j}[k]}{1+d_{j}^{+}[k]}, (1)
zi[k+1]\displaystyle z_{i}[k+1] =j𝒩i[k]{i}zj[k]1+dj+[k],\displaystyle=\sum_{j\in\mathcal{N}_{i}^{-}[k]\cup\{i\}}\frac{z_{j}[k]}{1+d_{j}^{+}[k]},

where yi[0]=xi[0]y_{i}[0]=x_{i}[0] and zi[0]=1z_{i}[0]=1 for i𝒱i\in\mathcal{V}. The algorithm requires each node ii to know its out-degree di+[k]d_{i}^{+}[k], and transmit to each out-neighbor the values

y¯i[k]:=yi[k]1+di+[k],z¯i[k]:=zi[k]1+di+[k].\overline{y}_{i}[k]\medspace\medspace:=\frac{y_{i}[k]}{1+d_{i}^{+}[k]},\medspace\medspace\overline{z}_{i}[k]\medspace\medspace:=\frac{z_{i}[k]}{1+d_{i}^{+}[k]}. (2)

Then, by (1), these out-neighbors take the sum of received values as their new values.

At each time kk, node ii calculates the ratio

ri[k]:=yi[k]zi[k].r_{i}[k]:=\frac{y_{i}[k]}{z_{i}[k]}.

Under some joint connectivity assumptions on the union of the underlying graphs in a certain time window, it was reported in, e.g., [24] that ri[k]r_{i}[k] asymptotically converges to the average of the initial values, i.e.,

limkri[k]=jyj[0]jzj[0]=X¯,i𝒱.\lim_{k\to\infty}r_{i}[k]=\frac{\sum_{j}y_{j}[0]}{\sum_{j}z_{j}[0]}=\overline{X},\medspace\medspace\forall i\in\mathcal{V}. (3)

Now, the running-sum ratio consensus algorithm is a variation of the push-sum algorithm used to overcome packet drops or unknown out degrees [11]. It can be summarized as follows. At each time kk, node ii does not send y¯i[k]\overline{y}_{i}[k], z¯i[k]\overline{z}_{i}[k] in (2) to its out-neighbors. Instead, it sends the so-called yy and zz running sums denoted by λ\lambda and γ\gamma, respectively. The two values contain the information of y¯i[k]\overline{y}_{i}[k] and z¯i[k]\overline{z}_{i}[k], and are defined as

λi[k+1]:=t=0ky¯i[t],γi[k+1]:=t=0kz¯i[t].\lambda_{i}[k+1]\medspace\medspace:=\sum\limits_{t=0}\limits^{k}{\overline{y}_{i}[t]},\medspace\medspace\gamma_{i}[k+1]\medspace\medspace:=\sum\limits_{t=0}\limits^{k}{\overline{z}_{i}[t]}. (4)

Therefore, an out-neighbor obtains node ii’s values y¯i[k]\overline{y}_{i}[k], z¯i[k]\overline{z}_{i}[k] by taking the difference of two consecutive λi[k]\lambda_{i}[k], γi[k]\gamma_{i}[k] as

y¯i[k]\displaystyle\overline{y}_{i}[k] =λi[k+1]λi[k],\displaystyle=\lambda_{i}[k+1]-\lambda_{i}[k],
z¯i[k]\displaystyle\overline{z}_{i}[k] =γi[k+1]γi[k].\displaystyle=\gamma_{i}[k+1]-\gamma_{i}[k].

Thus, the running-sum algorithm can achieve average consensus as the push-sum algorithm does, with additional bookkeeping procedures at each node.

Next, we formally outline the structure of the running-sum ratio consensus algorithm [11]. At each time kk, node ii maintains two kinds of values: (i) the running-sum values λi[k]\lambda_{i}[k] and γi[k]\gamma_{i}[k] of its own; and (ii) the two incoming running-sums from each in-neighbor jj. More specifically, node ii uses δij[k]\delta_{ij}[k] and ωij[k]\omega_{ij}[k] to keep track of the yy and zz running sums from node jj, respectively. They are given as

δij[k]\displaystyle\delta_{ij}[k] =λj[k],δij[0]=0,\displaystyle=\lambda_{j}[k],\medspace\medspace\delta_{ij}[0]=0, (5)
ωij[k]\displaystyle\omega_{ij}[k] =γj[k],ωij[0]=0.\displaystyle=\gamma_{j}[k],\medspace\medspace\omega_{ij}[0]=0.

2.3 Update Rule and Threat Model

We now introduce the model of the adversaries and the general structure of the proposed resilient algorithm. First, the node set 𝒱\mathcal{V} is partitioned into the set of normal nodes 𝒩\mathcal{N} and the set of adversary nodes 𝒜\mathcal{A}. The latter set is unknown to the normal nodes at time k=0k=0. The adversary nodes in 𝒜\mathcal{A} try to prevent the normal nodes in 𝒩\mathcal{N} from reaching average consensus. All algorithms in this paper are synchronous.

In our problem setting, the adversary nodes can be quite powerful. We assume that they may behave arbitrarily, deviating from the protocols with which the normal nodes are equipped. Here, we define the threat model of this paper; see also [19, 22, 36, 39].

Definition 1

(ff-total / ff-local set) The set of adversary nodes 𝒜\mathcal{A} is said to be ff-total if it contains at most ff nodes, i.e., |𝒜|f\left|\mathcal{A}\right|\leq f. Similarly, it is said to be ff-local if for any normal node i𝒩i\in\mathcal{N}, it has at most ff adversary in-neighbors, i.e., |𝒩i𝒜|f,i𝒩\left|\mathcal{N}_{i}^{-}\cap\mathcal{A}\right|\leq f,\forall i\in\mathcal{N}.

Definition 2

(Malicious nodes) An adversary node i𝒜i\in\mathcal{A} is said to be malicious if it changes its own value arbitrarily and sends the same value222It may also decide not to make a transmission at any time. This corresponds to the crash model [1]. to its neighbors at each transmission.

In this paper, we focus on the malicious model. This model is reasonable in applications such as wireless sensor networks and robotic networks, where neighbors’ information is obtained by broadcast communication or vision sensors [27]. This model is different from the Byzantine model, which is well-studied in computer science [1]. Specifically, a Byzantine node can send different values to its different neighbors. Here, we define a connectivity notion for directed graphs. A directed graph 𝒢\mathcal{G} is said to be kk^{\prime}-strongly connected if after removing any set of nodes satisfying the (k1k^{\prime}-1)-local model, the remaining digraph is strongly connected.

As mentioned in the Introduction, the proposed algorithm for resilient average consensus is based on detection of the malicious nodes in the network. To this end, each normal node ii is equipped with a detection algorithm to monitor the behaviors of its own neighbors. The output of such an algorithm will be the set of malicious nodes known or detected by node ii by time kk and is denoted by 𝒜i[k]\mathcal{A}_{i}[k].

The overall structure of the proposed algorithm is as follows. At each time kk, each normal node ii forms an information set denoted by Φi[k]\Phi_{i}[k]. This set will be shared with its out-neighbors, who will make use of it for their averaging and detection algorithms. The exact contents of the information sets will be given in the next subsection. Specifically, node ii conducts the four steps given below at time k+1k+1:

1. Transmit the information set Φi[k]\Phi_{i}[k] (described in (6) later) and the detection information 𝒜i[k]\mathcal{A}_{i}[k] to all its out-neighbors j𝒩i+j\in\mathcal{N}_{i}^{+}.

2. Receive the information sets Φj[k]\Phi_{j}[k] and the detection information 𝒜j[k]\mathcal{A}_{j}[k] from all in-neighbors j𝒩ij\in\mathcal{N}_{i}^{-}.

3. Detect neighbors according to the detection algorithm to obtain 𝒜i[k+1]\mathcal{A}_{i}[k+1].

4. Update xi[k+1]x_{i}[k+1] according to the resilient average consensus algorithm.

The RAC algorithm in Step 4 will be outlined in Section 3 whereas the detection algorithm in Step 3 will be given in Sections 4 and 5.

2.4 Detection of Adversaries and Information Sets

We now describe the general approach for our detection algorithms, based on the ideas from [36]. As mentioned above, each normal node monitors its neighboring nodes and checks if any inconsistencies can be found in their behaviors. In particular, our approach employs two-hop communication among the nodes. That is, each node sends the information received from its direct in-neighbors to its out-neighbors together with its own information. We assume that each node receives information from its two-hop in-neighbors via a sufficient number of different paths. Then, if any of its direct in-neigbors make changes in the information to be passed on, there will be inconsistencies in the data, which can lead to detections of misbehaviors. To formalize this approach, in this subsection, we first introduce the key notion of information sets of the nodes and then provide assumptions regarding these sets for both normal and malicious nodes.

Information sets define the data exchanged within the network for performing detection and averaging. Node ii’s information set Φi[k]\Phi_{i}[k] to be broadcasted at time k+1k+1 is

Φi[k]=\displaystyle\Phi_{i}[k]= (𝒜i[k],(i,δii[k+1|k],ωii[k+1|k]),\displaystyle\Big{(}\mathcal{A}_{i}[k],(i,\delta_{ii}[k+1|k],\omega_{ii}[k+1|k]), (6)
{(j,δij[k|k],ωij[k|k])}j𝒩i{i}).\displaystyle\{(j,\delta_{ij}[k|k],\omega_{ij}[k|k])\}_{j\in\mathcal{N}_{i}^{-}\cup\{i\}}\Big{)}.

It has three parts. The first is the set 𝒜i[k]\mathcal{A}_{i}[k] of adversaries detected by node ii by time kk. The second and the third are node ii’s own and its in-neighbors’ information. We use the notation δii[k+1|k]\delta_{ii}[k+1|k] to indicate that this value is in the set Φi[k]\Phi_{i}[k] from time kk. Note that Φi[k1]\Phi_{i}[k-1] and Φi[k]\Phi_{i}[k] contain δii[k|k1]\delta_{ii}[k|k-1] and δii[k|k]\delta_{ii}[k|k], respectively, and if node ii is malicious, these values may be different.

Next, we introduce assumptions on the nodes’ knowledge and the attacks generated by the malicious nodes.

Assumption 1

Each node i𝒩i\in\mathcal{N} has access to the information sets received from its in-neighbors. It knows the indices and topology of its two-hop in-neighbors and those of its direct out-neighbors.

Assumption 2

Each node i𝒜i\in\mathcal{A} may have all the information of the entire network including the topology and state values of all nodes and may cooperate with other malicious nodes even if no edges exist. It can manipulate its own information set in (6) and broadcasts the same set to out-neighbors.

By Assumption 1, each normal node has only partial knowledge about the network. To perform detection based on two-hop communication, normal nodes are aware of the topology of two-hop in-neighbors. This setting may be justified in sensor networks when the nodes are geographically fixed and the network topology remain the same. Similar settings are studied in [23, 39, 38]. We should highlight that this assumption can be met relatively easily and is of low cost. In MSR-based resilient consensus algorithms [32, 19], it is sufficient that fault-free nodes have access to the information only from their one-hop neighbors. Clearly, this requirement is weaker than Assumption 1, but MSR-based algorithms are not capable to detect malicious agents (though they can avoid their influences). Also, in contrast, each fault-free node must know the topology of the entire network in related works based on observer-based detection [20], [25], multi-hop communication [1], [31], and Byzantine agreement [43].

On the other hand, in Assumption 2, since two-hop communication is employed, a malicious node may modify not only its own states but also those received from its in-neighbors, which are part of its information set. This means that there are more options in terms of attacks compared to, e.g., the MSR-based algorithms. However, we emphasize that such attacks can be detected. For example, a malicious node may add or delete some pairs of agent IDs and values in its information set. It may also decide to remove information from some of its in-neighbors. Since the normal agents have the knowledge of up to their two-hop in-neighbors, attacks will be found by their direct out-neighbors. Moreover, in the case that a malicious node adopts an ID of another node, such attacks can be detected too [36]. Therefore, our approach does not assume that each node should identify the senders of incoming messages, which is imposed in [24, 38].

3 Resilient Average Consensus

In this section, we define the RAC problem and introduce our algorithm with an emphasis on the averaging part.

3.1 Problem Statement

Consider a time-invariant directed graph 𝒢=(𝒱,)\mathcal{G}=(\mathcal{V},\mathcal{E}). In our resilient average consensus (RAC) problem, the goal is to design distributed algorithms that allow normal agents to eventually converge to the average value of their initial states, i.e.,

xi[k]X¯𝒩0:=i𝒩0xi[0]|𝒩0|ask,i𝒩,x_{i}[k]\to\overline{X}_{\mathcal{N}_{0}}:=\frac{\sum_{i\in\mathcal{N}_{0}}x_{i}[0]}{|\mathcal{N}_{0}|}\medspace\medspace\textup{as}\medspace\medspace k\to\infty,\medspace\medspace\forall i\in\mathcal{N}, (7)

regardless of the adversarial actions taken by the nodes in 𝒜\mathcal{A}. In (7), the resilient average consensus is not defined on the true set 𝒩\mathcal{N} of normal agents but on the set 𝒩0𝒩\mathcal{N}_{0}\supseteq\mathcal{N} of all nodes that behave properly over time. This is justified for the case where an adversary node acts as normal for all times. In this case, such an adversary agent’s value is included in the average computing since there is no way to detect such adversary nodes. See Section 5.5 for more discussions.

Similar problems have been studied in related works [22, 23, 38]. However, our approach has advantages over these works in different aspects as we discuss in due course.

3.2 Overview of the RAC Algorithm

To solve the RAC problem in a distributed iterative fashion, normal agents must know whether their neighbors are malicious or normal. Thereafter, they only interact with normal ones to obtain the desired average. Hence, our RAC algorithm contains two parts: (i) detection and (ii) averaging. The detection algorithm guarantees that each normal node can detect any malicious in-/out-neighbors. On the other hand, the averaging algorithm needs to ensure that each normal node can remove the erroneous effects received from malicious neighbors by the time those neighbors are detected as malicious.

In this section, we first present the averaging algorithm based on the RAC approach of [24], where each normal node is assumed to have access to the correct detection information of all normal nodes in the network. Note that the detection approach is not discussed in [24]. For ease of presentation, we assume that every normal node can obtain the correct detection of malicious neighbors by a certain time kck_{c}. Our detection algorithms presented later in Sections 4 and 5 are tailored for working with this averaging algorithm and realize the important function of correct detection.

Recall that in the running-sum algorithm, each agent maintains two variables λ\lambda and γ\gamma to record the sum of its own yy and zz values from the initial time. This feature makes it a good basis for our RAC algorithm. Moreover, for the running-sum algorithm to achieve average consensus, the adjacency matrix needs to be column stochastic, which is easy to realize in directed networks. In contrast, the related resilient averaging works for undirected networks [23, 38] are based on average consensus via linear iterations [2], which require the adjacency matrix to be doubly stochastic. However, it may be difficult to design such an adjacency matrix for directed networks, and even tougher for time-varying networks.

Next, we introduce the major steps of our RAC algorithm. At each time kk, each normal node ii utilizes the detection algorithm to update its detection information regarding in-/out-neighbors in 𝒜i[k]\mathcal{A}_{i}[k]. Then it updates the set of non-faulty in-neighbors as i[k]=𝒩i𝒜i[k]\mathcal{M}_{i}^{-}[k]=\mathcal{N}_{i}^{-}\setminus\mathcal{A}_{i}[k] and updates the set of non-faulty out-neighbors as i+[k]=𝒩i+𝒜i[k]\mathcal{M}_{i}^{+}[k]=\mathcal{N}_{i}^{+}\setminus\mathcal{A}_{i}[k]. Simultaneously, the out-degree is updated by di+[k]=|i+[k]|=|𝒩i+𝒜i[k]|d_{i}^{+}[k]=|\mathcal{M}_{i}^{+}[k]|=|\mathcal{N}_{i}^{+}\setminus\mathcal{A}_{i}[k]|. Given the new 𝒜i[k]\mathcal{A}_{i}[k], node ii updates its yy and zz using only the running sums from the in-neighbors in i[k]{i}\mathcal{M}_{i}^{-}[k]\cup\{i\}:

yi[k]\displaystyle y_{i}[k] =ji[k]{i}(δij[k]δij[k1]),\displaystyle=\sum\limits_{j\in\mathcal{M}_{i}^{-}[k]\cup\{i\}}{(\delta_{ij}[k]-\delta_{ij}[k-1])}, (8)
zi[k]\displaystyle z_{i}[k] =ji[k]{i}(ωij[k]ωij[k1]).\displaystyle=\sum\limits_{j\in\mathcal{M}_{i}^{-}[k]\cup\{i\}}{(\omega_{ij}[k]-\omega_{ij}[k-1])}.

By the assumption that every normal node obtains the correct detection of malicious neighbors by time kck_{c}, eq. (8) constrains the averaging within only normal nodes for time k>kck>k_{c}. Therefore, the running-sum algorithm on normal nodes achieves ratio consensus of the values of normal agents at time kck_{c} if the subgraph of normal nodes (i.e., the normal network) is strongly connected. However, due to possible erroneous effects from malicious neighbors, the sum of values of normal agents at time kck_{c} may not be the sum of initial values of normal agents. Thus, if the erroneous effects from malicious neighbors can be subtracted by normal agents and normal agents’ values sent to malicious neighbors can be compensated precisely, then normal agents can recover the sum of their initial values and achieve resilient average consensus.

3.3 Removing Malicious Effects Based on Detection

In this part, we introduce how the normal agents conduct the subtraction of in-coming malicious values and compensation of out-going normal values, respectively. The two actions are different for the cases where in-neighbors or out-neighbors are detected as malicious for the first time. Note that the actions taken by each node ii are based on its own detection 𝒜i[k]\mathcal{A}_{i}[k].

Case 1: A malicious in-neighbor jj is detected for the first time at time kk. In this case, node ii not only ignores node jj’s values for updating as in (8) but also removes the effects received from node jj so far, i.e., δij[k],ωij[k]\delta_{ij}[k],\omega_{ij}[k] in (5). This subtraction of in-coming malicious values has to be done for each in-neighbor jj in the set Δi[k]=i[k1]i[k]\Delta\mathcal{M}_{i}^{-}[k]=\mathcal{M}_{i}^{-}[k-1]\setminus\mathcal{M}_{i}^{-}[k], which consists of node ii’s in-neighbors that are detected as malicious at time kk. Specifically, we replace yi[k]y_{i}[k] and zi[k]z_{i}[k], respectively, with

yi[k]\displaystyle y_{i}[k] =yi[k]jΔi[k]δij[k1],\displaystyle=\medspace\medspace y_{i}[k]-\sum\limits_{j\in\Delta\mathcal{M}_{i}^{-}[k]}{\delta_{ij}[k-1]}, (9)
zi[k]\displaystyle z_{i}[k] =zi[k]jΔi[k]ωij[k1].\displaystyle=\medspace\medspace z_{i}[k]-\sum\limits_{j\in\Delta\mathcal{M}_{i}^{-}[k]}{\omega_{ij}[k-1]}.

Case 2: A malicious out-neighbor qq is detected for the first time at time kk. In this case, node ii not only decreases its out-degree by one (di+[k]=di+[k1]1d_{i}^{+}[k]=d_{i}^{+}[k-1]-1) but also compensates for all its own values sent to node qq while qq was considered normal. It does so by adding to its yy and zz values its own yy and zz running sums (λi[k],γi[k]\lambda_{i}[k],\gamma_{i}[k]), respectively. Similar to Case 1, this adjustment has to be done for every out-neighbor that is detected as malicious at time kk. Let Δi+[k]=i+[k1]i+[k]\Delta\mathcal{M}_{i}^{+}[k]=\mathcal{M}_{i}^{+}[k-1]\setminus\mathcal{M}_{i}^{+}[k] be the set of node ii’s out-neighbors that are detected as malicious at time kk. Then yi[k]y_{i}[k] and zi[k]z_{i}[k] are updated as

yi[k]\displaystyle y_{i}[k] =yi[k]+|Δi+[k]|λi[k],\displaystyle=\medspace\medspace y_{i}[k]+|\Delta\mathcal{M}_{i}^{+}[k]|\lambda_{i}[k], (10)
zi[k]\displaystyle z_{i}[k] =zi[k]+|Δi+[k]|γi[k].\displaystyle=\medspace\medspace z_{i}[k]+|\Delta\mathcal{M}_{i}^{+}[k]|\gamma_{i}[k].

This is needed because, e.g., λi[k]\lambda_{i}[k] is the cumulative yy values that were sent to any malicious out-neighbor by time kk.

3.4 Convergence Analysis

Finally, we are ready to present our RAC algorithm in Algorithm 1. By Assumption 1, node ii knows the sets 𝒩i\mathcal{N}_{i}^{-}, 𝒩i+\mathcal{N}_{i}^{+}. Moreover, through our detection algorithms, node ii keeps track of the set of misbehaving in-/out-neighbors 𝒜i[k]\mathcal{A}_{i}[k] that it detected by time kk. Then, by following the above two processes for malicious in-/out-neighbors, resilient average consensus can be achieved with Algorithm 1. The following proposition is the main convergence result for this algorithm from [24] with some enhanced applicability.

Proposition 1

Consider the directed network 𝒢=(𝒱,)\mathcal{G}=(\mathcal{V},\mathcal{E}), where each node i𝒱i\in\mathcal{V} has an initial value xi[0]x_{i}[0]. Under Assumptions 1 and 2, if each normal node can detect all the malicious in- and out-neighbors, and the normal network is strongly connected, then the normal nodes executing Algorithm 1 converge to the average of their initial values given by X¯𝒩0=i𝒩0xi[0]|𝒩0|\overline{X}_{\mathcal{N}_{0}}=\frac{\sum_{i\in\mathcal{N}_{0}}x_{i}[0]}{|\mathcal{N}_{0}|} in (7) as kk\to\infty.

Remark 1

The convergence results of this proposition have appeared in [24]; hence, we omit the proof for brevity. The key idea is to show that the sums of the yy and zz values of normal agents remain invariant at all times. Moreover, it is emphasized that we have improved the results for the averaging algorithm by relaxing the required assumptions as well as justifying the case where adversary agents act normally at all times. Specifically, the results in [24] need the assumption that each transmission of any node i𝒱i\in\mathcal{V} is associated with a unique node ID that allows the receiver to identify the sender. In contrast, we have discussed in Section 2.4 that we do not assume that each malicious node must send its real ID as the misbehavior of changing ID can be detected by our algorithms. Besides, the work [24] requires the assumption that each normal node knows the correct detection information of all normal nodes in the network after time kck_{c}. However, each normal node is supposed to detect only the malicious in- and out-neighbors in Proposition 1, which can save storage resources. We will show later that such detection can be realized through our fully distributed detection algorithm.

In the following sections, we present two detection algorithms for our RAC algorithm, which are redesigned based on the two-hop detection approach in our previous work [36]. In Section 4, we propose the sharing detection algorithm for undirected networks. In Section 5, we present the fully distributed detection algorithm for general directed networks. The latter algorithm is fully distributed and is efficient in storage usage compared to the related works [22, 24, 38], where each normal agent must obtain the correct detection of all malicious or normal agents in the network.

Input: Node ii knows xi[0]x_{i}[0], 𝒩i\mathcal{N}_{i}^{-}, 𝒩i+\mathcal{N}_{i}^{+} by Assumption 1.
1
2 Initialization:
3       Node ii initializes 𝒜i[0]=𝒜i[1]=\mathcal{A}_{i}[0]=\mathcal{A}_{i}[1]=\emptyset, and yi[0]=xi[0]y_{i}[0]=x_{i}[0], λi[0]=0\lambda_{i}[0]=0, δij[0]=0\delta_{ij}[0]=0, j𝒩i\forall j\in\mathcal{N}_{i}^{-},  zi[0]=1z_{i}[0]=1, γi[0]=0\gamma_{i}[0]=0, ωij[0]=0\omega_{ij}[0]=0, j𝒩i\forall j\in\mathcal{N}_{i}^{-}.
4      At k=1k=1, send λi[1]\lambda_{i}[1], γi[1]\gamma_{i}[1] using eq. (4) to q𝒩i+\forall q\in\mathcal{N}_{i}^{+} and receive δij[1]\delta_{ij}[1], ωij[1]\omega_{ij}[1] from j𝒩i\forall j\in\mathcal{N}_{i}^{-}.
5      Obtain yi[1]y_{i}[1], zi[1]z_{i}[1] using eq. (1).
6      Obtain λi[2]\lambda_{i}[2], γi[2]\gamma_{i}[2] using eq. (4).
7
8for  k2k\geq 2  do
9      
10      Transmit: Φi[k1]\Phi_{i}[k-1] to q𝒩i+\forall q\in\mathcal{N}_{i}^{+}.
11      Receive: Φj[k1]\Phi_{j}[k-1] from j𝒩i\forall j\in\mathcal{N}_{i}^{-}.
12      Detect: in-/out-neighbors according to the Detection Algorithm to obtain 𝒜i[k]\mathcal{A}_{i}[k].
13       Update using detection of in-neighbors:
14             Set i[k]=𝒩i𝒜i[k]\mathcal{M}_{i}^{-}[k]=\mathcal{N}_{i}^{-}\setminus\mathcal{A}_{i}[k].
15            In Case 1:
16            For each j𝒩i{i}j\in\mathcal{N}_{i}^{-}\cup\{i\}, set
17            δij[k]={λj[k],ji[k]{i},0,otherwise.\delta_{ij}[k]=\left\{\begin{aligned} &\lambda_{j}[k],&&\forall j\in\mathcal{M}_{i}^{-}[k]\cup\{i\},\\ &0,&&\textup{otherwise}.\end{aligned}\right.
18            ωij[k]={γj[k],ji[k]{i},0,otherwise.\omega_{ij}[k]=\left\{\begin{aligned} &\gamma_{j}[k],&&\forall j\in\mathcal{M}_{i}^{-}[k]\cup\{i\},\\ &0,&&\textup{otherwise}.\end{aligned}\right.
19      
20       Compute:
21            
22            yi[k]=j𝒩i{i}(δij[k]δij[k1])y_{i}[k]\medspace\medspace=\sum\limits_{j\in\mathcal{N}_{i}^{-}\cup\{i\}}{(\delta_{ij}[k]-\delta_{ij}[k-1])},
23            zi[k]=j𝒩i{i}(ωij[k]ωij[k1])z_{i}[k]\medspace\medspace=\sum\limits_{j\in\mathcal{N}_{i}^{-}\cup\{i\}}{(\omega_{ij}[k]-\omega_{ij}[k-1])}.
24      
25       Update using detection of out-neighbors:
26             Set i+[k]=𝒩i+𝒜i[k]\mathcal{M}_{i}^{+}[k]=\mathcal{N}_{i}^{+}\setminus\mathcal{A}_{i}[k].
27            Set di+[k]=|i+[k]|d_{i}^{+}[k]=|\mathcal{M}_{i}^{+}[k]|.
28            Set Δi+[k]=i+[k1]i+[k]\Delta\mathcal{M}_{i}^{+}[k]=\mathcal{M}_{i}^{+}[k-1]\setminus\mathcal{M}_{i}^{+}[k].
29            In Case 2:
30            yi[k]=yi[k]+|Δi+[k]|λi[k]y_{i}[k]\medspace\medspace=\medspace\medspace y_{i}[k]+|\Delta\mathcal{M}_{i}^{+}[k]|\lambda_{i}[k],
31            zi[k]=zi[k]+|Δi+[k]|γi[k]z_{i}[k]\medspace\medspace=\medspace\medspace z_{i}[k]+|\Delta\mathcal{M}_{i}^{+}[k]|\gamma_{i}[k].
32      
33       Compute:
34            
35            λi[k+1]=λi[k]+yi[k]/(1+di+[k])\lambda_{i}[k+1]\medspace\medspace=\medspace\medspace\lambda_{i}[k]+y_{i}[k]/(1+d_{i}^{+}[k]),
36            γi[k+1]=γi[k]+zi[k]/(1+di+[k])\gamma_{i}[k+1]\medspace\medspace=\medspace\medspace\gamma_{i}[k]+z_{i}[k]/(1+d_{i}^{+}[k]).
37      
      Output: ri[k]=yi[k]/zi[k]r_{i}[k]=y_{i}[k]/z_{i}[k]
38      
39
Algorithm 1 Resilient Average Consensus Algorithm

4 Sharing Detection in Undirected Networks

We introduce our first distributed detection algorithm to be presented as Algorithm 2, where the normal nodes are capable to detect malicious neighbors by using the two-hop information in undirected networks. It provides the basis for the two-hop detection in an adversarial environment, motivated by the works [37, 36].

4.1 Detection Algorithm Design

Input: Φj[k1],j𝒩i{i}\Phi_{j}[k-1],\forall j\in\mathcal{N}_{i}^{-}\cup\{i\}
1
2 Initialization:
3      
4      Node ii follows the initialization in Algorithm 1.
5      At k=1k=1, let 𝒜i[1]\mathcal{A}_{i}[1] include the IDs of in-neighbors not sending initial values to ii.
6      Let 𝒞i[1]={δij[1],ωij[1],j𝒩i{i}}\mathcal{C}_{i}[1]=\{\delta_{ij}[1],\omega_{ij}[1],\forall j\in\mathcal{N}_{i}^{-}\cup\{i\}\} be the initial check set.
7
8for  k2k\geq 2  do
9      
10      Let 𝒜i[k]=v𝒩𝒜v[k1]\mathcal{A}_{i}[k]=\bigcup_{v\in\mathcal{N}}\mathcal{A}_{v}[k-1] by Assumption 3.  
11      for  ji[k]j\in\mathcal{M}_{i}^{-}[k]  do
12            
13            (Step 1) if 𝒜j[k1]𝒜i[k]\mathcal{A}_{j}[k-1]\neq\mathcal{A}_{i}[k]  then
14                   let j𝒜i[k]j\in\mathcal{A}_{i}[k].
15            
16            (Step 2) if any ID in Φj[k1]\Phi_{j}[k-1] 𝒩j{j}\notin\mathcal{N}_{j}^{-}\cup\{j\}  then
17                   let j𝒜i[k]j\in\mathcal{A}_{i}[k].
18            
19            (Step 3) if any of δjh[k1|k1]\delta_{jh}[k-1|k-1] or ωjh[k1|k1]\omega_{jh}[k-1|k-1] in Φj[k1]\Phi_{j}[k-1] is not equal to the corresponding value in 𝒞i[k1]\mathcal{C}_{i}[k-1]  then
20                   let j𝒜i[k]j\in\mathcal{A}_{i}[k].
21            
22            (Step 4) if any of δjj[k|k1]\delta_{jj}[k|k-1] or ωjj[k|k1]\omega_{jj}[k|k-1] in Φj[k1]\Phi_{j}[k-1] is not equal to the reconstructed λj[k]\lambda_{j}^{\prime}[k] or γj[k]\gamma_{j}^{\prime}[k] by node ii  then
23                   let j𝒜i[k]j\in\mathcal{A}_{i}[k].
24            
25      
      Output: 𝒜i[k]\mathcal{A}_{i}[k]
26      
27      Store: δjj[k|k1]\delta_{jj}[k|k-1] and ωjj[k|k1]\omega_{jj}[k|k-1] from Φj[k1]\Phi_{j}[k-1], j𝒩i{i}j\in\mathcal{N}_{i}^{-}\cup\{i\}, into 𝒞i[k]\mathcal{C}_{i}[k].
28
Algorithm 2 Sharing Detection Algorithm

For this algorithm, the sharing detection function below is needed for the communication among the nodes when events of detecting adversaries occur.

Assumption 3

Once a malicious node is detected by any normal node at any time step, its ID will be securely notified to all nodes within the same time step.

This assumption also appeared in [37, 36] for resilient consensus. As we reported in [36], the sharing detection function can be realized by introducing fault-free mobile nodes which are appropriately distributed throughout the network and are capable to immediately verify if the detection reports from a node is true or false. Note that the deployment of such mobile agents is only for verification of detection reports instead of detecting malicious agents by themselves.

The sharing function is crucial for Algorithm 2 since it is necessary for detecting malicious nodes that are neighboring and cooperating with each other. We must emphasize that if we do not have this function for Algorithm 2, it can only handle the case where no neighboring malicious nodes exist (i.e., any malicious node is surrounded by normal nodes only). This is exactly the case studied in related works [23, 39, 38].

We now present Algorithm 2. To ensure that all nodes follow the correct averaging in Algorithm 1, the normal nodes check consistency among the neighbors’ information sets. In Algorithm 2, step 1 is to guarantee that each normal node should not use the values from the nodes detected as malicious already. Moreover, it ensures that a node does not falsely claim another node being malicious. Step 2 is to prevent the malicious nodes from faking any neighbors. Step 3 is to enforce the normal nodes not to modify the values received from their neighbors. Finally, step 4 is to guarantee that the neighbors follow the averaging in Algorithm 1.

4.2 Necessary Graph Structure for Algorithm 2

Refer to caption
Refer to caption
Figure 1: Illustration of the graph condition for Algorithm 2: (a) There is at least one common normal neighbor between any pair of neighboring malicious nodes. (b) There are at least f1f-1 common neighbors between any pair of neighbors under the ff-total model.

In this part, we provide the necessary graph condition for Algorithm 2. We can observe that a malicious node can be detected if there is at least one normal node among its neighbors that monitors its behavior. However, such detection may fail if neighboring malicious nodes cooperate with each other. Hence, it is critical that one or more normal nodes are present as their common neighbors. We illustrate this graph structure in Fig. 1(a). Here, nodes ii and jj are malicious. They can cooperate as follows: Node ii manipulates δij[k|k]\delta_{ij}[k|k] in its information set, and node jj manipulates δji[k|k]\delta_{ji}[k|k] in its information set. If there is no normal node having access to the information sets of both nodes ii and jj, such an attack will not be detected. In contrast, the detection works if there is a common neighbor hh of nodes ii and jj.

Next, we state the necessary and sufficient graph condition for Algorithm 2 to detect all the misbehaving agents.

Lemma 1

Consider the undirected graph 𝒢=(𝒱,)\mathcal{G}=(\mathcal{V},\mathcal{E}). Algorithm 2 detects every pair of neighboring misbehaving nodes if and only if they have at least one normal common neighbor.

Proof 4.1.

We can show similarly to Lemma 8 in [36] except that the update check in step 4 in Algorithm 2 is more complex than the general consensus protocol used there. Here, we provide a sketch of the proof. In the undirected network using Algorithm 2, each normal node ii can at least verify its own values δji[k|k]\delta_{ji}[k|k] and ωji[k|k]\omega_{ji}[k|k] in Φj[k]\Phi_{j}[k], j𝒩ij\in\mathcal{N}_{i}^{-}. If a malicious node is only surrounded by normal agents, then it cannot change any δji[k|k]\delta_{ji}[k|k] and ωji[k|k]\omega_{ji}[k|k] values from neighbors. Moreover, normal neighbor ii can reconstruct λj[k+1]\lambda_{j}[k+1] or γj[k+1]\gamma_{j}[k+1] through the averaging part in Algorithm 1 to check if node jj is following the averaging or not. Thus, misbehaving node jj will be detected by all normal neighbors. In the case of neighboring malicious nodes, they can modify the values from each other, but this is also detected by the normal common neighbor of them as discussed before.

Given that the malicious nodes are unknown and possibly cooperate with each other to launch attacks, we should impose a connectivity requirement so that the condition in Lemma 1 holds for any possible combination of pairs of neighboring malicious nodes in the network. The following proposition is the main result of this section.

Proposition 4.2.

Consider the undirected network 𝒢=(𝒱,)\mathcal{G}=(\mathcal{V},\mathcal{E}) under the ff-total malicious model. Suppose that Assumptions 1, 2, and 3 hold. Then, for Algorithm 1 with Detection Algorithm 2, the following hold:

(a) All malicious nodes that behave against the averaging in Algorithm 1 are detected if and only if for every pair of neighboring nodes, they have at least f1f-1 common neighbors.

(b) Under the condition of (a), normal nodes achieve resilient average consensus if 𝒢\mathcal{G} is (f+1f+1)-connected.

We proved in [36] that under the ff-total model, the graph condition in Lemma 1 is equivalent to condition (a) in Proposition 4.2. Moreover, condition (b) in Proposition 4.2 guarantees that the normal network is connected. Thus, normal nodes using Algorithm 1 with Detection Algorithm 2 can achieve resilient average consensus as we proved in Proposition 1.

As we will further explain in Section 6.2, the graph condition for Algorithm 2 does not require dense graph structures. However, this feature is achieved at the cost of additional authentication from the secure mobile agents.

5 Fully Distributed Detection in Directed Networks

In this section, we present our second distributed detection algorithm as Algorithm 3. It is fully distributed and operates without outside authentication for resilient average consensus in general directed networks. This important feature is realized through majority voting [41, 40] and requiring a denser graph structure. Moreover, we prove a necessary and sufficient graph condition for Algorithm 3 to properly function.

Refer to caption
Refer to caption
Figure 2: Illustration of the graph condition for node hh being detectable by node ii in Definition 5.3.

5.1 Detection Algorithm Design

In the last section, we have seen that node ii’s information set consists of two parts that need to be investigated by its out-neighbors: (i) the current value y¯i[k]\overline{y}_{i}[k] (z¯i[k]\overline{z}_{i}[k]) to check if it is updated according to Algorithm 1; (ii) the past values δij[k]\delta_{ij}[k] (ωij[k]\omega_{ij}[k]) used as inputs for updates to check if they are equal to the true values of the corresponding nodes. In Algorithm 2, normal node ii can check whether part of the past values are manipulated in the information sets of its neighbors. More specifically, node ii knows the true past values of its direct neighbors. Then, using the sharing detection function, node ii can report a malicious node (or a pair of neighboring malicious nodes) if any values known by itself are manipulated.

However, to achieve fully distributed detection without any outside authentication, node ii should be able to independently verify whether any entries of the past values are manipulated in the information sets of its in-neighbors. We have seen that node ii can directly obtain the original value and the detection information of an in-neighbor jj. For other values that node ii cannot directly obtain, we impose a certain graph structure so that it can access the original value and the detection information of a two-hop in-neighbor hh through majority voting. Specifically, if node ii receives mm values of node hh, among the mm values, if more than m/2m/2 values are the same, then node ii will take it as the true value of node hh. In computer science, such redundancy schemes are common strategies to enhance the security and reliability of systems [41, 40].

Next, we formally introduce the notion of detectable nodes to indicate the kind of nodes that can be detected by node ii using Algorithm 3.

Definition 5.3.

In the directed graph 𝒢=(𝒱,)\mathcal{G}=(\mathcal{V},\mathcal{E}) under the ff-local malicious model, node hh is said to be detectable by node ii if one of the following conditions holds:

  • h𝒩ih\in\mathcal{N}_{i}^{-};

  • there are at least 2f+12f+1 two-hop paths from hh to ii.

We illustrate the above graph condition in Fig. 2. Here, we also say that there is a detectable path from node hh to node ii if node hh is detectable by node ii. To achieve fully distributed detection, we need to impose a certain graph structure so that each node can have access to the necessary information used in its neighbors’ updates (see Fig. 3). We introduce the graph condition for Algorithm 3 as follows.

Assumption 4

A directed graph 𝒢=(𝒱,)\mathcal{G}=(\mathcal{V},\mathcal{E}) under the ff-local malicious model satisfies all the following conditions for i𝒱\forall i\in\mathcal{V}:

  1. 1.

    any two-hop in-neighbor hh is detectable by ii;

  2. 2.

    any out-neighbor qq is detectable by ii;

  3. 3.

    any out-neighbor ll of in-neighbor jj is detectable by ii.

We will refer to conditions 1)–3) together as the graph condition for Algorithm 3.

Refer to caption
Figure 3: Illustration of the graph condition for Algorithm 3.

Although the above graph condition may require a locally dense graph structure, such a graph does not necessarily have a small diameter. This means that the path length of the shortest path between any two nodes may not be small. In fact, we can construct graphs satisfying the conditions in large scales. See the examples in Section 5.4.

We now present Algorithm 3. Each normal node ii performs majority voting on two things: the nodes’ values and the detection information. Since we consider the ff-local model in this section, if node ii receives the same information from at least f+1f+1 distinct in-neighbors, it considers this information trustable. Additionally, node ii keeps a local set (only accessible to ii) for the detection of two-hop in-neighbors as 𝒜i2[k]\mathcal{A}_{i}^{2}[k] at time kk. After obtaining the true values of its one-hop in-neighbors j𝒩i\forall j\in\mathcal{N}_{i}^{-} and two-hop in-neighbors h𝒩i2\forall h\in\mathcal{N}_{i}^{2-}, it follows similar procedures as the ones in Algorithm 2.

Input: Φj[k1],j𝒩i{i}\Phi_{j}[k-1],\forall j\in\mathcal{N}_{i}^{-}\cup\{i\}
1
2 Initialization:
3      
4      Node ii follows the initialization in Algorithm 2.
5      Let 𝒜i2[1]=\mathcal{A}_{i}^{2}[1]=\emptyset.
6
7for  k2k\geq 2  do
8      
9       Majority voting:
10            
11            Let 𝒞i[k1]=𝒞i[k1]{δjh[k1|k1]\mathcal{C}_{i}[k-1]=\mathcal{C}_{i}[k-1]\cup\{\delta_{jh}[k-1|k-1], ωjh[k1|k1],h𝒩i2}\omega_{jh}[k-1|k-1],\forall h\in\mathcal{N}_{i}^{2-}\}, i.e., the majority values of hh from Φj[k1]\Phi_{j}[k-1], j𝒩ij\in\mathcal{N}_{i}^{-}.
12            Let 𝒜i[k]=𝒜i[k1]\mathcal{A}_{i}[k]=\mathcal{A}_{i}[k-1], 𝒜i2[k]=𝒜i2[k1]\mathcal{A}_{i}^{2}[k]=\mathcal{A}_{i}^{2}[k-1].
13            for  ID mm\in majority of j𝒩i𝒜j[k1],\bigcup_{j\in\mathcal{N}_{i}^{-}}\mathcal{A}_{j}[k-1], do
14                   if  m𝒩i𝒩i+m\in\mathcal{N}_{i}^{-}\cup\mathcal{N}_{i}^{+}  then
15                         let m𝒜i[k]m\in\mathcal{A}_{i}[k], the faulty reporter j𝒜i[k]j^{\prime}\in\mathcal{A}_{i}[k];
16                  else
17                        let m𝒜i2[k]m\in\mathcal{A}_{i}^{2}[k], j𝒜i[k]j^{\prime}\in\mathcal{A}_{i}[k].
18                  
19            
20      
21      for  ji[k]j\in\mathcal{M}_{i}^{-}[k]  do
22            
23            (Step 1a) if 𝒩j𝒜j[k1]\mathcal{N}_{j}^{-}\cap\mathcal{A}_{j}[k-1] contains any different detection of the nodes in 𝒜i[k]𝒜i2[k]\mathcal{A}_{i}[k]\cup\mathcal{A}_{i}^{2}[k]  then
24                   let j𝒜i[k]j\in\mathcal{A}_{i}[k].
25            
26            (Step 1b) if 𝒩j2𝒜j[k1]\mathcal{N}_{j}^{2-}\cap\mathcal{A}_{j}[k-1] contains any ID not in 𝒜i[k]𝒜i2[k]\mathcal{A}_{i}[k]\cup\mathcal{A}_{i}^{2}[k] or it does not contain a same ID in 𝒜i[k]𝒜i2[k]\mathcal{A}_{i}[k]\cup\mathcal{A}_{i}^{2}[k] for the second time  then
27                   let j𝒜i[k]j\in\mathcal{A}_{i}[k].
28            
29            Steps 2-4 in Algorithm 2
      Output: 𝒜i[k]\mathcal{A}_{i}[k]
30       Store: δjj[k|k1]\delta_{jj}[k|k-1] and ωjj[k|k1]\omega_{jj}[k|k-1] from Φj[k1]\Phi_{j}[k-1], j𝒩i{i}j\in\mathcal{N}_{i}^{-}\cup\{i\}, into 𝒞i[k]\mathcal{C}_{i}[k].
Algorithm 3 Fully Distributed Detection Algorithm

5.2 Necessary Graph Structure for Algorithm 3

In Algorithm 3, we must impose the connectivity requirement in Assumption 4 on every node and its in-/out-neighbors. This enables the detection to be guaranteed for any combination of nodes being malicious in the network. The following theorem is the main result of this section.

Theorem 5.4.

Consider the directed network 𝒢=(𝒱,)\mathcal{G}=(\mathcal{V},\mathcal{E}) under the ff-local malicious model. Suppose that Assumptions 1 and 2 hold. Then, for Algorithm 1 with Detection Algorithm 3, the following hold:

(a) Each normal node detects all malicious nodes in its out-neighbors and in-neighbors within two hops that behave against the averaging in Algorithm 1 if and only if 𝒢\mathcal{G} satisfies the condition for Algorithm 3 (in Assumption 4).

(b) Under the condition of (a), normal nodes achieve resilient average consensus if 𝒢\mathcal{G} is (f+1)(f+1)-strongly connected.

Proof 5.5.

(a) Necessity: We prove condition 1) by contradiction; conditions 2) and 3) follow by a similar proof. Suppose that there is a node h𝒩jh\in\mathcal{N}_{j}^{-} with h𝒩ih\notin\mathcal{N}_{i}^{-}, and that there are at most 2f2f two-hop paths from node hh to node ii including the path containing node jj. Suppose that node jj is malicious and there are f1f-1 malicious middle nodes in the paths from hh to ii (by the assumption of the ff-local model). Then, in the worst case, node ii could get ff copies of the true value of δhh[k]\delta_{hh}[k] from the ff normal middle nodes. In the mean time, node ii also gets ff copies of an identical false value of δhh[k]\delta_{hh}[k] from the ff malicious middle nodes (including jj). Thus, node ii cannot get majority regarding the true value of δhh[k]\delta_{hh}[k]. Thus, it cannot detect node jj’s manipulation on δjh[k|k]\delta_{jh}[k|k] in Φj[k]\Phi_{j}[k].

Sufficiency: By Definition 5.3, if an out-neighbor qq of node ii is detectable by node ii, then node qq becomes a direct in-neighbor or a two-hop in-neighbor of ii. Therefore, the detection of node qq is the same as the detection of node jj or hh below.

We must show that node ii can confirm the true value of every entry of the information set Φj[k]\Phi_{j}[k] of in-neighbor jj by three major steps at time k+1k+1. See the illustration in Fig. 3. First, it can obtain the true values δhh[k]\delta_{hh}[k], ωhh[k]\omega_{hh}[k] of every neighbor h𝒩jh\in\mathcal{N}_{j}^{-} (i.e., ii’s two-hop in-neighbor hh) from the previous time kk. Second, node ii can obtain the correct detection of h𝒩jh\in\mathcal{N}_{j}^{-} before the detection loop at time kk. Moreover, node ii can obtain the correct detection of jj’s out-neighbor ll depending on the corresponding case (ll can be a two-hop in-neighbor or a direct in-neighbor of ii by condition 3) in Assumption 4). Then we can prove that node ii will detect node jj at time k+1k+1 if node jj sends out faulty Φj[k]\Phi_{j}[k].

Depending on how the detectable path is formed, consider two cases for node ii: (i) nodes hh and ll are direct in-neighbors of ii; (ii) nodes hh and ll are two-hop in-neighbors of ii.

(i) In the case where h,l𝒩ih,l\in\mathcal{N}_{i}^{-}, it is clear that node ii can receive the true values δhh[k]\delta_{hh}[k] and ωhh[k]\omega_{hh}[k] from Φh[k1]\Phi_{h}[k-1]. Moreover, it can have the correct detection of its direct in-neighbors hh and ll before time kk.

(ii) Suppose that h𝒩ih\notin\mathcal{N}_{i}^{-}, and there are at least 2f+12f+1 two-hop paths from hh to ii. In this case, there is some normal node p𝒩h+𝒩ip\in\mathcal{N}_{h}^{+}\cap\mathcal{N}_{i}^{-} which carries the true values δhh[k]\delta_{hh}[k] and ωhh[k]\omega_{hh}[k] in its information set Φp[k]\Phi_{p}[k]. Then, node ii can get the true values δhh[k]\delta_{hh}[k] and ωhh[k]\omega_{hh}[k] since the majority of the 2f+12f+1 paths from hh to ii contain nodes as pp by the ff-local model.

For node ii to obtain the correct detection of two-hop in-neighbors hh and ll, it follows a similar analysis. We look at the case for hh. If hh transmits faulty Φh[k1]\Phi_{h}[k-1], then it is detected by its one-hop neighbors at time kk. Recall that there are at least 2f+12f+1 directed two-hop paths from hh to ii. Thus, under the ff-local model, node ii can obtain the correct detection of hh by majority voting before the detection loop of time k+1k+1.

Therefore, node ii knows the true values δhh[k]\delta_{hh}[k], ωhh[k]\omega_{hh}[k] and obtains the correct detection of its two-hop in-neighbors hh, ll before running the detection loop at time k+1k+1. Thus if node j𝒩ij\in\mathcal{N}_{i}^{-} sends out faulty Φj[k]\Phi_{j}[k] by possible manipulation including modifying δjh[k]\delta_{jh}[k] and/or ωjh[k]\omega_{jh}[k] in Φj[k]\Phi_{j}[k], or by sending false detection information of nodes hh and ll, then node ii will detect. Note that when out-neighbor ll is a two-hop in-neighbor of jj, the detection of ll is included in 𝒜j[k+1]\mathcal{A}_{j}[k+1] and is sent to node ii in Φj[k+1]\Phi_{j}[k+1]. Therefore, step 1b in Algorithm 3 is designed to handle this case. This procedure will not cause problems since the removal of malicious neighbors can be asynchronous at each normal agent.

Next, we show that node ii can verify if node jj updates δjj[k+1]\delta_{jj}[k+1] and ωjj[k+1]\omega_{jj}[k+1] in Φj[k]\Phi_{j}[k] by the averaging in Algorithm 1 or not. This is done by reconstructing λj[k+1]\lambda_{j}^{\prime}[k+1] and γj[k+1]\gamma_{j}^{\prime}[k+1] and checking whether ϵλ=ϵγ=0\epsilon_{\lambda}=\epsilon_{\gamma}=0, where

δjj[k+1]\displaystyle\delta_{jj}[k+1] =λj[k+1]+ϵλ,\displaystyle=\lambda_{j}^{\prime}[k+1]+\epsilon_{\lambda},
ωjj[k+1]\displaystyle\omega_{jj}[k+1] =γj[k+1]+ϵγ.\displaystyle=\gamma_{j}^{\prime}[k+1]+\epsilon_{\gamma}.

As shown above, node ii can verify 𝒜j[k]\mathcal{A}_{j}[k] in Φj[k]\Phi_{j}[k]. Thus,

λj[k]\displaystyle\lambda_{j}^{\prime}[k] =δjj[k1|k1]\displaystyle=\medspace\medspace\delta_{jj}[k-1|k-1]
+yj[k1]+|Δj+[k1]|δjj[k1|k1]1+dj+[k1],\displaystyle\medspace\medspace\medspace+\frac{y_{j}^{\prime}[k-1]+|\Delta\mathcal{M}_{j}^{+}[k-1]|\delta_{jj}[k-1|k-1]}{1+d_{j}^{+}[k-1]},
γj[k]\displaystyle\gamma_{j}^{\prime}[k] =ωjj[k1|k1]\displaystyle=\medspace\medspace\omega_{jj}[k-1|k-1]
+zj[k1]+|Δj+[k1]|ωjj[k1|k1]1+dj+[k1],\displaystyle\medspace\medspace\medspace+\frac{z_{j}^{\prime}[k-1]+|\Delta\mathcal{M}_{j}^{+}[k-1]|\omega_{jj}[k-1|k-1]}{1+d_{j}^{+}[k-1]},

where δjj[k1|k1]\delta_{jj}[k-1|k-1], ωjj[k1|k1]\omega_{jj}[k-1|k-1], Δj+[k1]\Delta\mathcal{M}_{j}^{+}[k-1], dj+[k1]d_{j}^{+}[k-1] are from Φj[k1]\Phi_{j}[k-1] with 𝒜j[k1]\mathcal{A}_{j}[k-1]. Moreover, yj[k1]y_{j}^{\prime}[k-1] and zj[k1]z_{j}^{\prime}[k-1] are obtained by node ii through

yj[k1]\displaystyle y_{j}^{\prime}[k-1] =(δjj[k|k1]δjj[k1|k1])\displaystyle=\medspace\medspace(\delta_{jj}[k|k-1]-\delta_{jj}[k-1|k-1])
×(1+dj+[k1]),\displaystyle\medspace\medspace\medspace\times(1+d_{j}^{+}[k-1]),
zj[k1]\displaystyle z_{j}^{\prime}[k-1] =(ωjj[k|k1]ωjj[k1|k1])\displaystyle=\medspace\medspace(\omega_{jj}[k|k-1]-\omega_{jj}[k-1|k-1])
×(1+dj+[k1]).\displaystyle\medspace\medspace\medspace\times(1+d_{j}^{+}[k-1]).

We also note that node ii has access to the true values δhh[k]\delta_{hh}[k] and ωhh[k]\omega_{hh}[k]. Besides, node ii knows δjh[k1]=δhh[k1]\delta_{jh}[k-1]=\delta_{hh}[k-1] and ωjh[k1]=ωhh[k1]\omega_{jh}[k-1]=\omega_{hh}[k-1] from Φj[k1]\Phi_{j}[k-1]. Otherwise, node jj would have been detected at time kk. Thus, we have

λj[k+1]\displaystyle\lambda_{j}^{\prime}[k+1] =λj[k]+h𝒩j{j}(δjh[k]δjh[k1]),\displaystyle=\medspace\medspace\lambda_{j}^{\prime}[k]+\sum\limits_{h\in\mathcal{N}_{j}^{-}\cup\{j\}}{(\delta_{jh}[k]-\delta_{jh}[k-1])},
γj[k+1]\displaystyle\gamma_{j}^{\prime}[k+1] =γj[k]+h𝒩j{j}(ωjh[k]ωjh[k1]),\displaystyle=\medspace\medspace\gamma_{j}^{\prime}[k]+\sum\limits_{h\in\mathcal{N}_{j}^{-}\cup\{j\}}{(\omega_{jh}[k]-\omega_{jh}[k-1])},

where δjh[k]=0\delta_{jh}[k]=0 and ωjh[k]=0\omega_{jh}[k]=0 for h𝒜j[k]h\in\mathcal{A}_{j}[k]. Then node ii can compare δjj[k+1]\delta_{jj}[k+1] (and ωjj[k+1]\omega_{jj}[k+1]) in Φj[k]\Phi_{j}[k] with λj[k+1]\lambda_{j}^{\prime}[k+1] (and γj[k+1]\gamma_{j}^{\prime}[k+1]) and checks if node jj follows the averaging in Algorithm 1 or not.

(b) A malicious node will be detected immediately once it manipulates its information set. Thus, misbehavior of any malicious node cannot affect normal nodes since normal nodes exclude values from detected malicious nodes in Algorithm 1. Moreover, since 𝒢\mathcal{G} is (f+1)(f+1)-strongly connected, the normal network is strongly connected after removing the ff-local adversary node set. Therefore, resilient average consensus is achieved as shown in Proposition 1.

Here, we show that the graph satisfying the condition for Algorithm 3 has the minimum in-degree as 2f+12f+1. It indicates that we need to make the minimum in-degree no less than 2f+12f+1 when we design a desirable network topology. Conversely, it is straightforward that a graph with minimum in-degree less than 2f+12f+1 does not meet our condition. We formally state the property in the next lemma for general strongly connected digraphs since strong connectivity is necessary for achieving average consensus in directed graphs [11]. Moreover, we can confirm that a complete graph 𝒦n\mathcal{K}_{n} satisfies the condition for Algorithm 3. To avoid trivial cases, we consider n>3n>3 in the following discussions.

Lemma 5.6.

If a strongly connected and incomplete digraph 𝒢=(𝒱,)\mathcal{G}=(\mathcal{V},\mathcal{E}) (under the ff-local model) satisfies the condition for Algorithm 3 in Assumption 4, then 𝒢\mathcal{G} has the minimum in-degree no less than 2f+12f+1.

The proof of Lemma 5.6 can be found in the Appendix.

We note that the detection condition (a) for Algorithm 3 is not sufficient to guarantee strong connectivity of the graph. A simple counter example is a graph with two disconnected complete subgraphs. Clearly, the whole graph is not connected while it meets the detection condition for Algorithm 3. This observation reveals that the consensus condition (b) guaranteeing the normal network to be strongly connected is also critical for our RAC algorithm.

5.3 Graph Condition in Undirected Networks

For the special case of undirected networks, the condition for Algorithm 3 can be simplified as follows.

Lemma 5.7.

An undirected graph 𝒢=(𝒱,)\mathcal{G}=(\mathcal{V},\mathcal{E}) satisfies the condition for Algorithm 3 in Assumption 4 if for each node i𝒱i\in\mathcal{V}, it holds that any two-hop in-neighbor hh of node ii is detectable by node ii.

Proof 5.8.

We can easily observe that condition 2) in Assumption 4 is satisfied automatically in undirected networks. Moreover, it holds that an out-neighbor ll of node ii’s in-neighbor jj is a two-hop in-neighbor of node ii in undirected networks. Therefore, condition 3) can be derived if condition 1) holds in an undirected network.

Next, we show that for undirected networks, the condition for Algorithm 3 and connectivity of the graph together guarantee that the normal network is connected after removing the ff-local malicious node set.

Proposition 5.9.

Consider the undirected graph 𝒢=(𝒱,)\mathcal{G}=(\mathcal{V},\mathcal{E}) under the ff-local model. If (i) 𝒢\mathcal{G} is connected and (ii) for any node i𝒱i\in\mathcal{V}, it holds that any two-hop in-neighbor hh of node ii is detectable by node ii, then the normal network induced by the normal agents in 𝒩𝒱\mathcal{N}\subseteq\mathcal{V} is connected.

The proof of Proposition 5.9 can be found in the Appendix.

We can see from the above two results that both the detection condition (a) and consensus condition (b) for undirected networks are simplified compared to the conditions in Theorem 5.4 for directed networks. We formally state the conditions for undirected networks as follows, which can be proved by Lemma 5.7 and Proposition 5.9.

Theorem 5.10.

Consider the undirected network 𝒢=(𝒱,)\mathcal{G}=(\mathcal{V},\mathcal{E}) under the ff-local malicious model. Suppose that Assumptions 1 and 2 hold. Then, for Algorithm 1 with Detection Algorithm 3, the following hold:

(a) Each normal node detects all malicious neighbors within two hops that behave against the averaging in Algorithm 1 if and only if for any node i𝒱i\in\mathcal{V}, it holds that any two-hop neighbor hh of node ii is detectable by node ii.

(b) Under the condition of (a), normal nodes achieve resilient average consensus if 𝒢\mathcal{G} is connected.

5.4 Construction of Graphs Satisfying the Condition

Refer to caption
(a) The four-layer directed graph which contains an undirected graph after removing the red directed edges.
Refer to caption
(b) The five-layer directed graph.
Figure 4: Large-scale graphs satisfying the condition for Algorithm 3 under the ff-local model. Here, we set f=1f=1 for illustration.

In this part, we present some example graphs satisfying the conditions for our algorithms. Furthermore, we present a systematic way to construct large-scale graphs that meet the condition for Algorithm 3.

We first give example graphs satisfying the conditions in Theorem 5.4. The network in Fig. 4(a) satisfies the conditions for Algorithm 3 under the 11-local model. Moreover, there is a characteristic four-layer structure. We can extend this idea to the cases with any ff. Each layer should have 2f+12f+1 nodes for the ff-local model. Each node in one layer should be connected with every node in the neighbor layers and have no connection with the nodes in its own layer. This structure can also have many layers as long as the ff-local set is met for i𝒩\forall i\in\mathcal{N}. A similar way for constructing large-scale directed networks for Algorithm 3 is presented in Fig. 4(b). From these examples, we can conclude that the unbalanced directed graphs can also meet the condition for Algorithm 3. Simulations of Algorithm 3 in these networks will be given in Section 6.

Node ii is said to be a full access node if it is an out-neighbor of all other nodes in the network [36]. Notice that such a node can detect any malicious node in the network. We can enhance the performance of Algorithm 3 (and Algorithm 2) by incorporating nodes with such characteristics. However, we emphasize that we do not assume such full access nodes to be normal. As long as the conditions for Algorithm 3 (or Algorithm 2) are met, a full access node can also be detected by its normal neighbors when it misbehaves. This result can be easily proved by Theorem 5.4 and is given as follows.

Corollary 5.11.

A normal full access node using Algorithm 3 (or Algorithm 2) detects any node behaving against the averaging in Algorithm 1 in the network.

As an example, the 5-node undirected network in Fig. 5(a) could tolerate two malicious nodes when the conditions for 11-local are met except for the full access node 1. In the same graph, if only node 1 becomes malicious and the conditions for 11-local are also met for other nodes, then RAC is still guaranteed. As another example, in the 8-node directed network in Fig. 8, normal nodes using Algorithm 1 with Detection Algorithm 3 can achieve resilient average consensus even in the presence of 5 malicious nodes. More details are discussed in the numerical examples.

The following corollary states the maximum tolerable number of malicious nodes in a network applying our algorithms. It can be directly proved from Theorem 5.4.

Corollary 5.12.

In the complete graph 𝒦n\mathcal{K}_{n}, normal nodes using Algorithm 1 with Algorithm 3 (or Algorithm 2) achieve resilient average consensus in the presence of fn2f\leq n-2 malicious nodes in the network.

Refer to caption
Refer to caption
Figure 5: Small-scale graphs satisfying the condition for Algorithm 3 under the 11-local model (except for the full access node 1 in (a)).
Remark 5.13.

From above, we see that it is relatively easy to check whether a graph meets the condition in Assumption 4 since there is no combinatorial process in the checking. In particular, the verification of the detectable condition in Definition 5.3 is very simple. Further, the checking on the condition in Assumption 4 requires less than n(n1)n(n-1) times of such verification (usually much less for sparse graphs). In contrast, the verification of robustness of a graph in resilient consensus works (e.g., [19, 32, 35]) involves combinatorial processes and is computationally NP-hard. Moreover, we have proved a much simpler condition for undirected networks using our algorithm, which is suitable for the deployment in large-scale networks. Besides, the proposed systematic way for constructing the desirable large-scale graphs can also facilitate the deployment of our algorithm in various applications.

5.5 Discussions and Comparisons with Related Works

We discuss the differences between the resilient consensus (RC) algorithm from [36] and RAC Algorithm 3 for directed networks. In [36], normal nodes achieve resilient consensus by monitoring their in-neighbors. However, the situation becomes much more complex for Algorithm 3. Aside from the detection of in-neighbors, node ii should also be able to detect each out-neighbor. This is because for solving the RAC problem, node ii should not send its yy and zz values to the malicious out-neighbor(s) so that the normal network can accurately preserve the “mass” of normal nodes only and achieve averaging as shown in Theorem 5.4. The detection condition for Algorithm 3 hence requires denser graphs than the one for the RC algorithm [36]. Moreover, notice that the necessary condition for resilient average consensus is the strong connectivity of the graph. It requires each node to have at least one out-going edge. In contrast, the necessary condition for resilient consensus is that there is at least one rooted spanning tree in the graph. Therefore, the necessary condition for Algorithm 3 is stricter than the one for the RC algorithm [36].

The recent work [22] proposed a certified propagation algorithm (CPA)-type broadcast and retrieval approach for the RAC problem. There, each normal node broadcasts its initial value to all the nodes in the network through relaying by neighbors (i.e., the flooding technique). Then, the normal node confirms another node’s initial value if it receives more than ff copies of the value of the same node, which is similar to the CPA approach [44]. Lastly, the normal node converges to the average of values from the nodes which it has confirmed.

We must note that this kind of approaches for each node to verify and store the initial values of all the normal agents become infeasible in large-scale networks, as it consumes intensive storage and computation for each single node to monitor the whole network. Compared to these approaches, our iterative detection algorithm is more efficient, especially in large-scale networks. Specifically, the storage needed on each node for our algorithm is modest as each node stores only local information of its in-neighbors within two hops.

It is challenging for our algorithms, as well as any other algorithms [21, 20, 37, 39] to identify adversarial nodes that adopt extreme initial values but behave according to the proposed algorithms as if they were normal nodes. Clearly, such nodes are indistinguishable from normal nodes with extreme initial values. To mitigate the impact of such adversary nodes, we can set a safety interval [ymin,ymax][y_{\min},y_{\max}] (recall that yi[0]=xi[0]y_{i}[0]=x_{i}[0]) for normal nodes so that neighbors taking initial values outside this interval are considered malicious [36].

We conduct some comparisons between Algorithms 2 and 3 for the case of undirected networks. Recall that Algorithm 2 is for the ff-total model while Algorithm 3 is for the ff-local model. We first note that the ff-local model contains the ff-total model and is more adversarial in the sense that more than ff malicious agents in total may be in the entire network under the ff-local model. The reason is that if the graph condition for Algorithm 2 is satisfied under the ff-local model, then it cannot guarantee that there is a normal neighbor of any pair of adjacent malicious nodes (Lemma 1). Here is a simple counter example. Consider the 5-node network in Fig. 5(a) with malicious node set 𝒜={1,4,5}\mathcal{A}=\{1,4,5\}. It satisfies the f1f-1 common neighbors condition for Algorithm 2 under the 22-local model. Yet, it does not meet the condition in Lemma 1. However, we must note that this phenomenon is not in presence for Algorithm 3 since the condition in Theorem 5.10 has required the necessary condition for each node to independently detect the malicious neighbors.

6 Numerical Examples

We present numerical examples to verify the efficacy of RAC Algorithm 1 with Detection Algorithms 2 and 3.

6.1 Simulations with Directed Networks

In this part, we provide the simulations of Algorithm 3 in three directed networks of different scales.

1) Small Directed Network: Consider the 6-node network in Fig. 5(b). It meets the graph condition (Assumption 4) for Algorithm 3 under the 1-local model. Moreover, it is 2-strongly connected (i.e., the remaining graph is strongly connected after the removal of the 1-local malicious node set). Hence, it meets the requirements in Theorem 5.4.

Refer to caption
(a) Under attacks with less edges.
Refer to caption
(b) Under attacks.
Figure 6: Time responses of Algorithm 3 in the 6-node network in Fig. 5(b).

Set the initial values x[0]=[9 7 1 3 4 6]Tx[0]=[9\ 7\ 1\ 3\ 4\ 6]^{T} and the adversary node set 𝒜={6}\mathcal{A}=\{6\}. Malicious node 6 is indicated in red in Fig. 5(b). First, we show that the detection condition for Algorithm 3 is critical for the success of our RAC algorithm. Suppose that three undirected edges (1,4)(1,4), (2,5)(2,5), and (3,6)(3,6) are removed from the network, i.e., the condition for Algorithm 3 is not satisfied. The simulation under attacks for the above case with less edges is displayed in Fig. 6(a). Until time k=2k=2, malicious node 6 follows the averaging in Algorithm 1 to avoid being detected. Then it manipulates its yy values through changing the past values of node 2 while not manipulating other entries of its information set. We can see in Fig. 6(a) that resilient average consensus is not achieved by Algorithm 3 with less edges. This is because only nodes 2 and 4 can detect the above misbehavior of node 6. Normal nodes 1, 3 and 5 are misled by node 6 due to the lack of necessary graph structure to obtain the correct value of node 2.

Next, we apply Algorithm 3 in the 6-node network as presented in Fig. 5(b), where the condition for Algorithm 3 is met. Consider the same initial states and the same attacks for the network. The simulation result is presented in Fig. 6(b). Malicious node 6 launches attacks as before, however, this misbehavior is soon detected by its normal out-neighbors. The normal nodes then compensate the erroneous effects received from node 6 and start to form consensus among normal neighbors only. Lastly, the normal nodes are able to reach the average of their initial values X¯𝒩=i𝒩xi[0]|𝒩|=4.8\overline{X}_{\mathcal{N}}=\frac{\sum_{i\in\mathcal{N}}x_{i}[0]}{|\mathcal{N}|}=4.8, and resilient average consensus is reached using Algorithm 3.

Refer to caption
(a) No attack.
Refer to caption
(b) Under attacks.
Figure 7: Time responses of Algorithm 3 in the 14-node network in Fig. 4(b).

2) Medium-scale Directed Network: Next, consider the 14-node network in Fig. 4(b) constructed using the method in Section 5.4. It satisfies the condition for Algorithm 3 under the 1-local model and is 2-strongly connected.

Let the initial values x[0]=[11 2 9 3 2 10 1 4 6 9 7 5 14 8]Tx[0]=[11\ 2\ 9\ 3\ 2\ 10\ 1\ 4\ 6\ 9\ 7\ 5\ 14\ 8]^{T} and 𝒜={2,14}\mathcal{A}=\{2,14\} in Fig. 4(b). The time responses of nodes under no attack are shown in Fig. 7(a), where all nodes using Algorithm 3 reach the average of their initial values X¯=x[0]n=6.5\overline{X}=\frac{\sum x[0]}{n}=6.5. Here, the lines not in red represent the values of normal nodes. Then, the time responses of nodes under attacks are displayed in Fig. 7(b). There, until time k=3k=3, malicious nodes 2 and 14 pretend to be normal by following the averaging. Then node 14 changes its own value to a fixed value and is detected by its normal out-neighbors at the next time step. In the meantime, node 2 keeps concealing itself. At time k=11k=11, node 2 and normal nodes almost reach the average of their initial values (around 6.385). However, node 2 starts to manipulate its yy value through changing the past values of its in-neighbors in its information set. Such an attack is also quickly detected and normal nodes remove the effects received from node 2 until then. Finally, the normal nodes reach the average of their initial values X¯𝒩=6.75\overline{X}_{\mathcal{N}}=6.75, and resilient average consensus is attained. Moreover, the convergence of Algorithm 3 is quick even in the presence of malicious attacks.

Refer to caption
Figure 8: The 8-node network satisfying the condition for Algorithm 3 under the 11-local model.
Refer to caption
Figure 9: Time responses of Algorithm 3 in the 8-node network in Fig. 8.

3) Over Half of the Nodes Turn Malicious: We conduct the simulation of Algorithm 3 under an extremely adversarial case, where over half of the nodes in the network turn malicious. Consider the 8-node network in Fig. 8 with 𝒜={3,4,5,6,7}\mathcal{A}=\{3,4,5,6,7\}. It is almost complete except that there are 4 directed edges from node 2. Moreover, it satisfies the condition for Algorithm 3 under the 1-local model for non-full access node 2.

Set the initial values x[0]=[3 15 9 8 4 7 1 12]Tx[0]=[3\ 15\ 9\ 8\ 4\ 7\ 1\ 12]^{T}. The simulation result is presented in Fig. 9. All malicious nodes simultaneously launch attacks at time k=3k=3 by manipulating their values arbitrarily. However, these attacks are soon detected. In Fig. 9, the normal nodes eventually reach the average of their initial values X¯𝒩=10\overline{X}_{\mathcal{N}}=10. Therefore, we can conclude that resilient average consensus is still guaranteed using Algorithm 3 despite the erroneous effects from 5 malicious nodes.

6.2 Simulations with Large-scale Undirected Networks

1) Simulation of Algorithm 2: Here, we show the effectiveness of Algorithm 2 by conducting a simulation in the 5-node undirected network in Fig. 5(a) with initial states x[0]=[8 6 1 3 9]Tx[0]=[8\ 6\ 1\ 3\ 9]^{T}. It is 3-connected and with at least one common neighbor for every pair of neighbors. Given these properties, Proposition 4.2 indicates that resilient average consensus can be achieved using Algorithm 2 under the 2-total malicious model. Let the adversary set 𝒜={4,5}\mathcal{A}=\{4,5\}. The simulation result under attacks is shown in Fig. 10. Malicious node 5 first attacks other agents by transmitting arbitrary values at time k=4k=4 and it is soon detected by its normal neighbors. At time k=13k=13, nodes 1, 2, 3 and 4 almost reach the average of their initial values (i.e., 4.5). However, node 4 starts to manipulate its own value. Such an attack is also quickly detected. Then the normal nodes reach the average of their initial values X¯𝒩=5\overline{X}_{\mathcal{N}}=5.

2) Simulation of Algorithm 3: In this part, we carry out the simulation of Algorithm 3 in a large-scale network, which is constructed by the method proposed in Section 5.4. Specifically, we consider the 30-node network in Fig. 11. It has a 10-layer structure satisfying the condition for Algorithm 3 under the 1-local model. The malicious nodes are indicated in red with 𝒜={3,6,15,18,27,30}\mathcal{A}=\{3,6,15,18,27,30\}.

Refer to caption
Figure 10: Time responses of Algorithm 2 in the 5-node network in Fig. 5(a).
Refer to caption
Figure 11: A large-scale network satisfying the condition for Algorithm 3 under the 1-local model.

Let the initial values x[0]=[8 7 5 3 2 11 1 4 6 9x[0]=[8\ 7\ 5\ 3\ 2\ 11\ 1\ 4\ 6\ 9\ 10 12 11 13 14 3 5 2 8 7 5 3 2 11 1 4 6 9 10 12]T10\ 12\ 11\ 13\ 14\ 3\ 5\ 2\ 8\ 7\ 5\ 3\ 2\ 11\ 1\ 4\ 6\ 9\ 10\ 12]^{T}. The simulation results of Algorithm 3 without and with attacks are shown in Figs. 12(a) and (b), respectively. One can see in Fig. 12(a) that all nodes achieve average consensus X¯=6.8\overline{X}=6.8 using Algorithm 3 although the convergence is slow due to the large network size. As for the results of nodes under attacks, it shows in Fig. 12(b) that at time k=9k=9, all the 6 malicious nodes start to manipulate their values through cooperating with their malicious neighbors and changing the past values of each other in their information sets. Such attacks are soon detected by their normal neighbors. Thereafter, the normal nodes reach the average of their initial values X¯𝒩=6.4166\overline{X}_{\mathcal{N}}=6.4166. The RAC problem is solved by Algorithm 3 in the presence of 6 malicious nodes. Note that Algorithm 3 can still guarantee resilient average consensus if any one of the nodes become malicious in each one of the 6 layers containing malicious nodes currently. This is because the malicious nodes also satisfy the 1-local model in this case. We finally emphasize that none of the methods in [23, 38, 39] can handle the above case of neighboring malicious nodes.

Refer to caption
(a) No attack.
Refer to caption
(b) Under attacks.
Figure 12: Time responses of Algorithm 3 in the 10-layer network in Fig. 11.

7 Conclusion

In this paper, we have investigated the problem of resilient average consensus in the presence of adversaries. We have proposed a distributed iterative detection and averaging algorithm for normal agents to achieve resilient average consensus in general directed topologies. For the detection part, we have proposed two distributed algorithms and the second one can achieve fully distributed detection of malicious agents. For the averaging part, it can precisely preserve the sum of the initial values of normal agents. Moreover, we have fully characterized the network requirement for the algorithms to successfully achieve resilient average consensus. Compared to the existing detection approaches, our method is the only one that can handle the case of neighboring malicious agents. Besides, we have solved the resilient average consensus in directed networks, whereas the existing detection approaches studied undirected networks only. Moreover, in comparison with the existing secure broadcast and retrieval approach [22], our algorithm can save storage as each agent keeps only the values of two-hop neighbors. In the end, we have provided extensive numerical examples to show the effectiveness of the proposed algorithms.

In future works, we are interested in applying our algorithms to various applications of average consensus where security needs to be enhanced, e.g., the economic dispatch problem and the PageRank problem.

\appendices

Appendix

Proof of Lemma 5.6

Proof 7.14.

The proof is shown in two stages. First, we show that the clique structure (see the examples in Fig. 13(a)) is the minimum subgraph not having any node with in-degree more than 2f2f while satisfying the condition for Algorithm 3. Due to the ff-local model, each node must have at least ff in-neighbors. It is obvious that if any node uses the majority voting structure (i.e., 2f+12f+1 two-hop paths) to obtain the original value of a two-hop in-neighbor or an out-neighbor, then such a node will have at least 2f+12f+1 in-neighbors. Consider node ii with ff in-neighbors. By the above discussion, it has undirected edges to the ff in-neighbors, which results in these in-neighbors being two-hop in-neighbors to each other. Thus, by the same argument, there must be undirected edges between them. Therefore, the clique is the only structure satisfying the condition for Algorithm 3 while not having any node with in-degree more than 2f2f.

Next, we show the minimum in-degree of the whole graph. Since the graph is strongly connected, there exist bi-directional edges (one undirected edge or two separate directed edges) connecting two subgraphs. For example, we take the undirected edge between nodes ii and jj in Fig. 13(b). Then other nodes in the right subgraph become two-hop neighbors of node ii. By similar discussions as above, there exist undirected edges between node ii and all the neighbors of node jj (as indicated by the blue edges in the figure). As a result, node ii has 2f+12f+1 in-neighbors. Moreover, we can check that all the rest of nodes also have 2f+12f+1 in-neighbors to fulfill the condition for Algorithm 3. We conclude that the whole graph has the minimum in-degree no less than 2f+12f+1.

Refer to caption
Refer to caption
Figure 13: (a) Examples: A clique is a complete subgraph. (b) Illustration for the minimum in-degree.

Proof of Proposition 5.9

Proof 7.15.

We first show that for any node i𝒱i\in\mathcal{V}, there must exist the minimum subgraph containing node ii as the one in Fig. 14(a). Recall from Lemma. 5.6 that an undirected graph satisfying the condition for Algorithm 3 has minimum in-degree no less than 2f+12f+1. In Fig. 14(a), we set f=1f=1 for illustration. The edges in blue and black represent the detectable path and the communication edge, respectively. Note that this subgraph also includes the middle nodes on the detectable path (not shown in the figure for convenience) if such path is not constructed by an undirected communication edge. It can be observed that in a minimum subgraph, after removing any node set being ff-local, the remaining graph is connected. This means that there is at least one path connecting any two nodes in the remaining graph.

Now, consider any two minimum subgraphs with node sets 𝒱1\mathcal{V}_{1} and 𝒱2\mathcal{V}_{2} (see Fig. 14(b)). There must be at least one edge between them since the whole graph is connected by assumption. There are three subcases for placing such an edge. These are between (i) ii and j1j_{1}, (ii) ii and jj, (iii) i1i_{1} and j1j_{1}. (Without loss of generality, select j1j_{1} as one of jj’s neighbors.) In cases (i) and (ii), node j1j_{1} or jj becomes a direct neighbor of node ii. Thus, node j1j_{1} or jj is connected with any node in 𝒱1\mathcal{V}_{1} after removing an ff-local node set in 𝒱1\mathcal{V}_{1}. Since node j1j_{1} or jj is also connected with any node in 𝒱2\mathcal{V}_{2} after the removal, we can conclude that in cases (i) and (ii), any node in 𝒱2\mathcal{V}_{2} is connected with any node in 𝒱1\mathcal{V}_{1} after removing an ff-local node set in the whole graph.

In case (iii), nodes i1i_{1} and j1j_{1} become neighbors (see Fig. 14(b)). There should be detectable paths between i1i_{1} and jj and also between ii and j1j_{1}. If any of the two paths is constructed by an undirected communication edge, the result is the same as the one in case (i). So we examine the case where both paths are constructed by 2f+12f+1 two-hop communication paths. Suppose that node ii is connected to node j1j_{1} through nodes i4i_{4}, i5i_{5}, and i6i_{6}. The three nodes become two-hop neighbors of node jj and there should be detectable paths to node jj. In this case, even if we remove the ff-local model consisting of both nodes i1i_{1} and j1j_{1}, node ii and node jj are connected with each other, and so are the rest of the nodes in 𝒱1\mathcal{V}_{1} and 𝒱2\mathcal{V}_{2}. Note that removing both nodes i1i_{1} and j1j_{1} does not violate the ff-local model since they do not have common normal neighbors. Finally, since the malicious set is ff-local, we conclude that the normal network induced by the normal agents in 𝒩𝒱\mathcal{N}\subseteq\mathcal{V} is connected.

Refer to caption
Refer to caption
Figure 14: (a) The minimum subgraph of an undirected graph satisfying the condition for Algorithm 3. (b) Illustration for two connected subgraphs 𝒱1\mathcal{V}_{1} and 𝒱2\mathcal{V}_{2}.

References

References

  • [1] N. A. Lynch, Distributed Algorithms. Morgan Kaufmann, 1996.
  • [2] R. Olfati-Saber, J. A. Fax, and R. M. Murray, “Consensus and cooperation in networked multi-agent systems,” Proc. IEEE, vol. 95, no. 1, pp. 215–233, 2007.
  • [3] F. Bullo, J. Cortés, and S. Martinez, Distributed Control of Robotic Networks: A Mathematical Approach to Motion Coordination Algorithms. Princeton University Press, 2009.
  • [4] M. Mesbahi and M. Egerstedt, Graph Theoretic Methods in Multi-agent Networks. Princeton University Press, 2010.
  • [5] G. S. Seyboth, D. V. Dimarogonas, and K. H. Johansson, “Event-based broadcasting for multi-agent average consensus,” Automatica, vol. 49, no. 1, pp. 245–252, 2013.
  • [6] K. Cai and H. Ishii, “Quantized consensus and averaging on gossip digraphs,” IEEE Transactions on Automatic Control, vol. 56, no. 9, pp. 2087–2100, 2011.
  • [7] K. Cai and H. Ishii, “Average consensus on general strongly connected digraphs,” Automatica, vol. 48, no. 11, pp. 2750–2761, 2012.
  • [8] M. E. Chamie, J. Liu, and T. Basar, “Design and analysis of distributed averaging with quantized communication,” IEEE Transactions on Automatic Control, vol. 61, no. 12, pp. 3870–3884, 2016.
  • [9] T. C. Aysal, M. J. Coates, and M. G. Rabbat, “Distributed average consensus with dithered quantization,” IEEE Transactions on Signal Processing, vol. 56, no. 10, pp. 4905–4918, 2008.
  • [10] A. Olshevsky and J. N. Tsitsiklis, “Convergence speed in distributed consensus and averaging,” SIAM Journal on Control and Optimization, vol. 48, no. 1, pp. 33–55, 2009.
  • [11] C. N. Hadjicostis, N. H. Vaidya, and A. D. Dominguez-Garcia, “Robust distributed average consensus via exchange of running sums,” IEEE Transactions on Automatic Control, vol. 61, no. 6, pp. 1492–1507, 2016.
  • [12] M. Zhu and S. Martinez, “Discrete-time dynamic average consensus,” Automatica, vol. 46, no. 2, pp. 322–329, 2010.
  • [13] E. Montijano, J. I. Montijano, C. Sagues, and S. Martinez, “Robust discrete time dynamic average consensus,” Automatica, vol. 50, no. 12, pp. 3131–3138, 2014.
  • [14] S. Yang, S. Tan, and J. Xu, “Consensus based approach for economic dispatch problem in a smart grid,” IEEE Transactions on Power Systems, vol. 28, no. 4, pp. 4416–4426, 2013.
  • [15] H. Ishii and R. Tempo, “The PageRank problem, multiagent consensus, and web aggregation: A systems and control viewpoint,” IEEE Control Systems Magazine, vol. 34, no. 3, pp. 34–53, 2014.
  • [16] H. Ishii and A. Suzuki, “Distributed randomized algorithms for PageRank computation: Recent advances,” in T. Basar (editor), Uncertainty in Complex Networked Systems: In Honor of Roberto Tempo, Birkhauser, pp. 419-447, 2018.
  • [17] L. Yuan and H. Ishii, “Event-triggered approximate Byzantine consensus with multi-hop communication,” IEEE Transactions on Signal Processing, vol. 71, pp. 1742–1754, 2023.
  • [18] S. Sundaram and B. Gharesifard, “Distributed optimization under adversarial nodes,” IEEE Transactions on Automatic Control, vol. 64, no. 3, pp. 1063–1076, 2018.
  • [19] H. J. LeBlanc, H. Zhang, X. Koutsoukos, and S. Sundaram, “Resilient asymptotic consensus in robust networks,” IEEE Journal on Selected Areas in Communications, vol. 31, no. 4, pp. 766–781, 2013.
  • [20] F. Pasqualetti, A. Bicchi, and F. Bullo, “Consensus computation in unreliable networks: A system theoretic approach,” IEEE Transactions on Automatic Control, vol. 57, no. 1, pp. 90–104, 2012.
  • [21] A. Fagiolini, F. Babboni, and A. Bicchi, “Dynamic distributed intrusion detection for secure multi-robot systems,” in Proc. IEEE Int. Conf. Robotics and Autom., 2009, pp. 2723–2728.
  • [22] S. M. Dibaji, M. Safi, and H. Ishii, “Resilient distributed averaging,” in Proc. American Control Conference, 2019, pp. 96–101.
  • [23] W. Zheng, Z. He, J. He, and C. Zhao, “Accurate resilient average consensus via detection and compensation,” in Proc. IEEE Conference on Decision and Control, 2021, pp. 5502–5507.
  • [24] C. N. Hadjicostis and A. D. Dominguez-Garcia, “Trustworthy distributed average consensus,” in Proc. IEEE Conference on Decision and Control, 2022, pp. 7403–7408.
  • [25] S. Sundaram and C. N. Hadjicostis, “Distributed function calculation via linear iterative strategies in the presence of malicious agents,” IEEE Transactions on Automatic Control, vol. 56, no. 7, pp. 1495–1508, 2011.
  • [26] A. Teixeira, D. Pérez, H. Sandberg, and K. H. Johansson, “Attack models and scenarios for networked control systems,” in Proc. 1st International Conference on High Confidence Networked Systems, 2012, pp. 55–64.
  • [27] A. Goldsmith, Wireless Communications. Cambridge University Press, 2005
  • [28] M. Azadmanesh and R. Kieckhafer, “Asynchronous approximate agreement in partially connected networks,” International Journal of Parallel and Distributed Systems and Networks, vol. 5, no. 1, pp. 26–34, 2002.
  • [29] N. H. Vaidya, L. Tseng, and G. Liang, “Iterative approximate Byzantine consensus in arbitrary directed graphs,” in Proc. ACM Symposium on Principles of Distributed Computing, 2012, pp. 365–374.
  • [30] L. Yuan and H. Ishii, “Resilient consensus with multi-hop communication,” in Proc. IEEE Conference on Decision and Control, 2021, pp. 2696–2701.
  • [31] L. Su and N. H. Vaidya, “Reaching approximate Byzantine consensus with multi-hop communication,” Information and Computation, vol. 255, pp. 352–368, 2017.
  • [32] S. M. Dibaji, H. Ishii, and R. Tempo, “Resilient randomized quantized consensus,” IEEE Transactions on Automatic Control, vol. 63, no. 8, pp. 2508–2522, 2018.
  • [33] L. Yuan and H. Ishii, “Asynchronous approximate Byzantine consensus via multi-hop communication,” in Proc. American Control Conference, 2022, pp. 755–760.
  • [34] D. M. Senejohnny, S. Sundaram, C. De Persis, and P. Tesi, “Resilience against misbehaving nodes in asynchronous networks,” Automatica, vol. 104, pp. 26–33, 2019.
  • [35] H. Ishii, Y. Wang, and S. Feng, “An overview on multi-agent consensus under adversarial attacks,” Annual Reviews in Control, vol. 53, pp. 252–272, 2022.
  • [36] L. Yuan and H. Ishii, “Secure consensus with distributed detection via two-hop communication,” Automatica, vol. 131, no. 109775, 2021.
  • [37] C. Zhao, J. He, and J. Chen, “Resilient consensus with mobile detectors against malicious attacks,” IEEE Transactions on Signal and Information Processing over Networks, vol. 4, no. 1, pp. 60–69, 2018.
  • [38] C. N. Hadjicostis and A. D. Dominguez-Garcia, “Trustworthy distributed average consensus based on locally assessed trust evaluations,” arXiv preprint, arXiv:2309.00920, 2023.
  • [39] C. N. Hadjicostis and A. D. Dominguez-Garcia, “Identification of malicious activity in distributed average consensus via non-concurrent checking,” IEEE Control Systems Letters, vol. 7, pp. 1927–1932, 2023.
  • [40] B. Parhami, “Voting algorithms,” IEEE Transactions on Reliability, vol. 43, no. 4, pp. 617–629, 1994.
  • [41] R. E. Blahut, Theory and Practice of Error Control Codes. Addison-Wesley, 1983.
  • [42] D. Kempe, A. Dobra, and J. Gehrke, “Gossip-based computation of aggregate information,” in Proc. Annual IEEE Symposium on Foundations of Computer Science, 2003, pp. 482–491.
  • [43] L. Tseng and N. H. Vaidya, “Fault-tolerant consensus in directed graphs,” in Proc. ACM Symposium on Principles of Distributed Computing, 2015, pp. 451–460.
  • [44] L. Tseng, N. Vaidya, and V. Bhandari, “Broadcast using certified propagation algorithm in presence of Byzantine faults,” Information Processing Letters, vol. 115, no. 4, pp. 512–514, 2015.
{IEEEbiography}

[[Uncaptioned image]]Liwei Yuan (Member) received the B.E. degree in Electrical Engineering and Automation from Tsinghua University, China, in 2017, and the Ph.D. degree in Computer Science from Tokyo Institute of Technology, Japan, in 2022. He is currently a Postdoctoral Researcher in the College of Electrical and Information Engineering at Hunan University, Changsha, China. His current research focuses on security in multi-agent systems and distributed algorithms.

{IEEEbiography}

[[Uncaptioned image]]Hideaki Ishii (M’02-SM’12-F’21) received the M.Eng. degree in applied systems science from Kyoto University, Kyoto, Japan, in 1998, and the Ph.D. degree in electrical and computer engineering from the University of Toronto, Toronto, ON, Canada, in 2002. He was a Postdoctoral Research Associate at the University of Illinois at Urbana-Champaign, Urbana, IL, USA, from 2001 to 2004, and a Research Associate at The University of Tokyo, Tokyo, Japan, from 2004 to 2007. He was an Associate Professor and Professor at the Department of Computer Science, Tokyo Institute of Technology, Yokohama, Japan in 2007–2024. Currently, he is a Professor at the Department of Information Physics and Computing at The University of Tokyo, Tokyo, Japan. He was a Humboldt Research Fellow at the University of Stuttgart in 2014–2015. He has also held visiting positions at CNR-IEIIT at the Politecnico di Torino, the Technical University of Berlin, and the City University of Hong Kong. His research interests include networked control systems, multiagent systems, distributed algorithms, and cyber-security of control systems.

Dr. Ishii has served as an Associate Editor for Automatica, the IEEE Control Systems Letters, the IEEE Transactions on Automatic Control, the IEEE Transactions on Control of Network Systems, and the Mathematics of Control, Signals, and Systems. He was a Vice President for the IEEE Control Systems Society (CSS) in 2022–2023. He was the Chair of the IFAC Coordinating Committee on Systems and Signals in 2017–2023. He served as the IPC Chair for the IFAC World Congress 2023 held in Yokohama, Japan. He received the IEEE Control Systems Magazine Outstanding Paper Award in 2015. Dr. Ishii is an IEEE Fellow.