This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

HI-GVF: Shared Control based on Human-Influenced Guiding Vector Fields for Human-multi-robot Cooperation

Pengming Zhu, Zongtan Zhou, Weijia Yao, Wei Dai, Zhiwen Zeng, and Huimin Lu This work was supported by the National Science Foundation of China under Grant 62203460, U22A2059, Major Project of the Natural Science Foundation of Hunan Province (No. 2021JC0004).Pengming Zhu, Zongtan Zhou, Wei Dai, Zhiwen Zeng, Huimin Lu are with the College of Intelligence Science and Technology, National University of Defense Technology, Changsha 410073, China (e-mail: zhupengming@nudt.edu.cn; narcz@163.com; weidai_nudt@foxmail.com; zengzhiwen@nudt.edu.cn; lhmnew@nudt.edu.cn).Weijia Yao is with School of Robotics, Hunan University, Changsha 410082, China (e-mail: wjyao@hnu.edu.cn).
Abstract

Human-multi-robot shared control leverages human decision-making and robotic autonomy to enhance human-robot collaboration. While widely studied, existing systems often adopt a leader-follower model, limiting robot autonomy to some extent. Besides, a human is required to directly participate in the motion control of robots through teleoperation, which significantly burdens the operator. To alleviate these two issues, we propose a layered shared control computing framework using human-influenced guiding vector fields (HI-GVF) for human-robot collaboration. HI-GVF guides the multi-robot system along a desired path specified by the human. Then, an intention field is designed to merge the human and robot intentions, accelerating the propagation of the human intention within the multi-robot system. Moreover, we give the stability analysis of the proposed model and use collision avoidance based on safety barrier certificates to fine-tune the velocity. Eventually, considering the firefighting task as an example scenario, we conduct simulations and experiments using multiple human-robot interfaces (brain-computer interface, myoelectric wristband, eye-tracking), and the results demonstrate that our proposed approach boosts the effectiveness and performance of the task.

Index Terms:
Human-robot interaction, multi-robot systems cooperation, shared control.

I Introduction

Multi-robot systems (MRS) have attracted much attention from the research community, and they are widely used in tasks such as monitoring [1], firefighting [2], search and rescue [3], source seeking [4], etc. Despite that considerable efforts have been made by researchers to enhance the autonomy of distributed multi-robot systems, there are still several challenges due to, e.g., perception inaccuracies, communication delays, and limited decision-making capabilities of robots [5]. Therefore, practical multi-robot systems often require human operators’ supervision and intervention, which can help assess and find potential targets, identify threats, react to emergencies and especially, provide intelligent guidance when planning algorithms fail. In many cases, human partners can operate remotely using the shared control technology to influence multi-robot systems, which can perform collaborative tasks with guaranteed safety autonomously when human influence is absent and thus alleviate the physical and mental workload on humans.

Shared control methods used in remote operating systems are usually based on mixed inputs [6, 7] or haptic feedback [8, 9]. There are various human-multi-robot interaction systems. For instance, a task-dependent graphical user multi-touch interface [10, 11], a gesture-based interface [12, 13] of a mobile handheld device [11], and a multi-modal interface with gesture, voice, and vision [14]. As noted in several reviews [15, 16], the effectiveness of such a team is primarily influenced by the type of information operators receive from the robots and the tools they have available to exert control. In some cases, multiple modalities are necessary for the expression of human intentions. Assume a scenario similar to Fig. 1, where an operator and a group of robots collaborate on a mission together, where the robots perform tasks in the rubble, and the operator provides a more comprehensive view via image transmission from the UAV. Humans hold remote operating equipment/tools and, therefore, need other methods to send commands to robots, such as covert hand gestures and smart helmets that can capture EEG signals as well as eye movement information. In this way, the operator can provide the robots with more information about the environment, where the target is (e.g., threats that need to be removed, trapped persons), and, if necessary, the operator even needs to be given a certain trajectory that will lead the robot to a certain area.

Refer to caption
Figure 1: Human and multi-robot collaboration.

Based on this motivation, this paper proposes a layered shared control computing framework based on human-influenced guiding vector fields (HI-GVF) for multi-robot systems, which is shown in Fig. 2. The blocks of different colors represent different research contents. Robots acquire different world models by sensors for calculation and decision-making, as shown in Fig. 2 (gray block).

Human operators acquire the world models not based on direct observations of the environment but on a virtual environment based on remote images. This method gives humans a broader view of the situation, which is useful for integrated planning. The operator can then give human intention by plotting a trajectory for robots or selecting a priority target area, as shown in Fig. 2 (green block). Robot intention and human intention are fused through a shared control framework, as shown in Fig. 2 (blue block). The components are explained in detail in subsequent sections.

The main contributions of this paper are as follows:

  • 1)

    We propose a shared control framework based on HI-GVF for human-multi-robot cooperation, without requiring humans to intervene at every moment, and the operator can intuitively give a certain desired path as an input to guide robots.

  • 2)

    We design a layered shared control framework, where an intention field is utilized to merge the robot intention generated by the robot local controller and the human intention generated by HI-GVF in the upper layer, and a policy-blending model is used to fuse various VFs in the lower layer.

  • 3)

    We give a stability analysis of the proposed method and use an obstacle avoidance method based on optimization in a minimally invasive way, which ensures stability in obstacle avoidance for the whole formation.

  • 4)

    We conduct simulations and physical experiments using multiple human-robot interfaces (brain-computer interface, myoelectric wristband, eye-tracking) based on self-designed human-multi-robot interaction system to verify the effectiveness of the proposed method.

Figure 2: The framework of the proposed human-influenced guiding vector field shared control method. The human operator and each robot are independent units, ultimately acting on the MRS through shared control. Different colored blocks are used to distinguish the different research contents in this figure.

II preliminary: guiding vector field

In this section, we introduce preliminary knowledge about guiding vector field. Then, we utilize HI-GVF to describe human intention, which is used to influence the collective behavior of the multi-robot system. The HI-GVF is derived based on a desired path given by the human operator, who expects a robot to follow the desired path for a specified period of time according to the human intention.

II-A GVF

To derive the guiding vector field, we first suppose that the desired path can be described by an implicit function:

𝒫={ξ2:ϕ(ξ)=0},\mathcal{P}=\{\xi\in\mathbb{R}^{2}:\phi(\xi)=0\}, (1)

where ϕ:2\phi:\mathbb{R}^{2}\to\mathbb{R} is twice continuously differentiable, and ξ=(x,y)2\xi=(x,y)\in\mathbb{R}^{2} represents the two-dimensional position. Then a guiding vector field proposed in [17] to solve the path-following problem is defined by:

𝒳(ξ)=γEϕ(ξ)kϕ(ξ)ϕ(ξ),\mathcal{X}(\xi)=\gamma E\nabla\phi(\xi)-k\phi(\xi)\nabla\phi(\xi), (2)

where EE is the 9090^{\circ} rotation matrix [0110]\left[\begin{smallmatrix}0&-1\\ 1&0\end{smallmatrix}\right], γ{1,1}\gamma\in\{1,-1\} is used to specify the propagation direction along the desired path, kk is an adjustable parameter. The first term of (2) is the tangential component, enabling the robot to move along the desired path, while the second term of (2) is the orthogonal component, assisting the robot to move closer to the desired path. Therefore, intuitively, the guiding vector field can guide the robot to move toward and along the desired path at the same time.

Refer to caption
Figure 3: The visualization of a guiding vector field. The corresponding desired path (red curve) is the circle described by choosing ϕ(x,y)=x2+y2R2\phi(x,y)=x^{2}+y^{2}-R^{2}, where RR is the circle radius, in (1). Each blue arrow represents a vector of the corresponding vector field at the position

.

II-B HI-GVF

When the desired path is occluded by obstacles, a robot needs to deviate from the desired path to avoid obstacles, and then return to the desired path. To address this problem, we design a smooth composite vector field, consisting of vector fields generated by the desired path and obstacles and integrated with two bump functions [18].

II-B1 Reactive boundary and repulsive boundary

Regardless of the shapes (circle or bar) of the obstacles, we design some boundaries to enclose each obstacle, such that avoiding collisions with the obstacles is simplified as avoiding collisions with the boundaries, as shown in Fig. 4(a). The reactive boundary \mathcal{R} and the repulsive boundary 𝒬\mathcal{Q} around the circle obstacle or bar barrier are defined as follows:

i={ξ2:φi(ξ)=0},\mathcal{R}_{i}=\{\xi\in\mathbb{R}^{2}:\varphi_{i}(\xi)=0\}, (3)
𝒬i={ξ2:φi(ξ)=ci},\mathcal{Q}_{i}=\{\xi\in\mathbb{R}^{2}:\varphi_{i}(\xi)=c_{i}\}, (4)

for ii\in\mathcal{I}, where \mathcal{I} represents the set of a finite number of obstacles, φi:2\varphi_{i}:\mathbb{R}^{2}\to\mathbb{R} is a twice continuously differentiable function, and ci0c_{i}\neq 0 is a given constant. The reactive boundary \mathcal{R} divides the workspace into the interior, denoted by in{}^{\text{in}}\mathcal{R}, and the exterior, denoted by ex{}^{\text{ex}}\mathcal{R}, as shown in Fig. 4(b). For convenience, in{}^{\text{in}}\mathcal{R} is also called the reactive area, and ex{}^{\text{ex}}\mathcal{R} is the non-reactive area. When the robot enters the reactive area, it is able to detect and react to the obstacle. Similarly, we define 𝒬in{}^{\text{in}}\mathcal{Q} and 𝒬ex{}^{\text{ex}}\mathcal{Q} to represent the repulsive area and the non-repulsive area, respectively, as shown in Fig. 4(c). The repulsive area is a region where the robot is forbidden to enter; otherwise collision with the obstacle would happen. Besides, we use ()¯\overline{(\cdot)} to represent the area containing the corresponding boundary (i.e., the closure of a given set). For instance, in¯\overline{{}^{\text{in}}\mathcal{R}} denotes the reactive area in{}^{\text{in}}\mathcal{R} containing the reactive boundary \mathcal{R}.

Refer to caption
(a) Desired path 𝒫\mathcal{P}, and the obstacle (black disk) with reactive boundary \mathcal{R} and repulsive boundary 𝒬\mathcal{Q}
Refer to caption
(b) Reactive area in{}^{\text{in}}\mathcal{R} (green elliptic disk) and non-reactive area ex{}^{\text{ex}}\mathcal{R}
Refer to caption
(c) Repulsive area 𝒬in{}^{\text{in}}\mathcal{Q} (pink elliptic disk) and non-repulsive area 𝒬ex{}^{\text{ex}}\mathcal{Q}
Figure 4: Illustration of the reactive and repulsive boundary and area of one obstacle.

For the desired path in (1) and the reactive boundary in (3), following (2), we can obtain the corresponding vector fields 𝒳𝒫,𝒳i\mathcal{X}_{\mathcal{P}},\mathcal{X}_{\mathcal{R}_{i}} with the constants γp,γi,kp,kri\gamma_{p},\gamma_{i},k_{p},k_{r_{i}} sharing the same meanings as those in (2).

II-B2 Zero-in and zero-out bump function

To smooth the motion of the robot as it passes through the reactive area where a robot is able to sense the obstacle, we use two smooth bump functions to generate a composite vector field so that the velocity does not change abruptly near the reactive boundary.

For a reactive boundary \mathcal{R} and a repulsive boundary 𝒬\mathcal{Q}, we can define smooth bump function i\sqcap_{\mathcal{R}_{i}}, 𝒬i\sqcup_{\mathcal{Q}_{i}} as follows:

i(ξ)={1ξ𝒬iin¯Zi(ξ)ξ𝒬iexiin0ξiex¯,\sqcap_{\mathcal{R}_{i}}(\xi)=\left\{\begin{array}[]{ll}1&\xi\in\overline{{}^{\text{in}}\mathcal{Q}_{i}}\\ Z_{i}(\xi)&\xi\in{{}^{\text{ex}}\mathcal{Q}_{i}}\cap{{}^{\text{in}}\mathcal{R}_{i}}\\ 0&\xi\in\overline{{}^{\text{ex}}\mathcal{R}_{i}}\end{array}\right., (5)
𝒬i(ξ)={0ξ𝒬iin¯Si(ξ)ξ𝒬iexiin1ξiex¯,\sqcup_{\mathcal{Q}_{i}}(\xi)=\left\{\begin{array}[]{ll}0&\xi\in\overline{{}^{\text{in}}\mathcal{Q}_{i}}\\ S_{i}(\xi)&\xi\in{{}^{\text{ex}}\mathcal{Q}_{i}}\cap{{}^{\text{in}}\mathcal{R}_{i}}\\ 1&\xi\in\overline{{}^{\text{ex}}\mathcal{R}_{i}}\end{array}\right., (6)

where Zi:𝒬iexiin(0,1)Z_{i}:{{}^{\text{ex}}\mathcal{Q}_{i}}\cap{{}^{\text{in}}\mathcal{R}_{i}}\rightarrow(0,1) and Si:𝒬iexiin(0,1)S_{i}:{{}^{\text{ex}}\mathcal{Q}_{i}}\cap{{}^{\text{in}}\mathcal{R}_{i}}\rightarrow(0,1) are smooth functions.

Refer to caption
(a) zero-out bump function i\sqcap_{\mathcal{R}_{i}}
Refer to caption
(b) zero-in bump function 𝒬i\sqcup_{\mathcal{Q}_{i}}
Figure 5: Illustration of smooth zero-out and zero-in bump functions.

Intuitively, we call i\sqcap_{\mathcal{R}_{i}} (see Fig. 5(a)) a smooth zero-out function with respect to the reactive boundary i\mathcal{R}_{i}, and 𝒬i\sqcup_{\mathcal{Q}_{i}} (see Fig. 5(b)) a smooth zero-in function with respect to the repulsive boundary 𝒬i{\mathcal{Q}_{i}}. Evidently, Zi(ξ)Z_{i}(\xi) approaches smoothly value 1 when ξ\xi approaches 𝒬i\mathcal{Q}_{i}, and approaches smoothly value 0 when ξ\xi approaches i\mathcal{R}_{i}, and the converse applies for Si(ξ)S_{i}(\xi).

II-B3 Composite vector field

Using the smooth zero-out and zero-in bump functions to blend two vector fields, we can obtain a composite vector field 𝒳c\mathcal{X}_{c} as follows:

𝒳c(ξ)=(i𝒬i(ξ))𝒳^𝒫(ξ)+i(i(ξ)𝒳^i(ξ)),\mathcal{X}_{c}(\xi)=\left(\prod_{i\in\mathcal{I}}\sqcup_{\mathcal{Q}_{i}}(\xi)\right)\hat{\mathcal{X}}_{\mathcal{P}}(\xi)+\sum_{i\in\mathcal{I}}\left(\sqcap_{\mathcal{R}_{i}}(\xi)\hat{\mathcal{X}}_{\mathcal{R}_{i}}(\xi)\right), (7)

where ()^\hat{(\cdot)} is the normalization notation (i.e., 𝒗^=𝒗/𝒗\hat{\boldsymbol{v}}=\boldsymbol{v}/\|\boldsymbol{v}\| for a non-zero vector). In our paper, it is assumed that different reactive areas do not overlap, which is a common assumption; if they overlap, then one can redefine the reactive boundaries such that this assumption is satisfied (e.g., by shrinking the sizes of the boundaries or by merging two boundaries into a large one). Therefore, according to (5) and (6), (7) is equivalent to:

𝒳c(ξ)={𝒳^i(ξ)ξ𝒬iin¯Si(ξ)𝒳^𝒫(ξ)+Zi(ξ)𝒳^i(ξ)ξ𝒬iexiin𝒳^𝒫(ξ)ξiex¯.\mathcal{X}_{c}(\xi)=\left\{\begin{array}[]{ll}\hat{\mathcal{X}}_{\mathcal{R}_{i}}(\xi)&\xi\in\overline{{}^{\text{in}}\mathcal{Q}_{i}}\\ S_{i}(\xi)\hat{\mathcal{X}}_{\mathcal{P}}(\xi)+Z_{i}(\xi)\hat{\mathcal{X}}_{\mathcal{R}_{i}}(\xi)&\xi\in{{}^{\text{ex}}\mathcal{Q}_{i}}\cap{{}^{\text{in}}}\mathcal{R}_{i}\\ \hat{\mathcal{X}}_{\mathcal{P}}(\xi)&\xi\in\overline{{}^{\text{ex}}\mathcal{R}_{i}}\end{array}\right.. (8)
Refer to caption
(a) VF 𝒳^𝒫\hat{\mathcal{X}}_{\mathcal{P}} for desired path 𝒫\mathcal{P}
Refer to caption
(b) VF 𝒳^i\hat{\mathcal{X}}_{\mathcal{R}_{i}} for reactive boundary i\mathcal{R}_{i}
Refer to caption
(c) VF 𝒬i(ξ)𝒳^𝒫(ξ)\sqcup_{\mathcal{Q}_{i}}(\xi)\hat{\mathcal{X}}_{\mathcal{P}}(\xi)
Refer to caption
(d) VF i(ξ)𝒳^i(ξ)\sqcap_{\mathcal{R}_{i}}(\xi)\hat{\mathcal{X}}_{\mathcal{R}_{i}}(\xi)
Refer to caption
(e) composite VF 𝒬i(ξ)𝒳^𝒫(ξ)+i(ξ)𝒳^i(ξ)\sqcup_{\mathcal{Q}_{i}}(\xi)\hat{\mathcal{X}}_{\mathcal{P}}(\xi)+\sqcap_{\mathcal{R}_{i}}(\xi)\hat{\mathcal{X}}_{\mathcal{R}_{i}}(\xi)
Figure 6: Illustration of construction of the composite vector field.

The illustration of the construction of the composite vector field is shown in Fig. 6. The red solid line represents the desired path 𝒫\mathcal{P}, the green dotted line represents the reactive boundary i\mathcal{R}_{i}, and the purple dotted line represents the repulsive boundary 𝒬i\mathcal{Q}_{i}. Each arrow in the subfigures represents a vector of the corresponding vector field at the position. In Fig. 6(e), the green arrows belong to the vector field 𝒳^i\hat{\mathcal{X}}_{\mathcal{R}_{i}} generated by the reactive boundary, and are all in 𝒬iin¯\overline{{}^{\text{in}}\mathcal{Q}_{i}}. The red arrows belong to the vector field 𝒳^𝒫\hat{\mathcal{X}}_{\mathcal{P}} generated by the desired path, and are all in iex¯\overline{{}^{\text{ex}}\mathcal{R}_{i}}. The blue arrows belong to the mixed vector field, the sum of 𝒳^𝒫\hat{\mathcal{X}}_{\mathcal{P}} and 𝒳^i\hat{\mathcal{X}}_{\mathcal{R}_{i}}, and are all in the mixed area 𝒬iexiin{{}^{\text{ex}}\mathcal{Q}_{i}}\cap{{}^{\text{in}}}\mathcal{R}_{i}. More discussion can be found in our previous work [18].

From this, the human intention 𝒗ih:22\boldsymbol{v}^{h}_{i}:\mathbb{R}^{2}\to\mathbb{R}^{2} for the ii-th robot is defined to be the following “zero-in” guiding vector field:

𝒗ih(ξ)=(j𝒬j(ξ))𝒳^𝒫(ξ).\boldsymbol{v}^{h}_{i}(\xi)=\left(\prod_{j\in\mathcal{I}}\sqcup_{\mathcal{Q}_{j}}(\xi)\right)\hat{\mathcal{X}}_{\mathcal{P}}(\xi). (9)

Notably, it is very difficult for human operators to establish stable communication with each robot because of wireless communication bandwidth limitations and environmental interference. Therefore, in this paper, we assume that the human operator influences only one of the robots in the formation, and the human intention is propagated through the communication links among the robots and thus indirectly influences the entire MRS. The human operator can intuitively draw a desired path via the mouse, the touch screen, or the eye-tracking device, etc., and can choose which robot is affected.

III methodology: multi-robot local control

Without human intervention, the robot should autonomously perform tasks through its own perception and decision-making in firefighting scenarios111In fact, our human-robot shared control approach can be used in a wide range of studies not limited to firefighting missions, and this paper uses firefighting missions as an illustrative example.. For such tasks, we assume that 1) a single robot has limited firefighting capability, and a single fire source requires multiple robots to work simultaneously to extinguish it. 2) individual robots should not leave the formation for a long period of time or long distances to maintain communication with others. In this section, we design a local controller for multi-robot coordination, including two subtasks: target selection (robot intention) and formation control, and these subtasks are accomplished by using different vector fields that will be designed in the following subsections. For convenience, we use the notation ab\mathbb{Z}_{a}^{b} to denote the set of integers {a,a+1,,b}\{a,a+1,\dots,b\} for a<ba<b.

III-A Robot Intention

Finding and prioritizing the most threatening fire source to extinguish is the primary task for the MRS. Each robot prioritizes the fire to be extinguished based on its perception of the fire situation, and generates the robot intention 𝒗ti\boldsymbol{v}_{t}^{i}, which points to the largest fire source observed by the ii-th robot.

Refer to caption
Figure 7: The environment influences the observed fire values.

Suppose there are NN homogeneous omnidirectional robots, with positions denoted by 𝒑ri2,i1N\boldsymbol{p}_{r_{i}}\in\mathbb{R}^{2},i\in\mathbb{Z}_{1}^{N}, MM fire sources, with positions denoted by 𝒑si2,i1M\boldsymbol{p}_{s_{i}}\in\mathbb{R}^{2},i\in\mathbb{Z}_{1}^{M}, PP circle obstacles, with positions denoted by 𝒑oi2,i1P\boldsymbol{p}_{o_{i}}\in\mathbb{R}^{2},i\in\mathbb{Z}_{1}^{P}, and QQ barriers, with positions denoted by 𝒑bi2,i1Q\boldsymbol{p}_{b_{i}}\in\mathbb{R}^{2},i\in\mathbb{Z}_{1}^{Q}, in the two-dimensional space. In this paper, the observed severity of the jj-th fire source by the ii-th robot is described by the observed fire value θ(𝒑ri,𝒑sj)0\theta(\boldsymbol{p}_{r_{i}},\boldsymbol{p}_{s_{j}})\geq 0. This value increases if the fire becomes more severe and decreases if obstacles obstruct the robot’s observation. It is worth emphasizing that, in practice, the observed fire value is directly measured by the onboard sensors. However, for simplicity, we use a circle centered at the fire source to represent the severity of the fire; the more severe the fire, the larger the radius. Then the observed fire value is approximated as the angle calculated by subtracting the angle of view obscured by the obstacle (α\alpha in Fig. 7) from the angle of view of the fire (β\beta in Fig. 7) when there are no obstacles. Note that if α>β\alpha>\beta, i.e., the fire source is totally obstructed by obstacles, then the observed fire value is zero. Fig. 7 graphically illustrates how the fire value is determined in an environment with fire sources, robots and obstacles. Therefore, the ii-th robot can select the most severe fire source as its prioritized target sis^{i}, which is defined as:

si=argmaxj1Mθ(𝒑ri,𝒑sj).s^{i}=\mathop{\arg\max}\limits_{j\in\mathbb{Z}_{1}^{M}}\theta(\boldsymbol{p}_{r_{i}},\boldsymbol{p}_{s_{j}}). (10)

In some cases, robots may not choose the same target. For example, in Fig. 7, two robots choose ➀ as their prioritized target while one robot has a different target: ➁. Then, similar to method (2), we define a vector field for the ii-th robot based on the target sis^{i}:

𝒳it(ξ)=θ(𝒑ri,𝒑si)ϕsi(ξ)ϕsi(ξ),\mathcal{X}^{t}_{i}(\xi)=-\theta(\boldsymbol{p}_{r_{i}},\boldsymbol{p}_{s^{i}})\phi_{s^{i}}(\xi)\nabla\phi_{s^{i}}(\xi), (11)

where ϕsi(ξ)=|ξ𝒑si|2Rsi2\phi_{s^{i}}(\xi)=|\xi-\boldsymbol{p}_{s^{i}}|^{2}-{R_{s^{i}}}^{2}, and 𝒑si\boldsymbol{p}_{s^{i}}, RsiR_{s^{i}} represent the position and radius of the prioritized target fire source sis^{i}, respectively. Note that the tangential component is removed and only the orthogonal component remains. Therefore, following this vector field (11), the robot would move directly towards the fire target.

III-B Formation Control

As some robots have different intentions, it may cause them to be separated from the team for too long or too far away, thus breaking away from the formation and resulting in a failure to work together to accomplish the mission. To address this problem, we adopt the distance-based formation control method to maintain the formation of MRS while moving.

Then, using (2), we can define a composite vector field for the ii-th robot based on the neighbors of the ii-th robot:

𝒳if(ξ)=j𝒩i|ξ𝒑rjdij|ϕ^rj(ξ),\mathcal{X}_{i}^{f}(\xi)=\sum_{j\in\mathcal{N}_{i}}-\left|\|\xi-\boldsymbol{p}_{r_{j}}\|-d_{ij}\right|\nabla\hat{\phi}_{r_{j}}(\xi), (12)

where dijd_{ij} represents the desired distance between the ii-th and jj-th robot, ϕrj(ξ)=ξ𝒑rj2dij2\phi_{r_{j}}(\xi)=\|\xi-\boldsymbol{p}_{r_{j}}\|^{2}-d_{ij}^{2}, ()^\hat{(\cdot)} is the normalization notation (i.e., it normalizes the gradient ϕrj\nabla\phi_{r_{j}}), and 𝒩i\mathcal{N}_{i} represents the neighbor set of the ii-th robot222If a robot has a (bidirectional) communication link with the ii-th robot, then it is a neighbor of the ii-th robot. For example, when a communication range RcR_{c} is set, other robots within the range are neighbors of ii-th robot. Human operators can also initialize the communication connection topology among robots before the task execution.. Due to the constraints of formation control, robots with greater robot intention will dominate when each robot moves to its selected target. Therefore, the target selection of the entire MRS will eventually become consistent.

Then, similar to (9), the VF induced by the target (i.e., the robot intention), and the VF for formation control are defined respectively as follows:

𝒗it(ξ)=(j1M+P+Q𝒬j(ξ))𝒳it(ξ),\boldsymbol{v}_{i}^{t}(\xi)=\left(\prod_{j\in\mathbb{Z}_{1}^{M+P+Q}}\sqcup_{\mathcal{Q}_{j}}(\xi)\right)\mathcal{X}_{i}^{t}(\xi), (13)
𝒗if(ξ)=(j1M+P+Q𝒬j(ξ))𝒳if(ξ).\boldsymbol{v}_{i}^{f}(\xi)=\left(\prod_{j\in\mathbb{Z}_{1}^{M+P+Q}}\sqcup_{\mathcal{Q}_{j}}(\xi)\right)\mathcal{X}_{i}^{f}(\xi). (14)

IV methodology: shared control for the human-multi-robot

So far, we have obtained the VFs based on the local controller and HI-GVF. Then, using a shared control method, the human intention and the robot intention can be reasonably fused to generate more appropriate actions, which can improve the execution efficiency of the system. In this section, a layered shared control framework is proposed to propagate the human intention and blend the VFs, effectively influencing the entire MRS by controlling one robot.

IV-A Upper Layer: Intention Field Model

In the upper layer of the framework, we use an intention field model to allow the human operator to control one robot and to propagate the intention among the robots, thus affecting the whole system. As stated in the preceding sections, 𝒗it2\boldsymbol{v}^{t}_{i}\in\mathbb{R}^{2} obtained by (13) represents the robot intention, and 𝒗ih2\boldsymbol{v}^{h}_{i}\in\mathbb{R}^{2} obtained by (9) represents the human intention received by the human-influenced robot. In each control cycle, the ii-th robot updates the shared intention 𝒗is\boldsymbol{v}_{i}^{s} at time step k+1k+1 as follows333For simplicity, the notation of the shared intention 𝒗is(ξ(k+1))\boldsymbol{v}_{i}^{s}(\xi(k+1)) at time step k+1k+1 is simplified as 𝒗is(k+1)\boldsymbol{v}_{i}^{s}(k+1); the same simplification is used for other notation. :

𝒗is(k+1)=\displaystyle\boldsymbol{v}_{i}^{s}(k+1)= ω0𝒗is(k)+ω1(𝒗it(k)𝒗is(k))\displaystyle-\omega_{0}\boldsymbol{v}_{i}^{s}(k)+\omega_{1}\left(\boldsymbol{v}_{i}^{t}(k)-\boldsymbol{v}_{i}^{s}(k)\right) (15)
+ω2j𝒩iΦ(𝒗js(k)𝒗is(k))\displaystyle+\omega_{2}\sum_{j\in\mathcal{N}_{i}}\Phi\left(\boldsymbol{v}_{j}^{s}(k)-\boldsymbol{v}_{i}^{s}(k)\right)
+ω3Ψi(𝒗ih(k)𝒗is(k)),\displaystyle+\omega_{3}\Psi_{i}\left(\boldsymbol{v}_{i}^{h}(k)-\boldsymbol{v}_{i}^{s}(k)\right),

where 𝒗is(0)=𝟎\boldsymbol{v}_{i}^{s}(0)=\boldsymbol{0}, ω0,ω1,ω2,ω3+\omega_{0},\omega_{1},\omega_{2},\omega_{3}\in\mathbb{R}_{+} are weighted coefficients that regulate the degree of influence of the robot intention, the neighbor intention, and the human intention in the update process, and Ψi\Psi_{i} is an indicator function that Ψi=1\Psi_{i}=1 holds when the ii-th robot has received the human intention 𝒗hi\boldsymbol{v}_{h}^{i} and Ψi=0\Psi_{i}=0 otherwise, and Φ()\Phi(\cdot) is a dead zone function defined as:

Φ(𝒙)={(𝒙ϵ)𝒙/𝒙,𝒙>ϵ𝟎,𝒙ϵ,\Phi(\boldsymbol{x})=\left\{\begin{array}[]{lll}(\|\boldsymbol{x}\|-\epsilon)\boldsymbol{x}/\|\boldsymbol{x}\|&,&\|\boldsymbol{x}\|>\epsilon\\ \mathbf{0}&,&\|\boldsymbol{x}\|\leq\epsilon\end{array},\right. (16)

where ϵ0\epsilon\geq 0. The intention field model can help the human operator to control effectively by abstracting the human input as an intention field value and fusing it with the robot intention. The shared intention can be viewed as a virtual control intention generated by robot intentions and human intentions. Assuming that the desired robot motion speed is 𝒞>0\mathcal{C}>0, the final shared intention can be represented as

𝒗^is=𝒞𝒗is/𝒗is.\hat{\boldsymbol{v}}_{i}^{s}=\mathcal{C}\boldsymbol{v}_{i}^{s}/\|\boldsymbol{v}_{i}^{s}\|. (17)

The model has the following properties:

  • 1)

    Local similarity. If the ii-th robot is close to the jj-th robot, then 𝒗si\boldsymbol{v}_{s}^{i} and 𝒗sj\boldsymbol{v}_{s}^{j} have similar values.

  • 2)

    Space separation. If the ii-th robot is far from the jj-th robot, then variation in 𝒗si\boldsymbol{v}_{s}^{i} has little impact on 𝒗sj\boldsymbol{v}_{s}^{j}.

Local similarity allows nearby robots to be controlled simultaneously by a human operator, while space separation enables different parts of the MRS to be controlled separately. With the intention field model, each robot updates itself under the influence of its neighbors and the human operator. When one robot receives the human intention, the inter-robot network implicitly propagates that intention to each robot. This result is fundamentally different from the “leader-follower” model, in which followers just react passively to the actions of their neighbors. The stability analysis of the intention field can be referred to in the Appendix-A.

IV-B Lower Layer: Policy-Blending Model

In the lower layer of the framework, a consensus network is used to blend the shared intention generated by the upper layer and VF of formation control generated by the local controller. We can obtain the final velocity 𝒗i\boldsymbol{v}^{i} as follows:

𝒗i=λi𝒗^is+(1λi)𝒗if,\boldsymbol{v}_{i}=\lambda_{i}\hat{\boldsymbol{v}}_{i}^{s}+(1-\lambda_{i})\boldsymbol{v}_{i}^{f}, (18)

where λi=λ(𝒗is,𝒗if)\lambda_{i}=\lambda(\|\boldsymbol{v}_{i}^{s}\|,\|\boldsymbol{v}_{i}^{f}\|), 𝒗is\boldsymbol{v}_{i}^{s}, 𝒗if\boldsymbol{v}_{i}^{f} and 𝒗^is\hat{\boldsymbol{v}}_{i}^{s} are obtained by (15), (14) and (17), respectively.

The weighting function λ:+×+[0,1]\lambda:\mathbb{R}^{+}\times\mathbb{R}^{+}\to[0,1] need to satisfy the following conditions:

  • 1)

    λ\lambda is continuous, non-negative, and λ(0,0)=0\lambda(0,0)=0.

  • 2)

    Let fb(a)=λ(a,b)f_{b}(a)=\lambda(a,b), then fb(a)f_{b}(a) is strictly increasing.

  • 3)

    Let fa(b)=λ(a,b)f_{a}(b)=\lambda(a,b), then fa(b)f_{a}(b) is strictly decreasing.

  • 4)

    limafb(a)=1\lim_{a\to\infty}f_{b}(a)=1 and limbfa(b)=0\lim_{b\to\infty}f_{a}(b)=0

We design a function that satisfies these conditions

λ(𝒗is,𝒗if)=ks𝒗isks𝒗is+kf𝒗if\lambda(\|\boldsymbol{v}_{i}^{s}\|,\|\boldsymbol{v}_{i}^{f}\|)=\frac{k_{s}\|\boldsymbol{v}_{i}^{s}\|}{k_{s}\|\boldsymbol{v}_{i}^{s}\|+k_{f}\|\boldsymbol{v}_{i}^{f}\|} (19)

where ks,kfR+k_{s},k_{f}\in R^{+} are constants.

The final velocity 𝒗i\boldsymbol{v}_{i} integrates all the VFs so that with suitable policy-blending, the MRS can maintain its formation and avoid obstacles while reaching the target. The stability analysis of the consensus network can be referred to in the Appendix-B.

V collision avoidance based on safety barrier certificates

An analysis of the stability of the whole formation when human and target intentions are used as inputs has already been given in the previous section. However, when the formation performs obstacle avoidance, it is difficult to ensure the stability of the whole formation anymore. Therefore, it is very important and necessary to adopt suitable collision avoidance methods. We use a safety barrier certificates method [19] based on optimization in a minimally invasive way, which ensures stability in obstacle avoidance for formation.

V-A collision avoidance among robots

Although formation control can keep robots at a distance from each other to some extent, it does not guarantee that robots will not collide with each other at all. Therefore, we still need to consider collision avoidance within the formation. By first defining a set of safe states, we use the control barrier function (CBF) to formally guarantee forward invariance of the desired set, i.e., if the system starts in the safe set, it stays in the safe set. A CBF is similar to a Control Lyapunov Function in that they ensure certain properties of the system without the need to explicitly compute the forward reachable set.

For i,j1N,iji,j\in\mathbb{Z}_{1}^{N},i\neq j, we define a function

𝔥ij(𝒑ri,𝒑rj)=𝒑ri𝒑rj2Rr2,\mathfrak{h}_{ij}(\boldsymbol{p}_{r_{i}},\boldsymbol{p}_{r_{j}})=\|\boldsymbol{p}_{r_{i}}-\boldsymbol{p}_{r_{j}}\|^{2}-R_{r}^{2}, (20)

where Rr>0R_{r}>0 is the safe distance among robots. The function 𝔥ij\mathfrak{h}_{ij} reflects whether two robots maintain a safe distance. Therefore, the pairwise safe set can be defined as

𝔖ij={(𝒑ri,𝒑rj)|𝔥ij(𝒑ri,𝒑rj)0ij}.\mathfrak{S}_{ij}=\{(\boldsymbol{p}_{r_{i}},\boldsymbol{p}_{r_{j}})|\mathfrak{h}_{ij}(\boldsymbol{p}_{r_{i}},\boldsymbol{p}_{r_{j}})\geq 0\quad\forall i\neq j\}. (21)

Using 𝔥ij\mathfrak{h}_{ij}, we can define a barrier function

Bij(𝒑ri,𝒑rj)=1𝔥ij(𝒑ri,𝒑rj).B_{ij}(\boldsymbol{p}_{r_{i}},\boldsymbol{p}_{r_{j}})=\frac{1}{\mathfrak{h}_{ij}(\boldsymbol{p}_{r_{i}},\boldsymbol{p}_{r_{j}})}. (22)

We use the optimization method to adjust the final velocity 𝒗i\boldsymbol{v}_{i} so that the new velocity 𝒗^i\hat{\boldsymbol{v}}_{i} enables collision-avoidance. Specifically, the new velocity 𝒗^i\hat{\boldsymbol{v}}_{i} should subject to the constraint B˙ijα/Bij\dot{B}_{ij}\leq\alpha/B_{ij}, where α>0\alpha>0 is a constant. Using (22), we can obtain B˙ij=𝔥˙ij/𝔥ij2=2(𝒑ri𝒑rj)(𝒗^i𝒗^j)/𝔥ij2\dot{B}_{ij}=-\dot{\mathfrak{h}}_{ij}/\mathfrak{h}_{ij}^{2}=-2(\boldsymbol{p}_{r_{i}}-\boldsymbol{p}_{r_{j}})^{\top}(\hat{\boldsymbol{v}}_{i}-\hat{\boldsymbol{v}}_{j})/\mathfrak{h}_{ij}^{2}, so the constraint B˙ijα/Bij\dot{B}_{ij}\leq\alpha/B_{ij} can be rewritten as

(𝒑ri𝒑rj)𝒗^i+(𝒑ri𝒑rj)𝒗^jα2𝔥ij3.-(\boldsymbol{p}_{r_{i}}-\boldsymbol{p}_{r_{j}})^{\top}\hat{\boldsymbol{v}}_{i}+(\boldsymbol{p}_{r_{i}}-\boldsymbol{p}_{r_{j}})^{\top}\hat{\boldsymbol{v}}_{j}\leq\frac{\alpha}{2}\mathfrak{h}_{ij}^{3}. (23)

In order to ensure that the main control objectives are met to the greatest extent possible, the collision avoidance strategy should modify the target velocity as little as possible. Therefore, we calculate the new velocity 𝒗^=(𝒗^1,,𝒗^N)2N\hat{\boldsymbol{v}}=(\hat{\boldsymbol{v}}_{1},\cdots,\hat{\boldsymbol{v}}_{N})\in\mathbb{R}^{2N} by the quadratic program (QP) as follows:

𝒗^=argmin𝒗^2N\displaystyle\hat{\boldsymbol{v}}^{*}=\underset{\hat{\boldsymbol{v}}\in\mathbb{R}^{2N}}{\text{argmin}} i=1N𝒗^i𝒗i2\displaystyle\sum_{i=1}^{N}\left\|\hat{\boldsymbol{v}}_{i}-\boldsymbol{v}_{i}\right\|^{2} (24)
s.t. (23),ij\displaystyle\eqref{constraint},\quad\forall i\neq j

This optimization problem is centralized as it requires the positions and desired velocities of all robots. Further, we can improve it to a distributed form for each robot i1Ni\in\mathbb{Z}_{1}^{N}:

𝒗^i=argmin𝒗^i2\displaystyle\hat{\boldsymbol{v}}_{i}^{*}=\underset{\hat{\boldsymbol{v}}_{i}\in\mathbb{R}^{2}}{\text{argmin}} 𝒗^i𝒗i2\displaystyle\left\|\hat{\boldsymbol{v}}_{i}-\boldsymbol{v}_{i}\right\|^{2} (25)
s.t. (𝒑ri𝒑rj)𝒗^iα4𝔥ij3,j𝒩i\displaystyle-(\boldsymbol{p}_{r_{i}}-\boldsymbol{p}_{r_{j}})^{\top}\hat{\boldsymbol{v}}_{i}\leq\frac{\alpha}{4}\mathfrak{h}_{ij}^{3},\quad\forall j\in\mathcal{N}_{i}

Therefore, robot ii only considers the positions of its neighbors, and does not need to get the desired velocity of other robots.

V-B Collision avoidance between robots and obstacles

Note that the safety barrier certificates can also deal with static or moving obstacles by treating obstacles as agents with no inputs. Similar to the method of collision avoidance among robots, the robot needs to maintain a safe distance between obstacles. Therefore, a function can be designed to ensure that the robot ii does not collide with the obstacle 𝒐k\boldsymbol{o}_{k}

𝔥¯ik(𝒑ri,𝒑ok)=𝒑ri𝒑ok2Ro2\bar{\mathfrak{h}}_{ik}(\boldsymbol{p}_{r_{i}},\boldsymbol{p}_{o_{k}})=\|\boldsymbol{p}_{r_{i}}-\boldsymbol{p}_{o_{k}}\|^{2}-R_{o}^{2} (26)

where RoR_{o} is the safe distance between the robot and the obstacle.

Similar to (23), we can obtain the constraint of robot avoiding obstacles:

(𝒑ri𝒑ok)𝒗^iβ2𝔥¯ij3,-(\boldsymbol{p}_{r_{i}}-\boldsymbol{p}_{o_{k}})^{\top}\hat{\boldsymbol{v}}_{i}\leq\frac{\beta}{2}\bar{\mathfrak{h}}_{ij}^{3}, (27)

where β>0\beta>0 is a constant. Then, the new velocity 𝒗^i\hat{\boldsymbol{v}}_{i} can be calculated by the QP as follows:

𝒗^i=argmin𝒗^i2\displaystyle\hat{\boldsymbol{v}}_{i}^{*}=\underset{\hat{\boldsymbol{v}}_{i}\in\mathbb{R}^{2}}{\text{argmin}} 𝒗^i𝒗i2\displaystyle\left\|\hat{\boldsymbol{v}}_{i}-\boldsymbol{v}_{i}\right\|^{2} (28)
s.t. (𝒑ri𝒑rj)𝒗^iα4𝔥ij3,j𝒩i\displaystyle-(\boldsymbol{p}_{r_{i}}-\boldsymbol{p}_{r_{j}})^{\top}\hat{\boldsymbol{v}}_{i}\leq\frac{\alpha}{4}\mathfrak{h}_{ij}^{3},\quad\forall j\in\mathcal{N}_{i}
(𝒑ri𝒑ok)𝒗^iβ2𝔥¯ik3.\displaystyle-(\boldsymbol{p}_{r_{i}}-\boldsymbol{p}_{o_{k}})^{\top}\hat{\boldsymbol{v}}_{i}\leq\frac{\beta}{2}\bar{\mathfrak{h}}_{ik}^{3}.

It is worth noting that the collision avoidance behavior may affect the convergence of the robot formation, and even deadlocks may occur at certain positions (the deadlock problem is still an intractable challenge at this point), resulting in robots not being able to move. Thus, to ensure that multi-robot cooperative tasks and collision avoidance behavior are feasible, we assume that the required parameters have been properly designed, that the initial Euclidean distances between the robots are all greater than the safe distance RrR_{r}, and that the initial distances between robots and obstacles are all greater than the safe distance RoR_{o}. When the robots are unable to move due to the deadlock problem, we temporarily disengage the robots from the deadlock position by increasing the human intent.

VI Validation

To verify the effectiveness of our proposed method, we designed a series of experiments based on simulation and physical MRS platforms, respectively.

Given the function φi\varphi_{i} and cic_{i} as in (4). We design the function SiS_{i} and ZiZ_{i} as follows:

Si(ξ)=f1(ξ)f1(ξ)+f2(ξ)S_{i}(\xi)=\frac{f_{1}(\xi)}{f_{1}(\xi)+f_{2}(\xi)} (29)
Zi(ξ)=f2(ξ)f1(ξ)+f2(ξ)Z_{i}(\xi)=\frac{f_{2}(\xi)}{f_{1}(\xi)+f_{2}(\xi)} (30)

where f1f_{1} and f2f_{2} are defined as follows:

f1(ξ)={0ξ{φ(ξ)ci}exp(l1ciφ(ξ))ξ{φ(ξ)>ci}f2(ξ)={exp(l2φ(ξ))ξ{φ(ξ)<0}0ξ{φ(ξ)0}\begin{array}[]{c}f_{1}(\xi)=\left\{\begin{array}[]{ll}0&\xi\in\{\varphi(\xi)\leq c_{i}\}\\ \exp\left(\frac{l_{1}}{c_{i}-\varphi(\xi)}\right)&\xi\in\{\varphi(\xi)>c_{i}\}\end{array}\right.\\ f_{2}(\xi)=\left\{\begin{array}[]{ll}\exp\left(\frac{l_{2}}{\varphi(\xi)}\right)&\xi\in\{\varphi(\xi)<0\}\\ 0&\xi\in\{\varphi(\xi)\geq 0\}\end{array}\right.\end{array} (31)

where l1>0l_{1}>0, l2>0l_{2}>0 are used to change the decaying or increasing rate.

Refer to caption
Figure 8: The details of simulation environment.
Refer to caption
(a) MRS follows the desired path
Refer to caption
(b) MRS follows the desired path with changing affected robots
Refer to caption
(c) MRS follows the desired path when encountering obstacles
Refer to caption
Figure 9: The robot trajectories when MRS follows the desired path.
Refer to caption
(a) MRS performs task without human intervention
Refer to caption
(b) Changing target priority with human
Refer to caption
(c) Finding undetected target
Refer to caption
Figure 10: MRS perform task with human intervention.

VI-A Simulation

We design a human-computer interface based on QT for simulation experiments as shown in Fig. 8. The operator can arbitrarily initialize the robot position and connection, modify parameters, etc. We set up the following three types of experiments to verify the effectiveness of the algorithm.

VI-A1 MRS follows the desired path without robot intention

Given the desired path as a circle with center (250,300) and radius 250, the initial positions of the three robots are (590,260), (540,190), (490,260), and (910,500), (870,560), (960,560), which are located inside and outside the circle, respectively.

Case 1: MRS follows the desired path. Selecting a certain robot to be affected, this affected robot is able to follow the desired path from inside or outside the circle. Besides, the other robots can keep moving in formation, thus generating similar trajectories, as shown in Fig. 9(a).

Case 2: MRS follows the desired path with changing affected robots. When the MRS follows the desired path, the human can choose any robot affected. Thanks to the influence of the intention field, the entire MRS can respond quickly and continue to follow the desired path, as shown in Fig. 9(b).

Case 3: MRS follows the desired path when encountering obstacles. In this case, there are bar barriers and circle obstacles in the environment. The robots will temporarily deviate from the desired path and change formation to avoid obstacles, which can be irregularly shaped, as shown in Fig. 9(c).

VI-A2 Shared control of human intervention

For firefighting tasks, we consider the following scenario where human intervention can make the MRS more efficient to reduce the damage caused by fire. There are five fire sources, several circle obstacles, and a bar barrier in the environment. The initial setup of these experiments is shown in TABLE I. The fire source radius increases with time (1m/s), and the time required to extinguish the fire is proportional to the size of the fire; the area covered by the ii-th fire source is denoted by the loss area a(si)a(s_{i}). The ultimate goal of the MRS in extinguishing the fire is to minimize the total loss area.

TABLE I: Initial setup of simulation experiments.
Robot(Position) (700,300)   (670,350)   (730,350)
Fire source(Position)[Size]
(500,250)[20]    (300,450)[30]   (170,120)[50]
(650,600)[10]    (1100,300)[40]
Obstacles(Position)[Size]
(300,200)[20]    (670,450)[20]   (480,400)[20]
(150,450)[20]    (650,140)[20]

Case 1: MRS changes target priority with human intervention. The robots choose the next priority fire to be extinguished based on their observations. The MRS will maintain the formation and make adjustments when one robot encounters obstacles while performing the task. During the firefighting process, the robots are distributed to the appropriate locations according to the size of the fire. After extinguishing fire source 1, the robots choose fire source 2 over the actual more threatening fire source 3 due to the obstruction caused by obstacle 1. When humans intervene, the robot will be guided to change its trajectory to discover and select the fire source 3, as shown in Fig. 10. The results in Table II show that the area of damage caused by the fire is reduced after human intervention.

Case 2: MRS finds undetected target with human intervention. After the robots extinguish fire source 2, there is still a fire source left in the environment. Because of the barriers, the robot could not detect the fire until the fire spreads further, which would cause more damage. At this point, human intervention can guide the robot to detect the target early and reduce the damage, as shown in Table II.

TABLE II: comparison of the loss areas with different situations.
Loss area Without human With human
Robot autonomous
operation
Change target
priority
Find undetected
target
a(s1)a(s_{1}) 3512.61 3512.61 3512.61
a(s2)a(s_{2}) 7026.29 8639.3 7026.29
a(s3)a(s_{3}) 16203.8 14491.39 16203.8
a(s4)a(s_{4}) 4646.9 4273 4646.9
a(s5)a(s_{5}) 20157.5 17335.6 17920.2
Sum 51547.1 48253.9 49309.8

VI-B Human-computer interaction methods

As we mentioned earlier, we assume that in a scenario of human-multi-robot collaboration, humans obtain a global view through the UAV, thus guiding the robots to accomplish the task together using multiple human-robot interactions. We briefly overview the brain-computer interfaces, myoelectric wristbands, and eye-tracking devices used.

VI-B1 brain-computer interface

The brain-computer interface (BCI) is a unique form of human-computer interaction that uses brain activity to communicate with the external world [20]. Researchers typically capture brainwave signals using non-invasive electrodes placed on the scalp. Unlike traditional methods, BCI generates commands directly from EEG signals, reflecting users’ choices and uncertainties [21]. Therefore, BCI is ideal for controlling intelligent robots.

Our experiment uses the NeuSen W wireless EEG system, which is portable, stable, and well-shielded, equipped with a nine-axis motion sensor for both controlled and natural experiments. The hardware includes a 64-channel EEG cap and an amplifier, with electrodes positioned according to the international 10-20 system.

VI-B2 myoelectric wristbands

The EMG interface captures electromyography signals from the skin to interpret human intent. Surface-measured muscle contractions range from tens to hundreds of microvolts. Although weak and prone to noise, EMG signals directly reflect body activity, making them an advanced HCI for predicting user intentions [22]. Extensive research has utilized EMG to control robots [23, 24, 25, 13].

We have developed an EMG acquisition device that contains an array of standard medical snap buttons compatible with dry Ag/AgCl and gel electrodes. These snap buttons are arranged longitudinally in pairs along the arm to detect distinct muscle activities. Our wristband supports eight channels for surface EMG acquisition. This multi-channel design improves sensitivity and fidelity, enabling the detection of subtle hand gestures and fine motions.

VI-B3 eye-tracking device

The eye-tracker monitors and records human eye movements using infrared light and camera technology. This data can be analyzed to understand users’ intentions, predict future actions, and gain insights into their cognitive states [26]. Eye-tracking techniques have been widely used to recognize human intentions for indirect input or task assistance [27, 28, 29].

Intent prediction through eye movements is natural and seamless, serving as an efficient interface. Our system uses a simple Tobii eye-tracker to capture eye movements. The eye-tracker captures gaze positions over time, allowing us to derive gaze trajectories and construct heat maps that illustrate attention distribution and estimate cognitive workload.

VI-C Physical Multi-robot Experiment

Based on our self-designed human-multi-robot interaction system, we conduct physical robotics experiments. All participants were provided with written informed consent to participate in the experiment in accordance with the Declaration of Helsinki. We consider the following two scenarios: (1) The desired path is given in advance, and the robot moves along the path motion and deviates from the path when the target is found. (2) While the robot is executing the task, a human intervenes to change the execution sequence.

VI-C1 case 1

When the robot cannot directly observe the target, it is slow to search for it by directly exploring the environment. At this point, the human operator can utilize the information provided by the UAV to guide the robot’s movement, given the trajectory, which can improve the efficiency of the task. Consider a scenario like the one in Fig. 11, where the environment has circular and bar-shaped obstacles, simulating objects such as general obstacles and walls in real situations.The robot should avoid collisions with obstacles as well as other robots at all times during the task. Besides, it is assumed that the robot has a limited perceptual range and can only detect the target when it is close to it. The red line is the human’s given path. When the robots find the target, they break away from the path, and when the task is completed, the robots can still be guided to the given path.

Figure 11: physical experimental scenario for case 1.

The experimental result is shown in Fig. 12. The robots achieve the obstacle avoidance requirement throughout, even when the desired path crosses barriers. Besides, The robots are able to follow the desired path and reach the target position for operation.

Figure 12: Robot trajectories of case 1.

VI-C2 case 2

A human operator intervenes online in real-time with a robot based on a human-multi-robot interaction interface. In this case, the eye-tracker is used to capture the gaze to draw a desired path, the BCI interface is used to select the robot to be guided, and gestures are used to switch the task progress. The experimental equipment is shown in Fig. 13. Although the operator in the picture is holding the drone remote control in his hand, it is only to simulate that the operator can use tiny gestures for ground robot control even though his hands are manipulating the drone. In fact, we do not use the remote control for the whole experiment. It is worth noting that the entire task process does not require using traditional interaction methods such as mouse and keyboard, which provides a flexible and feasible solution for human-multi-robot collaboration.

Refer to caption
Figure 13: Various HCI devices used in the experiment.
Figure 14: Physical experimental scenario for case 2.

The experimental scenario is shown in Fig. 14. In this case, the robots are not able to observe targrt 3, which has the highest threat level, due to the occlusion of the barrier. The robots operate autonomously with an execution sequence target 2 \rightarrow target 1 \rightarrow target 3, which leads to great damage. At this point, the human operator guides the robot around the obstacle to prioritize the elimination of target 3. Fig. 15(a)-(f) show the motion of the robot in sequence. The direction of the red arrow in the figure represents the robot’s current direction of motion. In Fig. 15(a), the robots are close to target 2, and target 2 is a greater threat than target 1, so the robots prioritize movement toward target 2. When target 2 is eliminated, the robots can only observe target 1 due to the obstacles, so the robots autonomously move towards target 1 (see Fig. 15(b)). However, in reality, the threat level of target 3 is greater. At this point, the human operator guides the robot along the desired path by first pausing the robot’s motion through hand gestures, giving the desired path through eye movements, and selecting the affected robot through a brain-computer interface (see Fig. 15(c)). When the robots move to the end of the desired path, the robots move toward target 2 (see Fig. 15(d)), and later move towards target 1 avoiding obstacles. The robot trajectories are show in Fig. 16.

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Refer to caption
(e)
Refer to caption
(f)
Figure 15: The execution process with human intervention.
Figure 16: Physical experimental scenario for case 2.

VII Conclusion

Multi-robot teams can robustly accomplish large-scale complex missions. However, robots often need human involvement to solve tough dilemmas. In this context, a shared control method integrating human intentions within a multi-robot team can improve the execution capabilities of multi-robot systems. In this paper, a novel shared control method is proposed that allows humans to use desired paths (or certain areas) as inputs without requiring real-time human involvement in the motion control of robots, thus reducing the burden on human operators. In our method, with human-influenced guiding vector fields utilized to describe human intentions, a hierarchical shared control framework is proposed to propagate and merge the robot and human intentions. In addition, a novel human-multi-robot interaction system based on multi-touch screens is built for a physical demonstration to validate the proposed algorithm. Besides, we use a variety of human-robot interactions (brain-computer interface, myoelectric wristband, eye-tracking) to intervene in the robot’s task progression. The results in different scenarios show that our method promotes the effectiveness and efficiency for completing the tasks.

Appendix

VII-A The stability analysis of the intentional field

Consider the robot as a first-order model, i.e., 𝒙˙i=𝒗i\dot{\boldsymbol{x}}_{i}=\boldsymbol{v}_{i}. Equation (15) can be rewritten as follows:

𝒗is(k+1)=\displaystyle\boldsymbol{v}_{i}^{s}(k+1)= ω0𝒗is(k)+ω1(𝒗it(k)𝒗is(k))\displaystyle-\omega_{0}\boldsymbol{v}_{i}^{s}(k)+\omega_{1}\left(\boldsymbol{v}_{i}^{t}(k)-\boldsymbol{v}_{i}^{s}(k)\right) (32)
+ω2j𝒩iΦ(𝒗js(k)𝒗is(k))\displaystyle+\omega_{2}\sum_{j\in\mathcal{N}_{i}}\Phi\left(\boldsymbol{v}_{j}^{s}(k)-\boldsymbol{v}_{i}^{s}(k)\right)
+ω3Ψi(𝒗ih(k)𝒗is(k)).\displaystyle+\omega_{3}\Psi_{i}\left(\boldsymbol{v}_{i}^{h}(k)-\boldsymbol{v}_{i}^{s}(k)\right).

For the whole robot formation, the state can be represented as 𝒙=[𝒙1,,𝒙N]2N\boldsymbol{x}=[\boldsymbol{x}_{1}^{\top},\cdots,\boldsymbol{x}_{N}^{\top}]^{\top}\in\mathbb{R}^{2N}. Let Ψ=diag(Ψi)I22N×2N\Psi=\text{diag}({\Psi_{i}})\otimes I_{2}\in\mathbb{R}^{2N\times 2N}, Φ(𝒙)=[ϕ(𝒙1),,ϕ(𝒙N)]\Phi(\boldsymbol{x})=[\phi(\boldsymbol{x}_{1})^{\top},\cdots,\phi(\boldsymbol{x}_{N})^{\top}]^{\top}, 𝒗s=[𝒗1s,,𝒗Ns]\boldsymbol{v}_{s}=[{\boldsymbol{v}^{s}_{1}}^{\top},\cdots,{\boldsymbol{v}^{s}_{N}}^{\top}]^{\top}, 𝒗t=[𝒗1t,,𝒗Nt]\boldsymbol{v}_{t}=[{\boldsymbol{v}^{t}_{1}}^{\top},\cdots,{\boldsymbol{v}^{t}_{N}}^{\top}]^{\top}, 𝒗h=[𝒗1h,,𝒗Nh]\boldsymbol{v}_{h}=[{\boldsymbol{v}^{h}_{1}}^{\top},\cdots,{\boldsymbol{v}^{h}_{N}}^{\top}]^{\top}, where \otimes is the Kronecker product. Then, (32) becomes

𝒗s=ω0𝒗s+ω1(𝒗t𝒗s)ω2D~Φ(D~𝒗s)+ω3Ψ(𝒗h𝒗s),\boldsymbol{v}_{s}=-\omega_{0}\boldsymbol{v}_{s}+\omega_{1}(\boldsymbol{v}_{t}-\boldsymbol{v}_{s})-\omega_{2}\widetilde{D}\Phi(\widetilde{D}^{\top}\boldsymbol{v}_{s})+\omega_{3}\Psi(\boldsymbol{v}_{h}-\boldsymbol{v}_{s}), (33)

where D~=DI2\widetilde{D}=D\otimes I_{2}, D is the incidence matrix. We use the Lyapunov method for stability analysis of the intention field model.

Theorem 1.

The intention field model with 𝐯t\boldsymbol{v}_{t} and 𝐯s\boldsymbol{v}_{s} as the input, and 𝐯s\boldsymbol{v}_{s} as the state, is input-to-state stable. Moreover, let 𝐯s=lim supt𝐯s(t)2\|\boldsymbol{v}_{s}\|_{\infty}=\limsup\limits_{t\to\infty}\|\boldsymbol{v}_{s}(t)\|_{2}, γt=12ω1ω0\gamma_{t}=\frac{1}{2}\sqrt{\frac{\omega_{1}}{\omega_{0}}}, and γh=12ω3ω0\gamma_{h}=\frac{1}{2}\sqrt{\frac{\omega_{3}}{\omega_{0}}}, we can obtain

𝒗sγt𝒗t+γh𝒗h.\|\boldsymbol{v}_{s}\|_{\infty}\leq\gamma_{t}\|\boldsymbol{v}_{t}\|_{\infty}+\gamma_{h}\|\boldsymbol{v}_{h}\|_{\infty}. (34)
Proof.

Let V(𝒗s)=12𝒗s𝒗sV(\boldsymbol{v}_{s})=\frac{1}{2}\boldsymbol{v}_{s}^{\top}\boldsymbol{v}_{s}, from (32), we can obtain

V˙=\displaystyle\dot{V}= ω0𝒗s2+ω1𝒗sΨ(𝒗t𝒗s)\displaystyle-\omega_{0}\|\boldsymbol{v}_{s}\|^{2}+\omega_{1}\boldsymbol{v}_{s}^{\top}\Psi(\boldsymbol{v}_{t}-\boldsymbol{v}_{s}) (35)
ω2𝒗sD~Φ(D~𝒗s)+ω3𝒗sΨ(𝒗h𝒗s).\displaystyle-\omega_{2}\boldsymbol{v}_{s}^{\top}\widetilde{D}\Phi(\widetilde{D}^{\top}\boldsymbol{v}_{s})+\omega_{3}\boldsymbol{v}_{s}^{\top}\Psi(\boldsymbol{v}_{h}-\boldsymbol{v}_{s}).

More generally, suppose robots 1,,N1,\cdots,N, all have their own target, thus from 𝒗is𝒗it𝒗is2+14𝒗it2\boldsymbol{v}^{s}_{i}{{}^{\top}}\boldsymbol{v}^{t}_{i}\leq\|\boldsymbol{v}^{s}_{i}\|^{2}+\frac{1}{4}\|\boldsymbol{v}^{t}_{i}\|^{2}, we have

𝒗sΨ(𝒗t𝒗s)=i=1N(𝒗is𝒗it𝒗is2)𝒗t24.\boldsymbol{v}_{s}^{\top}\Psi(\boldsymbol{v}_{t}-\boldsymbol{v}_{s})=\sum_{i=1}^{N}(\boldsymbol{v}_{i}^{s}{{}^{\top}}\boldsymbol{v}^{t}_{i}-\|\boldsymbol{v}^{s}_{i}\|^{2})\leq\frac{\|\boldsymbol{v}_{t}\|^{2}}{4}. (36)

Likewise, we have

𝒗sΨ(𝒗h𝒗s)=i=1N(𝒗is𝒗ih𝒗is2)𝒗h24.\boldsymbol{v}_{s}^{\top}\Psi(\boldsymbol{v}_{h}-\boldsymbol{v}_{s})=\sum_{i=1}^{N}(\boldsymbol{v}^{s}_{i}{{}^{\top}}\boldsymbol{v}^{h}_{i}-\|\boldsymbol{v}^{s}_{i}\|^{2})\leq\frac{\|\boldsymbol{v}_{h}\|^{2}}{4}. (37)

Since 𝒗sD~Φ(D~𝒗s)0\boldsymbol{v}_{s}^{\top}\widetilde{D}\Phi(\widetilde{D}^{\top}\boldsymbol{v}_{s})\geq 0, from (35) we have

V˙\displaystyle\dot{V} ω0𝒗s2+14ω1𝒗t2+14ω3𝒗h2\displaystyle\leq-\omega_{0}\|\boldsymbol{v}_{s}\|^{2}+\frac{1}{4}\omega_{1}\|\boldsymbol{v}_{t}\|^{2}+\frac{1}{4}\omega_{3}\|\boldsymbol{v}_{h}\|^{2} (38)
ω0𝒗s2+14(ω1𝒗t+ω3𝒗h)2.\displaystyle\leq-\omega_{0}\|\boldsymbol{v}_{s}\|^{2}+\frac{1}{4}(\sqrt{\omega_{1}}\|\boldsymbol{v}_{t}\|+\sqrt{\omega_{3}}\|\boldsymbol{v}_{h}\|)^{2}.

From this, we can get that the intention model is input-to-state stable when 𝒗s12ω1ω0𝒗t+12ω3ω0𝒗h.\|\boldsymbol{v}_{s}\|_{\infty}\leq\frac{1}{2}\sqrt{\frac{\omega_{1}}{\omega_{0}}}\|\boldsymbol{v}_{t}\|_{\infty}+\frac{1}{2}\sqrt{\frac{\omega_{3}}{\omega_{0}}}\|\boldsymbol{v}_{h}\|_{\infty}.

VII-B The stability analysis of the consensus network

According to the consensus theory [30], the system dynamics of the consensus network can be represented as

𝒙˙=L~𝒙+Δ\boldsymbol{\dot{x}}=-\widetilde{L}\boldsymbol{x}+\Delta (39)
𝒆˙=D~D~𝒆+D~Δ\boldsymbol{\dot{e}}=-\widetilde{D}^{\top}\widetilde{D}\boldsymbol{e}+\widetilde{D}^{\top}\Delta (40)

where 𝒙=[𝒙1,,𝒙N]2N\boldsymbol{x}=[\boldsymbol{x}_{1}^{\top},\cdots,\boldsymbol{x}_{N}^{\top}]^{\top}\in\mathbb{R}^{2N} is the state set; Δ=[Δ1,,ΔN]\Delta=[\Delta_{1}^{\top},\cdots,\Delta_{N}^{\top}], Δi=𝒗i𝒗if=λi(𝒗^is𝒗if)\Delta_{i}=\boldsymbol{v}_{i}-\boldsymbol{v}_{i}^{f}=\lambda_{i}(\hat{\boldsymbol{v}}_{i}^{s}-\boldsymbol{v}_{i}^{f}); L~=LI2\widetilde{L}=L\otimes I_{2}, LL is the Laplacian matrix; 𝒆=D~𝒙\boldsymbol{e}=\widetilde{D}^{\top}\boldsymbol{x} is the consensus error. And we can easily obtain Lemma 1.

Lemma 1.

Given a connected graph 𝒢(𝒱,)\mathcal{G}(\mathcal{V},\mathcal{E}), D~D~\widetilde{D}^{\top}\widetilde{D} is positive definite on space span{D~}\mathrm{span}\{\widetilde{D}^{\top}\}. Therefore, for 𝐞,δ>0\forall\boldsymbol{e},\exists\delta>0, such that 𝐞D~D~𝐞δ𝐞2\boldsymbol{e}^{\top}\widetilde{D}^{\top}\widetilde{D}\boldsymbol{e}\geq\delta\|\boldsymbol{e}\|^{2}.

For the designed λ\lambda function in (19), we have Lemma 2.

Lemma 2.

For λ(a,b)\lambda(a,b), there exists f¯b(a)𝒦\bar{f}_{b}(a)\in\mathcal{K}, when b>f¯b(a)b>\bar{f}_{b}(a) such that a>0,c0\forall a>0,\forall c\geq 0,

λ(a,b)(b+c)b12b2.\lambda(a,b)(b+c)b\leq\frac{1}{2}b^{2}. (41)
Proof.

Let ga(b)=λ(a,b)(1+cb)g_{a}(b)=\lambda(a,b)(1+\frac{c}{b}), then we can easily obtain that ga(b)g_{a}(b) is strictly decreasing in bb, and limbga(b)=0\lim_{b\rightarrow\infty}g_{a}(b)=0, limb0ga(b)=\lim_{b\rightarrow 0}g_{a}(b)=\infty. Therefore, for a>0\forall a>0, there exists bb^{*} such that ga(b)=12g_{a}(b^{*})=\frac{1}{2}, and for bb\forall b\geq b^{*}, ga(b)12g_{a}(b)\leq\frac{1}{2}. ∎

Theorem 2.

The consensus network with 𝐯^s\hat{\boldsymbol{v}}_{s} as the input, and 𝐞\boldsymbol{e} as the state, is input-to-state stable. Moreover, let 𝐞=lim supt𝐞(t)\|\boldsymbol{e}\|_{\infty}=\limsup\limits_{t\to\infty}\|\boldsymbol{e}(t)\|, γe=12ω1ω0\gamma_{e}=\frac{1}{2}\sqrt{\frac{\omega_{1}}{\omega_{0}}}, we can obtain

𝒆γe𝒗^s.\|\boldsymbol{e}\|_{\infty}\leq\gamma_{e}\|\hat{\boldsymbol{v}}_{s}\|_{\infty}. (42)
Proof.

Let V(𝒆)=12𝒆𝒆V(\boldsymbol{e})=\frac{1}{2}\boldsymbol{e}^{\top}\boldsymbol{e}, from (40), we can obtain

V˙(e)=𝒆D~D~𝒆+𝒆D~Δ=𝒗f2𝒗fΔ,\dot{V}(e)=-\boldsymbol{e}^{\top}\widetilde{D}^{\top}\widetilde{D}\boldsymbol{e}+\boldsymbol{e}^{\top}\widetilde{D}^{\top}\Delta=-\|\boldsymbol{v}_{f}\|^{2}-\boldsymbol{v}_{f}^{\top}\Delta, (43)

where 𝒗f=D~𝒆=[𝒗1f,,𝒗Nf]\boldsymbol{v}_{f}=-\widetilde{D}\boldsymbol{e}=[{\boldsymbol{v}_{1}^{f}}^{\top},\cdots,{\boldsymbol{v}_{N}^{f}}^{\top}]^{\top}.

From (17), we can obtain 𝒗^is𝒞\hat{\boldsymbol{v}}_{i}^{s}\equiv\mathcal{C}. Since Δi=𝒗i𝒗if=λi(𝒗^is𝒗if)\Delta_{i}=\boldsymbol{v}_{i}-\boldsymbol{v}_{i}^{f}=\lambda_{i}(\hat{\boldsymbol{v}}_{i}^{s}-\boldsymbol{v}_{i}^{f}), we obtain that

𝒗fΔ=i𝒱𝒗ifΔii𝒱λi(𝒞+𝒗if)𝒗if.-\boldsymbol{v}_{f}^{\top}\Delta=-\sum_{i\in\mathcal{V}}{\boldsymbol{v}_{i}^{f}}^{\top}\Delta_{i}\leq\sum_{i\in\mathcal{V}}\lambda_{i}(\mathcal{C}+\|\boldsymbol{v}_{i}^{f}\|)\|\boldsymbol{v}_{i}^{f}\|. (44)

From Lemma 2, it follows that i𝒱\forall i\in\mathcal{V},

𝒗ifΔi{𝒗if2/2 if 𝒗iff¯b(𝒗is)(𝒞+f¯b(𝒗is)f¯b(𝒗is) if 𝒗if<f¯b(𝒗is)-{\boldsymbol{v}_{i}^{f}}^{\top}\Delta_{i}\leq\left\{\begin{array}[]{ll}\|\boldsymbol{v}_{i}^{f}\|^{2}/2&\text{ if }\|\boldsymbol{v}_{i}^{f}\|\geq\bar{f}_{b}\left(\|\boldsymbol{v}_{i}^{s}\|\right)\\ \left(\mathcal{C}+\bar{f}_{b}\left(\|\boldsymbol{v}_{i}^{s}\|\right)\bar{f}_{b}\left(\left\|\boldsymbol{v}_{i}^{s}\right\|\right)\right.&\text{ if }\|\boldsymbol{v}_{i}^{f}\|<\bar{f}_{b}\left(\|\boldsymbol{v}_{i}^{s}\|\right)\end{array}\right. (45)

Let f^s=(𝒞+f¯b)f¯b\hat{f}_{s}=\left(\mathcal{C}+\bar{f}_{b}\right)\bar{f}_{b}, then f^s𝒦\hat{f}_{s}\in\mathcal{K}_{\infty}. Then it follows that i𝒱,𝒗ifΔi𝒗if2/2+f^s(𝒗is)\forall i\in\mathcal{V},-{\boldsymbol{v}_{i}^{f}}^{\top}\Delta_{i}\leq\|\boldsymbol{v}_{i}^{f}\|^{2}/2+\hat{f}_{s}\left(\left\|\boldsymbol{v}_{i}^{s}\right\|\right). Since i𝒱f^s(𝒗is)\sum_{i\in\mathcal{V}}\hat{f}_{s}\left(\|\boldsymbol{v}_{i}^{s}\|\right)\leq Nf^s(𝒗s)N\hat{f}_{s}(\|\boldsymbol{v}_{s}\|), from (43) and Lemma 1, we have

V˙(𝒆)\displaystyle\dot{V}(\boldsymbol{e}) =𝒗f2𝒗fΔ\displaystyle=-\|\boldsymbol{v}_{f}\|^{2}-\boldsymbol{v}_{f}^{\top}\Delta (46)
12𝒗f2+f^s(𝒗s)12δ𝒆2+f^s(𝒗s).\displaystyle\leq-\frac{1}{2}\|\boldsymbol{v}_{f}\|^{2}+\hat{f}_{s}(\|\boldsymbol{v}_{s}\|)\leq-\frac{1}{2}\delta\|\boldsymbol{e}\|^{2}+\hat{f}_{s}(\|\boldsymbol{v}_{s}\|).

Thus the system is input-to-state stable, with a bound that 𝒆(2Nf^s(𝒗s)/δ)1/2\|\boldsymbol{e}\|_{\infty}\leq\left(2N\hat{f}_{s}\left(\|\boldsymbol{v}_{s}\|_{\infty}\right)/\delta\right)^{1/2}

Combining Theorem 1 and Theorem 2, we obtain corollary:

Corollary 1.

The hierarchical control system with inputs human intention 𝐯h\boldsymbol{v}_{h} and robot intention 𝐯t\boldsymbol{v}_{t} and state consensus error 𝐞\boldsymbol{e} is input-state stable and satisfies

𝒆(2Nf^s(12ω1ω0𝒗t+12ω3ω0𝒗h)/δ)1/2.\|\boldsymbol{e}\|_{\infty}\leq\left(2N\hat{f}_{s}\left(\frac{1}{2}\sqrt{\frac{\omega_{1}}{\omega_{0}}}\|\boldsymbol{v}_{t}\|_{\infty}+\frac{1}{2}\sqrt{\frac{\omega_{3}}{\omega_{0}}}\|\boldsymbol{v}_{h}\|_{\infty}\right)/\delta\right)^{1/2}. (47)

References

  • [1] S. L. Smith and D. Rus, “Multi-robot monitoring in dynamic environments with guaranteed currency of observations,” in 49th IEEE Conference on Decision and Control (CDC), 2010, pp. 514–521.
  • [2] J. Saez-Pons, L. Alboul, J. Penders, and L. Nomdedeu, “Multi-robot team formation control in the guardians project,” Industrial Robot: An International Journal, vol. 37, no. 4, pp. 372–383, 2010.
  • [3] Y. Liu and G. Nejat, “Robotic urban search and rescue: A survey from the control perspective,” Journal of Intelligent and Robotic Systems, vol. 72, no. 2, pp. 147–165, 2013.
  • [4] R. Fabbiano, F. Garin, and C. Canudas-de Wit, “Distributed source seeking without global position information,” IEEE Transactions on Control of Network Systems, vol. 5, no. 1, pp. 228–238, 2018.
  • [5] W. Dai, H. Lu, J. Xiao, and Z. Zheng, “Task allocation without communication based on incomplete information game theory for multi-robot systems,” Journal of Intelligent and Robotic Systems, vol. 94, no. 3, pp. 841–856, 2019.
  • [6] L. Saleh, P. Chevrel, F. Claveau, J.-F. Lafay, and F. Mars, “Shared steering control between a driver and an automation: Stability in the presence of driver behavior uncertainty,” IEEE Transactions on Intelligent Transportation Systems, vol. 14, no. 2, pp. 974–983, 2013.
  • [7] R. Krzysiak and S. Butail, “Information-based control of robots in search-and-rescue missions with human prior knowledge,” IEEE Transactions on Human-Machine Systems, vol. 52, no. 1, pp. 52–63, 2022.
  • [8] Y. Zhu, C. Yang, Z. Tu, Y. Ling, and Y. Chen, “A haptic shared control architecture for tracking of a moving object,” IEEE Transactions on Industrial Electronics, vol. 70, no. 5, pp. 5034–5043, 2023.
  • [9] D. Sieber, S. Musić, and S. Hirche, “Multi-robot manipulation controlled by a human with haptic feedback,” in 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2015, pp. 2440–2446.
  • [10] N. Ayanian, A. Spielberg, M. Arbesfeld, J. Strauss, and D. Rus, “Controlling a team of robots with a single input,” in 2014 IEEE International Conference on Robotics and Automation (ICRA), 2014, pp. 1755–1762.
  • [11] J. Patel, Y. Xu, and C. Pinciroli, “Mixed-granularity human-swarm interaction,” in 2019 International Conference on Robotics and Automation (ICRA), 2019, pp. 1059–1065.
  • [12] G. Podevijn, R. O’Grady, Y. S. G. Nashed, and M. Dorigo, “Gesturing at subswarms: Towards direct human control of robot swarms,” in Towards Autonomous Robotic Systems, 2014, pp. 390–403.
  • [13] A. Suresh and S. Martínez, “Human-swarm interactions for formation control using interpreters,” International Journal of Control, Automation and Systems, vol. 18, no. 8, pp. 2131–2144, 2020.
  • [14] B. Gromov, L. M. Gambardella, and G. A. Di Caro, “Wearable multi-modal interface for human multi-robot interaction,” in 2016 IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR), 2016, pp. 240–245.
  • [15] J. Y. Chen and M. J. Barnes, “Human–agent teaming for multirobot control: A review of human factors issues,” IEEE Transactions on Human-Machine Systems, vol. 44, no. 1, pp. 13–29, 2014.
  • [16] A. Kolling, P. Walker, N. Chakraborty, K. Sycara, and M. Lewis, “Human interaction with robot swarms: A survey,” IEEE Transactions on Human-Machine Systems, vol. 46, no. 1, pp. 9–26, 2015.
  • [17] Y. A. Kapitanyuk, A. V. Proskurnikov, and M. Cao, “A guiding vector-field algorithm for path-following control of nonholonomic mobile robots,” IEEE Transactions on Control Systems Technology, vol. 26, no. 4, pp. 1372–1385, 2017.
  • [18] W. Yao, B. Lin, B. D. O. Anderson, and M. Cao, “Guiding vector fields for following occluded paths,” IEEE Transactions on Automatic Control, vol. 67, no. 8, pp. 4091–4106, 2022.
  • [19] L. Wang, A. D. Ames, and M. Egerstedt, “Safety barrier certificates for collisions-free multirobot systems,” IEEE Transactions on Robotics, vol. 33, no. 3, pp. 661–674, 2017.
  • [20] J. R. Wolpaw, N. Birbaumer, D. J. McFarland, G. Pfurtscheller, and T. M. Vaughan, “Brain–computer interfaces for communication and control,” Clinical Neurophysiology, vol. 113, no. 6, pp. 767–791, 2002.
  • [21] F. Lotte, L. Bougrain, A. Cichocki, M. Clerc, M. Congedo, A. Rakotomamonjy, and F. Yger, “A review of classification algorithms for eeg-based brain–computer interfaces: a 10 year update,” Journal of Neural Engineering, vol. 15, no. 3, p. 031005, 2018.
  • [22] M. Atzori, A. Gijsberts, C. Castellini, B. Caputo, A.-G. M. Hager, S. Elsig, G. Giatsidis, F. Bassetto, and H. Müller, “Electromyography data for non-invasive naturally-controlled robotic hand prostheses,” Scientific Data, vol. 1, no. 1, p. 140053, 2014.
  • [23] M. T. Wolf, C. Assad, M. T. Vernacchia, J. Fromm, and H. L. Jethani, “Gesture-based robot control with variable autonomy from the jpl biosleeve,” in 2013 IEEE International Conference on Robotics and Automation, 2013, pp. 1160–1165.
  • [24] A. S. Kundu, O. Mazumder, P. K. Lenka, and S. Bhaumik, “Hand gesture recognition based omnidirectional wheelchair control using imu and emg sensors,” Journal of Intelligent & Robotic Systems, vol. 91, no. 3, pp. 529–541, 2018.
  • [25] J. Luo, Z. Lin, Y. Li, and C. Yang, “A teleoperation framework for mobile robots based on shared control,” IEEE Robotics and Automation Letters, vol. 5, no. 2, pp. 377–384, 2020.
  • [26] F. Koochaki and L. Najafizadeh, “Predicting intention through eye gaze patterns,” in 2018 IEEE Biomedical Circuits and Systems Conference (BioCAS), 2018, pp. 1–4.
  • [27] S. B. Shafiei, S. Shadpour, F. Sasangohar, J. L. Mohler, K. Attwood, and Z. Jing, “Development of performance and learning rate evaluation models in robot-assisted surgery using electroencephalography and eye-tracking,” npj Science of Learning, vol. 9, no. 1, p. 3, 2024.
  • [28] R. M. Aronson, T. Santini, T. C. Kübler, E. Kasneci, S. Srinivasa, and H. Admoni, “Eye-hand behavior in human-robot shared manipulation.” ACM/IEEE International Conference on Human-Robot Interaction, pp. 4–13, 2018.
  • [29] H. Chien-Ming and B. Mutlu, “Anticipatory robot control for efficient human-robot collaboration.” ACM/IEEE International Conference on Human-Robot Interaction, pp. 83–90, 2016.
  • [30] F. L. Lewis, H. Zhang, K. Hengster-Movric, and A. Das, Cooperative control of multi-agent systems: optimal and adaptive design approaches.   Springer Science & Business Media, 2013.