This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

UAV-Assisted Multi-Task Federated Learning with Task Knowledge Sharing

Yubo Yang, Tao Yang, Xiaofeng Wu, and Bo Hu School of Information Science and Technology, Fudan University, Shanghai, China
Email: yangyb24@m.fudan.edu.cn, {taoyang, xiaofengwu, bohu}@fudan.edu.cn
Abstract

The rapid development of Unmanned aerial vehicles (UAVs) technology has spawned a wide variety of applications, such as emergency communications, regional surveillance, and disaster relief. Due to their limited battery capacity and processing power, multiple UAVs are often required for complex tasks. In such cases, a control center is crucial for coordinating their activities, which fits well with the federated learning (FL) framework. However, conventional FL approaches often focus on a single task, ignoring the potential of training multiple related tasks simultaneously. In this paper, we propose a UAV-assisted multi-task federated learning scheme, in which data collected by multiple UAVs can be used to train multiple related tasks concurrently. The scheme facilitates the training process by sharing feature extractors across related tasks and introduces a task attention mechanism to balance task performance and encourage knowledge sharing. To provide an analytical description of training performance, the convergence analysis of the proposed scheme is performed. Additionally, the optimal bandwidth allocation for UAVs under limited bandwidth conditions is derived to minimize communication time. Meanwhile, a UAV-EV association strategy based on coalition formation game is proposed. Simulation results validate the effectiveness of the proposed scheme in enhancing multi-task performance and training speed.

I Introduction

Unmanned Aerial Vehicle (UAV)-assisted networks are seen as an important enabler of post-5G and upcoming 6G networks. Due to their flexibility and line-of-sight (LOS) communication links, UAVs have great potential in providing emergency communications and wide-area on-demand data collection[1]. In addition, UAVs can be used to train popular machine learning models on-board, which is crucial for tasks such as trajectory planning and object recognition[2]. In a multi-UAV system, UAVs operate in a coordinated manner to support reliable and efficient communications or complete specific tasks[3][4]. UAV coordination is typically assisted by a ground base station. For instance, in an emergency rescue scenario with poor network coverage, a swarm of UAVs is controlled by a ground emergency vehicle (EV). However, since UAVs are deployed in various areas and often perform data collection dynamically[5], given the limited communication resources and energy capacity of UAVs, it is unrealistic to collect and transmit raw data from each UAV to the EV, as well as to frequently shuttle between the incident scene and the EV for charging. Given these challenges, federated learning (FL) has emerged as a promising framework for distributed data processing and model training. Taking advantage of the capabilities of the UAV data platform, it enables collaborative learning among multiple UAVs, thereby enhancing model training efficiency and accuracy[6].

Currently, most existing studies focus on single-task learning, while some multi-task FL either treats each individual model as a separate task [7][8] or learns multiple unrelated tasks concurrently [9][10]. This differs from the scenario considered in this paper. Actually, the data collected by UAVs may be used for multiple related tasks. For example, in emergency rescue scenarios, UAV images can support tasks like disaster level identification, crowd density estimation, and road feasibility analysis. Therefore, the multi-task scheme considered in this paper refers to using data collected by UAVs to complete multiple related but different tasks simultaneously. Using the traditional single-task learning framework to learn each task independently is inefficient[11]. Since multiple tasks are trained on the same image dataset, they share common underlying features and may have potential correlations, which allows for effective knowledge sharing among tasks and enables a reduction in resource consumption for training.

Studies have shown that sharing feature extractors across correlated tasks can enhance both the training speed and robustness of models[12]. However, these findings are limited to centralized training scenarios. In our scenario, each UAV is deployed in a different area or assigned a distinct but related task, leading to variations in local data collection. Integrating data features collected by multiple UAVs assigned to different tasks significantly enhances the robustness and generalization of the feature extractors. This approach allows the extractor to learn from a broader range of areas, thus improving feature extraction performance, accelerating model convergence, and boosting overall performance through mutual reinforcement among tasks[13]. In summary, the contributions of this paper are as follows:

  • We propose a novel UAV-assisted multi-task federated learning scheme, where the data collected by the UAVs can be used to train multiple related tasks, and the system performance and robustness can be improved by sharing knowledge among related tasks.

  • To balance the task performance and enhance the overall effectiveness of multiple tasks, we propose a task attention mechanism to capture the dynamic importance of each task. Specifically, we introduce the task shapley value to facilitate knowledge sharing among tasks.

  • Considering the dynamic environment of the network and resource constraints, we derived an expression for optimal bandwidth allocation for UAVs and introduced a UAV-EV association mechanism based on coalition formation game. Simulation results demonstrate the effectiveness of the proposed scheme.

II System Model

Refer to caption
Figure 1: UAV assisted multi-task federated learning system.

II-A Federated Learning System

The system consists of MM EVs indexed by ={1,2,,M}\mathcal{M}=\{1,2,\cdots,M\} and NN UAVs indexed by 𝒩={1,2,,N}\mathcal{N}=\{1,2,\cdots,N\}. Each EV is assigned a different task and can interact with other EVs to share task knowledge. Additionally, each EV can collect a small amount of data for model validation. Each UAV nn has a local dataset 𝒟n\mathcal{D}_{n} with Dn=|𝒟n|D_{n}=\left|\mathcal{D}_{n}\right| data samples, and the data can be used for MM tasks. Since the same image dataset is used for multiple tasks, the underlying image features required by multiple tasks may be correlated. To enhance feature extraction performance, we allow multiple tasks to share feature extractors while maintaining task-specific predictors. The entire dataset is denoted by 𝒟={𝒟n}n=1N\mathcal{D}=\cup\left\{\mathcal{D}_{n}\right\}_{n=1}^{N} with a total number of samples D=n=1NDnD=\sum_{n=1}^{N}D_{n}. Given a data sample (𝒙,𝒚)𝒟(\boldsymbol{x},\boldsymbol{y})\in\mathcal{D}, where 𝒙d\boldsymbol{x}\in\mathbb{R}^{d} is the dd-dimensional input data vector, 𝒚=[y1,y2,,yM]M\boldsymbol{y}=[y_{1},y_{2},\cdots,y_{M}]\in\mathbb{R}^{M} is the corresponding MM-dimensional ground-truth label and each dimension corresponds to a specific task. Let fm,n(𝒙,𝒚;𝒘m,n)f_{m,n}(\boldsymbol{x},\boldsymbol{y};\boldsymbol{w}_{m,n}) denote the sample-wise loss function for task mm on UAV nn, 𝒘m,n\boldsymbol{w}_{m,n} can be split into two parts [𝒘m,ns,𝒘m,nu][\boldsymbol{w}_{m,n}^{s},\boldsymbol{w}_{m,n}^{u}], where 𝒘m,ns\boldsymbol{w}_{m,n}^{s} represents the shared feature extractor and 𝒘m,nu\boldsymbol{w}_{m,n}^{u} represents the unique layer of task mm. Thus, the local loss function that UAV nn measures the model error for task mm on its local dataset is given by

Fm,n(𝒘m,n)=1Dn(𝒙,𝒚)𝒟nfm,n(𝒙,𝒚;𝒘m,n){F_{m,n}(\boldsymbol{w}_{m,n})=\frac{1}{D_{n}}\sum_{(\boldsymbol{x},\boldsymbol{y})\in\mathcal{D}_{n}}f_{m,n}(\boldsymbol{x},\boldsymbol{y};\boldsymbol{w}_{m,n})} (1)

Accordingly, the global loss function for task mm is

Fm(𝒘m)=n=1NDnDFm,n(𝒘m)F_{m}(\boldsymbol{w}_{m})=\sum\nolimits_{n=1}^{N}\frac{D_{n}}{D}F_{m,n}(\boldsymbol{w}_{m}) (2)

We use 𝒘=[𝒘1,𝒘2,𝒘M]\boldsymbol{w}=[\boldsymbol{w}_{1},\boldsymbol{w}_{2},\cdots\boldsymbol{w}_{M}] to represent the set of all task parameters. For simplicity, we omit the subscript mm from ff and use fn(𝒙,𝒚;𝒘m,n)f_{n}(\boldsymbol{x},\boldsymbol{y};\boldsymbol{w}_{m,n}) to denote the loss function of task mm on UAV nn. Each UAV can participate in different tasks in different rounds. However, due to the limited computing resources and energy, each UAV can only participate in one training task per round. Considering the goal to improve the average performance of all tasks, the objective function of the entire network is

F(𝒘)=1Mm=1MFm(𝒘m)F(\boldsymbol{w})=\frac{1}{M}\sum\nolimits_{m=1}^{M}F_{m}(\boldsymbol{w}_{m}) (3)

To achieve the objective outlined in (3), the training steps of the proposed scheme are given as follows:

\bullet (Step 1 UAV-EV association and Task Model Broadcast) At the beginning of each round, each UAV is assigned to an EV through the UAV-EV association algorithm. Each EV broadcasts its latest model parameters 𝒘m,t\boldsymbol{w}_{m,t} to its associated UAVs. Let βm,n,t{0,1}\beta_{m,n,t}\in\{0,1\} denote the association indicator of UAV nn in rount tt, where βm,n,t=1\beta_{m,n,t}=1 indicates that UAV nn is associated with EV mm, βm,n,t=0\beta_{m,n,t}=0 otherwise. We use 𝜷𝒎,𝒕=[βm,1,t,βm,2,t,,βm,N,t]TN,m\boldsymbol{\beta_{m,t}}=[\beta_{m,1,t},\beta_{m,2,t},\cdots,\beta_{m,N,t}]^{T}\in\mathbb{R}^{N},\forall m\in\mathcal{M} to represent the UAV association of task mm in round tt. And 𝜷t=[𝜷𝟏,𝒕;𝜷𝟐,𝒕;;𝜷𝑴,𝒕]T\boldsymbol{\beta}_{t}=[\boldsymbol{\beta_{1,t}};\boldsymbol{\beta_{2,t}};\cdots;\boldsymbol{\beta_{M,t}}]^{T} can be used to represent the overall UAV-EV association in round tt.

\bullet (Step 2 Local Model Training) Each UAV nn initializes 𝒘m,n,t0=𝒘m,t\boldsymbol{w}_{m,n,t}^{0}=\boldsymbol{w}_{m,t} and performs KK steps stochastic gradient descent (SGD) to update the shared layer and the task-specific layer. 𝒘m,n,ts\boldsymbol{w}_{m,n,t}^{s} and 𝒘m,n,tu\boldsymbol{w}_{m,n,t}^{u} are updated as follows:

𝒘m,n,ts,k+1=𝒘m,n,ts,kηsf~n(𝒘m,n,ts,k,𝒘m,n,tu,k)\boldsymbol{w}_{m,n,t}^{s,k+1}=\boldsymbol{w}_{m,n,t}^{s,k}-\eta\nabla_{s}\widetilde{f}_{n}(\boldsymbol{w}_{m,n,t}^{s,k},\boldsymbol{w}_{m,n,t}^{u,k}) (4)
𝒘m,n,tu,k+1=𝒘m,n,tu,kηuf~n(𝒘m,n,ts,k,𝒘m,n,tu,k)\boldsymbol{w}_{m,n,t}^{u,k+1}=\boldsymbol{w}_{m,n,t}^{u,k}-\eta\nabla_{u}\widetilde{f}_{n}(\boldsymbol{w}_{m,n,t}^{s,k},\boldsymbol{w}_{m,n,t}^{u,k}) (5)

where 𝒘m,n,ts,k\boldsymbol{w}_{m,n,t}^{s,k} and 𝒘m,n,tu,k\boldsymbol{w}_{m,n,t}^{u,k} are the feature extractor layer and task-specific layer parameters of task mm on UAV nn in the kk-th local iteration of the tt-th round, respectively. η>0\eta>0 is the learning rate.

In (4), the stochastic gradient sf~n(𝒘m,n,ts,k,𝒘m,n,tu,k)\nabla_{s}\widetilde{f}_{n}(\boldsymbol{w}_{m,n,t}^{s,k},\boldsymbol{w}_{m,n,t}^{u,k}) is given by sf~n(𝒘m,n,ts,k,𝒘m,n,tu,k)=1Lb(𝒙,𝒚)ζn,thsfn(𝒙,𝒚;𝒘m,n,tk)\nabla_{s}\widetilde{f}_{n}(\boldsymbol{w}_{m,n,t}^{s,k},\boldsymbol{w}_{m,n,t}^{u,k})=\frac{1}{L_{b}}\sum_{(\boldsymbol{x},\boldsymbol{y})\in\zeta_{n,t}^{h}}\nabla_{s}f_{n}\left(\boldsymbol{x},\boldsymbol{y};\boldsymbol{w}_{m,n,t}^{k}\right), where ζn,tk\zeta_{n,t}^{k} is a mini-batch data uniformly sampled from DnD_{n} with Lb=|ζn,tk|L_{b}=\left|\zeta_{n,t}^{k}\right| data samples. The update form of uf~n(𝒘m,n,ts,k,𝒘m,n,tu,k)\nabla_{u}\widetilde{f}_{n}(\boldsymbol{w}_{m,n,t}^{s,k},\boldsymbol{w}_{m,n,t}^{u,k}) is similar and is omitted here.

\bullet (Step 3 Model Update &\& Parameter Uploading ) After local model training, each UAV nn uploads its cumulative local stochastic gradient 𝑮~m,n,t\tilde{\boldsymbol{G}}_{m,n,t} to its corresponding EV. 𝑮~m,n,t\tilde{\boldsymbol{G}}_{m,n,t} is given by 𝑮~m,n,t=k=0K1f~n(𝒘m,n,ts,k,𝒘m,n,tu,k)=1η(𝒘m,n,t0𝒘m,n,tK)\tilde{\boldsymbol{G}}_{m,n,t}=\sum_{k=0}^{K-1}\nabla\widetilde{f}_{n}(\boldsymbol{w}_{m,n,t}^{s,k},\boldsymbol{w}_{m,n,t}^{u,k})=\frac{1}{\eta}\left(\boldsymbol{w}_{m,n,t}^{0}-\boldsymbol{w}_{m,n,t}^{K}\right) Note that the cumulative gradient can be expressed as 𝑮~m,n,t=[𝑮~m,n,ts,𝑮~m,n,tu]\tilde{\boldsymbol{G}}_{m,n,t}=[\tilde{\boldsymbol{G}}_{m,n,t}^{s},\tilde{\boldsymbol{G}}_{m,n,t}^{u}], where 𝑮~m,n,ts\tilde{\boldsymbol{G}}_{m,n,t}^{s} and 𝑮~m,n,tu\tilde{\boldsymbol{G}}_{m,n,t}^{u} represent the feature extractor gradient and task-specific layer gradient respectively.

\bullet (Step 4 EV Model Aggregation and Knowledge Sharing) Each EV aggregates the latest gradient and updates the global model of task mm. Specifically, the feature extractor is updated as 𝒘^m,t+1s=𝒘m,tsηn=1Nβm,n,tDn𝑮~m,n,tsn=1Nβm,n,tDn\hat{\boldsymbol{w}}_{m,t+1}^{s}=\boldsymbol{w}_{m,t}^{s}-\eta\frac{{\textstyle\sum\nolimits_{n=1}^{N}}\beta_{m,n,t}D_{n}\tilde{\boldsymbol{G}}_{m,n,t}^{s}}{\sum\nolimits_{n=1}^{N}\beta_{m,n,t}D_{n}} and the task predictor is updated as 𝒘m,t+1u=𝒘m,tuηn=1Nβm,n,tDn𝑮~m,n,tun=1Nβm,n,tDn\boldsymbol{w}_{m,t+1}^{u}=\boldsymbol{w}_{m,t}^{u}-\eta\frac{{\textstyle\sum\nolimits_{n=1}^{N}}\beta_{m,n,t}D_{n}\tilde{\boldsymbol{G}}_{m,n,t}^{u}}{\sum\nolimits_{n=1}^{N}\beta_{m,n,t}D_{n}}. Subsequently, the EVs interact to share knowledge by transmitting 𝒘^m,t+1s\hat{\boldsymbol{w}}_{m,t+1}^{s} to one another for further fusion of the feature extractor parameters. Given that EVs possess relatively sufficient transmission resources and computing capabilities, we disregard any delays and energy consumption associated with this parameter exchange. The fused feature extractor can then be expressed as: 𝒘t+1s=m=1MDm,t𝒘^m,t+1sm=1MDm,t\boldsymbol{w}_{t+1}^{s}=\frac{{\textstyle\sum\nolimits_{m=1}^{M}}D_{m,t}\hat{\boldsymbol{w}}_{m,t+1}^{s}}{\sum\nolimits_{m=1}^{M}D_{m,t}}, where Dm,t=n=1Nβm,n,tDnD_{m,t}=\sum\nolimits_{n=1}^{N}\beta_{m,n,t}D_{n}.

II-B Computation Model

Let Cm,nC_{m,n} denotes the number of CPU cycles required to process one data sample when UAV nn participates in the training of task mm, fnf_{n} represents the computation capability of UAV nn. Thus, if UAV nn participate in the training of task mm, the computational time of local training of is given by Tm,n,tcomp=KLbCm,nfnT_{m,n,t}^{comp}=\frac{KL_{b}C_{m,n}}{f_{n}} and the corresponding energy consumption of UAV nn is Em,n,tcomp=ςKLbCm,n(fn)2E_{m,n,t}^{comp}=\varsigma KL_{b}C_{m,n}\left(f_{n}\right)^{2}, where ς\varsigma is the energy coefficient, which depends on the chip architecture.

II-C Communication Model

The air-to-ground communication link is modeled as a probabilistic superposition of LOS and NLOS channels based on the method in [14]. For the sake of brevity, the details are omitted here. We consider using the frequency division multiple access (FDMA) techniques for communication. Each UAV can be allocated a certain proportion γn,t\gamma_{n,t} of uplink bandwidth resources. Denote pn,tp_{n,t} as the transmit power of UAV nn in round tt, and it has a maximum transmission power limit pn,maxp_{n,max}. Thus, the achievable transmission rate between UAV nn and EV mm in round tt is rm,n,t=γn,tBlog2(1+pn,thm,n,tγn,tBN0)r_{m,n,t}=\gamma_{n,t}B\log_{2}\left(1+\frac{p_{n,t}h_{m,n,t}}{\gamma_{n,t}BN_{0}}\right). N0N_{0} is the noise power spectral density. Let QmQ_{m} denote the size of the gradient for task mm, if UAV nn is scheduled to participate in task mm in round tt, its transmission time is given by Tm,n,tcomm=Qmrm,n,tT_{m,n,t}^{comm}=\frac{Q_{m}}{r_{m,n,t}} The corresponding energy consumption of UAV nn for transmission is En,tcomm=pn,tTm,n,tcommE_{n,t}^{comm}=p_{n,t}T_{m,n,t}^{comm}.

III PROBLEM FORMULATION AND CONVERGENCE ANALYSIS

III-A Problem Formulation

In order to minimize the expected average global loss of all tasks and reduce the total training time subject to the constrained resources of UAVs, we need to jointly optimize UAV-EV association, bandwidth allocation, and UAV transmission power. The problem is formulated as follows:

P1:\displaystyle P_{1}: min{𝜷t,𝜸t,𝒑t}t=0T1λ𝔼[F(𝒘T)]+(1λ)Ttotal\displaystyle\min_{\left\{\boldsymbol{\beta}_{t},\boldsymbol{\gamma}_{t},\boldsymbol{p}_{t}\right\}_{t=0}^{T-1}}\lambda\mathbb{E}[F(\boldsymbol{w}_{T})]+(1-\lambda)T_{\text{total}} (6)
s.t. En,tcomp+En,tcommEn,max,n𝒩,t,\displaystyle E_{n,t}^{comp}+E_{n,t}^{comm}\leq E_{n,\text{max}},\quad\forall n\in\mathcal{N},\forall t, (6a)
βm,n,t{0,1},m,n𝒩,t,\displaystyle\beta_{m,n,t}\in\{0,1\},\quad\forall m\in\mathcal{M},\forall n\in\mathcal{N},\forall t, (6b)
m=1Mβm,n,t=1,n𝒩,t,\displaystyle\sum\nolimits_{m=1}^{M}\beta_{m,n,t}=1,\quad\forall n\in\mathcal{N},\forall t, (6c)
n=1Nβm,n,tδm,m,t,\displaystyle\sum\nolimits_{n=1}^{N}\beta_{m,n,t}\geq\delta_{m},\quad\forall m\in\mathcal{M},\forall t, (6d)
0γn,t1,n𝒩,t,\displaystyle 0\leq\gamma_{n,t}\leq 1,\quad\forall n\in\mathcal{N},\forall t, (6e)
n=1Nγn,t1,t,\displaystyle\sum\nolimits_{n=1}^{N}\gamma_{n,t}\leq 1,\quad\forall t, (6f)
0pn,tpn,max,n𝒩,t.\displaystyle 0\leq p_{n,t}\leq p_{n,\text{max}},\quad\forall n\in\mathcal{N},\forall t. (6g)

The constraint in (6a) ensures that the energy consumption of each UAV in each round cannot exceed the maximum energy limit En,maxE_{n,\text{max}} per round. Constraints (6b)-(6d) imposes restrictions on UAV-EV associations. Constraints (6e) and (6f) regulate the bandwidth allocation, and (6g) limits the transmission power of each UAV.

Since the objective function of problem P1P_{1} involves the neural network’s loss function F(𝒘t)F(\boldsymbol{w}_{t}), the factors affecting the network’s performance are still unknown. Consequently, problem P1P_{1} is not yet solvable. To address this, we first conduct the convergence analysis of the proposed scheme to identify the factors affecting task performance, and then formulate the problem into a solvable suboptimal form.

III-B Convergence Analysis

Assumption 1. For each UAV nn participating in each task mm, the stochastic gradient is unbiased and variance bounded

𝔼[sf~n(𝒘m,n,tk)]=sfn(𝒘m,n,tk)\mathbb{E}\left[\nabla_{s}\widetilde{f}_{n}(\boldsymbol{w}_{m,n,t}^{k})\right]=\nabla_{s}f_{n}(\boldsymbol{w}_{m,n,t}^{k}) (7)
𝔼[sf~n(𝒘m,n,tk)sfn(𝒘m,n,tk)2]σs2\mathbb{E}\left[\left\|\nabla_{s}\widetilde{f}_{n}(\boldsymbol{w}_{m,n,t}^{k})-\nabla_{s}f_{n}(\boldsymbol{w}_{m,n,t}^{k})\right\|^{2}\right]\leq\sigma_{s}^{2} (8)

The task-specific layer properties are similar, which are omitted here.

Assumption 2. The expected squared norm of local gradients for each task mm and UAV nn is uniformly bounded

𝔼[sfn(𝒘m,n,tk)2]ϵs2, 𝔼[ufn(𝒘m,n,tk)2]ϵu2.\mathbb{E}\left[\left\|\nabla_{s}f_{n}(\boldsymbol{w}_{m,n,t}^{k})\right\|^{2}\right]\leq\epsilon_{s}^{2},\text{ }\mathbb{E}\left[\left\|\nabla_{u}f_{n}(\boldsymbol{w}_{m,n,t}^{k})\right\|^{2}\right]\leq\epsilon_{u}^{2}. (9)

Assumption 3. For each task mm and UAV nn, there exist Ls,Lu,LsuL_{s},L_{u},L_{su} and LusL_{us} such that:

\bullet sfn(𝒘m,ts,𝒘m,tu)\text{ }\nabla_{s}f_{n}(\boldsymbol{w}_{m,t}^{s},\boldsymbol{w}_{m,t}^{u}) is LsL_{s}-Lipschitz continuous with 𝒘m,ts\boldsymbol{w}_{m,t}^{s} and LsuL_{su}-Lipschitz continuous with 𝒘m,tu\boldsymbol{w}_{m,t}^{u}:

sfn(𝒘¯s,𝒘m,tu)sfn(𝒘ˇs,𝒘m,tu)Ls𝒘¯s𝒘ˇs||\nabla_{s}f_{n}(\bar{\boldsymbol{w}}^{s},\boldsymbol{w}_{m,t}^{u})-\nabla_{s}f_{n}(\check{\boldsymbol{w}}^{s},\boldsymbol{w}_{m,t}^{u})||\leq L_{s}||\bar{\boldsymbol{w}}^{s}-\check{\boldsymbol{w}}^{s}|| (10)
sfn(𝒘m,ts,𝒘¯u)sfn(𝒘m,ts,𝒘ˇu)Lsu𝒘¯u𝒘ˇu||\nabla_{s}f_{n}(\boldsymbol{w}_{m,t}^{s},\bar{\boldsymbol{w}}^{u})-\nabla_{s}f_{n}(\boldsymbol{w}_{m,t}^{s},\check{\boldsymbol{w}}^{u})||\leq L_{su}||\bar{\boldsymbol{w}}^{u}-\check{\boldsymbol{w}}^{u}|| (11)

\bullet ufn(𝒘m,ts,𝒘m,tu)\text{ }\nabla_{u}f_{n}(\boldsymbol{w}_{m,t}^{s},\boldsymbol{w}_{m,t}^{u}) is LuL_{u}-Lipschitz continuous with 𝒘m,tu\boldsymbol{w}_{m,t}^{u} and LusL_{us}-Lipschitz continuous with 𝒘m,ts\boldsymbol{w}_{m,t}^{s}, the detail is omitted here.

Based on the above assumptions, we give the upper bound of the difference in loss function between two consecutive rounds of task mm:

Theorem 1.

Given the UAV-EV association βt\beta_{t} in round tt, when the learning rate η<2K2(Lu+2Lsu)\eta<\frac{2}{K^{2}(L_{u}+2L_{su})}, the difference in the loss function for task mm between two consecutive rounds is bounded by:

Fm(𝒘t+1s,𝒘m,t+1u)Fm(𝒘ts,𝒘m,tu)\displaystyle F_{m}(\boldsymbol{w}_{t+1}^{s},\boldsymbol{w}_{m,t+1}^{u})-F_{m}(\boldsymbol{w}_{t}^{s},\boldsymbol{w}_{m,t}^{u}) (12)
ηK2[𝔼sFm(𝒘m,t)2+𝔼uFm(𝒘m,t)2]\displaystyle\leq-\frac{\eta K}{2}\left[\mathbb{E}\left\|\nabla_{s}F_{m}(\boldsymbol{w}_{m,t})\right\|^{2}+\mathbb{E}\left\|\nabla_{u}F_{m}(\boldsymbol{w}_{m,t})\right\|^{2}\right]
+(Dn=1Nβm,n,tDn)2Ω1D2\displaystyle+\frac{\left(D-\sum\nolimits_{n=1}^{N}\beta_{m,n,t}D_{n}\right)^{2}\Omega_{1}}{D^{2}}
+(Dn=1Nβm,n,tDn)Ω2D+Ω3\displaystyle+\frac{\left(D-\sum\nolimits_{n=1}^{N}\beta_{m,n,t}D_{n}\right)\Omega_{2}}{D}+\Omega_{3}

where Ω1=(4ηK+16K2Lsu)ϵs2+4ηKϵu2\Omega_{1}=(4\eta K+16K^{2}L_{su})\epsilon_{s}^{2}+4\eta K\epsilon_{u}^{2}, Ω2=8K2σsLsuϵs+4Kϵs2\Omega_{2}=8K^{2}\sigma_{s}L_{su}\epsilon_{s}+4K\epsilon_{s}^{2}, Ω3=Kσs(ϵs+2KσsLsu)+ηK2Ω4\Omega_{3}=K\sigma_{s}(\epsilon_{s}+2K\sigma_{s}L_{su})+\frac{\eta K}{2}\Omega_{4} and Ω4=((K+1)(Luϵu+Lusϵs)2+σu)2+((K+1)(Lsϵs+Lsuϵu)2+σs)2\Omega_{4}=\left(\frac{\left(K+1\right)\left(L_{u}\epsilon_{u}+L_{us}\epsilon_{s}\right)}{2}+\sigma_{u}\right)^{2}+\left(\frac{\left(K+1\right)\left(L_{s}\epsilon_{s}+L_{su}\epsilon_{u}\right)}{2}+\sigma_{s}\right)^{2}

The proof is omitted here due to space limitations.

According to Theorem 1, it can be seen that task learning performance is mainly affected by the amount of associated data. Due to the dynamic UAV-EV association and the varying complexity of tasks, the training progress differs from task to task. Given the diminishing marginal effect observed in neural network training, improving overall task performance calls for greater emphasis on tasks with slower training progress. Besides, due to the differences in the quality of UAV data collected in different regions, the feature extractors of EVs associated with these high-quality data UAVs usually have better feature extraction capabilities. Through feature extractor fusion, these EVs usually bring higher performance improvements to other EVs. Therefore, EVs with a strong performance history should receive greater focus. Based on the above analysis, each task should have a dynamic task weight in different rounds. Therefore, to address problem P1, we reformulate it into the following form.

P2:\displaystyle P_{2}: max𝜷t,𝜸t,𝒑tλm=1Mαmn=1Nβm,n,tρn(1λ)Tt\displaystyle\max_{\boldsymbol{\beta}_{t},\boldsymbol{\gamma}_{t},\boldsymbol{p}_{t}}\lambda\sum_{m=1}^{M}\alpha_{m}\sum_{n=1}^{N}\beta_{m,n,t}\rho_{n}-(1-\lambda)T_{t} (13)
s.t. (6a6g)\displaystyle(6a-6g) (13a)

where ρn=DnD\rho_{n}=\frac{D_{n}}{D} represents the sample size ratio of UAV nn and this problem will be addressed in section IV.

IV Proposed Algorithm

IV-A Task Attention Mechanism

In this section, we propose a task-level attention mechanism to dynamically characterize the weights of each task in each round.

IV-A1 Task Performance Balance

In a multi-task learning system, achieving a balanced performance across tasks is key to improving the system’s overall performance. Excessive performance differences among tasks may lead to longer delays and unnecessary waste of resources. To this end, we propose a dynamic weight adjustment mechanism based on historical loss, Specifically, each EV mm maintains the parameter Γm,t\Gamma_{m,t} to record the current weighted cumulative loss value of its own task and Γm,t=ϖΓm,t1+(1ϖ)lossm,t1\Gamma_{m,t}=\varpi\Gamma_{m,t-1}+(1-\varpi)loss_{m,t-1}, where lossm,t1loss_{m,t-1} represents the loss of task mm when the t1t-1 round of training is completed, and ϖ\varpi is the parameter that balances the importance of historical cumulative loss and current loss.

Then, the loss based weight Ψ~m,t\tilde{\Psi}_{m,t} of each task mm can be obtained as Ψ~m,t=eΓm,tm=1MeΓm,t\tilde{\Psi}_{m,t}=\frac{e^{\Gamma_{m,t}}}{\sum_{m=1}^{M}e^{\Gamma_{m,t}}}.

IV-A2 Task Shapley Value

In this part, the concept of the Shapley value is used to quantify the marginal contribution of task mm in enhancing the performance of all tasks. We use CC to represent the EV subset, and utility function vm(C)v_{m}(C) represents the accuracy of task mm after updating the feature extractor with the gradient of EV contained in CC. Then, the shapley value of task mm in round tt can be calculated as

ϕm,t=1M2i=1Mj=0M1|C|=j,mC1(M1j)(vi(C{m})vi(C))\phi_{m,t}=\frac{1}{M^{2}}\sum_{\begin{subarray}{c}i=1\end{subarray}}^{M}\sum_{j=0}^{M-1}\sum_{\begin{subarray}{c}|C|=j,\\ m\notin C\end{subarray}}\frac{1}{\binom{M-1}{j}}(v_{i}(C\cup\{m\})-v_{i}(C)) (14)

Then, the cumulative Shapley value of each task mm is updated to m,t=κm,t1+(1κ)ϕm,t\mathcal{I}_{m,t}=\kappa\mathcal{I}_{m,t-1}+(1-\kappa)\phi_{m,t}, κ\kappa is a balancing factor similar to ϖ\varpi. To promote knowledge sharing among tasks and accelerate the training process of each task, tasks with high shapley values should receive greater focus, so the task shapley value based weight Ψ^m,t\hat{\Psi}_{m,t} can be calculated as Ψ^m,t=em,tm=1Mem,t\hat{\Psi}_{m,t}=\frac{e^{\mathcal{I}_{m,t}}}{\sum_{m=1}^{M}e^{\mathcal{I}_{m,t}}}.

Finally, we can get the weight of each task in round tt based on the task attention mechanism as Ψm,t=Ψ~m,t+Ψ^m,tm=1M(Ψ~m,t+Ψ^m,t)\Psi_{m,t}=\frac{\tilde{\Psi}_{m,t}+\hat{\Psi}_{m,t}}{\sum_{m=1}^{M}(\tilde{\Psi}_{m,t}+\hat{\Psi}_{m,t})}.

IV-B Optimal Bandwidth Allocation

Given the UAV-EV association 𝜷𝒕\boldsymbol{\beta_{t}}, problem P2P_{2} can be simplified as

minγt,ptTt\displaystyle\min_{\gamma_{t},p_{t}}T_{t} (15)
s.t. (6a),(6e6g)\displaystyle(6a),(6e-6g) (15a)

We first determine the transmission power pn,tp_{n,t} for each UAV. From the previous communication model, we know that rn,tpn,t>0\frac{\partial r_{n,t}}{\partial p_{n,t}}>0. To reduce transmission time, each UAV should operate at its highest possible transmission rate within the limits of its energy constraints. Therefore, the transmission power for each UAV is set to pn,t=pn,maxp_{n,t}=p_{n,\text{max}}. Consequently, problem (14) can be rewritten as:

minγtTtmax\displaystyle\min_{\gamma_{t}}T_{t}^{max} (16)
s.t. Tn,tcomp+Tn,tcommTtmax,n𝒩,t\displaystyle T_{n,t}^{comp}+T_{n,t}^{comm}\leq T_{t}^{max},\forall n\in\mathcal{N},\forall t (16a)
(6e6f)\displaystyle(6e-6f) (16b)

where Ttmax=maxn{Tn,tcomp+Tn,tcomm}T_{t}^{\text{max}}=\max_{n}\{T_{n,t}^{\text{comp}}+T_{n,t}^{\text{comm}}\} and the optimal solution of (16) is presented in Theorem 2 as follows.

Theorem 2.

Given the UAV-EV association βt\beta_{t}, the optimal bandwidth allocation ratio for each UAV is

γn,t=dn,tln2(TtTn,tcomp)(𝒲(χn,teχn,t)+χn,t)\gamma_{n,t}=\frac{d_{n,t}\ln 2}{\left(T_{t}^{*}-T_{n,t}^{\mathrm{comp}}\right)\left(\mathcal{W}\left(-\chi_{n,t}e^{-\chi_{n,t}}\right)+\chi_{n,t}\right)} (17)

where χn,tdn,tN0ln2(TtTn,tcomp)Pn,thn,t2\chi_{n,t}\triangleq\frac{d_{n,t}N_{0}\mathrm{ln}2}{\left(T_{t}^{*}-T_{n,t}^{\mathrm{comp}}\right)P_{n,t}h_{n,t}^{2}}, 𝒲()\mathcal{W(\cdot)} is the Lambert-W function and TtT_{t}^{*} is the optimal solution to problem (16)(16) and satisfies

n=1Ndn,tln2(TtTn,tcomp)(𝒲(χn,teχn,t)+χn,t)=1\sum_{n=1}^{N}\frac{d_{n,t}\ln 2}{\left(T_{t}^{*}-T_{n,t}^{\mathrm{comp}}\right)\left(\mathcal{W}\left(-\chi_{n,t}e^{-\chi_{n,t}}\right)+\chi_{n,t}\right)}=1 (18)

The proof is omitted here due to space limitations.

Since the optimal bandwidth allocation requires solving the Lambert-WW function, to obtain a closed-form expression just through parameter iteration computation explicitly is almost impossible. Therefore, we adopt a bisection method to give the solution numerically.

IV-C Optimal UAV-EV Association

Once the association between the UAV and the EV is determined, the optimal bandwidth allocation ratio and the minimum round time TtT_{t}^{*} are determined. Substituting the minimum time TtT_{t}^{*} obtained in the previous section into equation (16), we can transform the problem as

maxβtλm=1Mαmtn=1Nβm,ntρn(1λ)T\displaystyle\max_{\beta_{t}}\lambda\sum_{m=1}^{M}\alpha_{m}^{t}\sum_{n=1}^{N}\beta_{m,n}^{t}\rho_{n}-(1-\lambda)T^{*} (19)
s.t. (6b6d)\displaystyle(6b-6d) (19a)

Since the objective function does not have a clear analytical expression at this time and the constraints contain binary constraint variables, the problem is a non-convex optimization problem. In order to optimize the association between UAVs and EVs in an efficient and feasible way, we use the coalition formation game to obtain a suboptimal solution to the problem.

First we introduce some essential concepts and definitions in the following part.

Definition 1.

All NN UAVs can be divided into MM clusters, expressed as 𝒮={Sii=1,2,M}\mathcal{S}=\left\{S_{i}\mid i=1,2,\cdots M\right\} and thus constitute a UAV-EV association strategy. We have SiSj=ϕ,i,j=1,2,MS_{i}\cap S_{j}=\phi,\forall i,j=1,2,\cdots M and i=1,MSi=𝒩\bigcup_{i=1,\cdots M}S_{i}=\mathcal{N}.

Definition 2.

Given the current UAV-EV association 𝒮\mathcal{S}, the utility is given by (13) and denoted as U(𝒮)U(\mathcal{S}).

Definition 3.

Given two different UAV-EV association strategies S1S_{1} and S2S_{2}, we define a preference order as S1S2S_{1}\succ S_{2} if and only if U(S1)>U(S2)U(S_{1})>U(S_{2}).

Definition 4.

A UAV transferring adjustment by nn means that UAV nSmn\in S_{m} with |Sm|>δm\left|S_{m}\right|>\delta_{m} retreats its current training group SmS_{m} and joins another training group SmS_{-m}. Causing the network utility to change from U(S1)U(S_{1}) to U(S2)U(S_{2}) and U(S2)>U(S1)U(S_{2})>U(S_{1}).

Definition 5.

A UAV exchanging adjustment means that UAV n1Sm1n_{1}\in S_{m_{1}} and n2Sm2n_{2}\in S_{m_{2}} are switched to each other’s training group. Causing the network utility to change from U(S1)U(S_{1}) to U(S2)U(S_{2}) and U(S2)>U(S1)U(S_{2})>U(S_{1}).

Definition 6.

A UAV-EV association strategy SS^{*} is at a stable system point if no EV mm and UAV nn can change its association strategy to obtain a higher global network utility.

Based on the above definition, we can solve (19) by continuously adjusting the UAV-EV association policy SS to obtain a higher utility according to the preference order. The association policy adjustment will terminate when a stable association SS^{*} is reached, where no UAV in the system has the motivation to leave its currently associated EV and join another EV.

After the UAV-EV association algorithm converges, the optimal UAV-EV association and resource allocation strategy will be broadcast to all UAVs. All the involved UAVs will participate in the task of its associated EV and execute local training.

V Numerical Results

We consider a cellular network with a coverage radius of 500mm, where NN UAVs and MM EVs are randomly distributed in the network.

V-A Datasets and local models

This work utilizes the MNIST dataset for simulation validation and some modifications are performed to make it suitable for multi-task training. The digits are first rotated, and the rotation angle is selected from the preset angle vector [0,36,,324][0,36,\cdots,324]. In addition, the digits and background are dyed differently to simulate the different backgrounds of the images collected by UAVs. For the modified MNIST dataset, we use it to complete two tasks: digit recognition and rotation angle recognition. Both tasks require information about the shape and contours of the digits, so there are similarities in the underlying tasks. For convenience, we use the same predictor for both tasks, but the results in this article can be easily extended to the case of different predictors. The model shared layers contain 3 convolutional layers, each followed by a batch normalization layer. Each task retains its own 4 fully connected layers.

V-B Hyperparameters and baselines

We set N=10N=10, M=2M=2, and distribute training data to each UAV based on two dirichlet distributions. The parameter α1\alpha_{1} controls the size distribution of the UAV dataset, and the parameter α2\alpha_{2} controls the proportional distribution of the digit categories and rotation angles. The remaining network parameters are shown in Table I\mathrm{I}. The three baseline strategies are set as follows: Strategy 1 shares the feature extractor, but UAV-EV associations are randomly associated. Strategy 2 shares the feature extractor, but each UAV is associated with its nearest EV. Strategy 3 does not share the feature extractor, and UAV-EV associations are randomly associated, which is similar to the FedAvg algorithm.

TABLE I: Parameter settings
o 0.45X[c]—X[b]—X[c]—X[m] Parameter Value Parameter Value
NN 10 MM 2
BB 2MHz LbL_{b} 500
N0N_{0} -174dBm HH 100m
η\eta 0.01 κ\kappa 0.8
ϖ\varpi 0.8 fnf_{n} [1,2][1,2]GHz

V-C Performance Comparison

Accuracy Comparison In this part, energy and time constraints are not considered, so we set λ=1\lambda=1. Figure 2(a)2(a) shows the average accuracy of the two tasks as the training progresses over multiple rounds. The results show that the proposed algorithm is better than the three baseline strategies. When compared to Strategy 1, the proposed algorithm achieves a 3.2%3.2\% improvement in final accuracy, demonstrating the effectiveness of the UAV-EV association algorithm. Moreover, Strategy 1 shows a 4.5%4.5\% accuracy improvement compared to Strategy 3, further confirming the effectiveness of sharing feature extractors among related tasks. Strategy 2 outperforms Strategies 1 and 3 in early training accuracy by accelerating model convergence through frequent interactions between each EV and its nearest UAVs. However, this approach weakens generalization, resulting in lower final accuracy. Figure 2(b)2(b) resents the average task performance variance across all rounds for the four strategies, demonstrating that the proposed algorithm achieves a better performance balance compared to the baseline strategies.

Refer to caption
(a) Average Accuracy
Refer to caption
(b) Variance
Figure 2: Average Task Accuracy and Variance on MNIST

Additionally, to investigate the impact of varying degrees of non-iid UAV data, we compare the proposed scheme with Strategy 1 to demonstrate the effectiveness of the UAV-EV association algorithm, and with Strategy 3 to highlight the benefits of sharing feature extractors. Figure 3(a)3(a) shows the percentage performance improvement of the proposed scheme relative to the other two strategies as α2\alpha_{2} varies, with α1=1\alpha_{1}=1 fixed. The results show that improvement decreases as α2\alpha_{2} increases, indicating the proposed scheme performs better with higher levels of non-iid data. This confirms that sharing feature extractors enhances the model’s robustness and generalization in non-iid scenarios.

Training time comparison To verify the effectiveness of the proposed algorithm in reducing total time consumption, we set λ=0.5\lambda=0.5 and compare the proposed algorithm with four benchmark strategies. The four benchmark strategies adopt random or nearest association, average or optimal bandwidth allocation respectively. The results in Figure 3(b)3(b) show that, except for the Distance-Opt strategy with the goal of time minimization, the proposed algorithm is significantly better than the other strategies under different numbers of UAVs, proving the effectiveness and optimality of the proposed joint bandwidth allocation and UAV-EV association algorithm.

Refer to caption
(a) Accuracy Improvement
Refer to caption
(b) Total Time
Figure 3: Accuracy Improvement and Total Time Consumption Comparison

VI Conclusion and Future Work

In this paper, we introduced a multi-task federated learning scheme based on task knowledge sharing. By sharing feature extractors among multiple related tasks, the overall task training acceleration and performance improvement were achieved. In addition, by analyzing the convergence of the algorithm, the joint UAV bandwidth allocation and UAV-EV association algorithm were designed to enhance training performance and reduce training time for each round. The simulation results confirm the effectiveness of the proposed algorithm. In the future, we plan to explore the underlying mechanisms for improving task performance through shared feature extractors among related tasks. Additionally, we will investigate effective methods for quantifying task correlations and design strategies to encourage collaboration in scenarios with dynamic task relevance.

References

  • [1] H. Yang, J. Zhao, Z. Xiong, K. -Y. Lam, S. Sun, and L. Xiao, “Privacy-Preserving Federated Learning for UAV-Enabled Networks: Learning-Based Joint Scheduling and Resource Management,” IEEE J. Sel. Areas Commun., vol. 39, no. 10, pp. 3144–3159, 2021.
  • [2] Z. Cui, T. Yang, X. Wu, H. Feng and B. Hu, “The Data Value Based Asynchronous Federated Learning for UAV Swarm Under Unstable Communication Scenarios,” IEEE Trans. Mob. Comput., vol. 23, no. 6, pp. 7165-7179, 2024.
  • [3] W. He, H. Yao, T. Mai, F. Wang and M. Guizani, “Three-Stage Stackelberg Game Enabled Clustered Federated Learning in Heterogeneous UAV Swarms,” IEEE Trans. Veh. Technol., vol. 72, no. 7, pp. 9366-9380, 2023.
  • [4] T. Wang, X. Huang, Y. Wu, L. Qian, B. Lin and Z. Su, “UAV Swarm-Assisted Two-Tier Hierarchical Federated Learning,” IEEE Trans. Netw. Sci. Eng., vol. 11, no. 1, pp. 943-956, 2024.
  • [5] H. Zhang and L. Hanzo, “Federated Learning Assisted Multi-UAV Networks,” IEEE Trans. Veh. Technol., vol. 69, no. 11, pp. 14104-14109, 2020.
  • [6] M. Fu, Y. Shi and Y. Zhou, “Federated Learning via Unmanned Aerial Vehicle,” IEEE Trans. Wirel. Commun., vol. 23, no. 4, pp. 2884-2900, 2024.
  • [7] V. Smith, C. K. Chiang, M. Sanjabi, et al., “Federated Multi-Task Learning,” in Proc. Adv. Neural Inf. Process. Syst. (NeurIPS), Long Beach, CA, USA, Dec. 2017, pp. 4427–4437.
  • [8] H. Ma, X. Yuan, Z. Ding, D. Fan and J. Fang, “Over-the-Air Federated Multi-Task Learning via Model Sparsification, Random Compression, and Turbo Compressed Sensing,” IEEE Trans. Wirel. Commun., vol. 22, no. 7, pp. 4974-4988, 2023.
  • [9] J. Yang, J. Jia, T. Deng, and M. Dong, “Efficient Scheduling for Multi-Job Vertical Federated Learning,” in Proc. IEEE Int. Conf. Commun. (ICC), Denver, CO, USA, 2024.
  • [10] C. Zhou, J. Liu, J. Jia, et al., “Efficient Device Scheduling with Multi-Job Federated Learning,” in Proc. AAAI Conf. Artif. Intell. (AAAI), Vancouver, Canada, Feb. 2022, pp. 9971–9979.
  • [11] J. Liu et al., “Multi-Job Intelligent Scheduling With Cross-Device Federated Learning,” IEEE Trans. Parallel Distrib. Syst., vol. 34, no. 2, pp. 535-551, 2023.
  • [12] X. Sun, R. Panda, R. Feris, et al., “Adashare: Learning What to Share for Efficient Deep Multi-Task Learning,” in Proc. Adv. Neural Inf. Process. Syst. (NeurIPS), Vancouver, Canada, Dec. 2020, pp. 8728–8740.
  • [13] Z. Chen, W. Yi, A. Nallanathan, and G. Y. Li, “Is Partial Model Aggregation Energy-Efficient for Federated Learning Enabled Wireless Networks?,” in Proc. IEEE Int. Conf. Commun. (ICC), Rome, Italy, 2023.
  • [14] X. Liu, Y. Deng and T. Mahmoodi, “Wireless Distributed Learning: A New Hybrid Split and Federated Learning Approach,” IEEE Trans. Wirel. Commun., vol. 22, no. 4, pp. 2650-2665, 2023.