This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Generalizable Learning Reconstruction for Accelerating MR Imaging via Federated Neural Architecture Search

Ruoyou Wu, Cheng Li, Juan Zou, Shanshan Wang Manuscript received August 23, 2023. (Corresponding author: Shanshan Wang.)Ruoyou Wu is with the Paul C. Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China, and with the Peng Cheng Laboratory, Shenzhen 518055, China, and also with the University of Chinese Academy of Sciences, Beijing 100049, China (e-mail: ry.wu@siat.ac.cn).Cheng Li is with the Pual C. Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China (e-mail: cheng.li6@siat.ac.cn).Juan Zou is with the School of Physics and Optoelectronics, Xiangtan University, Xiangtan 411105, China (e-mail: zjuan@xtu.edu.cn).Shanshan Wang is with the Paul C. Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China, and also with the Peng Cheng Laboratory, Shenzhen 518055, China (e-mail: sophiasswang@hotmail.com).
Abstract

Heterogeneous data captured by different scanning devices and imaging protocols can affect the generalization performance of the deep learning magnetic resonance (MR) reconstruction model. While a centralized training model is effective in mitigating this problem, it raises concerns about privacy protection. Federated learning is a distributed training paradigm that can utilize multi-institutional data for collaborative training without sharing data. However, existing federated learning MR image reconstruction methods rely on models designed manually by experts, which are complex and computational expensive, suffering from performance degradation when facing heterogeneous data distributions. In addition, these methods give inadequate consideration to fairness issues, namely, ensuring that the model’s training does not introduce bias towards any specific dataset’s distribution. To this end, this paper proposes a generalizable federated neural architecture search framework for accelerating MR imaging (GAutoMRI). Specifically, automatic neural architecture search is investigated for effective and efficient neural network representation learning of MR images from different centers. Furthermore, we design a fairness adjustment approach that can enable the model to learn features fairly from inconsistent distributions of different devices and centers, and thus enforce the model generalize to the unseen center. Extensive experiments show that our proposed GAutoMRI has better performances and generalization ability compared with six state-of-the-art federated learning methods. Moreover, the GAutoMRI model is significantly more lightweight, making it an efficient choice for MR image reconstruction tasks. The code will be made available at https://github.com/ternencewu123/GAutoMRI.

Index Terms:
Magnetic resonance imaging (MRI), neural architecture search (NAS), federated learning, fairness.
publicationid: pubid: 0000–0000/00$00.00 © 2023 IEEE

I Introduction

Magnetic resonance imaging (MRI) is a non-invasive imaging modality that can effectively assess the structure, morphology, and function of soft tissues. It plays a crucial role in clinical diagnosis and disease evaluation due to its ability to provide detailed and comprehensive information. However, acquiring fully-sampled kk-space data is usually time-consuming due to the physical limitations of the scanning device[1]. One possible way to improve the imaging speed is to reconstruct high-quality images from undersampled data[2]. Recently, deep learning has shown great potential in accelerating MR image reconstruction[3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16]. Deep learning offers the advantage of fast online reconstruction through offline learning from large datasets. However, due to factors like varying scanning devices, imaging protocols, and other discrepancies, the distribution of data among different centers can become highly heterogeneous. As a consequence, generalizing a model trained in one center to other centers becomes a challenging task, and ensuring the stability of the model becomes equally difficult. This could lead to biased performance at different centers. One conventional approach is to aggregate data from multiple centers for centralized learning[17]. However, collecting data from multiple centers can lead to significant resource consumption and may leak patient privacy[18, 19].

To address this limitation, federated learning emerges as a promising framework that allows multi-institutional collaborative learning while protecting patient privacy[20, 21, 22, 23, 24, 25, 26, 27, 28, 29]. In federated learning, each client independently trains a local model using its computational resources and private data. Subsequently, the server collects these locally trained models from all clients and employs an aggregation approach to get the server model. Finally, the server distributes the aggregated model to the clients for the subsequent communication rounds[21]. The classical federated learning method FedAvg[21] is effective for client model aggregation with the same distribution, but in MRI, the data distribution of different clients may present heterogeneity caused by different tissue contrasts, imaging protocols, and other factors. This may cause models trained by federated learning to suffer performance degradation. To this end, several federated learning MRI methods have been proposed to alleviate this problem. In[26], the similarity of latent-space representations is improved by repeatedly aligning the source and target sites. In [27], client specificity is maintained by dividing the reconstruction model into a globally shared encoder and a locally personalized decoder. While existing federated learning MRI reconstruction methods have achieved promising performance, the reconstruction models of these methods typically utilize hand-designed network architectures, which are often complex and require more computational resources. Additionally, these methods also rarely consider the fairness of the algorithm, which involve an obvious performance gap of the model across different centers, especially in the case of heterogeneous data distributions. This makes it difficult to generalize the trained model to other centers[29].

Neural Architecture Search (NAS)[30] can achieve better performance by automatically searching for the optimal network architecture while reducing the consumption of computational resources. A typical NAS architecture consists of three parts: search space, search approach, and evaluation approach. Recently, there have been some works on utilizing NAS alone for MR image reconstruction[31, 32]. All of these methods have yielded improved results, achieving better performance with fewer model parameters when compared to the existing hand-designed models. However, these methods don’t consider heterogeneous data distributions from different centers.

To address the above mentioned challenges, we propose a generalizable federated neural architecture search framework for accelerating MR imaging. Our main contributions can be summarized as follows:

1) A generalizable federated neural architecture search framework (GAutoMRI) is proposed for accurate under-sampled MR image reconstruction with lightweight models, using multi-institutional data in a privacy-preserving manner.

2) A fairness adjustment approach is designed to enhance the learning capabilities of the model across different centers, enabling the model to better generalize to both in-distribution and out-of-distribution data.

3) Extensive experimental validations have been performed on three public and one in-house data set with different undersampling patterns, imaging modalities, and scanning devices. Results show that our proposed GAutoMRI has better performances and generalization capability compared with six state-of-the-art federated learning methods with significantly lightweight models.

II Related Work

II-A MR Image Reconstruction

Reconstructing high-quality MR images from undersampling kk-space data is a typical ill-posed inverse problem. Deep learning-based methods learn the mapping relationship from undersampled to fully-sampled data offline and then enable fast online reconstruction[3]. These methods can avoid manual fine-tuning of optimization parameters. Deep learning-based MR image reconstruction methods can be broadly categorized into data-driven[5, 7, 9, 11, 12, 13, 14, 15, 16] and model-driven approaches[4, 6, 8, 10]. Data-driven methods typically rely on large amounts of data and directly learn the mapping relationship from undersampled data to fully-sampled data. However, the networks could be totally black-box. Model-driven approaches formulate an optimization problem and then unfold the iterative optimization algorithm into the network. The network architecture is constructed based on iterative steps, leading to a slightly longer reconstruction time. While these methods have made certain contributions to the development of MR image reconstruction, they all require the collection of large amounts of data for centralized training. This undoubtedly increases labor costs and resource consumption, and it also brings a serious issue—data privacy leakage. We utilize a distributed collaborative training scheme to effectively avoid the problems of data collection and patient privacy leakage.

II-B Federated Learning

Federated learning is a distributed collaborative training paradigm that protects data privacy by training shared model without sharing data[20, 21, 22, 23, 24, 25, 26, 27, 28, 29]. One of the most classical federated learning methods is FedAvg[21], which obtains the global model through an average aggregation method, and it obtains a relatively better performance when the data distributions of different centers are similar. However, in real-world scenarios, there is often statistical heterogeneity among different centers, and the traditional federated learning methods are difficult to adapt to such requirements. Some works start to reconcile the differences between global and local models [22, 23], aiming to obtain a global model with better performance. For example, Li et al.,[22] facilitate the learning of the global model by imposing an additional constraint on the client’s objective function. In addition, there are also some proposed personalized federated learning methods to preserve the specificity of local data. For example, Zhang et al.,[25] proposed a local adaptive aggregation approach to improve the performance of local clients by adaptively aggregating global and local models. As for magnetic resonance imaging, the problem of data heterogeneity may also exist in different centers due to scanning devices, imaging protocols, and other issues[33]. Some methods start to solve the problems of feature drift and domain generalization. For example, Guo et al.,[26] reduces the data heterogeneity of clients by repeatedly aligning the hidden feature distributions between the source and target domains. It selects a target site in each communication round and then aligns it with the representation of the remaining sites, which adds more communication consumption. In addition, Feng et al.,[27] effectively preserve local client-specific features by dividing the reconstruction model into a globally shared encoder and a locally personalized decoder. These approaches effectively mitigate the privacy leakage problem, but their model architectures are hand-designed by experts, which are often complex and require more computational resources. We utilize an automatically differentiable architecture search method to obtain an optimal model architecture, effectively reducing the number of model parameters and the need for computational resources.

II-C Neural Architecture Search

Neural architecture search can obtain optimal network architectures by employing an automated network architecture search approach while keeping fewer model parameters[30]. Based on different search strategies, neural architecture search algorithms can be broadly categorized into three types: reinforcement learning (RL)-based[34, 35], evolutionary algorithm (EA)-based[36, 37], and gradient-based (GD) methods[38, 39]. Reinforcement learning-based and evolutionary algorithm-based methods are gradient-free methods, in contrast to gradient-based methods that tend to be faster in the searching process and consume slightly fewer computational resources[39]. DARTS is the most classical gradient-based method, which scales the selection of candidate operations into continuous variables and then utilizes efficient gradient back-propagation for learning, effectively reducing the search time[38]. In addition, there is also some NAS-based work in MR image reconstruction[31, 32]. For example, Yan et al., [31] used NAS for compressed sensing MR image reconstruction, achieving enhanced performance. These NAS-based MR image reconstruction methods are used for centralized training scenarios but have not yet been explored in federated learning scenarios. In this paper, we utilize federated learning and neural architecture search for MR image reconstruction and improve the generalization performance for both in-distribution and out-of-distribution data through a fairness adjustment approach.

III Methodology

III-A DL-based MR Image Reconstruction

The goal of MR image reconstruction is to reconstruct a high-quality image 𝐱N(M<N)\mathbf{x}\in\mathbb{C}^{N}(M<N) from undersampled kk-space data 𝐛M\mathbf{b}\in\mathbb{C}^{M}, formulated as:

𝐛=𝐀𝐱+ϵ\mathbf{b}=\mathbf{Ax}+\mathbf{\epsilon} (1)

where 𝐀M×N\mathbf{A}\in\mathbb{C}^{M\times N} denotes the encoding matrix and ϵM\mathbf{\epsilon}\in\mathbb{C}^{M} denotes the measurement noise. Solving 𝐱\mathbf{x} from Eq. (1) is an ill-posed inverse problem that can be expressed as the following optimization problem:

min𝐱12𝐀𝐱𝐛+λR(𝐱)\min_{\mathbf{x}}\frac{1}{2}\left\|\mathbf{Ax}-\mathbf{b}\right\|+\lambda R(\mathbf{x}) (2)

where the first term is the data fidelity, the second term is the regularization term, and λ\lambda is the regularization parameter used to balance the tradeoff between the data fidelity term and the regularization term. Based on the method of model unfolding[8], Eq. (2) can be transformed into the following alternating iteration process:

{𝐳j=fθ(𝐱j)𝐱j+1=(𝐀H𝐀+λ𝐈)1(𝐀H𝐛+λ𝐳j)\left\{\begin{array}[]{lr}\mathbf{z}^{j}=f_{\theta}\left(\mathbf{x}^{j}\right)&\\ \mathbf{x}^{j+1}=\left(\mathbf{A}^{H}\mathbf{A}+\lambda\mathbf{I}\right)^{-1}\left(\mathbf{A}^{H}\mathbf{b}+\lambda\mathbf{z}^{j}\right)\end{array}\right. (3)

where fθf_{\theta} denotes the deep neural network used for denoising, 𝐀H\mathbf{A}^{H} denotes the conjugate operator of 𝐀\mathbf{A}, and jj denotes the iterative control variable, λ\lambda is learnable.

III-B Details of GAutoMRI

The overall architecture of GAutoMRI is shown in Fig. 1, which consists of two phases: the searching phase and the training phase. In the searching phase, we adjusted the search space according to the reconstruction task, details of which can be found in III.B.2 and Fig. 2, and the whole search process draws on the method in [40] to obtain the optimal network architecture. In the training phase, the architecture obtained in the searching phase is employed as the base model. In addition, we introduce a fairness adjustment approach to improve the generalization of the global model over out-of-distribution data. The specific search process and fairness adjustment approach will be described later.

Refer to caption
Figure 1: The overall architecture of our proposed GAutoMRI. It consists of two phases: (a) Searching phase. The aim of this phase is to obtain the optimal architecture using the federated neural architecture search method and use the searched architecture for the training phase. In this paper, we utilize the approach outlined in [40] to achieve this, while adjusting the search space to suit our specific MR image reconstruction task. (b) Training phase. The optimal architecture obtained in the searching phase is utilized as the base model. In addition, we introduce a fairness adjustment approach to improve the generalization of the global model over out-of-distribution data, as detailed in III.C.

III-B1 Preliminaries

In the classical federated learning setting, the model architecture is fixed. It is assumed that there are CC clients or hospitals, each with a dataset Dc={(xci,bci)}i=1NcD_{c}=\{(x_{c}^{i},b_{c}^{i})\}_{i=1}^{N_{c}}, which is non IID. The goal of federated learning is to learn a global model on dataset D={D1,D2,,DC}D=\{D_{1},D_{2},...,D_{C}\} in a distributed training manner, which can be expressed as:

minθ𝔏(θ)=c=1CNcN{𝔼(𝐱c,𝐛c)𝒟cc(f(𝐛c;θ))}\min_{\theta}\mathfrak{L(\theta)}=\sum_{c=1}^{C}\frac{N_{c}}{N}\{\mathbb{E}_{(\mathbf{x}_{c},\mathbf{b}_{c})\in\mathcal{D}_{c}}\ell_{c}(f(\mathbf{b}_{c};\theta))\} (4)

where NcN_{c} and NN denote the number of samples for client cc and the total number of samples, respectively; θ\theta denotes network parameters; c()\ell_{c}(\cdot) denotes the local loss on client cc. For federated neural architecture search, it is necessary to optimize both model parameters θ\theta and architecture parameters α\alpha. Formally, the objective function is defined as:

minθ,αc=1CNcN{𝔼(𝐱c,𝐛c)𝒟cc(f(𝐛c;θ,α))}\min_{\theta,\alpha}\sum_{c=1}^{C}\frac{N_{c}}{N}\{\mathbb{E}_{(\mathbf{x}_{c},\mathbf{b}_{c})\in\mathcal{D}_{c}}\ell_{c}(f(\mathbf{b}_{c};\theta,\alpha))\} (5)

Our goal is to search the optimal architecture parameter α\alpha and the corresponding model parameters θ\theta by utilizing a differentiable architecture search method, making it fit the heterogeneous data distribution, thereby improving the generalization performance of out-of-distribution data.

III-B2 Search Space

Typically, a NAS method consists of three consecutive parts: search space, search approach, and performance evaluation approach[30]. Our search space follows the mixed-operation search space defined in DARTS[38], as shown in Fig. 2. We designed our candidate set of operations, denoted as 𝒪\mathcal{O}, based on the number of parameters and test performance of the target architecture, aiming to make it more suitable to our reconstruction task. The set includes: 1) Standard convolution, which extracts multi-scale information using convolution kernels of different sizes: std_conv_3×3std\_conv\_3\times 3, std_conv_5×5std\_conv\_5\times 5, std_conv_7×7std\_conv\_7\times 7; 2) Depthwise separable convolution, reducing parameters and computation while improving feature extraction and representation ability: sep_conv_3×3sep\_conv\_3\times 3, sep_conv_5×5sep\_conv\_5\times 5 and sep_conv_7×7sep\_conv\_7\times 7; 3) Dilated convolution, enlarging receptive field and reducing the number of parameters: dil_2_conv_3×3dil\_2\_conv\_3\times 3 and dil_3_conv_3×3dil\_3\_conv\_3\times 3. An edge between two nodes in Fig. 2(b) represents one of these candidate operations. Inside the cell, to make the search space continuous by relaxing the categorical candidate operations between two nodes, mixed operations using the softmaxsoftmax function over all candidate operations are performed:

ϕ¯(i,j)(x)=o=1|𝒪|exp(αo(i,j))o=1|𝒪|exp(αo(i,j))ϕo(i,j)(x)\bar{\phi}^{(i,j)}(x)=\sum_{o=1}^{|\mathcal{O}|}\frac{exp(\alpha_{o}^{(i,j)})}{{\textstyle\sum_{o^{{}^{\prime}}=1}^{|\mathcal{O}|}}exp(\alpha_{o^{{}^{\prime}}}^{(i,j)})}\phi_{o}^{(i,j)}(x) (6)

where ϕ¯(i,j)(x)\bar{\phi}^{(i,j)}(x) denotes the mixed operation for a pair of nodes (i,j)(i,j), |𝒪||\mathcal{O}| denotes the number of candidate operations, αo(i,j)\alpha_{o}^{(i,j)} denotes the weight of the otho^{th} operation for a pair of node (i,j)(i,j), and ϕo(i,j)(x)\phi_{o}^{(i,j)}(x) denotes the otho^{th} operation for a pair of node (i,j)(i,j). The aim of architecture search is to learn the optimal encoding α={α(i,j)}\alpha=\{\alpha^{(i,j)}\} of the architecture.

Refer to caption
Figure 2: The overall structure of search space. (a) The structure of denoiser module fθ()f_{\theta}(\cdot) in our reconstruction network. The searched cells are stacked to form our basic architecture, with each cell connected to the two consecutive cells that follow it. (b) The search architecture of the cell. PnP_{n} denotes the node in the cell.
1
Input : The datasets from CC clients: 𝒟1,𝒟2,,𝒟C={(𝐛i,𝐱i)i=1,,NC}\mathcal{D}_{1},\mathcal{D}_{2},\cdots,\mathcal{D}_{C}=\{(\mathbf{b}^{i},\mathbf{x}^{i})\mid i=1,\cdots,N_{C}\}; communication rounds: RR; local epochs: ZZ; learning rate for α\alpha: ηα\eta_{\alpha}; learning rate for θ\theta: ηθ\eta_{\theta} total sample size: NN; tradeoff coefficient: λ\lambda ;
2
3ServerExecution :
4       Initialize global parameters α0\alpha^{0} and θ0\theta^{0};
5       for r=1,2,,Rr=1,2,\cdots,R do
6             for c=1,2,,Cc=1,2,\cdots,C do
7                   Get global parameters αcr\alpha_{c}^{r}, θcr\theta_{c}^{r} \leftarrow αr\alpha^{r}, θr\theta^{r} ;
8                   α^cr\hat{\alpha}_{c}^{r}, θ^cr\hat{\theta}_{c}^{r} \leftarrow LocalSearch(c, αcr\alpha_{c}^{r}, θcr\theta_{c}^{r});
9                  
10             end for
11            αr+1c=1CNcNα^cr\alpha^{r+1}\leftarrow{\textstyle\sum_{c=1}^{C}\frac{N_{c}}{N}\hat{\alpha}_{c}^{r}} ;
12             θr+1c=1CNcNθ^cr\theta^{r+1}\leftarrow{\textstyle\sum_{c=1}^{C}\frac{N_{c}}{N}\hat{\theta}_{c}^{r}} ;
13            
14       end for
15      return αr+1\alpha^{r+1} and θr+1\theta^{r+1} ;
16      
17
18LocalSearch (c,θ,αc,\theta,\alpha):
19       for z=1,2,,Zz=1,2,\cdots,Z do
20             Get the updated parameters α𝔏tr(θcr,z,αcr,z)\bigtriangledown_{\alpha}\mathfrak{L}_{tr}(\theta_{c}^{r,z},\alpha_{c}^{r,z}) and α𝔏val(θcr,z,αcr,z)\bigtriangledown_{\alpha}\mathfrak{L}_{val}(\theta_{c}^{r,z},\alpha_{c}^{r,z}) ;
21             Update α^cr,z+1αcr,zηα(α𝔏tr(θcr,z,αcr,z)+\hat{\alpha}_{c}^{r,z+1}\leftarrow\alpha_{c}^{r,z}-\eta_{\alpha}(\bigtriangledown_{\alpha}\mathfrak{L}_{tr}(\theta_{c}^{r,z},\alpha_{c}^{r,z})+
22             λα𝔏val(θcr,z,αcr,z))\lambda\bigtriangledown_{\alpha}\mathfrak{L}_{val}(\theta_{c}^{r,z},\alpha_{c}^{r,z})) ;
23             Get the updated parameters θ𝔏tr(θcr,z,αcr,z)\bigtriangledown_{\theta}\mathfrak{L}_{tr}(\theta_{c}^{r,z},\alpha_{c}^{r,z}) ;
24             Update θ^cr,z+1θcr,zηθ(θ𝔏tr(θcr,z,αcr,z))\hat{\theta}_{c}^{r,z+1}\leftarrow\theta_{c}^{r,z}-\eta_{\theta}(\bigtriangledown_{\theta}\mathfrak{L}_{tr}(\theta_{c}^{r,z},\alpha_{c}^{r,z})) ;
25            
26       end for
27      return α^cr\hat{\alpha}_{c}^{r} and θ^cr\hat{\theta}_{c}^{r} ;
28      
29
Algorithm 1 GAutoMRI-Searcher follows in [40]

III-B3 Search Process

During the search process, to optimize the objective function in Eq. (5), we refer to the method in [40] to find the optimal architecture with a distributed neural architecture search algorithm. The details of the search process are shown in Algorithm. 1. First, the initialized global architecture parameters α0\alpha^{0} and model parameters θ0\theta^{0} are distributed to each local client; then each client updates the architecture parameters αcr\alpha_{c}^{r} and model parameters θcr\theta_{c}^{r} using the local private data as follows:

{α^cr,z+1αcr,zηα(α𝔏tr(θcr,z,αcr,z)+λα𝔏val(θcr,z,αcr,z))θ^cr,z+1θcr,zηθ(θ𝔏tr(θcr,z,αcr,z))\left\{\begin{array}[]{lr}\hat{\alpha}_{c}^{r,z+1}\leftarrow\alpha_{c}^{r,z}-\eta_{\alpha}(\bigtriangledown_{\alpha}\mathfrak{L}_{tr}(\theta_{c}^{r,z},\alpha_{c}^{r,z})+\\ \lambda\bigtriangledown_{\alpha}\mathfrak{L}_{val}(\theta_{c}^{r,z},\alpha_{c}^{r,z}))\\ \hat{\theta}_{c}^{r,z+1}\leftarrow\theta_{c}^{r,z}-\eta_{\theta}(\bigtriangledown_{\theta}\mathfrak{L}_{tr}(\theta_{c}^{r,z},\alpha_{c}^{r,z}))\end{array}\right. (7)

where rr and zz denote global communication rounds and local training epochs, respectively. 𝔏tr(θcr,z,αcr,z)\mathfrak{L}_{tr}(\theta_{c}^{r,z},\alpha_{c}^{r,z}) and 𝔏val(θcr,z,αcr,z)\mathfrak{L}_{val}(\theta_{c}^{r,z},\alpha_{c}^{r,z}) denote the local training loss and validation loss with respect to the architecture parameters αcr,z\alpha_{c}^{r,z} and model parameters θcr,z\theta_{c}^{r,z}, respectively. ηα\eta_{\alpha} and ηθ\eta_{\theta} denote the learning rate of architecture parameters α\alpha and model parameters θ\theta, respectively.

After completing the local training, each client sends the architecture parameters α^cr\hat{\alpha}_{c}^{r} and model parameters θ^cr\hat{\theta}_{c}^{r} to the server. The server receives the parameters from each client and performs the aggregation operation as follows:

{αr+1c=1CNcNα^crθr+1c=1CNcNθ^cr\left\{\begin{array}[]{lr}\alpha^{r+1}\leftarrow\sum_{c=1}^{C}\frac{N_{c}}{N}\hat{\alpha}_{c}^{r}\\ \theta^{r+1}\leftarrow\sum_{c=1}^{C}\frac{N_{c}}{N}\hat{\theta}_{c}^{r}\end{array}\right. (8)

where αr+1\alpha^{r+1} and θr+1\theta^{r+1} denote the global architecture parameters and model parameters for the rthr^{th} communication round, respectively. Then, the server distributes the global parameters αr+1\alpha^{r+1} and θr+1\theta^{r+1} to each client for the next communication round. After RR communication rounds, we can obtain the optimal architecture parameters and model parameters without sharing local private data.

1
Input : The datasets from CC clients: 𝒟1,𝒟2,,𝒟C={(𝐛i,𝐱i)i=1,,NC}\mathcal{D}_{1},\mathcal{D}_{2},\cdots,\mathcal{D}_{C}=\{(\mathbf{b}^{i},\mathbf{x}^{i})\mid i=1,\cdots,N_{C}\}; communication rounds: RR; local epochs: ZZ; learning rate for θ\theta: ηθ\eta_{\theta}; initialized aggregation weights: ac0=1Ca_{c}^{0}=\frac{1}{C} ; Incremental control parameters: γ\gamma;
2
3ServerExecution :
4       Initialize global parameters θ0\theta^{0} ;
5       for r=1,2,,Rr=1,2,\cdots,R do
6             for c=1,2,,Cc=1,2,\cdots,C do
7                   Get global parameters θcrθr\theta_{c}^{r}\leftarrow\theta^{r};
8                   𝒢Dc(θr)\mathcal{G}_{D_{c}}(\theta^{r}), θ^cr\hat{\theta}_{c}^{r} \leftarrow LocalSearch(c,θcrc,\theta_{c}^{r});
9                  
10             end for
11            Update the aggregation weights acra_{c}^{r} based on Eq. (10) and Eq. (11);
12            
13            Aggregate θr+1c=1Cacrθ^cr\theta^{r+1}\leftarrow\sum_{c=1}^{C}a_{c}^{r}\cdot\hat{\theta}_{c}^{r};
14            
15       end for
16      return θr+1\theta^{r+1}
17
18LocalSearch (c, θ\theta):
19       Get the updated parameters θr\theta^{r};
20       Compute 𝒢Dc(θr)\mathcal{G}_{D_{c}}(\theta^{r}) based on Eq. (9);
21       for z=1,2,,Zz=1,2,\cdots,Z do
22             Get the updated gradients θ𝔏tr(𝐱c,𝐛c;θcr,z)\bigtriangledown_{\theta}\mathfrak{L}_{tr}(\mathbf{x}_{c},\mathbf{b}_{c};\theta_{c}^{r,z});
23            
24            Update θ^cr,z+1θcr,zηθ(θ𝔏tr(𝐱c,𝐛c;θcr,z))\hat{\theta}_{c}^{r,z+1}\leftarrow\theta_{c}^{r,z}-\eta_{\theta}(\bigtriangledown_{\theta}\mathfrak{L}_{tr}(\mathbf{x}_{c},\mathbf{b}_{c};\theta_{c}^{r,z}));
25            
26       end for
27      return 𝒢Dc(θr)\mathcal{G}_{D_{c}}(\theta^{r}) and θ^cr\hat{\theta}_{c}^{r}  
28
Algorithm 2 GAutoMRI-Trainer

III-C Fairness Adjustment for FL

After the searching phase, we can obtain an optimal model architecture, which is subsequently utilized for the training phase. However, some existing federated MR image reconstruction methods seldom consider inter-client fairness issues, leading to performance degradation of the trained global model on out-of-distribution data, especially when faced with statistical heterogeneity among clients. Intuitively, if the global model does not discriminate in performance on known clients (i.e., the performance gap is small on each client), then it logically should exhibit respectable performance even when faced with out-of-distribution data. Based on this notion, we introduce a fairness adjustment approach that utilizes the difference between the empirical risks of the global and local models in each communication round. This approach dynamically adjusts the aggregation weights of the clients, aiming to progressively diminish the performance gap of the global model across the known clients. As a result, this approach gradually improves the generalization performance of the global model on out-of-distribution data. Formally, the empirical risk gap between global and local models can be expressed as:

𝒢Dc(θr)=Dc(θr)Dc(θ^cr1),c=1,2,,C\mathcal{G}_{D_{c}}(\theta^{r})=\mathbf{\ell}_{D_{c}}(\theta^{r})-\mathbf{\ell}_{D_{c}}(\hat{\theta}_{c}^{r-1}),c=1,2,\cdots,C (9)

where Dc(θr)\ell_{D_{c}}(\theta^{r}) and Dc(θ^cr1)\ell_{D_{c}}(\hat{\theta}_{c}^{r-1}) denote the empirical risk of the global and local model, respectively. After completing local model training, the client sends the empirical risk gap 𝒢Dc(θr)\mathcal{G}_{D_{c}}(\theta^{r}) and local model parameters θ^cr\hat{\theta}_{c}^{r} to the server. After the server receives 𝒢Dc(θr)\mathcal{G}_{D_{c}}(\theta^{r}) and θ^cr\hat{\theta}_{c}^{r} from each client, it first updates the aggregation weight of each client according to the fairness adjustment approach, which is formulated as:

{βcr=acr1 if 𝒢Dc(θr)0βcr=acr1+γ𝒢Dc(θr)max(𝒢Dc´(θr)) if 𝒢Dc(θr)>0\begin{array}[]{l}\begin{cases}\beta_{c}^{r}=a_{c}^{r-1}&\text{ if }\mathcal{G}_{D_{c}}(\theta^{r})\leq 0\\ \beta_{c}^{r}=a_{c}^{r-1}+\gamma\frac{\mathcal{G}_{D_{c}}(\theta^{r})}{\max(\mathcal{G}_{D_{\acute{c}}}(\theta^{r}))}&\text{ if }\mathcal{G}_{D_{c}}(\theta^{r})>0\end{cases}\end{array} (10)
acr=βcrc=1Cβcr,c=1Cacr=1a_{c}^{r}=\frac{\beta_{c}^{r}}{{\textstyle\sum_{c=1}^{C}}\beta_{c}^{r}},\qquad\sum_{c=1}^{C}a_{c}^{r}=1 (11)

where c´={1,2,,C}\acute{c}=\{1,2,\cdots,C\}, acr1a_{c}^{r-1} denotes the aggregation weight of the cthc^{th} client at the (r1)th(r-1)^{th} communication round, acra_{c}^{r} is the weight updated by the fairness adjustment approach, γ\gamma is utilized to control the magnitude of the weight increment.

In each communication round, for each client cc, if the empirical risk gap 𝒢Dc(θr)0\mathcal{G}_{D_{c}}(\theta^{r})\leq 0, we do not change its aggregation weight; if the empirical risk gap 𝒢Dc(θr)>0\mathcal{G}_{D_{c}}(\theta^{r})>0, which indicates that the global model does not generalize very well for local client, we increase its aggregation weight to reduce the empirical risk gap. In this way, we can continuously adjust the empirical risk gap on each client, and finally make the empirical risk gap smaller on each client, so that the global model can adapt to out-of-distribution data. The details of the training phase are shown in Algorithm. 2.

IV Experiments

IV-A Experimental Setup

IV-A1 Datasets

In this paper, we use three public datasets (fastMRI[33], cc359111https://sites.google.com/view/calgary-campinas-dataset/mr-reconstruction-
challenge
, MoDL-Brain222https://github.com/hkaggarwal/modl) and one In-house dataset[15]. Details of the datasets are provided as follows: 1) fastMRI: a large public MRI dataset where we use only the T1 and T2 data of the brain, each including 285 subjects respectively. Among them, the T2 data is only used for validation experiments of different contrast; 2) cc359: there are 35 T1-weighted MR scans acquired on a clinical MR scanner (Discovery MR 750; General Electric (GE) Healthcare, Waukesha, WI); 3) MoDL-Brain: there are 524 slices in total, and this dataset is only used for validation experiments of unseen center; 4) In-house: There are T1, T2, and PD data for 22 subjects acquired in a United Imaging system, uMR 790. All data were divided into training, validation, and testing sets with the ratio of 7: 1: 2. Fig. 3 shows the t-SNE visualization results of the distribution of the four datasets, from which it can be seen that the distribution of the four datasets is still quite different, which also shows the problem of statistical heterogeneity of the client. And the difference in data distribution between fastMRI T1 and T2 is somewhat smaller.

Refer to caption
Figure 3: t-SNE visualizations of data distribution from four datasets: fastMRI, cc359, MoDL-Brain, and In-house.

IV-A2 Competing Methods

To validate our proposed method, we compare it with six state-of-the-art FL algorithms, including: 1) FedAvg[21], a classical FL algorithm that is trained by average aggregation method; 2) FedProx[22], learns a global model by adding a regularization term to the local objective function; 3) MOON[23], an FL algorithm that corrects for local update through contrastive learning of model representations; 4) AM[24], a Fourier-based augmentation method, which was used in [24], we removed the segmentation task; 5) FedMRI[27], a personalized FL algorithm that adds a weighted contrastive regularization term to each client; 6) FedALA[25], learns personalized client model by local adaptive aggregation method.

IV-A3 Implementation Details

All the networks were trained using the PyTorch framework with one NVIDIA RTX A6000 GPU (with 48GB memory). The whole process is divided into two phases. In the searching phase, the Adam optimizer with a learning rate of 1e-4 is used to update architecture parameters, and the weight decay is set to 1e-3. Besides, the Adam optimizer with a learning rate of 1e-3 is used to update local model parameters. We set the batch size to 4. The models are trained with 50 global communication rounds and 5 local epochs for each round. In the training phase, the AdamW optimizer is adopted. We set the batch size and initial learning rate to 24 and 1e-3, respectively. Networks are trained with 150 global communication rounds and 5 local epochs for each round.

IV-B In-Distribution Results

To fully compare the performance of our method, we set up two scenarios: In-Distribution and Out-of-Distribution (different sampling pattern, acceleration rate, contrast, and unseen center) experiments. In this section, only the results of the In-Distribution scenario are presented, and the results of the Out-of-Distribution scenario will be presented in IV. C. The In-Distribution experiments are conducted using a 1D random 4×\times mask.

Table I provides the quantitative results of the different methods under In-Distribution scenario. From Table I, we can see that our method is superior to other FL methods in terms of PSNR and SSIM metrics. Compared with the classical FedAvg method, the average value of the PSNR metric of our GAutoMRI improves from 32.3288 dB to 35.5253 dB, and the average value of the SSIM metric increases from 0.8958 to 0.9384. This is primarily attributed to the fact that FedAvg uses an average aggregation approach, which does not adapt well to the statistical heterogeneity of the client. Comparatively, personalized federated learning methods (FedMRI and FedALA) perform better than FedAvg. They exhibit improved adaptability to local data distributions by maintaining client specificity. The In-house dataset comprises data with three different contrasts, resulting in greater heterogeneity, as illustrated in Fig. 3. Therefore, the performance of all methods on this dataset is slightly lower compared to their performance on the other two datasets. However, relatively speaking, our method still maintains a better performance, proving the superiority of our methods. Furthermore, our fairness adjustment approach contributes to achieving an obviously improved performance balance for each client.

In addition to the quantitative results, the qualitative results of the different methods are also presented in Fig. 4. The reference images and reconstruction results from the comparative methods are displayed from left to right, with the corresponding error maps shown in the second and fourth rows. The corresponding quantitative metrics are also provided in the upper-right corner of the reconstructed images. From Fig. 4, our method obtains the best reconstruction quality and the smallest error, which also indicates that our method is superior to other methods.

TABLE I: Quantitative results of different methods with respect to In-Distribution scenario. Bold numbers indicate our results. Detailed analyses are described in III.B.
method In-Distribution
fastMRI T1 cc359 In-house avg
PSNR SSIM PSNR SSIM PSNR SSIM PSNR SSIM
FedAvg[21] 35.7327 0.9551 31.6436 0.9060 29.6101 0.8262 32.3288 0.8958
FedProx[22] 35.7656 0.9544 31.6642 0.9039 29.9135 0.8317 32.4478 0.8967
MOON[23] 35.4647 0.9547 31.7809 0.9088 26.0787 0.7490 31.1081 0.8708
AM[24] 35.8460 0.9559 32.2660 0.9125 27.7329 0.7926 31.9483 0.8870
FedMRI[27] 36.3148 0.9588 31.7741 0.9070 30.7310 0.8638 32.9400 0.9099
FedALA[25] 36.5881 0.9603 31.9418 0.9076 31.0660 0.8695 33.1986 0.9125
GAutoMRI 39.5964 0.9767 35.1023 0.9506 31.8779 0.8880 35.5253 0.9384
Refer to caption
Figure 4: Qualitative reconstruction results of different methods under In-Distribution scenario. From left to right, the eight images corresponding to the reference image, the reconstructed images of FedAvg, FedProx, MOON, AM, FedMRI, FedALA, and our GAutoMRI, respectively. The second and fourth rows plot the corresponding error maps.

IV-C Out-of-Distribution Results

To demonstrate the performance of our proposed GAutoMRI on out-of-distribution data, we compare the model performance at different sampling pattern (1D random 4×\times for training; 1D equispaced 4×\times for testing), acceleration rate (1D random 4×\times for training; 1D random 6×\times for testing), contrast (fastMRI T2) and unseen center (MoDL-Brain). We divided it into two parts: 1) different sampling pattern and acceleration rate, and 2) different contrast and unseen center.

IV-C1 Different Sampling Pattern and Acceleration Rate

Table II shows the quantitative results at different sampling pattern and acceleration rate. In terms of PSNR and SSIM values, our proposed method exhibits superior performance compared to the other methods. Compared to FedAvg, PSNR and SSIM improved from 27.8763 dB and 0.8271 to 30.0927 dB and 0.8602 under different sampling pattern; and PSNR and SSIM improved from 28.6256 dB and 0.8403 to 31.8182 dB and 0.8938 under different acceleration rate. This indicates that our approach demonstrates obviously improved performance on out-of-distribution data, and also shows that our approach has better generalization performance. The In-house dataset consists of three modalities, resulting in a notably heightened level of data heterogeneity. Consequently, the performance of all methods is expected to exhibit a slight reduction. However, the performance of our method remains the best.

TABLE II: Quantitative results of different methods with respect to Out-of-Distribution scenario (sampling pattern and acceleration rate). Bold numbers indicate our results. Detailed analyses are described in III.C.
method Out-of-Distribution
Sampling pattern Acceleration rate
PSNR SSIM PSNR SSIM
FedAvg[21] 27.8763 0.8271 28.6256 0.8403
FedProx[22] 27.9759 0.8279 28.7533 0.8425
MOON[23] 27.0243 0.7992 27.7452 0.8134
AM[24] 27.7492 0.8226 28.7145 0.8401
FedMRI[27] 28.1829 0.8438 28.8346 0.8531
FedALA[25] 28.0714 0.8470 28.9034 0.8526
GAutoMRI 30.0927 0.8602 31.8182 0.8938
Refer to caption
Figure 5: Qualitative reconstruction results of different methods under Out-of-Distribution scenario (different contrast and unseen center). From left to right, the seven images corresponding to the reference image, the reconstructed images of FedAvg, FedProx, MOON, AM, FedALA, and our GAutoMRI, respectively. The second and fourth rows plot the corresponding error maps.

IV-C2 Different Contrast and Unseen Center

Table III shows the quantitative results of the generalization performance at different contrast and unseen center, and the experimental results show that the performance of our method is the best. Compared to FedAvg, under different contrast, PSNR improved from 29.7607 dB to 34.6154 dB, and SSIM improved from 0.9126 to 0.9647. Under the unseen center, PSNR improved from 29.0923 dB to 31.5570 dB, and SSIM improved from 0.7879 to 0.8834. As depicted in Fig. 3, the disparity between the data distributions of fastMRI T1 and T2 is relatively small, resulting in somewhat lesser performance degradation for all methods in scenarios involving different contrast. Generalizing models trained on heterogeneous data distributions to the unseen center is a more challenging task. However, our GAutoMRI also maintains a better performance, attributed to the fairness adjustment approach we introduced. This approach enhances generalization under the unseen center by mitigating the performance gap of the global model across known centers. AM is a domain generalization method, and its performance is the best except for our method. Furthermore, we also compared the parameter size of individual client models. As indicated in Table III, the model obtained through the neural architecture search exhibits a notably reduced parameter size, which indicates that our approach is more efficient. In addition, we visualized the reconstructed images and the corresponding error maps for the different methods, as shown in Fig. 5. From the error maps, it can be seen that the error of our method is minimized, indicating our method outperforms the other methods. Overall, our proposed method has better generalization performance.

TABLE III: Quantitative results of different methods with respect to Out-of-Distribution scenario (different contrast and unseen center). Bold numbers indicate our results. Detailed analyses are described in III.C.
method Out-of-Distribution
Different contrast Unseen Center #Params/M
PSNR SSIM PSNR SSIM
FedAvg[21] 29.7607 0.9126 29.0923 0.7879 7.76
FedProx[22] 29.8297 0.9107 29.0932 0.7840 7.76
MOON[23] 30.0428 0.9174 29.2461 0.7891 41.44
AM[24] 30.3382 0.9206 30.1058 0.8168 7.76
FedALA[25] 29.7476 0.9132 29.0459 0.7842 7.76
GAutoMRI 34.6154 0.9647 31.5570 0.8834 0.0487

IV-D Ablation Studies

In this section, we conduct ablation studies and utilize a radar chart to analyze the performance differences among different methods within the source domain.

IV-D1 Ablation Studies of Fairness Adjustment Approach

Table IV shows the quantitative results of the ablation studies conducted on the fairness adjustment (FA) approach under the In-Distribution scenario. where w/o(FA) denotes the result without the fairness adjustment approach. Compared to the hand-designed model, the performance of the model obtained using the neural architecture search exhibits an improved performance. The PSNR and SSIM metrics have been improved from 33.8465 dB and 0.9126 to 35.2859 dB and 0.9342, respectively. GAutoMRI requires only about 18% of the computational and parameter sizes of the baseline model, underscoring the efficiency of the searched model. Upon incorporating the fairness adjustment approach, there is a clear improvement in the average metrics, even though there is a slight decline in performance on the cc359 dataset. This is because increasing the aggregation weight of one client may affect the aggregation weight of other clients, resulting in slight performance degradation, which is acceptable.

In addition, we explored the role of FA on the generalization performance of the model. Figs. 6 and 7 show the quantitative and qualitative results of FA on unseen center, respectively. From the quantitative metrics, when the fairness adjustment approach is added, the performance on the unseen center is improved, which verifies the facilitating effect of FA on improving the generalization of the model. From the error maps of the qualitative results, when FA is introduced, the error of the reconstructed image is reduced. Overall, the fairness adjustment approach promotes the generalization performance of the model.

TABLE IV: Ablation study about fairness adjustment approach under In-Distribution scenario (without (w/o)).
method fastMRI T1 cc359 In-house avg FLOPs/G #Param/M
PSNR SSIM PSNR SSIM PSNR SSIM PSNR SSIM
baseline 37.4104 0.9598 33.8985 0.9320 30.2305 0.8461 33.8465 0.9126 85.6503 0.2614
w/o(FA) 39.5233 0.9778 35.1219 0.9520 31.2125 0.8729 35.2859 0.9342 15.7444 0.0487
GAutoMRI 39.5964 0.9767 35.1023 0.9506 31.8779 0.8880 35.5255 0.9384 - -
Refer to caption
Figure 6: Quantitative results of ablation experiment under Out-of-Distribution scenario (unseen center).
Refer to caption
Figure 7: Qualitative results of ablation experiment under Out-of-Distribution scenario (unseen center).

IV-D2 Testing Errors in Source Domain

This section focuses on analyzing the fairness of the different methods, and Fig. 8 shows a radar chart of the testing errors of the different methods under In-Distribution scenario. The chart features five axes that symbolize the testing errors of each method across the three datasets, as well as the mean and standard deviation of these testing errors. The magnitude of the standard deviation is utilized to portray the fairness of the model, as depicted in Fig. 8, the standard deviation of our method is the smallest and the area enclosed by the pentagon is also the smallest, indicating the better fairness of our proposed GAutoMRI. The contribution of FA to model fairness can also be observed.

Refer to caption
Figure 8: Radar chart of the testing errors of the global model under In-Distribution scenario.

V Conclusion

In this paper, we focus on addressing the generalization issue of deep learning MR image reconstruction models towards heterogeneous data captured by different scanning devices and imaging protocols, using multi-institutional data in a privacy-preserving manner. Specifically, we propose a generalizable federated neural architecture search framework, GAutoMRI. In the framework, we first design an adapted search space for undersampled MR image reconstruction task, and then learn an optimal model from heterogeneous data distributions with differentiable federated neural architecture search. In addition, we design a fairness adjustment approach that can enable the model to learn features fairly from inconsistent distributions of different devices. Extensive experiments on different sampling patterns, acceleration rates, contrasts, and unseen centers demonstrate that our proposed GAutoMRI has better generalization and performances compared to six state-of-the-art federated learning methods.

Acknowledgments

This research was partly supported by the National Natural Science Foundation of China (62222118, U22A2040), Guangdong Provincial Key Laboratory of Artificial Intelligence in Medical Image Analysis and Application (2022B1212010011), Shenzhen Science and Technology Program (RCYX20210706092104034, JCYJ20220531100213029), Key Laboratory for Magnetic Resonance and Multimodality Imaging of Guangdong Province (2023B1212060052), and the Major Key Project of Peng Cheng Laboratory (PCL2023AS1-2).

References

  • [1] Y. Korkmaz, S. U. Dar, M. Yurt, M. Özbey, and T. Cukur, “Unsupervised MRI reconstruction via zero-shot learned adversarial transformers,” IEEE Transactions on Medical Imaging, vol. 41, no. 7, pp. 1747–1763, 2022.
  • [2] C. M. Sandino, J. Y. Cheng, F. Chen, M. Mardani, J. M. Pauly, and S. S. Vasanawala, “Compressed sensing: From research to clinical practice with deep neural networks: Shortening scan times for magnetic resonance imaging,” IEEE signal processing magazine, vol. 37, no. 1, pp. 117–127, 2020.
  • [3] S. Wang, Z. Su, L. Ying, X. Peng, S. Zhu, F. Liang, D. Feng, and D. Liang, “Accelerating magnetic resonance imaging via deep learning,” in 2016 IEEE 13th international symposium on biomedical imaging (ISBI).   IEEE, 2016, pp. 514–517.
  • [4] K. Hammernik, T. Klatzer, E. Kobler, M. P. Recht, D. K. Sodickson, T. Pock, and F. Knoll, “Learning a variational network for reconstruction of accelerated MRI data,” Magnetic resonance in medicine, vol. 79, no. 6, pp. 3055–3071, 2018.
  • [5] M. Mardani, E. Gong, J. Y. Cheng, S. S. Vasanawala, G. Zaharchuk, L. Xing, and J. M. Pauly, “Deep generative adversarial neural networks for compressive sensing MRI,” IEEE transactions on medical imaging, vol. 38, no. 1, pp. 167–179, 2018.
  • [6] Y. Yang, J. Sun, H. Li, and Z. Xu, “ADMM-CSNet: A deep learning approach for image compressive sensing,” IEEE transactions on pattern analysis and machine intelligence, vol. 42, no. 3, pp. 521–538, 2018.
  • [7] T. M. Quan, T. Nguyen-Duc, and W.-K. Jeong, “Compressed sensing MRI reconstruction using a generative adversarial network with a cyclic loss,” IEEE transactions on medical imaging, vol. 37, no. 6, pp. 1488–1497, 2018.
  • [8] H. K. Aggarwal, M. P. Mani, and M. Jacob, “MoDL: Model-based deep learning architecture for inverse problems,” IEEE transactions on medical imaging, vol. 38, no. 2, pp. 394–405, 2018.
  • [9] B. Zhu, J. Z. Liu, S. F. Cauley, B. R. Rosen, and M. S. Rosen, “Image reconstruction by domain-transform manifold learning,” Nature, vol. 555, no. 7697, pp. 487–492, 2018.
  • [10] C. Qin, J. Schlemper, J. Caballero, A. N. Price, J. V. Hajnal, and D. Rueckert, “Convolutional recurrent neural networks for dynamic MR image reconstruction,” IEEE transactions on medical imaging, vol. 38, no. 1, pp. 280–290, 2018.
  • [11] Y. Han, J. Yoo, H. H. Kim, H. J. Shin, K. Sung, and J. C. Ye, “Deep learning with domain adaptation for accelerated projection-reconstruction MR,” Magnetic resonance in medicine, vol. 80, no. 3, pp. 1189–1205, 2018.
  • [12] M. Akçakaya, S. Moeller, S. Weingärtner, and K. Uğurbil, “Scan-specific robust artificial-neural-networks for k-space interpolation (RAKI) reconstruction: Database-free deep learning for fast imaging,” Magnetic resonance in medicine, vol. 81, no. 1, pp. 439–453, 2019.
  • [13] S. Wang, T. Xiao, Q. Liu, and H. Zheng, “Deep learning for fast MR imaging: a review for learning reconstruction from incomplete k-space data,” Biomedical Signal Processing and Control, vol. 68, p. 102579, 2021.
  • [14] D. Lee, J. Yoo, S. Tak, and J. C. Ye, “Deep residual learning for accelerated MRI using magnitude and phase networks,” IEEE Transactions on Biomedical Engineering, vol. 65, no. 9, pp. 1985–1995, 2018.
  • [15] S. Wang, H. Cheng, L. Ying, T. Xiao, Z. Ke, H. Zheng, and D. Liang, “DeepcomplexMRI: Exploiting deep residual network for fast parallel MR imaging with complex convolution,” Magnetic Resonance Imaging, vol. 68, pp. 136–147, 2020.
  • [16] T. Eo, Y. Jun, T. Kim, J. Jang, H.-J. Lee, and D. Hwang, “KIKI-net: cross-domain convolutional neural networks for reconstructing undersampled magnetic resonance images,” Magnetic resonance in medicine, vol. 80, no. 5, pp. 2188–2201, 2018.
  • [17] D. Liang, J. Cheng, Z. Ke, and L. Ying, “Deep magnetic resonance image reconstruction: Inverse problems meet neural networks,” IEEE Signal Processing Magazine, vol. 37, no. 1, pp. 141–151, 2020.
  • [18] G. A. Kaissis, M. R. Makowski, D. Rückert, and R. F. Braren, “Secure, privacy-preserving and federated machine learning in medical imaging,” Nature Machine Intelligence, vol. 2, no. 6, pp. 305–311, 2020.
  • [19] G. Kaissis, A. Ziller, J. Passerat-Palmbach, T. Ryffel, D. Usynin, A. Trask, I. Lima Jr, J. Mancuso, F. Jungmann, M.-M. Steinborn et al., “End-to-end privacy preserving deep learning on multi-institutional medical imaging,” Nature Machine Intelligence, vol. 3, no. 6, pp. 473–484, 2021.
  • [20] T. Li, A. K. Sahu, A. Talwalkar, and V. Smith, “Federated learning: Challenges, methods, and future directions,” IEEE signal processing magazine, vol. 37, no. 3, pp. 50–60, 2020.
  • [21] B. McMahan, E. Moore, D. Ramage, S. Hampson, and B. A. y Arcas, “Communication-efficient learning of deep networks from decentralized data,” in Artificial intelligence and statistics.   PMLR, 2017, pp. 1273–1282.
  • [22] T. Li, A. K. Sahu, M. Zaheer, M. Sanjabi, A. Talwalkar, and V. Smith, “Federated optimization in heterogeneous networks,” Proceedings of Machine learning and systems, vol. 2, pp. 429–450, 2020.
  • [23] Q. Li, B. He, and D. Song, “Model-contrastive federated learning,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 10 713–10 722.
  • [24] Q. Liu, C. Chen, J. Qin, Q. Dou, and P.-A. Heng, “Feddg: Federated domain generalization on medical image segmentation via episodic learning in continuous frequency space,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 1013–1023.
  • [25] J. Zhang, Y. Hua, H. Wang, T. Song, Z. Xue, R. Ma, and H. Guan, “Fedala: Adaptive local aggregation for personalized federated learning,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 37, no. 9, 2023, pp. 11 237–11 244.
  • [26] P. Guo, P. Wang, J. Zhou, S. Jiang, and V. M. Patel, “Multi-institutional collaborations for improving deep learning-based magnetic resonance image reconstruction using federated learning,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 2423–2432.
  • [27] C.-M. Feng, Y. Yan, S. Wang, Y. Xu, L. Shao, and H. Fu, “Specificity-preserving federated learning for MR image reconstruction,” IEEE Transactions on Medical Imaging, 2022.
  • [28] G. Elmas, S. U. Dar, Y. Korkmaz, E. Ceyani, B. Susam, M. Ozbey, S. Avestimehr, and T. Çukur, “Federated learning of generative image priors for MRI reconstruction,” IEEE Transactions on Medical Imaging, 2022.
  • [29] Z. Wang, X. Fan, J. Qi, C. Wen, C. Wang, and R. Yu, “Federated learning with fair averaging,” in Proceedings of International Joint Conference on Artificial Intelligence, 2021.
  • [30] T. Elsken, J. H. Metzen, and F. Hutter, “Neural architecture search: A survey,” The Journal of Machine Learning Research, vol. 20, no. 1, pp. 1997–2017, 2019.
  • [31] J. Yan, S. Chen, Y. Zhang, and X. Li, “Neural architecture search for compressed sensing magnetic resonance image reconstruction,” Computerized Medical Imaging and Graphics, vol. 85, p. 101784, 2020.
  • [32] Q. Huang, D. Yang, Y. Xian, P. Wu, J. Yi, H. Qu, and D. Metaxas, “Enhanced MRI reconstruction network using neural architecture search,” in Machine Learning in Medical Imaging: 11th International Workshop, MLMI 2020, Held in Conjunction with MICCAI 2020, Lima, Peru, October 4, 2020, Proceedings 11.   Springer, 2020, pp. 634–643.
  • [33] F. Knoll, J. Zbontar, A. Sriram, M. J. Muckley, M. Bruno, A. Defazio, M. Parente, K. J. Geras, J. Katsnelson, H. Chandarana et al., “fastMRI: A publicly available raw k-space and dicom dataset of knee images for accelerated MR image reconstruction using machine learning,” Radiology: Artificial Intelligence, vol. 2, no. 1, p. e190007, 2020.
  • [34] I. Bello, B. Zoph, V. Vasudevan, and Q. V. Le, “Neural optimizer search with reinforcement learning,” in International Conference on Machine Learning.   PMLR, 2017, pp. 459–468.
  • [35] H. Pham, M. Guan, B. Zoph, Q. Le, and J. Dean, “Efficient neural architecture search via parameters sharing,” in International conference on machine learning.   PMLR, 2018, pp. 4095–4104.
  • [36] Z. Yang, Y. Wang, X. Chen, B. Shi, C. Xu, C. Xu, Q. Tian, and C. Xu, “Cars: Continuous evolution for efficient neural architecture search,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 1829–1838.
  • [37] Y. Sun, B. Xue, M. Zhang, G. G. Yen, and J. Lv, “Automatically designing CNN architectures using the genetic algorithm for image classification,” IEEE transactions on cybernetics, vol. 50, no. 9, pp. 3840–3854, 2020.
  • [38] H. Liu, K. Simonyan, and Y. Yang, “DARTS: Differentiable architecture search,” in International Conference on Learning Representations, 2018.
  • [39] C. He, H. Ye, L. Shen, and T. Zhang, “Milenas: Efficient neural architecture search via mixed-level reformulation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 11 993–12 002.
  • [40] C. He, M. Annavaram, and S. Avestimehr, “Towards non-iid and invisible data with FedNAS: Federated deep learning via neural architecture search,” in CVPR 2020 workshop on neural architecture search and beyond for representation learning, 2020.