Abstract
Secure and trusted cross-platform knowledge sharing is significant for modern intelligent data analysis. To address the trade-off problems between privacy and utility in complex federated learning, a novel differentially private federated learning framework is proposed. First, the impact of data heterogeneity of participants on global model accuracy is analyzed quantitatively based on 1-Wasserstein distance. Then, we design a multilevel and multiparticipant dynamic allocation method of privacy budget to reduce the injected noise, and the utility can be improved efficiently. Finally, they are integrated, and a novel adaptive differentially private federated learning algorithm (A-DPFL) is designed. Comprehensive experiments on redefined non-I.I.D MNIST and CIFAR-10 datasets are conducted, and the results demonstrate the superiority of model accuracy, convergence, and robustness.
1. Introduction
Nowadays, artificial intelligence faces two main challenges: data silos and privacy concerns [1, 2]. Meanwhile, with the widely used edge and Internet of things devices, the federated learning framework attacks extensive attentions, which can train the global model based on the mode of decentralization and cooperation. However, there are still severe privacy protection problems in federated learning, although models are trained by sharing updates (such as gradient information) rather than raw data [3, 4]. Recent studies show that, by analyzing the parameter differences between training and uploading, privacy can still be leaked to a certain extent, such as the weight of neural network training [5, 6]. Federated learning requires different participants to upload and aggregate parameters repeatedly to train the global model. This process leads to more privacy information disclosure. Therefore, personal privacy can be obtained through the model-inversion attack.
Differential privacy (DP) provides a strict, quantifiable, and context-independent privacy protection method for machine learning. For its information theory guarantee, DP is also widely used to enhance data privacy with its simplicity and low cost [7–10]. However, as the fundamental challenge of privacy-preserved methods, DP mechanisms inevitably cause model performance degradation and system utility loss. Traditional differential privacy injects bounded noise into model to protect privacy, and privacy budget is a crucial factor in measuring the level of protection and the amount of noise. Existing differentially private methods allocate the same amount of privacy budget for each participant and each iteration in model updating, which leads to ubiquitous trade-off problems between privacy and utility [7, 9, 10]. This problem is mainly because federated learning is a complex training process with many participants and iterative updating, aggregation, and broadcasting. The guarantee of its privacy comes from adding a certain amount of noise at each communication. Therefore, the methods which allocate the privacy budgets equally will lead to severe utility damage.
In global model training, allocating privacy budgets for different participants and iterative process is an essential means of utility optimization under certain privacy protection levels. However, how to allocate the privacy budget is a crucial issue. At the participants’ selected level, due to personalized use and collection of local data for each participant, the assumption of independent and identically distributed (I.I.D) of underlying data may not exist. Nonindependent and identically distributed (non-I.I.D) makes of data heterogeneity for federated learning. Data heterogeneity, also called statistical heterogeneity, refers to the heterogeneity of the underlying data distribution for various participants [1, 2, 9]. The existing theoretical research only concentrates on proofing the convergence and privacy protection level. The analysis for the internal relationship between the utility of federated learning and data heterogeneity is insufficient. Therefore, it is a sort of theoretical basis for privacy budget allocation that utility improvement oriented. At the iterative training process level, for each participant, with the convergence of the local model, the parameter update speed slows down, and adding the same amount of noise will have a significant impact on the convergence performance. In addition, at a certain level of global privacy, adjusting privacy budgets in different training stages will improve accuracy. In general practice, a similar strategy is applied to deep learning, which reduces the learning rate as the training goes on, rather than using a constant one. Therefore, in model convergence, adding gradually reduced noise is an effective way to improve the performance of a federated learning system.
Works on the trade-off between privacy and utility in federated model training become the research hot spot, and many state-of-the-art differentially private algorithms have been proposed [11–17]. Geyer et al. [11] proposed a DP-SGD algorithm. It analyzed the convergence performance boundary, which showed that the privacy protected level and data size affected the convergence performance. Zhang et al. [12] improved DP-SGD by tracking the privacy loss. It obtained an accurate estimation of the overall privacy loss. Kerkouche et al. [13] proposed an effective federated learning algorithm for privacy protection under a given privacy level and plenty of customers involved. Wei et al. [14] made theoretical research on differential privacy level and convergence performance. However, it does not study the allocation strategy. Authors of [15–17] proposed personalized privacy budget allocation methods to improve the accuracy of the federated learning according to the different contributions of participants to the global accuracy under a scenario of unbalanced data. Although these methods are promising, there remain several issues of the method design. First, these methods only consider the privacy budget allocation technology for different participants, but they do not consider the privacy budget cost on the process of the global model iteration. Second, inevitable existing data heterogeneity in real federated scenarios brings new challenges to the theoretical analysis. The existing research is limited to the research of technical methods, and there is a lack of method design based on theoretical research.
In order to solve problems of utility optimization in federated learning, a novel differentially private federated learning framework is proposed in this paper. The impact of data heterogeneity of participants on global model accuracy is first analyzed. Then, a multilevel and multiparticipant privacy budget dynamic allocation method is designed. Based on the proposed methods, a novel adaptive differentially private federated learning algorithm (A-DPFL) is also devised to balance privacy and utility. The main contributions of this paper are three folds:(1)We quantitatively analyze the impact of the participants’ data heterogeneity on the global model accuracy based on 1-Wasserstein distance, and then, we propose an appropriate participant evaluation method for participants’ selection in federated learning to improve the system’s effectiveness.(2)A multilevel and multiparticipant privacy budget dynamic allocation method is designed. At the global iteration process, the noise scale adjustment method is proposed to inject less noise in pace with the convergence of the global model, and the global optimal accuracy is gained at a local optimum position. Moreover, in each training iteration, participants’ active selection is achieved to obtain better utility.(3)A new federated learning algorithm with differential privacy is proposed. It dynamically allocates the privacy budget for multiple participants and multiple iterations in view of the data heterogeneity of participants in a federated system. It can improve system utility while keeping privacy and adaptively balance privacy and utility.
2. Related Works
2.1. Utility Optimization of Federated Learning
Many works have been done to improve the utility of global models in federated learning, such as global accuracy, convergence property, time consumption, and robustness. These studies are mainly based on the following different assumptions: number of participants, data heterogeneity, communication properties, the properties of loss functions, and the bound of gradient noise. In the device-based federated learning system, some personalized federated learning studies are derived for system heterogeneity [18–24]. They involve the optimization objectives of federated systems such as distributed mobile user interaction, communication costs in mass distribution, unbalanced data distribution, and device reliability. The key directions of this kind of research can be classified as asynchronous communication [18], scheduling policy [21], and model compression [24].
Federated learning is generally based on gradient information, such as gradient descent (GD) and stochastic gradient descent (SGD). Among them, the most used is the federated average (FedAvg) [23], which updated the average of local stochastic gradient descent. The convergence of FedAvg performs well when data follows the I.I.D hypothesis. However, research shows that even in a classical federated scenario, the underlying data does not necessarily conform to the I.I.D. It demonstrates that data heterogeneity may lead to model divergence. Some distributed gradient descent algorithms (Dist. GD) and related variants of distributed stochastic gradient descent (Dist. SGD) have been proposed, which achieve local updates similar to the federated average. For data heterogeneity, FedProx was proposed recently [24]. FedProx made a slight modification to FedAvg to ensure convergence in both theory and practice. Many advanced optimization methods are also available for federated system optimization [25–28].
The above studies achieve good results in optimizing the federated algorithm, but the algorithms only concern one or two utility indicators. Nevertheless, the method proposed in this paper boosts the possibility of improving more comprehensive utility indicators through the active selection of participants. In addition, these existing algorithms have insufficient research on privacy issues. Besides the utility loss caused by federated model training, privacy protection also brings accuracy loss and convergence performance decline additionally. At last, since the utility optimization will be different when performing privacy protection, the theoretical research on the utility optimization of privacy-preserved federated learning is still insufficient.
2.2. Privacy-Preserved Federated Leaning
A series of theoretical and methodological research on privacy-preserved federated learning has been conducted, and many research results have been obtained. Privacy in federated learning can be divided into global differential privacy [14] (GDP) and local differential privacy (LDP) [7]. Global privacy requires that the model updates generated in each round be privacy-protected for all untrusted third parties except the central server. In contrast, local privacy further requires that the update be privacy-protected for the server. The privacy protected method research in federated learning continues and extends in traditional machine learning, which are mainly based on multiparty secure computing [29, 30] and differential privacy [7, 31, 32]. Among them, multiparty secure computing is a kind of the lossless method, which can maintain the original accuracy and make a strong privacy guarantee. However, the approach results significant additional communication costs. Considering the high communication cost in the federated system, differential privacy has low system overhead. Existing studies include federated learning algorithms that satisfy LDP [7], differentially private stochastic gradient descent algorithm (DP-SGD) [31], and metalearning with DP [33].
However, as the fundamental challenge of privacy-preserved methods, DP inevitably causes utility loss. Focusing on the trade-off between privacy and utility in the model training process, many state-of-the-art differentially private federated algorithms have been proposed [11–17]. There are still several issues of the method design. First, these methods only consider the privacy budget allocation for different participants, but they do not consider the privacy budget cost on the process of the global model iteration. Second, inevitable data heterogeneity in real federated scenarios brings new challenges on both the theoretical analysis and pro. The existing research is limited to technical methods, and there lacks theoretical research.
3. Preliminary
3.1. Differential Privacy
Differential privacy (DP) designs a mechanism to add noise to the target dataset so that the statistical information loss of the released dataset and the original dataset is in a small range, which ensures that the modification of an individual record in the dataset does not have a significant impact on the statistical results.
Definition 1. (differential privacy). For any dataset and differing on at most one record and for any possible sanitized dataset , a random mechanism satisfies -differential privacy ifwhere refers to the privacy budget that controls the privacy level of the mechanism . The lower represents the higher privacy level.
Definition 2. (global sensitivity). For any function , for all and differing in at most one record, the global sensitivity of is
Definition 3. (mechanism). Mechanism is associated with the global sensitivity. This measures the maximal change on the result of query when removing one record from the dataset . Let be a database, where a randomized function is a (randomized) perturbation mechanism on , if the output follows a conditional distribution .
Definition 4. (Gaussian mechanism). A privacy mechanism guarantees -differential privacy if, for any two adjacent dataset and , the following inequality holds:Gaussian mechanism is used to guarantee -differential privacy. We present the following DP mechanism by adding artificial Gaussian noise , where .
3.2. Federated Learning
The concept of federated learning was first proposed by Google [34]. It is an algorithm framework for building machine learning models, which generally includes the following features: two or more participants jointly train a global model; each participant has some local data that can be used to train the global model; during the global model training, data are retained locally by participants; during local parameter communicating, privacy protection can be used to prevent privacy leakage; the accuracy of the federated model is an optimal approximation of the accuracy of the ideal model constructed from the centralized data. There are two implicit concerns, include the utility of the global model and privacy issues in the communication process.
In application scenarios, federated learning can be divided into horizontal federated learning, vertical federated learning, and federated transfer learning. Among them, vertical and transfer federated learning devote to train different local models for aggregation. These kinds of research focus on multitask learning and metalearning, and this is different from classical federated learning in nature [35]. In a classical federated learning scenario, samples from different participants are introduced into the same feature space shared by the dataset, called horizontal federated learning, also known as sample-based federated learning. Figure 1 illustrates a classical federated learning system, in which participants can be organizations, nodes and clients, and terminals or devices in the Internet of Things according to different application scenarios.

4. Problem Setting
In this paper, we take the horizontal federated learning system as the research object, so vertical federated learning and federated transfer learning can be regarded as exceptional cases of horizontal federated learning. Aiming at the problem of decreasing system utility caused by the privacy protection method and the new challenges in theoretical analysis and the method design caused by the data heterogeneity of federated learning, we focus on how to inject an appropriate differential privacy noise in the global model training to achieve privacy protection and optimize utility through Collaborator’s privacy budget dynamic allocation. In this section, we first formalize the federated learning system (Section 4.1) and then analyze the possible attack in the federated system (Section 4.2), and finally, a global differential privacy model is designed to solve the privacy issues setting above (Section 4.3).
4.1. Formalization of Federated Learning
We consider a federated learning system, as shown in Figure 2, which trains the global model by introducing different samples from local datasets into the common feature space, namely, . The system consists of a trusted collaborator and participants. represents local dataset of participant . Defining as the loss function of participant , federated learning can be formalized as the following optimization problem:

In order to optimize problem 4, the federated learning process of training a global model is designed as follows:(1)Local model training: the participants conduct a round of parameters’ training according to the local data and upload the local parameters of the current round to the collaborator(2)Secure aggregating: the collaborator performs secure aggregation for the current round of local parameters uploaded by the participants(3)Global parameter broadcasting: the collaborator broadcasts the aggregation parameters of the current round to the participants(4)Local model updating: the participants update the model according to the global parameters of the current round and conduct the next round of training
4.2. Threat Model
This paper analyzes and assumes the following two possible attacks based on the above-federated learning scenario. The collaborator (server) is assumed to be honest; however, external adversaries aim to obtain participants’ private information, as demonstrated by model-inversion attacks. Moreover, the participants are assumed to be semihonest; adversaries can perform poison attacks on the federated system by uploading malicious parameters to interfere with the aggregation of the global model.
4.3. Global Differential Privacy Model
According to the possible threat model and solving the optimization problem formalized mentioned above, we design a differentially private federated learning framework that satisfies global differential privacy for the federated system and optimizes the utility by dynamic allocating differential privacy budgets, as shown in Figure 2. By injecting Gaussian noise to local parameters, the designed federated system achieves -differential privacy. Furthermore, the honest collaborator conducts the privacy budgets dynamic allocation according to the utility performance of participants (data heterogeneity) and iteration process. Meanwhile, if the participant’s evaluation is below a certain threshold, it signifies the participant may be an adversary, and the collaborator refuses to update its parameters against poison attacks.
5. Utility Optimization of DPFL Framework
5.1. Framework Outline
In the paper, a novel differentially private federated learning framework is proposed to provide global differential privacy and optimize the system utility for the federated system. Including this framework, the participants’ evaluation method is proposed based on analyzing the impact of the data heterogeneity of participants on the global model accuracy, a multilevel and multiparticipant privacy budget dynamic allocation method is designed, and a novel adaptive differentially private federated learning algorithm (A-DPFL) is also devised to balance the privacy and utility. The research framework proposed in this paper mainly proposes an appropriate solution to the problem of equal allocation of privacy budget by existing methods and designs a privacy allocation scheme for multiple participants and multiple iterations in complex federal systems. It mainly includes participant selection in a federated system with non-I.I.D and dynamic adjustment of privacy parameters in global iteration. The following is an explanation of the scheme design at each level.
Due to the general data heterogeneity in the complex federated system and the significant impact of data heterogeneity on the accuracy and convergence performance of the global model [18, 22, 24], the participants make a significant difference in the contributions of the global training model. Therefore, participants’ active selection is an effective means to improve the utility. Allocating privacy budgets for the selected participants will improve the utility under the premise of a privacy guarantee. Our method quantitatively analyzes the impact of the participant on the global model accuracy based on 1-Wasserstein. It proposes an appropriate participant evaluation method used as the basis for participants selection to improve the system’s utility. We design a general approach to participants’ selection. We take improving global accuracy as an example to expand the description in this paper. In fact, we can also choose participants with better convergence performance or other utility indicators to train the global model.
At present, federated learning with differential privacy is realized by adding quantitative random noise to the parameters in each iteration to protect privacy. However, as the model converges, the parameter update speed slows down, and adding the same amount of noise will significantly impact the final convergence of the model. Moreover, one main concern of model training is the accuracy achieve under a certain private level. Maintaining certain global private levels and reducing the noise injected as the iteration process will achieve higher accuracy. The practice of general deep learning applies a similar strategy to achieve higher accuracy, which reduces the learning rate as the training progresses instead of using a constant learning rate in all periods [36, 37]. Therefore, to approach the local optimum better and achieve higher model accuracy, gradually adding reduced noise during the model convergence is an effective means to improve the utility of the federated learning system. We design a privacy budget dynamic allocation method that adjusts the noise scale dynamically as the model converges so that less noise is injected to achieve better model performance under a certain level of privacy.
In the following section, we conduct a theoretical analysis of participants’ statistical heterogeneity in the federated system (Section 5.2) and then design a multilevel and multiparticipant privacy budget dynamic allocation method (Section 5.3) to allocate privacy budget adaptively in the federated learning training process. Finally, we propose an adaptive differentially private federated learning algorithm (Section 5.4) to optimize the utility of federated learning. The designed framework achieves a good balance between privacy and utility for federated learning by these necessary means of the framework.
5.2. Data Heterogeneity Analysis of Participants
To theoretically analyze the participants’ data heterogeneity in federated learning, we firstly declare the federated system’s data heterogeneity, then quantitatively analyze the effect of participant data heterogeneity on the global accuracy, and finally propose the participants’ evaluation method.
5.2.1. Data Heterogeneity Declaration
Many factors cause non-I.I.D; among them, most extensive research focuses on label distribution skew [38, 39]. Since participants are associated with specific geographic areas, the label distribution of local data is quite different from each other. For example, kangaroos are only found in Australia or zoos, and each specific face only appears in a few places in the world. Therefore, we declare data heterogeneity in federated learning as the labels skewed distribution of different participants and as
If formalizing the federated learning as a supervised learning, we have the expected loss under the local data distribution of the participant :where the loss function is used to measure the error of local model parameters judging the true label based on the given input . is the probability distribution of the local sample space , and data heterogeneity is manifested as a skewed distribution of . More intuitively, assuming that the value of the classification problem is discrete, the participant’s expected loss can be formalized as
5.2.2. Quantitative Analysis between Data Heterogeneity and Global Accuracy
According to the setting of the research scope, the participants of the federated system in this paper have the same loss function, which is recorded as
Then, the loss function of the participant can be recorded as
According to the formalized definition of the research problem, that is, equation (7), it can be concluded that the global accuracy is closely related to the similarity between the local distribution and the global distribution. 1-Wasserstein distance is one of the effective methods to measure the similarity of two probability distributions. When is distance function, is the set of measures of the marginal distributions and in the space . The 1-Wasserstein distance of the two probability distributions and in the matrix space can be defined in
Therefore, assuming that, for any , has continuity and smoothness, we quantitatively analyze and prove the influence of the 1-Wasserstein distance between the local data distribution and the global data distribution on the global accuracy, which can be summarized and simply proofed as
5.2.3. Participants’ Evaluation Method
Based on the above theoretical research on the impact of participants’ local distribution on global accuracy, we propose the following methods for evaluating participant data heterogeneity. The participant evaluation method based on 1-Wasserstein distance is evaluating the data heterogeneity of participants by solving equation (12). The value is the average degree of data heterogeneity for all participants. Then, comparing the values of , is defined as evaluation score :
5.3. Dynamic Allocation Method of Privacy Budgets
Since the complex federated learning differs from the previous machine learning, we design a multilevel and multiparticipant privacy budget dynamic allocation method to improve the utility. It reduces noise intake from two levels and improves the utility of the federated system. First, in one round of the iterative process, according to the heterogeneity evaluation of the participants, that is, the degree of contribution to the global model, it allocates privacy budget for selected participants. Second, according to the degree of convergence in the global iteration process, it allocates the privacy budget for global iteration by dynamically adjusting the noise scale.
5.3.1. Participants Active Selection
In practice, it is unrealistic for global participants in the federated learning system to participate in each iteration. In theory, at the expense of system time utility and robustness, increasing the number of participants can achieve better global accuracy. However, in one round of iteration, the introduction of differential privacy protection operation makes greater amount of noise with more participants, resulting in a decrease in global accuracy. Therefore, the number of participants of each iteration has a significant impact on the utility of the federated system. Meanwhile, according to the data heterogeneity of participants, selecting different participants to participate in each round of iteration will have a different impact on the system’s utility. Based on the above analysis, we design a participants’ active selection method that actively allocates the privacy budget to the selected participants to reduce global noise intake and obtain better utility.
The specific method of allocating privacy budgets for participants in one round of iteration is based on the participant’s probability of being selected. The probability of participant being selected is denoted as , and is calculated according to evaluation result . The probability of global participants satisfies
According to the ranking of the probability of being selected, the decision of the participants in round is made, see equation (14). Then, the participants are allocated privacy budgets for this round to achieve privacy protection:
5.3.2. Noise Scale Dynamic Adjustment
Theoretically, when all global participants participate in training, the convergence performance of the federated system is related to the number of iterations and the noise level. As increases, the convergence performance improves, but the accumulated noise level also increases, which leads the convergence performance decrease. Therefore, this paper designs a privacy budget allocation method, i.e., noise scale dynamic adjustment method; under a certain global privacy level, adjusting the noise scale dynamically as each round of convergence changing and reducing the noise scale dynamically by verifying the accuracy of the current model. We can use the calculating average of the local accuracy as the verification accuracy. Another way is to use a public dataset accessible to the server, monitor the accuracy of the model, and reduce the noise scale when the accuracy stops increasing. Every time the increase in verification accuracy is lower than the threshold, the noise scale is reduced by times, until the total privacy budget is exhausted. Here, is the noise scale adjusting parameter. The iterative round of performing the inspection mentioned above operations is called dynamic adjustment round. When the accuracy improvement between two dynamic adjustment rounds is less than the threshold , the server will reduce the noise parameter to . If the accuracy increase is less than the threshold , the noise scale remains unchanged. Subsequently, the updated noise scale will be used in the subsequent training until the next dynamic adjustment round is reached. The noise scale update equation is as follows:where and .
5.4. Adaptive Differentially Private Federated Learning
In this section, an adaptive differentially private federated learning algorithm (A-DPFL) is proposed combining the above theories and methods. By injecting Gaussian noise to the local weights to provide global differential privacy, it aggregates the updating private local weights for federated global model training. The multilevel and multiparticipant privacy budget dynamic allocation method is performed in the traditional federated learning process to reduce the noise injected from the perspectives of participants selecting and noise scale adjusting, thereby improving the utility. Algorithm 1 presents the detail of A-DPFL.
|
|
|
|
The main steps are described as follows. Comparing with the traditional process of DPFL, after initialization, a participant selection step is added, and after several rounds of aggregation, noise scale adjustment step is carried out.(1)Initialization: the collaborator initializes global model parameters such as learning rate, batch size, and noise parameter.(2)Participant selection: the collaborator evaluates the available participants and, according to the optimal indicator, decides who is selected to join in the current training round of federated learning.(3)Parameter broadcast: the collaborator broadcasts the global model and the noise parameters of the current round to the selected participants. The participants synchronously update the local model as the global one.(4)Local model update: the participant privately and locally performs a round of model training by adding random Gaussian noise to the local weights based on the local data.(5)Local models upload and aggregation: the selected participants upload the private local model to the collaborator. After that, the collaborator aggregates the received weights as a new global model.(6)Noise scale adjustment: the collaborator obtains verification accuracy to verify the global model convergence situation and decides whether adjusting the noise parameter or not in the current predetermined dynamic adjustment round. If so, an adjusted noise parameter will be broadcast in the next training round.
Algorithm 1 includes four significant steps: participant selection, local model update, local models’ upload and aggregation, and noise scale adjustment. The following is a detailed description of these steps.
5.4.1. Participant Selection
Algorithm 2 presents the participant active selection process.
In the participant selection step, the participant evaluation method described in Section 5.2.3 is utilized to determine the contribution to the federated model training. Equation (12) shows the detail of evaluation. We calculate the probability of participant i being selected according to its evaluation result and sort them and select the participants whose probability of being selected is higher than a certain probability threshold and add them to current round of iteration.
5.4.2. Local Model Update
In the participant local model update step, each participant selected in the iteration will train a local model for a specified number of times to obtain the updated local model. In A-DPFL, DP-SGD is utilized, and Gaussian noise is added to grantee privacy. Algorithm 3 presents the process in detail.
According to DP-SGD, it adjusts the model weights to minimize the error function of the model. We also adopt the idea of the minibatch gradient descent method and randomly extract a batch of data with the number of B from all the local data of the participants for model training. The gradient firstly is calculated based on the training local data with its label :where and represents the gradient of the loss function calculated by training data to the weight in the th iteration.
Gaussian noise is added to the gradient calculated by the data in the current batch, and then, the average gradient of the current batch is calculated:
where represents Gaussian noise with mean 0 and variance . is the key parameter to control the noise scale. The greater the , the greater the variance of the normal distribution and the greater the amount of Gaussian noise.
Following adding noise, the gradient is clipped, as shown in equation (18). Gradient clipping ensures that the second norm of the gradient does not exceed the clipping threshold . If , the value of will not change. If , the value of will be reduced to . Gradient clipping is a common step in deep learning, which can effectively prevent the phenomenon of gradient explosion that results in poor model convergence performance:
Finally, we have the processed gradient and updated the model according to the learning rate to obtain the local parameters after training:
5.4.3. Local Models’ Upload and Aggregation
The collaborator receives the local weight uploaded by each participant and aggregates to update the global model as equation (20). In the step, FedAvg is considered to aggregate local weights as a new global model:
5.4.4. Noise Scale Adjustment
In this step, the noise scale dynamic adjustment method presented in Section 5.3.2 is utilized to dynamically adjust the noise scale . In the dynamic adjustment round, the collaborator obtains verification accuracy and compares with the verification accuracy of the last dynamic adjustment round. If the difference is less than the set threshold , it can be considered that the current global model is gradually converging. Therefore, to achieve better model accuracy, it is necessary to reduce the allocation of privacy budget by decreasing noise parameter times. If the difference is greater than the threshold, will remain unchanged and continue training with current noise scale. The algorithm description is shown in Algorithm 4.
6. Evaluation
In this section, we evaluate the performance of our method and against data heterogeneity and different parameters. We first compare convergence properties and the accuracy properties between our method and classical method and then study the impact of data heterogeneity and the trade-off between accuracy and privacy.
6.1. Experimental Setting
This paper used TensorFlow-Federated, a federated learning library in TensorFlow. We develop the algorithm secondarily based on Python and built a real federated learning environment by deploying the algorithm to multiple edge devices. In the experimental process, two classical datasets commonly used in deep learning, the MNIST dataset [40] and CIFAR-10 dataset [41], were chosen for the experiments:(i)MNIST: the dataset contains 70,000 gray-scale images of handwritten digits, divided into 60,000 training images and 10,000 test images. The size of each image is 28 ∗ 28, and the labels are numbers between 0 and 9.(ii)CIFAR-10: this dataset contains ten categories of RGB color images, divided into 50,000 training images and 10,000 test images. The size of each image is 32 ∗ 32. The images have the following ten categories: airplane, car, bird, cat, deer, dog, frog, horse, boat, and truck.
To simulate the characteristics of participants’ data heterogeneity in the real environment, we performed the non-I.I.D setting common in the bottom data of federated learning, i.e., the label skewed distribution setting. The skewness of data labels refers to the inhomogeneous degree in the division of labels among different participants. In the I.I.D setting, the data distribution of different labels is uniformly distributed, while in the non-I.I.D setting of label skewness, some labels are present only in some participants. Because partitions are tied to particular geo-regions, the distribution of labels varies across partitions. For example, a 20% skewness means that 20% of the participants’ data are divided by labels and 80% of the data are randomly selected with different labels. In contrast, a 100% skewness means that all the participants’ data are of the same label. We first preprocessed all the training data that were read and assigned data of different label skewness to all participants during the experiments. Taking the MNIST dataset as an example, the data distribution of the participants is shown in Figure 3. Participant 1 has 80% label skewness and 80% of the training data labels are 1; participant 2 has 50% label skewness and half of the training data labels are 4; participant has 0 label skewness and the data labels it owns are more uniformly distributed.

Due to convolutional neural networks’ excellent feature extraction ability, convolutional neural networks were used as the structure of the neural network of the participant’s local model in the experiments. In the experiments for both datasets, the convolutional neural network with two convolutional layers was used, and its specific structure was set up as follows: a convolutional layer with the convolutional kernel of 5 ∗ 5, the output channel number of 32, and a pooling layer of 2 ∗ 2 and next was a fully connected layer with 1024 nodes, ReLU as the activation function, and an output layer with 10 nodes.
Some of the parameters used in the experiments are shown in Table 1. The CIFAR-10 dataset is an RGB three-channel color image with a larger total data volume compared to MNIST. Due to hardware limitations, the number of participants was set to a smaller number of 100 in the experiments on the CIFAR-10 dataset. In training on the CIFAR-10 dataset, there were more model parameters. In order to make the gradient after gradient cropping keeps the same direction as the original gradient as much as possible, the gradient cropping threshold needed to be set to a larger value. In this paper, the gradient cropping threshold was set to 7.0 in the experiments on the CIFAR-10 dataset.
6.2. Utility evaluation
In order to verify the global model accuracy and convergence performance of the algorithm designed in this paper, the comparison experiments were conducted to compare the designed algorithm (A-DPFL) with the classical differentially private federated learning optimization algorithm (CL-FL) [11], and the differentially private federated learning algorithm without optimization (DPFL). Also, a nonprivate federated learning algorithm was included to observe the effect of the differential privacy operation on the accuracy. The comprehensive experiments were conducted on the algorithm using MNIST and CIFAR-10 datasets to verify the effectiveness of the designed algorithm on different image classification datasets. Figures 4(a) and 4(b) present the variation of global model accuracy with increasing iteration rounds on the MNIST dataset and CIFAR-10.

(a)

(b)
The experimental results on the two datasets show that the nonprivate federated learning algorithm has the highest global accuracy on different datasets, and the global accuracy increases more steadily during the convergence process. It indicates that the privacy-protected operation will have some impact on the federated system performance. The global model accuracies of the remaining private algorithms are lower than that of the nonprivate algorithms, and the global model accuracy fluctuates to some extent during the convergence process because of adding random noise. Federated learning for privacy protection optimization algorithms, such as the classical CL-FL algorithm, will improve the system utility to some extent due to the personalized allocation of the privacy budget considered. In contrast, the traditional DPFL algorithm differs significantly from the optimization algorithm in terms of global accuracy and convergence performance. However, the convergence performance (convergence speed) and the global model accuracy of our method (A-DPFL) outperform CL-FL on both datasets due to the dynamic allocation method of privacy budget designed. It autonomously selects participants with higher model accuracy contribution and reduces the noise intake as training proceeds, thus enhancing the utility of federated learning from two levels.
6.3. Impact of Data Heterogeneity
When performing federated learning, it is often desirable that each participant has I.I.D local data so that each participant’s local model can be adequately trained to obtain better global model accuracy after aggregation. However, there is usually non-I.I.D of local data due to the differences in device usage. In this set of experiments, we simulated the case of label skewed distribution of the participants’ data and observed the difference of the global accuracy under different data heterogeneity settings. Table 2 shows the comparison of the global accuracy of A-DPFL on two datasets.
As shown explicitly, the global model accuracy still retains high usability in data label skewed distribution, although there is some degradation compared to the ideal case. The experiment illustrates that data heterogeneity impacts the training accuracy and verifies that our method has good robustness in a data heterogeneity practical scenario.
6.4. Trade-Off between Accuracy and Privacy
The necessary means of our method that significantly impacts the global model utility enhancement and privacy utility balance is the dynamic adjustment of the noise scale. In this section, we design experiments to show the changes in the accuracy enhancement with the convergence process and the sensitivity of the global accuracy to the noise to verify the effectiveness of our method.
6.4.1. Noise Scale Dynamic Adjustment Method Evaluation
Firstly, to verify the noise scale dynamic adjustment method proposed in this paper, the change of noise scale during the training process of the designed algorithm using the MNIST dataset is observed in this paper. The trend of noise scale with the increase of training rounds is shown in Figure 5.

It can be observed that the time to keep the noise scale constant in each round decreases as the training proceeds. In the beginning, is kept at 1.0 for 20 training rounds; then, it is kept at 0.7 for 4 rounds, 0.49 for 4 rounds, and 0.34 for 7 rounds, and the following holding time is gradually shortened. The reason for this phenomenon is that, as the global model converges, the speed of accuracy rise of the global model on the validation dataset gradually decreases, and the amount of validation accuracy rise does not exceed the given threshold for more time, resulting in a frequent adjustment of the noise scale . At the same time, as the noise scale decreases faster, the noise intake decreases during training. The experiment indicates that the global model can obtain better global model accuracy in the process of convergence to the local optimum, proving the effectiveness of the method designed in this paper.
6.4.2. Impact of Noise Scale Adjusting Parameter
In our method, the noise scale adjusting parameter controls the noise variation and effects on global accuracy. In noise scale dynamic adjustment method design, determines the speed of noise size reduction with training rounds, so weighs the level of privacy protection against the global model accuracy. In the last experiments, we measured the accuracy of the global model corresponding to , keeping other parameters’ constant. This paper conducted experiments with ranging from 0.3 to 0.9 and observed the trend of global model accuracy with on the MNIST dataset. The specific results are shown in Figure 6.

As we can observe from the figure, the global model accuracy decreases as increases. An increase in represents a slow reduction in the noise scale, which means that more noise is added during the training process, causing a decrease in the global model accuracy. However, the decrease of will make the noise decrease too fast and reduce the privacy level. Through the experiments, we can find that the global model accuracy decreases slowly when the value of is in the range of 0.3–0.7, but the decreasing trend speeds up significantly when it is in the range of 0.8–0.9. According to the above analysis, is a more suitable value to balance privacy protection and global model accuracy. In other experiments of this paper, the value of is also adopted.
7. Conclusion and Future Works
This study makes research on utility optimization of federated learning with differential privacy. A novel framework is designed for solving the inherent problem of balancing between privacy and utility in privacy-preserved machine learning. In order to deal with utility damage through equal allocation to each participant and iteration in federated learning, a multilevel and multiparticipant privacy budgets’ dynamic allocation method based on theoretical analysis of data heterogeneity is proposed. Implementing the framework of this study, a novel adaptive differentially private federated learning algorithm (A-DPFL) is also devised, and comprehensive experiments are conducted to demonstrate the improvement of the global accuracy and convergence speed and the robustness in non-I.I.D scenario on the MNIST and CIFAR-10 dataset.
Our future works will focus on these aspects. (1) More comparison algorithms should be added to verify the method’s effectiveness. (2) In the complex federated learning, due to devices’ performance deference, system heterogeneity should be considered to improve the system robustness.
Data Availability
The mnist and cifar-10 datasets used to support the findings of this study are available at tensorflow/g3doc/tutorials/mnist/ and http://www.cs.toronto.edu/∼kriz/cifar-10-python.tar.gz.
Conflicts of Interest
The authors declare no conflicts of interest.
Acknowledgments
This work was supported by the National Natural Science Foundation of China under Grant No. 62102074, the Natural Science Foundation of Liaoning Province of China under Grant No. 2020-MS-091, and the Fundamental Research Funds for the Central Universities of China under Grant No. N2017015.