• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    ADC-DL:Communication-Efficient Distributed Learning with Hierarchical Clustering and Adaptive Dataset Condensation

    2022-12-09 09:49:54ZhipengGaoYanYangChenZhaoZijiaMo
    China Communications 2022年12期

    Zhipeng Gao,Yan Yang,Chen Zhao,Zijia Mo

    State Key Laboratory of Networking and Switching Technology,Beijing University of Posts and Telecommunications,Beijing 100876,China

    Abstract: The rapid growth of modern mobile devices leads to a large number of distributed data,which is extremely valuable for learning models.Unfortunately,model training by collecting all these original data to a centralized cloud server is not applicable due to data privacy and communication costs concerns,hindering artificial intelligence from empowering mobile devices.Moreover,these data are not identically and independently distributed (Non-IID)caused by their different context,which will deteriorate the performance of the model.To address these issues,we propose a novel Distributed Learning algorithm based on hierarchical clustering and Adaptive Dataset Condensation,named ADC-DL,which learns a shared model by collecting the synthetic samples generated on each device.To tackle the heterogeneity of data distribution,we propose an entropy topsis comprehensive tiering model for hierarchical clustering,which distinguishes clients in terms of their data characteristics.Subsequently,synthetic dummy samples are generated based on the hierarchical structure utilizing adaptive dataset condensation.The procedure of dataset condensation can be adjusted adaptively according to the tier of the client.Extensive experiments demonstrate that the performance of our ADC-DL is more outstanding in prediction accuracy and communication costs compared with existing algorithms.

    Keywords: distributed learning;Non-IID data partition;hierarchical clustering;adaptive dataset condensation

    I.INTRODUCTION

    With the dramatic development of the Internet of Things(IoT)in recent years,many sensor-driven intelligent systems such as smart home,smart transportation,smart building[1]are becoming more and more prevalent in our lives.These intelligent systems are superior to humans in terms of safety,timeliness and energy conservation,and have substantially improved the productivity and convenience of society [2].In most intelligent systems,machine learning is desired to be implemented to train a massive amount of data from multiple data sources.However,data deposited in various devices and locations is constrained in sharing with servers or other devices due to privacy security and expensive communication costs.[3].Therefore,there is a tendency to explore solutions for implementing machine learning into distributed data scenarios with these limitations.

    There are two predominant methods that empower clients to securely share their local information[4].In the first approach represented by federated learning[5],data sources share the models or gradients learned on local data.In the other approach represented by coreset construction[4,6],data sources share the local summaries of their original data,which are aggregated to train a global model.Neither of these two methods requires the clients to share their data in a direct manner[7].Although extensive previous research on these two approaches has achieved data security and reduction of communication costs to a certain extent,there are still many deficiencies for further investigation.For federated learning,studies have demonstrated that the publicly shared gradients will leak information of the original data[8,9],which renders it inappropriate for frequent transmission of the enormous parameters in modern deep neural networks.In previous studies of the second approach,coreset is the most pervasive method for generating local summaries.However,such methods have the following vulnerabilities:1) they rely on the selections of the center of clustering and the presence of representative samples,which does not guarantee that the generated local summaries are sufficiently valid.2) some of these methods are less effective when the local datasets in clients are heterogeneous as they blindly determine the local centers[4].

    In terms of accuracy,communication overhead and data privacy,each of these two approaches has its own strengths.Generally,federated learning can achieve slightly better accuracy than coreset-based learning.However,modern neural networks are usually of relatively huge parameter size.Their frequent transmission will result in a massive communication overhead.Therefore,coreset-based learning significantly outperforms federated learning in communication overhead.In terms of security,a study has been conducted to measure the degree of privacy leakage of two methods by membership Inference attack [7],which indicates that when we need a target model of the maximum accuracy,federated learning is more desirable in terms of privacy protection.However,if we can tolerate a mild loss of target model accuracy,coreset-based learning becomes preferable because it can achieve the same accuracy as federated learning with less privacy leakage and a much lower communication cost.In summary,each of these two methods has its own strengths and weaknesses,and the desirable method is different in various situations.

    As mentioned above,a critical vulnerability of existing approaches is the performance degradation when data distribution is not identically and independently,which is a considerable challenge in the edge-based learning scenario[10].Numerous distributed learning algorithms have demonstrated superior performance in situations where the data in clients are independent and identically distributed(IID).However,data is typically unbalanced and Non-IID,as it is originated in different contexts.Most of the existing distributed learning algorithms(e.g.,[5,11—14])assume the data partitions are IID or have conducted extremely limited research on Non-IID data partitions.Therefore,addressing the challenge of Non-IID quagmire is substantial and imperative in distributed learning.

    In this paper,we propose ADC-DL,a communication-efficient distributed learning method withhierarchical clusteringandAdaptive Dataset Condensation.ADC-DL generates the local summaries by adaptive dataset condensation,which is not dependent on the presence of representative samples,ensuring the validity of the local summaries.In ADC-DL,the edge clients train representative synthetic samples adaptively based on the hierarchical structure.Subsequently,all the clients transmit their synthetic samples to the cloud server for global model training.On the one hand,ADC-DL adjusts the size of local summaries adaptively based on our proposed comprehensive hierarchical model,which improves the model performance in the situation of Non-IID data partitions.On the other hand,it is communication-efficient and privacy-preserving by transmitting the synthetic dummy summaries instead of the original data.

    We are the first to employ dataset condensation to solve the Non-IID dilemma in the scenario of distributed data.Our main contributions are illustrated as follows:

    ? We design a distributed learning approach with hierarchical clustering and adaptive dataset condensation,which is communication-efficient,privacy-preserving and appropriate for Non-IID data partitions.

    ? We propose the entropy topsis comprehensive tiering model for hierarchical clustering to distinguish clients according to their data characteristics.The implementation and purpose of our tiering model are innovative in comparison to existing tiering methods.

    ? We propose the adaptive dataset condensation to adjust the size of synthetic samples based on the hierarchical structure,which improves the performance of ADC-DL on Non-IID data partitions.

    ? We evaluate the performance of ADC-DL by conducting extensive experiments on various datasets and data distribution schemes,which confirm that our ADC-DL outperforms other mainstream methods.

    The remaining of the article is organized as follows.In Section II,we introduce the related work.The principle and framework of ADC-DL we proposed,especially the entropy topsis comprehensive tiering model and adaptive dataset condensation,are presented in Section III.In Section IV,we evaluate the performance of ADC-DL compared with other distributed training methods.Section V concludes the paper.

    II.RELATED WORK

    Distributed learning is one of the most valuable and promising directions of research in large-scale machine learning [15],which avoids the necessity of gathering all the original data into a single server for central processing,saving time and energy[16].Originally,it is implemented by computing and transmitting the output of local models or models themselves[17].Kittler et al.design a series of efficient classifier combination rules (e.g.,majority vote rule) to combine all the local outputs to produce a unique result [18].Wolpert and David propose to employ a high-level model for combining the low-level classifiers to improve the accuracy of distributed learning[19].However,such methods need to define a uniform representation into which the different classifiers are translated,which is challenging to conduct.

    Sharing models and sharing representative summaries are two of the most predominant approaches in recent studies.McMahan et al.propose the most prevalent algorithms in federated learning called federated averaging(FedAvg)[5].FedAvg has been certified to work well empirically,especially for nonconvex problems.However,the accuracy of FedAvg is dramatically degraded when the data distribution is Non-IID.

    In attempting to solve the quagmire of Non-IID data partitions,Zhao et al.investigated the causes of global model degradation in data heterogeneity scenarios at first[20].Simultaneously they design a strategy to improve the model performance on Non-IID data through the establishment of a small subset of data that is globally shared among all the edge devices.Experiments indicate that the accuracy of global model is improved.However,this method will consequently result in additional communication costs.

    Research on data summarization has inspired the approach of collecting data summaries in distributed learning.A variety of data summarization techniques have been proposed in literature and are currently implemented in practice(e.g.,coresets,sketches[21,22]).Many researchers have adopted this approach to distributed learning recently.Lu et al.develop a robust coreset construction algorithm based on k-means clustering to reduce the communication overhead in distributed learning [4].However,the model accuracy of this algorithm is not satisfactory.

    Our proposed ADC-DL is a novel distributed learning algorithm based on data summarization.Effectively improve the accuracy of the central model under the premise of privacy security and low communication costs,especially in the setting of Non-iid data distribution.

    III.THE DESIGN OF ADC-DL

    In this section,we present the architecture overview of our proposed ADC-DL first.Subsequently,we illustrate the two most essential components in ADC-DL:model of hierarchical clustering and adaptive dataset condensation.

    3.1 Overview

    LetDkandNkindicate the original dataset and condensed dataset of clientkrespectively,where the number of samples inNkis considerably smaller than that inDk.Our ADC-DL learns a mapping Ψ:Dk →Nkby minimizing theθDandθNthat are calculated for training loss over original datasetDkand condensed datasetNkwhile adaptively adjusting the number of samples inNkaccording to the hierarchical structure.Eventually,the server aggregates all theNkand constructs the global model.

    The framework of ADC-DL primarily consists of the following three parts: hierarchical clustering,adaptive data condensation,and aggregation of all the condensed data.The architecture overview is displayed in Figure 1.The main procedure of ADC-DL can be summarized as follows:

    ?Step 1:Hierarchical Clustering[Figure 1(a)].All the clients calculate the characteristics of their data distribution locally.The client data scoreSkand the hierarchical structure are subsequently acquired by the entropy topsis comprehensive tiering model.

    Figure 1.Overview of the ADC-DL framework,which consists of three parts: (a) hierarchical clustering: evaluate and acquire the client data score and the hierarchical structure by entropy topsis comprehensive tiering model.(b) Adaptive data condensation: condense the data of client adaptively into a small set of synthetic samples.(c)Aggregation of all the condensed data: The server trains the model on the aggregated data and shares it with all the clients.

    ?Step 2:Global Model Downloading and Adaptive Dataset Condensation[Figure 1(b)].The server initializes the global model parameters and distributes the model to all the clients.After receiving the global model from the server,each participated client respectively condenses the original data.The size of synthetic samples is adaptively determined subject to the position of client in hierarchical structure and their data distribution characteristics.

    ?Step 3:Aggregation of synthetic samples and Model Training[Figure 1(c)].The server aggregates all the synthetic samples and trains the initialized model.At last,the trained model is transmitted to all the participated clients.

    The first two steps will be discussed in detail in the following subsections.After completing these two steps,the server aggregate all the synthetic samples into a sequence,on which trains and distributes the global model to clients.

    Algorithm 1 depicts the training process of ADCDL.At the beginning of training,the server initializes the parameters of global model and shares the model with all the clients.Lines 2—3 obtain the hierarchical structure by the entropy topsis tiering model.In lines 4—7,each client calculates the synthetic samplesNkparallelly and transmits them to the server.The implementation details of line 5 are shown in Algorithm 2.Line 8 aggregates all the synthetic samples,and line 10 is performed to update the parameters of global modelθi+1,ηis the learning rate.

    Algorithm 1.Training Process of ADC-DL.The M clients are indexed by k.E is the local training numbers of server.Pθ0 is the probability distribution over randomly initialized weights.

    Algorithm 2.Adaptive dataset condensation.

    3.2 Model of Hierarchical Clustering

    As shown in Figure 2,ADC-DL stratifies all the clients to facilitate the adaptive dataset condensation in the following subsection.

    Figure 2.Hierarchical clustering of the clients.

    Our idea of hierarchical clustering was primarily inspired by the following observations: 1) Clients with more extensive data samples contribute more to the trained global model eventually.2) Clients with diverse degrees of Non-IID and imbalance have various performances in the model training.Our method of hierarchical clustering achieves a comprehensive consideration of these factors.

    There have been some approaches proposed in recent studies to layering clients.Zhang et al.stratify the clients according to the computational speed of edge devices for the implementation of asynchronous federated learning [23].Briggs et al.stratify the clients by evaluating the similarity of the global model to the local model[24].Our approach is highly distinct from these existing approaches.On the one hand,the basis of stratification is different.Our stratification is on the basis of the data distribution characteristics of each client.On the other hand,the purpose is different.The purpose of our layering is to provide the foundation for subsequent adaptive dataset condensation,reflecting the suitability of individual clients for dataset condensation.

    With the above analysis,we primarily take the following three specific factors into consideration when layering the clients: (1) the Non-IID level of data in client,(2)the number of data samples in client,(3)the balance level of data in client.All these factors will have a substantial impact on the process of adaptive condensation on the client-side.

    According to [20],when all the clients start from the same initialization as the centralized settings,the Non-IID degree of client k can be simplified as:

    whereCis the number of classes.Earth mover’s distance(EMD)has been proved to be a good metric for quantifying the Non-IID degree of data [20].Consequently,we follow this definition and employ EMD to indicate Non-IID degree in our paper.The amount of data samples in client is an influential factor affecting the size of the condensed dataset.We apply|Dk|to represent it.The balance degree of data in client is another factor to consider.We perceive that the client with balanced data is favorable for model training.We take the distance between the data of client k and the balanced distributed data as balanced distance(BD),

    Subsequently,We set up the Entropy Topsis comprehensive tiering model to consider these factors.Entropy Topsis is the combination of entropy weight method and Topsis,which can objectively weigh various factors and evaluate the quality of each sample by approximating the ideal solution[25].To construct a uniform metric across clients,we introduce the definition ofclient data score Sk,which represents the evaluation of suitability for model training.After the standardization of the three factors of EMDk,|Dk|and BDkof each client,calculate the entropy of each factor:

    whereMis the amount of clients.The weight of each factor is:

    We employxkito represent the weighted values of clients.Then the ideal clientC+and the negative clientC?can be determined as:

    whereI={|D|}associated with the bigger the better factor,I′={EMD,BD}associated with the smaller the better factors.Then we can evaluate the data score of all the clients by calculating the relative closeness of each client as follows:

    According toSk,participated clients can be partitioned into multiple tiers :{tier1,tier2,...,tierN},where the clients intier1are the most ideal clients for training global model and the clients intierNare the most unsuitable clients for training global model.

    3.3 Adaptive Dataset Condensation

    Inspired from DC [26],we propose adaptive dataset condensation,which adaptively adjusts the condensation process according to the tier of client so as to be applicable to Non-IID data.

    The goal of dataset condensation is to generate a small set of synthetic samples,which can achieve comparable performance with the original dataset for a specific model,as shown in Figure 3.This goal can be formulated as a minimization problem between two sets of gradients computed for training loss over the original and synthetic datasets.

    Figure 3.The purpose of dataset condensation is to generate a small set of synthetic samples,which can achieve comparable performance with original data for a specific model.

    Suppose the client k has|D|pairs of training samples and labels,on which we want to acquire a neural network with parametersθ.Generally,θcan be learned by minimizing the cost function on the training set:

    wherelis the loss function.Our goal is to acquire a synthetic datasetNwith a much smaller number of samples than the raw dataset.The model parametersθNtrained onNcan be learned as follows:

    We expect that the test accuracy of the model withθDandθNcan be comparable.To accomplish this expectation,the model trained onNshould converge to a similar solution with the model trained onDin the parameter space.Therefore,the goal can be formulated as:whered(·,·)is the distance function.

    Furthermore,to achieve the optimal solution,it is imperative that not only to make the finalθNclose toθD,but also to make all theθNin the whole optimization process follow a similar path toθD,that is,theθNof each optimization round close toθD.Therefore,the formulation in Eq.(10)can be simplified as:

    where T is the total number of optimized steps.

    Thus,we can acquire the condensed datasetN,which can accomplish comparable performance to the original dataset for a specific model.Note that the condensed datasetNis synthetic summaries for training the specific deep neural network rather than the subset of original samples.Consequently,the raw data of clients will never be exposed during the transmission of these synthetic samples,which adequately protects the privacy of clients.

    The size of synthetic samples of each client is a crucial parameter during the process of condensation,which has a substantial impact on the performance of the model trained on the synthetic samples.It is evident that the test accuracy will be improved with the increase of|N|because the synthetic dataset will contain more information from the original dataset.However,the communication costs will increase simultaneously with the enlargement of|N|.Therefore,we propose that the size of synthetic samples should be adaptively adjusted according to the data characteris-tics and tier of client.Specifically,we propose Eq.(12)as follows to accurately calculate the number of synthetic samples.

    where the parameterλiare determined according to the hierarchical structure,λi ∈[0.2,1].aandbrepresent the range of client data scoreSk,the lousiest client score isaand the most desirable client score isb.The other two parametersμ1andμ2affect the range of|Nk|for each client.The experiment in Section IV proves that this way for determining|Nk|is quite effective.

    The implementation of adaptive dataset condensation is presented in Algorithm 2.Lines 2—3 initialize the synthetic samplesNk,where the size ofNkis determined by thetierkand Eq.(12).In lines 4—5,lis the number of outer-loop steps of iteration andtis the number of inner-loop steps of iteration.In lines 6—13,the stochastic gradient descent optimization is performed to updateNk.Line 15 updates the parameters of the model trained onNk.Line 18 returns the finalNk.

    IV.EXPERIMENTS

    In this section,we compare the performance of ADCDL with a variety of other distributed training algorithms.We evaluate the prediction accuracy and communication costs respectively based on the public datasets.

    4.1 Experimental Settings

    Datasets.The datasets used in our experiments include: MNIST [27] with 60000 images for training and 10000 images for testing in 10 classes,Fashion-MNIST [28] with 60000 images for training and 10000 images for testing in 10 classes,SVHN[29] with more than 600000 real-world images in 10 classes,and CIFAR10 [30] with 50000 images for training and 10000 images for testing in 10 classes.In order to analyze the performance of ADC-DL under different data distribution,we investigate four schemes with different data settings of clients: (a)IID distribution,where data from each client are independently identically distributed.(b) Mildly skewed Non-IID distribution,the data are sorted by class and then divided into 70 partitions.Each client receives 1—7 partitions randomly.(c) Highly skewed Non-IID distribution,analogous to schemes b,except that dataset is divided into 20 partitions and each client receives 1 or 2 partitions.(d) Extremely skewed Non-IID distribution,where each client contains data from only one class.The major distinction in these schemes is the degree of data heterogeneity and balance.In these schemes,(a) is the scheme of independent and identical distribution among each client’s data.Most of the distributed learning algorithms can achieve satisfactory performance under this setting.The degree of data heterogeneity in scheme (b) to (d) increases,rendering it progressively tricky for distributed algorithms to achieve favorable performance.We apply our dataset setting on 30 clients for evaluation in our experiment.

    Model.Regarding the models,we employ standard deep network architectures of ConvNet[31]in our approach.CNN is a prevalent architecture that is typically deployed for image recognition.It can be denoted as[W,N,A,P]×D,whereDis the number of duplicate blocks,each block has a convolutional Layer withWfilters,a normalization layerN,an activation layerAand a pooling layerP.In our setting,D is assigned to be 3.Each block has 128 filters followed by ReLU and AvgPooling modules.A linear classifier is placed at the end of ConvNet.

    Experimental platform and Hyperparameters.We use python of version 3.8.3 and Pytorch of version 1.6.0 to construct our distributed learning framework.And we use a single GPU,GeForce RTX 2080 Ti,for hardware acceleration.In the process of adaptive dataset condensation for each client,we set the learning rate for updating synthetic samplesηimg=0.1,the learning rate for updating network parametersηnet=0.01,batch size for raw datasetbs1=256,batch size for training networksbs2=256,the training iterationit=500.In determining|Nk|adaptively,empirically,we set the parameters in Eq.(12)μ1=0.3 andμ2=9.7 to keep the size of each client’s synthetic dataset within a reasonable range.

    4.2 Results

    In this subsection,we conduct experiments to evaluate the prediction performance under various datasets and schemes.We compare ADC-DL with two mainstream privacy-protected distributed training methods,sharing models and sharing local summaries.Subsequently,we investigate the convergence performance of ADC-DL when partial clients are out of order.At last,we perform a set of ablation experiments to verify the improvement of our ADC-DL.

    Comparison to federated averaging.First,we compare ADC-DL with the distributed learning method of federated averaging(FedAvg).We employ the setting of federated averaging in[32],which conducts a global aggregation at each training epoch.All the clients transmit their local model to the server after each epoch of local gradient descent,and the server aggregates the models by weighted average.Then distributes the aggregated models to clients.In this setting of federated learning,the final aggregated model is equivalent to the model trained on the union of all the local data.[32].

    The prediction accuracy and training epochs of ADC-DL and FedAvg are presented in Table 1.It can be expected that ADC-DL achieves superior performance than FedAvg in most cases in terms of test accuracy and training epochs.Particularly,ADC-DL adjusts the size of synthetic samples according to the data distribution,making it more appropriate for the scenario with data distribution skew.

    Table 1.Prediction performance comparison between ADC-DL and FedAvg.

    Table 2.Results of ablation experiments for the adaptive scheme.The communication costs for methods with fixed|Nk|are constant for each average EMD.ADC-DL implements an excellent trade-off between accuracy and communication overhead compare to the methods without adaptive components.

    Comparison to coreset methods.Previously,data summaries in distributed learning were generally constructed by coreset.Therefore,we evaluate ADC-DL together with benchmarks including the RCC-kmeans algorithm in [4] and its distribution extensions,i.e.,CDCC[6]and DRCC on different data distribution of four datasets(MNIST,FashionMNIST,SVHN and CIFAR10)to demonstrate the advantages of ADC-DL.

    Intuitively,each data sample in the coreset represents a set of data samples from the raw dataset[3].RCC-kmeans is a centralized clustering algorithm with no communication between each client during the computation of coreset points.CDCC and DRCC are the distributed clustering algorithm.They perform local cluster analysis on individual clients at first.Then transmit partial clustering results as output to other clients and aggregate them into final clustering results.Our algorithm has a remarkable superiority compared to these methods.The above algorithm calculates a generalized data summary for a distributed dataset,considering only the characteristics of the data itself but not the task model structure.Our ADC-DL calculates the data summary for a specific model,which is not valid for other models but can achieve better performance for the particular model.We parameterize CDCC with k=2 according to the evaluated k-means problem[6].Other parameters of coreset following the same setting as[4]to achieve an ideal result.Figure 4 shows the accuracy results measured by ADC-DL and benchmarks.

    Figure 4.Evaluation of ADC-DL and benchmarks on different datasets.

    Figure 5.(a)Convergence performance with respect to the proportion of working devices on IID data partition of MNIST.A small proportion of devices failures only have a slight impact on the prediction accuracy.(b)Convergence performance with respect to the proportion of working devices on highly-skewed Non-IID data partition of MNIST.Devices failures cause a distinct decline in prediction accuracy.

    Figure 6.The model accuracy with respect to the communication costs on MNIST when data distribution is Non-IID.The X-axis is in log-scale.

    As Figure 4 illustrates,the prediction performance of ADC-DL is remarkably higher than that of distributed learning based on sharing coreset in all cases and this dominance is more pronounced on the simpler dataset.In general,the model accuracy degrades as the degree of non-iid of data gradually increases in scheme (a) to (d).Compared with other algorithms,the accuracy decline of ADC-DL with the increase of Non-IID level is acceptable.RCC-kmeans is the best performing algorithm except for ADC-DL,and its accuracy is basically constant as the data distribution changes.As for CDCC algorithm,which blindly selects k coreset points to be local centers at each client,it severely suffers from the highly heterogeneous data when the classification task is challenging [4].Although DRCC can customize the configuration to the local dataset and achieve robust performance,its prediction performance is less than ideal.Summing up the above analysis,it can be demonstrated that ADCDL is relatively outstanding in terms of accuracy and the robustness of performance.

    Figure 7.The model accuracy with respect to the communication costs on CIFAR10 when data distribution is Non-IID.The X-axis is in log-scale.

    Performance evaluation when parts of clients malfunction.In distributed settings,there are a variety of distinct differences among clients in terms of network battery power and connectivity,which will lead to exacerbating challenges such as straggler [33].In case that parts of devices get out of order and are offline from the system,these devices will fail to calculate and share their synthetic samples.Thus,the prediction accuracy of the global model will be considerably affected.We evaluate the performance of ADC-DL in the face of this challenge,testing the accuracy for the devices dropping rate of 0%,10%,...,90% respectively.To reveal the relationship between model accuracy and dropping rate more clearly,we use 3Dgraph to observe.As shown in Figure 5,the two coordinates on the horizontal plane represent the epoch of model training and the proportion of devices working in order,the vertical coordinate represents the model accuracy.It can be seen that ADC-DL can achieve an acceptable result even if some clients fail in the case of data distribution is IID.The impact of device offline on model performance is even pronounced when the data distribution is Non-IID.The model accuracy will deteriorate dramatically to below 80%when more than 20%of the devices are out of order,which is unacceptable for the classification task.

    Comparison of the accuracy vs.the communication costs.We show the comparison of the model accuracy vs.uploading bytes in communication on MNIST and CIFAR10 in Figure 6 and Figure 7.The benchmarks include FedAvg,CDCC,DRCC mentioned above and another communication-efficient distributed learning method named SAPS-PSGD [34].The experimental results show that the ADC-DL we proposed can condense data of clients effectively according to the data distribution to achieve a target model accuracy with fewer communication costs on both MNIST and CIFAR10.

    On Non-IID data distribution of MNIST as shown in Figure 6,to achieve 91.6% test accuracy,SAPSPSGD requires about 92MB communication overhead and FedAvg requires about 286.6MB.While ADC-DL we proposed needs only 24.1MB,which is about 3.8×and 11.9×smaller than SAPS-PSGD and FedAvg,respectively.CDCC and DRCC can only achieve around 80% accuracy with more than 100MB.On Non-IID data distribution of CIFAR10 as shown in Figure 7,to achieve 60%test accuracy,ADC-DL requires 68MB,which is 1.64×and 5.5×smaller than SAPSPSGD(109MB) and FedAvg(372.1MB),respectively.Algorithms of CDCC and DRCC have higher communication costs and they are not able to achieve acceptable accuracy.Through the above analysis,we can conclude that the adaptive scheme based on the hierarchical structure leads to lower communication costs than the SAPS-PSGD,FedAvg and CDCC/DRCC algorithm.

    Ablation experiment of our improvement.To further validate the effectiveness of our adaptive scheme,we conduct a set of ablation experiments to contrast the result obtained from ADC-DL with those obtained by directly applying DC to distributed learning.We set|Nk|of all the clients to 10,100,500 respectively for dataset condensation as the control experiment.The prediction accuracy and communication costs are computed on MNIST for each average EMD.

    Table 2 presents the accuracy and communication costs of each algorithm as the average EMD rises.Generally,the increase of average EMD leads to a noticeable degradation in the performance of the algorithm based on dataset condensation,which is because of the imbalance and variable quality of each class in the union of synthetic samples.It can be observed that when simply applying DC to distributed learning,there is a remarkable improvement (about 4%?7%)in model accuracy as|Nk|rises from 10 to 100.However,when|Nk|rises from 100 to 500,the model accuracy improves just a little bit (less than 1%) while the communication costs increase approximately five times.Our ADC-DL dramatically lowers the communication overhead while achieves an acceptable accuracy.That is,our algorithm implements an excellent trade-off between accuracy and communication overhead.In general,it can be revealed that our proposed scheme of adaptively adjusting the number of synthesized samples is effective and outperforms the current algorithm in terms of accuracy and communication overhead.

    V.CONCLUSION

    In this paper,we propose a communication-efficient distributed learning algorithm (ADC-DL) with hierarchical clustering and adaptive dataset condensation.We construct a hierarchical structure to better tackle the challenge of data heterogeneous.Our proposed adaptive dataset condensation ensures data privacy and low communication cost in the process of distributed learning.Experiments on different datasets and schemes of data distribution demonstrate that our ADC-DL outperforms other algorithms in terms of prediction accuracy and communication costs.In addition,the ablation experiments are performed to verify the superiority of the adaptive scheme compared to DC.In this paper,we condense the client data by class.In future research,we plan to assign each synthetic sample a soft label to increase the condensation ratio and reduce the communication costs further.

    ACKNOWLEDGEMENT

    This work is supported by the General Program of National Natural Science Foundation of China(62072049).

    国产成人精品久久二区二区91| 一级毛片精品| 麻豆成人av在线观看| 欧美日韩黄片免| 中文字幕人成人乱码亚洲影| 一个人免费在线观看的高清视频| 午夜激情福利司机影院| 99久久99久久久精品蜜桃| 妹子高潮喷水视频| 波多野结衣高清作品| 51午夜福利影视在线观看| bbb黄色大片| 午夜免费鲁丝| 亚洲熟女毛片儿| 欧美在线黄色| 亚洲国产欧美一区二区综合| 美女 人体艺术 gogo| 精品国产乱子伦一区二区三区| 欧美成人午夜精品| 国产精品免费一区二区三区在线| 麻豆av在线久日| 成熟少妇高潮喷水视频| 久久精品亚洲精品国产色婷小说| 成人三级黄色视频| 久久精品国产亚洲av高清一级| e午夜精品久久久久久久| 亚洲中文日韩欧美视频| 国产亚洲精品第一综合不卡| 丝袜美腿诱惑在线| 中出人妻视频一区二区| 免费无遮挡裸体视频| 国产在线观看jvid| 欧美在线一区亚洲| 国产精品 欧美亚洲| 成人国产综合亚洲| 男女下面进入的视频免费午夜 | 禁无遮挡网站| 成人欧美大片| 久久精品人妻少妇| 黑丝袜美女国产一区| 18禁黄网站禁片午夜丰满| 国产99久久九九免费精品| 欧美亚洲日本最大视频资源| 国产三级黄色录像| 免费在线观看亚洲国产| 欧美绝顶高潮抽搐喷水| 日本 av在线| 欧美zozozo另类| 成年女人毛片免费观看观看9| 91九色精品人成在线观看| 亚洲专区中文字幕在线| 叶爱在线成人免费视频播放| 亚洲男人的天堂狠狠| 亚洲国产精品合色在线| 免费电影在线观看免费观看| 麻豆成人午夜福利视频| 亚洲av片天天在线观看| 无遮挡黄片免费观看| 丝袜美腿诱惑在线| 法律面前人人平等表现在哪些方面| 国产精品国产高清国产av| 精华霜和精华液先用哪个| 18禁美女被吸乳视频| 亚洲成国产人片在线观看| 亚洲一区二区三区不卡视频| 国产不卡一卡二| 亚洲国产精品成人综合色| 成人亚洲精品av一区二区| 亚洲国产精品合色在线| 久久精品国产亚洲av香蕉五月| 日韩av在线大香蕉| 最近最新中文字幕大全电影3 | 18禁美女被吸乳视频| 老司机午夜福利在线观看视频| 国产成+人综合+亚洲专区| 亚洲专区中文字幕在线| 18禁美女被吸乳视频| 午夜福利成人在线免费观看| 久久香蕉精品热| 久久久久久久久久黄片| 午夜福利视频1000在线观看| 久久国产亚洲av麻豆专区| 欧美又色又爽又黄视频| 精品久久久久久久毛片微露脸| 精品一区二区三区四区五区乱码| 一本综合久久免费| 婷婷六月久久综合丁香| 欧美成人一区二区免费高清观看 | 在线看三级毛片| 亚洲av成人一区二区三| 日韩精品免费视频一区二区三区| 午夜福利视频1000在线观看| 精品卡一卡二卡四卡免费| 日本五十路高清| 久久中文字幕人妻熟女| 国产久久久一区二区三区| 一边摸一边做爽爽视频免费| 国产一区二区激情短视频| 久久久精品国产亚洲av高清涩受| 91九色精品人成在线观看| 精华霜和精华液先用哪个| 久久亚洲真实| 看免费av毛片| 精品久久久久久,| а√天堂www在线а√下载| 欧美另类亚洲清纯唯美| 不卡一级毛片| 天天添夜夜摸| 免费看日本二区| 免费无遮挡裸体视频| 国产私拍福利视频在线观看| 国产成人av教育| 亚洲国产日韩欧美精品在线观看 | 观看免费一级毛片| 亚洲精品中文字幕在线视频| 欧美性猛交╳xxx乱大交人| 亚洲一区高清亚洲精品| 精品不卡国产一区二区三区| 久久久久久久午夜电影| 成熟少妇高潮喷水视频| 日韩欧美 国产精品| 国产一区二区三区在线臀色熟女| 日韩欧美在线二视频| 亚洲 欧美 日韩 在线 免费| av在线天堂中文字幕| 久久久久国产一级毛片高清牌| 亚洲av五月六月丁香网| 久久天堂一区二区三区四区| 亚洲av电影不卡..在线观看| 精品少妇一区二区三区视频日本电影| 久久久精品欧美日韩精品| 亚洲成国产人片在线观看| 婷婷精品国产亚洲av| 国产精品免费一区二区三区在线| 国产97色在线日韩免费| 亚洲色图av天堂| 亚洲精品色激情综合| 免费在线观看黄色视频的| 搡老熟女国产l中国老女人| 国产一区在线观看成人免费| 亚洲色图 男人天堂 中文字幕| 一级毛片高清免费大全| 男女下面进入的视频免费午夜 | 亚洲国产日韩欧美精品在线观看 | 18美女黄网站色大片免费观看| 很黄的视频免费| 一本大道久久a久久精品| 日韩精品免费视频一区二区三区| 国产极品粉嫩免费观看在线| 亚洲欧美精品综合一区二区三区| a级毛片a级免费在线| 性欧美人与动物交配| 欧美国产精品va在线观看不卡| 国产免费av片在线观看野外av| 两性夫妻黄色片| 热re99久久国产66热| 美女大奶头视频| 精品久久久久久久末码| 91九色精品人成在线观看| 国产高清激情床上av| 成人手机av| 窝窝影院91人妻| 亚洲美女黄片视频| 99在线视频只有这里精品首页| 亚洲国产欧美网| 一a级毛片在线观看| 禁无遮挡网站| 女性被躁到高潮视频| 18禁国产床啪视频网站| 亚洲成国产人片在线观看| 国产亚洲av高清不卡| 女性生殖器流出的白浆| 亚洲第一电影网av| 村上凉子中文字幕在线| 国产精品久久视频播放| 女人爽到高潮嗷嗷叫在线视频| 草草在线视频免费看| 在线观看日韩欧美| 在线观看66精品国产| 亚洲中文av在线| 亚洲av电影不卡..在线观看| 久久婷婷成人综合色麻豆| 少妇裸体淫交视频免费看高清 | 精品卡一卡二卡四卡免费| 久久精品夜夜夜夜夜久久蜜豆 | 成人国语在线视频| 国产高清视频在线播放一区| 久久亚洲真实| 在线观看舔阴道视频| 他把我摸到了高潮在线观看| 十分钟在线观看高清视频www| 精品不卡国产一区二区三区| 亚洲性夜色夜夜综合| 久久伊人香网站| 一区二区三区高清视频在线| 啪啪无遮挡十八禁网站| 久久青草综合色| 亚洲第一av免费看| 操出白浆在线播放| 日韩中文字幕欧美一区二区| 国产精品av久久久久免费| 欧美中文综合在线视频| 黑人巨大精品欧美一区二区mp4| 精品国产一区二区三区四区第35| 一区二区三区国产精品乱码| 一级毛片精品| 麻豆久久精品国产亚洲av| 国产精品久久久久久精品电影 | 成熟少妇高潮喷水视频| 日本熟妇午夜| 99国产极品粉嫩在线观看| 久久中文看片网| av超薄肉色丝袜交足视频| 日韩欧美在线二视频| 国产亚洲av高清不卡| 国产av一区二区精品久久| 午夜福利在线观看吧| 日本三级黄在线观看| 亚洲午夜精品一区,二区,三区| 人人妻人人澡欧美一区二区| tocl精华| 成人国产综合亚洲| 日韩欧美免费精品| 免费一级毛片在线播放高清视频| 亚洲男人天堂网一区| 国产日本99.免费观看| 国产精品亚洲一级av第二区| 国产亚洲精品久久久久5区| 国产成人av激情在线播放| 波多野结衣av一区二区av| 青草久久国产| 欧美日韩瑟瑟在线播放| 中文字幕人妻丝袜一区二区| 午夜成年电影在线免费观看| 波多野结衣av一区二区av| 国产欧美日韩一区二区精品| 国产精品99久久99久久久不卡| 亚洲熟妇熟女久久| 成人18禁在线播放| 神马国产精品三级电影在线观看 | 正在播放国产对白刺激| 此物有八面人人有两片| 成年人黄色毛片网站| 亚洲自偷自拍图片 自拍| 久久中文看片网| 免费看日本二区| 黄色 视频免费看| 欧美性猛交╳xxx乱大交人| 精品久久久久久久久久免费视频| 久久亚洲精品不卡| 日本撒尿小便嘘嘘汇集6| 日本成人三级电影网站| 免费看美女性在线毛片视频| 窝窝影院91人妻| 国产蜜桃级精品一区二区三区| 热99re8久久精品国产| 日本成人三级电影网站| 免费女性裸体啪啪无遮挡网站| 欧美一级a爱片免费观看看 | 亚洲成av人片免费观看| 国产精品1区2区在线观看.| 亚洲无线在线观看| 成人一区二区视频在线观看| 久久天堂一区二区三区四区| 亚洲国产精品久久男人天堂| 欧美av亚洲av综合av国产av| 美女 人体艺术 gogo| 国产亚洲精品久久久久久毛片| 亚洲av第一区精品v没综合| 免费在线观看日本一区| 亚洲男人天堂网一区| 特大巨黑吊av在线直播 | 午夜成年电影在线免费观看| 亚洲国产精品sss在线观看| 日本a在线网址| 精品久久久久久成人av| 久久人人精品亚洲av| av福利片在线| 国产一区二区三区视频了| 色播亚洲综合网| 窝窝影院91人妻| 国产精品,欧美在线| 久久草成人影院| 妹子高潮喷水视频| 日韩高清综合在线| 成人国产综合亚洲| 黄频高清免费视频| 久久精品亚洲精品国产色婷小说| 国产午夜精品久久久久久| 亚洲天堂国产精品一区在线| 国产精品香港三级国产av潘金莲| 久久精品国产99精品国产亚洲性色| 亚洲成人免费电影在线观看| 丝袜人妻中文字幕| 国产黄片美女视频| 母亲3免费完整高清在线观看| 国产成人精品久久二区二区免费| 天堂√8在线中文| 高清在线国产一区| 日韩欧美一区二区三区在线观看| 亚洲国产中文字幕在线视频| 美女午夜性视频免费| 桃色一区二区三区在线观看| 韩国精品一区二区三区| 午夜老司机福利片| 99久久久亚洲精品蜜臀av| 麻豆av在线久日| 天天一区二区日本电影三级| 日本a在线网址| 中文字幕高清在线视频| 国产三级黄色录像| 窝窝影院91人妻| 88av欧美| 人人澡人人妻人| 久久精品国产亚洲av香蕉五月| 少妇的丰满在线观看| 自线自在国产av| 午夜精品在线福利| 色哟哟哟哟哟哟| 又大又爽又粗| 国产精品一区二区精品视频观看| 日本 欧美在线| 国产免费男女视频| 中文字幕另类日韩欧美亚洲嫩草| 久久精品国产综合久久久| 欧美激情极品国产一区二区三区| 人妻丰满熟妇av一区二区三区| 久久天堂一区二区三区四区| 国产高清videossex| www日本黄色视频网| 非洲黑人性xxxx精品又粗又长| 老熟妇仑乱视频hdxx| 午夜福利视频1000在线观看| 亚洲中文字幕一区二区三区有码在线看 | 人人妻,人人澡人人爽秒播| 欧美色视频一区免费| 黄网站色视频无遮挡免费观看| 亚洲aⅴ乱码一区二区在线播放 | 久热这里只有精品99| 精品欧美一区二区三区在线| 日韩成人在线观看一区二区三区| 性欧美人与动物交配| 国产精品影院久久| 国产97色在线日韩免费| 亚洲自拍偷在线| 国产亚洲欧美98| 伊人久久大香线蕉亚洲五| 99精品久久久久人妻精品| 妹子高潮喷水视频| 麻豆成人午夜福利视频| 一区二区三区高清视频在线| 无人区码免费观看不卡| 神马国产精品三级电影在线观看 | 在线国产一区二区在线| 色综合站精品国产| 韩国av一区二区三区四区| 国产午夜精品久久久久久| 亚洲 欧美一区二区三区| 国产成年人精品一区二区| 黑人欧美特级aaaaaa片| 男女床上黄色一级片免费看| 免费观看精品视频网站| 在线天堂中文资源库| 黄色a级毛片大全视频| 亚洲,欧美精品.| 制服人妻中文乱码| 国产私拍福利视频在线观看| 50天的宝宝边吃奶边哭怎么回事| 亚洲熟妇中文字幕五十中出| 国产1区2区3区精品| 久久久久久久午夜电影| 9191精品国产免费久久| 91av网站免费观看| 欧美绝顶高潮抽搐喷水| 一个人观看的视频www高清免费观看 | tocl精华| 午夜视频精品福利| 日本撒尿小便嘘嘘汇集6| 女人高潮潮喷娇喘18禁视频| 国产精品亚洲av一区麻豆| 日韩成人在线观看一区二区三区| 婷婷丁香在线五月| 久久婷婷人人爽人人干人人爱| 国产成人啪精品午夜网站| 日本 av在线| 欧洲精品卡2卡3卡4卡5卡区| 亚洲,欧美精品.| 久久国产精品人妻蜜桃| 免费高清视频大片| 一级毛片高清免费大全| 成在线人永久免费视频| 国产精品亚洲美女久久久| 精品国产乱码久久久久久男人| 日韩国内少妇激情av| 精品免费久久久久久久清纯| 亚洲午夜精品一区,二区,三区| 亚洲aⅴ乱码一区二区在线播放 | 国产精品一区二区三区四区久久 | 午夜福利一区二区在线看| 亚洲精品国产一区二区精华液| 午夜福利欧美成人| 级片在线观看| 岛国视频午夜一区免费看| 怎么达到女性高潮| 视频区欧美日本亚洲| 精品卡一卡二卡四卡免费| 成人免费观看视频高清| 熟妇人妻久久中文字幕3abv| 久久精品国产亚洲av高清一级| 亚洲午夜精品一区,二区,三区| 啪啪无遮挡十八禁网站| 精品国产超薄肉色丝袜足j| 午夜精品久久久久久毛片777| 每晚都被弄得嗷嗷叫到高潮| 一进一出抽搐动态| 99国产极品粉嫩在线观看| 少妇的丰满在线观看| 国产真人三级小视频在线观看| 国产精品久久电影中文字幕| 午夜免费观看网址| 丁香六月欧美| 日韩视频一区二区在线观看| 亚洲自偷自拍图片 自拍| 色播亚洲综合网| 一级a爱视频在线免费观看| 国产成人影院久久av| 欧美三级亚洲精品| 久久久精品国产亚洲av高清涩受| 天天躁狠狠躁夜夜躁狠狠躁| 日韩欧美在线二视频| 午夜免费观看网址| 国产精品野战在线观看| 免费高清在线观看日韩| 老熟妇乱子伦视频在线观看| 嫁个100分男人电影在线观看| 欧美日韩黄片免| 亚洲一区二区三区色噜噜| 青草久久国产| 国产精品影院久久| 亚洲欧美精品综合久久99| 国产又爽黄色视频| 级片在线观看| 美女高潮到喷水免费观看| 91在线观看av| 国产精品 国内视频| 精品国产亚洲在线| 人人妻人人澡人人看| 最好的美女福利视频网| 亚洲精品国产一区二区精华液| 18禁国产床啪视频网站| 丁香六月欧美| 久久精品国产亚洲av高清一级| 少妇 在线观看| 69av精品久久久久久| 黑人巨大精品欧美一区二区mp4| 最近在线观看免费完整版| 好男人电影高清在线观看| 性欧美人与动物交配| 午夜a级毛片| 淫妇啪啪啪对白视频| 视频区欧美日本亚洲| 成人永久免费在线观看视频| 中国美女看黄片| 国产一级毛片七仙女欲春2 | 国产精品久久久人人做人人爽| 制服人妻中文乱码| 国产精品爽爽va在线观看网站 | 亚洲狠狠婷婷综合久久图片| 亚洲va日本ⅴa欧美va伊人久久| 1024视频免费在线观看| 少妇粗大呻吟视频| 久久久久国产一级毛片高清牌| 午夜福利一区二区在线看| 中文字幕久久专区| 久久中文字幕一级| 在线观看免费视频日本深夜| 可以在线观看的亚洲视频| 日本五十路高清| 日韩欧美在线二视频| 国产精品二区激情视频| 99国产精品99久久久久| 色播亚洲综合网| 精品国产亚洲在线| 国产97色在线日韩免费| 少妇被粗大的猛进出69影院| 日日爽夜夜爽网站| 亚洲七黄色美女视频| 日本成人三级电影网站| 久久这里只有精品19| 国产成人系列免费观看| 久久午夜综合久久蜜桃| 黑丝袜美女国产一区| 手机成人av网站| 亚洲色图 男人天堂 中文字幕| 99国产极品粉嫩在线观看| 午夜成年电影在线免费观看| 亚洲精品美女久久av网站| 美国免费a级毛片| 色精品久久人妻99蜜桃| 欧美一区二区精品小视频在线| 亚洲第一欧美日韩一区二区三区| 1024香蕉在线观看| 黄片播放在线免费| 变态另类丝袜制服| 9191精品国产免费久久| 欧美成人一区二区免费高清观看 | 观看免费一级毛片| 精品久久久久久久毛片微露脸| 婷婷亚洲欧美| 精品熟女少妇八av免费久了| 一级a爱片免费观看的视频| 久久午夜综合久久蜜桃| 久久中文字幕人妻熟女| 在线观看免费视频日本深夜| 亚洲欧美激情综合另类| 香蕉av资源在线| 久久国产亚洲av麻豆专区| 亚洲国产精品成人综合色| 成人av一区二区三区在线看| 成年女人毛片免费观看观看9| 久久久久久久久久黄片| 久久久久免费精品人妻一区二区 | 亚洲色图av天堂| 99在线视频只有这里精品首页| 97人妻精品一区二区三区麻豆 | 日韩欧美一区视频在线观看| 精品久久久久久久人妻蜜臀av| 大型av网站在线播放| 午夜视频精品福利| 国产成人啪精品午夜网站| 在线天堂中文资源库| 成人国产综合亚洲| 中文字幕人妻熟女乱码| av欧美777| 人妻丰满熟妇av一区二区三区| 51午夜福利影视在线观看| 老汉色∧v一级毛片| 最新在线观看一区二区三区| 黄片小视频在线播放| 91九色精品人成在线观看| 一本久久中文字幕| 欧美日本亚洲视频在线播放| 亚洲天堂国产精品一区在线| 香蕉丝袜av| 夜夜看夜夜爽夜夜摸| 首页视频小说图片口味搜索| 十八禁网站免费在线| 不卡av一区二区三区| 午夜精品在线福利| 午夜福利欧美成人| 国产精品电影一区二区三区| 99国产精品一区二区蜜桃av| 日韩欧美 国产精品| 亚洲国产中文字幕在线视频| 露出奶头的视频| 一级黄色大片毛片| 成年免费大片在线观看| 久久中文看片网| 免费在线观看视频国产中文字幕亚洲| 最近在线观看免费完整版| 国产一区在线观看成人免费| 亚洲av中文字字幕乱码综合 | 国内少妇人妻偷人精品xxx网站 | 中出人妻视频一区二区| 十八禁网站免费在线| 美女国产高潮福利片在线看| 可以在线观看的亚洲视频| 少妇被粗大的猛进出69影院| 一本综合久久免费| 久久午夜亚洲精品久久| www日本黄色视频网| 成人一区二区视频在线观看| 中亚洲国语对白在线视频| 人人妻人人澡欧美一区二区| 男女那种视频在线观看| 国内毛片毛片毛片毛片毛片| 成人特级黄色片久久久久久久| 久久久国产成人精品二区| 99在线人妻在线中文字幕| www日本在线高清视频| 女警被强在线播放| 日本 av在线| 欧美日本视频| 后天国语完整版免费观看| 亚洲国产看品久久| 波多野结衣巨乳人妻| 一本久久中文字幕| 国产在线观看jvid| 成人精品一区二区免费| 亚洲精品久久国产高清桃花| 村上凉子中文字幕在线| 国产精品电影一区二区三区| 亚洲国产欧美一区二区综合| 国产野战对白在线观看| 久久久久久久久免费视频了| 欧美成人性av电影在线观看| 此物有八面人人有两片| 久久国产精品人妻蜜桃| 69av精品久久久久久| 精品人妻1区二区| 91大片在线观看| av视频在线观看入口| 12—13女人毛片做爰片一| 国产片内射在线| 香蕉丝袜av| 99久久久亚洲精品蜜臀av| 国产高清videossex| 亚洲精品中文字幕一二三四区| 老熟妇仑乱视频hdxx| 久久午夜综合久久蜜桃| 国产成人精品久久二区二区91| 自线自在国产av| 国产成人av激情在线播放| 色精品久久人妻99蜜桃| 午夜影院日韩av|