• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Improving Federated Learning through Abnormal Client Detection and Incentive

    2024-01-20 13:01:00HongleGuoYingchiMaoXiaomingHeBentengZhangTianfuPangandPingPing

    Hongle Guo ,Yingchi Mao,★ ,Xiaoming He ,Benteng Zhang ,Tianfu Pang and Ping Ping

    1College of Computer and Information,Hohai University,Nanjing,211100,China

    2Key Laboratory of Water Big Data Technology of Ministry of Water Resources,Hohai University,Nanjing,211100,China

    ABSTRACT Data sharing and privacy protection are made possible by federated learning,which allows for continuous model parameter sharing between several clients and a central server.Multiple reliable and high-quality clients must participate in practical applications for the federated learning global model to be accurate,but because the clients are independent,the central server cannot fully control their behavior.The central server has no way of knowing the correctness of the model parameters provided by each client in this round,so clients may purposefully or unwittingly submit anomalous data,leading to abnormal behavior,such as becoming malicious attackers or defective clients.To reduce their negative consequences,it is crucial to quickly detect these abnormalities and incentivize them.In this paper,we propose a Federated Learning framework for Detecting and Incentivizing Abnormal Clients (FL-DIAC) to accomplish efficient and security federated learning.We build a detector that introduces an auto-encoder for anomaly detection and use it to perform anomaly identification and prevent the involvement of abnormal clients,in particular for the anomaly client detection problem.Among them,before the model parameters are input to the detector,we propose a Fourier transform-based anomaly data detection method for dimensionality reduction in order to reduce the computational complexity.Additionally,we create a credit scorebased incentive structure to encourage clients to participate in training in order to make clients actively participate.Three training models (CNN,MLP,and ResNet-18) and three datasets (MNIST,Fashion MNIST,and CIFAR-10)have been used in experiments.According to theoretical analysis and experimental findings,the FL-DIAC is superior to other federated learning schemes of the same type in terms of effectiveness.

    KEYWORDS Federated learning;abnormal clients;incentive;credit score;abnormal score;detection

    1 Introduction

    With Internet of Things (IoT) technology continuing to evolve,a huge amount of data is being generated at the edge and Machine Learning (ML) is being used as a powerful analytical tool in Internet of Things scenarios[1].Traditional machine learning frameworks collect data from different information sources,and then the data are shared with the central server for processing,after which the machine learning algorithms are trained in the central server.However,this centralized framework has two drawbacks.Firstly,data is shared with the central server,and client privacy may be compromised.Secondly,due to the huge amount of shared data,the communication and time overhead become expensive.

    In this regard,the development of federated learning both safeguards client privacy and minimizes information transmitted [2].Instead of sharing the original data,the federated learning model is trained locally,and then the model parameters are shared with the central server.As a result,the client’s original data is left locally,and much less information is transmitted over the network.The Federated Averaging algorithm(FedAvg)is one of the most widely used federated learning algorithms at the moment[3].

    The FedAvg framework was proposed by McMahan et al.This framework allows the averaging of different client aggregations.However,in practice,it is difficult to ensure that all clients are normal throughout the training process,considering that federated learning usually requires thousands of rounds of communication to converge and the number of clients involved in training is relatively large.Additionally,unlike distributed machine learning,federated learning does not allow the central server to view data provided by clients or control client behavior.Therefore,clients may experience abnormal behavior in federated learning which is initially referred to as a Byzantine attack.In this paper,we call this abnormal client behavior,as shown in Fig.1.Abnormal client behavior may be caused by a malicious attacker’s attack,or it may be due to the client’s own will.It is now especially crucial to reduce the impact of abnormal clients.

    However,the traditional Byzantine fault-tolerant algorithms are based on defense,they employ untargeted defense against attackers at the expense of honest clients,and this approach tends to reduce model accuracy.Therefore,there is a lack of effective anomalous client detection schemes to prevent anomalous clients from reducing the model training rate.In addition,if abnormal clients are detected,previous work has been to discard them directly,but discarding too many clients leads to missing information as well as unfairness problems,which results in low global model prediction performance.The client’s identity as the normal client and the abnormal client are interchangeable.Such clients are active during roundtbut may become abnormal clients att+1.Or they are abnormal clients in roundtbut become normal clients again in roundt+1.Then,effective incentives are needed to leave more clients to provide quality data.

    Therefore,to efficiently execute federated learning in response to the presence of problems with abnormal clients,we propose a Federated Learning scheme for Detecting and Incentivizing Abnormal Clients (FL-DIAC).Specifically,we propose a Detecting Anomalous Clients Module Based on Detectors(DACM-D).This module uses a pre-trained anomaly detection model to detect anomalous client behavior and eliminate its adverse effects.We apply the FedAvg algorithm to aggregate model weight updates.Since the model weights of deep learning models can easily be too large,we use a downscaling technique to generate a proxy for local model weight updates on the server for anomaly detection.In addition,to encourage more clients to participate in federated learning,we propose an incentive mechanism called Client Incentive Module Based on Credit Score(CIM-CS)to encourage clients to participate in training.

    The following are the main contributions of this paper:

    · To achieve the efficiency and security of federated learning,we propose a scheme to detect anomaly clients and design incentive mechanisms to improve federated learning.Specifically,an approach for detecting abnormal clients based on the detector that uses an anomaly detection model to detect abnormal client behavior and eliminate its negative effects is proposed.Additionally,to reduce the computational complexity,a Fourier transform-based anomaly data detection method for latitude reduction is proposed.

    · In addition,an incentive method based on credit score is proposed,which introduces a multidimensional reverse auction to encourage more clients to participate in the training.

    · Three data sets(MNIST,Fashion MNIST,and CIFAR-10)and three training models(CNN,MLP,and ResNet-18)were set in our experiments.The experimental results demonstrate that the FL-DIAC scheme can increase the client participation rate and ensure the accuracy of model training compared to the same type of federated learning schemes.

    The rest of this paper is organized as follows.We review related work in Section 2.In Section 3.1,we introduce the overall framework of FL-DIAC.We present the DACM-D model of FL-DIAC in detail in Section 3.2 and the CIM-CS model of FL-DIAC in detail in Section 3.3.In Section 4,we conduct extensive experiments to evaluate the performance of FL-DIAC.Finally,we conclude the paper and guide our future work in Section 5.

    2 Related Work

    In this section,we review the related works on detecting anomalous clients and client incentives in order to relate our study to existing research.

    2.1 Detecting Abnormal Clients

    The defense of anomalous clients in the context of federated learning has received a lot of attention.The majority of works are defensive in nature.GeoMed[4],Krum[5],and Trimmed Mean[6]are a few examples.A gradient update method for distribution learning of heterogeneous distributed data that is resistant to Byzantine attacks is also introduced by Li et al.[7]called RSA.Under federated learning,a distributed backdoor defense technique is proposed by Shen et al.[8].In federated learning,distributed backdoor assaults are a major issue that is mostly addressed by this study.However,the existing approaches employ untargeted defense against attackers at the expense of honest clients,and this approach tends to reduce model accuracy.

    2.2 Incentive Abnormal Clients

    To incentivize more clients to take part in training for federated learning,several works of literature have designed incentive mechanisms from different perspectives to increase client engagement.Ding et al.[9] chosed a dimension reduction technology to consider the optimal pricing scheme based on the variability of the client information possessed by the server,and this scheme solves the information asymmetry problem between the client and the server.For clients with different data privacy preserving needs,Wu et al.[10]classified the contribution and privacy overheads of the client and established the payment mechanism corresponding to the privacy type through contract theory.Sarikaya et al.[11] analyzed the reasons for the heterogeneous impact of federated learning clients and proposed an incentive mechanism based on a master-slave type to balance the time delay of each iteration.Ding et al.[12] proposed a multidimensional contract scheme to build the best optimal incentive mechanism on the server side to obtain the best data volume and shortest communication time by solving the optimal reward for various data and communication volumes.This scheme takes into account different multidimensional privacy information of clients including training overheads,transmission delay,and data volume.In order to allow model transmission and free trading,Feng et al.[13]built a cooperative communication platform based on relay networks,where clients are rewarded for acting as relay nodes,and models are passed to the server over a cooperative relay network.

    The auction mechanism is also applied to federated learning.However,this may degrade the performance of federated learning due to heterogeneity between different clients.A brand-new multidimensional incentive paradigm for federated learning was put forth by Zeng et al.[14].The best policy for each client is determined using game theory,and the server and client are jointly parameterized to choose the best client to train the model.

    Another useful tool for assessing data contributions is contract theory.In a scenario involving multiple clients,Lim et al.[15]proposed a tiered incentive architecture.They create incentives between clients and users using contract theory,and they are simply rewarded for the actual contribution of their marginal clients.

    Additionally,Zhan et al.[16]suggested a theory based on an incentive mechanism for a federated learning scheme that integrates distributed machine learning with swarm intelligence perception for large data analysis on mobile clients.This incentive mechanism adjusts the quantity of the data used in training.The platform begins by distributing a task and its associated reward.Each edge client determines its degree of engagement,or the volume of training data,to maximize its utility,taking into account the rewards it receives and the energy expenses.The edge client’s decision-making dilemma is treated as a non-cooperative game in order to establish a Nash equilibrium.

    Deep Reinforcement Learning(DRL)is another effective strategy.Zhan et al.[17,18]suggested an incentive mechanism for DRL-based federated learning by combining deep reinforcement learning and game theory.This study characterizes the shared between the central server and the clients as a Starkelberg game to encourage clients to take part in model training.The client chooses the amount of data to participate in the training as a follower when the central server publishes a training assignment and announces a total reward for the task leader.

    Zhang et al.[19] used the Starkelberg game and auction theory,respectively,to study central server-centric and client-centric crowdsourcing.To encourage users to employ inter-device communication,Li et al.[20]suggested incentive schemes.They took into account two alternative informational settings,one where clients have access to information about all other clients and the other where clients only have access to their own information.Zhan et al.[21]created online and offline methodologies as well as incentive mechanisms for opportunity networks.

    While the above initiatives work to focus on incentive-related concerns,they only take into account one aspect of pricing and contribution.

    In summary,traditional Byzantine tolerance algorithms are defense-based and the performance of such schemes is inefficient.The crux of the performance degradation is that the existing methods employ untargeted defenses to defend against attackers at the expense of honest clients.Abnormal clients continue to exist.Therefore,the main objective of this paper is to detect abnormal clients and eliminate the impact of abnormal clients on accuracy.In addition,we can precisely build a model of each participant’s contribution when designing incentives in these areas.Furthermore,when designing the federated learning incentive mechanism,it is important to estimate the value of training data for each client,and cut costs.So,using the aforementioned two factors pricing and contribution,we will develop the incentive mechanism in this study.

    3 Scheme Architecture

    We introduce each FL-DIAC component in this section.

    3.1 Scheme Model

    The FL-DIAC architecture comprises two physical bodies: the client and the central server,as depicted in Fig.2.The central server manages the gathering and aggregating parameters that the client uploads and the client is the IoT’s data owner.Moreover,FL-DIAC is made up of two components.Specifically,while aggregating models,the central server is unable to identify the anomalous clients since standard federated learning cannot detect the abnormalities of model parameters submitted by clients.We propose Detecting Anomalous Clients Module Based on Detectors(DACM-D)to achieve the objective that anomalous clients can be detected,maximize the performance of the global model,and reflect its credit score for each client to support the subsequent incentive for clients to participate in training.Additionally,we propose a credit score-based client incentive mechanism called Client Incentive Module Based on Credit Score (CIM-CS) to encourage abnormal clients to actively take part in the next epoch of training.For ease of reading,the article’s most significant notations are included in Table 1 below.

    The complete illustration of the FL-DIAC framework is shown in Fig.3.Step 1: some clients download the global model parameters from the central server;Step 2: the parameters uploaded by the local clients are detected by the Detecting Anomalous Clients Module based on Detector,and the normal clients are selected;Step 3: the abnormal clients are incentivized by the Client Incentive Module based on Credit Score,which makes them actively participate in the training;Step 4: the central server performs the aggregation update.The above iterations are repeated for each round until the set number of iteration rounds or the desired model accuracy is reached.

    Below is a description of the particulars of each of these two modules.

    Figure 2:Architecture of the FL-DIAC

    Table 1: Notation

    Figure 3:The complete illustration of the FL-DIAC

    3.2 Detecting Anomalous Clients Module Based on Detectors

    This article makes the following assumptions about the traditional federated learning framework.Specifically,in this scenario,the central server is fully trusted,but anomalies can occur on the client side.In addition,the standard verification set on the central server side is selected according to the needs of the task model owner.This part of the data is not accessible to the client.The encryption and decryption process of model transmission is omitted in the experiments because the encryption and decryption are not relevant to the scenario proposed in this paper.

    In Fig.4,DACM-D has two entities:Nclients and a central server.Each client receives the model parameters sent from the central server and then trains the model using its local dataset and uploads the parameters of the new training model to the server.Unlike traditional federated learning,we design a detector in the server to calculate a credit score for each model parameter uploaded by the client.The detector is a deep artificial neural network.The credit score of each client is calculated by the detector.Specifically,as shown in DACM-D,the server collects the local model parameters uploaded by each client and then uses the model parameters as the input to the detector,whose output is the credit score of each local model.In addition,before the model parameters are input to the detector,we propose a Fourier transform-based anomaly data detection method for dimensionality reduction to reduce the computational complexity.The central server accepts all model parameters and selects those with high credit scores for aggregation according to the federated average method to update the global model.Algorithm 1 displays the Detector-based algorithm for abnormal client detection.The following is a more thorough description.

    InitializationThe central server establishes the federated learning architecture and initializes the global model’s parameters at random.The initialized model parametersw0are then transmitted to each client by the central server.

    In upload phaseIn the training roundt,each client is trained using its local data combined with the downloaded newest shared parametersand updated to obtain the newest local parametersThen,each client uploads the newest local model parametersto the central server.At this point,the minimize the empirical lossFis calculated about clientiin epocht.

    whereDidenotes the client’s data set.|Di|denotes the amount of samples.

    The client-side update method uses Stochastic Gradient Descent(SGD)[22].

    whereδ?Fdenotes the gradient of the loss function andηis the learning rate.

    DetectionThe central server collects the model parameters shared by each client.Those parameters are introduced to the detector to determine each local model’s credit score.We can dynamically select the clients involved in federated learning based on the credit scores of the model parameters to maximize the accuracy of federated learning.

    Specifically,in DACM-D,the shared newest local model parameters by the client are used as input to the central server detector.On the central server,we use an already-trained autoencoder model to detect abnormal parameters from the client.Autoencoder can reconstruct and restore normal data,but it cannot restore anomalous data as well.So it is used for anomaly detection[23].And autoencoder is better at processing high-dimensional data[24].

    Our detection-based approach’s main tenet is to first determine each client’s anomaly score using an anomaly detection model,and then determine each client’s credit score using the anomaly score,where the number of clients takes part in federated learning training isK,the amount of data each client isnk,the local model parameter of each client isand the epoch ist+1.

    The following equation provides the formula for the model parameters in the FedAvg framework:

    wherenkdenotes the number of clients currently involved in training,andndenotes the amount of data for all,wt+1denotes the global model weight update,and we havenk=n.

    In this case,the more data from the client involved in the training,the greater the weighting.Therefore,in Eq.(4),we useinstead ofto obtain the following equation:

    The clientkuses the abnormal scorein epocht+1 to get the credit scorecalculated as follows:

    For the detection of abnormal clients,our scheme considers both the abnormal score and the data score.The hyperparameterLin Eq.(6)controls how muchinfluences the calculation ofLis a parameter of>1.If the data volume is very unevenly distributed,we can increase the value ofL.If one client owns a sizable chunk of the data,Lshould have a high value.In our scheme,we propose an anomaly detection model based on autoencoder[25]to obtain anomaly scores.

    The training data for the autoencoder isD=which represents the parameters after the last epoch of central model aggregation.Therefore,the subscript is-1.Using this datasetD,the autoencoder is already trained.The input of the autoencoder is the model parameters uploaded by each clientThe output of the autoencoder is[26].Given by is the reconstruction error of thei.

    Afterward,we can get the anomaly score.In the epocht+1,clientkreceivedas

    In the federated learning framework,the clients participating in the training are often in the tens of thousands.As a result,the latitude of the parameters input to the autoencoder is substantial,which can lead to high computational complexity for anomaly client detection[27].To reduce the computational complexity,Ghosh et al.[28]used a random element extraction method for dimensionality reduction,and this dimensionality reduction method may leave some normal model parameters of the client unextracted,thus causing a decrease in the quasi-deficiency rate.Therefore,we propose a Fourier transform-based anomaly data detection method.We first detect the input model parameters and kick out the clients that are particularly abnormal,thus generating low-dimensional vectors that are then used as input to the detector[29].

    The Fourier transform based on the anomaly data detection method is described as follows.

    The Fourier transform converts digital signals in the time domain into frequency domain signals.We consider each round of shared model parameters as a signal withnclients corresponding tonsampled values.We put thenclients through the Fourier transform to observe the differences between each of these model parameters,find the anomalies in them,and determine the anomalous model parameters.

    The Fast Fourier Transform is a commonly used method,and we use the Fast Fourier Transform to complete the transformation.First,we take thenmodel parameters according to the requirements of Fast Fourier Transform,the number must meet 2m,take the closest 64,and if the number does not reach 64 is taken as 0,this process is called“data signal expression”.Second,for each household to carry out Fast Fourier Transform transformation,this process directly uses the Fast Fourier Transform algorithm,without any improvement to the finally,the results are counted to find the abnormal model for kicking out.

    After obtaining the client’s abnormal score,the credit score of each client is calculated by using the Eq.(6).Then we use the threshold value to select the anomalous clients,and this process is the same as the two-mechanism classification process.Here,we will set a threshold valueThis threshold can be either the median or the mean.For any client whose model parameter has an anomaly scoregreater than the threshold value for this epoch,this model parameter is not selected for aggregation in this epoch.Therefore,the client providing this model parameter is temporarily defined as an anomaly client in this epoch.The credit scoreof the client is set to zero.

    AggregationThe parameter server collects the client’s local model parameters for each epoch and aggregates them by FedAvg to get the latest global model.Higher aggregation weights will be applied to models that were trained using high-value data.The global model parameters that have been combined are then forwarded to each client.The following updates are made to the newest model parameters:

    Finally,each client downloads the latest model parameters from the central server for each epoch and proceeds to the next epoch until the model converges to the optimal accuracy rate.

    3.3 Client Incentive Module Based on Credit Score

    In the previous subsection,our goal is to detect anomalous clients and obtain a credit score for each client.Due to the presence of anomalous clients,clients suffer from privacy breaches of local data,and such clients do not have sufficient rewards for clients who may not want to participate or share their models.For the abnormal clients.The previous solution was to remove the abnormal client,but dropping the client leads to missing information as well as an unfairness problem,which leads to low performance of global model prediction.The client’s identity is switched between normal and abnormal clients.Such clients are active in round t but have the possibility to become abnormal clients at t+1.Or they are abnormal clients in round t but become normal clients again in round t+1.Therefore,to incentivize abnormal clients to actively take part in federated learning,we propose a credit score-based client incentive mechanism and encourage them to stay passionate in training through rewards.More clients are attracted to providing high-quality models.

    The proposed credit score-based incentive method for federated learning clients is shown in Fig.5.In the CIM-CS,a multidimensional reverse auction[30]is introduced in the incentive module in order to incentivize clients.The goal of CIM-CS is to prevent abnormal clients from being discarded or dropping out of training early while reducing the incentive overheads by preventing overhead explosion during the multidimensional reverse auction.

    Figure 5:Client incentive module based on credit score(CIM-CS)

    As shown in Fig.6,in a one-dimensional reverse auction,because only one attribute determines the reward,it is possible that some clients that provide abnormal parameters will be the eventual winners,and these clients always are judged as the winners.Then,the client that always loses can possibly drop out of federated learning training.In addition,some clients are always judged to be losers because they are not positive at one time.

    Figure 6:One-dimensional reverse auction

    To make competition fair and to prevent excessive overheads of auctions,a sufficient number of clients should continuously participate in the reverse auction of CIM-CS.To this end,the proposed CIM-CS incentive module combines virtual coupons and credit scores calculated by the DACM-D module to get a novel federated learning incentive mechanism.

    To boost the likelihood of winning the following auction round,clientiwill be given a virtual coupon as compensation for losing the previous round.Virtual coupons are expressed asD={d1,d2,···,dn}and has the following formula:

    whereγis the number of virtual coupons andis the credit score provided by the FLDAC-D module.Thus,whenever clientifails in the auction round,the number of couponsγweighted by the credit scoreis added to the virtual coupondi.More virtual coupons are available to creditworthy clients.

    The virtual coupondiis set to zero whenever the clientiwins or withdraws from the previous epoch of auctions.Virtual coupons increase the probability of the client winning this epoch being selected.

    We distinguish between two bid prices:the actual bid price and the ranking price.Clientiproposes the actual bid pricebi.The equation can be used to calculate the ranked pricepi.

    In the proposed incentive module,the ranking pricepiis used to select the winner in each round of the auction,and CIM-CS increases the bidder’s probability of winning by using virtual coupons to reduce the ranking price.

    Even participants with higher bids can become winners through continuous participation (as shown in Fig.7).Therefore,CIM-CS encourages normal clients to continuously participate in training and abnormal clients to actively participate in training.

    3.4 Complexity Analysis

    For the DACM-D algorithm,the client upload parameters need to be traversed and its complexity isO(n).Formdimensional data,ndenotes the data length.Then the complexity of the fast Fourier transform isO(m·n·logn).The complexity of the computation of the abnormal score isO(nlog(n)).The complexity of the client credit score evaluation phase isO(nlog(n)).The complexity of the CIMCS algorithm is alsoO(nlog(n)).

    Figure 7:Combining virtual coupons with multi-dimensional reverse auctions

    4 Experimental Evaluation

    In this section,we use simulation settings to evaluate FL-DIAC’s performance.In-depth descriptions are provided of the datasets,models,benchmarks,and experimental setup used in the experiments.The analysis and outcomes of the experiment are given.

    4.1 Dataset and Model

    In this section,we evaluate the performance of FL-DIAC using three widely used learning models,including Convolutional Neural Network (CNN) [31],Multi-Layer Perceptron (MLP) [32],and ResNet18[33].Because CNN and Multi-Layer Perceptron(MLP)have good image feature extraction capabilities,CNN and MLP are used to identify the MNIST images and the Fashion MNIST face database.The softmax cross-entropy as the loss function is used.The above three models were trained using the following datasets:as shown in Table 2,MNIST[34],Fashion MNIST[35],and CIFAR-10[36]datasets,respectively.Because these above datasets reflect the characteristics of IoT device data.We choose the accuracy of the model on the test set as an indicator to demonstrate the effectiveness of FL-DIAC.

    Table 2: Size of MNIST,fashion MNIST,and CIFAR-10 datasets

    MNISTis a significant collection of handwritten numbers that the National Institute of Standards and Technology has gathered.10,000 photos and labels make up the test set,while 60,000 images and labels make up the training set.

    Fashion MNISTis meant to be a more difficult replacement for the first MNSIT dataset.A balanced selection of 10 distinct classes,each with 7,000 samples,is present.With 60,000 training samples and 10,000 test samples,the dataset has 70,000 samples altogether.

    CIFAR-10is a subset of the 80 million micro picture collection that has been tagged.There are 60,000 32×32 color photos total,divided into 10 categories with 6,000 images each.There are 1,000 photographs in each category,with 50,000 training images and 10,000 test images.

    4.2 Benchmark

    To validate the performance of our proposed FL-DIAC,the following reasonable benchmarks were used.

    FedAvg[3]is a popular model aggregation algorithm for federated learning that is effective and commonly used in federated learning frameworks today.The quantity of data samples utilized for training determines the aggregate weights provided to model updates in this algorithm.

    Median[37]is a Byzantine robust federated learning,which is based on mean aggregation at the central server.

    Krum[5].The closest majority is chosen as the aggregated model after the customers are sorted according to the geometric distance of the customer model update distribution.

    Bulyan[38]aggregates the leftover clients after sorting the clients by geometric distance.

    4.3 Experimental Settings

    The simulation experiment is run in Python 1.4.0 using Pytorch to mimic the FL-DIAC on a server with an Intel(R)Core(TM)i7-10875H CPU running at 2.30 GHz and 32 GB of RAM.

    The training data is dispersed across the clients using a non-IID distribution in order to adapt the MNIST,Fashion MNIST,and CIFAR-10 datasets to the federated environment.The Dirichlet distribution is used to generate cross-category federated datasets for different clients,where the parameterα∈[0,+∞)reflects the degree of heterogeneity of the generated federated datasets,where a largerαindicates a more consistent distribution of data between clients,andα=0 indicates that only one category of data is assigned to a single client.In this paper,a dataset with heterogeneity ofα=10 is selected for the experiment.Therefore,to simulate a situation where each client only has incomplete knowledge,we at random assign instances with fewer labels to each client.Additionally,the presence of anomalous clients,a quick labeling of training data,and other factors could result in the training samples of federated learning clients having corrupted samples.To increase the federated task model’s accuracy,it is very helpful to automatically identify datasets containing corrupted samples.In this simulation experiment,to simulate the presence of abnormal clients,we added different amounts of noise to the model parameters shared by different clients.We take into account the sign-flipping,additive noise,and gradient ascent of three adversarial assault models.A sign-flipping assault changes the sign of the model parameters.The additive noise attack increases the model parameters with Gaussian noise.The abnormal clients launch a local gradient climb attack rather than a gradient descent attack.

    4.4 Experimental Analysis

    4.4.1PerformanceAnalysisofDACM-D

    In order to validate the performance of the DACM-D,the DACM-D focuses on detecting abnormal clients and reducing the training time.Therefore,we will compare it with the same type of methods in terms of accuracy and training time.

    As shown in Figs.8a-8c,9a-9c,10a-10c,the proposed module (DACM-D) outperforms the baseline schemes: Median scheme,Krum scheme,and Bulyan scheme in terms of model accuracy under the attacks of sign-flipping,additive noise,and gradient ascent.In the absence of abnormal clients,our proposed detector-based model DACM-D has approximately the same accuracy as the FedAvg algorithm.In the absence of any detection or defense,the FedAvg algorithm has the lowest model accuracy,which is even lower than 0.1.Although Median is better at resolving unusual client issues,the proposed detector-based model DACM-D improves the model accuracy under symbol flipping attacks by 0.01 or more than Median.Moreover,as described in[37],Median cannot defend against a large number of anomalous clients.The Krum algorithm is a federated learning scheme based on the Euclidean distance,and it is an algorithm that can still converge in the presence of abnormal clients.As a result,Krum was unable to resolve the problems that existed with the abnormal client.

    Figure 8:Model performance under sign-flipping((a):MNIST,(b):Fashion MNIST,(c):CIFAR-10)

    Figure 9:Model performance under additive noise((a):MNIST,(b):Fashion MNIST,(c):CIFAR-10)

    As shown in Table 3,the overall distribution of the training time for the four schemes remains constant under the attacks of sign-flipping,additive noise,and gradient ascent attacks.The proposed module DACM-D uses the least training time in this paper,which is mainly due to the fact that DACM-D takes into account a variety of computationally efficient clients and the low computational complexity of the detection method.The median scheme,and Krum scheme all showed an increase in training time compared to the FedAvg scheme.The above experimental results also show that DACMD is more suitable for resource-constrained mobile edge environments.

    As shown in Fig.11a,the accuracy of the model with all clients selected for training is the lowest.The model accuracy of the Fourier transform-based abnormal data detection method is higher than that of the method without the use of Fourier transform.Similarly,it is derived from Fig.11b that the training time by the Fourier transform-based anomaly data detection method is also minimal.Therefore,the Fourier transform-based abnormal data detection method proposed in this paper is effective.

    Figure 10:Model performance under gradient ascent((a):MNIST,(b):Fashion MNIST,(c):CIFAR-10)

    Table 3: Training time(s)comparison with different schemes

    Figure 11: Comparison of performance of modules with and without Fourier transforms ((a):Accuracy,(b):Training time)

    4.4.2PerformanceAnalysisofCIM-CS

    The main purpose of CIM-CS is to incentivize clients to participate in training.Therefore,to validate the performance of CIM-CS,the number of clients participating in training is compared as a metric.

    In this section,numerous experiments are carried out to evaluate how well the proposed incentive mechanism functions in federated learning.The quantity of participants produced by training the MNIST,Fashion MNIST,and CIFAR-10 datasets,respectively,is depicted in Figs.12a-12c.The y-axis in this graph indicates how many clients participated in each communication round,while the x-axis lists the communication rounds.Figs.12a-12c show that the CIM-CS stabilizes more quickly than the random mechanism.There are also more clients participating in the CIM-CS than in the random mechanism.

    Figure 12:The number of clients per communication round selected by CIM-CS on three datasets((a):MNIST,(b):Fashion MNIST,(c):CIFAR-10)

    4.4.3PerformanceAnalysisofFL-DIAC

    Similarly,we evaluate the FL-DIAC scheme performance by comparing the accuracy of each scheme.

    We trained the MNIST dataset using the CNN model,the Fashion MNIST dataset using MLP,and the CIFAT-10 dataset using ResNet-18,respectively.From Figs.13a-13c,we can see that the proposed FL-DIAC scheme is compared with the FedAvg,the Median scheme,the Krum scheme,and the Bulyan scheme.The accuracy of FL-DIAC,the Median scheme,the Krum scheme,and the Bulyan scheme are higher than the federated average,where FL-DIAC has the highest accuracy rate.The above results can directly indicate that the performance of FL-DIAC is expected to be better.

    Figure 13: Performance comparison with different schemes ((a): MNIST,(b): Fashion MNIST,(c):CIFAR-10)

    In conclusion,we evaluate the proposed scheme on various benchmark datasets.Our findings demonstrate that in comparison to conventional schemes,our federated learning framework for Detecting and Incentivizing Abnormal Clients (FL-DIAC) scheme dramatically lowers the number of abnormal clients and enhances model performance.Additionally,our reward system successfully entices abnormal clients to take part in subsequent training sessions.

    5 Conclusion

    Federated learning is becoming increasingly important as the computational power of remote edge devices and local data privacy increase.However,issues with attacked clients in the federated learning framework still affect model accuracy.And for abnormal clients,previous work was chosen to be discarded,which will lead to the problem of decreasing model accuracy.Therefore,in this paper,we proposed a Federated Learning framework for Detecting and Incentivizing Abnormal Clients (FL-DIAC) to deal with the problem of identifying and incentivizing abnormal clients in federated learning.Specifically,on the one hand,an abnormal client detection model(DACM-D)was constructed.An autoencoder for abnormal detection was introduced in this model.In the federated learning framework,tens of thousands of clients are often involved during training.As a result,the range of parameters input to the autoencoder is substantial,which can lead to high computational complexity for abnormal client detection.In order to reduce the computational complexity,we also proposed a Fourier transform-based abnormal detection method for dimensionality reduction.On the other hand,the Incentive Model Based on Credit Score(CIM-CS)was constructed and this model was used to incentivize clients to attend the training in order to encourage the active participation of abnormal clients.The experimental results show that the FL-DIAC scheme outperforms other algorithms of the same type in terms of model accuracy.The DACM-D model has higher accuracy than its counterparts under attacks of symbol flipping,additive noise,and gradient rise.DACM-D is more suitable for resource-constrained mobile edge environments.The Fourier transform-based anomaly data detection method proposed in this paper is effective.CIM-CS is efficient in incentivizing clients.

    Although this paper solved some problems,due to the large number of clients,timely and fasttracking of abnormal clients becomes a new problem.Future research should focus on introducing blockchain technology to secure federated learning and trace back anomalous clients.

    Acknowledgement:The authors wish to express their appreciation to the reviewers and journal editors for their helpful suggestions which greatly improved the presentation of this paper.The authors are grateful for the support of Fundamental Research Funds for the Central Universities(No.B220203006).

    Funding Statement:This work is supported by Key Research and Development Program of China(No.2022YFC3005401),Key Research and Development Program of Yunnan Province,China(Nos.202203AA080009,202202AF080003),Science and Technology Achievement Transformation Program of Jiangsu Province,China (BA2021002),Fundamental Research Funds for the Central Universities(Nos.B220203006,B210203024).

    Author Contributions:The authors confirm contribution to the paper as follows: study conception and design:Hongle Guo,Yingchi Mao;data collection:Xiaoming He;analysis and interpretation of results: Hongle Guo,Yingchi Mao,Ping Ping;draft manuscript preparation: Hongle Guo,Benteng Zhang,Tianfu Pang.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:In this paper,we use the MNIST dataset.The URL of the dataset is http://yann.lecun.com/exdb/mnist/.We use the Fashion MNIST dataset.The URL of the dataset is https://github.com/zalandoresearch/fashion-mnist.We use the CIFAR-10 dataset.The URL of the dataset is http://www.cs.toronto.edu/kriz/cifar.html.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    亚洲自偷自拍三级| 久久久久久九九精品二区国产| 久久精品国产a三级三级三级| 日韩不卡一区二区三区视频在线| 成人免费观看视频高清| 小蜜桃在线观看免费完整版高清| 国产91av在线免费观看| 简卡轻食公司| 男插女下体视频免费在线播放| 校园人妻丝袜中文字幕| 国产高清国产精品国产三级 | 亚洲不卡免费看| 免费在线观看成人毛片| 在线 av 中文字幕| 久久久久性生活片| 欧美 日韩 精品 国产| 别揉我奶头 嗯啊视频| 国产综合懂色| 国产日韩欧美在线精品| 国产免费又黄又爽又色| 中文字幕人妻熟人妻熟丝袜美| av黄色大香蕉| 成人综合一区亚洲| 久热这里只有精品99| 国产伦在线观看视频一区| 欧美日本视频| 国产精品不卡视频一区二区| 九草在线视频观看| 国产免费一级a男人的天堂| 少妇人妻一区二区三区视频| av国产免费在线观看| 亚洲av.av天堂| 精品久久国产蜜桃| 欧美激情在线99| 日本爱情动作片www.在线观看| 69人妻影院| 美女内射精品一级片tv| 中文乱码字字幕精品一区二区三区| 国产美女午夜福利| 性色av一级| 涩涩av久久男人的天堂| 美女被艹到高潮喷水动态| 18+在线观看网站| 久久久久久久精品精品| av一本久久久久| 精品一区在线观看国产| 精品一区在线观看国产| 欧美激情在线99| 婷婷色综合www| 亚洲人与动物交配视频| 久久久久国产精品人妻一区二区| 少妇 在线观看| 国产一级毛片在线| 国产亚洲av片在线观看秒播厂| 亚洲丝袜综合中文字幕| 特级一级黄色大片| 网址你懂的国产日韩在线| 亚洲电影在线观看av| 国内少妇人妻偷人精品xxx网站| 欧美亚洲 丝袜 人妻 在线| 久久久成人免费电影| a级毛片免费高清观看在线播放| 一本一本综合久久| 亚洲在线观看片| 国产黄片美女视频| 亚洲精品自拍成人| 99久久精品热视频| 国产精品人妻久久久影院| 夫妻午夜视频| 色视频在线一区二区三区| 最后的刺客免费高清国语| 青春草国产在线视频| 亚洲精品一区蜜桃| 亚洲性久久影院| 夫妻性生交免费视频一级片| 亚洲成人精品中文字幕电影| 亚洲精品久久久久久婷婷小说| 亚洲熟女精品中文字幕| 一级片'在线观看视频| 在线观看国产h片| 国产视频内射| 久久久久国产网址| 一本色道久久久久久精品综合| 亚洲国产最新在线播放| 成年女人在线观看亚洲视频 | 六月丁香七月| 国产综合懂色| 永久网站在线| 成人特级av手机在线观看| 国产精品嫩草影院av在线观看| 丝袜喷水一区| 免费av不卡在线播放| 精品国产乱码久久久久久小说| 日韩强制内射视频| 成人亚洲精品av一区二区| 日韩,欧美,国产一区二区三区| 亚洲精品久久久久久婷婷小说| 日韩强制内射视频| 99热这里只有精品一区| 国产免费福利视频在线观看| 青青草视频在线视频观看| 久久久久久久大尺度免费视频| 国产成人午夜福利电影在线观看| 五月开心婷婷网| 午夜日本视频在线| 亚洲欧美日韩卡通动漫| 尤物成人国产欧美一区二区三区| 美女脱内裤让男人舔精品视频| 日产精品乱码卡一卡2卡三| 哪个播放器可以免费观看大片| 国产精品av视频在线免费观看| 一个人观看的视频www高清免费观看| 久久久久久久国产电影| 国产成人精品久久久久久| 男女边吃奶边做爰视频| 99热全是精品| 一级毛片aaaaaa免费看小| 内射极品少妇av片p| 国产女主播在线喷水免费视频网站| 亚洲精品色激情综合| 久久国内精品自在自线图片| 亚洲av中文字字幕乱码综合| 亚洲自拍偷在线| 免费观看无遮挡的男女| 真实男女啪啪啪动态图| 日韩av不卡免费在线播放| 国产综合精华液| 亚洲久久久久久中文字幕| 日本免费在线观看一区| 一级二级三级毛片免费看| 国产精品熟女久久久久浪| 亚洲人成网站在线播| 99re6热这里在线精品视频| 久久精品久久久久久噜噜老黄| 少妇猛男粗大的猛烈进出视频 | 日韩大片免费观看网站| 51国产日韩欧美| 日韩成人伦理影院| 少妇人妻久久综合中文| 精品人妻熟女av久视频| h日本视频在线播放| 少妇人妻 视频| 日本一二三区视频观看| 美女被艹到高潮喷水动态| 久久热精品热| 国产成人a∨麻豆精品| tube8黄色片| 久久久久久九九精品二区国产| 爱豆传媒免费全集在线观看| 亚洲成人久久爱视频| 美女被艹到高潮喷水动态| 中文字幕av成人在线电影| 一级毛片久久久久久久久女| 蜜桃久久精品国产亚洲av| 欧美激情久久久久久爽电影| 亚洲精品日韩在线中文字幕| 2018国产大陆天天弄谢| 国语对白做爰xxxⅹ性视频网站| 国产 精品1| 小蜜桃在线观看免费完整版高清| 全区人妻精品视频| 91久久精品电影网| 亚洲aⅴ乱码一区二区在线播放| 日本三级黄在线观看| 免费观看av网站的网址| 一本—道久久a久久精品蜜桃钙片 精品乱码久久久久久99久播 | 午夜福利视频精品| 老司机影院成人| 国产淫语在线视频| 少妇人妻一区二区三区视频| 六月丁香七月| 日韩在线高清观看一区二区三区| 欧美日韩精品成人综合77777| 69av精品久久久久久| 国产综合懂色| 精品人妻偷拍中文字幕| 午夜福利视频精品| 熟女av电影| 日本wwww免费看| 亚洲欧美中文字幕日韩二区| 少妇高潮的动态图| 夫妻午夜视频| 最近最新中文字幕大全电影3| 国产熟女欧美一区二区| a级毛片免费高清观看在线播放| 国产精品伦人一区二区| 久久久久久九九精品二区国产| av在线天堂中文字幕| 国产91av在线免费观看| 免费高清在线观看视频在线观看| videossex国产| 国产午夜精品一二区理论片| 在线观看一区二区三区| 麻豆精品久久久久久蜜桃| 成人亚洲精品一区在线观看 | 免费不卡的大黄色大毛片视频在线观看| 国产片特级美女逼逼视频| 久久久久久久久久成人| 日本猛色少妇xxxxx猛交久久| 久久精品久久久久久久性| 三级男女做爰猛烈吃奶摸视频| 亚洲av福利一区| 成人二区视频| 亚洲精品视频女| 一区二区三区免费毛片| 免费看光身美女| 亚洲av.av天堂| 麻豆乱淫一区二区| 人妻夜夜爽99麻豆av| 亚洲av电影在线观看一区二区三区 | 青春草视频在线免费观看| 午夜福利在线在线| 男女边吃奶边做爰视频| 九色成人免费人妻av| 精华霜和精华液先用哪个| 亚洲国产精品专区欧美| 色播亚洲综合网| 免费看日本二区| 国产精品人妻久久久久久| av播播在线观看一区| 久久精品综合一区二区三区| av专区在线播放| 国产亚洲5aaaaa淫片| 97超视频在线观看视频| 亚洲aⅴ乱码一区二区在线播放| 成人二区视频| 精品少妇黑人巨大在线播放| 日日啪夜夜撸| 久久影院123| 国产 精品1| 在线精品无人区一区二区三 | 国产午夜精品一二区理论片| 午夜福利在线在线| 精品一区二区免费观看| 一区二区三区免费毛片| 夜夜爽夜夜爽视频| 亚洲av国产av综合av卡| 18禁裸乳无遮挡动漫免费视频 | h日本视频在线播放| 亚洲精品,欧美精品| 黄色怎么调成土黄色| 免费黄网站久久成人精品| 久久久久久久久久人人人人人人| 少妇丰满av| 一二三四中文在线观看免费高清| 久久精品国产a三级三级三级| 久久精品熟女亚洲av麻豆精品| 人体艺术视频欧美日本| 日本猛色少妇xxxxx猛交久久| 成人无遮挡网站| 国产大屁股一区二区在线视频| 九草在线视频观看| 精华霜和精华液先用哪个| 综合色av麻豆| 久久99热6这里只有精品| 亚洲色图av天堂| 久久99热这里只有精品18| 一区二区三区四区激情视频| 嘟嘟电影网在线观看| 欧美+日韩+精品| 久久热精品热| a级毛片免费高清观看在线播放| 一本一本综合久久| 在线观看一区二区三区激情| 日韩av不卡免费在线播放| 国产精品精品国产色婷婷| 最近手机中文字幕大全| 男人添女人高潮全过程视频| 1000部很黄的大片| 免费观看a级毛片全部| 青青草视频在线视频观看| 欧美日韩亚洲高清精品| 内射极品少妇av片p| 欧美少妇被猛烈插入视频| 少妇 在线观看| 欧美高清成人免费视频www| 国产乱人视频| 成人特级av手机在线观看| 久久人人爽人人片av| 在线精品无人区一区二区三 | 制服丝袜香蕉在线| 亚洲国产av新网站| 一本一本综合久久| 中文字幕久久专区| 91aial.com中文字幕在线观看| 一级二级三级毛片免费看| 久热久热在线精品观看| 白带黄色成豆腐渣| 男女边摸边吃奶| 免费大片18禁| 久久99热6这里只有精品| 三级男女做爰猛烈吃奶摸视频| 菩萨蛮人人尽说江南好唐韦庄| 日韩欧美精品免费久久| 国产高清三级在线| 91久久精品电影网| 夜夜看夜夜爽夜夜摸| 99久久精品热视频| 亚洲国产色片| 国产精品99久久久久久久久| 亚洲自拍偷在线| 在线观看免费高清a一片| 精品久久久噜噜| 国产黄色免费在线视频| 综合色丁香网| 亚洲精品乱码久久久v下载方式| av天堂中文字幕网| 亚洲真实伦在线观看| 18禁裸乳无遮挡免费网站照片| 伊人久久国产一区二区| 国产高潮美女av| 精品一区在线观看国产| 国产乱人偷精品视频| 国内少妇人妻偷人精品xxx网站| 亚洲精品国产成人久久av| 国产乱来视频区| 久久久欧美国产精品| 亚洲怡红院男人天堂| videossex国产| 国产淫语在线视频| 亚洲成人精品中文字幕电影| 免费观看无遮挡的男女| 国产 一区 欧美 日韩| 大陆偷拍与自拍| 黄片wwwwww| 禁无遮挡网站| 18+在线观看网站| 国产国拍精品亚洲av在线观看| 久久精品国产a三级三级三级| 黄片无遮挡物在线观看| 热re99久久精品国产66热6| 久久精品人妻少妇| 精品人妻熟女av久视频| 欧美精品一区二区大全| 国产人妻一区二区三区在| 亚洲欧美一区二区三区黑人 | 亚洲av一区综合| 亚洲欧美成人精品一区二区| 舔av片在线| 精品视频人人做人人爽| 日韩av免费高清视频| 真实男女啪啪啪动态图| 99久久中文字幕三级久久日本| 在线亚洲精品国产二区图片欧美 | 精品99又大又爽又粗少妇毛片| 神马国产精品三级电影在线观看| 日韩 亚洲 欧美在线| 国产免费一级a男人的天堂| 亚洲av免费高清在线观看| 国产熟女欧美一区二区| 最近手机中文字幕大全| 亚洲av日韩在线播放| 亚洲av免费高清在线观看| 欧美三级亚洲精品| 国产在视频线精品| 美女脱内裤让男人舔精品视频| 亚洲色图综合在线观看| 国产白丝娇喘喷水9色精品| 最近中文字幕高清免费大全6| 免费看不卡的av| 国产精品爽爽va在线观看网站| 日韩电影二区| 丰满乱子伦码专区| 午夜视频国产福利| 中国国产av一级| 国产精品不卡视频一区二区| 少妇裸体淫交视频免费看高清| 中文在线观看免费www的网站| 久久精品久久久久久噜噜老黄| 又爽又黄无遮挡网站| 国产熟女欧美一区二区| 国产亚洲av嫩草精品影院| 五月开心婷婷网| 午夜福利视频精品| 人体艺术视频欧美日本| 久久久色成人| 免费观看av网站的网址| 亚洲欧美中文字幕日韩二区| 男人舔奶头视频| 国产精品久久久久久精品电影小说 | 99久久九九国产精品国产免费| 黄色怎么调成土黄色| 国产一区二区三区综合在线观看 | 亚洲国产精品成人久久小说| 1000部很黄的大片| 国产老妇伦熟女老妇高清| 国产女主播在线喷水免费视频网站| 一二三四中文在线观看免费高清| 女人久久www免费人成看片| 黄色怎么调成土黄色| 人妻制服诱惑在线中文字幕| 国产成人免费无遮挡视频| 久久久久久久久久久免费av| 一区二区三区乱码不卡18| 国产av码专区亚洲av| 国产亚洲av嫩草精品影院| 国产白丝娇喘喷水9色精品| 王馨瑶露胸无遮挡在线观看| 天堂中文最新版在线下载 | 91精品一卡2卡3卡4卡| 久久影院123| 在线精品无人区一区二区三 | 欧美3d第一页| 日本色播在线视频| 久久久久久久久久久丰满| 我的女老师完整版在线观看| 成年女人在线观看亚洲视频 | 日日摸夜夜添夜夜添av毛片| 综合色丁香网| .国产精品久久| 精品久久久精品久久久| 国产高潮美女av| 乱系列少妇在线播放| 亚洲精品自拍成人| 国产精品久久久久久久久免| 亚洲美女视频黄频| 亚洲图色成人| 又粗又硬又长又爽又黄的视频| 国产 一区 欧美 日韩| 我的女老师完整版在线观看| 成人一区二区视频在线观看| 又粗又硬又长又爽又黄的视频| 青春草视频在线免费观看| a级毛片免费高清观看在线播放| 成人欧美大片| 狂野欧美激情性xxxx在线观看| 国产午夜精品一二区理论片| 久久99热这里只频精品6学生| 日本熟妇午夜| av卡一久久| 99视频精品全部免费 在线| 国产精品嫩草影院av在线观看| 国产免费一级a男人的天堂| 最近2019中文字幕mv第一页| 亚洲精品日韩av片在线观看| 新久久久久国产一级毛片| 在线观看三级黄色| 免费看不卡的av| 亚洲av一区综合| 国产白丝娇喘喷水9色精品| 亚洲欧美清纯卡通| 久久久亚洲精品成人影院| 色视频www国产| 国产亚洲5aaaaa淫片| 女的被弄到高潮叫床怎么办| 中国三级夫妇交换| 看黄色毛片网站| 日本av手机在线免费观看| h日本视频在线播放| av卡一久久| 国产午夜福利久久久久久| av国产免费在线观看| 男插女下体视频免费在线播放| 国产精品久久久久久久电影| 久久久久国产精品人妻一区二区| 久久99蜜桃精品久久| 99热这里只有精品一区| 一本—道久久a久久精品蜜桃钙片 精品乱码久久久久久99久播 | 午夜福利在线在线| 美女高潮的动态| 美女主播在线视频| 免费黄频网站在线观看国产| 亚洲精品国产色婷婷电影| 日本熟妇午夜| 国模一区二区三区四区视频| 天天躁夜夜躁狠狠久久av| 婷婷色av中文字幕| 久久久久久伊人网av| 免费人成在线观看视频色| 免费大片18禁| 亚洲成色77777| 色婷婷久久久亚洲欧美| 天天躁夜夜躁狠狠久久av| 嘟嘟电影网在线观看| 亚洲av男天堂| 人人妻人人看人人澡| 中国国产av一级| 成人免费观看视频高清| 99久国产av精品国产电影| 国产精品国产三级专区第一集| 欧美3d第一页| 九九久久精品国产亚洲av麻豆| 免费黄频网站在线观看国产| av一本久久久久| 精品酒店卫生间| 在现免费观看毛片| 亚洲欧美日韩无卡精品| 免费黄频网站在线观看国产| 亚洲av电影在线观看一区二区三区 | 亚洲欧洲国产日韩| 一级毛片aaaaaa免费看小| 视频中文字幕在线观看| 美女脱内裤让男人舔精品视频| 国产片特级美女逼逼视频| 亚洲成人精品中文字幕电影| av线在线观看网站| 欧美少妇被猛烈插入视频| 黄色一级大片看看| 亚洲国产精品成人久久小说| 精华霜和精华液先用哪个| 亚洲久久久久久中文字幕| 免费不卡的大黄色大毛片视频在线观看| tube8黄色片| av一本久久久久| 看十八女毛片水多多多| 亚洲国产精品成人久久小说| 日韩中字成人| 免费人成在线观看视频色| 69人妻影院| 寂寞人妻少妇视频99o| 99热全是精品| 久久久久久伊人网av| 国产又色又爽无遮挡免| 国产免费一区二区三区四区乱码| 国产av码专区亚洲av| 免费在线观看成人毛片| av国产免费在线观看| 乱码一卡2卡4卡精品| 91久久精品国产一区二区三区| 亚洲,欧美,日韩| 亚洲国产精品成人综合色| 伦理电影大哥的女人| 国产老妇伦熟女老妇高清| 亚洲天堂av无毛| 精品久久久噜噜| 日本av手机在线免费观看| 永久网站在线| 女人被狂操c到高潮| 免费不卡的大黄色大毛片视频在线观看| 精品一区二区三卡| 黄色怎么调成土黄色| 欧美日韩综合久久久久久| 99热全是精品| 99久久中文字幕三级久久日本| 男女那种视频在线观看| 久久ye,这里只有精品| 三级经典国产精品| 亚洲成人久久爱视频| 国内少妇人妻偷人精品xxx网站| 美女高潮的动态| 亚洲综合色惰| 亚洲精品一二三| 插阴视频在线观看视频| 91aial.com中文字幕在线观看| 97在线人人人人妻| 深爱激情五月婷婷| 全区人妻精品视频| 亚洲精品第二区| 男女下面进入的视频免费午夜| tube8黄色片| 欧美一区二区亚洲| 亚洲av二区三区四区| 国产美女午夜福利| 亚洲精品成人久久久久久| 成人一区二区视频在线观看| 91在线精品国自产拍蜜月| 纵有疾风起免费观看全集完整版| 啦啦啦啦在线视频资源| 亚州av有码| 人妻一区二区av| 日本欧美国产在线视频| 久久精品国产自在天天线| 久久久精品免费免费高清| 国产免费一区二区三区四区乱码| av网站免费在线观看视频| 丝瓜视频免费看黄片| 91久久精品国产一区二区三区| 三级经典国产精品| 久久久久性生活片| 国产伦精品一区二区三区视频9| 午夜免费观看性视频| 六月丁香七月| 日本与韩国留学比较| 最近最新中文字幕大全电影3| 我要看日韩黄色一级片| 欧美xxⅹ黑人| 国产爽快片一区二区三区| 亚洲精品国产色婷婷电影| 高清午夜精品一区二区三区| 国产精品久久久久久久久免| 2018国产大陆天天弄谢| 久久女婷五月综合色啪小说 | 欧美日韩综合久久久久久| 啦啦啦啦在线视频资源| 国产国拍精品亚洲av在线观看| 免费大片黄手机在线观看| 国产黄色视频一区二区在线观看| 99热国产这里只有精品6| 久久久精品免费免费高清| 一区二区三区乱码不卡18| 日本色播在线视频| 可以在线观看毛片的网站| 好男人视频免费观看在线| 国产亚洲5aaaaa淫片| 国产色婷婷99| 91在线精品国自产拍蜜月| 亚洲精品自拍成人| 乱码一卡2卡4卡精品| 91在线精品国自产拍蜜月| 22中文网久久字幕| 国产日韩欧美亚洲二区| 国产综合精华液| 久久热精品热| 丝袜喷水一区| 国产高清有码在线观看视频| 日本黄大片高清| 18禁动态无遮挡网站| 97在线视频观看| 伦理电影大哥的女人| 国产亚洲一区二区精品| 大片电影免费在线观看免费| 在线a可以看的网站| 秋霞在线观看毛片| 日日撸夜夜添| www.av在线官网国产|