• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Privacy-Preserving Federated Deep Learning Diagnostic Method for Multi-Stage Diseases

    2024-03-23 08:17:08JinboYangHaiHuangLailaiYinJiaxingQuandWanjuanXie

    Jinbo Yang ,Hai Huang ,Lailai Yin ,Jiaxing Qu and Wanjuan Xie

    1School of Computer Science and Technology,Harbin University of Science and Technology,Harbin,150080,China

    2Shanghai Futures Information Technology Co.,Ltd.,Shanghai,201201,China

    3Heilongjiang Province Cyberspace Research Center,Harbin,150001,China

    4Information Network Engineering and Research Center,South China University of Technology,Guangzhou,510641,China

    ABSTRACT Diagnosing multi-stage diseases typically requires doctors to consider multiple data sources,including clinical symptoms,physical signs,biochemical test results,imaging findings,pathological examination data,and even genetic data.When applying machine learning modeling to predict and diagnose multi-stage diseases,several challenges need to be addressed.Firstly,the model needs to handle multimodal data,as the data used by doctors for diagnosis includes image data,natural language data,and structured data.Secondly,privacy of patients’data needs to be protected,as these data contain the most sensitive and private information.Lastly,considering the practicality of the model,the computational requirements should not be too high.To address these challenges,this paper proposes a privacy-preserving federated deep learning diagnostic method for multi-stage diseases.This method improves the forward and backward propagation processes of deep neural network modeling algorithms and introduces a homomorphic encryption step to design a federated modeling algorithm without the need for an arbiter.It also utilizes dedicated integrated circuits to implement the hardware Paillier algorithm,providing accelerated support for homomorphic encryption in modeling.Finally,this paper designs and conducts experiments to evaluate the proposed solution.The experimental results show that in privacy-preserving federated deep learning diagnostic modeling,the method in this paper achieves the same modeling performance as ordinary modeling without privacy protection,and has higher modeling speed compared to similar algorithms.

    KEYWORDS Vertical federation;homomorphic encryption;deep neural network;intelligent diagnosis;machine learning and big data

    1 Introduction

    In medicine,multi-stage diseases refer to diseases that can be divided into different stages or periods during their development process.These stages may have different pathological characteristics,clinical manifestations,and prognoses.The staging of multi-stage diseases can be based on physiological processes,pathological features,clinical symptoms,imaging results,and other aspects of the disease.The diagnosis of multi-stage diseases also requires multiple types of data to support it,such as clinical symptoms,physical signs,laboratory test data,imaging data,histopathological data,and genetic data.In addition,doctors need to have comprehensive medical knowledge and experience.To assist doctors in diagnosis,improve diagnostic accuracy,speed up the diagnosis process,and provide personalized treatment,artificial intelligence algorithms can be used for intelligent diagnosis and treatment.Traditional artificial intelligence algorithms require the consolidation of patient data from different sources for modeling and prediction,which can easily lead to the leakage of patients’medical records,genetic information,and other sensitive data.To protect patients’privacy data in intelligent diagnosis,we need to combine federated learning techniques to design a federated intelligent diagnosis method for multi-stage diseases.

    The concept of federated learning was first introduced by Google in 2016 [1].Since then,it has received continuous attention and become a research hotspot in the field of machine learning.It can realize the safe sharing of data between different organizations.Federal learning includes horizontal federated learning(HFL),vertical federated learning(VFL)and federal transfer learning(FTL).

    For vertical federated learning,it can be divided into two architectures:the architecture with coordinator and the architecture with de coordinator.Yang et al.[2]implemented a vertical federal logical regression algorithm with a coordinator,which approximated the loss function and gradient function with second-order Taylor,and then used homomorphic encryption to calculate privacy protection.This method uses the first-order random gradient descent algorithm,which requires a large number of communication rounds.Therefore,in order to reduce the communication cost,Yang et al.[2]proposed a vertical logic regression framework based on quasi Newton method.However,these two methods are aimed at binary classification.In order to expand vertical federated learning,Feng et al.[3]proposed a VFL framework for multiple participants and multiple classifications.Yang et al.[4]proposed a vertical logic regression framework for de coordinator,which effectively protects privacy and improves the accuracy of the classifier.Hardy et al.[5]proposed a three-party end-to-end logistic regression,which consists of a trusted arbiter and two other parties,where the coordinator’s tasks include computing the training loss and generating homomorphic encryption key pairs for privacy protection.

    However,the arbiter-based architecture poses privacy risks as the arbiter has the potential to leak information about the participating parties.Furthermore,the existing algorithms for arbiterfree architectures are mainly focused on conventional machine learning,such as logistic regression and boosting trees.However,the diagnosis of multi-stage diseases requires the integration of various multimodal data for inference.Therefore,this paper aims to investigate an arbiter-free algorithm for vertical deep neural networks,eliminating the need for a third-party arbiter and enabling collaborative participants to engage in federated training of deep neural networks.Our main contributions are as follows.

    This paper proposes a vertical federated deep neural network approach and provides a detailed description of its three-layer model’s secure forward and backward propagation processes.By introducing homomorphic encryption during the propagation process,it can operate without the involvement of a third-party arbiter.

    This paper proposes a hardware acceleration solution for our federated deep neural network is designed based on finite field arithmetic chips.This approach effectively addresses the challenges of multimodal data in the diagnosis of multi-stage diseases,the privacy leakage issue in vertical federated learning,and the computational burden of encryption operations in federated learning.

    The method referred to in this article is named the Vertical Deep Neural Network(VDNN).The article has conducted complexity analysis and security assessment of this method,demonstrating its feasibility and security.Through comparative experiments with the method in reference[6],this paper proves the better performance of the proposed method in terms of communication overhead and program execution time.The method presented in this paper achieves a more balanced computation load between the Guest and Host parties,resulting in improved performance.

    The rest of this paper is organized as follows.Section 2 introduces the preliminary knowledge of homomorphic encryption and the basic knowledge of deep neural networks.Section 3 discusses the specific method of vertical deep neural networks without an arbiter and provides complexity and security analysis.Then,in Section 4,we present a detailed comparative experiment between our federated modeling VDNN model and the model in reference[6]on commonly used datasets.Finally,Section 5 summarizes this work.

    2 Related Works

    Alzheimer’s disease is a typical multi-stage disease.Saleem et al.[7]summarized the application of machine learning in the diagnosis of Alzheimer’s disease and found that neuroimaging data,electroencephalogram(EEG)data,and genomic data can improve the accuracy of Alzheimer’s disease diagnosis.These data are large-scale and high-dimensional multimodal data,and ordinary machine learning methods struggle to handle such complex multimodal data.Therefore,neural networks that can directly process multimodal data are necessary for this type of multi-stage disease.Furthermore,to achieve a privacy-preserving intelligent diagnostic method,it is crucial to focus on implementing privacy-preserving deep neural networks.

    Liu et al.[8]presented the federated forest algorithm,which is a VFL method with coordinator based on random forest.The structure of the global federated forest model is stored in a decentralized manner,the central server retains the complete structure information of the global model,and the node information is stored in each participant in a decentralized manner.The method presented in this paper eliminates the central server,reducing the risk of data leakage.Zhang et al.[6]decomposed the forward and backward propagation of neural networks into four distinct steps and proposed a privacy-preserving architecture that enables collaborative parties to effectively federated train deep learning models.For ease of expression,in this paper,this method is referred to as the Fate Deep Neural Network(FDNN).

    Our work focuses on the forward and backward propagation processes of deep neural networks.We introduce encrypted noise in both the forward and backward propagation processes and use homomorphic encryption methods for parameter updates.This allows vertical deep neural networks to achieve lossless joint modeling while protecting data privacy.Additionally,our approach is insensitive to the neural network structure and can effectively solve various multi-class and non-linear problems.Furthermore,we implement a hardware-level Paillier algorithm that supports decimals and negative numbers based on a large finite field arithmetic chip,enhancing the security of encryption and decryption during the interaction process.

    3 Preliminaries

    3.1 Homomorphic Encryption

    In the vertical federated learning scenario,both the Guest and the Host in the collaborative modeling process each hold part of the private data,and the traditional encryption mechanism cannot perform computing operations on the undecrypted encrypted data.Therefore,in order to calculate the gradient required for model training,the Guest and Host need to disclose some private information.Homomorphic encryption(HE)[9,10],an asymmetric encryption technique widely used in privacy computing,can solve this problem by allowing any third party to operate on encrypted data.We perform calculations on the homomorphically encrypted data,and after decrypting the calculation results,the decrypted results match the results of the calculations performed directly on the plaintext.The concept of HE was first proposed in 1978[11]to protect the private data of banks.According to the number and types of ciphertext calculations,homomorphic encryption can be divided into: Partial Homomorphic Encryption (PHE),Somewhat Homomorphic Encryption (SHE),Fully Homomorphic Encryption(FHE).

    The PHE technique only supports a single type of homomorphic computation (additive or multiplicative homomorphism),RSA[12]is a widely used PHE algorithm that follows multiplicative homomorphism and its security is based on the factorization of the product of two large prime numbers question.GM proposed the first probabilistic public key encryption scheme Goldwasser-Micali[13],which is based on the quadratic residual difficulty problem with multiplicative homomorphism.In 1985,Elgama[14]proposed a new public key encryption scheme,which improved the original Diffie-Hellman key exchange algorithm,and its security was based on the discrete logarithm problem.Benaloh [15] proposed an extension of the GM algorithm,using the encryption mode of block encryption instead of bitwise encryption.In 1999,Paillier proposed a SHE algorithm based on the composite residual problem[16],which satisfies the additive homomorphism.This algorithm is also a commonly used algorithm in the follow-up research field.

    There are four definitions for the homomorphic properties of homomorphic encryption algorithms:

    Additive homomorphism:without needing to know the value ofx,y,Enc(x+y)can be calculated byEnc(x)andEnc(y),and satisfiesC(Enc(x),Enc(y))=Enc(x+y)that the functionCrepresents any operation.The“ENC”signifies the corresponding ciphertext after encryption.

    Multiplicative homomorphism: without needing to know the value ofx,y,Enc(x×y)can be calculated byEnc(x)andEnc(y),and it is satisfiedC(Enc(x),Enc(y))=Enc(x×y)that the functionCrepresents any operation.

    Hybrid multiplicative homomorphism:without needing to know the value ofx,Enc(x×y)can be calculated fromEnc(x)andy,and satisfyC(Enc(x),y)=Enc(x×y)that the function C represents any operation.

    Fully homomorphic encryption FHE scheme:if the HE encryption function Enc satisfies consistency for all Boolean circuits,then the scheme is called a fully homomorphic encryption FHE scheme.

    3.2 Paillier Encryption

    Among different homomorphic encryption algorithms,the SHE schemes has excellent performance in terms of execution efficiency and construction complexity,so it is widely used.Meanwhile,the vertical LR models mainly involve addition and multiplication when calculating gradients for parameter update.Therefore,this paper adopts the famous SHE algorithm Paillier [16],which is based on the difficult problem of compound residual classes and has been widely used in electronic voting and biometric applications,and its encryption and decryption efficiency can be controlled at the millisecond level,which can meet the encryption and decryption operations and ciphertext computing operations of ciphertext in this paper.The encryption and decryption mechanism of Paillier homomorphic encryption algorithm is as follows:

    Key generation process:the Paillier algorithm is a homomorphic encryption algorithm under the public key encryption system,so before using the Paillier algorithm,a pair of public and private keys needs to be constructed.First,randomly select two sums of large prime numberspandq,and the greatest common divisor ofpqand(p-1)(q-1)is 1.DefineN=pq,λ as the least common multiple of(p-1)(q-1).Then randomly select an integerg∈,let μ=(L(gλmod N2))-1,where the functionL(x)=(x-1)/N.The final pair of public key and private key is generated,where the public keyPkis represented as(N,g)and the private keySkis represented as λ.

    Encryption process:after generating the public/private key pair,we can use the Paillier public keyPk=(N,g)for encryption.First,define the plaintext to be encrypted bymPaillier 0≤m≤N.Select a random integerr,and match 0

    It is worth noting here that for the same plaintext,Paillier’s algorithm can get different ciphertexts,which makes it have the semantic security of ciphertexts.This is because in the encryption process,even if the same public keyPk=(N,g)is used for encryption,when we pick different random numbersr,the obtained ciphertext will also be different,but decryption can restore the original plaintextm.

    Decryption process:using the private keySk=λ,the ciphertextccan be decrypted to obtain the plaintext:

    Therefore,Paillier’s algorithm has additive homomorphism and mixed multiplicative homomorphism,that is,for a given ciphertextE(m1),E(m2)and integerk∈ZNhave:

    3.3 Deep Neural Network and Backpropagation

    Deep Neural Networks(DNN)simulate the working principles of the biological neural system.It treats inputs as electrical signals between neuron connections,where the importance of different connections corresponds to the weight values of different inputs.DNN can be divided into three categories:(1)Feedforward deep networks,constructed by multiple encoder layers,typical examples include multilayer perceptron and convolutional neural networks.(2) Feedback deep networks,constructed by multiple decoder layers,typical examples include HSC level sparse coding grid and deconvolutional networks.(3)Bidirectional deep networks,each layer can contain either an encoder or a decoder,or a combination of both,typical examples include stacked autoencoders,deep Boltzmann machines,and DBN deep belief networks.For feedforward deep networks,the training process of the model can be divided into two steps:forward propagation and backward propagation[17].

    Forward propagation is the prediction process of the DNN model for the input.The input information flows in the same direction,from the input layer through the hidden layers,and finally reaches the output layer.There are no closed loops in the network model structure.During the forward propagation process,each neuron in the neural network takes the dot product of the outputs of all the neurons in the upper layer and its own weight vector,and then passes through an activation function to obtain the output of that neuron.

    The backpropagation algorithm(BP)is the process of updating model parameters to minimize the value of the objective function.By using the chain rule of differentiation,the partial derivatives of the loss function with respect to each model parameter can be calculated,and these derivatives are then used to update the corresponding parameters.After iteratively performing forward propagation and backpropagation,the loss function gradually converges.Therefore,before using the BP algorithm,it is necessary to determine the loss function to measure the difference between the predicted labels of the model and the true labels of the samples.

    4 Vertical Deep Neural Network without Arbiter

    This paper proposes a vertical deep neural network algorithm,called Vertical DNN (VDNN),which can effectively achieve lossless joint modeling while protecting data privacy.VDNN is insensitive to the neural network structure and can address multi-class and non-linear problems.

    In the scenario of vertical federated linear modeling,the participating party Host possesses data featuresXH=x1,x2,...,xk,while the initiating party Guest possesses data featuresXG=xk+1,xk+2,...,xI,where k represents the number of data features of the participating party and(I-k)represents the number of data features of the initiating party.Both parties jointly train a vertical DNN model consisting of a bottom layer,an interaction layer,and a top layer,as shown in Fig.1.We can perform secure data exchange at the interaction layer to address the issue of information leakage during the forward and backward propagation processes.For the parameters in the neural network model,the Host party holds the interaction layer parametersWHand keeps them confidential from the Guest party.The Guest party holds the interaction layer parametersWGand the parameters of the top layer model,keeping them confidential from the Host party.Here,the parametersWG=w(k+1)1,...,w(k+1)J,w(k+2)1,...,w(k+2)J,...,wI1,...,wIJand the parametersWH=w11,...,w1J,w21,...,w2J,...,wk1,...,wkJ.J represents the total number of neurons in the input layer of the top layer model,andwijrepresents the linear coefficient from the i-th neuron in the interaction layer to the j-th neuron in the input layer of the top layer model.

    Figure 1:Vertical DNN model

    The secure forward propagation and secure backward propagation processes of the VDNN algorithm will be presented below,and their timing is illustrated in Figs.2 and 3.

    Figure 3:VDNN secure backward propagation timing diagram

    The improved secure forward propagation process is as follows:

    a) The Host and Guest parties respectively useXHandXGas inputs to their own bottom layer models and perform the forward propagation process of the bottom layer models,obtaining the outputsαHandαGof the bottom layer models.

    b) The Guest party computes the encrypted accumulated noise [εacc]Gand sends it to the Host party.

    c) The Host party receives and computes the encrypted weighted value(WH+[εacc]G)αHand sends it to the initiating party Guest.

    d) The Guest party decrypts the encrypted weighted value(WH+[εacc]G)αHusing the private key of the Host party,obtaining the true weighted value(WH+εacc)αHof the Host party.Therefore,the Guest party can compute the true weighted valueWα=(WH+εacc)αH+WGαGand use the computed result to obtain the output of the interaction layer’s activation function,and then continue with the forward propagation process of the top layer model.

    Improved secure backward propagation process is as follows:

    a) The Host party computes the encrypted data [αH]Hand WHH and sends them to the Guest party.

    b) The Guest party receives and computes the encrypted gradient value [gH]H=[?L/?WH]H=δinteract·αHHin the Host party’s interaction layer.It generates random noiseε,computes the encrypted gradient value with noise[gH]H+ε/η=δinteract·αHH+ε/η,and the encrypted bottom layer intermediate error b.[δbottom]H=δinteract·([WH]H+[εacc]H).It sends [gH]Hand [δbottom]Hto the Host party and updates the accumulated noise[εacc]H+=ε,where η is the learning rate.

    c) The Host party receives the encrypted gradient value with noise[gH]H=[?L/?WH]H+?/η=δinteract·αHH+?/η,decrypts it to obtain the gradient value with noisegH=?L/?WH+ε/η=δinteract·αH+ε/η,and updates the parameters in the Host party’s interaction layer using the decrypted result: [WH]H=[WH]H-η·gH.Then,it performs the bottom layer backward propagation process usingδbottom.

    d) The Guest party computes its own gradient value in the interaction layergG=?L/?WG=δinteract·αGand updates the parameters in the Host party’s interaction layer using the computed result:WG=WG-η·gG.Then,it performs the bottom layer backward propagation process usingδbottom=δinteract·WG.

    4.1 Hardware Acceleration for Paillier

    The computation of the Paillier algorithm involves a significant amount of finite field arithmetic,including modular addition,modular multiplication,and modular exponentiation of large integers and prime numbers.When the key length of Paillier is set to 1024 bits,the CPU needs to perform modular addition,modular multiplication,and modular exponentiation operations on 2048-bit data.Since the CPU is not specifically designed for large finite field arithmetic operations,these computations consume a substantial amount of CPU power and slow down the modeling process.Therefore,we propose the use of dedicated finite field arithmetic chips to accelerate the computations of the Paillier algorithm.

    The Paillier hardware module is shown in Fig.4.Our hardware module communicates with the host computer via the Peripheral Component Interconnect express (PCIe) interface.The PCIE communication module receives algorithm instructions and data from the host computer,performs byte order conversion to convert big-endian data to little-endian data,and then sends the data and instructions to the Paillier algorithm controller.

    The Paillier algorithm controller generates corresponding control byte streams and data byte streams based on the received encryption and decryption instructions,and sends the byte streams to the Montgomery algorithm hardware module.

    In the Montgomery algorithm hardware module,the module performs specific computational steps on the data based on the control bytes,thereby completing the Paillier encryption and decryption operations.

    Finally,the module returns the computed result data byte stream to the Paillier algorithm controller,which then sends it back to the host computer via the PCIE module.With this,we have successfully completed the hardware implementation of the Paillier algorithm.

    Figure 4:Hardware acceleration for Paillier

    4.2 Complexity Analysis

    Equations in display format are separated from the paragraphs of the text.Equations should be flushed to the left of the column.Equations should be made editable.Displayed equations should be numbered consecutively,using Arabic numbers in parentheses.See Eq.(1)for an example.The number should be aligned to the right margin.

    The VDNN model consists of a bottom layer,an interaction layer,and a top layer model.The bottom layer model involves participants executing the parameter update process locally,using their own local data features as input.This layer plays a role in extracting data features and reducing data dimensions.The output of the bottom layer model serves as the input for the interaction layer model.Therefore,unlike the VLR algorithm,the VDNN model is not sensitive to the feature dimensions of the input dataset.Additionally,the top layer model is not sensitive to the DNN model architecture.It can choose any loss function,activation function,and optimizer.This is because during the backpropagation process of the top layer model,according to the chain rule of differentiation,the error of each layer is related to the error and weight of the previous layer.

    During the training of the entire vertical federated model,participants perform encrypted computations and exchanges in the interaction layer.Therefore,the computational complexity and communication efficiency of the vertical deep neural network(VDNN)algorithm depend to some extent on the model structure of the interaction layer,specifically the values ofinteract_inandinteract_out,which represent the number of input and output nodes in the interaction layer,respectively.The following analysis will focus on the computational complexity and communication complexity of the interaction layer in the VDNN algorithm.

    For ease of description,let us assume that the batch size for each iteration isbatch_size=n,the number of output nodes in the bottom layer model isbottom_outG=bottom_outH=m,theinteract_inof the interaction layer model is 2m,andinteract_outisl.Letαandβrepresent the computational cost of performing one 1024-bit exponentiation and one 2048-bit multiplication operation during the encryption and decryption process,respectively.Letγandδrepresent the computational cost of performing one homomorphic addition and one homomorphic scalar multiplication operation.

    In the secure forward propagation algorithm,the computational cost of encryption and decryption operations for the Guest ismlα+(m+n)lβ.In the secure backward propagation algorithm,the computational cost of encryption and decryption operations for the Host is(n+l)mα+(n+2l)mβ.Throughout the VDNN algorithm,the Guest needs to perform 2mlhomomorphic addition operations and 2nmlhomomorphic scalar multiplication operations,while the Host needs to performmlhomomorphic addition operations andnmlhomomorphic scalar multiplication operations.

    In the FDNN algorithm,the Host is responsible for encryption and decryption operations on intermediate results,while the Guest performs homomorphic computations in the ciphertext domain.The computational time complexity of both algorithms is shown in Table 1.It can be seen that the VDNN algorithm has a more balanced computational time complexity,with comparable computational tasks for both the Guest and the Host.This eliminates the need for both parties to wait for each other’s computation results during the model update process,resulting in shorter program execution time.

    Table 1:Time complexity of VDNN algorithm

    In each iteration of the VDNN algorithm,the number of ciphertexts that need to be transmitted between participants is 2nm+3ml+nl.It can be observed that the communication cost and computational cost of the VDNN algorithm are influenced by the parametersm,l,andn,but are not sensitive to the input data dimension.At the same time,the number of output nodes m in the bottom layer model for both the Guest and the Host also affects the feature extraction performance of the bottom layer model,which in turn affects the classification performance of the VDNN model.Therefore,when setting the hyperparametersbottom_outG,bottom_outH,andinteract_out,a balance needs to be struck between communication efficiency and classification performance.

    4.3 Security Analysis

    In the forward propagation and backward propagation processes,the outputs of the bottom layer model,αHandαG,as well as the intermediate errorδinteractin the interaction layer,need to be encrypted or noise-added before being sent to other participants.Otherwise,malicious attackers may collect these intermediate results and infer the privacy information of other participants.In this section,we will analyze the security of the vertical deep neural network algorithm VDNN from two aspects.

    Security of Guest’s label data and feature data.Secure forward propagation and secure backward propagation constitute the iterative update process of the VDNN algorithm.Step 2 of secure forward propagation and step 3 of secure backward propagation involve sending data from the Guest to the Host.In step 2 of secure forward propagation,the Host learns the encrypted accumulated noise [εacc]G,removes the noise from its own parameters,and sends the real encrypted weighted value(WH+[εacc]G)αHto the initiating party.Since the Host does not know the Guest’s private key information,it cannot decrypt the ciphertext.In step 3 of secure backward propagation,after decrypting the received encrypted values,the Host can obtain=?L/?WH+ε/η=δinteract·αH+ε/ηandδbottomH=δinteract·,where δinteractis the intermediate error calculated by the Guest in the interaction layer,εis the random noise added by the Guest,andεaccis the cumulative value ofε.All three values are kept secret from the Host,so the Host cannot determine the value of δinteractby solving a system of linear equations.

    Security of Host’s feature data.Step 3 of secure forward propagation and step 2 of secure backward propagation involve sending data from the Host to the Guest.In step 3 of secure forward propagation,the Guest decrypts and obtains the real weighted value(WH+εacc)αHfrom the Host.Since αHis the output of the Host’s bottom layer model and the Guest does not know the value ofWH,the Guest cannot infer any additional information.In step 2 of secure backward propagation,the Host calculates the encrypted data[αH]Handto send to the initiating party.Since the Guest does not have the Host’s private key information,it cannot decrypt the data and can only perform ciphertext computations,ensuring security.

    5 Experimental Results and Analysis

    5.1 Experimental Datasets

    To test the performance of the VDNN methods,experiments were conducted on two widely used datasets.Each dataset was divided into two parts based on their features,with Guest and Host having a subset of features each,and only Guest having access to the labels.The datasets are:

    Dataset 1:Default credit dataset,which collected credit card data from a Taiwan bank from April to September 2005.It is a binary classification problem on default payment,with 30,000 samples and 24 data features,including payment history,demographic factors,credit data,billing information,etc.The dataset is split longitudinally,with Guest having 13 data features and labels,and Host having 10 data features.

    Dataset 2:Vehicle scale dataset,which is divided into four categories based on the vehicle contour data features.It consists of 846 data samples and 18 data features related to vehicle contours.After vertical splitting the dataset,the Guest has 9 data features and labels,while the Host has 9 data features.

    5.2 Experimental Design

    The paper implements the vertical deep neural network algorithm VDNN without an arbiter and the heterogeneous neural network algorithm FDNN from [6] as a control using TensorFlow.In the neural network model,bottom_outGandbottom_outHrepresent the number of output nodes in the bottom layer models of the Guest and Host,respectively.interact_inrepresents the number of input nodes in the interaction layer,and interact_out represents the number of output nodes in the interaction layer.Therefore,interact_in=bottom_outG+bottom_outH.The paper setsbottom_outG=bottom_outH=6 for the bottom layer models,interact_in=12,andinteract_out=4 for the interaction layer model.The optimizer chosen is Nadam,with hyperparameterslr=0.001 andclipnorm=1.For dataset 1,the batch size is set to 500,and the maximum number of iterations isMAX=5.For dataset 2,the batch size is 846,and the maximum number of iterations isMAX=500.

    The experimental environment for the models was a computer with 3.10 GHz (8 CPUs) and 16 GB RAM.Different models and hyperparameters can achieve different classification results for classification tasks,but this paper focused on the overall system and did not focus on the performance of the models themselves.Therefore,the same hyperparameters were used to compare different solutions horizontally.

    Common evaluation metrics for machine learning models include accuracy,precision,recall,and F1-Score,etc.In addition,as a federated learning model,continuous communication is often required among the participating parties to exchange model update information.Therefore,the total communication volume is also an important metric to consider,as it will affect the network cost and minimum bandwidth required for deploying the model in practice.Therefore,this paper evaluates and compares the performance of the VDNN and FDNN algorithms based on metrics such as accuracy,AUC,F1-Score,program running time,and the total transmission volume (TV) between the participating parties during the model training process.

    The total transmission volumeTV=SendG+ReceiveG.whereSendGandReceiveGrepresent the total amount of data sent and received by the Guest,respectively,measured in MB.A smaller value of the total transmission volume indicates higher communication efficiency of the algorithm.

    5.3 Experimental Analysis

    We compared and analyzed the performance of the VDNN and FDNN algorithms based on metrics such as accuracy,AUC,F1-Score,program running time,and the total transmission volume(TV) between the participating parties during the model training process.The results showed that under the same model parameters and experimental environment,the VDNN method achieved better classification performance,faster running time,and higher communication efficiency.Table 2 presents the classification comparison results of the VDNN and FDNN models on Dataset 1 and Dataset 2.Figs.5–7 show the convergence curves of accuracy,loss function,and F1-Score for both models during the model training process.It can be observed that the VDNN model avoided local optima on Dataset 2,resulting in higher classification accuracy.From Table 3,it can be seen that the VDNN model required 6 interactions between the Guest and Host for each iteration,while the FDNN model required 7 interactions.The total transmission volume(TV)and program running time of the VDNN model were both superior to those of the FDNN model.

    Table 2:Experimental results of VDNN model

    Table 3:Experimental results of performance comparison

    Figure 5:Comparison of ACC curve for VDNN

    Figure 6:Comparison of loss curve for VDNN

    Figure 7:Comparison of F1 curve for VDNN

    6 Conclusions

    This paper firstly introduces the concept of homomorphic encryption algorithm,and then focuses on the key generation,encryption and decryption process and homomorphic properties of Paillier encryption algorithm,and gives the correctness proof of Paillier algorithm.Furthermore,this paper introduces the forward and backward propagation algorithms of deep neural networks,detailing the steps of adding noise and applying homomorphic encryption in both forward and backward propagation.The security and complexity of the method are analyzed.Finally,comparative experiments are conducted between the proposed method and the FDNN method.The experimental results demonstrate that the VDNN method achieves better classification performance,faster runtime,and superior communication efficiency.The VDNN method can be applied to federated modeling of multimodal data in medical scenarios,which is highly beneficial for intelligent diagnosis of multi-stage diseases.However,even with chip acceleration,the modeling speed of the VDNN method still lags behind that of traditional centralized data modeling.Future work should focus on further enhancing the computational speed of vertical federated modeling.

    Acknowledgement:The authors thank the support from the central government and the Harbin Manufacturing Technology Innovation Talent Project.

    Funding Statement:This research was funded by the National Natural Science Foundation,China(No.62172123),the Key Research and Development Program of Heilongjiang(Grant No.2022ZX01A36),the Special Projects for the Central Government to Guide the Development of Local Science and Technology,China(No.ZY20B11),the Harbin Manufacturing Technology Innovation Talent Project(No.CXRC20221104236).

    Author Contributions:The authors confirm contribution to the paper as follows:Conceptualization,Jinbo Yang and Lailai Yin;Data curation,Wanjuan Xie;Formal analysis,Jinbo Yang;Funding acquisition,Jiaxing Qu;Investigation,Wanjuan Xie;Methodology,Jinbo Yang and Lailai Yin;Project administration,Jiaxing Qu;Resources,Hai Huang;Software,Jinbo Yang and Lailai Yin;Supervision,Hai Huang;Validation,Jinbo Yang;Writing–original draft,Jinbo Yang;Writing–review &editing,Hai Huang.

    Availability of Data and Materials:The data comes from publicly available datasets on Kaggle,including the Default of Credit Card Clients Dataset and the vehicle scale Data Set.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    国产高清激情床上av| 一本一本综合久久| 中文字幕人妻熟人妻熟丝袜美 | 国产伦人伦偷精品视频| 亚洲成a人片在线一区二区| 久久国产精品人妻蜜桃| 麻豆成人av在线观看| 精品久久久久久,| 琪琪午夜伦伦电影理论片6080| 亚洲精华国产精华精| 两人在一起打扑克的视频| 可以在线观看毛片的网站| 变态另类成人亚洲欧美熟女| 亚洲美女视频黄频| 精品久久久久久久末码| 精品人妻一区二区三区麻豆 | 久久久久久久午夜电影| 无人区码免费观看不卡| 国产欧美日韩精品一区二区| 亚洲一区二区三区色噜噜| 亚洲片人在线观看| 久久天躁狠狠躁夜夜2o2o| 国产单亲对白刺激| 两性午夜刺激爽爽歪歪视频在线观看| 亚洲精品一区av在线观看| 19禁男女啪啪无遮挡网站| 男插女下体视频免费在线播放| 熟女少妇亚洲综合色aaa.| 久久久久免费精品人妻一区二区| 老汉色∧v一级毛片| 香蕉av资源在线| 丝袜美腿在线中文| 内地一区二区视频在线| 熟女少妇亚洲综合色aaa.| 国产探花极品一区二区| 一级a爱片免费观看的视频| 舔av片在线| 麻豆成人午夜福利视频| 国产成人av激情在线播放| 一级黄色大片毛片| www日本黄色视频网| 最近视频中文字幕2019在线8| 99精品久久久久人妻精品| xxx96com| av在线蜜桃| 久久精品国产清高在天天线| 18美女黄网站色大片免费观看| 老司机福利观看| 久久精品亚洲精品国产色婷小说| 97人妻精品一区二区三区麻豆| 听说在线观看完整版免费高清| 久99久视频精品免费| 午夜精品一区二区三区免费看| 午夜两性在线视频| 波多野结衣高清作品| 高清毛片免费观看视频网站| 国产高清videossex| 老司机在亚洲福利影院| 亚洲激情在线av| 18禁裸乳无遮挡免费网站照片| 国产久久久一区二区三区| 成人高潮视频无遮挡免费网站| 村上凉子中文字幕在线| 国产高清有码在线观看视频| 精品午夜福利视频在线观看一区| 日本熟妇午夜| www日本黄色视频网| 日日干狠狠操夜夜爽| 国产单亲对白刺激| 久久久久国产精品人妻aⅴ院| 免费看a级黄色片| 午夜福利免费观看在线| 久久久国产成人精品二区| 久久久久免费精品人妻一区二区| 一区福利在线观看| a在线观看视频网站| 高清在线国产一区| 国产一区二区三区视频了| 亚洲av免费在线观看| 嫁个100分男人电影在线观看| 天堂影院成人在线观看| 悠悠久久av| 一二三四社区在线视频社区8| 亚洲国产色片| 精品久久久久久久人妻蜜臀av| 国产爱豆传媒在线观看| 色综合站精品国产| 很黄的视频免费| 一区二区三区国产精品乱码| 村上凉子中文字幕在线| 精品久久久久久久末码| 亚洲人成网站高清观看| 悠悠久久av| 久久久久久久午夜电影| 美女高潮的动态| 国产亚洲精品一区二区www| 日本五十路高清| 日本免费a在线| 亚洲第一欧美日韩一区二区三区| 亚洲av二区三区四区| 国产伦精品一区二区三区视频9 | 免费看光身美女| av国产免费在线观看| 男女视频在线观看网站免费| 免费在线观看日本一区| 伊人久久大香线蕉亚洲五| 亚洲性夜色夜夜综合| 国产亚洲欧美98| 欧美绝顶高潮抽搐喷水| 香蕉av资源在线| 久久午夜亚洲精品久久| 韩国av一区二区三区四区| 人人妻人人澡欧美一区二区| 久久久久久久精品吃奶| 国产成人av激情在线播放| 午夜福利高清视频| 少妇裸体淫交视频免费看高清| 免费看日本二区| 国产毛片a区久久久久| 国产主播在线观看一区二区| 亚洲人成电影免费在线| 久久精品国产亚洲av涩爱 | а√天堂www在线а√下载| 少妇的丰满在线观看| 欧美激情久久久久久爽电影| 久久精品影院6| 一进一出抽搐动态| 国产亚洲精品一区二区www| 国产一区二区在线观看日韩 | 国产免费男女视频| 国产主播在线观看一区二区| 日韩亚洲欧美综合| 午夜日韩欧美国产| 男人舔女人下体高潮全视频| 麻豆成人午夜福利视频| 欧美日韩国产亚洲二区| 丰满的人妻完整版| 少妇人妻精品综合一区二区 | 中文亚洲av片在线观看爽| 日本精品一区二区三区蜜桃| 久99久视频精品免费| 亚洲无线观看免费| 国产免费av片在线观看野外av| 男人的好看免费观看在线视频| 成人欧美大片| 国产午夜精品久久久久久一区二区三区 | 国产淫片久久久久久久久 | 岛国在线观看网站| 亚洲熟妇熟女久久| 欧美一级毛片孕妇| av天堂中文字幕网| 亚洲无线在线观看| 亚洲不卡免费看| 亚洲国产高清在线一区二区三| 在线免费观看的www视频| 亚洲国产色片| 99久久九九国产精品国产免费| 91麻豆精品激情在线观看国产| 国产精品嫩草影院av在线观看 | x7x7x7水蜜桃| 亚洲最大成人手机在线| 国产亚洲精品一区二区www| 国内揄拍国产精品人妻在线| 18禁黄网站禁片午夜丰满| 国产伦一二天堂av在线观看| 久久中文看片网| 熟女人妻精品中文字幕| 内射极品少妇av片p| 男女做爰动态图高潮gif福利片| 国产精华一区二区三区| 欧美日本视频| 美女黄网站色视频| 成年女人毛片免费观看观看9| 变态另类成人亚洲欧美熟女| 偷拍熟女少妇极品色| 久久草成人影院| 国产毛片a区久久久久| 中文字幕人妻丝袜一区二区| 欧美丝袜亚洲另类 | 日日摸夜夜添夜夜添小说| 亚洲欧美日韩无卡精品| 成人性生交大片免费视频hd| 国产精品一区二区三区四区免费观看 | 久久性视频一级片| 又黄又粗又硬又大视频| 一二三四社区在线视频社区8| 又爽又黄无遮挡网站| 一级黄色大片毛片| 天天添夜夜摸| 亚洲欧美日韩高清在线视频| 午夜福利欧美成人| 别揉我奶头~嗯~啊~动态视频| 久久精品亚洲精品国产色婷小说| 九色成人免费人妻av| 国产单亲对白刺激| 国产乱人视频| 亚洲精品久久国产高清桃花| 真人做人爱边吃奶动态| 国产不卡一卡二| 亚洲av免费高清在线观看| 无限看片的www在线观看| 草草在线视频免费看| 99在线视频只有这里精品首页| 伊人久久精品亚洲午夜| 有码 亚洲区| 91麻豆av在线| 91字幕亚洲| 免费人成视频x8x8入口观看| 9191精品国产免费久久| 免费高清视频大片| 日本三级黄在线观看| 欧美一区二区精品小视频在线| 一夜夜www| 九色国产91popny在线| 久久久久亚洲av毛片大全| 老鸭窝网址在线观看| 国产激情欧美一区二区| 九九久久精品国产亚洲av麻豆| 欧美日韩一级在线毛片| 精品不卡国产一区二区三区| 日本免费a在线| 国产色婷婷99| 亚洲美女视频黄频| 99久久精品国产亚洲精品| 三级男女做爰猛烈吃奶摸视频| 欧美日韩中文字幕国产精品一区二区三区| 精品久久久久久久久久久久久| 男人舔奶头视频| 最后的刺客免费高清国语| 中文字幕av在线有码专区| 国产不卡一卡二| 欧美日韩一级在线毛片| 成年女人看的毛片在线观看| 久久天躁狠狠躁夜夜2o2o| 亚洲午夜理论影院| 国产成年人精品一区二区| 99久久综合精品五月天人人| 老熟妇仑乱视频hdxx| 欧美激情在线99| 成人一区二区视频在线观看| 最新中文字幕久久久久| 三级毛片av免费| 久久香蕉精品热| 天美传媒精品一区二区| 免费一级毛片在线播放高清视频| av欧美777| 很黄的视频免费| 亚洲国产高清在线一区二区三| 精品免费久久久久久久清纯| 午夜视频国产福利| 男女床上黄色一级片免费看| 中文字幕人妻熟人妻熟丝袜美 | 国产 一区 欧美 日韩| 午夜影院日韩av| 欧美色视频一区免费| 亚洲色图av天堂| 最近视频中文字幕2019在线8| xxx96com| 男女做爰动态图高潮gif福利片| xxxwww97欧美| 在线观看av片永久免费下载| 欧美乱妇无乱码| 欧美日韩精品网址| 深爱激情五月婷婷| 小蜜桃在线观看免费完整版高清| 亚洲精品美女久久久久99蜜臀| 特大巨黑吊av在线直播| ponron亚洲| 天天躁日日操中文字幕| 免费看光身美女| 亚洲无线在线观看| 99久久精品一区二区三区| 久久99热这里只有精品18| 最新在线观看一区二区三区| 日韩人妻高清精品专区| 午夜福利视频1000在线观看| 婷婷精品国产亚洲av| 听说在线观看完整版免费高清| 中文字幕久久专区| 亚洲18禁久久av| 欧美极品一区二区三区四区| 免费看a级黄色片| 成人性生交大片免费视频hd| 午夜福利在线观看免费完整高清在 | 久久这里只有精品中国| 精品日产1卡2卡| 97碰自拍视频| 黄色片一级片一级黄色片| 午夜两性在线视频| 日韩高清综合在线| 小说图片视频综合网站| 中国美女看黄片| 国产免费男女视频| 超碰av人人做人人爽久久 | 亚洲激情在线av| 欧美极品一区二区三区四区| 啦啦啦免费观看视频1| bbb黄色大片| 一级黄色大片毛片| or卡值多少钱| 99久久成人亚洲精品观看| 国产一级毛片七仙女欲春2| 久久精品夜夜夜夜夜久久蜜豆| 国产精品久久久久久人妻精品电影| 国产精品永久免费网站| 免费在线观看影片大全网站| 中文字幕人妻丝袜一区二区| 国产精品美女特级片免费视频播放器| 99久久精品一区二区三区| 国产野战对白在线观看| 欧美激情在线99| 两个人看的免费小视频| 国产免费一级a男人的天堂| 成人特级av手机在线观看| 三级国产精品欧美在线观看| 国产成人福利小说| 很黄的视频免费| 国产精品国产高清国产av| 色综合亚洲欧美另类图片| 一级作爱视频免费观看| 香蕉久久夜色| 国产蜜桃级精品一区二区三区| 麻豆国产av国片精品| 搞女人的毛片| 国产高清激情床上av| 嫁个100分男人电影在线观看| 好看av亚洲va欧美ⅴa在| 有码 亚洲区| 国产成人av激情在线播放| 高清日韩中文字幕在线| 老师上课跳d突然被开到最大视频 久久午夜综合久久蜜桃 | 国产精品av视频在线免费观看| 99久国产av精品| 日韩欧美精品v在线| 非洲黑人性xxxx精品又粗又长| ponron亚洲| 亚洲欧美日韩高清专用| 三级国产精品欧美在线观看| 麻豆成人午夜福利视频| 日本免费一区二区三区高清不卡| 99在线视频只有这里精品首页| 欧美三级亚洲精品| 亚洲五月婷婷丁香| 日本黄色视频三级网站网址| 日本一二三区视频观看| 日本黄色视频三级网站网址| tocl精华| 免费在线观看影片大全网站| 国产成人aa在线观看| 精品一区二区三区人妻视频| 别揉我奶头~嗯~啊~动态视频| 三级国产精品欧美在线观看| 亚洲人成电影免费在线| 伊人久久大香线蕉亚洲五| 操出白浆在线播放| 91字幕亚洲| 一进一出抽搐动态| av中文乱码字幕在线| 国内精品久久久久久久电影| 婷婷精品国产亚洲av在线| 18禁黄网站禁片午夜丰满| 精品人妻偷拍中文字幕| aaaaa片日本免费| 午夜福利在线在线| 性欧美人与动物交配| 午夜福利在线在线| 首页视频小说图片口味搜索| 亚洲国产精品999在线| 成年免费大片在线观看| av国产免费在线观看| 99久国产av精品| www日本黄色视频网| 亚洲av不卡在线观看| 男人舔女人下体高潮全视频| 黄色丝袜av网址大全| 一进一出好大好爽视频| 精品一区二区三区av网在线观看| 少妇的逼好多水| 美女被艹到高潮喷水动态| 国产一区二区激情短视频| 免费人成在线观看视频色| 有码 亚洲区| 亚洲av中文字字幕乱码综合| 露出奶头的视频| 国产亚洲精品av在线| or卡值多少钱| 亚洲va日本ⅴa欧美va伊人久久| 狠狠狠狠99中文字幕| 99国产精品一区二区三区| 免费大片18禁| 天天一区二区日本电影三级| 91久久精品国产一区二区成人 | 亚洲性夜色夜夜综合| 国产亚洲精品一区二区www| 观看美女的网站| 国产精品99久久久久久久久| 免费搜索国产男女视频| 亚洲中文字幕日韩| 1024手机看黄色片| 波多野结衣高清无吗| 一级黄片播放器| 人人妻,人人澡人人爽秒播| 精品无人区乱码1区二区| av中文乱码字幕在线| 欧美中文日本在线观看视频| 欧美bdsm另类| 国产av一区在线观看免费| 黄片小视频在线播放| 国内精品久久久久精免费| 19禁男女啪啪无遮挡网站| 成人精品一区二区免费| 国产伦人伦偷精品视频| 免费av不卡在线播放| 久久久久九九精品影院| 国内精品美女久久久久久| 国产真实乱freesex| 亚洲专区中文字幕在线| 桃色一区二区三区在线观看| 少妇丰满av| 精品久久久久久久久久久久久| 亚洲av美国av| 久久久久久久久中文| 国内毛片毛片毛片毛片毛片| 老司机在亚洲福利影院| 成年人黄色毛片网站| 我的老师免费观看完整版| 女警被强在线播放| 国产成人系列免费观看| 黄色丝袜av网址大全| 五月玫瑰六月丁香| 国产亚洲精品综合一区在线观看| 51午夜福利影视在线观看| 欧美激情久久久久久爽电影| 男人的好看免费观看在线视频| 悠悠久久av| 免费在线观看成人毛片| 久久精品国产自在天天线| av黄色大香蕉| 国产三级黄色录像| 一个人看视频在线观看www免费 | 日韩精品青青久久久久久| aaaaa片日本免费| 18禁黄网站禁片免费观看直播| 99热这里只有精品一区| 国产高清视频在线观看网站| 91在线观看av| 母亲3免费完整高清在线观看| 一级黄色大片毛片| 精品熟女少妇八av免费久了| 成人三级黄色视频| 丰满人妻熟妇乱又伦精品不卡| 亚洲18禁久久av| 色综合婷婷激情| 制服丝袜大香蕉在线| 日韩精品中文字幕看吧| 免费观看的影片在线观看| 大型黄色视频在线免费观看| 国产单亲对白刺激| 国产91精品成人一区二区三区| 国产色婷婷99| 日韩欧美免费精品| 女同久久另类99精品国产91| 国产成人aa在线观看| 少妇人妻一区二区三区视频| 欧美日韩福利视频一区二区| 免费在线观看影片大全网站| 国产97色在线日韩免费| 日本 av在线| 欧美大码av| 国产精品日韩av在线免费观看| 91九色精品人成在线观看| 国产精品久久久久久人妻精品电影| 小说图片视频综合网站| 国产美女午夜福利| eeuss影院久久| 午夜久久久久精精品| 亚洲精品国产精品久久久不卡| 香蕉久久夜色| 性色avwww在线观看| 日韩免费av在线播放| 日本黄色片子视频| 久久久久久久久大av| av专区在线播放| 亚洲av第一区精品v没综合| 白带黄色成豆腐渣| tocl精华| 成人三级黄色视频| 两个人看的免费小视频| 舔av片在线| 婷婷精品国产亚洲av| 久久久国产成人免费| 亚洲欧美日韩无卡精品| 久久人妻av系列| or卡值多少钱| 在线观看66精品国产| 久久久国产成人免费| 亚洲欧美日韩无卡精品| 黑人欧美特级aaaaaa片| 免费看美女性在线毛片视频| 久久香蕉精品热| 欧美日韩一级在线毛片| 久久精品国产自在天天线| 久久久精品大字幕| 国产伦人伦偷精品视频| 动漫黄色视频在线观看| 性欧美人与动物交配| 色吧在线观看| 美女黄网站色视频| 亚洲午夜理论影院| 中国美女看黄片| 嫩草影视91久久| 欧洲精品卡2卡3卡4卡5卡区| 女生性感内裤真人,穿戴方法视频| 国产精品乱码一区二三区的特点| 久久午夜亚洲精品久久| 亚洲国产精品999在线| 国产高清三级在线| 亚洲av免费在线观看| 精品国内亚洲2022精品成人| 亚洲成人中文字幕在线播放| 久久99热这里只有精品18| svipshipincom国产片| 国内揄拍国产精品人妻在线| 两人在一起打扑克的视频| 亚洲美女黄片视频| 久久久久性生活片| av在线蜜桃| 午夜精品在线福利| 悠悠久久av| 成人性生交大片免费视频hd| 国产免费一级a男人的天堂| 亚洲av免费高清在线观看| 国产免费一级a男人的天堂| 国产真实乱freesex| 亚洲 国产 在线| 免费在线观看成人毛片| 最好的美女福利视频网| 欧美丝袜亚洲另类 | 亚洲人与动物交配视频| 一进一出抽搐gif免费好疼| 一级作爱视频免费观看| 好男人电影高清在线观看| 国产中年淑女户外野战色| 女人高潮潮喷娇喘18禁视频| 性欧美人与动物交配| 3wmmmm亚洲av在线观看| 校园春色视频在线观看| 国产欧美日韩一区二区三| 精品久久久久久久久久免费视频| 国产成人福利小说| 9191精品国产免费久久| 亚洲黑人精品在线| 午夜两性在线视频| 亚洲av免费高清在线观看| 色在线成人网| 欧美一级a爱片免费观看看| 午夜亚洲福利在线播放| 两个人的视频大全免费| 亚洲国产中文字幕在线视频| 90打野战视频偷拍视频| 亚洲人成网站在线播放欧美日韩| 女人被狂操c到高潮| 亚洲熟妇中文字幕五十中出| 国产精品国产高清国产av| 亚洲人与动物交配视频| 久久久久久久久大av| 欧美日韩国产亚洲二区| 1024手机看黄色片| 国产主播在线观看一区二区| 免费av观看视频| 午夜老司机福利剧场| 免费看a级黄色片| 国产成人av激情在线播放| av视频在线观看入口| 久久精品人妻少妇| 午夜久久久久精精品| 天美传媒精品一区二区| 国产精品综合久久久久久久免费| 亚洲人成网站在线播| 久久精品综合一区二区三区| 免费av不卡在线播放| 国产69精品久久久久777片| 亚洲av美国av| 国产精品一区二区三区四区免费观看 | www日本在线高清视频| 一进一出抽搐动态| 18禁国产床啪视频网站| 国产精品乱码一区二三区的特点| 国产成人aa在线观看| 国产私拍福利视频在线观看| 日韩国内少妇激情av| 国产精品美女特级片免费视频播放器| 久久久久久久久大av| 99国产精品一区二区三区| 国产成人aa在线观看| 免费电影在线观看免费观看| 欧美精品啪啪一区二区三区| 少妇人妻精品综合一区二区 | or卡值多少钱| 精品一区二区三区视频在线观看免费| 亚洲欧美精品综合久久99| 亚洲无线在线观看| 精品国产超薄肉色丝袜足j| 亚洲人成网站在线播| av在线天堂中文字幕| 国产黄片美女视频| 麻豆国产av国片精品| 日韩成人在线观看一区二区三区| 一本一本综合久久| 又黄又粗又硬又大视频| а√天堂www在线а√下载| 亚洲人成网站在线播放欧美日韩| 熟女人妻精品中文字幕| 美女大奶头视频| 国产成人啪精品午夜网站|