• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    An Innovative K-Anonymity Privacy-Preserving Algorithm to Improve Data Availability in the Context of Big Data

    2024-05-25 14:43:08LinlinYuanTiantianZhangYulingChenYuxiangYangandHuangLi
    Computers Materials&Continua 2024年4期

    Linlin Yuan ,Tiantian Zhang ,Yuling Chen ,Yuxiang Yang and Huang Li

    1State Key Laboratory of Public Big Data,College of Computer Science and Technology,Guizhou University,Guiyang,550025,China

    2College of Information Engineering,Guizhou Open University,Guiyang,550025,China

    3Guizhou Academy of Tobacco Science,Guiyang,550025,China

    ABSTRACT The development of technologies such as big data and blockchain has brought convenience to life,but at the same time,privacy and security issues are becoming more and more prominent.The K-anonymity algorithm is an effective and low computational complexity privacy-preserving algorithm that can safeguard users’privacy by anonymizing big data.However,the algorithm currently suffers from the problem of focusing only on improving user privacy while ignoring data availability.In addition,ignoring the impact of quasi-identified attributes on sensitive attributes causes the usability of the processed data on statistical analysis to be reduced.Based on this,we propose a new K-anonymity algorithm to solve the privacy security problem in the context of big data,while guaranteeing improved data usability.Specifically,we construct a new information loss function based on the information quantity theory.Considering that different quasi-identification attributes have different impacts on sensitive attributes,we set weights for each quasi-identification attribute when designing the information loss function.In addition,to reduce information loss,we improve K-anonymity in two ways.First,we make the loss of information smaller than in the original table while guaranteeing privacy based on common artificial intelligence algorithms,i.e.,greedy algorithm and 2-means clustering algorithm.In addition,we improve the 2-means clustering algorithm by designing a mean-center method to select the initial center of mass.Meanwhile,we design the K-anonymity algorithm of this scheme based on the constructed information loss function,the improved 2-means clustering algorithm,and the greedy algorithm,which reduces the information loss.Finally,we experimentally demonstrate the effectiveness of the algorithm in improving the effect of 2-means clustering and reducing information loss.

    KEYWORDS Blockchain;big data;K-anonymity;2-means clustering;greedy algorithm;mean-center method

    1 Introduction

    With the rapid development of blockchain technology,it has offered Internet of Things (IoT)systems a more efficient and stable decentralized method for data storage and management.But at the same time,it has also raised a series of network security issues.For example,data privacy[1–3],validation techniques [4],and transmission reliability [5] are difficult to guarantee,which will lead to the leakage of user privacy.Nowadays,we generate a large amount of data every day,which contains great application value[6]but also implies sensitive information,such as location information[7–10]and transaction information[11].Once this data is stolen by adversaries,it will possibly cause the leakage of user privacy.To solve the privacy security problems in networks in the context of blockchain and IoT integration,privacy-preserving algorithms are essential in addition to techniques such as authentication[8],distributed storage and computing[12–15],big data management[16,17],and anomaly detection [18].The K-anonymity algorithm [19] is an effective privacy-preserving algorithm with lower time complexity compared with commonly used cryptographic algorithms such as homomorphic encryption [20].In addition,compared to the commonly used differential privacy techniques,this algorithm does not need to add noise[21]to sensitive information,thus preserving the specific values of sensitive attributes for the subsequent statistical analysis process.The K-anonymity algorithm obtains less accurate datasets through steganography and generalization techniques so that at least K quasi-identified attributes in a certain sub-table have the same value,which in turn ensures that an adversary cannot use these quasi-identified attributes to link to a particular record[22].By anonymizing the data with a K-anonymization algorithm before storing[23],publishing[24],or transmitting[25],the privacy of users can be secured to a certain extent.For example,by generalizing the age property in a data table from 25 to interval[20,30],an adversary will not be able to guess the exact value of the age directly.However,the algorithm can lead to reduced usability of the data after generalization,which will be detrimental to the subsequent use of the data for statistical analysis[26]or forecasting.When the data owners get the data,they want to perform statistical analysis or prediction on this data.For example,they want to know the age distribution characteristics of the population in a certain region based on these data,but if the loss of information after generalization is too large,such as generalizing age from 50 to 100[10],then the data owner will not be able to arrive at an accurate statistical analysis or prediction result based on these generalized values.Therefore,how to process the data with the K-anonymity algorithm to improve the usability of the data while safeguarding privacy is the main problem to be solved in this paper.

    The K-anonymity algorithm consists of three processes: attribute classification,clustering,and generalization [27].In clustering,the more similar the data in an equivalence class,the smaller the generalized interval or value,and the greater the availability of the generalized data [28].Therefore,it is crucial to know how to perform equivalence class classification based on the similarity between tuples.The most common equivalence class partitioning method is the k-means clustering algorithm,which is a classical artificial intelligence algorithm.This algorithm guarantees that tuples belonging to the same equivalence class are as similar as possible by aggregating tuples that are close to each other to improve the usability of the data after generalization [29,30].k-means clustering algorithm as an artificial intelligence[31]algorithm has the advantages of simplicity,speed,and low time complexity[32].However,this clustering algorithm itself is unstable due to the randomness of the initial cluster center selection [33] and thus is not conducive to achieving a stable K-anonymity effect.To address the initial clustering selection problem of the k-means algorithm,Bhattacharya et al.explored the k-means++algorithm and suggested that a better clustering effect can be achieved when the initial prime centers are as far away from each other as possible [34],however,the algorithm still suffers from randomness in the selection of the first prime center.The clustering algorithm in this paper is improved based on the above clustering algorithm,and the comparison of the improved effect is shown in Table 1.

    Table 1: Comparison of the effect of different clustering algorithms

    Moreover,the greedy algorithm,as a common algorithm in the field of deep learning[35],is now also applied in K-anonymization algorithms to improve the usability of the data.Aghdam et al.applied a greedy algorithm in the K-anonymization algorithm to achieve high data utility[36].Tang et al.proposed to achieve clustering and generalization of data by a greedy algorithm and binary classification as a way to fully exploit the value of data and safeguard user privacy [37].Based on this,we apply the greedy algorithm in the K-anonymity algorithm to achieve lower information loss.The specific contributions are as follows:

    1.We designed the information loss function for the K-anonymity algorithm.Based on the information quantity theory,we represent the information loss after the generalization of quasiidentified attributes,and also calculate the influence degree for each quasi-identified attribute on sensitive attributes,and set the weight value for each quasi-identified attribute according to the influence degree to improve the reasonableness of information loss assessment.

    2.We improve the 2-means clustering algorithm.We designed a mean-center method to select the initial center of mass for clustering,which improves the effect of 2-means clustering and guarantees the stability of the clustering effect at the same time.

    3.Based on the 2-means clustering algorithm and greedy algorithm,we propose an improved Kanonymity algorithm to reduce information loss while guaranteeing privacy.Finally,we experimentally demonstrate the effectiveness of the algorithm in improving the effect of 2-means clustering and reducing information loss.

    The algorithm is innovative because it improves on the original algorithm by clustering the clusters to achieve more stable results and less loss of information.In addition,it introduces the amount of information to design the information loss function and uses the greedy algorithm to design the anonymization algorithm which again reduces the information loss.The overall structure block diagram of the algorithm in this paper is shown in Fig.1.

    Figure 1: Block diagram of the overall structure of the algorithm

    2 Basic Knowledge

    This section will introduce the basics covered in this paper,including the introduction to the Kanonymity algorithm,the amount of information,and the k-means clustering algorithm.

    2.1 K-Anonymity

    The development of smart IoT [38] allows people and machines to be interconnected through the network,which also generates a huge amount of data that contains users’privacy.K-anonymity algorithms [39] can solve the privacy leakage problem by ensuring that each record in the dataset cannot be distinguished from other K-1 individuals for sensitive attributes.Through generalization and steganography techniques,less accurate data is obtained such that there are at least K records for the same quasi-identifier,making it impossible for an observer to link records through the quasi-identifier[40].K-anonymization algorithms are used in the context of big data [41] because of their low time complexity.Its processing process consists of attribute classification,clustering,and generalization.Attribute classification refers to the classification of attributes in a data table into identifying attributes,quasi-identifying attributes,and sensitive attributes.Identifying attributes are attributes that uniquely identify the user and are deleted directly for them.Sensitive attributes are attributes that the user does not want to be known by others,and they are not processed.The rest of the attributes are quasiidentifiable attributes,these attributes cannot be used alone to identify a user,such as age and gender.However,an adversary may infer sensitive information about a user based on some quasi-identifying attributes.For example,an adversary,with some background knowledge,may be able to locate a specific user based on information such as age,gender,and region of a record,so the quasi-identifiable attributes are generalized to protect the user’s privacy.Clustering refers to clustering tuples with high similarity into one class.Generalization refers to the generalization of attribute values of the same class.

    2.2 Amount of Information

    The amount of information[42]is the amount of information needed to select one event from N equally likely events.Shannon’s information theory applies probability to describe uncertainty.The smaller the probability of an event appearing,the more uncertainty there is,and the more information there is,and vice versa.After generalizing the data,the smaller the probability of an adversary inferring the true value through the generalized interval,the more information is indicated,but also the greater the loss of information.Therefore,we use the amount of information to measure the loss of information after the generalization of data tables.The Information volume of informationxis denoted as:

    whereP(x)denotes the probability of occurrence ofx.The log in the information quantity function of this paper is all taken to a base of 2.

    2.3 k-means Clustering Algorithm

    k-means clustering is a common algorithm in artificial intelligence and belongs to unsupervised machine learning algorithms.The algorithm has appeared in application scenarios such as the Internet of Things and cloud computing[43]and is used to solve clustering problems.The k-means clustering algorithm puts similar objects together according to some rules.Distance is a measure of similarity.The closer the distance of each object in a class,the better the clustering will be.In k-means clustering,k denotes the number of classes that need to be determined first before clustering.In this paper,we use binary classification for clustering,so k is 2.The steps of this algorithm are as follows[44]:

    Step 1: Select the initialized k samples as the initial clustering centers notated asO=O1,O2,...,Ok;

    Step 2:For each sampleriin the dataset,calculate its distance to each initial clustering center and assign it to the class corresponding to the clustering center with the smallest distance;

    Step 3:For each classclassi,recalculate the clustering centerOi;

    Step 4:Repeat Step 2 and Step 3 above until the set number of iterations is reached or the clustering centers no longer change.

    In this paper,we use binary classification for clustering,so k is 2.

    3 Algorithm Design

    In this section,we introduce the K-anonymity algorithm in this paper.Specifically,we introduce the designed information loss function and the 2-means artificial intelligence clustering algorithm.Finally,we apply the information loss function designed in this paper to the K-anonymity algorithm and construct the K-anonymity privacy-preserving algorithm in this paper based on two artificial intelligence algorithms,i.e.,the greedy algorithm and the improved 2-means clustering algorithm.The parameters of this paper are defined in Table 2.

    Table 2: Parameter definition

    3.1 Information Loss Function Based on the Amount of Information

    In this paper,we measure the information loss after generalization based on the amount of information.In addition,there are associations between attributes[45],and different quasi-identified attributes have different degrees of influence on sensitive attributes.In this section,we assign a weight to each quasi-identified attribute based on its influence on the sensitive attribute and introduce the weight into the information loss function.When generalizing the classified quasi-identified attributes,for numerical quasi-identifiers,the original data value is replaced by the minimum value field in the equivalence set.For example,generalize age 10 to interval[5,15].For Subtype quasi-identifiers,they are generalized to a minimum value that is larger than the original quasi-identifier value.For example,generalizing gender male as gender unknown.

    For a numerical quasi-identification attribute,the amount of information loss for thetth numerical quasi-identification attribute value of a data table after a certain generalization is denoted as:

    wherestmaxdenotes the maximum value under thetth numeric attribute in the generalized table,stmindenotes the minimum value under thetth numeric attribute in the generalized table,|stmax-stmin+1|denotes the number of values that can be taken for thetth numeric attribute in the generalized table,and|Rt|denotes the number of values contained in the total value field of thetth numeric attribute,|Rt|≥1.nis the number of records in this generalized data table.x=0,1,2,...symbolizes the step size of the value taken by the numeric attribute,if the value of the attribute takes a step size of 1,such as the age attribute,thenx=0,if the step size is 0.1,thenx=1.

    For subtype quasi-identification attributes,assuming that the attribute weights are calculated in the order of numeric attributes first and subtype attributes second,and that there aremnumeric attributes,the amount of information loss for theτth subtype quasi-identification attribute of a generalized data table is denoted as:

    whereleaf(Tτ) denotes the number of leaf nodes under the generalized value of theτth subtype attribute in this generalization table,andleaf(TτR) denotes the total number of leaf nodes in the classification tree where theτth subtype attribute is located,leaf(TτR)≥1.nis the number of records in the data table.

    The amount of information lost in a particular generalized table after generalization is denoted as:

    whereγt,γτ+mrepresent the attribute weights,and assuming that the attribute weights are calculated in the order of numerical attributes first and subtype attributes second,we have:

    The total information loss after anonymization is equal to the sum of the information loss of each generalization table.

    3.2 The 2-Means Improvement Algorithm Based on the Mean-Center Method

    Before generalizing the data by a K-anonymization algorithm,to improve the usability of the anonymized data,the data with high similarity needs to be clustered into one class by a clustering algorithm.In this paper,an improved 2-means clustering algorithm is used for clustering.Next,we present the definition of the distance between data and tuples when applying the 2-means algorithm for clustering,and the mean-center method designed for initial center of mass selection.

    3.2.1 Definition of Distance between Data and Tuples

    (1)Distance between data

    For a certain numerical quasi-identification attribute,letRbe a finite numerical domain,and the distance between any valuessi,sj∈Rin thetth numerical quasi-identification attribute is denoted as:

    wheresmaxdenotes the maximum value of the total value domain of the attribute andsmindenotes the minimum value of the total value domain of the attribute.

    For a subtype quasi-identification attribute,letTRbe a classification treeR.Assuming that the attribute weights are calculated in the order of numerical attributes first and subtype attributes second and that there are m numerical attributes,the distance between any valuessi,sj∈Rin theτth subtype quasi-identification attribute is denoted as:

    whereleafdenotes the number of leaf nodes under the tree rooted by the smallest common ancestor ofsiandsjin the classification tree,andleaf(TR)denotes the total number of leaf nodes of the classification treeT.

    (2)Distance between tuples

    The distance between each data tuple is equal to the sum of the distances between each quasiidentified attribute in the tuple.Denote the tupleTU={N1,N2...Nm,C1,C2...Cn,S},whereN1,N2...Nmdenotesmnumerical quasi-identified attributes in a record,C1,C2...Cndenotesnsubtyped quasi-identified attributes in a record,andSis a sensitive attribute in a record,then the distance between tupleTUiandTUjin this scheme can be expressed as:

    3.2.2 Selection of the Center of Mass

    To improve the clustering effect and ensure the stability of the clustering effect,we propose a mean-center method to select the initial center of mass for the 2-means algorithm.The algorithm is influenced by the idea of the farthest distance in the k-means++algorithm.It aims to find the farthest point in the class and to avoid the problem of unstable effects caused by randomness.This section describes how to choose the initial center of mass and how to update the center of mass in this paper.

    (1)Selection of the initial center of mass mean center method

    In this paper,2-means clustering is used for classification,and two initial centers of mass need to be selected for each clustering.In previous studies,the most common methods for selecting two centers of mass are the random selection method and the farthest distance method.The random selection method can affect the effect of clustering and make the clustering results unstable.The farthest distance method refers to choosing two points as far as possible from each other as the center of mass to make the clustering as effective as possible[46].However,to fix two centers of mass,we need each node to be compared with other nodes,which have high time complexity.Based on this,we designed a mean center of mass method to select the initial center of mass.It can avoid the unstable clustering results and also make the selected initial center of mass as far as possible.The method is designed as shown in Algorithm 1.

    When this scheme performs calculations on tuples,it refers to calculations on quasi-identifiers only.For numeric quasi-identifier attributes,the mean value of thetth numeric attribute in the data table can be expressed as:

    wheresidenotes the value of thetth numeric attribute under theith tuple andnis the number of tuples in the table.

    For subtype quasi-identification attributes,direct averaging operations cannot be performed like numerical attributes.Therefore,in this paper,we design the mean value representation method for the quasi-identified attributes of subtypes by combining the property that the mean value is the smallest sum of squares of each number.Specifically,the mean value of theτth subtype attribute in the data table can be expressed as:

    wheresandsiare the values under a quasi-identified attribute in the equivalence class.

    (2)Renewal of the center of mass

    The k-means algorithm,after classification,for each class to re-select the clustering center and repeatedly iterate the clustering process until the required number of iterations is reached or the location of the clustering center no longer changes when the clustering ends[47].In this process,we calculate the mean value of the data as the updated center of mass.The updated center of mass is denoted as:

    3.3 K-Anonymity Algorithm Based on Greedy Algorithm and Improved 2-Means Clustering

    To make the K-anonymization algorithm improve the usability of the data while safeguarding user privacy,we use the greedy algorithm and 2-means clustering to cluster and generalize the dataset.The specific K-anonymization algorithm is designed as follows:

    (1) Determination of identification attribute (La),quasi-identification attribute (QLa),and sensitive attribute(S)

    We consider attributes that uniquely identify a user as quasi-identifying attributes and remove them directly.No processing is done for sensitive attributes.Other attributes,considered as quasiidentifying attributes,are generalized.

    (2)Construct equivalence classes and generalize

    In this paper,we construct an equivalence class that minimizes the loss of information under the condition of privacy protection.To guarantee the privacy of the data,based on the definition of the K-anonymity algorithm,the processed data table should satisfy the following equation:

    wheretidenotes theith record in the user data table andDT′denotes the data table after the clustering and generalization process.To achieve a certain goal,algorithms often need to satisfy a minimum threshold in their design.To guarantee that the amount of information loss is as small as possible,the algorithm should also satisfy the following equation:

    whereIL(DT′)indicates the degree of information loss in the data tableDT′after the generalization process.Based on this,we classify the original data based on the improved 2-means algorithm and greedy algorithm.The overall algorithm flow is shown in Algorithm 2.

    4 Simulation Experiments and Results

    In this section,we mainly introduce the data set used for the experiments and the preprocessing process of the data,analyze the contribution of quasi-identified attributes,and verify that the 2-means algorithm designed in this paper can achieve better clustering results and the improved K-anonymity algorithm can better reduce information loss.

    4.1 Experimental Data Set

    This experiment uses the Adult dataset from the UCI machine learning repository,which contains partial U.S.Census data with over 30,000 records and is widely used in privacy protection and data mining research.The dataset has been de-identified with user identification attributes,and the remaining 15 attributes cover sub-type attributes and numerical attributes,such as age,job category,gender,etc.In addition,among these attributes,the attribute education-num corresponds to the attribute education,so only one attribute is kept in this paper for the experiments.Finally,we randomly selected 1000 data in the dataset,set income as a sensitive attribute,and set the remaining 13 attributes as quasi-identifying attributes for the simulation experiments in this paper.

    4.2 Data Pre-processing

    In this paper,the data was pre-processed before the experiments,including three processes data missing value processing,data standardization and normalization,and outlier processing.

    4.2.1 Missing Value Handling

    There are some missing values in the dataset,and before starting the experiment,this paper first deals with these missing values.For numerical attributes,the missing values are filled with the average of the data under the attribute.For the sub-type quasi-identified attribute,the missing values are filled with the plural of the values under the attribute,and then the complete experimental data set is obtained.

    4.2.2 Data Standardization and Normalization

    When clustering data using the k-means algorithm,the data needs to be standardized and normalized to achieve accurate clustering results.In this paper,the data are standardized using the Z-score[48].Specifically,the data are processed by the following equation:

    wheremean(x)refers to the mean ofxandrefers to the standard deviation ofx.In addition,we use the maximum-minimum normalization to normalize the data.The formula is as follows:

    wherexmaxrepresents the maximum value ofxin the data andxminrepresents the minimum value ofxin the data.After processing,the data values in the data set are in the interval[0,1].

    4.2.3 Outlier Processing

    Outliers [49] can affect the effectiveness of k-means clustering.We use the 3σmethod to detect outliers in the data,which is based on the principle of 3σof normal distribution to detect outlier points.If the data is in the interval [μ-3σ,μ+3σ],it is a normal point,and conversely,it is an outlier.Specifically,the following formula can be used to determine whether a point is an outlier.

    whereμrefers to the mean of the data andσrefers to the standard deviation of the data.For outlier points existing in the data set,they are not taken into account in the process of calculating the mean value to select the initial center of mass.

    4.3 Quasi-Identification Attribute Impact Degree Analysis

    The weights of each attribute are taken into account when we design the information loss,and they represent the degree of influence of the quasi-identified attribute on the sensitive attribute.In the experiments,we use a machine learning[50]algorithm,namely logistic regression,to obtain the degree of influence of each quasi-identified attribute on the sensitive attribute according to the characteristics of the attributes,and the results are shown in Fig.2.The algorithm can analyze the influence of the independent variables on the dependent variable.First,we construct a logistic regression model to obtain the coefficients of each independent variable in the model and then calculate the contribution of each quasi-identified attribute to the sensitive attribute based on the coefficients.Finally,the contribution degrees are processed so that the contribution degrees of each quasi-identified attribute are summed to 1,and the best combination of the weights of each quasi-identified attribute is obtained.

    Figure 2: Degree of influence of each quasi-identified attribute on sensitive attributes

    We use the absolute value of the influence degree as the weight value of each quasi-identified attribute.Table 3 shows the attribute name,influence degree,and absolute value of the influence degree corresponding to each attribute number.

    Table 3: Each quasi-identification attribute and its influence degree

    4.4 2-Means Algorithm Improvement Effect

    To verify the effect of 2-means algorithm improvement,simulation experiments are set up in this paper to compare the clustering effect of this scheme with the clustering effect of the randomized center of mass method.When measuring the clustering effect,we use the contour coefficient method,which is a clustering evaluation index used to evaluate the effect of data clustering.Its value ranges from [–1,1],the higher the similarity of the data within a class and the lower the similarity of the data between classes,the larger the contour coefficient will be,which means the better the clustering effect.To enhance the accuracy of the experiment,30 comparative experiments were conducted in this scheme,as shown in Fig.3 for the comparison of the contour coefficients under 30 trials.

    Figure 3: Comparison of profile coefficients under different tests

    As can be seen from Fig.3,in each experiment,the contour coefficients of the improved 2-means method in this paper are greater than or equal to the 2-means method under random prime selection.Therefore,it has been proved that our method has a better clustering effect.In addition,it can be seen that the improved 2-means method adopted in this scheme obtains a stable clustering effect because the center of mass is determined.That is,compared with the 2-means method under random center of mass selection,our method is both stable and improves the clustering effect.

    To demonstrate the magnitude of the clustering effect improvement,we compare the average contour coefficients under a different number of experiments.In general,the more the number of experiments,the closer to the real situation,as can be seen from Fig.4,our method has a greater improvement in the average contour coefficient when the number of experiments is 30.To show the improvement of the clustering effect more specifically,the average of the improvement rate of our algorithm under 30 experiments is calculated in this paper,and from the calculation results,the clustering effect of our algorithm has improved by 124.92%over the 2-means method under the random center of mass,which achieves a good improvement effect.

    Figure 4: Comparison of the average profile coefficients under different number of tests

    4.5 K-Anonymity Algorithm Improvement Effect

    To verify the overall improvement of the K-anonymization algorithm in this scheme,simulation experiments are designed in this paper.First,we compared the information loss when using this scheme with the information loss from the direct generalization of the original data table.In addition,to verify the effect of the improvement of the 2-means algorithm on the degree of information loss of the whole K-anonymization algorithm,the degree of information loss using our method was compared with the degree of information loss using the randomized center of mass method.

    (1)Overall information loss of K-anonymity algorithm

    Fig.5 shows the change of information loss with different K values.As the K value increases,the information loss of the data roughly shows an increasing trend.Because the larger the K value is,generally the more data in the generalization table,the longer the length of the generalized interval will be,and the information loss will be larger.In addition,it can be seen from Fig.5 that using the method in this paper,the information loss is reduced,and the data availability is improved compared to the original.

    Figure 5: Comparison of average information loss with the original table under different K

    (2)Comparison of information loss under different mass center selection methods

    To verify the effectiveness of this paper’s mean-centered method in reducing the information loss of K-anonymity,we conducted 30 experiments to compare the information loss of K-anonymity with this paper’s method under the randomized center of mass method with different K and took the average of the 30 experiments as the final result.As shown in Fig.6,the method of this paper achieves a better reduction of information loss at different K.

    In addition,in this paper,K=4,8,12,and 16 are uniformly chosen,and the comparison between the randomized center of mass method and our method in terms of information loss for 30 experiments at specific K values is plotted.As shown in Fig.7,it can be seen that using the method in this paper can not only improve the information loss of the K-anonymity algorithm but also keep the information loss stable.

    Figure 6: Comparison of average information loss with random center of mass method at different K

    Figure 7: Comparison of information loss under different sub-experiments

    5 Conclusion

    We propose an improved K-anonymity algorithm to solve the problem of user privacy disclosure in the context of blockchain and IoT integration.This algorithm can improve the availability of data,and because of the improvement of the initial centroid selection method of the clustering algorithm,more stable results are achieved.In the future,privacy security issues in distributed storage and computing[51]can be considered,and combining the K-anonymization algorithm of this paper with distributed storage and techniques could solve privacy security issues in more scenarios.In addition,we use binary classification for clustering and design the initial center of mass selection algorithm only for the case of two initial centers of mass selection,which can be subsequently extended to solve the problem of the initial center of mass selection in multi-classification.

    Acknowledgement:We sincerely thank the editors and reviewers for their comments on this paper.

    Funding Statement:This research was supported by Foundation of National Natural Science Foundation of China (62202118),and Scientific and Technological Research Projects from Guizhou Education Department([2023]003),and Guizhou Provincial Department of Science and Technology Hundred Levels of Innovative Talents Project (GCC[2023]018),and Top Technology Talent Project from Guizhou Education Department([2022]073).

    Author Contributions:The authors confirm contribution to the paper as follows:study conception and design: Linlin Yuan;data collection: Tiantian Zhang;analysis and interpretation of results: Yuling Chen;draft manuscript preparation:Yuxiang Yang,Huang Li.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:This paper uses the Adult dataset from the UCI machine learning repository,which contains partial U.S.Census data with over 30,000 records and is widely used in privacy protection and data mining research.These data were derived from the following resources available in the public domain:https://archive.ics.uci.edu/ml/datasets/Adult.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    亚洲图色成人| 天堂8中文在线网| 五月伊人婷婷丁香| 欧美亚洲 丝袜 人妻 在线| 日韩一本色道免费dvd| 成人毛片60女人毛片免费| 国产av精品麻豆| 久久鲁丝午夜福利片| 亚洲性久久影院| 亚洲国产日韩一区二区| 黑丝袜美女国产一区| 久久国产精品大桥未久av | 欧美成人午夜免费资源| 欧美 日韩 精品 国产| 啦啦啦啦在线视频资源| 狂野欧美激情性bbbbbb| 久久人人爽人人片av| 亚洲欧美一区二区三区国产| 久久精品夜色国产| 亚洲成人一二三区av| 亚洲四区av| 成年人午夜在线观看视频| 久久久久精品性色| 久久精品夜色国产| 中文字幕久久专区| 国产免费一级a男人的天堂| 久久久久久久久大av| 在线观看免费高清a一片| 国产精品蜜桃在线观看| av有码第一页| 精品一区二区免费观看| 国内精品宾馆在线| 精品一品国产午夜福利视频| 亚洲一区二区三区欧美精品| 一级二级三级毛片免费看| 欧美xxxx性猛交bbbb| 免费人成在线观看视频色| 亚洲欧美精品自产自拍| 一级a做视频免费观看| 男人和女人高潮做爰伦理| 新久久久久国产一级毛片| 波野结衣二区三区在线| 爱豆传媒免费全集在线观看| 18禁在线播放成人免费| 性色avwww在线观看| 人妻少妇偷人精品九色| 国产精品一区www在线观看| 中文乱码字字幕精品一区二区三区| 亚洲高清免费不卡视频| 国产精品一区www在线观看| 99热这里只有精品一区| videossex国产| 亚洲精品国产成人久久av| 国产片特级美女逼逼视频| 久久久久久久大尺度免费视频| 国产91av在线免费观看| 久热这里只有精品99| av.在线天堂| 国产精品偷伦视频观看了| 亚洲国产av新网站| 亚洲色图综合在线观看| 99热全是精品| 久久久久久久国产电影| 亚洲第一av免费看| 国产一区二区三区综合在线观看 | 成人国产麻豆网| a级毛片免费高清观看在线播放| 国产成人免费无遮挡视频| 六月丁香七月| 国产精品久久久久久精品电影小说| 99久久人妻综合| 人妻系列 视频| 免费人成在线观看视频色| 国产精品一区二区性色av| 国产男人的电影天堂91| 欧美精品亚洲一区二区| 欧美高清成人免费视频www| freevideosex欧美| 亚洲欧美日韩卡通动漫| tube8黄色片| 午夜福利在线观看免费完整高清在| www.色视频.com| 精品国产国语对白av| 亚洲精品一区蜜桃| 欧美精品一区二区大全| 美女xxoo啪啪120秒动态图| 国产中年淑女户外野战色| 国产又色又爽无遮挡免| 日本午夜av视频| 国产av国产精品国产| 国产欧美日韩一区二区三区在线 | 街头女战士在线观看网站| 国产精品女同一区二区软件| 国产爽快片一区二区三区| 性高湖久久久久久久久免费观看| 又大又黄又爽视频免费| 2021少妇久久久久久久久久久| 乱人伦中国视频| 精品国产露脸久久av麻豆| 草草在线视频免费看| 免费大片18禁| 亚洲四区av| 亚洲av综合色区一区| 久久这里有精品视频免费| 亚洲欧美日韩东京热| 国产女主播在线喷水免费视频网站| 91久久精品电影网| 久久久亚洲精品成人影院| 五月伊人婷婷丁香| 亚洲自偷自拍三级| 亚洲欧洲日产国产| 少妇猛男粗大的猛烈进出视频| 亚洲va在线va天堂va国产| 久久久久国产精品人妻一区二区| 亚洲四区av| 亚洲成人一二三区av| 欧美97在线视频| 国产免费福利视频在线观看| 亚洲国产精品一区三区| 国产在线一区二区三区精| 精品国产一区二区三区久久久樱花| 亚洲av.av天堂| 一区二区三区免费毛片| 三上悠亚av全集在线观看 | 一级毛片 在线播放| 丰满人妻一区二区三区视频av| 男女边吃奶边做爰视频| 好男人视频免费观看在线| 99热全是精品| 免费黄网站久久成人精品| 人人妻人人看人人澡| 色94色欧美一区二区| 蜜臀久久99精品久久宅男| 桃花免费在线播放| 一级毛片电影观看| 亚洲成人手机| 久久精品国产亚洲网站| 欧美成人午夜免费资源| 在线观看免费视频网站a站| 大香蕉久久网| 国产极品天堂在线| av又黄又爽大尺度在线免费看| 成人特级av手机在线观看| 夜夜骑夜夜射夜夜干| 精品亚洲乱码少妇综合久久| 欧美日本中文国产一区发布| a级毛片在线看网站| 亚洲精品色激情综合| 最近2019中文字幕mv第一页| 色94色欧美一区二区| 亚洲国产欧美在线一区| 久热久热在线精品观看| 国产高清不卡午夜福利| 狂野欧美激情性xxxx在线观看| 天堂俺去俺来也www色官网| 国产精品一区二区三区四区免费观看| 精品午夜福利在线看| 乱系列少妇在线播放| 伦精品一区二区三区| 亚洲欧洲国产日韩| 免费黄色在线免费观看| h日本视频在线播放| 又粗又硬又长又爽又黄的视频| 日韩 亚洲 欧美在线| 亚洲在久久综合| 我的女老师完整版在线观看| 性色av一级| 三级经典国产精品| 男女国产视频网站| 久久久a久久爽久久v久久| 久久久久久久久久久久大奶| 国产精品久久久久久av不卡| 免费看不卡的av| 少妇丰满av| 老司机影院成人| 熟女av电影| 视频区图区小说| 一本一本综合久久| 伦理电影大哥的女人| 午夜激情久久久久久久| 女的被弄到高潮叫床怎么办| 男女啪啪激烈高潮av片| 亚洲,一卡二卡三卡| av在线老鸭窝| 国产一区二区在线观看日韩| 水蜜桃什么品种好| 成人亚洲欧美一区二区av| 精品久久国产蜜桃| av线在线观看网站| 国产黄频视频在线观看| 亚洲精品国产av成人精品| 成人午夜精彩视频在线观看| 亚洲av男天堂| 一级av片app| 你懂的网址亚洲精品在线观看| 男女边吃奶边做爰视频| 高清在线视频一区二区三区| av播播在线观看一区| 一区二区三区乱码不卡18| 少妇的逼水好多| 久久精品久久久久久噜噜老黄| 美女cb高潮喷水在线观看| 人人妻人人看人人澡| 国产一区二区在线观看av| 丝袜脚勾引网站| 日韩,欧美,国产一区二区三区| 日本91视频免费播放| 亚洲欧美一区二区三区黑人 | 九草在线视频观看| av一本久久久久| 黑人巨大精品欧美一区二区蜜桃 | 黄色配什么色好看| 丰满乱子伦码专区| 国产高清三级在线| 日韩精品有码人妻一区| 建设人人有责人人尽责人人享有的| 女性生殖器流出的白浆| 国产日韩一区二区三区精品不卡 | 精品一区二区三卡| 亚洲伊人久久精品综合| 亚洲国产欧美日韩在线播放 | 色婷婷久久久亚洲欧美| 成人亚洲精品一区在线观看| 人人妻人人添人人爽欧美一区卜| 国产成人精品一,二区| 国产一区二区三区av在线| 一本大道久久a久久精品| 国产伦理片在线播放av一区| 精品午夜福利在线看| 久热久热在线精品观看| 日产精品乱码卡一卡2卡三| 性色avwww在线观看| 人人妻人人爽人人添夜夜欢视频 | 中文字幕亚洲精品专区| 婷婷色麻豆天堂久久| 精品久久久精品久久久| 国产在线免费精品| 丰满迷人的少妇在线观看| 乱码一卡2卡4卡精品| 欧美亚洲 丝袜 人妻 在线| 一本久久精品| 亚洲经典国产精华液单| 久久综合国产亚洲精品| 成人综合一区亚洲| 日韩不卡一区二区三区视频在线| 一级黄片播放器| 精品国产一区二区三区久久久樱花| 丰满乱子伦码专区| 丁香六月天网| 国产极品粉嫩免费观看在线 | 国产精品伦人一区二区| 男女边摸边吃奶| 欧美日韩国产mv在线观看视频| 国产又色又爽无遮挡免| 大片电影免费在线观看免费| 新久久久久国产一级毛片| 欧美亚洲 丝袜 人妻 在线| 大香蕉97超碰在线| 欧美日韩综合久久久久久| 亚洲情色 制服丝袜| 国产免费又黄又爽又色| 国产成人免费观看mmmm| 内地一区二区视频在线| 国产 一区精品| 久久韩国三级中文字幕| 丝袜喷水一区| 亚洲av欧美aⅴ国产| 国产男女内射视频| 人妻系列 视频| 波野结衣二区三区在线| 久久人人爽人人爽人人片va| 97精品久久久久久久久久精品| 一个人免费看片子| 精品人妻偷拍中文字幕| 久久久久久久精品精品| 亚洲av在线观看美女高潮| 全区人妻精品视频| 欧美精品国产亚洲| 国产精品一二三区在线看| 亚洲欧美日韩东京热| 国产一区二区在线观看av| av女优亚洲男人天堂| 少妇的逼水好多| 亚洲一区二区三区欧美精品| 欧美国产精品一级二级三级 | 全区人妻精品视频| 久久久国产一区二区| 水蜜桃什么品种好| 国产成人午夜福利电影在线观看| av在线观看视频网站免费| 亚洲,一卡二卡三卡| 免费黄频网站在线观看国产| 99精国产麻豆久久婷婷| 97在线视频观看| 一级av片app| 伦理电影大哥的女人| 欧美日韩视频精品一区| 人人妻人人添人人爽欧美一区卜| 秋霞在线观看毛片| 中国美白少妇内射xxxbb| 22中文网久久字幕| 五月玫瑰六月丁香| 日韩亚洲欧美综合| 久久毛片免费看一区二区三区| 寂寞人妻少妇视频99o| 少妇的逼好多水| 亚洲人与动物交配视频| 亚洲人成网站在线观看播放| 免费观看a级毛片全部| 99久久综合免费| 日韩视频在线欧美| 国产精品偷伦视频观看了| 久久精品久久精品一区二区三区| 欧美老熟妇乱子伦牲交| 成人特级av手机在线观看| 日本欧美视频一区| 亚洲av男天堂| 国产亚洲av片在线观看秒播厂| 一区二区三区精品91| 日韩 亚洲 欧美在线| 亚洲欧美一区二区三区黑人 | 免费观看av网站的网址| 国产亚洲精品久久久com| 亚洲经典国产精华液单| 久久午夜福利片| 黄色一级大片看看| 少妇猛男粗大的猛烈进出视频| 国产伦理片在线播放av一区| 久久综合国产亚洲精品| 内射极品少妇av片p| 欧美另类一区| 免费人成在线观看视频色| 日韩中文字幕视频在线看片| av在线app专区| 人人妻人人澡人人看| 女的被弄到高潮叫床怎么办| 少妇丰满av| 极品教师在线视频| 五月伊人婷婷丁香| 国产爽快片一区二区三区| 欧美日韩精品成人综合77777| 久久人人爽人人片av| a级毛片在线看网站| 下体分泌物呈黄色| 美女大奶头黄色视频| 国产片特级美女逼逼视频| 亚洲欧美精品自产自拍| 性色avwww在线观看| 亚洲成色77777| 免费观看无遮挡的男女| 内射极品少妇av片p| 免费大片18禁| 人妻制服诱惑在线中文字幕| 国产成人免费无遮挡视频| 久久久久国产网址| 少妇猛男粗大的猛烈进出视频| 欧美人与善性xxx| 久久综合国产亚洲精品| 视频区图区小说| 亚洲av综合色区一区| 免费人妻精品一区二区三区视频| 少妇 在线观看| 91午夜精品亚洲一区二区三区| 色婷婷久久久亚洲欧美| 伊人亚洲综合成人网| 久久久久网色| 精品人妻偷拍中文字幕| 99热全是精品| 亚洲熟女精品中文字幕| 亚洲人成网站在线播| av福利片在线| 亚洲国产精品国产精品| 成人国产av品久久久| 乱码一卡2卡4卡精品| 色婷婷av一区二区三区视频| 久久久a久久爽久久v久久| 一本—道久久a久久精品蜜桃钙片| 欧美成人午夜免费资源| av免费在线看不卡| 香蕉精品网在线| 边亲边吃奶的免费视频| 视频中文字幕在线观看| 亚洲三级黄色毛片| 亚洲美女黄色视频免费看| 国产免费视频播放在线视频| 日韩电影二区| 少妇人妻一区二区三区视频| 亚洲综合精品二区| 日韩在线高清观看一区二区三区| 搡女人真爽免费视频火全软件| 亚洲欧美中文字幕日韩二区| 少妇高潮的动态图| 高清av免费在线| 国产欧美日韩一区二区三区在线 | 国产黄色免费在线视频| 一级毛片我不卡| 成人午夜精彩视频在线观看| 日日撸夜夜添| 欧美3d第一页| 高清黄色对白视频在线免费看 | 免费看光身美女| 一级片'在线观看视频| 91精品国产九色| 美女大奶头黄色视频| 国模一区二区三区四区视频| 亚洲精品一二三| 亚洲国产欧美在线一区| 自线自在国产av| 69精品国产乱码久久久| 丰满饥渴人妻一区二区三| 91久久精品电影网| 青春草国产在线视频| 街头女战士在线观看网站| 欧美成人精品欧美一级黄| 日韩中字成人| videos熟女内射| 在线观看三级黄色| 国产av一区二区精品久久| 我要看日韩黄色一级片| 国产欧美亚洲国产| 一区二区三区乱码不卡18| 大话2 男鬼变身卡| 少妇 在线观看| 亚洲欧美精品专区久久| 亚洲不卡免费看| 亚洲美女视频黄频| 少妇精品久久久久久久| 国产亚洲午夜精品一区二区久久| 国产成人午夜福利电影在线观看| 一级二级三级毛片免费看| 日韩,欧美,国产一区二区三区| 久久精品久久精品一区二区三区| 日韩,欧美,国产一区二区三区| 免费观看性生交大片5| 伦理电影大哥的女人| 午夜老司机福利剧场| 亚洲第一区二区三区不卡| 亚洲国产色片| av在线观看视频网站免费| 新久久久久国产一级毛片| 老司机影院毛片| 一区二区三区免费毛片| 成人毛片60女人毛片免费| 免费在线观看成人毛片| 日韩中字成人| 国产色爽女视频免费观看| 精品久久久久久久久av| 亚洲久久久国产精品| 日日爽夜夜爽网站| 黑丝袜美女国产一区| 日本欧美国产在线视频| 国产极品天堂在线| 自线自在国产av| 亚州av有码| av视频免费观看在线观看| 日韩成人av中文字幕在线观看| 精品人妻熟女毛片av久久网站| 免费观看性生交大片5| 亚洲成人av在线免费| 精品一区二区三区视频在线| 国产精品成人在线| 久久精品国产a三级三级三级| av免费在线看不卡| 亚洲人成网站在线观看播放| 欧美xxxx性猛交bbbb| 一级,二级,三级黄色视频| 国产精品久久久久久精品古装| 最黄视频免费看| 日日啪夜夜爽| 多毛熟女@视频| 不卡视频在线观看欧美| 最近手机中文字幕大全| 亚洲精品视频女| 亚洲国产成人一精品久久久| 大码成人一级视频| a级片在线免费高清观看视频| 丝袜喷水一区| 精品一区在线观看国产| av有码第一页| av女优亚洲男人天堂| 国产精品福利在线免费观看| 久久精品国产鲁丝片午夜精品| 精品国产一区二区久久| h视频一区二区三区| 久久久久久久精品精品| 欧美日韩一区二区视频在线观看视频在线| 色婷婷久久久亚洲欧美| 中文欧美无线码| 十八禁高潮呻吟视频 | 久久久国产欧美日韩av| 亚洲美女视频黄频| 精品99又大又爽又粗少妇毛片| 久久99蜜桃精品久久| 亚洲人与动物交配视频| 成年美女黄网站色视频大全免费 | 最近手机中文字幕大全| 人体艺术视频欧美日本| av国产久精品久网站免费入址| 亚洲欧美日韩东京热| 视频中文字幕在线观看| 国产精品99久久久久久久久| 美女xxoo啪啪120秒动态图| 久久久久久久久久成人| 青青草视频在线视频观看| 久久av网站| 一级毛片久久久久久久久女| 亚洲精品一二三| 美女xxoo啪啪120秒动态图| 麻豆成人av视频| 天堂8中文在线网| 伊人久久精品亚洲午夜| 一级毛片我不卡| 这个男人来自地球电影免费观看 | 女人久久www免费人成看片| 在线 av 中文字幕| 成人黄色视频免费在线看| av在线老鸭窝| 亚洲情色 制服丝袜| 最近中文字幕2019免费版| 亚洲人与动物交配视频| 日本wwww免费看| 国产成人精品婷婷| 久久精品国产亚洲av涩爱| 春色校园在线视频观看| 国产精品福利在线免费观看| 免费黄频网站在线观看国产| 国产综合精华液| 在线天堂最新版资源| 国产视频内射| 亚洲av男天堂| 午夜福利,免费看| 国语对白做爰xxxⅹ性视频网站| 国内少妇人妻偷人精品xxx网站| 丰满饥渴人妻一区二区三| 我的女老师完整版在线观看| 尾随美女入室| 亚洲综合精品二区| 欧美精品一区二区免费开放| 在线观看一区二区三区激情| 99热这里只有是精品50| 国产亚洲91精品色在线| 精品视频人人做人人爽| 国产男女内射视频| 久久毛片免费看一区二区三区| 男女边吃奶边做爰视频| 日韩在线高清观看一区二区三区| 伦理电影免费视频| 中文资源天堂在线| 嫩草影院新地址| 欧美激情极品国产一区二区三区 | 中文天堂在线官网| 秋霞在线观看毛片| 日本av免费视频播放| 99热这里只有精品一区| 久久久久久伊人网av| 精品久久久噜噜| 97在线视频观看| 亚洲无线观看免费| 一二三四中文在线观看免费高清| 精品少妇黑人巨大在线播放| 一二三四中文在线观看免费高清| 国产成人精品福利久久| 春色校园在线视频观看| 哪个播放器可以免费观看大片| 国产色爽女视频免费观看| 又粗又硬又长又爽又黄的视频| 久久精品熟女亚洲av麻豆精品| 一边亲一边摸免费视频| 久久这里有精品视频免费| 噜噜噜噜噜久久久久久91| 亚洲国产精品999| av国产精品久久久久影院| 免费av不卡在线播放| 亚洲av免费高清在线观看| 22中文网久久字幕| 久久99热6这里只有精品| 久久精品久久久久久噜噜老黄| 欧美成人午夜免费资源| 免费不卡的大黄色大毛片视频在线观看| 久久久a久久爽久久v久久| 亚洲内射少妇av| 色5月婷婷丁香| 男男h啪啪无遮挡| 亚洲第一区二区三区不卡| 狂野欧美白嫩少妇大欣赏| 国产真实伦视频高清在线观看| 国产精品蜜桃在线观看| 一级毛片 在线播放| 国产伦精品一区二区三区四那| 久久久精品94久久精品| 欧美人与善性xxx| 六月丁香七月| 亚洲精品亚洲一区二区| 老女人水多毛片| 久久97久久精品| 亚洲av中文av极速乱| 又爽又黄a免费视频| 在线观看美女被高潮喷水网站| 永久网站在线| 久久久久久人妻| 一级二级三级毛片免费看| 国产成人精品无人区| 久久久久久久久久久丰满| 在线观看美女被高潮喷水网站| 亚洲欧美一区二区三区黑人 | 亚洲av.av天堂| 99re6热这里在线精品视频| 熟女人妻精品中文字幕| 最近最新中文字幕免费大全7| 久久久久久久久久成人| 简卡轻食公司| 桃花免费在线播放| 国产欧美亚洲国产| 国产精品国产三级国产专区5o| 色吧在线观看|