• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Evaluating the Risk of Disclosure and Utility in a Synthetic Dataset

    2021-12-14 09:59:20KangChengChenChiaMuYuandTooskaDargahi
    Computers Materials&Continua 2021年7期

    Kang-Cheng Chen,Chia-Mu Yuand Tooska Dargahi

    1Industrial Technology Research Institute,Hsinchu,310,Taiwan

    2National Chiao Tung University,Hsinchu,320,Taiwan

    3University of Salford,Manchester,M5 4WT,United Kingdom

    Abstract: The advancement of information technology has improved the delivery of fnancial services by the introduction of Financial Technology(FinTech).To enhance their customer satisfaction,Fintech companies leverage artifcial intelligence(AI)to collect fne-grained data about individuals,which enables them to provide more intelligent and customized services.However,although visions thereof promise to make customers’ lives easier, they also raise major security and privacy concerns for their users.Differential privacy(DP)is a common privacy-preserving data publishing technique that is proved to ensure a high level of privacy preservation.However,an important concern arises from the trade-off between the data utility the risk of data disclosure(RoD), which has not been well investigated.In this paper, to address this challenge,we propose data-dependent approaches for evaluating whether the suffcient privacy is guaranteed in differentially private data release.At the same time, by taking into account the utility of the differentially private synthetic dataset, we present a data-dependent algorithm that, through a curve ftting technique, measures the error of the statistical result imposed to the original dataset due to the injection of random noise.Moreover, we also propose a method that ensures a proper privacy budget, i.e., ?will be chosen so as to maintain the trade-off between the privacy and utility.Our comprehensive experimental analysis proves both the effciency and estimation accuracy of the proposed algorithms.

    Keywords: Differential privacy; risk of disclosure; privacy; utility

    1 Introduction

    Financial Technology (FinTech) concept has evolved as a result of integrating innovative technologies into fnancial services, e.g., AI and big data, Blockchain and mobile payment technologies, to provide better fnancial services [1].Investments in FinTech industry is trending upward, such that by September 2020 the global investment in Fintech was $25.6 Billion, reported by KPMG [2].However, security and privacy of the users’ data is among the main concerns of the FinTech users [3].Data governance and user data privacy preservation are reported as the most important challenges in FinTech due to the accessibility of user data by suppliers and third parties [3].Some fnancial institutions rely on honest but curious FinTech service providers which might be interested in accessing sensitive attributes of users’ data.Specially, in the case of small and medium businesses, which provide personalized fnancial services to their customers,with no background knowledge in security and data privacy, protection of users’ data becomes even more challenging.

    The “open data”concept and data sharing for banking industry has been promoted by several countries, including the UK.A report by the Open Data Institute, discusses the benefts of sharing anonymized bank account data with the public and suggests that such data release could improve customers decision making [4].User data are usually shared with the data consumers via data release; herein, we consider scenarios in which data are released, i.e., where tabular data with numeric values (which could be related to user’s bank transactions, stock investments, etc.)are to be published (or shared) by the data owner.However, data release often presents the problem of individual privacy breach, and there has always been a debate between data privacy and openness, reported by Deloitte [5].A real-world example thereof is that, by using publicly available information, researchers from the University of Melbourne were able to re-identify seven prominent Australians in an open medical dataset [6].Furthermore, researchers from Imperial College London found that it would be possible to correctly re-identify 99.98% of Americans in any dataset by using 15 demographic attributes [7].Other relevant privacy incidents were also reported [8,9].All of these incidents provide evidence of privacy leakage because of improper data release.

    Recently, the United States and Europe launched new privacy regulations such as the California Consumer Privacy Act (CCPA) and General Data Protection Regulation (GDPR) to strictly control the manner in which personal data are used, stored, exchanged, and even deleted by data collectors (e.g., corporations).Attempts to assist law enforcement have given rise to a strong demand for the development of privacy-preserving data release (PPDR) algorithms, together with the quantitative assessment of privacy risk.Given an original dataset (the dataset to be released),PPDR aims to convert the original dataset into a sanitized dataset (or a private dataset) such that privacy leakage by using the sanitized dataset is controllable and then publish the sanitized dataset.In the past, the former demands could be satisfed by conventional approaches such ask-anonymity,l-diversity, andt-closeness.However, these approaches have shortcomings in terms of syntactic privacy defnition and the diffculty in distinguishing between quasi-identifers and sensitive attributes (the so-called QI fallacy), and therefore are no longer candidates for PPDR.In contrast, differential privacy (DP) [10] can be viewed as a de-facto privacy notion, and many differentially private data release (DPDR) algorithms [11] have been proposed and even used in practice.Note that DPDR can be considered as a special type of PPDR with DP as a necessary privatization technique.

    Although it promises to maintain a balance between privacy and data utility, the privacy guarantee of DPDR is, in fact, only slightly more explainable.Therefore, in the case of DPDR,it is diffcult to choose an appropriate confguration for the inherent privacy parameter, privacy budget?.More specifcally, DP uses an independent parameter?that controls the magnitude of the injected noise, yet the selection of?such that the data utility is maximized remains problematic.On the other hand, although the value of?affects the magnitude of noise, it has no direct relevance to the risks of data disclosure, such as the probability of re-identifying a particular individual.In other words, the choice of?such that the privacy is meaningfully protected still needs to be investigated.In practice, there is no clear recommendation by the regulatory institutions in Fintech industry on the preferred anonymization technique which could address the challenge of preserving privacy while providing openness.This might be due to the unclarity of privacy guarantee versus utility in the existing DPDR algorithms.Thus, a strong demand exists to develop novel measures for the risk of disclosure (RoD) and utility for DPDR.

    1.1 Related Work

    In this section we present a brief review of studies on the differentially private data release and the risk of disclosure.

    1.1.1 Differentially Private Data Release(DPDR)

    Privacy-preserving data release (PPDR) methods have been introduced to address the challenge of the trade-off between privacy and utility of a released dataset.Several anonymization and privacy-enhancing techniques have been proposed in this regard.The most popular technique isk-anonymity [12], which uses generalization and suppression to obfuscate each record between at leastk?1 other similar records.Due to vulnerability of thek-anonymity against sensitive attribute disclosure where attackers have background knowledge,l-diversity [13] andt-closeness [14] are proposed to further diversify the record values.However, all of these techniques are proven to be theoretically and empirically insuffcient for privacy protection.

    Differential privacy (DP) [10] is another mainstream privacy preserving technique which aims to generate an obfuscated dataset where addition or removal of a single record does not affect the result of the performed analysis on that dataset.Since the introduction of DP, several DPDR algorithms have been proposed.Here, we place particular emphasis on the synthetic dataset approach in DPDR.Namely, the data owner generates and publishes a synthetic dataset that is statistically similar to the original dataset (i.e., the dataset to be released).It should be noted that, since 2020, the U.S.Census Bureau has started to release census data by using the synthetic dataset approach [15].DPDR can be categorized into two types:Parametric and non-parametric.The former relies on the hypothesis that each record in the original dataset is sampled from a hidden data distribution.In this sense, DPDR identifes the data distribution, injects noise into the data distribution, and repeatedly samples the noisy data distribution.The dataset constructed in this manner is, in fact, not relevant to the original dataset, even though they share a similar data distribution, and can protect individual privacy.The latter converts the original dataset into a contingency table, where i.i.d.noise is added to each cell.The noisy contingency table is then converted to the corresponding dataset representation.This dataset can be released without privacy concerns because each record can claim plausible deniability.

    Two examples in the category of parametric DPDR are PrivBayes [16] and JTree [17].In particular, PrivBayes creates a differentially private but high-dimensional synthetic datasetDby generateing a low-dimensional Bayesian networkN.PrivBayes is composed of three steps:1) Network learning, where ak-degree Bayesian networkNis constructed over the attributes in the original high-dimensional datasetOusing an(?/2)-DP algorithm; herekrefers to a small value dependent on the affordable memory size and computational load.2) Distribution learning:an(?/2)-DP algorithm is used to generate a set of conditional distributions, such that each attribute-parent (AP) inNhas a noisy conditional distribution.3) Data synthesis:Nanddnoisy conditional distributions are used to obtain an approximation of the input distribution, and then from the approximate distribution, we sample tuples to generate a synthetic datasetD.

    JTree [17] proposes to use Markov random feld, together with a sampling technique, to model the joint distribution of the input data.Similar to PrivBayes, JTree consists of four steps:1) Generate the dependency graph:The goal of this step is to calculate the pairwise correlation between attributes through the sparse vector technique (SVT), leading to a dependency graph.2) Generate attribute clusters:Once the pairwise correlation between attributes is computed, we use junction tree algorithm to generate a set of cliques from the above dependency graph.These attribute cliques will be used to derive noisy marginals with the minimum error.3) Generate noisy marginals:We generate a differentially private marginal table for each attribute cluster.After that, we also apply consistency constraints to each differentially private marginal table, as a postprocessing, to enhance the data utility.4) Produce a synthetic dataset:From these differentially private marginal tables, we can effciently generate a synthetic dataset while satisfying differential privacy.Other methods in the category of parametric DPDR include DP-GAN [18–20],GANObfuscator [21], and PATE-GAN [22].

    On the other hand, Priview [23] and DPSynthesizer [24] are two representative examples in the category of non-parametric DPDR.Priview and DPSynthesizer are similar in that they frst generate different marginal contingency tables.The main difference between parametric and nonparametric DPDR lies in the fact that the former assumes a hidden data distribution, whereas the latter processes the corresponding contingency table directly.Specifcally, noise is applied to each cell of the contingency tables to derive the noisy table.Noisy marginal contingency tables are combined to reconstruct the potentially high-dimensional dataset, followed by a sophisticated design of the post-processing step for further utility enhancement.Other methods in the category of non-parametric DPDR include DPCube [25], DPCopula [26], and DPWavelet [27].

    1.1.2 Risk of Disclosure(RoD)Metrics

    Not much attention has been paid to develop RoD, although DP has its own theoretical foundation for privacy.The research gap arises because the privacy of DP has been hidden in the corresponding defnition, according to which the query results only differ negligibly from those of neighboring datasets.However, in the real world setting, the user prefers to know whether (s)he will be re-identifed.Moreover, the user wants to know what kind of information is protected,what the corresponding privacy level is, and the potentially negative impact of the perturbation for the statistical analysis tasks.On the other hand, although many DPDR algorithms have emerged,because of the lack of a clear and understandable defnition of RoD, we know that the choice of?is critical but it hinders the practical deployment of DPDR systems.Thus, we eager to develop an RoD to quantitatively answer questions such as what kind of information is protected, what the corresponding privacy level is, and the potentially negative impact of the perturbation for the statistical analysis tasks properly.

    To make the privacy notion easy to understand by layperson, Lee et al.[28] made the frst step to have a friendly defnition of the RoD.They adopt the cryptographic game approach defne the RoD.Specifcally, given a datasetOwithmrecords, the trusted server randomly determines,by tossing a coin, whether a recordr∈Owill be deleted.LetDbe the resulting dataset after the deletion of the chosen record.The attacker’s objective is to determine whether the recordrexists.Here, the attacker is assumed to be equipped with an arbitrary knowledge of the datasetsOandD.In this sense, Lee and Clifton formulated the attacker as a Bayesian attacker, which means that the attacker is aimed to maximize the probability of guessing correctly by using both the prior and posterior knowledge ofO,D, andr.

    Hsu et al.[29], on the other hand, propose to choose the parameter?based on an economic perspective.The idea behind their work is based on the observation that a normal user has a fnancial incentive to contribute sensitive information (if the third party or even attacker provide the fnancial compensation).Their economics-inspired solution [29] can calculate an proper?by striking a balance between the accuracy of the released data and?.In addition, Naldi et al.[30]calculated the parameter?according to estimation theory.More specifcally, Naldi et al.put their emphasis only on the counting query (as the counting query leads to the minimal sensitivity) and defne their RoD.Their solution is similar to the solution in this paper.Nonetheless, the restriction on the counting query implies the limited practicality.Tsou et al.[31] also presented an RoD and their RoD is defned by restricting the confdence level of the magnitude of Laplace noise.With an observation that the data owner who wants to evaluate the RoD is only in possession of an original datasetOand a candidate differentially private synthetic datasetD.With another observation that the magnitude of Laplace noise can be bounded with high probability, the value range of the Laplace noise can then be estimated with a certain confdence level.As a result, the estimated value range of the Laplace noise can be used to determine the level of distortion for the record values, and this also implies the privacy level.

    1.2 Problem Statement

    The assessment of the RoD and data utility of the DP synthetic dataset presents the following two diffculties.

    ? RoD requires good explainability for layman users in terms of a privacy guarantee, while simultaneously allowing quantitative interpretation to enable the privacy effect of different DPDR algorithms to be compared.In particular, although the privacy budget?in DP has a mathematical explanation, it is diffcult for layman users to comprehend the meaning behind the defnition.Moreover, the privacy budget is inconsistent with the requirements in the current privacy regulations such as GDPR and CCPA, because the relation between?and legal terms such as “single-out” remains to be investigated.

    ? Usually, it is necessary to generate a synthetic dataset and then estimate the corresponding privacy level by calculating the RoD.Nonetheless, this process requires an uncertain number of iterative steps until the pre-defned privacy level is reached, leading to ineffciency in synthetic dataset generation.Thus, the aim is to develop a solution that can effciently estimate the privacy level of the DP synthetic dataset.

    Though the methods in Section 1.1 can be used to determine?(or to quantify the privacy level), all of them have inherent limitations as mentioned previously.In particular, two of these studies [28,29] can apply only in the case of interactive DP, where the data analyst keeps interacting with the server and receives query results from the server.Nevertheless, as interactive DP suffers from the privacy budget completion problem, the DPDR in this paper only considers non-interactive DP (see Section 2.1), which results in the publication of the differentially privacy dataset allowing an arbitrary number of queries.Thus, the studies [28,29] are not applicable to the assessment of RoD.Moreover, though the method in [30] is somewhat related to non-interactive DP, its application is limited to the counting queries.Sarathy et al.[32] also have a similar limitation since their work only applies to numerical data.Lastly, Tsou et al.[31] method works only when the synthetic dataset is synthesized with the injection of Laplacian noise.Unfortunately,this is not always the case, because Laplacian noise leads to synthetic dataset with awful utility and therefore the data owner might choose alternatives.The design of PriBayes, JTree, Priview,and DPSynthesizer (in Section 1.1.1) similarly indicate that a more sophisticated design is used for DPDR, further limiting the applicability of this work [31].

    1.3 Contribution

    Our work makes the following two contributions:

    ? We defne a notion for evaluating the risk of disclosure (RoD) particularly for the DP synthetic dataset.The state-of-the-art DPDR algorithms decouple the original dataset from the synthetic dataset which makes it diffcult to evaluate the RoD.However, we strive to quantify the RoD without making unrealistic assumptions.

    ? We propose a framework for effciently determining the privacy budget?in DP, using the curve ftting approach, taking into consideration the desired trade-off between the data utility and privacy.

    2 Preliminaries

    2.1 Differential Privacy(DP)

    The scenario in our consideration is a trusted data owner with a statistical database.The database stores a sensitive dataset.The database constructs and publishes a differentially private synthetic dataset for the public.In this sense, DP has been a de facto standard for protecting not only the privacy in the interactive (statistical database) query framework but also the (noninteractive) data release framework (see below).

    There are two different kinds of DP scenarios, interactive and non-interactive.In the former,a dedicated and trusted server is located between the data analyst, who issues queries to the server(the data owner), and server (data owner).The server is responsible for answering the queries issued by data analyst.However, to avoid information leakage from query results, before forwarding the query result, the server will perturb it.Obviously, the interactive setting is cumbersome because in reality the data owner needs to setup a dedicated server.On the contrary, in the latter,the server (data owner) simply releases a privatized dataset to the public after the sanitization of dataset.During the whole process, no further interaction with anyone is needed.The synthetic dataset approach is a representative for non-interactive DP.Throughout the paper, we consider the non-interactive setting of DP (i.e., DPDR) unless stated otherwise.

    Let?andMbe a positive real number and a randomized algorithm with the dataset as the input, respectively.We claim thatMis?-DP if, for all neighboring datasetsD1andD2that differ at most one single record (e.g., the data of one person), and all subsetsSof the image ofM,

    where the parameter?can be adjusted according to the tradeoff between utility and privacy;a higher?implies lower privacy.Therefore,?is also called theprivacy budgetbecause, the number of query responses is positively proportional the privacy loss.From the above defnition, we can also know that DP provides a cryptographic privacy guarantee (from indistinguishability point of view) that the presence or absence of a specifc record will not affect the algorithm signifcantly.From attacker’s point of view, (s)he cannot tell whether a specifc record exists given the access of the algorithm output.

    DP can be achieved by injecting a zero-mean Laplace noise [33].Specifcally, the noise sampled from a zero-mean Laplace distribution is added to perturb the query answer.Then, the data analyst only receives the noisy query answer.With two parameters on the mean and variance,Laplace distribution is determined jointly by?and global sensitivity,

    of the query functionq; that is, for any queryqand mechanismM,

    is?-DP, whereLapis a random variable that follows a zero-mean Laplace distribution.Apparently, as stated above,?determines the tradeoff between privacy and data utility.DP features the sequential composition theorem, which states that by querying the datasetktimes,if each noisy response satisfes?-DP, then all thekqueries achievek?-DP together.In addition,DP involves post-processing, which states that any kind of data-independent processing of a noisy answer (?-DP) does not compromise its privacy guarantee.

    2.2 Voronoi Diagram

    In our proposed algorithm in Section 3.2.3, we take advantage of the Voronoi Diagram, a mathematical concept which refers to partitioning a plane into adjacent regions called Voronoi cells to cover a fnite set of points [34].The defnition of a Voronoi diagram is as follows [34,35]:LetXbe a set ofnpoints (called sites or generators) in the plane.For two distinct pointsx,y∈Xthe Voronoi region/cell associated toxis the set of all points inthe planethat are closer toxthan to any other point in the plane (i.e.the nearest neighbor to the point).In other words,the region associated toxis all the points in the plane lying in all of the dominances ofy,i.e.,region(x)=dominance(x,y), wheredominance(x,y)={p∈R2|l2(p,x)≤l2(p,y)},wherel2is the Euclidean distance.Due to specifc geometrical structure of Voronoi diagrams and their simplicity in visual perception, they have been used by several research studies, such as fle searching, scheduling and clustering [34].Recently, the Voronoi diagram has also been used for preserving location privacy in various research studies [36–38].In [36] the authors propose a privacy preserving model for mobile crowd computing to hide users in a cloaked area based on the Voronoi diagram.This paper takes advantage of the Voronoi diagram to provide k-anonymity for users in each Voronoi cell.In another study, Bi et al [38] combine local differential privacy and Voronoi diagram in order to preserve privacy in edge computing.

    Compared to the state-of-the-art, in this paper we adopt Voronoi diagram in a completely different manner for evaluating the RoD in a differentially private synthetic dataset.

    3 Proposed Approach for Evaluating Risk of Disclosure

    In the following, we consider the setting of an original datasetOand the corresponding differentially private synthetic datasetD, both sizedm×nand with numeric attributesA1,A2,...,An.Each record inOrepresents personal information; a concrete example ofOis a medical dataset,where each record corresponds to the diagnosis of a patient.We do not assume a particular DPDR algorithm unless stated otherwise.The DPDR algorithms in Section 1.1.1 are all available for consideration.Our goal is to develop friendly notions of both RoD and utility, ensuring that these notions can easily quantify the RoD and utility ofD, given the access toOand the satisfaction of?-DP.

    First, we discuss a simple metric for RoD in Section 3.1.In Section 3.2, we present four privacy notions.After that, we claim that the combined notion would be the best by justifying its self-explainability.Thereafter, we present our solution of how to quickly evaluate the utility of the synthetic dataset, given a specifc privacy level, in Section 3.3.At last, we present in Section 3.4 a unifed framework of calculating the maximal privacy budget?by jointly considering utility and RoD.

    3.1 Straightforward Metric for RoD

    Irrespective of the approach that was followed to generate the synthetic dataset, each record inDcould be “fake”; i.e., the existence or non-existence of a record in the synthetic dataset does not indicate the existence status of this record in the original dataset.In theory, even an attacker with abundant background knowledge cannot infer the individual information inO.Specifcally, the privacy foundation lies in the fact thatDis no longer a transformed version ofO, and one cannot linkOandD.Nonetheless, in reality, layman users are still concerned about the probability of the attacker re-identifying an individual from the synthetic dataset.We term this phenomenon the scapegoat effect.In particular, the scapegoat effect states that despite the fact that the information about an individualxinOwill almost certainly not appear inD, because a recordinDcould be suffciently similar toxand the attacker only has partial knowledge ofx, the attacker will(falsely) identifyasx.We claim the importance of the scapegoat effect because this is similar to the skewness attack inl-diversity [14].In other words, an innocent individual might be accused of a crime if they are misclassifed as either being or not being in the dataset.

    Considering that most people may be concerned about the probability of an individual being re-identifed by an attacker, given a synthetic dataset, a straightforward method for assessing the privacy level of the synthetic dataset would be to calculate the hitting rate, which is defned as the ratio of the number of overlapping records to the total numbermof records in both datasets.Despite its conceptual simplicity, the use of the hitting rate incurs two problems.

    ? First, because of the curse of dimensionality, the density of the data points in a highdimensional space is usually low.This could easily result in a very low hitting rate and an overestimated level of privacy guarantee.

    ? Second, such an assessment metric leads to a trivial algorithm for a zero hitting rate.For example, a synthetic dataset could be constructed by applying a tiny (and non-DP)amount of noise to each record in the original dataset.Owing to the noise, the records in the original and synthetic datasets do not overlap, leading to a zero hitting rate and an overestimated level of privacy.

    3.2 Our Proposed Methods for RoD

    In this section, we develop friendly privacy notions (or say friendly RoD estimation) from a distance point of view.More specifcally, as we know from the literature, in the DPDR scenario,the original datasetOis already decoupled from the synthetic datasetD.A consequence is that there is no sense to connect betweenOandD.This also implies the diffculty in defning the appropriate RoD.However, the previous work [28–31] sought different ways to create the linkage betweenOandD, as the linkage between them is the most straightforward way for human understanding.Unfortunately, the privacy notion based on the linkage inevitably incurs security faw, especially in the sense that such a linkage does not exist.

    In the following, taking the scapegoat effect into consideration, a distance-based framework,(y,p)-coverage, is frst proposed in Section 3.2.1 to minimize the scapegoat effect and to reconcile the privacy level and decoupling ofOandD.The idea behind(y,p)-coverage is inspired by the observation that, even without the knowledge of the linkage betweenOandD, the only strategy left for the attacker is still to seek the closest record inOas the candidate original record inD.However,(y,p)-coverage is not suitable for measuring RoD because it has two parameters and does not have total order (described in more detail at the end of Section 3.2.1).Subsequently, we propose RoD metrics to measure the risk of re-identifcation.

    3.2.1 (y,p)-Coverage

    Here, we propose the notion of(y,p)-coverage as a framework for evaluating RoD.The idea behind(y,p)-coverage is that, the attacker would exhaustively search for a candidate matched record inO, given access toD.In particular, givenDias theith record ofD(ith row ofD),due to the lack of pre-knowledge, the attacker’s research range would be in the neighborhood of a specifc recordDi.In this sense, the corresponding privacy metric can be formulated as a minimum weight bipartite matching problem in graph theory.Also from graph theory, we know that one can use the Hungarian Algorithm to handle minimum weight bipartite matching problem and so one can conduct the same algorithm to evaluate the RoD with the observation thatOandDwould be of the same size.In particular, Algorithm 1 is the pseudocode of the above idea,whose aim is to test if the distance-based(y,p)-coverage is fulflled.GivenOandD, we construct a complete weighted bipartite graphG=(V,E,W), where

    In the graphG, each edge has an edge weight that indicates the dissimilarity between the respective records; hence, the edge weights are defned as

    The graphGhas 2mvertices, each of which corresponds to a record fromO.Thus, each vertex can also be seen as ann-dimensional row vector.Under such a construction,Gis completely bipartite as all the vertices fromOare connected to all those fromD.No direct edge for any pair of vertices both of which are forO(andD) exists.We also note that, the notation ‖·‖ denotes the norm.Here, while many norms can be used, an implicit assumption in this paper is that we always choose to use the Euclidean distance for ‖Oi?Dj‖=‖(χ1,...,χn)‖ for certainχ1,...,χn.However, the other norm can also be used as an alternative.

    Given a matchingM, let its incidence vector bex, wherexij=1 if(i,j)∈Mand 0 otherwise,and the perfect matching of the minimum weight is a subset of edge weights such that

    wherecij=wij.Once the Hungarian algorithm is performed, we can derive the perfect matching and the corresponding edge weightω, whereωis anm-dimensional vector and theith entry ofωdenotes an edge weight of the minimum weighted bipartite matching forDi.We then calculate the number of weights less than or equal toyas a countζ,

    whereIωi≤ydenotes an indicator function withIωi≤y=1 in the case ofωi≤y, andIωi≤y=0 otherwise, given a user-defned weighty.Subsequently, we calculatep′=ζ/m.With the probabilitypfrom the user,Dis supposed to fulfll(y,p)-coverage ifp′≤p.

    Algorithm 1: (y,p)-coverage Input:User-defned weight: y Input:User-defned probability: p Input:Original dataset: O Input:DP synthetic dataset: D Output:(Yes/No) Whether D fuflls (y,p)-coverage 1: M ←HUNGARIAN ALGORITHM(O,D);2: ξ =∑mi=1 Iωi≤y;3: p′=ζ/m 4:return (p′≤p) ? fulflled:not fulflled;

    Despite its conceptual simplicity and theoretical foundation,(y,p)-coverage cannot be used for assessing RoD because(y,p)-coverage has two parametersyandpand does not have total order.Note that the purpose of developing the RoD metric is to enable layman users to conveniently choose the privacy budget?in DP.However, when the notion of(y,p)-coverage is used,it may be diffcult to determine which of, for example,(3,0.5)-coverage and(4,0.4)-coverage,improves the privacy.Hence, the following sections defne additional privacy notions with total order to enable an RoD comparison.

    3.2.2 y-Privacy

    An underlying assumption for the(y,p)-coverage that the attacker conducts a search to look for a bijective matching betweenOandDby using the Hungarian algorithm.However,this assumption may fail in reality.To ft the reality setting, one can relax such an assumption,ensuring that the attacker instead performs an exhaustive search to fnd a matching betweenOandD.In this process, two records inDmay happen to be matched against the same record inO.In(y,p)-coverage, we only keep one matching and get rid of another matching, which does not make sense.Therefore, in this section, we propose to usey-privacy to overcome the above limitation.While(y,p)-coverage can be seen as an average-case analysis,y-privacy is featured by its focuses on the worst-case analysis.

    Algorithm 2 is proposed to achievey-privacy; more specifcally, it is used for verifying whether a given dataset satisfesy-privacy.In what follows, we consider the case ofn=1 with integral values inOto ease the representation.We will relax this assumption later.However, our implementation is a bit different from the above description.Instead, we in Algorithm 2 turn our focus to fnding the mapping, instead of the matching, betweenOandDwith the minimum incurred noise.First, we fnd the minimum valuefor each recordDiinDsuch that [Di±contains one original record inO.This ensures that an original record is within the attacker’s search range.Then, for eachiny′, we calculate

    The above equation indicates that, becauseycan be seen as all of the possibilities, when the attacker sees a record, it needs to be verifed whether this was a brute-force guess.Consequently,a loweryimplies a downgrade privacy.Thus, Eq.(8) can also be seen as the probability that the attacker successfully makes a correct guess on an original record inOwithin the range[Di?y,Di+y], given that a recordDi∈Dis always at most 1/(2y+1).One can also choosey′in such a way that the median ofy′is selected asyto strike a balance between the privacy and utility.In comparison, the choice ofy′goes back to the average-case analysis, and choosing the minimumy′asyhas a similar favor of(y,p)-coverage.

    ?

    One can see that Algorithm 2 can also apply to the case ofn≥2, once we properly defne the operation [Di±, becauseDiisn-dimensional.The defnition can be derived by considering‖Di?x‖≤yfor alln-dimensional vectorsx.The same patch can be used in the operation [Di±even if the record values are foating numbers.

    Under the framework of(y,p)-coverage,y-privacy considers a general attack strategy and provides a worst-case guarantee.Compared to(y,p)-coverage,y-privacy has total order, and it is easy to see thaty-privacy is better thany′-privacy wheny≥y′in terms of the privacy guarantee.Intuitively, the above argument thaty-privacy is better thany′-privacy wheny≥y′also indicates that each record in the synthetic dataset satisfyingy-privacy will be at leasty-far away from the closest record in the original dataset, in contrast to the synthetic dataset, which satisfesy′-privacy.However,y-privacy is still not practical when adense dataset(consisting of a huge number of records) is considered.Specifcally, the common weakness ofy-privacy and(y,p)-coverage is that when the records in a dataset are seriously dense, the parameteryiny-privacy and(y,p)-coverage should be very small.This can be understood by the fact that the records are close to each other both before and after the noise injection.It turns out that the parameterybecomes less meaningful as a privacy level.

    3.2.3 Voronoi Privacy

    The notion ofy-privacy can also be generalized to consider its extreme condition.In other words, becausey-privacy considers ay-radius ball centered at each data point and considers the number of data points inOcovered by thisy-radius ball, we can follow this perspective and consider they-radius balls centered at all data points inO.The rationale behind the above consideration is to determine the arrangement of the dataset with the optimaly-privacy.Expanding the radius of ally-radius balls ultimately leads to a Voronoi diagram [39].As explained in Section 2.2,this diagram is a partition of a multi-dimensional space into regions close to each of a given set of data points.Note that a Voronoi diagram can be effciently constructed for two-dimensional data points, but for high-dimensional data points this would only be possible by using approximation algorithms [40,41].The Voronoi diagram is characterized by the fact that, for each data point, a corresponding region consisting of all points of the multi-dimensional space exists closer to that data point than to any other.In other words, all positions within a Voronoi cell are more inclined to be classifed as a particular data point.

    From the perspective of RoD, we then have an interpretation that, in terms ofy-privacy,each record inDcannot be located at these positions within the Voronoi cell; otherwise, an attacker who fnds such a record inDis more inclined to link to a particular record inO.The above argument lies in the theoretical foundation of Voronoi privacy.Algorithm 3 shows the calculation of Voronoi privacy, given access toOandD.In particular, the rationale behind Voronoi privacy is to derive the optimal privatized dataset(in terms of privacy) frst, and then calculate the distance betweenDandas an RoD metric.A larger distance implies a higher level of dissimilarity betweenDandand therefore a lower risk of data closure.

    ?

    In this sense, frst, Algorithm 3 constructs an empty datasetThe subsequent procedures gradually add records to, making it optimal in terms of privacy.Then, Algorithm 3 constructs a Voronoi diagram fromObecause the data owner would like to know the corresponding optimal privatized dataset.As mentioned previously, approximation algorithms [40,41] might be candidates for this task.Once the data owner has a Voronoi diagram fromO, the collection of data points on the Voronoi diagram constitutes the optimal privatized dataset.Thus, an infnite number of optimal privatized datasets are available.Here, we aim to fnd the optimal privatized dataset with the optimal data utility.Considering that more perturbation on the record inOimplies lower data utility, for each data point inO, the closest point on Voronoi edges would have been identifed.The data owner collects all these data points asThereafter, the data owner calculates the distance betweenDandas an RoD metric.We particularly mention that different choices of the Distance function are possible in the implementation, depending on the domain of the dataset.In general, thel2distance (Euclidean distance) can be used, whereas the earth mover distance(EMD) could also be used if the data owner was interested in quantitatively measuring the data utility in terms of machine-learning applications.

    3.2.4 p-Privacy

    Based on the observation that dependency among attributes of the dataset can be a characteristic of the privacy, Algorithm 4 defnes a novel privacy metric, calledp-privacy.This is due to the fact that, in reality, the attacker will not perform a pure random guess; instead, the attacker would make educated guesses according to the distribution ofO.The diffculty for the attacker is that (s)he does possessO.However, becauseDandOhave the similar distribution according to the DPDR, the attacker can still make educated guesses by considering only the distribution ofD.Inspired by this observation, we know that further reducing the futile combinations in the general case ofn≥2 would be necessary.Thus, by computing the correlation among attributes (similar to JTree), our frst step is to construct the dependency graphG.This step would be different from the exhaustive search iny-privacy and(y,p)-coverage.After deriving a dependency graph,we consider each linked part as a clique and obtain a clique setC.We calculateDCi, whereDCiare records with values only for the attributes inCifor each cliqueCiinC.LetUbe the set ofDCi.Then, we produce a candidate tableFwithcombinations by merging eachDCiinUto.The candidate tableFcan be seen as the records that more likely to be the records inO.Subsequently, after a comparison betweenFandO, one can fnd a countξ, where the records ofFbelong toO, and then obtain the attack probabilityp,

    ?

    In the table below, we assume an exemplar original datasetOand synthetic datasetD.Based on this assumption, we show howp-privacy works to serve as a friendly privacy notion.

    ?

    We haveC={A1,A2,(A3,A4,A5)} after the lines 1 and 2 of Algorithm 4.Next, in the line 3 of Algorithm 4, we obtainU={(3,4,8),(1,5,7),[(1,2,3),(2,3,4)]}.In the line 4, we derive the following tableF.

    ?

    In TableF, both (3, 5, 1, 2, 3) and (8, 1, 2, 3, 4) exist inO.Therefore, in step 5,ξwill be 2.Finally, in the line 6 of Algorithm 4,p=

    3.2.5 Data-Driven Approach for Determining ?

    Although how to determine a proper privacy level (notion) is presented in Section 3.2.1~3.2.4,we still eager to develop a method for choosing an proper?in DP for a given privacy level.In other words, in the previous sections, we only have a friendly explanation on the privacy but still need a concrete method to determine?.In the following, based on the curve ftting technique, we propose algorithms to obtain satisfactory values for?.

    Baseline Approach for Determining ?.Inspired by JTree [42], Algorithm 5 adopts a similar strategy from JTree to determine the?that satisfes the user’s risk and utility requirements.Apparently, if theOis uniformly distributed, only a small amount of noise will be needed to reach the desirable privacy level, because each record has the low sensitivity.However, if the data distribution of the original dataset is not uniform, additional noise is needed to protect the sensitive records, as stated in Section 2.1.

    More specifcally, the data owner can determine the sensitive records and the variance of Laplacian noise, given the marginal tables in JTree.Thus, we construct the corresponding dependency graph and calculate the marginal tables.Then, once we have a user-defned probabilitypthat represents the preferable utility and privacy levels, the value of?can be derived from the equation

    The above procedures are similar to those in JTree, except that some operations are ignored.Moreover, as count queries are the fundamental operation that can have minimal impact on the function output, the global sensitivityΔ(f)is 1.So, the 95% confdence interval(μ+2σ)ofis used in our consideration as the maximum value that representsp.This choice enables us to determine a satisfactory?via Algorithm 5.

    ?

    Unfortunately, Algorithm 5 poses certain problems, such as the possibility that, because different kinds of DP noise injection mechanisms could be used, one cannot expect that the?retrieved from Algorithm 5 can be suitable for all of DP noise injection mechanisms.On the other hand, as the utility is closely related to?, the choice of?is also critical in improving the data utility.This makes it necessary to develop a more accurate method for estimating?.

    Data-Driven Approach for Determining ?.The simplest idea to determine?is an iterative algorithm; i.e., we generate a synthetic dataset with a chosen?and see whether the utility goes to be what we want.This is a theoretically feasible solution but is very ineffcient, especially in the case of an uncertain number of iterations.Therefore, curve ftting, a data-driven approach, is proposed to learn the correlation between privacy level and?.The curve bridges between privacy and utility.So, once we derive the curve,?can be calculated instantly, given the desired level of utility.The remaining question is how we can derive the curve.The corresponding privacy levels can indeed be obtained after generating a large number of differentially private synthetic datasets with different?values.Thereafter, the curve can be learned on the basis of the learned privacy levels.However, when learning the curve, although this process enables the best ftted coeffcients,we still need to determine the type of curve.Initially, we choose exponential and polynomial curves.After that, we also choose reciprocal curves as an alternative.

    One can see from Fig.1 that the reciprocal curve of degree 2 results in the best ft.The predictions in Tab.1 are quite close to the real risk distances.

    3.3 Evaluating Utility

    3.3.1 Baseline Method for Evaluating Utility

    As mentioned in the previous sections, even though the synthetic dataset has already achieved the required privacy level, usually the data utility will be sacrifced.So, the objective of data owner is to maximize the data utility subject to the required privacy.Deriving an explicit formulation for privacy and utility is complex; instead, we resort to data-driven approach.A simple idea for deriving the utility is to iteratively generate different synthetic datasets and then determine the corresponding utility.This operation is very time-consuming.In the worst case, one needs an infnite amount of time to try an infnite number of combinations of parameters.As a result, an algorithm capable of effciently learning the utility of a given synthetic dataset is desired.

    Figure 1:Curve ftting for RoD estimation

    Table 1:Predicted RoD and the real RoD

    The statistics such as the mean, standard deviation, and variance can be seen as the most popular statistics used by the data analysts and so the metrics for evaluating data utility.In what follows, for fair comparison, after the use of synthetic datasetD, we also used these metrics to estimate the error of the result.The error of variance introduced by the synthetic datasetDcan be formulated as

    As a whole, to evaluate the variance error of the entire dataset, what we can do is to sum up the errors for each record,

    Obviously, when the synthetic dataset has smaller estimation error, it also leads to better data utility.The analysis of other statistical measures is also consistent to the ones derived from the above formulas.As a result, we used these statistical measures because of the following two reasons.First, it is because of their popularity and simplicity.Second, we can also learn an approximate distribution from these measurements.Moreover, when there are huge errors for these simple statistics, it would defnitely lead to catastrophic utility loss for the other complex statistical measures.

    3.3.2 Data-Driven Method for Evaluating Utility

    Usually, calculating the estimation error from the synthetic dataset is to calculate Eqs.(11)and (12) over the differentially private synthetic datasets with?={0.01,0.1,1,10} is the most straightforward method that we start to try.For example, Fig.2 where the input dataset is a 5×1e6 health dataset with fve attributes HEIGHT, WEIGHT, BMI, DBP, SBP1DBP is diastolic blood pressure and SBP is systolic blood pressure.The height and weight are generated by sampling from a normal distribution manually.Meanwhile,the BMI is calculated from the height and weight.Eventually,DBP and SBP are generated from BMI with some noise with small magnitude.shows the errors incurred by different settings of?.

    Iterating the above process of choosing a?, generating a synthetic dataset, and then calculating the utility would be highly ineffcient.This is due to the fact that the data owner might want to further improve the utility of the current version of the synthetic dataset.Thus, the data owner will iterate the above process again and again until a satisfactory version appears.As a whole, we decided to generate synthetic datasets for pre-determined?, and then estimate their errors only.After that, our plan is to use these information to ft a curve bridging the privacy and utility.In particular, we propose using a curve ftting, a data-driven approach, as a surrogate method to learn the correlation between?and utility measures such as the error of the mean, standard deviation, and variance.Once we have such a curve from curve ftting, we can indeed calculate?very quickly, given the desired level of utility or vice versa.

    Specifcally, in the case of?var,?varcan be obtained after generating a large number of synthetic datasets with different values of?.Thereafter, the curve could be learned using the obtained values of?var.

    However, when performing curve ftting, although this could be used to learn the coeffcients that best ft the chosen curve, one factor that we can have freedom to choose is the type of curve.Initially, the two intuitive choices are exponential and polynomial.A more counterintuitive one would be reciprocal curves.We, however, found that the reciprocal curve with the following form:

    In fact, Eq.(14) has room to be improved so as to offer better prediction results.In our consideration, we aim to calculate the errors in the cases of?= {0.01,0.05,0.1,0.5,1,5,10}.Nevertheless, only three errors are calculated for the cases of?={0.01,0.5,10}.Afterwards, we learn the curve based on these three errors.These results are shown in Fig.4, where the real statistics and predicted statistics in Tab.2 are matched almost perfectly.

    Figure 2:Different methods for the analysis of the error of variance.(a) Each attribute, (b) Log10,(c) average

    Figure 3:Curve prediction for statistical measures including mean, standard deviation, and variance.(a) Mean (b) standard deviation (c) variance

    Despite the seemingly satisfactory results in Fig.4, once we scrutinize Tab.2, we will fnd that there are negative predicted values of?={5,10}, and this result is not pleasing.Once fxing the shape of the curve ftting and reciprocal curve, we found that the main reason for predictability degradation of the ftted curve can be attributed to the insuffcient degree of the reciprocal (or the other used) curve.Consequently, when we slightly increase the degree of the reciprocal curve,we obtain

    Here, after the comparison among the results in both Tab.3 and Fig.5, one can know immediately that the predicted errors are matched against the real error values, with a curve newly learned from the data.

    Figure 4:Curve prediction for statistical measures including mean, standard deviation, and variance.(a) Mean (b) Standard deviation (c) variance

    Table 2:Comparison between the predicted and real variance errors

    Figure 5:Difference between ftted reciprocal curves with the degree 1 and 2

    Table 3:Comparison of the predicted and real variance errors after applying Eq.(15)

    3.4 Jointly Evaluating RoD and Utility

    The data utility results when varyingdin the Voronoi privacy,yin they-privacy, andpin thep-privacy, are provided in Figs.6–8, respectively.Obviously, as the RoD increases, the data utility is not maintained.This is because additional perturbation is added to the original dataset and therefore the synthetic dataset is generated from a data distribution with more noise.

    Figure 6:Voronoi privacy vs.accuracy (a) SVM (b) LR (c) MLP

    One can know that the extension from the aforementioned data-driven approaches for evaluating both the utility and RoD to a data-driven approach for determining the privacy budget?with a joint consideration of the utility and RoD can be easily achieved.In essence, from Sections 3.2 and 3.3 we will learn different (reciprocal) curves; one for the privacy level and another for utility.While the curves for the privacy and utility are unrelated at the frst glance, if we consider the DP defnition and the way of how to achieve DP, they, in fact, are correlated to each other.In this sense, multidimensional curve ftting will be an alternative for a more complex setting; i.e., it would be a candidate to be used over the curves learned from Sections 3.2 and 3.3 so as to learn a higher dimensional curve for the privacy level and utility.Since the resulting higher dimensional curve will have a parameter?, after a simple calculation when the other parameters are fxed, the privacy budget?can be determined with a joint consideration of the utility and RoD.

    Figure 7: y-privacy vs.accuracy (a) SVM (b) LR (c) MLP

    Figure 8: p-privacy vs.accuracy (a) SVM (b) LR (c) MLP

    4 Conclusion

    In this paper, we proposed a number of friendly privacy notions to measure the RoD.We also developed curve ftting-based approach to determine the privacy budget?in a data-driven manner with the joint consideration of the RoD and utility.This approach enables novice users to grab the idea behind the level of privacy protection and the data utility.As a result, these users would be able to determine an appropriate privacy budget?for DPDR, depending on the amount of privacy risk they would be prepared to tolerate and the desired utility.

    Funding Statement:The work of Chia-Mu Yu has been initiated within the project MOST 110-2636-E-009-018 of Ministry of Science and Technology, Taiwan https://www.most.gov.tw/.Tooska Dargahi is supported by the UK Royal Society Award (Grant Number IECR3183047,https://royalsociety.org/).

    Conficts of Interest:The authors declare that they have no conficts of interest to report regarding the present study.

    街头女战士在线观看网站| 精品一区二区三卡| 亚洲五月色婷婷综合| 日韩亚洲欧美综合| 国产免费视频播放在线视频| 夫妻午夜视频| 亚洲av.av天堂| 国产一区二区三区av在线| 亚洲经典国产精华液单| 久久久精品免费免费高清| 一个人免费看片子| 亚洲成人av在线免费| 国产色爽女视频免费观看| 少妇人妻精品综合一区二区| 毛片一级片免费看久久久久| 黑人欧美特级aaaaaa片| 亚洲av.av天堂| 人妻人人澡人人爽人人| 亚洲国产欧美在线一区| 最新的欧美精品一区二区| 国产一区二区在线观看av| 2018国产大陆天天弄谢| 男男h啪啪无遮挡| 韩国av在线不卡| 国产高清三级在线| 色94色欧美一区二区| 蜜臀久久99精品久久宅男| 国产精品人妻久久久影院| 视频中文字幕在线观看| 久久久久国产精品人妻一区二区| 九九爱精品视频在线观看| 欧美人与善性xxx| 日韩av在线免费看完整版不卡| 国产欧美另类精品又又久久亚洲欧美| 午夜福利影视在线免费观看| 国产高清三级在线| 高清黄色对白视频在线免费看| 精品一区二区免费观看| av国产久精品久网站免费入址| 三级国产精品片| 在线观看人妻少妇| av有码第一页| 日韩不卡一区二区三区视频在线| 日韩成人av中文字幕在线观看| 午夜激情福利司机影院| 80岁老熟妇乱子伦牲交| 欧美精品人与动牲交sv欧美| 97在线视频观看| 99九九在线精品视频| 中文字幕亚洲精品专区| 国产成人精品福利久久| 亚洲欧美成人综合另类久久久| 综合色丁香网| 国产精品蜜桃在线观看| 九草在线视频观看| 一区二区av电影网| av卡一久久| 久热久热在线精品观看| 国产在线视频一区二区| 69精品国产乱码久久久| 免费观看无遮挡的男女| av在线app专区| 免费大片黄手机在线观看| 超色免费av| 伦理电影免费视频| 国产无遮挡羞羞视频在线观看| 日韩在线高清观看一区二区三区| 在线观看免费日韩欧美大片 | 成人免费观看视频高清| 99热网站在线观看| 一本久久精品| 大又大粗又爽又黄少妇毛片口| 天堂中文最新版在线下载| 一级毛片 在线播放| 精品国产国语对白av| 内地一区二区视频在线| 九九在线视频观看精品| 五月玫瑰六月丁香| av不卡在线播放| a 毛片基地| 日韩视频在线欧美| 视频在线观看一区二区三区| 亚洲欧洲精品一区二区精品久久久 | 国产精品人妻久久久影院| 我要看黄色一级片免费的| 久久久精品区二区三区| 女性生殖器流出的白浆| 我的女老师完整版在线观看| 在线 av 中文字幕| 日韩成人伦理影院| 欧美人与性动交α欧美精品济南到 | 妹子高潮喷水视频| 午夜免费男女啪啪视频观看| 久久99一区二区三区| 只有这里有精品99| 亚洲精品av麻豆狂野| 亚洲国产欧美日韩在线播放| .国产精品久久| 国产一区二区在线观看av| 精品国产国语对白av| 新久久久久国产一级毛片| 99精国产麻豆久久婷婷| 黑人高潮一二区| 国产亚洲精品久久久com| av在线老鸭窝| 久久久久久久久久成人| 日韩电影二区| av不卡在线播放| 制服人妻中文乱码| av又黄又爽大尺度在线免费看| 色吧在线观看| 99久久中文字幕三级久久日本| av又黄又爽大尺度在线免费看| 国产亚洲最大av| 黑人欧美特级aaaaaa片| 亚洲精品av麻豆狂野| av国产久精品久网站免费入址| 久久国内精品自在自线图片| 国产老妇伦熟女老妇高清| 在线观看www视频免费| 成年av动漫网址| 夜夜爽夜夜爽视频| 国产白丝娇喘喷水9色精品| 18在线观看网站| 一级片'在线观看视频| 国模一区二区三区四区视频| 18禁动态无遮挡网站| 国产又色又爽无遮挡免| 国产免费福利视频在线观看| 国产成人免费观看mmmm| 91久久精品国产一区二区成人| 男人爽女人下面视频在线观看| 精品人妻一区二区三区麻豆| 性高湖久久久久久久久免费观看| 国产精品国产三级专区第一集| 久久久久精品久久久久真实原创| 国产男人的电影天堂91| 国产黄片视频在线免费观看| 午夜福利影视在线免费观看| 国产乱来视频区| 久久av网站| av免费观看日本| 国产免费视频播放在线视频| 黄色欧美视频在线观看| 五月伊人婷婷丁香| 我要看黄色一级片免费的| h视频一区二区三区| av在线观看视频网站免费| 欧美日韩视频高清一区二区三区二| 女的被弄到高潮叫床怎么办| 欧美xxⅹ黑人| 日本爱情动作片www.在线观看| 久久久久久人妻| 狠狠婷婷综合久久久久久88av| 亚洲av男天堂| 秋霞在线观看毛片| 女性生殖器流出的白浆| 多毛熟女@视频| 蜜桃国产av成人99| 天堂中文最新版在线下载| 久久久精品94久久精品| 亚洲精品乱码久久久v下载方式| 国产熟女欧美一区二区| 久久久久人妻精品一区果冻| 精品亚洲乱码少妇综合久久| 最新的欧美精品一区二区| 国精品久久久久久国模美| 十八禁高潮呻吟视频| 大话2 男鬼变身卡| 国产精品女同一区二区软件| 国产亚洲av片在线观看秒播厂| 国产成人精品无人区| 97在线视频观看| 欧美成人精品欧美一级黄| 亚洲人成77777在线视频| 大码成人一级视频| 亚洲成人一二三区av| 成人亚洲精品一区在线观看| 国产片内射在线| 亚洲精品中文字幕在线视频| 国产免费福利视频在线观看| 日日摸夜夜添夜夜爱| 午夜福利视频在线观看免费| 中文欧美无线码| 91精品国产国语对白视频| 免费黄频网站在线观看国产| 中文欧美无线码| 亚洲国产最新在线播放| 亚洲性久久影院| 亚洲av男天堂| 国产69精品久久久久777片| 日韩熟女老妇一区二区性免费视频| 777米奇影视久久| 人妻系列 视频| 国产精品一区二区三区四区免费观看| 欧美少妇被猛烈插入视频| 国产精品久久久久久精品古装| 免费av不卡在线播放| 有码 亚洲区| 国产无遮挡羞羞视频在线观看| 美女脱内裤让男人舔精品视频| a级毛片免费高清观看在线播放| 三上悠亚av全集在线观看| 99国产综合亚洲精品| 国产精品三级大全| 欧美日韩亚洲高清精品| 国产女主播在线喷水免费视频网站| 日韩视频在线欧美| 插逼视频在线观看| 亚洲av.av天堂| 亚洲精品国产色婷婷电影| 不卡视频在线观看欧美| 青春草国产在线视频| 国产一级毛片在线| 99国产精品免费福利视频| av播播在线观看一区| 亚洲av.av天堂| 成人影院久久| 国产日韩欧美视频二区| 一个人免费看片子| 一级毛片黄色毛片免费观看视频| 又大又黄又爽视频免费| 日韩成人av中文字幕在线观看| 中文乱码字字幕精品一区二区三区| 久久久久网色| 精品久久久噜噜| 国产色爽女视频免费观看| 考比视频在线观看| 乱码一卡2卡4卡精品| 欧美成人精品欧美一级黄| 精品视频人人做人人爽| 女性生殖器流出的白浆| 成人黄色视频免费在线看| 精品一区在线观看国产| 久久久久精品性色| 免费黄频网站在线观看国产| 久久久久久久久久久免费av| 伊人亚洲综合成人网| 亚洲欧洲日产国产| 熟女人妻精品中文字幕| 成年女人在线观看亚洲视频| 性高湖久久久久久久久免费观看| 26uuu在线亚洲综合色| 日本av手机在线免费观看| 久久鲁丝午夜福利片| 一级爰片在线观看| videosex国产| 少妇被粗大猛烈的视频| 久久人人爽av亚洲精品天堂| 国产片特级美女逼逼视频| 日韩精品免费视频一区二区三区 | 自线自在国产av| 伊人久久精品亚洲午夜| av福利片在线| 在线观看免费高清a一片| 亚洲国产成人一精品久久久| 国产一区二区三区综合在线观看 | 欧美日韩精品成人综合77777| 美女xxoo啪啪120秒动态图| 美女视频免费永久观看网站| 超色免费av| 免费日韩欧美在线观看| 成人亚洲欧美一区二区av| 中国国产av一级| 天天影视国产精品| 精品一区二区免费观看| 91精品国产九色| 久久久亚洲精品成人影院| 午夜福利影视在线免费观看| 老司机影院毛片| 一级二级三级毛片免费看| 久久久久久久亚洲中文字幕| 精品久久久久久电影网| 亚洲国产成人一精品久久久| 国产极品天堂在线| 午夜激情av网站| 在线天堂最新版资源| 久久97久久精品| 日韩成人伦理影院| 搡老乐熟女国产| 欧美精品高潮呻吟av久久| 欧美日韩精品成人综合77777| 免费黄色在线免费观看| 天堂中文最新版在线下载| 精品少妇内射三级| 看免费成人av毛片| 青春草亚洲视频在线观看| 精品久久久噜噜| 精品一区二区免费观看| av线在线观看网站| 日韩不卡一区二区三区视频在线| 午夜精品国产一区二区电影| 欧美日韩视频精品一区| 午夜激情av网站| 国产熟女欧美一区二区| 18禁在线无遮挡免费观看视频| 亚洲,欧美,日韩| 男女边吃奶边做爰视频| 亚洲三级黄色毛片| 一级片'在线观看视频| 精品亚洲成a人片在线观看| 国产 一区精品| 简卡轻食公司| 高清黄色对白视频在线免费看| 99久久综合免费| 不卡视频在线观看欧美| 校园人妻丝袜中文字幕| 精品人妻熟女毛片av久久网站| 黑人猛操日本美女一级片| 免费少妇av软件| 日日撸夜夜添| 国产深夜福利视频在线观看| 精品国产一区二区久久| 18禁观看日本| 少妇被粗大猛烈的视频| 热re99久久精品国产66热6| 国产永久视频网站| 五月天丁香电影| 国产av精品麻豆| 毛片一级片免费看久久久久| av有码第一页| 精品少妇内射三级| 国产男女超爽视频在线观看| 美女内射精品一级片tv| 国产高清不卡午夜福利| 精品卡一卡二卡四卡免费| 日本黄色日本黄色录像| 久久久久久久精品精品| 欧美人与善性xxx| 久久午夜综合久久蜜桃| 午夜福利,免费看| 黑人猛操日本美女一级片| 亚洲三级黄色毛片| 国产成人精品婷婷| 秋霞伦理黄片| 男女边吃奶边做爰视频| 精品视频人人做人人爽| 夫妻午夜视频| 国产精品人妻久久久久久| 免费av中文字幕在线| 日韩精品免费视频一区二区三区 | 久久ye,这里只有精品| 精品一品国产午夜福利视频| 国产精品免费大片| 热99久久久久精品小说推荐| 成人18禁高潮啪啪吃奶动态图 | 成人国语在线视频| 国产亚洲av片在线观看秒播厂| 色婷婷av一区二区三区视频| 91成人精品电影| 日本vs欧美在线观看视频| 国产深夜福利视频在线观看| 99热网站在线观看| 欧美精品高潮呻吟av久久| 18禁裸乳无遮挡动漫免费视频| 国产乱人偷精品视频| 最新中文字幕久久久久| 免费黄网站久久成人精品| 26uuu在线亚洲综合色| 国产黄片视频在线免费观看| 免费看光身美女| 亚洲av电影在线观看一区二区三区| 一级毛片aaaaaa免费看小| 3wmmmm亚洲av在线观看| 老司机影院成人| 成年人免费黄色播放视频| 亚洲精品日韩av片在线观看| 观看美女的网站| av在线老鸭窝| 亚洲av中文av极速乱| 在线观看国产h片| 高清在线视频一区二区三区| av网站免费在线观看视频| 校园人妻丝袜中文字幕| 丁香六月天网| 免费人妻精品一区二区三区视频| 在线观看免费高清a一片| 一级毛片aaaaaa免费看小| 亚洲精品日韩在线中文字幕| 亚洲婷婷狠狠爱综合网| 黑人巨大精品欧美一区二区蜜桃 | 亚洲欧美成人精品一区二区| 亚洲av.av天堂| 国产深夜福利视频在线观看| 最近最新中文字幕免费大全7| 久久久a久久爽久久v久久| 各种免费的搞黄视频| 在线观看一区二区三区激情| 大片电影免费在线观看免费| 久久这里有精品视频免费| 久久久精品94久久精品| 伦精品一区二区三区| 搡老乐熟女国产| 国产毛片在线视频| 免费观看性生交大片5| www.av在线官网国产| 在线观看免费高清a一片| 国产在线视频一区二区| 精品久久久久久久久av| 18禁动态无遮挡网站| 亚洲精华国产精华液的使用体验| 黄片无遮挡物在线观看| 一级二级三级毛片免费看| 亚洲av在线观看美女高潮| 国产黄色免费在线视频| 啦啦啦视频在线资源免费观看| 18+在线观看网站| 免费播放大片免费观看视频在线观看| av女优亚洲男人天堂| 成人午夜精彩视频在线观看| 久久久久精品久久久久真实原创| 18禁在线无遮挡免费观看视频| 免费人成在线观看视频色| 日韩中字成人| 国产精品嫩草影院av在线观看| 麻豆成人av视频| 色婷婷av一区二区三区视频| 一级黄片播放器| 大话2 男鬼变身卡| 卡戴珊不雅视频在线播放| 大香蕉久久网| 日韩成人av中文字幕在线观看| 久久国产精品男人的天堂亚洲 | 18禁在线无遮挡免费观看视频| 国产av码专区亚洲av| 丰满少妇做爰视频| 交换朋友夫妻互换小说| 日本黄色片子视频| 国产精品秋霞免费鲁丝片| 99久久人妻综合| 十分钟在线观看高清视频www| 国产免费又黄又爽又色| 激情五月婷婷亚洲| 不卡视频在线观看欧美| 一级毛片我不卡| 国产免费福利视频在线观看| 国产日韩欧美视频二区| 久久久久精品久久久久真实原创| 久久久国产欧美日韩av| 丰满乱子伦码专区| 熟女av电影| 免费观看无遮挡的男女| av.在线天堂| 水蜜桃什么品种好| 黄片无遮挡物在线观看| 高清毛片免费看| 爱豆传媒免费全集在线观看| 国产又色又爽无遮挡免| 中文字幕av电影在线播放| 插逼视频在线观看| 最近的中文字幕免费完整| 久久久午夜欧美精品| 嘟嘟电影网在线观看| 成人国产麻豆网| 黑人巨大精品欧美一区二区蜜桃 | 国产精品久久久久成人av| 久久99一区二区三区| 国产日韩欧美视频二区| 国产一区二区在线观看日韩| 国产成人精品无人区| 午夜福利影视在线免费观看| 亚洲精品乱码久久久久久按摩| 亚洲欧美中文字幕日韩二区| 久久久亚洲精品成人影院| 亚洲av综合色区一区| 亚洲欧美成人综合另类久久久| 少妇精品久久久久久久| 激情五月婷婷亚洲| 欧美激情极品国产一区二区三区 | av线在线观看网站| 亚洲精品乱久久久久久| 亚洲精品一二三| 91精品国产国语对白视频| 久久人妻熟女aⅴ| 国产黄色视频一区二区在线观看| 汤姆久久久久久久影院中文字幕| 亚洲精品国产av蜜桃| 桃花免费在线播放| 亚洲人成网站在线观看播放| 色视频在线一区二区三区| 国产高清不卡午夜福利| 欧美精品亚洲一区二区| 婷婷色综合大香蕉| 精品久久久噜噜| 丰满迷人的少妇在线观看| 亚洲精品美女久久av网站| 亚洲精品国产av成人精品| 精品人妻偷拍中文字幕| 午夜精品国产一区二区电影| 亚洲综合色网址| 91精品一卡2卡3卡4卡| 啦啦啦视频在线资源免费观看| 久久人妻熟女aⅴ| 免费黄色在线免费观看| 亚洲欧美成人综合另类久久久| 日韩视频在线欧美| 免费大片18禁| 一级毛片黄色毛片免费观看视频| 亚洲精品第二区| 少妇人妻 视频| 免费人成在线观看视频色| 久久人人爽av亚洲精品天堂| 99视频精品全部免费 在线| 国产老妇伦熟女老妇高清| 国产一区二区三区av在线| 国产亚洲精品第一综合不卡 | 精品一区二区三区视频在线| 只有这里有精品99| videos熟女内射| 国产精品免费大片| 黄色配什么色好看| 一本一本综合久久| 免费高清在线观看视频在线观看| 久久精品国产鲁丝片午夜精品| 九色成人免费人妻av| 国产精品99久久久久久久久| 欧美 亚洲 国产 日韩一| 男女啪啪激烈高潮av片| 欧美成人午夜免费资源| 午夜久久久在线观看| 不卡视频在线观看欧美| 国产精品嫩草影院av在线观看| 日本与韩国留学比较| videos熟女内射| 国产片内射在线| .国产精品久久| 国产黄色视频一区二区在线观看| 少妇被粗大猛烈的视频| 成年女人在线观看亚洲视频| 久久国产亚洲av麻豆专区| 免费观看av网站的网址| 国产精品一二三区在线看| 亚洲伊人久久精品综合| 少妇被粗大的猛进出69影院 | 国产精品嫩草影院av在线观看| 国产精品偷伦视频观看了| 国产亚洲av片在线观看秒播厂| 亚洲av电影在线观看一区二区三区| 波野结衣二区三区在线| 热99国产精品久久久久久7| 国产一区二区三区av在线| 欧美精品高潮呻吟av久久| 天堂8中文在线网| 全区人妻精品视频| a级毛色黄片| 成人国产av品久久久| 一本—道久久a久久精品蜜桃钙片| 国产不卡av网站在线观看| 国模一区二区三区四区视频| 免费高清在线观看视频在线观看| 亚洲精品一区蜜桃| 国产高清国产精品国产三级| 国产探花极品一区二区| 亚洲av免费高清在线观看| av播播在线观看一区| 精品亚洲成国产av| 视频中文字幕在线观看| 制服丝袜香蕉在线| 中文字幕制服av| 国产精品欧美亚洲77777| 国产欧美亚洲国产| 老司机影院成人| 精品国产乱码久久久久久小说| 美女中出高潮动态图| 国精品久久久久久国模美| 人妻夜夜爽99麻豆av| 中文字幕免费在线视频6| 国国产精品蜜臀av免费| 亚洲在久久综合| 一本大道久久a久久精品| 街头女战士在线观看网站| 丝袜喷水一区| 2018国产大陆天天弄谢| 久久毛片免费看一区二区三区| 啦啦啦视频在线资源免费观看| 91久久精品电影网| 如何舔出高潮| 十分钟在线观看高清视频www| 成人综合一区亚洲| 亚洲,一卡二卡三卡| 日本黄大片高清| 狠狠精品人妻久久久久久综合| 有码 亚洲区| 成人亚洲精品一区在线观看| 三级国产精品片| 五月玫瑰六月丁香| 成年人免费黄色播放视频| 亚洲欧美一区二区三区国产| 久久久久久久久大av| 午夜福利,免费看| 老司机影院成人| 少妇人妻 视频| 日韩在线高清观看一区二区三区| 国产在视频线精品| 日韩中字成人| 日韩欧美一区视频在线观看| 美女内射精品一级片tv| 2022亚洲国产成人精品| 亚洲精品中文字幕在线视频| 成人国产麻豆网| 国产色婷婷99| av天堂久久9| 欧美变态另类bdsm刘玥| 国产片内射在线| 欧美激情 高清一区二区三区| 成人漫画全彩无遮挡| 五月开心婷婷网| 亚洲高清免费不卡视频| 人妻夜夜爽99麻豆av| 韩国av在线不卡| 国产 一区精品| 久久精品国产亚洲av天美| 日韩精品有码人妻一区| 免费黄频网站在线观看国产| 少妇的逼好多水|