• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Data-Based Filters for Non-Gaussian Dynamic Systems With Unknown Output Noise Covariance

    2024-04-15 09:36:36ElhamJavanfarandMehdiRahmani
    IEEE/CAA Journal of Automatica Sinica 2024年4期

    Elham Javanfar and Mehdi Rahmani ,,

    Abstract—This paper proposes linear and nonlinear filters for a non-Gaussian dynamic system with an unknown nominal covariance of the output noise.The challenge of designing a suitable filter in the presence of an unknown covariance matrix is addressed by focusing on the output data set of the system.Considering that data generated from a Gaussian distribution exhibit ellipsoidal scattering, we first propose the weighted sum of norms (SON)clustering method that prioritizes nearby points, reduces distant point influence, and lowers computational cost.Then, by introducing the weighted maximum likelihood, we propose a semi-definite program (SDP) to detect outliers and reduce their impacts on each cluster.Detecting these weights paves the way to obtain an appropriate covariance of the output noise.Next, two filtering approaches are presented: a cluster-based robust linear filter using the maximum a posterior (MAP) estimation and a clusterbased robust nonlinear filter assuming that output noise distribution stems from some Gaussian noise resources according to the ellipsoidal clusters.At last, simulation results demonstrate the effectiveness of our proposed filtering approaches.

    I.INTRODUCTION

    DUE to science and technology’s rapid advances, a significant number of data are being generated in various engineering fields, including but not limited to satellite-based remote sensors, time-series systems, and telecommunication data [1].This has made it imperative to analyze and process big data in contemporary engineering design, particularly in the areas of modeling, control, and estimation.

    By appearing complex dynamics in different real-world systems such as robotics, aerospace, transportation, power grid,etc., we face significant uncertainties and less knowledge in our designs [2].In this regard, traditional methods and principles in controller design, system monitoring, and performance evaluation are challenging or infeasible.Most approaches depend on accurate physical and dynamic models and complete information about design parameters.Obtaining an exact model is hard or impossible for complex systems.Different methods to model systems can be involved in four main categories: analytical, numerical, data-driven, and hybrid models[3].In the era of machine learning, data-driven approaches offer powerful tools for identifying dynamical systems without requiring a deep understanding of the model structure[4], [5].

    Data-based methods in dynamic systems have been presented primarily in control problems and rarely in state estimation.These methods play a crucial role in system identification and balancing the lack of knowledge about essential design parameters.During the last decades since the Industrial Revolution, system identification has been a critical element in most practical complex designs.Moreover, in some cases, despite access to the system model, many conventional estimation and control structures may not be applicable due to a lack of essential information.Recently, advancements in computational capabilities, iterative learning, reinforcement learning, and deep learning have given rise to new online and offline approaches to compensate for the problems mentioned above [6]-[9].

    Estimation in dynamic systems is performed to obtain approximations of the system parameters using information from a model and any available measurements.Among all estimation and filtering methods, the Kalman filter is pervasive.It is a filter that uses the Bayesian rule to express the posterior probability in terms of the likelihood and the prior distributions [10].The classical Kalman filtering theory has two main assumptions.The first is the accuracy of prior knowledge of the system model and statistical noise features,and the second is the Gaussianity of noises.In many practical systems, noises are non-Gaussian due to environmental conditions, sensor failure, manufacturing activities, etc.Heavytailed noises are most important among these different kinds of noises [11], [12].Various filters have been designed to work out the filtering problem against heavy-tailed noises,e.g., see [13]-[18].Although these filters have suitable performance, they have limitations like knowing the accurate nominal relevant covariance matrices or comparative threshold.Since non-Gaussian heavy-tailed noises increase the covariance value of a distribution [19], filters need some considerations to compensate for it.Lack of exact models and essential knowledge about the system and noise features degrade the performance of filters; therefore, data-driven filtering methods are gaining popularity.In [20], a direct data-driven filter has been designed with no mathematical model for linear time-invariant dynamic systems with bounded disturbances and noises.The authors use set membership to estimate the sets of solutions.Furthermore, using quantized measurements,[21] proposes a data-driven filter to minimize the worst-case estimation error in the presence of bounded noises.Also, it introduces an L2- L∞approximately-optimal worst-case filter through linear programming technique.Recently, data-driven unknown input observer and state estimator for linear timeinvariant systems are investigated in [22].

    Considering all of the mentioned points, data-driven techniques have not been effectively applied in state estimation problems.Moreover, despite having information about dynamic models, conventional filtering methods are sometimes inefficient because of severe environmental noise conditions and lack of statistical features.Generally, covariance matrix estimation methods can be divided into four major groups including correlation method [23], the maximum-likelihood method [24], the covariance matching method [25] and the Bayesian method [26].Unfortunately, not only do most of these approaches assume that noises have Gaussian distribution but also they suffer from some problems such as sensitivity to outliers, reaching non-invertible covariance matrices,etc.Therefore, there is a solid need to design an appropriate filter for dynamic systems in the presence of non-Gaussian noises with less or no information about noise covariances.For non-Gaussian systems, a finite set of higher-order moments of the state and measurement noises is obtained in[27] using the correlation measurement difference method,such that the observable matrix is full rank.Moreover, some recent works such as [28], [29] present a set of suitable filters against inaccurate covariances of the process and measurement noises for non-Gaussian dynamic systems using the variational Bayesian method.Since these filters are designed based on the probability density functions, they are more sensitive to initial values of distributions’ parameters.Considering all stated problems for obtaining noise covariance, using systems output data can be an effective way to compensate for deficit knowledge.This motivates us to propose a data-based filter for time-variant non-Gaussian systems without depending on the output nominal noise and accurate process noise covariances.In the proposed approach, we assume that system output data are accessible.

    Grouping data based on their likeness is an approved concept in various science fields.In statistics, however, it can be done based on two different situations where there is prior information to gain more about the group structure or not.Unavailable information necessitates unsupervised learning tools or clustering algorithms.In other words, the problem of dividing a given set of data points with high uniformity within the groups and low diversity between groups is called clustering.Clustering is ubiquitous in machine learning, pattern recognition, statistics, image processing, and biology.Some important clustering algorithms are hierarchical clustering,Gaussian mixture models (GMMs), and K-mean clustering[30].Each of these methods bears some disadvantages.Time complexity and nonexisting mathematical objectives are primary defects of hierarchical clustering [31].Long computation time, falling into local optimum, and deciding are the dominant shortages of Gaussian mixture models [32], [33].Also, sensitivity to the initial condition and considerably different clustering results are the focal paucities of K-means clustering [34], [35].These methods are generally beset by local minima, which are sometimes significantly suboptimal.Recently, sum-of-norms (SON) clustering has been introduced that ensures a unique global minimizer [36], [37] and covers all the problems mentioned earlier.

    Clustering the system data plays a vital role in our proposed filtering scheme.The shape representation of a cluster is also vital in preserving the data features.Ellipsoidal clusters are common because many data observations are normally distributed [38].We cluster the data using the sum-of-norms clustering method according to this characteristic and the above-discussed advantages.In this method, because of choosing a threshold value as a meter for Euclidian distances,each cluster may suffer from some outliers.This changes the ellipsoidal shape due to pushing the cluster center closer to the outliers.We propose a semi-definite program based on the weighted maximum likelihood estimation (MLE) to decrease their bad effects.By reducing the outlier’s effect, we find the robust covariance of each cluster as the covariance of output noise for data belonging to that cluster.Finally, we propose cluster-based linear and nonlinear filters.To sum it up, the goals and contributions of the proposed approach can be listed as follows.

    1) To design a data-based filter against heavy-tailed noises with unknown output and inaccurate process noise covariance.

    2) To present the idea of ellipsoidal clustering to compensate for less knowledge about noises’ statistical features.

    3) To suggest the weighted SON clustering to improve regularization and the performance of the conventional SON.

    4) To propose the weighted MLE to decrease the effect of outliers in each cluster and keep the clusters’ ellipsoidal shape.

    5) To present two data-based filtering approaches, including cluster-based linear and cluster-based non-linear filters.

    The remainder of the paper is organized as follows.Section II formally reviews some prerequisite and briefly refers to the main problem.The clustering steps, a new SDP to reduce the effects of outliers, and the proposed cluster-based filters are discussed in Section III.The notable specifications and features of the proposed filtering approaches are presented in Section IV.The simulation results are given in Section V before concluding the paper in Section VI.

    Notations: The paper uses the following standard notations.Rmand Rm×rsignify them-dimensional Euclidean space and the set of allm×rreal matrices, respectively.N (·) designates the multivariate Gaussian probability density function.0 andI represent zero and identity matric∏es with appropriate dimensions, respectively.Furthermore, shows the product operation.The symbol “*” in matrices stands for the symmetric terms.Also, l og(·) indicates the natural logarithm operation.

    II.PRELIMINARIES AND PROBLEM FORMULATION

    The concepts listed below will be used to achieve our key goals.

    A. Sum-of-Norms Clustering

    We are interested in dividing a set of observations,Rd, into different clusters such that the close points to each other are assigned to the same cluster based on the Euclidian meter.We do not know the number of clusters, and it is unnecessary to be large.Assume that each cluster has a centroid in μjand is a subset of Rd.The SON clustering problem is presented as follows [37]:

    in whichp≥1, and λ>0 can be regarded as a parameter that controls the trade-off between the first term in (1) and the number of clusters.The term ofpresents the sum-of-squares error, where μjis the centroid of the cluster containingxj.If the corresponding μ’s are the same, two differentx’s belong to the same cluster.This is the result of the second term in (1), which is a regularization term.In addition,we choosep=2 in the proposed approach, but other choices are possible.After finding the center of the clusters using the optimization problem (1), the data are fitted to each cluster based on the spatial threshold.

    B. Multivariate Gaussian Distribution

    The Gaussian distribution, also known as the normal distribution, resembles a symmetrical bell shape.Letxbe a random vector on Rp.It has the following probability density function:

    where Ξ ∈Rp×pis the positive definite covariance matrix, andμis the mean.

    Remark 1: (x-μ)TΞ-1(x-μ), is a square of Mahalanobis distance.It corresponds to the actual probability of the occurrence of the observation.

    C. Maximum Likelihood Estimation

    Maximum likelihood estimation is a popular way of obtaining practical estimators.Cramer Rao Lower Bound (CRLB),Gaussian PDF, and unbiasedness are among MLE’s asymptotic properties.Consider a set of i.i.d of data pointsY∈Rm×n,containingnobservations, which have a Gaussian distribution with mean μ and covariance Ξ.The likelihood function, under the normality assumption, can be written as

    When dealing with large data sets, we frequently seek statistical and mathematical models to simplify their presentations.One of the first questions we ask is whether the data can be fitted with a normal distribution.This entails estimating the normal distribution’s mean and covariance.They are usually computed using the conventional MLE method by the following problem:

    In this problem, the best estimates of the mean and variance are obtained by taking the partial derivative with respect toμand Ξ of the log-likelihood function and setting it to zero.As a result, we get

    According to the normality condition, the log-likelihood function consists of the sum of the squared Mahalanobis distances.Consequently, if outlier data exist, the mean and covariance are pushed toward the outliers.

    Remark 2: Maximizing the logarithm of the likelihood cost function is equivalent to minimizing the Mahalanobis distance.Outliers make the Mahalanobis distance large.

    D. Kalman Filter

    The Kalman filter is the most famous state estimator for linear Gaussian systems, but its performance is degraded in the presence of non-Gaussian noises.It can be derived from Bayesian recursive relations.In this regard, prediction and filtering steps are achieved as follows:

    Prediction:

    Filter:

    where the gain matrix is given by

    E. System Model

    We assume that output measured data are produced from the following linear state-space dynamic model:

    wherexk∈Rnxandyk∈Rnyare, respectively, the state and the measurement signals.Ak∈Rnx×nxandCk∈Rny×nxare dynamic and output matrices, respectively.Process noise,vk, is a non-Gaussian noise vector with zero mean and inaccurate nominal covariance,Qk.Also, measurement noise sequence,wk, is a non-Gaussian noise vector with zero mean and unknown nominal covariance matrix,Rk.It is remarkable that the process noise stems from internal factors while output noise comes from external sources; therefore, the process e xperiences less intense noise than the system’s output.Moreover, all measurements up to and including timekare presented by yk.We assume that the initial conditionx0and the system’s noises are mutually independent, satisfying the following relation:

    wherexˉ0is the expectation ofx0.

    This paper aims to design data-based filters for the output data set,y, produced by (9).It is assumed that the output noise nominal covariance is unknown.In the presence of outliers,conventional MLE estimators are affected by both good and bad observations.To compensate for this defect, after applying the proposed clustering method, we try to detect outliers in each cluster to decrease their effects on changing the shape of clusters and the covariances.Then, we present a linear filter based on the moving horizon estimation technique by restating the conventional MAP estimation problem for a measurement data set.Moreover, considering that data in each ellipsoidal cluster originated from a Gaussian distribution, we assume that there are εlellipsoidal clusters with Gaussian specifications for the output noise.By doing so, we propose a novel data-based nonlinear filter.

    III.MAIN RESULTS

    We intend to design a suitable filter for non-Gaussian linear dynamic systems with unknown output and inaccurate process noise covariance.Regarding our idea of output measurements clustering, first, we propose the weighted SON clustering method that improves the conventional SON’s performance in a large data set.

    A. Weighted Sum-of-Norms Clustering

    We propose the following weighted SON clustering to mitigate the influence of distances between cluster centers on the clustering performance and enhance the computational efficiency of the conventional SON clustering method (1):

    where ζi j≥0 is a weighting parameter.

    Using a constant weight for each point in the objective function of clustering can lead to suboptimal results because distant points with low similarity values would have the same impact as nearby points with high similarity values.To address this issue, we introduce the concept that the affinity weight, ζi j, quantifies the similarity betweenxjandxi, and assigns higher weights to nearby points and lower weights to distant ones.This approach can also reduce the impact of noise and outliers and consequently, result in more accurate and robust clustering.This requires that the weight ζi jbe obtained from a similarity function based on a statistical similarity measure criterion.The similarity function is typically a strictly monotonically decreasing continuous function in the[0 ∞)range with a positive second-order derivative.Remark 3: The similarity measure criterion provides freedom in selecting the similarity function.One of the most widely recognized functions is the Gaussian kernel.?

    We know that the data in the same cluster have similar statistical features.Hence, to obtain output noise covariance,first, we decide to cover a set ofnpoints yk, by some ellipses(?1,?2,...,ε?)using the weighted SON clustering.Now, at the first step, Algorithm 1 is presented for clustering purpose.

    Algorithm 1 Clustering Steps Required:{X j}Nj=1 ∈Rd f Data point and similarity function.Steps:1) Set parameter in (11).2) Run optimization (11).3) Find mean and covariance of each cluster based on (5).4) Detect unsuitable clusters and re-cluster them.5) Find new mean and covariance of each cluster by (5).λ>0

    Remark 4: To have more efficiency in Algorithm 1, a reclustering process is welcome for unsuitable clusters.In this regard, clusters with a small number of data are removed and combined with the other clusters, and clusters with large estimated covariances are divided into some smaller clusters.By doing so, a balance is made between the number of clusters and their number of data.

    B. Detecting and Reducing the Effect of Outliers

    The presented clustering algorithm results in some ellipsoidal clusters for output data of the system.According to the fact that Gaussian data have ellipsoidal scattering, we will encourage the use of the filters developed for Gaussian systems for each cluster.Note that the obtained clusters may suffer from some outlier data.Outliers can change the covariance of that cluster and create remarkable bias in the conventional MLE.For this reason, we are supposed to make a difference between outliers and good data to improve the performance of the MLE.In this regard, we need to minimize the sum of the smallest squared Mahalanobis distances while outliers with larger squared values are excluded.Under such a circumstance, we introduce a variable ωiin the range of [0 1],so that if the squared Mahalanobis distance is larger, the corresponding ωiis smaller.This approach is useful in state estimation to detect and decrease the effect of outliers.According to this concept, we consider the following weighted MLE:

    We can rewrite the cost function in (12) by replacing the likelihood function (3) and using the features of the logarithm function as follows:

    One crucial point is that the above log-likelihood function is not concave or convex.

    If we fix the parameter ω in the optimization problem (12),we can easily obtain the mean and covariance variables by the following lemma.

    Lemma 1: Consider the weighted MLE problem (12), the mean and covariance, μ and Ξ, are obtained by

    Proof: By taking partial derivations of (12) with respect toμand Ξ and setting them tos zero, it is straightforward to obtain(14).

    Due to a lack of knowledge of good and bad observations and the problem’s non-convexity, we cannot determine the appropriate weights and utilize the weighted MLE.Iterative algorithms, like iteratively reweighted least squares [39], can solve this problem by updating the weights in inverse proportion to Mahalanobis distances.However, slow convergence and trap into local optimal solutions are some of their weaknesses.In the following, with the help of a suitable objective function and introducing probability criterion on all observations, we will present an optimization problem to set the outliers with smaller weights for solving the above-mentioned issues.To this end, considering the following relations:

    and using the fact that outliers have low probability and large Mahalanobis distances, we encourage using probability of occurrence as weights to decrease the effect of outliers in our proposed method.We introduce a positive weight vectorwithTherefore, inspiring Remark 2, the mean and covariances of the weighted MLE (14), and the squared Mahalanobis distances (15), we present the following optimization problem to obtain optimal values ofχ andd:

    In (16), the loss function must be monotonically nondecreasing function of χianddi.This function should be chosen such that χiis inversely proportional to the corresponding Mahalanobis distance,di.Doing so allows all outlier observations with largedito receive small weight χi.On the other hand, we ought to choose a suitable cost function such that the sum of all Mahalanobis distances is small.To satisfy all of these conditions, in the optimization problem (16), we propose the cost function as

    In what follows, we present a theorem to obtain the best values of weights, χi, which have an essential role in outliers detection and decreasing their effects.

    Theorem 1: The optimization problem (16) with the cost function (17) can be reformulated as the following SDP problem:

    Proof:For the cost function (17), assume an upper bound,δi, such that.Thus, minimization of this cost can be formulated as follows:

    The first constraint can be reformulated as

    Now, by substituting the weighted mean and covariance from the constraints of (16), the above equation can be simplified as

    Finally, applying the Schur complement equivalence, the SDP problem in (18) is obtained.■

    Remark 5: Since outliers are destructive in different domains and applications and cause some limitations, the proposed SDP problem in (18) can detect outliers in each Gaussian data set without needing extra designs.

    Until now, we have presented a method to cluster the data set and an optimization problem to compute the weights in the MLE for reducing the effect of outliers.Note that we can obtain the covariance of each cluster using two methods, the conventional MLE and the weighted MLE.The second method is robust against outliers thanks to obtaining the covariance according to reducing their impact.Now, we are ready to introduce our proposed filtering approaches.

    C. Cluster-Based Linear Filter

    Since the proposed approach incorporates the new weighted SON clustering and the new weighted MLE, we deal with a data set belonging to Gaussian distributions with different statistical features in each cluster.Given this circumstance, in this section, instead of using the existing recursive filters, such as the Kalman filter, to estimate or predict the states according to the available information at each instant, we intend to estimate the entire states at once using the whole data set.

    Theorem 2: The state estimations for the output data set,y1:N, belonging to the system (9), can be obtained as

    where

    Proof: Based on the Bayesian method and using the fact that process and measurement noises are independent, we have

    Thus, the posterior is given by

    By rearranging the terms in (23) and using the defined matrices in (21), the optimization problem (22) can be written as

    Consequently, taking partial derivation with respect to X,the estimation, X?, in (21) is obtained.

    Since (21) involves the inverse of the matrix (O?TO?), this matrix grows in size as the number of data or dimension increases.To solve this unfavorable issue, we propose the following optimization problem to obtain the moving horizon estimation (MHE) with a window length ofNw:

    Consequently, the steps of the proposed cluster-based robust moving horizon estimation are presented in Algorithm 2.

    Remark 6: In Algorithm 2, Theorem 1 determines the best weights and detects outliers; subsequently, the elimination of outliers when computing the robust covariance matrices of each cluster results in a filter that is robust against outliers.Therefore, we call it a cluster-based robust MHE.We have a simple (non-robust) cluster-based MHE if we omit step 2 and determine the covariance matrix by the conventional MLE.

    Algorithm 2 Cluster-Based Robust Moving Horizon Estimation Required:{X j}Nj=1 ∈Rd Nw Data Point and window length.Steps:1) Run Algorithm 1.2) Detect outliers of each cluster using Theorem 1.3) Calculate robust covariance matrix of each cluster.yk 4) Determine cluster of each.5) Obtain the state estimation from (25).

    Remark 7: Equation (21) is the closed-form solution of the optimization problem (25) in a complete intervalNw=N.

    D. Cluster-Based Non-Linear Filter

    As we discussed, each ellipsoidal set’s data originates from a Gaussian distribution in the SON clustering.Since we cluster the output data with εlindependent ellipsoidal set and considering this assumption that the intensity of the process noise is less than output noise, we can assume that there existεlGaussian resources for the output noise distribution.Considering all clusters, the total occurrence probability of the output noise is equal to the sum of the occurrence probability of each Gaussian resource.This is another idea to design a new filter.We show that this filter would be non-linear.To this end, we introduce the following probability density function for each noise resource:

    Using above explanations and substituting (27) into the Bayesian rule result in

    The denominator in (28) involves the integration of the sum of likelihoods and priors for each noise distribution.This ensures that the total probability of all states ofx0giveny0equals one.In this vein, the density function (28) can be written as (29).

    For simplicity, we assume

    Using the Bayesian rule to inspire the strides of obtaining the Kalman filter, and after setting out the results, the proposed filter structure consists of the prediction and filtering steps as follows:

    Prediction:

    Filter:

    Moreover, the filter gain can be obtained as

    Now, after computing the estimation of each cluster, to obtain the unique estimation for the state,k, we have

    Moreover, using (35) and (36), covariances of estimation error,Pkand, are given by

    and

    According to the main filter’s relation (35), this filtering approach consists of a bank of εlKalman filters.Although the prediction and filtering relations in (31), (32) and (33) are linear similar to the Kalman filter, because of using the nonlinear parameters ζkiand αkiin (33) for obtaining state estimationx?kin (35), the proposed filter in this part is nonlinear.

    As the last step, Algorithm 3 is presented for the proposed cluster-based robust nonlinear filter.We have a simple cluster-based nonlinear filter by removing the second step in Algorithm 3 and using the conventional MLE.

    Algorithm 3 Cluster-Based Robust Non-Linear Filter Required:{X j}Nj=1 ∈Rd εl Data Point and.Steps:1) Run Algorithm 1.2) Detect outliers of each cluster using Theorem 1.3) Calculate robust covariance matrix of each cluster.k ←1 for to N do? ←1 εl for to do 5) Compute the parameters in the prediction and filtering steps from (32) and (33).end for 6) Obtain the state estimation by (35).end for

    IV.DISCUSSION ON THE RESULTS

    In the development of filters, a bias in the estimate can arise due to heavy-tailed noises.To compensate for this bias, the effects of outliers should be reduced.Based on the above concepts, our proposed cluster-based robust filters have a better performance thanthe proposedcluster-basedfilters,dueto decreasing thefiltergain (34) andthe matrix F?in (21)which consequently improves the filter performance.However, various factors can affect the performance of moving horizon estimators such as window length, initial state, covariance values in the beginning of window, time-variant or time-invariant characteristics of the systems’ dynamic, etc.In light of the points mentioned above, we cannot explicitly claim which of the proposed filters is the best.

    The impact of removing outliers (from a cluster or data set)on statistical covariance can vary depending on the distribution of the remaining data.The covariance decreases significantly when the outliers are large and far from the other data.These outliers can lead to significant variability and distortion in the covariance estimate.On the other hand, outliers consistent with the general trend of the data have different impacts on covariance if removed.It should be noted that the number of data in each cluster can also affect this phenomenon.

    The number of measurement data in this process depends on the system complexity, dispersion, and constraints.Although clustering is possible for small number of data, the accuracy of covariance estimation and consequently, the precision of filtering decreases.Besides, highly dispersed data set are unreliable for clustering, while excessive data leads to a single ellipsoidal cluster due to the central limit theorem.Therefore,we suggest to check the data scatter and collect the measurement data such that the dispersion and empty positions are minimized in the whole data set.

    In the proposed methods, different parameters appear in the optimization problem, filtering, and clustering.The parameters in the first and second groups are computed precisely.However, in the weighted SON clustering problem, the weights ζi jare obtained using the similarity functions.In Remark 3, the Gaussian kernel has been introduced as one of the most famous functions.Also,pis another parameter and determines which norm is used in the clustering.We know that different norms emphasize different aspects of the data,leading to variations in the clustering outcome.The proposed approach optimally uses the Euclidean norm (p=2) to capture the overall distance between data points without any bias or focusing on specific characteristic of data.

    The proposed filtering algorithms are offline; therefore, they do not depend on the time properties of dynamic systems, i.e.,our proposed data-based filters can be applied to both timevariant and time-invariant systems.

    V.SIMULATION STUDY

    In this section, we want to study the performance and effectiveness of the proposed filters, including cluster-based MHE(C-B MHE), cluster-based robust MHE (C-B RMHE), clusterbased non-linear filter (C-B NF), and cluster-based robust non-linear filter (C-B RNF).For this purpose, we exert these filters on the time-varying and practical system subject to heavy-tailed noises.Since no completely related filters exist in the literature, we compare our proposed filters with those designed for non-Gaussian systems with known parameters.

    A. Example 1: Three-Tank System

    We consider a three-tank system with the schematic shown in Fig.1.The dynamics of this system can be described as follows:

    whereQ1andQ2are the flow rates of pumps 1 and 2,qLiis the leakage flow rate of the tanki(i=1,2,3),hiis the level of the tanki,Acdenotes the cross-sectional area of the connecting pipe, andqmn(m≠n) is the flow rate from tankmto tankn.Parameters of the system and their numerical values are tabulated in Table I.

    Fig.1.Three-tank system.

    TABLE ITHREE-TANK SYSTEM PARAMETERS

    Using the system parameters in Table I with operating points (Q1=5.5×10-5,Q2=3.4×10-5m3/s), and (h1=0.4,h2=0.23,h3=0.31m), and after linearization and discretization of the model (39), we reach the state space model of the system with the following matrices:

    Process and measurement noises are non-Gaussian with the following distributions:

    whereQ=R=0.035.

    The performance of the proposed filters is compared to the maximum correntropy Kalman filter (MCKF) [13] as a famous filter in the presence of heavy-tailed noises, the conventionalMHEwithwindowlengthNw=10,and the conventionalKalmanfilter.Sincethe dimensionofthe matrixO?in(21) is extensive, to have more analysis, we introduce clusterbased and cluster-based robust Kalman filters using the conventional Kalman filter with estimated covariance matrix using MLE and weighted MLE, respectively.Fig.2 shows different filters’ state estimation and MSE.Our proposed databased filters perform better than the Kalman filter, the MCKF,and the conventional MHE designed with known and true parameters.The main reason for this deficiency is that the Kalman filter is a minimum mean square filter and is sensitive to non-Gaussian noises; therefore, its performance degrades against these noises.Also, as a prior filter against non-Gaussian noises, the maximum correntropy Kalman filter cannot guarantee a compelling performance against these noises.Moreover, although the conventional MHE uses data in a window to obtain estimations and has a weak performance against non-Gaussian noises, it outperforms the conventional Kalman filter and MCKF.This fact brings about the proposed cluster-based MHE and the proposed cluster-based robust MHE has a better response than the cluster-based Kalman filter and cluster-based robust Kalman filter.In contrast, data clustering, outlier detection, and data-based filtering equip the proposed approaches to perform well against non-Gaussian noises.

    Fig.2.State estimations and MSEs in Example 1 (C-B, R, and N are abbreviation of cluster-based, robust, and non-linear).

    Fig.3.Error ellipse (Iso-Contour) of the each cluster with outliers in Example 2.

    B. Example 2: Time-Variant System

    To show the usefulness of our proposed filters for time-variant systems, we assume that output data set of the following non-Gaussian time-variant system (taken from [40] with some modification) are available:

    Process and measurement noises are non-Gaussian with the

    following distributions:

    whereQ=R=0.15.

    We have six ellipsoidal clusters after running Algorithm 1.Fig.3 shows the iso-contour of these six ellipsoidal clusters along with the outliers of each cluster.Clusters’ shapes without outliers emphasize the proposed optimization’s proper performance.It is worth noting that if the axes of ellipsoidal are parallel to coordinate axes, the obtained covariances are diagonal.Moreover, the results of mean square error of states for the proposed data-based filters are compared in Table II.Likewise, our proposed filters have better performance than the others.

    TABLE II MEAN SQUARE ERROR OF DIFFERENT FILTERS FOR NOISES IN (43)

    Based on the results in Table II, the robust moving horizon estimation with a whole window performs better than other methods.This phenomenon is because MHE with a large window length in slow varying dynamic systems can capture more relevant information for estimation, and consequently,has a more reliable performance.

    VI.CONCLUSION

    In this paper, we aimed to develop data-based filters for linear dynamic systems against non-Gaussian heavy-tailed noises under the unknown output noise covariance condition.Inspiring that Gaussian data sets have ellipsoidal scattering, we clustered the output data set of a non-Gaussian dynamic system using the proposed weighted SON clustering method.The proposed approach can improve the performance of the SON clustering method by focusing on closely spaced points and reducing the computational cost.Outliers in each cluster can change the clusters’ ellipsoidal shape and affect the filter’s performance.To address this issue, we proposed an SDP problem based on the weighted maximum likelihood to detect outliers in each cluster and obtain a robust covariance matrix.We then developed four different filters using two approaches.In the first method, we presented the cluster-based MHE.Provided that output noise covariance is computed by the conventional MLE and the proposed SDP problem, we have a simple cluster-based filter and a robust cluster-based filter, respectively.In the second approach, given εlclusters, we assumed that there are εlGaussian resources for the output noise distribution with their specific statistician features.This idea led us to extract a non-linear filter structure, presenting a clusterbased non-linear filter and a cluster-based robust non-linear filter, depending on how the covariance matrix is computed.Finally, we verified the performance of our proposed filters through simulation results on a practical system and a timevariant system.

    国产欧美日韩一区二区三| 两人在一起打扑克的视频| 亚洲天堂国产精品一区在线| 欧美性长视频在线观看| 熟女电影av网| 中文亚洲av片在线观看爽| 婷婷精品国产亚洲av| 黑人巨大精品欧美一区二区mp4| 麻豆成人av在线观看| 嫩草影视91久久| 两个人免费观看高清视频| 国产精品久久久久久亚洲av鲁大| 精品一区二区三区视频在线观看免费| 国产精品综合久久久久久久免费| 一级a爱片免费观看的视频| 亚洲av美国av| 听说在线观看完整版免费高清| 丝袜人妻中文字幕| 欧美色欧美亚洲另类二区| 黑人欧美特级aaaaaa片| 91九色精品人成在线观看| 国产视频一区二区在线看| 欧美日韩黄片免| 亚洲自拍偷在线| 亚洲熟妇熟女久久| 久久亚洲真实| 别揉我奶头~嗯~啊~动态视频| 给我免费播放毛片高清在线观看| 欧美日韩一级在线毛片| 在线播放国产精品三级| 日日夜夜操网爽| 婷婷亚洲欧美| 久久久久精品国产欧美久久久| 校园春色视频在线观看| 啦啦啦韩国在线观看视频| 国产av麻豆久久久久久久| 真人做人爱边吃奶动态| 高潮久久久久久久久久久不卡| 亚洲av成人精品一区久久| 色av中文字幕| 国产亚洲欧美98| 又爽又黄无遮挡网站| 制服人妻中文乱码| 制服人妻中文乱码| 好看av亚洲va欧美ⅴa在| 成人特级黄色片久久久久久久| 国内少妇人妻偷人精品xxx网站 | 欧美日韩瑟瑟在线播放| 亚洲午夜精品一区,二区,三区| 在线十欧美十亚洲十日本专区| 久久精品亚洲精品国产色婷小说| 精品第一国产精品| 国产片内射在线| 久久国产精品人妻蜜桃| 亚洲av中文字字幕乱码综合| 日本熟妇午夜| 亚洲熟女毛片儿| 此物有八面人人有两片| 丁香欧美五月| 1024手机看黄色片| 国产成人av教育| 国产久久久一区二区三区| 国产亚洲精品综合一区在线观看 | 中文字幕熟女人妻在线| 国产69精品久久久久777片 | avwww免费| 亚洲aⅴ乱码一区二区在线播放 | 91成年电影在线观看| 女警被强在线播放| 久久久久免费精品人妻一区二区| 一区福利在线观看| 香蕉国产在线看| 国产探花在线观看一区二区| 人人妻人人澡欧美一区二区| 亚洲国产精品成人综合色| 人妻久久中文字幕网| 亚洲午夜精品一区,二区,三区| 国内揄拍国产精品人妻在线| 欧美成狂野欧美在线观看| 久久国产精品人妻蜜桃| 男人舔奶头视频| 俄罗斯特黄特色一大片| 特大巨黑吊av在线直播| 亚洲精品一卡2卡三卡4卡5卡| 色综合亚洲欧美另类图片| 最近最新免费中文字幕在线| 脱女人内裤的视频| 精品人妻1区二区| 在线观看美女被高潮喷水网站 | 精品电影一区二区在线| 性色av乱码一区二区三区2| 国产v大片淫在线免费观看| 久久久久国产精品人妻aⅴ院| 男人舔女人下体高潮全视频| 欧美乱妇无乱码| 国产探花在线观看一区二区| 亚洲av成人不卡在线观看播放网| 日韩av在线大香蕉| 淫妇啪啪啪对白视频| 国产高清视频在线播放一区| 欧美日韩乱码在线| 日韩欧美在线二视频| 国产视频一区二区在线看| 日本精品一区二区三区蜜桃| 欧美日韩福利视频一区二区| 欧美绝顶高潮抽搐喷水| 久久99热这里只有精品18| 高清毛片免费观看视频网站| 99精品在免费线老司机午夜| 国产成人系列免费观看| 男插女下体视频免费在线播放| 免费在线观看日本一区| 18禁黄网站禁片免费观看直播| 免费人成视频x8x8入口观看| 高潮久久久久久久久久久不卡| 好男人电影高清在线观看| 久久久精品欧美日韩精品| 久久久国产成人精品二区| 曰老女人黄片| 亚洲精品美女久久av网站| 国产成人av教育| 久久精品亚洲精品国产色婷小说| 亚洲美女视频黄频| 在线观看免费午夜福利视频| 亚洲一区中文字幕在线| 欧美午夜高清在线| 久久久久性生活片| 国产精品免费视频内射| 亚洲18禁久久av| bbb黄色大片| 特级一级黄色大片| 亚洲精品色激情综合| 国产精品美女特级片免费视频播放器 | 日日爽夜夜爽网站| 国产一区二区在线观看日韩 | 国产精品免费一区二区三区在线| 亚洲人成伊人成综合网2020| 亚洲专区国产一区二区| 国产精品香港三级国产av潘金莲| 久久精品综合一区二区三区| 国产黄a三级三级三级人| 久久久国产精品麻豆| 欧美日韩亚洲综合一区二区三区_| 99在线视频只有这里精品首页| 国产真实乱freesex| 国产激情久久老熟女| www.熟女人妻精品国产| 欧美久久黑人一区二区| 国产亚洲精品一区二区www| 黄色 视频免费看| 中文字幕人成人乱码亚洲影| 一个人免费在线观看的高清视频| 欧美zozozo另类| 岛国在线免费视频观看| 男人的好看免费观看在线视频 | 伊人久久大香线蕉亚洲五| 欧美黄色片欧美黄色片| 男女之事视频高清在线观看| 女警被强在线播放| 国产探花在线观看一区二区| 亚洲熟女毛片儿| 国产精品久久久久久精品电影| 51午夜福利影视在线观看| 国产日本99.免费观看| 精品国内亚洲2022精品成人| 麻豆成人av在线观看| 国产一区在线观看成人免费| 国产精品 国内视频| 看免费av毛片| 一级a爱片免费观看的视频| 日韩高清综合在线| 香蕉久久夜色| 十八禁网站免费在线| 最新在线观看一区二区三区| 色哟哟哟哟哟哟| 日韩欧美免费精品| 看免费av毛片| 色综合亚洲欧美另类图片| 日本一区二区免费在线视频| 黑人巨大精品欧美一区二区mp4| 国产三级黄色录像| 欧美极品一区二区三区四区| ponron亚洲| 亚洲av第一区精品v没综合| 制服人妻中文乱码| 五月伊人婷婷丁香| 午夜激情福利司机影院| 手机成人av网站| 久久精品影院6| 久久久国产成人精品二区| 国产99白浆流出| АⅤ资源中文在线天堂| 亚洲国产精品久久男人天堂| 村上凉子中文字幕在线| 中国美女看黄片| 男男h啪啪无遮挡| av在线播放免费不卡| tocl精华| 亚洲专区国产一区二区| 制服丝袜大香蕉在线| 久久国产精品人妻蜜桃| 在线观看一区二区三区| 欧美色欧美亚洲另类二区| 麻豆成人av在线观看| 国产精品爽爽va在线观看网站| www.精华液| 男人舔女人的私密视频| 禁无遮挡网站| 国产伦一二天堂av在线观看| 国产一区二区激情短视频| 亚洲精品国产一区二区精华液| 国产高清视频在线播放一区| 十八禁人妻一区二区| 欧美日韩国产亚洲二区| av有码第一页| 亚洲av成人精品一区久久| 精品久久久久久,| 午夜福利视频1000在线观看| 欧美日韩精品网址| 欧美日韩亚洲综合一区二区三区_| 最新美女视频免费是黄的| av中文乱码字幕在线| 日韩大码丰满熟妇| 日韩精品中文字幕看吧| 欧美性猛交╳xxx乱大交人| 亚洲天堂国产精品一区在线| 村上凉子中文字幕在线| 18禁黄网站禁片午夜丰满| 国语自产精品视频在线第100页| 国产精品日韩av在线免费观看| 国产精品久久电影中文字幕| 午夜影院日韩av| 两性夫妻黄色片| 超碰成人久久| a级毛片a级免费在线| 国产片内射在线| 岛国在线免费视频观看| 韩国av一区二区三区四区| 18禁国产床啪视频网站| 美女扒开内裤让男人捅视频| 在线视频色国产色| 老司机午夜十八禁免费视频| 色哟哟哟哟哟哟| 亚洲黑人精品在线| 不卡av一区二区三区| 天堂√8在线中文| 亚洲avbb在线观看| 村上凉子中文字幕在线| 非洲黑人性xxxx精品又粗又长| 久热爱精品视频在线9| 国产精品精品国产色婷婷| 淫妇啪啪啪对白视频| avwww免费| 在线看三级毛片| 不卡一级毛片| 嫁个100分男人电影在线观看| 国产成+人综合+亚洲专区| 欧美黑人欧美精品刺激| 久久久久精品国产欧美久久久| 久久国产乱子伦精品免费另类| 真人一进一出gif抽搐免费| 成年版毛片免费区| 亚洲av第一区精品v没综合| 亚洲av中文字字幕乱码综合| АⅤ资源中文在线天堂| 亚洲av成人精品一区久久| 中文字幕人成人乱码亚洲影| 久久精品国产99精品国产亚洲性色| 黄色 视频免费看| 色精品久久人妻99蜜桃| 曰老女人黄片| 搡老妇女老女人老熟妇| 韩国av一区二区三区四区| 长腿黑丝高跟| 久久久久久久久免费视频了| avwww免费| av片东京热男人的天堂| 亚洲 国产 在线| 国产亚洲av嫩草精品影院| 亚洲av五月六月丁香网| 欧美成人性av电影在线观看| 老司机深夜福利视频在线观看| 九色成人免费人妻av| 天堂影院成人在线观看| 久久婷婷人人爽人人干人人爱| 国产亚洲欧美98| 老熟妇仑乱视频hdxx| 午夜福利在线观看吧| 欧美乱色亚洲激情| 欧美日韩亚洲综合一区二区三区_| 亚洲 欧美一区二区三区| 欧美午夜高清在线| 久久久久精品国产欧美久久久| 美女免费视频网站| 性欧美人与动物交配| 亚洲全国av大片| 毛片女人毛片| 日本成人三级电影网站| 夜夜爽天天搞| 白带黄色成豆腐渣| 人成视频在线观看免费观看| 亚洲va日本ⅴa欧美va伊人久久| 91成年电影在线观看| 啦啦啦观看免费观看视频高清| 日韩欧美一区二区三区在线观看| 久久伊人香网站| 国产精品久久久av美女十八| 少妇的丰满在线观看| 欧美又色又爽又黄视频| 黄色片一级片一级黄色片| 亚洲中文字幕一区二区三区有码在线看 | 一边摸一边抽搐一进一小说| 婷婷精品国产亚洲av| 18禁美女被吸乳视频| 国产亚洲精品综合一区在线观看 | 亚洲va日本ⅴa欧美va伊人久久| or卡值多少钱| 99久久99久久久精品蜜桃| 亚洲av成人av| 日韩欧美国产在线观看| 欧美午夜高清在线| 制服人妻中文乱码| 日韩欧美三级三区| 一本精品99久久精品77| 国产私拍福利视频在线观看| 制服人妻中文乱码| 久久久久久久久久黄片| 国产精品一区二区免费欧美| 精品乱码久久久久久99久播| 最新美女视频免费是黄的| 黄色女人牲交| 国产激情久久老熟女| www日本黄色视频网| 久久久久久亚洲精品国产蜜桃av| 全区人妻精品视频| 国产黄色小视频在线观看| 国产成人精品久久二区二区免费| 宅男免费午夜| 久久草成人影院| 嫩草影视91久久| 免费在线观看黄色视频的| 亚洲欧美精品综合一区二区三区| 变态另类丝袜制服| 成年女人毛片免费观看观看9| 不卡av一区二区三区| 久久精品国产清高在天天线| 欧美乱妇无乱码| 亚洲国产日韩欧美精品在线观看 | 成人亚洲精品av一区二区| 999久久久精品免费观看国产| 天堂动漫精品| 91麻豆av在线| 老司机在亚洲福利影院| 国产精品九九99| 精品欧美一区二区三区在线| 亚洲专区中文字幕在线| 午夜福利在线观看吧| 成人一区二区视频在线观看| 国产av又大| netflix在线观看网站| 免费电影在线观看免费观看| 国产成人av教育| 久久国产精品影院| 国产一区二区三区在线臀色熟女| 在线播放国产精品三级| 国产精品久久久久久亚洲av鲁大| 久久亚洲真实| 国产精品野战在线观看| 国产在线观看jvid| 日韩高清综合在线| 国产探花在线观看一区二区| 久久婷婷人人爽人人干人人爱| 中文字幕精品亚洲无线码一区| 搡老熟女国产l中国老女人| 日本 av在线| 欧美日韩福利视频一区二区| 日韩国内少妇激情av| 91麻豆精品激情在线观看国产| 久久这里只有精品19| 女同久久另类99精品国产91| 色噜噜av男人的天堂激情| 国产真实乱freesex| 欧美激情久久久久久爽电影| 国产精品一区二区三区四区久久| 国产69精品久久久久777片 | 久久久久久免费高清国产稀缺| 午夜亚洲福利在线播放| 夜夜躁狠狠躁天天躁| 一a级毛片在线观看| 精品熟女少妇八av免费久了| 老司机在亚洲福利影院| 黄色视频,在线免费观看| 黑人欧美特级aaaaaa片| 国产真人三级小视频在线观看| av福利片在线| 黑人巨大精品欧美一区二区mp4| 香蕉丝袜av| 啦啦啦韩国在线观看视频| 日韩三级视频一区二区三区| 国产亚洲精品综合一区在线观看 | 又紧又爽又黄一区二区| 俺也久久电影网| 国产一级毛片七仙女欲春2| 欧美人与性动交α欧美精品济南到| 少妇粗大呻吟视频| 久久久精品国产亚洲av高清涩受| 夜夜看夜夜爽夜夜摸| 狠狠狠狠99中文字幕| 久久 成人 亚洲| 在线观看一区二区三区| 国产成人aa在线观看| 国产精品免费视频内射| 91av网站免费观看| 亚洲中文日韩欧美视频| 天堂√8在线中文| 精品久久久久久久久久免费视频| av在线播放免费不卡| 亚洲成a人片在线一区二区| 男人舔女人的私密视频| 桃红色精品国产亚洲av| 精品福利观看| 国产伦人伦偷精品视频| 天堂av国产一区二区熟女人妻 | 91国产中文字幕| 曰老女人黄片| 国产精品日韩av在线免费观看| 色av中文字幕| 国产野战对白在线观看| 欧美色视频一区免费| 777久久人妻少妇嫩草av网站| 欧美日韩亚洲综合一区二区三区_| 99精品在免费线老司机午夜| 亚洲熟妇中文字幕五十中出| 精品久久久久久久久久久久久| 国内揄拍国产精品人妻在线| 色综合欧美亚洲国产小说| 欧美av亚洲av综合av国产av| 精品无人区乱码1区二区| 午夜精品久久久久久毛片777| 欧美国产日韩亚洲一区| 亚洲乱码一区二区免费版| 好男人在线观看高清免费视频| 亚洲色图av天堂| 亚洲九九香蕉| 国内精品久久久久精免费| 99久久精品热视频| av片东京热男人的天堂| 国产精品综合久久久久久久免费| 99riav亚洲国产免费| 亚洲午夜精品一区,二区,三区| 久久久精品欧美日韩精品| 黄色a级毛片大全视频| 国产三级在线视频| 热99re8久久精品国产| 欧美日韩黄片免| 精品久久久久久久末码| 超碰成人久久| 亚洲性夜色夜夜综合| 成人欧美大片| 国产一区二区三区在线臀色熟女| 日韩精品中文字幕看吧| 成人手机av| 天天躁夜夜躁狠狠躁躁| 精品日产1卡2卡| 免费电影在线观看免费观看| 看黄色毛片网站| 最近最新中文字幕大全免费视频| 久久精品91蜜桃| 亚洲人与动物交配视频| 中文字幕人成人乱码亚洲影| 亚洲av成人精品一区久久| 国产精品爽爽va在线观看网站| 亚洲国产精品合色在线| 香蕉国产在线看| 婷婷精品国产亚洲av在线| 亚洲avbb在线观看| www.999成人在线观看| 日韩欧美在线乱码| 老鸭窝网址在线观看| 亚洲精品在线美女| 亚洲免费av在线视频| 久久婷婷人人爽人人干人人爱| 啦啦啦观看免费观看视频高清| 国产精品精品国产色婷婷| 欧美日韩精品网址| 亚洲人成电影免费在线| 欧美在线黄色| 精品国产乱子伦一区二区三区| 欧美性猛交╳xxx乱大交人| 老司机在亚洲福利影院| 成人av在线播放网站| 欧美av亚洲av综合av国产av| 久久热在线av| 亚洲性夜色夜夜综合| 国产精品1区2区在线观看.| 国产欧美日韩一区二区精品| 国产在线观看jvid| 黄色a级毛片大全视频| 俄罗斯特黄特色一大片| 亚洲人成网站在线播放欧美日韩| 精品无人区乱码1区二区| 亚洲片人在线观看| 日日爽夜夜爽网站| 69av精品久久久久久| 999精品在线视频| 亚洲一区高清亚洲精品| 美女黄网站色视频| 香蕉久久夜色| 男插女下体视频免费在线播放| 亚洲精品在线观看二区| 国产高清videossex| 九九热线精品视视频播放| 成年版毛片免费区| 亚洲成人久久爱视频| 国产又色又爽无遮挡免费看| 欧美最黄视频在线播放免费| 岛国在线免费视频观看| 12—13女人毛片做爰片一| 久久草成人影院| 欧美黑人欧美精品刺激| 欧美绝顶高潮抽搐喷水| 亚洲色图 男人天堂 中文字幕| 久久国产乱子伦精品免费另类| 成人18禁高潮啪啪吃奶动态图| 淫秽高清视频在线观看| 小说图片视频综合网站| 99久久精品国产亚洲精品| 黄色女人牲交| 9191精品国产免费久久| 一边摸一边抽搐一进一小说| av天堂在线播放| 亚洲欧洲精品一区二区精品久久久| 90打野战视频偷拍视频| 在线a可以看的网站| 日本一本二区三区精品| 欧美日韩亚洲综合一区二区三区_| 国产亚洲精品久久久久久毛片| 欧美色欧美亚洲另类二区| 久久天堂一区二区三区四区| 女人爽到高潮嗷嗷叫在线视频| 18禁裸乳无遮挡免费网站照片| 久久久久亚洲av毛片大全| 国内久久婷婷六月综合欲色啪| 久久久国产精品麻豆| 三级男女做爰猛烈吃奶摸视频| 国产麻豆成人av免费视频| 久久精品国产亚洲av香蕉五月| 免费一级毛片在线播放高清视频| 老鸭窝网址在线观看| 国产三级中文精品| 国产人伦9x9x在线观看| 久久人妻福利社区极品人妻图片| 国产99白浆流出| 婷婷丁香在线五月| 床上黄色一级片| 欧美日韩精品网址| 亚洲人成伊人成综合网2020| 免费电影在线观看免费观看| 亚洲自偷自拍图片 自拍| 久久午夜亚洲精品久久| 五月伊人婷婷丁香| 黄色毛片三级朝国网站| a级毛片在线看网站| 国产三级中文精品| 国产又黄又爽又无遮挡在线| 午夜亚洲福利在线播放| 日韩欧美 国产精品| 欧洲精品卡2卡3卡4卡5卡区| 久久久精品欧美日韩精品| 亚洲 欧美 日韩 在线 免费| 99久久综合精品五月天人人| av有码第一页| 精品少妇一区二区三区视频日本电影| aaaaa片日本免费| 亚洲男人天堂网一区| 欧美精品啪啪一区二区三区| bbb黄色大片| 国产精品一及| 无限看片的www在线观看| 亚洲欧美日韩高清专用| 久久精品aⅴ一区二区三区四区| 国产亚洲精品综合一区在线观看 | 999精品在线视频| www.精华液| 极品教师在线免费播放| 老汉色∧v一级毛片| 欧美中文综合在线视频| or卡值多少钱| 妹子高潮喷水视频| 99热这里只有精品一区 | 久久亚洲真实| 最好的美女福利视频网| 精品国产乱子伦一区二区三区| 99久久无色码亚洲精品果冻| 女人高潮潮喷娇喘18禁视频| 日韩大尺度精品在线看网址| a级毛片在线看网站| 久久久国产欧美日韩av| 啦啦啦观看免费观看视频高清| 欧美黑人巨大hd| 男女午夜视频在线观看| 免费观看精品视频网站| 1024视频免费在线观看| 欧美av亚洲av综合av国产av| 国产精品野战在线观看| 桃红色精品国产亚洲av| av国产免费在线观看| 看黄色毛片网站| 欧美成狂野欧美在线观看| 草草在线视频免费看| 99国产精品99久久久久| 男人的好看免费观看在线视频 | 夜夜躁狠狠躁天天躁| 男女视频在线观看网站免费 | 亚洲天堂国产精品一区在线|