• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Anomaly Detection Algorithm of Power System Based on Graph Structure and Anomaly Attention

    2024-05-25 14:39:54YifanGaoJiemingZhangZhanchenChenandXianchaoChen
    Computers Materials&Continua 2024年4期

    Yifan Gao,Jieming Zhang,Zhanchen Chen and Xianchao Chen

    Zhaoqing Power Supply Bureau of Guangdong Power Grid Co.,Ltd.,Zhaoqing,526060,China

    ABSTRACT In this paper,we propose a novel anomaly detection method for data centers based on a combination of graph structure and abnormal attention mechanism.The method leverages the sensor monitoring data from target power substations to construct multidimensional time series.These time series are subsequently transformed into graph structures,and corresponding adjacency matrices are obtained.By incorporating the adjacency matrices and additional weights associated with the graph structure,an aggregation matrix is derived.The aggregation matrix is then fed into a pre-trained graph convolutional neural network(GCN)to extract graph structure features.Moreover,both the multidimensional time series segments and the graph structure features are inputted into a pretrained anomaly detection model,resulting in corresponding anomaly detection results that help identify abnormal data.The anomaly detection model consists of a multi-level encoder-decoder module,wherein each level includes a transformer encoder and decoder based on correlation differences.The attention module in the encoding layer adopts an abnormal attention module with a dual-branch structure.Experimental results demonstrate that our proposed method significantly improves the accuracy and stability of anomaly detection.

    KEYWORDS Anomaly detection;transformer;graph structure

    1 Introduction

    As China’s computer technology and artificial intelligence technology continue to mature,more and more industrial equipment to pursue the realization of industrial intelligence.As the core hub of our production and life,the power system is developing rapidly in the direction of large capacity,ultra-high voltage and long distance,and the requirements for safe and reliable operation of power equipment are becoming more and more strict.Therefore,the research and application of fault diagnosis technology for power equipment is of great practical significance.How to improve the accuracy and speed of the fault diagnosis system,so that the system can quickly and effectively find faults,locate faults,and analyze faults,is an important issue in power grid protection.A number of scholars and industry practitioners are currently working on the detection of anomalies in equipment,which can be defined as the problem of finding instances or models in data that deviate from normal behaviors[1].Depending on the application area,these‘deviations’can also be referred to as anomalies,outliers or singularities.Anomaly detection is used in a wide variety of fields,such as cyber security[2],anomaly detection in industrial systems[3],and network security[4]among others.One of the main reasons why anomaly detection is important is that anomalies often indicate important,critical and hard-to-capture information that is crucial to a company.Sheng[5]used multi-sensors to monitor the three main parameters of temperature,smoke and firelight,and uses multi-sensor fusion technology to solve the contradiction between alarm sensitivity and false alarm rate.Liu et al.[6]analyzed the sound data and vibration data of communication room equipment to determine whether the current health status of communication room equipment is abnormal;and predict the development tendency of the health status of communication room equipment and the time point of reaching the dangerous level.Li [7] used a fuzzy neural network combining neural networks and fuzzy theory to analyze the relationship between the various influencing factors and the equipment status in the power dispatch system,and to accurately describe the operation status of the equipment.The above methods do not work well for the detection of high-dimensional time series data.How to capture complex inter-sensor relationships and detect and interpret anomalies that deviate from these relationships is a challenging problem today? In addition to this,existing anomaly detection methods have several drawbacks,such as incorrect data collection,difficulty in parameter tuning,and the need for annotated datasets for training[8].Recently,deep learning methods have enabled improved anomaly detection in high-dimensional datasets.Transformers [9] has achieved excellent performance in many tasks in natural language processing and computer vision,and as a result an increasing number of researchers are applying Transformers to anomaly detection.Yin et al.[10] proposed an integrated model of convolutional neural networks (CNN) and recursive autoencoders for anomaly detection and used two-stage sliding windows in data preprocessing to learn data features,with results showing better performance on multiple classification metrics.Li et al.[11]used self-supervised models to construct a high-performance defect detection model that detects unknown anomalous patterns in images in the absence of anomalous data.Xu et al.[12]used the Transformer model and incorporated a very large and very small strategy to amplify,the gap between normal and abnormal data points to improve the correct detection rate of the model.To improve the feature extraction of models for non-linear variables,graphs are widely used to describe real-world objects and their interactions.Graph Neural Networks(GNN)[13–16]as de facto models for analyzing graph structured data,are highly sensitive to the quality of a given graph structure.This paper introduces an algorithm for anomaly detection in power systems based on graph structure and anomaly attention[17–19]which uses an unsupervised deep learning approach to detect anomalies in power equipment,omitting the step of manually annotating the dataset.The anomalous data and states are detected by exploiting correlations between multidimensional time series variables.Its validation results show that multidimensional variables can effectively capture the relationships and correlations between equipment operating states,avoiding the problems of limited prediction accuracy and lack of stability associated with single-dimensional data.Its contributions are summarized below:

    ? In this paper,a graph structure is established to represent multidimensional time series as directed graphs,enabling the model to learn the dependencies and correlations between multidimensional time data more effectively;

    ? In terms of technology,we propose a power system anomaly detection algorithm based on graph structure and anomaly concern to evaluate the healthy operation status of power plant rooms and to conduct unsupervised research on early faults;

    ? Based on the traditional transformer model,the Coding-Decoding Anomaly Detection Model(CD-ADM)model proposed in this paper uses a network structure with multiple encoders in series,which solves the problem that there may be too many data features and the anomaly detection capability of the model decreases.

    The remainder of the paper is organized as follows.We describe the methodology in Section 2,which includes the utilization of graph structures,graph convolutional networks,and an abnormal attention mechanism.Section 3 outlines the model’s flow,from data preprocessing to the final output.In Section 4,we present the results of our experiments,including evaluation metrics and performance comparisons.Finally,we provide discussion and conclusions in Sections 5 and 6.

    2 Materials and Methods

    2.1 Graph Structure

    The text graph structure is a more complex data structure than linear tables and tree structures,which can transform data from two-dimensional to multidimensional representation.A node can be connected to any node,and the structure formed by connecting all nodes is called a graph structure.In the graph structure,there may be an association between any two nodes,that is,there are various adjacency relationships between nodes.The graph structure is usually expressed as:

    where G represents a graph,V is the set of all nodes in the graph,and E is the set of edges between nodes.

    In a power machine room that needs to detect faults,sensors are used to read data such as main bearing temperature,fan speed,N1 CPU temperature,PCH,power supply voltage,etc.,to form a multidimensional time series as the input vector X,(1) Multivariate time series: Multivariate time.The sequence can be expressed as:

    wherexi∈RTis a univariate time series of length T,and N is the total number of signals,that is,the number of sensors.y∈Yrepresents the corresponding label,Y={1,2,...,K} represents normal data with k categories.(2)Abnormality:If the labelyaof the sample(xa,ya)does not belong to any predefined class,that is,it is abnormal data.Think of sensors as nodes in the graph structure,and the correlations between sensors as edges between graph structures,as shown in Fig.1.

    The definition of a graph isG=(V,E),where V represents the set of nodes and E represents the set of edges.

    where N represents the number of sensors.

    whereeirepresents the correlation hypothesis between each sensor and the remaining sensors,and the value range is(0–1).

    Since the model does not know the edge information in the graph structure when it is initialized,that is,E is unknown,we randomly initialize E and put it and the adjacency matrix in the model’s graph structure into the graph feature extraction network for training.Here we use the adjacency matrix A to represent the graph structure in a matrix.The sensors corresponding to the multidimensional time series are regarded as nodes in the graph structure,the correlation between sensors is regarded as the edges between nodes in the graph structure,and an adjacency matrix is constructed based on the correlation between nodes in the graph structure.

    Figure 1: Schematic diagram of graph structure

    First,for each sensor node,select relevant nodes to form a candidate node set:which does not contain i.

    In order to express the correlation between sensor node j and candidate node i,the correlation measurement relationship is introduced,expressed byCorji.

    where the first part is cosine correlation,which is used to measure the correlation between nodes in space,and the latter part is P probability correlation,xi,xjwhich is used to measure the correlation of nodes in time.Assume that N-dimensional time data obeys Gaussian distribution,where N is the number of sensors,sensor node j and candidate node i are respectivelyxi,xj,then

    whereμ,σrepresent the mean and variance,respectively.In the formula,Corjirepresents the correlation between nodesxi,xj,eiis the correlation hypothesis between the nodexiand other nodes,the value range is (0–1),prepresents the probability that the node data appears together in the multidimensional l time series in the multidimensional l time series,p(xi)is the probability that node data appears in a multidimensional time series.Ciis the set of related nodes of node.Combining the two dimensions of space and time,the dependence between nodes can be expressed more effectively and the topological structure of the nodes can be refined.The larger the result ofCorji,the more similar the two nodes are.The elements of the adjacency matrixAjiare expressed as:

    Using Eq.(6),first calculate the correlation between sensor i and candidate node j,and then select the first k values,(N 3

    Merge the adjacency matrix A with the additional weight E of the graph structure to obtain the aggregation matrix ?A,that is:

    whereα∈(0,1)is the preset graph structure complexity coefficient,which is selected according to the complexity of the actual network.

    2.2 Graph Convolutional Network(GCN)

    Graph convolutional neural network[20–22],is a type of neural network that convolves the graph structure and is essentially a feature extractor.Since the graph structure is more complex and irregular than the tree structure,it can be regarded as a kind of infinite dimensional data,and it does not have translation invariance during data processing.Therefore,CNN and Recurrent Neural Network(RNN)cannot be used for feature extraction.However,because nodes in the graph structure are easily affected by neighboring nodes,they will also affect other nearby nodes.Under the influence of this interdependence,the nodes will reach a final equilibrium state.Taking advantage of this feature,graph convolutional neural networks can complete feature extraction of graph structures.This article builds a lightweight convolutional neural network to extract features of the graph structure.The parameters of the convolution kernel here are 3 ?3.The input of the network is an aggregation matrix,which is an N-order square matrix,where N represents the number of sensors.That is,the dimensions of multidimensional time series.Due to the lightweight characteristics of the convolutional network,the network is easy to transplant and embed.The graph convolutional neural network shown in Fig.2,can be used to learn the correlation of adjacent nodes,that is,to build a feature extractor based on the aggregation matrix to learn the information in the graph structure and connect the nodes with surrounding nodes.

    Figure 2: Graph convolutional neural network for feature extraction of graph structures

    The input aggregation matrix is ?A,its dimension is (N ?N ?T),N is the number of sensors,T is the time series of signals intercepted through the sliding window,the sequence length is T,and the parameters of the convolution kernel are 3 ?3.After adding the pooling layer,the features are compressed through downsampling,the parameters of the fully connected layer are reduced,and the complexity of the model is reduced.Finally,through the fully connected layer,each node of the adjacency matrix is represented as a weighted sum of related nodes.Where,the i of each node,after passing through the convolution pooling-fully connected layer,is comprehensively expressed as:

    Through the convolutional neural network extracted from graph structure features,the model output is the representation of N nodes,that isthe loss function is expressed as:

    2.3 Abnormal Attention Mechanism

    Literature [12] improved the structure of the transformer and proposed two methods for calculating correlation,namely a priori correlation and a sequence correlation.The prior correlation mainly represents the correlation between each time point and adjacent time points,while the sequence correlation The focus is on the correlation between each time point and the entire sequence.In view of this,the network structure diagram used in the proposed model is shown in Fig.3.Coding-Decoding Anomaly Detection Model (CD-ADM).That is,the multidimensional time series perceived by the power system is intercepted through a sliding window,and a time series of length T is intercepted.An encoder based on the anomaly attention mechanism is used for feature extraction,and then the decoder based on the multi-head attention mechanism is used for prediction.

    The reconstruction error of its output is:

    whereXl∈RN?d,l∈ {1,...,L} represents the output of the L-th layer anddmodelchannel.The initial inputX0=Embedding(X) represents the original sequence of the embedding.Zl∈RN×dmodelrepresents the hidden representation of layer L.The encoder has an unusual attention to two branch structures internally.For the prior correlation,a learnable Gaussian kernel is used to calculate the prior correlation relative to the temporal distance.Since the Gaussian kernel function has only one peak,and the closer the distance to the center point,the closer the similarity is to 1,and vice versa.Taking advantage of this feature,the attention mechanism that introduces the Gaussian kernel function will pay more attention to the features of nearby nodes.Here,a learnable parameter is embedded into the Gaussian kernel function so that the prior correlation adapts to various time series patterns,such as abnormal segments of different lengths.The sequence correlation branch learns correlations from the original sequence and can adaptively find the most effective correlations.The abnormal time series attention model parameters of layer L are:The initial value of Q,K,V is:

    Figure 3: Abnormal attention mechanism structural mode

    Due to the rarity of equipment anomalies in power equipment rooms and the dominance of the normal mode,it is difficult to establish a strong correlation between anomalies and the entire sequence.Anomalous correlations should be clustered at adjacent time points,which are more likely to contain similar anomalous patterns due to continuity.This adjacent correlation bias is called a priori correlation.The present invention calculates it and records it as P,and uses the learnable Gaussian kernel r to calculate it.The calculation method is as follows:

    wherer∈RN×1N is the length of the time series involved in the calculation,i,j correspond to different time points,that is,the association weight from the i-th time point to the j-th time point is calculated by the Gaussian kernel:

    Rescale(.)represents the division by row sum operation,which is used to transform the associated weight into a discrete distribution.The model structure of the decoder is shown in Fig.4 below.

    Figure 4: The model structure of the decoder

    The reconstruction error is:

    As an unsupervised task,we employ a reconstruction loss to optimize our model.The reconstruction loss will guide the series of correlations to find the most informative correlation.To further amplify the differences between normal and abnormal time points,we also use additive losses to amplify the correlation differences.The loss function is:

    where e is the additional weight of each node,w is the parameter of the neural network,h(t) is the output obtained after the adjacency matrix of the graph structure passes through the graph neural network,indicating the sequence obtained by the anomaly detection module based on the self-attention mechanism reconstruction loss.We calculated an anomaly score for each time series on the sensor.Compare the model prediction results with the observed time series,and calculate the error value of sensor i at time t:

    Set the threshold to the maximum value of the validation set data,that is,exceed the maximum anomaly score of the validation data set.If A(t),exceeds a fixed threshold,the time is marked as an anomaly.We use the data of the previous few days at the current moment for training modeling,and then use the data of the current day as the test set to obtain the anomaly score.This improvement also achieves the ability to detect device anomalies in real time.

    3 Equations and Mathematical Expressions

    In the experiment,the operating health status of the power computer room is displayed in the form of data,which is used as an evaluation indicator for computer room maintenance.Through a reasonable diagnosis mechanism and status operation monitoring mechanism,the probability of failure in the power machine room will be effectively reduced.The proposed model constructs the following six parts for the abnormal diagnosis process,as shown below: 1.Perceive the signal;2.Perform preprocessing operations on multidimensional l time series signals and convert them into graph structures;3.Input the adjacency matrix obtained from the graph structure into the graph neural network,and the model output is a weighted set result between multidimensional time series variables;4.Pass the original signal through a sliding window of length T and input it into the encoder based on the self-attention mechanism.Multiple encoders are connected in series,and then input into the decoder containing the multi-head attention module,and then the model output is the reconstruction error;5.Establish a threshold and define anomaly scores exceeding the threshold as anomaly sequences;6.After an abnormality occurs,start the early warning mode.The abnormal diagnosis flow chart construction is shown in Fig.5.

    Figure 5: Anomaly detection flow chart

    4 Results

    First,the abnormal model training of the power computer room data set and the verification result analysis are performed.Figs.6 and 7 below respectively show the verification results of the(long shortterm memory)LSTM model and the CD-ADM model on the power computer room test set.The blue solid line represents the real data,the orange dotted line represents the predicted data,and the residual value of the two.It is the green curve.

    In Figs.6 and 7,the LSTM model test results show smaller fitting errors,while CD-ADM shows a more obvious residual amount.It is impossible to judge the pros and cons of the two methods just from the above picture.Next,real anomaly points are added,and the prediction error(green curve)is used as the evaluation criterion to determine equipment anomalies.The anomaly detection results in Fig.8,below are obtained:

    Figure 6: LSTM model test results chart

    Figure 7: CD-ADM model test results char

    The pink curve in Fig.8,above is the anomaly score curve of the CD-ADM model,the green is the anomaly score curve of the LSTM model,and the blue dotted line is the artificially set anomaly judgment threshold.If it is higher than the blue dotted line,it is considered an anomaly.The real points are abnormal alarms detected by the model,and the red crosses are the actual fault occurrence points of the data.The above figure shows that the CD-ADM model can fully learn the characteristics that the device health status should display,and it will also show a gradual upward trend when it is closer to the abnormal point.This is what this experiment hopes the model can learn.Important information.In the first section of the curve,the CD-ADM model and the LSTM model show similar characteristics;in the second section,the former shows a gradually increasing trend than the latter and generates multiple times before the first shutdown for maintenance,warning information,and also showed similar characteristics before the second shutdown;in the third paragraph,it can be seen that the improved model clearly shows an increasing trend,but the trend cannot be seen in the unimproved model.The upward trend shown in the CD-ADM model provides valuable information about the operational health of the equipment,because failures caused by this slow wear are often progressive and avoidable.Figs.9 and 10 show the fitting performance of the CD-ADM model and LSTM model on normal datasets in power rooms.Comparing the prediction error curves,except for some differences in the time period when the real curve fluctuates,the overall prediction modeling accuracy and fitting effect are maintained very well.

    Figure 8: Comparison chart of anomaly detection between LSTM model and CD-ADM model

    Figure 9: The fitting of CD-ADM on normal data sets in power rooms

    Figure 10: The fitting of LSTM on normal data sets in power rooms

    5 Discussion

    In this study,we proposed a novel approach for abnormal warning detection in power equipment rooms based on graph convolutional attention mechanism.Our results demonstrate that the use of graph convolutional networks with attention mechanism yields promising performance in identifying abnormal patterns within the power equipment rooms.The model showed high accuracy in detecting various anomalies such as temperature fluctuations,humidity deviations,and equipment malfunctions.One of the key strengths of our approach lies in its ability to capture complex relationships and dependencies among different sensors and equipment within the power equipment rooms.By leveraging the graph convolutional attention mechanism,the model can effectively learn and adapt to the dynamic interactions among the various data points,leading to improved abnormal detection capability.However,it is important to acknowledge the limitations of our study.One potential limitation is the dependency on the quality and quantity of the training data.Collecting sufficient and diverse datasets to represent all possible abnormal scenarios remains a challenge.Additionally,the computational complexity associated with graph convolutional networks may hinder real-time deployment in some practical settings.

    In future work,it would be beneficial to explore methods for enhancing the robustness of the model against noisy or incomplete data.Furthermore,investigating techniques to optimize the computational efficiency of the graph convolutional attention mechanism would facilitate its implementation in real-time abnormal warning systems.Additionally,the integration of multi-modal data sources and the incorporation of domain knowledge could further improve the model’s performance in capturing complex abnormal patterns.

    6 Conclusions

    This article uses the encoding-decoding anomaly detection model(CD-ADM)to reproduce the process of abnormal warning in the computer room.This paper proposes the application of deep learning to intelligent detection methods of power equipment room equipment faults,and the use of unsupervised deep learning methods to reduce the labeling of data sets,which greatly saves labor costs and thus improves traditional detection methods.This paper proposes a graph structure to establish correlations between multidimensional data.Traditional anomaly detection methods do not know which sensors are related to each other,so it is difficult to build sensor data with many potential correlations.In addition,traditional graph neural networks use the same model to establish the graph structure for each node,which limits the flexibility of the model.Therefore,we improve the graphstructured feature learning network,add additional weights to each edge,and select the k value for training according to the complexity of the model.Therefore,we can accurately understand the interdependencies between sensors.That is,the multidimensional time series is represented by a graph structure,and the aggregation matrix obtained by the graph structure is input to the feature learning network of the graph structure.Taking advantage of the lightweight characteristics of the network,it is easy to transplant and embed the network,realize nonlinear transformation,enhance the expression ability of the model,and perform end-to-end learning of node feature information and structural information.Finally,this method can effectively learn the correlation between multidimensional time variables,establish the topology of time series,and convert from two-dimensional space to multidimensional space.In addition,based on the self-attention mechanism,the single-branch selfattention module is changed to a dual-branch anomaly attention detection module to improve the model’s ability to distinguish between normal data and abnormal data.In the prediction time series module,in order to avoid the“over-fitting”phenomenon,the gradient disappears and the prediction results are often volatile,which makes the prediction performance of the model unstable.The proposed model can effectively solve the above problems by connecting multiple encoders in series.

    Acknowledgement:Thanks to Prof.Dr.Jiande Zhang from Nanjing Institute of Technology for his comments.The authors thanks to all individuals who contributed to the planning,implementation,editing,and reporting of this work but are not listed as authors.Their invaluable support played a significant role in the success of this research.

    Funding Statement:This paper was funded by the Science and Technology Project of China Southern Power Grid Company,Ltd.(031200KK52200003),the National Natural Science Foundation of China(Nos.62371253,52278119).

    Author Contributions:Conceptualization,Yifan Gao;Data curation,Xianchao Chen;Formal analysis,Yifan Gao;Investigation,Zhanchen Chen;Methodology,Yifan Gao,Jieming Zhang and Xianchao Chen;Software,Yifan Gao;Supervision,Xianchao Chen;Validation,Zhanchen Chen;Writing–original draft,Jieming Zhang;Writing–review editing,Zhanchen Chen.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:Due to the nature of this research,participants of this study did not agree for their data to be shared publicly,so supporting data is not available.We provide the author’s contact information so that readers in need can get in touch with us and learn more about relevant information.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    亚洲国产欧美人成| 精品电影一区二区在线| 中文字幕av在线有码专区| 久久久久精品国产欧美久久久| 免费电影在线观看免费观看| 嫩草影视91久久| 亚洲色图av天堂| 国产成人欧美在线观看| 日韩高清综合在线| 亚洲精品粉嫩美女一区| 欧美又色又爽又黄视频| 老熟妇仑乱视频hdxx| 欧美大码av| 午夜激情欧美在线| 亚洲av一区综合| 高清日韩中文字幕在线| 国产av在哪里看| 国产视频一区二区在线看| 国内精品久久久久精免费| 国产亚洲欧美98| 一本久久中文字幕| 麻豆国产97在线/欧美| 在线观看一区二区三区| 最好的美女福利视频网| 欧美绝顶高潮抽搐喷水| 美女被艹到高潮喷水动态| 国产久久久一区二区三区| 欧美极品一区二区三区四区| 99久久精品国产亚洲精品| 麻豆成人午夜福利视频| 亚洲最大成人手机在线| 成人性生交大片免费视频hd| www.www免费av| 中亚洲国语对白在线视频| 丰满人妻一区二区三区视频av | 蜜桃亚洲精品一区二区三区| 午夜精品在线福利| 一个人观看的视频www高清免费观看| 3wmmmm亚洲av在线观看| 成人av一区二区三区在线看| 亚洲av二区三区四区| 欧美一区二区精品小视频在线| 国产视频内射| 禁无遮挡网站| 久久精品国产清高在天天线| 每晚都被弄得嗷嗷叫到高潮| 又黄又粗又硬又大视频| 亚洲av不卡在线观看| 叶爱在线成人免费视频播放| 日本 av在线| 大型黄色视频在线免费观看| 午夜两性在线视频| 可以在线观看毛片的网站| 狂野欧美激情性xxxx| 最近最新中文字幕大全免费视频| 欧美又色又爽又黄视频| 亚洲电影在线观看av| 亚洲av五月六月丁香网| 欧美日韩瑟瑟在线播放| 又黄又爽又免费观看的视频| 成人精品一区二区免费| 欧美乱色亚洲激情| 午夜福利在线观看免费完整高清在 | 色精品久久人妻99蜜桃| 一二三四社区在线视频社区8| 国产伦精品一区二区三区视频9 | 99热只有精品国产| 91av网一区二区| 久久久久久人人人人人| 成人永久免费在线观看视频| 日日摸夜夜添夜夜添小说| 夜夜夜夜夜久久久久| 久久久久久大精品| 日韩成人在线观看一区二区三区| 免费一级毛片在线播放高清视频| 亚洲国产中文字幕在线视频| 成熟少妇高潮喷水视频| 女警被强在线播放| 欧美色视频一区免费| 桃红色精品国产亚洲av| 男人和女人高潮做爰伦理| 母亲3免费完整高清在线观看| av在线天堂中文字幕| 亚洲美女视频黄频| av在线蜜桃| 搞女人的毛片| 午夜视频国产福利| 真实男女啪啪啪动态图| 午夜精品在线福利| 免费观看人在逋| АⅤ资源中文在线天堂| 色吧在线观看| av国产免费在线观看| www日本黄色视频网| 日韩中文字幕欧美一区二区| 91在线观看av| 级片在线观看| 亚洲人成伊人成综合网2020| 国模一区二区三区四区视频| 国产一区二区在线av高清观看| 婷婷精品国产亚洲av在线| 一区二区三区激情视频| 在线观看66精品国产| 日韩精品中文字幕看吧| 亚洲精品一卡2卡三卡4卡5卡| 波多野结衣巨乳人妻| 中文资源天堂在线| 在线国产一区二区在线| 亚洲av中文字字幕乱码综合| 无限看片的www在线观看| 久久久久免费精品人妻一区二区| 色视频www国产| 久久亚洲真实| 制服丝袜大香蕉在线| 舔av片在线| 亚洲狠狠婷婷综合久久图片| 高清日韩中文字幕在线| 99热只有精品国产| 久久久久久国产a免费观看| 欧美黑人欧美精品刺激| 久久香蕉国产精品| 日本免费a在线| 亚洲激情在线av| 精品国产三级普通话版| 老汉色∧v一级毛片| 又黄又粗又硬又大视频| 国产aⅴ精品一区二区三区波| 给我免费播放毛片高清在线观看| 国产伦在线观看视频一区| 国产aⅴ精品一区二区三区波| 高清在线国产一区| 国产色爽女视频免费观看| 可以在线观看毛片的网站| 真人做人爱边吃奶动态| 制服人妻中文乱码| 高清在线国产一区| xxx96com| 色综合婷婷激情| 好看av亚洲va欧美ⅴa在| 免费看光身美女| 亚洲专区国产一区二区| 国产极品精品免费视频能看的| 99精品在免费线老司机午夜| 久久久久久久久中文| 国产毛片a区久久久久| 国产乱人伦免费视频| 一区二区三区激情视频| 午夜两性在线视频| 国产成人系列免费观看| 亚洲熟妇熟女久久| 国产精品香港三级国产av潘金莲| 国产精品综合久久久久久久免费| 韩国av一区二区三区四区| 成人一区二区视频在线观看| 脱女人内裤的视频| 看免费av毛片| 国产精品99久久久久久久久| av福利片在线观看| 在线观看一区二区三区| 日韩欧美一区二区三区在线观看| 男人舔女人下体高潮全视频| 国产欧美日韩精品亚洲av| 国产黄a三级三级三级人| 嫩草影院精品99| 亚洲五月婷婷丁香| 国产精品美女特级片免费视频播放器| 老鸭窝网址在线观看| 国产美女午夜福利| 级片在线观看| 小蜜桃在线观看免费完整版高清| 欧美日本视频| 久久国产精品影院| 日韩欧美 国产精品| 男女午夜视频在线观看| 一进一出好大好爽视频| 一个人免费在线观看的高清视频| 亚洲激情在线av| 国产av不卡久久| 深爱激情五月婷婷| 精品乱码久久久久久99久播| 国产男靠女视频免费网站| av国产免费在线观看| 亚洲内射少妇av| 亚洲无线在线观看| 亚洲欧美一区二区三区黑人| 精品欧美国产一区二区三| av女优亚洲男人天堂| 欧美激情在线99| 色精品久久人妻99蜜桃| 18禁黄网站禁片午夜丰满| 亚洲欧美日韩高清专用| 亚洲欧美激情综合另类| 色av中文字幕| 老司机福利观看| 成人18禁在线播放| 可以在线观看毛片的网站| 日本黄色视频三级网站网址| 网址你懂的国产日韩在线| 99视频精品全部免费 在线| 午夜福利高清视频| 黄色日韩在线| 9191精品国产免费久久| 亚洲av五月六月丁香网| 国产精品久久久久久亚洲av鲁大| 无限看片的www在线观看| 麻豆国产av国片精品| 欧美成人免费av一区二区三区| 欧美zozozo另类| 床上黄色一级片| 久久精品综合一区二区三区| 床上黄色一级片| 亚洲国产精品999在线| 我的老师免费观看完整版| 欧美乱色亚洲激情| 一区二区三区国产精品乱码| 国产高清三级在线| 久久国产精品人妻蜜桃| 亚洲国产中文字幕在线视频| 99久久综合精品五月天人人| 国产真实乱freesex| 国产精品久久视频播放| а√天堂www在线а√下载| 久久久国产精品麻豆| 天堂√8在线中文| 国产精品av视频在线免费观看| 99久久久亚洲精品蜜臀av| 国产黄片美女视频| 中文资源天堂在线| 18+在线观看网站| 哪里可以看免费的av片| 一进一出好大好爽视频| 男女之事视频高清在线观看| 亚洲熟妇中文字幕五十中出| 日本免费a在线| 久久久色成人| 18禁黄网站禁片午夜丰满| 亚洲欧美精品综合久久99| 丁香六月欧美| 又爽又黄无遮挡网站| 一级黄片播放器| 国产成人a区在线观看| 亚洲中文字幕日韩| 高清日韩中文字幕在线| 90打野战视频偷拍视频| 中国美女看黄片| 51国产日韩欧美| 一级毛片高清免费大全| 每晚都被弄得嗷嗷叫到高潮| 香蕉av资源在线| 最近最新免费中文字幕在线| 淫妇啪啪啪对白视频| 欧美三级亚洲精品| 少妇丰满av| 亚洲精品国产精品久久久不卡| 国产高清视频在线观看网站| 亚洲熟妇熟女久久| 欧美一区二区亚洲| 日韩av在线大香蕉| 久久九九热精品免费| 一卡2卡三卡四卡精品乱码亚洲| 日本三级黄在线观看| 欧美最新免费一区二区三区 | 国产单亲对白刺激| 麻豆一二三区av精品| 白带黄色成豆腐渣| 国产男靠女视频免费网站| 久久久久久国产a免费观看| 日韩高清综合在线| 亚洲av电影不卡..在线观看| 在线十欧美十亚洲十日本专区| 久久精品91蜜桃| www.色视频.com| 啪啪无遮挡十八禁网站| 成年女人毛片免费观看观看9| 欧美一级a爱片免费观看看| 日韩欧美 国产精品| 香蕉av资源在线| www.www免费av| 国产精品永久免费网站| 久久精品国产亚洲av涩爱 | 亚洲av一区综合| 性色avwww在线观看| 亚洲成a人片在线一区二区| 一本精品99久久精品77| 久久久久久久亚洲中文字幕 | 最新美女视频免费是黄的| 亚洲国产日韩欧美精品在线观看 | 母亲3免费完整高清在线观看| 日韩有码中文字幕| 国产又黄又爽又无遮挡在线| 久久精品亚洲精品国产色婷小说| 婷婷亚洲欧美| 国产爱豆传媒在线观看| 国产主播在线观看一区二区| 亚洲欧美一区二区三区黑人| av女优亚洲男人天堂| 日日摸夜夜添夜夜添小说| 成人特级黄色片久久久久久久| 久久这里只有精品中国| 91在线精品国自产拍蜜月 | 变态另类成人亚洲欧美熟女| 日本 av在线| 99久久无色码亚洲精品果冻| 久久伊人香网站| 手机成人av网站| 精品国产美女av久久久久小说| 岛国在线免费视频观看| 国产成人啪精品午夜网站| 欧美日韩精品网址| 伊人久久大香线蕉亚洲五| 中文字幕熟女人妻在线| 岛国在线免费视频观看| 亚洲欧美日韩卡通动漫| 无人区码免费观看不卡| 黑人欧美特级aaaaaa片| 国产午夜精品论理片| 国产黄片美女视频| 身体一侧抽搐| 日本五十路高清| 国产真实伦视频高清在线观看 | 国产精品爽爽va在线观看网站| 在线免费观看的www视频| 色综合站精品国产| 亚洲五月婷婷丁香| 国产精品 欧美亚洲| 国内精品一区二区在线观看| av中文乱码字幕在线| 丰满的人妻完整版| 99久久成人亚洲精品观看| 久久久久久久亚洲中文字幕 | 老司机午夜福利在线观看视频| 国内久久婷婷六月综合欲色啪| 国产亚洲精品综合一区在线观看| 国产乱人伦免费视频| 亚洲国产精品999在线| 亚洲av日韩精品久久久久久密| www.www免费av| 亚洲男人的天堂狠狠| 国产亚洲精品一区二区www| 欧美三级亚洲精品| 草草在线视频免费看| 性色av乱码一区二区三区2| 欧美成狂野欧美在线观看| 欧美日韩亚洲国产一区二区在线观看| 久久久久久久久中文| 精品国产超薄肉色丝袜足j| 国产av不卡久久| 亚洲人与动物交配视频| 在线观看舔阴道视频| 午夜福利在线观看免费完整高清在 | 国产一区二区激情短视频| 一夜夜www| 老司机福利观看| 精品福利观看| 19禁男女啪啪无遮挡网站| 首页视频小说图片口味搜索| 亚洲专区国产一区二区| 一进一出好大好爽视频| 国产美女午夜福利| 少妇高潮的动态图| 国产欧美日韩一区二区三| 女警被强在线播放| 韩国av一区二区三区四区| 少妇丰满av| 俄罗斯特黄特色一大片| 久久伊人香网站| 成人av一区二区三区在线看| 国产三级黄色录像| 黄色片一级片一级黄色片| a级一级毛片免费在线观看| 国产高潮美女av| 久久久久久国产a免费观看| x7x7x7水蜜桃| 欧美乱色亚洲激情| 中文字幕人成人乱码亚洲影| 国产精品久久久久久人妻精品电影| 国产精品亚洲av一区麻豆| 老司机深夜福利视频在线观看| 日韩人妻高清精品专区| 午夜福利免费观看在线| 日本撒尿小便嘘嘘汇集6| 国产伦在线观看视频一区| 午夜a级毛片| 99riav亚洲国产免费| 欧美一级a爱片免费观看看| 欧美一区二区精品小视频在线| 国产亚洲欧美在线一区二区| 99久久精品国产亚洲精品| 国产精品98久久久久久宅男小说| 国产一区二区在线观看日韩 | 51午夜福利影视在线观看| 国产v大片淫在线免费观看| 亚洲专区国产一区二区| 中文字幕人妻丝袜一区二区| 可以在线观看的亚洲视频| 在线观看舔阴道视频| 97人妻精品一区二区三区麻豆| 欧美成人性av电影在线观看| 欧美中文综合在线视频| 亚洲人成网站在线播放欧美日韩| 精品乱码久久久久久99久播| 免费搜索国产男女视频| 在线看三级毛片| 日韩精品中文字幕看吧| 神马国产精品三级电影在线观看| 亚洲18禁久久av| 少妇人妻精品综合一区二区 | 日韩欧美精品v在线| 丝袜美腿在线中文| 精品一区二区三区av网在线观看| 国产午夜福利久久久久久| 国产一区二区三区在线臀色熟女| 国产麻豆成人av免费视频| 久久精品国产亚洲av涩爱 | 99热精品在线国产| 午夜福利在线在线| а√天堂www在线а√下载| 尤物成人国产欧美一区二区三区| 日本成人三级电影网站| 亚洲国产欧洲综合997久久,| 在线看三级毛片| 婷婷亚洲欧美| 亚洲av免费在线观看| 一进一出抽搐动态| 搞女人的毛片| 亚洲激情在线av| 女警被强在线播放| 欧美精品啪啪一区二区三区| 黄片小视频在线播放| 非洲黑人性xxxx精品又粗又长| 午夜日韩欧美国产| 国产不卡一卡二| 一本一本综合久久| 欧美乱码精品一区二区三区| 久久婷婷人人爽人人干人人爱| 成人无遮挡网站| 人人妻人人看人人澡| 欧美bdsm另类| tocl精华| 欧美日韩国产亚洲二区| 国产精品av视频在线免费观看| 免费看十八禁软件| 看黄色毛片网站| 男女下面进入的视频免费午夜| 青草久久国产| 搡老熟女国产l中国老女人| 99久久精品热视频| 成人性生交大片免费视频hd| 99国产精品一区二区三区| 深夜精品福利| 亚洲一区高清亚洲精品| 国产精品亚洲一级av第二区| 亚洲无线在线观看| 亚洲精品亚洲一区二区| 欧美zozozo另类| 国产精品电影一区二区三区| 国产真人三级小视频在线观看| 亚洲五月天丁香| 色综合婷婷激情| 久久精品国产亚洲av香蕉五月| bbb黄色大片| 又黄又粗又硬又大视频| 97超视频在线观看视频| 97超级碰碰碰精品色视频在线观看| 日韩精品中文字幕看吧| 国产国拍精品亚洲av在线观看 | 久久久精品大字幕| av中文乱码字幕在线| 亚洲精品久久国产高清桃花| 午夜免费男女啪啪视频观看 | 97人妻精品一区二区三区麻豆| 日本 欧美在线| 少妇的逼水好多| 免费av毛片视频| 欧美3d第一页| 999久久久精品免费观看国产| 19禁男女啪啪无遮挡网站| 色综合欧美亚洲国产小说| 亚洲自拍偷在线| 日本撒尿小便嘘嘘汇集6| 国产午夜福利久久久久久| 两个人视频免费观看高清| 精品国内亚洲2022精品成人| 18禁黄网站禁片免费观看直播| 色精品久久人妻99蜜桃| 欧美高清成人免费视频www| 久久久久久久久大av| 欧美3d第一页| 悠悠久久av| 精品久久久久久久末码| 欧美乱色亚洲激情| 十八禁网站免费在线| 亚洲avbb在线观看| 91字幕亚洲| 久久久久九九精品影院| 日韩国内少妇激情av| 女生性感内裤真人,穿戴方法视频| 亚洲第一欧美日韩一区二区三区| 夜夜看夜夜爽夜夜摸| 男女之事视频高清在线观看| 国产色婷婷99| av欧美777| 国产视频内射| 每晚都被弄得嗷嗷叫到高潮| 啪啪无遮挡十八禁网站| 看黄色毛片网站| 亚洲,欧美精品.| 亚洲avbb在线观看| 午夜免费成人在线视频| 国产淫片久久久久久久久 | 午夜激情福利司机影院| 亚洲 欧美 日韩 在线 免费| 女人十人毛片免费观看3o分钟| 中文字幕久久专区| 国产免费av片在线观看野外av| 无人区码免费观看不卡| 九九在线视频观看精品| 18禁美女被吸乳视频| 看片在线看免费视频| 国产精品,欧美在线| 老司机福利观看| 国产精品久久久久久亚洲av鲁大| 国产精品久久视频播放| 欧美绝顶高潮抽搐喷水| 国产精品日韩av在线免费观看| 亚洲av不卡在线观看| av在线蜜桃| 中国美女看黄片| 欧美精品啪啪一区二区三区| 国内毛片毛片毛片毛片毛片| 国产欧美日韩精品一区二区| 日韩欧美三级三区| 亚洲男人的天堂狠狠| 无遮挡黄片免费观看| 国产中年淑女户外野战色| 亚洲一区二区三区色噜噜| 天天躁日日操中文字幕| 特级一级黄色大片| 亚洲无线在线观看| 啦啦啦韩国在线观看视频| 精品乱码久久久久久99久播| 蜜桃久久精品国产亚洲av| 亚洲国产精品999在线| 在线免费观看不下载黄p国产 | 中国美女看黄片| 在线十欧美十亚洲十日本专区| 国产一区二区亚洲精品在线观看| 亚洲第一电影网av| 国产一区在线观看成人免费| 国产精品亚洲av一区麻豆| 好男人电影高清在线观看| 男人和女人高潮做爰伦理| av福利片在线观看| 久久久久久久精品吃奶| 欧美乱色亚洲激情| 听说在线观看完整版免费高清| 亚洲av成人精品一区久久| 国产欧美日韩一区二区三| 亚洲国产中文字幕在线视频| 村上凉子中文字幕在线| 一级黄片播放器| 好男人电影高清在线观看| 99精品久久久久人妻精品| 亚洲中文字幕日韩| 亚洲精品久久国产高清桃花| 国产黄片美女视频| av视频在线观看入口| 久久精品国产亚洲av涩爱 | 成人一区二区视频在线观看| 欧美黄色片欧美黄色片| www.熟女人妻精品国产| 男女下面进入的视频免费午夜| 日本撒尿小便嘘嘘汇集6| 国内少妇人妻偷人精品xxx网站| 给我免费播放毛片高清在线观看| 日韩精品中文字幕看吧| 91在线精品国自产拍蜜月 | 亚洲av美国av| 国产在视频线在精品| 久久精品夜夜夜夜夜久久蜜豆| 最新在线观看一区二区三区| 亚洲精品亚洲一区二区| 国产真实乱freesex| 一区二区三区高清视频在线| 国产欧美日韩一区二区精品| 国产一区二区三区视频了| 久久久色成人| 1000部很黄的大片| 欧美色视频一区免费| 三级国产精品欧美在线观看| 国产精品嫩草影院av在线观看 | 最新中文字幕久久久久| 床上黄色一级片| 人人妻人人看人人澡| 亚洲狠狠婷婷综合久久图片| 人人妻人人看人人澡| 舔av片在线| 国产精品亚洲美女久久久| 超碰av人人做人人爽久久 | 夜夜看夜夜爽夜夜摸| 一本久久中文字幕| 亚洲精品在线美女| 久久久色成人| 久久久久久九九精品二区国产| 午夜精品在线福利| 1000部很黄的大片| 亚洲人与动物交配视频| 亚洲精品亚洲一区二区| 国内精品久久久久久久电影| 精品人妻1区二区| 成年免费大片在线观看| 少妇人妻一区二区三区视频| 亚洲欧美精品综合久久99| 亚洲片人在线观看| 精品人妻一区二区三区麻豆 |