• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Online Rail Fastener Detection Based on YOLO Network

    2022-11-11 10:49:38JunLiXinyiQiuYifeiWeiMeiSongandXiaojunWang
    Computers Materials&Continua 2022年9期

    Jun Li,Xinyi Qiu,Yifei Wei,*,Mei Song and Xiaojun Wang

    1Beijing Key Laboratory of Work Safety Intelligent Monitoring,School of Electronic Engineering,Beijing University of Posts and Telecommunications,Beijing,100876,China

    2Dublin City University,Dublin,9,Ireland

    Abstract: Traveling by high-speed rail and railway transportation have become an important part of people’s life and social production.Track is the basic equipment of railway transportation, and its performance directly affects the service lifetime of railway lines and vehicles.The anomaly detection of rail fasteners is in a priority, while the traditional manual method is extremely inefficient and dangerous to workers.Therefore, this paper introduces efficient computer vision into the railway detection system not only to locate the normal fasteners, but also to recognize the fasteners states.To be more specific, this paper mainly studies the rail fastener detection based on improved You can Only Look Once version 5(YOLOv5)network, and completes the real-time classification of fastener states.The improved YOLOv5 network proposed contains five sections, which are Input, Backbone, Neck, Head Detector and a read-only Few-shot Example Learning module.The main purpose of this project is to improve the detection precision and shorten the detection time.Ultimately,the rail fastener detection system proposed in this paper is confirmed to be superior to other advanced algorithms.This model achieves on-line fastener detection by completing the“sampling-detection-recognition-warning”cycle of a single sample before the next image is sampled.Specifically,the mean average precision of model reaches 94.6%.And the model proposed reaches the speed of 12 ms per image in the deployment environment of NVIDIA GTX1080Ti GPU.

    Keywords: Fastener detection; deep learning; state recognition; real-time classification

    1 Introduction

    The intelligent revolution based on neural network has given new power to many industrial fields,such as automatic control system,intelligent data cache policy for mobile edge networks based on deep Q-learning method[1],heartbeat classification in electronic health using deep learning algorithm[2],cognitive image steganography protocols integrated with machine learning[3],introducing automatic learning algorithms for user scheduling and resource allocation in wireless networks [4].As an important branch in the field of artificial intelligence, object detection technology has been applied more and more in daily life, such as automatic driving, pedestrian detection, large-scale scene recognition, etc.Therefore, the paper investigates the application of object detection technology in railway safety inspection,and makes the railroad track flaws detection more efficient and ensure the safety of railway transportation.While improving the accuracy of the algorithm,researchers also need to expand the scope of target detection technology and super-resolution image data matching[5].Track safety has always been a major theme of research all over the world, and the fastener is one of the important parts of the track.The state of fastener determines the rail safety,a missing or incomplete fastener will cause a significant hidden risk to the possessions of owner,and even may appear a major safety accident.

    Early detection of rail fastener mainly accomplished by manual inspecting.Workers made inspection tour on a regular time to find out whether there were abnormal phenomena.But there are many drawbacks of artificial detection:(1)Human eye fatigue occurs over a long period of time,so the overall reliability is low.(2)Nowadays the running speed of railway transportation is fast,there is no extra time for manual inspection.(3)The safety of the workers cannot be guaranteed.Apart from the most traditional manual methods, there are some automatic detection methods for fasteners internationally.Automated testing methods mainly include:rail detection based on ultrasonic method[6],fastener detection based on the eddy current testing method[7],fastener detection based on pressure detecting[8],fastener detection based on computer vision[9],etc.The first three methods are based on traditional industrial testing methods,but due to difficulties in physical layer,the detection accuracy and detection speed cannot achieve good results.Furthermore,the implementation cost of pressure detecting and other methods is too high.Therefore,more attention is focused on computer vision technology,and this paper is also based on this direction.Currently,commonly used computer vision detectors are mainly divided into one-stage detectors and two-stage detectors.The two-stage detectors are mainly RCNN (Regions with Convolutional Neural Network features)series, such as FAST-RCNN [10], Faster-RCNN [11] et al.The first step is to generate the suggestion box through object proposal and extract the content in the suggestion box through Backbone.The second step is to use linear classifier such as Supported Vector Machine(SVM).RCNN series have tedious training steps and slow training process.Although they have high detection accuracy,some industry that needs to be lightweight to deployed on mobile terminals and real-time to response issues quickly.As for the one-stage-detection, Single Shot Multi-Box Detector (SSD)[12], which shows a great balance between accuracy and speed, is based on position regression for depth learning image detection.The basic size of the prior box in the SSD network cannot be learned directly, but needs to be set manually.However,the size and shape of the prior box used by each feature layer are exactly different,resulting in the debugging process relying heavily on researchers’experience.YOLO series [13] have better performance with 24 convolution layers which are responsible for extracting features and 2 fully connected layers.YOLO series as the fastest detection algorithm are often used in various fields.Finally,Non-Maximum Suppression(NMS)[14]is used to solve the problem that there are multiple repeating frames in the same target.In any detector,whether one-stage or two-stage,the most timeconsuming step is the feature extraction[15].

    The method based on deep learning is gradually applied to automatic detection of rail fasteners.Li [16] proposed an improved SSD fastener positioning algorithm using non-maximum weighted suppression method to replace the original suppression method.After adding residual network to avoid manually adjust the prior box parameters,the improved SSD algorithm reached 90.2%accuracy and speed of 36 FPS.Long[17]used Faster RCNN to build a high-speed rail fastener detection system,but its network structure had a huge amount of computation,resulting in a slow detection speed.The authors in[18]used YOLO network to locate fasteners,but the recognition effect for small targets and the generalization ability were both average.In addition,the detection speed of this paper is only 54 FPS,and the test dataset is too small to prove the true performance of its model.

    According to the present research situation, this paper proposes an On-line Fastener Detection method based on computer vision, which mainly uses image processing algorithm to locate and identify fasteners for track images collected from real scenes.Based on YOLO network,the proposed model can effectively detect railway fasteners in different states.This paper adds read-only Few-Shot Example Learning model to the algorithm based on YOLOv5 detection framework.In addition,the details of YOLOv5 network are adjusted to optimize the method for the practical problems in this paper.This method is robust to almost all complex situations faced by high-speed railway system.

    2 Theory Model

    In the process of training the object detection model, it is necessary to carefully study the composition of the loss function.In practice, this paper mainly uses two loss calculation methods:Cross Entropy Loss[19]and Focal Loss[20].

    Eq.(1)is the calculation method of Cross Entropy function.Cross Entropy Loss is mainly used to judge the loss caused by object classification,and this function will be used to calculate the loss in the classical binary classification problem.

    In Eq.(1)Nclsrepresents the number of anchors selected in training, andLclsin Eq.(2)is the logarithmic loss caused by probability that the training process model judges that whether the anchor box area belongs to the target or not.However,in the practical application of object detection,multiple classifications are often encountered, so the output should be K+1 dimension, where K is the total number of categories and 1 is whether the prediction frame is the background.piis the probability of the prediction bounding box belongs to classi.pi*indicates whether there is an object in the predicted bonding box.

    With the above formula, this paper can basically solve the loss caused by classification in some simple scenes.However, in some complex specific scenes, there are some similar samples that are difficult to distinguish.In addition,most of the one-stage detectors enjoy the speed increase brought by the generation of region proposal network,and at the same time,they will also be affected by the decline of receiving accuracy.Therefore,Focal Loss is used as in Eq.(3).

    In the above formula, the parameterγis the focusing parameter, andγ≥ 0.(1-pt) is an adjustment factor, which can adjust the weight of samples that are difficult to classify in the loss,and make the model optimization stage focus on such samples.For example,if a sample is classified wrong,ptwill be low, then the coefficient(1-pt) of Eq.(3)will nearly reach 1.So it will have little influence on the original classification of this sample.It can be seen that when the model classifies fewshot samples incorrectly, its parameters will increase, thus increasing the weight of the lost samples.In addition, by adjusting the parameterαt, YOLO network can control the weight of positive and negative classification samples in the loss,thus improving the overall accuracy of the model.

    After the training dataset enters YOLO network,the output of feature map with fixed structure is generated, including the specific coordinates and size of the predicted bounding box, and the confidence of its classification.The loss is calculated with the output and the manually labeled information,and then the regression optimization of related algorithms is carried out for the overall loss after calculating the loss function.YOLOv5 defines the loss function as Eq.(4)below.

    YOLOv5 loss calculation process is mainly divided into two parts,namely regression loss caused by bounding box position (the first item in Eq.(4))and classification loss.The classification loss is divided into three parts, namely, the loss caused by whether the predicted bounding box contains object(the second and third items in Eq.(4))and the loss caused by the predicted category probability of objects(the fourth item in Eq.(4)).

    The first item is the loss of the predicted bounding box position,which respectively represent the loss caused the differences between the predicted bounding box and the ground truth bounding box.represents whether thejthpredicted bounding box in theithgrid is responsible for object prediction.Bis the number of the predicted bounding boxes,and theS2represents the two-dimensional space build by n*n grid.YOLOv5 uses Intersection over Union (IoU)to measure the loss between the ground truth value and the predicted value, specifically, through the value of intersection area of predicted bounding box and ground truth bounding box divided by the union area.Only when one box predictor is responsible to some ground truth box,the model would punish the coordinate error of the box.

    The latter three items are about the loss caused by classification probability with Binary Cross Entropy Loss.represents that there is no object in thejthpredicted bounding box in theithgrid.In the formula,is the Intersection over Union(IoU)calculation between the predicted bounding box and the ground truth bounding box.The second and third item are the confidence loss of different situation.The second item judges whether the grid contains objects.if there are objects in the grid,tends to 1,and the prediction will be more accurate.The third item is used to reduce the proportion of a grid without objects in the overall loss,and the polynomial coefficientis set to 0.

    The last item is the category probability of each bounding box using Binary Cross Entropy.1objirepresents whether the center point of an object is in gridi.Pjiis the probability of the prediction bounding box belongs to classcandis defined similarly to.To sum up,YOLO network assigns different loss weight to different tasks according to their importance.For example,the coordinates of predicted bounding box are the most significant issue,so the weight coefficient of position lossλcoordis set to 5.When it occurs to the boxes without object,the confidence loss weightλnoobjis set to 0.5.And for the boxes with object,the loss weight will be normally 1.

    YOLOv5 draws lessons from the multi-scale feature recognition of SSD and the concept of anchor in two-stage detector,and generates multiple anchors in the same position,so it can generate multiple prediction bounding boxes and improve the recognition ability of small objects.In practical application,the appropriate backbone will be selected as the support according to the needs,so that the detection of the network is faster and more accurate,and at the same time,it can meet the requirements of lightweight.

    3 Network Structure

    In this paper, an On-line Railway Fastener Detection system based on improved YOLOv5 is proposed.The data sampling module transmits the acquired depth map to the server through UDP communication in real time.The proposed system performs fastener detecting and state prediction online,and continues to transmit information such as predicted bounding box,confidence level and object category to the total server.The model structure is shown in Fig.1.

    Figure 1:The improved YOLOv5 network of fastener detection

    As mentioned above,the data image of the input part of this paper is PNG format with the size of 1280*1000.After image scaling and filling, the data of 640*640*1 is obtained and transmit into the Backbone module in Fig.1.There is a down-sampling mechanism in Backbone, which makes the three features with specific sizes generated in Head detector,80*80*27,40*40*27 and 20*20*27,respectively.27 stands for(5+4)*3,where 5 stands for the five-dimensional information of the length and width, center coordinates and confidence of the output detection boxes, 4 stands for the four fastener categories in this paper,and 3 stands for the number of anchors established in YOLOv5.The model proposed will be briefly described from five parts in Fig.1.

    3.1 Input

    3.1.1 Improved Mosaic Data Augmentation

    The Input structure of YOLOv5 adopts Mosaic data enhancement[21]which refers to Yun et al.[22] algorithm.Different from the commonly used datasets, due to engineering standards, all kinds of fasteners have the same size.Therefore,this model has improved the data augmentation model of YOLOv5 to some extent, and cancelled the scaling feature of Mosaic augmentation with different proportion of each picture.As shown in Fig.2, CutMix method randomly put the two samples together,while Mosaic data augmentation method can cut out and scale four original images.

    Figure 2:Schematic diagram of three different data augmentation algorithms

    The improved Mosaic fixed the ratio of each picture to the new picture after being cut randomly.Considering that the size of the retainer target in this mission is about 250*250, it is about 100*100 in the new image after Mosaic augmentation, which is not a small target.Therefore, based on the improvements mentioned above, this paper tries to combine 9 samples together for Mosaic augmentation, instead of collecting only 4 images.Practically speaking, as shown in Fig.2, Mosaic augmentation is to randomly select a point near the center of the new image to determine the x and y axes of the matching process of four images.In the improved method mentioned in this paper,the stitching criterion of nine images is to select two points randomly to determine two groups of x and y axes.

    Network trained with the improved method converges faster and requires less training time.In this method,the information in the nine pictures will be presented in one picture and fed into the network as a training image.This method makes the network structure lighter and the Input part runs faster.While the detection results are not much different from Mosaic data augmentation,and the robustness of the model is better.

    3.1.2 Adaptive Image Scaling

    According to different application fields,there are a variety of image sizes.The commonly used method of computer vision is to scale the original image according to the standard size and then send image after processing to the detection network.In this paper, the images of railway track fastener with the size of 1280*1000 are scaled and filled into 640*640*1.

    3.2 Backbone

    In this paper,the Backbone part of the model mainly uses Focus,CSPDarknet and SPP structures.Fig.3 shows the simplified Backbone structure,and its core content will be briefly described below.

    Figure 3:Simplified structure of YOLOv5 network backbone

    3.2.1 Focus

    The first innovative structure of Backbone is Focus structure,which is mainly used to slice images.In this paper, the original picture of 640*640*1 can be input into a feature map of 320*320*4 after slicing by using YOLOv5s structure.Then, a convolution operation with 32 convolution kernels is performed again,and a feature map of 320*320*4 is obtained.The YOLOv5s model this paper used has Focus structure with size of 3*3 convolution kernel and its output channel is 32.The purpose of Focus module is to reduce the params and FLOPs,so as to achieve the effect of speed.

    3.2.2 CSPDarknet

    The system adopted in this paper follows the Backbone module of YOLOv5 model and uses CSPDarknet as the Backbone to extract abundant information features from the input images.Its core contents are DarknetConv2D, Batch Normalization and Leaky ReLU.CSPNet focuses on solving the problem of repeated gradient information of network optimization existed in other large-scale CNN framework Backbone,and integrates the gradient changes into the feature graph from beginning to end, thus reducing the model parameters and FLOPS values, ensuring the reasoning speed and accuracy,and reducing the model size.

    3.2.3 SSP

    Unlike R-CNN, Spatial Pyramid Pooling (SPP)Network first convolves the whole picture and then gets a feature map.Then, each candidate region is mapped with feature map to get the feature vector of each candidate region.Because these size difference of feature vectors,an SSP layer is added.As shown in Fig.4, the specific application of SPP network in YOLO is to concat the feature maps through the MaxPooling layer of convolution kernel 5,9 and 13 respectively.The SSP layer can receive feature map input of random size,and output feature vectors of fixed size,then pass them to the full connection layer.By using SPP module,Backbone is more effective to increase the receiving range of trunk features.More importantly,SPP module helps Backbone separate the contextual features.

    Figure 4:Spatial Pyramid Pooling network architecture in YOLO network

    3.3 Neck

    Inspired by Feature Pyramid Network (FPN)[23], the Neck continues the structure of the originalYOLOV5 using Path Aggregation Network (PAN)[24].FPN mainly solves the multi-scale problem in object detection.ResNet greatly improves the performance of small object detection after adopting the simple network connection as FPN.As shown in Fig.5,FPN is roughly divided into three parts,namely,bottom-up sampling and top-down up-sampling and lateral connection.Each layer of the network through these three structures will have strong semantic information,and can well meet the requirements of speed and memory.

    Figure 5:The concrete structure of Neck module in YOLOv5 network

    PAN accesses behind FPN network in a bottom-up manner as the bottom-up path augmentation in Fig.5, and transmits the strong positioning features of the lower layer.The FPN layer conveys strong semantic features from top to bottom,while the PAN layer conveys strong positioning features from bottom to top.Specifically,as in FPN backbone,when down-sampling is carried out,new feature map is the element-wise addition result of feature down-sampling and the upper layer’s feature map of new size.The smallest feature map is directly copied from the 20*20 size feature map of the upper layer.The bottom-up path is the same as FPN backbone with up-sampling operation.Consequently,Neck structure aggregates parameters of different detection layers from different trunk layers.

    3.4 Head

    In this paper,YOLOv3 head is used in the same model as YOLOv5,and the improved YOLOv5 model uses Distance-IoU loss function to replace the original IoU loss as in the first item in Eq.(4)mentioned before.

    Distance-IoU is a generalized IoU algorithm, and the expression of DIoU is shown in Eq.(5),in which the latter term is a penalty term.As in Eq.(5),bandbgtrepresent the center point of the predicted bounding box and the center point of the ground truth box respectively, that is, the blue point and the orange point in Fig.6.Whileρ2(·)represents Euclidean distance,andcrepresents the length of black line in Fig.5,that is,the diagonal length of minimum circulated matrix of these two boxes.

    The penalty term is used to minimize the distance between the center points of two bounding boxes.Experiments show that D-IoU has good convergence speed and effect,and it also can be used in NMS calculation,considering not only the overlapping area but also the distance between the center points.When the target frame wraps the prediction frame, the distance between the two frames is directly measured,so DIoU_Loss converges faster.Accordingly,the on-line rail fastener positioning system in this paper defines the non-maximum suppression of the prediction box as DIoU_NMS.

    Figure 6:DIoU loss for bounding box regression

    3.5 FEL Module

    As the dataset in this paper is an imbalanced long-tailed dataset, this paper introduces a readonly Few-shot Example Learning(FEL)module as in Fig.1.Structurally,FEL model is as same as YOLOv5’s original network.The difference between FEL model and original YOLOv5 network is about the input data.As for the rail fasteners in Fig.7, these types of fasteners cannot be detected easily due to the small number of samples during training.Therefore,the improved YOLOv5 model inputs these few-shot data into the FEL module.

    The conventional YOLOv5 network carries out forward propagation,which propagates forward and backward propagations in FEL module with few shot samples as input.FEL module synchronizes the residuals learned from few shot samples to the regular YOLOv5 network for weight update.In the testing and detecting stages,only the conventional modules run,which does not affect the testing speed of the improved network.

    Figure 7:Few-shot examples in rail fastener dataset

    4 Experiment Results

    In this paper,the threshold of IoU is chosen as 0.8,and the experiment is running on two NVIDIA GTX1080Ti GPUs.In this paper, inference time and mean Average Precision (mAP)are used to measure network performance.

    One conventional rail data after brightness adjustment is shown in Fig.8.It is clearly that the rail edge of most data is in the middle of the picture,and the fasteners are symmetrically distributed beside the rail edge.It can be seen from Fig.9a that the normalized coordinates of the center of the detection bounding boxes predicted by the model are distributed on both sides of the central track as mentioned above, while the size information of the detection bounding box (normalized length and width)is shown in Fig.9b,showing a large-scale centralized and sporadic abnormal size distribution.

    Figure 8:A railway data image in JPG format after brightness adjustment

    In this paper, four different advanced algorithms are selected for comparison, which are Histogram of Oriented Gradient (HOG)+SVM, Faster R-CNN, MobileNet-SSD, MobileDets and original YOLOv5.To break through the speed bottleneck of region proposal,Faster R-CNN directly uses CNN to generate region proposal by introducing RPN, and shares the convolution layer with CNN in the second stage.HOG is a kind of edge feature, which makes use of the orientation and intensity information of the edge, and is then widely used in visual target detection such as vehicle detection and license plate detection.The combination of HOG and SVM classifier is a common method for locating and detecting rail fasteners in China.The core of SSD method is to use a small convolution filter to predict the category scores and position offsets of a set of default bounding boxes fixed on the feature map.MobileDets [25] is a lightweight object detection algorithm that can be deployed on mobile devices,its most prominent advantage is that it reduced FLOPs and Params to a new level.

    Figure 9:The distribution of predicted detection boxes in the dataset images

    When the model proposed in this paper is applied to compare with baseline in the rail fastener dataset, as shown in Tab.1, the model proposed is superior in two different verification criteria with higher discrimination precision and shorter sample recognition time.In addition to the model proposed in this paper, the original YOLOv5 has obvious advantages in the inference time, while Faster R-CNN confirms its detection accuracy as a two-stage detector with the second highest mAP.MobileDets, known for its lightness, is indeed superior to improved YOLOv5 network proposed in inference time,but its mAP is slightly inferior to the model in this paper.

    Table 1: Fastener detection performance of the improved YOLOv5 network

    In order to clarify the influence of improvements proposed in this paper,there are several ablation experiments results.As shown in Tab.1,the three improvements proposed in this paper can improve the mAP of the model in rail Fastener data set to a certain extent.As shown in Tab.1,the most significant method to promote mAP is the FEL Model proposed in this paper, which has made great progress(0.7%)by increasing the network’s attention to the few-shot examples.DIoU can achieve better detection accuracy (0.1%)in fastener detection than GIoU.Improved Mosaic data augmentation is also helpful to mAP improvement, but its main purpose is to accelerate the convergence speed of the network in training, which will be verified in future work.This paper improves the structure of YOLOv5 to adapt to the rail fastener dataset, and then receives the best performance than other advanced algorithms in the same period.

    5 Conclusion

    In this paper, a rail fastener detection system based on improved YOLOv5 is proposed.The FEL model is introduced into YOLOv5 network structure, which enhances the detecting ability of the network for the few-shot samples.Compared with other similar algorithms,the Input and Head modules are improved on the basis of YOLOv5, and better results are obtained.For the specific application scenario[27],the model in this paper gives the whole rail anomaly detection system more relaxed judgment time.This model makes the “sampling-detection-recognition-warning”cycle of a single picture complete before the next picture is collected.The sampling time of a single picture in this paper is about 53 ms,and the shorter detection time means the establishment of a more mature so-called real-time rail fastener anomaly detection system.

    Funding Statement:This work was supported by the National Natural Science Foundation of China(61871046,SM,http://www.nsfc.gov.cn/).

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    国产69精品久久久久777片| 久久精品国产亚洲av涩爱 | 亚洲精品日韩av片在线观看 | 一二三四社区在线视频社区8| 又粗又爽又猛毛片免费看| 好看av亚洲va欧美ⅴa在| 在线观看舔阴道视频| x7x7x7水蜜桃| 最近在线观看免费完整版| 日韩有码中文字幕| 一本久久中文字幕| 日本 欧美在线| 亚洲欧美日韩高清专用| 国内精品久久久久久久电影| 国产精品久久久久久久久免 | 日本精品一区二区三区蜜桃| 一本久久中文字幕| 精品久久久久久久久久免费视频| 亚洲精品在线美女| 少妇的逼好多水| 亚洲五月天丁香| 一卡2卡三卡四卡精品乱码亚洲| 国产午夜精品论理片| 神马国产精品三级电影在线观看| 黄色视频,在线免费观看| 99精品久久久久人妻精品| 久久人妻av系列| 九九在线视频观看精品| 日韩欧美国产一区二区入口| 国产亚洲精品一区二区www| 久久香蕉国产精品| 国产91精品成人一区二区三区| 午夜福利欧美成人| 国产一区二区在线观看日韩 | 国产亚洲欧美在线一区二区| 欧美一区二区国产精品久久精品| 深爱激情五月婷婷| 性色av乱码一区二区三区2| 亚洲av熟女| 一区二区三区免费毛片| tocl精华| 欧美3d第一页| 淫妇啪啪啪对白视频| 免费无遮挡裸体视频| 最新中文字幕久久久久| 在线观看午夜福利视频| 啦啦啦韩国在线观看视频| 人妻夜夜爽99麻豆av| 亚洲熟妇中文字幕五十中出| 可以在线观看的亚洲视频| 日日摸夜夜添夜夜添小说| 精品国产亚洲在线| 国产三级黄色录像| 国产麻豆成人av免费视频| 一级毛片高清免费大全| 国产真实伦视频高清在线观看 | 久久久久久久午夜电影| 夜夜夜夜夜久久久久| 亚洲午夜理论影院| 亚洲精品一区av在线观看| 天天添夜夜摸| 欧美成人免费av一区二区三区| 国产免费男女视频| 色噜噜av男人的天堂激情| 啪啪无遮挡十八禁网站| 国产成人av激情在线播放| av在线蜜桃| 99在线人妻在线中文字幕| 日韩亚洲欧美综合| 9191精品国产免费久久| 中文字幕人成人乱码亚洲影| 国产精品一区二区三区四区久久| 免费看美女性在线毛片视频| 淫秽高清视频在线观看| 国产91精品成人一区二区三区| 免费高清视频大片| 中文字幕熟女人妻在线| 日韩av在线大香蕉| 在线观看午夜福利视频| 成人av在线播放网站| 国产成人av教育| 首页视频小说图片口味搜索| 欧美极品一区二区三区四区| 色尼玛亚洲综合影院| 亚洲国产日韩欧美精品在线观看 | 成年人黄色毛片网站| 久久久久久久亚洲中文字幕 | 亚洲精品成人久久久久久| 国产真实伦视频高清在线观看 | 级片在线观看| 国产一区二区三区视频了| www.999成人在线观看| 在线观看免费午夜福利视频| 日韩国内少妇激情av| 国内揄拍国产精品人妻在线| tocl精华| 搞女人的毛片| 国产精品免费一区二区三区在线| 麻豆国产av国片精品| 国产精品电影一区二区三区| 国产真实伦视频高清在线观看 | 俄罗斯特黄特色一大片| 亚洲不卡免费看| 一级黄色大片毛片| www.999成人在线观看| 女警被强在线播放| www日本在线高清视频| av欧美777| 最后的刺客免费高清国语| 制服人妻中文乱码| 精品人妻1区二区| 一夜夜www| 久久伊人香网站| 精品福利观看| 床上黄色一级片| 老司机午夜十八禁免费视频| av福利片在线观看| 国产精品日韩av在线免费观看| av片东京热男人的天堂| 岛国在线观看网站| 老汉色∧v一级毛片| 国产欧美日韩一区二区三| 免费观看的影片在线观看| 欧美最黄视频在线播放免费| 中文字幕人成人乱码亚洲影| 欧美激情在线99| 精品无人区乱码1区二区| xxxwww97欧美| 色尼玛亚洲综合影院| 成人精品一区二区免费| 成人高潮视频无遮挡免费网站| 麻豆久久精品国产亚洲av| 97人妻精品一区二区三区麻豆| 亚洲国产欧洲综合997久久,| 99国产精品一区二区蜜桃av| 欧美色视频一区免费| 亚洲人成网站在线播| 啦啦啦韩国在线观看视频| 欧美日韩瑟瑟在线播放| 国产精品久久久久久亚洲av鲁大| 美女黄网站色视频| 欧美成狂野欧美在线观看| 人妻丰满熟妇av一区二区三区| 成人国产一区最新在线观看| 亚洲精品影视一区二区三区av| 观看免费一级毛片| 人人妻人人澡欧美一区二区| 黄片大片在线免费观看| 99久久精品一区二区三区| 欧美+亚洲+日韩+国产| 国产成年人精品一区二区| 久99久视频精品免费| 好看av亚洲va欧美ⅴa在| 久久精品国产自在天天线| 色精品久久人妻99蜜桃| 18+在线观看网站| 91av网一区二区| 中国美女看黄片| 脱女人内裤的视频| 2021天堂中文幕一二区在线观| 日日干狠狠操夜夜爽| 两个人看的免费小视频| 日韩欧美国产一区二区入口| 一边摸一边抽搐一进一小说| 成人永久免费在线观看视频| 宅男免费午夜| 两人在一起打扑克的视频| 成人国产综合亚洲| 色吧在线观看| 很黄的视频免费| 岛国在线观看网站| 久久久久久国产a免费观看| 亚洲第一欧美日韩一区二区三区| 麻豆久久精品国产亚洲av| 深爱激情五月婷婷| 淫妇啪啪啪对白视频| 欧美又色又爽又黄视频| 国产成人影院久久av| 日日摸夜夜添夜夜添小说| 欧美av亚洲av综合av国产av| 好男人在线观看高清免费视频| 国产亚洲精品一区二区www| 丰满的人妻完整版| av片东京热男人的天堂| 午夜福利在线观看免费完整高清在 | 久久这里只有精品中国| 狂野欧美白嫩少妇大欣赏| 日本熟妇午夜| 国产黄a三级三级三级人| www国产在线视频色| 麻豆成人av在线观看| 九色成人免费人妻av| 久久草成人影院| 在线国产一区二区在线| 亚洲成人久久性| 精品国产三级普通话版| 91在线精品国自产拍蜜月 | 久久欧美精品欧美久久欧美| 国产成人福利小说| 国产欧美日韩一区二区精品| 真实男女啪啪啪动态图| 免费av不卡在线播放| 男女做爰动态图高潮gif福利片| 男女之事视频高清在线观看| 亚洲成a人片在线一区二区| 亚洲国产欧美网| 欧美日本亚洲视频在线播放| 免费观看的影片在线观看| 久久精品国产综合久久久| 色噜噜av男人的天堂激情| tocl精华| 国产老妇女一区| 亚洲无线在线观看| 精品久久久久久成人av| 国产高清视频在线播放一区| 欧美黑人巨大hd| 国产欧美日韩精品一区二区| 丰满的人妻完整版| 亚洲av不卡在线观看| 午夜福利在线观看免费完整高清在 | 国产精品 国内视频| 国产色婷婷99| 日韩欧美一区二区三区在线观看| 99久久精品国产亚洲精品| 国产亚洲精品久久久久久毛片| 久久久国产精品麻豆| 欧美极品一区二区三区四区| 国产乱人视频| 男插女下体视频免费在线播放| 内射极品少妇av片p| 99精品欧美一区二区三区四区| 少妇的丰满在线观看| 久久精品91蜜桃| а√天堂www在线а√下载| 99热这里只有是精品50| 国产蜜桃级精品一区二区三区| 熟女人妻精品中文字幕| 全区人妻精品视频| 亚洲aⅴ乱码一区二区在线播放| 免费一级毛片在线播放高清视频| 久久中文看片网| 国产视频内射| ponron亚洲| 免费搜索国产男女视频| 特级一级黄色大片| 久久久国产精品麻豆| 午夜久久久久精精品| 国产男靠女视频免费网站| 亚洲最大成人手机在线| 色老头精品视频在线观看| 欧美日韩中文字幕国产精品一区二区三区| 色尼玛亚洲综合影院| 国产三级在线视频| 亚洲人与动物交配视频| av在线蜜桃| 免费搜索国产男女视频| 伊人久久精品亚洲午夜| 舔av片在线| 亚洲av成人精品一区久久| 成人鲁丝片一二三区免费| 不卡一级毛片| 99久久无色码亚洲精品果冻| 亚洲国产日韩欧美精品在线观看 | 免费一级毛片在线播放高清视频| 亚洲18禁久久av| 国产v大片淫在线免费观看| 久久精品人妻少妇| 亚洲av免费高清在线观看| 免费无遮挡裸体视频| 噜噜噜噜噜久久久久久91| 久久久久久久亚洲中文字幕 | 一个人观看的视频www高清免费观看| 免费在线观看日本一区| 久久久成人免费电影| 两人在一起打扑克的视频| 国产一区二区三区视频了| 亚洲精华国产精华精| 久久久久免费精品人妻一区二区| 琪琪午夜伦伦电影理论片6080| 中文字幕人妻熟人妻熟丝袜美 | 精品一区二区三区av网在线观看| www日本黄色视频网| 在线观看免费午夜福利视频| 悠悠久久av| 久久天躁狠狠躁夜夜2o2o| 又黄又粗又硬又大视频| 亚洲狠狠婷婷综合久久图片| 精品久久久久久久人妻蜜臀av| 久久久久久久久大av| 色av中文字幕| 搡老岳熟女国产| 国产日本99.免费观看| 亚洲电影在线观看av| 丝袜美腿在线中文| 免费看光身美女| 噜噜噜噜噜久久久久久91| 夜夜躁狠狠躁天天躁| 99久久精品热视频| 亚洲不卡免费看| 神马国产精品三级电影在线观看| 中亚洲国语对白在线视频| 亚洲第一欧美日韩一区二区三区| 精品熟女少妇八av免费久了| 看黄色毛片网站| 午夜福利欧美成人| 免费看十八禁软件| 中文资源天堂在线| 色吧在线观看| 国产精品美女特级片免费视频播放器| 中文在线观看免费www的网站| 国产精品久久久久久久电影 | 桃色一区二区三区在线观看| 欧美不卡视频在线免费观看| 国产精品一区二区三区四区久久| 精品无人区乱码1区二区| 亚洲色图av天堂| 五月玫瑰六月丁香| 精品无人区乱码1区二区| h日本视频在线播放| 欧美黑人欧美精品刺激| 一级黄片播放器| 国产精品98久久久久久宅男小说| 深夜精品福利| 国产中年淑女户外野战色| 伊人久久精品亚洲午夜| 国产精品一及| 三级毛片av免费| 国产精品亚洲一级av第二区| 午夜免费激情av| 国产精品99久久99久久久不卡| 成人午夜高清在线视频| 亚洲色图av天堂| 男女之事视频高清在线观看| 两性午夜刺激爽爽歪歪视频在线观看| 制服丝袜大香蕉在线| 午夜免费男女啪啪视频观看 | 国产精品乱码一区二三区的特点| 一个人免费在线观看电影| 亚洲国产欧美网| 国产精品三级大全| 亚洲av中文字字幕乱码综合| 女生性感内裤真人,穿戴方法视频| 国产色婷婷99| 九色成人免费人妻av| 天堂网av新在线| 欧美午夜高清在线| 日韩欧美国产在线观看| 深夜精品福利| 日韩大尺度精品在线看网址| 亚洲国产欧美人成| 国产激情偷乱视频一区二区| 亚洲熟妇熟女久久| 99热精品在线国产| 亚洲国产欧美人成| 9191精品国产免费久久| 俄罗斯特黄特色一大片| 一个人看视频在线观看www免费 | 成人特级黄色片久久久久久久| 久久精品91无色码中文字幕| 免费看光身美女| 亚洲精品成人久久久久久| 一区二区三区高清视频在线| 麻豆国产97在线/欧美| www国产在线视频色| 嫩草影视91久久| 成人av在线播放网站| 悠悠久久av| 久久久色成人| 伊人久久大香线蕉亚洲五| 日韩有码中文字幕| 日韩中文字幕欧美一区二区| 欧美一区二区国产精品久久精品| 欧美丝袜亚洲另类 | 一卡2卡三卡四卡精品乱码亚洲| 啪啪无遮挡十八禁网站| 日本熟妇午夜| 欧美在线一区亚洲| 国产视频一区二区在线看| 欧美成人一区二区免费高清观看| 亚洲精品一卡2卡三卡4卡5卡| 夜夜爽天天搞| 国产精品久久久久久久电影 | 亚洲 国产 在线| 亚洲人与动物交配视频| 99国产极品粉嫩在线观看| 午夜激情福利司机影院| 亚洲精品在线观看二区| 女人被狂操c到高潮| 欧美日本亚洲视频在线播放| 亚洲av不卡在线观看| 日韩 欧美 亚洲 中文字幕| 亚洲久久久久久中文字幕| 国产精品香港三级国产av潘金莲| 老师上课跳d突然被开到最大视频 久久午夜综合久久蜜桃 | 一二三四社区在线视频社区8| 国产成人av激情在线播放| 久久中文看片网| 黑人欧美特级aaaaaa片| 国产黄a三级三级三级人| 美女被艹到高潮喷水动态| 久久久久久久午夜电影| 香蕉久久夜色| 女人十人毛片免费观看3o分钟| 搡老熟女国产l中国老女人| 免费电影在线观看免费观看| 国产v大片淫在线免费观看| 午夜福利高清视频| 噜噜噜噜噜久久久久久91| 国产精品日韩av在线免费观看| 狂野欧美激情性xxxx| 2021天堂中文幕一二区在线观| 热99re8久久精品国产| 变态另类成人亚洲欧美熟女| 久久精品影院6| xxx96com| 97超视频在线观看视频| 亚洲人成伊人成综合网2020| 亚洲成人免费电影在线观看| 国产一区二区在线观看日韩 | 欧美乱色亚洲激情| 19禁男女啪啪无遮挡网站| 国产色爽女视频免费观看| 一a级毛片在线观看| 国产精品久久久久久久久免 | 欧美三级亚洲精品| 欧美在线一区亚洲| 成人高潮视频无遮挡免费网站| 精品乱码久久久久久99久播| xxxwww97欧美| 一区二区三区高清视频在线| 欧美在线一区亚洲| 又黄又爽又免费观看的视频| 舔av片在线| 国产激情偷乱视频一区二区| 久久这里只有精品中国| 欧美3d第一页| 久久久久亚洲av毛片大全| 美女cb高潮喷水在线观看| 少妇高潮的动态图| 免费看美女性在线毛片视频| 亚洲在线观看片| 日本 av在线| 99久国产av精品| 99国产极品粉嫩在线观看| 久99久视频精品免费| 熟女电影av网| 夜夜爽天天搞| 又黄又爽又免费观看的视频| 嫩草影院精品99| 中文在线观看免费www的网站| 久久精品夜夜夜夜夜久久蜜豆| 18禁裸乳无遮挡免费网站照片| a级一级毛片免费在线观看| 男女做爰动态图高潮gif福利片| 国产亚洲欧美98| 午夜两性在线视频| 国产精品 国内视频| 国产精品乱码一区二三区的特点| 中文字幕熟女人妻在线| 精品久久久久久久人妻蜜臀av| 床上黄色一级片| 中文在线观看免费www的网站| 国产精品久久久人人做人人爽| 欧美中文综合在线视频| 搡女人真爽免费视频火全软件 | 黄色视频,在线免费观看| 级片在线观看| 欧美午夜高清在线| 法律面前人人平等表现在哪些方面| 国产野战对白在线观看| 久久久久性生活片| 国产高潮美女av| 夜夜爽天天搞| 欧美乱码精品一区二区三区| 19禁男女啪啪无遮挡网站| 久久久精品大字幕| 美女黄网站色视频| 中文在线观看免费www的网站| 亚洲美女视频黄频| 国产精品久久久久久久久免 | 亚洲成人免费电影在线观看| 性色avwww在线观看| 黄色成人免费大全| 69av精品久久久久久| 内射极品少妇av片p| 黄色视频,在线免费观看| 亚洲中文日韩欧美视频| 欧美一区二区国产精品久久精品| 级片在线观看| 99久久精品热视频| 亚洲男人的天堂狠狠| av在线蜜桃| 久99久视频精品免费| av在线蜜桃| 国产淫片久久久久久久久 | 岛国在线观看网站| 成人午夜高清在线视频| 老司机深夜福利视频在线观看| 欧美性猛交╳xxx乱大交人| 成人永久免费在线观看视频| 极品教师在线免费播放| 亚洲中文日韩欧美视频| 国产精品爽爽va在线观看网站| 国产av麻豆久久久久久久| 欧美黑人欧美精品刺激| 观看免费一级毛片| 成人av一区二区三区在线看| 可以在线观看的亚洲视频| 嫩草影视91久久| 国产视频内射| 3wmmmm亚洲av在线观看| 亚洲男人的天堂狠狠| 婷婷六月久久综合丁香| 精品久久久久久成人av| 久久久久久久久大av| 久久久久久久亚洲中文字幕 | 国产精品三级大全| 丁香六月欧美| 久久久久久大精品| 少妇的逼水好多| 成人特级av手机在线观看| 成人av在线播放网站| 久久久久久久精品吃奶| 欧美在线一区亚洲| 国产在线精品亚洲第一网站| 国产野战对白在线观看| 久久久国产成人免费| 国产探花极品一区二区| 国产成人福利小说| 国产又黄又爽又无遮挡在线| 国产一区二区在线观看日韩 | 国产成人a区在线观看| 久久精品夜夜夜夜夜久久蜜豆| 色av中文字幕| 热99在线观看视频| 午夜视频国产福利| 69av精品久久久久久| 12—13女人毛片做爰片一| 特大巨黑吊av在线直播| 亚洲av免费在线观看| 精品久久久久久久人妻蜜臀av| av视频在线观看入口| 精品国产超薄肉色丝袜足j| 国产美女午夜福利| 热99在线观看视频| 可以在线观看毛片的网站| 亚洲av熟女| 亚洲成人久久性| 校园春色视频在线观看| 免费高清视频大片| 亚洲最大成人中文| 给我免费播放毛片高清在线观看| 午夜影院日韩av| 国产色爽女视频免费观看| 亚洲狠狠婷婷综合久久图片| 在线观看免费视频日本深夜| 精品无人区乱码1区二区| 他把我摸到了高潮在线观看| 久久久久久久久中文| 亚洲精品久久国产高清桃花| 久久精品人妻少妇| 日本在线视频免费播放| 成年女人永久免费观看视频| 精品福利观看| 熟女人妻精品中文字幕| 热99在线观看视频| 精品久久久久久久人妻蜜臀av| 18美女黄网站色大片免费观看| 国内少妇人妻偷人精品xxx网站| 午夜福利成人在线免费观看| 国产三级黄色录像| 激情在线观看视频在线高清| 美女 人体艺术 gogo| 日本 av在线| 午夜免费男女啪啪视频观看 | 久久久久久九九精品二区国产| 特大巨黑吊av在线直播| 国产亚洲欧美在线一区二区| 99久久九九国产精品国产免费| 操出白浆在线播放| 男女下面进入的视频免费午夜| 真人做人爱边吃奶动态| 亚洲黑人精品在线| 成人国产一区最新在线观看| 日韩欧美在线二视频| 淫秽高清视频在线观看| 国产亚洲精品久久久久久毛片| 亚洲人成伊人成综合网2020| www.色视频.com| 怎么达到女性高潮| 久久久久久大精品| 91在线观看av| 叶爱在线成人免费视频播放| 香蕉久久夜色| 免费看美女性在线毛片视频| 色老头精品视频在线观看| 日本一二三区视频观看| 亚洲成人精品中文字幕电影| 国产探花在线观看一区二区| 成人午夜高清在线视频| 97碰自拍视频| 亚洲av成人精品一区久久| 亚洲欧美一区二区三区黑人| 国产高清视频在线播放一区| 午夜福利成人在线免费观看| 中文字幕熟女人妻在线| 午夜两性在线视频| 一卡2卡三卡四卡精品乱码亚洲| 床上黄色一级片| 麻豆成人av在线观看| 免费在线观看成人毛片| 美女高潮喷水抽搐中文字幕| 日韩精品中文字幕看吧| 国产精品美女特级片免费视频播放器| 免费av毛片视频|