• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Semantic segmentation method of road scene based on Deeplabv3+ and attention mechanism

    2021-12-21 13:34:26BAIYanqiongZHENGYufuTIANHong

    BAI Yanqiong, ZHENG Yufu, TIAN Hong

    (School of Electronic and Information Engineering, Lanzhou Jiaotong University, Lanzhou 730070, China)

    Abstract: In the study of automatic driving, understanding the road scene is a key to improve driving safety. The semantic segmentation method could divide the image into different areas associated with semantic categories in accordance with the pixel level, so as to help vehicles to perceive and obtain the surrounding road environment information, which would improve driving safety. Deeplabv3+ is the current popular semantic segmentation model. There are phenomena that small targets are missed and similar objects are easily misjudged during its semantic segmentation tasks, which leads to rough segmentation boundary and reduces semantic accuracy. This study focuses on the issue, based on the Deeplabv3+ network structure and combined with the attention mechanism, to increase the weight of the segmentation area, and then proposes an improved Deeplabv3+ fusion attention mechanism for road scene semantic segmentation method. First, a group of parallel position attention module and channel attention module are introduced on the Deeplabv3+ encoding end to capture more spatial context information and high-level semantic information. Then, an attention mechanism is introduced to restore the spatial detail information, and the data shall be normalized in order to accelerate the convergence speed of the model at the decoding end. The effects of model segmentation with different attention-introducing mechanisms are compared and tested on CamVid and Cityscapes datasets. The experimental results show that the mean Intersection over Unons of the improved model segmentation accuracies on the two datasets are boosted by 6.88% and 2.58%, respectively, which is better than using Deeplabv3+. This method does not significantly increase the amount of network calculation and complexity, and has a good balance of speed and accuracy.

    Key words: autonomous driving; road scene; semantic segmentation; Deeplabv3+; attention mechanism

    0 Introduction

    The technology of automatic driving requires vehicles to simulate drivers understanding the relationship of the traffic participants and respond to complex traffic environments, such as pedestrians, vehicles, obstacles, etc. Semantic segmentation is one of the most commonly used methods in the study of automatic driving visual perception task, which can divide various interest target areas of the image at the pixel level, label different objects of the image according to semantic categories, and then obtain an image with pixel semantic annotation[1-2].

    In recent years, segmentation accuracy and speed of segmentation methods based on deep learning have been significantly improved. In 2015, a Fully Convolutional Network (FCN)[3]was proposed, which implemented an end-to-end segmentation method that supported to input image in any size. Since that, the method of applying convolutional neural network to semantic segmentation has become the mainstream and made good effects[4]. In order to reduce the loss of spatial information caused by down-sampling and pooling operations to affect the segmentation effect of the model, the encoding and decoding network model was proposed in Ref.[5] that used a decoder to recover the detailed information of image features and perform feature extraction. For example, the SegNet network proposed in Ref.[6] and the U-Net network proposed in Ref.[7], both of them used the encoder and decoder structure to capture rich spatial information. DeepLabv3+[8]was proposed in Ref.[6] that added a simple and effective decoding module to the DeepLabv3 network[9], which allowed it to capture sufficient spatial information in the shallow layer, helped the model to recover the target details, and had obtained a good segmentation effect.

    In computer vision tasks, the attention mechanism (AM)[10], i.e. a problem-solving method, is proposed by imitating human attention, which has been applied in many tasks. For example, Ref.[11] introduced the AM into FCN, and proposed a Dual Attention Network (DANet), which made ideal segmentation effects in scene segmentation tasks. In addition, the AM can be contained behind one or more layers as an operator, and recognize important features in the image.

    The currently popular Deeplabv3+ semantic segmentation model captures high-level semantic information through an encoder and recovers spatial detail information by a decoder, which has relatively a good performance about segmenting complex and diverse urban road scene images. However, there are some problems in local details, such as rough boundary segmentation, small objects being ignored and misjudgment of objects with similar shapes. In order to solve the above problems, this study uses Deeplabv3+ as the basic network structure, introduces the AM, and proposes a road scene image semantic segmentation method that integrates the attention mechanism. The attention weight is calculated and assigned through the AM to guide feature learning. Among them, the position attention module (PAM) can capture the spatial dependence between any two positions in the feature map, so that any two similar position features can be improved mutually. Moreover, the similarity affects the allocation of attention weights, and then the characteristics of specific positions in the feature map are determined by weighting and summing all the position characteristics. The channel attention module (CAM) can capture the correlation of features on any two channels, and improve the characteristics of each channel feature by weighted sum. Then, the outputs of the two attention modules are integrated to enhance the feature representation of the feature map. Finally, the Batch normalization (BN)[9]operation is used to normalize the data in order to achieve refined output results.

    1 Network achitecture

    1.1 Deeplabv3+ network junction

    Deeplabv3+ is a typical network framework for semantic segmentation, which is developed on the basis of Deeplabv1-3. First of all, Deeplabv1[12]uses atrous convolution to clarify and control the resolution of feature responses in deep convolutional neural networks. It uses the method of reducing down-sampling operations and increasing the network’s receptive field to obtain dense feature maps, but its performance for multi-scale segmentation is poor. In order to cover the shortcomings of Deeplabv1, Deeplabv2[13]uses Atrous Spatial Pyramid Pooling (ASPP) structure with multiple sampling rates and effective field filters to segment targets of multiple scales. To improve segmentation accuracy, Deeplabv3 uses image-level features to enhance the ASPP module, which captures longer distance information, and merges global context information. In addition, it also introduces the BN operation to facilitate training. Furthermore, Deeplabv3+ adds a simple and efficient decoder module on Deeplabv3 to optimize target boundary segmentation results with end-to-end training. Compared with Deeplabv3, Deeplabv3+ in the encoder and decoder structure can control arbitrarily the resolution of extracted features through atrous convolution to improve image segmentation effect and achieve a balance between speed and accuracy. Besides, the DeepLabv3+ network has achieved ideal results on many datasets, such as PASCAL VOC2012, Cityscapes and other public datasets. The network structure is shown in Fig.1.

    Fig.1 Deeplabv3+ network structure

    The encoder part mainly uses Xception and ASPP modules for feature extraction. Xception is a DCNN network containing input and output. In order to increase the receptive field of the network, the ASPP module firstly performs 1×1 convolutional compression of the feature map, and uses 3×3 atrous convolution with expansion rates of 6, 12, and 18 to realize learning multi-scale features at the same time. It can not only reduce the down-sampling operation, but also obtain more context information by using the global average pooling layer to capture the global information. So that the target boundary information of the segmentation feature map can be captured, and finally the segmentation of multi-scale targets is achieved and the segmentation effect is improved. Then the feature maps output by the ASPP module would be integrated, and the features are compressed by a 1×1 convolution operation. In the end, the high-level feature maps would be output.

    In the decoder part, in order to avoid the low-level features containing more channels than the output encoding features, 1×1 convolution operation is adopted for the features output by the Xception module to reduce the number of low-level feature channels. Performing bilinear interpolation up-sampling operation on the high-level features output by the encoding end, and it is integrated with the low-level features after the 1×1 convolution operation to strengthen the recovery of the boundary information of the target part. Then, 3×3 convolution operation is performed on it as input to restore the details and spatial information of the feature map. After bilinear interpolation up-sampling operation, the final segmentation image would be obtained.

    1.2 Attention mechanism

    AM[11]can be understood as a resource allocation mechanism that redistributes the resources according to the importance of the attention object. In computer vision, the resource to be allocated by the AM refers to the weight, which is obtained by the high-level feature map containing rich semantic information and the low-level feature map containing global context information. This study mainly introduces the PAM and the CAM to better capture the context information of the channel and space dimensions, and then improve the segmentation effect of the model.

    1.2.1 Position attention module

    PAM[11]selectively aggregates the features of each position according to the weighted sum of the features of all positions. The feature between two points has the characteristics that similar features are related to each other no matter how far apart, so linking the correlation between any two features can enhance the expression of their respective feature. It is very important in scene understanding tasks to enhance and distinguish feature representation. The PAM can establish rich context relationships on local features, and encode more context information as local features, thereby enhance their representation capabilities. The work flow of the PAM is shown in Fig.2.

    Fig.2 Position attention module

    As shown in Fig.2, the local feature matrixMis obtained through the backbone network,M∈RC×H×W. First, after convolution operation on matrixM, two new feature matricesXandYare generated, respectively, {X,Y}∈RC×H×W. Then transforming the dimensions of the matricesXandY, {X,Y}∈RC×N, whereN=H×Wis the number of pixels. Finally, the transposed matrixXTof the matrixXand the matrixYare matrix multiplied, and the softmax layer is used to calculate the space attention graphS,S∈RN×N. The calculation process is expressed as[11]

    (1)

    wheresjiis the impact factor of thei-th position on thej-th position. The more similar the feature representation between the two positions, the correlation and the effect have also increased.Xiis thei-th positional element of the matrixX,X∈RC×N, andYjis thej-th positional element of the matrixY,Y∈RC×N.

    At the same time, the matrixM∈RC×H×Wis subjected to convolution operation to obtain a new feature matrixZ,Z∈RC×H×W, and the dimension is converted toZ∈RC×N. Then the transformed matrixZand the transposed matrixSTof matrixSare performed matrix multiplication, and the result dimension is converted toRC×H×W. The converted result is multiplied by the scale parameterαand summed with the matrixMto obtain theαis final matrixP,P∈RC×H×W.The initial value ofαis 0, and the learning parameters of the distribution weights is gradually learned. The final output is expressed as[11]

    (2)

    whereZiis thei-th positional element of the matrixZ. It can be seen from Eq.(2) that the output featurePat each position, besides its original feature. It also aggregates the features of all positions, so that the network will not lose the original features information even if it has not learned new features. According to the spatial attention mapS, the PAM can selectively aggregate the context information, capture the global information, and use the similarity between the semantic features to improve the intra-class compact and semantic consistency.

    1.2.2 Channel attention module

    CAM[11]is a reallocation of resources among convolutional channels. Each convolution kernel of each layer of convolutional network corresponds to a feature channel. The channel map of each high-level feature can be regarded as a class-specific response, and different semantic responses are related to each other. The interdependence between each channel mapping is used to learn the correlation of each channel feature map, and improve the specific semantic feature representation. Therefore, through introducing the CAM, the interdependence among channel models can be clarified, and the feature maps with less correlation can be adjusted according to the degree of dependence, and then more useful information will be obtained. The work flow of the CAM is shown in Fig.3.

    Fig.3 Channel attention module

    The CAM first converts the feature matrixM,M∈RC×H×W, the dimension toM∈RC×N, whereN=H×Wis the number of pixels. Then the converted matrixMand its transposed matrixMTare subjected to matrix multiplication, and the channel attention mapFis obtained through the softmax layerF∈RC×C. The calculation process is expressed as[11]

    (3)

    fjiis the impact factor of thei-th channel on thej-th channel. The transposed matrixFTof the matrixFis subjected to matrix multiplication with the transformed matrixM, and the dimension of the result is converted toRC×H×W. Then the transformed result is multiplied by the proportional parameterβ, and finally summed with the matrixMto obtain the final output resultP,P∈RC×H×W, where β is the proportional parameter for gradually learning weights from 0. The final output is expressed as[11]

    (4)

    It can be seen from Eq.(4) that the final result featurePof the corresponding channel of each convolution kernel is the integration of the features of all channels with the local features obtained from the original backbone network. The spatial attention module can use the correlation and dependence between the spatial information of all channels to readjust the feature map to enhance the discriminability of features.

    1.3 Improved Deeplabv3+ network structure

    In order to increase the receptive field of the network, the Deeplabv3+ network uses atrous convolution with different expansion rates to replace the down-sampling operation. However, the atrous convolution is discrete sampling on the feature map, and the lack of correlation between the sampling points makes it produce a grid effect[14-15]. Discrete sampling will achieve better results for obtaining the semantic information of large targets, but some local information will still be lost. For small targets, when the expansion rate is set to be large, the learned features lack relevance due to the large interval, and then the wrong feature information may be learned, which makes segmentation boundaries unclear and misjudge.

    To reduce the impact of grid effect on the segmentation accuracy, a group of parallel PAM and CAM are added after the Xception module of the Deeplabv3+ encoding end. These modules can capture global context information and high-level semantic information to enhance the presentation of the output feature map. After the output results of the two attention modules are converted by convolution, the element summation is performed to complete fusion of the output features of the two attention modules. To further refine the output results, PAM and CAM are added to the convolutional layer after the 3×3 convolution operation on the decoding end. These attention modules are used to simulate the interdependence of spatial and channel dimension semantic information, to capture global context information and enhance detailed features. Then, the original size feature map is restored through a 4-fold linear up-sampling operation. Finally, the refined segmentation result will be output. The improved Deeplabv3+ network model structure is shown in Fig.4.

    Fig.4 Deeplabv3+ network structure with attention mechanism

    The feature maps in Fig.4 are processed by PAM and CAM, respectively, and their features are integrated together after convolution conversion, and then the result is used as the input for subsequent operations to obtain a more refined segmentation result. In order to further refine the output results, BN operations on the output results of each attention module can speed up training, improve network generalization capabilities, accelerate model convergence and prevent gradients from disappearing or exploding.

    2 Results and discussion

    2.1 Experimental parameters

    2.1.1 Datasets and experimental environment configuration

    The CamVid dataset[16]is a public dataset with rich objects and fine semantic annotation. Most of the data pictures are taken from the perspective of driving cars with cameras, and it contains a large number of high-quality and high-resolution color video images of various urban driving scenes. The CamVid dataset is divided into 11 semantic categories, including road, building, sky, tree, sidewalk, car, pole, symbol and fence. The dataset contains 701 pieces of high-precision labeling image, 367 of them are used for training, 233 pieces are used for testing and 101 pieces are used for verification.

    The Cityscapes dataset[17]is a publicly available large-scale urban road scene dataset. It consisted of video pictures taken on the streets of 50 different cities under different time and weather conditions. The dataset contains 5 000 high-quality pixel set annotated pictures, and 2 975 of them are used for training, 500 of them are used for testing, and 1 525 of them are used for verification. In addition, it contains 20 000 pictures with rough labels. The experiment in this study only uses 5 000 pictures with fine labels to verify the effectiveness of the improved model. Cityscapes dataset defines a total of 30 target categories from the perspective of frequency and practicality, which is mainly divided into 8 semantic categories, including flat, construction, nature, vehicle, sky, object, human and void. The machine environment configuration of the experiment in this study is shown in Table 1.

    Table 1 Laboratory environment configuration

    2.1.2 Experimental procedure

    In order to verify the segmentation performance of the improved network model, under the same experimental environment conditions, this study has performed a test on six network models, respectively, including Deeplabv3+, Deeplabv3+ with PAM, Deeplabv3+ with CAM, Deeplabv3+ with PAM and CAM (series structure), Deeplabv3+ with PAM and CAM (parallel structure), Deeplabv3+ with PAM and CAM (parallel structure) and BN operation. The segmentation results of Deeplabv3+ on CamVid dataset and Cityscapes dataset were used as the criterion to compare the effectiveness of the segmentation performance of each network model.

    2.1.3 Evaluation Index

    In semantic segmentation, intersection over union (IoU)[18-20]describes the overlap ratio of the image prediction area and the real area. The more overlapped areas or the less misjudged areas will make the IoU value increase. Pixel accuracy (PA) represents the ratio between correctly classified pixels and total pixels. This study uses the mean intersection over union (mIoU) and mean pixel accuracy (mPA) as the standard to measure the segmentation accuracy of the network model before and after the improvement. The larger the values of mIoU and mPA are, the better the model segmentation effect will be.

    2.2 Results on CamVid dataset

    2.2.1 Comparison of model segmentation effect before and after improvement

    In the experiment, the training times of each network model were set as 30 000 times, the batch_size value was set as 2, and the initial learning rate was set as 1e-4. In order to verify the superiority of the improved model segmentation effect and segmentation speed, the mIoU value, mPA value of each model in the experiment and the average time required to generate a predicted picture are statistically compared. The statistical results are shown in Table 2.

    Table 2 Segmentation effect of different models on CamVid

    It can be seen from the values of mIoU and mPA in Table 2 that the network structure model after the introduction of PAM and PCM are increased. It can improve the segmentation effect to varying degrees, but compared to other models, the parallel structure model has a better segmentation effect. From the results of the average time for each model to generate a predicted picture, it can be seen that the addition of the attention module hardly increases prediction time of the picture. After BN operation, the processing time of the model will be increased a little, and the average duration of a predicted image will be increased by 59 ms. However, at the same time, the corresponding segmentation mIoU value and mPA value will be increased by 0.72% and 1.08%, respectively. Compared with the improvement in accuracy, the increase in prediction time can be ignored. Fig.5 shows the segmentation results of the model on CamVid dataset when PAM and CAM are added separately.

    Fig.5 Comparison of visualization results of models with different attention modules on CamVid. (a) Original image;(b) Without attention module;(c) With CAM;(d) With PAM

    2.2.2 Impact of batch_size on segmentation accuracy

    Taking into account that the size of batch_size in deep learning will have a certain impact on the segmentation results. In this study, the value of batch_size was set to 1, 2, 4, and 8 for experiments under the improved network structure to count the corresponding mIoU values. The results are shown in Table 3.

    Table 3 Segmentation effect of model under different batch_size

    It can be seen from Table 3 that as the batch_size value increases, the segmentation effect of the model will be improved to a certain extent. However, it is found that when the batch_size is larger in the experiment process, it will occupy a higher video memory and require more hardware equipment. Moreover, to achieve the same segmentation accuracy, the training time of the model will be greatly increased. When the batch_size is small, it will significantly reduce the accuracy of the BN layer statistics, which makes the model difficult to converge. In this study, after a large number of experiments comprehensively compared the experimental results under different batch_size values, choosing to use the batch_size value of 2 to count the segmentation effects of each model. The segmentation speed and effectiveness of the improved network model were verified, and good results were obtained.

    2.2.3 Comparison with classic networks

    There are common semantic segmentation models including SegNet, FCN and Deeplabv3+. The segmentation accuracy mIoU values of these network models on CamVid dataset are compared, and the results are shown in Table 4.

    Table 4 Segmentation accuracy of several network models

    It can be seen from Table 4 that the Deeplabv3+ network model integrated with the AM proposed in this study has better segmentation accuracy. The mIoU value of the improved network model on CamVid dataset is 6.88% higher than the original Deeplabv3+, and 12.39% higher than the mIoU value of the SegNet network structure. The experimental results show that the AM can capture more contextual and semantic information, enhance the feature representation ability, and then improve the segmentation effect of the model.

    Fig.6 is part of the segmentation results of the network model on CamVid dataset before and after the improvement.

    Fig.6 Different segmentation effects of model before and after improvement on CamVid

    It can be seen from the Fig.6 that compared with the segmentation results of the original Deeplabv3+ model, the segmentation effect of the network model after integrating the AM is significantly improved. For example, the model integrated with the AM has a clearer segmentation of the outline of the cyclist on the road, and the problem of missed judgments such as telephone poles at traffic lights on the side of the road and bicycles parked on the sidewalk has also been improved. The colors represented by different semantic categories in this experiment are different from the original tags, but they do not affect the judgment of the model segmentation effect, as it is shown in the color description of each semantic category in Fig.6.

    2.3 Results on Cityscapes dataset

    2.3.1 Comparison of model segmentation effect before and after improvement

    For the Cityscapes dataset, the initial learning rate was set to 1e-4, and the batch_size value was set to 4. The segmentation results under each network model are counted to verify the effectiveness of the improved model. The statistical results are shown in Table 5.

    Table 5 Segmentation effect of different models on Cityscapes

    It can be seen from Table 5 that on the Cityscapes dataset, compared with the Deeplabv3+ model, the introduction of the AM can still improve the segmentation accuracy. Adding the BN operation will increase the time for the model to generate a predicted picture, but its mIoU value has increased by 0.51% compared without using the BN operation, and its mPA value has also increased by 0.81%. Therefore, compared to the improvement in accuracy, the time-consuming cost can be ignored. Fig.7 is the change curves of loss value on CamVid dataset and Cityscapes dataset before and after BN operation. It can be seen that the model is easier to converge after BN operation.

    (a) On CamVid

    Fig.8 shows the segmentation effect of the model when PAM and CAM are integrated. It can be seen that the segmentation effect of the model integrated with PAM is slightly better than with CAM, and the segmentation effect of the model integrating PAM and CAM is better than Deeplabv3+.

    Fig.8 Comparison of visualization results of models with different attention modules on Cityscapes. (a) Original image; (b) Without attention module; (c) With CAM; (d) With PAM

    2.3.2 Comparison with classic networks

    By comparing the mIoU values of SegNet, FCN, Deeplabv3+ and the improved network on Cityscapes dataset, the results are shown in Table 6. The experimental results show that the improved network model has a better segmentation effect.

    Table 6 Segmentation accuracy of several network models

    Fig.9 shows the partial segmentation results of the model before and after the improvement on Cityscapes. It can be seen that the improved model has a certain improvement in the problems of small objects being missed and misjudgment of objects with similar shapes. However, there are also some problems. For example, in the first picture of Fig.9, the segmentation result of the building misjudged as a telegraph pole was slightly improved compared with the model before the improvement, but the classification is still not correct, and the pedestrian pushing a stroller was judged as a cyclist in the last picture. Therefore, these problems need to be further improved in the future research.

    Fig.9 Different segmentation effects of model before and after improvement on Cityscapes

    3 Conclusions

    Aiming at the image semantic segmentation technology in the field of automatic driving, a road scene semantic segmentation method combining the attention mechanism is proposed in this study. With Deeplabv3+ as the basic network model, PAM and CAM are introduced in both encoding and decoding. A parallel structure is adopted for the two attention modules to capture more context and semantic information in spatial dimension and channel dimension, and finally the refined results are output.

    The experimental results show that the improved network structure on CamVid dataset and Cityscapes dataset improves mIoU by 6.88% and 2.58%, respectively, compared with Deeplabv3+. Although BN operation will increase the time for the model to generate predicted pictures, compared with the improvement of segmentation accuracy, the time cost can be ignored. It can also accelerate the convergence speed of the model, and better solve the problems of fuzzy segmentation boundary, misjudgment and missed judgment under Deeplabv3+ network structure.

    The disadvantage of this study is that the model segmentation effect has been improved to a certain extent, but the segmentation time has not reduced, and the segmentation result will be affected by batch_size. In future studies, it is necessary to reduce the network complexity and shorten the training time while ensuring the model segmentation effect, so as to achieve a balance between the segmentation speed and accuracy in the actual driving scene, thereby ensuring the driving safety of vehicles.

    少妇的逼好多水| 少妇的逼水好多| 亚洲av不卡在线观看| 亚洲av免费高清在线观看| 天堂动漫精品| 久久人人精品亚洲av| 精品国产亚洲在线| 欧美精品啪啪一区二区三区| 亚洲av中文字字幕乱码综合| 亚洲欧美精品综合久久99| 国产成人av教育| 日本 欧美在线| 宅男免费午夜| 亚洲美女黄片视频| 12—13女人毛片做爰片一| 日本成人三级电影网站| 免费一级毛片在线播放高清视频| 欧美国产日韩亚洲一区| 黄色片一级片一级黄色片| 九色成人免费人妻av| 宅男免费午夜| 嫩草影视91久久| 麻豆成人av在线观看| 日韩精品中文字幕看吧| 欧美日韩国产亚洲二区| 亚洲在线观看片| 男女那种视频在线观看| 99久久成人亚洲精品观看| 免费看日本二区| 久久久久性生活片| 亚洲欧美日韩高清专用| 久久性视频一级片| 久久精品91无色码中文字幕| av在线天堂中文字幕| 国产精品一及| 最新中文字幕久久久久| www日本黄色视频网| www日本黄色视频网| 国产真人三级小视频在线观看| 99国产极品粉嫩在线观看| 国产高清视频在线播放一区| 成年版毛片免费区| 伊人久久精品亚洲午夜| 午夜日韩欧美国产| 亚洲在线自拍视频| 99久久成人亚洲精品观看| 国产久久久一区二区三区| 人妻久久中文字幕网| 操出白浆在线播放| 男人的好看免费观看在线视频| 色综合亚洲欧美另类图片| 日本免费一区二区三区高清不卡| 精品一区二区三区人妻视频| 在线观看66精品国产| 欧美日本视频| 亚洲成av人片在线播放无| 国产色婷婷99| 天堂网av新在线| 老师上课跳d突然被开到最大视频 久久午夜综合久久蜜桃 | 欧美国产日韩亚洲一区| 老司机福利观看| e午夜精品久久久久久久| 欧美日韩一级在线毛片| 男插女下体视频免费在线播放| 国产高清有码在线观看视频| 欧美性猛交黑人性爽| 国产精品三级大全| 99热只有精品国产| 最新中文字幕久久久久| 精品国产三级普通话版| 久久婷婷人人爽人人干人人爱| 日本 av在线| 亚洲国产精品sss在线观看| 国产午夜精品久久久久久一区二区三区 | 国产午夜精品久久久久久一区二区三区 | 黄片大片在线免费观看| 欧美成人a在线观看| 丝袜美腿在线中文| 日韩高清综合在线| 十八禁人妻一区二区| 禁无遮挡网站| 久99久视频精品免费| 1024手机看黄色片| 在线看三级毛片| 免费看日本二区| 高清在线国产一区| 一本精品99久久精品77| 国产aⅴ精品一区二区三区波| 久久人妻av系列| xxx96com| 亚洲av成人精品一区久久| 黑人欧美特级aaaaaa片| 久久国产精品人妻蜜桃| 午夜精品一区二区三区免费看| 啦啦啦观看免费观看视频高清| 亚洲欧美日韩高清在线视频| 美女被艹到高潮喷水动态| 国产探花在线观看一区二区| 国产探花在线观看一区二区| 色综合婷婷激情| 无限看片的www在线观看| 好看av亚洲va欧美ⅴa在| 国产伦在线观看视频一区| 老司机深夜福利视频在线观看| 欧美+日韩+精品| 亚洲va日本ⅴa欧美va伊人久久| 成年版毛片免费区| 亚洲精品一卡2卡三卡4卡5卡| 一区二区三区激情视频| 亚洲国产高清在线一区二区三| 全区人妻精品视频| 精品国产亚洲在线| 成年女人看的毛片在线观看| 国产亚洲精品久久久久久毛片| 国产极品精品免费视频能看的| 亚洲国产高清在线一区二区三| 三级国产精品欧美在线观看| 亚洲 欧美 日韩 在线 免费| 日韩免费av在线播放| 在线观看av片永久免费下载| 男人舔奶头视频| 久久欧美精品欧美久久欧美| 成人国产一区最新在线观看| 亚洲真实伦在线观看| 久久国产乱子伦精品免费另类| 欧美日本亚洲视频在线播放| 久久精品国产自在天天线| 啦啦啦免费观看视频1| 国产一区二区在线观看日韩 | 成年女人永久免费观看视频| 亚洲精品国产精品久久久不卡| 亚洲内射少妇av| 天天添夜夜摸| 国产精品女同一区二区软件 | 亚洲在线自拍视频| 国产精品电影一区二区三区| 免费av观看视频| 亚洲欧美日韩东京热| 搞女人的毛片| www日本在线高清视频| 757午夜福利合集在线观看| 99久久九九国产精品国产免费| 亚洲av二区三区四区| 久久香蕉精品热| 不卡一级毛片| 我的老师免费观看完整版| 老汉色∧v一级毛片| 亚洲精品亚洲一区二区| 中文字幕人成人乱码亚洲影| 亚洲美女视频黄频| 亚洲av美国av| 久久久久久大精品| 国产精品精品国产色婷婷| 丝袜美腿在线中文| 午夜福利在线观看吧| 十八禁人妻一区二区| 高清毛片免费观看视频网站| 亚洲人与动物交配视频| 美女 人体艺术 gogo| 日本撒尿小便嘘嘘汇集6| 精品日产1卡2卡| 国产精品久久电影中文字幕| 亚洲精品乱码久久久v下载方式 | 亚洲成人免费电影在线观看| 一进一出抽搐gif免费好疼| 男人舔奶头视频| 国产高清激情床上av| 国产一区在线观看成人免费| 男人舔女人下体高潮全视频| 色吧在线观看| 90打野战视频偷拍视频| 亚洲七黄色美女视频| 俺也久久电影网| 1024手机看黄色片| 黄色片一级片一级黄色片| 中出人妻视频一区二区| 亚洲精品乱码久久久v下载方式 | 久久人人精品亚洲av| 一级a爱片免费观看的视频| 国产91精品成人一区二区三区| 久久香蕉国产精品| 欧美xxxx黑人xx丫x性爽| a在线观看视频网站| av女优亚洲男人天堂| 人妻久久中文字幕网| 中文字幕久久专区| 亚洲一区二区三区不卡视频| 精品国产亚洲在线| 亚洲av成人不卡在线观看播放网| 国产av在哪里看| 日本 av在线| 国产精品日韩av在线免费观看| 国产黄色小视频在线观看| 18禁裸乳无遮挡免费网站照片| 可以在线观看的亚洲视频| 色综合欧美亚洲国产小说| 亚洲av成人精品一区久久| 国产成+人综合+亚洲专区| 老熟妇乱子伦视频在线观看| 久久精品91蜜桃| 亚洲中文字幕一区二区三区有码在线看| 日韩欧美精品v在线| 国产免费av片在线观看野外av| 国产成人福利小说| 亚洲av熟女| 最近在线观看免费完整版| 精品日产1卡2卡| 看免费av毛片| www日本在线高清视频| 日本免费一区二区三区高清不卡| 丁香欧美五月| 国产成年人精品一区二区| 国产一区二区三区在线臀色熟女| 18美女黄网站色大片免费观看| 免费在线观看成人毛片| 老司机福利观看| 成人精品一区二区免费| 黄色日韩在线| 午夜福利在线在线| 19禁男女啪啪无遮挡网站| а√天堂www在线а√下载| 成年人黄色毛片网站| 99久久综合精品五月天人人| 观看免费一级毛片| 91在线观看av| 在线观看免费视频日本深夜| 人人妻,人人澡人人爽秒播| 九九久久精品国产亚洲av麻豆| 宅男免费午夜| 在线观看美女被高潮喷水网站 | 久久国产精品影院| 在线观看av片永久免费下载| 亚洲片人在线观看| 国产黄a三级三级三级人| 免费大片18禁| 成人无遮挡网站| 可以在线观看的亚洲视频| 精品午夜福利视频在线观看一区| 亚洲第一电影网av| 成人一区二区视频在线观看| 国产精品一区二区三区四区免费观看 | 久久午夜亚洲精品久久| 免费在线观看亚洲国产| 丰满人妻熟妇乱又伦精品不卡| 国产午夜精品论理片| 欧美av亚洲av综合av国产av| 国产精品久久久久久精品电影| 天天添夜夜摸| 成人性生交大片免费视频hd| 色尼玛亚洲综合影院| 好看av亚洲va欧美ⅴa在| 成年女人看的毛片在线观看| 欧美av亚洲av综合av国产av| www.999成人在线观看| 久久天躁狠狠躁夜夜2o2o| 91麻豆av在线| 天堂av国产一区二区熟女人妻| 色老头精品视频在线观看| 婷婷精品国产亚洲av在线| 搡女人真爽免费视频火全软件 | 欧美绝顶高潮抽搐喷水| 亚洲一区二区三区色噜噜| 搡女人真爽免费视频火全软件 | 国语自产精品视频在线第100页| 两个人的视频大全免费| 午夜影院日韩av| 变态另类丝袜制服| 一边摸一边抽搐一进一小说| xxxwww97欧美| 国产欧美日韩精品一区二区| 手机成人av网站| 又粗又爽又猛毛片免费看| 国产主播在线观看一区二区| 变态另类丝袜制服| 国模一区二区三区四区视频| 99热这里只有精品一区| 91九色精品人成在线观看| 动漫黄色视频在线观看| 男人舔女人下体高潮全视频| 露出奶头的视频| 男人的好看免费观看在线视频| 国产精品 欧美亚洲| 老司机深夜福利视频在线观看| 好男人电影高清在线观看| 亚洲,欧美精品.| 一个人观看的视频www高清免费观看| 欧美中文综合在线视频| 国产精品亚洲一级av第二区| 亚洲精品成人久久久久久| 午夜福利免费观看在线| 国产真实乱freesex| 久久人妻av系列| 欧美日韩综合久久久久久 | 宅男免费午夜| 欧美性猛交黑人性爽| 国产高清有码在线观看视频| 超碰av人人做人人爽久久 | 欧美日本亚洲视频在线播放| 亚洲av电影不卡..在线观看| 日韩欧美精品v在线| 男女做爰动态图高潮gif福利片| 国产三级黄色录像| 男插女下体视频免费在线播放| 中文在线观看免费www的网站| 波野结衣二区三区在线 | 九色成人免费人妻av| 久久久久九九精品影院| 99久久综合精品五月天人人| 在线播放无遮挡| 国产一级毛片七仙女欲春2| 五月伊人婷婷丁香| 观看美女的网站| e午夜精品久久久久久久| 中文字幕高清在线视频| 麻豆久久精品国产亚洲av| 黄片小视频在线播放| 最近最新中文字幕大全免费视频| 桃色一区二区三区在线观看| 久久精品国产综合久久久| 午夜福利在线在线| 国产三级黄色录像| 国产淫片久久久久久久久 | av在线天堂中文字幕| 别揉我奶头~嗯~啊~动态视频| 久久伊人香网站| 国产精品乱码一区二三区的特点| 日本五十路高清| 国产三级黄色录像| 国产成人aa在线观看| 一区二区三区国产精品乱码| 夜夜看夜夜爽夜夜摸| 18禁国产床啪视频网站| 国产亚洲av嫩草精品影院| 色尼玛亚洲综合影院| 国产精品一区二区三区四区免费观看 | 欧美日韩一级在线毛片| 色尼玛亚洲综合影院| 亚洲色图av天堂| avwww免费| 丝袜美腿在线中文| 蜜桃亚洲精品一区二区三区| 在线播放无遮挡| 国产三级在线视频| 欧美性猛交╳xxx乱大交人| 在线免费观看不下载黄p国产 | www.色视频.com| 免费人成视频x8x8入口观看| 亚洲人成网站在线播| 网址你懂的国产日韩在线| 国产黄a三级三级三级人| 免费观看的影片在线观看| 国产精品一区二区免费欧美| 欧美一区二区精品小视频在线| 淫秽高清视频在线观看| 欧美色欧美亚洲另类二区| 老司机深夜福利视频在线观看| 国产一区在线观看成人免费| 亚洲av熟女| 日本在线视频免费播放| 全区人妻精品视频| 99热只有精品国产| 日韩有码中文字幕| 可以在线观看毛片的网站| 亚洲中文字幕一区二区三区有码在线看| 欧美午夜高清在线| 中亚洲国语对白在线视频| 中文在线观看免费www的网站| 亚洲中文字幕日韩| 老汉色av国产亚洲站长工具| 操出白浆在线播放| 人人妻人人看人人澡| 日本与韩国留学比较| 欧美日韩乱码在线| 中出人妻视频一区二区| 免费无遮挡裸体视频| 亚洲国产精品sss在线观看| 久久久久久久精品吃奶| 国产精品 欧美亚洲| 小说图片视频综合网站| 99精品欧美一区二区三区四区| 成人性生交大片免费视频hd| a级一级毛片免费在线观看| 亚洲精品在线观看二区| av在线天堂中文字幕| 色av中文字幕| 精品福利观看| 亚洲激情在线av| 亚洲最大成人手机在线| 日韩大尺度精品在线看网址| 欧美+亚洲+日韩+国产| 最近最新中文字幕大全免费视频| 亚洲人成网站在线播| 亚洲中文字幕日韩| 亚洲精品美女久久久久99蜜臀| 久99久视频精品免费| 亚洲中文日韩欧美视频| 一级a爱片免费观看的视频| 成人午夜高清在线视频| 色尼玛亚洲综合影院| 欧美日韩黄片免| 久久久久久国产a免费观看| 欧美日韩一级在线毛片| 免费一级毛片在线播放高清视频| 成人三级黄色视频| 午夜老司机福利剧场| 国产激情欧美一区二区| 免费在线观看影片大全网站| 波多野结衣高清作品| 中文字幕精品亚洲无线码一区| 搞女人的毛片| 国产精品av视频在线免费观看| 母亲3免费完整高清在线观看| 欧美大码av| 亚洲一区二区三区色噜噜| 黄色视频,在线免费观看| 国产v大片淫在线免费观看| 国产老妇女一区| 国产国拍精品亚洲av在线观看 | 熟女少妇亚洲综合色aaa.| 又黄又爽又免费观看的视频| 高清日韩中文字幕在线| 麻豆国产av国片精品| 波多野结衣巨乳人妻| 欧美在线黄色| 国模一区二区三区四区视频| 午夜视频国产福利| 国产野战对白在线观看| 午夜福利在线观看吧| 真人做人爱边吃奶动态| 在线观看一区二区三区| 男插女下体视频免费在线播放| 一级毛片女人18水好多| 亚洲美女黄片视频| 久久久久国内视频| 99热精品在线国产| 人人妻人人澡欧美一区二区| 91在线精品国自产拍蜜月 | 精品一区二区三区人妻视频| 亚洲熟妇中文字幕五十中出| 非洲黑人性xxxx精品又粗又长| 国产亚洲欧美在线一区二区| 天天添夜夜摸| 亚洲人成网站在线播| 精品一区二区三区视频在线观看免费| 十八禁网站免费在线| 国产精品久久久久久久久免 | 国产成人啪精品午夜网站| 日本黄色视频三级网站网址| 久久久精品欧美日韩精品| 欧洲精品卡2卡3卡4卡5卡区| 国产一区二区三区视频了| 最近视频中文字幕2019在线8| 亚洲国产精品久久男人天堂| 老熟妇仑乱视频hdxx| 色播亚洲综合网| 波多野结衣巨乳人妻| 久久99热这里只有精品18| 日本撒尿小便嘘嘘汇集6| 女同久久另类99精品国产91| 听说在线观看完整版免费高清| 在线播放国产精品三级| 中文字幕av成人在线电影| 一区二区三区免费毛片| 欧美高清成人免费视频www| 草草在线视频免费看| 在线免费观看不下载黄p国产 | 一个人观看的视频www高清免费观看| 亚洲av中文字字幕乱码综合| 国产精品久久久久久精品电影| 国内精品久久久久久久电影| 成年女人看的毛片在线观看| а√天堂www在线а√下载| 国产成人aa在线观看| 岛国在线观看网站| 国产成人福利小说| 亚洲不卡免费看| 此物有八面人人有两片| 国模一区二区三区四区视频| netflix在线观看网站| 亚洲,欧美精品.| 欧美3d第一页| 12—13女人毛片做爰片一| 在线观看美女被高潮喷水网站 | 日本免费a在线| 99热精品在线国产| 国产免费av片在线观看野外av| 国产视频内射| 免费在线观看影片大全网站| 国产精品久久久久久人妻精品电影| 99热精品在线国产| 精品国产三级普通话版| 亚洲人成伊人成综合网2020| 在线天堂最新版资源| 99国产精品一区二区蜜桃av| 免费一级毛片在线播放高清视频| 免费av毛片视频| 老司机午夜福利在线观看视频| 国内少妇人妻偷人精品xxx网站| 欧美成人一区二区免费高清观看| 天堂av国产一区二区熟女人妻| 色吧在线观看| 精品午夜福利视频在线观看一区| 久久欧美精品欧美久久欧美| 一边摸一边抽搐一进一小说| 亚洲最大成人中文| 国产三级黄色录像| 淫妇啪啪啪对白视频| 国产主播在线观看一区二区| 国产97色在线日韩免费| 国产精品综合久久久久久久免费| 俺也久久电影网| 婷婷亚洲欧美| 免费大片18禁| 国产av麻豆久久久久久久| 久久精品国产清高在天天线| 欧美日韩亚洲国产一区二区在线观看| 日本五十路高清| 欧美日韩福利视频一区二区| 在线观看66精品国产| 日日干狠狠操夜夜爽| 国产午夜精品论理片| 国产国拍精品亚洲av在线观看 | 激情在线观看视频在线高清| 香蕉丝袜av| 在线观看舔阴道视频| 国产亚洲欧美在线一区二区| 夜夜躁狠狠躁天天躁| 国产精品,欧美在线| 高清在线国产一区| av国产免费在线观看| e午夜精品久久久久久久| 97碰自拍视频| 亚洲人成电影免费在线| 欧美三级亚洲精品| 黄片大片在线免费观看| 免费av不卡在线播放| 免费看十八禁软件| 亚洲最大成人中文| 亚洲性夜色夜夜综合| 天堂av国产一区二区熟女人妻| 亚洲国产欧洲综合997久久,| 一本精品99久久精品77| 一进一出好大好爽视频| 在线播放国产精品三级| 国产主播在线观看一区二区| 色av中文字幕| av国产免费在线观看| 免费看日本二区| 波多野结衣高清无吗| 麻豆国产av国片精品| 999久久久精品免费观看国产| 老鸭窝网址在线观看| 亚洲国产精品合色在线| 亚洲av一区综合| 少妇的逼水好多| 久久久久久大精品| 窝窝影院91人妻| www.999成人在线观看| 九色国产91popny在线| 精品国产美女av久久久久小说| 99久久成人亚洲精品观看| 国产精品久久久人人做人人爽| 午夜福利免费观看在线| 天堂影院成人在线观看| 一a级毛片在线观看| 老司机福利观看| 国产精品香港三级国产av潘金莲| 91麻豆av在线| 美女免费视频网站| 国产69精品久久久久777片| 国产精品久久久久久亚洲av鲁大| 亚洲成人久久爱视频| 波多野结衣高清无吗| 可以在线观看的亚洲视频| 日本免费一区二区三区高清不卡| 女警被强在线播放| 嫩草影视91久久| 99久久精品国产亚洲精品| 亚洲真实伦在线观看| 18禁黄网站禁片午夜丰满| 日韩人妻高清精品专区| aaaaa片日本免费| 欧美又色又爽又黄视频| aaaaa片日本免费| 午夜影院日韩av| 女人被狂操c到高潮| 久久人人精品亚洲av| 性欧美人与动物交配| 美女大奶头视频| 成人亚洲精品av一区二区| 男人舔女人下体高潮全视频| 亚洲第一电影网av| 精品国产美女av久久久久小说| 尤物成人国产欧美一区二区三区| 美女黄网站色视频| 夜夜躁狠狠躁天天躁| 亚洲熟妇熟女久久| 天堂av国产一区二区熟女人妻| 色综合站精品国产| 香蕉久久夜色| 我要搜黄色片| 成人高潮视频无遮挡免费网站| 两个人视频免费观看高清| 欧美最新免费一区二区三区 | 欧美在线黄色| 欧美国产日韩亚洲一区| 色吧在线观看| 看免费av毛片| av在线天堂中文字幕| 久久久久国产精品人妻aⅴ院| 最近最新免费中文字幕在线| 白带黄色成豆腐渣| 欧美成人一区二区免费高清观看| 国产99白浆流出| 99精品久久久久人妻精品|