• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Land cover classification in a mixed forest-grassland ecosystem using LResU-net and UAV imagery

    2022-06-04 11:43:22ChongZhangLiZhangBessieZhangJingqianSunShikuiDongXueyanWangYaxinLiJianXuWenkaiChuYanweiDongPeiWang
    Journal of Forestry Research 2022年3期

    Chong Zhang·Li Zhang·Bessie Y.J.Zhang·Jingqian Sun·Shikui Dong·Xueyan Wang·Yaxin Li·Jian Xu·Wenkai Chu·Yanwei Dong·Pei Wang

    Abstract Using an unmanned aerial vehicle (UAV) paired with image semantic segmentation to classify land cover within natural vegetation can promote the development of forest and grassland field.Semantic segmentation normally excels in medical and building classification,but its usefulness in mixed forest-grassland ecosystems in semi-arid to semi-humid climates is unknown.This study proposes a new semantic segmentation network of LResU-net in which residual convolution unit (RCU) and loop convolution unit(LCU) are added to the U-net framework to classify images of different land covers generated by UAV high resolution.The selected model enhanced classification accuracy by increasing gradient mapping via RCU and modifying the size of convolution layers via LCU as well as reducing convolution kernels.To achieve this objective,a group of orthophotos were taken at an altitude of 260 m for testing in a natural forest-grassland ecosystem of Keyouqianqi,Inner Mongolia,China,and compared the results with those of three other network models (U-net,ResU-net and LU-net).The results show that both the highest kappa coefficient(0.86) and the highest overall accuracy (93.7%) resulted from LResU-net,and the value of most land covers provided by the producer’s and user’s accuracy generated in LResUnet exceeded 0.85.The pixel-area ratio approach was used to calculate the real areas of 10 different land covers where grasslands were 67.3%.The analysis of the effect of RCU and LCU on the model training performance indicates that the time of each epoch was shortened from U-net (358 s) to LResU-net (282 s).In addition,in order to classify areas that are not distinguishable,unclassified areas were defined and their impact on classification.LResU-net generated significantly more accurate results than the other three models and was regarded as the most appropriate approach to classify land cover in mixed forest-grassland ecosystems.

    Keywords UAV images·Semantic segmentation·LResU-net·Land cover classification

    Introduction

    As one of the world’s largest renewable natural resources,mixed forest-grassland resources directly affect the development of agriculture,forestry and other industries (Langley et al.2001;Ma et al.2010).According to Scurlock et al.(2002) and Dong et al.(2017b),mixed forest-grassland ecosystems are approximately 3.2 billion hectares,accounting for 40% of the total land area.Using remote sensing to classify land cover in a mixed forest-grassland ecosystem can provide detailed grassland and woodland information over large areas (Fang Fang et al.2010).

    In the past decade,UAV image analysis technology has been widely applied to the identification and classification in forest and grassland resource surveys (Chen 2019).It has increasingly become an opportunity to attach high resolution cameras (Huseyin et al.2019),LiDAR (Yang et al.2020),thermal infrared (Crusiol et al.2019) and hyperspectral cameras (Clark et al.2018) on UAV to better collect field information for land classification.Christian and Christiane(2014) compared forest point cloud data collected from UAV images and airborne LiDAR and concluded that more information was captured through UAV image data.Zhang et al.(2020a) used aerial hyperspectral images to classify tree species on forest farms in China,and obtained an accuracy of 93.1%.However,hyperspectral imaging may be limited when used on grassland areas with low-level color contrast as it creates a large amount of redundant data (Grigorieva et al.2020).In addition,wind has considerable influence on LiDAR data which leads to noise and ghost points around the detected targets (Yun et al.2016;Xu et al.2018).Therefore,using UAV high resolution cameras is one of the most preferred methods to classify land cover in a mixed forestgrassland ecosystem.

    Traditional segmentation methods of remote sensing involve pixel-based segmentation (Bhadoria et al.2020)object-based analysis (José et al.2013),and random forest segmentation (Fei et al.2015).The analysis of pixel-based segmentation aims only at the color information among pixels,ignoring the semantic information of the classified objects,giving a poor performance in multi-object classification (Zhang et al.2 020c).Numerous researchers have studied forestry classification algorithms based on a combination of object-based analysis,random forest and manual feature extraction.Ke et al.(2010) applied an object-based approach to evaluate the synergism in high spatial resolution multispectral imagery and low-posting-density LiDAR data for forest species classification.Random forest segmentation was applied to classify tree species using satellite images of temperate forests in Austria,and the overall accuracy was 82% (Immitzer et al.2012).In practice,the above approaches require extensive manual marking which contributes to a waste of human resources for high accuracy feature extraction (Wolf and Bochum 2013;Dalponte et al.2015).

    With the development of deep learning and convolutional neural networks (CNN) (Zhang et al.2020b;Lou et al.2021),numerous semantic segmentation algorithms exist for automatic classification (Fu and Qu 2018;Braga et al.2020).U-Net (Ronneberger et al.2015),is a semantic segmentation model based on a fully convolutional network and was initially used for biomedical image segmentation (Dong et al.2017a;Rad et al.2020).In comparison to other deep learning networks such as fully convolutional networks(FCN) (Long et al.2015) and Densenet (Huang et al.2017),U-Net has the overwhelming advantage of overall accuracy using a small number of data sets (Liu et al.2020).In this context,U-net was used to extract complex terrain features to classify hills and ridges of the Loess Plateau in China (Li et al.2020a).Due to unsurpassed reliability and excellent segmentation quality,some researchers have applied U-net to train hyperspectral satellite images and obtain the distribution of trees in the Sahara and Sahel regions of West Africa (Brandt et al.2020).

    Numerous studies have indicated that defects occur during U-net’s feature extraction process (Freudenberg et al.2019;Cao and Zhang 2020;Li et al.2020b).Since U-net’s downsampling depends on a stack modules of Conv-BN-ReLU(CBR),this may cause extraction scales to vary at different depths,leading to more exaggerated classification errors(Cicek et al.2016).In an effort to correct the defects,the following improvements have been made:

    (1) Replace CBR modules with RCU of ResNet (He et al.2016).ResNet directly connects the encoder and the decoder in the sample and has a capability to prevent the loss of the encoded information within different layers (Zahangir et al.2017).For example,a building-extraction algorithm based on ResNet’s remote imaging demonstrated an outstanding performance in an urban setting (Xu et al.2018).

    (2) Add LCU to down-sampling in feature extraction.LUnet is a combination of U-net and LCU,where the number of convolution at each layer of the network was increased while the convolutional dimension was shorten.Alom et al.(2018)proposed a recurrent convolutional neural network based on U-net structure that exhibited superior performance on skin cancer segmentation tasks.

    However,the above methods performed well in the form of binary classification for medical and urban building domains,but the ability to classify land cover in a complex forest and grassland ecosystem remains a major challenge.As such,in this study,applying the improved U-net model to achieve accurate land cover classification in a mixed forest-grassland ecosystem is proposed.The objectives of this study include the following:(1) To propose an LResU-net model applicable to land cover classification based on U-net framework as well as a combination of RCU and LCU;(2) To evaluate the classification accuracy of U-net,ResU-net,LU-net and LResU-net in a mixed forest-grassland ecosystem;and,(3) To calculate the actual areas of various land covers using the best model of this study.

    Materials and methods

    Study area

    The study area (Fig.1) is located near the green Lv Shui Animal Breeding Farm of Horqin,Xing’an League,Inner Mongolia Autonomous Region at 46°42?51″ N and 120°30?1″ E with an altitude of 230–300 m.The local climate is midtemperate,semi-arid continental monsoon within an average annual temperature of 13 °C,rainfall of 420 mm,and humidity of 18%.The area consists of forests,grasslands and cultivated lands,and provides a variety of land covers such as natural grasslands,trees,roads,rivers,and buildings.

    Fig.1 a and b Study area:natural grasslands near Xing’an League,Inner Mongolia Autonomous Region; c synthesis orthophoto using UAV image; d UAV flight path generated from satellite planning route

    Field survey and acquisition of UAV image data

    The field investigation,September 23rd to 28th,2020 was near the Lv Shui Animal Breeding Farm.And involved the determination of land cover classes and UAV image data collection.The river bed has been eroded over many years and so some land cover is not identifiable.Aerial images were taken by the DJI Mavic 2 Pro drone equipped with Suha’s one-inch 20-megapixel CMOS sensor (Table 1).The flight airspace was 1210 m×600 m at an altitude of 260 m in which the overlap of flight paths was 85% and a side overlap of 80%.A total of 798 photos was produced.

    Table 1 DJI Mavic 2 Pro UAV flight parameters

    Data preprocessing

    The usual way to obtain an orthophoto map is three dimensional (3D) reconstruction over the entire study area.The steps of 3D reconstruction are (Fig.2).First,applying the structure from motion (SfM) algorithm achieved detection and matching of the feature points to obtain the sparse point clouds.Second,based on sparse point clouds,the dense point cloud was systematically acquired using multi-view stereo (MVS) algorithm.Third,a 3D imagery(Fig.3) was generated in a way of surface reconstruction and texture mapping upon dense point cloud.

    Fig.2 3D reconstruction work flow based on OpenMVG library and OpenMVS library

    Fig.3 An orthophoto derived from the 3D imagery using the Context Capture platform;details of roads,rivers and buildings are displayed in high resolution 3D imagery

    During the reconstruction process,open multiple view geometry (OpenMVG) library was y used for the SfM algorithm to get sparse point clouds.The subsequent procedures,including MVS,surface reconstruction and texture mapping,were implemented with the open multiple view stereo (OpenMVS) library.Finally,the 3D reconstructed model was compressed to an orthophoto on the Context Capture platform.The complete orthophoto with a pixel resolution of 20,167×13,534 was used to establish a high resolution data set for land cover classification.

    Production of data sets

    To prevent data loss,the data sets were produced by overlapping and cutting the entire orthophoto.The orthophoto with resolutions of 20,167×13,534 pixels,was first reshaped into an original dataset in which the image resolution was modified to 1024×1024 pixels.Based on the ratio of 6:2:2,the data sets were then divided into training,validation and test sets to ensure the mutual independence in data and to maintain the robustness of the model.For expending the data sets as well as reducing the performance requirements to graphics processing unit(GPU),images of 128×128 pixel were obtained from the original data sets by cutting with step of 64.The training,validation and test set were assigned to13145,4598 and 4596 images (Table 2).

    Table 2 Land cover classification classes and related data sets

    Table 3 Computer hardware attributes

    In the field investigation,some complex land covers were difficult to define as a category,such as the mixed landscape of swamps and the eroded lands around rivers.However,since the image grid is very large,it was difficult to label the whole image without any gaps;therefore,unclassified areas were defined as one of the label categories.According to visual interpretation,ten different categories of land cover were recognized using different colors as image classification objectives (Fig.4).The original orthophoto and labeled image were then respectively divided into samples and objectives in training set,verification set and test set by Photoshop software.

    Fig.4 a Map of the drone’s orthophoto; b Map of ten different category labels

    LResU-net network

    The backbone of LResU-net (Fig.5) is a combination of sampling characteristics in ResU-net and LU-net.On the one hand,due to encoder layer of U-net being relatively shallow,LCU was added to down-sampling in the feature extraction.In comparison to U-net,LCU increases model depth as well as achieve the improvement of sample details during feature extraction.On the other hand,advances in the closed-loop feedback mapping function of RCU effectively avoided the problem of gradient overflow and disappearance,i.e.,when the network’s loss rate reached the lowest value,ResU-net ensures that the network of the next layer still works in the most optimal state.

    Fig.5 a Backbone of the U-net feature extraction structure was a common CBR module (conv->BN->ReLU); b backbone of the ResU-net feature extraction structure with RCU added; c backbone of the LU-net feature extraction structure with LCU added; d backbone of the LResU-net feature extraction structure with RCU and LCU added

    At the same time,the number of convolution kernels was modified from 64 → 128 → 256 → 512 → 1024 to 32 → 64 → 128 → 256 → 512 to decrease the overall kernel count to 50% of U-net in the whole training process.According to previous studies (Liang and Hu 2015;Alom et al.2018),when the loop step of the LCU was 3,the featureextraction effect and training time were optimal;the entire LResU-net structure is shown in Fig.6.

    Fig.6 LResU-net model structure with the number of convolution kernels halved

    Comparison with U-net

    There are three differences of LResU-net from U-net:

    1) The original feature extraction backbone CBR has been abandoned and replaced with RCU of Resnet.

    2) Aiming at the feature extraction structure encoding–decoding,LCU was added to the network.Meanwhile,the number of convolutions at each layer can be changed quantitatively in accordance with the difficulty of feature extraction.

    3) The total convolution kernels was shorten by half in the training progress.

    Three advantages of LResU-net compared with U-Net as follows.

    1) Using the modified network solved the problem of gradient overflow and gradient disappearance.

    2) Reducing the rate of misjudgment of image segmentation with low color contrast by applying an improvement of accuracy in detailed feature extraction.

    3) The training time was shortened in approaches of optimizing network parameters and reducing redundant convolution kernels.

    Loss function and accuracy evaluation index

    Loss function estimated the inconsistency between the classification data of the model and the reference data during network training progress.The pixel set and the category set,respectively,are defined asi{1,2,…,N} andc{1,2…,M},the image set becomes.After feature extraction,the probability of different categories of pixels becomes a M-dimensional tensorsand the extent of[0,1],thereby,resulting in a multi-category,cross-entropy loss function:

    In this study,producer’s and user’s accuracy (Story and Congalton 1986;Olofsson et al.2013;Shao et al.2019),as well as a kappa coefficient,were used to evaluate classification accuracy.

    Network training

    To facilitate a performance comparison among different networks,four different models were used to train and predict,U-net,ResU-net,LU-net,and LResU-net.The learning rate was adjusted to 1×e–4and the batch size was 32.A total of 60 epoch with 18,000 steps allowed the model accuracy to reach the maximum in the training process.For the software platform,tensorflow-gpu 1.15 and keras 2.3.1 based on a Linux operating system was used as the learning framework,and all code was written by python.For the hardware platform-(Table 3),an Intel Xeon E5-2650 processor,a Nvidia GTX-1070 GPU,a 2 T ROM and 4 of 8 GB RAM were used to train and test.

    Table 4 Classification evaluation coefficients of four models with unclassified areas

    Table 5 Classification evaluation coefficients of four models without unclassified area

    Large-scale remote sensing imagery prediction and real area calculation

    Given that memory overflow may be caused if the entire orthophoto is directly inputted to the model to predict,all images were cropped into a group of 128×128 image slices.After prediction of the slices,a composite imagery was spliced using these images in accordance with the order they were cropped.However,the splice approach of clipping-prediction-splicing can result in obvious segmentation edges.An alternate method of clipping overlapping images and ignoring edges (Wang et al.2020) may mitigate this,i.e.,apply the area ratio between the ignored edge image and the stitched image to calculate the overlapping area size among the image slices.Of real area calculation,the ratio of UAV real flight area and the pixel area was used to calculate thearea of ten different land covers where the flight region was 70.54 ha.

    Results

    Accuracy of land cover classification using different networks

    The reference data were derived from the pixel area of each land cover in the labeled orthophoto,and the classified data from the predicted pixel area of each land cover using different models.

    Kappa coefficient and overall accuracy

    The first step is to make an accurate assessment of the different models for image classification,including U-net,ResUnet,LU-net,and LResU-net.Table 4 and Table 5,respectively,show kappa coefficient and classification accuracy with and without undefined areas in the whole data sets.The undefined area has a strong impact on accuracy assessment because the unclassified area was predicted in other land covers.

    On the basis of the separate data analysis in Table 4,the raise of the kappa coefficient and overall accuracy generated by ResU-net and LU-net indicates that both RCU and LCU played a positive role in modifying U-net.At the same time,the accuracy assessment of LResU-net was obviously higher,which also reflects a positive effect on the combination of RCU and LCU.

    When not involving the unclassified areas,the variation tendency of Table 4 and Table 5 are consistent,i.e.,the kappa coefficient and overall accuracy had an improvement to different extents on the modified model adding RCU and LCU.As expected,the optimum performance of accuracy assessment of LResU-net (kappa coefficient=0.86,overall accuracy=93.7%) was found in a test set,which is attributed to the advancement of ResU-net and LU-net.

    Producer’s and user’s accuracy derived from LResU-net

    The producer’s and user’s accuracy obtained from the LResU-net model is presented in Table 6.For most categories,both demonstrate highly favorable results.For example,trees occupy the highest value (producer’s=0.98 and user’s=0.93),and the harvested crop second (producer’s=0.94 and user’s=0.91).However,there are obvious differences between producer’s and user’s accuracy among some categories,including harvested grassland (producer’s=0.43 and user’s=0.99),and river (producer’s=0.85 and user’s=0.96).Such results are due to some undefined areas that were classified to above classes.

    Land cover classification results from different networks

    Figure 7 shows the graphs of land cover classification based on the four different network models.Compared with the results of U-net (Fig.7 a),noise and misjudgment rate of ResU-net (Fig.7 b) were slightly reduced,and the capacity of building classification significantly strengthened.Similarly,LU-net (Fig.7 c) was superior to U-net in overall classification performance.Even with some noise in the harvested grassland,the classification capacity for road,river,and building was better than that of U-net.As for the results of LResU-net (Fig.7 d),it exceeded others in classification performance,especially noise suppression from the grassland,harvested grassland,tree,and river.

    Fig.7 a U-net model prediction result; b ResU-net model prediction result; c LU-net model prediction result; d LResU-net model prediction result

    The real area of various land covers

    The outcome of various land covers is presented in Table 7.Regardless of unclassified area,the differences of the classification and the reference data for various land covers was insignificant.According to the results of the unclassified area,grassland (38.7%,area=27.3 ha) is the largest proportion of the area,followed by harvested grassland (28.7%,area=20.3 ha).The entire grassland area accounted for 67.4% of the study area.The smallest area was buildings(0.2%,area=4.8 ha).The proportion of forest area was 8.0%,which was average among all land covers.

    Table 6 Population error matrix involving producer’s and user’s accuracy

    Table 7 Areas of each land cover in the reference data and the classified data

    Discussion

    Effect of unclassified areas on classification results

    Unclassified areas will change the attribute of land cover.According to Fig.8 and Table 7,about 50% of unclassified areas was likely to have the same reference data in label,which indicated that some unclassified areas are subject to distinctive features and attributes,for example,swamp.Amid other unclassified areas,the regions attached fissures were predicted to be correct land cover,but other areas were grassland,harvested grassland and forest.It is attributed to the same or similar features between the unclassified and above classes in LResU-net’s vision.

    Fig.8 a Orthophoto map fused with label image; b orthophoto map fused with LResU-net model prediction image

    Performance of RCU and LCU on the model training process

    In line with the Table 8,the total parameters and training time in ResU-net were greater than that of U-net,whichresulted from an addition of RCU.This was consistent with previous studies on improvements of U-net (Alom et al.2018;Rad et al.2020).In contrast,as the convolutional dimension decreased,LU-net significantly reduced parameters and time.As RCU and LCU are combined,the number of parameters and training time in LResU-net (25.11 million,282 s) were slightly lower than U-net (31.05 million,358 s).

    Table 8 Four different model parameters and training time in each epoch

    Curves of accuracy and loss (Fig.9) show the overall error between the predicted data and the reference data during the training process.When training is near the 55th epoch,accuracy and loss tend to be steady and hardly need the supplement of more epochs.Thus,all training stopped at the 60th epoch.In addition,it can be seen that ResU-net provided the fastest convergence rate in comparison with the other networks,which is attributed to the decline of the encoded loss in different layers.

    Fig.9 a Training accuracy curves of four different models; b training loss curves of four different models

    Comparison of user’s accuracy on train and test sets using different network models

    Figure 10 a,b shows that the user’s accuracy of ResU-net and LU-net had a similar improvement in grasslands,crops,and buildings,which did not include unclassified areas.However,it is not as precise as the classification of harvested crops and river,indicating that the modification based separately on RCU or LCU still had some defects on classification in mixed forest-grassland ecosystems.At the same time,LRes-Unet produced the highest user’s accuracy,proving the positive effect of the combination of RCU and LCU on classification of land cover.Figure 10,d,proves that the above statement was still valid for unclassified areas and also recognizes the influence of unclassified area on each classification based on the four models.

    Effect of background area on overall accuracy

    Because of a few areas,the background effects had been ignored in previous studies (Cao and Zhang 2020;Zhang et al.2020c).In this study,the impacts of background on classification can be analyzed by the producer’s and user’s accuracy (Table 6).The accuracy (producer’s=0.90,user’s=0.94) of background area was higher than that of other classes,which led to a false improvement of overall accuracy.However,the results in Fig.10 exhibit drastically different user’s accuracy of background area based on U-net and ResU-net model.This difference was linked to the effect of LCU which can deepen the depth of image feature points and further improve classification accuracy.

    Fig.10 a User’s accuracy without unclassified area on training set using four network models; b user’s accuracy without unclassified area on test set using four network models; c user’s accuracy with unclassified area on training set using four network models; d user’s accuracy with unclassified area on test set using four network models

    Failure classification

    Figure 11 illustrates the error of classifying trees under shadows which is the most common classification failure in data sets.Environment problems from sampling and image mosaics were the main factors deteriorating the classification performance.Under low light or shadow conditions,the image features of some land cover change and further weaken similarities with other land covers in color level.In addition,the orthophoto obtained from the 3D reconstructed model may produce a blurry edge for images (Skabek et al.2020),which is likely to destroy classification.

    Conclusions

    Classifying land cover in a mixed forest-grassland ecosystem is a significant use of remote sensing technology,particularly from unmanned aerial vehicles (UAV)to manage forests and grasslands.This study presents a new method,LResU-net,to do land cover classification based on U-net,residual convolution and loop convolution network.On the basis of U-net,it adds RCU and LCU on U-net approach to improve the model and reduces the number of parameters and training time.Compared with other networks (U-net,ResU-net,LU-net),LResUnet has higher Kappa coefficients and greater accuracy in the entire data sets.The analysis of producer’s and user’s

    accuracy indicates that LResU-net had the favorable performance in various land covers.The result of classification was affected by unclassified areas,and a solution to some unclassified lands was found.The area of various land covers,which can be used for statistics and analysis of landform was calculated.However,this study does not include height data and future research should use the 3D reconstructed model to study height data of land cover classification.

    Fig.11 a UAV remote sensing image; b label image; c results of U-net classification; d results of LResU-net classification

    Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License,which permits use,sharing,adaptation,distribution and reproduction in any medium or format,as long as you give appropriate credit to the original author(s) and the source,provide a link to the Creative Commons licence,and indicate if changes were made.The images or other third party material in this article are included in the article’s Creative Commons licence,unless indicated otherwise in a credit line to the material.If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use,you will need to obtain permission directly from the copyright holder.To view a copy of this licence,visit http://creativecommons.org/licenses/by/4.0/.

    亚洲精品中文字幕在线视频| 一本大道久久a久久精品| 777久久人妻少妇嫩草av网站| 国产在线一区二区三区精| 99re6热这里在线精品视频| 一级,二级,三级黄色视频| 在线av久久热| 亚洲一区二区三区不卡视频| 自线自在国产av| 亚洲av成人av| 天堂√8在线中文| 99国产精品一区二区蜜桃av | 国产精品欧美亚洲77777| 看黄色毛片网站| 精品一区二区三区视频在线观看免费 | 亚洲精品久久午夜乱码| 麻豆乱淫一区二区| 国产淫语在线视频| 亚洲国产欧美一区二区综合| 国产高清videossex| 一本综合久久免费| 亚洲全国av大片| a级片在线免费高清观看视频| cao死你这个sao货| 十分钟在线观看高清视频www| 日日摸夜夜添夜夜添小说| 超碰97精品在线观看| 少妇被粗大的猛进出69影院| 久久亚洲真实| 久久久久精品国产欧美久久久| 色婷婷久久久亚洲欧美| 国产淫语在线视频| a在线观看视频网站| 国产精品永久免费网站| 亚洲一区中文字幕在线| 午夜福利一区二区在线看| 亚洲aⅴ乱码一区二区在线播放 | 国产精品一区二区精品视频观看| 91麻豆精品激情在线观看国产 | 国产高清videossex| 黄色成人免费大全| av视频免费观看在线观看| cao死你这个sao货| 男女高潮啪啪啪动态图| 国产亚洲一区二区精品| 五月开心婷婷网| 别揉我奶头~嗯~啊~动态视频| 下体分泌物呈黄色| 久久亚洲精品不卡| 男人的好看免费观看在线视频 | 午夜两性在线视频| 久久久久精品人妻al黑| 精品无人区乱码1区二区| 成人精品一区二区免费| 黑人巨大精品欧美一区二区蜜桃| 亚洲九九香蕉| 国产成人免费观看mmmm| 又黄又爽又免费观看的视频| 久久精品国产清高在天天线| 一级作爱视频免费观看| 亚洲av成人一区二区三| 国产男靠女视频免费网站| 午夜福利在线免费观看网站| 欧美丝袜亚洲另类 | 别揉我奶头~嗯~啊~动态视频| 中文字幕最新亚洲高清| 免费在线观看视频国产中文字幕亚洲| 韩国精品一区二区三区| 五月开心婷婷网| 热99re8久久精品国产| 91精品国产国语对白视频| 午夜精品国产一区二区电影| 国产av精品麻豆| 国产亚洲欧美98| 好看av亚洲va欧美ⅴa在| 国产精品影院久久| 亚洲av熟女| 欧美日韩亚洲国产一区二区在线观看 | 日韩制服丝袜自拍偷拍| 国产av精品麻豆| 久久久国产成人精品二区 | 日韩成人在线观看一区二区三区| 黄色视频,在线免费观看| 亚洲欧美激情综合另类| 757午夜福利合集在线观看| 一级a爱片免费观看的视频| 国产亚洲精品久久久久5区| 久久久久视频综合| 黄色成人免费大全| 在线天堂中文资源库| 免费在线观看黄色视频的| 男男h啪啪无遮挡| 男女午夜视频在线观看| 夜夜爽天天搞| 精品少妇久久久久久888优播| 老汉色∧v一级毛片| 人人妻人人爽人人添夜夜欢视频| 操出白浆在线播放| 99久久国产精品久久久| 美女福利国产在线| 国产不卡一卡二| a级毛片在线看网站| 91大片在线观看| 在线观看www视频免费| 亚洲av美国av| 极品人妻少妇av视频| 91九色精品人成在线观看| 亚洲精品在线观看二区| 日韩人妻精品一区2区三区| 中文字幕最新亚洲高清| 三上悠亚av全集在线观看| 天天添夜夜摸| 首页视频小说图片口味搜索| 下体分泌物呈黄色| 中文字幕人妻丝袜一区二区| 两人在一起打扑克的视频| 不卡av一区二区三区| 久久精品aⅴ一区二区三区四区| 中国美女看黄片| 亚洲在线自拍视频| 国产野战对白在线观看| 色精品久久人妻99蜜桃| 久久国产亚洲av麻豆专区| 婷婷精品国产亚洲av在线 | 在线观看66精品国产| 美女高潮喷水抽搐中文字幕| 色婷婷av一区二区三区视频| 久久国产精品影院| 女性生殖器流出的白浆| 自拍欧美九色日韩亚洲蝌蚪91| 97人妻天天添夜夜摸| 欧美激情 高清一区二区三区| 欧美成狂野欧美在线观看| 真人做人爱边吃奶动态| 一区福利在线观看| 久久九九热精品免费| 中文字幕av电影在线播放| 日本五十路高清| 大型av网站在线播放| 亚洲精品国产色婷婷电影| 身体一侧抽搐| 男女之事视频高清在线观看| 午夜免费鲁丝| 丁香欧美五月| 久久久国产成人免费| 色播在线永久视频| 在线永久观看黄色视频| 三级毛片av免费| 高清欧美精品videossex| 久久精品熟女亚洲av麻豆精品| 日韩大码丰满熟妇| 天天躁日日躁夜夜躁夜夜| 国产精品香港三级国产av潘金莲| 午夜精品久久久久久毛片777| 丝袜美足系列| 18禁国产床啪视频网站| 色婷婷av一区二区三区视频| 国产乱人伦免费视频| 亚洲精品美女久久久久99蜜臀| bbb黄色大片| 国产精品久久久久成人av| 人妻 亚洲 视频| 欧美日韩视频精品一区| 香蕉久久夜色| 99re6热这里在线精品视频| 性少妇av在线| 日韩中文字幕欧美一区二区| 丝袜美腿诱惑在线| 国产在线精品亚洲第一网站| 在线视频色国产色| 波多野结衣一区麻豆| 欧美日韩国产mv在线观看视频| 美女午夜性视频免费| 超色免费av| 99久久99久久久精品蜜桃| 午夜福利乱码中文字幕| 三级毛片av免费| 色老头精品视频在线观看| 亚洲第一青青草原| 午夜福利欧美成人| 国产国语露脸激情在线看| 国产精品乱码一区二三区的特点 | avwww免费| 中文欧美无线码| 色综合婷婷激情| 久久精品熟女亚洲av麻豆精品| 波多野结衣一区麻豆| 欧美乱色亚洲激情| 久久久久精品人妻al黑| 精品午夜福利视频在线观看一区| 欧美亚洲日本最大视频资源| av网站免费在线观看视频| 亚洲欧美日韩高清在线视频| 国产淫语在线视频| 亚洲欧美激情综合另类| 精品国产超薄肉色丝袜足j| 黄色视频,在线免费观看| 亚洲精品国产色婷婷电影| 老熟女久久久| 最新的欧美精品一区二区| 啦啦啦在线免费观看视频4| 亚洲欧美色中文字幕在线| 久久精品亚洲av国产电影网| 在线观看舔阴道视频| 精品人妻在线不人妻| 中文字幕人妻熟女乱码| 国产精品1区2区在线观看. | 久久 成人 亚洲| 日韩精品免费视频一区二区三区| 亚洲欧美激情综合另类| 丝袜在线中文字幕| 国产精品久久视频播放| 国产精品久久久久成人av| 国产精品二区激情视频| www.精华液| 伊人久久大香线蕉亚洲五| 一级片免费观看大全| 日日摸夜夜添夜夜添小说| 亚洲片人在线观看| 久久久国产成人精品二区 | 欧美日本中文国产一区发布| 免费av中文字幕在线| 国产在线一区二区三区精| 这个男人来自地球电影免费观看| 国产精品乱码一区二三区的特点 | 国产一卡二卡三卡精品| 精品午夜福利视频在线观看一区| 乱人伦中国视频| 午夜福利在线观看吧| 亚洲中文日韩欧美视频| 99精品欧美一区二区三区四区| 首页视频小说图片口味搜索| 黑人巨大精品欧美一区二区蜜桃| 男人操女人黄网站| 又紧又爽又黄一区二区| 日本a在线网址| 99精品在免费线老司机午夜| 日韩欧美三级三区| 欧美另类亚洲清纯唯美| 国产野战对白在线观看| 别揉我奶头~嗯~啊~动态视频| 欧美日韩精品网址| 亚洲精品中文字幕一二三四区| 日日夜夜操网爽| 成年动漫av网址| 视频区图区小说| 亚洲精品成人av观看孕妇| 国产精品av久久久久免费| 久久人人爽av亚洲精品天堂| 80岁老熟妇乱子伦牲交| 色综合欧美亚洲国产小说| 久久久久久人人人人人| 在线观看免费日韩欧美大片| 午夜激情av网站| 国产精品一区二区在线观看99| 久久ye,这里只有精品| 无遮挡黄片免费观看| 久久久国产成人免费| 国产精品1区2区在线观看. | 国产亚洲精品一区二区www | 777久久人妻少妇嫩草av网站| 国产男女超爽视频在线观看| 久久人人爽av亚洲精品天堂| 亚洲精品一二三| 国产成人欧美在线观看 | 亚洲精品乱久久久久久| 久久精品国产a三级三级三级| 91国产中文字幕| 老司机亚洲免费影院| 久久香蕉国产精品| 亚洲精品久久午夜乱码| 一本一本久久a久久精品综合妖精| 人妻久久中文字幕网| 黑人巨大精品欧美一区二区蜜桃| 欧美日韩一级在线毛片| 妹子高潮喷水视频| 丝袜人妻中文字幕| av免费在线观看网站| 精品国产乱码久久久久久男人| 久久青草综合色| 国产主播在线观看一区二区| 久久久久国产一级毛片高清牌| 伊人久久大香线蕉亚洲五| 国产精华一区二区三区| 日韩人妻精品一区2区三区| 咕卡用的链子| 90打野战视频偷拍视频| 亚洲色图综合在线观看| 亚洲久久久国产精品| 午夜久久久在线观看| 精品亚洲成a人片在线观看| tocl精华| 国产亚洲av高清不卡| 久久精品国产亚洲av高清一级| 精品高清国产在线一区| 欧美激情极品国产一区二区三区| 免费在线观看日本一区| a级片在线免费高清观看视频| 一级片'在线观看视频| 亚洲少妇的诱惑av| 国产麻豆69| 不卡一级毛片| 欧美乱妇无乱码| 国产精品1区2区在线观看. | 超碰成人久久| 性色av乱码一区二区三区2| 真人做人爱边吃奶动态| 国产精品.久久久| 国产成人av激情在线播放| 国产av一区二区精品久久| 少妇的丰满在线观看| 韩国精品一区二区三区| 国产成人一区二区三区免费视频网站| 自线自在国产av| 久久久国产成人免费| 中文亚洲av片在线观看爽 | 三上悠亚av全集在线观看| 国产色视频综合| 又紧又爽又黄一区二区| 欧美精品人与动牲交sv欧美| 亚洲色图av天堂| 欧美日韩乱码在线| 天天躁日日躁夜夜躁夜夜| 久久久国产精品麻豆| 日日夜夜操网爽| 日本精品一区二区三区蜜桃| 国产在线精品亚洲第一网站| 亚洲成av片中文字幕在线观看| 久久午夜综合久久蜜桃| 欧美黑人精品巨大| 国产精品 欧美亚洲| 日本欧美视频一区| 18禁裸乳无遮挡动漫免费视频| 老熟女久久久| 亚洲成a人片在线一区二区| 亚洲专区中文字幕在线| 国产成人av教育| 丝袜在线中文字幕| 久久午夜亚洲精品久久| 女人被躁到高潮嗷嗷叫费观| 国产极品粉嫩免费观看在线| 嫩草影视91久久| 别揉我奶头~嗯~啊~动态视频| 999久久久精品免费观看国产| 久久久久视频综合| 亚洲第一av免费看| 人妻久久中文字幕网| 丁香欧美五月| 亚洲avbb在线观看| 精品一品国产午夜福利视频| 色尼玛亚洲综合影院| 成年动漫av网址| 国产亚洲欧美精品永久| 成年动漫av网址| 国产男女内射视频| 久久久精品免费免费高清| 自拍欧美九色日韩亚洲蝌蚪91| 在线观看一区二区三区激情| 老司机福利观看| 日本五十路高清| 精品无人区乱码1区二区| 成人国产一区最新在线观看| 99热网站在线观看| 波多野结衣av一区二区av| 夜夜夜夜夜久久久久| 女人高潮潮喷娇喘18禁视频| 亚洲免费av在线视频| 人人妻,人人澡人人爽秒播| 久久这里只有精品19| 天天躁夜夜躁狠狠躁躁| 日本vs欧美在线观看视频| 久久人人97超碰香蕉20202| bbb黄色大片| 欧美+亚洲+日韩+国产| 无人区码免费观看不卡| videos熟女内射| 免费不卡黄色视频| 久久国产乱子伦精品免费另类| 夜夜夜夜夜久久久久| 亚洲精品在线观看二区| 国产成人欧美在线观看 | 国产精品免费大片| 久久精品国产a三级三级三级| 欧美精品一区二区免费开放| 每晚都被弄得嗷嗷叫到高潮| 亚洲国产欧美一区二区综合| 久久精品亚洲精品国产色婷小说| 日本a在线网址| 757午夜福利合集在线观看| 大型av网站在线播放| 老熟妇乱子伦视频在线观看| 久久精品91无色码中文字幕| 精品国产超薄肉色丝袜足j| 后天国语完整版免费观看| 欧美 日韩 精品 国产| 国产欧美日韩综合在线一区二区| 人人妻,人人澡人人爽秒播| 两人在一起打扑克的视频| 波多野结衣av一区二区av| 欧美一级毛片孕妇| 国产精品久久久人人做人人爽| 午夜成年电影在线免费观看| 在线免费观看的www视频| 国产精品一区二区免费欧美| 日本黄色日本黄色录像| 亚洲成人手机| 51午夜福利影视在线观看| 热99国产精品久久久久久7| 亚洲aⅴ乱码一区二区在线播放 | 亚洲久久久国产精品| 亚洲 欧美一区二区三区| 五月开心婷婷网| 91在线观看av| 亚洲精品中文字幕一二三四区| 亚洲午夜精品一区,二区,三区| 亚洲av欧美aⅴ国产| 亚洲精品一卡2卡三卡4卡5卡| 99riav亚洲国产免费| 不卡一级毛片| x7x7x7水蜜桃| 日韩 欧美 亚洲 中文字幕| 熟女少妇亚洲综合色aaa.| 一区二区日韩欧美中文字幕| 精品久久久久久久久久免费视频 | 午夜亚洲福利在线播放| 老熟妇乱子伦视频在线观看| 国产精品偷伦视频观看了| 12—13女人毛片做爰片一| 国产麻豆69| 在线视频色国产色| 高清在线国产一区| 精品人妻熟女毛片av久久网站| 亚洲 欧美一区二区三区| 国产成人精品在线电影| 女性被躁到高潮视频| 日韩人妻精品一区2区三区| 亚洲中文字幕日韩| bbb黄色大片| 免费黄频网站在线观看国产| 欧美国产精品一级二级三级| 99精品欧美一区二区三区四区| 久久 成人 亚洲| 亚洲九九香蕉| av免费在线观看网站| 成人手机av| 中文字幕制服av| 国产成人免费观看mmmm| 人妻 亚洲 视频| 精品一区二区三区av网在线观看| 成人免费观看视频高清| 欧美精品啪啪一区二区三区| a级毛片在线看网站| 久久精品熟女亚洲av麻豆精品| 欧美成人免费av一区二区三区 | 久久久久视频综合| 一进一出好大好爽视频| 一级毛片精品| 18禁美女被吸乳视频| 777久久人妻少妇嫩草av网站| 一区二区三区精品91| 999久久久国产精品视频| 亚洲一区高清亚洲精品| 一区二区日韩欧美中文字幕| 免费观看精品视频网站| 免费观看人在逋| 九色亚洲精品在线播放| 69精品国产乱码久久久| 极品少妇高潮喷水抽搐| www.精华液| 天堂动漫精品| 日韩免费高清中文字幕av| 人成视频在线观看免费观看| 婷婷成人精品国产| 欧美黄色淫秽网站| 亚洲视频免费观看视频| 热re99久久国产66热| 在线看a的网站| 国产精品99久久99久久久不卡| 两性夫妻黄色片| 欧美+亚洲+日韩+国产| 欧美一级毛片孕妇| 亚洲av美国av| 91九色精品人成在线观看| 久久亚洲真实| 国产高清视频在线播放一区| 欧美一级毛片孕妇| 男女床上黄色一级片免费看| 黑人操中国人逼视频| 老鸭窝网址在线观看| 999久久久精品免费观看国产| 免费在线观看亚洲国产| 国产主播在线观看一区二区| 日日夜夜操网爽| 男女之事视频高清在线观看| 9热在线视频观看99| 亚洲人成电影观看| av线在线观看网站| 免费久久久久久久精品成人欧美视频| www.999成人在线观看| 亚洲一区二区三区不卡视频| 最新的欧美精品一区二区| 大型黄色视频在线免费观看| 免费久久久久久久精品成人欧美视频| 日本黄色日本黄色录像| 免费日韩欧美在线观看| 青草久久国产| 女人高潮潮喷娇喘18禁视频| 国产高清videossex| videosex国产| 亚洲精品一卡2卡三卡4卡5卡| 欧美日韩一级在线毛片| 美女午夜性视频免费| 亚洲欧美日韩高清在线视频| 久久人妻熟女aⅴ| 一区二区三区激情视频| 久久久久久亚洲精品国产蜜桃av| 日韩欧美一区二区三区在线观看 | 深夜精品福利| 女人爽到高潮嗷嗷叫在线视频| 中亚洲国语对白在线视频| 久久久久久久国产电影| 一级片'在线观看视频| 久久久久精品人妻al黑| 757午夜福利合集在线观看| 欧美日韩视频精品一区| 亚洲第一av免费看| 精品国产乱子伦一区二区三区| 久久中文字幕人妻熟女| 午夜两性在线视频| 最新在线观看一区二区三区| 亚洲精品一二三| 欧美日韩国产mv在线观看视频| 免费看十八禁软件| 丝瓜视频免费看黄片| 国产男女内射视频| 国产精品.久久久| 午夜激情av网站| bbb黄色大片| 久热这里只有精品99| 日日爽夜夜爽网站| 69av精品久久久久久| 老司机靠b影院| 成年女人毛片免费观看观看9 | 法律面前人人平等表现在哪些方面| 日韩三级视频一区二区三区| 手机成人av网站| 国产欧美日韩一区二区精品| 亚洲国产看品久久| 免费一级毛片在线播放高清视频 | 亚洲七黄色美女视频| 亚洲三区欧美一区| 亚洲免费av在线视频| 欧美黄色片欧美黄色片| 99国产精品一区二区三区| 成人黄色视频免费在线看| 国产xxxxx性猛交| 亚洲欧美日韩高清在线视频| 51午夜福利影视在线观看| 美女视频免费永久观看网站| 高清av免费在线| 国产亚洲欧美在线一区二区| 亚洲专区国产一区二区| e午夜精品久久久久久久| 午夜福利,免费看| 欧美人与性动交α欧美软件| 免费观看a级毛片全部| av欧美777| 看片在线看免费视频| 1024视频免费在线观看| 一进一出抽搐gif免费好疼 | 午夜两性在线视频| 十分钟在线观看高清视频www| 亚洲久久久国产精品| 亚洲一区中文字幕在线| 国产成人一区二区三区免费视频网站| 久久天堂一区二区三区四区| 无限看片的www在线观看| 一区在线观看完整版| 亚洲精品久久成人aⅴ小说| 色综合婷婷激情| 在线观看免费日韩欧美大片| 国产又爽黄色视频| 在线观看免费高清a一片| 又大又爽又粗| 亚洲美女黄片视频| 脱女人内裤的视频| 老鸭窝网址在线观看| av中文乱码字幕在线| 人人妻人人添人人爽欧美一区卜| 久久99一区二区三区| 欧美+亚洲+日韩+国产| 久久人妻av系列| 久久99一区二区三区| 欧美+亚洲+日韩+国产| 久久国产精品人妻蜜桃| 国产成人精品无人区| 老熟女久久久| 淫妇啪啪啪对白视频| 啪啪无遮挡十八禁网站| 午夜免费成人在线视频| 精品一区二区三区av网在线观看| 人成视频在线观看免费观看| 国产男靠女视频免费网站| 亚洲精品久久午夜乱码| 色精品久久人妻99蜜桃| 午夜福利视频在线观看免费| av国产精品久久久久影院| 色精品久久人妻99蜜桃| 波多野结衣av一区二区av| 啦啦啦在线免费观看视频4| 18在线观看网站| 啦啦啦免费观看视频1| 亚洲国产欧美网| tocl精华| 日韩欧美一区二区三区在线观看 |