• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Enhanced segmentation of gastrointestinal polyps from capsule endoscopy images with artifacts using ensemble learning

    2022-11-23 05:25:46JunXiaoZhouZhanYangDingHaoXiShouJunDaiZhiQiangFengJunYanLiWeiXuHongWang
    World Journal of Gastroenterology 2022年41期

    Jun-Xiao Zhou, Zhan Yang, Ding-Hao Xi, Shou-Jun Dai, Zhi-Qiang Feng, Jun-Yan Li, Wei Xu, Hong Wang

    Abstract

    Key Words: Artifacts; Capsule endoscopy; Polyps; Ensemble learning; Segmentation; Robustness

    INTRODUCTION

    Colorectal cancer (CRC) is the second leading cause of death in the United States[1]. In China, an estimated 1101653 new cancer cases and 709529 cancer deaths from gastric cancer and CRC will occur in 2022, placing China first worldwide because of its large population[1]. Although other gastrointestinal lesions, such as erosions and ulcers, can also develop into cancers, most gastrointestinal cancers arise from precancerous polyps, which are the most common lesions found on endoscopy[2]. Therefore, early detection and removal of gastrointestinal polyps under endoscopy are critical for preventing gastrointestinal cancers[3-6]. Traditional gastroenteroscopy is widely used for the clinical assessment of gastrointestinal lesions. However, there are still some deficiencies, such as invasiveness and incomplete inspection of the site[7]. Additionally, some patients with small bowel diseases who have contraindications or are averse to undergoing gastroenteroscopy are more likely to use safer and non-invasive capsule endoscopy (CE) for visual examination of the digestive tract[8,9]. CE usually takes 8-12 h, which is not only time-consuming but also highly operator-dependent[10,11]. Otherwise, deep learning (DL)has greatly improved the sensitivity and specificity of CE for polyp detection while saving time[12].Studies have indicated that for every 1% increase in the detection rate of colorectal adenoma, the risk of CRC can decrease by 3%[4]. However, inadequate intestinal cleansing can produce various artifacts,such as motion blur, specular reflections, bubbles, and debris (Figure 1), which can interfere with image reading, reduce the detection rate of polyps, cause patients to miss treatment, and increase the risk of tumor development[13,14]. In addition, high-quality and clear standard datasets that rarely appear in clinical practice are often used in these studies[15-17], and the intestinal lumen is often fully dilated in these images. This is significantly different from CE images with natural contraction of the intestinal lumen, which can present various artifacts (Figure 1). Therefore, these methods are often less effective in clinical practice. Hence, identifying gastrointestinal polyps and other lesions to the maximum extent when the gastrointestinal tract is insufficiently cleansed and dilated with interference factors, such as fecal residue, cloudy liquid, and bubbles in the lumen, is one of the biggest challenges in the application of artificial intelligence (AI) for CE in clinical practice and is of great concern to clinicians.

    Figure 1 Examples of artifact-infested endoscopy images used in our current study. A: Presence of cloudy liquid and specular reflections; B:Presence of bubbles and low contrast between lesion tissue and normal tissues; C: Presence of fecal residue; D: Presence of bubbles and specular reflections.

    Figure 2 Semantic segmentation problem in the field of computer vision.

    Currently, DL is a popular topic in the field of AI. It is based on the construction of computational models by simulating the neural network structure of the human brain[18]. Semantic segmentation is a part of DL algorithms that segments different objects according to each marked pixel in an image[19](Figure 2). Some studies have proposed semantic segmentation models for medical images, such as SegNet[20], U-Net[21], Attention-UNet[22], Resnet-UNet[23], and HarDMSEG[24]. These studies have shown the significant superiority of various types of medical image semantic segmentation, as well as the feasibility of these models in tests with standard datasets. To improve the robustness of these models, researchers have begun to apply ensemble learning to medical image segmentation, not through a single model, but by combining several basic models to ensure the best prediction performance[25-27]. However, AI currently has limited ability to identify intestinal lesions with insufficient cleansing. For example, the detection rate of polyps in CE with a clean intestinal tract is significantly higher than that in CE with a dirty intestinal tract. In clinical practice, intestinal cleansing is not always performed well, and may not generate a clean image. Additionally, each patient has factors that can affect the identification by AI, such as insufficient intestinal distension, intestinal fecal residues,liquid residues, and air bubbles, resulting in the insufficient actual use of AI in clinical studies and low reliability.

    In the present study, we combined semantic segmentation and ensemble learning methods for the first time to analyze CE images with artifacts. We then compared the performance of the ensemble and single models to further improve the detection rate of polyps. Our results demonstrate that ensemble learning can be used to reduce the influence of artifacts, which has a significant auxiliary role in the detection of polyps in clinical practice. To the best of our knowledge, this is the first study to propose the use of ensemble learning and semantic segmentation to reduce the negative impact of artifacts on model performance in clinical practice. Overall, our current findings have instructive significance for improving the analysis of medical images with artifacts in clinics.

    MATERIALS AND METHODS

    This retrospective study was approved by the Ethics Committee of Guangzhou First People’s Hospital.All images were collected from videos of Ankon. This study has no conflicts of interest and did not receive any funding.

    Data preparation

    We collected 277 polyp CE images with artifacts selected from 5760 h of videos from 480 patients suffering from gastrointestinal disorders who received CE at Guangzhou First People’s Hospital from January 2016 to December 2019. The selection criteria for the experimental images were as follows: (1)The lumen on the picture was in a natural contraction state; (2) Images of the digestive tract with polyps; and (3) Artifacts in the lumen, such as feces, motion blur, specular reflections, bubbles, and debris. The polyps in these experimental images were verified for authenticity by using a large number of clear videos and photos containing the polyps or double-balloon enteroscopy. Additionally, to ensure the accuracy and rigor of the data annotation, the image data were obtained by an experienced gastroenterologist who watched the video recordings, extracted the frames where the polyps were captured through ES Navi, and annotated the pixel points of the polyp lesions using Labelme. Next, the annotated polyp profiles were carefully reviewed by two other experienced gastroenterologists. The processing time for each patient’s video was approximately 4-5 h. Before applying the dataset in the experiments, we cut off the black boxes of the images that typeset the patient’s name and other information to obtain 512 × 512 images.

    The other class of data comprehended publicly available high-quality datasets with images that rarely have artifacts and included the CVC_Colon[16] dataset (created by the Computer Vision Center and Computer Science Department, Universitat Autònoma de Barcelona) and the CVC_Clinic[17] dataset(captured by the Hospital Clinic, Barcelona, Spain and labeled by the Computer Vision Center,Barcelona, Spain). CVC_Colon provided 380 colonoscopy images containing polyps with a frame size of 500 × 574 pixels. Similarly, the CVC-Clinic contained 612 still images with a size of 288 × 384 from 29 different sequences. Both datasets are frequently used in gastrointestinal endoscopic computer-assisted polyp detection studies, and several representative studies have used these datasets in their experiments.

    When using these datasets, we cropped or padded the edges of the images for two reasons. First,black edges or information, such as patient and time on the edges of the images, have no effect on the polyp region segmentation. Second and the main reason is that when we cross-sectionally compared various base learners in our experiments, the convolution and pooling designs of some of them were found to be more suitable for images whose length and width were both divisible by powers of two.Therefore, to minimize the changes in the hyperparameters of these base learners, we cropped or padded the input images to match the model hyperparameter design. Finally, the GZ_Capcam dataset contains 277 images of size 512 × 512, the CVC_Clinic contains 612 images of size 288 × 384, and the CVC_Colon dataset contains 380 images of size 512 × 576; all images are eight-bit three-channel color images[17]. All images used in this study contained at least one polyp class, including the standard datasets.

    Snapshot ensemble method

    In supervised learning problems, we always expect to obtain models that perform well and are stable in all aspects; however, owing to the presence of randomness, the trained models are not always ideal, and the models obtained always have prediction preferences. The main goal of ensemble learning is to combine weak models to build a more integrated and comprehensive model that integrates the strengths of weak models. The snapshot ensemble method is a type of ensemble learning for DL models and was used in the present study[28].

    In the DL method, the model parameters are adjusted according to the gradient of the objective function, as shown in Formula 1:

    The parameters of the model take a step in the direction of the gradient descent at each iteration, and the size of the step depends on both the size of the current gradient and the learning rate, as shown in Formula 1, where θtdenotes the model parameters in time step t and α denotes the learning rate.Usually, to speed up convergence and prevent DL models from repeatedly jumping at different local optima during training, the learning rate decays as the number of iterations increases, eventually causing the model to fall into a certain local optimum and not jump out. The core idea of the snapshot ensemble method is to restart the learning rate when it decays to less than a certain threshold so that the model jumps out of the current local optimum and finds a new local optimum nearby and converges,and Formula 2 and Figure 3 show the specific changes in the learning rate:

    Figure 3 Learning rate for each epoch in the snapshot ensemble learning, with γ and α0 set for 0.3, M for 10, and T for 75.

    Where α0, γ, M, and T represent the initial learning rate, learning rate decay rate, number of epochs per learning rate decay, and number of epochs per learning rate restart cycle, respectively. In the snapshot ensemble method, the model that is at the local optimum before each restart learning rate is recorded as a weak model, and in the end, the prediction results of multiple weak models are integrated by ensemble voting. In the process of training, we set the number of learning rate restart cycles to 13, the learning rate decay rate to 0.3, and the number of epochs per learning rate decay to 10 and perform a total of 75 epochs in each cycle,i.e., 0.3 for γ, 10 for M, and 75 for T in Formula 2. In other words, the learning rate is reduced to 0.3 of the previous value every 10 epochs of training and reverts to the initial learning rate setting of 0.3 after 75 epochs. The model parameters that perform best on the validation set are retained in these 75 epochs as the parameters of the weak model. The entire training process lasted for 13 cycles, that is, we ended up with 13 weak learners. In the integration phase of weak models, we selected three, five, and seven weak learners with the best performance on the validation set and obtained the prediction results of the ensemble model by vote ensemble. All computational processes,including data pre-processing, model training, validation, and testing, were performed through Python programming. We built the model using PyTorch, and all experiments were based on an NVIDIA Titan V GPU. Figure 4 shows the change in validation loss in the experiment with the UNet model on the CVC_Colon dataset. The light pink line indicates the epochs of the restart learning rate, and the red points indicate the epochs of preserving the weak models.

    State-of-the-art segmentation models

    To show that the ensemble classification is effective in improving the segmentation in comparison with the single model when dealing with medical images with artifacts, and to illustrate the generality of its enhancement effect, we used five existing state-of-the-art (SOTA) segmentation models as base learners;SegNet[20], which is proposed to solve the deep network model of image semantic segmentation for autonomous driving or intelligent robots, and is mainly based on full convolutional networks; U-Net[21], which performs well on neuron structure segmentation datasets with only a small number of annotations, and is a basic solution for medical image analysis of small datasets; Attention-UNet[22],which is an improved model based on U-Net, and achieves performance beyond that of the U-Net model for semantic segmentation of human organs on abdominal three-dimensional computed tomography scans; ResNet-UNet[23], which is also an improved version of U-Net, and gets outstanding performance on the public challenge of identifying pneumothorax diseases on chest X-rays; and HarDMSEG[24], which is an efficient image segmentation model, and achieves SOTA level in terms of both computational efficiency and analytical accuracy, in comparison experiments to illustrate that ensemble learning method improves their analysis capability in the face of images with artifacts.

    Setup of comparison experiments

    First, we randomly divided the experimental and public data into training (195 images), validation (41 images), and testing (41 images) sets. The model was trained using the training set, and the best model was selected for the final test on the validation set to ensure that the model did not overfit the final test data. Finally, we tested the model using the testing set. For each model and dataset pair, multiple cycles of the learning rate restart were performed during the training phase. The best-performing model,evaluated using the validation set data in each restart learning rate cycle, was retained as a weak model.Finally, every weak model and the strong model comprehending several of the best weak models were evaluated using the validation set and tested using the testing set. Figure 5 shows the overall experimental design.

    Figure 4 Validation loss in training.

    Outcome measures

    The intersection over union (IoU) and dice coefficients (Figure 6) are the most widely used metrics for semantic segmentation problems[29-31]. Both metrics measure the similarity between the sets of real and predicted regions. The calculation process is illustrated in Figure 6, where the area of intersection denotes the number of pixels in the intersection between the prediction area and ground truth, and the area of the union denotes that of the union. These two metrics were used to assess the performance of the segmentation models.

    RESULTS

    In summary, we performed two sets of comparison experiments using SOTA base models for the two types of datasets. In the first experiment, we compared the performance of the ensemble learning model with that of single models on a dataset with artifacts. In the second experiment, we compared the performance of the single models with that of the ensemble model on high-quality datasets without artifacts. Finally, we compared the differences between the improvements of the ensemble learning method for datasets with and without artifacts.

    Comparison between the ensemble model and single models

    First, we compared the performance of the ensemble learning model with that of single models on CE images for all the five aforementioned base learners. A total of 41 images from the test dataset of GZ_Capcam were used for the final test. These test images were used only in the final testing phase to avoid data leakage and the consequent erroneous evaluation of the models. To illustrate that the ensemble learning model improves the performance of the single model on the artifact-infected dataset,we replicated all base learners mentioned in the previous section to illustrate the robustness of the conclusions in this study.

    For U-Net, three test samples in the GZ-Hospital dataset were selected to compare the performances of the single and ensemble models (Figure 7). We can see that the semantic segmentation model was affected by different noises, such as stains, blurs, and light-dark variations in the sample images, leading to results that were not always clear. However, the performance of the ensemble model often met or exceeded the best results of a single model, indicating that a model constructed based on ensemble learning can effectively mitigate the effects of artifacts on the performance of the semantic segmentation model.

    Figure 5 Train-test workflow of the proposed method.

    Figure 6 Illustration of intersection over union and dice metrics. IoU: Intersection over union; FN: False negative; TP: True positive; FP: False positive.

    The results for the GZ_Capcam dataset are presented in Table 1, which includes images rich in artifacts. The IoU and dice metrics were calculated, as previously described. The performances of the single and ensemble models on the test set are presented in Supplementary Table 1. The results for all five basic learners on the CE dataset showed that the ensemble model outperformed the single models.Compared with single models, specifically, on the dataset with artifacts, the ensemble learning models with SegNet, U-Net, Attention-UNet, Resnet-UNet, and HarDMSEG as the base learners improved the detection by 0.08%, 7.01%, 3.88%, 5.13%, and 2.22%, respectively, using the IoU metric, and 1.71%,4.93%, 1.40%, 2.86%, and 0.61%, respectively, using the dice metric. Overall, the ensemble model outperformed the single models. The performance of a truly single model, that is, a model obtained from a single training validation, was consistently worse than that of the ensemble model, as shown in the results for the weak models excluding the best one.

    Comparisons using datasets without artifacts

    Similarly, we checked the performance of the single and ensemble models using standard datasets(Figure 8, Tables 2 and 3). The performances of the single and ensemble models on the test set are presented in Supplementary Tables 2 and 3. By comparing the results presented in Figures 7 and 8, we found that the ensemble learning method can improve the robustness of the semantic segmentation model when the dataset is affected by artifacts.

    Table 1 Comparison between the single and ensemble models on GZ_Capcam (artifact-affected) datasets

    Table 2 Comparison between the single and ensemble models on CVC_Colon (clear) datasets

    Table 3 Comparison between the single and ensemble models on CVC_Clinic (clear) datasets

    DISCUSSION

    In the present study, we demonstrated that current computer-aided medical image analysis methods performed poorly in the presence of artifacts that were previously ignored. Nevertheless, almost every patient presents with insufficient intestinal cleansing. Thus, we used ensemble learning to improve the existing AI models and enhance their robustness in dealing with images with artifacts. Previous studies have extensively analyzed and concluded that integrated learning methods improve the robustness of medical image classification in a credible manner[32]. By improving the segmentation performance of the model, we can separate polyps more accurately from surrounding tissues, which can improve the detection probability of polyps and aid in monitoring the size of polyps in patients with unresectable polyps[33-36]. Semantic segmentation provides pixel-level classification and clearer polyp boundaries,which are also crucial in surgical procedures or radiofrequency ablation, and is expected to be used for real-time detection of polyp boundaries in surgical resection under gastroenteroscopy to assist polyp resection[33,36]. More in-depth studies have shown that the noise immunity of single models is weaker than that of integrated learning models[37], and clinical images, such as the CE images used in this study, are not always perfect in terms of image quality.

    Figure 7 Comparison between the ensemble and single models on artifact-affected images. Origin represents the images taken by the endoscope.Ground truth represents the anomalous locations marked by experts based on the original images. Val. 1st best to Val. 5th best represents the results of the five bestperforming single models evaluated on the validation set. 5-ensemble represents the results of the ensemble model based on these single models.

    Figure 8 Comparison between the ensemble and single models on standard (clear) images. Ground truth represents the anomalous locations marked by experts based on the original images. Val. 1st best to Val. 5th best represents the results of the five best-performing single models evaluated on the validation set. 5-ensemble represents the results of the ensemble model based on these single models.

    We used CE image datasets as samples, mainly because CE is an increasingly widely used and safe form of endoscopy but also has many artifacts[15,38]. The ensemble learning approach was tested for 15 pairs, consisting of three datasets and five SOTA segmentation models. The results showed that for CE images with various artifacts, ensemble learning improved the analytical performance of AI models.Herein, we demonstrated that ensemble learning can reduce the influence of artifacts on the semantic segmentation of CE images, which might also apply to other medical images.

    In general, artifacts are prevalent in medical images and seriously challenge the performance of existing computer-aided diagnosis (CAD) models; therefore, this study discusses the enhancement of ensemble learning methods for CAD models to analyze images with artifacts, mainly using CE images as an example in the experiments. In addition, our experiments did not involve the injection of a priori knowledge of gastroenterology; in other words, the use of the ensemble learning approach mentioned in this paper does not imply any additional workload or workflow reordering. The only additional cost associated with the method is the computational resources. Additionally, from the perspective of DL,better model performance often relies on more model parameters and computational resources.Methods that already use DL models can easily apply ensemble learning methods to improve model performance without the need for additional workflow tuning. It is worth mentioning that, although the ensemble learning approach can improve the robustness of CAD image analysis models, misuse may lead to a less-than-expected improvement in the model’s effectiveness, mainly because the essence of ensemble learning is to reduce model variance, and when the variance of a single model is already very low, the improvement brought by ensemble learning may be very limited.

    From the experimental results, although the ensemble learning approach improves the performance of the segmentation model on the dataset with artifacts, there are still false-positive and false-negative cases. On the one hand, the main reason for false-negative cases was that the model confuses normalcolor polyps with normal gastrointestinal folds or confuses abnormal-color polyps with artifacts, such as yellow bubbles. However, the main cause of false-positive cases was that some artifacts or normal folds had a high similarity with polyps in the image, which led the model to misidentify them as polyps.Overall, the main reason for segmentation errors is that the color and texture are highly confusing, and we will further attempt to improve the ability of the model to distinguish polyps, normal tissues, and artifacts in a subsequent study.

    In clinical practice, video frames can be completely infested with artifacts, making the content of the image simply unrecognizable. Therefore, the appearance of these frames is inevitable in clinical practice.In the present study, we confirmed the authenticity of polyps in pictures with artifacts by using more images, videos, and other inspection methods. Thus, we solved the dilemma of applying AI to these medical images. However, our study has some limitations. For example, the images were insufficient and did not involve lesions other than polyps.

    We believe that the direction of feature AI for CE imaging research lies in making existing computer models better serve clinical diagnosis in a practical sense rather than letting these methods stay in the laboratory. CE is commonly used to examine digestive diseases. In addition to polyps, many digestive diseases can be detected using CE. Thus, AI for CE imaging can be considered to enrich the diagnosis,localization, and grading of more forms of the disease, such as ulcers and erosions, to assist doctors in more refined disease research and diagnosis. In the future, we will validate the ensemble learning method in clinical practice to demonstrate that it can improve the detection rate of polyps in CE in the clinic and evaluate the potential of this method for other types of medical images or lesions[39].

    CONCLUSION

    Ensemble learning can improve the semantic segmentation performance of AI models on CE images with artifacts.

    ARTICLE HIGHLIGHTS

    Research conclusions

    The ensemble learning method can improve the performance of semantic segmentation of polyps in CE images with artifacts.

    Research perspectives

    We will validate other digestive tract lesions and other medical images and perform real-time detection during endoscopic and surgical procedures.

    ACKNOWLEDGEMENTS

    The authors would like to acknowledge Ai-Hua Lin for assistance with statistical analysis.

    FOOTNOTES

    Author contributions:Zhou JX collected and compiled the data; Yang Z and Xi DH were responsible for coding the model; Zhou JX, Yang Z, and Li J drafted the manuscript; Dai SJ, Feng ZQ, and Wang H performed the capsule endoscopy examinations and reviewed the images; Xu W and Wang H contributed to the study design and manuscript writing.

    Institutional review board statement:This study was reviewed and approved by the Ethics Committee of the Guangzhou First People’s Hospital.

    Informed consent statement:Patients were not required to give informed consent to the study because the analysis used anonymous clinical data that were obtained after each patient agreed to treatment by written consent.

    Conflict-of-interest statement:All the authors report no relevant conflicts of interest for this article.

    Data sharing statement:No additional data are available.

    Open-Access:This article is an open-access article that was selected by an in-house editor and fully peer-reviewed by external reviewers. It is distributed in accordance with the Creative Commons Attribution NonCommercial (CC BYNC 4.0) license, which permits others to distribute, remix, adapt, build upon this work non-commercially, and license their derivative works on different terms, provided the original work is properly cited and the use is noncommercial. See: https://creativecommons.org/Licenses/by-nc/4.0/

    Country/Territory of origin:China

    ORCID number:Jun-Xiao Zhou 0000-0002-3968-103X; Zhan Yang 0000-0002-7949-5653; Ding-Hao Xi 0000-0002-1685-8066; Shou-Jun Dai 0000-0002-6253-2585; Zhi-Qiang Feng 0000-0002-0478-7520; Jun-Yan Li 0000-0002-1392-5796; Wei Xu 0000-0001-7992-9063; Hong Wang 0000-0001-7299-1884.

    S-Editor:Wang JJ

    L-Editor:Wang TQ

    P-Editor:Wang JJ

    欧美日韩瑟瑟在线播放| 国产一区二区三区视频了| 每晚都被弄得嗷嗷叫到高潮| 欧美午夜高清在线| 一个人免费在线观看电影 | 欧美日韩亚洲国产一区二区在线观看| 欧美 亚洲 国产 日韩一| 日本免费一区二区三区高清不卡| 久久久久久久午夜电影| 欧美日韩一级在线毛片| 制服人妻中文乱码| 日本在线视频免费播放| 男女床上黄色一级片免费看| 波多野结衣高清无吗| 久久精品人妻少妇| 老司机在亚洲福利影院| 国产99白浆流出| 身体一侧抽搐| 全区人妻精品视频| 日韩中文字幕欧美一区二区| 在线观看www视频免费| 亚洲成人久久性| 好男人电影高清在线观看| 亚洲精品在线观看二区| 天天躁狠狠躁夜夜躁狠狠躁| av天堂在线播放| 97人妻精品一区二区三区麻豆| 亚洲国产精品sss在线观看| 国产高清激情床上av| 中文字幕熟女人妻在线| 国产真实乱freesex| 级片在线观看| 中文亚洲av片在线观看爽| 少妇人妻一区二区三区视频| 国产黄色小视频在线观看| 非洲黑人性xxxx精品又粗又长| 国内精品久久久久精免费| 成人特级黄色片久久久久久久| 国产不卡一卡二| 两个人视频免费观看高清| 首页视频小说图片口味搜索| 757午夜福利合集在线观看| 男女午夜视频在线观看| 性色av乱码一区二区三区2| 国产麻豆成人av免费视频| 50天的宝宝边吃奶边哭怎么回事| 国产精品综合久久久久久久免费| 哪里可以看免费的av片| 在线看三级毛片| 亚洲成av人片在线播放无| 18禁观看日本| 欧美日韩亚洲综合一区二区三区_| 久久这里只有精品19| 亚洲av中文字字幕乱码综合| 人人妻人人澡欧美一区二区| 亚洲七黄色美女视频| 欧美3d第一页| 午夜激情av网站| 亚洲片人在线观看| 高清毛片免费观看视频网站| 日韩大尺度精品在线看网址| 嫩草影视91久久| 国产亚洲欧美在线一区二区| 国产亚洲欧美98| 韩国av一区二区三区四区| 99国产极品粉嫩在线观看| 国产一区二区三区在线臀色熟女| 久久久国产成人精品二区| 国产伦在线观看视频一区| 此物有八面人人有两片| 欧美日韩乱码在线| 日本在线视频免费播放| 黄色 视频免费看| 长腿黑丝高跟| 可以在线观看的亚洲视频| 国产人伦9x9x在线观看| 中文亚洲av片在线观看爽| 他把我摸到了高潮在线观看| 在线观看免费日韩欧美大片| 亚洲人成网站高清观看| 三级国产精品欧美在线观看 | 琪琪午夜伦伦电影理论片6080| 欧美乱妇无乱码| 中文字幕最新亚洲高清| 国产亚洲av高清不卡| 国产高清激情床上av| 亚洲五月天丁香| 婷婷丁香在线五月| 老司机在亚洲福利影院| 亚洲成人国产一区在线观看| 亚洲一卡2卡3卡4卡5卡精品中文| 看片在线看免费视频| 好男人电影高清在线观看| 九九热线精品视视频播放| 免费在线观看亚洲国产| 亚洲中文日韩欧美视频| 国产真人三级小视频在线观看| 欧美黑人巨大hd| 法律面前人人平等表现在哪些方面| 欧美 亚洲 国产 日韩一| 日韩成人在线观看一区二区三区| av视频在线观看入口| 久久久久性生活片| 国产免费男女视频| 亚洲国产看品久久| 一边摸一边抽搐一进一小说| 日韩欧美国产在线观看| 亚洲 欧美 日韩 在线 免费| 免费高清视频大片| 18禁黄网站禁片午夜丰满| 精品久久久久久久久久免费视频| 欧美成人午夜精品| 久久精品影院6| 国产欧美日韩一区二区精品| 搡老熟女国产l中国老女人| 神马国产精品三级电影在线观看 | 国产精品1区2区在线观看.| 少妇人妻一区二区三区视频| 日韩大码丰满熟妇| 校园春色视频在线观看| 成人高潮视频无遮挡免费网站| 亚洲熟女毛片儿| 在线观看一区二区三区| 啪啪无遮挡十八禁网站| 怎么达到女性高潮| 村上凉子中文字幕在线| 午夜福利视频1000在线观看| 欧美高清成人免费视频www| 亚洲乱码一区二区免费版| 成人午夜高清在线视频| 最新美女视频免费是黄的| 成人av一区二区三区在线看| 69av精品久久久久久| 成人18禁高潮啪啪吃奶动态图| 麻豆久久精品国产亚洲av| 欧美一区二区国产精品久久精品 | 黄色 视频免费看| 精品久久久久久久毛片微露脸| 久久久久国产一级毛片高清牌| 欧美日韩精品网址| 亚洲中文av在线| 国产区一区二久久| 亚洲中文字幕日韩| 99国产综合亚洲精品| 国产亚洲欧美98| 亚洲国产欧美网| 18禁观看日本| 国产私拍福利视频在线观看| 国产91精品成人一区二区三区| 黄片小视频在线播放| 欧美zozozo另类| 黄色片一级片一级黄色片| 日本在线视频免费播放| 99久久99久久久精品蜜桃| 三级男女做爰猛烈吃奶摸视频| av天堂在线播放| 99在线人妻在线中文字幕| 一级片免费观看大全| 久久久久性生活片| 黑人欧美特级aaaaaa片| 亚洲午夜理论影院| 日本黄大片高清| av国产免费在线观看| 亚洲中文av在线| 亚洲五月婷婷丁香| 美女扒开内裤让男人捅视频| 黄色a级毛片大全视频| ponron亚洲| 露出奶头的视频| 狂野欧美激情性xxxx| 国产精品99久久99久久久不卡| 国产成年人精品一区二区| 日本一区二区免费在线视频| 白带黄色成豆腐渣| videosex国产| 两个人看的免费小视频| 搡老岳熟女国产| 青草久久国产| 搞女人的毛片| 哪里可以看免费的av片| 日本在线视频免费播放| 午夜影院日韩av| 亚洲五月天丁香| 看片在线看免费视频| www日本黄色视频网| 十八禁人妻一区二区| 亚洲国产欧美网| 国产免费av片在线观看野外av| 操出白浆在线播放| 国产视频一区二区在线看| 午夜福利免费观看在线| 最新美女视频免费是黄的| 女人高潮潮喷娇喘18禁视频| 啪啪无遮挡十八禁网站| 中文资源天堂在线| 99久久精品热视频| 亚洲欧美激情综合另类| 精品国产亚洲在线| 亚洲一码二码三码区别大吗| 免费在线观看日本一区| 国产免费av片在线观看野外av| bbb黄色大片| 色老头精品视频在线观看| 国产精品一区二区三区四区久久| 精品欧美一区二区三区在线| 久久久久久九九精品二区国产 | 精品少妇一区二区三区视频日本电影| 国产亚洲精品久久久久5区| 国产精品一区二区免费欧美| 美女午夜性视频免费| 午夜精品在线福利| 老熟妇乱子伦视频在线观看| 亚洲18禁久久av| 国产区一区二久久| 精品电影一区二区在线| 一本久久中文字幕| 少妇熟女aⅴ在线视频| 国产精品一区二区三区四区久久| 成人18禁在线播放| 色在线成人网| 99久久国产精品久久久| 亚洲国产高清在线一区二区三| 国产精品自产拍在线观看55亚洲| 午夜精品一区二区三区免费看| 精华霜和精华液先用哪个| 久久午夜亚洲精品久久| 亚洲午夜精品一区,二区,三区| 亚洲熟妇熟女久久| 高清在线国产一区| 国产又色又爽无遮挡免费看| 97超级碰碰碰精品色视频在线观看| 午夜两性在线视频| 69av精品久久久久久| 免费一级毛片在线播放高清视频| 深夜精品福利| 亚洲欧洲精品一区二区精品久久久| 波多野结衣高清无吗| 国产成人影院久久av| 91九色精品人成在线观看| 亚洲av美国av| 法律面前人人平等表现在哪些方面| 波多野结衣高清无吗| 午夜免费激情av| 超碰成人久久| 最近视频中文字幕2019在线8| 成人18禁在线播放| 村上凉子中文字幕在线| 亚洲人成伊人成综合网2020| 我的老师免费观看完整版| 欧美成人性av电影在线观看| 国产亚洲精品综合一区在线观看 | 99久久99久久久精品蜜桃| 亚洲成人国产一区在线观看| svipshipincom国产片| 日本五十路高清| 久久久久九九精品影院| 成人高潮视频无遮挡免费网站| 国产精品九九99| 级片在线观看| 国产精品av视频在线免费观看| 非洲黑人性xxxx精品又粗又长| 亚洲成人免费电影在线观看| 日韩av在线大香蕉| 老司机深夜福利视频在线观看| 哪里可以看免费的av片| 淫妇啪啪啪对白视频| 日韩 欧美 亚洲 中文字幕| 可以在线观看的亚洲视频| 一边摸一边做爽爽视频免费| 亚洲国产精品成人综合色| 9191精品国产免费久久| 两性午夜刺激爽爽歪歪视频在线观看 | 午夜精品在线福利| 午夜福利成人在线免费观看| 精品久久久久久久人妻蜜臀av| 国产亚洲av高清不卡| 88av欧美| 亚洲人成电影免费在线| 久久热在线av| 免费在线观看影片大全网站| 深夜精品福利| 女警被强在线播放| 一级毛片高清免费大全| 精品一区二区三区视频在线观看免费| 国产蜜桃级精品一区二区三区| 午夜免费激情av| 精品第一国产精品| 午夜两性在线视频| 九色成人免费人妻av| 成人av在线播放网站| 91在线观看av| 天天躁狠狠躁夜夜躁狠狠躁| 亚洲国产欧美一区二区综合| 亚洲天堂国产精品一区在线| 国产一区二区三区视频了| 国产三级黄色录像| 久久久精品国产亚洲av高清涩受| 一本大道久久a久久精品| 免费av毛片视频| 欧美人与性动交α欧美精品济南到| 在线a可以看的网站| 最近视频中文字幕2019在线8| 国产v大片淫在线免费观看| а√天堂www在线а√下载| 一个人观看的视频www高清免费观看 | 一边摸一边做爽爽视频免费| 老汉色av国产亚洲站长工具| 欧美色欧美亚洲另类二区| 国产精品乱码一区二三区的特点| 亚洲国产欧美人成| 免费一级毛片在线播放高清视频| 亚洲熟妇中文字幕五十中出| 国产在线精品亚洲第一网站| 国产私拍福利视频在线观看| 亚洲国产精品久久男人天堂| 美女午夜性视频免费| 欧美3d第一页| 成人三级黄色视频| 久久久国产成人精品二区| 亚洲精品在线观看二区| 97超级碰碰碰精品色视频在线观看| 国内少妇人妻偷人精品xxx网站 | 精品高清国产在线一区| 视频区欧美日本亚洲| 精品久久久久久久久久免费视频| 国产高清视频在线观看网站| 校园春色视频在线观看| 日本 av在线| 亚洲av电影不卡..在线观看| 国内揄拍国产精品人妻在线| √禁漫天堂资源中文www| 国产精品久久久av美女十八| av片东京热男人的天堂| 久久国产乱子伦精品免费另类| 99久久国产精品久久久| 国产精品影院久久| 在线看三级毛片| 国产三级中文精品| 非洲黑人性xxxx精品又粗又长| 精品一区二区三区视频在线观看免费| 亚洲色图 男人天堂 中文字幕| 黑人巨大精品欧美一区二区mp4| 男女之事视频高清在线观看| 亚洲国产精品合色在线| 一夜夜www| 又黄又粗又硬又大视频| 丝袜美腿诱惑在线| 亚洲va日本ⅴa欧美va伊人久久| 免费观看人在逋| 香蕉久久夜色| 国产伦人伦偷精品视频| 国产午夜精品论理片| 天天添夜夜摸| 老司机深夜福利视频在线观看| 麻豆国产97在线/欧美 | 国产蜜桃级精品一区二区三区| 亚洲精品国产一区二区精华液| 亚洲成人中文字幕在线播放| 蜜桃久久精品国产亚洲av| av欧美777| 精品久久久久久成人av| 国产主播在线观看一区二区| 亚洲国产精品久久男人天堂| 国产私拍福利视频在线观看| 国产成人av激情在线播放| 国产91精品成人一区二区三区| 级片在线观看| 十八禁网站免费在线| 啦啦啦观看免费观看视频高清| 亚洲成人国产一区在线观看| 亚洲欧美日韩高清在线视频| 哪里可以看免费的av片| 人人妻人人看人人澡| 国产一区二区三区在线臀色熟女| 又黄又粗又硬又大视频| 亚洲国产精品成人综合色| 亚洲色图av天堂| av超薄肉色丝袜交足视频| 白带黄色成豆腐渣| 日韩高清综合在线| 妹子高潮喷水视频| 最近在线观看免费完整版| 日本黄大片高清| 欧美人与性动交α欧美精品济南到| 每晚都被弄得嗷嗷叫到高潮| 精品无人区乱码1区二区| 午夜免费成人在线视频| 国产v大片淫在线免费观看| 国产真人三级小视频在线观看| 亚洲精品国产一区二区精华液| 久久精品aⅴ一区二区三区四区| 99热这里只有精品一区 | 麻豆国产97在线/欧美 | 国产精品日韩av在线免费观看| 中文字幕高清在线视频| 国内毛片毛片毛片毛片毛片| 国产精品国产高清国产av| 成人午夜高清在线视频| 精品少妇一区二区三区视频日本电影| 一本一本综合久久| 此物有八面人人有两片| 精品久久蜜臀av无| 国内揄拍国产精品人妻在线| 久久人妻av系列| 亚洲18禁久久av| 亚洲 欧美一区二区三区| 变态另类丝袜制服| 日韩欧美在线乱码| 国产高清有码在线观看视频 | 久久中文字幕人妻熟女| 国产成+人综合+亚洲专区| 制服丝袜大香蕉在线| 欧美激情久久久久久爽电影| 精品久久久久久久末码| 身体一侧抽搐| 在线看三级毛片| 国产野战对白在线观看| 国产97色在线日韩免费| 国产主播在线观看一区二区| 久久午夜综合久久蜜桃| АⅤ资源中文在线天堂| 国产熟女xx| 叶爱在线成人免费视频播放| 岛国在线观看网站| 国产精品 国内视频| 亚洲专区国产一区二区| 欧美成狂野欧美在线观看| 亚洲成人久久性| 精品无人区乱码1区二区| 色在线成人网| 在线播放国产精品三级| 黄色视频不卡| 免费一级毛片在线播放高清视频| 国产成人影院久久av| 女同久久另类99精品国产91| 国产黄a三级三级三级人| 黄色毛片三级朝国网站| 国产伦人伦偷精品视频| 日本成人三级电影网站| 在线观看免费午夜福利视频| 精品久久久久久,| 蜜桃久久精品国产亚洲av| 丰满人妻一区二区三区视频av | 十八禁人妻一区二区| 色哟哟哟哟哟哟| 国产精品美女特级片免费视频播放器 | 日韩 欧美 亚洲 中文字幕| 国产av一区在线观看免费| www.999成人在线观看| 精品高清国产在线一区| 亚洲av熟女| 一级黄色大片毛片| 亚洲成人久久爱视频| 国产男靠女视频免费网站| 亚洲av电影不卡..在线观看| 国产成年人精品一区二区| 天堂av国产一区二区熟女人妻 | 国产成人精品久久二区二区免费| 丝袜美腿诱惑在线| 变态另类成人亚洲欧美熟女| 真人一进一出gif抽搐免费| 人人妻,人人澡人人爽秒播| 久久久久国产一级毛片高清牌| 色综合亚洲欧美另类图片| 50天的宝宝边吃奶边哭怎么回事| 黄色视频,在线免费观看| 亚洲av五月六月丁香网| 国产av又大| 国产男靠女视频免费网站| 熟女少妇亚洲综合色aaa.| 亚洲男人天堂网一区| 午夜日韩欧美国产| 大型av网站在线播放| 亚洲 欧美 日韩 在线 免费| aaaaa片日本免费| 色噜噜av男人的天堂激情| 欧美久久黑人一区二区| 亚洲性夜色夜夜综合| 老司机在亚洲福利影院| 亚洲一卡2卡3卡4卡5卡精品中文| 成人三级黄色视频| 黑人欧美特级aaaaaa片| 俄罗斯特黄特色一大片| 日本成人三级电影网站| 黄色片一级片一级黄色片| 18禁黄网站禁片免费观看直播| www.熟女人妻精品国产| 久久久久久久午夜电影| 手机成人av网站| 国产欧美日韩一区二区精品| 成人三级黄色视频| 久久久久久久久免费视频了| 亚洲av电影在线进入| 亚洲国产精品合色在线| 国产欧美日韩一区二区精品| 美女大奶头视频| 伦理电影免费视频| 琪琪午夜伦伦电影理论片6080| 亚洲最大成人中文| 香蕉国产在线看| 亚洲美女黄片视频| 中文字幕最新亚洲高清| 久久久精品欧美日韩精品| 色老头精品视频在线观看| 精品欧美国产一区二区三| 久久久国产欧美日韩av| 国产aⅴ精品一区二区三区波| 亚洲无线在线观看| 国产乱人伦免费视频| 在线观看日韩欧美| 色噜噜av男人的天堂激情| 啦啦啦观看免费观看视频高清| 国产亚洲精品久久久久5区| 国产精品久久久久久久电影 | 后天国语完整版免费观看| 日韩成人在线观看一区二区三区| 日本一本二区三区精品| 国产精品av久久久久免费| 欧美成人午夜精品| 国产亚洲欧美98| 国产精品免费一区二区三区在线| 女人爽到高潮嗷嗷叫在线视频| 18美女黄网站色大片免费观看| 国产爱豆传媒在线观看 | 99在线人妻在线中文字幕| 日本熟妇午夜| 亚洲一卡2卡3卡4卡5卡精品中文| 国产一区在线观看成人免费| 啦啦啦韩国在线观看视频| 两个人视频免费观看高清| 久久婷婷成人综合色麻豆| 精品欧美一区二区三区在线| 老汉色av国产亚洲站长工具| 日韩精品中文字幕看吧| 黄片大片在线免费观看| 精品国产超薄肉色丝袜足j| 成人av在线播放网站| www.自偷自拍.com| 真人一进一出gif抽搐免费| 精品一区二区三区视频在线观看免费| 在线观看午夜福利视频| 亚洲中文字幕日韩| 在线观看免费日韩欧美大片| 夜夜看夜夜爽夜夜摸| 亚洲五月婷婷丁香| 不卡一级毛片| 久9热在线精品视频| 亚洲全国av大片| 午夜激情福利司机影院| 亚洲美女黄片视频| 色av中文字幕| 久久天堂一区二区三区四区| 国产精品一区二区免费欧美| 亚洲精品美女久久av网站| 男男h啪啪无遮挡| 国产精品av久久久久免费| 在线十欧美十亚洲十日本专区| 色播亚洲综合网| 欧美成人午夜精品| 国产欧美日韩一区二区精品| 97碰自拍视频| 狂野欧美激情性xxxx| 亚洲片人在线观看| 正在播放国产对白刺激| 亚洲18禁久久av| 中文字幕av在线有码专区| 欧美日韩乱码在线| 国产精品99久久99久久久不卡| 久久久久九九精品影院| 欧美av亚洲av综合av国产av| 欧美最黄视频在线播放免费| 亚洲aⅴ乱码一区二区在线播放 | 久久久国产成人精品二区| 亚洲av成人av| 国语自产精品视频在线第100页| 精品一区二区三区四区五区乱码| 国产精品免费一区二区三区在线| 两性午夜刺激爽爽歪歪视频在线观看 | 老熟妇乱子伦视频在线观看| 美女 人体艺术 gogo| 黄片小视频在线播放| 欧美性长视频在线观看| 女警被强在线播放| 亚洲成人国产一区在线观看| 亚洲va日本ⅴa欧美va伊人久久| 日韩有码中文字幕| 香蕉av资源在线| 欧美 亚洲 国产 日韩一| 欧美日韩福利视频一区二区| 国产精品久久久久久精品电影| 搡老熟女国产l中国老女人| 给我免费播放毛片高清在线观看| 亚洲国产看品久久| 91在线观看av| 亚洲欧美日韩高清专用| 亚洲美女视频黄频| 亚洲成av人片在线播放无| 一本一本综合久久| 男女做爰动态图高潮gif福利片| 好看av亚洲va欧美ⅴa在| 后天国语完整版免费观看| 久久精品国产亚洲av香蕉五月| 91九色精品人成在线观看| 麻豆成人午夜福利视频| 一进一出好大好爽视频| 国产伦在线观看视频一区| 99精品久久久久人妻精品| 午夜免费观看网址| 欧美乱色亚洲激情| 麻豆成人午夜福利视频| 一进一出好大好爽视频| 亚洲欧美日韩东京热| 精品一区二区三区四区五区乱码| 国产主播在线观看一区二区| 国产伦在线观看视频一区|