• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Comparative Study of Transfer Learning Models for Retinal Disease Diagnosis from Fundus Images

    2022-03-14 09:26:40KunthaPinJeeHoChangandYunyoungNam
    Computers Materials&Continua 2022年3期

    Kuntha Pin,Jee Ho Chang and Yunyoung Nam

    1Department of ICT Convergence,Soonchunhyang University,Asan,31538,Korea

    2Department of Ophthalmology, Soonchunhyang University Bucheon Hospital,Soonchunhyang University College of Medicine,Bucheon,420-767,Korea

    3Department of Computer Science and Engineering,Soonchunhyang University,Asan,31538,Korea

    Abstract:While the usage of digital ocular fundus image has been widespread in ophthalmology practice, the interpretation of the image has been still on the hands of the ophthalmologists which are quite costly.We explored a robust deep learning system that detects three major ocular diseases:diabetic retinopathy (DR), glaucoma (GLC), and age-related macular degeneration(AMD).The proposed method is composed of two steps.First, an initial quality evaluation in the classification system is proposed to filter out poorquality images to enhance its performance, a technique that has not been explored previously.Second,the transfer learning technique is used with various convolutional neural networks(CNN)models that automatically learn a thousand features in the digital retinal image,and are based on those features for diagnosing eye diseases.Comparison performance of many models is conducted to find the optimal model which fits with fundus classification.Among the different CNN models, DenseNet-201 outperforms others with an area under the receiver operating characteristic curve of 0.99.Furthermore,the corresponding specificities for healthy,DR,GLC,and AMD patients are found to be 89.52%,96.69%,89.58%,and 100%,respectively.These results demonstrate that the proposed method can reduce the time-consumption by automatically diagnosing multiple eye diseases using computer-aided assistance tools.

    Keywords: Multiclass classification; deep neural networks; glaucoma; agerelated macular degeneration; diabetic retinopathy; transfer learning; quality evaluation

    1 Introduction

    Diabetic retinopathy (DR), glaucoma (GLC), and age-related macular degeneration (AMD)are leading causes of vision loss worldwide, and their effects will continue to increase in the absence of rapid detection [1-3].DR, GLC, and AMD are common causes of blindness with different damage areas such as retinal, vascular, optic nerve, and macular.An estimated 64.3,425, and 11.26 million people had GLC, DR, and AMD, respectively in 2013-2020 [4-6]; these numbers are predicted to increase to 112 (GLC), 642 (DR), and 18.57 (AMD) million by 2040.Ocular fundus imaging plays a crucial role in the diagnosis and tracking of ophthalmologic diseases.The number of patients is rapidly increasing, causing a burden on clinics by requiring numerous ophthalmologists, specialized equipment, and health care services.This burden can be alleviated by using an automated system.In recent days, deep learning techniques have been applied for automatic detection of eye diseases [7-9].The detection results indicated high sensitivity and specificity in diagnosis.

    Although deep learning techniques have advanced to state-of-the-art levels in fundus image classification, a large amount of training data and computational time remain a constraint in the classification process.

    Transfer learning is the process of learning a new task through the transfer of knowledge from a previous task that has already been learned.It helps achieve better performance with less training time and an acceptable performance when working with limited data.In order to alleviate the aforementioned constraints, transfer learning algorithm plays an important role in the application of the features learned on one task to another (fundus images).Transfer learning can be employed to use the features acquired from previously trained neural networks on general images as knowledge in fundus image classification.Recently, one or two eye disease categories have been introduced by applying transfer learning as feature extraction and classifiers [10-12].

    The deep neural network (DNN) contains large layers and filters to calculate and distinguish various features of fundus images for identifying each eye disease.However, blurry, uneven illumination and excessively dark or bright pixels damage the fundus image and become hurdles for the automatic diagnosis systems, resulting in misdiagnosis or the inability to detect the disease.Therefore, the image quality becomes a critical property of fundus images and an indispensable initialization process in fundus image classification.In previous studies, integrated quality evaluations had not been conducted to eliminate poor quality images prior to the eye disease diagnosis process.Only quality appraisals or fundus classifications have been performed separately.Therefore, this is a new study on integrated quality assessment in multiclass fundus image classification.

    The aim of this study is to explore the usefulness of transfer learning techniques in identifying major ophthalmologic diseases, namely, DR, GLC, and AMD in normal eyes.In this study,image-preprocessing techniques are applied to prepare images before feeding to a neural network.Transfer learning techniques transfer the learned features trained on one task (thousands of general images) to classify the second task (fundus images).Finally, the results with and without the quality evaluation are compared.

    2 Related Work

    Various DNNs have been used in automatic fundus-photograph classification systems to detect eye diseases.A combination of different layers and operations establishes a convolutional neural network (CNN) architecture.Shankar et al.[13] developed a DNN named synergic deep learning to classify various severity levels of DR fundus images.Gómez-Valverde et al.[12] proposed a CNN that contains 13 layers for training and grading five levels of fundus images (no DR, mild DR, moderate DR, severe DR, and proliferative DR).Wan et al.[14] employed deep learning with transfer learning by using four different CNN models for training and extracting feature images to classify DR.The results of DR image classification showed the highest accuracy of 95.68% among the four models in their study.van Grinsven et al.[15] applied data augmentation to prevent overfitting when training fundus images using their proposed CNN model.The result of detecting hemorrhage in DR achieved an area under the receiver operating characteristic curve(AUC) of 0.972.

    GLC is the second major cause of vision loss [16].Ophthalmologists have used fundus photographs to examine changes in the optic disc, while GLC was present.Automated systems have been proposed to detect GLC in fundus images to obtain an accurate diagnosis by using a deep learning approach.Raghavendra et al.[17] proposed a DNN architecture contain 18 layers to distinguish between normal/healthy eye (NR) and GLC.The proposed DNN model showed an accuracy of 98.13%.In [18], transfer learning techniques were applied to assess only one eye disease, AMD or NR; and result achieved an AUC of 0.995 for glaucoma and non-glaucoma,which demonstrated that the diagnosis system performed adequately.

    3 Methods and Experiments

    3.1 Original Fundus Image Dataset

    In this study, retinal images were obtained from [19], in which, a new retinal image dataset,called the Diabetic Retinopathy Image Database (DRIMDB), was created.All retinal images of the DRIMDB were graded by an expert into three classes: 125 as good, 69 as bad, and 22 as outliers.The images classified as good qualified for automatic retinal image analysis.Images of all three grades were captured by using a Canon CF-60UVi Fundus Camera at a 60-degree field of view and stored in JPEG format with a resolution of 570×760 pixels.The performance of our proposed quality evaluation was assessed on the DRIMDB dataset, excluding outliers.

    Fundus images of eye diseases were obtained from the Soonchunhyang Bucheon Hospital(SCH-BH) dataset.Ophthalmologists classified all fundus images of the dataset into four classes:DR, AMD, GLC, and NR.From among a total of 1304 digital fundus images in the SCH-BH database, 312 were classified as DR, 254 as GLC, 119 as AMD, and 619 as NR.Furthermore,images were acquired by using a Kowa VX-10 digital fundus camera (Kowa Company.Ltd., Aichi,Japan) at a 60-degree field of view and a resolution of 8575×5695 pixels and 24 bits per pixel(standard RGB).The institutional review board approved this work (approval no.SCHBC 2019-02-008-002).

    3.2 Datasets and Image Preprocessing

    Fundus images of the DRIMDB were processed to resize the original image to a resolution of 299×299 pixels, which is a suitable size for inputting into the Inception-v3 model [20].As datasets are small, data augmentation has become a popular technique for producing various images in experiments to make the model more generalized [21].In this study, images were subjected to data augmentation techniques, such as horizontal flip, vertical flip, zoom-in, zoom-out, and contrast,for which the alpha parameter was 0.58-1.06.From among a total of 2910 augmentation images,1875 were good, and 1035 were bad.Furthermore, 80% of the augmented images were used for training and 20% for validation to perform the quality evaluation.

    Before images were input into the DNN model, each eye-disease fundus image was resized to 256×256 pixels to decrease the computation time.Moreover, another preprocessing technique,contrast limited adaptive histogram equalization (CLAHE), was applied to the resized images.In the CLAHE process, fundus images were converted to CIELAB color space to extract lightness using an open-source library feature [22].CLAHE was performed within a clip limit of 2 and a slicing grid of eight kernel sizes.

    The processed images from the SCH-BH fundus dataset were utilized for the experimentation and evaluation of the classification system.The size of the training and validation images was increased by random augmentation, such as horizontal flip, vertical flip, and brightness shift in the range of 0.5-1.25 [23].The augmentation-brightness process was repeated twice for DR and five times for AMD for each image to balance the data in each class.Each augmented image had the same label (DR, GLC, AMD, or NR) as the original input image from the source.

    3.3 The Proposed Deep Learning Model for Fundus Images Quality Evaluation

    The proposed deep learning model was based on the addition of four layers to the Inceptionv3 model.The Inception-v3 model comprises 48 layers, pre-trained on thousands of general images (1000 categories) from the ImageNet dataset.This structure was modified because the original structure was developed to classify 1000 categories.To employ Inception-v3 networks with quality evaluation (good and bad grade), four layers, namely, AveragePooling2D with a downscale of 8×8, Dropout (30%), Flatten, and Dense (2 nodes), were added, as shown in Fig.1.

    Figure 1: Inception-v3 model with attaching new layers

    3.4 Deep Learning Model and Transfer Learning

    To improve fundus image classifciation, various CNN architectures were used for comparison:VGG-19, VGG-16, MobileNet, MobileNet-V2, ResNet-50, InceptionResNet-V2, Inception-v3, and DenseNet-201.These architectures are commonly used to classify general images or medical images, and their results are comparable to those of other architectures (AlexNet and GoogLeNet)selected for the experiment [24-26].Five layers (Flatten:1 and Dense:4) were added to each CNN architecture as the classification layer to create a novel architecture for classifying fundus images.VGG-19, VGG-16, MobileNet, MobileNet-V2, ResNet-50, InceptionResNet-V2, Inception-v3, and DenseNet-201 were trained on a large general dataset (ImageNet database) for classifying 1000 categories of images, and the knowledge previously learned is the feature extraction for a new dataset, SCH-BH.Pre-training of deep learning trains only the newly added classification layers freezes all extraction layers, and uses previous work as knowledge for grading fundus images(see Fig.2); this takes less time than training all neural network layers.VGG-19 and VGG-16 architecture, as implemented here, comprises 19 layers with 54,269,380 training parameters and 16 layers with 48,959,684 training parameters, respectively.MobileNet and MobileNetV2 consist of 28 layers with 71,028,292 training parameters and 53 layers with 86,834,628 training parameters, respectively.The number of layers and parameters of Inception-v3 is similar to that of MobileNetV2, which contains 48 layers with 97,990,820 parameters.ResNet50 integrates 158,496,004 training parameters, which are larger than those of the other models used in this study.DenseNet201 is a larger layer with more training parameters than the models used, with 201 layers and 144,841,668 total training parameters.Tab.1 shows the total number of training parameters and the number of layers of each CNN architecture with the same input image size used in this study.

    Figure 2: Transfer learning for fundus images classification

    Table 1: Training parameters and number of layers of the CNN architectures used in this study

    Five layers are added to each of the CNN architectures as follows:

    o The native output parameters of the CNN architecture were converted to data in a onedimensional array for input to the next layer by using flattening, which is connected to the ultimate classification layers.

    o Three Dense layers with rectified linear (ReLu) activation functions were fused to perform a matrix-vector multiplication and produce an output with 1024, 512, and 256 dimensions,respectively.

    o The output of the previous layer (256 dimensions) is randomized, and 30% of the layer’s neurons are rejected to prevent overfitting.

    o After the Dropout layer, the Dense layer with ReLu was applied to produce an output of 128 dimensions.

    o Finally, the Dense layer is implemented with the softmax activation function to construct an output array of four dimensions, which indicate the probability of DR, GLC, AMD,and NR.

    3.5 Model Training and Selection

    In this experiment, two tasks were performed sequentially: first, fundus image quality evaluation was performed, and images were segregated into good and bad quality; second, the good quality images were used in fundus classification to diagnose three eye diseases.

    3.5.1 Fundus Image Quality Evaluation Training

    The implemented model is trained by fine-tuning the pre-trained Inception-v3 network.Transfer learning is applied, and the original weight of the pretraining is fine-tuned for two classes of quality classification.The training is conducted only on the four newly added layers.Furthermore,the augmentation images are split into 80% training, and 20% for validation, and the hyperparameters are set with a batch size of 32 images and 50 epochs.This evaluation is conducted to assess the grade of the fundus image (SCH-BH dataset) to distinguish between good and poor quality.

    3.5.2 Fundus Images Evaluation Training

    Each CNN model was trained and evaluated on two levels: the overall image of the SCH-BH and the good images after applying the quality assessment.The model was trained by fine-tuning the pre-trained CNN models.The CNN models are VGG-19, VGG-16, MobileNet, MobileNet-V2, ResNet-50, InceptionResNet-V2, Inception-v3, and DenseNet-201.There are many possible techniques for splitting the data during the testing and training phases.According to Breiman et al.[27], the K-fold cross-validation technique performed better than leave-one-out.Therefore,5-fold cross-validation was applied where the given dataset was split into five sections/folds, and testing was set at some point of each fold.In the first iteration, the first fold was used to test the model, and the rest were used to train the model.In the second iteration, the second fold was used as the testing set, while the rest served as the training set.The operation was repeated until each fold of the five folds completed the testing set.Furthermore, the hyperparameter for training was 1000 epochs with a batch size of 32 images.Early stopping was monitored on validation loss during training to avoid overfitting of the training deep learning model; the training stopped if the validation loss did not improve by 0.05 for 20 epochs.A summary of the hyperparameter configuration of the training CNN network is presented in Tab.2.This process was repeated for each architecture (VGG-19, VGG-16, MobileNet, MobileNet-V2, ResNet-50, InceptionResNet-V2, Inception-v3, and DenseNet-201), pre-training, early stopping, cross-validation fold, and hyperparameter settings.

    In order to perform the classification, TensorFlow [28] and Scikit-Learn [29] were employed to train and evaluate the proposed fundus image classification system.Each model was trained on a Windows 10 operating system with an Intel(R) Xeon (R) Silver 4114@2.20 GHz CPU, 192GB RAM, NVIDIA TITAN RTX 119 GB GPU.

    Table 2: Hyperparameter configuration

    3.6 Model Evaluation

    Accuracy, sensitivity, and specificity are statistical analyses that are widely used to evaluate the performance of a diagnostic system [30].The area under macro average of ROC curve (macro-AUC) and confusion matrices (overall classes) were calculated by using Python with 3.7.9 version and scikit-learn 0.24.1.The output prediction of one fundus image from the classification is the probability of DR, GLC, AMD, or NR.The confusion matrices of each class were extracted to assess the performance of each class of multiclass classification.The confusion matrices provide the number of true positives (TP), false positives (FP), true negatives (TN), and false negatives(FN).

    4 Results

    4.1 Fundus Image Quality Evaluation

    The fundus image quality evaluation was trained on an open dataset (DRIMDB), which obtained a better performance in training with an accuracy of 97.93% and validation accuracy of 96.97%.The quality of images in the SCH-BH dataset was evaluated using our proposed fundus image quality evaluation method before performing disease classification.Of the 1340 images processed through the quality assessment system, the proposed method distinguished 1207 as good quality or suitable images (DR: 295, GLC: 217, AMD: 114, NR: 581) and 97 as bad or of unsuitable grading.Images classified as good by our proposed algorithm contain normal color, sufficient features, and proper structure, which made them acceptable for analysis [31].The features, such as blood vessels, optic disc, and macula, on good images, are clearly expressed,as illustrated in Fig.3.The bad quality or ungradable fundus images are in four ungradable categories [32], blurry, lash artifact, poor optic disc visibility, and uneven illumination as shown in Fig.4.The ungradable predicted images, as shown in Figs.4a-4c, which are blurry images, lash artifact, and poor optic disc visibility, affect the presentation of features in fundus photographs,especially blood vessels, optic disc, and sight-threatening features.The demonstration of features helps identify various eye diseases.Fig.4d shows uneven illumination; some parts of the fundus images are dark, making all of the information insufficient for identification.

    Figure 3: Example of good quality images assessed by the proposed method.(a) DR class (b)GLC class (c) AMD class (d) normal class

    Figure 4: Example of poor-quality images assessed by the proposed method.(a) blurry images (b)lash artifact (c) poor optic disc visibility (d) uneven illumination

    4.2 Fundus Image Assessment Without Initial Fundus Image Quality Evaluation

    Fundus images of the SCH-BH dataset were used to improve multiclass classification.Tab.3 shows the performance of the different CNN architectures without quality evaluation in terms of accuracy, sensitivity, and specificity.MobileNet, ResNet-50, and DenseNet-201 models are the top-three models, outperforming five of the CNN models evaluated.MobileNet achieved 79.07% accuracy with 79.03% sensitivity and 94.90% specificity for DR, 37.37% sensitivity and 91.82% specificity for GLC, 39.13% sensitivity and 98.30% specificity for AMD, and 88.62%sensitivity and 80.00% specificity for the NR class.ResNet-50 achieved 78.68% accuracy, 82.26%sensitivity, and 94.41% specificity for DR, 40.59% sensitivity and 84.08% specificity for GLC,30.44% sensitivity and 98.72% specificity for AMD, and 84.55% sensitivity and 86.67% specificity for NR.For DenseNet-201%, 78.68% accuracy, 88.71% sensitivity, and 95.92% specificity for DR,32.38% sensitivity and 88.24% specificity for GLC, 34.78% sensitivity and 98.30% specificity for AMD, and 86.18% sensitivity and 81.48% specificity for the NR class.Moreover, we also evaluated the models’performances by utilizing the ROC curve (see Fig.4), along with the macro-AUC values.A comparison of the macro-AUC values showed that the VGG-19 with a macro-AUC of 0.94 performs better than other deep learning architectures.

    Table 3: Performances of fundus image classification without applying quality-evaluation.The highest values among eight models are denoted in bold

    4.3 Fundus Images Evaluation with Initial Quality Fundus Evaluation(Quality Images Filtering)

    All the qualified images that were evaluated by our proposed method were utilized to assess the performance of the eight CNN architectures.Tab.4 shows the performance of the different CNN architectures with quality evaluation in terms of accuracy, sensitivity, and specificity.Fig.5 illustrates the macro-ROC curve and AUC values of all models.The AUC was 0.99, 0.96, and 0.95 of DenseNet-201, ResNet-50, and MobileNet, respectively.These are the top three models in terms of AUC values.

    Table 4: Performances of fundus image classification with applying quality-evaluation.The highest values among eight models are denoted in bold

    Figure 5: ROC curves of fundus image evaluation without applying QET

    In addition, Tab.5 shows a comparison of the performance between applying quality evaluation and without applying.The performance of top-three models (MobileNet, ResNet50, and DenseNet201) of implementing quality assessment outperform fundus classification without performing quality assessment (except sensitivity for NR of MobileNet, and sensitivity for GLC and specificity for NR of ResNet50 model).The ResNet-50 achieved an average accuracy of 86.25% of the fuse quality evaluation, this accuracy being higher than the accuracy of multi-class classification with none assessing (78.68%).The multiclass classification with excluding filtering got an average accuracy of 77.91%, 77.13%, 79.07%, 72.87%, 78.68%, 75.58%, 71.71%, and 78.68% of VGG-19, VGG-16, for MobileNet, MobileNet-V2, ResNet-50, InceptionResNet-V2 model, Inception-V3, and DenseNet-201, respectively.For multi-class classification with filtering unacceptable images achieved average accuracy of 82.90%, 82.50%, 82.92%, 80.42%, 86.25%,77.92%, 79.16%, 85.83% of VGG-19, VGG-16, among MobileNet, MobileNet-V2, ResNet-50,InceptionResNet-V2, Inception-V3, and DenseNet-201 model, respectively.

    Table 5: Comparison performances of applying quality-evaluation (QET) and without QET.The higher values of comparison QET and without QET for each model are denoted in bold

    (Continued)

    Table 5: Continued

    5 Discussion and Conclusion

    The evaluation of the quality of fundus is proposed to assess the appropriate images for eye-diseases analysis.Poor quality images can be detected and excluded by the proposed method(as shown in the Results section), allowing deep learning to be used to analyze fundus images.By comparing macro-AUC values without QET (Fig.5) and the macro-AUC values with QET(Fig.6), it is found that the macro-AUC value of each CNN architecture is higher with QET.

    Figure 6: ROC curves of fundus image evaluation with applying QET

    This study is the first to simultaneously provide integrated quality assessment in multiclass fundus image classification and three eye disease classifications.The previous studies have only studied quality appraisals or fundus classification separately.Chea et al.[33] evaluated eye diseases based on deep learning technique (none transfer learning) for multiclass classification on a public dataset and obtained an average accuracy of 85.79% (on the ResNet-50 model),which outperforms other deep learning models (ResNet-101, ResNet-152, VGG-16, VGG-19)used with the manual filtering of different publishing datasets.This study obtained an average accuracy of 86.25% on the ResNet-50 model, which is higher than other models (VGG-19,VGG-16, MobileNet, MobileNet-V2, InceptionResNet-V2, Inception-v3, and DenseNet-201) on the SCH-BH dataset.ResNet-50 outperforms the others because it is an excellent model for image classification [34] and can learn and extract meaningful features of images.The feature extraction layers of ResNet-50 also perform well on fundus image classification.

    To assess the effectiveness of suitable-image assessment for fundus classification, the comparison performances between initial quality evaluation and native (without assessing quality)were examined.The results demonstrated that integrating quality evaluation in the fundus image classification system obtained better performance than without assessed quality for almost all CNN architectures used.DenseNet-201, without quality evaluation, achieved an overall accuracy of 78.68%, a sensitivity of 88.71% and a specificity of 95.92% for DR, a sensitivity of 32.38%and a specificity of 88.24% for GLC, the sensitivity of 34.78%, and specificity of 98.30% for AMD, and a sensitivity of 86.18% and a specificity of 81.48% for NR.For the same model,with quality assessment, the overall accuracy was 85.83%, 89.83% sensitivity and 96.69% specificity for DR, 36.46% sensitivity and 89.58% specificity for GLC, 72.73% sensitivity and 100%specificity for AMD, and 87.93% sensitivity and 89.52% specificity for NR (Tab.5).Therefore,quality assessment is possible by screening good images for the fundus diagnosis system.Poor quality/bad images affect the fundus diagnosis system, which can result in misdiagnosis.Thus,quality assessment contributes to improving the performance of multiclass fundus classification.

    DNNs were trained using large datasets to achieve better classification performance [35,36].Even with a limited number of fundus images, the classification was still highly accurate based on the training technique.Transfer learning technique transfers learned features trained on one task (thousands of general images) to classify the second task (fundus images).Training on the second task requires a significant training time with a small dataset.Vujosevic et al.[37]compared the classification performance of native and transfer learning and demonstrated that the transfer learning technique was better than native.The transfer learning technique contributes to an accurate evaluation.

    In conclusion, this study presents a multiclass classification of fundus images for detecting eye diseases (DR, GLC, and AMD) and healthy eyes (NR) based on a transfer learning approach.This study aims to propose an integrated quality evaluation with multiclass classification to eliminate bad quality images before performing fundus image classification, which can improve the classification performance.Integration quality evaluation in fundus classification can improve AUC performance from 0.91 to 0.99 on the DenseNet-201 model.

    Funding Statement:This work was supported by the National Research Foundation of Korea(NRF) grant funded by the Korea government (MSIT) (No.NRF-2021R1A2C1010362) and the Soonchunhyang University Research Fund.

    Conflicts of Interest:The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

    亚洲成人一二三区av| 国产精品一二三区在线看| 制服人妻中文乱码| 国产一区二区在线观看日韩| 日本-黄色视频高清免费观看| 伦精品一区二区三区| 91成人精品电影| 亚洲国产欧美日韩在线播放| 免费日韩欧美在线观看| 91在线精品国自产拍蜜月| 日韩av免费高清视频| 一区二区av电影网| 亚洲精品视频女| 国产成人精品在线电影| 免费人成在线观看视频色| 纯流量卡能插随身wifi吗| 热99国产精品久久久久久7| 久久久精品94久久精品| 男人操女人黄网站| 亚洲av二区三区四区| 夜夜爽夜夜爽视频| 少妇熟女欧美另类| 国产免费视频播放在线视频| 亚州av有码| 五月玫瑰六月丁香| 成人手机av| 欧美xxⅹ黑人| 亚洲精品美女久久av网站| 久久婷婷青草| 免费黄网站久久成人精品| 少妇熟女欧美另类| 成年人午夜在线观看视频| 久久韩国三级中文字幕| 久久午夜综合久久蜜桃| av免费在线看不卡| 3wmmmm亚洲av在线观看| 22中文网久久字幕| 全区人妻精品视频| 色94色欧美一区二区| 国产成人freesex在线| 国产又色又爽无遮挡免| 人妻 亚洲 视频| 免费少妇av软件| 考比视频在线观看| 国产精品人妻久久久久久| 国产av一区二区精品久久| 中文字幕久久专区| 国产不卡av网站在线观看| 中国美白少妇内射xxxbb| 两个人免费观看高清视频| 亚洲第一av免费看| 亚洲图色成人| 亚洲人成网站在线播| 欧美日韩亚洲高清精品| 久久久精品区二区三区| 最近2019中文字幕mv第一页| 我要看黄色一级片免费的| av国产久精品久网站免费入址| 久久97久久精品| 色视频在线一区二区三区| 久久影院123| 校园人妻丝袜中文字幕| 国产高清国产精品国产三级| 最近最新中文字幕免费大全7| 91久久精品国产一区二区三区| 男女国产视频网站| 一级a做视频免费观看| 黄色怎么调成土黄色| 亚洲欧美中文字幕日韩二区| 18禁动态无遮挡网站| 另类精品久久| 啦啦啦啦在线视频资源| 国产无遮挡羞羞视频在线观看| 狠狠精品人妻久久久久久综合| 另类精品久久| av黄色大香蕉| 91精品一卡2卡3卡4卡| 夫妻性生交免费视频一级片| 国产熟女午夜一区二区三区 | 一级毛片黄色毛片免费观看视频| 欧美精品一区二区大全| 欧美亚洲 丝袜 人妻 在线| 日日撸夜夜添| 91在线精品国自产拍蜜月| 久久久久国产精品人妻一区二区| 久久久精品区二区三区| 精品亚洲成国产av| 十八禁高潮呻吟视频| 亚洲国产毛片av蜜桃av| 日日摸夜夜添夜夜添av毛片| 91精品三级在线观看| 国产精品久久久久久久电影| 午夜日本视频在线| av福利片在线| 爱豆传媒免费全集在线观看| 国产亚洲午夜精品一区二区久久| 成人漫画全彩无遮挡| 国产免费视频播放在线视频| 欧美日韩av久久| 99热这里只有精品一区| av专区在线播放| 一个人看视频在线观看www免费| 免费高清在线观看日韩| 久久久亚洲精品成人影院| 久久久久精品久久久久真实原创| 国产欧美另类精品又又久久亚洲欧美| 亚洲精品国产色婷婷电影| 美女内射精品一级片tv| 国产精品蜜桃在线观看| 成人午夜精彩视频在线观看| 色网站视频免费| 欧美精品一区二区免费开放| 国产国拍精品亚洲av在线观看| 美女xxoo啪啪120秒动态图| 色婷婷久久久亚洲欧美| 久久久久久久大尺度免费视频| 日韩一区二区视频免费看| 99国产精品免费福利视频| 国产成人精品无人区| 日韩制服骚丝袜av| 久久久久人妻精品一区果冻| 99九九在线精品视频| 亚洲av综合色区一区| 亚洲国产毛片av蜜桃av| 色94色欧美一区二区| 国产黄色免费在线视频| 观看av在线不卡| 欧美另类一区| 春色校园在线视频观看| 国产精品蜜桃在线观看| 亚洲精品一二三| 欧美日韩一区二区视频在线观看视频在线| 成人手机av| 大码成人一级视频| 美女内射精品一级片tv| 国产精品久久久久久精品古装| 欧美激情国产日韩精品一区| 桃花免费在线播放| 人妻人人澡人人爽人人| 午夜精品国产一区二区电影| 91久久精品电影网| 亚洲国产欧美在线一区| 国产精品一区www在线观看| a级片在线免费高清观看视频| 亚洲天堂av无毛| 精品卡一卡二卡四卡免费| 久久久国产欧美日韩av| 大香蕉久久成人网| av不卡在线播放| av视频免费观看在线观看| 哪个播放器可以免费观看大片| 国产精品久久久久久av不卡| 精品一区二区三卡| 久久精品久久久久久噜噜老黄| 一本大道久久a久久精品| 欧美变态另类bdsm刘玥| 精品久久久久久久久亚洲| 国语对白做爰xxxⅹ性视频网站| 亚洲三级黄色毛片| 视频区图区小说| 日韩免费高清中文字幕av| av专区在线播放| 日本午夜av视频| 成人无遮挡网站| 一级毛片aaaaaa免费看小| 欧美亚洲 丝袜 人妻 在线| 国产成人a∨麻豆精品| 免费少妇av软件| 街头女战士在线观看网站| 国模一区二区三区四区视频| 国产淫语在线视频| 亚洲色图 男人天堂 中文字幕 | 亚洲综合精品二区| 亚洲国产精品999| 九九久久精品国产亚洲av麻豆| 久久久久精品性色| videosex国产| 国产精品一区二区在线不卡| av黄色大香蕉| 亚洲天堂av无毛| 99久久精品国产国产毛片| 国产精品不卡视频一区二区| 成人国产av品久久久| 美女主播在线视频| 十分钟在线观看高清视频www| 一区二区三区四区激情视频| 下体分泌物呈黄色| 日本欧美视频一区| 日韩熟女老妇一区二区性免费视频| 成人手机av| 中文字幕最新亚洲高清| 亚洲av免费高清在线观看| 亚洲av在线观看美女高潮| 欧美人与善性xxx| 男女边摸边吃奶| 欧美另类一区| 飞空精品影院首页| 22中文网久久字幕| 久久韩国三级中文字幕| 久久精品国产亚洲av香蕉五月 | 啦啦啦视频在线资源免费观看| 国产野战对白在线观看| 丁香欧美五月| 免费高清在线观看日韩| 久久午夜亚洲精品久久| 久久ye,这里只有精品| 欧美性长视频在线观看| 精品亚洲成国产av| 亚洲av电影在线进入| 亚洲一卡2卡3卡4卡5卡精品中文| 久久影院123| 欧美成人午夜精品| 亚洲av美国av| cao死你这个sao货| 午夜成年电影在线免费观看| 女性被躁到高潮视频| 午夜福利欧美成人| 香蕉丝袜av| 一本—道久久a久久精品蜜桃钙片| 天天躁日日躁夜夜躁夜夜| 国产亚洲精品第一综合不卡| 亚洲av第一区精品v没综合| 成人三级做爰电影| 欧美成人午夜精品| 欧美在线一区亚洲| 搡老熟女国产l中国老女人| 亚洲视频免费观看视频| 精品久久蜜臀av无| 婷婷成人精品国产| 国产成人影院久久av| av电影中文网址| 欧美黑人欧美精品刺激| 午夜久久久在线观看| 亚洲专区中文字幕在线| 成年人黄色毛片网站| 国产老妇伦熟女老妇高清| 国产午夜精品久久久久久| 国产男女内射视频| 国产精品久久久av美女十八| 国产麻豆69| 成人特级黄色片久久久久久久 | 少妇猛男粗大的猛烈进出视频| 日韩欧美一区视频在线观看| 精品久久久久久电影网| 丝袜在线中文字幕| 久久精品成人免费网站| 老熟妇乱子伦视频在线观看| 黄频高清免费视频| 91九色精品人成在线观看| 日本精品一区二区三区蜜桃| 久久ye,这里只有精品| tube8黄色片| 精品熟女少妇八av免费久了| 999久久久精品免费观看国产| 国产福利在线免费观看视频| 一区二区三区国产精品乱码| 久久久水蜜桃国产精品网| 亚洲avbb在线观看| 久久国产精品男人的天堂亚洲| bbb黄色大片| 黄网站色视频无遮挡免费观看| 天堂中文最新版在线下载| 一级毛片女人18水好多| 日本精品一区二区三区蜜桃| 最近最新中文字幕大全电影3 | 亚洲男人天堂网一区| 一级片'在线观看视频| 日韩欧美国产一区二区入口| 女警被强在线播放| 久久久久久免费高清国产稀缺| 久久99一区二区三区| 国产免费视频播放在线视频| 亚洲男人天堂网一区| 国产激情久久老熟女| 国产男靠女视频免费网站| 久久性视频一级片| 在线观看66精品国产| 精品免费久久久久久久清纯 | 亚洲伊人色综图| 大香蕉久久成人网| 亚洲欧美激情在线| 亚洲av欧美aⅴ国产| 一本久久精品| 俄罗斯特黄特色一大片| 国产成人免费无遮挡视频| 一区福利在线观看| 亚洲avbb在线观看| 精品视频人人做人人爽| 可以免费在线观看a视频的电影网站| 欧美 日韩 精品 国产| 久久久久国内视频| 九色亚洲精品在线播放| 亚洲九九香蕉| 丰满少妇做爰视频| 90打野战视频偷拍视频| 午夜福利视频精品| 性高湖久久久久久久久免费观看| 啦啦啦中文免费视频观看日本| 18禁国产床啪视频网站| 人妻久久中文字幕网| 国产野战对白在线观看| 无人区码免费观看不卡 | 18禁裸乳无遮挡动漫免费视频| 欧美日韩亚洲高清精品| 别揉我奶头~嗯~啊~动态视频| 少妇 在线观看| 免费女性裸体啪啪无遮挡网站| 丝袜美足系列| 一个人免费在线观看的高清视频| 日日摸夜夜添夜夜添小说| 欧美老熟妇乱子伦牲交| 成人影院久久| 宅男免费午夜| 久久午夜综合久久蜜桃| 亚洲第一青青草原| 最新美女视频免费是黄的| 日韩视频在线欧美| 国产亚洲欧美精品永久| 午夜福利欧美成人| 国产男女内射视频| 精品人妻在线不人妻| 亚洲综合色网址| 女人高潮潮喷娇喘18禁视频| 亚洲成a人片在线一区二区| 嫩草影视91久久| 后天国语完整版免费观看| 亚洲欧美激情在线| 国产精品偷伦视频观看了| 老司机在亚洲福利影院| 男女边摸边吃奶| 手机成人av网站| 狠狠狠狠99中文字幕| 无限看片的www在线观看| 大陆偷拍与自拍| 国产成人免费无遮挡视频| 国产一区二区三区综合在线观看| 久久热在线av| 丁香六月天网| 高清毛片免费观看视频网站 | 国产不卡一卡二| 高清毛片免费观看视频网站 | 99re6热这里在线精品视频| 69精品国产乱码久久久| 精品久久久精品久久久| 巨乳人妻的诱惑在线观看| 免费人妻精品一区二区三区视频| 淫妇啪啪啪对白视频| 中文字幕制服av| 在线看a的网站| 国产精品一区二区在线不卡| 久久久久久久大尺度免费视频| 18禁美女被吸乳视频| 亚洲 欧美一区二区三区| 精品少妇黑人巨大在线播放| 日日夜夜操网爽| 成年人免费黄色播放视频| 国产99久久九九免费精品| 一二三四社区在线视频社区8| 国产人伦9x9x在线观看| 亚洲av美国av| 日韩成人在线观看一区二区三区| 亚洲成国产人片在线观看| 亚洲国产成人一精品久久久| 午夜老司机福利片| av天堂在线播放| 操美女的视频在线观看| 波多野结衣av一区二区av| 18禁裸乳无遮挡动漫免费视频| 久久久国产欧美日韩av| 国产亚洲欧美精品永久| 久久毛片免费看一区二区三区| svipshipincom国产片| avwww免费| 国产成人欧美在线观看 | 99国产综合亚洲精品| 精品久久久精品久久久| 亚洲国产精品一区二区三区在线| 成年版毛片免费区| 亚洲人成电影观看| 欧美成人午夜精品| 国产黄频视频在线观看| 蜜桃国产av成人99| 无人区码免费观看不卡 | 色尼玛亚洲综合影院| 考比视频在线观看| 人人澡人人妻人| 十八禁网站网址无遮挡| 一区二区三区国产精品乱码| 欧美日韩亚洲综合一区二区三区_| 欧美乱码精品一区二区三区| 亚洲国产精品一区二区三区在线| 亚洲欧美精品综合一区二区三区| a级毛片在线看网站| 黑人巨大精品欧美一区二区mp4| 久久精品成人免费网站| 美女视频免费永久观看网站| 99香蕉大伊视频| 久久人人97超碰香蕉20202| 亚洲国产av影院在线观看| 老熟女久久久| 18禁裸乳无遮挡动漫免费视频| 国产日韩欧美在线精品| av一本久久久久| 欧美黄色片欧美黄色片| 国产1区2区3区精品| 国产一区二区三区在线臀色熟女 | 在线观看免费高清a一片| 一个人免费在线观看的高清视频| 国产在视频线精品| 男女免费视频国产| av免费在线观看网站| 嫁个100分男人电影在线观看| 欧美成人午夜精品| 亚洲人成电影免费在线| 精品亚洲成国产av| 国产成人啪精品午夜网站| 人妻 亚洲 视频| 正在播放国产对白刺激| 国产欧美亚洲国产| 国产在视频线精品| 久久免费观看电影| 国产成人欧美| 女人精品久久久久毛片| 久久亚洲精品不卡| 我的亚洲天堂| 女人久久www免费人成看片| 在线观看舔阴道视频| 少妇裸体淫交视频免费看高清 | 国产高清videossex| 亚洲成人国产一区在线观看| 免费一级毛片在线播放高清视频 | 亚洲欧美精品综合一区二区三区| 高清毛片免费观看视频网站 | 看免费av毛片| 国产一区二区三区在线臀色熟女 | 精品卡一卡二卡四卡免费| 国产av精品麻豆| 在线十欧美十亚洲十日本专区| 母亲3免费完整高清在线观看| 亚洲成人免费电影在线观看| 黄色片一级片一级黄色片| 丁香六月天网| 丰满人妻熟妇乱又伦精品不卡| 亚洲第一青青草原| 国产免费视频播放在线视频| 大陆偷拍与自拍| 国产精品国产高清国产av | 丰满人妻熟妇乱又伦精品不卡| a级片在线免费高清观看视频| 国精品久久久久久国模美| videosex国产| 丰满迷人的少妇在线观看| 另类亚洲欧美激情| 成人国产av品久久久| 国产色视频综合| 成人国产一区最新在线观看| 久久久久视频综合| 欧美老熟妇乱子伦牲交| 亚洲中文日韩欧美视频| 黄片播放在线免费| 美女午夜性视频免费| 久久ye,这里只有精品| 国产在线视频一区二区| 又黄又粗又硬又大视频| 欧美精品一区二区大全| 妹子高潮喷水视频| 亚洲免费av在线视频| 久久精品aⅴ一区二区三区四区| 天堂8中文在线网| 午夜福利免费观看在线| 国产精品熟女久久久久浪| 99riav亚洲国产免费| 欧美日韩黄片免| aaaaa片日本免费| 男女无遮挡免费网站观看| 欧美大码av| 性高湖久久久久久久久免费观看| 女性被躁到高潮视频| 国产黄频视频在线观看| 极品少妇高潮喷水抽搐| 国产激情久久老熟女| 99国产精品免费福利视频| 欧美日韩国产mv在线观看视频| av网站免费在线观看视频| 999精品在线视频| av片东京热男人的天堂| 少妇被粗大的猛进出69影院| 少妇猛男粗大的猛烈进出视频| 国产免费福利视频在线观看| 亚洲中文字幕日韩| 精品久久久久久电影网| 两个人免费观看高清视频| av国产精品久久久久影院| 少妇裸体淫交视频免费看高清 | 91老司机精品| 亚洲九九香蕉| 黄网站色视频无遮挡免费观看| 国产麻豆69| 亚洲精品久久成人aⅴ小说| 亚洲精品久久午夜乱码| 亚洲精品中文字幕一二三四区 | 女人精品久久久久毛片| 1024香蕉在线观看| 99九九在线精品视频| 中亚洲国语对白在线视频| 少妇粗大呻吟视频| 成人三级做爰电影| av欧美777| 国产片内射在线| 一本一本久久a久久精品综合妖精| √禁漫天堂资源中文www| 亚洲第一av免费看| 在线av久久热| 国产一区有黄有色的免费视频| 日韩大片免费观看网站| 免费少妇av软件| 久久久精品国产亚洲av高清涩受| 自线自在国产av| 久久久精品国产亚洲av高清涩受| 嫩草影视91久久| 日本vs欧美在线观看视频| 韩国精品一区二区三区| 无人区码免费观看不卡 | 日韩有码中文字幕| 蜜桃国产av成人99| 欧美人与性动交α欧美精品济南到| 少妇粗大呻吟视频| 99久久国产精品久久久| 人人妻,人人澡人人爽秒播| 日本撒尿小便嘘嘘汇集6| 日韩成人在线观看一区二区三区| 国产片内射在线| 国产男女超爽视频在线观看| 国产精品亚洲一级av第二区| 两人在一起打扑克的视频| 欧美激情 高清一区二区三区| 欧美日韩国产mv在线观看视频| 久久久久国内视频| 国产男靠女视频免费网站| 久久午夜综合久久蜜桃| videosex国产| 免费在线观看视频国产中文字幕亚洲| 欧美日韩亚洲国产一区二区在线观看 | 蜜桃国产av成人99| 国产极品粉嫩免费观看在线| av天堂在线播放| 成年女人毛片免费观看观看9 | 日本黄色视频三级网站网址 | 精品熟女少妇八av免费久了| 欧美日韩亚洲高清精品| 无人区码免费观看不卡 | 国产区一区二久久| 国产又爽黄色视频| 色尼玛亚洲综合影院| 午夜福利,免费看| 国产99久久九九免费精品| 91成年电影在线观看| 欧美人与性动交α欧美软件| 亚洲av欧美aⅴ国产| 亚洲五月色婷婷综合| 后天国语完整版免费观看| 久久人人97超碰香蕉20202| 首页视频小说图片口味搜索| 国产不卡一卡二| 中文字幕最新亚洲高清| 捣出白浆h1v1| 9191精品国产免费久久| 最近最新中文字幕大全免费视频| 久久国产精品影院| 久久精品亚洲av国产电影网| 999久久久国产精品视频| 女人精品久久久久毛片| 99精品欧美一区二区三区四区| 国产精品久久久久成人av| 两性夫妻黄色片| 国产精品香港三级国产av潘金莲| 国产成人精品久久二区二区免费| 亚洲av日韩在线播放| 女人被躁到高潮嗷嗷叫费观| 丰满人妻熟妇乱又伦精品不卡| 叶爱在线成人免费视频播放| 国产成人欧美| 丰满人妻熟妇乱又伦精品不卡| 俄罗斯特黄特色一大片| 黄色 视频免费看| 免费av中文字幕在线| 老司机影院毛片| 黑人巨大精品欧美一区二区mp4| 亚洲成国产人片在线观看| 国产欧美亚洲国产| 侵犯人妻中文字幕一二三四区| 女警被强在线播放| a级片在线免费高清观看视频| 国产精品熟女久久久久浪| 黄色片一级片一级黄色片| 亚洲第一青青草原| a级毛片黄视频| 日本vs欧美在线观看视频| 精品少妇黑人巨大在线播放| 国产精品一区二区精品视频观看| 国产区一区二久久| 国产午夜精品久久久久久| 下体分泌物呈黄色| 国产区一区二久久| 欧美激情极品国产一区二区三区| 在线观看免费视频日本深夜| 可以免费在线观看a视频的电影网站| 五月开心婷婷网| 女同久久另类99精品国产91| 日韩中文字幕视频在线看片| 啦啦啦 在线观看视频| 国产精品亚洲一级av第二区| 亚洲综合色网址| 啦啦啦 在线观看视频| 黄色视频在线播放观看不卡|