• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Hybrid Segmentation Scheme for Skin Features Extraction Using Dermoscopy Images

    2021-12-10 11:55:54JehyeokRewHyungjoonKimandEenjunHwang
    Computers Materials&Continua 2021年10期

    Jehyeok Rew,Hyungjoon Kim and Eenjun Hwang

    School of Electrical Engineering,Korea University,145,Anam-ro,Seongbuk-gu,Seoul,Korea

    Abstract:Objective and quantitative assessment of skin conditions is essential for cosmeceutical studies and research on skin aging and skin regeneration.Various handcraft-based image processing methods have been proposed to evaluate skin conditions objectively,but they have unavoidable disadvantages when used to analyze skin features accurately.This study proposes a hybrid segmentation scheme consisting of Deeplab v3+with an Inception-ResNet-v2 backbone,LightGBM,and morphological processing(MP)to overcome the shortcomings of handcraft-based approaches.First,we apply Deeplab v3+with an Inception-ResNet-v2 backbone for pixel segmentation of skin wrinkles and cells.Then,LightGBM and MP are used to enhance the pixel segmentation quality.Finally,we determine several skin features based on the results of wrinkle and cell segmentation.Our proposed segmentation scheme achieved a mean accuracy of 0.854,mean of intersection over union of 0.749,and mean boundary F1 score of 0.852,which achieved 1.1%,6.7%,and 14.8%improvement over the panoptic-based semantic segmentation method,respectively.

    Keywords:Image segmentation;skin texture;feature extraction;dermoscopy image

    1 Introduction

    Skin texture is the outermost indicator of the skin condition or the progression level of skin aging.Skin texture depends on various intrinsic and extrinsic factors that affect the health of the skin layers[1–5].For instance,prolonged exposure to sunlight and excessive smoking can adversely affect skin layers,as the skin loses moisture and becomes dry,which is then reflected in the skin texture.Therefore,tracking and observing changes in skin texture can help intuitively understand changes in skin health.

    However,objective assessment of skin condition has been regarded as challenging because dermatologists have traditionally diagnosed the skin condition by examining the skin texture with the naked eye,making the diagnostic results subjective.To diagnose the skin condition objectively,two-general approaches have been proposed:three-dimensional topography analysis(3DTA)and two-dimensional image analysis(2DIA).

    3DTA mainly analyzes depth-related skin features such as maximal,mean,minimum depth of roughness,and smoothness[6–9].This usually requires silicon replicas of the skin surface and optical equipment.Therefore,3DTA suffers from significant problems such as long processing time,complexity of analysis,and high cost.In contrast,most 2DIA methods focus on analyzing the structural shapes of skin texture using dermoscopy devices[10–17].The 2DIA methods aim to extract diverse representative visual features from dermoscopy images,such as irregular polygons and borderlines that are regarded as cells and wrinkles,respectively.Skin features include wrinkle width,wrinkle length,and the density of cells that can be used as essential criteria to evaluate skin conditions or estimate the degree of skin aging[12–14].For decades,2DIA methods have generally been used for skin condition diagnosis in terms of time and cost efficiency,but 2DIA methods,too,suffer from various limitations:

    ?The image preprocessing steps of 2DIA methods,such as contrast enhancement,noise removal,and color histogram equalization,have been designed differently for different datasets.

    ?Most 2DIA methods have opted for a handcrafted feature-based method,but handcrafted features are generally not robust and are computationally intensive due to their high dimensionality.

    ?Traditional 2DIA methods must choose which features are important in a given dataset.Therefore,substantial engineering experience and dermatology knowledge are required to obtain successful results.

    Deep learning is one alternative technique to handcrafted feature-based approaches because it can learn visual features automatically to solve a specific task,such as object classification and target segmentation.In many research fields,a convolutional neural network(CNN)-based model,which is a deep learning technique,has been reported to be extremely effective compared with handcraft-based approaches.We believe that applying a deep learning model overcomes the disadvantage of the 2DIA methods in accurate skin feature extraction.In this study,we address three major issues,with the following main contributions:

    ?We performed wrinkle and cell segmentation using Deeplab v3+[18]with an Inception-ResNet-v2 backbone[19],which is a popular CNN model.By using this approach,we overcome the serious limitations of traditional handcrafted feature-based methods.

    ?We improved the quality of wrinkle and cell segmentation by using LightGBM[20]and morphological processing(MP),which can accurately extract skin features.

    ?We demonstrated the effectiveness of our proposed scheme by extracting various skin features from the segmentation results and comparing them with other existing methods in terms of accuracy.

    The remainder of this paper is organized as follows.Related works of skin image analysis are introduced in Section 2.The major differences between previous works and our proposed scheme are described in Section 3.Finally,the experimental results and conclusions are presented in Sections 4 and 5,respectively.

    2 Related Works

    This section presents several studies that have been conducted to assess skin conditions.Since the 2000s,most skin image analyses have been performed through molding replica analysis based on the 3DTA method[21–23].Masuda et al.[21]proposed a 3D morphological characterization method for molding replica analysis.Using 3D measurement equipment and a surface analysis method,they extracted and analyzed several skin texture features such as depth,width,and length of skin furrows.Friedman et al.[22]conducted a 3D skin topology analysis,a non-invasive approach for identifying facial rhytides and acne scars.They utilized a micro-topology imaging system and a charge-coupled device(CCD)camera to record the skin surface topology.They concluded that 3D skin topology analysis could allow fast and quantitative assessment of skin conditions.Molding replica and 3D topology analysis methods have been regarded as useful approaches,but their requirements for expensive equipment and substantial analytic experience make their widespread application difficult.

    Recently,2DIA-based approaches have attracted much attention[15–17,24–28].Zou et al.[17]suggested an objective modeling approach for skin surface characterization.They developed new measurement parameters,including polygons on the skin texture and the average area of the polygons detected.They argued that skin surface analysis using image processing techniques is useful for quantitatively expressing the skin condition.Cula et al.[24]proposed an automatic wrinkle detection algorithm to extract the orientation of polygon borders and bidirectional histogram features of illumination.Facial wrinkle features were extracted using contrast stretching and Gabor filtering,and the wrinkle types were classified to identify the subjects.Tanaka et al.[25]applied the cross-binarization method to detect skin wrinkles.Image binarization was first conducted for accurate wrinkle detection,and subsequently,the straight-line matching algorithm was applied to measure the length of skin wrinkles.Razalli et al.[26]proposed a wrinkle detection method to determine the relationship between aging and changes in wrinkle shape.The Hessian-based filter(HBF)was applied to extract the facial features,and then,the method was evaluated using the FG-NET database.Choi et al.[27]proposed a framework for skin texture extraction and skin aging trend estimation.The depth of wrinkles,wrinkle length and width,and the number of closed polygons were calculated as a feature set.Then,a support vector machine model was utilized to estimate the trend of skin aging.A skin aging estimation accuracy of more than 90%was achieved in a dermatologist’s blind test.

    More recently,using the versatility of machine learning techniques,breakthrough attempts have been made to detect or diagnose critical diseases such as melanoma,lesions,and COVID-19[29–33].Afza et al.[29]proposed an optimal color feature selection scheme for skin lesion classification.They used contrast stretching,color feature extraction,and entropy-controlled feature optimization.To improve the performance of skin lesion classification,they proposed a hierarchical 3-step super-pixels with deep learning–based framework that includes a pre-trained deep learning model(ResNet-50)and a new optimization technique called Grasshopper[30].Khan et al.[31]proposed an intelligent framework that includes Mask R-CNN,ResNet with feature pyramid network(FPN),and 24-layered CNN for localizing and classifying multi-class skin lesions.They used pre-trained DenseNet,entropy-controlled least square support vector machine and extreme learning machine(ELM)techniques to improve the accuracy of skin lesion classification[32].

    To summarize,3DTA,2DIA,and machine learning methods have been used to objectively assess skin conditions and diagnose critical diseases such as melanoma and skin lesions.3DTA is an excellent approach to accurately analyze skin conditions,but it requires expensive equipment and dermatological knowledge and experience.Although 2DIA is a cost-effective alternative to 3DTA,it is difficult to guarantee its applicability as most 2DIA methods follow a handcrafted feature-based method that makes it difficult to obtain robust features and rely heavily on an engineer’s experience and knowledge[34].To overcome the shortcomings of the 3DTA and 2DIA methods,we propose a scheme that performs image segmentation based on a deep learning model and extracts accurate skin features from the segmentation result.

    3 Methodology

    3.1 Comparison with Previous 2DIA Methods

    In this section,we show the difference between the traditional 2DIA approaches and the proposed scheme.Traditional 2DIA methods for skin feature extraction first perform some preprocessing steps to eliminate light interference,noise,and distortion to obtain skin features of reasonable quality[10–17].

    The skin image is then converted into a binary image,which is regarded as the groundwork for identifying wrinkle and cell areas.Fig.1 presents the typical structure of traditional 2DIA methods with the preprocessing steps in the yellow box.Image cropping eliminates the harmful effects of vignetting because a luminous source of the skin image from a dermoscopy device is typically concentrated around its center[10–13].Contrast stretching and histogram equalization can minimize light interference.Generally,noise filtering and morphological transformation improve the effectiveness of preprocessing.The watershed transform method[35]performs wrinkle and cell segmentation by calculating the intensity difference in a grayscale image and draws borderlines based on the intensity difference.Finally,wrinkles and cells are obtained using the watershed transform.

    Figure 1:Typical structure of traditional 2DIA methods for skin feature extraction

    Fig.2 presents an illustration of our scheme.Unlike traditional approaches,our scheme adopts a CNN,LightGBM,and MP as preprocessing units.The CNN is organized by multiple convolutional layers that perform image classification and segmentation tasks.Convolutional layers reduce the images into a form that is easier to process without losing features.This is critical for achieving good predictions.Each convolutional layer extracts valuable visual features,considering image distortion,noise,and brightness.Therefore,well-designed convolutional layers can extract reasonable visual features that represent the semantic information of the image.For this,the Deeplab v3+ with Inception-ResNet-V2 backbone model,a popular CNN model,was used to replace several preprocessing steps in the traditional approach.In addition,to enhance the quality of binarization,LightGBM and MP were used together.Manually,the ground truth was made based on the annotated wrinkle and cell class for every pixel in the dermoscopy image,and these features were utilized to train the Inception-ResNet-V2 model[19].Subsequently,we applied watershed transformation and performed wrinkle and cell feature extraction.

    Figure 2:Overall structure of the proposed scheme for skin feature extraction

    Figure 3:Segmentation examples(a)Input images(b)Ground truth

    3.2 Wrinkle and Cell Segmentation

    In this section,we present details of the wrinkle and cell segmentation process with the organized CNN architecture and LightGBM.Figs.3a and 3b present the input skin images captured by a commercial dermoscopy device and their ground truth,respectively.We defined that the black pixels represent the wrinkle class and the white pixels indicate the cell class.

    Fig.4 presents the overall flow of our proposed scheme.In this study,a pre-trained Inception-ResNet-v2 model obtained from the MathWorks repository was used.Inception-ResNet-v2 showed the best performance in the ImageNet Large Scale Visual Recognition Challenge(ILSVRC)2015,and its performance in object recognition and segmentation is better than other inception models,such as Inception-ResNet-v1,Inception-v3,and Inception-v4.

    Figure 4:Overall flow of our proposed scheme

    Inception-ResNet-v2 is a combination of residual connections and an inception module.The front module of Inception-ResNet-v2 is a stem layer,which generates a 35×35×256 feature map using five convolutions of 3×3,single max-pooling of 3×3,and one convolution of 1×1 from the 299×299×3 input image.In general,the stem layer is used to maximize the performance of inception neural networks.In the figure,the pink boxes indicate the various types of inception modules.Each inception module acts as multiple convolution filters,such as 1×1 and 3×3,which are designed for efficient visual feature extraction.Fig.5 compares the various schemas of the Inception-ResNet-A,Inception-ResNet-B,and Inception-ResNet-C modules.Each module extracts general and local visual features simultaneously and reduces the computation resources.This is because the inception modules replace anyn×nconvolution by a 1×nconvolution followed by ann×1 convolution.This form can significantly reduce computational cost asnincreases.The green box represents the reduction modules of Inception-ResNet-v2.Fig.6 shows detailed schemas of the reduction-A and reduction-B modules.The reduction-A module reduces the 35×35×256 feature map to a 17×17×896 feature map,and reduction-B reduces the 17×17×896 feature map to an 8×8×1792 feature map.These two modules reduce the size of the features from the previous layer and pass them to the next layer.Atrous spatial pyramid pooling(ASPP)extracts the high-level features from the results of the Inception-ResNet-C module and fed them to the decoder module.The decoder infers the semantic information of each pixel by using the extracted representative features.The DeepLab v3+ decoder module was used to infer the wrinkle and cell area of the skin texture image.The DeepLab v3+ decoder performs up-sampling twice,which restores the dimensions to the input image size.During up-sampling,the encoder module first concatenates the corresponding low-level features from the stem layer.Then,the encoder module up-samples by a factor of four to infer the pixel segmentation.The last layer is the soft-max layer,which outputs a vector that represents the probability distribution of the potential outcomes.This layer takes the output of multiple dimensional vectors and rescales them in the range from 0 to 1,with a sum of one.Through the arguments of the maxima(argmax)function,the segmentation results for all pixels can be obtained.If the argmax function calculates the probability as zero,it is predicted as a cell class.In contrast,if the argmax function calculates the probability as one,it is predicted as a wrinkle class.

    Figure 5:Detail schema of Inception-ResNet-v2 network(a)Inception-RestNet-A module(b)Inception-ResNet-B module(c)Inception-ResNet-C module

    Figure 6:Detailed schema of Inception-ResNet-v2 network(a)Reduction-A module(b)Reduction-B module

    To verify the segmentation performance of Inception-ResNet-v2,the soft-max activation layer was visualized,as presented in Fig.7c.The white pixels belong to the cell class and black pixels belong to the wrinkle class.In the figure,the red boxes indicate false segmentation candidates.

    Figure 7:Example of soft-max activation and prediction result(a)Input image(b)Ground truth(c)soft-max activation(d)Prediction result from the Deeplab v3+ with Inception-ResNet-v2 backbone

    To alleviate the false segmentation problem,a tree-based ensemble framework,namely,Light-GBM,was used.LightGBM has obvious advantages,such as high training speed,low memory usage,and high prediction accuracy,when processing large-scale data.In this case,LightGBM was used to improve the segmentation quality by considering the output probabilities of the soft-max layer as input.For this,we sequentially calculated the probabilities of the soft-max layer extracted from the training images.The overall probability values were then accumulated and reshaped to a flattened form.If one image has 299(width)× 299(height)× 2(class probabilities)soft-max probabilities,the probabilities were reshaped to 89,401(flatten)× 2(class probabilities).Using class labels of the ground truth,we trained all reshaped-probabilities as the input of LightGBM.The segmentation improvement is shown in the experimental section.

    3.3 Morphological Processing

    Although CNN and LightGBM show good performance in wrinkle and cell segmentation,their results are not always satisfactory.For instance,the segmentation includes areas wherein it is unclear whether they contain wrinkles or cells,which degrades the reliability of the segmentation results[36–39].Another problem is false segmentation pixels,which appear as noise.Most skin-related 2DIA studies attempted to eliminate noise using median and linear filters,such as Gaussian and Wiener[40].These filtering methods often lead to structural changes.To avoid this,an MP method was used.Morphological processing has the advantage of removing a single-pixel line,dot,or tiny pixel structures while maintaining the structural shape.MP performs shrinking and growing processes.The shrinking process rounds large structures and the growing process removes small structures.To achieve this,the dilation and erosion processes were used together.For instance,for the segmentation result illustrated in Fig.8a obtained by means of CNN and LightGBM methods,Fig.8b shows the result of MP,in which false segmented areas were removed or merged.

    Figure 8:Examples of result(a)Original segmentation(b)Improvement using MP(c)Ground truth

    3.4 Skin Texture Feature Extraction

    After segmentation,various skin features are extracted from the segmentation results.In this study,four skin features that are essential for evaluating skin condition and aging were considered.They are the length of wrinkle lines,wrinkle width,number of cells,and area of detected cells.The length of wrinkle lines is easily identified using the skeleton pixels classified as wrinkle class.To calculate the wrinkle width,we used Algorithm 1.The number of cells and their area were calculated using the polygon mesh detection algorithm(PMDA),which is presented in Algorithm 2[12].Fig.9 depicts the steps for skin texture feature extraction.

    Algorithm 1:W Input:Segment Output:Wrinkl Wrinkl rinkle feature extraction ed wrinkle SW e width Ww,e length Wl Binary wrinkle skeleton S ←watershed_transform(SW);Wrinkle contour WC ←canny_edge_detection(SW);Angle A ←PCA_angle(S);for each point p in S and each angle a in A do Measurement line ML ←find_orthognal(a,p)[Start point s,end point e]←find_start_end_points(ML,WC);localWrinkleWidth += euclidean_distance(s,e);end for Wl=cnt(S);Ww= localWrinkleWidth/cnt(ML);return Ww,Wl

    Algorithm 2:Polyg Input:Segmented on mesh detection wrinkle SW Output:Number of cell Cn,Average area of detected cell Cavg_area Binary wrinkle skeleton S ←watershed_transform(SW);

    for each poi do recursive localCell nt p in S ly_seek_4_neighbors();Area++;if p belongs to image border then localCellArea = 0;end if Cn++;totalCellArea += localCellArea;while not reaching the borderlines end for Cavg_area= totalCellArea/Cn;return Cn,Cavg_area

    Figure 9:Steps for skin texture feature extraction(a)Input image(b)Segmentation result(c)Wrinkle-related feature extraction(d)Cell-related feature extraction using PMDA

    4 Results and Discussion

    4.1 Dataset and Model Configuration

    To construct a skin image dataset,50× magnified facial images from 365 healthy subjects were collected.All measurements were conducted in a room at a temperature of 23 ± 3°C and humidity of 50 ± 10%.Fig.10 shows the sample images of our dataset.In the experiment,none of the subjects applied any makeup on the face.

    To train the Inception-ResNet-v2 model,the resolution of the input images was first set to 299×299.Then,data augmentation was conducted using random vertical flip,random horizon flip,10% zooms,and 10% shear changes.The stochastic gradient descent with a momentum(SGDM),learning rate of 1×10?3,and L2 regularization were then used to configure the training strategy.For verification,80% of augmented facial images were used as training and validation sets,and the remaining 20% as the test set.To train LightGBM,the following conditions were set:1×10?3learning rate,5,000 iterations with early stopping,10 max depth branching,1,024 max leaves,and the gradient-boosted decision trees(GBDT)method.10-fold cross-validation was conducted to assess the LightGBM performance.The ratio of the training set,validation set,and test set is equal to that of the Inception-ResNet-v2.

    Figure 10:Sample images from the dataset

    4.2 Performance Evaluation of Segmentation

    In the evaluation of segmentation performance,three metrics were used:mean accuracy(MA),mean of intersection over union(MIOU),and mean of boundary F1 score(MBF)[41].MA,which can be defined by Eq.(1),indicates the percentage of correctly identified pixels as wrinkle class or cell class.For each class,accuracy is the ratio of precisely matched pixels to the total number of pixels in that class,according to the ground truth.In Eq.(2),TP,FN,andFPrepresent the number of true positives,the number of false negatives,and the number of false positives,respectively.In addition,iis the total number of images.

    MIOU is another popular metric for measuring the performance of image segmentation and can be expressed as Eq.(2).IOU indicates the overlap percentage between the prediction pixels and the ground truth pixels.

    The last metric is the MBF score between the segmentation result and the ground truth.MBF,which is usually used to evaluate the contour matching,indicates how close the boundary of the segmentation matches the boundary of the ground truth.Eq.(3)shows how to calculate the MBF.In Eq.(3),PcandRcindicate the precision and recall of classc,respectively,andBcpsandBcgtindicate the boundary binary map of the predicted segmentation and the ground truth in classc,respectively.In Eq.(3),“[[]]” is the Iverson bracket notation,where[[z]]= 1 ifz= true and 0 otherwise.In addition,d()represents the Euclidean distance measured in pixels.In Eq.(4),BFis defined as the harmonic mean of the precision and recall values with a distance error tolerance.Finally,we obtained the MBF by averaging the per-image BF scores.

    In the experiment,we considered diverse segmentation methods including Choi’s method,U-Net,SegNet,Deeplab v3+ with ResNet backbone families,and Panoptic-Deeplab with ResNet-101 backbone for comparison.Tab.1 compares their wrinkle and cell class segmentation performances.As is clear from the table,the higher the score,the better the segmentation result.Our proposed scheme showed the best segmentation performance compared with other methods.It means that our proposed scheme significantly improved the segmentation quality in terms of finding the borderlines and regions of wrinkle and cell.Choi’s method presented the worst performance because it is basically a handcraft-based method.In contrast,the Deeplab v3+ models with ResNet backbone families showed better performance than U-Net and SegNet.Panoptic-Deeplab scored higher than Deeplab v3+ models in MA evaluation,but it showed relatively lower accuracy in wrinkle and cell boundary matching than did Deeplab v3+ models.Fig.11 presents the example of segmentation results.The first and second columns show the input skin texture image and the ground truth of the segmentation task,respectively.From the third to the last column,the segmentation results of different CNN models are presented.When the boundaries of the wrinkles and cells are difficult to distinguish in the image,SegNet and two Deeplab v3+models showed relatively low segmentation performance compared with the proposed model.

    Table 1:Comparisons of wrinkle and cell segmentation results

    Figure 11:Segmentation results of SegNet[43],Deeplab v3+(Resnet-18 backbone),Deeplab v3+(Resnet-50 backbone)[18],and the proposed scheme

    4.3 Performance Evaluation for Skin Texture Feature Extraction

    In this experiment,we evaluated the accuracy of skin feature extraction using four popular metrics.First,the accuracy of wrinkle line extraction is defined using Eq.(5).In this equation,DWPandWSEGrepresent the detected wrinkle pixels and ground truth pixels of the wrinkle,respectively.To calculate the accuracy,the equation first counts the correctly overlapped pixels between theDWPandWSEG,and then,it divides them by the total number of detected wrinkle pixels.

    To evaluate the accuracy of wrinkle width extraction,the mean absolute percentage error(MAPE)was used,which can be calculated using Eq.(6).In this equation,MLindicates the number of measurement lines,andAWWiandEWWiindicate the actual length and estimated length of theith measurement line on the wrinkle width,respectively.

    Eqs.(7)and(8)show how to count valid cells and calculate their accuracy.To count valid cells,the weighted distance was calculated between the detected cellsDCand ground truth cellsGC.Then,the IOU was calculated between theDCandGC.If the IOU between theith matchedDCandGCis greater than 0.6,it is counted as a valid cell.

    Finally,the MAPE of the valid cell area was measured using Eq.(9).Here,VCis the number of valid cells,andAGCiandAVCiare the areas of theith ground truth cell and valid cell,respectively.

    Tab.2 shows the performance comparison of skin feature extraction.The same skin feature extraction methods were used for each segmentation model.CNN-based models such as SegNet,U-Net,Deeplab v3+,and Panoptic-Deeplab showed better performance than did Choi’s model,which is based on the handcrafted features.The proposed scheme showed the best performance in extracting skin features,and U-Net presented the worst performance.Despite the fact that Choi’s model used a handcrafted method,it performed better than the U-Net model because it was mainly designed to extract wrinkles and cells from dermoscopy images.

    Table 2:Performance comparisons of skin texture feature extraction

    5 Conclusions

    In this study,a hybrid segmentation scheme for skin feature extraction using Deeplab v3+with an Inception-ResNet-v2 backbone,LightGBM,and MP was proposed.As traditional 2DIA approaches for skin analysis use handcraft-based methods,it is difficult to obtain satisfactory results using them.To alleviate this problem,based on deep neural networks,which can perform efficient feature extraction from deep and complex networks,the hybrid approach was used to improve segmentation quality and skin feature extraction accuracy.To validate the effectiveness of the proposed scheme,extensive comparisons with other popular models were performed using diverse evaluation metrics.From the experimental results,the proposed scheme was confirmed to outperform the handcraft-based method and other popular CNN models under every evaluation metric.It is our belief that the proposed scheme can be used in diverse skin-related applications,such as skin damage estimation,skin condition assessment,and skin aging estimation.

    Funding Statement:This work was supported by the National Research Foundation of Korea(NRF)grant funded by the Korea government(MSIT)(No.2020R1F1A1074885)and was supported by the Brain Korea 21 Project in 2021(No.4199990114242).

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    成人国产一区最新在线观看 | 久久国产亚洲av麻豆专区| 国语对白做爰xxxⅹ性视频网站| 热99国产精品久久久久久7| 999精品在线视频| 免费看不卡的av| 国产精品av久久久久免费| 亚洲天堂av无毛| 久久久久精品国产欧美久久久 | 一级黄色大片毛片| 在线观看免费视频网站a站| 国产91精品成人一区二区三区 | 色94色欧美一区二区| 国产一区二区三区av在线| 男女国产视频网站| 亚洲国产最新在线播放| 亚洲图色成人| 大香蕉久久网| 青草久久国产| 高潮久久久久久久久久久不卡| 一二三四在线观看免费中文在| 国产成人系列免费观看| 亚洲av日韩在线播放| 亚洲av男天堂| 2021少妇久久久久久久久久久| 99久久人妻综合| 日韩,欧美,国产一区二区三区| 丝袜人妻中文字幕| 国产亚洲午夜精品一区二区久久| 午夜激情av网站| 日本欧美视频一区| 美女大奶头黄色视频| 啦啦啦在线观看免费高清www| 丰满少妇做爰视频| 精品人妻一区二区三区麻豆| 免费观看人在逋| 免费不卡黄色视频| 又黄又粗又硬又大视频| 蜜桃国产av成人99| av有码第一页| 国产成人91sexporn| 国产成人免费无遮挡视频| 国产精品久久久久久精品古装| 欧美老熟妇乱子伦牲交| 成年人黄色毛片网站| 欧美中文综合在线视频| 国产1区2区3区精品| 欧美人与性动交α欧美软件| 免费在线观看影片大全网站 | 1024视频免费在线观看| 中文精品一卡2卡3卡4更新| 欧美少妇被猛烈插入视频| 亚洲国产欧美网| 国产在线一区二区三区精| 在线观看免费视频网站a站| 国产一卡二卡三卡精品| 亚洲国产精品一区二区三区在线| 在线观看免费日韩欧美大片| 99热网站在线观看| 国产激情久久老熟女| 亚洲 欧美一区二区三区| 99国产精品一区二区三区| 免费看av在线观看网站| 99久久人妻综合| 在线天堂中文资源库| 在线看a的网站| 日本av免费视频播放| 在线观看免费午夜福利视频| 欧美成狂野欧美在线观看| 精品熟女少妇八av免费久了| 亚洲欧美成人综合另类久久久| 免费不卡黄色视频| 搡老乐熟女国产| 国精品久久久久久国模美| 99久久99久久久精品蜜桃| 99久久综合免费| 久久国产精品影院| 亚洲国产成人一精品久久久| 亚洲国产毛片av蜜桃av| h视频一区二区三区| 国产精品久久久人人做人人爽| 深夜精品福利| www.999成人在线观看| 夫妻性生交免费视频一级片| 欧美在线一区亚洲| 免费看十八禁软件| 只有这里有精品99| 精品熟女少妇八av免费久了| 一级毛片 在线播放| 国产一区亚洲一区在线观看| 啦啦啦 在线观看视频| 黄色视频不卡| 日韩av在线免费看完整版不卡| 99热国产这里只有精品6| 国产精品免费大片| 成人影院久久| 亚洲中文日韩欧美视频| 热re99久久国产66热| 欧美成人午夜精品| 久久鲁丝午夜福利片| av福利片在线| 一本久久精品| 日韩欧美一区视频在线观看| 免费高清在线观看视频在线观看| 久久99热这里只频精品6学生| 晚上一个人看的免费电影| 久久久国产精品麻豆| 亚洲伊人久久精品综合| 国产高清视频在线播放一区 | 五月开心婷婷网| 欧美精品一区二区免费开放| 美国免费a级毛片| 日本欧美视频一区| 两性夫妻黄色片| 脱女人内裤的视频| 亚洲av在线观看美女高潮| 叶爱在线成人免费视频播放| 欧美精品av麻豆av| 午夜精品国产一区二区电影| 国产有黄有色有爽视频| 国产高清videossex| 国产精品久久久久久精品古装| 大香蕉久久网| 婷婷丁香在线五月| 亚洲国产欧美日韩在线播放| 久久中文字幕一级| 久久国产精品人妻蜜桃| 人人妻,人人澡人人爽秒播 | 一本久久精品| 人人妻人人爽人人添夜夜欢视频| 操美女的视频在线观看| 亚洲欧美中文字幕日韩二区| www.av在线官网国产| 少妇人妻久久综合中文| 97精品久久久久久久久久精品| 久久亚洲国产成人精品v| 欧美精品一区二区免费开放| 亚洲精品国产av成人精品| 美女午夜性视频免费| 最近最新中文字幕大全免费视频 | 啦啦啦 在线观看视频| 99久久精品国产亚洲精品| 精品卡一卡二卡四卡免费| 亚洲国产毛片av蜜桃av| 国产一区二区三区av在线| 在线观看免费高清a一片| 一个人免费看片子| 国产色视频综合| 少妇猛男粗大的猛烈进出视频| 97在线人人人人妻| 伊人亚洲综合成人网| 欧美少妇被猛烈插入视频| 黄色 视频免费看| 美女高潮到喷水免费观看| 国产一区二区在线观看av| 亚洲欧美清纯卡通| 欧美黄色片欧美黄色片| 亚洲国产av新网站| 老司机在亚洲福利影院| 啦啦啦 在线观看视频| 别揉我奶头~嗯~啊~动态视频 | 亚洲少妇的诱惑av| 色精品久久人妻99蜜桃| 天天躁日日躁夜夜躁夜夜| 男人添女人高潮全过程视频| 亚洲av成人精品一二三区| 国产片特级美女逼逼视频| 18禁国产床啪视频网站| 桃花免费在线播放| 国产无遮挡羞羞视频在线观看| 视频区欧美日本亚洲| 大码成人一级视频| 成年av动漫网址| 狠狠婷婷综合久久久久久88av| av片东京热男人的天堂| 99热全是精品| 免费观看av网站的网址| 99国产精品一区二区蜜桃av | 久久精品久久精品一区二区三区| 一二三四在线观看免费中文在| 男人舔女人的私密视频| 国产成人影院久久av| 国产高清videossex| 人人澡人人妻人| 黄色视频在线播放观看不卡| 男女高潮啪啪啪动态图| 亚洲国产av影院在线观看| 大片免费播放器 马上看| 欧美+亚洲+日韩+国产| www.av在线官网国产| 欧美日韩一级在线毛片| 91老司机精品| 热99国产精品久久久久久7| 亚洲少妇的诱惑av| 国产高清不卡午夜福利| 后天国语完整版免费观看| 一本—道久久a久久精品蜜桃钙片| 又粗又硬又长又爽又黄的视频| 黄色毛片三级朝国网站| 黄色视频在线播放观看不卡| 国产精品麻豆人妻色哟哟久久| 中文字幕人妻丝袜一区二区| 男的添女的下面高潮视频| 最黄视频免费看| 久久影院123| a 毛片基地| 999久久久国产精品视频| 在线观看人妻少妇| 啦啦啦 在线观看视频| 亚洲av欧美aⅴ国产| 成年女人毛片免费观看观看9 | 久久精品国产亚洲av涩爱| 天天操日日干夜夜撸| 日韩伦理黄色片| 在线观看国产h片| 精品国产国语对白av| 亚洲av综合色区一区| av欧美777| 性高湖久久久久久久久免费观看| 久久人妻福利社区极品人妻图片 | 纯流量卡能插随身wifi吗| 性色av乱码一区二区三区2| 亚洲欧洲精品一区二区精品久久久| 久久久久久久国产电影| 欧美精品一区二区大全| 啦啦啦视频在线资源免费观看| 国产亚洲精品第一综合不卡| 亚洲美女黄色视频免费看| 99九九在线精品视频| 久久亚洲精品不卡| 桃花免费在线播放| 国产精品一二三区在线看| 丝袜美腿诱惑在线| 我的亚洲天堂| 青春草视频在线免费观看| 欧美精品人与动牲交sv欧美| 男女高潮啪啪啪动态图| 日韩 亚洲 欧美在线| 男人添女人高潮全过程视频| 婷婷色麻豆天堂久久| 欧美少妇被猛烈插入视频| 亚洲欧美精品综合一区二区三区| 亚洲视频免费观看视频| 亚洲熟女精品中文字幕| 男男h啪啪无遮挡| 亚洲国产成人一精品久久久| 国产av精品麻豆| 日韩一卡2卡3卡4卡2021年| 国产成人免费无遮挡视频| 妹子高潮喷水视频| 九草在线视频观看| 日本wwww免费看| 啦啦啦视频在线资源免费观看| 久久久久久人人人人人| 夫妻性生交免费视频一级片| 一本—道久久a久久精品蜜桃钙片| 精品久久久精品久久久| av欧美777| 伊人亚洲综合成人网| 日本猛色少妇xxxxx猛交久久| 中文乱码字字幕精品一区二区三区| 少妇精品久久久久久久| 亚洲黑人精品在线| 久久午夜综合久久蜜桃| 18禁观看日本| 香蕉丝袜av| 十分钟在线观看高清视频www| 精品国产国语对白av| 涩涩av久久男人的天堂| 亚洲国产欧美一区二区综合| 欧美久久黑人一区二区| 国产爽快片一区二区三区| 日韩一本色道免费dvd| 国产免费视频播放在线视频| 亚洲视频免费观看视频| 亚洲欧洲日产国产| 搡老岳熟女国产| 日韩伦理黄色片| 国产日韩欧美视频二区| 日韩一本色道免费dvd| 婷婷色综合www| 亚洲国产精品999| 日韩中文字幕视频在线看片| 亚洲精品一区蜜桃| 欧美人与性动交α欧美精品济南到| 国产精品一区二区精品视频观看| 日本a在线网址| 国精品久久久久久国模美| 国产又色又爽无遮挡免| 777米奇影视久久| 999久久久国产精品视频| 十八禁人妻一区二区| 国产精品熟女久久久久浪| 日韩熟女老妇一区二区性免费视频| 亚洲av电影在线观看一区二区三区| 无遮挡黄片免费观看| 在线观看免费午夜福利视频| 妹子高潮喷水视频| 91国产中文字幕| 亚洲久久久国产精品| 高清av免费在线| 黑人巨大精品欧美一区二区蜜桃| 国产一区二区三区av在线| 亚洲欧美精品综合一区二区三区| 人成视频在线观看免费观看| 亚洲精品日本国产第一区| 女人久久www免费人成看片| 久久国产精品人妻蜜桃| 亚洲精品一二三| 国产男女内射视频| 久久精品熟女亚洲av麻豆精品| 日本vs欧美在线观看视频| 妹子高潮喷水视频| 亚洲美女黄色视频免费看| 97精品久久久久久久久久精品| 成人亚洲精品一区在线观看| 丁香六月欧美| 国产三级黄色录像| av国产精品久久久久影院| 1024香蕉在线观看| 在线观看国产h片| 国产主播在线观看一区二区 | 国产午夜精品一二区理论片| 欧美少妇被猛烈插入视频| 欧美在线黄色| 亚洲av男天堂| 在线观看人妻少妇| 十分钟在线观看高清视频www| 好男人电影高清在线观看| 香蕉国产在线看| 九色亚洲精品在线播放| 国产福利在线免费观看视频| 国产老妇伦熟女老妇高清| 久久国产精品大桥未久av| xxxhd国产人妻xxx| 99re6热这里在线精品视频| 欧美少妇被猛烈插入视频| 精品一区二区三卡| 熟女少妇亚洲综合色aaa.| 精品人妻一区二区三区麻豆| 七月丁香在线播放| 无遮挡黄片免费观看| 少妇人妻久久综合中文| 国产日韩欧美亚洲二区| 午夜老司机福利片| 成人国产一区最新在线观看 | 午夜福利视频在线观看免费| 首页视频小说图片口味搜索 | 免费久久久久久久精品成人欧美视频| 亚洲国产看品久久| 大片电影免费在线观看免费| 精品一品国产午夜福利视频| 欧美日韩亚洲国产一区二区在线观看 | 欧美日韩黄片免| 十八禁人妻一区二区| 侵犯人妻中文字幕一二三四区| 久久99一区二区三区| 真人做人爱边吃奶动态| 色94色欧美一区二区| 真人做人爱边吃奶动态| 精品久久久久久久毛片微露脸 | av不卡在线播放| 欧美激情高清一区二区三区| 咕卡用的链子| 亚洲男人天堂网一区| 久久久久久亚洲精品国产蜜桃av| 亚洲成人免费电影在线观看 | 大片免费播放器 马上看| av欧美777| 欧美日韩福利视频一区二区| 水蜜桃什么品种好| 精品国产一区二区三区久久久樱花| √禁漫天堂资源中文www| 黄色 视频免费看| 亚洲国产精品一区二区三区在线| 午夜福利视频精品| 蜜桃在线观看..| 极品人妻少妇av视频| 狂野欧美激情性bbbbbb| 欧美日韩综合久久久久久| 夫妻午夜视频| 国产在线一区二区三区精| 美女大奶头黄色视频| 在线观看免费高清a一片| av视频免费观看在线观看| 777米奇影视久久| 男女无遮挡免费网站观看| 午夜久久久在线观看| 免费少妇av软件| 久久久国产一区二区| h视频一区二区三区| 欧美国产精品一级二级三级| 黄片小视频在线播放| 久久国产精品人妻蜜桃| 在线天堂中文资源库| 脱女人内裤的视频| 老司机影院毛片| 国产成人精品久久二区二区免费| 黑丝袜美女国产一区| 亚洲精品国产av成人精品| 成年美女黄网站色视频大全免费| 超碰成人久久| 国产福利在线免费观看视频| 精品少妇一区二区三区视频日本电影| 制服诱惑二区| 在线亚洲精品国产二区图片欧美| 精品久久久久久电影网| 欧美精品高潮呻吟av久久| 久久久久国产一级毛片高清牌| 在线观看一区二区三区激情| 久久久国产一区二区| 精品人妻在线不人妻| 一级毛片电影观看| 中文字幕人妻丝袜一区二区| 国产免费视频播放在线视频| 免费观看人在逋| 岛国毛片在线播放| 国产亚洲av片在线观看秒播厂| 久久精品国产亚洲av涩爱| 日日爽夜夜爽网站| 国产精品.久久久| 搡老乐熟女国产| 亚洲国产看品久久| 欧美大码av| 国产成人一区二区在线| 精品少妇久久久久久888优播| 美女高潮到喷水免费观看| 9191精品国产免费久久| 在线看a的网站| 亚洲专区国产一区二区| 久9热在线精品视频| 少妇粗大呻吟视频| 国产成人av教育| 日本91视频免费播放| 国产精品人妻久久久影院| 人妻 亚洲 视频| 国产精品 欧美亚洲| kizo精华| 久久综合国产亚洲精品| 捣出白浆h1v1| 亚洲中文日韩欧美视频| 在线看a的网站| 97精品久久久久久久久久精品| 青草久久国产| 女人被躁到高潮嗷嗷叫费观| 性少妇av在线| 国产精品麻豆人妻色哟哟久久| 亚洲av成人精品一二三区| 亚洲av在线观看美女高潮| 精品欧美一区二区三区在线| 熟女av电影| 天堂8中文在线网| 黄色一级大片看看| 精品久久久精品久久久| 天天躁夜夜躁狠狠久久av| 在线观看www视频免费| 久久精品久久久久久噜噜老黄| av在线app专区| 久热这里只有精品99| 激情五月婷婷亚洲| 久久精品国产亚洲av涩爱| 色播在线永久视频| 交换朋友夫妻互换小说| 一区二区三区激情视频| 亚洲欧美成人综合另类久久久| 涩涩av久久男人的天堂| 国产成人啪精品午夜网站| 深夜精品福利| 一级毛片我不卡| 男女下面插进去视频免费观看| 亚洲成人免费电影在线观看 | 久久人人爽人人片av| 日本色播在线视频| 午夜福利视频在线观看免费| 国产高清videossex| 欧美激情极品国产一区二区三区| 99国产精品一区二区三区| 最新的欧美精品一区二区| 老司机靠b影院| 人人妻人人澡人人看| 亚洲国产精品国产精品| 自拍欧美九色日韩亚洲蝌蚪91| 国产日韩欧美在线精品| 男女国产视频网站| 日韩熟女老妇一区二区性免费视频| 午夜91福利影院| 午夜福利在线免费观看网站| 日韩免费高清中文字幕av| 国产亚洲精品久久久久5区| 777久久人妻少妇嫩草av网站| 久久精品成人免费网站| 另类精品久久| 国产极品粉嫩免费观看在线| 男女无遮挡免费网站观看| 我的亚洲天堂| e午夜精品久久久久久久| 国产一区有黄有色的免费视频| 每晚都被弄得嗷嗷叫到高潮| 日韩人妻精品一区2区三区| 在线 av 中文字幕| 国产亚洲欧美在线一区二区| 国产成人免费无遮挡视频| 久久鲁丝午夜福利片| 日日摸夜夜添夜夜爱| 欧美黄色片欧美黄色片| 在线观看国产h片| 亚洲精品久久成人aⅴ小说| 狂野欧美激情性xxxx| 老熟女久久久| 又大又黄又爽视频免费| 欧美 日韩 精品 国产| av网站免费在线观看视频| a级毛片黄视频| 亚洲七黄色美女视频| av电影中文网址| 中文字幕另类日韩欧美亚洲嫩草| 婷婷色综合www| 成年美女黄网站色视频大全免费| 又黄又粗又硬又大视频| 最近中文字幕2019免费版| 一区二区三区乱码不卡18| 人人妻人人澡人人爽人人夜夜| 精品少妇内射三级| 免费高清在线观看日韩| 亚洲国产最新在线播放| 天天添夜夜摸| 一本综合久久免费| 男女之事视频高清在线观看 | 国产爽快片一区二区三区| bbb黄色大片| 好男人电影高清在线观看| 亚洲欧美成人综合另类久久久| 亚洲av成人不卡在线观看播放网 | 亚洲免费av在线视频| 十八禁高潮呻吟视频| 午夜av观看不卡| 亚洲成人手机| 亚洲成色77777| 亚洲精品第二区| 欧美日韩视频精品一区| 18禁观看日本| 欧美少妇被猛烈插入视频| videosex国产| 亚洲精品一卡2卡三卡4卡5卡 | 女人精品久久久久毛片| 午夜福利,免费看| 欧美 亚洲 国产 日韩一| 只有这里有精品99| 少妇精品久久久久久久| 亚洲精品久久午夜乱码| 男女下面插进去视频免费观看| 纯流量卡能插随身wifi吗| 国产亚洲午夜精品一区二区久久| 国产精品九九99| 我的亚洲天堂| cao死你这个sao货| 日韩制服丝袜自拍偷拍| 日韩大片免费观看网站| 午夜福利,免费看| 2021少妇久久久久久久久久久| 成人黄色视频免费在线看| 搡老岳熟女国产| 亚洲精品av麻豆狂野| 久久久亚洲精品成人影院| 自线自在国产av| 国产成人啪精品午夜网站| 久久女婷五月综合色啪小说| 久久久久国产精品人妻一区二区| 免费在线观看黄色视频的| 赤兔流量卡办理| 色播在线永久视频| 国产黄频视频在线观看| 波多野结衣av一区二区av| 免费在线观看日本一区| cao死你这个sao货| 黑人巨大精品欧美一区二区蜜桃| 中国国产av一级| 久久精品国产亚洲av高清一级| 久久久久久久国产电影| 欧美精品av麻豆av| 男女之事视频高清在线观看 | 免费人妻精品一区二区三区视频| 国产精品麻豆人妻色哟哟久久| av一本久久久久| 亚洲精品国产av蜜桃| 欧美精品av麻豆av| 性色av乱码一区二区三区2| 少妇 在线观看| 亚洲成国产人片在线观看| 19禁男女啪啪无遮挡网站| 久久国产精品影院| 女人精品久久久久毛片| 国产欧美日韩精品亚洲av| 国产精品一区二区精品视频观看| 亚洲av男天堂| 成年人免费黄色播放视频| 国产精品熟女久久久久浪| 婷婷色麻豆天堂久久| 国产精品久久久av美女十八| 国产成人免费无遮挡视频| 日韩大码丰满熟妇| 国产真人三级小视频在线观看| 久久99热这里只频精品6学生| 欧美精品一区二区大全| 丝瓜视频免费看黄片| 精品一区在线观看国产| 51午夜福利影视在线观看| 中文字幕亚洲精品专区| 精品一区在线观看国产| 男女之事视频高清在线观看 | 香蕉国产在线看| 一区福利在线观看| 国产免费一区二区三区四区乱码| 日本av免费视频播放| 午夜福利一区二区在线看|