• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Skin Lesion Segmentation and Classification Using Conventional and Deep Learning Based Framework

    2022-08-24 03:26:56AminaBibiMuhamamdAttiqueKhanMuhammadYounusJavedUsmanTariqByeongGwonKangYunyoungNamRehamMostafaandRashaSakr
    Computers Materials&Continua 2022年5期

    Amina Bibi,Muhamamd Attique Khan,Muhammad Younus Javed,Usman Tariq,Byeong-Gwon Kang,Yunyoung Nam,*,Reham R.Mostafa and Rasha H.Sakr

    1Department of Computer Science,HITEC University,Taxila,Pakistan

    2College of Computer Engineering and Sciences,Prince Sattam Bin Abdulaziz University,Al-Khraj,Saudi Arabia

    3Department of ICT Convergence,Soonchunhyang University,Asan,31538,Korea

    4Information Systems Department,F(xiàn)aculty of Computers and Information Sciences,Mansoura University,Mansoura,35516,Egypt

    5Computer Science Department,F(xiàn)aculty of Computers and Information Sciences,Mansoura University,Mansoura,35516,Egypt

    Abstract:Background:In medical image analysis,the diagnosis of skin lesions remains a challenging task.Skin lesion is a common type of skin cancer that exists worldwide.Dermoscopy is one of the latest technologies used for the diagnosis of skin cancer.Challenges:Many computerized methods have been introduced in the literature to classify skin cancers.However,challenges remain such as imbalanced datasets,low contrast lesions,and the extraction of irrelevant or redundant features.Proposed Work:In this study,a new technique is proposed based on the conventional and deep learning framework.The proposed framework consists of two major tasks:lesion segmentation and classification.In the lesion segmentation task,contrast is initially improved by the fusion of two filtering techniques and then performed a color transformation to color lesion area color discrimination.Subsequently,the best channel is selected and the lesion map is computed,which is further converted into a binary form using a thresholding function.In the lesion classification task,two pre-trained CNN models were modified and trained using transfer learning.Deep features were extracted from both models and fused using canonical correlation analysis.During the fusion process,a few redundant features were also added,lowering classification accuracy.A new technique called maximum entropy score-based selection(MESbS)is proposed as a solution to this issue.The features selected through this approach are fed into a cubic support vector machine(C-SVM)for the final classification.Results:The experimental process was conducted on two datasets:ISIC 2017 and HAM10000.The ISIC 2017 dataset was used for the lesion segmentation task, whereas the HAM10000 dataset was used for the classification task.The achieved accuracy for both datasets was 95.6% and 96.7%,respectively, which was higher than the existing techniques.

    Keywords: Skin cancer; lesion segmentation; deep learning; features fusion;classification

    1 Introduction

    Skin cancer is a popular research topic due to the high number of deaths and diagnosed cases [1].Cancer is a group of diseases characterized by unrestrained development and the spread of atypical cells.This may cause death if the expansion of irregular cells is not controlled.Skin carcinoma is an irregular expansion of skin cells that frequently appears on the skin when exposed to sunlight or ultraviolet rays.Skin cancer is a fatal disease that can be classified into two types:melanoma and benign (basal cell and squamous cell carcinoma).Benign is constantly retorting to treatment and hardly spreads to other skin tissues.Melanoma is a dangerous type of skin cancer that starts in the pigment cells.Skin cancer develops as a result of malignant lesions and accounts for approximately 75% of all deaths [2].

    In the United States of America, 2021 cases are reported to be 207,390, of which 106,110 are noninvasive and 101,280 are invasive, including 62,260 men and 43,850 women.The estimated death count in 2021 in the USA is 7,180, including 4600 men and 2580 women (https://www.cancer.org/content/dam/cancer-org/research/cancer-facts-and-statistics/annual-cancer-facts-and-figures/2021/cancer-facts-and-figures-2021.pdf).The number of cases reported in the United States of America in 2020 is 100,350, including 60,190 men and 40,160 women, with 6,850 deaths from melanoma, including 4,610 men and 2,240 women.Since 2019, the total number of skin cancer patients in the USA has been 192,310.The death count has been 7,230, including 4,740 men and 2,490 women.In 2020, it is estimated that over 16,221 novel cancer cases were analyzed in Australia, including 9,480 men and 6,741 women, with a death count of 1,375, including 891 men and 484 women (https://www.canceraustralia.gov.au/affected-cancer/cancer-types/melanoma/statistics).According to dermatologists, if a melanoma is not detected at a very early stage, it spreads to the entire body or nearby tissues.However, if detected early on, there is a good chance of survival [3].Melanoma has received a lot of attention from the research community because of its high mortality rate.

    Dermatologists have previously used the ABCDE rule, a seven-point check list, laser technology, and a few other methods [4].However, these methods require an expert dermatologist.In addition, manual inspection and diagnosis of skin cancer using these methods is difficult,time-consuming, and expensive.Therefore, it is essential to develop a computerized method for automated skin cancer segmentation and classification [5].Dermoscopy is a new technology for the diagnosis of skin cancer [6].Through dermoscopy technology, RGB images of the skin are captured and later analyzed by experts.A computerized method consists of the following steps:preprocessing of dermoscopic images, segmentation of skin lesions, feature extraction, and finally classification [7].Preprocessing is the step in which low-contrast images are enhanced and artifacts such as hair and noise can be removed through different dermoscopic image techniques [8].This step follows the segmentation step in which the lesion region is segmented based on the shape and color of the lesion, and irregularity of the border [9].Many techniques for lesion segmentation have been introduced in the literature.Some focused on traditional techniques, and few used convolutional neural networks (CNNs).Feature extraction is the third step used to represent an image [10].In this step, image features are extracted such as color, texture, shape, and name.Color is an important feature in skin cancer classification [11].These different features are fused later to obtain the maximum image information [12].However, one major disadvantage is high computational time required to complete this step.Many researchers have implemented feature selection techniques to select the most valuable features.The main purpose of this approach is to obtain maximum accuracy with less computational time.In addition, this step is useful for the redundancy of irrelevant features for classification [13,14].The final step is to classify the features.Features are classified using different classifiers in a relevant category, such as benign or malignant [5].

    More recently, deep learning models have been shown to significantly contribute to medical image analysis for both segmentation and classification [15,16].In deep learning, CNNs are used for classification as they are composed of several hidden layers such as convolutional, pooling,batch normalization, ReLU, and fully connected layers [17,18].CV studies have introduced many techniques for the segmentation and classification of skin lesions.Afza et al.[19] presented a hierarchical framework for skin lesion segmentation and classification.They began with a preprocessing step to enhance the quality of images before running a segmentation algorithm.Later, the ResNet50 model was fine-tuned, and features were extracted.The extracted features are refined using the grasshopper optimization algorithm, which is classified using the Na?ve Bayes algorithm.The experimental process was conducted on three dermoscopy datasets, and improved accuracy was achieved.Zhang et al.[20] presented an intelligent framework for multiclass skin lesion classification.In this method, the skin lesions were initially segmented using MASK RCNN.In the classification phase, they proposed 24 layered CNN model.Three datasets were used for the experimentation of the segmentation phase and the HAM10000 dataset was used for classification.On these datasets, the accuracy of the proposed method was improved.Akram et al.[21] presented a CAD system for skin lesion localization.They applied a de-correlation operation at the initial step and then passed it to the MASK RCNN for lesion segmentation.In the next step, the DenseNet201 pre-trained model is modified, and features are extracted from the two layers.The extracted features were fused and further refined using a selection block.The experimental process was conducted on dermoscopy datasets, and improved performance was achieved.Alom et al.[22]introduced a deep learning architecture for the segmentation of skin lesions.In this model, the best features are initially selected to better represent the lesion region, and then inception RCNN was applied for the final lesion classification.Dermoscopy datasets were employed for evaluation and achieved improved accuracy.Thomas et al.[23] applied interpretable CNN models for the classification of skin cancers.In this method, the outer padding was applied in the first step and then iterated through overlapping tiles.The next step segments the lesion, which later crops for the final segmentation.Al-Masni et al.[24] presented a two-stage deep learning framework for skin lesion segmentation and classification.The segmentation was performed using a fully resolved CNN (FrCNN), and four pre-trained networks were considered for the final classification.Sikkandar et al.[25] presented a computerized method for the segmentation and classification of skin lesions using traditional techniques.The authors combined the performance of the GrabCut and Neuro Fuzzy (NF) classifier for the final classification.In the preprocessing step, top-hat filtering and in-painting techniques were applied.In the later step, the GrabCut algorithm was applied to the segmentation task.In the feature extraction phase, deep learning features are extracted and finally classified using the NF classifier.A mutual bootstrap method was also presented in [26] for skin lesion classification.

    The methods discussed above have some limitations that affect the performance of skin lesion segmentation and classification.The following are the major issues:i) the presence of hair bubbles and irrelevant areas not required for detecting accurate skin lesions; ii) low contrast skin lesions are a factor for inaccurate lesion segmentation; iii) knowledge of useful feature extraction for the accurate classification of skin lesion types; iv) presence of irrelevant features that mislead correct classification; v) manual inspection of skin lesions is time consuming, and vi) accuracy is always dependent on an expert.In this work, we proposed a new computerized method by amalgamating traditional and deep learning methods.The proposed method includes contrast enhancement of dermoscopic images, segmentation of skin lesions, deep learning feature extraction and fusion,selection of the best features, and classification.Our major contributions are as follows:

    ·A contrast enhancement approach was implemented based on the fusion of the haze reduction approach and fast local Laplacian filters.The fusion process followed the HSV color transformation.

    ·The best channel is selected based on the probability value, and then a saliency map is constructed, which is later converted into a binary form using a threshold function.

    ·Two modified pre-trained models, MobileNet V2 and VGG16, were trained on dermoscopic datasets using transfer learning.Later, the features were extracted from the dense layers.

    ·Canonical correlation-based features were fused and later refined using the maximum entropy score-based selection(MESbS).

    The remainder of this article is organized as follows:the proposed methodology is presented in Section 2, the results are detailed in Section 3.Finally, the conclusions are presented in Section 4.

    2 Proposed Methodology

    The proposed method comprises two main tasks:lesion segmentation and classification.For lesion segmentation, a hybrid contrast enhancement technique was proposed, and the best channel was selected based on the histogram.Subsequently, an activation function was proposed to construct a saliency map.In the later stage, the threshold function is applied to convert the image into binary form, which is then mapped onto the original image for final detection.For classification,two pre-trained models were modified and trained through transfer learning.The features were extracted from both models and fused using canonical correlation analysis (CCA).Subsequently,the fused vector was further refined using the highest entropy score.Finally, multiple classifiers were used for the classification of selected features.Several datasets were used for the experimental process, and the results were obtained in visual and numeric form.The detailed architecture of the proposed methodology is illustrated in Fig.1.

    2.1 Lesion Segmentation Task

    As shown in Fig.1, the proposed method performs two tasks:lesion segmentation and classification.The lesion segmentation task is described in this section.Here, a hybrid method was initially proposed for contrast enhancement of the original dermoscopy images.Then, an HSV color transformation was applied and the best channel was selected based on the activation function.Subsequently, a lesion map was constructed based on the selected channel.The resultant lesion map was finally converted into binary form based on a threshold function.The details of each step are as follows:

    Hybrid Contrast Enhancement:The first step was the hybrid contrast enhancement.Here,the image quality is enhanced and bubbles are removed.For this purpose, two techniques were implemented, and the resultant information was fused into one image.First, a haze reduction technique was implemented to clear the boundaries of the lesion region.AssumeU(x) is the input image,S(x) is the hazy image, andY(x) is the medium of transmission.Seff(x) is the image affected by haze and is represented as follows:

    Figure 1:Proposed parallel architecture of skin lesion segmentation and classification

    This image is affected by reflected light represented as follows:

    Here,Ωrepresents a local patch with its origin atX.After this method, an estimation of transmissionY(x) is required before proceeding.Second, a fast local Laplacian filter was implemented to smooth the image and emphasize the edges.The local Laplacian can be defined as follows:

    where,Z(I) is a sample function, wherehis the reference value,γcontrols the amount of increasing and decreasing value,αcontrols the dynamic range compression and expansion, andwzdefines the threshold function.

    whererrepresents distribution time that how much time the process will run anduis the weighted function which is 1.

    wheretrepresents number of iterations have been performed,A4represents the 4-neighborhood oft, andPshows the input image.Mathematically, it is defined as follows:

    whereHωxandHωzare the Gaussian kernels.After that HSV color transformation is applied.HSV consists of three channels such as hue, saturation, and value.Through this transformation,the image is refined in terms of lesion colors.The visual results of this transformation are showing in Fig.2b.From this figure, we extracted the Hue channel for lesion map construction.Mathematically, this channel is defined as follows:

    where,R′=R/255,G′=G/255,B′=B/255,Amax=max(R′,G′,B′), and ?=Amax-Amin.In the next step, an activation function was constructed based on the multiplication function after which a lesion map was constructed, this was later converted into a binary form using a threshold function.Mathematically, the activation and threshold functions are defined as follows:

    Figure 2:Proposed lesion segmentation task results.(a) Original image; (b) Enhanced image; (c)Binary segmented image; (d) Mapped image; (e) Localized image

    Here, 0.4 is computed based on the mean value of all computed pixels of H.The visual results of the threshold function are shown in Fig.2.In this figure, the binary images are shown in (c),whereas the lesion mapped and final localized images are illustrated in (d) and (e), respectively.The final localized images are compared with the ground truth images for the final evaluation process.The numerical segmentation results are presented in Tab.1.The ISIC 2017 dataset was used for the experimental process, and an average dice rate of 95.6% was achieved.For each image, three parameters were computed:dice, Jaccard distance, and Jaccard index.

    Table 1:Sample numerical results of lesion segmentation task

    2.2 Skin Lesion Classification

    During this phase, skin lesions are classified into relevant categories such as melanoma, bkl,and others.For classification, the features were extracted from the input images.Feature extraction is an important step in pattern recognition, and many descriptors have been extracted from the literature.More recently, deep learning has shown success in the classification of medical infections [27,28].A CNN is a deep learning method used for feature extraction [29].A simple CNN model consists of many layers, such as a convolutional layer, ReLU layer, pooling, normalization,fully connected, and softmax.

    VGG16—VGG-16 contains N number of fully connected layers, where N = 1, 2, 3....ThePNunits are in the Nth layer forNST= 224,NX= 224, andNXH= 3.The dataset is represented byγand the training sample is shown by∈γ.Eachis a real number .

    where r(.) represents the activation function ReLu6.ST expresses the no of rows, X symbolizes the number of columns, and XH symbolizes number of channels.Represents the bias vector andn(1)express the weights of the first level which is defined as below:

    The output of the first layer is used as the input of the second layer and so on.This is shown in the mathematical form below:

    wheren(2)∈RN(2)×N(1)andn(2)∈RN(2)×N(1).Soφ(z)represents the last fully connected layer that is used for high level feature extraction.Mathematically expression of last layer is shown as below:

    Visually, the architecture of VGG16 is showing in Fig.3.

    Figure 3:Architecture of VGG-16 CNN model

    In Fig.3, the original architecture includes a total of 16 layers; the first 13 layers are convolutional and the final three are fully connected.The output was generated using softmax.In this study, we modified the VGG-16 pre-trained CNN model for skin cancer classification.For this purpose, the last layer was removed, and a new layer that included seven classes of skin carcinoma was added.These classes are known as the target labels.Transfer learning was then applied to transfer the knowledge of the original model to the target model and obtain a new customized CNN model.This model can be used for feature extraction.The modified architecture of the VGG-16 model is shown in Fig.4.

    MobileNetV2—MobileNet V2 is a CNN model designed specifically for portable and resourceconstrained circumstances.It is founded on an upturned residual structure in which the connections of the residual structure are linked to the bottleneck layers [30].There are 153 layers in MobileNet V2, and the size of the input layer ish×w×k, whereh=224,w=224, andkrepresents the channels, three of which are in the first layer.There are two types of residual blocks in MobileNet V2, with strides 1 and 2.These blocks had three types of layers and were used for downsizing.The first layer is 1×1 convolutions with ReLU6, where ReLU6 is an activation function.It is min (max(x, 0), 6).The second layer is a depthwise convolution used to crop the unnecessary information, and the third layer is a 1×1 convolution, but without nonlinearity.Each layer has batch normalization and activation functions, but the third layer has only batch normalization because the output of this layer has less dimension, and by using ReLU6, the performance will decrease [31].The convolutional block of MobileNet V2 is shown in Fig.5.In the basic architecture of MobileNet V2, there is a convolution layer with 32 filters, followed by 19 residual bottleneck layers.The detailed architecture is presented in Tab.2.

    Figure 4:Fine-tuned VGG-16 model for lesion classification

    Figure 5:MobileNetV2 convolutional blocks [31]

    In the original architecture, there were 153 layers.The output is generated in the last layer.In our work, we used the MobileNet-V2 pre-trained CNN model for skin cancer classification.For this purpose, the original architecture was fine-tuned and the last layer was removed.Subsequently,a new layer was added that includes seven skin classes.These classes are known as the target labels.Subsequently, transfer learning (TL) was used to transfer the knowledge of the original model to the target model and obtain a new customized CNN model.The TL process is discussed in the next section.After the training process, features are extracted from the feature layer(convolutional layer).

    Table 2:Architecture of mobileNetV2 [31]

    2.2.1 Transfer Learning for Feature Extraction

    Transfer learning is a technique that transfers information from a pre-trained model to a modified CNN model for a new task.The primary objective was to obtain the result for the target problem with better performance [32].Given a source domain Dsand target domain as DT,the learning task isTs,andTt.Transfer learning assists the learning of the target predictive functionF(t) in the target domain with knowledge in the source domain and learning task, whereDs≠DtandTs≠Tt.Fig.6 illustrates the TL process for the modified VGG16 model for skin lesion feature extraction.Fig.6 illustrates that the source data ?Dare from the Imagenet dataset,the source model is VGG16, represented by ?Modand the source labels ?Lare 1000.The target dataψτare the HAM10000 dataset, the target model is modified VGG16, and the target labels are seven, represented byψL.Through TL, the weights, and parameters of the VGG16 model are transferred to the modified VGG16 model, whereas the following condition holds.

    Similarly, this process was performed for the modified MobileNet V2 CNN model.In this study, the MobileNet V2 model was used as the source model and the modified MobileNet V2 model was used as the target model (Fig.6).After training both modified models, the deep learning features were extracted from the FC7 layer (modified VGG16) and convolutional layer(modified MobileNet V2).The extracted feature vector sizes of both vectors wereN×4096 andN×1056, respectively.

    Figure 6:Transfer learning process for VGG16 CNN model

    2.2.2 Features Fusion and Selection

    Feature fusion is an important research area and many techniques have been introduced for the fusion of two or more feature vectors [33].The most useful fusion techniques are serial-based,parallel, and correlation-based approaches.In this study, we used the CCA approach [34] for the fusion of both extracted feature vectors.Using CCA, a fused vector is obtained withN×1750 dimensions.However, after the fusion process, we determined that some features had repeated and had to be removed from the final vector.For this purpose, a new method called the MESbS has been proposed.In this approach, initially, the entropy vector is computed using the fused feature vector (column-based).Then, the entropy vector is sorted in descending order.Subsequently, we computed the mean value of the entropy vector and used this value as a threshold function for selecting the best features.Mathematically, this process is defined as follows:

    where,EnFVrepresents the entropy feature vector,μis the mean entropy value, andFuncis the final threshold function.Through this function, the featureFv(i) values greater than the mean value are considered for the final selection, and the remaining features are discarded.Lastly,the final selected features are classified using a multiclass SVM classifier with a one-against-all method.

    3 Experimental Results and Discussion

    3.1 Experimental Setup

    This section presents the experimental process for the proposed classification process.The HAM10000 dataset [35] was used.This dataset consists of approximately 10,000 dermoscopic images in RGB format.A total of seven skin lesion classes, Bkl, Bcc, Vasc, Akiec, Nevi, Mel,and Df.This dataset is highly imbalanced because of the high variation in the number of sample images in each class.Many classifiers are used to compare the accuracy of the proposed method on a cubic SVM.To train the classifiers, a 70:30 approach was used.This ratio indicates that 70% of the images were considered for the training process and 30% for the testing process.The recall rate (TPR), precision rate (PPV), FNR, AUC, accuracy, and time were calculated for each classifier in the evaluation process.All experiments were conducted in MATLAB 2020b on a system with an Intel(R) Core(TM) i5-7200u CPU running at 2.50 and 2.7 GHz, with 16 GB RAM, and an 8 GB graphics card.

    3.2 Proposed Classification Results

    This section presents the proposed classification results in a numerical and confusion matrix.The results were obtained using four different experiments.The first experiment extracted features from the modified VGG16 CNN model and used them in the experimental process.The results are presented in Tab.3.In this table, it can be observed that the cubic SVM showed a better accuracy of 78.2%, whereas the computational time of this classifier was approximately 468 s.The minimum computational time of this experiment was 83.230 s for the Fine KNN classifier.The recall rate of the cubic SVM was 78.2%.

    Table 3:Classification results using modified VGG16 CNN model

    In the second experiment, features were extracted from the modified MobileNet V2 CNN model and used in the experimental process.The results presented in Tab.4 show that the best accuracy of 82.1% was achieved on the cubic SVM.This classifier performed better than the other classifiers listed in this table.The computational time of the cubic SVM was approximately 91 s,whereas the minimum noted time was 20 s for the linear discriminate classifier.The recall rate of the cubic SVM was 82.1%.This table illustrates that the correct prediction accuracy of each class is better than that of the confusion matrix of the modified VGG16 (Tab.4).In addition,the accuracy of this experiment was improved compared to Experiment 1.

    Table 4:Classification results using modified mobileNetV2 model

    In this experiment, we fused the features of both models using CCA.The results are presented in Tab.5 which shows that the maximum accuracy achieved is 82.8% on the cubic SVM.The other calculated evolution measures include a recall rate of 82.1%, a precision rate of 82.97%, an FNR of 17.03%, and an AUC value of 0.97.The computational time of this classifier is 988.07 s.The recall rate of the cubic SVM is 82.81%, as shown in Fig.7.The minimum time required for this experiment was approximately 245 s.From Tab.5, it can be observed that the accuracy of all classifiers increases slightly, however, the execution time increases significantly.This indicates that there are many redundant features included in the fused vector, which degrades the classification accuracy.

    Table 5:Classification result of fused models

    Figure 7:Confusion matrix of cubic SVM for fused features of both models

    In this experiment, features were selected based on the MESbS approach; the results are detailed in Tab.6.In Tab.6, it can be observed that the top-attained accuracy is 96.7% on cubic SVM, whereas the additional calculated measures have a recall rate of 88.31%, a precision rate of 94.48%, an FNR of 5.52%, and an AUC value of 0.98.The computational time is 51.771 s,which is significantly minimized compared to Experiments 1 and 3.The recall rate of the cubic SVM was 88.31%, as shown in Fig.8.From Fig.8, it can be observed that the correct prediction accuracy of each skin lesion class is considerably higher than that of the first three experiments.In addition, the overall computational time of this experiment decreased.Hence, based on the results, we can demonstrate that the proposed method outperforms the proposed framework.A fair comparison was also conducted with the recent techniques, given in Tab.7, which shows the proposed framework outclass for multiclass lesion classification.

    Table 6:Skin lesion classification results using proposed framework

    Figure 8:Confusion matrix of cubic SVM using proposed framework

    Table 7:Comparison of the proposed method with recent techniques

    4 Conclusion

    A conventional and deep learning-based framework is proposed in this study for skin lesion segmentation and classification using dermoscopy images.Two tasks were performed.In the first task, conventional techniques-based skin lesions were segmented.The contrast of lesions was improved for accurate lesion map creation.The accurate lesion map creation process improves segmentation accuracy.The segmentation performance was evaluated on the ISIC 2017 dataset and achieved an accuracy of 95.6%.In the classification tasks, VGG16 and MobileNet V2 CNN models were fine-tuned and trained through TL on dermoscopic images.These models performed better according to recent studies in the medical image processing field.The features were extracted from these fine-tuned trained CNN models and fused using the CCA approach.The main purpose of fusion in this study was to increase image information.However, some redundant features were also added during the fusion process.The redundant features have an impact on classification accuracy.Therefore, we propose MESbS, a novel feature selection method.This method selects the features and classifies them using the C-SVM classifier.The results of our experiments demonstrate better accuracy than the existing techniques.We conclude that the lesion contrast enhancement step improves segmentation accuracy.In addition, the selection of the best features increases classification accuracy and minimizes execution time.Future studies will focus on the CNN for lesion segmentation and provide segmented lesions to modified models for useful feature extraction.

    Funding Statement:This research was supported by Korea Institute for Advancement of Technology (KIAT) grant funded by the Korea Government (MOTIE) (P0012724, The Competency Development Program for Industry Specialist) and the Soonchunhyang University Research Fund.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    欧美日韩一级在线毛片| 国产成人啪精品午夜网站| 精品一区在线观看国产| 免费一级毛片在线播放高清视频 | 一二三四在线观看免费中文在| 亚洲 国产 在线| 亚洲精品一二三| 亚洲av日韩在线播放| 国产亚洲精品久久久久5区| 免费少妇av软件| 国产成人精品在线电影| 热99re8久久精品国产| 天堂8中文在线网| 国产精品99久久99久久久不卡| 久久久国产欧美日韩av| 午夜福利免费观看在线| 欧美日韩一级在线毛片| 黄色怎么调成土黄色| 黄色视频在线播放观看不卡| 丝瓜视频免费看黄片| 欧美午夜高清在线| 国产精品偷伦视频观看了| 国产男人的电影天堂91| 亚洲国产欧美网| 亚洲精品成人av观看孕妇| 波多野结衣av一区二区av| 欧美日韩中文字幕国产精品一区二区三区 | 国产精品99久久99久久久不卡| 丁香六月天网| 亚洲av日韩精品久久久久久密| 91国产中文字幕| 亚洲avbb在线观看| 美国免费a级毛片| 国产欧美日韩一区二区三区在线| 国产精品久久久人人做人人爽| 久久人妻福利社区极品人妻图片| 久久香蕉激情| 一区二区三区四区激情视频| 热re99久久精品国产66热6| 精品一品国产午夜福利视频| 99精国产麻豆久久婷婷| 精品国产一区二区久久| 精品国内亚洲2022精品成人 | 亚洲国产精品一区三区| 一边摸一边做爽爽视频免费| 国产精品香港三级国产av潘金莲| 叶爱在线成人免费视频播放| 美女主播在线视频| 午夜免费鲁丝| 久久人人爽人人片av| 欧美av亚洲av综合av国产av| 国产亚洲欧美在线一区二区| 亚洲五月婷婷丁香| 一级毛片女人18水好多| 亚洲欧美激情在线| 日韩欧美一区视频在线观看| 成人亚洲精品一区在线观看| 丝袜人妻中文字幕| 国产精品一区二区免费欧美 | 久久久久网色| 国产免费av片在线观看野外av| 51午夜福利影视在线观看| 亚洲一卡2卡3卡4卡5卡精品中文| 亚洲欧洲精品一区二区精品久久久| 免费日韩欧美在线观看| 青春草视频在线免费观看| 欧美xxⅹ黑人| 亚洲avbb在线观看| 久久人人爽人人片av| 免费黄频网站在线观看国产| 叶爱在线成人免费视频播放| 亚洲伊人久久精品综合| 国产在视频线精品| 精品亚洲成a人片在线观看| www.av在线官网国产| 欧美日韩视频精品一区| 亚洲精品av麻豆狂野| 久久久精品区二区三区| 成年人免费黄色播放视频| 在线观看免费日韩欧美大片| 成人影院久久| 国产无遮挡羞羞视频在线观看| 久久久欧美国产精品| 亚洲国产欧美一区二区综合| 高清av免费在线| 国产成+人综合+亚洲专区| 国产精品影院久久| 亚洲九九香蕉| 色综合欧美亚洲国产小说| 两性午夜刺激爽爽歪歪视频在线观看 | 亚洲国产欧美一区二区综合| 婷婷丁香在线五月| 曰老女人黄片| 久久中文看片网| 久久精品成人免费网站| 两个人看的免费小视频| 国产亚洲av高清不卡| 午夜福利一区二区在线看| 啪啪无遮挡十八禁网站| 青春草亚洲视频在线观看| 国产精品免费视频内射| 人人妻人人爽人人添夜夜欢视频| 在线观看免费日韩欧美大片| 国产精品一区二区免费欧美 | 各种免费的搞黄视频| 国产精品久久久久成人av| 国产又爽黄色视频| 亚洲国产精品一区二区三区在线| 大片电影免费在线观看免费| 一二三四在线观看免费中文在| 精品一品国产午夜福利视频| 最新的欧美精品一区二区| 国产精品一区二区免费欧美 | 夜夜夜夜夜久久久久| 亚洲av成人一区二区三| 国产精品香港三级国产av潘金莲| 亚洲激情五月婷婷啪啪| 人人澡人人妻人| 亚洲精品在线美女| 免费高清在线观看日韩| 国产激情久久老熟女| 一二三四在线观看免费中文在| 精品视频人人做人人爽| 黄片大片在线免费观看| 亚洲精品久久午夜乱码| 国产欧美日韩一区二区三区在线| 午夜福利免费观看在线| 亚洲综合色网址| 精品福利观看| 日韩,欧美,国产一区二区三区| 在线观看免费日韩欧美大片| 亚洲欧美日韩高清在线视频 | 欧美在线黄色| a级毛片在线看网站| 亚洲国产精品一区三区| 日韩制服丝袜自拍偷拍| 亚洲精品中文字幕一二三四区 | 热99国产精品久久久久久7| 青春草视频在线免费观看| 大型av网站在线播放| 成人三级做爰电影| 亚洲三区欧美一区| 性高湖久久久久久久久免费观看| 99久久国产精品久久久| 大片免费播放器 马上看| 日韩欧美国产一区二区入口| 国产成人精品在线电影| 777米奇影视久久| 我要看黄色一级片免费的| 天天影视国产精品| av超薄肉色丝袜交足视频| 国产精品亚洲av一区麻豆| 午夜福利影视在线免费观看| 欧美97在线视频| 一级,二级,三级黄色视频| 久久人人爽av亚洲精品天堂| 国产欧美日韩一区二区三区在线| 久久中文字幕一级| 青春草亚洲视频在线观看| 纯流量卡能插随身wifi吗| 99精国产麻豆久久婷婷| 久久影院123| 黄色 视频免费看| 人人妻人人澡人人爽人人夜夜| 各种免费的搞黄视频| 国产精品1区2区在线观看. | 国产无遮挡羞羞视频在线观看| 国产亚洲欧美在线一区二区| 欧美一级毛片孕妇| 久久久久久免费高清国产稀缺| 国产精品九九99| e午夜精品久久久久久久| 男人爽女人下面视频在线观看| 欧美老熟妇乱子伦牲交| 99久久精品国产亚洲精品| 一边摸一边抽搐一进一出视频| 久久中文看片网| 国产成人一区二区三区免费视频网站| 老司机影院成人| 男人添女人高潮全过程视频| 99精品欧美一区二区三区四区| 久久99热这里只频精品6学生| 国产亚洲精品一区二区www | 这个男人来自地球电影免费观看| 日韩中文字幕视频在线看片| 热99久久久久精品小说推荐| 亚洲久久久国产精品| 成人亚洲精品一区在线观看| 国产黄色免费在线视频| 精品一区二区三区四区五区乱码| 亚洲综合色网址| 亚洲中文日韩欧美视频| 亚洲欧美精品综合一区二区三区| 亚洲伊人色综图| 午夜福利乱码中文字幕| 国产精品av久久久久免费| 欧美日韩视频精品一区| 午夜激情久久久久久久| 99九九在线精品视频| 动漫黄色视频在线观看| 国产野战对白在线观看| 一区二区三区精品91| 亚洲 欧美一区二区三区| 十八禁高潮呻吟视频| 亚洲欧美一区二区三区黑人| av天堂久久9| 亚洲国产成人一精品久久久| 叶爱在线成人免费视频播放| 高清在线国产一区| 97在线人人人人妻| 久久免费观看电影| 99九九在线精品视频| 少妇被粗大的猛进出69影院| 欧美日韩福利视频一区二区| 国产成人a∨麻豆精品| 少妇 在线观看| 18禁裸乳无遮挡动漫免费视频| 国产精品欧美亚洲77777| 国产免费一区二区三区四区乱码| 久久久久久免费高清国产稀缺| 人妻一区二区av| 国产色视频综合| 多毛熟女@视频| 五月天丁香电影| 啦啦啦免费观看视频1| 肉色欧美久久久久久久蜜桃| 老司机午夜十八禁免费视频| 91精品国产国语对白视频| 啦啦啦啦在线视频资源| 在线永久观看黄色视频| 国产精品av久久久久免费| 十分钟在线观看高清视频www| 九色亚洲精品在线播放| 免费高清在线观看日韩| 成人18禁高潮啪啪吃奶动态图| 日本五十路高清| 伊人亚洲综合成人网| 一级毛片精品| 久久亚洲国产成人精品v| 精品久久久久久电影网| 飞空精品影院首页| 汤姆久久久久久久影院中文字幕| 美国免费a级毛片| 午夜福利,免费看| 各种免费的搞黄视频| 999精品在线视频| 狠狠狠狠99中文字幕| xxxhd国产人妻xxx| kizo精华| 成人三级做爰电影| 精品人妻一区二区三区麻豆| 成人手机av| 又黄又粗又硬又大视频| 亚洲中文av在线| 精品一区在线观看国产| 欧美在线黄色| 三级毛片av免费| 国产精品1区2区在线观看. | 老司机影院成人| 捣出白浆h1v1| 久久久精品94久久精品| 在线观看人妻少妇| 免费在线观看黄色视频的| 成年人午夜在线观看视频| 午夜福利,免费看| 一本色道久久久久久精品综合| 亚洲三区欧美一区| 精品福利永久在线观看| 亚洲中文日韩欧美视频| 亚洲成av片中文字幕在线观看| 国产区一区二久久| 三上悠亚av全集在线观看| 母亲3免费完整高清在线观看| 美女福利国产在线| 人妻一区二区av| 老司机福利观看| 在线亚洲精品国产二区图片欧美| 中国美女看黄片| 搡老熟女国产l中国老女人| 午夜福利,免费看| 一级片'在线观看视频| 伊人久久大香线蕉亚洲五| 又大又爽又粗| 老熟妇乱子伦视频在线观看 | 黑人欧美特级aaaaaa片| 国产日韩欧美在线精品| 日本一区二区免费在线视频| 久久国产精品影院| 亚洲av欧美aⅴ国产| 国产在视频线精品| 丝袜美腿诱惑在线| 久久国产亚洲av麻豆专区| 国产91精品成人一区二区三区 | 手机成人av网站| 亚洲性夜色夜夜综合| 日日夜夜操网爽| 国产一卡二卡三卡精品| 伦理电影免费视频| 另类亚洲欧美激情| 国产成人啪精品午夜网站| 波多野结衣av一区二区av| 亚洲欧美成人综合另类久久久| 建设人人有责人人尽责人人享有的| 男女下面插进去视频免费观看| 亚洲av成人一区二区三| 亚洲人成77777在线视频| 高潮久久久久久久久久久不卡| 午夜福利乱码中文字幕| 亚洲欧洲日产国产| 一区福利在线观看| 午夜激情久久久久久久| 韩国精品一区二区三区| 啪啪无遮挡十八禁网站| 精品一区二区三区av网在线观看 | 人人澡人人妻人| 91字幕亚洲| 国产三级黄色录像| 三级毛片av免费| netflix在线观看网站| 午夜影院在线不卡| 国产精品 国内视频| 国产精品一区二区在线不卡| 老汉色∧v一级毛片| 国产精品.久久久| 18禁黄网站禁片午夜丰满| 精品少妇久久久久久888优播| 欧美 亚洲 国产 日韩一| 香蕉丝袜av| 操出白浆在线播放| 欧美精品亚洲一区二区| 女性生殖器流出的白浆| 久久久久久久精品精品| 欧美av亚洲av综合av国产av| 最新的欧美精品一区二区| 日韩 亚洲 欧美在线| www.999成人在线观看| 后天国语完整版免费观看| 亚洲精品第二区| 视频区欧美日本亚洲| av又黄又爽大尺度在线免费看| 精品一区在线观看国产| 中文字幕制服av| 天堂中文最新版在线下载| 一区二区三区乱码不卡18| 国产淫语在线视频| 麻豆国产av国片精品| 亚洲 欧美一区二区三区| 国产不卡av网站在线观看| 亚洲熟女精品中文字幕| 9191精品国产免费久久| 国产欧美亚洲国产| 国产亚洲欧美精品永久| 丝袜在线中文字幕| 亚洲美女黄色视频免费看| 女人爽到高潮嗷嗷叫在线视频| 国产xxxxx性猛交| 亚洲第一欧美日韩一区二区三区 | 日日夜夜操网爽| 亚洲欧美色中文字幕在线| 又大又爽又粗| 日韩中文字幕欧美一区二区| 久久av网站| 国产色视频综合| 成人黄色视频免费在线看| 各种免费的搞黄视频| 亚洲成人手机| 五月开心婷婷网| 美女扒开内裤让男人捅视频| 老熟妇仑乱视频hdxx| 久久九九热精品免费| 不卡av一区二区三区| 十八禁网站网址无遮挡| xxxhd国产人妻xxx| 男人添女人高潮全过程视频| 国产熟女午夜一区二区三区| 不卡一级毛片| av网站免费在线观看视频| 我的亚洲天堂| 亚洲精品美女久久av网站| 成人国产av品久久久| 咕卡用的链子| 99国产精品免费福利视频| 高潮久久久久久久久久久不卡| 欧美av亚洲av综合av国产av| 国产黄色免费在线视频| 黑人巨大精品欧美一区二区mp4| 黄色视频,在线免费观看| 18禁裸乳无遮挡动漫免费视频| 老汉色∧v一级毛片| 欧美日本中文国产一区发布| 精品久久久久久电影网| 50天的宝宝边吃奶边哭怎么回事| 亚洲第一av免费看| 久久精品国产亚洲av高清一级| 男人操女人黄网站| 欧美精品av麻豆av| 久久久久久人人人人人| 中文字幕另类日韩欧美亚洲嫩草| 窝窝影院91人妻| 无遮挡黄片免费观看| 在线av久久热| 久久久久久人人人人人| 色精品久久人妻99蜜桃| 性色av一级| 男女之事视频高清在线观看| 一区二区三区四区激情视频| 亚洲第一青青草原| 欧美另类亚洲清纯唯美| 午夜影院在线不卡| 国产激情久久老熟女| 成人国产一区最新在线观看| 精品国产一区二区久久| 久久久久久人人人人人| 91国产中文字幕| 每晚都被弄得嗷嗷叫到高潮| 一级,二级,三级黄色视频| 久久国产亚洲av麻豆专区| 国产成人精品在线电影| 精品国产一区二区久久| 国产成人一区二区三区免费视频网站| 成人三级做爰电影| 亚洲欧美清纯卡通| 午夜免费成人在线视频| a在线观看视频网站| 欧美久久黑人一区二区| 国产淫语在线视频| 青春草亚洲视频在线观看| 精品国产一区二区三区四区第35| av片东京热男人的天堂| 亚洲国产看品久久| 国产成人免费观看mmmm| 日韩欧美一区视频在线观看| 国产精品亚洲av一区麻豆| 黄色 视频免费看| 久久精品久久久久久噜噜老黄| 亚洲三区欧美一区| 男女下面插进去视频免费观看| 久久久久国产精品人妻一区二区| 国产精品av久久久久免费| 我要看黄色一级片免费的| 国产欧美亚洲国产| 黄色a级毛片大全视频| svipshipincom国产片| 性色av一级| 精品免费久久久久久久清纯 | 国产麻豆69| 韩国精品一区二区三区| 亚洲精品日韩在线中文字幕| 久久久久视频综合| 日本五十路高清| 老司机福利观看| 久久人人97超碰香蕉20202| 一本色道久久久久久精品综合| 亚洲国产欧美日韩在线播放| 十分钟在线观看高清视频www| 国产又爽黄色视频| 亚洲国产欧美在线一区| 亚洲,欧美精品.| √禁漫天堂资源中文www| 深夜精品福利| 久久久久精品国产欧美久久久 | 午夜福利乱码中文字幕| 久久久久久久久免费视频了| 午夜福利一区二区在线看| 人人妻人人添人人爽欧美一区卜| 国产精品国产av在线观看| 天堂8中文在线网| 国产在线视频一区二区| 亚洲精品久久午夜乱码| 欧美激情高清一区二区三区| 丝袜美足系列| 啪啪无遮挡十八禁网站| 午夜福利影视在线免费观看| 又大又爽又粗| 少妇粗大呻吟视频| 成年人黄色毛片网站| 日本五十路高清| 黄色视频在线播放观看不卡| 亚洲色图 男人天堂 中文字幕| 欧美+亚洲+日韩+国产| 亚洲国产欧美在线一区| 亚洲,欧美精品.| 国产高清国产精品国产三级| 亚洲第一青青草原| 国产成人啪精品午夜网站| 亚洲精品美女久久久久99蜜臀| 男女国产视频网站| www.av在线官网国产| 欧美日韩国产mv在线观看视频| 精品久久久久久久毛片微露脸 | 大陆偷拍与自拍| 亚洲中文日韩欧美视频| 亚洲黑人精品在线| 精品国产乱子伦一区二区三区 | 99香蕉大伊视频| 精品亚洲成a人片在线观看| 国产激情久久老熟女| 日本五十路高清| 亚洲成国产人片在线观看| 亚洲成人国产一区在线观看| 欧美国产精品一级二级三级| 性高湖久久久久久久久免费观看| 久久99热这里只频精品6学生| 亚洲性夜色夜夜综合| 老司机在亚洲福利影院| 亚洲一码二码三码区别大吗| 黄网站色视频无遮挡免费观看| 丝袜人妻中文字幕| 精品人妻1区二区| 久9热在线精品视频| 99久久人妻综合| 国产色视频综合| 极品少妇高潮喷水抽搐| 桃红色精品国产亚洲av| 黑人巨大精品欧美一区二区mp4| 国产免费福利视频在线观看| 啦啦啦啦在线视频资源| 亚洲熟女精品中文字幕| 天天躁狠狠躁夜夜躁狠狠躁| 日韩大片免费观看网站| 男人添女人高潮全过程视频| 亚洲精品自拍成人| 国产高清国产精品国产三级| 国产男女内射视频| 美女国产高潮福利片在线看| 考比视频在线观看| 亚洲美女黄色视频免费看| 国产成人av激情在线播放| 免费黄频网站在线观看国产| 交换朋友夫妻互换小说| 黄色a级毛片大全视频| 一区福利在线观看| 成年女人毛片免费观看观看9 | 欧美成狂野欧美在线观看| 欧美成人午夜精品| 大陆偷拍与自拍| 亚洲精品国产色婷婷电影| 久久精品国产a三级三级三级| 777米奇影视久久| 亚洲国产毛片av蜜桃av| 十八禁人妻一区二区| 婷婷成人精品国产| 亚洲色图 男人天堂 中文字幕| 亚洲国产毛片av蜜桃av| 制服诱惑二区| 国产日韩欧美在线精品| 久久青草综合色| 欧美日本中文国产一区发布| 亚洲国产精品成人久久小说| 中文字幕另类日韩欧美亚洲嫩草| 久久久精品区二区三区| av又黄又爽大尺度在线免费看| 精品一区二区三卡| tocl精华| 黄色毛片三级朝国网站| 欧美xxⅹ黑人| 久久 成人 亚洲| 亚洲黑人精品在线| 少妇粗大呻吟视频| 精品国产一区二区久久| 久久亚洲精品不卡| 久久99热这里只频精品6学生| 女性生殖器流出的白浆| 欧美精品啪啪一区二区三区 | 中国美女看黄片| 99久久人妻综合| 久久九九热精品免费| 国产真人三级小视频在线观看| 亚洲国产中文字幕在线视频| 成人免费观看视频高清| 女性被躁到高潮视频| 人成视频在线观看免费观看| 每晚都被弄得嗷嗷叫到高潮| 国产黄频视频在线观看| 纯流量卡能插随身wifi吗| 亚洲av男天堂| 深夜精品福利| 亚洲精品久久午夜乱码| 俄罗斯特黄特色一大片| 超碰成人久久| 如日韩欧美国产精品一区二区三区| 成人av一区二区三区在线看 | 精品人妻1区二区| 三级毛片av免费| 国产精品久久久人人做人人爽| www.精华液| 99国产综合亚洲精品| 精品高清国产在线一区| 欧美黑人精品巨大| 在线观看免费高清a一片| 久久精品人人爽人人爽视色| 亚洲精品第二区| 黄片大片在线免费观看| 精品一区二区三卡| 欧美中文综合在线视频| 啦啦啦免费观看视频1| 一级,二级,三级黄色视频| 久热爱精品视频在线9| 亚洲精品自拍成人| 午夜免费成人在线视频| 爱豆传媒免费全集在线观看| 99re6热这里在线精品视频| 日本av手机在线免费观看| 男女免费视频国产| 国产老妇伦熟女老妇高清| 99精品久久久久人妻精品| 99国产极品粉嫩在线观看| 99热网站在线观看| kizo精华| 人人妻人人添人人爽欧美一区卜| 十八禁网站网址无遮挡| 美女福利国产在线| 性色av乱码一区二区三区2| 国产精品99久久99久久久不卡| 嫁个100分男人电影在线观看|