• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    A New Method for Diagnosis of Leukemia Utilizing a Hybrid DL-ML Approach for Binary and Multi-Class Classification on a Limited-Sized Database

    2024-01-20 13:01:40NilkanthMukundDeshpandeShilpaGiteBiswajeetPradhanAbdullahAlamriandChangWookLee

    Nilkanth Mukund Deshpande ,Shilpa Gite ,Biswajeet Pradhan ,Abdullah Alamri and Chang-Wook Lee

    1Department of Electronics&Telecommunication,Lavale,Symbiosis Institute of Technology,Symbiosis International(Deemed University),Pune,Maharashtra,412115,India

    2Electronics&Telecommunication,Dr.Vithalrao Vikhe Patil College of Engineering,Ahmednagar,Maharashtra,414111,India

    3Artificial Intelligence and Machine Learning Department,Symbiosis Institute of Technology,Symbiosis International(Deemed)University,Pune,412115,India

    4Symbiosis Centre of Applied AI(SCAAI),Symbiosis International(Deemed)University,Pune,412115,India

    5Centre for Advanced Modelling and Geospatial Information Systems(CAMGIS),School of Civil and Environmental Engineering,Faculty of Engineering&IT,University of Technology Sydney,Sydney,Australia

    6Earth Observation Centre,Institute of Climate Change,Universiti Kebangsaan Malaysia,Bangi,Selangor,43600,Malaysia

    7Department of Geology&Geophysics,College of Science,King Saud University,P.O.Box 2455,Riyadh,11451,Saudi Arabia

    8Department of Science Education,Kangwon National University,Chuncheon-si,24341,Korea

    ABSTRACT Infection of leukemia in humans causes many complications in its later stages.It impairs bone marrow’s ability to produce blood.Morphological diagnosis of human blood cells is a well-known and well-proven technique for diagnosis in this case.The binary classification is employed to distinguish between normal and leukemiainfected cells.In addition,various subtypes of leukemia require different treatments.These sub-classes must also be detected to obtain an accurate diagnosis of the type of leukemia.This entails using multi-class classification to determine the leukemia subtype.This is usually done using a microscopic examination of these blood cells.Due to the requirement of a trained pathologist,the decision process is critical,which leads to the development of an automated software framework for diagnosis.Researchers utilized state-of-the-art machine learning approaches,such as Support Vector Machine (SVM),Random Forest (RF),Na?ve Bayes,K-Nearest Neighbor (KNN),and others,to provide limited accuracies of classification.More advanced deep-learning methods are also utilized.Due to constrained dataset sizes,these approaches result in over-fitting,reducing their outstanding performances.This study introduces a deep learning-machine learning combined approach for leukemia diagnosis.It uses deep transfer learning frameworks to extract and classify features using state-of-the-art machine learning classifiers.The transfer learning frameworks such as VGGNet,Xception,InceptionResV2,Densenet,and ResNet are employed as feature extractors.The extracted features are given to RF and XGBoost classifiers for the binary and multi-class classification of leukemia cells.For the experimentation,a very popular ALL-IDB dataset is used,approaching a maximum accuracy of 100%.A private real images dataset with three subclasses of leukemia images,including Acute Myloid Leukemia(AML),Chronic Lymphocytic Leukemia(CLL),and Chronic Myloid Leukemia(CML),is also employed to generalize the system.This dataset achieves an impressive multi-class classification accuracy of 97.08%.The proposed approach is robust and generalized by a standardized dataset and the real image dataset with a limited sample size(520 images).Hence,this method can be explored further for leukemia diagnosis having a limited number of dataset samples.

    KEYWORDS Leukemia diagnosis;deep learning;machine learning;random forest;XGBoost

    1 Introduction

    The ability of bone marrow to form blood results in the production of various blood components.Blood,an essential human body component,keeps the human alive.It consists of various cells such as White Blood Cells (WBC),Red Blood Cells (RBC),plasma (a fluid made up of various gases and electrolytes),and others [1].An infection in the body alters blood components regarding their counts,size,shape,and others.These changes must be analyzed to diagnose the particular infection[2].Leukemia is an infection that causes an increase in the number of blast cells in the blood[3].These are immature WBC blasts.Because of the higher percentage of blasts,RBC has very little space to flow through the blood.This reduces the percentage of RBC,resulting in anemia and other abnormalities[4].The leukemia is broadly classified as chronic and acute,with lymphoblastic and myeloid leukemia being sub-classified[5].Fig.1 depicts the various blood cells with and without leukemia infection.

    Figure 1:(a)Normal cells,and(b)Leukemia cells[6]

    The best way to diagnose leukemia is to examine the morphology of white blood cells under a microscope [7].Such diagnosis requires a highly trained and experienced pathologist [8],as the shape,size,and presence of Auer rods in the blood sample influence the decision regarding leukemia infection.This complexity in diagnostic decisions has prompted many researchers to explore the issue and develop software solutions for diagnostic purposes[9].Various methods such as SVM[10],N?ive Bayes[11],decision trees[12],random forest[13],XGBoost[14],AdaBoost[15],CNN[16],and other software algorithms are employed by researchers for diagnostic decisions.

    Traditional image processing techniques like thresholding[17],edge detection[18],and contour detection[19]are employed to detect and segment leukemia.These techniques offer optimal accuracies compared to machine learning and deep learning methods[20].However,to enhance system robustness and improve diagnostic accuracy,more advanced techniques,including machine learning and deep learning,are employed [21].Due to the limited number of sample images in medical datasets,deep learning techniques can lead to overfitting[22],affecting detection accuracy[23].However,there exists a trade-off between accuracy and interpretability with machine learning[24].

    In order to address these challenges,this study presents a hybrid approach to harness the strengths of both machine learning and deep learning models.In many strategies,machine learning or deep learning classifiers are used exclusively.Given the limited sample images,deep learning classification can result in overfitting,thereby limiting accuracy.Thus,instead of relying solely on deep learning,the approach is hybridized with cutting-edge machine learning algorithms.Transfer learning frameworks are employed solely for feature extraction.These features are then passed to machine learning algorithms for the final classification.This approach is innovative and robust for limited dataset images,achieving an accuracy of 100% for the standard dataset (ALL-IDB) and approaching 96%for multi-class classification of the real-image-dataset.

    The paper’s second section introduces the proposed methodology and dataset,followed by the results and discussion section.The conclusion is then presented,accompanied by remarks on future work directions.

    2 Related Work

    There are different traditional image processing and machine learning approaches employed by researchers for leukemia detection,such as support vector machine(SVM)[25-28],SVM and k-means clustering[29],K-nearest neighbor(KNN)and Na?ve Bayes[30],KNN[31],Zack algorithm[32,33],and deep learning approaches such as CNN[34],deep learning nets[35],CNN plus SVM approach[36,37],AlexNet [38,39],Ensembles of CNN [40],a special CNN called ConVNet [41],and ANN[42].Singhal et al.[43]used a support vector machine(SVM)classifier for leukemia detection.This work utilized features including geometry and local binary pattern(LBP),which achieved accuracies of 88.79%and 89.72%,respectively.

    Mohamed et al.[44] presented a study on the diagnosis of sub-classes of leukemia,viz.(ALL and AML) and Myeloma.Features were extracted via methods such as scale-invariant feature transform(SIFT),speeded-up robust features(SURF),and oriented FAST and rotated BRIEF(ORB)approaches.These different features were combined and passed to classifiers,such as SVM,K-nearest neighbor(KNN),and random forest(RF).The accuracy of their approach reached 94.3%using the RF classifier.Here,pre-processing and segmentation stages should be included for better results.After this,different features such as area,perimeter,statistical feature,and gray-level-co-occurrence-matrix are utilized,and these were hand-crafted.Instead of using hand-crafted features,more sophisticated deep learning feature extractors could improve performance.Also,DL feature extractors could be tested for better results without segmentation.

    Mohapatra et al.[45] used an ensemble classifier in which the data were segmented before classification.Afterward,feature extraction was performed considering different morphological and textural features.Their results were compared with state-of-the-art machine learning algorithms(Na?ve Bayes,KNN,multilayer perceptron,SVM,radial basis functional network) and were found superior with an accuracy of 94.73%.This work produces sub-images consisting of a single malignantinfected cell.Hence,during the analysis of real images,this work might fail to demonstrate its robustness and generalization.Additionally,this work experiments on binary classification,but in a practical scenario,there is a need to classify different sub-classes of leukemia,necessitating multi-class classification.

    Mishra et al.[46] proposed a feature extractor based on the gray-level co-occurrence matrix(GLCM).Probabilistic principal component analysis was also used,followed by an RF classifier.This method used marker-based segmentation before the application of RF,which can be removed when DL methods are applied.This method deals only with binary classification,considering ALL and normal cells.Multi-class classification with leukemia sub-classes is not addressed in this work.Das et al.[47]employed feature extraction based on the gray-level-run-length matrix(GLRLM)and GLCM texture features,and they also considered color and shape features.The most prominent features were selected based on PCA and then used with an SVM for binary leukemia classification.This method segments sub-images based on which leukemia decisions are made.When a larger dataset is used,this sub-image segmentation takes longer,increasing the time required for classifier decisions.Real image datasets contain many leukocyte blasts,making it challenging to segment each via sub-image segmentation.The number of blasts determines the disease’s severity.Hence,applying this method to a real-image dataset may challenge the achievement of good accuracies.

    Abdeldaim et al.[48]proposed a system that extracted different features,including shape,color,texture,and hybrid features.After normalizing these features,several state-of-the-art machine learning algorithms were used for final classification,such as Na?ve Bayes with Gaussian distribution,decision tree,SVM with RBF kernel,and KNN.In their work,only binary classification is considered,and sub-classes of leukemia are not addressed.In research,Mandal et al.[49] used a decision tree(DT) with Gradient Boosting for the classification of Acute Lymphoblastic Leukemia (ALL).This work did not measure the accuracy metric;instead,the F1-score was considered with sensitivity and specificity.The maximum F1-score was 85%,which could be improved further.Mishra et al.[50]conducted experimentation using a discrete orthogonal S-transform and linear discriminant analysis(LDA)feature extractors.The final classification was done with the AdaBoost algorithm.This work addresses only the binary classification of leukemia,detecting normal and infected cells.Different leukemia subtypes also need to be predicted in a practical scenario,and multi-class classification must be implemented.Al-Jaboriy et al.[51] used artificial neural networks (ANN) with a four-moment statistical feature extraction stage.The accuracy obtained in this technique was 97%,but the algorithm should be generalized by applying it to a real-image dataset and considering multi-class classification to identify the sub-classes of leukemia.

    In 2020,Banik et al.[52] applied a convolutional neural network (CNN)-based approach for detecting white blood cells in microscopic blood images.In this method,the features from the first and last layers were combined to enhance performance.Pre-processing and segmentation were conducted before extracting features via the CNN.This approach should be generalized by considering a realimage dataset.Honnalgere et al.[53]used the VGG16 framework with transfer learning from a pretrained model on the ImageNet dataset.Batch normalization and data augmentation helped achieve a larger sample size.Their experiments resulted in a precision,recall,and F1-score of 91.70%,91.75%,and 91.70%,respectively.Shah et al.[54]introduced a method that combined CNN and a recurrent neural network (RNN),achieving an accuracy of 86.6%.Although decent,this can be improved.Yu et al.[55] explored this area with a CNN-based approach,considering and combining different frameworks such as ResNet50[56],InceptionV3[57],VGG16[58],VGG19[59],and Xception.Their methods achieved an accuracy of 88.5%.Pan et al.[60] proposed an approach using a pre-trained RNN with various feature extraction and combination stages,claiming an F1-score of 92.50%.Marzahl et al.[61]employed the ResNet18 deep learning framework to detect leukemia and normal cells.With advanced dataset augmentation,they achieved an F1 score of 87.46%.Still,there is a need for more reliable and robust methods,especially for smaller datasets.The methods by[54-60]achieved accuracy up to 92.50%,and the same researchers[54-60]tested binary classification on the standard ALL-IDB datasets.Hence,it is necessary to generalize and fortify these methods by applying them to a real-image dataset with multi-class classification.

    In a separate work,Mittal et al.[62]presented a review of leukemia detection using blood smear images.They explored blood components,databases,leukemia types,and various methodologies for leukemia detection.Different steps,such as pre-processing,segmentation,feature extraction,and classification methods employed by various researchers,are reviewed in this study.The challenges faced by classifiers,including large training data,achieving good accuracy,generalization,and reproducibility,are highlighted in this survey.The need for data augmentation is also discussed.

    The need for a large dataset and augmentation,as mentioned by Mittal et al.[62],can be addressed by employing a hybrid approach that combines deep learning and machine learning.The number of required dataset images can be reduced if deep learning is used for feature extraction and machine learning for the final classification.

    Although the methods discussed in the previous section achieved reasonable accuracies,there is still room for improvement.While deep learning algorithms offer higher accuracies compared to traditional algorithms [63],the challenge lies in the dataset’s image size.For leukemia,there is a limitation on the number of images in the dataset.Consequently,deep learning algorithms might face over-fitting during training.Thus,this study introduces a solution using a hybrid machine learningdeep learning(ML-DL)combination.Features are extracted through a deep learning framework,and classification is executed using RF and XGBoost algorithms.This experiment achieved notable results using the publicly available dataset(ALL-IDB)and a private real-image dataset.

    The ALL-IDB1 dataset contains 109 images,while the ALL-IDB2 has 260.For multi-class classification,there are 528 images divided into three sub-classes.Since deep learning is employed solely for automatic feature extraction and not for classification,there is no risk of overfitting during experimentation.Although data augmentation is typically employed to increase sample sizes,achieving the desired accuracy with the available samples meant that augmentation was unnecessary in this study.The proposed approach,combining deep learning and machine learning,promises enhanced performance for datasets with limited images,applicable to binary and multi-class classification.

    3 Proposed Methodology

    Fig.2 shows the proposed hybrid approach methodology for the diagnosis of leukemia via binary classification and multi-class classification.

    3.1 Image Loading

    Images from the dataset are loaded to process them and propagate in further stages for classification.ALL-IDB dataset is used for the processing.

    3.2 Feature Extraction

    There are different frameworks available with convolutional neural networks.Some popular frameworks,including VGG16,VGG19,ResNet50,Xception,Inception-ResNet-v2,and Densenet,are employed for feature extraction.These architectures are explained below.

    3.2.1VGGNet[64]

    CNNs’depth was increased with VGG to improve model performance.Visual Geometry Group,or VGG,is a typical deep CNN architecture with several layers.Two classical models were developed as VGG,namely VGG16 and VGG19.Fig.3 depicts a typical VGG16 architecture.Very tiny convolutional filters are used in the construction of the VGG network.A total of sixteen layers,of which three are fully connected,and thirteen convolutional layers are present in the VGG16 network.The VGGNet accepts input of size 224×224-sized images.

    Figure 2:Proposed methodology of leukemia classification

    Convolutionallayers: This layer uses a receptive field of 3×3 to record two movements,viz.left to right and up to down.Additionally,1×1 convolution filters are employed for the linear transformation of the input.The next part is a ReLU unit,a progressive step over AlexNet regarding training time [65].The convolution stride is set to 1 pixel to maintain spatial resolution.The stride represents the number of pixel shifts across the input matrix.

    Hiddenlayers:These layers primarily make use of ReLU.Local Response Normalization(LRN)is commonly avoided when using VGG because it increases memory usage and training time.Furthermore,it does not affect overall accuracy.

    Fully-connectedlayers:It has three layers that are fully connected.The first two levels have 4096 channels each.The third layer has 1000 channels,one for each class.

    3.2.2ResNet[66]

    ResNet50,as shown in Fig.4,is a variant of a typical Reset model.It has 50 layers in total.The model consists of 48 convolutional,one maximum pool,and one average pooling layer.It has one convolution layer having a 7×7 kernel size and distinct kernels of 64.All layers have a stride size of 2.After this,there is a max pooling layer with a size of stride 2.Another convolution contains a 1×1 layer,with 64 kernel,then a 3×3 sized kernel,which continues with a 64 kernel and a 1×1,256 kernel.This phase has nine layers.Following that,there is a kernel of 1×1,128.After this,3×3,128 kernel and a kernel of 1×1,512 size at the final stage.

    This stage was repeated four times,which resulted in 12 layers.After this stage,there are different combinations,including 1×1,3×3,and 1×1,with convolution filter sizes of 256,256,and 512,respectively,repeated six times.The sequence of 1×1,3×3,and 1×1 was repeated with convolution filter sizes of 512,512,and 2048,respectively.A fully connected layer of 1000 nodes comes at the last with the Softmax function.

    Figure 3:VGG16 architecture

    3.2.3XceptionNet[67]

    This has depth-wise separable convolutions.Xception is another name for an“extreme”version of an Inception module.Xception is an abbreviation for “extreme inception”,it uses the Inception principles.There is a use of 1×1 convolutions to compress the original input in Inception,and different filters were used on each depth space from each input space.Xception simply reverses this step.This method is nearly identical to the depth-wise separable convolution used in designing the neural network.There is one more distinction between Inception and Xception.The presence or absence of a nonlinearity following the initial operation.A ReLU nonlinearity follows both operations in the Inception model,whereas Xception does not.The data is processed after passing through the entry flow,through the middle flow(which repeats eight times),and through the exit flow.Fig.5 shows the detailed architecture of XceptionNet.

    Figure 4:ResNet architecture

    3.2.4Inception-ResNet-v2[68]

    Inception-ResNet-v2 was trained with a vast number of images from the ImageNet database.The 164-layer network can classify images into numerous additional object categories,allowing it to learn detailed representations of features for a wide range of images.The input to the network is a 299×299 image,and it produces class probabilities at the output.This network is a combination of Inception and residual connections[69].It has a residual connection with a combination of convolutional filters of multiple sizes.Due to the use of the residual connections,the degradation issue was compensated,and training time was also reduced to some extent.Fig.6 shows the basic network architecture of Inception-ResNet-v2.

    3.2.5Densenet[70]

    Fig.7 shows a Densenet architecture consisting of low level layers,high level layers,upsampling layers,and integration layers.Densenet has shorter connections between the layers,making it efficient during training[71].Each layer is deeply connected to the other,ensuring the free flow of information within the network.Every layer receives input from previous layers to maintain the feed-forward nature.This input is then passed on to subsequent layers via its feature maps.There is the concatenation of features of different layers.So,the ithlayer has I input and comprises feature maps from all previous convolutional blocks.All subsequent‘I-i’layers receive their feature maps.In contrast to traditional deep learning architectures,this introduces ‘(I(I+1))/2’connections into the network.It requires fewer parameters than traditional CNN because it does not need to learn unimportant feature maps.Densenet has two crucial blocks besides the basic convolutional and pooling layers.They are known as Transition Layers and Dense Blocks.

    Figure 5:Xception architecture

    Figure 6:Inception-ResNet-Version-2 architecture

    After the feature extraction,popular state-of-the-art,classical machine learning algorithms are employed for classification.RF and XGBoost classifiers are utilized after the feature extraction via deep learning frameworks.

    3.3 Classification

    For this stage,RF and XGBoost classification techniques are employed.

    3.3.1RF

    These models predict output by combining the outcomes of multiple regression decision trees[72].The architecture of RF is shown in Fig.8.It has many decision trees,each with its training dataset.As a result,this algorithm performs better in classification problems than a decision tree[73].The pruning method selection and the branching criteria are regarded as the two most important factors[74].This algorithm is determined by the number of created trees and the utilized samples by a particular node.Each tree is generated independently as a random vector sampled from input data,while the forest’s tree distribution remaining constant.The forest predictions are averaged using bootstrap aggregation and random feature selection[75].

    Figure 7:Densenet architecture

    Figure 8:RF architecture

    3.3.2XGBoost

    Extreme Gradient Boosting (XGBoost) is a scalable,distributed gradient-boosted decision tree(GBDT)machine learning library[76].It works with many problems,including classification,regression,and ranking.This algorithm supports parallel tree boosting[77].Gradient-boosted decision trees are implemented in XGBoost,which are dominant in many cases [78].This algorithm sequentially generates decision trees [79].Weights are crucial in XGBoost [80].The architecture of XGBoost is shown in Fig.9.In this algorithm,the weights are assigned to each independent variable.These weights are then transferred into the decision tree,which predicts the outcome.The weight of variables incorrectly predicted by the tree is increased,and these variables are input into the second decision tree.Then,individual classifiers and predictors are combined to create a more robust and precise model.It can solve regression,classification,ranking,and user-defined prediction problems[81].

    Figure 9:XGBoost architecture

    3.3.3HyperparameterTuningofModel

    As explored in Subsection 3.3.2,two state-of-the-art machine learning models are employed for final classification,including RF and XGBoost.The following steps were used for the tuning of hyperparameter of the classification models:

    1.RF is tuned with n_estimators(no.of trees).It has experimented with the number of trees with values of 1 to 100.

    2.For getting the best value of n_estimaters,the GridSearchCV object is used.It performs an exhaustive search over the hyperparameter grid,fitting and evaluating the RF model for the specified values of trees.

    3.Cross-validation is employed to estimate the model’s performance,and the value of n_estimater obtained is 24.Hence,fine-tuning is done to get the value of 24 trees.

    4.XGBoost is tuned for n_estimators,maximum depth,and Gamma.These parameters are tuned by considering n_estimators,Maximum depth,and Gamma.

    5.With the GridSearchCV object,the parameters are obtained after the tuning with the following values.In order to find the best values of these parameters,these parameters are tested with values as n_estimators with 10,20,40,and 100;max_depth:3,4,and 5;Gamma with the values as 0,0.15,0.30,0.50,and 1.

    6.Finally,the values of these parameters obtained for the best accuracies with cross-validation of 5 for the XGBoost are as follows:

    i) N_estimators=10

    ii) Maximum Depth=3

    iii) Gamma=0

    3.4 Performance Metrics

    Classification of normal and abnormal cell images and multi-class classification are the prime motivations of the presented work.Different metrics are employed to evaluate the performance of the proposed system.Following are the parameters utilized for accuracy measurement:

    Accuracy[82]:

    where TP=true positive,TN=true negative,FP=false positive,FN=false negative.

    Recall[83]:

    Precision[84]:

    Confusion matrix[86]:

    A confusion matrix is often employed to demonstrate classifier performance based on the four values listed above (TP,FP,TN,and FN).When these are plotted against each other,a confusion matrix is formed.

    4 Datasets Used

    4.1 ALL-IDB1 and ALL-IDB2

    This study utilizes A very popular dataset employed by many researchers,i.e.,ALL-IDB1 and ALL-IDB2.Figs.10a and 10b show images from the ALL-IDB1 dataset,while Figs.10c and 10d show images from the ALL-IDB2 dataset.This database is used for acute leukemia analysis and research.This dataset has two subtypes:ALL-IDB1 and ALL-IDB2.All images were captured using a Canon Powershot G5 camera.The microscope’s magnification range was between 300 and 500.JPEG images with a 24-bit color depth were used.The ALL-IDB1 dataset contains 109 images with a total element count of 3,900 and a resolution of 2592*1944.There are 510 lymphoblasts in total.The ALL-IDB2 dataset consists of 260 images with a total element count of 257 and a resolution of 257×257.It has a total of 130 lymphoblasts.

    Figure 10:(a)ALL-IDB1 infected image;(b)ALL-IDB1 normal image;(c)ALL-IDB2 infected image;(d)ALL-IDB2 normal image

    4.2 Private Real Images Dataset

    In addition to this dataset,a real image dataset is also utilized for the experiment.The dataset is a multi-class dataset with three sub-classes of leukemia,namely acute myeloid leukemia (AML),(Chronic myeloid leukemia) CML,and (Chronic lymphocytic leukemia) CLL.The AML class contains 181 images,while the CLL and CML have 166 and 173 images,respectively.Figs.11a-11c show images from the dataset with CLL,CML,and AML types,respectively.This dataset contains 520 images of blood slides from three classes.This dataset was provided by Nidan Diagnostic in Ahmednagar,India.

    5 Results

    The experimentation was conducted in two ways.A binary classification was performed on ALLIDB1 and ALL-IDB2 datasets,which output metrics showing the normal cell and infected cells’classification.Subsequently,a real-image dataset consisting of different sub-classes of leukemia(AML,CML,and CLL)was tested,yielding a multi-class classification.The following sections discuss this in detail.

    Figure 11:Images from the real dataset:(a)CLL;(b)CML;(c)AML

    5.1 Binary Classification

    Performance metrics considering ALL-IDB1 and ALL-IDB2 datasets for binary classification of leukemia cells and normal cells are shown in this section.

    5.1.1VGG16FeatureExtractor

    Figs.12a-12d show the confusion matrix as an outcome of VGG16 as a feature extractor and RF,and XGBoost classifier for ALL-IDB1 and 2 datasets.

    Figure 12:Confusion matrix considering VGG16 as feature extractor:(a)VGG16 plus RF for ALLIDB1 dataset;(b)VGG16 plus RF for ALL-IDB2 dataset;(c)VGG16 plus XGBoost for ALL-IDB1 dataset;(d)VGG16 plus RF for ALL-IDB2 dataset

    Table 1 shows performance measures using VGG16 as the feature extractor and two classical machine learning algorithms for final classification.RF and XGBoost are employed for classification.Different measures are presented,including precision,recall,F1-score,and accuracy.ALL-IDB1 and ALL-IDB2 both are applied for the analysis.Precision varies from 0.87 to 0.96 for both parts of the dataset.Recall takes the value from 0.86 to 0.96,while the F1-score varies from 0.86 to 0.95 in the case of RF and XGBoost with ALL-IDB1 and ALL-IDB2.The accuracy reaches a peak of 95.45%for the ALL-IDB1 dataset.

    Table 1: Metrics show VGG16 as a feature extractor

    5.1.2VGG19FeatureExtractor

    Figs.13a-13d show the confusion matrices as outcomes of VGG19 used as a feature extractor,combined with RF and XGBoost classifiers,for the ALL-IDB1 and 2 datasets.

    As indicated in Table 2,VGG19 is used as a feature extractor,and RF and XGBoost are employed for final classification.The precision obtained in this case is 0.88 to 0.92,which is a good and permissible value.Recall also varies from 0.87 to 0.91,and the score range is from 0.87 to 0.91.The maximum accuracy value is over 90%for both datasets and the machine learning classifiers.

    Figure 13:Confusion matrix considering VGG19 as feature extractor:(a)VGG19 plus RF for ALLIDB1 dataset;(b)VGG19 plus RF for ALL-IDB2 dataset;(c)VGG19 plus XGBoost for ALL-IDB1 dataset;(d)VGG19 plus RF for ALL-IDB2 dataset

    Table 2: Metrics show VGG19 as a feature extractor

    5.1.3XceptionasaFeatureExtractor

    Figs.14a-14d show the confusion matrix as an outcome of Xception as a feature extractor and RF and XGBoost classifiers for ALL-IDB1 and two datasets.

    Table 3 shows the results obtained utilizing Xceptionnet as a feature extractor,with RF and XGBoost as the final classifiers.Parameters,such as precision,recall,and F1-score,vary from 0.90 to 1.As far as accuracy is concerned,it ranges from 90%to 100%in the cases mentioned above.

    5.1.4ResNetFeatureExtractor

    Figs.15a-15d show the confusion matrices as outcomes of ResNet50 used as a feature extractor,in combination with RF and XGBoost classifiers,for the ALL-IDB1 and ALL-IDB2 datasets.

    ResNet50 is used as a feature extractor using a transfer learning approach in the next case.After extracting features,they are passed through RF and XGBoost algorithms.Different metrics obtained with these approaches are listed in Table 4.Precision,recall,and F1-score achieve a maximum value of 0.86,while the maximum accuracy is 86%.

    5.1.5InceptionResNetV2FeatureExtractor

    Figs.16a-16d show the confusion matrices as outcomes ofInceptionResNetV2used as a feature extractor,in combination with RF and XGBoost classifiers,for the ALL-IDB1 and ALL-IDB2 datasets.

    There is a popular framework consisting of Inception-Restnet-Version-2,which is used as a feature extractor,and both RF and XGBoost as the final stage of the classifier.Table 5 shows the different performance measures in this case.It indicates that precision,recall,and F1-score approaches a maximum of 0.96 and a minimum of 0.86.The maximum and minimum accuracy values are 95%and 86%,respectively.In addition to these networks,Denset is employed to extract features,which are passed through RF and XGBoost algorithms for the final classification.As shown in Table 6,Inception-ResNet-V2 is utilized as a feature extractor,producing a maximum precision of 94%,a maximum recall value obtained of 96%,and the F1-score achieves a maximum value of 95%.For this approach,accuracy ranges from 88%to 95%.

    Figure 14:Confusion matrix considering Xception as feature extractor:(a)Xception plus RF for ALLIDB1 dataset;(b) Xception plus RF for ALL-IDB2 dataset;(c) Xception plus XGBoost for ALLIDB1 dataset;(d)Xception plus RF for ALL-IDB2 dataset

    Table 3: Metrics show Xception as a feature extractor

    Figure 15:Confusion matrix considering ResNet as feature extractor:(a)ResNet plus RF for ALLIDB1 dataset;(b)ResNet plus RF for ALL-IDB2 dataset;(c)ResNet plus XGBoost for ALL-IDB1 dataset;(d)ResNet plus RF for ALL-IDB2 dataset

    Table 4: Metrics showing ResNet50 as a feature extractor

    5.1.6DensenetFeatureExtractor

    Figs.17a-17d show the confusion matrices as outcomes ofDensenetused as a feature extractor,in combination with RF and XGBoost classifiers,for ALL-IDB1 and ALL-IDB2 datasets.

    Fig.18 shows the graphical representation of accuracies obtained by different approaches for binary classification of leukemia,considering datasets ALL-IDB1 and ALL-IDB2.It indicates that the accuracies of the XGBoost classifier with different feature extractor is greater than RF in most cases,except Densenet and Inception-ResNet-V2 as feature extractors.It approaches 100%in Xception and VGG19 as feature extractors.

    5.2 Multi-Class Classification

    Performance metrics consider real image datasets for leukemia sub-classes multi-class classification with Acute Myloid Leukemia (AML),Chronic Myloid Leukemia (CML),and Chronic Lymphocytic Leukemia (CLL) types.Figs.19a-19f show the confusion matrices as outcomes ofdifferentfeatureextractorswithRF classifiers for the ALL-IDB1 and ALL-IDB2 datasets.

    Figure 16: Confusion matrix considering Inception-ResNet-V2 as feature extractor: (a) Inception-ResNet-V2 plus RF for ALL-IDB1 dataset;(b)Inception-ResNet-V2 plus RF for ALL-IDB2 dataset;(c)Inception-ResNet-V2 plus XGBoost for ALL-IDB1 dataset;and(d)Inception-ResNet-V2 plus RF for ALL-IDB2 dataset

    Table 5: Metrics show Inception-ResNet-Version-2 as a feature extractor

    Table 6: Metrics show Densenet as a feature extractor

    Figure 17:Confusion matrix considering Densenet as feature extractor:(a)Densenet plus RF for ALLIDB1 dataset;(b) Densenet plus RF for ALL-IDB2 dataset;(c) Densenet plus XGBoost for ALLIDB1 dataset;(d)Densenet plus RF for ALL-IDB2 dataset

    Table 7 shows different performance metrics with RF for real-image multi-class datasets.All these results are obtained utilizing the real image dataset.Precision value achieves a maximum value of 96%for Densenet and VGG16 feature extractors,while recall achieves a maximum value of 96%for Densenet F1-score,which also achieves a 9%maximum value for Densenet.A good and acceptable range is obtained as the accuracies are observed in this part of the experimentation.It ranges from 88%to 96%.

    Figs.20a-20f show the confusion matrices as outcomes ofdifferentfeatureextractorswithXGBoost classifiers for the ALL-IDB1 and ALL-IDB2 datasets.

    Table 8 presents the performance metrics for XGBoost classification and different deep learning frameworks,including VGG16,VGG19,Xception,Inception-ResNet-V2,ResNet50,and Densenet.Precision takes the minimum value of 94% and a maximum of 95% by observing all the feature extractors’weighted and macro average values.Recall has a maximum value of 89%and a minimum of 97%.Weighted and macro average values of the F1-score range from 89% to 97% while moving through all the feature extractors.A maximum accuracy of 97.08% is obtained with this part of the experimentation.

    Figure 18: Comparison of average accuracies of different approaches for: (a) ALL-IDB1;(b) ALLIDB2 dataset

    Fig.21 compares the accuracies of various feature extractors with RF and XGBoost classifiers.In this Figs.1-6 represent different feature extractors: VGG16,VGG19,Xception,ResNet50,InceprionResNetV2,and Densenet,respectively.The Xception and XGBoost combination achieved the maximum accuracy.

    Table 9 compares the accuracy obtained by the proposed approaches with Ahmed et al.[87].The study employed CNN classifier and other classical machine learning classifiers,including SVM,Na?ve Bayes,decision trees,and KNN.With the same dataset,the accuracies in these approaches varied from 50%to 88%,with the highest accuracy for the CNN classifier.The accuracies obtained by Singhal et al.[43],Patel et al.[88],Rawat et al.[89],and Shahin et al.[90] are 89.72%,93.75%,89.80%,and 96.01%,respectively.The accuracies vary from 74%to 100%compared to the proposed approaches.

    Figure 19:Confusion matrix of different feature extractors with RF classifier for real image dataset:(a)VGG16 plus RF;(b)VGG19 plus RF;(c)Xception plus RF;(d)ResNet50 plus RF;(e)Inception-ResNet-V2 plus RF;(f)Densenet plus RF

    Table 7: Metrics showing performance with different feature extractors and RF classifiers

    Figure 20: Confusion matrix of different feature extractors with XGBoost classifier for real image dataset: (a) VGG16 plus XGBoost;(b) VGG19 plus XGBoost;(c) Xception plus XGBoost;(d)ResNet50 plus XGBoost;(e)Inception-ResNet-V2 plus XGBoost;(f)Densenet plus XGBoost

    Table 8:Metrics showing performance metrics with different feature extractors and XGBoost classifiers

    Figure 21:Accuracy comparison of different feature extractors for real image datasets

    Table 9: Comparison of accuracies of proposed approaches with state-of-the-arts models from published literature

    6 Discussions

    Leukemia is a disease that proves life-threatening in many cases at its later stages of infection.It has different sub-types,which differ in treatment guidelines.This variability in sub-types has motivated the work for multi-class leukemia cell classification.The designed framework classifies infected and normal cells in binary classification on a standard ALL-IDB dataset.Additionally,a real-image dataset consisting of leukemia sub-class images was tested with this framework for predicting the subtype of leukemia,including AML,CML,and CLL.

    The proposed approaches here are hybrid,consisting of a deep learning-machine learning(DLML)architecture.Popular deep learning frameworks are utilized for feature extraction with a transfer learning approach.Transfer learning frameworks VGG16,VGG19,Xception,Inception-ResNet-V2,Densenet,and ResNet are used for this purpose.If these frameworks are used for training,they may lead to over-fitting due to the limited number of dataset images.Additionally,deep learning training consumes more time as the number of epochs increases.This challenge is addressed by using these frameworks solely for feature extraction.After feature extraction,the final classification is performed by state-of-the-art machine learning models.RF and XGBoost are applied for the final classification,and various performance metrics,including precision,recall,F1-score,and accuracy,are considered for model evaluation.

    Precision is a crucial metric for measuring classification results.It represents the true positive predictions against all positive predictions from the model’s testing.Various DL frameworks are combined with RF and XGBoost classifiers.Precision values obtained for VGG16,VGG19,Xception,Inception-ResNet-V2,Densenet,and ResNet feature extractors hybridized with RF for the ALLIDB1 dataset are 0.96,0.92,1,0.95,0.87,and 0.87,respectively.For the ALL-IDB2 dataset,precision values are 0.88,0.88,0.91,0.88,0.74,and 0.88 for the above respective feature extractors.With the XGBoost classifier,the feature extractor frameworks VGG16,VGG19,Xception,Inception-ResNet-V2,Densenet,and ResNet achieved precision values of 0.95,1,1,0.86,0.95,and 0.88 for the ALLIDB1 dataset.For the ALL-IDB2 dataset,precision values are 0.89,0.91,0.91,0.88,0.89,and 0.85 for the respective feature extractors.A notable observation from the comparison of precision values is that the XGBoost classifier consistently delivered higher precision than the RF in most cases.

    Recall is another critical measure,indicating positive predictions of an experiment against true positives and false negatives.Ideally,recall should approach a value of 1,ensuring no false negatives during predictions.Recall values,calculated by considering various feature extractors used in this experimentation,for VGG16,VGG19,Xception,Inception-ResNet-V2,Densenet,and ResNet hybridized with RF for the ALL-IDB1 dataset are 0.95,0.92,1,0.96,0.82,and 0.86,respectively,with the RF classifier.For the ALL-IDB2 dataset,these values are 0.85,0.87,0.91,0.89,0.74,and 0.88 for the same feature extractors.With the XGBoost classifier,the same feature extractor frameworks achieved recall values of 0.96,1,1,0.87,0.96,and 0.88 for ALL-IDB1.ALL-IDB2 recall values are 0.89,0.90,0.91,0.88,0.89,and 0.84 for the respective frameworks.As previously observed in precision,the XGBoost classifier generally outperformed the RF in recall across most feature extractors.

    The F1-score metric is also important when evaluating the performance of machine learning and deep learning models.As there is a trade-off between precision and recall,if one increases,the other decreases.To measure the performance of any model,the F1-score can prove to be a valuable measure alongside precision and recall.It essentially provides the harmonic mean of precision and recall.This metric considers both precision and recall,addressing false positives and false negatives.For all employed feature extractors,the F1-score is calculated.For the ALL-IDB1 dataset and the RF classifier,the feature extractors VGG16,VGG19,Xception,Inception-ResNet-V2,Densenet,and ResNet achieved F1-scores of 0.95,0.91,1,0.95,0.82,and 0.86,respectively.ALL-IDB2 was used for the same steps of experimentation,yielding F1-scores of 0.86,0.87,0.91,0.88,0.74,and 0.88 for the respective feature extractors.Additionally,the XGBoost classifier replaced RF,using the same feature extractors.The F1-scores obtained for VGG16,VGG19,Xception,Inception-ResNet-V2,Densenet,and ResNet were 0.95,1,1,0.86,0.95,and 0.86 for ALL-IDB1.When ALL-IDB2 was tested with the same feature extractors and XGBoost classification,the observed F1-scores were 0.89,0.91,0.91,0.88,0.88,and 0.84 for VGG16,VGG19,Xception,Inception-ResNet-V2,Densenet,and ResNet,respectively.

    Another critical parameter for evaluating model performance is accuracy,which offers the number of correct predictions relative to the total predictions.For the ALL-IDB1 and ALL-IDB2 datasets,accuracy was calculated using various feature extractors and classifiers.The accuracy values for feature extractors VGG16,VGG19,Xception,Inception-ResNet-V2,Densenet,and ResNet with RF classifiers were 0.95,0.909,1,0.88,0.818,and 0.86 for the ALL-IDB1 dataset.For the ALLIDB2 dataset,the accuracy values were 0.86,0.87,0.907,0.9545,0.74,and 0.876 for these feature extractors.XGBoost was another classifier employed in this study.The accuracy values for the frameworks above were 0.95,1,1,0.876,0.95,and 0.86 for the ALL-IDB1 dataset.For the ALLIDB2 dataset,the accuracy values were 0.89,0.907,0.907,0.86,0.88,and 0.846 for VGG16,VGG19,Xception,Inception-ResNet-V2,Densenet,and ResNet,respectively.Accuracy generally improved with XGBoost over RF for most of the feature extractors.

    The proposed approach aims to be generalized and robust by testing on a real-image dataset with multiple classes.AML,CLL,and CML sub-types of leukemia are diagnosed in the second step of this work.Precision,recall,F1-score,and accuracy are the metrics for performance evaluation in this phase.The values for feature extractors VGG16,VGG19,Xception,Inception-ResNet-V2,Densenet,and ResNet were 0.88,0.96,0.95,0.86,0.96,and 0.88 with the RF classifier.For the XGBoost classifier,these values were 0.94,0.94,0.97,0.92,0.95,and 0.89 for the same feature extractors.For recall,the real-image dataset yielded values of 0.88,0.96,0.95,0.86,0.96,and 0.86 with RF.Using XGBoost,the recall values were 0.94,0.94,0.97,0.92,0.95,and 0.89 for these feature extractors.The F1-score was also calculated using both RF and XGBoost classifiers.The values were 0.88,0.96,0.95,0.86,0.96,and 0.86 with RF,and 0.94,0.9,0.97,0.92,0.95,and 0.89 with XGBoost for VGG16,VGG19,Xception,Inception-ResNet-V2,Densenet,and ResNet,respectively.Accuracy,a crucial evaluation metric,was calculated for all feature extractors,yielding values of 0.88,0.96,0.95,0.86,0.96,and 0.86 with RF,and 0.94,0.9,0.97,0.92,0.95,and 0.89 with XGBoost.The precision,recall,F1-score,and accuracy metrics showed improvement with XGBoost over RF for most classifiers.

    There are different state-of-the-art methods presented and experimented on by researchers for the binary classification of leukemia cells.Rawat et al.[89] employed classical machine learning algorithms,including Na?ve Bayes,Decision Tree,SVM,and KNN,with accuracies ranging from 50% to a maximum of 88% for the CNN approach.Singhal et al.[43] utilized the SVM classifier with LBP features,yielding an accuracy of 89.72%.Additionally,Singh et al.[67] and Ali et al.[66]utilized the SVM and SVM with shape and texture features,respectively,and achieved accuracies of 93.75%and 89.80%on the same ALL-IDB dataset.Separately,Shahin et al.[90]applied WBCsNet,Alexnet,and Lenet,achieving accuracies between 88%and 96%.In our experimentation,the presented frameworks achieved an accuracy of 100% on the ALL-IDB dataset for binary classification.The proposed approach was tested on a real-image dataset consisting of 520 images of leukemia subclasses,including AML,CML,and CLL.For this multi-class classification,the approaches achieved an accuracy of 97.08%.

    However,there are some concerns in this domain.The primary challenge with deep learning approaches is the dataset size.Achieving desirable accuracy and performance metrics requires a vast number of images.In medical imaging,there is a consistent constraint on the number of available images.There are fewer images in the datasets we employed,ALL-IDB1 and ALL-IDB2.This dataset is highly popular among researchers for leukemia diagnosis and is widely used.Another real-image dataset used in this study comprises approximately 520 images.Due to the limited number of images,relying solely on deep learning may not yield optimal results.Consequently,a hybrid approach is proposed.This method uses established deep learning frameworks for feature extraction,while classical machine learning algorithms handle the final classification.In detecting leukemia,which is binary classification,our proposed method has shown promising precision,recall,F1-score,and accuracy results.Another challenge in deep learning is the increased training time as the number of epochs grows.However,in our method,DL classifiers are only used for feature extraction,meaning their trainable parameters are considered zero across all frameworks.As a result,training and model fitting occur solely in the final RF/XGBoost classification stage.

    While various combinations of methods yield impressive accuracies and other metrics,Xception Net combined with RF and XGBoost achieved the highest accuracy at 100%.In other instances,an intriguing observation was that accuracy increased with XGBoost over RF classification.Observing the real-image dataset has proven suitable for our methods,delivering impressive accuracies across different strategies and peaking at 97.08%for multi-class classification with real dataset images.In this study,leukemia classification occurs in two phases.First,using the publicly available ALL-IDB1 and two datasets,binary classification distinguishes between normal and infected images.In the second phase,multi-class classification uses the real-image dataset to identify three leukemia sub-classes:AML,ALL,and CLL.With the real-image accuracy nearing 97%,this method can be deemed robust and generalized.As such,the proposed method might be valuable to researchers and professionals in this field for leukemia sub-class predictions and expedited,accurate diagnoses.

    7 Conclusions

    A hybrid DL-ML framework is implemented for the diagnosis of leukemia and classification.Classification is conducted in two ways: binary classification,with output predictions of cells as normal or infected,and multi-class classification using a real-image dataset.Three sub-classes of leukemia,AML,CML,and CLL,are considered.When comparing machine learning to deep learning classification,deep learning is consistently ahead in terms of accuracy.Regarding leukemia diagnosis,the primary constraint is dataset size.The widely used dataset,ALL-IDB,has a limited number of images.Hence,the deep learning framework may become overfitted and deviate from its remarkable accuracy.The same issue arises with the real-image dataset used,which contains approximately 520 images.Although this dataset is large,the deep learning framework requires more time to train the network,especially when many trainable parameters exist.Popular frameworks like Xception,ResNet,VGGnet,and Densenet possess many trainable parameters,leading to extended training times as the number of epochs increases.Therefore,a hybrid approach combining deep learning and machine learning is adopted for classification.Features are extracted using a deep learning classifier,including convolutional neural networks and other frameworks such as VGGNet,Xception,Inception-ResNet-V2,Densenet,and ResNet.Final classification is performed using state-of-the-art machine learning classifiers,with RF and XGBoost chosen for the task.The proposed approach demonstrates an improvement in leukemia diagnosis performance.Testing accuracies reached nearly 100%for the ALLIDB dataset and 96%for the real-image dataset in the current experimentation.

    Acknowledgement:Thanks to two anonymous reviewers for their valuable comments on the earlier version,which helped us improve the manuscript’s overall quality.

    Funding Statement:The study is supported by the Centre for Advanced Modelling and Geospatial Information Systems(CAMGIS),the University of Technology Sydney,the Ministry of Education of the Republic of Korea,and the National Research Foundation of Korea(NRF-2023R1A2C1007742)and in part by the Researchers Supporting Project Number RSP-2023/14,King Saud University.

    Author Contributions:The authors confirm their contribution to the paper as follows:study conception and design: Nilkanth Mukund Deshpande and Shilpa Gite;data collection: Nilkanth Mukund Deshpande;analysis and interpretation of results:Nilkanth Mukund Deshpande;validation:Shilpa Gite and Biswajeet Pradhan;visualization:Shilpa Gite and Biswajeet Pradhan;project administration:Shilpa Gite and Biswajeet Pradhan;draft manuscript preparation: Nilkanth Mukund Deshpande,Shilpa Gite,Biswajeet Pradhan Abdullah Alamri,Chang-Wook Lee;funding: Biswajeet Pradhan,Abdullah Alamri,Chang-Wook Lee.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The data that support the findings of this study are available from the authors(Deshpande N.M.and Gite S.G.)upon reasonable request.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    .国产精品久久| 大香蕉久久网| 99热这里只有是精品在线观看| 欧美日韩国产mv在线观看视频 | 成人二区视频| 成人二区视频| 特大巨黑吊av在线直播| 国产精品久久久久久精品古装| 51国产日韩欧美| 韩国高清视频一区二区三区| 欧美极品一区二区三区四区| av在线app专区| 美女中出高潮动态图| 国产午夜精品一二区理论片| 一个人看视频在线观看www免费| 性高湖久久久久久久久免费观看| 黑丝袜美女国产一区| 久久女婷五月综合色啪小说| 狂野欧美白嫩少妇大欣赏| 青青草视频在线视频观看| 久久精品久久精品一区二区三区| 亚洲va在线va天堂va国产| 久久人人爽av亚洲精品天堂 | 免费看光身美女| 欧美高清性xxxxhd video| 中文乱码字字幕精品一区二区三区| 深爱激情五月婷婷| 有码 亚洲区| 中国三级夫妇交换| 五月伊人婷婷丁香| 永久网站在线| 大话2 男鬼变身卡| 夫妻性生交免费视频一级片| 亚洲国产精品国产精品| 国产精品久久久久久精品电影小说 | 男的添女的下面高潮视频| 国产又色又爽无遮挡免| 国产精品蜜桃在线观看| 高清欧美精品videossex| 亚洲精品国产成人久久av| 99热国产这里只有精品6| 成人毛片a级毛片在线播放| 少妇的逼水好多| 日韩一区二区视频免费看| 不卡视频在线观看欧美| 国产男女超爽视频在线观看| 国产高清不卡午夜福利| 国产av精品麻豆| 日韩电影二区| 韩国高清视频一区二区三区| 成人无遮挡网站| 欧美日韩国产mv在线观看视频 | 国产在线视频一区二区| 亚洲精品中文字幕在线视频 | 又大又黄又爽视频免费| 欧美日韩在线观看h| 精华霜和精华液先用哪个| 成人毛片a级毛片在线播放| 青春草视频在线免费观看| 99热这里只有是精品在线观看| 18禁裸乳无遮挡免费网站照片| 99久国产av精品国产电影| 日韩亚洲欧美综合| 丝袜脚勾引网站| 欧美另类一区| 久久国产乱子免费精品| 一个人看的www免费观看视频| 亚洲经典国产精华液单| 网址你懂的国产日韩在线| 国产av码专区亚洲av| 亚洲电影在线观看av| 亚洲av欧美aⅴ国产| 久久99精品国语久久久| av天堂中文字幕网| 国产精品一区二区在线观看99| 日本-黄色视频高清免费观看| 日日摸夜夜添夜夜添av毛片| 亚洲丝袜综合中文字幕| 赤兔流量卡办理| 国产伦理片在线播放av一区| 夫妻性生交免费视频一级片| 男人舔奶头视频| 蜜臀久久99精品久久宅男| 亚洲在久久综合| 国产亚洲av片在线观看秒播厂| 晚上一个人看的免费电影| 97在线人人人人妻| 国产午夜精品一二区理论片| 九九在线视频观看精品| 蜜桃在线观看..| 亚洲av综合色区一区| 干丝袜人妻中文字幕| 午夜精品国产一区二区电影| 99热这里只有精品一区| 亚洲人与动物交配视频| 18禁裸乳无遮挡免费网站照片| 国产高清有码在线观看视频| 久久久久人妻精品一区果冻| 高清av免费在线| 久久久国产一区二区| 男女下面进入的视频免费午夜| 草草在线视频免费看| 亚洲精品乱久久久久久| 丰满人妻一区二区三区视频av| 亚洲av福利一区| 日本-黄色视频高清免费观看| 伦理电影大哥的女人| 亚洲精华国产精华液的使用体验| 又黄又爽又刺激的免费视频.| 午夜精品国产一区二区电影| 中文字幕免费在线视频6| 狂野欧美白嫩少妇大欣赏| 午夜激情福利司机影院| 国产精品久久久久久久久免| 亚洲图色成人| 亚洲国产成人一精品久久久| 韩国av在线不卡| 亚洲精品第二区| 最近的中文字幕免费完整| 性色av一级| 国产无遮挡羞羞视频在线观看| 麻豆成人午夜福利视频| 中文乱码字字幕精品一区二区三区| 久久人人爽人人片av| 欧美高清性xxxxhd video| 久久久久久久亚洲中文字幕| 日本黄色日本黄色录像| 国产伦理片在线播放av一区| 伊人久久国产一区二区| 中文字幕亚洲精品专区| 亚洲精品国产av蜜桃| 久久精品国产a三级三级三级| 美女中出高潮动态图| 国产精品一区二区三区四区免费观看| 成人免费观看视频高清| 大又大粗又爽又黄少妇毛片口| 99热这里只有精品一区| 国产91av在线免费观看| 国产乱人偷精品视频| 亚洲国产精品国产精品| 日韩一区二区视频免费看| 亚洲色图av天堂| 午夜福利高清视频| 全区人妻精品视频| 久久久亚洲精品成人影院| 99久久精品热视频| av又黄又爽大尺度在线免费看| 欧美bdsm另类| 成人国产av品久久久| a级毛片免费高清观看在线播放| av又黄又爽大尺度在线免费看| 欧美一区二区亚洲| 一级毛片我不卡| 少妇裸体淫交视频免费看高清| 久久ye,这里只有精品| 亚洲av电影在线观看一区二区三区| 欧美性感艳星| 久久综合国产亚洲精品| av一本久久久久| 欧美日韩视频高清一区二区三区二| 99久国产av精品国产电影| 嘟嘟电影网在线观看| 毛片女人毛片| 高清不卡的av网站| 视频中文字幕在线观看| 久久影院123| 欧美一区二区亚洲| 国产精品一区二区在线不卡| 男女边吃奶边做爰视频| 国产男女内射视频| 2018国产大陆天天弄谢| 国产男女内射视频| a级毛片免费高清观看在线播放| 欧美另类一区| 搡老乐熟女国产| 欧美国产精品一级二级三级 | 丰满人妻一区二区三区视频av| 久久影院123| 成人影院久久| 青青草视频在线视频观看| 亚洲色图综合在线观看| 国产人妻一区二区三区在| av黄色大香蕉| 国产午夜精品一二区理论片| 亚洲色图av天堂| 少妇人妻一区二区三区视频| 国产精品一区二区在线不卡| 下体分泌物呈黄色| 91精品国产九色| 国产欧美日韩一区二区三区在线 | 亚洲av中文av极速乱| 国产成人91sexporn| 成人漫画全彩无遮挡| 亚洲欧洲日产国产| 久久久久久久亚洲中文字幕| 国产淫片久久久久久久久| 热99国产精品久久久久久7| 内地一区二区视频在线| 97超碰精品成人国产| 日本免费在线观看一区| 亚洲精品亚洲一区二区| 国产黄色免费在线视频| 亚洲精品久久久久久婷婷小说| 国产一区二区三区av在线| 国产亚洲欧美精品永久| 久久综合国产亚洲精品| 水蜜桃什么品种好| 精品一区二区三卡| 欧美三级亚洲精品| 国产精品秋霞免费鲁丝片| 久久久久久久精品精品| 日日摸夜夜添夜夜爱| 在现免费观看毛片| 尤物成人国产欧美一区二区三区| 午夜免费男女啪啪视频观看| 国产在线视频一区二区| 我的老师免费观看完整版| 一级毛片 在线播放| 亚洲aⅴ乱码一区二区在线播放| 久久99热6这里只有精品| 赤兔流量卡办理| 全区人妻精品视频| 交换朋友夫妻互换小说| av网站免费在线观看视频| 成人黄色视频免费在线看| 国产精品爽爽va在线观看网站| 国产精品久久久久久精品电影小说 | 国产 精品1| 免费在线观看成人毛片| 婷婷色av中文字幕| 国产亚洲av片在线观看秒播厂| 国产av一区二区精品久久 | 亚洲精品乱码久久久久久按摩| 久久久久精品久久久久真实原创| 一本色道久久久久久精品综合| 免费av中文字幕在线| 欧美日韩视频高清一区二区三区二| 精品久久久精品久久久| 九九爱精品视频在线观看| 久久久久久久久久人人人人人人| 免费高清在线观看视频在线观看| 视频区图区小说| 久久久久视频综合| 青春草亚洲视频在线观看| 久久国产亚洲av麻豆专区| 在现免费观看毛片| 久久精品久久久久久久性| 国产深夜福利视频在线观看| 日韩视频在线欧美| 欧美老熟妇乱子伦牲交| 大香蕉久久网| 日韩欧美 国产精品| 欧美日韩精品成人综合77777| 亚州av有码| 欧美一区二区亚洲| 一区二区三区乱码不卡18| 干丝袜人妻中文字幕| 精品人妻一区二区三区麻豆| 久久久欧美国产精品| 午夜福利高清视频| 午夜激情久久久久久久| 热99国产精品久久久久久7| 国产日韩欧美亚洲二区| 久久这里有精品视频免费| 九草在线视频观看| 国产女主播在线喷水免费视频网站| 亚洲av男天堂| 一二三四中文在线观看免费高清| 热re99久久精品国产66热6| 日日摸夜夜添夜夜添av毛片| 1000部很黄的大片| 亚洲怡红院男人天堂| 一级毛片久久久久久久久女| 三级国产精品片| 久久av网站| 91aial.com中文字幕在线观看| 天美传媒精品一区二区| 黄色视频在线播放观看不卡| 欧美极品一区二区三区四区| 国产成人精品一,二区| 久久久欧美国产精品| 在线观看av片永久免费下载| 搡老乐熟女国产| 日韩一区二区三区影片| 97超碰精品成人国产| 久久国产乱子免费精品| 国产成人a∨麻豆精品| 麻豆乱淫一区二区| 亚洲精品中文字幕在线视频 | 久久青草综合色| 国产精品嫩草影院av在线观看| 久久久a久久爽久久v久久| 肉色欧美久久久久久久蜜桃| 麻豆国产97在线/欧美| 简卡轻食公司| 男人添女人高潮全过程视频| 91午夜精品亚洲一区二区三区| 成人影院久久| 热re99久久精品国产66热6| 成年女人在线观看亚洲视频| 国产精品.久久久| 又大又黄又爽视频免费| 大香蕉97超碰在线| 午夜免费观看性视频| 国产有黄有色有爽视频| 视频区图区小说| 王馨瑶露胸无遮挡在线观看| 久久国产精品男人的天堂亚洲 | 日韩大片免费观看网站| 日韩成人av中文字幕在线观看| 热99国产精品久久久久久7| 嘟嘟电影网在线观看| 精品酒店卫生间| 亚洲怡红院男人天堂| 亚洲av免费高清在线观看| 秋霞伦理黄片| 日韩av不卡免费在线播放| 免费播放大片免费观看视频在线观看| 舔av片在线| 日本欧美国产在线视频| 亚洲不卡免费看| 一个人看视频在线观看www免费| 亚洲精品日韩在线中文字幕| 大片电影免费在线观看免费| 亚洲va在线va天堂va国产| 亚洲av在线观看美女高潮| 中文资源天堂在线| 汤姆久久久久久久影院中文字幕| 亚洲,一卡二卡三卡| av黄色大香蕉| 黄色日韩在线| 亚洲欧美日韩无卡精品| 国国产精品蜜臀av免费| 国模一区二区三区四区视频| 美女视频免费永久观看网站| 97超碰精品成人国产| 国产永久视频网站| 一本色道久久久久久精品综合| 美女福利国产在线 | 老女人水多毛片| 国产精品一区二区在线观看99| 一区在线观看完整版| 日韩强制内射视频| 女性被躁到高潮视频| 观看免费一级毛片| 老司机影院毛片| 免费播放大片免费观看视频在线观看| 一级毛片久久久久久久久女| 亚洲欧美日韩另类电影网站 | 成人亚洲欧美一区二区av| 26uuu在线亚洲综合色| 校园人妻丝袜中文字幕| 国产精品国产三级国产专区5o| 国产伦精品一区二区三区四那| 男女免费视频国产| 国产一区亚洲一区在线观看| 午夜福利影视在线免费观看| 丝袜喷水一区| 老司机影院毛片| 美女主播在线视频| 99久久精品国产国产毛片| 国产精品国产av在线观看| 婷婷色综合大香蕉| 91精品伊人久久大香线蕉| 黄色欧美视频在线观看| 亚洲精品日韩在线中文字幕| 麻豆国产97在线/欧美| 成年人午夜在线观看视频| 最近最新中文字幕大全电影3| 国产av码专区亚洲av| 自拍欧美九色日韩亚洲蝌蚪91 | 欧美丝袜亚洲另类| 久久久久久久久久久免费av| 夫妻午夜视频| 中文字幕制服av| 国产精品一区二区在线观看99| 精品国产乱码久久久久久小说| 欧美日韩国产mv在线观看视频 | 一本色道久久久久久精品综合| 韩国av在线不卡| 伦精品一区二区三区| 日本wwww免费看| 久久韩国三级中文字幕| 啦啦啦啦在线视频资源| 亚洲最大成人中文| 国产精品成人在线| 99九九线精品视频在线观看视频| 欧美极品一区二区三区四区| 免费观看在线日韩| 青春草亚洲视频在线观看| 国产深夜福利视频在线观看| 高清黄色对白视频在线免费看 | 少妇被粗大猛烈的视频| av在线播放精品| 直男gayav资源| 一级毛片黄色毛片免费观看视频| 国产国拍精品亚洲av在线观看| 麻豆精品久久久久久蜜桃| .国产精品久久| 国产视频首页在线观看| 99热这里只有精品一区| 免费观看在线日韩| 亚洲自偷自拍三级| 免费观看的影片在线观看| 欧美变态另类bdsm刘玥| 五月伊人婷婷丁香| 人妻少妇偷人精品九色| 日日摸夜夜添夜夜添av毛片| 日韩不卡一区二区三区视频在线| 久久国内精品自在自线图片| 日韩成人伦理影院| 一区二区av电影网| 亚洲欧美中文字幕日韩二区| 不卡视频在线观看欧美| 国产男女内射视频| 蜜桃久久精品国产亚洲av| 全区人妻精品视频| 男女国产视频网站| 国产无遮挡羞羞视频在线观看| av福利片在线观看| 男女国产视频网站| 久久久a久久爽久久v久久| 欧美精品一区二区免费开放| 最后的刺客免费高清国语| 亚洲精华国产精华液的使用体验| 亚洲伊人久久精品综合| 韩国av在线不卡| 久久国产精品大桥未久av | av网站免费在线观看视频| 国产精品女同一区二区软件| 欧美日韩亚洲高清精品| 国产av精品麻豆| a级毛片免费高清观看在线播放| 香蕉精品网在线| 国产熟女欧美一区二区| av一本久久久久| 熟妇人妻不卡中文字幕| 欧美日韩一区二区视频在线观看视频在线| 一本一本综合久久| 熟女人妻精品中文字幕| 欧美+日韩+精品| 久久99热这里只有精品18| 国产精品嫩草影院av在线观看| 久久毛片免费看一区二区三区| 亚洲精品,欧美精品| 午夜福利视频精品| 日本欧美视频一区| 国产高清不卡午夜福利| 亚洲欧洲日产国产| 国产v大片淫在线免费观看| 黄色日韩在线| 国产精品免费大片| 麻豆乱淫一区二区| .国产精品久久| 国产精品免费大片| 91精品国产九色| 在线看a的网站| 日日摸夜夜添夜夜爱| 极品少妇高潮喷水抽搐| 国产精品国产三级专区第一集| 日日摸夜夜添夜夜爱| 亚洲欧美成人精品一区二区| 一本色道久久久久久精品综合| 久久久久久久大尺度免费视频| 久久99蜜桃精品久久| 在现免费观看毛片| 成人免费观看视频高清| 大片电影免费在线观看免费| 男人狂女人下面高潮的视频| 免费观看在线日韩| 亚洲,一卡二卡三卡| 人人妻人人看人人澡| 五月玫瑰六月丁香| 久久精品国产鲁丝片午夜精品| 少妇人妻一区二区三区视频| 久久久久精品性色| 高清黄色对白视频在线免费看 | 国产综合精华液| 少妇精品久久久久久久| 美女主播在线视频| 久久久久性生活片| 22中文网久久字幕| 亚洲精品日本国产第一区| 老司机影院毛片| 最近中文字幕高清免费大全6| 久久亚洲国产成人精品v| 国产亚洲欧美精品永久| 国产欧美日韩一区二区三区在线 | 毛片一级片免费看久久久久| 精品亚洲成国产av| 国产 一区精品| 久久久久国产精品人妻一区二区| 亚洲欧美一区二区三区黑人 | 十八禁网站网址无遮挡 | www.色视频.com| 久久精品久久久久久久性| 久久久久久九九精品二区国产| 毛片一级片免费看久久久久| 久久av网站| 久久久精品94久久精品| 精品视频人人做人人爽| 97超视频在线观看视频| 国产在线一区二区三区精| 精华霜和精华液先用哪个| 国产精品一二三区在线看| 中文字幕av成人在线电影| 插阴视频在线观看视频| 国产一区二区三区综合在线观看 | 免费观看a级毛片全部| 国产精品久久久久久精品电影小说 | 亚洲欧美精品自产自拍| 在线观看免费高清a一片| 国产在线视频一区二区| 欧美性感艳星| 欧美老熟妇乱子伦牲交| 久久久国产一区二区| 尤物成人国产欧美一区二区三区| 久久99蜜桃精品久久| 亚洲精品日本国产第一区| 一个人看的www免费观看视频| 国产精品精品国产色婷婷| 国产伦精品一区二区三区四那| 啦啦啦在线观看免费高清www| 亚洲国产毛片av蜜桃av| 精品久久久久久久久亚洲| 妹子高潮喷水视频| 国产女主播在线喷水免费视频网站| 国产免费视频播放在线视频| 国产精品不卡视频一区二区| 一边亲一边摸免费视频| 最近最新中文字幕大全电影3| 最黄视频免费看| 成年免费大片在线观看| 亚洲av欧美aⅴ国产| 我要看黄色一级片免费的| 精品酒店卫生间| 青青草视频在线视频观看| a级一级毛片免费在线观看| 大码成人一级视频| av在线app专区| 成人国产麻豆网| 91精品国产九色| av在线观看视频网站免费| 黄色视频在线播放观看不卡| 日韩av不卡免费在线播放| a级毛色黄片| 成人黄色视频免费在线看| 男人狂女人下面高潮的视频| 成人黄色视频免费在线看| 国产精品伦人一区二区| 91午夜精品亚洲一区二区三区| videossex国产| 中文资源天堂在线| 97在线视频观看| 国产成人精品一,二区| 菩萨蛮人人尽说江南好唐韦庄| 亚洲精品乱码久久久v下载方式| 最新中文字幕久久久久| 国产精品蜜桃在线观看| 日本色播在线视频| 热re99久久精品国产66热6| 91午夜精品亚洲一区二区三区| 色婷婷久久久亚洲欧美| 精品国产一区二区三区久久久樱花 | 少妇被粗大猛烈的视频| 久久精品久久久久久久性| freevideosex欧美| 国产精品一区www在线观看| 国产免费福利视频在线观看| 午夜福利影视在线免费观看| av福利片在线观看| 国产成人a∨麻豆精品| 久久久久国产网址| 观看美女的网站| 国产毛片在线视频| 国产精品精品国产色婷婷| 久久毛片免费看一区二区三区| 国产精品三级大全| h视频一区二区三区| 一个人看的www免费观看视频| 国产av国产精品国产| 十八禁网站网址无遮挡 | 久久久国产一区二区| 精品一区二区三卡| 久久人人爽人人爽人人片va| 国产成人精品一,二区| 精品午夜福利在线看| 国产色爽女视频免费观看| av免费在线看不卡| 日日啪夜夜爽| 一本色道久久久久久精品综合| 男女无遮挡免费网站观看| 日韩强制内射视频| 亚洲最大成人中文| 欧美xxxx黑人xx丫x性爽| 亚洲欧美一区二区三区黑人 | 欧美日韩一区二区视频在线观看视频在线| 好男人视频免费观看在线| 国产精品一区二区在线观看99| 插阴视频在线观看视频| 三级国产精品欧美在线观看| 日韩 亚洲 欧美在线| av在线观看视频网站免费| 中文字幕免费在线视频6| 爱豆传媒免费全集在线观看| 国产伦精品一区二区三区四那| 激情 狠狠 欧美| 久久久久国产网址| 能在线免费看毛片的网站| 寂寞人妻少妇视频99o| 日韩 亚洲 欧美在线| 日韩av在线免费看完整版不卡| 久久久久性生活片| 免费人妻精品一区二区三区视频| 男人狂女人下面高潮的视频|