• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Automatic Detection of COVID-19 Infection Using Chest X-Ray Images Through Transfer Learning

    2021-04-14 06:54:04EleneFirmezaOhataGabrielMaiaBezerraJoVictorSouzadasChagasAlosioVieiraLiraNetoAdrianoBessaAlbuquerqueVictorHugodeAlbuquerqueSeniorMemberIEEEandPedroPedrosaRebouasFilhoMemberIEEE
    IEEE/CAA Journal of Automatica Sinica 2021年1期

    Elene Firmeza Ohata, Gabriel Maia Bezerra, Jo?o Victor Souza das Chagas, Aloísio Vieira Lira Neto, Adriano Bessa Albuquerque, Victor Hugo C. de Albuquerque, Senior Member, IEEE, and Pedro Pedrosa Rebou?as Filho, Member, IEEE

    Abstract—The new coronavirus (COVID-19), declared by the World Health Organization as a pandemic, has infected more than 1 million people and killed more than 50 thousand. An infection caused by COVID-19 can develop into pneumonia,which can be detected by a chest X-ray exam and should be treated appropriately. In this work, we propose an automatic detection method for COVID-19 infection based on chest X-ray images. The datasets constructed for this study are composed of 194 X-ray images of patients diagnosed with coronavirus and 194 X-ray images of healthy patients. Since few images of patients with COVID-19 are publicly available, we apply the concept of transfer learning for this task. We use different architectures of convolutional neural networks (CNNs) trained on ImageNet, and adapt them to behave as feature extractors for the X-ray images.Then, the CNNs are combined with consolidated machine learning methods, such as k-Nearest Neighbor, Bayes, Random Forest, multilayer perceptron (MLP), and support vector machine (SVM). The results show that, for one of the datasets, the extractor-classifier pair with the best performance is the MobileNet architecture with the SVM classifier using a linear kernel, which achieves an accuracy and an F1-score of 98.5%. For the other dataset, the best pair is DenseNet201 with MLP,achieving an accuracy and an F1-score of 95.6%. Thus, the proposed approach demonstrates efficiency in detecting COVID-19 in X-ray images.

    I. INTRODUCTION

    THE COVID-19 pandemic has become a severe health problem being at the center of media cover since December, 2019 [1], [2]. In about 74% of the cases, the COVID-19 causes mild (18%) or moderate (56%) symptoms[3]. However, the remainder of the cases range from critical(20%) to severe (6%) [3]. As of today (2020-04-03), the total number of registered cases is approximately 1 015 667, with 53 200 deaths worldwide and 212 991 cases where a full recovery was achieved. Moreover, the number of active cases is 749 476 [4], [5].

    The main symptoms presented by suspected infections include respiratory distress, fever, and cough. The virus may also cause pneumonia in more aggressive infections. Besides pneumonia, the infection can lead to severe acute respiratory syndrome, septic shock, multi-organ failure, and, ultimately,death [6]. Studies showed that men (about 60%) were more affected than women (about 40%), and that there were, up to this point, no significant death rates in children younger than nine years old [2]. Despite being developed, many first world countries have been facing a collapse of the healthcare system due to escalating demand for intensive care units simultaneously [7], [8].

    Virus tests take less and less time as new technologies are developed world-wide. The diagnostic of COVID-19 infections involves a chest scan to verify the lung condition, in such a way that, if the patient shows pneumonia in the scans,they are deemed to have a COVID-19 infection. This method allows authorities to isolate and treat affected patients in a timely and affirmative fashion [9].

    One of the available methods to detect pneumonia is a computed tomography scan of the chest (CT scan). Automated image analysis based on artificial intelligence is being developed to detect, quantify, and monitor COVID-19 infections, as well as to separate healthy lungs from diseased ones [10]. Ke et al. [11] use the image’s basic characteristics and analyze neural network co-working with heuristic algorithms. The method is divided into the following steps:first, an initial analysis of the possibility of detecting respiratory disease through basic descriptors with a neural network, then, the use of heuristic algorithms for the rapid detection of affected lung tissues, since the possibility of detection is considerable. Poap et al. [12] and Shan et al. [13]developed segmentation studies based on a heuristic and a deep learning method, respectively. These studies seek to segment all-region on a lung that presented infection, isolating the sick region from the rest and thus conduct studies of patterns unique to that region, helping to identify the region in a new sample. Xu et al. [14] aspired to develop an early screening model that was capable of differentiating COVID-19 pneumonia, Influenza-A pneumonia, and healthy lungs using CT scan images and deep learning techniques. A study by Wang et al. [15] developed a deep learning method based on the changes presented in COVID-19 patients’ CT scans that can acquire graphical features and provide clinical diagnosis much faster than waiting for the pathogen test.Following the same basis, Chouhan et al. [16] use different neural network models pre-trained on ImageNet to extract exam characteristics. These characteristics are used to obtain individual classification results for each network. The combination of each network’s results uses the majority vote,so the diagnosis corresponds to the class that achieved the highest number of votes.

    X-ray is an imaging technique that is used to investigate fractures, bone displacement, pneumonia, and tumor. X-rays have been used for many decades and provide an astonishingly fast way of seeing the lungs and, therefore, can be a helpful tool in the detection of COVID-19 infections [8],[17]. They are capable of generating images that show lung damage, such as from pneumonia caused by the SARS-CoV-2 virus [18]. Since X-rays are very fast and cheap, they can help to triage patients in places where the healthcare system has collapsed or in places that are far from major centers with access to more complex technologies. Furthermore, there are portable X-ray devices that can be easily transported to where it is needed [18]. CT scans make use of the principles of X-ray in an advanced manner to examine the soft structures of the body. It is also used to obtain clearer images of organs and soft tissues [19]. On the other hand, X-rays use less radiation[20], thus using an X-ray is faster, less harmful, and presents lower cost than a CT scan. Narin et al. [8] proposed an automatic detection of COVID-19 using chest X-rays and CNNs. Apostolopoulos et al. [17] also proposed the automatic detection of the disease but analyzing three classes: COVID-19, common pneumonia, and normal conditions.

    In this paper, we propose an automatic system to classify chest X-ray images as from COVID-19 patients or healthy patients using transfer learning with convolution neural networks (CNNs). We performed 144 experiments, which are a combination of 12 CNNs and six classifiers in two datasets.The results show that MobileNet combined with support vector machines (SVM) (Linear) achieved the highest accuracy (98.462%) in one dataset, and the combination DenseNet201 with MLP achieved 95.64% in the second dataset, showing the effectiveness of the proposed approach.

    This paper is organized as follows: Section II presents the transfer learning method. Section III describes the proposed methodology of the approach, detailing the dataset, the steps for the feature extraction and classification, and the metrics used to evaluate the approach. Section IV discusses the results. Lastly, Section V presents the conclusion and future works.

    II. TRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS

    Transfer learning is a method that utilizes the knowledge acquired by a CNN from a specific problem to solve a distinct but similar task. This transferred knowledge is used in a new dataset, whose size is usually smaller than the adequate size to train a CNN from scratch [21].

    In deep learning, this method requires an initial training of a CNN for a given task, using large datasets. The availability of a sizable dataset is the main factor to ensure the success of the method since the CNN can learn to extract the most significant features of a sample. The CNN is deemed suitable for transfer learning if it is found to be able to extract the most important image features [22].

    Then, in the transfer learning, the CNN is used to analyze a new dataset of a different nature and extract its features according to the knowledge acquired in the first training. One common strategy to exploit the capabilities of the pre-trained CNN is called feature extraction via transfer learning [23].This approach means that the CNN will retain its architecture and weights between its layers; therefore, the CNN is used only as a feature extractor. The features are later used in a second network/classifier that will process its classification.

    The transfer learning approach is mostly used to work around computational costs of training a network from scratch or to keep the feature extractor trained during the first task. In medical applications, the most accepted practice of transfer learning is to utilize the CNNs that achieved the best results in the ImageNet large scale visual recognition challenge(ILSVRC) [24], which assesses algorithms for object detection and classification in large scales. The use of large datasets for initial training of the network enables high performance in smaller datasets. This performance is linked to various extraction parameters that are typically not allowed as they cause overfitting of the network [25]. That said, feature extraction performed with transfer learning allows a large number of features to be extracted by generalizing the problem and avoiding excessive adjustments [26].

    The use of transfer learning also allows the use of the internet of things (IoT) systems to classify medical images.For example, Dourado Jr et al. [27] proposed an IoT system to detect a stroke in CT images. Rodrigues et al. [28] used the system proposed by Dourado Jr et al. [27] to classify EEG signals.

    The transfer learning method is used in the feature extraction step for the COVID-19 detection. The process is detailed in Section III-B.

    III. METHODOLOGY

    In this section, we present the proposed methodology for classifying an X-ray as being of a healthy patient or a patient affected by COVID-19. First, we describe the datasets of images used in this study. Then, we explain the process of feature extraction, which is based on the transfer learning theory. After that, we present the classification techniques applied and the steps of their training process. Lastly, we define the metrics we use to evaluate the results and to compare it to other approaches. Fig. 1 presents the infographics of the proposed approach; each step is explained in the next subsections.

    A. Datasets

    Fig. 1. Infographics of the proposed approach.

    In our study, we use frontal-view chest X-ray images. Only posterior-anterior (PA) and anterior-posterior (AP) X-ray views were collected. We divided the samples into two classes: X-ray images of patients diagnosed with COVID-19 and X-ray images of healthy patients. For better evaluation of the proposed method, we built two datasets: Dataset A and Dataset B. Both datasets have the same images for the COVID-19 class, but they have different images for the healthy class. In both datasets, the classes are balanced,consisting of 194 images for each class or 388 images for each dataset.

    In Dataset A, the COVID-19 class is composed of 194 chest X-ray images of patients diagnosed with COVID-19, which were collected from different sources [29], [30]. Both these sources were accessed on 2020-03-31. They consist of compilations of X-ray images taken from different papers,databases, and other sources. For this dataset, we collected the set of chest X-ray images of healthy patients from the “Chest X-ray Images (Pneumonia)” challenge available on Kaggle[31]. We randomly selected 194 samples from the X-ray images labeled as “normal” , which correspond to healthy patients. This source was chosen since it has been commonly used in related works that propose methods of detecting COVID-19 in X-rays [17], [8]. However, all the X-ray images from this source are of pediatric patients. Since the X-ray images of the COVID-19 class are mostly of adult patients,we built another dataset with patients of a similar age range.

    In Dataset B, as previously mentioned, the COVID-19 class images are the same as in Dataset A. However, due to the age difference between healthy patients and patients with COVID-19 present in Dataset A, we collected chest X-ray images from a different source for Dataset B. For this dataset, we took images from the “NIH Chest X-rays” challenge organized by the National Institutes of Health and available on Kaggle [32].We randomly selected 194 images from the class of “no findings”, which correspond to healthy patients.

    All images from the datasets are either in the joint photographic experts group (JPG/JPEG) or in the portable network graphics (PNG) format. The image resolution is varied within the dataset, with resolutions as low as 249 by 255 pixels and as high as 3520 by 4280 pixels. However, all images were pre-processed using the resizing technique. In Table I, we can see the sizes to which the images were resized for each specific CNN. The equipment used to take the X-rays is also diverse and is often not determined; more information about it can be found on the image sources [29]-[32]. We present examples of images from the datasets in Fig. 2. An example of a chest X-ray of a patient with coronavirus disease with PA view is shown in Fig. 2(a) and AP view is shown in Fig. 2(b), an example of a chest X-ray of a healthy patient from Dataset A is shown in Fig. 2(c), and an example of a chest X-ray of a healthy patient from Dataset B is shown in Fig. 2(d).

    Fig. 2. Samples from the dataset used in this study. (a) X-ray with PA view of a patient with COVID-19; (b) X-ray with AP view of a patient with COVID-19; (c) X-ray of a healthy patient from Dataset A; (d) X-ray of a healthy patient from Dataset B.

    In order to have more input data to create a more generalized model, data augmentation was implemented. The augmentation of training set is a widespread technique in the literature [41]. The images were randomly selected to undergo a transformation. The most commonly used transformations for data augmentation are the affine transformations [42]. The affine transformations applied in this study were rotation,change in width, height, and magnification.

    B. Feature Extraction Steps

    For extracting features from the X-ray images, we use the transfer learning concept discussed in Section II. Firstly, we select different CNN architectures that achieved excellent performance on the ImageNet dataset. Secondly, we choose different configurations, previously trained on ImageNet, from the selected CNN architectures. Thirdly, we remove any fully connected layers from these configurations, leaving only convolutional and pooling layers. These two types of layers are responsible for extracting features from the image, while the fully connected ones are responsible for classifying the features and, consequently, the image. Thus, removing these layers is necessary to turn a CNN into a feature extractor.After this step, the new output of the adapted CNN is a set of features extracted from an input image.

    For each CNN configuration, we create a sub-dataset composed of sets of features extracted from each image of the original datasets. In order to build a sub-dataset, we first resize each image according to the input size required by the selected CNN. Then, each resized image is used as input to the CNN,and its set of features is extracted and stored in the corresponding sub-dataset. In Table I, we show all the CNN architectures and their respective configurations used. It is worth noting that InceptionResNetV2 [36] is a hybrid configuration, originated from Inception [34] and ResNet[35]. In Table I, we also present the input image size required by each configuration and the number of features extracted from a single image.

    C. Classification Steps

    In order to classify the X-ray images, we selected widely used machine learning methods in the literature: Bayes[43]-[45], random forest (RF) [46]-[48], multilayer perceptron (MLP) [49]-[51], k-nearest neighbors (kNN)[52]-[54], and SVM [55]-[57]. In the SVM classifier, we consider the linear and RBF kernels. It is noteworthy that these classifiers are from different types: kNN is instancebased, RF is based on the decision tree method, MLP is based on neural networks, SVM is based on finding an optimal hyperplane, and Bayes is based on probability and statistics.

    The classification is performed in three steps: i) model training, ii) model testing, and iii) repetition of processes i)and ii).

    Each sub-dataset is composed of features extracted from the extractors presented in Section III-B. These sub-datasets are divided into 80% for training and the remaining for testing.Furthermore, we applied the data augmentation in the training set of Dataset A. The number of images for train and test for each dataset is presented in Table II.

    TABLE II DATA SPLIT ACCORDING TO DATASET AND CLASS

    1) Model Training: In this step, we use 80% of the subdataset to perform the training of the model. We consider the setup for the hyperparameters presented in Table III to find the configuration of the classifiers on the training set. The classifiers that were configured for a random search perform a 20-iterations search. The hyperparameters for all classifiers,except for the Bayes classifier, are determined after 10-fold cross-validation. Then, each classifier has optimal hyperparameters, which are saved on the computer.

    2) Model Testing: In this step, we perform a test in the remaining 20% of the sub-dataset using the saved classifiers.The system determines one class for each sample of the subdataset. In addition, the metrics are calculated in this step.

    3) Repetition of Processes 1) and 2): The sub-datasets are randomly divided into other train and test sets. These sets are ensured to be different from the rest by the seed used. Then,we perform ten repetitions of steps 1) and 2).

    D. Evaluation Metrics

    We analyze the results of this paper utilizing the metrics:accuracy (Acc), F1-score , and false positive rate (FPR).Accuracy describes how often the model is classifying correctly. F1-score can be described as the harmonic mean of Sensitivity and Precision; this metric can provide a number that suggests an overall quality of the approach. FPR indicates a rate of healthy patients being wrongly classified. True positives (TP) indicates the number of instances that the model classified the images as COVID-19 correctly. False negative (FN) corresponds to the number of occasions that the COVID-19 images were misclassified as from a healthy patient. False positives (FP) points out the number of times that the model wrongly classified a healthy patient. True negatives (TN) informs the number of healthy patient images that were correctly classified. The equations for Acc,Sensitivity, Precision, FPR, and F1-score are presented on(1)-(5), respectively.

    TABLE III SETUP TO SEARCH FOR HYPERPARAMETERS OF THE CLASSIFIERS

    TABLE IV METRICS OBTAINED BY CLASSIFYING FEATURES EXTRACTED FROM DATASET A BY DIFFERENT CNN ARCHITECTURES OF THE TOP FIVE RESULTS

    In addition to the metrics already mentioned, we also analyze the training, extraction, and test times. The training time represents the length of the period it takes from the beginning of the classifier training to the moment it is ready to perform the classification. Extraction time measures how long the adapted CNN takes to output the attribute vector from the moment it receives the X-ray. Also, the test time is the duration it takes for the classifier to predict the image’s class after receiving its attribute vector. Thus, training time is vital during model building. After this step, the extraction and test times are more relevant. Their sum represents the classification time, which is the period between receiving the X-ray and returning its class.

    IV. RESULTS

    In this section, we investigate the results achieved by combining the features extracted by CNNs, applying transfer learning, and the classifiers. We executed 72 × 2 = 144 experiments, which are the combination of twelve CNNs and six classifiers in both datasets. The system infrastructure used was an Intel i7, 8 Gb of RAM, with Linux Ubuntu 16.04 system without a graphical processing unit (GPU).

    Table IV shows the metrics and their standard deviations of the top 5 results after the 10-iterations of the steps described in Section III-C in Dataset A; the full table can be found in the following link: https://bit.ly/2w9DEpT. We applied Friedman test on the accuracy results of Dataset A to test if there is no statistical difference among the results of classifiers. There is a significant effect at p <0.05 level found for the classifiers(F=30.73, p=0.001). This result is confirmed by the Nemenyi post-hoc test (C D=2.18), which is presented in Fig. 3. We can observe that there is no consistent indication of statistical differences between SVM (RBF), SVM (Linear),MLP, and RF for the 144 experiments evaluated.

    Analyzing Table IV, we can observe that all combinations in the top five achieved, reaching a minimum Acc of 98.205%and minimum F1-score of 98.205%. However, the combination that should be highlighted is MobileNet with SVM with the linear kernel, since it reached a maximum Acc of 98.462%, a maximum F1-score of 98.461%, and an FPR of 1.026%.

    Fig. 3. Result of Nemenyi test for the classifiers on Dataset A.

    In the full table of Dataset A provided in the link, we can observe that the CNNs architectures that achieved a minimum of 95% in Acc and F1-score independently of the classifier were: VGG16, DenseNet201, DenseNet169, and MobileNet,showing the effectiveness of the proposed approach. We can also observe that RF achieved the slowest test times independently of the CNN architecture, which is related to its high number of estimators.

    In Table IV, we can observe that MobileNet with SVM(Linear) reached a test time of 0.443 ms and an extraction time of 21 ms; then, this time is attractive for real-time implementations, since it would take, approximately 21.443 ms to define if an image is from class COVID-19 or Healthy.Furthermore, although the system infrastructure does not have GPU and its configuration is not high-end, the proposed approach achieved satisfactory extraction and training times,then clinics and hospitals do not need to acquire new equipment for a system to aid in the medical diagnosis.

    Table V presents the final confusion matrix of the MobileNet with SVM (Linear). It is a sum of the 10 confusion matrices from the 10-iterations of Section III-C. Therefore, the confusion matrix shows that even after 10-iterations, the combination MobileNet with SVM (Linear) did not have many FPs and FNs. In addition, we can observe on Table IV that this combination reached an FPR of 1.026 %. In a reallife application, this means that not many patients will be misclassified as not infected, then reducing the spread of the disease and allowing them to have the proper treatment. Also,patients without COVID-19 will rarely be submitted to more exams or admissions in dedicated locations to COVID-19,then decreasing the probability of their contamination.

    Table VI shows the top five results of Dataset A with data augmentation; the full table can be found in the following link: https://bit.ly/2w9DEpT. We can observe that with data augmentation, the highest Acc only increased 0.5%, with the pair DenseNet121 with MLP, which achieved 98.974%.Nonetheless, the proposed approach with data augmentation reached more combinations with 98% of Acc, showing that the use of the CNN to extract features is effective. Another point that can be observed is that the FPR rates are slightly lower in average than those achieved with the dataset without data augmentation, showing that the model was able to better generalize the problem.

    TABLE V FINAL CONFUSION MATRIX OF THE TEST SET FOR THE CLASSIFICATION OF CHEST X-RAY IMAGES AS HEALTHY OR COVID-19 FOR MOBILENET WITH SVM (LINEAR) FOR DATASET A

    Since most of the papers that are going to be compared in Section IV-A do not use the same source of healthy X-ray images as Dataset B, Table VII is a summary of the top five results, ordered by Acc; the tiebreaker was the F1-score. The complete results can be found in the following link:https://bit.ly/2w9DEpT.

    Analyzing Table VII , we can observe that even though images in Dataset B have similar contrasts and many artifacts in both classes, the transfer learning method combined with consolidated machine learning methods could achieve an Acc of 95.641% and an F1-score of 95.633% through the DenseNet201 architecture with the MLP classifier.

    Table VIII presents the confusion matrix for the combination DenseNet201 with MLP for Dataset B; it shows that the errors are balanced. Table IX displays the confusion matrix for the combination DenseNet201 with SVM (Linear)for Dataset B; even though the Acc of this combination is close to that of the combination DenseNet201 with MLP, the combination DenseNet201 with SVM (Linear) is not desirable because more COVID-19 patients are being classified as healthy, which could contribute to the spread of the disease.This difference can also be noted on FPR, where the FPR for DenseNet201 with SVM (Linear) is lower than the FPR forDenseNet201 with MLP. Some distinct characteristics of the exam define the method’s misclassification. For example,exams with high contrast variations and with angulations that do not centralize the critical region, are exams that deviate from the basic standards of Chest X-ray images promoting unsatisfactory results in the classification.

    TABLE VI METRICS OBTAINED BY CLASSIFYING FEATURES EXTRACTED FROM DATASET A WITH DATA AUGMENTATION BY DIFFERENT CNN ARCHITECTURES OF THE TOP FIVE RESULTS

    TABLE VII METRICS OBTAINED BY CLASSIFYING FEATURES EXTRACTED FROM DATASET B BY DIFFERENT CNN ARCHITECTURES OF THE TOP FIVE RESULTS

    TABLE VIII FINAL CONFUSION MATRIX OF THE TEST SET FOR THE CLASSIFICATION OF CHEST X-RAY IMAGES AS HEALTHY OR COVID-19 POSITIVE FOR DENSENET201 WITH MLP FOR DATASET B

    TABLE IX FINAL CONFUSION MATRIX OF THE TEST SET FOR THE CLASSIFICATION OF CHEST X-RAY IMAGES AS HEALTHY OR COVID-19 POSITIVE FOR DENSENET201 WITH SVM (LINEAR)FOR DATASET B

    Fig. 4 presents the features extracted by DenseNet201 from both datasets using the t-distributed stochastic neighbor embedding (t-SNE) technique [58]. We can observe that,although DenseNet201 is not in the top five of Dataset A, a hyperplane can separate the classes, which justifies the three combinations with SVM (Linear) in the top five of this dataset. As mentioned in Section III-A, the healthy X-ray images from Dataset A are of pediatric patients, which explains the easily distinguishable cluster formation of the features from this class. In contrast, when analyzing Dataset B, the healthy class features are scattered, which indicates a wider variety of sources in the original dataset. Also, this dataset needs a classifier that can classify non-linear data,which justifies MLP as the best classifier for this dataset.

    A. Comparison to Related Works

    As shown in Table X, we compare the proposed approach with the studies of Apostolopoulos et al. [17], Narin et al. [8],and Ozturk et al. [59]. Apostolopoulos et al. [17] and Narin et al. [8] used chest X-rays images and the transfer learning method, but they used a second network for the classification process. Furthermore, those papers used the same sources for the creation of the datasets as the proposed approach.

    Fig. 4. Visualization of the features extracted by DenseNet201 from both datasets using t-SNE.

    Apostolopoulos et al. [17] used images from the datasets“Chest X-ray images (Pneumonia)” [31] and “COVID-19 image data collection” [29]. From these datasets, they selected 50 images of each class, reaching an Acc and F1-score of 98%. However, this dataset is highly selected to be images for educational purposes [60]. In addition, they used a high-end GPU (Tesla K80), which could be impracticable for many clinics, hospitals, and countries. We tested our approach in this dataset, and it achieved 100% in many combinations of CNNs and classifiers.

    Narin et al. [8] used a dataset with images from “COVID-19 image data collection” [29], “Chest X-ray images(Pneumonia)” [31], and “COVID-19 X rays” [61], totaling 224 images for the COVID-19 condition, 504 for the normal condition and 700 for common pneumonia. They achieved an Acc of 98.75% when analyzing 2 classes and 93.48% for 3 classes. However, since their dataset is heavily unbalanced,this tends to increase the accuracy.

    Ozturk et al. [59] used 127 images from the dataset“COVID-19 image data collection” [29] that were diagnosed with COVID-19. In addition, they used 500 images for normal condition and 500 for pneumonia from dataset “ChestX-ray8”[32]; thus they used the same sources as the Dataset B of the proposed approach. They reached 98.08% in Acc and 95.51%in F1-score, when making a binary classification using a modified CNN based on you only look once (YOLO); thismodified CNN is called DarkCovidNet. Although they achieved an Acc 3% higher, their classes are also unbalanced,which can contribute for better metrics.

    TABLE X COMPARISON TO OTHER METHODS OF CLASSIFICATION OF CHEST X-RAYS IMAGES TO DETECT COVID-19

    V. CONCLUSION AND FUTURE WORKS

    Early detection of patients with the new coronavirus is crucial for choosing the right treatment and for preventing the quick spread of the disease. Our results show that the use of CNNs to extract features, applying the transfer learning concept, and then classifying these features with consolidated machine learning methods is an effective way to classify Xray images as in normal conditions or positive for COVID-19.For Dataset A, the MobileNet with SVM (Linear) combination had the best performance, achieving a mean Acc of 98.462% and a mean F1-score of 98.461%. In addition, it was able to classify a new image in only 0.443 ± 0.011 ms,proving to not only be accurate but fast as well. For Dataset B,the pair with the best performance was DenseNet201 with MLP, reaching a mean Acc of 95.641% and a mean F1-score of 95.633%. Although it had slightly lower Acc and F1-score,it classified an image in only 0.282 ± 0.154 ms, which is faster than the best combination in Dataset A.

    The proposed method has not undergone a clinical study.Thus, it does not replace a medical diagnosis since a more thorough investigation could be done with a larger dataset.Under those circumstances, our work contributes to the possibility of an accurate, automatic, fast, and inexpensive method for assisting in the diagnosis of COVID-19 through chest X-ray images.

    For future work, we intend to increase the size of the dataset by adding new X-ray images of patients with COVID-19, as soon as these images are available, and by adding X-ray exams of other lung-related diseases, thus reassuring the efficiency of the proposed approach. Besides, we aim to test the proposed method using an imbalanced dataset. We also intend to integrate our method into a free online platform of image classification, such as LINDA [27]. This way, hospitals and medical clinics around the world would be able to identify diseases in chest X-ray images without the need for building their classification platform. Furthermore, we aim to compare the proposed approach with methods based on fine-tuning,and train a network from scratch.

    波多野结衣巨乳人妻| 日韩国内少妇激情av| 亚洲国产色片| 最新中文字幕久久久久| 内射极品少妇av片p| 亚洲av不卡在线观看| 精品免费久久久久久久清纯| 久久精品夜夜夜夜夜久久蜜豆| av福利片在线观看| 高清在线视频一区二区三区 | 日韩三级伦理在线观看| 国产爱豆传媒在线观看| 欧美激情久久久久久爽电影| 99热这里只有是精品在线观看| 91在线精品国自产拍蜜月| 国产精品久久久久久精品电影| 欧美三级亚洲精品| 国内精品一区二区在线观看| 国产亚洲av片在线观看秒播厂 | .国产精品久久| 成年av动漫网址| 99久久精品国产国产毛片| 久久久久久久午夜电影| 国产三级在线视频| 看十八女毛片水多多多| 麻豆一二三区av精品| 九九爱精品视频在线观看| 99久久精品国产国产毛片| 禁无遮挡网站| 欧美性猛交╳xxx乱大交人| 色综合亚洲欧美另类图片| 久久久亚洲精品成人影院| 亚洲av二区三区四区| 国产伦精品一区二区三区四那| 欧美变态另类bdsm刘玥| 国产探花极品一区二区| 日韩成人伦理影院| 精品人妻熟女av久视频| 国产在视频线在精品| 亚洲丝袜综合中文字幕| 黄色日韩在线| 欧美一区二区精品小视频在线| 亚洲av中文av极速乱| 久久精品国产自在天天线| 国产精品一及| 舔av片在线| 99久久无色码亚洲精品果冻| 国产精品一二三区在线看| 精品人妻一区二区三区麻豆| av线在线观看网站| 精品少妇黑人巨大在线播放 | 日韩三级伦理在线观看| 丝袜美腿在线中文| 久久久久网色| 99久久精品国产国产毛片| 成人高潮视频无遮挡免费网站| 黄色欧美视频在线观看| 亚洲美女搞黄在线观看| 免费看日本二区| 国产精华一区二区三区| 女人十人毛片免费观看3o分钟| a级毛片免费高清观看在线播放| 中文精品一卡2卡3卡4更新| 成人三级黄色视频| ponron亚洲| 在线天堂最新版资源| 99久久九九国产精品国产免费| 久久久久久久亚洲中文字幕| 亚洲国产精品久久男人天堂| 嘟嘟电影网在线观看| 麻豆乱淫一区二区| 国产在视频线精品| 91在线精品国自产拍蜜月| 偷拍熟女少妇极品色| a级毛片免费高清观看在线播放| 国产欧美另类精品又又久久亚洲欧美| 亚洲欧洲国产日韩| 村上凉子中文字幕在线| 1000部很黄的大片| 精品国产露脸久久av麻豆 | 日本wwww免费看| 1000部很黄的大片| 人体艺术视频欧美日本| 少妇熟女aⅴ在线视频| 亚洲精华国产精华液的使用体验| 国产精品野战在线观看| 成人一区二区视频在线观看| 成人欧美大片| 天堂网av新在线| 国产久久久一区二区三区| 国产乱人偷精品视频| 精品99又大又爽又粗少妇毛片| 一级av片app| 日本黄色片子视频| 少妇熟女欧美另类| 久久久久性生活片| 久久久久性生活片| 成人av在线播放网站| 六月丁香七月| 一边摸一边抽搐一进一小说| 日本av手机在线免费观看| 久久久久久国产a免费观看| АⅤ资源中文在线天堂| 亚洲国产欧洲综合997久久,| 热99在线观看视频| 国产av在哪里看| 欧美又色又爽又黄视频| 日日啪夜夜撸| 自拍偷自拍亚洲精品老妇| 两个人的视频大全免费| 夜夜看夜夜爽夜夜摸| 熟女电影av网| 嫩草影院精品99| 国产精品女同一区二区软件| 欧美一区二区亚洲| or卡值多少钱| 国产精品爽爽va在线观看网站| 男女边吃奶边做爰视频| 床上黄色一级片| 免费不卡的大黄色大毛片视频在线观看 | 午夜福利在线在线| 少妇丰满av| 国产高潮美女av| 国产午夜精品久久久久久一区二区三区| 日韩精品青青久久久久久| 男女啪啪激烈高潮av片| 国产高潮美女av| 中文字幕av成人在线电影| 中文天堂在线官网| 午夜福利高清视频| 国产乱人视频| 在线观看av片永久免费下载| 久久精品国产鲁丝片午夜精品| 在线a可以看的网站| 国产精品久久久久久av不卡| 精品人妻一区二区三区麻豆| 91狼人影院| 一夜夜www| 日本五十路高清| 人人妻人人澡欧美一区二区| 亚洲第一区二区三区不卡| 99视频精品全部免费 在线| 成人亚洲欧美一区二区av| 一级黄色大片毛片| 国产成人精品久久久久久| 人妻夜夜爽99麻豆av| 日本黄大片高清| 国产爱豆传媒在线观看| 久久久午夜欧美精品| 天天一区二区日本电影三级| 中文字幕av在线有码专区| 国产单亲对白刺激| 国产精品一区www在线观看| 在线免费十八禁| 国产精品一二三区在线看| 麻豆av噜噜一区二区三区| 国产单亲对白刺激| 中文字幕制服av| 天堂影院成人在线观看| 日韩视频在线欧美| 中文资源天堂在线| 成人高潮视频无遮挡免费网站| 国产探花极品一区二区| 国语自产精品视频在线第100页| av福利片在线观看| 中文亚洲av片在线观看爽| 国产一级毛片七仙女欲春2| 亚洲精品,欧美精品| 亚洲国产精品sss在线观看| 欧美日韩在线观看h| 久久国产乱子免费精品| 久久婷婷人人爽人人干人人爱| 草草在线视频免费看| 在线观看一区二区三区| 国产不卡一卡二| 日本wwww免费看| 国产亚洲91精品色在线| 91久久精品国产一区二区三区| 欧美一区二区亚洲| 91在线精品国自产拍蜜月| 国语对白做爰xxxⅹ性视频网站| 国产精品久久久久久精品电影小说 | 人体艺术视频欧美日本| 黄色一级大片看看| 人体艺术视频欧美日本| 美女脱内裤让男人舔精品视频| 日本黄大片高清| 国产av在哪里看| 亚洲国产精品sss在线观看| 亚洲怡红院男人天堂| 日韩精品有码人妻一区| 黄片无遮挡物在线观看| 精品一区二区三区人妻视频| 国产免费一级a男人的天堂| 国产亚洲5aaaaa淫片| 日本wwww免费看| eeuss影院久久| 纵有疾风起免费观看全集完整版 | 精品无人区乱码1区二区| 欧美成人精品欧美一级黄| 亚洲天堂国产精品一区在线| 亚洲av中文字字幕乱码综合| 伦理电影大哥的女人| 亚洲av日韩在线播放| 国产午夜精品久久久久久一区二区三区| 亚洲av免费在线观看| 午夜亚洲福利在线播放| av国产久精品久网站免费入址| 亚洲欧美日韩高清专用| 国产一级毛片在线| 最近最新中文字幕大全电影3| 国产成人午夜福利电影在线观看| 国产精品不卡视频一区二区| 精品酒店卫生间| 亚洲欧美精品自产自拍| 国产亚洲5aaaaa淫片| 热99在线观看视频| 亚洲av福利一区| 视频中文字幕在线观看| 午夜免费激情av| 国产中年淑女户外野战色| 午夜精品国产一区二区电影 | 特级一级黄色大片| 色网站视频免费| 日韩精品青青久久久久久| 久久亚洲精品不卡| 又粗又硬又长又爽又黄的视频| 乱人视频在线观看| 久久午夜福利片| 能在线免费观看的黄片| 免费黄网站久久成人精品| 一级爰片在线观看| 久久亚洲国产成人精品v| 精品不卡国产一区二区三区| 日韩av在线免费看完整版不卡| 国产不卡一卡二| 国产精品一区二区三区四区久久| 女人十人毛片免费观看3o分钟| 一级毛片我不卡| 狂野欧美白嫩少妇大欣赏| 欧美高清性xxxxhd video| 精品久久久噜噜| av在线亚洲专区| 精品不卡国产一区二区三区| 国产极品天堂在线| 日韩制服骚丝袜av| 免费播放大片免费观看视频在线观看 | 成人午夜精彩视频在线观看| 国产成人精品一,二区| 亚洲人成网站在线播| 亚洲图色成人| 日本黄色片子视频| av黄色大香蕉| 欧美高清成人免费视频www| 欧美日韩综合久久久久久| 1000部很黄的大片| 亚洲欧美成人综合另类久久久 | 99热这里只有精品一区| 夜夜爽夜夜爽视频| 国产探花极品一区二区| 国内精品美女久久久久久| 亚洲精品,欧美精品| 亚洲国产精品专区欧美| 中文字幕熟女人妻在线| 一个人观看的视频www高清免费观看| 午夜爱爱视频在线播放| 国产久久久一区二区三区| 国产淫语在线视频| 日韩,欧美,国产一区二区三区 | 国产成人a∨麻豆精品| 亚洲欧美一区二区三区国产| 久久久久久久国产电影| 亚洲欧美成人精品一区二区| 91久久精品电影网| 99久久精品热视频| 免费观看人在逋| 国产精品熟女久久久久浪| 亚洲av电影不卡..在线观看| 天堂av国产一区二区熟女人妻| h日本视频在线播放| 亚洲精品,欧美精品| 久久久久久久久久黄片| 国产黄片视频在线免费观看| 亚洲综合精品二区| 一级av片app| 天堂av国产一区二区熟女人妻| 老司机影院毛片| 亚洲精品亚洲一区二区| 久久久久久伊人网av| 免费看美女性在线毛片视频| 国产真实乱freesex| 九九爱精品视频在线观看| 两性午夜刺激爽爽歪歪视频在线观看| 三级国产精品欧美在线观看| av免费观看日本| 亚洲国产欧美人成| 亚洲精品国产av成人精品| 亚洲国产成人一精品久久久| 国产伦精品一区二区三区四那| 成人亚洲欧美一区二区av| 久久精品国产亚洲网站| 国产综合懂色| 伦理电影大哥的女人| 免费看a级黄色片| 色吧在线观看| 岛国在线免费视频观看| 国产美女午夜福利| 国产色爽女视频免费观看| 一个人看视频在线观看www免费| 国产午夜福利久久久久久| 色综合亚洲欧美另类图片| 少妇裸体淫交视频免费看高清| 美女cb高潮喷水在线观看| 日本三级黄在线观看| 亚洲成色77777| 精华霜和精华液先用哪个| 少妇的逼好多水| 日本一二三区视频观看| 尤物成人国产欧美一区二区三区| 久久鲁丝午夜福利片| 亚洲精品日韩在线中文字幕| 亚洲国产欧美人成| 国产淫片久久久久久久久| 又黄又爽又刺激的免费视频.| 在线观看66精品国产| 国产精品无大码| 亚洲精品国产成人久久av| 午夜爱爱视频在线播放| 午夜精品一区二区三区免费看| 人体艺术视频欧美日本| 亚洲av成人精品一二三区| 女人十人毛片免费观看3o分钟| 色视频www国产| 国产欧美日韩精品一区二区| 精品国产露脸久久av麻豆 | 国产又色又爽无遮挡免| 成人毛片a级毛片在线播放| 五月玫瑰六月丁香| 2021少妇久久久久久久久久久| 中文字幕人妻熟人妻熟丝袜美| 最近中文字幕2019免费版| 九九久久精品国产亚洲av麻豆| 91久久精品国产一区二区成人| 免费看av在线观看网站| 尾随美女入室| 欧美一区二区精品小视频在线| 五月玫瑰六月丁香| 国产亚洲精品久久久com| 尤物成人国产欧美一区二区三区| 免费无遮挡裸体视频| 亚洲自拍偷在线| 中文字幕熟女人妻在线| 春色校园在线视频观看| 久久精品国产鲁丝片午夜精品| 最近中文字幕2019免费版| 99久久精品热视频| 插阴视频在线观看视频| 91精品国产九色| 免费无遮挡裸体视频| 日本wwww免费看| 男插女下体视频免费在线播放| 丝袜喷水一区| 在线天堂最新版资源| 美女黄网站色视频| 99久久精品国产国产毛片| 青春草视频在线免费观看| 男女啪啪激烈高潮av片| 国产精品日韩av在线免费观看| 国产老妇伦熟女老妇高清| 又爽又黄a免费视频| 少妇熟女aⅴ在线视频| 亚洲欧美日韩无卡精品| 青青草视频在线视频观看| 超碰97精品在线观看| 欧美人与善性xxx| 精华霜和精华液先用哪个| 天美传媒精品一区二区| 美女高潮的动态| 一个人看视频在线观看www免费| 少妇人妻一区二区三区视频| 一级毛片电影观看 | 晚上一个人看的免费电影| 欧美日韩国产亚洲二区| 日韩精品青青久久久久久| 国产高清国产精品国产三级 | 久久久久久久午夜电影| 国产精品一区www在线观看| 欧美97在线视频| av.在线天堂| 国产精品久久久久久精品电影| 中文字幕熟女人妻在线| 免费在线观看成人毛片| kizo精华| 亚洲精品久久久久久婷婷小说 | av卡一久久| 蜜桃久久精品国产亚洲av| av天堂中文字幕网| 国产伦精品一区二区三区视频9| av视频在线观看入口| 中文字幕亚洲精品专区| 一个人看的www免费观看视频| 亚洲精品日韩在线中文字幕| 免费观看人在逋| 99热这里只有精品一区| 九草在线视频观看| 亚洲精品乱久久久久久| 禁无遮挡网站| 国产精品不卡视频一区二区| 免费观看a级毛片全部| 六月丁香七月| 亚洲在线自拍视频| 久久久亚洲精品成人影院| av视频在线观看入口| 国产亚洲av片在线观看秒播厂 | 亚洲不卡免费看| 人体艺术视频欧美日本| 国国产精品蜜臀av免费| 99热这里只有是精品在线观看| 国产成人freesex在线| 我的女老师完整版在线观看| 51国产日韩欧美| 岛国毛片在线播放| 国产男人的电影天堂91| 男人的好看免费观看在线视频| 美女被艹到高潮喷水动态| 亚洲乱码一区二区免费版| 亚洲五月天丁香| 欧美一级a爱片免费观看看| 国产精品无大码| 97热精品久久久久久| 波多野结衣巨乳人妻| 在线观看一区二区三区| 2021少妇久久久久久久久久久| 亚洲精华国产精华液的使用体验| 久久国产乱子免费精品| 午夜老司机福利剧场| 天堂av国产一区二区熟女人妻| 黄片wwwwww| 麻豆成人午夜福利视频| 免费看光身美女| 九九热线精品视视频播放| 久久久久久久国产电影| 免费观看在线日韩| 亚洲最大成人av| 内地一区二区视频在线| 免费观看在线日韩| 久久综合国产亚洲精品| 亚洲国产精品久久男人天堂| 永久免费av网站大全| 日本爱情动作片www.在线观看| 国产 一区 欧美 日韩| 日韩精品青青久久久久久| 日本爱情动作片www.在线观看| 色尼玛亚洲综合影院| 日韩欧美 国产精品| 久久99热这里只频精品6学生 | 99久久九九国产精品国产免费| 国产成人aa在线观看| 国产 一区 欧美 日韩| 亚洲精品久久久久久婷婷小说 | 人妻系列 视频| av福利片在线观看| 国产精品爽爽va在线观看网站| 久久精品国产亚洲av天美| 久久久久久九九精品二区国产| 日日干狠狠操夜夜爽| 免费观看精品视频网站| 91av网一区二区| 国产片特级美女逼逼视频| 人妻制服诱惑在线中文字幕| 国产欧美另类精品又又久久亚洲欧美| 色视频www国产| 国产一级毛片七仙女欲春2| 视频中文字幕在线观看| 国产成人a区在线观看| 只有这里有精品99| 午夜福利在线观看吧| 91精品一卡2卡3卡4卡| 免费电影在线观看免费观看| 国产中年淑女户外野战色| 一个人看的www免费观看视频| 91久久精品国产一区二区三区| 亚洲精品乱码久久久v下载方式| 午夜福利高清视频| 最近中文字幕高清免费大全6| 免费观看精品视频网站| 五月伊人婷婷丁香| 亚洲人与动物交配视频| 在线观看66精品国产| 淫秽高清视频在线观看| 最近最新中文字幕大全电影3| 村上凉子中文字幕在线| 在现免费观看毛片| 免费无遮挡裸体视频| 变态另类丝袜制服| 亚洲国产精品久久男人天堂| 欧美又色又爽又黄视频| 春色校园在线视频观看| 国产探花极品一区二区| 床上黄色一级片| 在线播放国产精品三级| 国产精品福利在线免费观看| 禁无遮挡网站| 精品酒店卫生间| 听说在线观看完整版免费高清| 亚洲av成人精品一区久久| 色噜噜av男人的天堂激情| 秋霞在线观看毛片| av在线蜜桃| 久久午夜福利片| 国产伦在线观看视频一区| 亚洲av福利一区| 日本三级黄在线观看| 欧美日韩在线观看h| 水蜜桃什么品种好| 国产午夜福利久久久久久| 亚洲国产精品成人综合色| 亚洲成av人片在线播放无| 日韩欧美三级三区| 免费黄网站久久成人精品| 2022亚洲国产成人精品| 午夜免费激情av| www.av在线官网国产| 中文精品一卡2卡3卡4更新| 欧美高清性xxxxhd video| 亚洲精品久久久久久婷婷小说 | 久久久久久久国产电影| 国产极品天堂在线| 中文乱码字字幕精品一区二区三区 | 我要搜黄色片| 久久久久久久久中文| 国产av不卡久久| 高清在线视频一区二区三区 | 亚洲18禁久久av| 精品99又大又爽又粗少妇毛片| 自拍偷自拍亚洲精品老妇| 男人和女人高潮做爰伦理| h日本视频在线播放| 校园人妻丝袜中文字幕| 国产又色又爽无遮挡免| 美女黄网站色视频| 成年av动漫网址| 国产一区二区三区av在线| 欧美xxxx性猛交bbbb| 1000部很黄的大片| 国产精华一区二区三区| h日本视频在线播放| 午夜亚洲福利在线播放| 波多野结衣巨乳人妻| 中文精品一卡2卡3卡4更新| 高清视频免费观看一区二区 | 长腿黑丝高跟| 又爽又黄无遮挡网站| 国产三级中文精品| 在现免费观看毛片| 小说图片视频综合网站| 国产黄色小视频在线观看| 国产欧美日韩精品一区二区| 久久精品影院6| 丝袜喷水一区| 国产私拍福利视频在线观看| 精品久久久久久成人av| 亚洲乱码一区二区免费版| 99热6这里只有精品| 中文字幕av成人在线电影| 精品欧美国产一区二区三| 国产熟女欧美一区二区| 亚洲精品国产av成人精品| 中文亚洲av片在线观看爽| 18禁在线无遮挡免费观看视频| 伦理电影大哥的女人| 观看美女的网站| 国产伦在线观看视频一区| 噜噜噜噜噜久久久久久91| 可以在线观看毛片的网站| 久久久久久大精品| av专区在线播放| 热99在线观看视频| 天堂影院成人在线观看| 久久99热这里只有精品18| 亚洲国产精品国产精品| 亚洲美女视频黄频| 中文字幕亚洲精品专区| 国产不卡一卡二| 亚洲精品乱码久久久v下载方式| 日韩精品有码人妻一区| 国产久久久一区二区三区| 免费播放大片免费观看视频在线观看 | 午夜亚洲福利在线播放| 一个人看视频在线观看www免费| 日韩成人av中文字幕在线观看| 熟妇人妻久久中文字幕3abv| 深爱激情五月婷婷| 特级一级黄色大片| 在线观看美女被高潮喷水网站| 中国美白少妇内射xxxbb| 淫秽高清视频在线观看| 尾随美女入室| 视频中文字幕在线观看| 免费看光身美女| 色播亚洲综合网| 波野结衣二区三区在线| 亚洲国产最新在线播放| 一级黄色大片毛片| 亚洲国产最新在线播放| 久久精品91蜜桃| 在现免费观看毛片| 成人毛片a级毛片在线播放| 日韩 亚洲 欧美在线| 亚洲av中文av极速乱| 久久精品久久久久久久性| 亚洲欧美日韩无卡精品| 午夜精品国产一区二区电影 | 超碰av人人做人人爽久久| 不卡视频在线观看欧美|