key: cord-1036007-hlspwdub authors: Uçar, Emine; Atila, Ümit; Uçar, Murat; Akyol, Kemal title: Automated detection of Covid-19 disease using deep fused features from chest radiography images date: 2021-06-11 journal: Biomed Signal Process Control DOI: 10.1016/j.bspc.2021.102862 sha: fedf5636dbfa359fd0a8d77c07fb08fedb71822c doc_id: 1036007 cord_uid: hlspwdub The health systems of many countries are desperate in the face of Covid-19, which has become a pandemic worldwide and caused the death of hundreds of thousands of people. In order to keep Covid-19, which has a very high propagation rate, under control, it is necessary to develop faster, low-cost and highly accurate methods, rather than costly Polymerase Chain Reaction test that can yield results in a few hours. In this study, a deep learning-based approach that can detect Covid-19 quickly and with high accuracy on X-ray images, which are common in every hospital and can be obtained at low cost, is proposed. Deep features are extracted from X-Ray images in RGB, CIE Lab and RGB CIE color spaces using DenseNet 121 and EfficientNet B0 pre-trained deep learning architectures and then obtained features are fed into a two-stage classifier approach. Each of the classifiers in the proposed approach performs binary classification. In the first stage, healthy and infected samples are separated, and in the second stage, infected samples are detected as Covid-19 or pneumonia. In the experiments, Bi-LSTM network and well-known ensemble approaches such as Gradient Boosting, Random Forest and Extreme Gradient Boosting were used as the classifier model and it was seen that the Bi-LSTM network had a superior performance than other classifiers with 92.489% accuracy. Advances in the field of artificial intelligence increase its importance in the interpretation of medical images in order to support the early detection, correct diagnosis and treatment of diseases [1] . The Covid-19 disease that occurred in Wuhan, China in December 2019 has spread rapidly and has become a pandemic. Investigating the causes and effects of the Covid-19 pandemic, which is a serious threat to human health, has become a focus for scientists and healthcare professionals. The effects of this disease on people are followed anxiously by the world. In addition to the damage caused by Covid-19 disease to the organs in the human body, many researches are conducted on their psychological effects [2] [3] [4] . Researchers are constantly making efforts to control this epidemic and to find possible solutions in their fields [5] [6] [7] . One of the important steps in the fight against Covid-19 is to ensure that the infected patients can be screened effectively and thus they can be isolated and treated. Real-time reverse transcription polymerase chain reaction is the main screening method currently used for scanning [8, 9] . As an alternative to this method, the researchers [10, 11] stated that chest radiography images may be useful in Covid-19 detection. Studies have reported that patients with Covid-19 symptoms have mist-darkened spots in their lungs that can separate these patients from Covid-19 noninfected individuals [11, 12] . Therefore, systems based on chest radiology are considered an effective material for the detection and classification of Covid-19. The method used in the acquisition of these images is the use of computerized tomography scan (CT-Scan) and X-rays in a hospital with medical equipment. Since most of the hospitals have CT imaging machines, the systems developed based on these images can be useful in order to test many patients quickly in hospitals where there are no test kits or in limited numbers. Moreover, diagnosis and interpretation of Covid-19 disease using chest CT images requires additional time for a field specialist, and an increase in the workload on field specialist due to densities in hospitals may result in unintentional erroneous decisions. Today, in many countries, researchers from different disciplines are doing research to fight Covid-19, as well as other researchers are conducting many experimental studies to detect Covid-19 from chest radiograph images. In recent years, the increase in the speed and capacities of CPUs and GPUs has enabled the development of new high-performance computing models that can directly process raw data without the need for features [13] . Deep neural network architectures with multiple layers and neurons can efficiently perform high-complexity tasks such as voice and image recognition by processing large-size data. The use of deep learning models in the diagnosis and classification of diseases from medical image is quite common [14] [15] [16] [17] [18] [19] . Extracting the features needed to perform classification in classical machine learning methods involves complex processes and must be done carefully. Moreover, hand-designed feature extraction directly affects classification performance. For this reason, many researchers [20] [21] [22] [23] [24] [25] [26] [27] [28] carried out deep learning-based studies for accurate and reliable detection of Covid-19. In this study, we propose a deep learning model that can be used in the design of an expert system that can automatically classify Covid-19, pneumonia and no-finding cases. This model extracts deep features on different color spaces such as RGB, CIE Lab and RGB-CIE obtained from chest radiography images using DenseNet and EfficientNet transfer learning approaches and performs classification with Bidirectional Long Short-Term Memory (Bi-LSTM) network. RGB-CIE color space is one of many RGB color spaces distinguished by a specific monochromatic main color group [29] . The most prominent feature of the CIE-Lab color space is the smooth change of the color space in terms of perception. This color space defines all colors that the human eye can perceive, and is more useful in digital image processing than RGB color space for image sharpening and removing artifacts from the image [30] . This study examines the effects of deep features on image classification which reveal the different characteristics of mentioned color spaces, and also evaluates the contribution of combining deep features obtained from different color spaces to classification accuracy. With this study, it is aimed to have a model that may contribute to the detection of Covid-19 at low cost, low error and high speed, especially in hospitals without diagnostic kits. The contributions of this paper are as follows: 1. The proposed approach performs direct detection of Covid-19 disease on chest radiograph images. 2. The approach we propose realizes 3-class classification with high accuracy without the need for handcrafted features. 3. Features that are automatically extracted from X-Ray images in different color spaces with deep transfer learning architectures are fused to increase the classification accuracy. 4. The proposed approach can be used as an efficient expert system to support field experts, as it can quickly perform direct classification on X-ray images that are easily obtained in hospitals. The rest of this work is organized as follows. Section 2 gives a literature on deep-learning based Covid-19 detection. Section 3 describes the dataset and presents proposed model used for Covid-19 detection. Experiments are presented in Section 4. The results obtained are given and discussed in Section 5. Finally, the study is concluded in Section 6. Khan et al. [20] The main focus of this study is to reveal a stable and efficient model that successfully detects Covid-19 disease from X-Ray images. To this aim, radiography images belonging to Covid-19, pneumonia and no-finding classes were used. This model extracts features from chest radiography images RGB, CIE Lab and RGB CIE color spaces using DenseNet121 and EfficientNetB0 pre-trained deep learning models and performs two-stage classification on the deep fused features. In the first stage, the images are handled in two classes as no-finding and others (Covid-19 and pneumonia). The images labeled as patients in the first stage are passed to second stage for detection of Covid-19 or pneumonia. Figure 1 shows the block diagram of the proposed model. As stated in the introduction section, the main goal of the study is to create a low-cost, fast-running diagnostic model with a low error rate. In this context, DenseNet121 (8,062,504) and EfficientNetB0 (5,330,571) models, which have much lower number of parameters compared to other state-of-theart deep learning architectures, were selected in the feature extraction part of the proposed model. While most of the state-of-the-art models experience losses in the features obtained from the image as they progress through the layers, in DenseNet121 architecture, each layer connects to the next layers and thus layers can access the properties of the previous layers. EfficientNet architecture, on the other hand, can reduce the size of the model by performing compound scaling, thus obtaining more efficient results. These two architectural studies have come to the fore as the most appropriate options for the goal of realizing the desired fast and efficient disease diagnosis. X-ray images used in the study belong to Covid-19, pneumonia and no-finding cases, and were collected from different resources [31, 32] . In the data sets, there are 1125 images in RGB color space in different sizes, and while 125 of these images belong to the Covid-19 class, there are 500 images belonging to each of the other classes. In order to achieve balanced data distribution, which is an important issue in machine learning, Covid-19 images were increased from 125 to 500 by applying data augmentation. Image rotation at 30-60-90 angles was chosen as the data augmentation method. When neural networks are trained, there is a decrease in feature maps due to convolution and subsampling processes. At the same time, there are losses in the image feature in the transition between layers. DenseNet 121 architecture was proposed by Huang et al. [33] for more effective use of features extracted from images. In this architecture, each layer is connected to the other layers in a feed-forward manner. In this way, any layer can access the property information of all previous layers. In addition, DenseNet's other advantages stand out as lightening the lost angle problem and generously reducing the number of parameters. The EfficientNet model, which reaches 84.4% accuracy with 66M parameter calculation load in the ImageNet classification problem, can be considered as a group of convolutional neural network models. EfficientNet includes 8 models between B0-B7 and as this model number grows, the number of parameters calculated does not increase much, while the accuracy increases remarkably. The purpose of deep learning is to obtain more efficient models with least number of parameters. Unlike other state-of-the-art deep learning models, the EfficientNet shrinks the model by compound scaling which performs scaling in terms of depth, width and resolution to produce more efficient results. In compound scaling method, firstly a grid search is performed to find the relationship between different scaling dimensions of the baseline network under a fixed resource constraint. Therefore, suitable scaling factors are determined for depth, width and resolution dimensions. These factors are then applied to scale the baseline network to the desired target network [34] . LSTM networks have been developed due to the gradient vanishing problem in traditional RNNs [35] . A memory cell is used for learning and recall tasks in LSTM. This memory cell stores information about learning. Gate mechanisms are created with nonlinear activation functions in the memory cell. These mechanisms regulate the transmission of stored information to the next layer or forgetting it. Unlike the traditional LSTM structure, the bidirectional LSTM has two hidden layers (forward and backward) connected to the outputs. While the output layer in the LSTM network obtains the information from the input data at time t and the previous hidden state (t-1, t-2,…, t-N), bidirectional LSTM network also uses subsequent hidden state (t+1, t+2,…, t+N). Processing data in two-way hidden layers opposite each other provides additional information to the network including future data, which enables faster and better learning. There are 3 classes in the dataset used in this study, and the proposed approach realizes the classification with two independent models, each of which performs binary classification. The performances of the classifiers used in both stages of the proposed approach in this study were measured using different metrics such as Sensitivity (Sen), Specificity (Spe), Accuracy (Acc), Precision (Pre) and F1-score. The formulas for these metrics are given in Eq. 1-10 respectively. For a class k, Then, In addition to these metrics, the performance of the classifiers can also be evaluated with the Area Under the Receiver Operating Characteristic (AUC-ROC) metric. The ROC curve shows the relationship between True Positive Rate (TPR) and False Positive Rate (FPR). The TPR is the same with the sensitivity metric and gives the ratio of correctly predicted positives to all positives. FPR is the ratio of negatives that are incorrectly predicted as positive to all negative data and can be expressed as 1-specifity. If the area under the ROC curve (Area Under Curve-AUC) is close to 1, this indicates that the model has a high success in separating the classes, if it is close to 0, this indicates that the model has a low success in separating the classes. If AUC equals to 0.5 this indicates that the model does not have the ability to discriminate between classes and selects a class randomly at each time [36] . All models used in this study are compiled with GPU support. All experimental studies were conducted in Google cloud environment using 64-bitUbuntu 18 The X-ray images used in this study are in RGB color space. In addition to this color space, X-ray images in the dataset were converted into other color spaces such as LAB and CIE in order to examine the effect of using different color spaces on classification performance. DenseNet 121 and Efficient B0 pretrained deep learning models were used to extract deep features from X-ray images. All images were resized to 224x224 dimensions and normalized by the min-max method before inputting to these models. The feature size that each pre-trained model extracts on each image in the fully connected layer was 1x1000. In this context, DenseNet 121 and Efficient B0 architectures extracted 3000 features on 3 different color spaces. Finally, these features were fused to obtain 6000 features for each image. These features were normalized in the range between 0 and 1 with min-max normalization. Thus, 1500x6000 dimensional feature vector was extracted. Figure 1 demonstrates the proposed approach that discriminates Covid-19 from other cases through two stages. In the first stage, a classifier model (Model 1) distinguishes healthy ones from infected (Covid-19 and pneumonia). In the second stage, Covid-19 cases are distinguished from pneumonia by another classifier model (Model 2). These two models work independently of each other. In the proposed approach, while all images including Covid-19, pneumonia and healthy cases are given as input to Model 1, Model 2 only takes Covid-19 and pneumonia images as input. In this study, the k-fold cross validation technique where k was set to 5 was used to reveal and compare the performances of the algorithms more accurately and thus determine the best algorithm. In this context, as seen in Figure 3 , in each fold, 80% of the X-ray images were used for training and the rest of the images were used for testing. In the first stage, the classifiers in Model 1 were fed with a total of 1200 images including 400 healthy and 800 diseased for training. Besides, a total of 300 images, 100 from each class, were used to test the performance of Model 1. In the second stage, the Model 2 was fed by 800 diseased images for training, and then the performance of this model was tested on a total of 200 test images, including 100 images for each class of Covid-19 and pneumonia. Thus, Model 1 and Model 2 in the proposed approach were built. In the study, 6 feature vectors of 1x1000 were obtained from images in each color space using DenseNet and EfficientNet pre-trained models. For all classifiers except Bi-LSTM, 6 feature vectors were combined and a 1x6000 feature vector was used as input, while in the Bi-LSTM network, 1x1000 feature vectors from each color space were given as input to the model in 6 steps sequentially. Along with selecting the Bi-LSTM as a classifier in the proposed approach, experiments were also conducted with other machine learning algorithms to compare their performance with Bi-LSTM. In this context, trainings were carried out in both stages of the proposed approach by recent popular ensemble learning algorithms such as Random Forest (RF), Gradient Boosting (GB) and Extreme Gradient Boosting (XGboost). These algorithms were run with default parameter values defined in Scikit-learn library. Bi-LSTM network was built in Keras environment. For both stages of the proposed approach, the number of input layer units of Bi-LSTM, was set as 1000 and the number of output units as 2. The Bi-LSTM model was designed with 3 hidden layers and number of hidden layer units was set to 16. The hidden layer weights of the Bi-LSTM model were randomly initialized to have uniform distribution in the range between -1 and 1. Adam optimization method was chosen for Bi-LSTM network and the parameters of this method were set as B1 = 0.9 and B2 = 0.999. The learning rate was set as 5e-04 and decay as 0 and mean squared error (MSE) was used as the loss function of the network. The training of the Bi-LSTM network was completed in 100 epochs and the weights in the epoch, which obtained best validation accuracy, were stored and used in the final model to prevent overfitting. During the training, batch size was chosen as 200 since the number of samples per class in the training set is divisible with this value, and this results in more efficient use of memory. A dropout layer has been added to the outputs of all LSTM hidden layers to prevent overfitting. In addition, specific to the LSTM network, the recurrent dropout method which makes forward and backward dropout in each LSTM layer was also applied. For both dropout technique, parameter values were applied by being varied between 0.0 and 0.6 with step value of 0.1. In accordance with 5-fold cross-validation technique, each test data in the data set was used as validation data in the Bi-LSTM network. In both stages of the proposed approach, as a result of trial and error method, the Bi-LSTM network provided the best performance when dropout and recurrent dropout values were selected as 0.2 and 0.2, respectively. Accuracy / Loss curves obtained from training and validation sets during the training process with these values were presented in Figure 4 . In this study, a two-stage classification approach was presented on deep features obtained from X-ray images using DenseNet 121 and EfficientNet B0 pre-trained architectures. The focus of the study was to predict Covid-19 disease with an acceptable accuracy. For the detection of Covid-19 disease, Bi-LSTM network was used in both stages of the proposed approach. In addition, the performance of the Bi-LSTM network used in the proposed approach was compared with ensemble learning algorithms such as Random Forest (RF), Gradient Boosting (GB) and Extreme Gradient Boosting (XGboost). The approach proposed in this study performs 3-class classification with two separate models that each make binary classifications consecutively. Binary classification accuracies of the classifiers used in Model 1 and Model 2 in the proposed approach for RGB, CIE Lab and RGB CIE color spaces and as well as for the combined color space were summarized in Table 1 and Table 2 . It was seen that both Model 1 and Model 2 used in the proposed approach achieved the highest accuracies using Bi-LSTM in all color spaces. Moreover, when the successes of the classifiers in Model 1 and Model 2 for the combined color space were examined, it was seen that Bi-LSTM offers the highest accuracy compared to others. In addition, Table 3 summarizes the performance of the classifiers tested in the 3-class classification with the average Acc, Sen, Spe, F1-score and Pre values obtained based on the 5-fold cross-validation method. As can be seen in this table, the best results were obtained with 92.489% average accuracy when Bi-LSTM was used in both stages of the proposed two-step approach. Among other classifiers in experimental studies, GB and XGB algorithms showed the closest performances to Bi-LSTM. The values obtained by these two algorithms in the context of all evaluation metrics were lower than the Bi-LSTM method in the range between 2% and 3%. In this context, since the highest performances in the proposed approach were obtained with Bi-LSTM network, results given below in this section were based on Bi-LSTM. 3-class classification performance obtained for each fold was evaluated with the accuracy, sensitivity, specificity, precision and F1-score metrics, and the average classification performances of 5-fold were presented in Table 4 . Accordingly, the proposed approach presented an average of 88.933% sensitivity, 94.367% specificity, 89.028% precision, 88.760% F1-score and 92.489% accuracy in the 3-class classification. In addition, obtained confusion matrices for each fold and overlapped confusion matrix were shown in Figure 5 . As can be seen in the confusion matrices, the proposed approach classified Covid-19 patients more successfully in each fold compared to other classes. For example, when overlapped confusion matrix was examined, the proposed approach misclassified 15 out of 500 Covid-19 cases, while it misclassified 73 out of 500 pneumonia and 78 out of 500 no-finding samples. Therefore, the proposed model can detect Covid-19 case more successfully compared to other cases. In the first stage of the two-stage model proposed here, Sensitivity and Precision values decrease due to misclassifications of the samples in pneumonia class as no-finding or vice versa. When the overlapped confusion matrices (Figure 6a-b) of the Bi-LSTM were examined, it was seen that the number of misclassified samples in the no-finding and pneumonia classes was higher than Covid-19. Especially when the overlapped confusion matrices of Model 1 and Model 2 were compared, it is understood that Model 1 labeled 73 of 500 healthy images as diseased and 74 of 1000 diseased images as healthy. Moreover, it was seen that the performance of Model 2 in classifying Covid-19 and pneumonia images was higher than Model 1's performance in classifying healthy and diseased images. As can be observed here, the performance of the proposed approach was lower in distinguishing between pneumonia and no-finding compared to Covid-19. It should also be noted that there were very small differences between the values in the overlapped confusion matrix obtained as a result of the 3-class classification and the values in the overlapped confusion matrix obtained in the 2-class classification of Model 1 and Model 2. This was because, in the proposed approach, Model 2 was dependent on classification according to the labeling result from Model 1. In other words, healthy images that Model 1 has mistakenly classified as patients must be classified as Covid-19 or Pneumonia by Model 2 and the image that Model 1 classified as healthy was not handled by Model 2. Some of the samples that were misclassified by the proposed model were examined by a pulmonologist and the possible reasons that were thought to cause the images to be misclassified were explained. Figure 7 shows eight sample images that were misclassified by the proposed twostage model in the study. The images given in Figure 7 -a and Figure 7 -b were Covid-19 but misclassified as Pneumonia. In Figure 7 -a, the faint infiltrative area in the left paracardiac region led to misclassification. In Figure 7 -b, existence of indistinct infiltrations in bilateral basals led to misclassification. The images given in Figure 7 -c and Figure 7 -d belong to no-finding case but misclassified as Pneumonia. In Figure 7 -c, scattered dense views in the right lower mediobasal region led to incorrect classification. The X-Ray image shown in Figure 7 -d was taken without adequate inspiration and also has scattered dense views in the right lower mediobasal region. The Pneumonia image given in Figure 7 -e was misclassified as no-finding due to difficulties in discriminating the upper and lower right regions as widespread density, consolidation or ground glass. Also this image has ground glass infiltration in the lower left region. In Figure 7 -f, uncertain linear density at the bottom left caused misclassification. The images given in Figure 7 -g and Figure 7 -h belong to pneumonia class but misclassified as Covid-19. In Figure 7 -g, several cases such as cavity at the left apex, two nodular views adjacent to the right hilus, appearance of branule at the right apex led to misclassification. Finally in Figure 7 -h, diffuse reticular infiltrations in both lungs in the image, nodule in the upper right and nodular infiltrations on the left region led to misclassification. Figure 8 shows the class-based ROC curves and AUC values of the classifiers. As can be seen from the ROC curves in Figure 8 , the two-stage model we propose detects Covid-19 positive patients, which is the main goal of this study, with higher success compared to pneumonia and no-finding classes. As shown in the graphics, Bi-LSTM obtained the best AUC values for each class with 0.983 for Covid-19, 0.878 for pneumonia and 0.890 for no-finding. Table 5 analyzes the performances of the deep learning models created for the diagnosis of Covid-19 and the data sets used in the literature. As can be seen in the table, the studies were generally carried out with two or three classes. While Covid-19 and normal classes were used in the studies with two classes, Covid-19, pneumonia and normal classes were used in three-class studies. Only in the study of Khan et al., the pneumonia case was divided into viral and bacterial, and the study was carried out with four classes [20] . It stands out that different numbers of images belonging to the relevant classes are used in the studies. Therefore, direct comparison of the performance of the models proposed in these studies with the performance of our model would be misleading. While the images used in the modified ResNet50 model of Rahimzadeh et al. were CT images, the images used in other studies were X-Ray images. In addition, the number of images used in their study was much higher than the number of images used in other studies. Consequently, the success of their model was higher than other studies with an overall accuracy of 98.49% in two-class classification [25] . Among the two-class studies using X-ray images, the DarkCovidNet model of Öztürk et al achieved the highest accuracy rate with 98.08% [22] . On the other hand, among the three-class studies, CoroNet model proposed by Khan et al. achieved the highest accuracy with 95% [20] . Since the dataset we used in this study was the same as the dataset in the study of Ozturk et al. [22] , a direct comparison was performed only with this study. Ozturk et al. proposed a deep learning model named DarkNet for the diagnosis of Covid-19. The authors reported that the low number of Covid-19 images posed a disadvantage for their studies. In our study to overcome this disadvantage, we increased the number of images belonging to the Covid-19 class from 125 images to 500 by applying augmentation techniques and brought to the same number with the other two classes. As a result, while the DarkNet model presented by Ozturk et al. achieved an average accuracy of 91.35% which is calculated using their confusion matrix in 3-class classification, our proposed approach achieved an accuracy of 92.49% with the Bi-LSTM classifier, providing a slightly better performance. Developing a method that can diagnose Covid-19 disease, which has turned into a pandemic that threatens the whole world, quickly and accurately, at low cost, is very important to prevent the collapse of health systems. Today, the most common method used for the detection of Covid-19 is Polymerase Chain Reaction (PCR) test. Although this test is concluded within a few hours, it can still be considered slow, given the rate of spread of the Covid-19 virus. In addition, the fact that PCR is a high cost method limits the number of people to be tested in countries with low welfare and crowded populations. This gives the pandemic an opportunity to get out of control by causing it to spread faster in such countries. In this study, in order to avoid such limitations, a deep learning-based approach that can detect Covid-19 disease on X-ray-based images in a very short time without the need for any feature extraction method is proposed. In order to have a balanced dataset and thus to have increased diagnostic success, the number of images belonging to the Covid-19 class was increased to have equal number of samples with the other two classes using with various augmentation techniques. Secondly, for increasing the diagnostic success of the model, the features extracted with DenseNet 121 and EfficientNet B0 pre-trained models in different color spaces, such as RGB, CIE Lab and RGB-CIE were fused. In both stages of the two-step approach proposed in the study, Bi-LSTM provided the best performance with 92.489% accuracy compared to other ensemble methods. Although the model proposed in this study had difficulty in separating the images of pneumonia and no-finding cases, it stood out with its high success in the detection of Covid-19 disease, which is the starting point of this study. Since the proposed approach can be directly applied on X-ray images that can be easily obtained in almost every hospital, it avoids the disadvantages of Antibody test and PCR test methods mentioned above in terms of both cost and time. Thus, the proposed approach can help the hospital workflow in determining which patients need PCR test, and thus reduce hospital workload, rather than making a definitive Covid-19 diagnosis. We achieved acceptable diagnosis performance with the limited number of Covid-19 image, and it is planned to build a more stable and successful model by working with more Covid-19 images in the future study. Besides, the hybrid machine learning-based feature extraction approaches proposed in recent years can be used to extract important features from X-Ray images, thus increasing the classification success by overcoming the difficulties in distinguishing between pneumonia and no-finding classes [37, 38] . High-performance models to be developed in this way can contribute to slow down the spread of Covid-19, especially in hospitals lacking high-capacity devices. Artificial intelligence in healthcare: past, present and future Psychological effects of nurses and midwives due to COVID-19 outbreak: The case of Turkey Psychological Distress and Its Correlates Among COVID-19 Survivors During Early Convalescence Across Age Groups The psychological status of people affected by the COVID-19 outbreak in China What works and what does not work in response to COVID-19 prevention and control in Africa Containing the spread of coronavirus disease 2019 (COVID-19): Meteorological factors and control strategies Unique challenges to control the spread of COVID-19 in the Middle East Detection of SARS-CoV-2 in Different Types of Clinical Specimens others, Detection of 2019 novel coronavirus (2019-nCoV) by real-time RT-PCR Chest CT Findings in Coronavirus Disease-19 (COVID-19): Relationship to Duration of Infection Chest CT for typical 2019-nCoV pneumonia: relationship to negative RT-PCR testing Sensitivity of Chest CT for COVID-19: Comparison to RT-PCR Deep learning Automatic classification of pulmonary peri-fissural nodules in computed tomography using an ensemble of 2D views and a convolutional neural network out-of-the-box Holistic classification of CT attenuation patterns for interstitial lung diseases via deep convolutional neural networks Deep MRI brain extraction: A 3D convolutional neural network for skull stripping Automatic Segmentation of MR Brain Images With a Convolutional Neural Network Deep learning for neuroimaging: a validation study Latent feature representation with stacked auto-encoder for AD/MCI diagnosis CoroNet: A deep neural network for detection and diagnosis of COVID-19 from chest x-ray images Deep Transfer Learning Based Classification Model for COVID-19 Disease Automated detection of COVID-19 cases using deep neural networks with X-ray images Application of deep learning for fast detection of COVID-19 in X-Rays using nCOVnet Covid-19: automatic detection from x-ray images utilizing transfer learning with convolutional neural networks A Fully Automated Deep Learning-based Network For Detecting COVID-19 from a New And Large Lung CT Scan Dataset Automated diagnosis of COVID-19 with limited posteroanterior chest Xray images using fine-tuned deep neural networks Automatic Detection of Coronavirus Disease (COVID-19) Using Xray Images and Deep Convolutional Neural Networks COVID-Net: A Tailored Deep Convolutional Neural Network Design for Detection of COVID-19 Cases from Chest X-Ray Images The C.I.E. colorimetric standards and their use Perceptual difference in L* a* b* color space as the base for object colour identfication COVID-19 Image Data Collection: Prospective Predictions Are the Future ChestX-ray8: Hospital-Scale Chest X-Ray Database and Benchmarks on Weakly-Supervised Classification and Localization of Common Thorax Diseases Densely Connected Convolutional Networks Rethinking model scaling for convolutional neural networks, 36th Int. Conf. Mach. Learn. ICML Long Short-Term Memory An introduction to ROC analysis A novel automatic classification system based on hybrid unsupervised and supervised machine learning for electrospun nanofibers A novel hybrid machine learning approach for change detection in remote sensing images We specially thank to Assoc. Prof. Dr. Sertaç Arslan for his invaluable evaluations on misclassified image samples. The authors declare no conflict of interest. The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.  Deep fused features from X-Ray images in different color spaces were used. An enhanced two-stage model was proposed for Covid-19 detection.