key: cord-0739404-l86s5ss8 authors: Yaşar, Huseyin; Ceylan, Murat title: A novel study for automatic two-class COVID-19 diagnosis (between COVID-19 and Healthy, Pneumonia) on X-ray images using texture analysis and 2-D/3-D convolutional neural networks date: 2022-01-29 journal: Multimed Syst DOI: 10.1007/s00530-022-00892-z sha: 3ffaa4bd4959253988fa1b82288e27ce68f78a5a doc_id: 739404 cord_uid: l86s5ss8 The pandemic caused by the COVID-19 virus affects the world widely and heavily. When examining the CT, X-ray, and ultrasound images, radiologists must first determine whether there are signs of COVID-19 in the images. That is, COVID-19/Healthy detection is made. The second determination is the separation of pneumonia caused by the COVID-19 virus and pneumonia caused by a bacteria or virus other than COVID-19. This distinction is key in determining the treatment and isolation procedure to be applied to the patient. In this study, which aims to diagnose COVID-19 early using X-ray images, automatic two-class classification was carried out in four different titles: COVID-19/Healthy, COVID-19 Pneumonia/Bacterial Pneumonia, COVID-19 Pneumonia/Viral Pneumonia, and COVID-19 Pneumonia/Other Pneumonia. For this study, 3405 COVID-19, 2780 Bacterial Pneumonia, 1493 Viral Pneumonia, and 1989 Healthy images obtained by combining eight different data sets with open access were used. In the study, besides using the original X-ray images alone, classification results were obtained by accessing the images obtained using Local Binary Pattern (LBP) and Local Entropy (LE). The classification procedures were repeated for the images that were combined with the original images, LBP, and LE images in various combinations. 2-D CNN (Two-Dimensional Convolutional Neural Networks) and 3-D CNN (Three-Dimensional Convolutional Neural Networks) architectures were used as classifiers within the scope of the study. Mobilenetv2, Resnet101, and Googlenet architectures were used in the study as a 2-D CNN. A 24-layer 3-D CNN architecture has also been designed and used. Our study is the first to analyze the effect of diversification of input data type on classification results of 2-D/3-D CNN architectures. The results obtained within the scope of the study indicate that diversifying X-ray images with tissue analysis methods in the diagnosis of COVID-19 and including CNN input provides significant improvements in the results. Also, it is understood that the 3-D CNN architecture can be an important alternative to achieve a high classification result. The COVID-19 virus emerged in China in December 2019 and quickly spread to other countries. As a result of this rapid spread, the World Health Organization (WHO) announced in March that the disease had developed into a pandemic [1] . The pandemic still affects many countries deeply. Even in countries with developed health systems, there are occasional infrastructural deficiencies [2] . The main reason for this situation is the increase in the need for intensive care and respiratory support due to pneumonia caused by the virus [3] . As of the last week of November 2020, the number of people infected with the virus has reached 60 million. The number of deaths related to the virus approached 1.5 million [4] . Vaccine development studies against the COVID-19 virus are still ongoing. However, it is understood that the pandemic will affect people and countries economically, socially, and politically for a long time. The most common symptoms of the COVID-19 virus are fever, shortness of breath, and pneumonia [5] . The diagnosis of the COVID-19 virus is carried out with reverse transcription-polymerase chain reaction (RT-PCR). Generally, this test produces a result within a few hours [6] . One of the most important factors in controlling the COVID-19 outbreak is the rapid diagnosis and isolation of patients carrying the virus. The RT-PCR test takes hours to produce results, making it difficult to quickly isolate individuals suspected of carrying the virus. In addition, since the human factor in swabbing for this test is high, there are some problems producing the correct result. For this reason, the WHO requires not only the direct reporting of patients with positive RT-PCR tests but also the cases that are detected to carry this virus by radiological or clinical symptoms despite a negative RT-PCR result [7] . In academic studies on COVID- 19 , it has been shown that the virus causes interstitial involvement, lung opacities, bilateral ground-glass, and patchy opacity in the lungs [8] . This situation means that radiological imaging methods, such as CT, X-ray, and ultrasound, are important tools in the diagnosis of the COVID-19 virus. When they examine the CT, X-ray, and ultrasound images, the first job radiologist do is to determine whether there are signs of COVID-19 in the images. That is, COVID-19/Healthy detection is made. Then, they determine whether it is the separation of pneumonia caused by the COVID-19 virus or pneumonia caused by a bacteria or virus other than COVID-19. The result of this distinction is an important indicator when determining the treatment and isolation procedure to be applied to the patient. This distinction can be described as the COVID-19 Pneumonia/Bacterial, COVID-19 Pneumonia/Viral Pneumonia, and COVID-19 Pneumonia/Other Pneumonia classification. In the academic studies carried out, the different aspects of pneumonia caused by the COVID-19 virus from other types of pneumonia have been presented in detail [9] . Automated inspection of CT, X-ray, and ultrasound images is an important requirement due to the heavy workload imposed on the healthcare system and staff by the outbreak [10] . In addition, it is an important auxiliary factor in controlling the pandemic as it will speed up the detection of COVID-19. When the methods used in the studies given in Sect. 2 are examined in detail, they generally suggest different CNN architectures for classification. In this context, it is understood that they generally focus on revealing the effects of CNN architecture change on study results. In this study, the effect of diversification of CNN input images using texture feature methods instead of the CNN architecture used differently from previous studies was investigated. In this context, CNN architecture, which provides the best results in previous studies, was added to the study. In this study aiming at the early diagnosis of COVID-19 disease using X-ray images, an automatic two-class classification in four different titles as COVID-19/Healthy, COVID-19 Pneumonia/Bacterial Pneumonia, COVID-19 Pneumonia/Viral Pneumonia, and COVID-19 Pneumonia/ Other Pneumonia has been carried out. Within the scope of the study, 3405 COVID-19, 2780 Bacterial Pneumonia, 1493 Viral Pneumonia, and 1989 Healthy images obtained by combining eight different data sets with open access were used. In the study, besides the direct use of the original X-ray images alone, classification results were obtained using the images resulting from Local Binary Pattern (LBP) and Local Entropy (LE). The classification processes were repeated for the images obtained by combining these images in various ways. Classifiers used within the scope of the study were 2-D CNN (Two-Dimensional Convolutional Neural Networks) and 3-D CNN (Three-Dimensional Convolutional Neural Networks) architectures. As a 2-D CNN, Mobile-netv2, Resnet101, and Googlenet architectures were used in the study. As a 3-D CNN, a 24-layer 3-D CNN architecture has been designed. The results of the study indicate that as a result of the proposed methods, significant gains can be achieved in the diagnosis of COVID-19 using X-ray images. It is possible to summarize the innovative aspects and importance of our work with the following main headings. • Within the scope of our study, experiments were carried out with several images that have never been used in any study before. In this context, experiments were conducted with images approximately ten times more than the average number of COVID-19 images used in previous studies in our research. In addition, unlike other literature studies, more than 90% of the images in question are taken from the real world. • CNN architectures, which provide the highest results in previous studies, have also been added to our study. Experiments were carried out to diversify the input data for the mentioned CNN architectures. In this way, much more comprehensive and higher results have been obtained. • In addition to 2-D CNN architectures, similar operations were performed for a 3-D CNN architecture. The effects of diversifying the input data of the 3-D CNN architecture on the study results were revealed. In addition, the achievements of 2-D CNN architectures and 3-D CNN architectures are extensively compared. Some academic studies have been carried out to meet this need by disseminating and sharing radiological data on patients with COVID-19. In this context, two-class classification studies, carried out in 2020 and focused on the COVID-19/Healthy distinction, were examined. In the COVID-19/Healthy classification study performed by Tuncer et al. [11] , a total of 321 X-ray images, including 87 COVID-19 and 234 Healthy, were used. In the study in question, Residual Exemplar Local Binary Pattern, Iterative Relief, Decision Tree, Linear Discriminant, Support Vector Machine, k-Nearest Neighborhood, and Subspace Discriminant methods were used in classification procedures. In the study, train-test procedures were carried out using three different methods: tenfold cross-validation, 80% Train-20% Test, and 50% Train-50% Test. At the end of the study, sensitivity values varying between 0.8149 and 1.0000, specificity varying between 0.9380 and 1.0000, and accuracy values varying between 0.9049 and 0.9955 were obtained. Panwar et al. [12] conducted a COVID-19/Healthy classification study with a total of 284 X-ray images, including 142 COVID-19 and 142 Healthy, using the Convolutional Neural Network (CNN) (nCOVnet). The sensitivity was 0.9762, specificity was 0.7857, and accuracy was 0.881 as a result of training and testing using the 70% Train-30% Test. In the classification study performed by Ozturk et al. [13] , 625 X-ray images (125 COVID-19 and 500 Healthy) were used. In the study in which CNN (DarkNet) was used as a classifier, training-testing procedures were carried out using a fivefold cross-validation approach. At the end of the study, sensitivity was 0.9513, specificity 0.953, accuracy 0.9808, and F-1 Score 0.9651. Mohammed et al. [14] proposed a decision tree mechanism for COVID-19/Healthy classification. The decision tree in question is based on combining the results obtained using Naive Bayes, Neural Network, Support Vector Machine, Radial Basis Function, k-Nearest Neighbors, Stochastic Gradient Descent, Random Forests, Decision Tree, AdaBoost, and CN2 Rule Inducer Algorithm methods. As a result of the study performed using a total of 50 X-ray images, including 25 COVID-19 and 25 Healthy, sensitivity values varying between 0.706 and 0.974 and specificity values varying between 0.557 and 1.000 were obtained. In addition, accuracy values ranging from 0.620 to 0.987, F-1 Score values varying between 0.555 and 0.987, and Area Under Receiver-Operating Characteristic Curve (AUC) values varying between 0.800 and 0.988 were provided. In the study, the run time was also measured, and results were obtained at times ranging from 0.14 to 7.57 s per image. In a study conducted by Toraman et al. [15] , a total of 1281 X-ray images were used, including 231 COVID-19 and 1050 Healthy. The training and testing procedures of the study, in which the CNN (CapsNet) was used as a classifier, were carried out based on the tenfold crossvalidation method. At the end of the study, sensitivity values varying between 0.28 and 0.9742, specificity values varying between 0.8095 and 0.98, accuracy values varying between 0.4914 and 0.9724, and F-1 Score values varying between 0.55 and 0.9724 were provided. In the study, the run time per image was measured at values ranging from 16 to 500 s. In the classification study performed by Khan et al. [16] using 594 X-ray images (284 COVID-19 and 310 Healthy), the CNN [CoroNet (Xception]) classifier was used. As a result of the training and testing processes performed using the fourfold cross-validation method, sensitivity was calculated as 0.993, specificity 0.986, accuracy 0.990, and F-1 Score 0.985. In the COVID-19/Healthy classification study carried out by Yaşar and Ceylan [17] , two datasets containing 230 X-ray images (150 COVID-19 and 80 Healthy) and 476 X-ray images (150 COVID-19 and 326 Healthy) were used. In addition, the classification process was repeated by combining these two datasets (556 X-ray images [150 COVID-19 and 406 Healthy]). In the study, two different CNN architectures, 23-layer and 54-layer, were used with Local Binary Pattern. The Dual-Tree Complex Wavelet Transform was used to reduce the image dimensions. In addition, in the study, pipeline algorithms that enable combining the results obtained using or not using Local Binary Pattern are proposed. As a result of the training and testing processes performed using the 23-fold cross-validation and twofold cross-validation approach, sensitivity values varying between 0.9253 and 0.9947, specificity varying between 0.9725 and 1.0000, and accuracy values varying between 0.9745 and 0.9906 were obtained. In addition, F-1 Score values ranging from 0.9511 to 0.9881, and AUC values ranging from 0.9975 to 0.9997 were provided. In the study conducted by Waheed et al. [18] , CNN (VGG-16) and Synthetic Data Augmentation methods were used. A total of 1124 X-ray images (403 COVID-19 and 721 Healthy) were divided into two parts for training (932 X-ray images [331 COVID-19 and 601 Healthy]) and testing (192 X-ray images [72 COVID-19 and 120 Healthy]). As a result of the training-test processes, sensitivity values varying between 0.69 and 0.90, specificity between 0.95 and 0.97, and accuracy values varying between 0.85 and 0.95 were achieved. Duran-Lopez et al. [19] used CNN as the classifier in their COVID-19/Healthy classification study using a total of 6926 X-ray images, including 2589 COVID-19 and 4337 Healthy. As a result of the training and testing procedures performed using the fivefold cross-validation approach, sensitivity was 0.9253, specificity 0.9633, accuracy 0.9443, F-1 Score 0.9314, and AUC 0.988. Vaid et al. [20] In the study conducted by Benbrahim et al. [21] , a total of 320 X-ray images were used, including 160 COVID-19 and 160 Healthy. The training and testing procedures of the study in which Transfer Learning with Convolutional Neural Networks (Inceptionv3 and ResNet50) were used as a classifier were performed according to the 70% Train-30% Test procedure. At the end of the study, sensitivity ranging between 0.9803 and 0.9811, accuracy varying between 0.9803 and 0.9901, and F-1 Score values varying between 0.9803 and 0.9901 were obtained. In the study conducted by Loey et al. [22] , Transfer Learning with CNN (Alexnet, Googlenet, and Resnet18) methods were used for classification. Within the scope of the study, 148 images (69 COVID-19 and 79 Healthy) were divided into two parts: 130 images for training (60 COVID-19 and 70 Healthy), and 18 images for testing (nine COVID-19 and nine Healthy). At the end of the study, sensitivity, specificity, and accuracy values were obtained as 1.000. Minaee et al. [23] suggested using Transfer Learning with CNN (ResNet18, ResNet50, SqueezeNet, and DenseNet-121) methods for COVID-19/Healthy classification. While performing the training and testing processes of the study, 5184 images (184 COVID-19 and 5000 Healthy) were divided into two parts: 2084 images (84 COVID-19 and 2000 Healthy) for training and 3100 images (100 COVID-19 and 3000 Healthy) for testing. At the end of the study, sensitivity was obtained as 0.98. Specificity values ranging from 0.751 to 0.929 have also been calculated. In the study conducted by Elaziz et al. [24] , two different datasets containing 1891 images (216 COVID-19 and 1675 Healthy) and 1560 images (219 COVID-19 and 1341 Healthy) were used. In the study where classification processes were performed using Fractional Multichannel Exponent Moments, Manta-Ray Foraging Optimization, and k-NN classifier, training and testing procedures were carried out according to 80% Train-20% Test procedures. At the end of the study, sensitivity values ranging from 0.9875 to 0.9891 and accuracy values between 0.9609 and 0.9809 were achieved. Martínez et al. [25] conducted a study using the CNN (Neural Architecture Search Network [NASNet]) classifier with 240 images (120 COVID-19 and 120 Healthy) to classify COVID-19/Healthy. As a result of the training and testing performed with the 70% Train-30% Test procedure, the sensitivity was 0.97, the accuracy 0.97, and the F-1 score was 0.97. In the study conducted by Mahmud et al. [26] using 610 images (305 COVID-19 and 305 Healthy) and Transfer Learning with CNN (Stacked Multi-Resolution CovXNet) classifier, training and testing procedures were carried out with a fivefold cross-validation approach. At the end of the study, sensitivity was 0.978, and specificity was 0.947. The accuracy was calculated as 0.974, F-1 Score 0.971, and AUC 0.969. Next, in this section, two-class classification studies were examined. These were conducted in 2020, focusing on the distinctions of COVID-19 Pneumonia/Viral Pneumonia, COVID-19 Pneumonia/Bacterial Pneumonia, and COVID-19 Pneumonia/Other Pneumonia. In the study conducted by Mahmud et al. [26] , the classification of COVID-19 Pneumonia/Viral Pneumonia and COVID-19 Pneumonia/Bacterial Pneumonia was also performed. Within the scope of the study, a total of 610 X-ray images (305 COVID-19 and 305 Viral Pneumonia) were used for the differentiation of COVID-19 Pneumonia/Viral Pneumonia. Similarly, for the differentiation of COVID-19 Pneumonia/Bacterial Pneumonia, a total of 610 X-ray images (305 COVID-19 and 305 Bacterial Pneumonia) were used. In the study, where Transfer Learning with CNN (Stacked Multi-Resolution CovXNet) was used as a classifier, training and testing procedures were carried out according to the fivefold cross-validation procedure. As a result of the study, the sensitivity for COVID-19 Pneumonia/Viral Pneumonia classification was calculated as 0.874, specificity 0.855, accuracy 0.873, F-1 Score 0.878, and AUC 0.921. For the COVID-19 Pneumonia/Bacterial Pneumonia classification, sensitivity was 0.944, specificity 0.933, accuracy 0.947, F-1 Score 0.939, and AUC 0.951. Horry et al. [27] proposed Transfer Learning with CNN (VGG-16, VGG-19, Xception, Inception ResNet, Incep-tionV3, NASNetLarge, Resnet 50V2, and DenseNet-121) methods for COVID-19 Pneumonia/Other Pneumonia classification. The study used 329 images (139 COVID-19 and 190 Other Pneumonia). Training and testing procedures were performed according to the 80% Train-20% Test procedure. At the end of the study, sensitivity ranging from 0.86 to 0.89 and F-1 Score values varying between 0.86 and 0.89 were provided. If a general analysis of the number of X-ray images used in the literature studies with detailed descriptions above is made, it is understood that 19 data sets used in 16 studies for COVID-19/Healthy classification contain an average of 304.7 COVID-19 and 903.9 Healthy images. Similarly, 305 COVID-19 and 305 Viral Pneumonia images were used for COVID-19 Pneumonia/Viral Pneumonia classification, 305 COVID-19 and 305 Bacterial Pneumonia images were used for COVID-19 Pneumonia/Bacterial Pneumonia classification, and 139 COVID-19 and 190 Other Pneumonia images were used for COVID-19 Pneumonia/Other Pneumonia classification, respectively. The average number of COVID-19 X-ray images used in the studies is around 300, which is acceptable for initial stage studies. However, it is still important to carry out studies using a much larger number of images to increase the reliability and stability of the results. In addition, the COVID-19 images used in these studies are generally metadata data collected from academic studies by Cohen et al. [28] . The use of real-world images in studies is a basic requirement for testing the proposed methods in the real world. In this context, eight data sets with open access have been combined to maximize the number of images used in the study. In this way, a comprehensive data set, including 3405 COVID-19, 2780 Bacterial Pneumonia, 1493 Viral Pneumonia, and 1989 Healthy images, was created. More than 90% of the images in question are taken directly from the real world. X-ray images used in the study were obtained by combining eight different open-access data sets. Numerical information of X-ray images taken from these datasets is given in Table 1 . In the study, two-class classifications were made under four different titles: COVID-19/Healthy, COVID-19 Pneumonia/Bacterial Pneumonia, COVID-19 Pneumonia/Viral Pneumonia, and COVID-19 Pneumonia/Other Pneumonia. As can be seen in Table 1 , a total of 5303 X-ray images, including 3405 COVID-19 and 1989 Healthy, were used for COVID-19/Healthy classification. In the COVID-19 Pneumonia/Bacterial Pneumonia classification, a total of 6187 X-ray images, including 3405 COVID-19 and 2782 Bacterial Pneumonia, were used. A total of 4898 X-ray images, including 3405 COVID-19 and 1493 Bacterial Pneumonia, were used for COVID-19 Pneumonia/Viral Pneumonia classification. For COVID-19 Pneumonia/Other Pneumonia, the last classification title, a total of 7680 X-ray images, including 3405 COVID-19 and 4275 Other Pneumonia, were used. Other Pneumonia X-ray images in this classification title were obtained by combining Bacterial Pneumonia and Viral Pneumonia X-ray images. More than 90% of the X-ray images ( [31] [32] [33] [34] [35] [36] [37] [38] [39] ) used in the study were taken from the real world. When the number of COVID-19 X-ray images used in the study is compared with the number of images used in the previous studies under the Sect. 1, the number of images used in the study is more than ten times the average number of images used in previous studies. Since images are collected from many datasets within the scope of the study, it is necessary to make the images suitable for use in the study by performing pre-processing. The images used in the study were recorded in four different formats: jpeg, jpg, png, and dicom. The sizes of these images range from 154 px × 124 px to 4064 px × 2992 px (The acronym "px" stands for pixels.). The bit depth of the images is 16, 24, 32, and 48. The images in question were first framed to cover the entire chest area, and unrelated areas were removed from the image. Then, the image was converted to gray-level format with a bit depth of 8-bit. Finally, the image dimensions have been resized to 224 px × 224 px. By applying these pre-processing processes to all X-ray images used in the study, a 9669 gray-level image matrix of 224 × 224 × 1 dimensions was obtained. In addition to the original images, texture feature images obtained using Local Binary Pattern (LBP) and Local Entropy (LE) methods were also used in the study. LBP is a non-parametrial feature extraction method in which the new spatial value of a pixel is compared with neighboring pixels [40] . LE displays the amount of uncertainty or randomness in an image using local histograms [41] . More detailed information about LBP and LE can be examined in the study conducted by Yaşar and Ceylan [42] on the diagnosis of COVID-19 using texture analysis methods on CT lung images. The software used in the study was created using Matlab 2020(b). The radius value used in obtaining LBP feature images was selected as 1. The entropyfilt function was used to obtain the LE feature images. Neighborhood (nhood) parameters were used as default in operations. More detailed information about the parameters used as the default can be examined through MathWorks [43] . In the scope of the study, the images obtained using feature analysis methods were resized to 224 px × 224 px in case of a change in size. A sample selected from each of the COVID-19 Pneumonia, Bacterial Pneumonia, Viral Pneumonia, and Healthy X-ray images was used in the study. The feature images obtained by applying the relevant texture analysis methods to these images are given in Fig. 1. Deep learning is an important sub-field of artificial intelligence. Generally, it generates new data and results from existing data using layers and machine learning. Convolutional Neural Network (CNN) is a deep learning approach to image processing that has been widely used recently. CNN consists of convolution layer, activation function, pooling layer, flattening, and fully connected layer components. The convolution layer is the layer on which convolution processes are performed on the image divided into as many parts as the number of filters. If the processed image is two-dimensional, convolution processes are done in two dimensions. If the images are three-dimensional, the convolution process can be done in three dimensions. In this context, the CNN architectures can be named as 2-D CNN (Two-Dimensional Convolutional Neural Networks) and 3-D CNN (Three-Dimensional Convolutional Neural Networks), respectively. 2-D versions of the CNN architectures of Mobilenetv2 [44] , Resnet101 [45] , and Googlenet [46] were also used within the scope of the study. However, some arrangements need to be made on these architectures to make them suitable for use in the study. The image input dimensions of Mobilenetv2, Resnet101, and Googlenet architectures are 224 × 224 × 3. In the study, the original images and the feature images obtained using LBP and LE were used directly and in combinations as the entrance image. For this reason, the input image dimensions of these architectures have been rearranged. In this context, the data entry sizes of Mobile-netv2, Resnet101, and Googlenet architectures were rearranged as 224 × 224 × 1, 224 × 224 × 2, and 224 × 224 × 3. In addition, the Fully Connected Layer output sizes of these architectures are 1000. As a two-class classification was carried out within the scope of the study, the final Fully Connected Layer output sizes of the architectures were rearranged as 2. Apart from these regulations, modification has been made on the 2-D CNN architectures. In the scope of work, 3-D CNN architecture was also used for COVID-19/Healthy, COVID-19 Pneumonia/Bacterial Pneumonia, COVID-19 Pneumonia/Viral Pneumonia, and COVID-19 Pneumonia/Other Pneumonia classification. The 3-D CNN architecture utilized within the scope of the study is a modified version of the 2-D CNN architecture previously used by Yaşar and Ceylan in studies [17, 42] and translated into 3-D. In this context, the two-dimensional image input layer, convolution layer, and maximum pool layer of the 2-D CNN architecture have been replaced with three-dimensional versions. In this context, the data entry size of the 3-D CNN architecture was set to 224 × 224 × 24 × 1. Also, in 3-D CNN architecture, unlike 2-D CNN architecture, a Dropout Layer is added just before the Fully Connected Layer as a precaution against the overfitting problem. The general working structure of this 24-layer 3-D CNN architecture is shown in Fig. 2 . In this CNN architecture, ReLU Layers and Batch Normalization Layers are used against the Vanishing Gradient problem. Against the overfitting problem, the Dropout Layer is used in the last part of the architecture. The training and testing procedures used in the study were carried out according to the fivefold cross-validation principle. No initial weight transfer was performed during the training. That is, all CNN architectures are trained from the beginning with randomly assigned weights. The software used in the study was created using Matlab 2020(b). Solver for training networks used in CNN training was chosen as stochastic gradient descent with momentum (sgdm). The maximum number of epochs is selected as 30, which is accepted as the default value by the program. The main reason for this choice is to avoid the overfitting problem. In other words, instead of choosing this value as high, an early stop was made using the default value. The size of the mini-batch is set to 64. As a precaution against overfitting problems, options for data shuffling are determined as every epoch. For the study, the initial learning rate was chosen as 0.01, which is accepted as the default value by the program. All other parameters are set as default values. It is possible to examine the descriptions of the parameters and the default values through Mathworks [47]. Confusion matrix and comparison parameters calculated using this matrix were used to evaluate the study results. In this context, confusion matrix (TP, FP, TN, and FN), In this study aiming at the early diagnosis of COVID-19 disease using X-ray images, automatic two-class classification has been carried out under four different titles: COVID- 19 In this study, the images were first arranged and standardized. X-ray images were framed to cover the lung area completely. In this way, irrelevant areas in the X-ray images were removed from the images. Then, the images were converted to gray-level format with a bit depth of 8-bit. Finally, the images were resized to 224 px × 224 px. Second, new feature images were obtained from the original images using texture analysis methods. In this context, two different texture analysis methods were used: Local Binary Pattern (LBP) and Local Entropy (LE). By applying these methods, new feature images with 8-bit depth and gray-level format were obtained. If the dimensions of the images obtained using feature analysis methods changed compared to the original images, the dimensions were resized to 224 px × 224 px. Third, classification processes were made using 2-D CNN. As a 2-D CNN classifier, Mobilenetv2, Resnet101, and Googlenet whose details and changes were previously described were used. 2-D CNN input images were created using the original image and texture feature images alone and in combination. In this context, classification results were created for six different input image combinations (Original, LBP, LE, Original + LBP, Original + LE, Original + LBP + LE). Fourth, classification processes were carried out using 3-D CNN. The input images of the 3-D CNN classifier were obtained by combining the Original, LBP, and LE images directly and in combination. The length of the third dimension of 3D input images is determined as 24. In the study, classification processes were carried out for six different input image combinations (Original × 24, LBP × 24, LE × 24, (Original + LBP) × 12, (Original + LE) × 12, (Original + LBP + LE) × 8). Training and testing procedures were carried out according to the fivefold cross-validation. No initial weight transfer was performed during the training. All CNN architectures were trained from the beginning with randomly assigned weights. In all the experiments carried out within the scope of the study, training options as previously detailed were used. Classification times per image were also measured. Measurement times were carried out over CPU time. MAT-LAB 2020(b) software and Intel (R) Xeon (R) CPU E5-2680 2.7 GHz (32 CPUs) computer with a 64 GB RAM were used in the study. All CPU time in the article refers to times in seconds. First, for this study, a COVID-19/Healthy classification was made. A total of 5303 X-ray images were used in the classification, including 3405 COVID-19 and 1989 Healthy. Training and testing procedures were carried out using fivefold cross-validation. Classification results obtained using 2-D Mobilenetv2, 2-D Resnet101, and 2-D Googlenet classifier are given in Tables 2, 3 Table 5 contains the classification results obtained using 24-layer 3-D CNN. The dimensions of the images given as input to 24-layer 3-D CNN are 224 × 224 × 24 × 1. Next, a COVID-19 Pneumonia/Bacterial Pneumonia classification was made. A total of 6187 X-ray images were Table 9 contains the classification results obtained using 24-layer 3-D CNN. The dimensions of the images given as input to 24-layer 3-D CNN are 224 × 224 × 24 × 1. The third classification of COVID-19 Pneumonia/Viral Pneumonia was made in this study. A total of 4898 X-ray images were used in the classification, including 3405 COVID-19 and 1493 Viral Pneumonia. Training and testing procedures were carried out using fivefold cross-validation. Classification results obtained using 2-D Mobilenetv2, 2-D Resnet101, and 2-D Googlenet classifier are given in Tables 10, 11, The last part of this study involved making the COVID-19 Pneumonia/Other Pneumonia classification. A total of 7680 X-ray images were used for classification, including 3405 COVID-19 and 4275 Other Pneumonia. Training and testing procedures were carried out using fivefold cross-validation. Classification results obtained using 2-D Mobilenetv2, 2-D Resnet101, and 2-D Googlenet classifier are given in Tables 14, 15, The results of the experiments performed for the classification of COVID-19/Healthy, COVID-19 Pneumonia/Bacterial Pneumonia, COVID-19 Pneumonia/Viral Pneumonia, and COVID-19 Pneumonia/Other Pneumonia are shared between Table 2 and Table 17 . In this section, the general evaluation of the results obtained has been made. For the COVID-19/Healthy classification, the comparison of the best results obtained using the feature images directly or in combination with the original image is summarized in Table 18 . Table 18 shows that the most successful results are obtained using 24-layer 3-D CNN. 2-D Resnet101, 2-D Mobilenetv2, and 2-D Googlenet architectures follow the 24-layer 3-D CNN architecture in terms of success, Table 18 are provided in Table 19 for the COVID-19 Pneumonia/Bacterial Pneumonia classification, Table 20 for the COVID-19 Pneumonia/Viral Pneumonia classification, and Table 21 for the COVID-19 Pneumonia/Other Pneumonia classification, respectively. When Tables 19, 20, 21 are examined, it is seen that the CNN architectural success order given for Table 18 is valid, although there are slight differences. In addition, it is understood that the most successful results were obtained using Original + LBP, Original + LE, Original + LBP + LE input image combinations, as well as using the original images directly. When comparing the run times, it is seen that 24-layer 3-D CNN, 2-D Resnet101, 2-D Mobilenetv2, and 2-D Googlenet are in order of costliest to the least costly. The comparison of the results obtained for the classification of COVID-19/Healthy within the scope of the study with the previous studies in the literature is given in Table 22 . The comparison of the results obtained for the classification of COVID-19 Pneumonia/Bacterial Pneumonia, COVID-19 Pneumonia/Viral Pneumonia, and COVID-19 Pneumonia/Other Pneumonia with previous studies in the literature can be seen in Table 23 . Detailed results have been obtained within the scope of the study, which recommends the automatic classification of X-ray chest images for early diagnosis of COVID-19. In this study, automatic two-class classification was carried out under four different titles: COVID-19/Healthy, COVID-19 Pneumonia/Bacterial Pneumonia, COVID-19 Pneumonia/Viral Pneumonia, and COVID-19 Pneumonia/ Other Pneumonia. The texture analysis methods Local Binary Pattern, Local Entropy (LE), 2-D CNN, and 3-D CNN classifier were used in the study. When the summary results in Table 18 and Table 21 are examined, it is seen that the diversification of CNN input images with texture analysis methods has a positive effect on the study results. In addition, these improvements in the study results occurred for all CNN classifiers used in the study. These improvements are due to the enrichment of the entrance images using the texture feature images in addition to the original images. Because the input images contain more information, the success achieved in classification processes is increasing. The 2-D CNN architectures used in the study are generally parallel or hybrid (a combination of series and parallel), not series. This situation prevents an increase in working times at the same rate despite the increase in the third color space dimensions of the input images. In other words, the enriching use of the entrance images has a very limited effect on working times. If making a general comparison between texture analysis methods within the scope of the study, using the original images in combination with the images obtained using LE and LBP (Original + LBP, Original + LE, Original + LBP + LE) is more effective in terms of increasing the results of the study. When the CNN classifiers are compared, it is understood that the 24-layer 3-D CNN classifier is more successful than the other 2-D CNN classifiers. However, due to the three-dimensional convolution processes and the increase in input image sizes, the result generation time of the 24-layer 3-D CNN classifier is the slowest among CNN classifiers. Table 22 , which compares the results obtained within the study for the COVID-19/Healthy classification and the results obtained in the previous studies, shows that important results were obtained in this study. The sensitivity parameter obtained within the study is higher than 14 of the 16 studies in which this parameter was calculated. Also, the specificity parameter is better than 9 of the 13 studies. When a comparison is made between the accuracy parameters, it is seen that the result obtained within the study is higher than 13 of the 15 studies in which this parameter was calculated. In addition, the F-1 Score parameter is higher than all ten studies in which this parameter was calculated, and the AUC parameter is higher than all four studies in which this parameter was calculated. In this study, the results obtained for the COVID-19/ Healthy classification fell behind the results obtained in the studies carried out by Tuncer et al. [11] , Loey et al. [22] , and Mohammed et al. [14] . However, to make a more detailed comparison, the number of images used in the studies and the methods used in training and testing processes should also be considered. The number of COVID-19 X-ray images used in the study was 40 times the number of images used in the study by Tuncer et al. [11] , 50 times the number of images used in the study by Loey et al. [22] , and 135 times the number of images used in the study by Mohammed et al. [14] . The fact that the number of COVID-19 X-ray images used in the study is quite high compared to other studies indicates that the results obtained are more stable. In addition, the CNN architectures used in the study carried out by Loey et al. [22] were also used as a classifier in this study. Another issue is that k-fold cross-validation was performed within the study. However, the training-testing procedure used in the study by Loey et al. [22] is different from the k-fold cross-validation. Table 23 compares the results obtained within the study for the COVID-19 Pneumonia/Bacterial Pneumonia, COVID-19 Pneumonia/Viral Pneumonia, and COVID-19 Pneumonia/Other Pneumonia classification, and the results obtained in the previous studies. The sensitivity, specificity, accuracy, F-1, and AUC results obtained within the scope of the study appear to be higher than the results obtained in the previous studies. It is important to consider that the number of COVID-19 X-ray images used in the study for these classification titles is quite high compared to other studies. In the studies to be carried out after this stage, the aim is to create more 3-D CNN architectures and to perform classification experiments. Another important work topic is to develop new Complex 2-D CNN and Complex 3-D CNN architectures and test their performance. Testing tissue analysis methods other than the tissue analysis method used within this study will help increase the success of the study. These further studies also aim to reveal results for multiclass classification (three-class and four-class). Author contributions HY: investigation, methodology, software, and writing-review and editing; MC: investigation, methodology, writing-review and editing, and supervision. Funding This research did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors. Conflict of interest Dr. Ceylan declares that he has no conflict of interest. Mr. Yasar declares that he has no conflict of interest. Ethical approval This article does not contain any studies with human participants or animals performed by any of the authors. COVID-19 and dengue, co-epidemics in Ecuador and other countries in Latin America: Pushing strained health care systems over the edge Treatment of COVID-19: old tricks for new challenges Clinical characteristics of COVID-19 patients with digestive symptoms in Hubei, China: a descriptive, cross-sectional, multicenter study Direct-RT-qPCR detection of SARS-CoV-2 without RNA extraction as part of a COVID-19 testing strategy: from sample to result in one hour Radiological findings from 81 patients with COVID-19 pneumonia in Wuhan, China: a descriptive study A comparative study on the clinical features of COVID-19 pneumonia to other pneumonias Artificial intelligence and COVID-19: deep learning approaches for diagnosis and treatment An automated Residual Exemplar Local Binary Pattern and iterative ReliefF based corona detection method using lung X-ray image Application of deep learning for fast detection of COVID-19 in X-rays using nCOVnet Automated detection of COVID-19 cases using deep neural networks with X-ray images Benchmarking methodology for selection of optimal COVID-19 diagnostic model based on entropy and TOPSIS methods Convolutional capsnet: A novel artificial neural network approach to detect COVID-19 disease from X-ray images using capsule networks Coronet: A deep neural network for detection and diagnosis of COVID-19 from chest x-ray images A new deep learning pipeline to detect COVID-19 on chest X-ray images using local binary pattern, dual tree complex wavelet transform and convolutional neural networks COVIDgan: Data augmentation using auxiliary classifier gan for improved COVID-19 detection COVID-XNet: A custom deep learning system to diagnose and locate COVID-19 in chest X-ray images Deep learning COVID-19 detection bias: accuracy through artificial intelligence Deep transfer learning with apache spark to detect COVID-19 in chest X-ray images Within the lack of chest COVID-19 X-ray dataset: a novel detection model based on GAN and deep transfer learning Deep-COVID: Predicting COVID-19 from chest x-ray images using deep transfer learning New machine learning method for image-based diagnosis of COVID-19 Performance evaluation of the NASNet convolutional network in the automatic identification of COVID-19 CovXNet: A multidilation convolutional neural network for automatic COVID-19 and other pneumonia detection from chest X-ray images with transferable multi-receptive feature optimization COVID-19 detection through transfer learning using multimodal imaging data COVID-19 image data collection COVID-net: A tailored deep convolutional neural network design for detection of COVID-19 cases from chest x-ray images Data from chest imaging with clinical and genomic correlates representing a rural COVID-19 positive population The cancer imaging archive (TCIA): maintaining and operating a public information repository BIMCV COVID-19+: a large annotated dataset of RX and CT images from COVID-19 patients Identifying medical diagnoses and treatable diseases by image-based deep learning Two public chest X-ray datasets for computer-aided screening of pulmonary diseases A comparative study of texture measures with classification based on featured distributions Image segmentation based on gray level and local relative entropy two dimensional histogram A novel comparative study for detection of COVID-19 on CT lung images using texture analysis, machine learning, and deep learning methods Proceedings of the IEEE conference on computer vision and pattern recognition Deep residual learning for image recognition Going deeper with convolutions Publisher's Note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations