key: cord-0883519-5zmu4r1y authors: nan title: Automated Diagnosis of COVID-19 Using Deep Features and Parameter Free BAT Optimization date: 2021-05-03 journal: IEEE J Transl Eng Health Med DOI: 10.1109/jtehm.2021.3077142 sha: 6e4fd8440a374b2099a088bf2eae9075ce8ac1cd doc_id: 883519 cord_uid: 5zmu4r1y Background: Accurate and fast diagnosis of COVID-19 is very important to manage the medical conditions of affected persons. The task is challenging owing to shortage and ineffectiveness of clinical testing kits. However, the existing problems can be improved by employing computational intelligent techniques on radiological images like CT-Scans (Computed Tomography) of lungs. Extensive research has been reported using deep learning models to diagnose the severity of COVID-19 from CT images. This has undoubtedly minimized the manual involvement in abnormality identification but reported detection accuracy is limited. Methods: The present work proposes an expert model based on deep features and Parameter Free BAT (PF-BAT) optimized Fuzzy K-nearest neighbor (PF-FKNN) classifier to diagnose novel coronavirus. In this proposed model, features are extracted from the fully connected layer of transfer learned MobileNetv2 followed by FKNN training. The hyperparameters of FKNN are fine-tuned using PF-BAT. Results: The experimental results on the benchmark COVID CT scan data reveal that the proposed algorithm attains a validation accuracy of 99.38% which is better than the existing state-of-the-art methods proposed in past. Conclusion: The proposed model will help in timely and accurate identification of the coronavirus at the various phases. Such kind of rapid diagnosis will assist clinicians to manage the healthcare condition of patients well and will help in speedy recovery from the diseases. Clinical and Translational Impact Statement— The proposed automated system can provide accurate and fast detection of COVID-19 signature from lung radiographs. Also, the usage of lighter MobileNetv2 architecture makes it practical for deployment in real-time. An outbreak of coronavirus infection (SARS-CoV-2) emerged in December 2019, and by the beginning of the year 2020, the World Health Organization (WHO) announced it as a global pandemic [1] - [3] . Globally, the confirmed coronavirus cases have reached 166 million by 22, May 2021 [4] and are continuously on an increase (https:// www.worldometers.info/coronavirus/). The medical experts and the researchers are working together for a better understanding of COVID-19 etiology. Novel strategies are being underway that can better control its spread. The conventional testing procedure is based on reverse transcription-polymerase chain reaction (RT-PCR) and nucleic acid sequencing from the virus. Although, RT-PCR being the gold standard, the procedure is time-consuming, needs to be re-iterated, and has considerable false-negative results. In such scenarios, CT-Scans of affected person plays an important role in better management of health condition. The variations like ground-glass opacities and pulmonary consolidation in CT images are an important biomarker for COVID-19 detection which can help in prompt identification of suspicious cases thereby saving crucial time and readily isolating the patient [5] , [6] . Also, the wide accessibility of CT scanners makes this task quicker. Adding to this, Machine learning (ML) and Deep learning (DL) methods are evolving rapidly that can lessen the workload of the medical experts by providing an automatic interpretation from a huge data sets [7] - [10] , [39] . Zhao et al [11] developed a transfer learned DenseNet model for the classification of CT scan images into COVID +ve and Normal categories. The researchers attained an accuracy value of 84.7% and an F1 score of 85.3% on a database encompassing 195 Normal and 275 COVID +ve scans. Kaur and Gandhi [12] introduced a transfer learning-based approach for COVID-19 diagnosis using ResNet50 and MobileNetv2. The system was validated on 250 COVID +ve and 246 normal scans. The authors attained an accuracy of 98.35%. Loey et al. [13] investigated five different transfer learned (DTL) models, i.e., GoogleNet, AlexNet, VGGNet16, VGGNet19, and ResNet50 for COVID-19 classification using the database provided by [11] . The researchers pooled these learned mathematical models with augmentation and Conditional Generative Adversarial Networks (CGAN). The authors established that ResNet 50 is the best, resulting in a test classification accuracy of 82.91%. Soares et al. [14] proposed the eXplainable Deep Learning approach (xDNN) to classify if the subject is infected by SAR-Cov-2 or not. They achieved an F-score value of 97.31% with an accuracy value of 97.38%. Pathak et al [15] proposed a deep bidirectional long short-term memory network with a Mixture Density model (DBM) for classification of CT scans as COVID + ve and normal. The hyperparameters of the DBM model were tuned using the Memetic Adaptive Differential Evolution (MADE)algorithm. The authors attained an accuracy of 98.37% and an F1-score of 98.14% under the train test ratios of 60: 40 over a dataset having 1252 + ve and 1230ve COVID scans. Surveying the literature reveals that the usage of the CT scans in diagnosing COVID-19 is drawing attention due to the inefficiency of the medical testing kits. Also manually inspecting each CT image in the entire volume is tiresome and challenging for the medical experts. As the number of cases are increasing by multi-folds daily, automatic Computer-aided diagnosis (CAD) systems is the need of the hour. Although CAD based on ML and DL are being developed and that have aided in speedy detection of the virus, however, not all the reported works are reproducible as the employed datasets are not available for public usage. Also, the DL requires large annotated training datasets for good detection results. To address this limitation, several works have employed transfer learned models on the benchmark dataset provided by Soares et al. [14] but the diagnosis accuracy is limited [12] , [14] , [15] .Encouraged by the benefit of the employing transfer learning on a small pathological database and to improvise the classification results over existing methods, the present work proposes a system based on the integration of MobineNetv2 architecture with Parameter Free BAT(PF-BAT) optimized Fuzzy KNN(FKNN) classifier for the automated classification of COVID-19 CT scans. Typically, the deep characteristic features are extracted from the fully connected ('new_ fc') layer of the transfer learned MobileNetv2 which are then fed to the FKNN whose hyper-parameters are fine-tuned via the PF-BAT. Summarizing the key contributions are as follows: • A novel diagnosis model is proposed based on the integration of deep features and one of the recent metaheuristic optimization algorithms. • The proposed model overcomes the drawback of the manual hyperparameter tuning via employing the PF-BAT optimization algorithm. • The proposed model attains an accuracy of 99.38% on the COVID-19 CT scan dataset that is better than the existing state of the art models. In this proposed work, we have considered the SARS-CoV-2 CT image database provided by Soares et al. [14] . The database consists of 2D CT images of 60 (28 females and 32 males) SARS-CoV-2 infected patients and images from 60 control subjects. Therefore, in total 2482 images are made available with approx. 1252 +ve and 1230 −ve for the virus [14] . The train-validation splits given in Table 1 , are as per the base paper without an indication whether they were subject independent or not. Example +ve and −ve COVID scans from the dataset are shown in Fig. 1 . This section provides the mathematical background for the blocks used in the proposed system. The structural design of MobileNetv2 is motivated by the MobileNetv1 [16] . It uses depth-wise separable convolutions as the building blocks. In comparison to the V1, it has two new additions: a) depthwise separable convolution, and b) inverted residuals [16] . The MobileNetv2 model is pretrained on ImageNet dataset with 1.4 million images and 1000 classes. The basic building block is shown below as Fig. 2 . Input, output dimensions at the different stages of the structure given in Fig. 2 . is mathematically represented in Table 2 . The general architecture of the model is summarized in Table 3 beginning with convolutional layer (having 32 filters) and subsequently followed by residual bottleneck layers which are 19 in number [16] . Depthwise separable convolution replaces conventional convolution via two processes. The first process works by applying a different convolution to every feature map which is known as feature map-wise convolution. The resultant feature maps are stacked together to be treated by the second process that is pointwise convolution. In the second process, all the feature maps undergo convolution using a kernel of size 1×1. Differing from the conventional convolution that involves the processing of the image across height, width, and channel dimensions simultaneously, the depthwise convolution processes the image via height and width dimensions in the first phase and via channel during the second phase. The computational cost of the convention convolution process(C Conven ) and depth-wise separable convolution(C Depth−seperable ) is mathematically given as: In the above equation p and q are input/output layer index, (h p , w p , d p ) are the height, width, and the count of input feature maps. Also k, d q denotes the kernel size and number of the output feature maps respectively. It is the lesser number of the residual connections between the first and the last feature maps that make MobileNetv2 memory efficient. MobileNetv2 is selected in the present paper as this model is faster than other deep models for the similar level of detection accuracy [16] . In comparison to V1 structure, it employs two times fewer operations, and 30% lesser parameters [16] , [17] . This enables its storage and implementation easier on a simple computing platform making it useful for applications in real-time [12] , [16] . FKNN classifier improves upon the conventional KNN classifier by adding the concept of fuzzy logic to the KNN. FKNN algorithm operates by allocating the membership as a mathematical relation of the exemplar distance vector from its k-neighbors and the corresponding neighbor's memberships in possible classes. For an exemplar vector, the fuzzy memberships corresponding to different classes are assigned according to the following formula [18] : where i = 1, 2 . . . . . . , C and j = 1, 2, . . . . . . , k. Also, C denotes the class count & k as the nearest neighbor number, b specifies the fuzzy strength metric, ||x − x j || as the distance (Euclidean) measure between an exemplar x and its j th closest neighbor, and mb ij specifies the degree of membership of exemplar x j from the training pool to the i th class. Various methods exist for defining mb ij and the popular among them are crisp membership and constrained fuzzy membership. where p j is the number of the exemplars that belong to the class j, in the K nearest training exemplars of x, and η ∈ [0, 1] signifies the bias parameter. Also, for binary class scenario mb ij must satisfy the following eq. (6) [18] where Depending upon the value of mb ij for every class under consideration, a test sample or exemplar is allocated to appropriate class to which it exhibits the maximum membership, i.e., The motivation behind using FKNN is its excellent performance over a wide variety of disease diagnosis problems like Thyroid [19] , Parkinson [20] , [21] and seizure [22] . PF-BAT is an improvement over the BAT algorithm developed by Yang [23] . The conventional BAT algorithm follows certain updation rules that guide them on their foraging behavior as specified below [23] : where x j is position, v j is velocity, t is the time step, freq j is the pulse frequency, B j is the loudness, and r j is the pulse emission rate (corresponding to j th BAT). δ and ε denotes the random numbers in the range [0, 1] and [−1, 1] respectively. Also, α and µ are fixed constants. The exploration and exploitation abilities of the BAT are impecunious [24] . To alleviate this constraint, the variation structure was introduced, motivated by the works in [25] , [26] . The new position update mechanism is given below [27] : Here, pbest j is the best solution previously identified by each bat and x * is the global best solution. As the improvised method removes the velocity update mechanism, so it is called as PFree BAT(PF-BAT). The proposed PF-BAT differs from the existing works by Fister et al. [28] as it eliminates the velocity update equation and doesnot involve any extensive experimental evaluation in determining the appropriate value of the five control parameters. Improved performance of the PF-BAT on standard benchmark mathematical functions and a variety of medical image classification tasks has been already demonstrated in the previous work [27] . Also, the superiority of PF-BAT enhanced FKNN over other machine learning algorithms over a broad variety of classification problems has already been established in our previous works [29] . In the present work, its performance has been investigated over COVID CT scan image dataset. In the present sub-section, the hyper-parameter tuning issue of the FKNN classifier is overcome via using the PF-BAT optimization algorithm as manual tuning is a time-intensive process. The proposed methodology is shown in Fig. 3 . As illustrated, the designed methodology is separated into four phases; 1) data preparation, 2) transfer learning, 3) feature extraction, and 4) classification. In the data preparation part, the database is distributed in 8:2 proportion (i.e., 80% of the whole images are employed for training and rest for validation). For better generalization and prevention of overfitting, data augmentation is employed over training scan images. Augmentation involves the following processes, i.e., random translation in pixel range [−3, 3] , random shear in the range from [−0.05, 0.05], and random rotation in the range [−10, 10]. In the second phase, a pre-trained MobileNetv2 [17] model is used which is fine-tuned on the CT scan images through the concept of transfer learning. It is evident that the first few layers of the pre-trained model comprise only edge and color-associated information, whereas more specific features are present in later layers. Hence, parameters of the initial layers require very less or no fine-tuning [12] , [30] . Based on these observations, we have fine-tuned the last three layers of MobileNetv2 by replacing them with a fully connected ('fc') layer, a softmax layer, and a classification output layer. The size of the 'fc' layer is equal to the number of classes (i.e., 2) in the new classification task [31] , [32] . Arithmetically: Let Model = {MobileNetv2} be the pre-trained architecture. Let (X input , Y output ) be the present CT scan image database; having 'N ' images with the set of labels as Y output = {y|y {Non − COVID; COVID}}. The training and validation pools are represented as (X train , Y train ) and (X val , Y val ). The training data is further allocated into mini-batches (n), such that (X i , Y i ) (X train , Y train ); i = 1, 2, . . . , N /n. Iterative optimization of the pre-trained model, m Model is carried out using 'n' for a specific count of epochs to decrease the loss, L by weight updation as given in eq. (15) . The l denotes the binary cross-entropy loss function and is represented in eq. (16) and m (x, w) is the mathematical function that maps a category 'y' for input feature 'x' and weight 'w'. In eq. (16), M is the class number, 'y' specifies whether target 'c' is the right classification for observation 'o', and 'p' is the probability that observation 'o' belongs to target class 'c'. Resolving, the equations will end up in a learned model. In the feature extraction phase, attributes are extricated from the learned MobileNetv2 model by taking activations onto the last fully connected layer ('new_ fc') of the learned model. In the classification phase, the extracted feature vector along with the corresponding targets is used for FKNN model training. The hyper-parameters of the FKNN model are fine-tuned via PF-BAT optimization using the validation accuracy as the fitness criteria (g) g = (Accuracy) validation The step-by-step procedure for hyper-parameter tuning is elicited in the form a simple illustrative example given as Algorithm 1. Parameter For the validation set, predict the class labels through trained FKNN using the following Update the generation value R4: Generation Number=1 S5: Update position for all bat's using the formula 13 Check if the generation end has reached. If yes go to S 14 else go to S 4 R13: No S14: Obtain the optimal value of hyper-parameters from gbest as k_opt= round(gbest (1)); m_opt=gbest (2)) R14: Example output at the end of the Generations k_opt=3; m_opt=1.0346; S: Specifies Step R: Specifies output Table 1 represents the total number of COVID scans used in training and validation. The scans are resized to 224×224×3, matching the dimensions of the Input layer of the learned model. The weight parameters are tuned via 'Adam' with the initial learning rate as 0.0006 and the transfer learned model is fitted for 40 epochs using a mini-batch size of 180. Moreover, binary cross-entropy is used as a loss function to improve the accuracy of diagnosis. The learned model is executed in MATLAB 19a and implemented on Intel Core i7-4500U CPU, 8 GB RAM, and 1.8 GHz processor. Additionally, metrics such as Recall, Precision, Accuracy, the area under the curve (AUC), and F1-Score are used for the performance evaluation of the proposed diagnosis system. They are mathematically defined as F-score is defined in (21) and AUC is calculated using the technique given in [33] . In the above equation β is taken as 1 We have presented the experimental results obtained using transfer learned MobileNetv2 model and proposed PF-BAT enhanced FKNN model (employing features from the 'new_ fc' layer) for the classification of COVID CT images from Non-COVID images. Firstly, the MobileNetv2 is trained on the COVID CT scan image database and thereafter discriminative features are extracted from the 'new_ fc' layer. The deep features from the 'new_ fc' layer are used to train an FKNN whose hyper-parameters are fine-tuned using the PF-BAT. Fig. 4 shows the training progress and loss curve for the transfer learned MobileNetv2 model for 40 epochs. Table 4 gives the results for learned MobileNetv2, and proposed method. The comparative tabular values show that the proposed PF-BAT enhanced FKNN method using features from the 'new_ fc' layer performs the best by achieving ceiling level validation accuracy of 99.38%, precision of 99.20%, recall of 99.60%, F1-score of 99.40%, and AUC of 99.58%. The FN + FP, i.e., 3, is less in the proposed method than in learned MobileNetv2, i.e., 6 resulting in small miss classification error. Fig. 6 shows the Occlusion Sensitivity visualizations for the proposed optimized FKNN classifier predictions. Visualizing the figure reveals that the proposed technique focuses on the image areas decisive for COVID-19 detection avoiding the false image edges and corners. Clearly, the activations are localized within the lungs. Fig. 7 (a) represents the AUC curve for the proposed model indicating that the model reaches sensitivity vs 1-specificity value close to 1, i.e., 99.58%. The increase in validation performance is attributed to the PF-BAT optimized classifier, where improvised BAT is used to fine tune hyper-parameters to reduce loss. Table 5 gives the optimized value of the nearest neighbor 'k' and fuzzy strength parameter 'b' that improved the validation accuracy to 99.38%. Fig. 7(b) shows the fitness vs the generation/iteration curve. Clearly, a surge in the validation accuracy is seen using the optimal value of the hyper-parameters for the FKNN classifier, i.e., 'k' = 3, and 'b' = 1.0346. Merely, in less than 20 iterations, the accuracy improved from 98.77% to 99.38%. TABLE 6. Comparison of the proposed method with the existing works in literature using the same database reported in [14] . Besides exploring the efficiency of the proposed PF-BAT-based FKNN for COVID detection, a comparative assessment is conducted with the state-of-the-art methods. The comparison has been constrained to the work reported on the same database [14] , [15] , [34] only. As evident from Table 6 , the proposed method yielded an accuracy of 99.38% and F-score 99.40% that is superior to the reported stateof-the-art techniques (accuracy of 97.38%, 97.23%, 98.37%, 98.99% and F-score of 97.31%, 97.89%, 98.14% respectively reported in literatures [14] , [15] , [34] ). Soares et al. [14] have proposed x-DNN model for coronavirus detection, considering attributes from the 'fc' layer of VGG-16. They have also tried to improve the computational complexity with various other models like GoogleNet, Alexnet, Decision Trees, and AdaBoost classifiers. Researcher in [15] have employed the Bi-LSTM-DBM for COVID classification. In this work, MADE is used to fine tune hyper-parameters of the DBM model, where the layer number, node number, clip gradients, learning rate, number of epochs, and batch size are finetuned. Clearly, MADE optimization surged the classification performance than using DBM alone. Differing from the studies reported in [14] and [15] , the proposed PF-BAT enhanced FKNN model offers the following advantages apart from rendering ceiling level of classification performance: 1) It eliminates the need to employ more than one pre-trained deep architectures for COVID classification, i.e., just MobileNetv2 compared to both VGG-16 and x-DNN used in [14] , 2) The dimensionality of the optimization problem is much lower, i.e., just two compared to six reported in [15] . Apart from this, experimentations were also done in using PF-FKNN on features extracted from trained VGG16 and VOLUME 9, 2021 GoogleNet models. For VGG16, the accuracy increased to 95.47% (using optimal value of 'k' and 'b' as 9 and 1.2613) in contrast to the base value of 94.96% reported in [14] . For GoogleNet the accuracy improvised to 94.24% (using optimal value of 'k and 'b as 12 and 6.1158) from the base value of 91.73%. In order to cross validate the performance of the proposed model on another dataset, the COVID19-CT images provided by He et al. [11] , [35] has been used in the given study. In this database, 349 +ve and 397 −ve COVID scans of different sizes are available. Training, test, and validation sets were created by dividing the data in the ratio of 0.6, 0.25, and 0.15 resulting in 191, 234; 98, 105; & 60, 58 +ve and −ve COVID scan images in the respective pools without any prior information (whether they were subject independent or not [11] , [35] ). We have followed the same data split ratio according to the base paper to get fair comparison with existing state of artwork. Table 7 summarizes experimental results on this dataset. The proposed model attains ceiling level of classification accuracy by achieving test accuracy and F1-score of 84.24% of 83.16% respectively. The obtained results are superior to other proposed models (e.g. CRNet [35] , CVR-Net [36] , MNasNet1.0 [37] , and Light CNN [38] , those yielded an accuracy of 73%, 78%, 81.77%, and 83% respectively). Apart from validating the performance of the proposed MobileNetv2+PF-BAT enhanced FKNN on the binary classification task (COVID Vs Non-COVID), the evaluation over larger multiclass data has also been carried out. The experimentations were done on multiclass dataset having 4173 CT scans from 210 patients [14] . The total scans were further divided into 758, 2168, and 1247 2D images with Healthy, COVID, and patients with other pulmonary conditions as labels as per the research conducted by Soares et al [14] . The overall test accuracy from this dataset using the proposed MobileNetv2+PF-BAT enhanced FKNN model is 89%, F1 score is 88%, Sensitivity is 89.41%, Specificity is 94.73%, Precision is 87.03%, and AUC is 96.94% under the train, test, and validation splits of 0.6, 0.2, and 0.2. Employing five-fold cross-validation strategy, the mean test accuracy of 95.99%, F1 score of 95.47%, Sensitivity of 95.54%, Specificity of 97.81%, Precision of 95.43%, and AUC of 97.77% is obtained. The validations over a large multiclass dataset also justify that the proposed approach can achieve a ceiling level of classification performance. In this present work, a novel model is proposed that automatically identifies the COVID +ve signature from the lung radiographs. In the proposed model, transfer learned MobileNetv2 is used as a feature extractor. The discriminative features extracted from the fully connected layer of the learned model are fed to the PF-BAT enhanced FKNN classifier. The hyperparameters of the FKNN have been optimized using the PF-BAT algorithm. Thereafter, the proposed model has been extensively validated on publicly available CT scan image datasets. The analysis on the datasets reveals that with the scheme of hyper-parameter optimization, an increase in the validation accuracy is obtained. The accuracy improves by 0.617%, 1.189%, and 3.851% respectively with optimization. The comparative analysis shows that the proposed model outperforms the existing state-of-the-art models. Consequently, the proposed model can work as a fast automated intelligent tool for assisting healthcare professionals in decision making. A novel coronavirus from patients with pneumonia in China Clinical course and risk factors for mortality of adult inpatients with COVID-19 in Wuhan, China: A retrospective cohort study Clinical features of patients infected with 2019 novel coronavirus in Wuhan, China Radiographic and CT features of viral pneumonia Chest CT findings in coronavirus disease-19 (COVID-19): Relationship to duration of infection Application of machine learning in diagnosis of COVID-19 through X-ray and CT images: A scoping review Rapid AI development cycle for the coronavirus (COVID-19) pandemic: Initial results for automated detection & patient monitoring using deep learning CT image analysis Deep learning-based image conversion of CT reconstruction kernels improves radiomics reproducibility for pulmonary nodules or masses A novel medical diagnosis model for COVID-19 infection detection based on deep features and Bayesian optimization COVID-CTdataset: A CT scan dataset about COVID-19 Automated diagnosis of COVID-19 from CT scans based on concatenation of Mobilenetv2 and ResNet50 features A deep transfer learning model with classical data augmentation and CGAN to detect COVID-19 from chest CT radiography digital images SARS-CoV-2 CT-scan dataset: A large dataset of real patients CT scans for SARS-CoV-2 identification,'' medRxiv Deep bidirectional classification model for COVID-19 disease infected patients MobileNetV2: Inverted residuals and linear bottlenecks Efficient convolutional neural networks for mobile vision applications A fuzzy K-nearest neighbor algorithm Design of an enhanced fuzzy K-nearest neighbor classifier based computer aided diagnostic system for thyroid disease Improving Parkinson's disease identification through evolutionary-based feature selection An efficient diagnosis system for detection of Parkinson's disease using fuzzy K-nearest neighbor approach Complex-valued distribution entropy and its application for seizure detection A new metaheuristic bat-inspired algorithm,'' in Nature Inspired Cooperative Strategies for Optimization A new modification approach on bat algorithm for solving optimization problems Hybrid bacterial foraging with parameter free PSO Hybrid particle swarm optimization algorithm with fine tuning operators A novel feature selection method for brain tumor MR image classification based on the Fisher criterion and parameter-free bat optimization Towards the development of a parameter-free bat algorithm An adaptive fuzzy K-nearest neighbor approach for MR brain tumor image classification using parameter free bat optimization algorithm How transferable are features in deep neural networks?'' in Proc Handwritten devanagari character classification using deep learning Pathological brain detection based on AlexNet and transfer learning ROC graphs: Notes and practical considerations for researchers COVID-19 detection in CT images with deep learning: A voting-based scheme and cross-datasets analysis Sample-efficient deep learning for COVID-19 diagnosis based on CT scans CVR-Net: A deep convolutional neural network for coronavirus recognition from chest radiography images COVID19 detection from radiographs: Is deep learning able to handle the crisis? A light CNN for detecting COVID-19 from CT scans of the chest Deep transfer learning-based automated detection of COVID-19 from lung CT scan slices The authors would like to thank IIT Delhi HPC facility for computational resources, especially for comparison and running the model over larger multiclass data.