key: cord-0822024-ud3j4q3u authors: Mondal, Arnab Kumar title: COVID-19 prognosis using limited chest X-ray images date: 2022-04-25 journal: Appl Soft Comput DOI: 10.1016/j.asoc.2022.108867 sha: 6e6f5d8ada9d9c6b94eba12938c11b33104d1367 doc_id: 822024 cord_uid: ud3j4q3u The COrona VIrus Disease 2019 (COVID-19) pandemic is an ongoing global pandemic that has claimed millions of lives till date. Detecting COVID-19 and isolating affected patients at an early stage is crucial to contain its rapid spread. Although accurate, the primary viral test ‘Reverse Transcription Polymerase Chain Reaction’ (RT-PCR) for COVID-19 diagnosis has an elaborate test kit, and the turnaround time is high. This has motivated the research community to develop CXR based automated COVID-19 diagnostic methodologies. However, COVID-19 being a novel disease, there is no annotated large-scale CXR dataset for this particular disease. To address the issue of limited data, we propose to exploit a large-scale CXR dataset collected in the pre-COVID era and train a deep neural network in a self-supervised fashion to extract CXR specific features. Further, we compute attention maps between the global and the local features of the backbone convolutional network while finetuning using a limited COVID-19 CXR dataset. We empirically demonstrate the effectiveness of the proposed method. We provide a thorough ablation study to understand the effect of each proposed component. Finally, we provide visualizations highlighting the critical patches instrumental to the predictive decision made by our model. These saliency maps are not only a stepping stone towards explainable AI but also aids radiologists in localizing the infected area. The COrona VIrus Disease 2019 (COVID- 19) , epicentered in Hubei Province of the People's Republic of China, spread so rapidly across the globe that the World Health Organization (WHO) declared COVID-19 a Public Health Emergency of International Concern on 30 January 2020, and finally a pandemic on 11 March 2020 [1] . It has caused a massive threat to global health with 174,918,667 cases of confirmed coronavirus and 3,782,490 deaths as of 12 June 2021. Once infected with COVID-19, one may experience fever, cough, and respiratory illness. Some may also experience shortness of breath, muscle or body aches, headache, loss of taste or smell, sore throat, and diarrhea [2, 3] . The virus can cause pneumonia or breathing problems in severe cases, leading to multi-organ failure and death [4] . Due to the exponential growth of COVID-19 patients, there is a shortage in supply of diagnostic kits, a limited number of beds in the hospitals to care for critical patients, a dearth of ventilators, scarcity in personal protective equipment (PPE) for healthcare personnel. Despite various preventive measures (such as complete lockdown) adopted by the government of different countries to contain the disease and delay the spread, several developed countries have faced a critical care crisis, and the health system has come to the verge of collapse. It is, therefore, of utmost importance to screen the positive COVID-19 patients accurately for efficient utilization of limited resources. Reverse Transcription Polymerase Chain Reaction (RT-PCR) [5, 6] is the most preferred viral test for COVID-19 detection due to its high sensitivity and specificity. However, the turn-around time of RT-PCR is high. * Corresponding author anz188380@iitd.ac.in (A.K. Mondal) ORCID(s): 0000-0001-7297-374X (A.K. Mondal) Consequently, chest radiography such as computerized tomography (CT) scan and X-ray imaging-based detection techniques have emerged as an alternative modality for screening COVID-19 patients. With these modalities, researchers have observed that COVID-19 patients' lungs exhibit ground-glass opacity and/or mixed ground-glass opacity and mixed consolidation that can separate COVID-19-positive cases from COVID-19-negative cases [7, 8] . In contrast to conventional diagnostic methods, X-ray offers several advantages as it is fast, can simultaneously analyze numerous cases, inexpensive and widely available. It can be very useful in hospitals with limited testing kits and resources. Deep Machine Learning has revolutionized the field of health care by accurately analyzing, identifying, and classifying patterns in medical images [9] . Artificial neural networks are able to diagnose a variety of illnesses with a high degree of accuracy. The reason for such success is that deep learning techniques do not rely on manual handcrafted features but rather learn features automatically from data itself. This allows the algorithm to be applicable on a broader variety of use cases than traditional machine learning methods and is also faster and more accurate in many cases. Motivated by the remarkable performance of CheXNet in Pneumonia detection from chest X-ray images, artificial intelligence (AI) researchers have put a lot of effort into designing machine learning (ML) algorithms for automated detection of COVID-19 using chest X-rays. However, the biggest challenge lies in the fact that COVID-19 being a novel disease, a limited number of sample images are available for training deep neural networks. Motivated by this, in this work, we propose a novel framework that can be trained using limited labeled data for COVID-19 detection using chest X-rays. In this work, our contributions are as follows. 1. We adopt a self-supervised training methodology to train a CXR feature extractor (a convolutional backbone network) on a large-scale chest X-ray dataset. 2. We design a local-global-attention-based classification network consisting of the pre-trained feature extractor, an attention block, and a classification head. 3. We empirically demonstrate the effectiveness of the proposed framework in the low data regime through extensive experimentation and ablation studies. 4. We present clinically interpretable saliency maps, which are helpful for disease localization and patient triage. The remainder of this paper is structured as follows: Section 2 provides an overview of related work; Section 3 describes the procedural and methodological stages of the development of this solution; Section 4 evaluates the proposed method and assesses the predictions; finally, Section 5 critically discusses the advantages and the limitations of the proposed framework. Several deep neural frameworks [10, 11, 12, 13, 14, 15] have been proposed in the past to identify different thoracic diseases such as Pneumonia using chest X-ray (CXR) images and surpassed average radiologist performance. ChestX-Ray8 [11] (later extended to constitute ChestX-ray14 dataset), and CheXpert [16] are two large-scale datasets of chest X-rays (CXR) that facilitate the training of deep neural networks (DNN) for automating the interpretation of a wide variety of thoracic diseases. ChexNet [16] is a deep neural network, built using DenseNet-121 [17] , for Pneumonia detection using chest X-ray images and it achieved excellent results surpassing average radiologist performance. Chest-Net [12] is another deep neural network for thoracic diseases diagnosis using chest radiography images. The authors in [14] propose to learn channel-wise, element-wise, and scalewise attention (triple attention) simultaneously to classify 14 thoracic diseases using chest radiography. Thorax-Net [15] is an attention regularized deep neural network for the classification of thoracic diseases on chest radiography. Motivated by this, the research community has examined the possibility of COVID-19 prognosis using CXR. The proposal in [18] leverages an enhanced cuckoo search algorithm to determine the most significant CXR features and train a -nearest neighbour (KNN) classifier to distinguish between COVID-19 positive and negative cases. In this work, features were extracted from X-ray images using standard feature extraction techniques such as Fractional Zernike Moments (FrZMs), Wavelet Transform (WT), Gabor Wavelet Transform (GW), and Gray Level Co-Occurrence Matrix (GLCM), followed by a fractional order cuckoo search method where the levy flight distribution was replaced with better suited heavy tailed distributions for selecting the most relevant features. Following feature selection, a KNN was used for classification. The work in [19] employs a new set of descriptors, Fractional Multichannel Exponent Moments (FrMEMs) to extract orthogonal moment features. Next, Manta Ray Foraging Optimization (MRFO) using Differential evolution (DE) is utlized to select the most relevant features. Finally, a -nearest neighbour (KNN) classifier is used for prediction. A novel shapedependent Fibonacci-p patterns-based feature descriptor is proposed in [20] for CXR features extraction, which are classified using conventional ML algorithms such as support vector machine (SVM), -nearest neighbour (KNN), Random Forest, AdaBoost, Gradient Tree Boosting, and Decision Trees. In [21] , the author uses Histogram of Oriented Gradients (HOG), Gray-Level Co-Occurrence Matrix (GLCM), Scale-Invariant Feature Transform (SIFT), and Local Binary Pattern (LBP) methods in the feature extraction phase. Next, Principle Component Analysis (PCA) is applied for feature selection. Finally, k-NN, SVM, Bag of Tree, and Kernel Extreme Learning Machine (K-ELM) are used for final classification. Many of the existing deep learning methods [22, 23, 24, 25, 26, 27, 28, 29] use the transfer learning approach by finetuning pre-trained networks such as ResNet-18 [30] or ResNet-50 [30] , DenseNet-121 [17] , InceptionV3 [31] , Xception [32] , etc., on COVID-19 CXR datasets. COVID-SDNet [28] combines segmentation, data-augmentation and data transformations together with a ResNet-50 [30] for inference. The authors in [28] define a novel three-stage segmentation-classification pipeline to solve a binary classification task between COVID-19 and non-COVID-19 CXR. First, the lung region is cropped from CXR using bounding box segmentation. Next, a GAN based class-inherent transformation network is employed to generate two class inherent transformations + and − from each input image . Finally, the transformed images are used to solve a fourclass classification problem using CNN with a Resnet-50 [30] backbone and an aggregation strategy is designed in order to obtain the final class. As the number of classes increase, so will the number of generators to be trained in stage two, which makes scaling inefficient for multi-class classification. In [24] an ensemble of off the shelf pretrained CNNs -InceptionV3 [31] , MobileNetV2 [33] , ResNet101 [30] , NASNet [34] and Xception [32] -is first fine tuned on the chest xray dataset. Their final layer representations are then stacked and then passed through a MLP for COVID-19 diagnosis. The Xception [32] backbone is used in CoroNet [22] for extracting CXR features which are classified using the MLP classification head. In [35] proposed DeepCoroNet where the CXR images are pre-processed using a sobel filter followed by marker-controlled watershed segmentation and then a deep LSTM network is used for classification. The work in [36] uses Google's Big Transfer models with DenseNet, InceptionV3 and Inception-ResNetV4 models for J o u r n a l P r e -p r o o f Journal Pre-proof COVID-19 Prognosis Using Limited Chest X-ray Images Covid-19 classification using chest X-rays. COVID-Net [37] proposes a custom architecture for CXR-based COVID-19 detection using a human-machine collaborative design strategy. However, limited COVID-19 samples restrict the generalizability of such large-capacity models. To address this issue, MAG-SD [38] employs a multi-scale attentionguided deep network to augment the data and formulates a new regularization term utilizing soft distance between predictions, to regularize the classifier from producing contradicted output for one target. An attention-based teacherstudent framework is proposed in [39] . The teacher network extracts global features and focuses on the infected regions to generate attention maps. An image fusion module transfers the attention knowledge to the student network. CHP-Net [40] involves a discrimination network for lung feature extraction to discriminate COVID-19 cases and a localization network to localize and assign the recognized X-ray images into the left lung, right lung or bipulmonary. In [41] a federated learning model is developed while keeping in mind the privacy of the patients. Individual hospitals or care centres are considered as nodes which have their own datasets and share a common diagnosis model provided by a central server. The individual nodes update the model according to the dataset that they have and their updated weights are averaged and the common server model is updated. In [42] a multimodal system is developed based on data consisting of breathing sounds and chest x-ray images. Sound data is converted to spectrograms and convolutional neural networks are used for analysis for both sound data and chest xray images. An InceptionV3 network is used followed by an MLP for covid-19 diagnosis. The authors in [43] propose a convolutional CapsNet for COVID-19 detection from chest X-ray images in binary as well as multi-class classification settings. xViTCOS [44] propose a vision transformer based deep neural classifier for COVID-19 prognosis. Supervised learning usually demands a large amount of labeled data. However, collecting quality annotated data is expensive, especially for medical applications. Moreover, COVID-19 being a novel disease, there is a scarcity of well-curated high volume datasets. Therefore, we propose to utilize a self-supervised training methodology to address this issue of data scarcity. In the first stage, we train a convolutional neural network on a large-scale CXR dataset, CheXpert [16] for extracting robust CXR features with selfsupervision. Next, we utilize limited COVID-19 CXR images to train a classification network that uses the pretrained backbone to extract local and global features, computes attention maps, and predicts the class label. The fundamental concept behind self-supervised learning is to design some auxiliary pre-text tasks such that the model discovers the underlying structure of the data while solving those tasks. Several state-of-the-art self-supervised methods [45, 46, 47, 48] rely on contrastive strategy to induce similarity between positive pairs (different augmented views of the same image) and dissimilarity between negative pairs (augmented views from different images). These methods, however, require either large batch size, memory bank, or custom mining strategies while selecting negative pairs. Bootstrap Your Own Latent (BYOL) [49] mitigates this issue associated with negative pair selection. In this work, we propose to use BYOL for representation learning. As illustrated in Figure 1 , BYOL consists of two neural networks, viz., online and target networks. These two networks interact and learn together. The online network consists of three sub-networks: an encoder ( ), a projector ( ), and a predictor ( ). denotes the set of trainable parameters of the online network. To break the symmetry between the online and target pipeline, the target network is comprised of two sub-networks: an encoder ( ) and a projector ( ). The parameters, of the target network are slow moving average of the online network parameters, i.e., where ∈ [0, 1] denotes the target decay rate. At the beginning of each training step, an original image, is drawn uniformly from the CheXpert [16] dataset. Next, two sets of randomly chosen transformations, 1 , 2 ∼  are applied on the original image, to obtain two distinct augmented views, 1 = 1 ( ), 2 = 2 ( ) of the underlying true image. During training, 1 is fed into the online network, and 2 is fed into the target network. The online network generates a representation, = ( 1 ), a projection, = ( ), and a prediction, = ( ). The target network produces a target representation, = ( 2 ) and a target projection, = ( ). Since the target network is derived from the online network, the online representations should be predictive of the target representations. Consequently, BYOL is trained to maximize the similarity between these two representations. Mathematically, the online network is trained to minimize the mean squared error between the normalized online prediction, and the normalized target projection, : To make the loss symmetric, next, 1 is passed through the target network and 2 is passed through the online network and loss is computed according to Equation 2. The total loss is now given as, The sub-script, in  implies that, only the online network is updated to minimize  , and the target network is updated as the exponential moving average as indicated in Equation 1. In the second stage of our proposed method, we utilize the pretrained backbone from the previous step and design a spatial attention network based on the local and the global features. Attention mechanism are widely adapted to enhance the performance of deep neural networks on various downstream tasks such as machine translation, text generation in natural language processing and object classification, image captioning, inpainting, etc., in computer vision. Attention in computer vision tasks can broadly be categorized into spatial attention [50, 51] that captures the local context and channel attention [52] that captures the global semantics. Several works [53, 54] consider a combination of both channel-wise and spatial attention. In this work, we adopt the soft trainable visual attention proposed in [55] . Figure 2 presents an overview of the attention mechanism. We extract local and global features using pretrained feature extractor backbone. 'Local features' refer to features extracted by some convolutional layer of the backbone network that have a limited receptive field. In other words the receptive field is a contiguous proper subset of the image (local). The contents of the 'local features' can be more specific to a certain region on the image, while 'global features' use the entire image as their information source. We insert three attention estimators after 'layer2', 'layer3', and 'layer4' (layer names as per PyTorch implementation) to capture coarse-to-fine attention maps at multiple levels. The local features extracted at these three layers together with the global feature at the penultimate 'avgpool' layer produce three attended encodings, which are concatenated and fed into a final classification head. Let,  = { 1 , 2 , ⋯ } denote the set of feature vectors extracted at a given convolutional layer . Where is the vector of output activations at the spatial location of total spatial locations in the layer. The global feature vector, , has the entire input image as the receptive field. Let  be the compatibility function that computes a scalar compatibility score between two vectors of equal dimension. Since the dimensionality of the local features and the global features do not match, we first project the low-dimensional local features to the high-dimensional space of . Next, the compatibility score function is employed to compute the compatibility scores as follows: where ∈ {1, 2, ⋯ , } and is a learnable vector. The compatibility scores are then normalized using the softmax function to compute the attention maps. Journal Pre-proof The attended representations are finally computed as follows: In this work, we concatenate the three representations obtained from three intermediate layers into a single vector, = [ 1 , 2 , 3 ] and feed to a linear classification head. In this section, we describe the dataset used in this work and discuss the experimental results. While some of the works [56] evaluate their proposed algorithm using private datasets, many other works [22, 37, 57] resort to publicly available datasets. In this work, we combine data from several publicly available repositories to create a custom dataset with four classes: Normal, Bacterial Pneumonia, Viral Pneumonia (non-COVID-19), COVID-19. As in [22] , we collected Normal, Bacterial Pneumonia, and non COVID-19 Viral Pneumonia chest X-ray images from the Kaggle repository 'Chest X-Ray Images (Pneumonia)' [58] , which is derived from [59] . Chest X-ray images of COVID-19 patients were obtained from the Kaggle repository 'COVIDx CXR-2' [60] , which is a combination of several publicly available resources [61, 62, 63, 64, 65, 66] . 'COVIDx CXR-2' [60] specifies only train-test split of the dataset. We hold out 20% training examples for automatic model selection based on its performance over the validation set. The validation set in the standard split of 'Chest X-Ray Images (Pneumonia)' [58] dataset contains only 8 images per class. To avoid a huge class imbalance in the validation set, we combine the training and validation examples and In our compiled dataset and the CheXpert [16] dataset, the images are of variable sizes. To address this issue, we resize all the images to a fixed size of 256 × 256. For training BYOL, we randomly choose an image from the CheXpert dataset, select a random patch and resize it to 224 × 224. Next, the image is flipped randomly horizontally with 0.5 probability. Apart from these spatial/geometric transformations of data, we apply appearance transformations on the image. Specifically, we apply a random color distortion transformation consisting of a random sequence of brightness, contrast, saturation, and hue adjustments [67, 68] . As noted in previous work [45] , stronger color jittering helps self-supervised algorithms learn better representation. We utilize PyTorch's standard implementation (torchvision.transforms.ColorJitter) for performing color distortion. Following [45] , we set brightness, contrast and saturation jitter factor uniformly from [0. Table 2 Comparison of performance of the proposed method on chest X-ray dataset against state-of-the-art methods. blur is applied to the patches, and the patches are normalized. We blur the image with 0.5 probability using a Gaussian kernel. We randomly sample ∈ [0.1, 2.0], and the kernel size is set to be 22. In the second stage of training, we randomly choose an image from the compiled CXR dataset, select a random patch and resize it to 224 × 224 with a random horizontal flip. Finally, the patches are normalized before feeding to the classifier. We center crop the image to 224 × 224 and normalize it before passing it to the classification network during inference. We use ResNet-50 [30] pretrained on ImageNet [69] as the online encoder, and the target encoder, . The projector networks, , are multi-layer perceptrons with a hidden layer consisting of 4096 neurons followed by batch normalization, ReLU activation, and an output layer of dimension 256. The predictor network, is architecturally the same as the projector. In our second stage of training, we modify the encoder block architecture to accommodate attention computation and initialize it with the pretrained weights from self-supervised training. We attach three attention estimators after 'layer2', 'layer3', and 'layer4' (layer names as per PyTorch implementation). The local features extracted at these three layers have dimensions (512, 28, 28), (1024, 14, 14) and (2048, 7, 7) respectively using 'channel first' representation. These three local features together with the global feature at the 'avgpool' layer produce three attended encodings. However, the global feature has a shape of (2048, 1), which causes shape incompatibility. To alleviate this issue, we use projector blocks consisting of 1 × 1 2-D convolution operations, which ensures that the channel dimension of local features matches the channel dimension of the global feature. Next, attention maps are computed using a linear combination of a local feature and the global features, 1 × 1 2-D convolution operations and softmax normalization. Finally, these attended embeddings are concatenated and classified using a linear classifier. For self-supervised training, we use a batch size of 256, Adam optimizer with a learning rate of 3 × 10 −4 , and the model is trained for 200 epochs. To train the classifier, we use a batch size of 128, Adam optimizer with an initial learning rate of 1 × 10 −4 with a cosine decay learning rate scheduler. Further, we use a global weight decay parameter of 5 × 10 −7 . For the self-supervised training we use 2 NVIDIA V100 GPU cards with 32GB memory and 5120 CUDA cores in parallel. One epoch approximately takes 1.5 hours to complete execution. For the finetuning stage, we use 1 NVIDIA V100 (32GB 5120 CUDA cores) GPU card and one epoch takes approximately 8 minutes to complete execution. To benchmark the proposed method against other stateof-the-art methodologies, we compute and report class-wise Precision (Positive Prediction Value), Recall (Sensitivity), F1 score, Specificity, Negative Prediction Value (NPV), and overall accuracy along with 95% confidence interval. Table 2 presents our findings. As can be seen from Table 2 , the proposed method achieves the best overall accuracy with best 95% confidence interval. Further the proposed method achieves the best precision for COVID-19 cases meaning the proposed classifier rarely label a COVID-19 negative sample as a positive sample. Moreover, the proposed method achieves the best recall score implying the classifier is able to J o u r n a l P r e -p r o o f Journal Pre-proof COVID-19 Prognosis Using Limited Chest X-ray Images (a) CoroNet [22] (b) COVID-Net [37] (c) Teacher Student Attention [39] (d) MAG-SD [38] (e) Proposed Method find most of the positive samples belonging to the COVID-19 class. The highest F1 score achieved by the proposed method indicates that the proposed method is the most balanced in terms of both precision and recall as compared to the baseline methods. Similarly, the proposed method achieves high specificity and NPV indicating that the false positive rate is low as well. Finally, from Figure 3 , it can be seen that the proposed method achieves best class-wise accuracy. In this section, we examine the impact of different training components proposed in this work. Specifically, we study the effect of pretraining on ImageNet [69] , self-supervised pretraining on CheXpert [16] and the attention map. Table 3 presents the findings. When a ResNet-50 architecture is trained on the COVID-19 CXR dataset from scratch, its performance is the worst. Transfer learning (ResNet-50 [30] pretrained using ImageNet [69] ) improves the model's classification performance. Attention mechanism provides a further boost in performance. Finally, self-supervised pretraining using ChexPert [16] helps the model extract useful CXR specific features and enhance the model's classification accuracy. visualizations one for each of bacterial pneumonia (Fig. 4a ), viral pneumonia (Fig. 4b) and COVID-19 (Fig. 4c ). This work introduces a method for automated COVID-19 prognosis using a limited amount of labeled COVID-19 CXR data. We have empirically demonstrated the effectiveness of the proposed method over existing SOTA methods as measured using various metrics such as precision, recall, F1 score, specificity, and NPV. While the proposed methodology is highly performant, it is not error-free as the CXR findings due to COVID-19 are not exclusive and overlap with other thoracic infections [70] . Therefore, to improve the efficiency of diagnosis and efficient resource utilization, we suggest the proposed method to be used in conjunction with RT-PCR, and first-line treatment may be initiated based on CXR findings while the RT-PCR test report is awaited. Despite, the great success achieved by deep learning models in different machine learning tasks, they are prone to various biases such as selection bias (distribution of training examples is not reflective of their real-world distribution), group attribution bias (tendency to generalize what is true of individuals to an entire group to which they belong) and so on. Therefor, to deploy the proposed method clinically, it is imperative to thoroughly evaluate the model through clinical trials to examine its generalization capabilities and stability. Although, the method proposed in this work is highly performant on multinational dataset (since the datasets used in this study were compiled from several repositories), to further improve the generalization ability of the proposed method, the model needs to be trained on a large, diverse, high-quality dataset. To conclude, preventing the spread of COVID-19 requires early diagnosis. While RT-PCR is highly accurate when the test is conducted appropriately, its turn-around time is high. Therefore, our proposed deep neural framework might be useful to initiate the first line treatment. Further, the proposed method, when used in conjuction with RT-PCR can be thought of as a complimentary diagnosis or a second opinion to ensure efficient utilization of limited resources. In our future work, we intend to extend this work to automate the analysis of infection severity. The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper. WHO updates on COVID-19 COVID-19 symptoms Coronavirus: covid-19 has killed more people than sars and mers combined, despite lower case fatality rate Detection of sars-cov-2 in different types of clinical specimens Detection of 2019 novel coronavirus (2019-ncov) by real-time rt-pcr Chest ct for typical coronavirus disease 2019 (covid-19) pneumonia: relationship to negative rt-pcr testing COVID-19 Prognosis Using Limited Chest X-ray Images Sensitivity of chest ct for covid-19: comparison to rt-pcr Deep learning in medical image analysis CheXNet: Radiologist-Level Pneumonia Detection on Chest X-Rays with Deep Learning Chestx-ray8: Hospital-scale chest x-ray database and benchmarks on weaklysupervised classification and localization of common thorax diseases Chestnet: A deep neural network for classification of thoracic diseases on chest radiography Jointly Learning Convolutional Representations to Compress Radiological Images and Classify Thoracic Diseases in the Compressed Domain Triple attention learning for classification of 14 thoracic diseases using chest radiography Thorax-net: An attention regularized deep neural network for classification of thoracic diseases on chest radiography CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison Densely connected convolutional networks COVID-19 X-ray images classification based on enhanced fractional-order cuckoo search optimizer using heavy-tailed distributions New machine learning method for image-based diagnosis of COVID-19 Automated Detection of COVID-19 Cases on Radiographs using Shape-Dependent Fibonacci-p Patterns A new approach for computer-aided detection of coronavirus (covid-19) from ct and x-ray images using machine learning methods CoroNet: A deep neural network for detection and diagnosis of COVID-19 from chest x-ray images Interpretable artificial intelligence framework for covid-19 screening on chest x-rays InstaCovNet-19: A deep learning classification model for the detection of COVID-19 patients using Chest X-ray Improving the performance of CNN to predict the likelihood of COVID-19 using chest X-ray images with preprocessing algorithms An optimized deep learning architecture for the diagnosis of covid-19 disease based on gravitational search optimization Diagnosis of Coronavirus Disease 2019 Pneumonia by Using Chest Radiography: Value of Artificial Intelligence COVIDGR Dataset and COVID-SDNet Methodology for Predicting COVID-19 Based on Chest X-Ray Images A deep learning approach to detect covid-19 coronavirus with x-ray images Deep residual learning for image recognition Rethinking the inception architecture for computer vision Xception: Deep learning with depthwise separable convolutions Mo-bilenetv2: Inverted residuals and linear bottlenecks Learning transferable architectures for scalable image recognition DeepCoroNet: A deep LSTM approach for automated detection of COVID-19 cases from chest X-ray images Coronavirus disease (covid-19) detection using x-ray images and enhanced densenet COVID-Net: a tailored deep convolutional neural network design for detection of COVID-19 cases from chest X-ray images Multiscale attention guided network for covid-19 diagnosis using chest x-ray images COVID-19 Automatic Diagnosis with Radiographic Imaging: Explainable Attention Transfer Deep Neural Networks Automatically discriminating and localizing COVID-19 from communityacquired pneumonia on chest X-rays Federated learning for covid-19 screening from chest x-ray images A deep-learning based multimodal COVID-19 Prognosis Using Limited Chest X-ray Images system for covid-19 diagnosis using breathing sounds and chest xray images Convolutional capsnet: A novel artificial neural network approach to detect covid-19 disease from xray images using capsule networks xvitcos: Explainable vision transformer based covid-19 screening using radiography A simple framework for contrastive learning of visual representations Big selfsupervised models are strong semi-supervised learners Momentum contrast for unsupervised visual representation learning Contrastive multiview coding Bootstrap your own latent -a new approach to self-supervised learning Show, attend and tell: Neural image caption generation with visual attention Visual7w: Grounded question answering in images Squeeze-and-Excitation Networks SCA-CNN: Spatial and channel-wise attention in convolutional networks for image captioning Proceedings of the European conference on computer vision (ECCV) Learn To Pay Attention ResGNet-C: A graph convolutional neural network for detection of COVID-19 COVID-19 image data collection Chest x-ray images Labeled optical coherence tomography (oct) and chest x-ray images for classification COVIDx CXR-2: Chest x-ray images for the detection of COVID-19 COVID-19 Image Data Collection: Prospective Predictions Are the Future Figure 1 COVID-19 Chest X-ray Dataset Initiative Actualmed COVID-19 Chest X-ray Dataset Initiative Can AI help in screening viral and COVID-19 Pneumonia? RSNA Pneumonia Detection Challenge: Can you build an algorithm that automatically detects potential pneumonia cases? Data from medical imaging data resource center (midrc) -rsna international covid radiology database (RICORD) release 1c -chest x-ray Some improvements on deep convolutional neural network based image classification Going deeper with convolutions ImageNet Large Scale Visual Recognition Challenge ACR recommendations for the use of chest radiography and computed tomography (CT) for suspected COVID-19 infection C-DOT he had the opportunity to participate in cutting-edge projects such as the Dense Wavelength Division Multiplexing (DWDM) and Packet Optical Transport Platform (P-OTP) We thank IIT Delhi HPC facility 1 for computational resources. Arnab Kumar Mondal: Conceptualization, Data curation, Methodology, Formal analysis, Validation, Software, Roles/Writing -original draft, Writing -review & editing.