key: cord-0899724-uqd30v7b authors: Shorfuzzaman, Mohammad; Hossain, M. Shamim title: MetaCOVID: A Siamese neural network framework with contrastive loss for n-shot diagnosis of COVID-19 patients date: 2020-10-17 journal: Pattern Recognit DOI: 10.1016/j.patcog.2020.107700 sha: 09631f1943efec16d8962ca0c042c63a40855fde doc_id: 899724 cord_uid: uqd30v7b Various AI functionalities such as pattern recognition and prediction can effectively be used to diagnose (recognize) and predict coronavirus disease 2019 (COVID-19) infections and propose timely response (remedial action) to minimize the spread and impact of the virus. Motivated by this, an AI system based on deep meta learning has been proposed in this research to accelerate analysis of chest X-ray (CXR) images in automatic detection of COVID-19 cases. We present a synergistic approach to integrate contrastive learning with a fine-tuned pre-trained ConvNet encoder to capture unbiased feature representations and leverage a Siamese network for final classification of COVID-19 cases. We validate the effectiveness of our proposed model using two publicly available datasets comprising images from normal, COVID-19 and other pneumonia infected categories. Our model achieves 95.6% accuracy and AUC of 0.97 in diagnosing COVID-19 from CXR images even with a limited number of training samples. The coronavirus disease 2019 initially identified in December 2019 in the city of Wuhan in China has rapidly spread throughout the world within a very short period of time resulting in an ongoing pandemic. Since the outbreak it has affected over two hundred countries and territories across the globe with more than 14.5 million cases reported. Fig. 1 depicts the global trend of COVID-19 as of July 20, 2020 including total number of confirmed, active, death, and recovered cases [1] . The outbreak was declared as Public Health Emergency of International Concern (PHEIC) by the World Health Organization (WHO) [2] on January 30, 2020. The virus is very contagious and primarily transmitted between people through close contact. To stop its rapid transmission, it is crucial to gain a good understanding of the genetic characteristics of the virus. The genome size of the virus varies from approximately 26 to 32 kilobases being one of the largest among single stranded RNA (ribonucleic acid) viruses. The average diameter of the particle of the virus is around 120 nm [3] . Various common symptoms are found in the infected patients such as cough, fever, shortness of breath, fatigue, loss of smell, and pneumonia. The complications include pneumonia, acute respiratory distress syndrome, and other infections. Precise and on time diagnosis are being hampered due to undiscovered treatment, scarcity of resources, and harsh conditions of laboratory environment. This has increased the challenges to curb the spread of the virus. Accurate and speedy identification of suspected patients at the early phase may possibly play a critical role in timely quarantine and progressive cure. Thus, swift identification of potential infection by coronavirus is incredibly crucial for timely control of epidemic and public health welfare. Identification of coronavirus infection is primarily done by nucleic acid test also called a PCR (polymerase chain reaction) test which examines for the existence of antibodies for an infection. However, results from recent studies show that this type of pathogenic laboratory testing though being a diagnostic gold standard suffers from limitations since it is timeconsuming and produces high false negative cases [5] . Furthermore, deploying COVID-19 tests at a large scale is very expensive and is not affordable by many developing and underdeveloped countries. Hence, development of Artificial Intelligence (AI) based diagnosis and testing methods will be very beneficial. Variety of such AI based medical applications include cancer detection [15] , diabetic retinopathy diagnosis [16] , multi-modal skin lesion classification [17] , polyp detection [23] and so on. In favor of this, researchers are taking global initiatives to use AI as a potentially powerful tool to come up with cost-effective and fast diagnostic procedures to control the ongoing epidemic [32] . The key research goals include COVID-19 transmission, its early diagnosis, development of effective treatment, and understanding its socio-economic impact [4] . Computer aided diagnosis (CAD) systems capable of processing chest X-ray (CXR) images and computed tomography (CT) scans along with state-of-the-art deep learning techniques could be very beneficial for the health professionals in diagnosing COVID-19 cases. Some studies in the literature have already demonstrated the effectiveness of using various deep learning techniques to identify positive COVID-19 cases from chest X-ray (CXR) images [27] and computed tomography (CT) scans [24] and to monitor the disease progress over time. Since deep learning algorithms generally require huge amount of training data to produce effective prediction results, the existing methods trained on limited training samples (due to the lack of large COVID-19 public dataset availability) are likely to suffer from model generalizability to new data. To alleviate this problem of data scarcity, researchers adopted various techniques such as data augmentation and generative adversarial network (GAN) [6, 7] . Nevertheless, these techniques are highly dependent on the appropriate selection of parameters. Various hand-tuned data augmentation technique suffers from over-fitting problem [6] whereas techniques related to generating images through GAN face challenges in emulating real patient data which leads to unanticipated bias during model testing [7] . Furthermore, some studies have adopted transfer learning technique by using various pre-trained CNN models. Since these CNN models are pretrained on a large non-medical dataset (i.e., ImageNet), substantial amount of fine-tuning which generally requires longer training period is necessary to produce promising diagnostic results. Lately, n-shot (specially, one-shot and few-shot) learning has gained immense popularity in research community for analyzing medical images with a limited sample size. In general terms, for example, one-shot learning refers to the task of classifying an image to a particular class given a single (or few) training samples of each class. Specifically, one or more samples from each image class are used to prepare (train) the model which in turn can classify unseen images in future. One of the meta learning models that recently gained success in implementing few-shot learning (especially one-shot) in various domains is Siamese network. In a Siamese network architecture, identical deep convolutional neural networks (CNNs) are trained to extract feature vectors discriminating between samples of each image class which are then contrasted to verify the similarity of the input images. This paper presents a trainable n-shot deep meta learning framework to classify COVID-19 cases with limited training CXR images. We use a fine-tuned CNN model called VGG16 [8] as backbone encoder network to generate feature embeddings from the input images and leverage pairwise contrastive loss function to adjust the network weights. More specifically, we have used CXR images from two public datasets to pre-train the embedding CNN network to generate feature representations that are used by the Siamese network which learns a metric space for n-shot classification of unseen images without any retraining. In summary, following are the contributions of our work: (a) A meta learning framework called MetaCOVID based on Siamese neural network is presented for diagnosis of COVID-19 patients from chest X-ray images, (b) The proposed work focuses on the benefit of using contrastive loss and n-shot learning in framework design, (c) A fine-tuned pre-trained VGG encoder is used to capture unbiased feature representations to improve feature embeddings from the input images, (d) The COVID-19 diagnosis problem is formulated as a k-way, n-shot classification problem where k and n represent the number of class labels and data samples used for model training, (e) Performance evaluation is presented to demonstrate the efficacy of the proposed framework with a limited dataset. The rest of the paper is organized as follows. Related work is presented in Section 2. Section 3 and 4 present method, dataset, and experiments with performance results. Finally, Section 5 concludes the paper with future work. Relevant to the proposed research, our literature study will largely contain existing research effort in the area of COVID-19 diagnosis using AI techniques. Deep learning which is a specialized form of machine learning in the domain of AI has shown great potential in medical image analysis during the last decade [9] . Substantial research has been conducted using deep learning in various medical fields such as disease prediction, diagnosis of pulmonary nodules, and classification of benign and malignant tumors and so on. According a recent study from the researchers at UN Global Pulse [10] , it is shown that AI applications can be as accurate as humans in detecting COVID-19 and offer faster and cheaper solutions in diagnosing the virus than standard test kits thus saving radiologists' valuable time. As part of this, researchers are primarily concentrating on techniques based on statistical learning for the detection of potential coronavirus infection from CXR images and computed tomography (CT scans). Some research initiatives in progress are provided below. A relatively earlier (in the beginning of the outbreak) effort done by a group of researchers in Renmin University of Wuhan, China [11] proposed an AI model for the diagnosis of COVID-19 cases using CT scans. The model uses UNet++ [12] architecture for coronavirus detection using CT scan features and makes use of more than 40,000 images from 106 patients for model training. Experimental results demonstrate that the radiologists' efforts in terms of time can substantially be decreased by using this model. Xu et al. [13] presented a deep learning model to screen coronavirus disease from viral pneumonia (of type Influenza-A) and normal cases with pulmonary CT scans. They have first Since it is relatively easier to find CXR images than CT scans especially in rural areas, they can be a viable alternative to CT images. Wang and Wong [18] proposed an AI system called COVID-Net to diagnose COVID-19 from chest X-ray images containing samples from healthy, COVID-19 and other pneumonia infected patients. The limitation of this study is that the authors trained and tested their model using an imbalanced dataset which contains very few (less than 100) COVID-19 images as opposed to about 16,000 images from healthy and other non-COVID pneumonia patients. Additionally, authors in [19] presented a similar study that uses CXR images to detect coronavirus infection through transfer learning mechanism. Chakraborty [20] also developed a CXR based model using deep neural network that can achieve significant performance improvement even when the size of the dataset is limited. Nevertheless, the model lacks generalizability and needs fine-tuning to produce more stable results. Another laudable effort made by researchers from Delft Imaging project [21] which developed an AI model for diagnosing COVID-19 from CXR images. Their model is called CAD4COVID which is built upon an existing AI model previously developed for diagnosing tuberculosis. It triages COVID-19 suspected patients. Hossain et al. [38] presented a healthcare framework based on 5G network that makes use of CXR and CT images for COVID-19. Abbas et al. [22] have also leveraged deep learning techniques for the diagnosis of COVID-19 from CXR images using CNN and other pre-trained models such as ResNet [25] . Besides, Karim et al. [26] have made contribution in interpreting their predictions by extracting critical features related to COVID-19 to gain deeper understanding. Although a recent report [28] has shown the success of a number of Chinese hospitals in deploying AI-assisted radiology technologies in combating COVID-19, radiologists have shown their concern [29] that the shortage of available data to train the AI diagnostic models is a major challenge. This is substantiated by the fact that large body of the AI models in the literature have used datasets containing limited COVID-19 samples. To address the challenge of limited training samples, He et al. [30] proposed a framework based on deep neural network that is able to attain significant improvement in accuracy in COVID-19 detection with a limited dataset consisting of CT scans. They have developed a synergistic approach to combine transfer learning with self-supervised contrastive learning to extract unbiased features to avoid overfitting problem. The experimental results demonstrate the superiority of their approach in comparison with several state-of-the-art models. In a subsequent effort, Chen et al. [31] develop a meta learning model with prototypical network to predict coronavirus infections from chest CT images that requires a small dataset for training. They have used momentum based contrastive learning to extract feature vectors form the input images that are used by the prototypical network to make predictions on unseen CT images for potential COVID-19 encounters. Validation results with two publicly available CT scans datasets suggest the effectiveness of their model compared to several other relevant methods. Existing techniques in the literature that use CXR images for COVID-19 diagnosis mostly use custom CNN architecture or pre-trained transfer learning models which require large training data to produce rich feature encoding. In contrast, we have proposed an end-to-end trainable nshot deep meta learning framework based on Siamese neural network to classify COVID-19 cases with limited training CXR images. Our proposed model is computationally efficient that can achieve better or the same level of performance as the pre-trained and other custom CNN models that require longer training time. Also, techniques in the literature mostly use de-facto categorical or binary cross entropy loss function. In contrast, we have used contrastive loss function which results in faster model convergence with fewer experiments and hyperparameter updates. Furthermore, most of the existing models use image augmentation to improve model generalizability even with prolonged training time. Alternatively, the proposed meta learning framework shows faster model convergence and greater generalizability by using contrastive loss function and appropriate hyper-parameter optimization such as learning schedule and regularization through dropout technique used in various layers in the model. The shortage of COVID-19 CXR images brings substantial challenges in developing tools for effectively diagnosing COVID-19 cases using deep learning based techniques. To tackle this challenge, we leverage deep meta learning and devise the diagnosis problem of COVID-19 from CXR images as a n-shot classification problem which can be considered as an instantiation of meta learning in the paradigm of supervised learning. Meta learning has recently emerged as a trending research area in the field of Artificial Intelligence (AI) and is believed to be a steppingstone for attaining Artificial General Intelligence (AGI). It is often referred to as "learning to learn" and is capable of learning new skills and generalizing to new tasks quickly by means of limited training samples. The aim of n-short learning is to make classification of unseen data given only a limited The problem of n-shot image classification can be defined as K-way N-shot episodic task where K denotes the number of target class labels in the dataset and N denotes the number of available images (samples) for each of the classes. Given, we have a dataset, D, we sample N data points (images) from each of the K classes present in our data set and we call it as support set. Similarly, we sample Q different images from each of the classes and call it as query set. The goal is to classify the images of the query set based on K classes and KN total images available in the support set. Figure 3 demonstrates an n-shot classification scenario in visual form. In our case, we have three different types CXR images in our dataset, namely, normal, COVID-19 positive, and non-COVID pneumonia cases. Hence, we view the diagnosis problem both as 3-class (images from all three categories) and 2-class (normal and COVID-19 positive images) classification problem. In this context, our COVID-19 diagnosis problem can be identified as three-way N-shot and two-way N-shot learning problem. This section presents the generic high-level architecture of our deep Siamese network (as shown in Fig. 4) for n-shot learning to diagnose COVID-19 cases which is widely known as metric learning-based approach to meta learning. We will describe all major components of the architecture including base CNN encoder for feature embedding, contrastive learning, and training strategies. In the metric-based meta learning setting, the goal is to learn the appropriate metric space. The fundamental concept is mostly related to nearest neighbors techniques and kernel density estimation where the predicted probability is calculated across a set of given output labels, y, as a weighted sum of labels of the examples from the support set. A kernel function, f ϴ , is used to create the weight which essentially estimates the similarity between two different input data points. The predicted probability over the samples of a support set, S, can be formulated as: The performance of a metric-based meta learning model largely depends on learning an appropriate kernel. A useful metric would represent the relationship among the inputs such as similarity in the latent space to facilitate highly accurate predictions. In our case, we want to learn the similarity between two images. For this purpose, a convolutional neural network is used to extract the features from two images and finds the similarity by computing the distance between features of these two images. This approach is widely used in metric-based learning algorithms such as Siamese networks that we have used in our meta learning framework besides other metric-based networks such as matching, relation and prototypical networks. As shown in the preceding diagram, a Siamese network contains two identical parallel networks both sharing the same weights and architecture where each of the networks accepts a different input image and the output from them are combined to make the final prediction. More specifically, the goal is to have two identical base neural networks that take an actual image and another candidate image as input and can learn a function to produce the similarity output between these two images. The concern is that how we can essentially train such a neural network encoder that can learn this similarity function. Ideally, a convolutional neural network could be used without any constraints. Apparently, it would be desirable to use a custom CNN model which is smaller and computationally efficient and can achieve the same level of performance in feature encoding as the pre-trained models. However, such a CNN model requires large training data to produce rich feature encoding. Since we have a limited dataset, we leverage the power of CNN models pre-trained on large ImageNet [39] data which in recent times have shown promising results in solving computer vision problems such as medical imaging. Hence, we have used a fine-tuned pre-trained VGG-16 [8] as base encoder to obtain feature embeddings from the input images to ultimately compute similarity among them. Let's consider that we have two input images, x 1 and x 2 . After passing the image first image, x 1 , through the top encoder, we receive a feature embedding of x 1 denoted as z 1 (x 1 ) = VGG16 (x 1 ) where z 1 (x 1 ) is the output generated from the average pooling layer. Similarly, the second image, x 2 is fed to the identical bottom encoder sharing the same weights, w, to get a different feature embedding of x 2 denoted as z 2 (x 2 ). Then, in the latent space of feature embeddings, we feed these two embeddings to an energy function, E, which will give us the similarity between the two inputs. We use L 1 component wise distance as our energy function which can be expressed as follows: ( 1 , 2 ) = ( 1 , 2 ) = ‖ 1 ( 1 , 2 ) − 2 ( 1 , 2 )‖ (2) The value of E will be smaller if the input images (x 1 and x 2 ) are similar and vice versa. In reality, if it is less than a supplied threshold value the images are similar and, if not, they are different. Finally, this distance value can be incorporated in loss function Through the use of an appropriate loss function, the base encoder network can learn parameters to obtain a better encoding of the input image. Since Siamese networks make binary classification by classifying if the input images are similar or not, using binary cross-entropy loss function would be a natural choice. However, we have also considered using contrastive loss function due to the nature of classification strategy adopted by Siamese networks based on the similarity of pairs of input images. Hence, we have used both the loss functions in this study for performance evaluation. Originally proposed by Hadsell et al. [33] , the contrastive loss function requires pairs of input samples as opposed to individual samples. The idea is that the base encoder In the preceding equation, the value of y is the true label, which will be 0 when the two input images are similar and 1 if they are dissimilar, and is the distance measure between feature embeddings of the input images. Now, y equals to 0 implies that the amount of loss contributed by similar pairs would be simplified to the first term only and is minimized. On the contrary if y=1then the loss will be simplified to the second term and is maximized to m, a hyperparameter called margin. Thus, when input pairs are dissimilar, and if their distance is greater than the margin, they do not incur a loss (as shown in Fig. 5 ) Binary cross entropy loss also called log loss is used to estimate the performance of a classifier with an output probability ranging from 0 to 1. The loss value will increase if the predicted probability deviates from the true label. This can be formulated as follows where y and p represent the class label and prediction probability, respectively: This loss function is used to train the network so that it can differentiate between similar and dissimilar images if we provide one training example from positive and negative categories and aggregate both the losses as below: Formally, we have in our hand a k-way, n-shot classification problem for classifying CXR images. We have our dataset, D, with a training (D train ) and a test (D test ) split. Now, the training set will contain n samples form each of k classes thus totaling k.n examples in D train and D test consists of several samples for evaluation. We train our model in an episodic fashion which dictates that in each episode, we sample a few data points from our meta-training dataset, D train , prepare our support set and query set, and train on the support set and test on the query set. So, over series of episodes, our model will learn how to learn from a smaller dataset. In this way, our model will learn to solve an unseen task by gaining experience through a series of training tasks. An example n-shot classification training scenario is given in Fig. 6 where we create episodes consisting of tasks each of which is defined by a support set and a query set containing sample images from meta-training dataset. Moreover, each of these tasks is similar to the test n-shot classification task containing images for support and query sets. Finally, the detailed training strategy of our model is shown in Algorithm 1. To demonstrate the effectiveness of our n-shot meta learning approach to COVID-19 diagnosis and inspect the effects of using different k-way, n-shot variants and loss functions, we extensively evaluate our proposed Siamese network model with two publicly available CXR datasets. Doctors frequently use CXR and CT scans for the diagnosis of various common diseases such as pneumonia, cancers, lung inflammation, and internal organ injuries. Given the fact that CXR imaging machines are available in nearly all hospitals we decide to use CXR images instead of using CT scans or other types of image data. In the subsequent sections, we will present the datasets used with preprocessing, experimental settings, and results with discussion. In this study, we use COVID- 19 Update parameter ϴ b with new weight, w end for end for a final balanced dataset containing a total of 678 CXR. We divide the dataset for generating Support set and Query set based on various values of n in n-shot classification task. To pretrain the base VGG-16 encoder to be used in generating feature embeddings on the input images to the Siamese network, we use the same dataset consisting of CXR images as mentioned above and split the dataset into training and validation sets with a ratio of 0.7:0.3. Table 1 shows the statistics of the dataset split. Due to the fact that the images in the dataset were collected from different locations with various clinical settings, the intensity and quality of images vary considerably. Nevertheless, we avoid extensive pre-processing of our CXR images in the dataset to gain improved generalization ability of our proposed Siamese network model. This in turn makes our model further robust to artifacts and noises present in the images while extracting feature embeddings from the input images. Thus, we only used few standard pre-processing tasks including image resizing, normalization, and histogram-equalization to optimize the model training method. The size of the CXR images in the dataset varies from 365 × 465 to 1125 × 859 pixels. Hence, we re-scale all images to a size of 100 × 100 pixels to get a consistent image dimension for the entire dataset. Additionally, we perform intensity normalization also called scaling which is an important pre-processing task to expedite model convergence by eliminating feature biases and attaining a uniform distribution for the dataset. We convert the image pixel values from [0, 255] to [0, 1] to obtain a standard normal distribution by using min-max normalization technique. Finally, we apply histogram equalization on the input images in all three RGB channels to improve image contrast. This is usually done by effectively stretching out the most often used intensity values which allows the areas with poorer local contrast to achieve a better contrast. The pre-trained base encoder network and the proposed Siamese network models are implemented using TensorFlow. To evaluate the efficacy of our proposed meta-learning model in diagnosing COVID-19 cases we took the following approach. We start by evaluating our model for various 3-way, n-shot learning settings with both contrastive and cross-entropy losses. A comparison of performance results is also done with different pre-trained CNN models such as Inception [36] , Xception [37] , Inception ResNet v2 [36] , VGG-16 [8] . Finally, we present the results obtained from 2-class (normal, COVID-19) variant of our classification problem with different n-shot learning settings. Since our method is based on multi-shot learning, we are interested in investigating how does model performance change with the number of shots. As such, we carry a number of experiments to observe the relationship between the number of shots and the performance. We also produce performance results (as shown in Table 3 ) with similar learning settings but with binary cross-entropy loss which is usually deemed to be a natural choice for classification problem. Generally, the performance results obtained with contrastive loss function seem to be better than the results obtained with cross-entropy loss function. This is due to the fact that our Siamese network model works based on similarity of pairs of images and contrastive loss function is reported in the literature to be more effective than crossentropy loss. Moreover, as shown in Fig. 7 Finally, to evaluate the effectiveness of our model in identifying COVID-19 cases from normal CXR images only (2-class problem) we perform experiments with similar settings in 3-class problem. Table 5 shows the performance results with contrastive loss and various nshot settings in classifying healthy and COVID-19 patients. As expected, the model shows better performance in all metrics for diagnosing COVID-19 patients in 2-class scenario. This study is one step towards better understanding of the dynamics of COVID-19 pandemic and proposing a state-of-the-art AI based solution for efficient and fast diagnosis system for COVID-19 infections which is the need of the time. The proposed research aims to achieve this through the integration of a meta learning network model with contrastive loss and pre-trained CNN encoder. Specifically, we use a fine-tuned pre-trained VGG-16 network encoder to capture unbiased feature representations that are robust to overfitting and leverage a Siamese network for final classification of COVID-19 cases. We show that our proposed model with contrastive loss and various n-shot learning settings offer a highly accurate yet practical solution for automatically diagnosing COVID-19 cases to accelerate line of treatment for patients. Our best model with 10-shot learning setting achieves an accuracy of 95.6% in diagnosing COVID-19 cases with impressive values of sensitivity (96.0%) and specificity (98.0%) which are deemed to be very critical performance estimates for applications in medical settings. Furthermore, our proposed model exhibits comparable or in some cases better performance than the studied fine-tuned pre-trained CNN models. This is promising due to the fact that our meta learning model is trained only with a limited sample (10) of training examples from each category of CXR images. Simultaneously, it is essential to pinpoint some of the shortcomings of this work which can possibly be tackled in future research. The major drawback is the inadequate interpretability of our model since effective diagnosis requires that results obtained from such interpretability study should be clinically verified by an expert radiologist. As an immediate future work, we plan to extend our work by producing qualitative results with the aid of a model interpretation tool to gain deeper understanding of what our model is learning from the input data during training and validation. We also plan to better tackle COVID-19 diagnosis problem as a multi-modal data fusion problem where various types of clinical data such as patient vitals, location, and population density will be used in addition to image data. COVID-19 Dashboard, CoronaBoard Statement on the second meeting of the International Health Regulations (2005) Emergency Committee regarding the outbreak of novel coronavirus (2019-nCoV)". World Health Organization COVID-19 infection: Origin, transmission, and characteristics of human coronaviruses B5G and Explainable Deep Learning Assisted Healthcare Vertical at the Edge: COVID-I9 Perspective A deep learning algorithm using CT images to screen for corona virus disease (COVID-19) Data augmentation using learned transformations for one-shot medical image segmentation Brain tumor segmentation using convolutional neural networks in MRI images Very deep convolutional networks for large-scale image recognition Deep learning and medical diagnosis Mapping the landscape of Artificial Intelligence applications against COVID-19 Deep learning-based model for detecting 2019 novel coronavirus pneumonia on high-resolution computed tomography: A prospective study. medRxiv Unet++: A nested U-net architecture for medical image segmentation Deep learning system to screen coronavirus disease 2019 pneumonia, Engineering Going deeper with convolutions Detection and classification of cancer in whole slide breast histopathology images using deep convolutional networks Exudate segmentation using fully convolutional neural networks and inception modules Multi-label classification of multi-modality skin lesion via hyper-connected convolutional neural network COVID-Net: A tailored deep convolutional neural network design for detection of COVID-19 cases from chest radiography images COVID-19: Automatic detection from X-ray images utilizing transfer learning with convolutional neural networks An attempt-Detection of COVID-19 presence from chest X-ray scans using CNN & class activation maps Delft Imaging, CAD4COVID: Triage for COVID-19 using artificial intelligence on chest X-rays Classification of COVID-19 in chest X-ray images using DeTraC deep convolutional neural network Polyp detection during colonoscopy using a regression-based convolutional neural network with a tracker AI augmentation of radiologist performance in distinguishing covid-19 from pneumonia of other etiology on chest CT Deep residual learning for image recognition Deep COVID explainer: Explainable COVID-19 predictions based on chest X-ray images Coronet: a deep neural network for detection and diagnosis of covid-19 from chest x-ray images Artificial Intelligence Assisted Radiology Technologies Aid COVID-19 Fight in China Debate flares over using AI to detect COVID-19 in lung scans Sample-efficient deep learning for COVID-19 diagnosis based on CT scans Momentum contrastive learning for few-shot COVID-19 diagnosis from chest CT images COVID-19 Networking Demand: An Auction-based Mechanism for Automated Selection of Edge Computing Services Dimensionality reduction by learning an invariant mapping COVID-19 image data collection Inception-v4, Inception-ResNet and the impact of residual connections on learning Xception: Deep learning with depthwise separable convolutions Explainable AI and mass surveillance systembased healthcare framework to combat COVID-I9 like pandemics ImageNet: A large-scale hierarchical image database The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.