key: cord-0025675-a17c5z9e authors: Thiagarajan, Jayaraman J.; Thopalli, Kowshik; Rajan, Deepta; Turaga, Pavan title: Training calibration-based counterfactual explainers for deep learning models in medical image analysis date: 2022-01-12 journal: Sci Rep DOI: 10.1038/s41598-021-04529-5 sha: 9e2e5a060e2488274dafa4022bdd28f569081ebf doc_id: 25675 cord_uid: a17c5z9e The rapid adoption of artificial intelligence methods in healthcare is coupled with the critical need for techniques to rigorously introspect models and thereby ensure that they behave reliably. This has led to the design of explainable AI techniques that uncover the relationships between discernible data signatures and model predictions. In this context, counterfactual explanations that synthesize small, interpretable changes to a given query while producing desired changes in model predictions have become popular. This under-constrained, inverse problem is vulnerable to introducing irrelevant feature manipulations, particularly when the model’s predictions are not well-calibrated. Hence, in this paper, we propose the TraCE (training calibration-based explainers) technique, which utilizes a novel uncertainty-based interval calibration strategy for reliably synthesizing counterfactuals. Given the wide-spread adoption of machine-learned solutions in radiology, our study focuses on deep models used for identifying anomalies in chest X-ray images. Using rigorous empirical studies, we demonstrate the superiority of TraCE explanations over several state-of-the-art baseline approaches, in terms of several widely adopted evaluation metrics. Our findings show that TraCE can be used to obtain a holistic understanding of deep models by enabling progressive exploration of decision boundaries, to detect shortcuts, and to infer relationships between patient attributes and disease severity. There is a growing interest in adopting artificial intelligence (AI) methods for critical decision-making, from diagnosing diseases to prescribing treatments and allocating resources, in healthcare [1] [2] [3] . However, in order to trust AI systems and to prioritize patient safety, it is imperative to ensure those methods are both accurate and reliable 4 . Examples of unreliable AI systems include a model that can produce highly confident predictions for patients presenting anomalies not seen in the training data, or a model accumulating evidence for a certain diagnosis based on uninformative regions in an image 5, 6 . This has strongly motivated the need to both reliably assess a model's confidence in its predictions [7] [8] [9] , and to enable rigorous introspection of its behavior 4,10-12 . To this end, uncertainty estimation methods are being adopted to determine the deficiencies of a model and/or the training data 13 . Meaningful uncertainties can play a crucial role in supporting practical objectives that range from assessing regimes of over (or under)-confidence and active data collection, to ultimately improving the predictive models themselves 14 . However, in practice, uncertainties are known to be challenging to communicate to decision-makers 15 , and the robustness of decisions with respect to uncertainties can vary considerably between use-cases 16 . Consequently, it could sometimes be more beneficial to implicitly leverage uncertainties when performing introspective analysis of machine-learned models. Model introspection approaches that attempt to explain the input-output relationships inferred by models are routinely used to understand and promote trust in AI solutions. While there has been a large body of work on building inherently explainable models (e.g., rule based systems), post-hoc explanation methods have become the modus operandi with modern deep learning systems 17 . In particular, local explanation methods are very popular as they provide a convenient way for users to introspect by generating local explanations specific to a given input (e.g., health record of a patient)-elucidate what features in the input data maximally support the prediction. Broadly, local explanation methods can be categorized into approximation and example-based approaches. The former class of methods begin by sampling in the vicinity of a query example and fit an explainer to the chosen set of samples (e.g., fit a linear model in LIME 18 or extract rules in ANCHORS 19 ). In contrast, example-based methods synthesize data samples in the vicinity of a query, such that the predictions for those samples align with a user-specified hypothesis. The data samples from the latter approach are referred to as counterfactual explanations 20, 21 . While counterfactual explanations provide more flexibility over feature importance estimation methods, user-studies have also demonstrated that counterfactuals can elicit meaningful insights into the data 22 . While it is common to utilize counterfactuals for causal reasoning, in the recent years, they have been found to be effective for scenario exploration even with predictive models 23, 24 . In its most generic form, for a given query x , one can pose counterfactual generation based on a predictive model F : X → Y as an optimization problem: where x is a counterfactual explanation for the query x (e.g., a medical image of a patient) and ȳ is the userspecified hypothesis about x (e.g., a certain diagnosis). Minimizing a suitable discrepancy d(., .) between x and x ensures that the underlying semantic content of x is preserved in the counterfactual (i.e., vicinity). Another important requirement to produce meaningful counterfactuals is that the generated x should lie close to the original data manifold M(X). When no tractable priors exist for M(X), it is common to perform this optimization in the latent space of a pre-trained generative model (e.g., variational autoencoders (VAE) 25 or generative adversarial networks (GAN) 26 ). Despite the effectiveness of such priors, when the model's predictions F(x) are poorly calibrated, i.e., prediction confidences are not indicative of the actual likelihood of correctness 9,27 , the optimization in Eq. (1) can still lead to bad quality explanations. Though different variants of the formulation in Eq. (1) have been considered in the literature 20 , the fundamental challenge with uncalibrated predictions still persists. We propose to circumvent this challenge by integrating prediction uncertainties into the counterfactual generation process. Proposed work. In this work, we propose TraCE (Training Calibration-based Explainers), an introspection method for deep medical imaging models, that effectively leverages uncertainties to generate meaningful counterfactual explanations for clinical image predictors. As illustrated in Fig. 1 , our framework is comprised of three key components: (1) an auto-encoding convolutional neural network to construct a low-dimensional, continuous latent space for the training data; (2) a predictive model that takes as input the latent representations and outputs the desired target attribute (e.g., diagnosis state, age etc.) along with its prediction uncertainty; and (3) a counterfactual optimization strategy that uses an uncertainty-based calibration objective to reliably elucidate the intricate relationships between image signatures and the target attribute. While our approach is flexible to support the use of any uncertainty estimator or prediction models that use explicit regularization to produce well-calibrated predictions, TraCE builds upon the recent Learn-by-Calibrating (LbC) technique 28 to obtain prediction intervals for both classification and regression settings. LbC jointly trains an auxiliary interval estimator alongside the predictor model using an interval calibration objective, and has been shown to be effective at recovering complex function mappings in scientific datasets. We first adapt LbC for multi-class classification problems and subsequently propose a counterfactual generation approach based on the estimated prediction intervals. When compared to interpretability techniques that provide saliency maps or feature importance scores to explain a specific decision 29 , TraCE enables progressive transition between different output states (e.g., normal → abnormal) through appropriate image manipulations and more importantly, allows optimization with both categorical-and continuous-valued target variables. Our key contributions can be summarized as follows: (1) a new calibration-based optimization approach that takes into account prediction uncertainties; (2) a generalized version of the recent Learn-by-Calibrating technique for classification settings; (3) novel analysis based on TraCE for progressive exploration of decision Figure 1 . An overview of TraCE applied for introspective analysis of chest X-ray (CXR)-based predictive models. In this example, we consider a binary classifier that has been trained to distinguish between normal and abnormal subjects (i.e., containing pneumonia-related anomalies). Since TraCE carries out the optimization in the latent space of a pre-trained auto-encoder model, we first transform a query image x (from the normal class) into its latent representation Z using the Encoder. Subsequently, we invoke the proposed calibration-driven optimization to obtain the counterfactual x in the latent space, such that the semantic discrepancy between z and z is minimized and the classifier's prediction changes to abnormal. Note that, the classifier is trained to output the probabilities for each of the classes along with the prediction intervals. Finally, the synthesized counterfactual z is transformed into the image-space ( x ) using the Decoder network. Results. While recent advances in ML such as deep learning have produced disruptive innovations in many fields, radiology is a prominent example. Conventionally, trained physicians visually assess medical images for characterization and monitoring of diseases. However, AI methods have been showed to be effective at automatically identifying complex signatures in imaging data and providing quantitative assessments of radiographic characteristics 30 . Motivated by this wide-spread adoption of AI tools in radiology [31] [32] [33] , our study focuses on detecting anomalies in chest X-ray (CXR) images. More specifically, we use images from the publicly available RSNA pneumonia detection challenge database (https:// www. kaggle. com/c/ rsna-pneum onia-detec tion-chall enge) in order to demonstrate the effectiveness of TraCE in performing introspective analysis. Given the diagnosis task of categorizing Chest X-ray (CXR) images into normal and abnormal groups (i.e., pneumonia-related anomalies), one can adopt a variety of ML solutions including deep neural networks to build classifiers. However, in practice, purely data-driven AI solutions can learn unintended shortcuts 34 (e.g., superficial correlations) instead of meaningful decision rules. Such models typically perform well on the observed data, including passing standard cross-validation tests, yet fail when deployed in the real-world 5, 6 . Hence, the foremost utility of TraCE is in validating that a predictive model has learned generalizable decision rules. To this end, we use TraCE to progressively generate counterfactuals with different levels of likelihood in assigning a patient to the abnormal group. From our results, we find that using TraCE consistently produces highly meaningful counterfactual evidences, wherein severity of abnormality (e.g., pneumonia) is characterized primarily by changes in the lung opacity. More importantly, the image manipulations are highly concentrated in the chest region of the subject, thus showing that the models did not pick shortcut inductive biases (e.g., scanner-specific features or background image pixels). Using rigorous empirical studies, we also show that TraCE outperforms existing baseline methods, in terms of several widely adopted evaluation metrics in counterfactual reasoning. Furthermore, we find that TraCE can effectively detect shortcuts (or unintended biases) in trained models and infer relationships between different attributes (for example, age and diagnosis state), thus enabling a holistic understanding of deep clinical models. Uncertainty estimation. The growing interest in employing machine learning (ML) based solutions to design diagnostic tools and to gain new insights into a host of medical conditions strongly emphasizes the need for a rigorous characterization of ML algorithms. In conventional statistics, uncertainty quantification (UQ) provides this characterization by studying the impact of different error sources on the prediction [35] [36] [37] . Consequently, several recent efforts have proposed to utilize prediction uncertainties in deep models to shed light onto when and how much to trust the predictions [38] [39] [40] . Some of the most popular uncertainty estimation methods today include: (1) Bayesian neural networks 37,41 : (2) methods that use the discrepancy between different models as a proxy for uncertainty, such as deep ensembles 42 and Monte-Carlo dropout that approximates Bayesian posteriors on the weight-space of a model 38 ; and (3) approaches that use a single model to estimate uncertainties, such as orthonormal certificates 43 , deterministic uncertainty quantification 44 , distance awareness 45 , depth uncertainty 46 , direct epistemic uncertainty prediction 47 and accuracy versus uncertainty calibration 48 . Prediction calibration. It has been reported in several studies that deep predictive models need not be inherently well-calibrated 27 , i.e., the confidences of a model in its predictions are not correlated to its accuracy. While uncertainties can be directly leveraged for a variety of downstream tasks including out-of-distribution detection and sequential sample selection, they have also been utilized for guiding models to produce well-calibrated predictions. In practice, these requirements are incorporated as regularization strategies to systematically adjust the predictions during training, most often leading to better performing models. For example, uncertainties from Monte-Carlo dropout 49 and direct error prediction 50 have been used to perform confidence calibration in deep classifiers. Similarly, the recently proposed Learn-by-Calibrating (LbC) approach 28 introduced an interval calibration objective based on uncertainty estimates for training deep regression models. Counterfactual generation in predictive models. Counterfactual (CF) explanations 20 that synthesize small, interpretable changes to a given image while producing desired changes in model predictions to support user-specified hypotheses (e.g., progressive change in predictions) have recently become popular. An important requirement to produce meaningful counterfactuals is to produce discernible local perturbations (for easy interpretability) while being realistic (close to the underlying data manifold). Consequently, existing approaches rely extensively on pre-trained generative models to synthesize plausible counterfactuals 20, 21, [51] [52] [53] . While the proposed TraCE framework also utilizes a pre-trained generative model, it fundamentally differs from existing approaches by employing uncertainty-based calibration for counterfactual optimization. Data. Our analysis uses CXR images available as public benchmark data for the tasks of predicting the diagnostic state and other patient attributes. In particular, our study uses the RSNA pneumonia detection challenge database, which is a collection of 30,000 CXR exams belonging to the NIH CXR14 benchmark dataset 54 , of which 15,000 exams show evidence for lung opacities related to pneumonia, consolidation and infiltration, and 7500 exams contain no findings (referred as normal). The CXR images in the dataset were annotated by six board-certified radiologists and additional information on the data curation process can be found in Ref. 55 . In addition to the diagnostic labels, this dataset contains age and gender information of the subjects. Note that, for www.nature.com/scientificreports/ this analysis, we used healthy control subjects from the RSNA pneumonia dataset to define the normal group and designed predictive models to discriminate them from patients presenting pneumonia-related anomalies in their CXR scans. We refer to the latter as the abnormal group. Evaluation metrics. We used the following metrics for a holistic evaluation of the counterfactual explanations obtained using TraCE and other baseline methods. (1) Validity For categorical attributes (as in classification problems), this metric measures the ratio of the counterfactuals that actually have the desired target attribute to the total number of counterfactuals generated (higher the better). In the case of continuous-valued attributes we measure the mean absolute percentage error (MAPE) between the desired and achieved target values (lower the better). where I denotes the identity function that returns 1 when the arguments match, and N is the total number of query samples used for evaluation. (2) Confidence In cases of categorical-valued targets (class labels), we compute the confidence P(ȳ i |x i ; F) (from softmax probabilities) of assigning the desired class ȳ i for a counterfactual x i (higher the better). (3) Sparsity Since we perform optimization directly in the latent space, measuring the amount of change in the images is a popular metric in the literature. We compute the sparsity metric as the ratio of the number of pixels altered to the total number of pixels. and T denotes the total number of pixels in the query x . In general, sparser changes to an image are more likely to preserve the inherent characteristics of the query image. (4) Proximity Recent works have considered the actionability of modified features by grounding them in the training data distribution. Following 56 , we measure the average ℓ 2 distance of each counterfactual to the K-nearest training samples in the latent space (lower the better) where N K (x; X) denotes the set of K nearest neighbors of the counterfactual x from the training data X, and E denotes the encoder network (see "Methods" section for details) to compute the latent representation. (5) Realism score We also employ this metric from the generative modeling literature 57 to evaluate the quality of images obtained using TraCE. While standard metrics such as the FID (Frechét Inception Distance) score or the precision/recall metrics are used to evaluate the overall quality of a population of generated images, they are not sufficient to assess individual images. Hence, we utilize the realism score introduced in Ref. 58 , which is high when the generated image is close to the true data manifold and decreases as the image moves further from the manifold. Denoting the feature vectors for the set of real images (used for training), obtained using a pre-trained classifier such as VGG-16, by the matrix r = [ψ 1 r , . . . , ψ N r ] and a generated image by ψ g , the realism score can be computed as follows: where ψ K r refers to the feature vector corresponding to the Kth nearest neighbor (w.r.t to r ) from the set N K (ψ j r ; r ). Given the rapid adoption of AI solutions in diagnosis and prognosis, it is critical to gain insights into black-box predictive models. In this study, we analyzed a predictive model that classifies CXR images into normal and abnormal groups, and used TraCE to synthesize counterfactuals for a given query image from the normal class to visualize the progression of disease severity. Such an analysis can reveal what image signatures are introduced by a predictive model to provide evidence for the abnormal class, and can be used by practitioners to verify if the model relies on meaningful decision rules or shortcuts (e.g., changes to the background) that cannot generalize. In our implementation of TraCE, we first constructed a low-dimensional latent space (100 dimensions) for the dataset of CXR images using a Wasserstein auto-encoder 59 . We subsequently learned the predictive model F D along with the interval estimator G D , using a modified version of the LbC algorithm (details in the "Methods" section). The hyper-parameters η 1 and η 2 in Eq. (11) are critical to trade-off between preserving the inherent semantics from query x and achieving the desired prediction. Hence, one can progressively transition from the normal to the abnormal class by fixing η 2 and gradually relaxing η 1 . Figure 2 illustrates the counterfactuals obtained using TraCE for multiple different examples from our benchmark dataset. More specifically, the query samples x correspond to CXR images from the normal class and we varied η 1 between 0.5 and 0.05, while setting η 2 = 0.5 and η 3 = 0.2 . These values were obtained using a standard www.nature.com/scientificreports/ www.nature.com/scientificreports/ By producing physically plausible evidences for crucial hypotheses, TraCE enables practitioners to effectively explore complex decision boundaries learned by deep predictive models. Comparing TraCE to baseline methods. In order to perform a quantitative evaluation of TraCE, we obtained counterfactuals for 500 randomly chosen images from a held-out test set (not used for training) and Table 1 presents a detailed comparison of different baseline methods (see "Methods" section for details). Note that, we chose the hyper-parameters η 1 , η 2 , η 3 such that the average discrepancy in the latent space is similar across all methods. The first striking observation is that, despite using the same pre-trained latent space for counterfactual optimization, all methods that incorporate explicit calibration strategies or uncertainty estimation consistently outperform the Vanilla model. More specifically, for similar levels of discrepancy in the latent space, TraCE achieves a significantly higher validity score of 0.88 as opposed to 0.69 of the Vanilla model, while inducing similar or lower amount of changes to the query (indicated by the sparsity and proximity metrics). Furthermore, our approach outperforms the results obtained with state-of-the-art uncertainty estimators and calibration strategies (in all the metrics), thus demonstrating its efficacy in generating counterfactual explanations. As discussed earlier, TraCE is applicable for predictive models outputting both categorical-and continuousvalued target variables. To demonstrate this, we considered only healthy control subjects from the RSNA dataset and designed a regressor to estimate their age attribute using their CXR images. Though the age prediction task is not necessarily relevant on its own in clinical diagnosis, as we will show next, such attribute estimators can be utilized for inferring relationships to the diagnosis state. For our evaluation, we used 500 randomly chosen test subjects whose age attribute was between 40 and 70 and set the desired value ȳ = 20 . From Table 2 , we notice that the proposed approach achieves lower validity (MAPE) scores, without compromising on the proximity metric, when compared to the other baselines. Interestingly, we find that changing the age attribute required the manipulation of much lesser number of pixels (low sparsity values) when compared to the diagnosis state. that they have the risk of inferring decision rules based on shortcuts, thereby limiting their utility in practice. Detecting such shortcuts is essential to both validate model behavior and to detect unintended biases (hospitalspecific or device-specific information) in the training data. In order to demonstrate the use of TraCE in detecting such shortcuts, we synthetically introduced a nuisance feature into images from the abnormal class-overlaid the text PNEUMONIA in the top-left corner of each image, and used TraCE to check if the model's decision was based on this nuisance feature. After training the Wasserstein autoencoder and the LbC model using the altered images, we selected query images from the normal group and generated the corresponding counterfactual evidences for the abnormal group. As illustrated in Fig. 3a- TraCE reveals attribute relationships. Motivated by the effectiveness of TraCE in producing counterfactuals for different types of target attributes, we next explored how counterfactual optimization can be used to study relationships between patient attributes, such as age and gender, and the diagnosis state. Note, this analysis is based on the assumption that the patient attribute can be directly estimated from the CXR images, and the inferred relationship does not necessarily imply causality. First, we study if the image signatures pertinent to the patient age attribute provides additional evidence for diagnosis state prediction. Given the age predictor, along with its interval estimator, (F A , G A ) and the diagnosis predictor (F D , G D ) , we constructed counterfactuals based on two independent hypotheses. Note, both predictors were constructed based on the same low-dimensional latent representations. More specifically, we provided the hypotheses ȳ A = 70 and ȳ D = abnormal for the two cases, and used TraCE to generate counterfactuals x A and x D that adhere to our hypotheses. We then estimated the age-specific and diagnosis-specific signatures introduced by TraCE: In order to check if there exists an apparent relationship between age and diagnosis state, we generated the hybrid counterfactual, Finally, we compared F D (x) − F D (x D ) to quantify if incorporating age-specific features into x D increased the disease severity (i.e., likelihood of being assigned to the abnormal class). An overview of this strategy is illustrated in Fig. 4 . Figure 5 shows the results for eight different normal subjects, wherein we find that there is an apparent increase in P(state = abnormal) when age-specific signatures are incorporated. Using 500 randomly chosen normal subjects, we estimated an average change of 0.09 ± 0.08 in F D (x) − F D (x D ) , thus indicating that the diagnosis predictor is sensitive to age-specific patterns. In practice, if such a relationship is expected, it is a strong validation for the model's behavior. On the other hand, if the attribute is a confounding variable, it becomes critical to retrain the model wherein this sensitivity is explicitly discouraged. Interestingly, when we repeated this analysis with the gender attribute, such a relationship was not apparent (see results in Fig. 6 ). In this section, we discuss in detail the methodology for performing calibration-driven counterfactual generation. All the methods presented were performed in accordance with the relevant guidelines and regulations. Constructing low-dimensional latent spaces. Given a set of samples from an unknown data distribution, our goal is to build a low-dimensional, continuous latent space that respects the true distribution, so that one can generate counterfactual representations in that space. A large class of generative modeling methods exist to construct such a latent space. In this work, we focus on Wasserstein autoencoders 59 since they have been found to outperform other variational autoencoder formulations, particularly in image datasets with low heterogeneity, e.g., scientific images from physics simulations 60 . This network is a composition of an encoder network E that transforms the input x into its latent code z , and a decoder network D that reconstructs the image. Additionally the encoder has the objective of matching the latent distribution of the training samples E P X [E(z | x)] to a pre-specified prior P Z . This helps us to sample from the prior as well as generate new unseen samples from the original data manifold M(X) after training such auto-encoding models. Wasserstein autoencoders thus have to minimize: (1) discrepancy cost D x between the original data distribution and the generated; (2) discrepancy cost D z between the latent distribution of the encoded training samples to that of a prior. Following standard WAE consturction, we employ the mean squared error (MSE) for D x and use the maximum mean discrepancy (MMD) to define D z . As shown in Fig. 7a , we also find that including another loss term to maximize the structural similarity (SSIM) 61 between the original and reconstructed images led to higher quality reconstructions. Training. All images were resized to 224 × 224 pixels, and treated as single channel images. With the latent space dimensionality fixed at 100, the encoder model was comprised of 4 convolutional layers, with the number of filters set to [16, 32, 64, 32] , followed by two fully connected layers with hidden units as 512 and 100. All convolutional layers used the kernel size (3, 3) and stride 2. The decoder consisted of two fully connected layers with 512 and 6272 hidden units followed by 4 transposed convolutional layers with channels [64, 32, 16, 1] respectively. ReLU non-linear activation was applied after every layer except for the last layer. We trained the models using the Adam 63 optimizer for 150 epochs with an initial learning rate of 1e − 3 and decreased it by factors 2, 5, 10 after 30, 50 and 100 epochs respectively. The three loss functions were assigned the weights [1, 0.5, 0] for the first 20 epochs and subsequently changed to [1, 0.1, 1] respectively until convergence. Predictive model design using LbC. While conventional metrics such as cross entropy (for categorical-valued outputs) and mean squared error (for continuous-valued outputs) are commonly used, it has been recently found that interval calibration is effective for obtaining accurate and well-calibrated predictive models 28 . Hence, in TraCE, we adapt the Learn-by-Calibrating approach to train classifier (or regression) models that map from the CXR latent space to a desired target variable. By design, LbC provides prediction intervals in lieu of point estimates for the response y , i.e., [ŷ − δ,ŷ + δ] . Here, δ is used to define the interval. Suppose that the likelihood for the true response y to be contained in the prediction interval is p(ŷ − δ ≤ y ≤ŷ + δ) , the intervals are www.nature.com/scientificreports/ considered to be well-calibrated if the likelihood matches the expected confidence level. For a confidence level α , we expect the interval to contain the true response for 100 × α% of realizations from p(x). Algorithm. The model is comprised of two modules F and G , implemented as neural networks, to produce estimates ŷ = F(z) and δ = G(z) respectively. For example, in the case of multi-class classification settings, ŷ ∈ R K is a vector of predicted logits for the K different classes. Since interval calibration is defined for continuous-valued targets, we adapt the loss function for training on the logits directly. To this end, we first transform the ground truth labels into logits. Note, for each sample, we allow a small non-zero probability (say 0.01) to all negative classes. As discussed earlier, suppose that the likelihood for the true y[k], k ∈ (1, . . . , K) to be contained in the interval is p(ŷ , the intervals are considered to be well-calibrated if the likelihood matches the confidence level. Denoting the parameters of the models F and G by θ and φ respectively, we use an alternating optimization strategy similar to 28 . In order to update φ , we use the empirical interval calibration error as the objective: where δ i = G(z i ; φ) , and the desired confidence level α (set to 0.9 in our experiments) is an input to the algorithm. When updating the parameters φ , we assume that the estimator F(.; θ) is known and fixed. Now, given the updated φ , we learn the parameters θ using the following hinge-loss objective: Figure 3 . Using TraCE to detect shortcuts in deep predictive models. In this experiment, we synthetically introduced a nuisance feature (overlaid the text PNEUMONIA in the top-left corner) into all images from the abnormal group, and used this data to train the predictive model. Given the entirely data-driven nature of machine-learned solutions, there is risk of inferring a decision rule based on this irrelevant feature in order to discriminate between normal and abnormal groups. (a-d) Here, we used randomly chosen query images from the normal class and generated counterfactuals for the abnormal class. In each case, we show the query image, the counterfactual explanation from TraCE and the absolute difference image between the two; (e, f) Here, we introduced the nuisance feature into CXR images from the abnormal group and synthesized counterfactuals for the normal class. We observe that TraCE can effectively detect such shortcuts-counterfactuals for changing the diagnosis state are predominantly based on manipulating the text on the top-left corner of the query images. www.nature.com/scientificreports/ Figure 4 . Using TraCE to infer relationships between a patient attribute (e.g., age) and disease states. For this analysis, we construct two independent predictive models, i.e., age and diagnosis state, and synthesize counterfactuals based on hypothesis on each of the predictions (e.g., patient age should be predicted as 70 while the diagnosis state should be abnormal. Finally, we combine the changes induced in the two counterfactuals, � A (x) and � D (x) respectively, and check if incorporating age-specific patterns strengthens the evidence for the abnormal class. Figure 5 . (a-h) Explanations generated using TraCE by introducing age-specific attributes into the counterfactuals synthesized for changing the diagnosis state of a normal subject to be abnormal. Interestingly, we find that there exists a correlation between the two attributes, as evidenced by the consistent increase in the likelihood P(state = abnormal) when compared to counterfactuals that rely only on patterns from the diagnosis state predictor. In each case, we highlight the changes � A (x), � D (x), (� A (x) + � D (x)) and display the likelihood P(state = abnormal). www.nature.com/scientificreports/ where ŷ i = F(z i ; θ) and τ is the margin parameter (set to 0.05 in our experiments). Intuitively, for a fixed φ , obtaining improved estimates for ŷ can increase the empirical calibration error in (8) by achieving higher likelihoods even for lower confidence levels. However, in the subsequent step of updating φ , we expect δ 's to become sharper in order to reduce the calibration error. We repeat the two steps (Eqs. 8 and 9) until convergence. Architecture. As showed in Fig. 7b Eq. (1) using the pre-trained predictor and interval estimator models from LbC. Our goal is to generate explanations to support a given hypothesis on the target variable-for example emulating high-confidence disease states given the CXR of a healthy subject. To this end, we first obtain the latent representation for the given query image x using the encoder, z = E(x) . We then use the pre-trained predictor ( F ) and interval estimator ( G ) models to (9) θ * = arg min 59 , and construct a low-dimensional, continuous latent space for CXR images. Note, we used a combination of maximum mean discrepancy (MMD), mean squared error (MSE) and structural similarity (SSIM) losses to train the network parameters; (b) next, we adapt the Learn-by-Calibrating 62 approach to train a classifier that takes as input the latent representation from the encoder and outputs a patient-specific attribute along with prediction intervals. www.nature.com/scientificreports/ generate the counterfactual z . Finally, the generated counterfactuals in the latent space are passed to the decoder to obtain a reconstruction in the image space, x = D(z) . We propose the following optimization to generate the counterfactual explanations: here, ȳ is the desired value for the target attribute (hypothesis), η 1 , η 2 , η 3 are hyper-parameters for weighting the different terms, and . 2 denotes the ℓ 2 norm. The first term ensures that the generated counterfactual is in the vicinity of the query sample x (in the latent space). The second term ensures that the expected target value is contained in the prediction interval (calibration), while the final term penalizes arbitrarily large intervals to avoid trivial solutions. The calibration objective L is implemented as a hinge-loss term: where the margin was fixed at τ = 0.05 . Choosing η 1 , η 2 , η 3 is essential to controlling the discrepancy between z and z , and ensuring that the prediction for the counterfactual is ȳ. Baselines. We considered a suite of baseline approaches for our empirical study and they differ by the strategies used for training the classifier, and counterfactual optimization. In particular, we investigate approaches that produce explicit uncertainty estimators as well as those that directly build well-calibrated predictors. However, note that, all methods perform their optimization in the same latent space. (1) Vanilla In this approach, we train the classifier with no explicit calibration or uncertainty estimation, and use the following formulation to generate the counterfactuals: where L ce denotes the cross entropy loss. (2) Mixup This is a popular augmentation strategy 64 that convexly combines random pairs of images and their labels, in order to temper overconfidence in predictions. Recently, in Ref. 65 , it was found that mixup regularization led to improved calibration in the resulting model. In mixup, the model is trained not only on the training data, but also using samples in the vicinity of each training sample: where x i and x j are randomly chosen samples with labels y i and y j . The parameter , drawn from a symmetric Beta distribution, determines the mixing ratio. Since this approach does not produce any uncertainty estimation, the counterfactual optimization is same as that of the Vanilla approach in Eq. (12). (3) MC dropout In this baseline, we train the classifier with dropout regularization and estimate the (epistemic) prediction uncertainty for any test sample by running multiple forward passes. Finally, we use the following heteroscedastic regression objective to implement uncertainty-based calibration during counterfactual optimization: Note, similar to the proposed approach, here we operate directly on the logits and the mean/variance estimates ( µˆz, σ 2 z ) are obtained using T (set to 5) forward passes with dropout. (4) Deep ensembles Deep ensembles form an important class of uncertainty estimation methods, wherein the model variance is used as a proxy for uncertainties. In this approach, we independently train M different models (with bootstrapping and different model initializations) with the same architecture. Subsequently, for any input sample x , we obtain the mean/variance estimates ( µˆz, σ 2 z ) by aggregating predictions from the M models. Finally, we employ the calibration objective in Eq. (14) to perform counterfactual optimization. While highly accurate and currently one of the best uncertainty estimation techniques, deep ensembles require training multiple models, which can become a computational bottleneck when training deep networks. (5) Uncertainty-Weighted confidence calibration (UWCC) The authors in Ref. 49 proposed to build calibrated classification models by augmenting a confidence-calibration term to the standard cross-entropy loss and weighting the two terms using the uncertainty measured via multiple stochastic inferences. Mathematically, here the first term denotes the cross-entropy loss, and the predictions P(ŷ i |z i ) are inferred using stochastic inferences for z i , while the variance ( α i ) in the predictions is used to balance the loss terms. More specifically, we perform T forward passes with dropout in the network and promote the softmax probabilities to be closer to an uniform distribution, i.e. high uncertainty, when the variance is large. The normalized variance α i is given by the mean of the Bhattacharyya coefficients between each of the T predictions and the mean prediction. Since the model is inherently calibrated during training, we do not measure the (10) z = arg min z η 1 �z −ẑ� 2 2 + η 2 L(ŷ, δ,ȳ) + η 3 δ, whereŷ = F(ẑ), δ = G(ẑ) (11) L(ŷ, δ,ȳ) = max 0, (ŷ − δ) −ȳ + τ + max 0,ȳ − (ŷ + δ) + τ , (12) z = arg min z η 1 �z −ẑ� 2 2 + η 2 L ce F(ẑ),ȳ , www.nature.com/scientificreports/ uncertainties at test time and hence use the optimization in Eq. (12) for generating counterfactuals. For the case of continuous-valued targets (i.e., regression tasks), we utilize the extension in Ref. 66 that performs heteroscedastic calibration of the MC dropout estimator during training. All datasets used in this were obtained from publicly released databases and pre-processed using open-source tool chains. We have added appropriate links to obtain the data as well as access the scripts for pre-processing, wherever applicable. Deep learning for healthcare applications based on physiological signals: A review Machine learning for medical diagnosis: history, state of the art and perspective Deep learning for healthcare: review, opportunities and challenges Opportunities and obstacles for deep learning in biology and medicine Common pitfalls and recommendations for using machine learning to detect and prognosticate for COVID-19 using chest radiographs and CT scans Prediction models for diagnosis and prognosis of COVID-19: Systematic review and critical appraisal On calibration of modern neural networks Leveraging uncertainty information from deep neural networks for disease detection Improving reliability of clinical models using prediction calibration Who wants accurate models? arguing for a different metrics to take classification models seriously What clinicians want: contextualizing explainable machine learning for clinical end use Understanding behavior of clinical models under domain shifts A survey of uncertainty in deep neural networks Building calibrated deep models via uncertainty matching with auxiliary interval predictors Second opinion needed: Communicating uncertainty in medical machine learning The negative consequences of failing to communicate uncertainties during a pandemic: The case of COVID-19 vaccines Explainable AI: Interpreting, Explaining and Visualizing Deep Learning Why should I trust you?": Explaining the predictions of any classifier Anchors: High-precision model-agnostic explanations Counterfactual explanations for machine learning: A review Explanation by progressive exaggeration Counterfactuals in explainable artificial intelligence (XAI): Evidence from human reasoning Gifsplanation via latent shift: A simple autoencoder approach to counterfactual generation for chest X-rays Using deep image priors to generate counterfactual explanations Auto-encoding variational bayes Generative adversarial nets Accurate uncertainties for deep learning using calibrated regression Designing accurate emulators for scientific processes using calibration-driven deep models Grad-cam: Visual explanations from deep networks via gradient-based localization Artificial intelligence in radiology COVID-19 identification in chest X-ray images on flat and hierarchical classification scenarios Identification of COVID-19 samples from chest X-ray images using deep learning: A comparison of transfer learning approaches Som-lwl method for identification of COVID-19 on chest X-rays Shortcut learning in deep neural networks Uncertainty Quantification: Theory, Implementation, and Applications Practical confidence and prediction intervals What uncertainties do we need in Bayesian deep learning for computer vision? Dropout as a Bayesian approximation: Representing model uncertainty in deep learning Simple and scalable predictive uncertainty estimation using deep ensembles Understanding deep neural networks through input uncertainties Weight uncertainty in neural network Simple and scalable predictive uncertainty estimation using deep ensembles Single-model uncertainties for deep learning Uncertainty estimation using a single deep deterministic neural network Simple and principled uncertainty estimation with deterministic deep learning via distance awareness Depth uncertainty in neural networks Direct epistemic uncertainty prediction Improving model calibration with accuracy versus uncertainty optimization Learning for single-shot confidence calibration in deep neural networks through stochastic inferences Accurate and robust feature importance estimation under distribution shifts Interpretable counterfactual explanations guided by prototypes Explanations based on the missing: Towards contrastive explanations with pertinent negatives Counterfactual visual explanations Hospital-scale chest X-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases Pneumonia dataset annotation methods Multi-objective counterfactual explanations. in Parallel Problem Solving from Nature-PPSN XVI Assessing generative models via precision and recall Improved precision and recall metric for assessing generative models Wasserstein auto-encoders Improved surrogates in inertial confinement fusion with manifold and cycle consistencies Image quality assessment: From error visibility to structural similarity Learn-by-calibrating: Using calibration as a training objective A method for stochastic optimization Mixup: Beyond empirical risk minimization On mixup training: Improved calibration and predictive uncertainty for deep neural networks Heteroscedastic calibration of uncertainty estimators in deep learning The software associated with this paper will be hosted through a public code repository (github). J.J.T. conceived the presented idea, and along with K.T. developed the overall formulation. While P.T. helped in setting the high-level objectives, D.R. contributed substantially in technical discussions through the course of this work. J.J.T. and K.T. implemented an initial version of the proposed framework, and with help of D.R., set up the empirical studies. D.R. was instrumental in preparing the dataset and validating the findings. While J.J.T. led the manuscript writing efforts, all the other authors contributed significantly to different sections. The authors declare no competing interests.