ACE: Explaining cluster from an adversarial perspective ACE: Explaining cluster from an adversarial perspective Yang Young Lu 1 Timothy C. Yu 2 Giancarlo Bonora 1 William Stafford Noble 1 3 Abstract A common workflow in single-cell RNA-seq anal- ysis is to project the data to a latent space, cluster the cells in that space, and identify sets of mark- er genes that explain the differences among the discovered clusters. A primary drawback to this three-step procedure is that each step is carried out independently, thereby neglecting the effects of the nonlinear embedding and inter-gene depen- dencies on the selection of marker genes. Here we propose an integrated deep learning frame- work, Adversarial Clustering Explanation (ACE), that bundles all three steps into a single workflow. The method thus moves away from the notion of “marker genes” to instead identify a panel of ex- planatory genes. This panel may include genes that are not only enriched but also depleted rela- tive to other cell types, as well as genes that exhib- it differences between closely related cell types. Empirically, we demonstrate that ACE is able to identify gene panels that are both highly discrimi- native and nonredundant, and we demonstrate the applicability of ACE to an image recognition task. 1. Introduction Single-cell sequencing technology has enabled the high- throughput interrogation of many aspects of genome biolo- gy, including gene expression, DNA methylation, histone modification, chromatin accessibility and genome 3D archi- tecture (Stuart & Satija, 2019) In each of these cases, the resulting high-dimensional data can be represented as a s- parse matrix in which rows correspond to cells and columns correspond to features of those cells (gene expression val- ues, methylation events, etc.). Empirical evidence suggests that this data resides on a low-dimensional manifold with latent semantic structure (Welch et al., 2017). Accordingly, 1Department of Genome Sciences, University of Washington, Seattle, WA 2Graduate Program in Molecular and Cellular Biology, University of Washington, Seattle, WA 3Paul G. Allen School of Computer Science and Engineering, University of Washington, Seattle, WA. Correspondence to: William Stafford Noble . Preliminary work. Under review. identifying groups of cells in terms of their inherent latent semantics and thereafter reasoning about the differences be- tween these groups is an important area of research (Plumb et al., 2020). In this study, we focus on the analysis of single cell RNA- seq (scRNA-seq) data. This is the most widely available type of single-cell sequencing data, and its analysis is chal- lenging not only because of the data’s high dimensionality but also due to noise, batch effects, and sparsity (Amodio et al., 2019). The scRNA-seq data itself is represented as a sparse, cell-by-gene matrix, typically with tens to hundreds of thousands of cells and tens of thousands of genes. A com- mon workflow in scRNA-seq analysis (Pliner et al., 2019) consists of three steps: (1) learn a compact representation of the data by projecting the cells to a lower-dimensional space; (2) identify groups of cells that are similar to each other in the low-dimensional representation, typically via clustering; and (3) characterize the differences in gene ex- pression among the groups, with the goal of understanding what biological processes are relevant to each group. Op- tionally, known “marker genes” may be used to assign cell type labels to the identified cell groups. A primary drawback to the above three-step procedure is that each step is carried out independently. Here, we pro- pose an integrated, deep learning framework for scRNA-seq analysis, Adversarial Clustering Explanation (ACE), that projects scRNA-seq data to a latent space, clusters the cells in that space, and identifies sets of genes that succinctly explain the differences among the discovered clusters (Fig- ure 1). At a high level, ACE first “neuralizes” the clustering procedure by reformulating it as a functionally equivalent multi-layer neural network (Kauffmann et al., 2019). In this way, in concatenation with a deep autoencoder that gen- erates the low-dimensional representation, ACE is able to attribute the cell’s group assignments all the way back to the input genes by leveraging gradient-based neural network explanation methods. Next, for each sample, ACE seeks small perturbations of its input gene expression profile that lead the neuralized clustering model to alter the group as- signments. These adversarial perturbations allow ACE to define a concise gene set signature for each cluster or pair of clusters. In particular, ACE attempts to answer the question, “For a given cell cluster, can we identify a subset of genes whose expression profiles are sufficient to identify members (which was not certified by peer review) is the author/funder. All rights reserved. No reuse allowed without permission. The copyright holder for this preprintthis version posted February 10, 2021. ; https://doi.org/10.1101/2021.02.08.428881doi: bioRxiv preprint https://doi.org/10.1101/2021.02.08.428881 Adversarial clustering explanation of this cluster?” We frame this problem as a ranking task, where thresholding the ranked list yields a set of explanatory genes. ACE’s joint modeling approach offers several benefits rela- tive to the existing state of the art. First, most existing meth- ods for the third step of the analysis pipeline—identifying genes associated with a given group of cells—treat each gene independently (Love et al., 2014). These approach- es ignore the dependencies among genes that are induced by gene networks, and often yield lists of genes that are highly redundant. ACE, in contrast, aims to find a smal- l set of genes that jointly explain a given cluster or pair of clusters. Second, most current methods identify genes associated with a group of cells without considering the nonlinear embedding model which maps the gene expres- sion to the low-dimensional representation where the groups are defined in the first place. To our knowledge, the only exception is the global counterfactual explanation (GCE) algorithm (Plumb et al., 2020), but that algorithm is limited to using a linear transformation. A third advantage of ACE’s integrated approach is its ability to take into account batch effects during the assignment of genes to clusters. Stan- dard nonlinear embedding methods, such as t-SNE (Van der Maaten & Hinton, 2008) and UMAP (McInnes & Healy, 2018; Becht et al., 2019), cannot take such structure into account and hence may lead to incorrect interpretation of the data (Amodio et al., 2019; Li et al., 2020). To address this problem, deep autoencoders with integrated denoising and batch correction can be used for scRNA-seq analysis (Lopez et al., 2018; Amodio et al., 2019; Li et al., 2020). We demonstrate below that batch effect structure can be usefully incorporated into the ACE model. A notable feature of ACE’s approach is that, by identify- ing genes jointly, the method moves away from the notion of a “marker gene” to instead identify a “gene panel”. As such, genes in the panel may not be solely enriched in a single cluster, but may together be predictive of the clus- ter. In particular, in addition to a ranking of genes, ACE assigns a Boolean to each gene indicating whether its inclu- sion in the panel is positive or negative, i.e., whether the gene’s expression is enriched or depleted relative to clus- ter membership. We have applied ACE to both simulated and real datasets to demonstrate its empirical utility. Our experiments demonstrate that ACE identifies gene panels that are highly discriminative and exhibit low redundancy. We further provide results suggesting that ACE is useful in domains beyond biology, such as image recognition. The Apache licensed source code of ACE (see submitted file) will be made publicly available upon acceptance. 2. Related work ACE falls into the paradigm of deep neural network interpre- tation methods, which have been developed primarily in the context of classification problems. These methods can be loosely categorized into three types: feature attribution meth- ods, counterfactual-based methods, and model-agnostic ap- proximation methods. Feature attribution methods assign an importance score to individual features so that higher scores indicate higher importance to the output prediction (Simonyan et al., 2013; Shrikumar et al., 2017; Lundberg & Lee, 2017). Counterfactual-based methods typically i- dentify the important subregions within an input sample by perturbing the subregions (by adding noise, rescaling (Sun- dararajan et al., 2017), blurring (Fong & Vedaldi, 2017), or inpainting (Chang et al., 2018)) and measuring the resulting changes in the predictions. Lastly, model-agnostic approxi- mation methods approximate the model being explained by using a simpler, surrogate function which is self-explainable (e.g., a sparse linear model, etc.) (Ribeiro et al., 2016). Recently, some interpretation methods have emerged to un- derstand models beyond classification tasks (Samek et al., 2020; Kauffmann et al., 2020; 2019), including the one we present in this paper for the purpose of cluster explanation. ACE’s perturbation approach draws inspiration from ad- versarial machine learning (Xu et al., 2020) where imper- ceivable perturbations are maliciously crafted to mislead a machine learning model to predict incorrect outputs. In particular, ACE’s approach is closest to the setting of a “white-box attack,” which assumes complete knowledge to the model, including its parameters, architecture, gradients, etc. (Szegedy et al., 2013; Kurakin et al., 2016; Madry et al., 2017; Carlini & Wagner, 2017). In contrast to these meth- ods, ACE re-purposes the malicious adversarial attack for a constructive purpose, identifying sets of genes that explain clusters in scRNA-seq data. ACE operates in concatenation with a deep autoencoder that generates the low-dimensional representation. In this paper, ACE uses SAUCIE (Amodio et al., 2019), a commonly- used scRNA-seq embedding method that incorporates batch correction. In principle, ACE is generalizable to any off-the- shelf scRNA-seq embedding methods, including SLICER (Welch et al., 2016), scVI (Way & Greene, 2018), scANVI (Xu et al., 2021), DESC (Li et al., 2020), and ItClust (Hu et al., 2020). 3. Approach 3.1. Problem setup We aim to carry out three analysis steps for a given scRNA- seq dataset, producing a low-dimensional representation of each cell’s expression profile, a cluster assignment for each cell, and a concise set of “explanatory genes” for each (which was not certified by peer review) is the author/funder. All rights reserved. No reuse allowed without permission. The copyright holder for this preprintthis version posted February 10, 2021. ; https://doi.org/10.1101/2021.02.08.428881doi: bioRxiv preprint https://doi.org/10.1101/2021.02.08.428881 Adversarial clustering explanation Genes C el ls Encoder Decoder Gene 1 ... Embeddings Gene 2 Gene 3 Gene p Cell ... Gene 1 ... Gene 2 Gene 3 Gene p Neuralized clusteringEncoderCell ... Source group assignment Target group assignment Gene p Rank Gene Score #1 Gene 1#2 Gene 2#3 Gene 3#4 ... ... ... Input: gene expression matrix Deep autoencoder learns low-dimensional representation Embedding clustering Clustering is neuralized and concatenated with the encoder Differentiation analysis by ACE Output: gene relevance + ... ... P er tu rb at io n vesus ... Figure 1. ACE workflow. ACE takes as input a single-cell gene expression matrix and learns a low-dimensional representation for each cell. Next, a neuralized version of the k-means algorithm is applied to the learned representation to identify cell groups. Finally, for pairs of groups of interest (either each group compared to its complement, or all pairs of groups), ACE seeks small perturbations of its input gene expression profile that lead the neuralized clustering model to alter the assignment from one group to the other. The workflow employs a combined objective function to induce the nonlinear embedding and clustering jointly. ACE produces as output the learned embedding, the cell group assignments, and a ranked list of explanatory genes for each cell group. cluster or pair of clusters. Let X = (x1,x2, · · · ,xn) T ∈ Rn×p be the normalized gene expression matrix obtained from a scRNA-seq experiment, where rows correspond to n cells and columns correspond to p genes. ACE relies on the following three components: (1) an autoencoder to learn a low-dimensional representation of the scRNA-seq data, (2) a neuralized clustering algorithm to identify groups of cells in the low-dimensional representation, and (3) an adversarial perturbation scheme to explain differences between groups by identifying explanatory gene sets. 3.2. Learning the low-dimensional representation Embedding scRNA-seq expression data into a low- dimensional space aims to capture the underlying structure of the data, based upon the assumption that the biological manifold on which cellular expression profiles lie is inher- ently low-dimensional. Specifically, ACE aims to learn a mapping f(·) : Rp 7→ Rd that transforms the cells from the high-dimensional input space Rp to a lower-dimensional embedding space Rd, where d � p. To accurately represent the data in Rd, we use an autoencoder consisting of two components, an encoder f(·) : Rp 7→ Rd and a decoder g(·) : Rd 7→ Rp. This autoencoder optimizes the generic loss min θ n∑ i=1 ‖xi −g(f(xi))‖ 2 2 (1) Finally, we denote Z = (z1,z2, · · · ,zn) T ∈ Rn×d as the low-dimensional representation obtained from the encoder, where zi ∈ Rd = f(xi) is the embedded representation of cell xi. The autoencoder in ACE can be extended in several impor- tant ways. For example, in some settings, Equation 1 is augmented with a task-specific regularizer Ω(X): min θ n∑ i=1 ‖xi −g(f(xi))‖ 2 2 + Ω(X). (2) As mentioned in Section 2, the scRNA-seq embedding method used by ACE, SAUCIE, encodes in Ω(X) a batch correction regularizer by using maximum mean discrepancy. In this paper, ACE uses SAUCIE coupled with a feature se- lection layer (Abid et al., 2019), with the aim of minimizing redundancy and facilitating selection of diverse explanatory gene sets. 3.3. Neuralizing the clustering step To carry out clustering in the low-dimensional space learned by the autoencoder, ACE uses a neuralized version of the k-means algorithm. This clustering step aims to partition Z ∈ Rn×d into C groups, where each group potentially corresponds to a distinct cell type. The standard k-means algorithm aims to minimize the fol- (which was not certified by peer review) is the author/funder. All rights reserved. No reuse allowed without permission. The copyright holder for this preprintthis version posted February 10, 2021. ; https://doi.org/10.1101/2021.02.08.428881doi: bioRxiv preprint https://doi.org/10.1101/2021.02.08.428881 Adversarial clustering explanation lowing objective function by identifying a set of group cen- troids { µc ∈ Rd : c = 1, 2, · · · ,C } : min ∑ ic δicoc(zi) (3) where δic indicates whether cell zi belongs to group c and the “outlierness” measure oc(zi) of cell zi relative to group c is defined as oc(zi) = ‖zi −µc‖ 2. Following Kauffmann et al. (2019), we neuralize the k- means algorithm by creating a neural network containing C modules, each with two layers. The architecture is mo- tivated by a soft assignment function that quantifies, for a particular cell zi and a specified group c, the group assign- ment probability score pc(zi) = exp(−βoc(zi))∑ k exp(−βok(zi)) (4) where the hyperparameter β controls the clustering fuzzi- ness. As β approaches infinity, Equation 4 approaches the indicator function for the closest centroid and thus reduces to hard clustering. To measure the confidence of group assignment, we use a logit function written as mc(zi) = log ( pc(zi) 1 −pc(zi) ) = β · β min k 6=c { ‖zi −µk‖ 2 −‖zi −µc‖ 2 } (5) where minβk 6=c{·} = − 1 β log ∑ exp(−β(·)) indicates a soft min-pooling layer. (See Kauffmann et al. (2019) for a detailed derivation.) The rationale for using the logit func- tion is that if there is as much confidence supporting the group membership as against it, then the confidence score mc(z) = 0. Additionally, Equation 5 has the following interpretation: the data point z belongs to the group c if and only if the distance to its centroid is smaller than the distance to all other competing groups. Equation 5 further decomposes into a two-layer neural network module: hck(zi) = w T ckzi + bck mc(zi) = β · β min k 6=c {hck(zi)} (6) where the first layer is a linear transformation layer with parameters wck = 2 · (µc −µk) and bck = ‖µk‖ 2 −‖µc‖ 2, and the second layer is the soft min-pooling layer introduced in Equation 5. ACE constructs one such module for each of the C clusters, as illustrated in Figure 1. 3.4. Explaining the groups ACE’s final step aims to induce, for each cluster identified by the neuralized k-mean algorithm, a ranking on genes such that highly ranked genes best explain that cluster. We consider two variants of this task: the one-vs-rest setting compares the group of interest Zs = f(Xs) ⊆ Z to its complement set Zt = f(Xt) ⊆ Z, where Xt = X\Xs; the one-vs-one setting compares one group of interest in Zs = f(Xs) ⊆ Z to a second group of interest Zt = f(Xt) ⊆ Z. In each setting, the goal is to identify the key differences between the source group Xs ⊆ X and the target group Xt ⊆ X in the input space, i.e., in terms of the genes. We treat this as a neural network explanation problem by finding the minimal perturbation within the group of interest, x ∈ Xs, that alters the group assignment from the source group s to the target group t. Specifically, we optimize an objective function that is a mixture of two terms: the first term is the difference between the current sample x and the perturbed sample x̂ = x + δ where δ ∈ Rp, and the second term quantifies the difference in group assignments induced by the perturbation. The objective function for the one-vs-one setting is min δ ‖δ‖1 + λ max(0,α + ms(x + δ) −mt(x + δ)) (7) where λ > 0 is a tradeoff coefficient to either encourage a small perturbation of x when small or a stronger alternation to the target group when large. The second term penalizes the situation where the group logit for the source group s is still larger than the target group t, up to a pre-specified margin α > 0. In this paper we fix α = 1.0. The difference between the current sample x and the potentially perturbed x̂ is measured by the L1 norm to encourage sparsity and non-redundancy. Note that Equation 7 assumes that the input expression matrix is normalized so that a perturbation added to one gene is equivalent to that same perturbation added to a different gene. Analogously, in the one-vs-rest case, the objective function for the optimization is min δ ‖δ‖1 +λ max(0,α+ms(x+δ)−max t6=s mt(x+δ)) (8) where the second term penalizes the situation in which the group logit for the source group s is larger than all non- source target groups. Finally, with the δ ∈ Rp obtained by optimizing either Equation 7 or Equation 8, ACE quantifies the importance of the ith gene relative to a perturbation from source group s to target group t as the absolute value of δi, thereby inducing a ranking in which highly ranked genes are more specific to the group of interest. 4. Baseline methods We compare ACE against six methodologically distinct base- line methods, each of which induces a ranking on genes in terms of group-specific importance, analogous to ACE. (which was not certified by peer review) is the author/funder. All rights reserved. No reuse allowed without permission. The copyright holder for this preprintthis version posted February 10, 2021. ; https://doi.org/10.1101/2021.02.08.428881doi: bioRxiv preprint https://doi.org/10.1101/2021.02.08.428881 Adversarial clustering explanation DESeq2 (Love et al., 2014) is a representative statistical hypothesis testing method that tests for differential gene expression based on a negative binomial model. The main caveat of DESeq2 is that it treats each gene as independent. The Jensen-Shannon Distance (JSD) (Cabili et al., 2011) is a representative distribution distance-based method which quantifies the specificity of a gene to a cell group. Similar to DESeq2, JSD considers each gene independently. Global counterfactual explanation (GCE) (Plumb et al., 2020) is a compressed sensing method that aims to identify consistent differences among all pairs of groups. Unlike ACE, GCE requires a linear embedding of the scRNA-seq data. The gene relevance score (GRS) (Angerer et al., 2020) is a gradient-based explanation method that aims to attribute a low-dimensional embedding back to the genes. The main limitations of GRS are two-fold. First, the embedding used in GRS is constrained to be a diffusion map, which is chosen specifically to make the gradient easy to calculate. Second, taking the gradient with respect to the embedding only indi- rectly measures the group differentiation compared to taking the gradient with respect to the group difference directly, as in ACE. SmoothGrad (Smilkov et al., 2017) and SHAP (Lundberg & Lee, 2017), which are designed primarily for classification problems, are two representative feature attribution methods. Each one computes an importance score that indicates each gene’s contribution to the clustering assignment. Smooth- Grad relies on knowledge to the model, whereas SHAP does not. 5. Results 5.1. Performance on simulated data To compare ACE to each of the baseline methods, we used a recently reported simulation method, SymSim (Zhang et al., 2019), to generate two synthetic scRNA-seq datasets: one “clean” dataset and one “complex” dataset. In both cases, we simulated many redundant genes, in order to adequately challenge methods that aim to detect a minimal set of informative genes. The simulation of the clean dataset uses a protocol similar to that of Plumb et al. (2020). We first used SymSim to generate a background matrix containing simulated counts from 500 cells, 2000 genes, and five distinct clusters. We then used this background matrix to construct our simu- lated dataset of 500 cells by 220 genes. The simulated data is comprised of three sets of genes: 20 causal genes, 100 dependent genes, and 100 noise genes. To select the causal genes, we identified all genes that are differentially expressed by SymSim’s criteria (nDiff-EVFgene > 0 and | log2 fold-change| > 0.8) between at least one pair of clus- ters, and we selected the 20 genes that exhibit the largest average fold-change across all pairs of clusters in which the gene was differentially expressed. A UMAP embedding on these causal genes alone confirms that they are jointly capa- ble of separating cells into their respective clusters (Fig. 2A). Next, we simulated 100 dependent genes, which are weight- ed sums of 1–10 randomly selected causal genes, with added gaussian noise. As such, a dependent gene is highly cor- related with a causal gene or with a linear combination of multiple causal genes. The weights were sampled from a continuous uniform distribution, U(0.01, 0.8), and the gaus- sian noise was sampled from N(0, 1). As expected, the dependent genes are also jointly capable of separating cells into their respective clusters (Fig. 2A). Lastly, we found all genes that were not differentially expressed between any cluster pair in the ground truth, and we randomly sampled 100 noise genes. These genes provide no explanation of the clustering structure (Fig. 2A). To simulate the complex dataset, we used SymSim to add dropout events and batch effects to the background ma- trix generated previously. We then selected the same exact causal and noise genes as in the clean dataset, and used the same exact random combinations and weights to generate the dependent genes. Thus, the clean and complex datasets contain the same 220 genes; however, the complex dataset enables us to gauge how robust ACE is to artifacts of tech- nical noise observed in real single-cell RNA-seq datasets (Fig. 2B). To compare the different gene ranking methods, we need to specify the ground truth cluster labels and a performance measure. We observe that the embedding representation learned by ACE exhibits clear cluster patterns even in the p- resence of dropout events and batch effects, and thus ACE’s k-means clustering is able to recover these clusters (Ap- pendix Figure A.1). Accordingly, to compare different meth- ods for inducing gene rankings, we provide ACE and each baseline method with the ground truth clustering labels from the original study (Zheng et al., 2017). ACE then calculates the group centroid used in Equation 3 by averaging the data points of the corresponding ground truth cluster. The em- bedding layer together with the group centroids are then used to build the neuralized clustering model (Equation 6). Each method produces gene rankings for every cluster in a one-vs-rest fashion. To measure how well a gene ranking captures clustering structure, we use the Jaccard distance to measure the similarity between a cell’s k nearest neighbors (k-NN) when using a subset of top-ranked genes and a cell’s k-NN when using all genes. To compute the k-NN, we use the Euclidean distance metric. The Jaccard distance is defined as JD(i) = 1 − Sfull ∩Ssub Sfull ∪Ssub (9) (which was not certified by peer review) is the author/funder. All rights reserved. No reuse allowed without permission. The copyright holder for this preprintthis version posted February 10, 2021. ; https://doi.org/10.1101/2021.02.08.428881doi: bioRxiv preprint https://doi.org/10.1101/2021.02.08.428881 Adversarial clustering explanation Figure 2. Comparing ACE to baseline methods on simulated scRNA-seq datasets. Each dataset consists of 20 causal genes, 100 dependent genes, and 100 noise genes. (A) UMAP embeddings of cells composing the clean dataset. Panels correspond to embeddings using the three subsets of genes (causal, dependent, and noise), as well as all of the genes together. (B) Same as panel A, but for the complex dataset. (C) Comparison of methods via Jaccard distance as a function of the number of genes in the ranking. ACE performs substantially better than each of the baseline methods on the clean dataset. The gray dashed line indicates the mean Jaccard distance achieved by the 20 causal genes alone. (D) Same as panel C but for the complex dataset. where Sfull represents cell i’s k-NN’s when using all genes, and Ssub represents cell i’s k-NN’s when using a subset of top-ranked genes. If the subset of top-ranked genes does a good job of explaining a cluster of cells, then Sfull ∩Ssub and Sfull ∪ Ssub should be nearly equal, and the Jaccard distance should approach 0. We select the gene ranking used to derive a subset of top-ranked genes based on the cell cluster assignment. For example, if the cell belongs in cluster 2, we use the cluster 2 vs. rest gene ranking. Thus, to obtain a global measure of how well a clustering structure is captured on a subset of top-ranked genes, we report the mean Jaccard distance across all cells. Our analysis shows that ACE considerably outperforms each of the baseline methods on the clean dataset, indicating that it is superior at identifying the minimal set of informative genes (Fig. 2B). Notably, ACE outperforms the mean Jac- card distance achieved by the causal genes alone before reaching 20 genes used, suggesting that the method success- fully identifies dependent genes that are more informative than individual causal genes. ACE also performs strongly on the complex dataset, though it appears to perform on par with SmoothGrad and SHAP) (Fig. 2D). Notably, these three methods —ACE, SHAP, and SmoothGrad —share a common feature, employing the SAUCIE framework that facilitates automatic batch effect correction, highlighting the utility of DNN-based dimensionality reduction and in- terpretation methods for single-cell RNA-seq applications. 5.2. Real data analysis We next applied ACE to a real dataset of peripheral blood mononuclear cells (PBMCs) (Zheng et al., 2017), repre- sented as a cell-by-gene log-normalized expression matrix containing 2638 cells and 1838 highly variable genes. The cells in the dataset were previously categorized into eight cell types, obtained by performing Louvain clustering (Blon- del et al., 2008) and annotating each cluster on the basis of differentially expressed marker genes. As shown in Fig- ure 3A and Appendix Figure A.2, ACE’s k-means clustering successfully recovers the reported cell types based upon the 10-dimensional embedding learned by SAUCIE. We first aimed to quantify the discriminative power of the top-ranked genes identified by ACE in comparison to the six baseline methods. To do this, we applied all the six baseline methods to the PBMC dataset using the groups identified by the k-means clustering based on the SAUCIE embedding. For each group of cells, we extracted the top- k group-specific genes reported by each method, where k ranges from 1%, 2%, · · · , 100% among all genes. Given the selected gene subset, we then trained a support vector machine (SVM) classifier with a radial basis function kernel to separate the target group from the remaining groups. The SVM training involves two hyperparameters, the regular- ization coefficient C and the bandwidth parameter σ. The σ parameter is adaptively chosen so that the training data is Z-score normalized, using the default settings in Scikit- learn (Pedregosa et al., 2011). The C parameter is selected (which was not certified by peer review) is the author/funder. All rights reserved. No reuse allowed without permission. The copyright holder for this preprintthis version posted February 10, 2021. ; https://doi.org/10.1101/2021.02.08.428881doi: bioRxiv preprint https://doi.org/10.1101/2021.02.08.428881 Adversarial clustering explanation Top 1% genes The intersection among different methods (B) (C) (D) # of genes included # of genes included From most to least important Important genes specific to CD4 T cell 0 5 10 0 4 8 UMAP1 U M A P 2 B cells CD14+ Monocytes CD4 T cells CD8 T cells Dendritic Cells FCGR3A+ Monocytes Megakaryocytes NK cells (A) From most to least important A U R O C P e a rs o n c o rr e la tio n Figure 3. Comparing ACE to baseline methods on PBMC dataset. (A) UMAP embedding of PBMC cells labelled by ACE’s k-means clustering assignment. (B) Classification performance of each method, as measured by AUROC, as a function of the number of genes in the set. Error bars correspond to the standard error of the mean of AUROC scores from each test split across different target groups. (C) Redundancy among the top k genes, as measured by Pearson correlation, as a function of k. Error bars correspond to the standard error of the mean calculated from the group-specific correlations. (D) The figure plots overlaps among the top 18 genes (corresponding to 1% of 1838 genes) identified by all seven methods with respect to the CD4 T cell cluster. by grid search from {5−5, 5−4, · · · , 50, · · · , 54, 55} . The classification performance, in terms of area under the receiv- er operating characteristic curve (AUROC), is evaluated by 3-fold stratified cross-validation, and an additional 3-fold cross-validation is applied within each training split to de- termine the optimal C hyperparameter. Finally, AUROC scores from each test split across different target groups are aggregated and reported, in terms of the mean and the stan- dard error of the mean. Two cell types—megakaryocytes and dendritic cells—are excluded due to insufficient sample size (< 50). As shown in Figure 3B, the top-ranked genes reported by ACE are among the most discriminative across all methods, particularly when the inclusion size is small (≤ 3%). The only method that yields superior performance is DESeq2. We next tested the redundancy of top-ranked genes, as it is desirable to identify diverse explanatory gene sets with minimum redundancy. Specifically, for each target group of cells, we calculate the Pearson correlations between all gene pairs within top k genes, for varying values of k. The mean and standard error of the mean of these correlations are computed within each group and then averaged across dif- ferent target groups. The results of this analysis (Figure 3C) suggest that the top-ranked genes reported by ACE are a- mong the least redundant across all methods. Other methods that exhibit low redundancy include GRS and the two meth- ods that use the same SAUCIE model (i.e., SmoothGrad and SHAP). In conjunction with the discriminative power analysis in Figure 3B, we conclude that ACE achieves a powerful combination of high discriminative power and low redundancy. Finally, to better understand how these methods differ from one another, we investigated the consistency among the top- ranked genes reported by each method. For this analysis, we focused on one particular group, CD4 T cells. We discover strong disagreement among the methods (Figure 3D). Sur- prisingly, no single gene is selected among the top 1% by all methods. Among all methods, ACE covers the most that are reported by at least one other method (14 out of 18 genes). The four genes that ACE uniquely identifies (red bar in Fig- ure 3D)—CCL5, GZMK, SPOCD1, and SNRNP27—are depleted rather than enriched relative to other cell types. It is worth mentioning that both CCL5 and GZMK are enriched in CD8 T cells (Thul et al., 2017), the closest cell type to CD4 T cell (Figure 3A). This observation suggests ACE identifies cells that exhibit highly discriminative changes in expression between two closely related cell types. In- deed, among ACE’s 18-gene panel, 15 genes are depleted rather than enriched, suggesting that much of CD4’s cell identity may be due to inhibition rather than activation of specific genes. In summary, ACE is able to move away from the notion of a “marker gene” to instead identify a highly discriminative, nonredundant gene panel. 5.3. Image analysis Although we developed ACE for application to scRNA-seq data, we hypothesized that the method would be useful in do- mains beyond biology. Explanation methods are potentially useful, for example, in the analysis of biomedical images, where the explanations can identify regions of the image responsible for assignment of the image to a particular phe- notypic category. As a proof of principle for this general domain, we applied ACE to the MNIST handwritten digits dataset (LeCun, 1998), with the aim of studying whether ACE can identify which pixels in a given image explain why the image was assigned to one digit versus another. Specif- ically, we solve the optimization problem for each input image in Equation 7, seeking an image-specific set of pixel modifications, subject to the constraint that the perturbed image pixel values are restricted to lie in the range [0, 1]. Note that this task is somewhat different from the scRNA- seq case: in the MNIST case, ACE finds a different set of explanatory pixels for each image, whereas in the scRNA- (which was not certified by peer review) is the author/funder. All rights reserved. No reuse allowed without permission. The copyright holder for this preprintthis version posted February 10, 2021. ; https://doi.org/10.1101/2021.02.08.428881doi: bioRxiv preprint https://doi.org/10.1101/2021.02.08.428881 Adversarial clustering explanation 0 8 1 4 1 7 2 3 3 8 4 5 4 9 5 0 5 6 5 8 6 0 7 0 7 1 7 2 7 9 8 0 8 3 8 5 9 0 9 8 -1 +1Perturbation range: Pixels in initial digit: Figure 4. Applying ACE to the MNIST dataset. ACE is able to explain 20 types of digit transitions in a pixel-wise manner. These digit transitions are chosen such that each digit category is covered at least once in both directions. seq case, ACE seeks a single set of genes that explains label differences across all cells in the dataset. ACE was applied to this dataset as follows. We used a sim- ple convolution neural network architecture containing two convolution layers, each with a modest filter size (5, 5), a modest number of filters (32) and ReLU activation, followed by a max pooling layer with a pool size (2, 2), a fully con- nected layer, and a softmax layer. The model was trained on the MNIST training set (60,000 examples) for 10 epochs, using Adam (Kingma & Ba, 2015) with an initial learning rate of 0.001. The network achieves 98.7% classification accuracy on the test test of 10,000 images. We observe that the embedding representation in the last pooling layer ex- hibits well-separated cluster patterns (Appendix Figure A.3). Since our goal is not to learn the cluster structure per se, for simplicity, we fixed the number of groups to be the number of digit categories (i.e., 10) and calculated the group cen- troid used in Equation 3 by averaging the data points of the corresponding category. The embedding layer together with the group centroids are then used to build the neuralized clustering model (Equation 6.) The results of this analysis show that ACE does a good job of identifying sets of pixels that accurately explain differ- ences between pairs of digits. We examined the pixel-wise explanations of 20 pairs of digits, randomly selected to cov- er each digit category at least once in both directions (Fig. 4). For example, to convert “8” to “5,” ACE disconnects the top right and bottom left of “8,” as expected. Similarly, to convert “8” to “3,” ACE disconnects the top left and bottom left of “8.” It is worth noting that the modifications intro- duced by ACE are inherently symmetric. For example, to convert “1” to “7” and back again, ACE suggests adding and removing the same part of “7.” 6. Discussion and conclusion In this work, we have proposed a deep learning-based scRNA-seq analysis pipeline, ACE, that projects scRNA- seq data to a latent space, clusters the cells in that space, and identifies sets of genes that succinctly explain the d- ifferences among the discovered clusters. Compared to existing state-of-the-art methods, ACE jointly takes into consideration both the nonlinear embedding of cells to a low-dimensional representation and the intrinsic dependen- cies among genes. As such, the method moves away from the notion of a “marker gene” to instead identify a panel of genes. This panel may include genes that are not only enriched but also depleted relative to other cell types, as well as genes that exhibit important differences between closely related cell types. Our experiments demonstrate that ACE identifies gene panels that are highly discriminative sets and exhibit low redundancy. We also provide results suggesting that ACE’s approach may be useful in domains beyond biology, such as image recognition. This work points to several promising directions for future research. In principle, ACE can be used in conjunction with any off-the-shelf scRNA-seq embedding method. Thus, empirical investigation of the utility of generalizing ACE to use embedders other than SAUCIE would be interest- ing. Another possible extension is to apply neuralization to alternative clustering algorithms. For example, in the con- text of scRNA-seq analysis the Louvain algorithm (Blondel et al., 2008) is commonly used and may be a good candidate for neuralization. A promising direction for future work is to provide confidence estimation for the top-ranked group- specific genes, in terms of q-values (Storey, 2003), with the help of the recently proposed knockoffs framework (Barber & Candès, 2015; Lu et al., 2018). (which was not certified by peer review) is the author/funder. All rights reserved. No reuse allowed without permission. The copyright holder for this preprintthis version posted February 10, 2021. ; https://doi.org/10.1101/2021.02.08.428881doi: bioRxiv preprint https://doi.org/10.1101/2021.02.08.428881 Adversarial clustering explanation References Abid, A., Balin, M. F., and Zou, J. Concrete autoencoders for differentiable feature selection and reconstruction. International Conference on Machine Learning, 2019. Amodio, M., Dijk, D. V., Srinivasan, K., Chen, W. S., Mohsen, H., Moon, K. R., Campbell, A., Zhao, Y., Wang, X., Venkataswamy, M., and Krishnaswamy, S. Exploring single-cell data with deep multitasking neural networks. Nature Methods, pp. 1–7, 2019. Angerer, P., Fischer, D. S., Theis, F. J., Scialdone, A., and Marr, C. Automatic identification of relevant genes from low-dimensional embeddings of single cell rnaseq data. Bioinformatics, 2020. Barber, R. F. and Candès, E. J. Controlling the false discov- ery rate via knockoffs. The Annals of Statistics, 43(5): 2055–2085, 2015. Becht, E., McInnes, L., Healy, J., Dutertre, C., Kwok, I. W. H., Ng, L. G., Ginhoux, F., and Newell, E. W. Dimen- sionality reduction for visualizing single-cell data using UMAP. Nature Biotechnology, 37(1):38–44, 2019. Blondel, V. D., Guillaume, J.-L., Lambiotte, R., and Lefeb- vre, E. Fast unfolding of communities in large net- works. Journal of Statistical Mechanics: Theory and Experiment, 2008(10):P10008, 2008. Cabili, M. N., Trapnell, C., Goff, L., Koziol, M., Tazon- Vega, B., Regev, A., and Rinn, J. L. Integrative annotation of human large intergenic noncoding RNAs reveals global properties and specific subclasses. Genes Dev, 25(18): 1915–1927, 2011. Carlini, N. and Wagner, D. Towards evaluating the robust- ness of neural networks. In 2017 IEEE Symposium on Security and Privacy (SP), pp. 39–57. IEEE, 2017. Chang, C., Creager, E., Goldenberg, A., and Duvenaud, D. Explaining image classifiers by counterfactual generation. arXiv preprint arXiv:1807.08024, 2018. Fong, R. and Vedaldi, A. Interpretable explanations of black boxes by meaningful perturbation. In Proceedings of the IEEE International Conference on Computer Vision, pp. 3429–3437, 2017. Hu, J., Li, X., Hu, G., Lyu, Y., Susztak, K., and Li, M. Iterative transfer learning with neural network for clus- tering and cell type classification in single-cell RNA-seq analysis. Nature Machine Intelligence, 2(10):607–618, 2020. Kauffmann, J., Esders, M., Montavon, G., Samek, W., and Müller, K. From clustering to cluster explanations via neural networks. arXiv preprint arXiv:1906.07633, 2019. Kauffmann, J., Müller, K., and Montavon, G. Towards ex- plaining anomalies: a deep taylor decomposition of one- class models. Pattern Recognition, 101:107198, 2020. Kingma, D. and Ba, J. Adam: A method for stochastic optimization. In Proceedings of the 3rd International Conference on Learning Representations, 2015. Kurakin, A., Goodfellow, I., and Bengio, S. Adversar- ial examples in the physical world. arXiv preprint arXiv:1607.02533, 2016. LeCun, Y. The MNIST database of handwritten digits. http://yann. lecun. com/exdb/mnist/, 1998. Li, X., Wang, K., Lyu, Y., Pan, H., Zhang, J., Stambo- lian, D., Susztak, K., Reilly, M. P., Hu, G., and Li, M. Deep learning enables accurate clustering with batch ef- fect removal in single-cell RNA-seq analysis. Nature Communications, 11(1):1–14, 2020. Lopez, R., Regier, J., Cole, M. B., Jordan, M. I., and Yosef, N. Deep generative modeling for single-cell transcrip- tomics. Nature Methods, 15(12):1053–1058, 2018. Love, M., Huker, W., and Anders, S. Moderated estimation of fold change and dispersion for RNA-seq data with deseq2. Genome Biology, 15(550), 2014. Lu, Y. Y., Fan, Y., Lv, J., and Noble, W. S. DeepPINK: reproducible feature selection in deep neural networks. In Advances in Neural Information Processing Systems, 2018. Lundberg, S. and Lee, S. A unified approach to interpret- ing model predictions. Advances in Neural Information Processing Systems, 2017. Madry, A., Makelov, A., Schmidt, L., Tsipras, D., and Vladu, A. Towards deep learning models resistant to ad- versarial attacks. arXiv preprint arXiv:1706.06083, 2017. McInnes, L. and Healy, J. UMAP: Uniform manifold approximation and projection for dimension reduction. arXiv, 2018. Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., Blondel, M., Prettenhofer, P., Weiss, R., Dubourg, V., Vanderplas, J., Passos, A., Cour- napeau, D., Brucher, M., Perrot, M., and Duchesnay, E. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825–2830, 2011. Pliner, H. A., Shendure, J., and Trapnell, C. Supervised clas- sification enables rapid annotation of cell atlases. Nature Methods, 16(10):983–986, 2019. (which was not certified by peer review) is the author/funder. All rights reserved. No reuse allowed without permission. The copyright holder for this preprintthis version posted February 10, 2021. ; https://doi.org/10.1101/2021.02.08.428881doi: bioRxiv preprint https://doi.org/10.1101/2021.02.08.428881 Adversarial clustering explanation Plumb, G., Terhorst, J., Sankararaman, S., and Talwalka- r, A. Explaining groups of points in low-dimensional representations. ICML, 2020. Ribeiro, M., Singh, S., and Guestrin, C. "why should i trust you?": Explaining the predictions of any classifier. In Proceedings of the 22Nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’16, pp. 1135–1144, New York, NY, USA, 2016. ACM. Samek, W., Montavon, G., Lapuschkin, S., Anders, C. J., and Müller, K. R. Toward interpretable machine learning: Transparent deep neural networks and beyond. arXiv preprint arXiv:2003.07631, 2020. Shrikumar, A., Greenside, P., Shcherbina, A., and Kunda- je, A. Learning important features through propagating activation differences. In International Conference on Machine Learning, 2017. Simonyan, K., Vedaldi, A., and Zisserman, A. Deep in- side convolutional networks: Visualising image clas- sification models and saliency maps. arXiv preprint arXiv:1312.6034, 2013. Smilkov, D., Thorat, N., Kim, B., Viégas, F., and Watten- berg, M. Smoothgrad: removing noise by adding noise. arXiv preprint arXiv:1706.03825, 2017. Storey, J. D. The positive false discovery rate: A bayesian interpretation and the q-value. The Annals of Statistics, 31(6):2013–2035, 2003. Stuart, T. and Satija, R. Integrative single-cell analysis. Nature Reviews Genetics, 20:252–272, 2019. Sundararajan, M., Taly, A., and Yan, Q. Axiomatic attribu- tion for deep networks. In International Conference on Machine Learning, 2017. Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan, D., Goodfellow, I., and Fergus, R. Intriguing properties of neural networks. arXiv preprint arXiv:1312.6199, 2013. Thul, P., Åkesson, L., Wiking, M., Mahdessian, D., Gelada- ki, A., Blal, H., Alm, T., Asplund, A., Björk, L., Breckels, L., et al. A subcellular map of the human proteome. Science, 356(6340), 2017. Van der Maaten, L. and Hinton, G. Visualizing data using t-SNE. Journal of Machine Learning Research, 9(2579- 2605):85, 2008. Way, G. and Greene, C. Bayesian deep learning for single- cell analysis. Nature Methods, 15(12):1009–1010, 2018. Welch, J., Hartemink, A., and Prins, J. SLICER: inferring branched, nonlinear cellular trajectories from single cell rna-seq data. Genome Biology, 17(1):1–15, 2016. Welch, J. D., Hartemink, A. J., and Prins, J. F. MATCHER: manifold alignment reveals correspondence between sin- gle cell transcriptome and epigenome dynamics. Genome biology, 18(1):138, 2017. Xu, C., Lopez, R., Mehlman, E., Regier, J., Jordan, M., and Yosef, N. Probabilistic harmonization and annotation of single-cell transcriptomics data with deep generative models. Molecular Systems Biology, 17(1):e9620, 2021. Xu, H., Ma, Y., Liu, D., Liu, H., Tang, J., and Jain, A. Adversarial attacks and defenses in images, graphs and text: A review. International Journal of Automation and Computing, 17(2):151–178, 2020. Zhang, X., Xu, C., and Yosef, N. Simulating multiple faceted variability in single cell RNA sequencing. Nature Communications, 10(1):1–16, 2019. Zheng, G. X. Y., Terry, J. M., Belgrader, P., Ryvkin, P., Bent, Z. W., Wilson, R., Ziraldo, S. B., Wheeler, T. D., McDermott, G. P., Zhu, J., Gregory, M. T., Shuga, J., Montesclaros, L., Underwood, J. G., Masquelier, D. A., Nishimura, S. Y., Schnall-Levin, M., Wyatt, P. W., Hind- son, C. M., Bharadwaj, R., Wong, A., Ness, K. D., Beppu, L. W., Deeg, H. J., McFarland, C., Loeb, K. R., Va- lente, W. J., Ericson, N. G., Stevens, E. A., Radich, J. P., Mikkelsen, T. S., Hindson, B. J., and Biela, J. H. Mas- sively parallel digital transcriptional profiling of single cells. Nature Communications, 8:14049, 2017. (which was not certified by peer review) is the author/funder. All rights reserved. No reuse allowed without permission. The copyright holder for this preprintthis version posted February 10, 2021. ; https://doi.org/10.1101/2021.02.08.428881doi: bioRxiv preprint https://doi.org/10.1101/2021.02.08.428881 Adversarial clustering explanation 75 50 25 0 25 50 75 UMAP1 80 60 40 20 0 20 40 60 80 UM AP 2 label=0 label=1 label=2 label=3 label=4 label=5 label=6 label=7 label=8 label=9 Figure A.3. The embedding representation in the last pooling layer of the convolutional neural network exhibits well-separated cluster patterns among 10 digits on the MNIST dataset. (which was not certified by peer review) is the author/funder. All rights reserved. No reuse allowed without permission. The copyright holder for this preprintthis version posted February 10, 2021. ; https://doi.org/10.1101/2021.02.08.428881doi: bioRxiv preprint https://doi.org/10.1101/2021.02.08.428881 Adversarial clustering explanation −5.0 −2.5 0.0 2.5 −6 −4 −2 0 2 4 UMAP1 U M A P 2 −4 −2 0 2 4 6 −5 0 5 10 UMAP1 U M A P 2 Group 1 2 3 4 5 (A) (B) Figure A.1. The embedding representation learned by SAUCIE exhibits well-separated cluster patterns on both (A) clean and (B) complex simulated scRNA-seq datasets. Figure A.2. The embedding representation learned by SAUCIE exhibits similar cluster patterns by using either (A) the Louvain algorithm or (B) k-means clustering on the PBMC dataset. (which was not certified by peer review) is the author/funder. All rights reserved. No reuse allowed without permission. The copyright holder for this preprintthis version posted February 10, 2021. ; https://doi.org/10.1101/2021.02.08.428881doi: bioRxiv preprint https://doi.org/10.1101/2021.02.08.428881