key: cord-0274328-kgt58dqb authors: Li, Yan-Chak; Wang, Linhua; Law, Jeffrey; Murali, T. M.; Pandey, Gaurav title: Integrating multimodal data through interpretable heterogeneous ensembles date: 2022-01-15 journal: bioRxiv DOI: 10.1101/2020.05.29.123497 sha: 4a2359c17d387ffd10082659d422c1051dc700a5 doc_id: 274328 cord_uid: kgt58dqb Motivation Integrating multimodal data represents an effective approach to predicting biomedical characteristics, such as protein functions and disease outcomes. However, existing data integration approaches do not sufficiently address the heterogeneous semantics of multimodal data. In particular, early and intermediate approaches that rely on a uniform integrated representation reinforce the consensus among the modalities, but may lose exclusive local information. The alternative late integration approach that can address this challenge has not been systematically studied for biomedical problems. Results We propose Ensemble Integration (EI) as a novel systematic realization of the late integration approach. EI infers local predictive models from the individual modalities using appropriate algorithms, and uses effective heterogeneous ensemble algorithms to integrate these local models into a global predictive model. We also propose a novel interpretation method for EI models. We tested EI on the problems of predicting protein function from multimodal STRING data, and mortality due to COVID-19 from multimodal data in electronic health records. We found that EI accomplished its goal of producing significantly more accurate predictions than each individual modality. It also performed better than several established early integration methods for each of these problems. The interpretation of representative EI predictive models for COVID-19 mortality identified several disease-relevant features, such as laboratory test (BUN and calcium) and vital sign (minimum oxygen saturation) measurements and demographics (age). These results demonstrated the effectiveness of the EI framework for biomedical data integration and predictive modeling. Availability Code and data are available at https://github.com/GauravPandeyLab/ensemble_integration. Contact gaurav.pandey@mssm.edu Multimodal data are a collection of diverse types of data that capture complementary aspects of a biomedical entity and its characteristics of interest (Boehm et al. (2021) ). For instance, a protein may be characterized by its amino acid sequence, three-dimensional structure, evolutionary history and interactions with other proteins. These different types of information may be used to infer the function(s) of the protein (Pandey et al. (2006) ). Similarly, Electronic Health Record (EHR) systems contain diverse types of clinical data, such as from questionnaires, imaging and laboratory tests, that collectively provide a comprehensive view of a patient's health (Jensen et al. (2012) ). These multimodal data are expected to be complementary. For instance, data from questionnaires provides baseline information of a patient's health, while laboratory test measurements indicate their current health state. Due to this complementarity, integrating these multimodal data can result in a more comprehensive understanding and accurate predictions of biomedical characteristics or outcomes, such as the functions of proteins and disease phenotypes of patients (Boehm et al. (2021) ; Krassowski et al. (2020) ; Hasin et al. (2017) ). However, the integration and predictive modeling of multimodal data are still challenging because of the heterogeneity of the individual data modalities, which are usually structured differently and have different semantics (Zitnik et al. (2019) ; Gligorijević and Pržulj (2015) ). For instance, gene expression data can be structured as matrices, each entry of which individually denotes the expression level of a gene in a sample. In contrast, string-formatted amino acid sequences are unstructured, since each position in these sequences needs to be studied in the context of its surrounding positions to infer its biological role. Several data types also present their own challenges, such as the high dimensionality of gene expression profiles, which can make their analysis difficult (Saeys et al. (2007) ). Similar issues exist with the multimodal data in EHR systems as well. Due to the varying semantics and challenges of the individual modalities, it has been challenging to identify a uniformly effective prediction method for diverse multimodal data (Zitnik et al. (2019) ). Several data integration methods have been proposed to address the challenge of heterogeneity among multimodal data (Zitnik et al. (2019) ; Gligorijević and Pržulj (2015) ). These methods can be generally categorized into three groups, namely early, intermediate and late integration (Fig. 1) . Early integration strategies (Fig. 1a ) first combine the multimodal data into a uniform intermediate representation, such as a network (Caldera et al. (2017) ; Wang et al. (2014) ), which is then used for prediction and other analysis purposes. Intermediate strategies (Fig. 1b) jointly model the multiple datasets and their elements, such as genes, proteins, etc., also through a uniform intermediate representation. These uniform representations reinforce the consensus among modalities, but may obfuscate the local signal exclusive to each individual modality, and thus adversely affect prediction performance (Greene and Cunningham (2009); Libbrecht and Noble (2015) ). In view of these limitations of the early and intermediate approaches, the late approach (Fig. 1c ) offers an alternative path to improved prediction performance by first deriving specialized local predictive models from the individual modalities, and aggregating those models (Zitnik et al. (2019) ; Gligorijević and Pržulj (2015) ). This approach has the potential to capture maximal information in the individual modalities into the local predictors, whose aggregation then builds consensus, thus effectively utilizing both the commonalities and diversity among the modalities. However, this approach has not been systematically implemented and studied for biomedical data integration and predictive modeling problems (Zitnik et al. (2019) ). In this work, we propose heterogeneous ensembles as a novel systematic realization of the late integration strategy to build effective integrative predictive models from multimodal biomedical data. These ensembles can aggregate an unrestricted number and types of local models, thus providing an effective framework for aggregating diverse information captured in these models (Whalen et al. (2016) ). Significantly, this approach differs from homogeneous ensembles, such as random forest (Breiman (2001) ) and boosting (Schapire and Freund (2013) ), which typically aggregate only one type of individual models, e.g., decision trees in random forests. Furthermore, homogeneous methods learn individual models as a part of the ensemble process, and thus cannot integrate models that have been derived independently a priori. The advantages of heterogeneous ensembles have been demonstrated in several biomedical applications, such as protein function prediction (Whalen et al. (2016) ; Wang et al. (2018) ; Stanescu and Pandey (2017) ), enhancing the predictive power of DREAM Challenges (Sieberts et al. (2016 (Sieberts et al. ( , 2021 ) and modeling infectious disease epidemics (Ray and Reich (2018) ). However, these applications were limited to individual (unimodal) datasets. In the current work, we leverage the flexibility of the heterogeneous ensemble framework to advance predictive modeling from multimodal data in an approach that we name Ensemble Integration (EI; Fig. 2 ). Specifically, EI uses the same heterogeneous ensemble methods as used for unimodal data to integrate local models derived from individual data types. Through this process, EI is capable of incorporating both the consensus and diversity among the individual modalities. A challenge associated with multimodal data integration using EI is that the ensembles generated may be complex and difficult to interpret, not unlike other sophisticated machine learning-based models (Doshi-Velez and Kim (2017)). The interpretation of these ensembles is important for understanding the rationale behind their predictions and generating trust for the user. To address this challenge, we also propose a novel interpretation framework for EI-based models built from multimodal data. We tested the prediction and interpretation capabilities of EI on two diverse and challenging problems: protein function prediction (PFP) (Szklarczyk et al. (2015) ). We also tested the effectiveness of EI for predicting the likelihood of a COVID-19 patient's mortality (death) from the disease using the multimodal data collected in a treating hospital's EHRs (Wynants et al. (2020) ). We compared EI's performance with those of established data integration methods that have been used for these problems. Finally, in addition to evaluating EI's prediction abilities, we also used our ensemble interpretation method to reveal the key features that contributed to the EI-based predictive model of COVID-19 mortality, and verified the relevance of these features from the literature. Here, we describe the methodologies and datasets used in our development and evaluation of the EI framework and other approaches. Although the EI framework ( Fig. 2) is general, it needs to be actualized using specific prediction and ensemble algorithms to be applied to the target multimodal data. Fig. 3 shows the implementation of EI tested in this work. First, predictive models were trained from each individual data modality; we refer to these as local models below. We used ten established binary classification algorithms implemented in Weka (Frank et al. (2005) ) for training local models. These algorithms were AdaBoost, Decision Tree (J48), Gradient Boosting, K-nearest Neighbors, Support Vector Machine (SVM), Random Forest, Logistic Regression (LR), Rulebased Classification (PART), Naive Bayes and Voted Perceptron. All the algorithms were used with their default parameters in Weka, with the exception of specifying C=0.001 for SVM and M=100 for LR to control time to convergence. To handle potential imbalance between the classes being predicted, the entities belonging to the majority negative class in each modality were randomly under-sampled to balance the number of positive and negative class entities before training the local models. The corresponding test sets retained the same class ratio. Next, using the base predictions generated by the local models, EI used the following heterogeneous ensemble methods to build the integrative predictive model. • Mean aggregation, which calculates the ensemble output as the mean of the base prediction scores. • The iterative ensemble selection method of Caruana et al. (2004 Caruana et al. ( , 2006 (CES), which starts with an empty set as the current ensemble. In each iteration, CES adds the local model that improves the current ensemble's prediction performance by the largest amount. The iterative process, which employs sampling of the local models with replacement, continues until there is no further gain in performance. • Stacking (Sesmero et al. (2015) ), which uses the base predictions as features to train a second-level predictive model (meta-predictor) as the final ensemble. We used eight established binary classification algorithms available in Python's sklearn library (Pedregosa et al. (2011) ) to train meta-predictors. These algorithms were AdaBoost, Decision Tree, Gradient Boosting, K-nearest Neighbors, SVM, Random Forest, LR and Naive Bayes. We also included the XGBoost classifier (Chen and Guestrin (2016) ) as a meta-prediction algorithm. All the algorithms were executed using their default parameter settings, with the exception of the linear kernel being used with SVM to control the runtime. Fig. 3 : Implementation of EI. We used ten standard binary classification algorithms, such as support vector machine (SVM), random forest (RF) and logistic regression (LR), as implemented in Weka (Frank et al. (2005) ), to derive local predictive models from each individual data modality. We then applied the stacking and ensemble selection methods to these local models to generate the final EI model. This model generated prediction scores for the entities and multimodal data of interest that were evaluated to assess the performance of the model. Finally, considering the bestperforming EI model as a representative, we used our interpretation method to identify the features that contributed the most substantially to the model's predictions. Thus, each execution of EI yielded eleven models, i.e., one each based on mean aggregation and CES, and nine based on stacking. These EI models then generated the final prediction scores for the entities of interest. These scores were evaluated to assess the models' performance. To aid the interpretation of EI models and build trust in them, we propose a novel method to identify the key features in the various data modalities that contribute the most to the model's predictions. This method first quantifies the contributions the features in the individual modalities (local features) make to the corresponding local models. It then quantifies the contribution of each local model to the EI ensemble. Finally, it combines these contributions to determine the most important features for the EI model. Algorithmically, the method calculates the local feature ranks (LFRs) from the local models and the local model ranks (LMRs) from the EI model (Algorithm 1) as follows. The LFRs for each local model are calculated based on the difference of the performance of the model when each feature is held out. This calculation is carried out using the ClassifierAttributeEval function in Weka. The calculation of LMRs depends on the type of ensemble algorithm used to build the EI model under consideration. If the model under consideration is based on mean aggregation, all the local models are assigned the same LMRs. If it is based on CES, the local models are assigned LMRs based on how many times they are included in the final ensemble. Finally, if the EI ensemble is based on stacking, LMRs are determined using the permutation importance of each constituent local model (Breiman (2001) ). This importance measure is calculated as the average change in performance of the ensemble when the local model's output is randomly permuted a hundred times and as many ensembles are retrained. This calculation is conducted using the permutation_importance function in the sklearn library (Pedregosa et al. (2011) ). LMRs are then determined based on the descending order of the permutation importance values. Due to the varying number of local models and features these ranks are calculated for, all the ranks are normalized into percentile ranks that 3 range from 1/n to 1, where n is the total number of local models in the ensemble or the total number of features in the modality being considered. Next, we combined these two lists of ranks to compute the feature ranking for the EI model as shown in Algorithm 2. We first calculated the product of the LMR and LFR for each valid pair of local model and feature. We then averaged these products for each local feature across all the local models to quantify the feature's overall contribution to the EI model. All the features in all the modalities were then sorted in ascending order of this average value to determine the final ranking of the features. Note that this interpretation method is applicable to any EI model. However, in order to focus its use in our experiments, we applied it only to the best-performing EI implementation for the outcome of interest (Fig. 3) . This implementation was used to train a representative EI model on the whole dataset, which was then interpreted. To evaluate the effectiveness of EI for biomedical problems, we tested the framework on two representative problems, one each from bioinformatics and clinical informatics. These problems were protein function prediction from STRING data (Szklarczyk et al. (2015) ), and COVID-19 mortality prediction from EHR data, respectively. Below, we provide details of these problems, as well as the associated multimodal datasets used. A variety of data modalities can be used to predict the functions of proteins (Pandey et al. (2006) ), and the most accurate predictions are often obtained by integrating these modalities (Zitnik et al. (2019) ). Thus, protein function prediction (PFP) is an ideal problem to test EI's effectiveness for integrating multimodal data and producing accurate predictions. We followed the most commonly used definition of protein functions as Gene Ontology (GO) terms (Radivojac et al. (2013) ). We predicted annotations of human proteins to these terms using the diverse multimodal datasets from the STRING database (Szklarczyk et al. (2015) ). These datasets consist of pairwise functional relationships between thousands of proteins from several species, including human. These relationships are derived from the following multi-omic data sources: 1. Protein-protein interactions (PPI). 2. Interactions in curated databases. 3. Co-expression relationships between proteins. 4. Co-location in the same genomic neighborhood. 5. Co-occurrence of orthologs in other genomes. 6. Fusion of orthologs in other genomes. We tested the effectiveness of EI for predicting protein function (GO term annotations) from these multimodal STRING datasets. For all the individual datasets, the adjacency vector of each protein was used as its feature vector for PFP model training and evaluation, since this representation has been shown to be effective for automated networkbased gene/protein classification (Liu et al. (2020) ). If a certain protein was not included in any of the STRING datasets, its corresponding feature vector was assigned to be all zeros before integration and prediction. EI was executed for each GO term individually on the multimodal datasets. We compared EI's performance with those of established networkbased early integration algorithms, namely Mashup (Cho et al. (2016) ) and deepNF (Gligorijević et al. (2018) ). To maintain consistency, the versions of the STRING datasets used in our experiments (version 9.1) were the same as those used to evaluate Mashup and deepNF. The same adjacency vector as feature vector representation was used for the integrated networks generated using these methods. We then applied the local modeling algorithms listed in Section 2.1 to these integrated networks to predict GO term annotations. Finally, to assess the value of integrating multimodal data, we also predicted annotations from each of the STRING datasets individually using the same heterogeneous ensemble algorithms as included in EI (Section 2.1). These alternate integrative prediction methods served as baselines that EI was compared to. EI and the baseline methods were used to predict the annotations of 17,058 human proteins to 2,125 GO Molecular Function and Biological Process terms that were used to annotate at least fifty human proteins in September, 2018. For each GO term, the proteins annotated to it were defined as positive examples, and those not annotated to the term nor its ancestors or descendants were defined as negative examples (Mostafavi et al. (2008) ). For each term, only the positively or negatively annotated proteins were used for the evaluation of EI and the baseline methods. 4 ), with deeper terms representing more specific functions and annotating fewer proteins. These properties of a GO term, namely its depth and the number of proteins annotated with it, have been shown to substantially influence the performance of PFP methods (Zhou et al. (2019) ). Thus, to compare the performance of EI and the baseline methods more comprehensively, we also assessed how their performance varied with these properties, which were calculated using the GOATOOLS package (version 1.0.3) (Klopfenstein et al. (2018) ). The depth of GO term t was defined as the shortest path from the root of the corresponding ontology to t. Since the number of proteins annotated to GO terms can vary over several orders of magnitude, we instead considered the information content of a term. For a term t, this quantity is defined as − log 10 (p(t)), where p(t) is the probability of a human protein being annotated with t. These properties collectively quantified the specificity and abundance of a GO term, and thus, are appropriate for examining PFP performance (Zhou et al. (2019) ). The COVID-19 pandemic has infected over 300 million individuals and caused over 5 million deaths globally, as per the JHU Coronavirus Resource Center (Dong et al. (2020) ). It thus represents the most serious public health threat the world has faced in a long time. For a patient being treated for this disease, it is immensely useful for healthcare providers to have an estimate of the patient's outcomes, since they can adapt the treatment accordingly (Wynants et al. (2020) ). The multimodal data collected from these patients into their EHRs, such as demographics, comorbidities and laboratory test measurements, represent useful sources of information that can be integrated and used for predicting these outcomes. We tested EI's ability to address this need. Specifically, we used EI to predict the likelihood of an individual dying from COVID-19 (mortality), the most serious outcome, for patients treated at Mount Sinai between March 15 and October 1, 2020 from the following data modalities in the EHR system: 1. Admission (23 features): The baseline information collected from patients at hospital admission, including age, race/ethnicity, sex, and some vital signs, such as oxygen saturation and body temperature. 2. Co-morbidities (23 features): The simultaneous presence of other common conditions in patients, such as asthma and obesity, that can affect COVID-19 trajectory and outcomes. 3. Vital signs (9 features): Clinically appropriate maximum and/or minimum values of heart rate, body temperature, respiratory rate, diastolic and systolic blood pressure, and oxygen saturation level that define a patient's clinical state during hospitalization. 4. Laboratory tests (44 features): Values of laboratory tests, such as the number of white blood cells and platelets in blood, conducted on samples taken from patients over the course of their hospitalization to monitor their internal health. Since vital signs and laboratory tests were measured multiple times for each patient, we used their respective first values recorded during the first 36 hours of hospitalization to enable early outcome prediction, as recommended by another study (Vaid et al. (2020) ). Only features with fewer than 30% missing values were included in the analysis. Any remaining missing values in each modality were imputed using the KNNImpute method (Troyanskaya et al. (2001) ) with K set to 5. The categorical features in all the modalities were transformed to numerical values by one-hot encoding. The continuous features were normalized into z-scores. The resultant number of features included in each modality are specified in the descriptions above. Using the above data, we first tested EI for predicting if a patient died from COVID-19 (mortality) over the course of their hospitalization. Furthermore, it is also often important for healthcare providers to know the most likely time window within which an event such as COVID-19 mortality is most likely to occur to plan for resources and adapt patient care (Rees et al. (2020) ). To address this need, we separated the full set of patients considered in the above experiments into three subgroups based on the number of days of hospital stay during which they passed away or were discharged, i.e., 0-3 days, 3-5 days and more than 5 days. We then evaluated the EI framework for predicting mortality within each of these patient subsets. Table 1 shows the details of the four resultant mortalityrelated outcomes studied in our experiments. In all the above experiments, we compared EI's performance with that of an early integration approach proposed by another study (Vaid et al. (2020) ). This approach concatenated the feature vectors from the individual modalities for a given patient, and built mortality prediction models using XGBoost (Chen and Guestrin (2016) ), which is considered the most effective method for tabular data (Shwartz-Ziv and Armon (2021)). The default parameters specified in the Python xgboost library (Chen and Guestrin (2016)) were used for this baseline method. Additionally, similar to the PFP experiments, we also considered heterogeneous ensembles built on the individual data modalities as alternative baselines. Finally, we used the novel method described in Section 2.2 to identify the features in the above data modalities that contributed the most to the EI predictive models for each of the above mortality outcomes. All the heterogeneous ensembles, both from EI and baseline approaches, were trained and evaluated in a five-fold nested cross-validation (Nested CV) setup (Whalen et al. (2016) ) in the above experiments. In this setup, the whole dataset is split into five outer folds, which are further divided into inner folds. The inner folds are used for training the local models, while the outer folds are used for training and evaluating the ensembles. Nested CV helps reduce overfitting during heterogeneous ensemble learning by separating the set of examples on which the local and ensemble models are trained and evaluated (Whalen et al. (2016) ). The Mashup and deepNF baselines in PFP, and XGBoost baseline in the COVID-19 mortality prediction experiments were executed in the standard 5-fold cross-validation setup. All the GO terms and COVID-19 mortality outcomes predicted in our experiments were unbalanced, as is common knowledge for PFP (Radivojac et al. (2013) ; Zhou et al. (2019) ) and can be seen from Table 1 . To assess the performance of the prediction methods for the more challenging minority positive class in each experiment, we used Fmax in all our evaluations. Fmax is the maximum value of F-measure across all prediction score thresholds among the combinations of precision 5 i i "output" -2022/1/15 -23:15 -page 6 -#6 i i i i i i and recall for the positive class, and has been recommended for PFP assessment by the CAFA initiative (Radivojac et al. (2013) ; Zhou et al. (2019) ). Furthermore, since our PFP experiments involved evaluations on over 2,000 GO terms, we also statistically compared the performances of EI and the baselines using Friedman and Nemenyi tests (Demšar (2006) ) to assess the overall performance of all the methods tested. Finally, as explained in the EI interpretation method (Section 2.2), performance assessment is also needed for calculating LM Rs in stackingbased ensembles, as well as for calculating LF Rs for features included in the local models. For both these calculations, we used the Area Under the Precision-Recall Curve (AUPRC), since it was the only class imbalanceaware performance measure available as an option in the sklearn and Weka functions used (Section 2.2). The code implementing all the above methods and the data used are available at https://github.com/GauravPandeyLab/ensemble_integration. Below, we describe and analyze the results obtained in our experiments on protein function and COVID-19 mortality prediction. Fig. 4 shows the distributions of the Fmax scores of all the protein function prediction (PFP) approaches tested in Section 2.3.1. For each approach, this figure shows a box plot denoting the distribution of one Fmax score for each GO term. This score is for the implementation of the approach that performed the best for that GO term, e.g., stacking using logistic regression for Ensemble Integration for GO:0000976 (transcription cis-regulatory region binding). These results show that the performance of Ensemble Integration (EI) was significantly better than those of Mashup and deepNF (Friedman-Nemenyi FDR < 2 × 10 −16 for both comparisons), as well as each of the individual STRING data modalities (Friedman-Nemenyi FDR < 1.03 × 10 −12 ). As explained in Section 2.3.1, we also examined how the performance of EI, deepNF and Mashup varied with the depth (Fig. 5a) and information content (Fig. 5b) of the GO terms included in our evaluation. Across all depths and information content levels, EI consistently performed better than both deepNF and Mashup, exemplified by the higher median Fmax values across all the GO term subsets. The distributions of the performances of the protein function prediction approaches tested in this work. This performance was measured in terms of the Fmax score for each of the 2,125 GO terms. For EI and the individual modalities, the score for the best performing heterogeneous ensemble method for each GO term is shown here. For the Mashup and deepNF early integration methods, the score of the best local modeling algorithm for each term is shown. Collectively, these results indicated that EI was able to achieve the goals of data integration and predictive modeling for PFP more effectively than other alternate approaches. Fig 6 shows the distributions of the performance of the various implementations of EI and heterogeneous ensembles derived from the individual EHR data modalities for predicting the mortality during hospitalization outcome. Note that, unlike Fig. 4 , the performance of all the implementations of each of the approaches are shown in the corresponding box plots, since these results are only for one outcome. Also shown is the performance of the baseline XGBoost method (dotted red line). As in the PFP experiments, EI performed significantly better than the individual modalities (Wilcoxon rank-sum FDR < 0.0032). While XGBoost's performance was comparable to that of EI's median performance, five of the eleven EI ensembles tested performed better than XGBoost. The best-performing EI scored a 0.0169 (2.69%) higher Fmax than XGBoost. These results show the slight advantage of EI over early integration for this problem. Next, we expanded this evaluation to all mortality outcomes in Table 1 . Fig. 7 shows the results of the best-performing implementations of EI and heterogeneous ensembles from the individual modalities, as well as XGBoost, for each of the outcomes. Overall, the predictions from all the methods were more accurate for patients with shorter hospital stays, since these were likely more critical or healthier patients, and thus were easier to distinguish from each other. Performance was lower for later time durations, since these patients weren't as critically ill or healthy, 6 i i "output" -2022/1/15 -23:15 -page 7 -#7 i i i i i i Table 2 . Ten highest contribution features identified from the best-performing EI model for predicting mortality over hospitalization due to . The table provides detailed information on the features, including their names, the modality they were included in, and their description. The features are sorted in terms of increasing rank product scores (third column), which is the EI interpretation metric our method calculates (Section 2.2.) Rank product score and thus were more difficult to distinguish from each other. Since these subsets constituted the majority of all the patients considered, the overall performance of the methods for the mortality over hospitalization outcome was also lower. Still, the comparative trends observed in Fig. 6 were also valid here for all the outcomes, include the time-dependent ones. Specifically, EI performed better than the individual modalities for all the outcomes. Moreover, some of the EI models had higher Fmax scores than XGBoost for the time-dependent outcomes: one model for mortality within 0-3 days, one model for the 3-5 days case, and two models for mortality after 5 days. Interestingly, all these models were based on stacking, indicating that this is a particularly effective EI approach. Collectively, these results show that EI has a slight advantage over other methods for this important clinical informatics problem. Table 1 . For Ensemble Integration (EI) and the heterogeneous ensembles derived from the individual modalities, the Fmax scores of the best-performing implementation for each outcome are shown here. Also shown are the performances of the XGBoost model for each outcome. Using the method described in Section 2.2, we also identified the ten features that contributed the most to the best-performing EI model for each of the COVID-19 mortality outcomes. Table 2 lists the ten features identified for the best-performing EI model (stacking with logistic regression) for the mortality over hospitalization outcome. These features included five from laboratory tests, three from admission, one from vital signs and one from co-morbidities. This distribution was consistent with the observation that that laboratory tests produced the most accurate predictions across all the modalities (Fig. 6 ). This trend was observed for the other time-dependent mortality outcomes as well, which shared several important features with those identified for the mortality during hospitalization outcome. Most of the features found exclusively for the time-dependent outcomes were also laboratory test measurements, such as of platelet count, the estimated glomerular filtration rate (eGFR) and troponin I. The only exceptions to this observation were the relevance of acute kidney injury (co-morbidity) for the mortality within 0-3 days and within 3-5 days outcomes, and the maximum respiratory rate measured within the first 36 hours for the mortality between 0-3 days outcome. 7 The results in the above subsections illustrate the utility of our EI framework for addressing biomedical prediction problems, as well as interpreting the predictive models generated by the framework. We proposed a novel framework named Ensemble Integration (EI) to perform data integration and predictive modeling on heterogeneous multimodal data. In contrast to the more commonly used early and intermediate data integration approaches, EI adopts the late integration approach. In EI, one or more local models are derived from the individual data modalities using appropriate algorithms. These local models are then integrated into a global predictive model using effective heterogeneous ensemble methods, such as stacking and ensemble selection. Thus, EI offers the flexibility of deriving individually effective local models from each of the data modalities, which addresses challenges related to the differing semantics of these modalities. The use of heterogeneous ensemble methods then enables the incorporation of both the consensus and diversity among the local models and individual data modalities. We tested EI for the diverse problems of protein function prediction (PFP) from multimodal STRING datasets and predicting mortality due to COVID-19 from multimodal EHR datasets. In both these experiments, we compared EI's performance with those obtained from the individual data modalities, as well as established early integration methods, specifically deepNF and Mashup for PFP, and XGBoost for COVID-19 mortality prediction. In all these experiments, EI performed significantly better than the individual data modalities, showing that it accomplishes its data integration goals successfully. EI also performed better than the early integration approaches, most likely due to its ability to aggregate the complementary information encapsulated in the local models, which may be lost in the uniform integrated representations used by early and intermediate strategies. We also proposed a novel interpretation method for EI models that ranks the features in the individual modalities in terms of their contributions to the EI model under consideration. We tested this method on the representative EI models constructed for predicting the COVID-19 mortality outcomes listed in Table 1 . We found that several of the ten most important features identified for each outcome were laboratory test measurements (Table 2) , which was consistent with the observation that this modality yielded the most accurate predictions among all the individual modalities (Fig. 7) . In particular, we found that the measurements of blood urea nitrogen (BUN) and calcium in patients' blood samples were key features for all the mortality outcomes. These findings were consistent with prior research on the relevance of these measurements to mortality due to COVID-19 (Basheer et al. (2021) ; Lippi et al. (2020) ). Patients' age and minimum oxygen saturation were also found to be important features for most of the outcomes, again consistent with prior evidence (Price-Haywood et al. Although our work provided strong evidence for the utility of the EI framework for challenging biomedical prediction problems, it also has some limitations, which offer avenues for future work. First, although EI is capable of integrating both structured and unstructured data modalities, as well as a variety of local models derived from these modalities, we only tested EI with structured datasets and standard classification algorithms used to derive local models from these datasets. In the future, it would be valuable to test EI with unstructured data and specialized local models as well, such as label propagation on network data (Cowen et al. (2017) ), convolutional neural networks (CNNs) derived from the biomedical images (Shen et al. (2017) ) and recurrent neural networks (RNNs) trained on sequential or time series data (Geraci et al. (2017) ). Furthermore, in our experimental evaluations, we only compared EI to the more commonly used early data integration approaches, namely Mashup and deepNF for PFP and XGBoost for COVID-19 mortality prediction. These comparisons should be expanded to intermediate integration approaches as well. Such expanded applications and evaluations will enable a more comprehensive assessment of the capabilities of EI. There are some limitations with the EI interpretation method and its application as well. First, since predictive model interpretation can be subjective, and hence, not scalable (Doshi-Velez and Kim (2017)), we only tested this method on the COVID-19 datasets and mortality outcomes, not the considerably more numerous (2125) and diverse GO terms considered in the PFP experiments. To interpret the EI models built for these terms, it would be useful to prioritize these terms based on their relevance to the biological topic of interest. Furthermore, the implementation of the proposed feature ranking method was based on AUPRC, since this was the only class imbalanceaware measure available in the Weka and sklearn functions used for determining local ranks of features and models respectively. However, since the basic principles of these rankings are general, they can be implemented with other performance measures, such as Fmax, as well using custom code. Also, due to the use of percentile ranks in the interpretation method, it is slightly biased in favor of modalities with larger sets of features. Thus, in addition to the clinical relevance of laboratory test measurements to monitoring patients' COVID-19 status, this bias also played a a role in these features being the highest ranked for all the mortality outcomes. This limitation can potentially be addressed by considering normalized versions of model and feature ranks. In conclusion, this work presented the novel EI framework, and evidence in support of its ability to address challenging biomedical prediction problems. We also discussed challenges with the current work, and avenues for future work. Gene ontology: tool for the unification of biology. the gene ontology consortium Clinical predictors of mortality and critical illness in patients with covid-19 pneumonia Characteristics and predictors of death among 4035 consecutively hospitalized patients with covid-19 in spain Harnessing multimodal data integration to advance precision oncology Random forests • Mathematical modelling • Mathematical modelling, Dynamics of brain activity at the systems level • Clinical and translational systems biology Ensemble selection from libraries of models Getting the most out of ensemble selection Xgboost: A scalable tree boosting system Compact integration of multi-network topology for functional analysis of genes Network propagation: a universal amplifier of genetic associations Statistical comparisons of classifiers over multiple data sets An interactive web-based dashboard to track COVID-19 in real time Towards a rigorous science of interpretable machine learning. arXiv: Machine Learning Weka: A machine learning workbench for data mining Applying deep neural networks to unstructured text notes in electronic medical records for phenotyping youth depression Methods for biological data integration: perspectives and challenges deepNF: deep network fusion for protein function prediction A matrix factorization approach for integrating multiple data views Multi-omics approaches to disease Mining electronic health records: towards better research applications and clinical care Goatools: A python library for gene ontology analyses State of the field in multi-omics research: From computational needs to data mining and sharing Machine learning applications in genetics and genomics Electrolyte imbalances in patients with severe coronavirus disease 2019 (COVID-19) Supervised learning is an accurate method for network-based gene classification GeneMANIA: a real-time multiple association network integration algorithm for predicting gene function Computational Approaches for Protein Function Prediction: A Survey Scikit-learn: Machine learning in python Hospitalization and mortality among black patients and white patients with covid-19 Prevalence and risk factors for delirium in critically ill patients with COVID-19 (COVID-D): a multicentre cohort study A large-scale evaluation of computational protein function prediction Prediction of infectious disease epidemics via weighted density ensembles COVID-19 length of hospital stay: a systematic review and data synthesis A review of feature selection techniques in bioinformatics Boosting: Foundations and algorithms Generating ensembles of heterogeneous classifiers using stacked generalization Deep learning in medical image analysis Tabular data: Deep learning is not all you need Crowdsourced assessment of common genetic contribution to predicting anti-tnf treatment response in rheumatoid arthritis Developing better digital health measures of parkinson's disease using free living data and a crowdsourced data analysis challenge Learning parsimonious ensembles for unbalanced computational genomics problems String v10: protein-protein interaction networks, integrated over the tree of life Expansion of the gene ontology knowledgebase and resources Missing value estimation methods for DNA microarrays Machine learning to predict mortality and critical events in a cohort of patients with covid-19 in new york city: Model development and validation Similarity network fusion for aggregating data types on a genomic scale Large-scale protein function prediction using heterogeneous ensembles Predicting protein function and other biomedical characteristics with heterogeneous ensembles Prediction models for diagnosis and prognosis of covid-19: systematic review and critical appraisal Clinical features of COVID-19 mortality: development and validation of a clinical prediction model The CAFA challenge reports improved protein function prediction and new functional annotations for hundreds of genes through experimental screens Machine learning for integrating data in biology and medicine: Principles, practice, and opportunities. Information Fusion We thank Akhil Vaid for his technical advice on our COVID-19 data processing and experiments, and Andrew DePass for testing the EI code. The study was enabled in part by computational resources provided by Scientific Computing at the Icahn School of Medicine at Mount Sinai. This work was supported by NIH grant# R01HG011407-01A1.