key: cord-0916571-vdt2y9yi authors: Wollenstein-Betech, Salomón; Cassandras, Christos G.; Paschalidis, Ioannis Ch. title: Personalized Predictive Models for Symptomatic COVID-19 Patients Using Basic Preconditions: Hospitalizations, Mortality, and the Need for an ICU or Ventilator date: 2020-08-22 journal: Int J Med Inform DOI: 10.1016/j.ijmedinf.2020.104258 sha: d1b8af6cf78c3a30a6ef7794d23798dd6331a513 doc_id: 916571 cord_uid: vdt2y9yi Abstract Background The rapid global spread of the SARS-CoV-2 virus has provoked a spike in demand for hospital care. Hospital systems across the world have been over-extended, including in Northern Italy, Ecuador, and New York City, and many other systems face similar challenges. As a result, decisions on how to best allocate very limited medical resources and design targeted policies for vulnerable subgroups have come to the forefront. Specifically, under consideration are decisions on who to test, who to admit into hospitals, who to treat in an Intensive Care Unit (ICU), and who to support with a ventilator. Given today’s ability to gather, share, analyze and process data, personalized predictive models based on demographics and information regarding prior conditions can be used to (1) help decision-makers allocate limited resources, when needed, (2) advise individuals how to better protect themselves given their risk profile, (3) differentiate social distancing guidelines based on risk, and (4) prioritize vaccinations once a vaccine becomes available. Objective To develop personalized models that predict the following events: (1) hospitalization, (2) mortality, (3) need for ICU, and (4) need for a ventilator. To predict hospitalization, it is assumed that one has access to a patient’s basic preconditions, which can be easily gathered without the need to be at a hospital and hence serve citizens and policy makers to assess individual risk during a pandemic. For the remaining models, different versions developed include different sets of a patient’s features, with some including information on how the disease is progressing (e.g., diagnosis of pneumonia). Materials and Methods National data from a publicly available repository, updated daily, containing information from approximately 91,000 patients in Mexico were used. The data for each patient include demographics, prior medical conditions, SARS-CoV-2 test results, hospitalization, mortality and whether a patient has developed pneumonia or not. Several classification methods were applied and compared, including robust versions of logistic regression, and support vector machines, as well as random forests and gradient boosted decision trees. Results Interpretable methods (logistic regression and support vector machines) perform just as well as more complex models in terms of accuracy and detection rates, with the additional benefit of elucidating variables on which the predictions are based. Classification accuracies reached 72%, 79%, 89%, and 90% for predicting hospitalization, mortality, need for ICU and need for a ventilator, respectively. The analysis reveals the most important preconditions for making the predictions. For the four models derived, these are: (1) for hospitalization:age, pregnancy, diabetes, gender, chronic renal insufficiency, and immunosuppression; (2) for mortality: age, immunosuppression, chronic renal insufficiency, obesity and diabetes; (3) for ICU need: development of pneumonia (if available), age, obesity, diabetes and hypertension; and (4) for ventilator need: ICU and pneumonia (if available), age, obesity, and hypertension. who should be tested (for the virus and/or antibodies) and at which frequency, (ii) providing more accurate estimates of who is more likely to be hospitalized and the type of care they may need, (iii) informing plans for staffing, resources, and prioritizing the level of care in extremely resource-constrained settings. Equally importantly, as societies adapt to the pandemic, predictive models can (i) assess individual risk so that social distancing measures can transition from "blanket" to more targeted (e.g., deciding who can return to work, who is advised to stay at home, who should be tested, etc.) and (ii) direct policy decisions on who should receive priority for vaccination, which will be critical as initial vaccine production may not suffice to vaccinate everybody. In an attempt to understand better the disease, several predictive models have been developed during this pandemic [6] . One of the limitations of all of these predictors is their highrisk of bias given their small sample sizes. In fact, out of the 66 predictors summarized in [6] , the mean and standard deviation of the sample size and test size used are 443.5 ± 560 and 155 ± 276 respectively. In turn, our work provides with a less biased predictor by employing a dataset which is 4.7 times larger than the biggest dataset reported in [6] . To develop predictive models, we leverage supervised machine learning methods that learn from given examples of predictive variables and associated outcomesthe so called training set. Performance is then evaluated on a separate test set. In the specific application of interest, we will focus on classification, a setting where the outcome is binary, e.g., someone is hospitalized or not. Many models have been used to predict a patient admission to a hospital, mortality and other health care applications based on comorbidities. Some examples include: predicting morbidity of patients with chronic obstructive pulmonary disease [7] , febrile neutropenia [8] , as well as classifying the hospitalization of patients with preconditions on diabetes [9] , heart disease [10, 11] , and hospital readmission for patients with mental or substance use disorders [12] . Recent advances in the machine learning literature have suggested that sparse classifiers, those that use few variables (e.g., 1-regularized Support Vector Machines), have stronger predictive power and generalize better on out-of-sample data points than very complex classifiers [13] . Related work has shown that regularization is equivalent to robustness, that is, learning models which are robust to the presence of outliers in the training set [14] . Moreover, the benefit of using sparse predictors is the enhanced interpretability they provide for both the model and the results. Construct data-driven predictive models using data from patients tested for SARS-CoV-2 to predict if a patient will (1) be hospitalized, (2) succumb to the disease, (3) need treatment in an ICU, and/or (4) need a ventilator. To train and test these classifiers we use a public dataset [15] made available by the Mexican government that contains individual information on: demographics (e.g., location), preconditions (e.g., hypertension) and outcomes (e.g., admission to an ICU) for every person who has been tested for SARS-CoV-2 in Mexico.  We provide descriptive statistics of the distribution of hospitalized and deceased patients given basic information on preconditions and demographics.  We develop interpretable models that not only predict the outcomes but also quantify the role of various variables in making these predictions.  The models we develop leverage data from Mexico. This can motivate additional work using the same data, while the models could be applicable to other Latin American countries with similar population characteristics. This adds to existing work using Electronic Health Records which has focused on patients in the US, Europe, or Asia. The remainder of the paper is organized as follows: In Section 2 we describe the data used accompanied by descriptive statistics and preprocessing procedures. In Section 3 we describe the binary supervised classification models used and the performance evaluation metrics employed. In Section 3, we present the main results. Discussion of the results can be found in Section 4 and Conclusions in Section 5. We use a dataset that has been open for the general public by the Mexican Government (and updated daily) [15] . These data include information about every person who has been tested for SARS-CoV-2 in Mexico. They include demographic information such as: Age, Location, Nationality, the use of an indigenous language; as well as information on prior medical conditions, including whether the patient has: diabetes, chronic obstructive pulmonary disease (COPD), asthma, immunosuppression (e.g., due to treatment for cancer or auto-immune conditions [16] ), hypertension, obesity, pregnancy, chronic renal failure, other prior diseases, and whether was or is using tobacco. In addition, the data report the dates on which the patient first noticed symptoms, the date when the patient arrived to a care unit, and the date when the patient was deceased (if applicable). Finally, it contains fields showing whether the patient was hospitalized, has pneumonia, needed a ventilator, was treated in an ICU, as well as the result of the SARS-CoV-2 test. To confirm a case, the Ministry of Health in Mexico requires that, in addition of being tested positive, the patient presents at least two of: cough, fever or headache, and at least one of: dyspnea, arthralgia, myalgia, odynophagia, rhinorrhea, conjunctivitis or chest pain. More technical details on the surveillance model used are provided in [17] . As of May 1st, 2020, the data contained more than 91,179 observations out of which more than 20,737 account for positive tests, around 15,000 tests were being processed, and the rest are negative test results. Table 2 -1 provides a more precise description of the dataset. We provide plots that help us observe trends in the data. We begin by disaggregating data into age groups. In the lower plot of Figure 1 the number of observations of patients having a positive test or waiting their result per age is shown. In addition, the upper bar plot denotes the percentage of the patients in a certain age range who have been hospitalized. This information is aligned with the current knowledge on COVID-19, which indicates that older people have higher risk of being hospitalized. Also, this plot suggests that the risk of being hospitalized increases linearly from the age of thirty up to seventy-five and then plateaus. We ran an ordinary linear regression (OLS) to calculate the rate at which the percentage of hospitalization increases for every additional year of age. The rate results to be 0.014 with an R 2 equal to 0.99. This suggests that the risk of hospitalization increases by approximately 1.4% for every year of age between 30 and 75 years old. Next, in Figure 2 we report the fraction of patients who have been hospitalized, deceased, needed an ICU or a ventilator given a certain precondition, e.g., in the upper-left box we divide the number of hospitalized patients with pneumonia by the total number of patients with pneumonia. We observe that for both hospitalizations and deaths, preconditions such as chronic renal insufficiency, COPD, diabetes, immunosuppression, cardiovascular disease and hypertension are critical. Nevertheless, even though this gives us information about the risk of a precondition, it does not include the sensitivity regarding how age and preconditions affect a patient with COVID-19. J o u r n a l P r e -p r o o f To complement the previous table, we report the percentage of the hospitalized by age group and by existing preconditions in Figure 3 . To that end, we create age groups for every five years and report results for groups with at least ten observations, otherwise the bin is left blank. On the top row of the table, we include the statistic for a patient without any preconditions. As an example, the top-left entry reports the ratio of the number of patients between 0-5 years old without preconditions who have been hospitalized divided by the number of patients between 0-5 years without preconditions who may or may not have been hospitalized. We observe that chronic renal insufficiency, diabetes, and immunosuppression are among the preconditions that are associated with a higher hospitalization rate. Finally, we present histograms reporting the lag times among various states of the disease for the Mexican population. For this analysis, we separate the data in three groups: individuals with ages between 0-20, 20-50, and patients over 50 years old. In Figure 4 (left), we plot the distribution of the number of days between the onset of symptoms and a subsequent hospitalization. Figure 4 (center) depicts the distribution of time (days) between hospital admission and death. Interestingly, we observe that a large portion of the patients who were hospitalized died the same day they were admitted. This could be explained either by a healthcare system working at capacity in which only seriously-ill patients are admitted or by the abrupt deterioration of a patient's condition [18, 19] and should be further investigated. The rest of the distribution behaves like the tail of a Weibull distribution with very few patients being hospitalized for more than three weeks. Finally, Figure 4 (right) shows the distribution of the number of days between the onset of symptoms and death (the mean is 9.8 days). We found a few outliers which are easily identified, for example, the pregnancy of male patients, the date of death of a patient being earlier than the day the patient was admitted to the hospital. Such data points were removed from the dataset. The data contain precondition features reported as categorical. Specifically, each of these precondition features takes the value yes, no, unknown or unspecified. We generate onehot encoding for all these features. One-hot encoding converts the categorical feature to multiple binary variables by creating auxiliary variables that help distinguish between the different categories of a feature. For the case of our data, one-hot encoding generates three binary variables for each specific precondition; these variables (as opposed to categories) are: no, unknown and unspecified. Then, for each observation, at most one of these variables will be active, pointing to the correct value for the original feature. If none of the three is active, then the value of the precondition is yes. We find and delete variables that are highly correlated since they, in general, provide similar information. Specifically, we compute pairwise correlations among the variables, and remove one variable from each highly correlated pair (using a threshold of 0.8 for the absolute correlation coefficient). We found that the correlated binary features were the ones corresponding to unknown or unspecified for preconditions. This is because observations that contain an unknown or unspecified value, typically have this same value for all preconditions (not just for one), indicating potential issues in data gathering. Hence, we remove all these auxiliary variables denoting unknown or unspecified preconditions. In this section, we briefly introduce the methodologies used to build the binary classifiers. For each model, we train the classifier using four different supervised classification methodologies: sparse Support Vector Machines (SVM), sparse Logistic Regression (LR), Random Forests (RF) and gradient boosted decision trees (XGBoost). For healthcare applications, the first two are preferable due to their interpretability. In turn, the last two are the state-of-the-art classification algorithms today and will serve as a basis to compare the accuracy of the interpretable methods with the non-interpretable benchmark models. Appendix B provides details on these methods, particularly because the robust/sparse LR and SVM formulations are not standard. Classifiers based on few variables are desirable because they have stronger predictive power, generalizing better out-of-sample, and offering enhanced interpretability [20, 21] . Aiming to reduce the number of variables, we employ a Recursive Feature Elimination (RFE) procedure [22] to find the variables that optimize a given performance metric. The general framework of this algorithm begins by building a classifier using all the features and computing an importance score for each predictor. In the case of Logistic Regression (or Linear SVM), we use as important score the absolute value (or magnitude) of the linear coefficient of feature . After this step, the least important feature (the one with the smallest | |) is deleted from the dataset. We repeat iteratively this process until we are left with one feature. Then, for each of these iterations we report the performance of the model (using cross-validation over the training set) and we pick the set of features that maximize this value. Additionally, at each iteration, we use the same cross-validation process to tune the hyper parameters of the classifier to achieve the best performance. In this work, we use LR to eliminate variables based on their coefficients as described earlier, as it gives a clear and interpretable meaning of the score for each variable. At each iteration we use a stratified ten-fold cross-validation (over the training set) to estimate the AUC performance until we are left with one variable. Finally, we pick the features for which we obtain the model with the maximum AUC value. This subset of variables is then used to train all the predictive models. The primary objective of learning a classifier is to maximize the prediction accuracy (or equivalently minimize a loss function), and in our health care setting offer interpretability of the results. We characterize the prediction accuracy of a classifier using two commonly used metrics: (1) the false positive (or false alarm) rate which measures how many patients were predicted to be in the positive class, e.g., hospitalized, while they truly were not, as a fraction of all negative class patients. In the medical literature, the term specificity is often used and it equals 1 minus the false positive rate; and (2) the detection rate that captures how many patients were predicted to be on the positive class while they truly were, as a fraction of all positive class patients. This term is often referred to as sensitivity or recall. Another term commonly used is precision defined as the ratio of true positives over true and false positives. A single metric that captures both types of error is the Area Under the Curve (AUC) of the Receiver Operating Characteristic (ROC). ROC plots the detection rate (or sensitivity or recall) over the false positive rate. A naïve random selection (assigning patients to classes randomly) has AUC of 0.5 while a perfect classifier an AUC of 1. To complement the AUC metric, we report the accuracy that computes the ratio of the number of correct predictions over all predictions within the test set. In addition to the ROC AUC and accuracy, and rather than evaluating the precision and recall for both the positive class and the negative class, we report a single metric, the weighted F1-score. Specifically, the F1-score is the harmonic mean of precision and recall and can be computed for both the positive and negative classes. The weighted F1-score is just an average of the per class F1-scores weighted according to the number of (test) samples in each class. We are interested in this performance metric because it is as important to accurately predict who is not likely to have a specific outcome (e.g., hospitalized), in addition to who will. For example, one can ease restrictions on those who are predicted to have lower risk. In fact, having more false positives corresponds to being more conservative with patients by assigning higher-risk profiles, and what is needed is striking the right balance between being conservative vs. having a lot of false positives. The weighted F1score is one appealing way of quantifying this trade-off. For individual variables and each different model, we also report the Odds Ratio (OR), which indicates how the odds of observing the outcome are scaled by having that variable take the value 1 (vs. 0), while controlling for all other variables in the model. We finally emphasize that all metrics we report are computed on a randomly selected test set of patients (i.e., out-of-sample) which corresponds to 30% of the observations and has not been used for training the models. In addition, all metrics were calculated using a discriminant classification threshold which was selected by optimizing the AUC and reported in Table 2 : Summary of results of all models using LR. We build binary classification models to predict hospitalization, mortality and the need for an ICU or ventilator. At a minimum, all models use a set of base features composed by: age, gender, diabetes, COPD, asthma, immunosuppression, hypertension, obesity, pregnancy, chronic renal failure, tobacco use, other disease, as well as the SARS-CoV-2 test result, which is either positive or pending (we exclude all negative cases to train our models). In this section, we provide a summary of the results while in Appendix A we provide all results. Our first model predicts if a patient who has tested positive or is waiting for the test result will be hospitalized given their base features. This model has a moderate accuracy for all methodologies employed which accounts for an AUC of 0.74 and an accuracy of classifying 72% of the observations correctly. An interesting observation is that SVM and LR performs better than RF and XGBoost. The coefficients of the SVM and LR models have the same trend and suggest that the features that contribute the most for predicting the hospitalization of a patient are: age over 80 (OR=3. have a much smaller impact. It is however possible that some of these variables have smaller coefficients because the effect is captured by another highly correlated variable (e.g., obesity and diabetes). We explore two models to predict mortality. The first model assumes we only know the base features of a patient whereas the second model includes variables that indicate if the patient has been hospitalized or not, has pneumonia, or has needed an ICU or ventilator. The reason to consider the first model is to have a classifier which identifies which patients are the most vulnerable prior to hospitalization, while the second model predicts the mortality of an individual in the hospital by using information on how the disease is progressing. In order to have a more balanced dataset and to detect better the deceased class, we ran this model only on the observations of patients who have been hospitalized and have been tested positive or are waiting for their test result. For the model which considers the case that only uses the base features of a patient (prior to attending a healthcare facility), we are able to predict with 79% accuracy and with an AUC equal to 0.63 the mortality of a patient. Moreover, when we include more information about the hospitalization, pneumonia, ICU, and ventilation, the classification task achieves a similar accuracy but a higher detection rate of 0.701 (an increase of ~12% in detection). Both interpretable models, LR and SVM, suggest that the variables that are critical for predicting mortality are the patient's age, gender, immunosuppression (OR=1.68), chronic renal insufficiency (OR=1.46), obesity (OR=1.4) and diabetes (OR=1.32). For the model that has more features, as expected, information about the need for ventilator and ICU are highly relevant when predicting mortality. Similar to the mortality case, we train two classification models to predict the need for an ICU. The first model predicts the need for an ICU bed using the base features and assumes that we don't know if the patient will or will not develop pneumonia. This might serve for planning purposes, as it will help us predict which individuals are more likely to need an ICU in case they contract SARS-CoV-2. This model achieves an accuracy of 89% with an AUC of 0.55 (XGBoost). Additionally, when we include information about the development of pneumonia, the AUC of the model increases by about 10% to 0.64, highlighting the importance of using the most recent information of a patient while predicting its outcome. In these cases, SVM and LR suggest that information on: age, development of pneumonia (OR=4.13), if available, diabetes (OR=1.23), obesity and hypertension are among the most important variables to predict the need for an ICU. In the same way as in the mortality and ICU models, we develop two models to predict the need for a mechanical ventilator given that a patient is either a confirmed or suspected COVID-19 case. The first model evaluates the situation prior to knowing if patient has developed pneumonia or needs an ICU. The accuracy reached by this model is higher than both the mortality and the ICU models, achieving an accuracy of 90% and an AUC of 0.58. In addition to this model, the second instance uses information about the development pneumonia and the admission to an ICU. As expected, this additional information is relevant for predicting ventilation need. It increases its accuracy to 92% and the AUC to 0.86. Moreover, both models classifying the need for a ventilator show that information on ICU (OR=15.5) and pneumonia (OR=9.1), if available, age, gender, chronic renal insufficiency (OR=1.5), obesity (OR=1.4), hypertension (OR=1.16) and diabetes (OR=1.12) are the most relevant features for predicting the need for a mechanical ventilator. To summarize and provide interpretability we report in " Table 2 : Summary of results of all models using LR." the performance metrics for all the models and in " Table 3 : Odds ratios for all models, considering LR-l1." the odds ratio for each model variables using LR. We observe that the coefficients of both interpretable models (SVM and LR) are consistent and have an accuracy comparable, or higher than RF and XGBoost. J o u r n a l P r e -p r o o f Overall, the models we develop range from moderately to significantly accurate. Predicting hospitalizations appears harder just based on the basic variables at our disposal, particularly considering all patients who have a positive test or with a test pending. Potential additional features are at play including state of health (measured through detailed lab results) and the viral load they were exposed to. Furthermore, a number of hospitalizations are driven by socioeconomic factors, e.g., the living arrangements of a patient and whether he/she can pose infection risk for many others. Still, an AUC of 0.75 is significantly better than random and the results could help tighten estimates on the number of hospitalizations expected. From an actionable and planning perspective, predicting ICU treatment and ventilator need are quite useful. These models can be quite accurate, achieving accuracies of 89% and 90%, respectively, when information on how the disease is progressing is taken into account (e.g., development of pneumonia). Similarly, the mortality model can achieve an accuracy of 76%. An interesting observation is that interpretable models (such as LR and SVM), when used in conjunction with robustness/regularization approaches and elaborate feature selection procedures, can lead to performance that is comparable, if not better than more complex and expensive classifiers. The significant advantage of the former models is that they are interpretable and provide information on which variables drive the predictions. This study has some limitations. It is important to emphasize that the dataset used in this work lacks critical information (such as lab results, vital signs, among others) to be able to provide a clinical understanding of COVID-19. Rather, the focus of this work is to help inform decisions on how to best allocate limited medical resources, and to help design targeted policies for vulnerable subgroups which might not have access to clinical and lab assessments. Interesting patterns can be observed in our results, motivating further research directions in resource allocation during a pandemic. For example, our results suggest that pregnancy is an important variable for predicting hospitalization but not mortality, ICU or ventilation, potentially indicating a bias towards being more conservative and hospitalizing pregnant women when they may not need it. Readers should also be aware that, due to the insufficient testing resources in Mexico, the dataset might be biased toward overestimating deaths. While the dataset may reflect all deaths, it does not include mild-moderate COVID-19 cases as these are never tested. Another limiting factor is that the dataset does not include specific dates at which hospitals discharge patients, which is of high importance to assess the utilization of medical equipment. Finally, to the extent that these risk models can be used to prioritize the use of resources, we understand that medical risk is not the only factor in making such decisions. Nevertheless, in order to quantify medical risk one can leverage the models presented in this work. We develop models to identify the medical risk of a patient with (or suspected for) COVID-19. We hope this work can help hospitals and policymakers to distribute more effectively their limited resources including tests, ICU beds and ventilators, as well as, to motivate countries and healthcare systems to standardize and share data with the medical informatics community. Moreover, we hope this research spreads the knowledge of the existence of this public dataset and motivates researchers to work with these data. Finally, we hope that risk models are taken into account to fine-tune social distancing advisories, moving from "blanket" to risk-based, as well as prioritizing vaccine distribution to the more vulnerable and to those who need to interact with the more vulnerable. For the sake of reproducibility and to facilitate the analysis for further research we have made our models and results available on a Github repository [23] . WHO announces COVID-19 outbreak a pandemic An interactive web-based dashboard to track COVID-19 in real time How Do Hospitals Decide Who Gets Treatment? -The New York Times The Hardest Questions Doctors May Face: Who Will Be Saved? Who Won't? -The New York Times Prediction models for diagnosis and prognosis of covid-19: systematic review and critical appraisal Predicting the morbidity of chronic obstructive pulmonary disease based on multiple locally weighted linear regression model with K-means clustering Predicting in-hospital mortality of patients with febrile neutropenia using machine learning models Predicting diabetes-related hospitalizations based on electronic health records Predicting Chronic Disease Hospitalizations from Electronic Health Records: An Interpretable Classification Approach Federated learning of predictive models from federated Electronic Health Records Predicting Hospital Readmission in Patients with Mental or Substance Use Disorders: A Machine Learning Approach Feature selection, L1 vs. L2 regularization, and rotational invariance A Robust Learning Approach for Regression Models Based on Distributionally Robust Optimization Calculadora de complicación de salud por COVID -19 Lineamiento estandarizado para la vigilancia epidemiológica y por laboratorio de la enfermedad respiratoria viral Clinical progression of patients with COVID-19 in Shanghai, China Clinical Characteristics of 138 Hospitalized Patients With Clinical Prediction Models: A Practical Approach to Development, Validation, and Updating Comparison of variable selection methods for clinical predictive modeling Gene selection for cancer classification using support vector machines Clinical Characteristics of Coronavirus Disease 2019 in China CT imaging features of 2019 novel coronavirus (2019-NCoV) Chest CT Findings in Coronavirus Disease-19 (COVID-19): Relationship to Duration of Infection Unveiling COVID-19 from Chest X-ray with deep learning: a hurdles race with small data Sensitivity of Chest CT for COVID-19: Comparison to RT-PCR Support-vector networks Pattern Recognition and Machine Learning (Information Science and Statistics) An Interior-Point Method for Large-Scale 1-Regularized Logistic Regression Random forests Classification and regression trees XGBoost: A scalable tree boosting system Tree Boosting With XGBoost -Why Does XGBoost Win "Every" Machine Learning Competition? Research partially supported by the NSF under grants IIS-1914792, DMS-1664644, and CNS-1645681, by the ONR under MURI grant N00014-19-1-2571, and by the NIH under grant 1R01GM135930.The authors would like to thank Diana Sverdlin-Lisker at the Massachusetts Institute of Technology for useful discussions. What this study added to the knowledge Due to the fast spread of COVID-19, a lot of attention has been devoted to measuring and predicting the spread. There have also been anecdotal reports on certain prior conditions that appear to lead to more severe disease.This work is among the first to use data to develop explicit models predicting hospitalization, ICU treatment, ventilator use, and mortality for individual patients.Most research related to COVID-19 has been done in countries and communities where the virus hit first. These include China, Italy, Spain, US.Our research focuses on the Mexican population, which has particular characteristics of interest to Latin American countries with similar socio-economic conditions and health care systems that may become more congested due to COVID-19. Most research related to COVID-19 that employs Machine Learning techniques has been focused on learning from complex data sources such as chest scans [24] [25] [26] [27] [28] .We focus on a basic set of preconditions that are known for the vast majority of the population without the need to attend a medical facility. Hence, the risk metrics we develop can be computed for anyone susceptible to COVID-19, helping to prioritize testing, care, and postsurge social distancing and vaccination policies. S.W.-B. co-designed and analyzed the methods, co-wrote the manuscript, performed the analysis, and produced results and figures. C.G.C co-led the study, co-designed the methods, and commented on the manuscript. I.C.P. co-led the study, co-designed the methods, and co-wrote the manuscript. J o u r n a l P r e -p r o o f The authors have no financial or personal relationships with other people or organizations that could inappropriately influence (bias) their work.J o u r n a l P r e -p r o o f