key: cord-0154152-0rcoj698 authors: Maitre, L'ea; Guimbaud, Jean-Baptiste; Warembourg, Charline; Guil-Oumrait, Nuria; Consortium, The Exposome Data Challenge Participant; Petrone, Paula Marcela; Chadeau-Hyam, Marc; Vrijheid, Martine; Gonzalez, Juan R.; Basagana, Xavier title: State-of-the-Art Methods for Exposure-Health Studies: results from the Exposome Data Challenge Event date: 2022-02-03 journal: nan DOI: nan sha: fd6c60202ad7ca261d4aaa82d5a650c9f7cb24b0 doc_id: 154152 cord_uid: 0rcoj698 The exposome recognizes that individuals are exposed simultaneously to a multitude of different environmental factors and takes a holistic approach to the discovery of etiological factors for disease. However, challenges arise when trying to quantify the health effects of complex exposure mixtures. Analytical challenges include dealing with high dimensionality, studying the combined effects of these exposures and their interactions, integrating causal pathways, and integrating omics layers. To tackle these challenges, ISGlobal Exposome Hub held a data challenge event open to researchers from all over the world and from all expertises. Analysts had a chance to compete and apply state-of-the-art methods on a common partially simulated exposome dataset (based on real case data from the HELIX project) with multiple correlated exposure variables (P>100) arising from general and personal environments at different time points, biological molecular data (multi-omics: DNA methylation, gene expression, proteins, metabolomics) and multiple clinical phenotypes in 1301 mother-child pairs. Most of the methods presented included feature selection or feature reduction to deal with the high dimensionality of the exposome dataset. Several approaches explicitly searched for combined effects of exposures and/or their interactions using linear index models or response surface methods, including Bayesian methods. Other methods dealt with the multi-omics dataset in mediation analyses using multiple-step approaches. Here we discuss the statistical models and provide the data and codes used, so that analysts have examples of implementation and can learn how to use these methods. Overall, the exposome data challenge presented a unique opportunity for researchers from different disciplines to create and share methods, setting a new standard for open science in the exposome and environmental health field. main advantage over traditional 'one-exposure-one-disease' study approaches is that it provides an unprecedented conceptual framework for the study of multiple environmental hazards (urban, chemical, lifestyle, social) and their combined effects. Given the increasing availability of complex environmental health data due to the emergence of new technologies (such as electronic health records, high throughput omics platforms, wearable sensors, etc), there is a need for more advanced statistical approaches that focus on complex mixtures of exposures. To address the numerous challenges that come with the analysis of such complex data and to promote interdisciplinary collaboration between researchers from around the world, ISGlobal hosted a 3-day online data challenge in April 2021 entitled "the Exposome Data Challenge Event". This event gathered a widely diverse scientific audience of 307 participants, including environmental epidemiologists, biostatisticians and computational scientists, to discuss state-of-the-art statistical methods for studying exposome-health associations. First, participants were invited to submit an abstract describing their team, the challenge(s) and the method they would apply on a common partially simulated exposome dataset (based on real case data from the HELIX project). The planning committee selected a total of 25 abstracts out of 39 based on method clarity, novelty, relevance for the exposome field and challenges presented. Second, the selected participants were invited to apply their method on the dataset during a month leading to the event. Third, they presented at the event their method's statistical background, type of research question(s) it best addressed, and their results. At the end of the event the committee and the audience voted for the best presentations based on clarity, novelty and relevance. Finally, the participants made their code available on the github account (González, 2021) of the event. Our main objective was to promote innovative statistical, data science, or other quantitative approaches to study the health effects of complex multi-dimensional exposures and high throughput omics measurements. To illustrate the methods discussed as part of this data challenge, it was important to create a publicly available real case scenario dataset (based on the HELIX project database). This dataset consisted of multiple correlated variables (P>100 exposure variables) arising from general and personal environments at different time points of early life, biological molecular data (multi-omics: DNA methylation, gene expression, proteins, metabolomics) and multiple clinical phenotypes. The population was drawn from the multi-center HELIX project, where the cohorts of origin represented the main confounding structure in the dataset (Maitre et al., 2018; Vrijheid et al., 2014) ). The participants were offered an opportunity to test their statistical methods of choice addressing one or several key challenges: a) the high dimensionality of the data, b) combined effect of exposure or mixtures, c) the omics data integration and the d) causal structure in the exposome. Participants were encouraged to accommodate in their approaches some of the particularities of the data (e.g. multi-cohort, count responses, categorical and continuous exposure variables, exposures measured at two time points, etc). Visualization of the results was also a key point across all the challenges listed above. This report outlines the approaches presented at the event, which represent useful computational, conceptual, and statistical models for analyzing high dimensional exposome datasets and health outcome associations. In collaboration with the event committee and the selected participants, we discuss the different techniques. The exposome data provided for this challenge came from the HELIX subcohort database and were partially simulated. The HELIX study (Maitre et al., 2018; Vrijheid et al., 2014) represents a collaborative project across six established and ongoing longitudinal population-based birth cohort studies in six European countries (France, Greece, Lithuania, Norway, Spain, and the United Kingdom). From the 31 472 mother-child pairs included in the cohorts, a subcohort of 1301 mother-child pairs were followed up with measurements of biomarkers, omics signatures and health outcomes at 6-11 years of age. The data provided for the challenge came from this subsample, but it was partially simulated to respond to policies of data anonymization for privacy protection in the cohorts. In detail, for the set of health outcomes and exposures, we conducted the following process: for each participant, a total of 50 random variables (different for each subject) were converted into missing values and then imputed (in successive rounds of 10 variables at a time) with the method of chained equations. Thus, for each participant, some of the values in the provided dataset were real and some were simulated, in a way that precluded knowing what is real data and what is simulated. Omics data instead were kept intact but annotations of genes and metabolites were not provided. We provided an imputed dataset in which all missing values in the original data were imputed by the chained equations method. Exposure data were transformed (e.g. logarithmic, square root, tertiles) to achieve symmetric distributions with a homogenous range of values (Annex 1). The original raw data are available on request subject to ethical and legislative review. The "HELIX Data External Data Request Procedures" are available with the data inventory in this website: http://www.projecthelix.eu/data-inventory. The datasets are available in the github repository of the challenge event and transcriptomic and Epigenomics through a FigShare account: https://figshare.com/account/home#/projects/98813. An overview of available data is shown in Figure 1 and the complete codebook description is available in Annex 1. It includes more than two hundred environmental exposure variables, 13 covariates, six health outcomes [body mass index (BMI), asthma, birth weight, neurobehaviour, intelligence quotient (IQ)], and omics data (serum metabolome, urine metabolome, gene expression, methylation). Exposure, covariate and health outcome datasets contain both continuous and categorical values; health outcome data additionally included count data. Exposure and omic datasets both included highly correlated features (correlation >0.8). Figure 1: Overview of the data available during the challenge. The data were based on the HELIX project which collected exposome, omics and health data from six mother-child cohorts across Europe in 1301 participants (Maitre et al., 2018) Challenges The main challenges that were addressed during the workshop were as follows: 1/ The high dimensionality of exposure data and, more precisely, methods to reduce it while minimizing the loss of useful information regarding health associations and combined effects. The exposome dataset available during the challenge consisted of a large number of environmental exposure variables and a multi-omics dataset (approximately 0.5 M features), with a high inter-variable correlation, but a small sample size (N = 1301), typical of exposome studies. 2/ The combined effect of exposure or mixtures. Researchers are interested in studying individual and combined effects of a large number of exposures together accounting for their potential interactions. Effects of environmental exposures may be small when taken individually, but their aggregation may lead to a significant alteration of the health outcome of interest, leading to cocktail effects that researchers want to investigate. 3/ Omics data integration. Omic data may be used, in addition to exposure data, in order to provide causal inference on the link between exposome and health. The challenge here was to incorporate one or several of the different omic layers available, with the purpose of finding patterns that can explain variations in one or more health outcomes and analysing how the exposome and the omes interact with regard to these outcomes. The method used in this context must be able to maximise omics data predictive power with very high dimensional data and small sample size (N<
p), applied some sort of dimensionality reduction techniques, as summarized in Table 2 . These techniques can be split in two categories: 1) feature extraction techniques (Khalid et al., 2014) , which consist in computing derived variables that are functions of the original ones and have a smaller dimension, but retain/extract most of the information contained in the original feature space, and 2) feature selection techniques which consist in selecting a subset of the original variable set, while keeping most of the information contained in the whole variable space. Among the feature extraction techniques that were used during the event, we can further define different types: 1) techniques that calculate summary measures (indices) based on the creation of weighted combinations of exposures that predict the health outcome; 2) feature projection techniques such as Principal Component Analysis (PCA) (Jolliffe, 1986) , Factor Analysis (FA), Principal Component Pursuit (PCP) (Candes et al., 2009) or Multiple Correspondence Analysis (MCA) (Blasius and Greenacre, 2006) to represent the data in a low dimensional space; and 3) techniques that provide clustering of participants sharing a similar exposome profile, which could be predictive of the outcome, i.e. supervised, or not. Feature selection approaches used during the event can also be divided into different types of selections: 1) based on correlation with the outcome of interest, using Pearson's correlation as a screening approach; 2) based on regularisation (LASSO, Elastic-net regression) by shrinking the less relevant features' coefficients; 3) based on feature importance which reflects the impact of a given feature on the model predictions through permutation (Altmann et al., 2010) , random forest (Breiman, 2001) , or regression coefficients. Other statistical tests exist for features selection (ex: chi-squared test, ANOVA, etc..) but were not used during this challenge. Some analysts made an a priori selection, by choosing to focus only on a particular subset of exposures, e.g. lifestyle exposure, based on prior knowledge for causal models (25) or model abilities (continuous-only-exposures). Ten studies implemented feature selection but not feature extraction, five implemented feature extraction but not feature selection, and nine applied both techniques (Table 2) . In the group of index models, Gennings (4) et al. presented the weighted quantile sum regression (WQSR) method (Carrico et al., 2015) , its history and recent extensions. WQSR builds a new index, which is a weighted average of the initial exposures (previously categorised into quantiles as a way to standardise the data and prevent the effect of influential observations). The new index is regressed against a health outcome, producing a single regression coefficient. The weights to build the index, which incorporate directionality constraints (e.g., all variables are expected to produce negative effects on the health outcome), are estimated simultaneously with the regression coefficients. This technique assumes additive effects of the different pollutants. Some presented extensions included models to produce strata-specific weights and regression coefficients, combining two indices (one for each directionality) in the same model, or using resampling to improve the properties of the method. Carli (6) et al. presented the application of Bayesian Group Index Regression (BGIR), a Bayesian equivalent to WQSR that does not use directionality constraints and allows multiple indices (based on groups of exposures) in the same model (Wheeler et al., 2021) . The method R package BayesGWQS is available here. In the application, they included several indices according to exposure families, and each exposure family could have two indices if the family contained both exposures that were positively and negatively correlated with the outcome. They conducted analyses separately by cohort and included serum metabolomics data as an additional group of exposures. Midya (2) et al. presented the application of LASSO-type Bayesian Weighted Quantile Sum Regression (LBWQSR), which is similar to BGIR, but introducing LASSO and Elastic net penalties to prevent overfitting (Xu and Ghosh, 2015) . In the group of response surface methods, Chattopadhyay (5) et al. presented a method to search for two-way non-linear interactions. Two-dimensional splines were used to capture the shape of the association for all pairs of continuous exposures. A Bayesian paradigm was used, with priors that allow shrinking terms to zero in the absence of interaction. Prior information on the direction of the interaction (synergistic vs. antagonistic) was also incorporated. . This approach combines the dimension reduction and interpretability of linear index models (such as WQSR and BGIR) and flexible exposure-response modelling of response surface methods (such as BKMR and PIMGP). With BMIM, the original exposures are reduced to a set of indices, as in BGIR. The approach simultaneously estimates the index component weights (with variable selection) and a potentially complex, high-dimensional exposure-response relationship between the indices and health outcome. Thus, it allows non-linear effects of the indices and interactions between indices. This presentation won one of two exposome data challenge prizes. Guimbaud (14) et al. linked the exposome with several health outcomes using several machine learning methods, namely multilayer perceptron, random forest, XGboost, support-vector machines (SVM), and Elastic net. They compared the prediction performance of the different techniques and used explainable AI by calculating SHAP (SHapley Additive exPlanations) values (Lundberg and Lee, 2017) to examine the impact of each exposure in the resulting models and their interactions with regards to the health outcomes. Zou (13) et al. conducted similar analyses, using in this case a deep neural network ensemble model, which was compared in terms of prediction accuracy to LASSO, SVM and random forest. Instead of SHAP values, they calculated a permutation-based feature importance test. Broséus (16) et al. studied the relationship between the exposome and child's BMI using a multistep approach. First, they ranked predictors by feature importance using multivariate ordinal random forests. Based on this metric, they used an arbitrary threshold to select the most pertinent exposures and, among these, they performed an analysis of exposure associations using: 1) an ordinal logistic regression model to obtain effect estimates and direction of association and 2) an MCA to obtain a graphical view of the clustering of exposures, adapted to categorical exposures. Finally, Yun Chen et al. (24) proposed a measure of Explained Variation to assess the performance of models, which can be estimated accurately without estimating the individual regression coefficients. Looking at explained variation can provide interesting insights, such as confidence intervals, the explanatory power of different exposure families or of interactions terms. Multi-stage modelling approaches were applied to model patterns in the exposome data prior to examining associations with the outcome. For example, Pearce et al. (12) linked the exposome to birth weight by applying the framework defined as Exposure Continuum Mapping (ECM). An exposure continuum map is a spatially organized map of exposome features that places similar exposure profiles close to each other and different ones are further apart. It is built in two steps: first they build a low dimensional (2D) representation of the data using Kohonen self-organising maps (Kohonen, 1982) in order to identify exposure profiles. Then, using information from this organised map, they used a Generalized Additive Model (GAM) to build a 3D exposure-response function that allows examination of a total mixture effect. Benavides et al. (8) linked the exposome with neurobehavior using a strategy that involved reducing the exposome via PCP (Gibson et al., 2021) and FA, in order to identify both consistent and unique exposure patterns, and then regress these lower dimensional patterns to the health outcome using generalized additive models (for the consistent patterns) and LASSO (for the unique patterns). Mistry (11) et al. linked the exposome to obesity. They used PCA and k-means clustering to identify exposure profiles in both the prenatal and postnatal periods, and then used logistic regression to assess the risk of obesity as a function of the transitions of individuals between prenatal and postnatal exposure clusters. Basu (23) et al. designed an iterative algorithm (COLRNS) that creates clusters of correlated exposures and then performs variable selection within the clusters to predict the health outcome while minimizing the error of the model. This method can also handle missing data. Several studies used one or more omics datasets as intermediate layers and conducted some kind of mediation or meet-in-the-middle analyses. Wang (Ziyue) et al. (10) studied the link between the exposome and asthma using transcriptome as an intermediate layer. In particular, they used a combination of 1) differential gene expression and gene set enrichment for asthma, 2) exposure selection in a model for asthma via Elastic net and calculation of exposure risk scores, and 3) highdimensional mediation analysis. Shen et al. (18) also conducted mediation analysis, in this case using several omics datasets (transcriptome, proteome, serum/urine metabolome) as potential mediators in the relationship between the exposome and several health outcomes (IQ, behavior, BMI). All models were fitted with linear mixed models and they used BH correction to correct for multiple comparisons. This presentation received the committee prize for integrating all the high dimensional omics and multiple outcomes while using informative visualisation of the results. Wang (Congrong) et al. (19) conducted a causal mediation analysis using multi-omics layers (transcriptome, proteome, serum/urine metabolome) as potential mediators of the relationship between the exposome and BMI. In this case, they used multi-omics factor analysis to reduce dimensionality of the omics layers and factor analysis to reduce the dimensionality of the exposome. Yu et al. (21) presented Gatekeepers (Yu et al., 2021) a new theory to assess exposure-metabolites associations. They identified some Gatekeepers in the data using the Pearson correlation and then studied their associations with asthma. Gatekeepers are metabolites associated with both exposures and other metabolites. They are presented as the bridge between the exposome and the metabolome. Brunius et al. (17) also implemented a meet-in-the middle approach in which they used the proteome, serum metabolites, urine metabolites, gene expression and methylation as middle layers between the exposome and health outcomes. In their analytical pipeline, Omics Modules for Exposome Health Associations (OMEXA), they used machine learning (MUVR Multivariate Methods with Unbiased Variable Selection, a predictive machine learning algorithm using an embedded recursive feature elimination mechanism within a repeated double cross validation procedure) to select the exposure variables related to the phenotypes available and partial correlation to further refine the selected list of exposures, adjusting for covariates. Then, they reused the same pipeline to select omic variables related to the selected exposures; and finally, they implemented a generalized linear model to link the selected omics with the phenotypes. They visualized the final results with triplots by projecting exposures, omics and phenotypes into two principal components. Zhao (Stratakis) et al. (9) studied the link between organochlorines and BMI, using proteomics and urine and serum metabolites as intermediate layers. They used the latent unknown clusters (LUCID) method, which found latent subgroups of subjects characterized by having at the same time distinguished BMI, distinguished omics profiles and distinguished exposure. The process includes variable selection with LASSO and results were visualized with a Sankey diagram. Finally, Wu et al. (7) linked the exposome with metabolomics, without including the health phenotypes. They used a bipartite graph model to represent pairwise associations between exposures and metabolites. From this graph they extracted subgraphs of concentrated most significant negative or positive association blocks (that were assessed using the pairwise Pearson correlation coefficients). Some researchers studied causal relationships between environmental exposure and health. Roscoe et al. (25) studied the relationship between air pollution and cognition, and used blood concentrations of metals (part of the exposome) as potential mediators. They used causal mediation analysis and quantile g-computation to assess mediated effects. Zugna et al. (20) also conducted mediation analyses, in this case they considered the exposome as a potential mediator in the association between socioeconomic position and birthweight. Thus, they studied a context with a high-dimensional mediator set, and their proposed analysis was based on interventional effects and penalized regression models (Loh et al., 2020) . Cáceres et al. (15) conducted an analysis trying to explain the differences in BMI between boys and girls (outcome variable). They inferred groups (clusters) of participants with specific exposomic profiles for which those differences were the highest. The underlying method used was a recent implementation (https://github.com/teff-package/teff) of random causal forests (Wager and Athey, 2017) . They also looked for omics markers (transcriptomic and methylomic) that were associated with differences between the exposome clusters. This presentation won the popular vote of the challenge. Most of the approaches which focused on child BMI confirmed previous HELIX publication results (Vrijheid et al., 2020) that childhood hexachlorobenzene (HCB) exposure is cross-sectionally associated with reduced childhood BMI z-score. Some studies also identified metals and PCBs (polychlorinated biphenyls, particularly PCB170) to be linked with BMI. In a WQS regression of 38 prenatal and postnatal chemicals, Gennings et al. (4) reported a significant negative association between the mixture index and child BMI. Higher weights belonged to the postnatal exposures, with HCB being the chemical with the highest contribution. Conversely, Midya (2) et al. found in a penalized group mixture BWQSR that prenatal organochlorine compounds (OCs) and metals were positively associated with child BMI, whereas PCBs were negatively associated. Within each group, the chemicals with the highest weight were: HCB (for OCs), As, Cd and Co (for metals), and PCB170 (for PCBs). (3) OCs, postnatal metals, prenatal water DBPs and the postnatal built environment. Among these, OCs were strongly negatively associated with child BMI z-score, and HCB was the chemical with the strongest effect in the index, followed by PCB170 and DDE. No interactions were observed except for prenatal water DBPs and postnatal OCs. Consistently, Broséus et al. (16) also identified postnatal HCB as the chemical exposure with the highest importance in multivariate ordinal random forests. After combining lifestyle and chemical exposures in an MCA, they found that Cu was associated with an increased risk of childhood overweight. Using the LUCID method, Yinqi Zhao et al. (9) identified protein signatures (IL-1beta, IL-6, insulin) giving insight into underlying mechanistic pathways of childhood obesity (eg., systemic inflammation, disturbed glucose metabolism). Finally, in multi-omics mediation analysis, Wang (Congrong) et al. (19) detected urine metabolites (e.g. phospholipids, TMAO, hippurate) that mediate the effect of maternal smoking and built environment on childhood BMI. This event brought together researchers from various disciplines to work on a common challenge: exposome data analysis. It established an overview of state-of-the-art methods currently used in the field and paved the way to interesting discussions and exchange of ideas. The first point of interest was the wide use of dimensionality reduction methods, especially feature selection, to deal with multivariate exposomic data (even without omics). Feature extraction methods were less used because they usually complicate the interpretation of the results if we are interested in the effect of a particular exposure on health. However, during the challenge, some interesting methods tried to analyse groups of correlated exposures as a way to reduce the dimensionality of the input while keeping the results interpretable. While there is still a use of linear methods (LASSO, ExWAS, Elastic-Net) to analyse exposome-health associations, there is a great interest in methods that can capture more complex non-linear relationships between the features and the health outcome. Among those methods, most of them were Bayesian methods that are able to approximate a wide class of functions while also giving credible intervals. The Bayesian paradigm is useful because it naturally penalizes complex models and it offers flexibility to incorporate a process of variable selection. It is worth noting that, even if models can capture complex high-dimensional surfaces, interpretation of such models is usually done using plots that show the effects of just one or two variables at a time. In fact, some studies suggest that structures based on four variables are at the limit of human ability to correctly process the information (Halford et al., 2005) . In this data challenge, a few studies detected two-way interactions in the data based on Bayesian methods (synergistic interactions, multiple index-based or additive regression tree-based) or machine learning models combined with XAI techniques such as SHAP (Lundberg and Lee, 2017) . The limitations and complexity of interpreting high-dimensional surfaces are addressed by changing the research question to that of estimating a mixture effect associated with health outcomes using weighted indices. The detection of a mixture effect is more powerful based on a single degree of freedom test. Detection of non-negligible weights provides evidence of important components. Methods have been developed to estimate quadratic associations with the weighted index where the significance of the quadratic term may be used as a goodness of fit for an assumption of linearity. Of course, the limitation of an index approach has been that interactions cannot be detected. Recently addressed, multiple index models allow for interactions between indices and provide a compromise between single or additive index models and high-dimensional exposure-response surface methods (McGee et al., 2021) . Despite their great predictive power, machine learning techniques were until now set aside in the context of environmental health studies because of their lack of interpretability. Black box models are able to capture more complex information from the data (e.g., complex interactions, non linear relationships etc.) than linear, regression-based models, and this can be key to accurately estimate the impact of the exposome on health. Black box methods used during the challenge included ensemble methods (such as random forests, xgboost), neural networks and support-vector machines. The method most often used was by far random forests. One can improve on the interpretability of black box models in several ways, here we will discuss two of them. The feature importance metric was the most popular technique during the challenge and is also the most used in the exposome field. It can be computed in several ways depending on the model: 1) impurity based feature importance (Breiman, 2001) for a model based on decision trees 2) permutation based feature importance (Altmann et al., 2010) and shapley values based feature importance (Harris et al., 2021) can be both applied on any model (model agnostic). Another approach is to use partial dependence plots (Zhao and Hastie, 2021 ) that allow visualizing partial associations between variables. A related example is the application of a combination of super-learner and g-estimation to assess the association between chemical pollutants and cognitive function (Oulhote et al., 2019) . Many of the machine learning methods used during the challenge were limited by the small sample size available (N=1301) and it is also worth mentioning that the data imputation could lead to spurious associations. Considering the analysis of omic data, all studies performed some sort of dimensionality reduction before applying different statistical analysis. This is a limitation of most existing multi-omic data integration approaches since they have not been implemented to deal with large matrices. Therefore, development of new integrative methods/tools must consider efficient handling of large data sets (Subramanian et al., 2020) . Most presentations studied the relationship between environmental exposures and health outcomes using omic data (single or multi-omics) as an intermediate layer in a mediation analysis fashion. Some other studies used Pearson correlation to study the relationships between omics and exposome data. Different combinations of statistical tools were proposed for omics data integration in analytical frameworks. These tools taken individually were not novel but their combination for integrating exposome, multi-omics and outcomes was strongly relevant and novel. We note that other methods previously used for exposome and omics data may also be of interest but were not presented during this event. These methods include dimension reduction derived from PCA such as partial least square (PLS) and its derivatives (sparse-PLS, group-PLS) (Chun and Keleş, 2010; Jain et al., 2018; Lenters et al., 2015) , canonical-based methods or network analysis (Bessonneau et al., 2021) . Causality in exposome-health models were also assessed through mediation analysis including omics data, g-computation, and causal random forest. We note that none of the methods presented, addressed the challenge to compare a priori approach with agnostic analyses that would perform variable selection treating all exposures in the same way, neither answered a large number of causal questions referring to different exposures using causal inference techniques for high-dimensional data. Results from mediation analyses using cross-sectional omics should be interpreted with caution since the omics markers might be a consequence of the health outcome or of the exposure (one or the other, not both). In particular serum metabolome data, which mainly includes information on lipid metabolism, are closely related to phenotypic outcomes such as BMI. Therefore, they should be expected to reflect more outcome classification (e.g. obesity subtypes) than an exposure effect. We note that this article does not provide an exhaustive list of potential methods to study the exposome. In addition, the provided dataset was limited in power to study exposure interaction, as it is the case for most epidemiological datasets with exposome data. In the future, a potential complementary approach would be the integration of a priori knowledge, such as from experimental toxicological data, to find chemical interaction or a priori group of exposures with similar targets. This knowledge-driven approach can also be used to reduce the dimensionality in the omics dataset. Future challenge for exposome data should include longitudinal datasets, as it is one of the key components in the exposome definition, the measure of cumulative effects across a lifespan, which was only partially covered here with the data available at two periods, pregnancy and childhood. The strength of this event was the application of various methods on the same, well-characterized HELIX dataset. Indeed, it was possible for approaches focusing on the same outcome, child BMI, to find the same chemicals as main predictors (PCBs, metals). Although biological interpretation from these analyses should be avoided due to the partially simulated nature of the data and the heterogeneity in the methods to deal with the confounder structure (e.g., multi-center structure), this exposome dataset could serve as a reference to test novel methods in the future. Finally, the work herein has resulted in computational algorithms with associated code made available to the community with an open-source licence, allowing for reproducible research and applications to other similar research questions based on exposome and even multi-omics data. This event was thought-provoking and highlighted the importance of networking between researchers in a multi-disciplinary environment. It fostered future collaborations at a time where interpersonal interaction is constrained by COVID pandemic, as well as giving visibility to researchers of various genders, backgrounds and career stages. Permutation importance: a corrected feature importance measure Gaussian graphical modeling of the serum exposome and metabolome reveals interactions between environmental chemicals and endogenous metabolites Multiple Correspondence Analysis and Related Methods. Multiple Correspondence Analysis and Random Forests Characterization of Weighted Quantile Sum Regression for Highly Correlated Data in a Risk Analysis Setting Sparse partial least squares regression for simultaneous dimension reduction and variable selection Per-and poly-fluoroalkyl substances and bone mineral density: Results from the Bayesian weighted quantile sum regression Principal Component Pursuit for Pattern Identification in Environmental Mixtures Joint Shapley values: a measure of joint feature importance How many variables can humans process? A multivariate approach to investigate the combined biological effects of multiple exposures Principal Component Analysis and Factor Analysis A quantile-based g-computation approach to addressing the effects of exposure mixtures A survey of feature selection and feature extraction techniques in machine learning Self-organized formation of topologically correct feature maps Phthalates, perfluoroalkyl acids, metals and organochlorines and reproductive function: a multipollutant assessment in Greenlandic, Polish and Ukrainian men Nonlinear mediation analysis with highdimensional mediators whose causal structure is unknown A Unified Approach to Interpreting Model Predictions Bayesian Multiple Index Models for Environmental Mixtures Permutation-based identification of important biomarkers for complex diseases via machine learning models Bagging and Deep Learning in Optimal Individualized Treatment Rules 27 Estimating Perinatal Critical Windows of Susceptibility to Environmental Mixtures via Structured Bayesian Regression Tree Pairs Joint and independent neurotoxic effects of early life exposures to a chemical mixture: A multi-pollutant approach combining ensemble learning and g-computation Exploring associations between prenatal exposure to multiple endocrine disruptors and birth weight with exposure continuum mapping A latent unknown clustering integrating multi-omics data (LUCID) with phenotypic traits Highly Scalable Bayesian Geostatistical Modeling via Meshed Gaussian Processes on Partitioned Domains Variable selection and validation in multivariate modelling Multi-omics Data Integration, Interpretation, and Its Application Early-Life Environmental Exposures and Childhood Obesity: An Exposome-Wide Approach The human early-life exposome (HELIX): project rationale and design Estimation and Inference of Heterogeneous Treatment Effects using Random Forests Bayesian Group Index Regression for Modeling Chemical Mixtures and Cancer Risk Bayesian Variable Selection and Estimation for Group Lasso Molecular Gatekeeper Discovery: Workflow for Linking Multiple Environmental Biomarkers to Metabolomics Causal Interpretations of Black-Box Models LM Data analysis: all the event participants Consortia The Exposome Data Challenge Participant Consortium is listed below in alphabetical order The administrative and communication team which did a tremendous job, Rodney Ortiz, Yvette Moya-Angeler and Aleix Cabrera. Congratulations to the winners of the challenge: Xiaotao Shen from Stanford University, Alejandro Caceresfrom ISGlobal and Glen McGee from the University of Waterloo. The data for the challenge were issued from a study from the European Community's Seventh