key: cord-0961474-p510o60l authors: Abdin, Adam F.; Fang, Yi-Ping; Caunhye, Aakil; Alem, Douglas; Barros, Anne; Zio, Enrico title: An optimization model for planning testing and control strategies to limit the spread of a pandemic – The case of COVID-19 date: 2021-11-06 journal: Eur J Oper Res DOI: 10.1016/j.ejor.2021.10.062 sha: 28beaa82ab4ee8e7a590a0d5f513a7a64a6e4e46 doc_id: 961474 cord_uid: p510o60l The global health crisis caused by the coronavirus SARS-CoV-2 has highlighted the importance of efficient disease detection and control strategies for minimizing the number of infections and deaths in the population and halting the spread of the pandemic. Countries have shown different preparedness levels for promptly implementing disease detection strategies, via mass testing and isolation of identified cases, which led to a largely varying impact of the outbreak on the populations and health-care systems. In this paper, we propose a new pandemic resource allocation model for allocating limited disease detection and control resources, in particular testing capacities, in order to limit the spread of a pandemic. The proposed model is a novel epidemiological compartmental model formulated as a non-linear programming model that is suitable to address the inherent non-linearity of an infectious disease progression within the population. A number of novel features are implemented in the model to take into account important disease characteristics, such as asymptomatic infection and the distinct risk levels of infection within different segments of the population. Moreover, a method is proposed to estimate the vulnerability level of the different communities impacted by the pandemic and to explicitly consider equity in the resource allocation problem. The model is validated against real data for a case study of COVID-19 outbreak in France and our results provide various insights on the optimal testing intervention time and level, and the impact of the optimal allocation of testing resources on the spread of the disease among regions. The results confirm the significance of the proposed modeling framework for informing policymakers on the best preparedness strategies against future infectious disease outbreaks. During the early days of January 2020, it was reported that a number of pneumonia cases, identified by the end of 2019 and, at the time, of an unknown aetiology, were, in fact, caused by a novel coronavirus later identified as SARS-CoV-2. In the months that followed, this novel coronavirus, which caused the disease popularly known as COVID-19, started to rapidly spread across the globe, resulting in an exponential increase in the number of infections and hospitalizations of severe cases. With hospitals in many countries reaching their maximum operational limits and with the rapidly increasing number of deaths caused by the virus, the World Health (country/region) to deploy and allocate the proper resources in order to limit the spread of an infectious disease. A notable example is the preparedness level seen in the case of the Republic of Korea. The Republic of Korea was successful in identifying a large number of cases very early on, as a result of implementing a massive testing strategy, reaching around 20,0 0 0 tests per day in less than 6 weeks since they identified the first case ( Sung et al., 2020 ) . The strategy was successful in significantly slowing down the spread of the disease without enforcing city-wide lock-downs or collapse of the national health-care system. In other cases, pandemic responses have been criticized as resulting in high peaks of infections and deaths during the early phase of the outbreak, particularly, due to the lack of identification of cases via mass testing and contact tracing ( Moatti, 2020 ) . Testing, case-isolation and control strategies have been shown to be effective measures to control the spread of infectious diseases, including the COVID-19 pandemic ( Aleta et Madubueze, Dachollom, & Onwubuya, 2020; Wells et al., 2020 ) . However, with regards to the actual levels of preparedness for rapidly implementing testing strategies and allocating testing capacities among different regions, policymakers face a number of planning challenges. Testing resources may be naturally limited at the onset of the outbreak of a new disease as testing technologies are still being developed. Other factors, such as distribution times, test administration times and the number of trained staff available to perform the procedure and interpret the results may also contribute to limit the capability to effectively allocate massive testing capacities. Furthermore, in the context of the COVID-19 outbreak, the decisions for controlling the pandemic spread should take into account disease specific attributes, among which, the difference in severity of symptoms and risk factors within the population (for example elders and those with underlying health conditions are more prone to develop severe symptoms compared to younger and healthier individuals Jordan, Adab, & Cheng, 2020 ) , the high rate of transmission of the disease among asymptomatic individuals ( Moghadas et al., 2020 ) and the impact of population mobility between regions ( Guan et al., 2020 ) for work or study purposes and particularly from infected-asymptomatic individuals who are harder to identify. Finally, a major challenge for policymakers for allocating limited control resources among regions is to ensure an equitable distribution of these resources. Despite the fact that there is little consensus about the meaning of equity, it is broadly accepted in public health that an equitable allocation of resources help to mitigate disparities and improve health outcomes among regions ( Braveman, 2006 ). An equitable public preparedness and management strategy does not solely depend on the character and severity of the disaster in and of itself, but also on the vulnerability of the particular communities impacted by it. Whereas public health disasters such as the COVID-19 pandemic do not, themselves, create the conditions of health access inequities in a population, they accentuate existing ones and make it so that vulnerable communities suffer the most ( Alberti, Lantz, & Wilkins, 2020 ) . The aim of this paper is to provide an optimization modeling framework to help policymakers optimally plan and evaluate decisions on allocating scarce disease detection and control resources in order to mitigate the spread of a pandemic. Since a pandemic progression within a population is an inherently non-linear phenomenon, the proposed model is an explicitly formulated nonlinear optimization model that captures key disease transmission characteristics and non-linear transition dynamics. In addition, the modeling framework proposed takes into account specific characteristics of the COVID-19 pandemic, most notably, the possibility of asymptomatic exposure and infection as well as the distinct risk levels of infection for different segments of the population. The spatio-temporal aspect of the decision making problem is fur-ther considered by accounting for population mobility between regions. Finally, the proposed modeling framework is extended to account for equity considerations for the exposed and vulnerable communities within the decision-making problem. The relevance of the modeling framework is validated with a real case study, first, by validating the model predictions against real pandemic control measures and the resulting disease transmission dynamics that occurred, and subsequently, by showing the significance of the results for a range of intervention policies. To the best of our knowledge, this study is the first attempt to formulate and address the problem of allocating testing and control resources as a Non-Linear Programming (NLP) model for controlling an infectious disease outbreak. There is a large body of literature that deals with modeling infectious diseases within a decision making framework . One of the most prominent modeling approaches that have proven its usefulness for modeling epidemics is the SIR compartmental model proposed by Kermack & McKendrick (1927) and its various extensions (see for example van den Driessche, 2008 ) . The basic SIR model separates the population in three groups: susceptible individuals (S) who can get the disease, those infected (I), and those removed from the population (R) due to recovery, death or immunity. In the majority of studies, these compartmental models are used within a simulation or differential equation framework, primarily to predict the spread of the disease within the population ( He, Peng, & Sun, 2020; Iwata & Miyakoshi, 2020; Meltzer et al., 2014; Mwalili, Kimathi, Ojiambo, Gathungu, & Mbogo, 2020 ) . Decision support for resource allocation in preparedness, mitigation and control of infectious diseases is another important application domain for operation research and management science. However, similarly, the majority of the existing literature is based on simulation models or simulation-optimization frameworks, which do not explore the full range of decisions and control policies to guarantee that the optimal ones are selected ( Dimitrov & Meyers, 2010 ) . Fewer studies considered compartmental models within mathematical optimization, such as linear, integer, non-linear, stochastic programming or optimal control for decision making in preparing against and controlling infectious diseases . Sun, DePuy, & Evans (2014) proposed linear optimization models for patient and resource allocation among hospitals in a health-care system during a pandemic influenza outbreak . Yarmand, Ivy, Denton, & Lloyd (2014) proposed a two-stage stochastic linear program to optimally allocate vaccines within a two-phase vaccination policy to contain an epidemic. proposed a theoretical analysis to derive decision support insights on mass vaccination against AH1N1 influenza pandemic. Büyüktahtakın, des Bordes, & Kıbı¸s (2018) presented an epidemics-logistics mixed-integer linear optimization framework for determining optimal intervention strategies in terms of treatment capacities and location for controlling the Ebola epidemic. Their proposed model is interesting in that it combines a number of important features within the optimization framework, such as spatially varying rates for disease transmission and the migration of population among regions, which are also considered in this work together with a number of additional important improvements. More recently, Madubueze et al. (2020) studied the effect of different control strategies for the case of the COVID-19 pandemic, in particular quarantine and isolation, as time-dependent interventions using an optimal control approach. The particular problem of allocating limited testing capacities and their impact on the dynamics of infectious disease is 2 ARTICLE IN PRESS JID: EOR [m5G; November 25, 2021; 11:20 ] To the best of our knowledge, no existing study addresses the optimization problem of multi-region multi-period allocation of testing capacity to prepare against and control the spread of a pandemic. Our approach also differs in a number of ways from existing epidemiological resource allocation problems, among which, the explicit modeling of asymptomatic exposure and infection, the separation of distinct risk levels of infection for different segments of the population, and the consideration of an equity-based objective function to derive the allocation decisions. Finally, the NLP model developed maintains the inherent non-linearity in pandemic progression dynamics, in particular with regards to the number of new infections and the positivity-rate of testing. In summary, our research makes several distinct contributions to the literature: (i) The paper proposes a novel non-linear programming (NLP) model based on an adapted compartmental SIR model for the optimal allocation of testing and control resources in a multi-spatial multi-temporal setting with the objective of minimizing the new infections in the population, as well as the cases that die without receiving hospitalization. (ii) The paper proposes novel extensions of the SIR compartmental model to take into account different levels of infection severity, asymptomatic transmission of the infectious disease, decisions related to testing symptomatic and asymptomatic individuals, case isolation modeling for confirmed cases and decisions related to hospitalization for severe cases, all within a single optimization framework. (iii) The paper proposes a novel formulation that considers the allocation of testing to asymptomatic and symptomatic individuals, and models the effect of positivity-rate of testing as a function of the ratio of the non-infected population to the total population. (iv) The paper considers an equity-driven objective function, based on the GINI index and weighed by the population vulnerability level, to generate equitable testing allocation plans. (v) The proposed model is validated against real data for a case study of the COVID-19 outbreak in France, and the results are illustrated for a number of scenarios featuring realistic assumptions. The results provide insights on the impact of testing intervention time and level on the control of the pandemic. Moreover, the results confirm the usefulness of the proposed tool to help inform the policymakers on the level of preparedness needed in terms of testing capacity to control a large scale infectious-disease outbreak. The proposed model describes the diffusion of the virus as a state transition among 12 compartments in each physical location and for each time period, as illustrated in Fig. 1 . In the Figure, the nodes represent the states of the individuals in the population impacted by the SARS-CoV-2 virus and the arcs represent the transition paths from one state to another, with the transition rate noted next to each arc. The state evolution of the population is either disease-dependent/clinical (solid arcs) or decision-dependent (dashed arcs). We divide the infected population into three groups according to the most prominent features found in the clinical literature related to the COVID-19 disease: • Infected asymptomatic ( IA ): this group is carrying the disease in its incubation period. COVID-19 was reported to be capable to be transmitted even in the incubation period. Given that this group is not symptomatic, in the beginning of a pandemic outbreak they are more likely to interact with other individuals without being noticed. [m5G; November 25, 2021; 11:20 ] significantly reduced compared to disease carrying individuals in the general population. Therefore, the disease transmission parameter for this group is significantly lower but non-zero. Furthermore, the disease-dependent state progression is separated in two groups. The first group (grey) starts with the untested susceptible individuals ( NA ), who become infected with transmis- x ) through community interactions with different infected subgroups and move to the untested-infected population ( IA , ISM , ISS ), as a consequence of these interactions. Within the infected subgroups, individuals progress through several disease severity states: from asymptomatic to mildly-symptomatic to severely-symptomatic ( IA → ISM → ISS ) with transition rates ( μ x and κ x ), respectively. The same transitions occur within the second group (red), which represents the same severity levels of infection, but for tested and confirmed individuals who have been identified and isolated ( IA , ISM , ISS ) and, therefore, can no longer infect others. Infected individuals from both groups can, also, transition to recovery if they have mild symptoms ( ISM , ISM → to R ), with rates ( ω A x ), or to death if they have severe symptoms ( ISS , ISS → D ), with rates ( ξ A x ). It is assumed in the proposed model that infected asymptomatic individuals (IA) will show some symptoms after the incubation period. Therefore, they will either transition to ISM or they will be isolated after being tested and transition to ISM from which they can recover. Notice that a proportion of infected asymptomatic individuals may recover without being tested nor showing any symptoms, but those individuals can not be identified since they were neither tested nor showed any symptoms. In this case we equivalently identity them as non-infected individuals, since, for modeling purposes, there is no way to quantify the proportion of these individuals without any clinical data. This should not impact the output of the model as the disease transition parameters are, first, calibrated with real data to reflect the actual disease progression in the population as discussed in Section (6.1) . Finally, the decision-dependent state progression concerns the testing and hospitalization states. If available capacities exist, asymptomatic and mildly-symptomatic individuals can get tested ( T A and T S, respectively) and move to the respective identified state. Similarly, based on available capacities, severely symptomatic individuals can either be accepted to hospitals ( H) or die without hospitalization. It should be noted that the proposed model considers a continuous relaxation of the integer state and decision variables for simplification and to allow a tractable solution of the proposed model. However, this is a common simplification in the literature for epidemiological simulation models based on differential equations for state transition and which has confirmed that this relaxation of integrality maintains high quality results. The formulation of the proposed model is discussed in details in the following section. Sets and indices: ξ B x Transition rate of severely symptomatic individuals who die after being hospitalized at location x . Various model parameters: (x,y ) , j \ IA [ mob] (x,y ) , j Number of non-infected \ infected asymptomatic individuals moving from location x to y at time j. Number of asymptomatic \ symptomatic individuals being tested at location x at time j. Decision variables: Total Asymptomatic \ Symptomatic individuals both infected and non-infected accepted for testing at location x time period j. [ TA ] x, j \ ISM JID: EOR [m5G; November 25, 2021; 11:20 ] locations and over the whole time horizon: The proposed objective function seeks to minimize the number of newly infected individuals instead of the cumulative number of infected individuals in the population since we consider heterogeneous infection levels that have different characteristics and duration of illness. If the cumulative number of infectious is considered this would lead to longer disease duration phases having a larger weight in the objective function with a priority to be minimized (for example priority to minimize the asymptomatic infectious phase if it lasts longer than severely symptomatic phase), which is not the desired outcome of the optimization. We, therefore, assume that the policymaker seeks to minimize the total number of newly infected individuals in population regardless of how long their symptoms will last. Furthermore, it should be noted that the penalty term is important to consider since, without setting the penalty of minimizing the number of individuals who do not receive hospitalization, the mathematical programming model does not have a preference between the last different removal states (recovery, death or hospitalization) as they all contribute to minimizing the number of infected individuals in the population. The only relevant factor in this case becomes the values of the transition parameters which are not enough to guarantee a logical outcome of the model where the preference is for individuals to receive hospitalization rather than to be left to die if it is a faster transition out of the infected population. Finally, as can be seen in Eq. (1) , the calculation of newly infected individuals results in a non-linear term in the objective function, that is a function of the disease transmission rates through the community interaction with infected individuals within the IA , ISM and ISS populations at the end of period j, and the ratio of the remainder susceptible population in the region. The set of constraints, grouped in Eq. (2) , define the initial conditions for the number of individuals in each of the states at each location and at the first time period: (2) It is important to consider the movement of the population between regions, as this impacts the disease transmission rates in the different regions. In this work, it is reasonably assumed that only asymptomatic individuals remain mobile between regions during a pandemic outbreak and that symptomatic individuals are confined in their respective regions until recovered (or in the worst-case deceased). It should be noted, however, that the asymptomatic population still contains infected individuals that may expose others to the disease and, therefore, may highly alter the infection rates between the regions ( Moghadas et al., 2020 ) . The mobility of the non-infected asymptomatic individuals is described in Eq. (3) as the product of the mobility ratio V from region x to the connected regions y , and the number of non-infected asymptomatic individuals at each region x and time j. Similarly, the mobility of the infected asymptomatic individuals is formulated as the product of the mobility ratio ρ from region x to con-nected regions y , and the number of infected asymptomatic individuals at each region x and time j, as shown in Eq. (4) : Notice that this formulation allows the direct implementation of different mobility policies, such as lock-downs, where mobility is restricted for some regions, by the proper setting of the mobility parameters (V and ρ) . It should be, also, noted that, in the proposed model, permanent population movement (immigration) between regions is not considered. Only temporary mobility of population for work or study is considered as reported in the official regional mobility data. It is assumed, therefore, that those individuals return back to their respective regions within the same day with a net mobility per region per day equal to 0. This means that the total population of each region ( T P x ) remains constant overtime, which is reasonable to consider for short to midterm planning models. Susceptible individuals Constraint (5) defines the number of susceptible individuals (non-infected asymptomatic NA individuals) at the end of period j + 1 to be equal to the number of susceptible individuals in the population in the previous period, plus the inwards-mobility to region x , minus the outwards-mobility from region x to connected regions y , minus newly infected individuals who transition to the infected state: Infectious individuals Constraint (6) describes the number of untested infected asymptomatic individuals ( IA ) at the end of period j + 1 to be equal to the number of IA in the previous period, plus the inwardsmobility from surrounding regions to location x of individuals with similar state, minus the outwards-mobility from location x , minus IA individuals accepted for testing ( IA [ TA ] ), minus the proportion of the IA population that transitions to the mildly-symptomatic state without getting tested, plus newly infected individuals due to community interactions, as described in Eq. (6) : It should be noted that, testing asymptomatic individuals, as is reflected in Eq. (6) , may be observed within a policy context such as contact tracing ( Braithwaite, Callender, Bullock, & Aldridge, 2020; MacIntyre, 2020; Park et al., 2020 ) or other specific policy implemented by the policymaker, which might impact the disease transmission dynamics. Constraint (7) describes the number of untested infected mildly symptomatic individuals ( ISM ) at the end of period j + 1 to be equal to the number of ISM in the previous period, plus the proportion of the IA population that transitions to the ISM state, minus the number of ISM individuals that are accepted for testing, minus the remainder of the ISM population that did not get tested and which either recovers (with transition rate ω A ) or develops severe symptoms (with transition rate κ): [m5G; November 25, 2021; 11:20 ] Symptomatic individuals are the primary candidates for testing and isolation measures. However, as will be described in Section 3.4 , it is important to consider the positive rate of testing, since there exists a possibility of testing individuals showing or reporting similar symptoms and not actually having the disease. Constraint (8) describes the number of untested infected severely symptomatic individuals ( ISS ) at the end of period j + 1 , equal to the number of ISS in the previous period, plus the proportion of the ISM population that develops severe symptoms, minus the number of ISM individuals that are hospitalized ( ISS [ H] ), minus the remainder of the ISS population that dies without being hospitalized due to lack of hospitalization resources, with transition rate ξ A : Notice that testing decisions for the symptomatic population do not consider those individuals in the severely symptomatic state. This is because it is assumed that severely symptomatic individuals require specialized care (through hospitalization) regardless of whether they have been tested for the infectious disease, or not. Moreover, as can be seen in the proposed formulation, when infected individuals are tested and isolated, hospitalized, recovered or deceased, they are removed from the infectious population, i.e., they can no longer contribute to the disease transmission. This is considered in Constraints (6) -(8) by reducing the number of infected individuals in the different categories by the respective transition proportion to one of those removed states. Regarding the tested population: constraint (9) describes the number of tested infected asymptomatic individuals ( IA ) at the end of period j + 1 , to be equal to the number of IA in the previous period, plus newly tested and confirmed individuals who were asymptomatic ( T A x, j ) and minus the IA population that develops mild symptoms and transitions to the ISM state: Constraint (10) describes the number of tested infected mildly symptomatic individuals ( ISM ) at the end of period j + 1 , to be equal to the number of ISM in the previous period, plus newly tested and confirmed individuals from the symptomatic population, plus the proportion of the IA population that develops mild symptoms, minus the proportion of the ISM population that recovers (with transition rate ω A ), or develops severe symptoms (with transition rate κ): x, j , ∀ x ∈ X, j ∈ J (11) Constraints (12) and (13) describe the number of individuals in testing states at j + 1 , equal to the number of individuals accepted for testing at the previous time period j, both for asymptomatic (12) and symptomatic (13) cases: Constraint (14) represents the total number of individuals hospitalized in location x at the end of period j + 1 , which is equal to the total hospitalized in period j plus severely symptomatic individuals (both untested and tested) accepted for hospitalization, minus the proportion of hospitalized individuals who recover (with transition rate ω B ) or die (with transition rate ξ B ). Recovered Constraint (15) provides the cumulative number of individuals in location x who recover with or without being hospitalized. The total number of recovered individuals at the end of period j + 1 is equal to the sum of: (i) the recovered individuals in the previous period; (ii) the proportion of infected individuals who recover after having showed mild symptoms but who did not get tested ; (iii) the proportion of infected individuals who recover after having showed mild symptoms and who have been tested ω A x ISM x, j ; and (iv) the proportion of infected individuals who recover after having been hospitalized ( ω B x H x, j ). Deceased Constraint (16) provides the cumulative number of individuals in location x who die with or without being hospitalized. The total number of deceased individuals at the end of period j + 1 is equal to the sum of: (i) the deceased individuals in the previous period; (ii) the proportion of individuals who die after being hospitalized ξ B x H x, j ; (iii) the proportion of severely symptomatic individuals who die without receiving hospitalization and who did not get tested ξ A x ISS x, j − ISS [ H] x, j ; (iv) the proportion of severely symptomatic individuals who die without receiving hospitalization and who were tested and isolated ξ A The resource allocation constraints that face the policymaker are presented in Eqs. (17) -(23) . Decisions regarding the spatial and temporal management of limited testing and control as well as hospitalization resources, are simultaneously optimized within a single optimization model, under the conditions developed by the disease transmission dynamics defined by Eqs. (2) -(16) . Constraint (17) limits the total number of individuals hospitalized to the number of available hospitalization capacity in location x at the end of period j. This is calculated as the minimum between the remaining capacity available to accommodate new severe cases at time j in location x ( HC x, j − H x, j ) and the total number of severely infected individuals requiring hospitalization ( ISS 6 ARTICLE IN PRESS JID: EOR [m5G; November 25, 2021; 11:20 ] and ISS ) if there is enough hospitalization resources: [ H] x, j + ISS Constraint (18) Testing limit Constraint (19) limits the total number of individuals tested to the number of available testing resources in location x at the end of period j. This is calculated as the minimum between the remaining testing capacity available after accepting asymptomatic and symptomatic individuals for testing ( T C x, j − T A x, j + T S x, j ) and the total number of the asymptomatic and mildly symptomatic individuals ( IA + ISM ), that are candidates for testing if there is enough testing resources: Constraint (20) ensures that the total testing done does not exceed the available testing capacity: Allocation of testing capacity Constraint (21) determines the total cumulative testing capacity available in location x up to period j, including any available initial testing resources T C 0 : Budget for allocating testing Constraints (22) and (23) limit the total testing resources available for distribution to the maximum level B T and ensure that testing capacities are available only at the start of the defined intervention date I start : x ∈ X All variables are continuous and non-negative, ∀ x ∈ X, j ∈ J Notice how, according to the mathematical model introduced thus far, the positivity-rate of the testing results is not considered, leading to an assumption that 100% of the total tested individuals will be positive (i.e. identified as infected). This is, evidently, not realistic: a high number of asymptomatic individuals who might be tested (because of contact tracing or other reasons) will not actually have the disease and, therefore, their test results will be negative. Similarly, a number of individuals showing symptoms might, in fact, have another illness and their test results may, therefore, also be negative. This is what is referred to as the positivity-rate of testing and it should be properly taken into consideration in the proposed disease transmission and resource allocation model. In order to consider the positivity-rate of testing, we introduce two new variables A [ TA ] x, j and S [ T S] x, j that serve to compute the effective number of infected asymptomatic and symptomatic individuals identified by testing, respectively. The following constraints are, therefore, added: Eq. (25a) ensures that the number of tests conducted on asymptomatic and symptomatic individuals is limited by the total available testing capacity. Eq. (25b) limits the number of tests conducted on asymptomatic individuals by the total asymptomatic population size. Eq. (25c) relates the total asymptomatic tests with its resulting number of positive cases by the positivity rate , which is the fraction of infected asymptomatic individuals over the total asymptomatic population. Eq. (25d) enforces that the number of tests conducted on (mild) symptomatic individuals is limited by the total mild symptomatic population size. Eq. (25e) relates the total symptomatic testings with its resulting number of positive cases by the positivity rate represents the population that shows similar symptoms but that is not infected by the specific disease. Finally, notice that, the objective function (1) and constraints (5), (6), (25c) and (25e) are non-convex and nonlinear. The resulting model described by Eqs. (1) -(25e) is, therefore, a Non-Linear Program (NLP). For ease of discussion, let us define the previously formulated objective function for each region and each time step (1) to be: Generating equitable plans require adapting the objective function to consider an equity measure between the various regions among which the resources are to be allocated. The measure O x, j is the total number of infected individuals that remain untested and those who die without receiving hospitalization. We complement this measure with an equity measure G, the Gini coefficient, which is a popular measure of inequity. To formulate the equity-driven objective function we follow the rationale developed in Eisenhandler & Tzur (2018) , wherein the objective function balances two fundamental principles of public management, effectiveness and equity. To achieve this, the equity-based objective function is defined as: JID: EOR [m5G; November 25, 2021; 11:20 ] where the GINI index G is measuring the weighed absolute difference in the allocation policy of two regions: and the weight ϑ x reflects the importance of each individual region, understood here as the vulnerability level of each region x . This leads to the following easily linearizable objective function: ϑ x can be calculated based on any number of vulnerability factors associated with region x . These factors are typically related to physical, social, economic, and/or environmental vulnerability indicators. For example, factors such as unemployment rate, poverty level or the proportion of physically vulnerable individuals in the population may be used to calculate the level ϑ x . However, socioeconomic criteria for allocating health-related resources may be difficult to validate and may often lead to controversial outcomes. In this study, we propose to calculate a vulnerability score based on a vulnerability factor which is the number of available hospital beds per capita in the respected region ( Ministère des solidarités et de la santé, 2020a ). A discussion on the choice of this vulnerability factor and its validation can be found in Appendix B. The normalized vulnerability score for each region x , is calculated as follows: in which ( η x ) is the selected vulnerability factor and ( T P x ) is the population size of each region. The NLP optimization model for pandemic preparedness introduced in Section 3.2 is applied to the real case study of the COVID-19 pandemic outbreak in France. To illustrate the usefulness of the model on a real application for the allocation of limited testing and control resources, a case study covering 3 interconnected regions in France is considered. Administratively, France is composed of 13 metropolitan regions as well as 5 overseas regions. The case study considers 3 of the major metropolitan regions, namely: (i) Ile-de-France (IdF): the most populous region and where the capital is located, (ii) Grand-Est (GE): one of the regions that were significantly impacted by the early spread of the disease and, (iii) Centre-Val de Loire (CVdL): a region which borders the most with regions in metropolitan France and may, therefore, be considered as an important connecting hub for the different regions. The choice of the geographical level (regional, country level, cities, etc.), as well as the number of regions considered in the case study, is primarily made to ensure the consistency of the data sources used as well as the clarity of evaluating and discussing the results obtained. The model, as formulated, can straightforwardly accommodate different geographical levels and numbers of locations as needed. France is reportedly the first country in Europe where COVID-19 cases were imported ( ECDC, 2020 ). At the time of writing, the number of confirmed cases and the number of deaths related to the COVID-19 infection in France have reached levels higher than 2.5 million cases and 60 thousand deaths ( WHO et al., 2020 ) . For the case study, we focus our attention on leveraging the NLP model developed in this work to optimally allocate the available testing capacities among the different regions in France, with the objective of achieving the minimal number of new infections and of deaths of individuals who do not receive medical attention. The data used for the numerical analysis includes disease transition parameters ( Table 1 ) , population and geographic data ( Table 2 ) and data describing the mobility of population between regions collected on a departmental level and aggregated for each region, as shown in Fig. 2 ( Table 3 ) . The disease transition parameters are described in Table 1 and are calculated based on clinical progression models (column 3) and generally known clinical values of the SARS-CoV-2 virus (column 4). These values are largely based on the sources cited in the Table. An exception to that are the transmission rates from susceptible to infected individuals ( σ ), which depend on location-specific attributes, such as the local policy implemented (confinement, travel bans, social distancing or other measures), the behavior of the population with regards to those policies ( Rajaonah & Zio, 2020 ) as well as other local factors. These values, therefore, need to be estimated or collected for the specific regions studied and under particular policy periods. In this study, we calibrate these values to reflect the disease transmission characteristics for Ile-de-France during the first de-confinement period (between the 1st of June to the 31st of October) and we ensure that the model results are validated for all other regions and under a variety of indicators. The model validation is thoroughly discussed in Section (6.1) . Note that we consider the same values for the disease transition parameters (recovery rate, death rate, disease progression rate, etc.) across regions. However, it is possible in other cases that these parameters differ among regions because of significant demographic or economic differences. Therefore, the model is formulated to allow the consideration of different values for the disease transmission parameters across regions, if needed. Table 2 summarizes information about the population size and information about the available hospitalization resources in each region (total number of beds in intensive care, reanimation and continuous care units) ( Ministère des solidarités et de la santé, 2020a ). An important data vector is the initial number of infected individuals in the population at the start of the planning horizon. The number of infected individuals per region is very difficult to know at the outbreak of a pandemic or without universal testing of all individuals; yet, it is important to initialize these numbers for the model state transitions to occur. We assume a relatively low number of initial infections per region in proportion to the population size and note that the only impact of varying these estimates is to shift the output along the time axis (high number of initial infections lead to earlier infections or deaths peaks, and low number of initial infections lead to delayed infections or death peaks Duffey & Zio, 2020 ) while preserving the exact transition proportions. This means that any conclusion on the results holds for any initialization of the parameters, with the exception of the observation time of occurrence. Finally, Table 3 summarizes the daily net movement between regions aggregated across all age segments and different traveling motivations, in particular work and study; the data have been obtained from INSEE (2019a,b) . The case studies consider daily progression and planning for the pandemic, in terms of allocation of testing resources under different resource availability levels and intervention start dates. Except for the choices related to the model validation illustrated in the next section; the analysis spans a horizon of 210 days (7 months) to allow for a planning comparison of a reasonably long period in which the decisions are relevant, and not too long that new information impact the model parameters initialization. In this section, we present the model validation and discuss its results with respect to actual data obtained for the selected regions. We, then, proceed to illustrate and analyze the results of optimal resource allocation obtained by the model under a variety of 8 JID: EOR [m5G; November 25, 2021; 11:20 ] ( Ferguson et al., 2020; Zhou et al., 2020 ) σ Transmission rate from susceptible to infected =0 . 21 σB =0 . 115 σC =0 . 06 (Calibrated for the disease progression in France. Further discussed in Section 6.1 ) σ = σ A · IA + σ B · ISM + σ C · ISS σA Population and regional specific data. ( Wächter & Biegler, 2006 ) , along with the HSL_MA97 sparse direct solver ( Hogg & Scott, 2011 ) , which is capable of handling large-sized instances. To handle the numerical complexity and stability of the non-linear model, the problem for each region is first solved independently and, then, the obtained solutions are used to warm-start the solver for the interconnected regional problem considering the movement of sus-9 ARTICLE IN PRESS JID: EOR [m5G; November 25, 2021; 11:20 ] ceptible and infected populations between regions. All solutions are reported only after the solver has terminated within the standard optimality tolerance on a desktop computer running LINUX OS with 12 cores, 3.2 GHz CPU and 32 gigabyte memory. All solutions were obtained within approximately 600 seconds of run time. It should be noted, however, that since the model is nonlinear and non-convex, global optimality is difficult to guarantee and, therefore, the solutions reported are considered to be only locally optimal. To ensure that the proposed NLP model is adequately capable of informing the pandemic preparedness decisions, the model is validated against the current pandemic outbreak data in France to accurately predict the real pandemic progression, given the formulation and the parameters settings. The model is validated by fixing the values of the decision variables for the daily number of tests conducted per region ( T S x, j + T A x, j ) and the daily hospitalization ( H x, j ) to the actual data observed for the region during a particular period and, then, solving the model to ensure that the observed disease state transition variables, such as the number of daily confirmed cases in each region, the daily cumulative number of deaths in each region and the daily number of positivity rate of testing per region, match the actual data. To achieve this, the regional specific parameter ( σ ) for the disease transmission between the different infections levels is first calibrated for one region and one output (in this case Ile-de-France, on the number of confirmed cases) and, then, the validation is performed as described. This means that the results for each region are validated on 3 different outputs (number of confirmed cases, number of deaths and positivity-rates of testing), including the effect of interdependent mobility between regions, leading to a validation that is expected to adequately cover the modeling choices and parameters settings. The choice of the time period selected for validation is critical, as different intervention policies have been applied to control the pandemic outbreak since its beginning. These measures include confinements, social-distancing and travel restrictions to different degrees in different periods. To ensure that the validation is consistent, the period chosen, therefore, should be consistent in the type of policies implemented and to the relevance of these policies. In this study, we chose to validate the model for the disease transmission under normal population movement (not under confinement) to ensure its relevance in a preparedness setting where no specific policies have been implemented. Ideally, this should be achieved using data from the onset of the pandemic outbreak, before any intervention measures have been implemented; however, these data are not clearly available for the COVID-19 pandemic, since most countries have implemented a confinement strategy very promptly at the onset of the outbreak. In France, the first confinement lasted from the 17th of March to the 11th of May, and a second confinement started on the 30th of October. Therefore, for the model validation, we chose the outbreak data for the period from the 1st of June to the 30th of October for the three selected regions Ile-de-France (Idf), Centre-Val de Loire (CVdL) and Grand-Est (GE) during the deconfinement period. Graphical representations of the real data used on testing ( Santé publique France, 2020b ) and hospitalization ( Santé publique France, 2020a ) are shown in Appendix A. Fig. 3 provides a visual comparison between the actual pandemic progression data and the model results for the daily number of confirmed COVID-19 cases ( Fig. 3 a) , the daily cumulative number of deaths ( Fig. 3 b) and the daily positivity-rate of testing ( Fig. 3 c) , during the period between June 1st, 2020 and October 30, 2020 in Ile-de-France (Idf), Centre-Val de Loire (CVdL) and Grand-Est (GE). The visual comparison demonstrates that the proposed model provides a good fit for the daily number of cases and the daily level of positivity-rate of testing in all three regions. For the daily number of deaths, it is seen that, while overall the model results provide a good-fit, there is a slight underestimation of the number of cases for Ile-de-France and Grand-Est, and an overestimation for Centre-Val de la Loire in particular for the cumulative number of deaths in the last two months of the studied period (September and October). It is reasonable to believe that a significant part of those misestimations is due to unreliability in data reported for the number of deaths, as it has been reported that in many instances those cases are reported with notable delays ( Santé publique France, 2020a ). It is noted, however, that the difference between the model output and the real data remains relatively low (the order of magnitude is in a few hundreds at most for the cumulative number of deaths as opposed to an order of magnitude of millions for the other indicators). The goodness-of-fit of the model results is confirmed quantitatively by computing the normalized mean absolute deviation ( n MAD), the normalized root mean squared error ( n RMSE) and the Explained Variance between the actual data and the model results, as summarized in Table 4 . It is shown that, with the exception of the number of deaths for CVdL discussed above, the error values ( n MAD and n RMSE) are very low and the Explained Variance is significant, with the lowest value being around 0.8 (maximum 1). In addition, it is shown that, for the number of deaths in the CVdL region, the error measures are low for the first three months of the data reported, with the deviation being mostly towards the end, which we note to be possibly, again, due to errors in real data reporting. In this section, the optimal decisions regarding the allocation of the available resources to control and minimize the impact of the pandemic is investigated. In particular, we focus on the allocation of testing resources and implementation of control measures, and their impact on reducing the progression of the disease transmission within the population, by properly identifying and isolating the infected cases. For this, the NLP model proposed is solved for different scenarios of total available testing capacity and different intervention starting dates. We consider a mid-term planning period of 7 months (210 days) since, in most cases, it is reasonable to assume that the daily allocation decisions will need to be updated after a given time, in response to the disease progression. For the hospitalization capacities, the allocation are also optimally obtained in the model; however, we do not consider investments in new hospitalization capacities and, instead, consider the real hospitalization capacity known for each region, as previously reported in Table 2c . The planning of new hospital investments can be straightforwardly investigated using the proposed model, by varying the parameter B H , in particular if longer planning periods are considered. For a limited testing capacity, the NLP model proposed seeks to allocate the testing resources optimally among the different regions to minimize the overall pandemic impact in terms of the number of infected and number of deceased individuals. In this section, we investigate the optimal allocation results for 5 intervention levels : no testing, 50 0 0, 10 0 0 0, 50,0 0 0 and 10 0,0 0 0 tests/day available to be distributed among all regions. Table 5 summarizes the optimal solution for allocating the available testing resources for each region and for all intervention levels. Table 5 also summarizes the impact of this optimal allocation and ARTICLE IN PRESS JID: EOR [m5G; November 25, 2021; 11:20 ] deployment of testing and control measures on the disease progression within the population, in terms of total number of infected individuals, total hospitalizations (limited by the available hospital capacities), cumulative deaths and recoveries. As seen in Table 5 , significant improvements are observed in terms of the number of people infected, hospitalized or deceased as the testing and control resources increase, regardless of the particular distribution among the regions. The "no testing" scenario performs considerably poor in terms of the total number of infected, hospitalized and deceased individuals. In this scenario, the number of infections in the population reaches more than 8M infections for IdF, 1.7M infections for CVdL and 3.6M infections for GE, which represent around 6 6.57%, 67.35% and 6 6.70% of the total population of IdF, CVdL and GE, respectively. Moreover, the results show that, in the absence of testing and control measures, around 3 million individuals in total (1.78M for IdF, 0.36M for CVdL and 11 JID: EOR [m5G; November 25, 2021; 11:20 ] 0.77M for GE) are expected to die, as seen in Table 5 . This represents a considerable 14% deaths of the population of each region. However, when testing and control interventions are introduced by increasing the total number of tests per day, the impact of the disease reduces considerably. Even by opting for the median scenario of 10,0 0 0 tests per day, the percentage of infections reduces to 2.2 M, 290 K and 471 K, which represents 18%, 11% and 8% of the population of IdF, CVdL and GE, respectively. Moreover, the number of deceased cases reduces to 112 K, 12 K and 18 K, representing much lower percentages, i.e. 0.9%, 0.47% and 0.34% of the population of IdF, CVdL and GE, compared to the average of 14% deaths observed when no intervention occurs. These improvements continue as the number of daily tests deployed is increasing, infections reaching levels lower than 2% and deaths reaching levels lower than 0.06% when 10 0,0 0 0 total tests per day are optimally allocated and implemented. The number of cases in hospitals reduces as more testing and control is introduced, although it maintains a lower improvement rate. This is because with limited hospitalization capacity, the maximum number of new daily hospital admissions cannot be surpassed until more places are vacated from patients under treatment. This results in hospitalization bottlenecks even if the number of severe cases requiring hospitalization are significantly reduced. In terms of how the testing resources are allocated, Table 5 shows that in all scenarios with non-zero testing capacity, the majority of the available resources are allocated to the highly populated region IdF. This is reasonable as more community interactions, disease transmission and mobility in and out of the region occur as a result of the higher number of population and, therefore, more resources would be needed to effectively manage this region. Interestingly, while this logic also holds when considering the other regions, the results in terms of the resources distribution are not directly correlated to the population size. Indeed, while CVdL has the lowest population size, for the highest intervention level T = 10 0,0 0 0/day, the optimal solution is to allocate more resources to this region compared to the more populous GE. This may be explained due to the impact of mobility between regions, which affects how fast the disease transmits within a population. While the impact of increasing the testing and control interventions is significant in terms of reducing the total number of infections within the population, it is also interesting to investigate how the intervention may impact the progression time of the infections within the population. Fig. 4 illustrates the time series for the evolution of the disease infection level in each region and for each testing and control intervention level. It is seen that increasing the number of tests per day does not only reduce the total number of infections, but it also reduces the peak level of the infections and delays its occurrence as shown in Fig. 4 . This can be particularly important if scarce testing and control resources are available, to extend the time available to prepare other medical in-terventions or to decide on a policy intervention in order to mitigate the worst outcomes. Finally, notice that at high testing and control levels (above 10,0 0 0 tests per day), the new infection rates are very low, with no apparent peak during the studied horizon. Similar results for the impact of increasing the testing and control measures on delaying the hospitals congestion can be seen in Fig. 5 . Fig. 5 illustrates the time series for daily hospitals occupancy in each region and for each intervention level. It is shown that, for all intervention levels, the hospitals eventually become congested and operated at their full capacities (the converging horizontal lines in the Figures). However, reaching this maximum occupancy limit can be effectively delayed by a number of months, by increasing the testing and control resources and optimally allocating them between regions. For example, for IdF region, delays in reaching hospital capacity limits of 3500 individuals receiving treatment at a given day is observed to be attained after around 4.5 months from the start of the pandemic for the scenario of T = 10,0 0 0 tests/day, compared to being reached only after 2 months for the scenario where no testing is introduced. This is even longer as the number of testing and control resources allocated increase and is observed for all regions, with hospitalization reaching their full capacity after 6 months for the highest testing intervention levels. Finally, out of the available daily testing and control resources allocated per region, the proposed model is capable to inform the optimal allocation of testing between asymptomatic vs symptomatic individuals to ensure the most effective outcomes. These results are shown in Fig. 6 for the average number of individuals tested from each population (symptomatic vs asymptomatic), at each region and for each intervention level considered. The results suggest that, for low intervention levels (T = 50 0 0 tests/day and T = 10,0 0 0 tests/day) the model allocates almost all the testing resources to the symptomatic individuals in all three regions (IdF, CVdL and GE). However, as the number of available testing and control resources increase, a significant proportion is allocated to testing the asymptomatic population. These results are reasonable since, for a limited number of tests, it is more effective to control individuals already showing signs and not to waste any testing capacity on non-infected individuals, which is highly probable within the asymptomatic population. Furthermore, Fig. 7 shows the daily percentage of asymptomatic vs symptomatic individuals in the population and compare them against the percentage of asymptomatic testing vs symptomatic testing performed. We illustrate two examples of scenarios of available testing capacities: 50 0 0 tests per day ( Fig. 7 a) and 10,0 0 0 tests per day ( Fig. 7 b) . It can be observed that, indeed, for the first case with very limited testing capacity (50 0 0 tests/day), all tests are performed for symptomatic individuals and none for asymptomatic individuals even if they represent a much higher percentage of the infected population. Similarly, the second case of 10,0 0 0 tests/day allocates 98% of testing to symptomatic individuals and 3% to asymptomatic individuals at the first few weeks. This is because the total daily test- JID: EOR [m5G; November 25, 2021; 11:20 ] JID: EOR [m5G; November 25, 2021; 11:20 ] ing capacity during this period exceeds the daily number of symptomatic individuals (who are endogenously given testing priority) and the remainder is used to test asymptomatic individuals. Starting around day 70, as the number of symptomatic individuals increase, 100% of the testing capacity is allocated for testing symptomatic individuals and none for asymptomatic. The results obtained in the previous section showed how the optimal allocation of testing and control resources lead to a significant reduction in the disease progression, the number of hospitalization and the number of deaths within the population. Moreover, different levels of resources availability was investigated and analyzed. Thus far, the investigation performed has assumed that the testing and control intervention capacities, if available, are deployed right at the onset of the disease outbreak. In many cases this may not be possible. In this section, we investigate the impact of delaying the intervention time of testing and control on the disease progression within the population. Three intervention times are considered: at the onset of the pandemic outbreak (day 0 similar to the results in the previous section), delayed by one month and by two months (day 30 and day 60, respectively). Moreover, for the different intervention start dates, we solve for all intervention levels considered in the previous section (from T = 50 0 0 to T = 10,0 0 0 tests per day). Fig. 8 compares the cumulative number of infections as a result of different testing and control intervention levels and dates. For each intervention level and intervention date, the results are broken down for each region besides providing the total values of all regions. As expected, the results confirm that delaying the intervention times, for the same intervention level, results in a higher number of infections per region and in total. Interestingly, for the highest intervention levels (T = 50,0 0 0 and T = 10 0,0 0 0 tests per day) delaying the intervention time results in significantly worse results compared to when intervention is performed earlier. Furthermore, the results confirm that in some cases delaying the in- Fig. 11 . Comparison between the cumulative number of cases (total of all regions) as a result of different intervention start dates (on day 0, day 30 and on day 60). The case of intervention level = 50 0 0 tests/day. tervention time would lead to an overall worse outcome than significantly reducing the intervention level but performing it as early as possible. This is clear in the comparison between the results of T = 50,0 0 0 and T = 10 0,0 0 0 for an intervention on day 60 (total of 8.47 million and 3.49 million cases of infection, respectively), compared to the outcome of T = 10,0 0 0 tests per day for an intervention on day 0 (total of 3.03 million cases of infection). Moreover, this analysis holds for the resulting number of deaths, as is illustrated in Fig. 9 , and highlight the usefulness of the proposed model in evaluating the trade-off between intervention level and time for effective preparedness strategies. Finally, we investigate how the optimal allocation strategy of testing and control resources change as a function of the intervention time. Fig. 10 illustrates the percentage of testing capacity allocated for each region when different intervention dates are considered and for all intervention levels. As seen in the Figure, for all intervention levels, as the intervention time is delayed, the results suggest that it would be optimal to increase the resources allocated to the GE region and reduce it for the other regions. For example, for the low resources scenario (T = 50 0 0/day), the optimal allocation for early intervention (at day 0) is 30% for IdF, 14% for CVdL and 56% for GE, whereas the optimal allocation for a delayed intervention (day 60) is 100% for the GE region and none for the others. The trend holds for all intervention levels; however, as more testing capacities are available, the proportion available for the other regions remains considerable. In this section, we investigate the impact of prohibiting mobility between regions on the spread of the disease. Computational experiments indicate that, indeed, prohibiting mobility between regions might be another way to limit the spread of the disease, in particular at the early stages of the intervention. For clarity, we consider a single intervention level (T = 50 0 0 tests per day) and compare the results for the 3 intervention dates previously considered. Fig. 11 compares the cumulative number of infections at each region, and their total for the scenario with mobility (Mobility) and with mobility prohibition (Wt/Mobility). Fig. 10 . Comparison between the optimal allocation decisions of testing resources among regions (in percentage) as a function of different intervention start dates (on day 0, day 30 and on day 60) and for each intervention level. Comparison between the evolution of the daily number of infections with and without regional mobility, for each region and for every intervention start date. The case of intervention level = 50 0 0 tests/day). The results suggest that combining both early intervention (day 0) and mobility restriction leads to a significant reduction in the number of cases by around 1.8 million cases as seen in Fig. 11 . The impact of mobility restriction is disproportional among the different regions, with regions which have a lower overall net mobility (CVdL and GE) benefiting more from this measure as opposed to the region with the highest net mobility (IdF), for which the mobility restriction actually leads to an increase in the local infections. Less significant impact for restricting mobility is observed as the testing and control intervention dates are delayed. This is clear for the insignificant impact on the number of cases when intervention is delayed by 2 months (intervention on day 60), even if mobility is restricted between regions. Furthermore, we can evaluate the impact of restricting mobility on the disease progression in terms of the daily number of infections at each region. Fig. 12 confirms that for the earliest intervention date (day 0), restricting mobility results in an earlier rise in the number of infections for the IdF region, but with no significant change for the peak infection level. For CVdL and GE, the daily number of infections reduces considerably (flattening the curve), as seen in the Figure. As intervention dates are delayed, the infection curves maintain their characteristics in terms of peak levels and shape, regardless of whether or not mobility between regions is restricted. This means that less flattening of the curves occur in these cases. Thus far, the results shown and analyzed do not consider equity in resource allocation, i.e., they seek to minimize the total new infections and untreated deaths in all regions, without accounting for the differences in the vulnerability levels of the regions. As mentioned earlier, an equity-driven objective considers these differences. The results in this section compare the non equity-driven results (for ease of discussion we call them the "standard" results) to the equity-driven results obtained by solving the NLP model proposed with the equity objective function (29) . Fig. 13 compares the standard objective results to the equitybased results, in terms of the objective function value ( Fig. 13 a) and in terms of testing capacity allocated for each region ( Fig. 13 b) , for the different intervention levels. As seen in the Figures, the following observations arise: (i) For the objective function, the equity-driven solution ( O EQUIT Y ), indeed, has a reduced absolute difference in the objective values across the regions and for all cases, as seen in Fig. 13 a. This is noted in the increase of the equity objective value (number of cases + number of untreated deaths) for the lowest impacted regions (CVdL and GE), compared to the standard solution ( O), and a respective decrease of the objective value for the highly impacted region (IdF). However, notice that, although the vulnerability levels of all regions are equivalent for all cases considered, the equity-driven solution does not have the same proportional impact on changing the objective function values for all the intervention levels. Indeed, the difference between the O and O EQUIT Y results is more pronounced for the lowest intervention levels (T = 50 0 0/days and T = 10,0 0 0/days) compared to the higher intervention levels (T = 50,0 0 0/day and T = 10 0,0 0 0/day) in which cases the solutions are almost identical. This is because, by taking into account the vulnerability factor, the equity-driven solution seeks to minimize the absolute difference between regions, which is much higher for the former cases (e.g. the difference in the O value between IdF and GE is [m5G; November 25, 2021; 11:20 ] Fig. 14. Comparison between the inequity level (GINI index score) between the standard solution and the equity-driven solution for the different intervention levels. around 7.6 million cases for T = 50 0 0/day) than for the latter ones (e.g. this difference is equal to just 130 thousand cases for T = 10 0,0 0 0 per day). (ii) To understand how the equitable solution is achieved by the optimal decisions, Fig. 13 b summarizes the allocation decisions of testing capacities for each region. The results show that, in order to achieve the equitable objective, the allocation of the testing capacities T C EQUIT Y shift from the lower impacted regions (CVdL and GE) to the higher impacted region (IdF) compared to the standard solution T C. With lower testing capacity, the proportion of tests allocated for the IdF region is higher than in cases where testing capacities are high, as previously explained. Furthermore, the effect of the equity-driven solution on the number of cases hospitalized and deceased follows the same trend as that observed for the objective value. These results are illustrated in Appendix C. Finally, the policymaker may evaluate the inequity level resulting from each allocation strategy by calculating the GINI score presented in Eq. (2c). Fig. 14 compares the inequity levels obtained by solving the allocation problem with the standard objective compared to the equity-driven objective. First, notice how, even for the standard solution, the inequity decreases as more resources are available. The inequity level starts at a level of 0.41 for the case of T = 50 0 0/day (1 being absolute inequity and 0 being absolute equity) and reduces constantly to an inequity level of 0.31 for the case of T = 10 0,0 0 0/day. Moreover, as expected, for all cases, the inequity level is lower for the solutions explicitly driven by the equitable objective, and remaining around a level of 0.27. Finally, notice that an inequity level remains in these solutions, since they still consider the effectiveness of reducing the global numbers of new infections and untreated deaths as a relevant objective, even if it maintains some level of inequity. This paper presented a novel non-linear programming (NLP) epidemic compartmental model for allocating limited testing and control resources to control the spread of an infectious disease. Our NLP framework contributes to the current epidemiological operational research literature by: (1) proposing a novel formulation for testing and control allocation decisions for both symptomatic and asymptomatic individuals in the population, and taking into account the positivity-rate of testing, (2) extending the pandemic compartmental model to consider different levels of infections severity and asymptomatic transmission of the disease, (3) developing a method to evaluate the vulnerability levels of the different impacted communities and a re-formulation of the objective function to consider equity in the allocation of testing and control resources based on the GINI index, (4) validating the possibility to solve the model for a wide variety of scenarios with explicit consideration of the non-linearity and showing the superiority of the results obtained by validation against real data. We illustrate the usefulness of the proposed modeling framework for informing policymakers about the optimal allocation of limited testing and control capacities, and their impact on the pandemic progression. In particular, we consider a case study for the COVID-19 pandemic outbreak in France and explore a wide range of possible intervention times and levels. Future research could address some of the limitations of this work, in particular with respect to the treatment of the inherent uncertainties in the disease transition parameters of the model, such as the rates of hospitalizations, recovery, severity of the disease or deaths, and their impact on the allocation decisions. Furthermore, future work may consider more details about hospitals. For example, people who need ward care, people who need to stay in the ICUs, and people who need ventilators in the ICU. This can allow the model to, also, directly consider the capacity allocation problem faced by the health-care system. Another interesting extension may consider the ability of the recovered population to become once again susceptible to the disease, which has been shown to be a possibility with the COVID-19 pandemic. Moreover, while the proposed model and case study were readily solvable with available commercial solvers, further extensions of the model may require the development of tailored solution algorithms, in particular if the number of variables and constraints increase significantly. Finally, such kind of modeling framework should be adopted by policymakers to inform decisions on actions and precautions for resilience to future pandemic outbreaks. Supplementary material associated with this article can be found, in the online version, at doi: 10.1016/j.ejor.2021.10.062. Equitable pandemic preparedness and rapid response: lessons from COVID-19 for pandemic health equity Modelling the impact of testing, contact tracing and household quarantine on second waves of COVID-19 Automated and partly automated contact tracing: A systematic review to inform the control of COVID-19. The Lancet Digital Health Health disparities and health equity: Concepts and measurement Optimal allocation of COVID-19 test kits among accredited testing centers in the philippines A new epidemics-logistics model: Insights into controlling the Ebola virus disease in west Africa Epidemics control and logistics operations: A review Mathematical approaches to infectious disease prediction and control An interactive web-based dashboard to track COVID-19 in real time. The Lancet Infectious Diseases Prediction of COVID-19 infection, transmission and recovery rates: A new analysis and global societal comparisons The humanitarian pickup and distribution problem Impact of non-pharmaceutical interventions (NPIs) to reduce COVID19 mortality and healthcare demand Transport effect of COVID-19 pandemic in France Seir modeling of the COVID-19 and its dynamics HSL_MA97: A bit-compatible multifrontal code for sparse symmetric systems The effect of large-scale anti-contagion policies on the COVID-19 pandemic INSEE, Institut national de la statistique et des études économiques migrations résidentielles en 2016 INSEE, Institut national de la statistique et des études économiques migrations résidentielles en 2016 A simulation on potential secondary spread of novel coronavirus in an exported country using a stochastic epidemic SEIR model COVID-19: Risk factors for severe disease and death A contribution to the mathematical theory of epidemics The incubation period of coronavirus disease 2019 (COVID-19) from publicly reported confirmed cases: Estimation and application Substantial undocumented infection facilitates the rapid dissemination of novel coronavirus (SARS-cov-2) Incubation period and other epidemiological characteristics of 2019 novel coronavirus infections with right truncation: a statistical analysis of publicly available case data Case isolation, contact tracing, and physical distancing are pillars of COVID-19 pandemic control, not optional choices. The Lancet Infectious Diseases Controlling the spread of COVID-19: Optimal control analysis Estimating the future number of cases in the ebola epidemic-liberia and sierra leone Nombre de lits de réanimation, de soins intensifs et de soins continus en france The french response to COVID-19: Intrinsic difficulties at the interface of science, public health, and policy The implications of silent transmission for the control of COVID-19 outbreaks Seir model for COVID-19 dynamics incorporating the environment and social distancing Mathematical modelling of the impact of testing, treatment and control of HIV transmission in Kenya Case fatality rate estimation of COVID-19 for European countries: Turkey's current scenario amidst a global pandemic; comparison of outbreaks with european countries Association of public health interventions with the epidemiology of the COVID-19 outbreak in Wuhan Contact tracing during coronavirus disease outbreak A deterministic resource scheduling model in epidemic control: A case study Risk perception biases and the resilience of ethics for complying to COVID-19-pandemic-related safety measures Estimating the infection and case fatality ratio for COVID-19 using age-adjusted data from the outbreak on the diamond princess cruise ship Données hospitalières relatives à l'épidémie de COVID-19 Données relatives aux résultats des tests virologiques COVID-19 Multi-objective optimization models for patient allocation during a pandemic influenza outbreak Preparedness and rapid implementation of external quality assessment helped quickly increase COVID-19 testing capacity in the republic of korea On the implementation of an interior-point filter line-search algorithm for large-scale nonlinear programming Incubation period of SARS-cov-2: A systematic review and meta-analysis Optimal COVID-19 quarantine and testing strategies Covid-19 weekly epidemiological update Virological assessment of hospitalized patients with COVID-2019 Characteristics of and important lessons from the coronavirus disease 2019 (COVID-19) outbreak in China: Summary of a report of 72 314 cases from the chinese center for disease control and prevention Why do countries respond differently to COVID-19? a comparative study of sweden, china, france, and japan. The American Review of Public Administration Optimal two-phase vaccine allocation to geographically different regions under uncertainty Clinical course and risk factors for mortality of adult inpatients with COVID-19 in Wuhan, China: A retrospective cohort study