key: cord- -u f kvg authors: broeck, wouter van den; gioannini, corrado; gonçalves, bruno; quaggiotto, marco; colizza, vittoria; vespignani, alessandro title: the gleamviz computational tool, a publicly available software to explore realistic epidemic spreading scenarios at the global scale date: - - journal: bmc infect dis doi: . / - - - sha: doc_id: cord_uid: u f kvg background: computational models play an increasingly important role in the assessment and control of public health crises, as demonstrated during the h n influenza pandemic. much research has been done in recent years in the development of sophisticated data-driven models for realistic computer-based simulations of infectious disease spreading. however, only a few computational tools are presently available for assessing scenarios, predicting epidemic evolutions, and managing health emergencies that can benefit a broad audience of users including policy makers and health institutions. results: we present "gleamviz", a publicly available software system that simulates the spread of emerging human-to-human infectious diseases across the world. the gleamviz tool comprises three components: the client application, the proxy middleware, and the simulation engine. the latter two components constitute the gleamviz server. the simulation engine leverages on the global epidemic and mobility (gleam) framework, a stochastic computational scheme that integrates worldwide high-resolution demographic and mobility data to simulate disease spread on the global scale. the gleamviz design aims at maximizing flexibility in defining the disease compartmental model and configuring the simulation scenario; it allows the user to set a variety of parameters including: compartment-specific features, transition values, and environmental effects. the output is a dynamic map and a corresponding set of charts that quantitatively describe the geo-temporal evolution of the disease. the software is designed as a client-server system. the multi-platform client, which can be installed on the user's local machine, is used to set up simulations that will be executed on the server, thus avoiding specific requirements for large computational capabilities on the user side. conclusions: the user-friendly graphical interface of the gleamviz tool, along with its high level of detail and the realism of its embedded modeling approach, opens up the platform to simulate realistic epidemic scenarios. these features make the gleamviz computational tool a convenient teaching/training tool as well as a first step toward the development of a computational tool aimed at facilitating the use and exploitation of computational models for the policy making and scenario analysis of infectious disease outbreaks. the h n influenza pandemic highlighted the importance of computational epidemic models for the real-time analysis of the health emergency related to the global spreading of new emerging infectious diseases [ ] [ ] [ ] . realistic computational models are highly complex and sophisticated, integrating substantial amounts of data that characterize the population and geographical context in order to attain superior accuracy, resolution, and predictive power [ ] [ ] [ ] [ ] [ ] [ ] [ ] . the challenge consists in developing models that are able to capture the complexity of the real world at various levels by taking advantage of current information technology to provide an in silico framework for testing control scenarios that can anticipate the unfolding of an epidemic. at the same time, these computational approaches should be translated into tools accessible by a broader set of users who are the main actors in the decision-making process of health policy, especially during an emergency like an influenza pandemic. the tradeoff between realistic and accurate descriptions of large-scale dynamics, flexibility, computational feasibility, ease of use, and accessibility of these tools creates a major challenge from both the theoretical and the computational points of view [ , , , , , ] . gleamviz is a client-server software system that can model the world-wide spread of epidemics for human transmissible diseases like influenzalike illnesses (ili), offering extensive flexibility in the design of the compartmental model and scenario setup, including computationally-optimized numerical simulations based on high-resolution global demographic and mobility data. gleamviz makes use of a stochastic and discrete computational scheme to model epidemic spread called "gleam" -global epidemic and mobility model, presented in previously published work [ , , ] which is based on a geo-referenced metapopulation approach that considers , subpopulations in countries of the world, as well as air travel flow connections and short-range commuting data. the software includes a client application with a graphical user interface (gui) for setting up and executing simulations, and retrieving and visualizing the results; the client application is publicly downloadable. the server application can be requested by public institutions and research centers; conditions of use and possible restrictions will be evaluated specifically. the tool is currently not suitable for the simulation of vector-borne diseases, infection transmission depending on local contact patterns such as sexually transmitted diseases and diseases with a time scale that would make demographic effects relevant. the tool, however, allows the introduction of mitigation policies at the global level. localized intervention in space or time can be implemented in the gleam model and their introduction in the gleamviz computational tool are planned for future releases. only a few computational tools are currently available to the public for the analysis and modeling of epidemics. these range from very simple spreadsheet-based models aimed at providing quick estimates for the number of patients and hospitalizations during a pandemic (see e.g. flusurge [ ] ) to more complicated tools based on increasingly sophisticated simulation approaches [ , , , , , ] . these tools differ in their underlying modeling approaches and in the implementation, flexibility, and accessibility of the software itself. influsim is a tool that provides a visual interface to simulate an epidemic with a deterministic compartmental model in a single population [ ] . the model includes age structure and explicit sojourn times with different stages in each compartment, extending an seir compartmentalization to include hospitalizations and intervention measures. the software provides the infectious disease dynamics and the user can set parameter values and add or remove interventions. however, no spatial structure or other forms of heterogeneity and stochasticity are considered in the model. on the other hand agent-based models describe the stochastic propagation of a disease at the individual level, thus taking into account the explicit social and spatial structure of the population under consideration. communityflu is a software tool that simulates the spread of influenza in a structured population of approximately , households with , persons [ ] . user interaction with the software is limited to the spreadsheet portion of the program, where one can choose the type of intervention and other parameters describing the disease and the population. a larger population is considered in flute, a publicly available tool for the stochastic simulation of an epidemic in the united states at the level of individuals [ ] . the model is based on a synthetic population, structured in a hierarchy of mixing social groups including households, household clusters, neighborhoods, and nation-wide communities. flute comes with a configuration file in text format that can be modified by an expert user to set various parameters such as the initiation of the epidemic, the reproductive number, and the interventions considered. no gui is provided, and the output of the simulations is given in the form of text files that must be analyzed through additional software. epifast involves a parallel algorithm implemented using a master-slave approach which allows for scalability on distributed memory systems, from the generation of synthetic population aggregated in mixing groups to the explicit representation of the contact patterns between individuals as they evolve in time [ ] . the epi-fast tool allows for the detailed representation and simulation of the disease on social contact networks among individuals that dynamically evolve in time and adapt to actions taken by individuals and public health interventions. the algorithm is coupled with a webbased gui and the middleware system didactic, which allows users to specify the simulation setup, execute the simulation, and visualize the results via plots. epidemic models and interventions are pre-configured, and the system can scale up to simulate a population of a large metropolitan area on the order of tens of millions of inhabitants. another class of models focuses on the global scale, by using a metapopulation approach in which the population is spatially structured into patches or subpopulations (e.g. cities) where individuals mix. these patches are connected by mobility patterns of individuals. in this vein two tools are currently available. the global epidemic model (gem) uses a metapopulation approach based on an airline network comprised of major metropolitan areas in the world for the stochastic simulation of an influenza-like illness [ ] . the tool consists of a java applet in which the user can simulate a hypothetical h n outbreak and test pre-configured intervention strategies. the compartmentalization is set to an seir model, and the parameterization can be modified in the full or stand-alone mode, but not currently in the java applet. the spatiotemporal epidemiological modeler (stem) is a modeling system for the simulation of the spread of an infectious disease in a spatially structured population [ ] . contrary to other approaches, stem is based on an extensible software platform, which promotes the contribution of data and algorithms by users. the resulting framework therefore merges datasets and approaches and its detail and realism depend on continuous developments and contributions. however, these are obtained from a variety of sources and are provided in different formats and standards, thus resulting in possible problems related to the integration and merging of datasets. such issues are left to the user to resolve. the existing tools described above thus offer the opportunity to use highly sophisticated data-driven approaches at the expense of flexibility and ease of use by non-experts on the one hand, or very simplified models with user-friendly guis and no specific computational requirements on the other. our approach aims at optimizing the balance of complex and sophisticated data-driven epidemic modeling at the global scale while maintaining an accessible computational speed and overall flexibility in the description of the simulation scenario, including the compartmental model, transition rates, intervention measures, and outbreak conditions by means of a user-friendly gui. in the gleamviz tool the setup of the simulations is highly flexible in that the user can design arbitrary disease compartmental models, thus allowing an extensive range of human-to-human infectious diseases and intervention strategies to be considered. the user interface has been designed in order to easily define both features specific to each compartment, such as the mobility of classes of individuals, and general environmental effects, such as seasonality for diseases like influenza. in addition, the user can define the initial settings that characterize the initial geographical and temporal conditions, the immunity profile of the population, and other parameters including but not limited to: the definition of an outbreak condition in a given country; the number of stochastic runs to be performed; and the total duration of each simulation. the tool allows the production of global spreading scenarios with geographical high resolution by just interacting with the graphic user interface. while an expert input would be required to interpret and discuss the results obtained with the software, the present computational platform facilitates the generation and analysis of scenarios from intensive data-driven simulations. the tool can be deployed both in training activities as well as to facilitate the use of large-scale computational modeling of infectious diseases in the discussion between modelers and public health stakeholders. the paper is organized as follows. the "implementation" section describes the software application architecture and its major components, including the computational model gleam. the "results and discussion" section presents in detail the gleamviz client and its components that allow for software-user interaction, including an application of the simulator to an influenza-like-illness scenario. the top-level architecture of the gleamviz tool comprises three components: the gleamviz client application, the gleamviz proxy middleware, and the simulation engine. the latter two components constitute the gleamviz server, as shown in figure . users interact with the gleamviz system by means of the client application, which provides graphical userinterfaces for designing and managing the simulations, as well as visualizing the results. the clients, however, do not themselves run the simulations. instead they establish a connection with the gleamviz proxy middleware to request the execution of a simulation by the server. multiple clients can use the same server concurrently. upon receipt of requests to run a simulation, the middleware starts the simulation engine instances required to execute the requests and monitors their status. once the simulations are completed, the gleamviz proxy middleware collects and manages the resulting simulation data to be served back to the clients. a schematic diagram of the workflow between client and server is shown in figure . this client-server model allows for full flexibility in its deployment; the client and server can be installed on the same machine, or on different machines connected by a local area network or the internet. the two-part decomposition of the server in terms of middleware and engines additionally allows for advanced high-volume setups in which the middleware server distributes the engine instances over a number of machines, such as those in a cluster or cloud. this architecture thus ensures high speed in large-scale simulations and does not rely on the cpu-specific availability accessible by the user. the gleamviz simulation engine uses a stochastic metapopulation approach [ ] [ ] [ ] , [ ] [ ] [ ] ] that considers data-driven schemes for the short-range and design the compartmental model of the infectious disease in the model builder. configure the simulation of the world-wide epidemic spreading in the simulation wizard. submit the simulation for execution by the engine on the server. inspect the results of a simulation in the interactive visualization. inspect all simulations and retrieve results in the simulations history. long-range mobility of individuals at the inter-population level, coupled with coarse-grained techniques to describe the infection dynamics within each subpopulation [ , ] . the basic mechanism for epidemic propagation occurs at multiple scales. individuals interact within each subpopulation and may contract the disease if an outbreak is taking place in that subpopulation. by travelling while infected, individuals can carry the pathogen to a non-infected region of the world, thus starting a new outbreak and shaping the spatial spread of the disease. the basic structure of gleam consists of three distinct layers -the population layer, the mobility layer, and the epidemic layer (see figure ) [ , ] . the population layer is based on the high-resolution population database of the gridded population of the world project by the socio-economic data and applications center (sedac) [ ] that estimates population with a granularity given by a lattice of cells covering the whole planet at a resolution of × minutes of arc. the mobility layer integrates short-range and longrange transportation data. long-range air travel mobility is based on travel flow data obtained from the international air transport association (iata [ ]) and the official airline guide (oag [ ] ) databases, which contain the list of worldwide airport pairs connected by direct flights and the number of available seats on any given connection [ ] . the combination of the population and mobility layers allows for the subdivision of the world into geo-referenced census areas obtained by a voronoi tessellation procedure around transportation hubs. these census areas define the subpopulations of the metapopulation modeling structure, identifying , subpopulations centered on iata airports in different countries. the model simulates the mobility of individuals between these subpopulations using a stochastic procedure defined by the airline transportation data [ ] . short-range mobility considers commuting patterns between adjacent subpopulations based on data collected and analyzed from more than countries in continents across the world [ ] . it is modeled with a time-scale separation approach that defines the effective force of infections in connected subpopulations [ , , ] . on top of the population and mobility layers lies the epidemic layer, which defines the disease and population dynamics. the infection dynamics takes place within each subpopulation and assumes a compartmentalization [ ] that the user can define according to the infectious disease under study and the intervention measures being considered. all transitions between compartments are modeled through binomial and multinomial processes to preserve the discrete and stochastic nature of the processes. the user can also specify the initial outbreak conditions that characterize the spreading scenario under study, enabling the seeding of the epidemic in any geographical census area in the world and defining the immunity profile of the population at initiation. seasonality effects are still an open problem in the transmission of ili diseases. in order to include the effect of seasonality on the observed pattern of ili diseases, we use a standard empirical approach in which population layer short-range mobility layer long-range mobility layer the short-range mobility layer covers commuting patterns between adjacent subpopulations based on data collected and analyzed from more than countries on continents across the world, modeled with a time-scale separation approach that defines the effective force of infections in connected subpopulations. the long-range mobility layer covers the air travel flow, measured in available seats between worldwide airport pairs connected by direct flights. seasonality is modeled by a forcing that reduces the basic reproductive number by a factor α min ranging from . to (no seasonality) [ ] . the forcing is described by a sinusoidal function of months-period that reaches its peak during winter time and its minimum during summer time in each hemisphere, with the two hemispheres with opposite phases. given the population and mobility data, infection dynamics parameters, and initial conditions, gleam performs the simulation of stochastic realizations of the worldwide unfolding of the epidemic. from these in silico epidemics a variety of information can be gathered, such as prevalence, morbidity, number of secondary cases, number of imported cases, hospitalized patients, amounts of drugs used, and other quantities for each subpopulation with a time resolution of day. gleam has been under continuous development since and during these years it has been used: to assess the role of short-range and long-range mobility in epidemic spread [ , , ] ; to retrospectively analyze the sars outbreak of - in order to investigate the predictive power of the model [ ] ; to explore global health strategies for controlling an emerging influenza pandemic with pharmaceutical interventions under logistical constraints [ ] ; and more recently to estimate the seasonal transmission potential of the h n influenza pandemic during the early phase of the outbreak to provide predictions for the activity peaks in the northern hemisphere [ , ] . the gleamviz simulation engine consists of a core that executes the simulations and a wrapper that prepares the execution based on the configuration relayed from the client by the gleamviz proxy middleware. the engine can perform either single-run or multi-run simulations. the single-run involves only a single stochastic realization for a given configuration setup and a random seed. the multi-run simulation involves a number of stochastic realizations as set by the user and performed by the core (see the following section), each with the same configuration but with a different random seed. the results of the multi-run simulation are then aggregated and statistically analyzed by the wrapper code. the simulation engine writes the results to files and uses lock files to signal its status to the middleware component. the core is written in c++, resulting in a fast and efficient engine that allows the execution of a single stochastic simulation of a -year epidemic with a standard seir model in a couple of minutes on a high-end desktop computer. the wrapper code is written in python [ ] . the server components can be installed on most unix-like operating systems such as linux, bsd, mac os x, etc. the gleamviz proxy middleware is the server component that mediates between clients and simulation engines. it accepts tcp connections from clients and handles requests relayed over these connections, providing client authorization management. a basic access control mechanism is implemented that associates a specific client with the simulations it launches by issuing a private simulation identifier key upon submission. users can only retrieve the results of the simulations they launched, or simulations for which they have obtained the simulation definition file -containing the private simulation identifier key-from the original submitter. upon receipt of a request to execute a simulation, the middleware sets up the proper system environment and then launches an instance of the simulation engine with the appropriate configuration and parameters according to the instructions received from the client. for singlerun simulations, the daily results are incrementally served back to the client while the simulation is being executed. this allows for the immediate visualization of the spreading pattern, as described in "visualization interface" subsection. for multi-run simulations the results are statistically analyzed after all runs are finished, and the client has to explicitly request the retrieval of the results once they become available. the gleamviz proxy server component can be configured to keep the simulation data indefinitely or to schedule the cleanup of old simulations after a certain period of time. multi-run metadata is stored in an internal object that is serialized on a system file, ensuring that authorization information is safely kept after a server shutdown or failure. the gleamviz proxy component additionally provides control features such as accepting administrative requests at runtime in order to manage stored simulations or to modify several configuration parameters like the number of simultaneous connections allowed, the number of simultaneous simulations per client, the session timeout, etc. the middleware server is written in python [ ] and uses the twisted matrix library suite [ ] for its networking functionality. client and server communicate using a special purpose protocol, which provides commands for session handling and simulation management. commands and data are binary encoded using adobe action message format (amf ) in order to minimize bandwidth needs. the gleamviz client is a desktop application by which users interact with the gleamviz tool. it provides guis for its four main functions: ) the design of compartmental models that define the infection dynamics; ) the configuration of the simulation parameters; ) the visualization of the simulation results; and ) the management of the user's collection of simulations. in the following section we describe these components in detail. the client was developed using the adobe air platform [ ] and the flex framework [ ] and can thus be deployed on diverse operating systems, including several windows versions, mac os x, and several common linux distributions. the gleamviz client has a built-in updating mechanism to check for the latest updates and developments and prompts the user to automatically download them. it also offers a menu of configuration options of the interface that allows the user to customize preferences about data storage, visualization options, the server connection, and others. the software system presented above is operated through the gleamviz client, which provides the user interface: the part of the tool actually experienced on the user side. the gleamviz client integrates different modules that allow the management of the entire process flow from the definition of the model to the visualization of the results. in the following we will describe the various components and provide the reader with a user study example. the model builder provides a visual modeling tool for designing arbitrary compartmental models, ranging from simple sir models to complex compartmentalization in which multiple interventions can be considered along with disease-associated complications and other effects. (an example can be found in previous work [ ] .) a snapshot of the model builder window is shown in figure . the models are represented as flow diagrams with stylized box shapes that represent compartments and directed edges that represent transitions, which is consistent with standard representations of compartmental models in the literature. through simple operations like 'click and drag' it is possible to create any structure with full flexibility in the design of the compartmentalization; the user is not restricted to a given set of pre-loaded compartments or transition dynamics. the interactive interface provided by the model builder enables the user to define the compartment label, the mobility constraints that apply (e.g. allowed/not allowed to travel by air or by ground), whether the compartment refers to clinical cases, as well as the color and position of their representation in the diagram (see figure ). this allows the user to model many kinds of human-to-human infectious diseases, in particular respiratory and influenza-like diseases. transitions individuals is equal to  si n , where n is the total size of the subpopulation. the gleam simulation engine considers discrete individuals. all its transition processes are both stochastic and discrete, and are modeled through binomial and multinomial processes. transitions can be visually added by dragging a marker from the source to the target compartment. spontaneous transitions are annotated with their rates, which can be modified interactively. infection transitions are accompanied with a representation of the infection's source compartment and the applicable rate (i.e. b in the example above), which can also be modified in an interactive way. the rates can be expressed in terms of a constant value or in terms of a variable whose value needs to be specified in the variables table, as shown in figure . the value can also be expressed by simple algebraic expressions. the client automatically checks for and reports inconsistencies in the model in order to assist the user in the design process (see bottom right window in figure ). models can be exported to xml files and stored locally, allowing the user to load a model later, modify it, and share it with other users. the diagram representation can be exported as a pdf or svg file for use in documentation or publications. a few examples of compartmental models are available for download from the simulator website. the simulation wizard provides a sequence of panels that leads the user through the definition of several configuration parameters that characterize the simulation. figure shows some of these panels. the consecutive steps of the configuration are as follows: •choice of the type of the simulation (panel a) the user is prompted with three options: create a new single-run simulation or a new multi-run simulation from scratch, or a new one based on a saved simulation previously stored in a file. •compartmental model selection and editing the user can design a new compartmental model, modify the current compartmental model (when deriving it from an existing simulation), or load a model compartmentalization from a file. •definition of the simulation parameters (panel c) the user is asked to specify various settings and parameter values for the simulation, including, e.g., the number of runs to perform (only accessible in the case of a multi-run), the initial date of the simulation, the length of the simulation (in terms of days), whether or not seasonality effects should be considered, the airplane occupancy rate, the commuting time, the conditions for the definition of an outbreak, and others. •initial assignment of the simulation (panel d) here the user assigns the initial distribution of the population amongst compartments, defining the immunity profile of the global population on the starting date. •definition of the outbreak start (panel e) this panel allows the user to define the initial conditions of the epidemic by selecting the city (or cities) seeded with the infection. •selection of output results (panel f) here the user selects the compartments that will constitute the output provided by the client at the end of the simulation. the corresponding data will be shown in the visualization window and made available for download. when all the above configuration settings are defined, the user can submit the simulation to the gleamviz server for execution. this will automatically add the simulation to the user's simulations history. it is furthermore possible to save the definition of the simulation setup to a local file, which can be imported again later or shared with other users. the simulations history is the main window of the client and provides an overview of the simulations that the user has designed and/or submitted, in addition to providing access to the model builder, the simulation wizard, and the visualization component. the overview panel shown in figure lists the simulation identifier, the submission date and time, the simulation type (i.e., single or multi-run), the execution status (i.e., initialized, start pending, started, aborted, complete, failed, or stop pending) and the results status (i.e., none, retrieve pending, retrieving, stop retrieve pending, complete, or stored locally). additional file provides a detailed explanation of all these values. a number of context-dependent command buttons are available once a simulation from the list is selected. those buttons allow the user to control the simulation execution, retrieve the results from the server and visualize them, clone and edit the simulation to perform a new execution, save the simulation definition or the output data to the local machine (in order to analyze the obtained data with other tools, for example), and remove the simulation. in addition to exporting the compartmental model (see the "model builder" subsection) the user can export a complete configuration of a simulation that includes the compartmental model and the entire simulation setup to a local file, which can be imported again later or shared with other users. once the execution of a simulation is finished and the results have been retrieved from the server, the client can display the results in the form of an interactive visualization of the geo-temporal evolution of the epidemic. this visualization consists of a temporal and geographic mapping of the results accompanied by a set of graphs (see figure ). the geographic mapping involves a zoomable multi-scale map on which the cells of the population layer are colored according to the number of new cases of the quantity that is being displayed. several visualization features can be customized by clicking on the gear icon and opening the settings widget. it is possible to zoom in and out and pan by means of the interface at the top left of the map. dragging the map with the mouse (on a location where there are no basin marks) can also pan the visualization. all the widgets and the graphs displayed over the map can be re-positioned according to the user's preferences by clicking and dragging the unused space in the title bar. the color coding of the map represents the number of cases on a particular day. the time evolution of the epidemic can be shown as a movie, or in the form of daily states by moving forward or backward by one day at a time. for single-run simulations it is also possible to show the airline transportation of the 'seeding' individuals by drawing the traveling edge between the origin and destination cities. in the case where the output quantity is a subset of the infectious compartments, the edges show the actual seeding of the infection. note that the evolution of the epidemic depends strongly on the model definition. for example, it is possible that some basins are infected by a latent individual that later develops the disease. in this case no seeding flight will be shown if only infectious compartments are selected as output. beside the geographical map, the visualization window displays two charts. one chart shows the number of new cases per , over time (incidence), and the other shows the cumulative number of new cases per , over time (size). for multi-run simulations, median values and corresponding % confidence intervals are shown. the menu above each chart combo lets the user choose the context for which the corresponding charts show incidence and size data. this context is either: global, one of three hemispheres, one continent, one region, one country, or one city. the currently selected day is marked by a vertical line in these plots, and the day number, counted from the initial date selected for the simulation, is shown by side of the time slider. here we present an example application of the gleamviz tool to study a realistic scenario for the mitigation of an emerging influenza pandemic. disease-control programs foresee the use of antiviral drugs for treatment and shortterm prophylaxis until a vaccine becomes available [ ] . the implementation of these interventions rely both on logistical constraints [ , ] -related, e.g., to the availability of drugs -and on the characteristics of the infection, including the severity of the disease and the virus's potential to develop resistance to the drugs [ ] . here we focus on the mitigation effects of systematic antiviral (av) treatment in delaying the activity peak and reducing attack rate [ ] [ ] [ ] , , , , ] , and assume that all countries have access to av stockpiles. we consider a scenario based on the h n influenza pandemic outbreak and feed the simulator with the set of parameters and initial conditions that have been estimated for that outbreak through a maximum likelihood estimate by using the gleam model [ ] . the results provided by the present example are not meant to be compared with those contained in the full analysis carried out with gleam [ ] due to the fact that in the figure the simulation results can be inspected in an interactive visualization of the geo-temporal evolution of the epidemic. the map shows the state of the epidemic on a particular day with infected population cells color-coded according to the number of new cases of the quantity that is being displayed. pop-ups provide more details upon request for each city basin. the zoomable multi-scale map allows the user to get a global overview, or to focus on a part of the world. the media-player-like interface at the bottom is used to select the day of interest, or show the evolution of the epidemic like a movie. two sets of charts on the right show the incidence curve and the cumulative size of the epidemics for selectable areas of interest. present example we do not consider additional mitigation strategies that were put in place during the early phase of the outbreak, such as the sanitary control measures implemented in mexico [ , ] , or the observed reduction in international travel to/from mexico [ ] . indeed, the current version of gleamviz does not allow for interventions that are geographically and/or temporally dependent. however, these features are currently under development and will be available in the next software release. for this reason the simulation scenario that we study in this application of the simulator does not aim to realistically reproduce the timing of the spreading pattern of the h n pandemic. the results reported here ought to be considered as an assessment of the mitigating impact of av treatment alone, based on the initial conditions estimated for the h n outbreak, and assuming the implementation of the same av protocol in all countries of the world. we adopt a seir-like compartmentalization to model influenza-like illnesses [ ] in which we include the systematic successful treatment of % of the symptomatic infectious individuals (see figure ). the efficacy of the figure compartmental structure in each subpopulation in the intervention scenario. a modified susceptible-latent-infectious-recovered model is considered, to take into account asymptomatic infections, traveling behavior while ill, and use of antiviral drugs as a pharmaceutical measure. in particular, infectious individuals are subdivided into: asymptomatic (infectious_a), symptomatic individuals who travel while ill (infectious_s_t), symptomatic individuals who restrict themselves from travel while ill (infectious_s_nt), symptomatic individuals who undergo the antiviral treatment (infectious_avt). a susceptible individual interacting with an infectious person may contract the illness with rate beta and enter the latent compartment where he/she is infected but not yet infectious. the infection rate is rescaled by a factor ra in case of asymptomatic infection [ , ] , and by a factor ravt in case of a treated infection. at the end of the latency period, of average duration equal to eps - , each latent individual becomes infectious, showing symptoms with probability -p a , whereas becoming asymptomatic with probability p a [ , ] . change in travelling behavior after the onset of symptoms is modeled with probability p t set to % that individuals would stop travelling when ill [ ] . infectious individuals recover permanently after an average infectious period mu - equal to . days. we assume the antiviral treatment regimen to be administered to a % fraction (i.e. pavt = . ) of the symptomatic infectious individuals within one day from the onset of symptoms, reducing the infectiousness and shortening the infectious period of day. [ , ] . av treatment is accounted for in the model by a % reduction in the transmissibility of the disease by an infected person under av treatment when av drugs are administered in a timely fashion [ , ] . we assume that the drugs are administered within day of the onset of symptoms and that the av treatment reduces the infectious period by day [ , ] . the scenario with av treatment is compared to the baseline case in which no intervention is considered, i.e. the probability of treatment is set equal to in all countries. the gleamviz simulation results are shown in figure where the incidence profiles in two different regions of the world, north america and western europe, are shown for both the baseline case and the intervention scenario with av treatment. the results refer to the median (solid line) and % reference range (shaded area) obtained from stochastic realizations of each scenario starting from the same initial conditions. the resulting incidence profiles of the baseline case peak at around mid-november and the end of november in the us and western europe, respectively. these results show an anticipated peak of activity for the northern hemisphere with respect to the expected peak time of seasonal influenza. in order to make a more accurate comparison with the surveillance data in these regions, we should rely on the predictions provided by models that can take into account the full spectrum of strategies that were put in place during the h n outbreak, viz. the predictions obtained by gleam [ ] . in the case of a rapid and efficient implementation of the av treatment protocol at the worldwide level, a delay of about weeks would be obtained in the regions under study, a result that could be essential in gaining time to deploy vaccination campaigns targeting high-risk groups and essential services. in addition, the gleamviz tool provides simulated results for the number of av drugs used during the evolution of the outbreak. if we assume treatment delivery and successful administration of the drugs to % of the symptomatic cases per day, the number of av drugs required at the activity peak in western europe would be . courses per , persons, and the size of the stockpile needed after the first year since the start of the pandemic would be about % of the population. again, we assume a homogeneous treatment protocol for all countries in the world; results may vary from country to country depending on the specific evolution of the pandemic at the national level. computer-based simulations provide an additional instrument for emerging infectious-disease preparedness and control, allowing the exploration of diverse scenarios and the evaluation of the impact and efficacy of various intervention strategies. here we have presented a computational tool for the simulation of emerging ili infectious diseases at the global scale based on a datadriven spatial epidemic and mobility model that offers an innovative solution in terms of flexibility, realism, and computational efficiency, and provides access to sophisticated computational models in teaching/training settings and in the use and exploitation of large-scale simulations in public health scenario analysis. project name: gleamviz simulator v . project homepage: http://www.gleamviz.org/simulator/ operating systems (client application): windows (xp, vista, ), mac os x, linux. programming language: c++ (gleamsim core), python (gleamproxy, gleamsim wrapper), action-script (gleamviz) other requirements (client application): adobe air runtime, at least mb of free disk space. license: saas baseline scenario scenario with av figure simulated incidence profiles for north america and western europe in the baseline case (left panels) and in the av treatment scenario (right panels). the plots are extracted from the gleamviz tool visualization. in the upper plots of each pair the curves and shaded areas correspond to the median and % reference range of stochastic runs, respectively. the lower curves show the cumulative size of the infection. the dashed vertical line marks the same date for each scenario, clearly showing the shift in the epidemic spreading due to the av treatment. any restrictions to use by non-academics: none. the server application can be requested by public institutions and research centers; conditions of use and possible restrictions will be evaluated specifically. additional file : the gleamviz computational tool: additional file. this file includes information for installing the gleamviz client and details of the features of its various components. the transmissibility and control of pandemic influenza a (h n ) virus potential for a global dynamic of influenza a (h n ) seasonal transmission potential and activity peaks of the new influenza a(h n ): a monte carlo likelihood analysis based on human mobility modelling disease outbreaks in realistic urban social networks epifast: a fast algorithm for large scale realistic epidemic simulations on distributed memory systems multiscale mobility networks and the spatial spreading of infectious diseases strategies for containing an emerging influenza pandemic in southeast asia mitigation strategies for pandemic influenza in the united states mitigation measures for pandemic influenza in italy: an individual based model considering different scenarios flute, a publicly available stochastic influenza epidemic simulation model the influenza pandemic preparedness planning tool influsim an extensible spatial and temporal epidemiological modelling system centers for disease control and prevention (cdc) modeling the spatial spread of infectious diseases: the global epidemic and mobility computational model centers for disease control and prevention (cdc) controlling pandemic flu: the value of international air travel restrictions a mathematical model for the global spread of influenza assessing the impact of airline travel on the geographic spread of pandemic influenza forecast and control of epidemics in a globalized world delaying the international spread of pandemic influenza modeling the worldwide spread of pandemic influenza: baseline case and containment interventions predictability and epidemic pathways in global outbreaks of infectious diseases: the sars case study socioeconomic data and applications center (sedac). columbia university the architecture of complex weighted networks estimating spatial coupling in epidemiological systems: a mechanistic approach a structured epidemic model incorporating geographic mobility among regions infectious diseases in humans the role of airline transportation network in the prediction and predictability of global epidemics the modeling of global epidemics: stochastic dynamics and predictability modeling vaccination campaigns and the fall/winter activity of the new a (h n ) influenza in the northern hemisphere python programming language twisted matrix networking engine adobe flex framework modeling the critical care demand and antibiotics resources needed during the fall wave of influenza a (h n ) pandemic world health organization: pandemic preparedness antiviral treatment for the control of pandemic influenza: some logistical constraints hedging against antiviral resistance during the next influenza pandemic using small stockpiles of an alternative chemotherapy containing pandemic influenza with antiviral agents containing pandemic influenza at the source potential impact of antiviral drug use during influenza pandemic modelling of the influenza a(h n )v outbreak in mexico city secretaría de comunicaciones y transportes the who rapid pandemic assessment collaboration: pandemic potential of a strain of influenza a(h n ): early findings we are grateful to the international air transport association for making the airline commercial flight database available to us. this work has been partially funded by the nih r -da award, the lilly endowment grant - and the dtra- - award to av; the ec-ict contract no. (epiwork) to av, vc, and wvdb; the ec-fet contract no. (dynanets) to av and vc; the erc ideas contract n.erc- -stg (epifor) to vc, cg, and mq. authors' contributions cg, wvdb and bg designed the software architecture. wvdb and mq developed the client application. bg implemented the gleam engine. cg developed the proxy middleware. cg, vwdb, vc and av drafted the manuscript. mq and bg helped draft the manuscript. av and vc conceived and coordinated the software project, designed and coordinated the application study. all authors read and approved the final manuscript.competing interests av is consulting and has a research agreement with abbott for the modeling of h n diffusion. the other authors have declared that no competing interests exist. key: cord- -paax pqh authors: henkel, jan; woodruff, maria a.; epari, devakara r.; steck, roland; glatt, vaida; dickinson, ian c.; choong, peter f. m.; schuetz, michael a.; hutmacher, dietmar w. title: bone regeneration based on tissue engineering conceptions — a st century perspective date: - - journal: bone res doi: . /br sha: doc_id: cord_uid: paax pqh the role of bone tissue engineering in the field of regenerative medicine has been the topic of substantial research over the past two decades. technological advances have improved orthopaedic implants and surgical techniques for bone reconstruction. however, improvements in surgical techniques to reconstruct bone have been limited by the paucity of autologous materials available and donor site morbidity. recent advances in the development of biomaterials have provided attractive alternatives to bone grafting expanding the surgical options for restoring the form and function of injured bone. specifically, novel bioactive (second generation) biomaterials have been developed that are characterised by controlled action and reaction to the host tissue environment, whilst exhibiting controlled chemical breakdown and resorption with an ultimate replacement by regenerating tissue. future generations of biomaterials (third generation) are designed to be not only osteoconductive but also osteoinductive, i.e. to stimulate regeneration of host tissues by combining tissue engineering and in situ tissue regeneration methods with a focus on novel applications. these techniques will lead to novel possibilities for tissue regeneration and repair. at present, tissue engineered constructs that may find future use as bone grafts for complex skeletal defects, whether from post-traumatic, degenerative, neoplastic or congenital/developmental “origin” require osseous reconstruction to ensure structural and functional integrity. engineering functional bone using combinations of cells, scaffolds and bioactive factors is a promising strategy and a particular feature for future development in the area of hybrid materials which are able to exhibit suitable biomimetic and mechanical properties. this review will discuss the state of the art in this field and what we can expect from future generations of bone regeneration concepts. after years of tissue engineering & regenerative medicine . and another years of . versions ( ) the era of tissue engineering . has begun. this review will des-care outcomes. today major reconstructive surgeries (due to trauma or tumour removal) are still limited by the paucity of autologous materials available and donor site morbidity. recent advances in the development of scaffold-based tissue engineering (te) have given the surgeon new options for restoring form and function. there are now bioactive biomaterials (second generation) available that elicit a controlled action and reaction to the host tissue environment with a controlled chemical breakdown and resorption to ultimately be replaced by regenerating tissue. third-generation biomaterials are now being designed to stimulate regeneration of living tissues using tissue engineering and in situ tissue regeneration methods. engineering functional bone using combinations of cells, scaffolds and bioactive factors are seen as a promising approach and these techniques will undoubtedly lead to ceaseless possibilities for tissue regeneration and repair. there are currently thousands of research papers and reviews available on bone tissue engineering, but there is still a major discrepancy between scientific research efforts on bone tissue engineering and the clinical application of such strategies. there is an evident lack of comprehensive reviews that cover both the scientific research aspect as well as the clinical translation and practical application of bone tissue engineering techniques. this review will therefore discuss the state of the art of scientific bone tissue engineering concepts and will also provide current approaches and future perspectives for the clinical application of bone tissue engineering. bone as an organ has next to its complex cellular composition a highly specialised organic-inorganic architecture which can be classified as micro-and nanocomposite tissue. its mineralised matrix consists of ) an organic phase (mainly collagen, % dry weight) responsible for its rigidity, viscoelasticity and toughness; ) a mineral phase of carbonated apatite ( % dry weight) for structural reinforcement, stiffness and mineral homeostasis; and ) other non-collagenous proteins that form a microenvironment stimulatory to cellular functions ( ) . bone tissue exhibits a distinct hierarchical structural organization of its constituents on numerous levels including macrostructure (cancellous and cortical bone), microstructure (harversian systems, osteons, single trabeculae), sub-microstructure (lamellae), nanostructure (fibrillar collagen and embedded minerals) and subnanostructure (molecular structure of constituent elements, such as mineral, collagen, and non-collagenous organic proteins) ( figure ) ( ) . macroscopically, bone consists of a dense hard cylindrical shell of cortical bone along the shaft of the bone that becomes thinner with greater distance from the centre of the shaft towards the articular surfaces. cortical bone encompasses increasing amounts of porous trabecular bone (also called cancellous or spongy bone) at the proximal and distal ends to optimise articular load transfer ( ) . in humans, trabecular bone has a porosity of - % with an average trabecular spacing of around mm and an average density of approximately . g·cm - ( ) ( ) ( ) . cortical bone has a much denser structure with a porosity of - % and an average density of . g·cm - ( , ) . on a microscopic scale, trabecular struts and dense cortical bone are composed of mineralized collagen fibres stacked parallel to form layers, called lamellae ( - µm thick) and then stacked in a± ° manner ( ) . in mature bone these lamellae wrap in concentric layers ( - lamellae) around a central part named haversian canal which containings nerve and blood vessels to form what is called an osteon (or a haversian system), a cylindrical structure running roughly parallel to the long axis of the bone ( ) . cancellous bone consists of interconnecting framework of rod and plate shaped trabeculae. on a nanostructural level, the most prominent structures are the collagen fibres, surrounded and infiltrated by mineral. at the sub-nanostructural level three main materials are bone crystals, collagen molecules, and non-collagenous organic proteins. for further details the reader is referred to ( ) . mineralised bone matrix is populated with four boneactive cells: osteoblasts, osteoclasts, osteocytes and bone lining cells. additional cell types are contained within the bone marrow that fills the central intramedullary canal of the bone shaft and intertrabecular spaces near the articular surfaces ( ) . bone has to be defined as an organ composed of different tissues and also serves as a mineral deposit affected and utilised by the body's endocrine system to regulate (among others) calcium and phosphate homeostasis in the circulating body fluids. furthermore, recent studies indicate that bone exerts an endocrine function itself by producing hormones that regulate phosphate and glucose homeostasis integrating the skeleton in the global mineral and nutrient homeostasis ( ) . bone is a highly dynamic form of connective tissue which undergoes continuous remodelling (the orchestrated removal of bone by osteoclasts followed by the formation of new bone by osteoblasts) to optimally adapt its structure to changing functional demands (mechanical loading, nutritional status etc.). from a material science point of view bone matrix is a composite material of a polymer-ceramic lamellar fibre-matrix and each of these design and material aspects influence the mechanical properties of the bone tissue ( ) . the mechanical properties depend on the bone composition (porosity, mineralisation etc.) as well as the structural organisation (trabecular or cortical bone architecture, collagen fibre orientation, fatigue damage etc.) ( ) . collagen possesses a young's modulus of - gpa and an ultimate tensile strength of - mpa, compared to the mineral hydroxyapatite which has a young's modulus of ~ gpa and an ultimate tensile strength of ~ mpa. the resulting mechanical properties of the two types of bone tissue, namely the cortical bone and cancellous bone, are shown in table . age and related changes in bone density have been reported to substantially influence the mechanical properties of cancellous bone ( ) . as outlined above, bone shows a distinct hierarchical structural organization and it is therefore important to also define the mechanical properties at microstructural levels ( table ) . although the cancellous and cortical bone may be of the same kind of material, the maturation of the cortical bone material may alter the mechanical properties at the microstructural level. bone tissue is also known to be mechano-receptive; both normal bone remodelling and fracture or defect healing are influenced by mechanical stimuli applied at the regenerating defect site and surrounding bone tissue ( ) ( ) ( ) ( ) . in contrast to most other organs in the human body, bone tissue is capable of true regeneration, i.e. healing without the formation of fibrotic scar tissue ( ) . during the healing process basic steps of fetal bone development are recapitulated and bone regenerated in this way does not differ structurally or mechanically from the surrounding undamaged bone tissue ( ) . however, despite this tremendous regenerative capacity, table mechanical properties of compact (cortical) and spongy (cancellous) bone. reproduced and modified from ( ) . dry specimen (submicrostructure) ( ) - % of all fractures are prone to delayed bony union or will progress towards a non-union and the development of a pseudarthrosis ( ) ( ) . together with large traumatic bone defects and extensive loss of bone substance after tumour resection or revision surgery after failed arthroplasties, these pathological conditions still represent a major challenge in today's clinical practice. the rangeof bone graft materials available to treatsuch problems in modern clinical practice essentially include autologous bone (from the same patient), allogeneic bone (from a donor), and demineralised bone matrices, as well as a wide range of synthetic bone substitute biomaterials such as metals, ceramics, polymers, and composite materials. during the last decades, tissue engineering strategies to restore clinical function have raised considerable scientific and commercial interest in the field of orthopaedic surgery as well as reconstructive and oromaxillofacial surgery. yet, the treatment of bone defects and the search for bone substitute materials is not just a modern day phenomenon, with its history reaching back through millennia. the quest for the most efficient way to substitute for lost bone and to develop the best bone replacement material has been pursued by humans for thousands of years. in peru, archaeologists discovered the skull of a tribal chief from bc in which a frontal bone defect (presumably from trepanation) had been covered with a mm-thick plate of hammered gold ( ) . trephined incan skulls have been found with plates made from shells, gourds, and silver or gold plates covering the defect areas ( ) . in a skull found in the ancient center of ishtkunui (armenia) from approx. bc, a mm diameter skull defect had been bridged with a piece of animal bone ( ). these pursuits are not limited to skull surgeries involving bone substitutes. ancient egyptians have been shown to have profound knowledge of orthopaedic und traumatological procedures with surgeons having implanted iron prostheses for knee joint replacement as early as bc, as analyses of preserved human mummies have revealed ( ). the first modern era report of a bone xenograft procedure is believed to be the dutch surgeon job janszoon van meekeren in ( - ). a skull defect of a russian nobleman was successfully treated with a bone xenograft taken from the calvaria of a deceased dog. the xenograft was reported to have become fully incorporated into the skull of the patient. in the s, plaster of paris (calcium sulphate) was used to fill bone cavities in patients suffering from tuberculosis ( ). attempts were also made to fill bone defects with cylinders made from ivory ( ). in the german surgeon phillips von walters described the first clinical use of a bone autograft to reconstruct skull defects in patients after trepanation ( ). walters successfully repaired trepanation holes, following surgery to relieve intracranial pressure, with pieces of bone taken from the patient's own head. in , scottish surgeon william macewen described the first allogenic bone grafting procedure: he used tibial bone wedges from three donors that had undergone surgery for skeletal deformity correction (caused by rickets) to reconstruct an infected humerus in a -year-old child ( ) major contributions leading to the development of modern day bone grafting procedures and bone substitutes have been made by ollier and barth in the late s. louis léopold ollier carried out extensive experiments to study the osteogenic properties of the periosteum and other various approaches to new bone formation, mainly in rabbit and dog models. he also meticulously reviewed the literature on bone regeneration available at that time and in he published his -page textbook 'traite experimentel et clinique de la regeneration des os et de la production artificielle du tissu osseux', in which he described the term 'bone graft' ("greffe osseuse") for the first time ( ) . in the german surgeon arthur barth published his treatise 'ueber histologische befunde nach knochenimplantationen' ('on histological findings after bone implantations') presenting his results of various bone grafting procedures involving the skull and long bones (humerus, forearm bones) of dogs and rabbits including histological assessment ( ) . today, both ollier's and barth's work are considered to be milestones in the development of present day bone grafting procedures and bone substitute materials. with the development of new orthopaedic techniques and increased numbers of joint replacement procedures (prostheses), the demand for bone grafts increased in the th century, leading to the opening of the first bone bank for allogenic bone grafts in new york in ( ) . but the risk of an immunological reaction from transplanted allogenic bone material was soon recognized and addressed in various studies ( ) ( ) . several procedures such as the use of hydrogen peroxide to macerate bone grafts ("kieler span") in the s and s to overcome antigenity were not successful ( ) ( ) . today, bone substitute materials such as (bovine) bone chips are routinely used in clinical practice after being pretreated to remove antigen structures. however, due to the processing steps necessary to abolish antigenicity, most of these grafts do not contain viable cells or growth factors and are therefore inferior to viable autologous bone graft options. when allografts with living cells are transplanted, there is a risk of transmitting viral and bacterial infections: transmission of human immunodeficiency virus (hiv), hepatitis c virus (hcv), human t-lymphozytic virus (htlv), unspecified hepatitis, tuberculosis and other bacteria has been documented (mainly) for allografts (mainly from those containing viable cells) ( ) . as early as , the work of the swiss h. matti proved the paramount meaning of autologous cancellous bone grafts for bone regeneration approaches ( ) . having conducted various experiments on the osteogenic potential of autologous and allogenic bone, schweiberer concluded in that the autologous transplant remains the only really reliable transplantation material of the future, if applied to bring about new bone formation or crucially to support the bridging bone defects ( ) . even though this statement was made more than years ago, it still remains valid today, when bone is still the second most transplanted material, second only to blood. worldwide more than . million bone grafts (either autografts or allografts) are performed each year ( ) . recent advances in technology and surgical pro-cedures have significantly increased the options for bone grafting material, with novel products designed to replace both the structural properties of bone, as well as promote faster integration and healing. the number of procedures requiring bone substitutes is increasing, and will continue to do so as the population ages and physical activity of the elderly population increases. therefore, while the current bone grafting market globally is estimated to be in excess of $ . billion us each year, it is expected to increase at a compound annual growth rate of - % ( ) . although the last decades have seen numerous innovations in bone substitute materials, the treatment of bone defects with autologous bone grafting material is still considered to be the 'gold standard' against which all other methods are compared ( ) . autologous bone combines all the properties desired in a bone grafting material: it provides a scaffold for the ingrowth of cells necessary for bone regeneration (=osteoconductive); it promotes the proliferation of stem cells and their differentiation into osteogenic cells (=osteoinductive) and it holds viable cells that can form new bone tissue (= osteogenic) ( , ) . however, the available volume of autologous bone graft from a patient is limited and an additional surgical procedure is required to harvest the grafting material which is associated with a significant risk of donor site morbidity. - % of autograft patients experience morbidity such as chronic pain or dysaesthesia at the graft-harvesting site ( ) . large bone defects (> cm) may be treated with bone segment transport or free vascularized bone transfer ( ) , as the use of an autologous bone graft alone is not recommended because of the risk of graft resorption despite good soft tissue coverage ( ) . the vascularised fibula autograft ( ) and the ilizarov method ( ) ( ) ( ) are the most commonly used treatment methods for larger bone defects; however, complications are common and the process can be laborious and painful for the patient as she/he may be required to use external fixation systems for up to one and half years ( , ( ) ( ) . the limitations of existing bone grafting procedures, either autologous or allogenic in nature, and the increased demand for bone grafts in limb salvage surgeries for bone tumours and in revision surgeries of failed arthroplasties have renewed the interest in bone substitute materials and alternative bone grafting procedures ( ) . in , masquelet and colleagues ( ) first described a new two-stage technique taking advantage of the body's immune response to foreign materials for bone reconstruction. the authors called it the 'concept of induced membranes' -soon to become known as the 'masquelet technique': in a first step, a radical debridement of necrotic bone and soft tissue is followed by the filling of the defect site with a polymethylmethacrylate (pmma) spacer and stabilisation with an external fixator. after the definitive healing of the soft tissue, a second procedure is performed - weeks later, when the pmma spacer is removed and a morcellised cancellous bone graft (from the iliac crest) is inserted into the cavitiy ( - ). the cement spacer was initially thought to prevent the collapse of the soft tissue into the bone defect and to prepare the space for bone reconstruction. however, it was soon discovered that the pmma spacer does not only serve as a place holder, but that a foreign body reaction to the spacer also induces the formation of a membrane that possesses highly desirable properties for bone regeneration ( - ): the induced membrane was shown to be richly vascularised in all layers; the inner membrane layer (facing the cement) composed of synovial like epithelium and the out part is made from fibroblasts, myoblasts and collagen. the induced membrane has also been shown to secrete various growth factors in a time-dependent manner: high concentrations of vascular endothelial growth factor (vegf) as well as transforming growth factor β (tgf β) are secreted as early as the second week after implantation of the pmma spacer; bone morphogenetic protein (bmp- ) concentration peaks at the fourth week. the induced membrane stimulates the prolifera-tion of bone marrow cells and differentiation towards an osteoblastic lineage. finally, clinical experience has shown that the cancellous bone inside the induced membrane is not subject to resorption by the body. ever since its introduction the 'induced membrane'-technique has been used very successfully in various clinical cases (see ( ) and references therein). however, the masquelet technique still requires the harvesting of an autologous bone graft, and with that come all the potential aforementioned complications. furthermore, the use of alternate bone substitute materials, such as hydroxyapatite tricalcium phosphate, in combination with the masquelet technique has so far yielded results inferior to the use the masquelet technique with autologous bone grafting material ( , ) . besides the masquelet technique, a more recent innovation has also significantly improved the clinical approach to restoring bone defects. the development of the reamer-irrigator-aspirator (ria © )-system (depuy-synthes) has given clinicians an alternative to iliac crest harvesting to retrieve bone grafting materials from patients: the ria system provides irrigation and aspiration during intramedullay reaming, allowing the harvesting of finely morselised autologous bone and bone marrow for surgical procedures requiring bone grafting material ( ). the ria was initially developed to lower the intramedullary pressure during the reaming of long bones to reduce the risk of fat embolisms and pulmonary complications such as the acute respiratory distress syndrom (ards), as well as to reduce local thermal necrosis of bone tissue ( ) ( ) . however, the finely morselised autologous bone and bone marrow that is collected by the ria has been shown to be rich in stem cells, osteogenic cells and growth factors and has been recognized to be a suitable bone graft alternative to the iliac crest autograft tissue ( ) ( ) . also, ria enables the harvesting of larger bone graft volumes compared to the iliac crest (approx. cm for the femur and cm for the tibia) ( , ) . furthermore, the risk of complications from the harvesting procedure has been reduced significantly (ria % vs. . % for iliac crest autografts) ( ) . since its introduction, the indications for use of ria have been further extended to include the treatment of postoperative osteomyelitis ( ) and the harvesting of mesenchymal stem cells (mscs) ( ) . the innovation driven by the ria systems was so significant, that the journal "injury" has dedicated a complete issue to the data available on ria and its applications recently ( ) . a systematic review on the reamer-irrigator-aspirator indications and clinical results has recently been published by cox et al ( ) . the masquelet technique as well as the ria-system is nowadays frequently used in clinical practice, in-dependently. however, the two techniques may also be combined to further improve their effectiveness when treating severe bone defects, for example in posttraumatic limb reconstruction ( ) . an example of a case from one of our author's clinical practice (m.s.) combining the use of masquelet technique and the use of the ria-system to treat a complex case of tibial nonunion is provided in figure . both the masquelet technique and the development of the ria-system represent significant improvements in today's clinical approach to bone reconstruction and regeneration. however, utilising these techniques, we have still not been able to replace autologous bone grafting in order to avoid surgical graft retrieval procedures with all the associated disadvantages. however, with research looking towards increasingly sophisticated bone tissue engineering techniques and their first clinical applications, the quest for developing improved bone substitute material advances to the next level. bone substitutes can be defined as "a synthetic, inorganic or biologically organic combination-biomaterialwhich can be inserted for the treatment of a bone defect instead of autogenous or allogenous bone" ( ) . this definition applies to numerous substances and a variety of materials have been used over time in an attempt to substitute bone tissue. although merely of historic interest and with no significance in modern therapies, the use of seashells, nuts, gourds and so forth show that humans have strived for bsm for thousands of years. with the introduction of tissue engineering and its clinical application the regenerative medicine in ( ) the modern day quest for bsms has undergone a significant change. the limitations of current clinical approaches have necessitated the development of alternative bone repair techniques and have driven the development of scaffold-based tissue engineering strategies. in the past, mostly inert bone substitute materials have been used, functioning mainly as space holders during the healing processes. now a paradigm shift has taken place towards the use of new 'intelligent' tissue engineering biomaterials that would support and even promote tissue re-growth ( ) . according to the "diamond concept" of bone tissue engineering ( ) ( ) , an ideal bone substitute material should offer an osteoinductive three-dimensional structure, contain osteogenic cells and osteoinductive factors, have sufficient mechanical properties and promote vascularisation. despite extensive research in the field of bone tissue engineering, apart from the "gold standard" figure clinical case combining the masquelet-technique and the ria-system to treat a tibial non-union. year old male acquired a gustillo b fracture of the right tibia and fibula and was treated with a stage procedure with locked plating and a free flap . the patient's progress was very slow and an implant failure occurred months post-operatively (a). the patient was then referred for the further management and underwent debridement of the non-union site on the distal tibia by lifting the flap (b). the size of the extensive bone defect is shown in b (intraoperative image of situs and x-ray image with retractor in defect site). additionally, a pmma bone cement spacer was inserted into the tibial defect as part of the masquelet technique. postop x-ray images after surgery with the pmma spacer (circles) in place (c). weeks later the pmma spacer was removed and the induced membrane at the defect site was packed with autologous cancellous bone graft obtained from the femur using the reamer-irrigator-aspirator (ria) technique. (d) shows assembled ria system, insert showing morselised autologous bone and bone marrow graft obtained. postop films after the second surgery (e). weeks after bone grafting the defect showed good healing and patient was able to fully bear weight as tolerated. over the following months x-ray images showed progressive bridging of the zone and he was able to return to work with light duties. he was reviewed again months post-surgery and had returned to work full-time and was walking long distances without any support (f). autograft bone, no currently available bsm can offer these properties in one single material. therefore, the fundamental concept underlying tissue engineering is to combine a scaffold or three-dimensional construct with living cells, and/or biologically active molecules to form a "tissue engineering construct" (tec), which promotes the repair and/or regeneration of tissues ( ) ( ) . currently used bsm can be classified into different subgroups according to their origin ( , ): ) bsm of natural origin this group consists of harvested autogenous bone grafts as well as allogenic bsm, such demineralised bone matrix, corticocancellous or cortical grafts, cancellous chips (from either cadavers or living donors) ( ) ( ) ( ) . xenogenic materials, for example porous natural bone hydroxyapatite from animal bones (bovine, equine, porcine etc.) are also part of this group ( ) . phytogenic materials such as bone-analogue calcium phosphate originally obtained from marine algae or coral derived materials, also fall into this category ( - ). this groups contains ceramics such as bioactive glasses ( ) , tricalciumphosphates (tcp) ( ) ( ) , hydroxyapatite (ha) ( ) ( ) ( ) and glass ionomer cements as well as calcium phosphate (cp) ceramics ( ) . metals such as titanium also belong to this group. furthermore polymers including polymethylmethacrylate (pmma), polylactides/ poliglycolides and copolymers as well as polycaprolactone (pcl) ( ) are summarised in this group ( , , ( ) ( ) . ) composite materials bsm combining different materials such as ceramics and polymers are referred to as composite materials ( , ( ) ( ) . by merging materials with different structural and biochemical properties into composite materials, the properties of composite materials can be modified to achieve more favourable characteristics, for instance with respect to biodegradability ( , ) . ) bsm combined with growth factors natural or recombinant growth factors such a bone morphogenic protein (bmp), platelet-derived growth factor (pdgf), transforming growth factor-ß (tgf-β), insulin-like growth-factor , vascular endothelial growth factor (vegf) and fibroblast growth factor can be added to increase the biological activity of bsm ( - ). for example, a composite material made of medicalgrade polycaprolactone-tricalcium phosphate (mpcl-tcp) scaffolds (combined with recombinant human bmp- ) has been demonstrated to completely bridge a critical-sized ( cm) tibial defect in a sheep model ( ) . ) bsm with living cells mesenchymal stem cells ( ) ( ) ( ) , bone marrow stromal cells ( ) ( ) , periosteal cells ( ) ( ) , osteoblasts ( ) and embryonic ( ) as well as adult stem cells ( ) have been used in bone tissue engineering ( , , ( ) ( ) ( ) ( ) . these cells can generate new tissue alone or can be used in combination with scaffold matrices. bsms can also be classified according to their properties of action. an overview of the currently available bsm for clinical (orthopaedic) use and their mode of action is given in table (reproduced from ( )). scaffolds serve as three-dimensional structures to guide cell migration, proliferation and differentiation. in load bearing tissues, it also serves as temporary mechanical support structure. scaffolds substitute for the function of the extracellular matrix and need to fulfil highly specific criteria. an ideal scaffold should be (i) three-dimensional and highly porous with an interconnected pore network for cell growth and flow transport of nutrients and metabolic waste; (ii) should have surface properties which are optimized for the attachment, migration, proliferation and differentiation of cell types of interest (depending on the targeted tissue); (iii) be biocompatible, not elicit an immune response and be biodegradable with a controllable degradation rate to compliment cell/tissue in-growth and maturation; (iv) its mechanical properties should match those of the tissue at the site of implantation and (v) the scaffold structure should be easily and efficiently reproducible in various shapes and sizes ( ) . biocompatibility represents the ability of a material to perform with an appropriate response in a specific application ( ) . as a general rule, scaffolds should be fabricated from materials that do not have the potential to elicit immunological or clinically detectable primary or secondary foreign body reactions ( ) . parallel to the formation of new tissue in vivo, the scaffold may undergo degradation via the release of by-products that are either biocompatible without proof of elimination from the body (biodegradable scaffolds) or can be eliminated through natural pathways from the body, either by simple filtration of by-products or after their metabolisation (bioresorbable scaffolds) ( ) . due to poor vascularisation or low metabolic activity, the capacity of the surrounding tissue to eliminate the by-products may be low leading to a build up of the by-products thereby causing local temporary disturbances ( ) : a massive in vivo release of acidic degradation by-products leading to inflammatory reactions has been reported for several ( ) ( ) ( ) . another example is the increase of osmotic pressure or ph caused by local fluid accumulation or transient sinus formation from fibre reinforced polyglycolide pins used in orthopaedic applications ( ) . it is also known that calcium phosphate biomaterial particles can cause inflammatory reactions after being implanted (although this inflammatory reaction may be considered desirable to a certain extent as it subsequently stimulates osteoprogenitor cell differentiation and bone matrix deposition) ( ) . these examples illustrate that potential problems related to biocompatibility in tissue engineering constructs for bone and cartilage applications may be related to the use of biodegradable, erodible and bioresorbable polymer scaffolds. therefore, it is important that the three dimensional tissue engineering construct (tec) is exposed at all times to sufficient quantities of neutral culture media when undertaking cell culture procedures, especially during the period where the mass loss of the polymer matrix occurs ( ) . for applications in vivo, it is of course not possible to expose the tec to neutral media, and one therefore has to carefully take into account the local specifications (ph, vascularisation, metabolic activity etc) of the tissue to be engineered when accessing biocompatibility of a tec. the design of tissue engineering scaffolds needs to consider physico-chemical properties, morphology and bio-mechanical properties as well as degradation kinetics. the scaffold structure is expected to guide the development of new bone formation by promoting attachment, migration, proliferation and differentiation of bone cells. parallel to tissue formation, the scaffold should also undergo degradation in order to allow for ultimate replacement of scaffold material with newly formed, tissue engineered bone. furthermore, the scaffold is also responsible for (temporal) mechanical support and stability at the tissue engineering site until the new bone is fully matured and is able to withstand mechanical load. as a general rule, the scaffold material should be sufficiently robust to resist changes in shape resulting from the introduction of cells into the scaffold (each of which should capable of exerting tractional forces) and from wound contraction forces that would be evoked during tissue healing in vivo ( ) . in order to achieve optimal results, it is therefore necessary to carefully balance the biomechanical properties of a scaffold with its degradation kinetics. a scaffold material has to be chosen that degrades and resorbs at a controlled rate, giving the tec sufficient mechanical stability at all times, but at the same time allowing new in vivo formed bone tissue to substitute for its structure. figure depicts the interdependence of molecular weight loss and mass loss of a slow degrading composite scaffold and also shows the corresponding stages of tissue regeneration ( ) . at the time of implantation the biomechanical properties of a scaffold should match the structural pro-perties of the tissue it is implanted into as closely as possible ( ) . it should possess sufficient structural integrity for the period until the engineered tissue ingrowth has replaced the slowly disappearing scaffold matrix with regards to mechanical properties. in bone tissue engineering the degradation and resorption kinetics of the scaffold have to be controlled in such a way that the bioresorbable scaffold retains its physical properties for at least months to enable cell and tissue remodelling to achieve stable biomechanical conditions and vascularisation at the defect site ( ) . apart from host anatomy and physiology, the type of tissue that is aimed to be engineered also has a profound influence on the degree of remodelling: in cancellous bone the remodelling takes - months, while cortical bone will take twice as long, approximately - months, to remodel ( ) . whether the tec will be part of a load bearing or non-load bearing site will also significantly influence the needs for mechanical stability of the tec as mechanical loading can directly affect the degradation behaviour as well ( ) . utilising orthopaedic implants to temporarily stabilise the defect area also influences the requirements for biomechanical stability of the tec significantly ( , ) . it is therefore crucial to meticulously select the scaffold material individually for each tissue engineering approach to tailor the mechanical properties and degradation kinetics exactly to the purpose of the specific tec ( ) . consequently, there is not one "ideal scaffold material" for all bone tissue engineering purposes, but the choice depends on the size, type and location of the bone tissue to be regenerated. the surface area of a scaffold represents the space where pivotal interactions between biomaterial and host tissue take place. the performance of a tec depends fundamentally on the interaction between biological fluids and the surface of the tec, and it is often mediated by proteins absorbed from the biological fluid ( ) . the initial events include the orientated adsorption of molecules from the surrounding fluid, creating a specific interface to which the cells and other factors respond to the macrostructure of the scaffold as well as the microtopography and chemical properties of the surface determine which molecules are adsorbed and how cells will attach and align themselves ( ) . the focal attachments made by the cells with their substrate then determines cell shape, which in turn transduces signals via the cytoskeleton to the nucleus resulting in expression of specific proteins which may be structural or signal-related and contribute towards the cell phenotype. due to technical progress, we are now able to manipulate materials at the atomic, molecular, and supramolecular level, and bulk materials and surfaces can be designed at a similar dimension to that of the nanometer constituent components of bone ( ): in natural bone, hydroxyapatite plates are approximately between nm in width and nm in length while collagen type is a triple helix nm in length, . nm in width and with a periodicity of nm ( ) . "nanomaterials" commonly refers to materials with basic structural units in the range - nm (nanostructured), crystalline solids with grain sizes between and nm (nanocrystals), individual layers or multilayer surface coatings in the range - nm (nanocoatings), extremely fine powders with an average particle size in the range - nm and fibres with a diameter in the range - nm (nanofibres) ( ) . the close proximity of the scale of these materials to the scale of natural bone composites makes the application of nanomaterials for bone tissue engineering a very promising strategy. surfaces with nanometer topography can promote the availability of amino acid and proteins for cell adhesion to a great extent, for example, the adsorption of fibronectin and vitronectin [two proteins known to enhance osteoblast and bone forming cell function ( ) ] can be significantly increased by decreasing the grain size on the scaffold/implant surface below nm ( ) . it has also been shown that calcium-mediated cell protein adsorption on nanophase material promotes unfolding of these proteins promoting bone cell adhesion and function ( ) . current literature supports the hypothesis that by creating surface topographies with characteristics that approximate the size of proteins, a certain control over protein adsorption and interactions will be possible. since the surface characteristics regarding roughness, topography and surface chemistry are then transcribed via the protein layer into information that is comprehensible for the cells ( ) , this will enable the fabrication of surface properties directly targeted at binding specific cell types. in vitro, osteoblast adhesion, proliferation and differentiation and calcium deposition is enhanced on nanomaterials with grain sizes less than nm ( ) ( ) . the adherence of osteoblasts has been shown to increase up to threefold when the surface is covered with nanophase titanium particles instead of conventional titanium particles ( ) . nano-and microporosiy has also been shown to promote osteogenic differentiation ( ) and osteogenesis ( ) . the use of nanomaterials to achieve better osteointegration of orthopaedic implants and for bone tissue engineering approaches has been extensively summarised in several recent reviews ( , - ) and will not be reviewed in its entirety here. however, it becomes clear that rough scaffold surfaces favour attachment, proliferation and differentiation of anchorage-dependent bone forming cells ( ) . osteogenic cells migrate to the scaffold surface through a fibrin clot initially established immediately after implantation of the tec from the haematoma caused by the surgical procedure ( ) . the migration causes retraction of the temporary fibrin matrix and, if not well secured, can lead to detachment of the fibrin from the scaffold during wound contraction leading to decreased migration of the osteogenic cells into the scaffold ( ) ( ) . with regards to surface chemistry, degradation properties and by-products (relating to ph, osmotic pressure, inflammatory reactions etc.) are of importance and have been briefly discussed already. in the following section, the role of calcium phosphate in the osteoinductivity of biomaterials will be summarized as an example of how surface chemistry may be manipulated to benefit scaffold properties. to date, most synthetic biomaterials that have been shown to be osteoinductive contained calcium phosphate underlining the crucial role of calcium and phosphate in osteoinduction properties of biomaterials ( ) . as summarised above, adequate porosity and pore size is crucial for bone tissue engineering scaffolds in order to allow sufficient vascularisation and enable a supply of body fluids throughout the tec. together with this nutrient supply, a release of calcium and phosphate ions from the biomaterial surface takes places and is believed to be the origin of bioactivity of calcium phosphate biomaterials ( ) ( ) ( ) . this process is followed by the precipitation of a biological carbonated apatite layer (that contains calcium-, phosphate-and other ions such as magnesium as well as proteins and other organic compounds) that occurs when the concentration of calcium and phosphate ions has reached super saturation level in the vicinity of the implant ( , ( ) ( ) . this bone-like biological carbonated apatite layer is thought to be physiological trigger for stem cells to differentiate down the osteogenic lineage or could induce the release of growth factors that complement this process ( ) . for biomaterials lacking calcium phosphate particles, the roughness of the surface is considered to act as a collection of nucleation sites for calcium phosphate precipitation from the hosts' body fluids, thereby forming a carbonated apatite layer. comparing calcium phosphate (cap) coated fibrous scaffolds (fibre diameter approx μm) made from medical grade polycaprolactone (mpcl) with non-coated mpcl-scaffolds, we have shown that cap-coating is beneficial for new bone formation in vitro, enhancing alkaline phosphatase activity and mineralisation within the scaffolds ( ) . interestingly, other research has shown that the implantation of highly soluble carbonated apatite ceramics alone did not result in bone induction in vivo ( ) , suggesting that a relatively stable surface (e.g. through a composite material that contains a less soluble phase) is needed for the facilitation of bone formation as discussed above (see "mechanical properties and degradation kinetics"). bone formation requires a stable biomaterial interface and therefore, too rapid in vivo dissolution of calcium phosphate materials has been shown to be unfavourable for the formation of new bone tissue ( ) ( ) . chai et al. and barradas et al. have recently reviewed the effects of calcium phosphate osteogenicity in bone tissue engineering ( , ) . further comprehensive reviews on the influence of surface topography and surface chemistry on cell attachment and proliferation for orthopaedic implants and bone tissue engineering are available ( , , , , ) . porosity is commonly defined as the percentage of void space in a so called cellular solid (the scaffold in bone tissue engineering applications) ( ) . using solid and porous particles of hydroxyapatite for the delivery of the growth factor bmp- , kuboki et al. showed that pores are crucial for bone tissue formation because they allow migration and proliferation of osteoblasts and mesenchymal cells, as well as vascularisation; no new bone formed on solid particles ( ) . a porous scaffold surface also improves mechanical interlocking between the implanted tecs and the surrounding natural bone tissue, providing greater mechanical stability at this crucial interface in tissue engineering ( ) . scaffold porosity and pore size relate to the surface area available for the adhesion and growth of cells both in vitro as well as in vivo and to the potential for host tissue ingrowth, including vasculature, to penetrate into the central regions of the scaffold architecture. in assessing the significance of porosity several in vivo studies have been conducted utilising hard scaffold materials such as calcium phosphate or titanium with defined porous characteristics ( ) . the majority of these studies indicate the importance of pore structure in facilitating bone growth. increase of porosity as well as pore size and spacing of pore interconnectivity has been found to positively influence bone formation in vivo, which is also correlated with scaffold surface area. pore interconnections smaller than μm were found to restrict vascular penetration and supplementation of a porous structure with macroscopic channels has been found to further enhance tissue penetration and bone formation ( , ) . interestingly, these results correlate well with the diameter of the physiological haversian systems in bone tissue that possess an approximate diameter of more than µm. the ability of new capillary blood vessels to grow into the tec is also related to the pore size, thereby directly influencing the rate of ingrowth of newly formed bone tissue into the tec: in vivo, larger pore sizes and higher porosity lead to a faster rate of neovascularisation, thereby promoting greater amounts of new bone formation via direct osteogenesis. in contrast, small pores favour hypoxic conditions and induce osteochondral formation before osteogenesis occurs ( ) . pores and pore interconnections should be at least microns in diameter to allow sufficient vascularisation. besides the actual macroporosity (pore size > µm) of the scaffold microporosity (pore size < µm) and pore wall roughness also have a large impact on osteogenic response: microporosity results in larger surface areas contributing to higher bone-inducing protein adsorption and to ion exchange and bone-like apatite formation by dissolution and re-precipitation ( , ) . as outlined above, sub-micron and nanometre surface roughness favours attachment, proliferation and differentiation of anchorage-dependent bone forming cells ( ) . although increased porosity and higher pore size facilitate bone ingrowth, it also compromises the structural integrity of the scaffold, and if the porosity becomes too high it may adversely affect the mechanical properties of the scaffold at the same time ( ) . in addition, the rate of degradation is influenced by the porosity and pore size (for biodegradable scaffolds). a higher pore surface area enhances interaction of the scaffold materials with host tissue and can thereby accelerate degradation by macrophages via oxidation and/or hydrolysis ( ) . therefore, scaffolds fabricated from biomaterials with a high degradation rate should not have high porosities (> %) in order to avoid compromise to the mechanical and structural integrity before adequate substitution by newly formed bone tissue. scaffolds made from slowly degrading biomaterials with robust mechanical properties can, in contrast, be highly porous ( ) . table illustrates mechanical properties and degradation kinetics in relation to the porosity for many commonly used composite scaffolds. this illustrates that there are a number of advantages and disadvantages associated with any changes made to the porosity or pore size of scaffolds. it is inevitable to find a balance between these pros and cons in order to tailor the scaffold properties ideally to the demands of the tissue engineering approach used. for comprehensive reviews on role of porosity and pore size in tissue engineering scaffolds, the reader is referred to two recently published reviews ( , ) . it becomes clear that a multitude of factors have to be taken into account when designing and fabricating scaffolds for bone tissue engineering. however, it is beyond the scope of this review to present all of them in detail and a number of comprehensive reviews have been published recently on this topic ( , , , , , ( ) ( ) . the three-dimensional design characteristics in combination with the material properties of a scaffold are crucial for bone tissue engineering purposes. not only does the scaffold structure need to be controlled on a macroscopic level (to achieve sufficient interposition of the scaffold into the defect site), but also on a microscopic level (to optimise tissue engineering properties with regards to osteoinduction, osteoconduction, osteogenesis and vascularisation as well as mechanical stability) and even down to nanostructural configuration (to optimise protein adsorption, cell adhesion, differentiation and proliferation related to desired tissue engineering characteristics of the tec). it is therefore necessary to exert strict control over the scaffold properties during the fabrication process. conventional techniques for scaffold fabrication include solvent casting and particulate leaching, gas foaming, fibre meshes and fibre bonding, phase separation, melt molding, emulsion freeze drying, solution casting and freeze drying ( ) . all of these techniques are subtractive in nature, meaning that parts of the fabricated scaffold are removed from the construct after the initial fabrication process in order to generate the desired three-dimensional characteristics. hence a number of limitations exist regarding these fabrication methods: conventional methods do not allow a precise control over pore size, pore geometry, pore interconnectivity or spatial distribution of pores and interconnecting channels of the scaffolds fabricated ( , ( ) ( ) . in addition, many of these techniques require the application of organic solvents and their residues can impose severe adverse effects on cells due to their potentially toxic and/or carcinogenic nature, reducing the biocompatibility of the scaffold significantly ( ) . the introduction of additive manufacturing (am) techniques into the field of bone tissue engineering has helped to overcome many of these restrictions ( , , ) . in am three-dimensional objects are created in a computer-controlled layer-by-layer fabrication process. in contrast to subtractive conventional methods of scaffold fabrication, this technique is additive in nature and does not involve removal of materials after the initial fabrication step. these techniques have also been named "rapid prototyping" or "solid free form fabrication" in the past, but in order to clearly distinguish them from conventional methods the latest astm standard now summarises all of these techniques under the term "additive manufacturing" ( ) . the basis for each am process is the design of a three-dimensional digital or in silico model of the scaffold to be produced. this computer model can either be created from scratch using "computer aided design" (cad) methods or can be generated using data from a d-scan of existing three-dimensional structures (such as the human skeleton) ( ) . the digital model is then converted into an stl-file that expresses the three-dimensional structure as the summary of multiple horizontal two-dimensional planes. using this stl-file an am-machine then creates the three-dimensional scaffold structure in a layer-bylayer fabrication method in which each layer is tightly connected to the previous layer to create a solid object. a number of different am techniques are currently applied using thermal, chemical, mechanical and/or optical processes to create the solid three-dimensional object ( ) . these methods include laser-based methods such as stereolithography (stl) and selective laser sintering (sls), printing-based applications (e.g. d-printing, wax-printing) and nozzle-based systems like melt extrusion/ fused deposition modeling (fdm) and bioplotting. the multitude of am techniques and their specifications were reviewed by several authors lately ( , , ( ) ( ) . am techniques have been used since the s in the telecommunication industry, in jewelry making and production of automobiles ( ) . from the s onwards, am was gradually introduced to the medical field as well ( ) : am was initially used to fabricate threedimensional models of bone pathologies in orthopaedic maxillofacial neurosurgical applications to plan surgical procedures and for haptic assessment during the surgery itself ( - ). with recent technical advances am is nowadays applied to make custom-made implants and surgical tools ( ) and to fabricate highly detailed, custom-made three-dimensional models for the indivi-dual patient (using data from ct, mri, spect etc.) to plan surgical approaches, specifically locate osteotomy sites, choose the correct implant and to predict functional and cosmetic outcomes of surgeries ( ) ( ) . thereby the operating time as well as the risk of complications has been reduced significantly. the application of am in bone tissue engineering represents a highly significant innovation that has drasticcally changes the way scaffolds are being fabricated; am has more or less become the new gold standard for scaffold manufacturing ( ) . the advantages of rapid prototyping processes include (but are not limited to) increased speed, customisation and efficiency. am technologies have relatively few process steps and involve little manual interaction, therefore, three-dimensional parts can be manufactured in hours and days instead of weeks and months. the direct nature of am allows the economical production of customized tissue engineering scaffolds. the products can be tailored to match the patient's needs and still sustain economic viability as compared to traditional techniques which must manufacture great numbers of devices. the conventional scaffold fabrication methods commonly limit the ability to form complex geometries and internal features. am methods reduce the design constraints and enable the fabrication of desired delicate features both inside and outside the scaffold. using stl, the am technique with the highest precision, for example objects at a scale of µm can be fabricated ( ) . a two-photon stl-technique to initiate the polymerisation can be used to pro-duce structures even at micrometer and sub-micrometer levels ( ) . am methods allow for variation of composition of two or more materials across the surface, interface, or bulk of the scaffold during the manufacturing. thereby, positional variations in physicochemical properties and surface characteristics can be created and utilized to promote locally specific tissue engineering signals. several am techniques operate without the use of toxic organic solvents. this is a significant benefit, since incomplete removal of solvents may lead to harmful residues that can affect adherence of cells, activity of incorporated biological agents or surrounding tissues as already described. am allows the control of scaffold porosity leading to the applications that may have areas of greater or lesser structural integrity and areas of encouraged blood flow due to increased porosity. fabricating devices and/or implants with differences in spatial distribution of porosities, pore sizes, mechanical and chemical properties can mimic the complex composition and architecture of natural bone tissue and thereby optimise bone tissue engineering techniques. in addition, scaffolds with gradients in porosity and pore sizes can be functionalised to allow vascularisation and direct osteogenesis in one area of the scaffold, while promoting osteochondral ossification in the other, which is an appealing approach to reproduce multiple tissues and tissue interfaces within one and the same biomaterial scaffold ( ) . table summarises the advantages of scaffolds designed and fabricated by am techniques. musculoskeletal conditions are highly prevalent and cause a large amount of pain, illness and disability to patients. these conditions are the second most common reason for consulting a general practitioner, accounting for almost % of the total cost of illness and up to % of primary care ( ) . in addition, the impact of musculoskeletal conditions is predicted to grow with the increasing incidence of lifestyle-related obesity, reduced physical fitness and increased road traffic accidents ( ) . the impact of bone trauma is significant-the consequences of failing to restore full function to an injured limb are dramatically demonstrated by the statistic that only % of patients suffering from severe open fractures of the tibia are able to resume full function and hence return to previous employment ( ) . along with trauma, tumour resection is another major cause of large bone defects. cancer is a major public health challenge, with one in four deaths in the united states currently due to this disease. recent statistics indicate that new cancer cases and deaths from cancer are projected to occur in the united states in ( ) . as outlined above, the number of procedures requiring bone implant material is increasing, and will continue to do so in our aging population and with deteriorating physical activity levels ( ) . the current bone grafting market already is estimated to be in excess of $ . billion each year and is expected to increase by - % per year ( ) . with the introduction of tissue engineering the hopes and expectations were extremely high to be able to substitute natural organs with similar (or even better) tissue engineered replacement organs. however, at the time it was stated that "few areas of technology will require more interdisciplinary research than tissue engineering" ( ) and this assessment holds true today. in the years to follow, numerous private and public institutes conducted scientific research and clinical translation efforts related to tissue engineering. at the beginning of , tissue engineering research and development was being pursued by scientists and support staff in more than start-up companies or business units with a combined annual expenditure of over $ million usd ( ) . the us national institutes of health (nih), accounting for the largest cumulative us federal research expenditures, has increased the funding in tissue engineering from . billion usd in the fiscal year to more than billion usd for the fiscal year ( ) . between and the number of papers published on tissue engineering and scaffolds per year increased by more than % and more than %, respectively ( ) . but despite the increasing research expenditure and the magnitude of discoveries and innovations in bone tissue engineering since its introduction more than three decades ago, the translation of these novel techniques into routine clinical applications on a large scale has still not taken place. as scott j. hollister has pointed out, there is, on the one hand, a stark contrast between the amount of tissue engineering research expenditures over the last years and the resulting numbers of products and sales figures. on the other hand, there is also a significant discrepancy between the complexities of intended tissue engineering therapies compared to the actual therapies that have reached clinical applications ( ) . this evident gap between research and clinical application/commercialisation is commonly termed the "valley of death" due to the large number of ventures that "die" between scientific technology development and actual commercialization due to lack of funds ( figure ) ( ) . the valley of death is particularly large for tissue engineering approaches because this field of research often utilises immensely cost intensive high-tech biotechnologies for technological development eating up large parts of the funding available, but then additionally faces the challenges of funding large scale preclinical studies and clinical studies to gain approval by regulatory bodies, demonstrate product safety and gain clinical acceptance ( ) ( ) ( ) . to bridge the gap between the bench and bedside, the scaffold is required to perform as a developmentally conducive extracellular niche, at a clinically relevant scale and in concordance with strict clinical (economic and manufacturing) prerequisites ( figure ) ( ) . in this context the scaffold facilitates for smaller and medium sized defects the entrapment of the hematoma and prevents it's "too early" contraction ( ) . for large and high-load bearing defects the scaffold can also deliver cells and/or growth factors to the site of damage and provides an appropriate template for new tissue formation. the scaffold should thus constitute a dynamically long-lasting yet degradable three-dimensional architecture, preferably serving as a functional tissue substitute which, over time, can be replaced by cell-derived tissue function. designing and manufacturing processes are believed to be the gatekeepers to translate tissue engineering research into clinical tissue engineering applications and concentration on the development of these entities will enable scaffolds to bridge the gap between research and clinical practice ( ) . one of the greatest difficulties in bridging the valley of death is to develop good manufacturing processes and scalable designs and to apply these in preclinical studies; for a description of the rationale and road map of how our multidisciplinary research team has addressed this first step to translate orthopaedic bone engineering from bench to bedside see below and refer to our recent publication ( ) . in order to take bone tissue engineering approaches from bench to bedside, it also imperative to meticulously assess the clinical demands for specific scaffold characteristics to achieve a broad and optimised range of clinical applications for the specific tissue engineering approach. a sophisticated bone tissue engineering technology will not necessarily have multiple clinical applications just because of its level of complexity, and defining specific clinical target applications remains one of the most underestimated challenges in the bridging the valley of death ( ) . there is often a great level of discrepancy between the clinical demands on a tissue engineering technique and the scientific realisation of such technique, hampering the clinical translation. thus a scaffold that is realistically targeted at bridging the valley of death should ( ): (i) meet fda approval (for further details on this topics see reviews by scott j. hollister and ) ( , ) ; (ii) allow for cost effective manufacturing processes; (iii) be sterilisable by industrial techniques; (iv) enable easy handling without extensive preparatory procedures in the operation theatre; (v) preferably, be radiographically distinguishable from newly formed tissue; and (vi) allow minimally invasive implantation ( ) ( ) . in targeting the translation of a (bone) tissue engineering approach from bench to bedside, there is a distinct hierarchy and sequence of the type of studies that need to be undertaken to promote the translation process ( ) : having identified clinical needs and based on fun- damental discoveries regarding biological mechanisms, a novel tissue engineering approach is designed and first studies are undertaken to characterise mechanical and chemical properties of the tec to be used. the next step involves feasibility and bioactivity testing and should be carried out in vitro and in vivo. in vitro assays using cell culture preparations are used to characterise the effects of materials on isolated cell function and for screening large numbers of compounds for biological activity, toxicity and immunogenicity ( ) ( ) . however, due to their nature using isolated cells, in vitro models are unavoidably limited in their capacity to reflect complex in vivo environments that the tec will be exposed to and are therefore inadequate to predict in vivo or clinical performances. therefore, in vivo models (that is animal models) are required in order to overcome the limitations of in vitro models to provide a reproducible approximation of the real life situation. in vivo feasibility testing is almost exclusively done in small animals, mainly in rodents and rabbits ( , ( ) ( ) ( ) . the advantages of small animal models include relatively easy standardisation of experimental conditions, fast bone turnover rates (=shorter periods of observation), similar lamellar bone architecture and similar cancellous bone thinning and fragility, similar remodelling rates and sites, common availability and relatively low costs for housing and maintenance. disadvantages of rodent and rabbit models include different skeletal loading patterns, open epiphyses at various growth plates up to the age of - months (or for lifetime in rats), minimal intra-cortical remodelling, the lack of harversian canal systems, a smaller proportion of cancellous bone to total bone mass and their relatively small size for testing of implants ( ) . whilst a large number of studies in rodents and rabbits have established proof of concept for bone tissue engineering strategies, scaling up to larger, more clinically relevant animal models has presented new challenges. quoting thomas a. einhorn, when conducting animal studies, one has to keep in mind that "in general, the best model system is the one which most closely mimics the clinical situation for which this technology is being developed, will not heal spontaneously unless the technology is used, and will not heal when another technology is used if that technology is less advanced than the one being tested" ( ) . the most effective animal models will therefore ) provide close resemblance of the clinical and biological environment and material properties, ) encompass highly standardised measurement methods providing objective parameters (qualitative and quantitative) to investigate the newly formed bone tissue and ) are able to detect and predict significant differences between the bone tissue engineering methods investigated ( ) . for clinical modelling and efficacy prediction of the tissue engineering strategy to be translated into clinical application, up-scaling to large animal models is therefore inevitable. thereby, the tissue engineering therapy can be delivered in the same (or similar) way in which it will be delivered in clinical settings utilising surgical techniques that match (or closely resemble) clinical methods at the site that matches the setting in which it will be used later as closely as possible ( ) . the advantage of large animal models (using nonhuman primates, dogs, cats, sheep, goats, pigs) is the closer resemblance of microarchitecture, bone physiology and biomechanical properties in humans. they encompass a well-developed haversian and trabecular bone remodelling, have greater skeletal surface to volume areas, show similar skeletal disuse atrophy, enable the use of implants and techniques similar to the ones used in humans and show highly localised bone fragility associated with stress shielding by implants. however, the use of large animal models has disadvantages as well, including the high cost and maintenance expenses, extensive housing and space requirements, relatively long life spans and lower bone turnover rates (making longer study periods necessary), difficulties in standardisation to generate large, homogenous samples for statistical testing as well as various ethical concerns depending on the species used (e.g. primates) ( ) . but despite several disadvantages, it is inevitable to perform the final pre-clinical in large animals, as realistically as possible, with relevant loading conditions and with similar surgical techniques as used in the final procedure in humans ( ) . large animal models provide mass and volume challenges for scaffold-based tissue engineering and require surgical fixation techniques that cannot be tested either in vitro or in small animal models ( ) . in general, preclinical translation testing is performed in large skeletally mature animals, the species most utilised are dog, sheep, goat and pig ( , ) . if sufficient preclinical evidence for the efficacy and safety of the new bone tissue engineering system has been generated utilising large animal models, clinical trials care undertaken to prove clinical significance and safety, ultimately leading to the translation of the technology into routine clinical practice. taking composite scaffold based bone tissue engineering from bench to bedside in accordance with the above outline rationale for translating bone tissue engineering research into clinical applications, during the last decade our interdisciplinary research team has focussed on the bench to bedside translation of a bone tissue engineering concept based on slowly biodegradable composite scaffolds made from medical grade polycaprolactone (mpcl) and calcium phosphates [hydroxyapatite (ha) and tricalcium phosphate (tcp)] ( , ) . detailed descriptions of the scaffold fabrication protocol can be found in our recent publications ( , , ( ) ( ) ( ) . the scaffolds have been shown in vitro to support cell attachment, migration and proliferation; degradation behaviour and tissue in-growth has also been extensively studied ( ) ( ) ( ) ( ) . we subsequently took the next step towards clinical translation by performing small animal studies using rat, mice and rabbit models ( ) ( ) ( ) . as reviewed in detail in reference ( ) , we were able to demonstrate the in vivo capability of our composite scaffolds in combination with growth factors or cells to promote bone regeneration within ectopic sites or critical sized cranial defects in the small animal models. studies in large animal models that closely resemble the clinical characteristics of human disease, with respect to defect size and mechanical loading, then became essential to advance the translation of this technology into the most difficult and challenging clinical applications in orthopaedic tumour and trauma surgery. the choice of a suitable large animal model depends on the ultimate clinical application, and consequently there is no such thing as "one gold standard animal model". over the last years, our research team has investigated the application of our composite scaffolds in several preclinical large animal models addressing different clinical applications: load-bearing, critical-sized ovine tibial defect model well-characterised, reproducible and clinically relevant animal models are essential to generate proof-ofprinciple pre-clinical data necessary to advance novel therapeutic strategies into clinical trial and practical application. our research group at the queensland university of technology (qut; brisbane, australia) has spent the last years developing a world-leading defect model to study pre-clinically different treatment options for cases of large volume segmental bone loss ( , ) . we have successfully established this cm critical-sized defect model in sheep tibiae to study the mpcl-tcp scaffold in combination with cells or growth factors including bone morphogenic proteins (bmps) ( ) ( ) . this model has not only generated a series of highly cited publications ( ) ( ) ( ) ( ) ( ) , but also has attracted large interest in the orthopaedic industry to be used as a preclinical test bed for their bone graft products under development. the model enables control of experimental conditions to allow for direct comparison of products against a library of benchmarks and gold standards we have developed over the last years (we have performed more than operations using this model todate). our preclinical tibial defect model developed at qut is one of the only available models internationally, which is suitable from both reproducibility and cost point of view for the evaluation of large segmental defect repair technologies in statistically powered study designs. we have chosen this critical sized segmental defect model of the tibia for our large animal model because tibial fractures represent the most common long bone fractures in humans and are often associated with significant loss of bone substance ( ) ( ) . also, tibial fractures result in high rates of non-unions or pseudarthroses ( , ) . from an orthopaedic surgeons point of view it can be argued that amongst all bone defects seen in the clinical practice, segmental defects of the tibia are often the most challenging graft sites. this owes to the grafts being required to bear loads close to physiological levels very soon after implantation, this is despite internal fixation, which often provides the necessary early stability, but also suffers from the poor soft tissue coverage (vascularisation issue) of the tibia compared to the femur. hence, in a bone engineering strategy for the treatment of segmental tibial defects, the scaffold must bear (or share) substantial loads immediately after implantation. the scaffold's mechanical properties (strength, modulus, toughness, and ductility) are determined both by the material properties of the bulk material and by its structure (macrostructure, microstructure, and nanostructure). matching the mechanical properties of a scaffold to the tibial graft environment is critically important so that progression of tissue healing is not limited by mechanical failure of the scaffold prior to successful tissue regeneration. similarly, because mechanical signals are important mediators of the differentiation of cell progenitors, a scaffold must create an appropriate stress environment throughout the site where new tissue is desired. hence, one of the greatest challenges in scaffold design for load bearing tibial defects is the control of the mechanical properties of the scaffold over time. by trialing our bone tissue engineering strategies in a tibial defect model, we will therefore address a highly relevant clinical problem and are creating valuable pre-clinical evidence for the translation from bench to bedside. with the cm critical defect being regenerated successfully by applying our mpcl-tcp scaffold in combination with bmp ( ), we are now investigating bone regeneration potentials in even larger sized tibial defects ( figure ). spinal fusion has been investigated in animal models for one hundred years now and a lot of the knowledge we have today on how spinal fusion progresses was gained through animal models ( ) ( ) . with regards to the above pictured rationale for translating bone tissue engineering approaches to clinical practice, it is of importance to note that the physical size of the sheep spine is adequate to allow spinal surgery to be carried out using the same implants and surgical approaches that are used in humans as well. also, sheep spines allow for an evaluation of the success of the study using fusion assessments commonly used in clinical practice. when considering spinal fusion in large animal models, it is apparent that due to the biomechanical properties of the spine a biped primate animal model [such as in ( ) ] should ideally preferred over a quadruped large animal model [for example ovine ( ) or porcine ( ) ]. but given the expenses and limited availability of primate testing as well as ethical concerns due to the close phylogenical relation, it is more feasible to trial large numbers of scaffold variations in the most appropriate quadruped large animal models and then evaluate the best performing scaffold in a primate model, if possible ( ) . we have outlined above that defining specific clinical target applications is a critical prerequisite for successful bone tissue engineering research that is meant to be translated into clinical practice. in accordance with this we have selected the thoracic spine for our animal model because we have identified idiopathic scoliosis as clinically highly relevant thoracic spine pathology. idiopathic scoliosis is a complex three-dimensional deformity affecting - % of the general population ( ) . scoliotic spine deformities include progressive coronal curvature, hypokyphosis or lordosis in the thoracic spine and vertebral rotation in the axial plane with posterior elements turned rotated toward the curve concavity. scoliotically deformed vertebral columns are prone to accelerated intervertebral disc degeneration, initiating more severe morphological changes of the affected vertebral joints and leading to chronic local, pseudoradicular, and radicular back pain ( ) . one of the critical aspects in surgical scoliosis deformity correction is bony fusion to achieve long-term stability ( ) . autologous bone grafting is still the gold standard to achieve spinal fusion and superior to other bone grafts for spinal fusion ( ) ( ) ( ) . nonetheless, the use of autologous bone grafting material has significant risks as outlined in detail above. a number of animal models for the use of tissue-engineered bone constructs in spinal fusion exists ( ) and the use of bone morphogenetic proteins for spinal fusion has been studied extensively ( , , ( ) ( ) . however, to the best of our knowledge, our ovine thoracic spine fusion model is the first existing preclinical large animal model on thoracic interverte-bral fusion allowing the assessment of tissue-engineering constructs such as biodegradable mpcl-cap scaffolds and recombinant human bone morphogenetic protein- (rhbmp ) as a bone graft substitute to promote bony fusion (figure ) ( ) . we have been able to show that radiological and histological results at -months post surgery indicated had comparable grades of fusion and evidenced new bone formation for the mpcl-cap scaffolds plus rhbmp- and autograft groups. the scaffold alone group, however, had lower grades of fusion in comparison to the other two groups. our results demonstrate the ability of this large animal model to trial various tissue engineering constructs against the current gold standard autograft treatment for spinal fusion in the same animal. in the future, we will be able to compare spinal fusion tissue engineering constructs in order to create statistically significant evidence for clinical translation of such techniques. the tibial diaphysis (c-d) and the periosteum is removed from the defect site and additionally also from cm of the adjacent bone proximally and distally. special care is taken not to damage the adjacent neurovascular bundle (e, bundle indicated by asterisk). the defect site is then stabilised using a hole dcp (synthes) (f). afterwards cm mpcl-tcp scaffold loaded with prp and rhbmp- is press fitted into the defect site to bridge the defect (g-h) and the plate is fixed in its final position. xray analysis at months after implantation (i) shows complete bridging of the defect site with newly formed radio-opaque mineralised tissue (in order to provide sufficient mechanical support, the scaffold is not fully degraded yet and scaffold struts appear as void inside the newly formed bone tissue). the interdisciplinary research group has evaluated and patented the parameters necessary to process medical grade polycaprolactone (mpcl) and mpcl composite scaffolds (containing hydroxyapatite or tricalciumphosphate) by fused deposition modeling ( ) . these "first generation scaffolds" have undergone more than years of studies in clinical settings and have gained federal drug administration (fda)-approval in and have also been successfully commercialised (www. osteoporeinternational.com). the scaffolds have been used highly successfully as burr whole plugs for cranioplasty ( ) and until today more than patients have received burr whole plugs, scaffolds for orbital floor reconstruction and other cranioplasties (figure ) ( ) . with their extensive, multidisciplinary approach the research team has achieved one of the rare examples of a highly successful bone tissue engineering approach bridging the gap between scientific research and clinical practice leading to significant innovations in clinical routines. as shown above, "second generation scaffolds" produced by fdm and based on composite materials have already been broadly studied in vitro plus in vivo in small animal models and are currently under preclinical evaluation in large animal studies conducted by our research group. available data so far clearly supports the view that further translation into clinical use will take place and that a broad spectrum of targeted clinical applications will exist for these novel techniques. our we herein propose that regenerative medicine . has commenced. we foresee that the complexity and great variety of large bone defects require an individualized, patient-specific approach with regards to surgical reconstruction in general and implant/tissue engineering selection in specific. we advocate that bone tissue engineering and bioengineering technology platforms, such as additive manufacturing approaches can be used even more substantially in bone grafting procedures to advance clinical approaches in general and for the benefit of individual patient in particular. the tremendous advantage of scaffolds made by additive manufacturing techniques such as fused deposition modeling (fdm) is the distinct control over the macroscopic and microscopic shape of the scaffold and thereby control over the shape of the entire tec in total. additive manufacturing enables the fabrication of highly structured scaffolds to optimise properties highly relevant in bone tissue engineering (osteoconductivity, osteoinductivity, osteogenicity, vascularisation, mechanical and chemical properties) on a micro-and nanometre scale. using high-resolution medical images of bone pathologies (acquired via ct, µct, mri, ultrasound, d digital photogrammy and other techniques) ( ), we are not only be able to fabricate patient-specific instrumentation ( ) ( ) ( ) , patient-specific conventional implants ( - ) or allografts ( ) , but also to realise custom-made tissue engineering constructs (tec) tailored specifically to the needs of each individual patient and the desired clinical application ( , , ) . we therefore predict that the commencing area of regenerative medicine . will hold a significant leap forward in terms of personalised medicine. we have already proven the clinical application of this concept by fabricating a custom-made bioactive mpcl-tcp implant via cad/fdm that was used clinically to successfully reconstruct a complex cranial defect ( ) . we have also recently provided a rationale for the use of cad/fdm and mpcl-tcp scaffolds in contributing to clinical therapy concepts after resection of musculoskeletal sarcoma (figures and ) ( ) . although it has to be mentioned that our approaches presented in this review are at different stages of clinical translation, their entity clearly represents a promising and highly significant century approach in taking bone tissue engineering strategies from bench to bedside and into the era of regenerative medicine . . in conclusion, the field of bone tissue engineering has significantly changed the millennia old quest by humans indicate osteotomy planes to achieve tumour free margins, after which, after which the cad model is virtually resected (e). a custom made scaffold to fit the defined defect is then created by mirroring the healthy side of the pelvis, adjusting the size of the scaffold accordingly and fabricating the scaffold from the virtual model using am techniques (f). flanges, intramedullary pegs and other details can be added to the porous scaffold structure to facilitate surgical fixation and to enhance its primary stability after implantation (g). images d-g reproduced with permission from ( ) , © the authors to optimise the treatment of bone defects and to identify suitable bone substitute materials. we have reviewed the historic development, current clinical therapy standards and their limitations as well as currently available bone substitute materials. we have also outlined current knowledge on scaffold properties required for primary stability and even load distribution is achieved by using an internal fixation device ( ) . secondary stability is achieved by osseointegration of both the fibula and the porous tissue engineering scaffold. over time, the scaffold is slowly replaced by ingrowing tissue engineered bone and the defect is completely bridged and regenerated ( ) . h partly reproduced with permission from ( ), © the authors. bone tissue engineering and the potential clinical applications as well as the difficulties in bridging the gap between research and clinical practice. although the clinical translation of these approaches has not taken place on a large scale yet, bone tissue engineering clearly holds the potential to overcome historic limitations and disadvantages associated with the use of the current gold-standard autologous bone graft. optimizing combinations of cells, scaffolds, and locally and systemically active stimuli will remain a complex process characterized by a highly interdependent set of variables with a large range of possible variations. consequently, these developments must also be nurtured and monitored by a combination of clinical experience, knowledge of basic biological principles, medical necessity, and commercial practicality. the responsibility for rational development is shared by the entire orthopaedic community (developers, vendors, and physicians). the need for objective and systematic assessment and reporting is made particularly urgent by the recent rapid addition of many new options for clinical use. by applying a complex interplay of st century technologies from various disciplines of scientific research, the gap between bone tissue engineering research and the translation into clinically available bone tissue engineering applications can successfully be bridged. flati g. chirurgia nella preistoria. parte i. provincia med aquila. regenerative medicine . nanostructured biomaterials for tissue engineering bone mechanical properties and the hierarchical structure of bone biomechanics of trabecular bone advanced biomaterials for skeletal tissue regeneration: instructive and smart functions form and function of bone comparison of microcomputed tomographic and microradiographic measurements of cortical bone porosity structure and development of the skeleton the skeleton as an endocrine organ bone as a ceramic composite material determinants of the mechanical properties of bones age-related bone changes an introduction to bioceramics the elastic modulus for bone the elastic moduli of human subchondral, trabecular, and cortical bone tissue and the size-dependency of cortical bone modulus elastic properties of human cortical and trabecular lamellar bone measured by nanoindentation effect of mechanical stability on fracture healing--an update biomechanics and tissue engineering mechanobiology of bone tissue mechanical conditions in the initial phase of bone healing the biology of fracture healing bone regeneration: current concepts and future directions enhancement of fracture healing risk factors contributing to fracture non-unions repairing holes in the head: a history of cranioplasty cranioplasty in prehistoric times traite experimental et clinique de la regeneration des os et de la production artificielle du tissu osseux histologische untersuchung über knochen implantationen the use of homogenous bone grafts; a preliminary report on the bone bank transplantation immunity in bone homografting immunologic aspects of bone transplantation the animal bone chip in the bone bank knochenregeneration mit knochenersatzmaterialien: eine tierexperimentelle studie. hefte unfallheilk adverse reactions and events related to musculoskeletal allografts: reviewed by the world health organisation project notify ueber die freie transplantation von knochenspongiosa. langenbecks arch clin chir experimental studies on bone transplantation with unchanged and denaturated bone substance. a contribution on causal osteogenesis bone graft substitutes: what are the options? bone grafts and bone graft substitutes in orthopaedic trauma surgery. a critical analysis autogenous bone graft: donor sites and techniques reamer-irrigator-aspirator bone graft and bi masquelet technique for segmental bone defect nonunions: a review of cases bone grafts: a radiologic, histologic, and biomechanical model comparing autografts, allografts, and free vascularized bone grafts long bone reconstruction with vascularized bone grafts limb-lengthening, skeletal reconstruction, and bone transport with the ilizarov method bone lengthening (distraction osteogenesis): a literature review ilizarov principles of deformity correction ilizarov bone transport treatment for tibial defects use of the ilizarov technique for treatment of non-union of the tibia associated with infection skeletal tissue regeneration: current approaches, challenges, and novel reconstructive strategies for an aging population effet biologique des membranes à corps etranger induites in situ sur la consolidation des greffes d'os spongieux masquelet technique for the treatment of bone defects: tips-tricks and future directions the concept of induced membrane for reconstruction of long bone defects induced membranes secrete growth factors including vascular and osteoinductive factors and could stimulate bone regeneration influences of induced membranes on heterotopic bone formation within an osteo-inductive complex treatment of large segmental bone defects with reamerirrigator-aspirator bone graft: technique and case series the influence of a one-step reamer-irrigator-aspirator technique on the intramedullary pressure in the pig femur bone graft harvest using a new intramedullary system a new minimally invasive technique for large volume bone graft harvest for treatment of fracture nonunions complications following autologous bone graft harvesting from the iliac crest and using the ria: a systematic review novel technique for medullary canal débridement in tibia and femur osteomyelitis osteogenic potential of reamer irrigator aspirator (ria) aspirate collected from patients undergoing hip arthroplasty reaming irrigator aspirator system: early experience of its multipurpose use reamer-irrigatoraspirator indications and clinical results: a systematic review using the bi-masquelet technique and reamer-irrigator-aspirator for post-traumatic foot reconstruction the use of bone substitutes in the treatment of bone defects-the clinical view and history current trends and future perspectives of bone substitute materials -from space holders to innovative biomaterials fracture healing: the diamond concept the diamond concept--open questions state of the art and future directions of scaffold-based bone engineering from a biomaterials perspective bone tissue engineering: from bench to bedside bone graft substitutes safety and efficacy of use of demineralised bone matrix in orthopaedic and trauma surgery demineralized bone matrix as an osteoinductive biomaterial and in vitro predictors of its biological potential the clinical use of allografts, demineralized bone matrices, synthetic bone graft substitutes and osteoinductive growth factors: a survey study a thorough physicochemical characterisation of calcium phosphate-based bone substitution materials in comparison to natural bone coralline hydroxyapatite bone graft substitute: a review of experimental studies and biomedical applications maxilla sinus grafting with marine algae derived bone forming material: a clinical report of long-term results three-dimensional glass-derived scaffolds for bone tissue engineering: current trends and forecasts for the future calcium salts bone regeneration scaffolds: a review article current application of beta-tricalcium phosphate composites in orthopaedics various preparation methods of highly porous hydroxyapatite/polymer nanoscale biocomposites for bone regeneration concepts of scaffold-based tissue engineering--the rationale to use solid free-form fabrication techniques biocomposites containing natural polymers and hydroxyapatite for bone tissue engineering calcium phosphate ceramic systems in growth factor and drug delivery for bone tissue engineering: a review the return of a forgotten polymer-polycaprolactone in the st century polymeric materials for bone and cartilage repair scaffolds in tissue engineering bone and cartilage bioactive composites for bone tissue engineering design of bone-integrating organic-inorganic composite suitable for bone repair current concepts of molecular aspects of bone healing bone tissue engineering: state of the art and future trends a tissue engineering solution for segmental defect regeneration in load-bearing long bones clinical application of human mesenchymal stromal cells for bone tissue engineering characterization, differentiation, and application in cell and gene therapy mesenchymal stem cells in musculoskeletal tissue engineering: a review of recent advances in clonal characterization of bone marrow derived stem cells and their application for bone regeneration bone marrow stromal cells (bone marrow-derived multipotent mesenchymal stromal cells) for bone tissue engineering: basic science to clinical translation current insights on the regenerative potential of the periosteum: molecular, cellular, and endogenous engineering approaches periosteal cells in bone tissue engineering osteoblasts in bone tissue engineering stem cells from umbilical cord and placenta for musculoskeletal tissue engineering adipose-derived stem cells for tissue repair and regeneration: ten years of research and a literature review bone tissue engineering: recent advances and challenges bone tissue engineering: current strategies and techniques--part ii: cell types cell sources for bone tissue engineering: insights from basic science assessing the value of autologous and allogeneic cells for regenerative medicine orthopaedic applications of bone graft & graft substitutes: a review mineralization processes in demineralized bone matrix grafts in human maxillary sinus floor elevations biological matrices and tissue reconstruction foreign-body reactions to fracture fixation implants of biodegradable synthetic polymers foreign body reactions to resorbable poly(l-lactide) bone plates and screws used for the fixation of unstable zygomatic fractures late degradation tissue response to poly(l-lactide) bone plates and screws inflammatory cell response to calcium phosphate biomaterial particles: an overview bone regeneration materials for the mandibular and craniofacial complex bone scaffolds: the role of mechanical stability and instrumentation relative influence of surface topography and surface chemistry on cell response to bone implant materials. part : physico-chemical effects role of material surfaces in regulating bone and cartilage cell response formation and function of bone specific proteins mediate enhanced osteoblast adhesion on nanophase ceramics mechanisms of enhanced osteoblast adhesion on nanophase alumina involve vitronectin osteoblast adhesion on nanophase ceramics nano rough micron patterned titanium for directing osteoblast morphology and adhesion the influence of surface microroughness and hydrophilicity of titanium on the up-regulation of tgfbeta/bmp signalling in osteoblasts the effects of microporosity on osteoinduction of calcium phosphate bone graft substitute biomaterials advances in bionanomaterials for bone tissue engineering development of nanomaterials for bone repair and regeneration tissue engineering -nanomaterials in the musculoskeletal system perspectives on the role of nanotechnology in bone tissue engineering a preliminary study on osteoinduction of two kinds of calcium phosphate ceramics in vitro modeling of the bone/implant interface osteoinduction, osteoconduction and osseointegration osteoinductive biomaterials--properties and relevance in bone repair bonding of bone to apatite-coated implants biological profile of calcium phosphate coatings early bone formation around calcium-ion-implanted titanium inserted into rat tibia basse-cathalinat b. evolution of the local calcium content around irradiated beta-tricalcium phosphate ceramic implants: in vivo study in the rabbit calcium phosphate-based osteoinductive materials effect of culture conditions and calcium phosphate coating on ectopic bone formation osteoinduction by biomaterials--physicochemical and structural influences osteoinductive biomaterials: current knowledge of properties, experimental models and biological mechanisms bone formation induced by calcium phosphate ceramics in soft tissue of dogs: a comparative study between porous alpha-tcp and beta-tcp current views on calcium phosphate osteogenicity and the translation into effective bone regeneration strategies relative influence of surface topography and surface chemistry on cell response to bone implant materials. part : biological aspects new perspectives in mercury porosimetry bmp-induced osteogenesis on the surface of hydroxyapatite with geometrically feasible and nonfeasible structures: topology of osteogenesis porosity of d biomaterial scaffolds and osteogenesis potential of ceramic materials as permanently implantable skeletal prostheses three-dimensional scaffolds for tissue engineering: role of porosity and pore size bone tissue engineering: current strategies and techniques--part i: scaffolds recent advances in bone tissue engineering scaffolds a review of rapid prototyping techniques for tissue engineering purposes scaffold-based tissue engineering: rationale for computer-aided design and solid free-form fabrication systems rapid prototyping in tissue engineering: challenges and potential making tissue engineering scaffolds work. review: the application of solid freeform fabrication technology to the production of tissue engineering scaffolds additive manufacturing of tissues and organs astm standard f - : standard terminology for additive manufacturing technologies image-guided tissue engineering porous scaffold design for tissue engineering printing and prototyping of tissues and scaffolds additive processing of polymers a review of rapid prototyping (rp) techniques in the medical and biomedical sector rapid prototyping for orthopaedic surgery prototyping for surgical and prosthetic treatment speedy skeletal prototype production to help diagnosis in orthopaedic and trauma surgery. methodology and examples of clinical applications clinical applications of physical d models derived from mdct data and created by rapid prototyping a review on stereolithography and its applications in biomedical engineering submicron stereolithography for the production of freely movable mechanisms by using single-photon polymerization development of the australian core competencies in musculoskeletal basic and clinical science project -phase cancer statistics the growth of tissue engineering support for tissue engineering and regenerative medicine by the national institutes of health scaffold engineering: a bridge to where? establishment of a preclinical ovine model for tibial segmental bone defect repair by applying bone tissue engineering strategies mapping the translational science policy 'valley of death' bridging the regeneration gap: stem cells, biomaterials and clinical translation in bone tissue engineering emerging rules for inducing organ regeneration scaffold translation: barriers between concept and clinic tissue engineering of bone: the reconstructive surgeon's point of view engineering bone: challenges and obstacles the design and use of animal models for translational research in bone tissue engineering and regenerative medicine high-throughput screening for analysis of in vitro toxicity high-throughput screening assays to discover smallmolecule inhibitors of protein interactions long bone defect models for tissue engineering applications: criteria for choice animal models for bone tissue engineering skeletal tissue engineeringfrom in vitro studies to large animal models clinically applied models of bone regeneration in tissue engineering research selection and development of preclinical models in fracture-healing research the return of a forgotten polymer: polycaprolactone in the st century fused deposition modeling of novel scaffold architectures for tissue engineering applications the stimulation of healing within a rat calvarial defect by mpcl-tcp/collagen scaffolds loaded with rhbmp- comparison of the degradation of polycaprolactone and polycaprolactone-(βtricalcium phosphate) scaffolds in alkaline medium evaluation of polycaprolactone scaffold degradation for months in vitro and in vivo dynamics of in vitro polymer degradation of polycaprolactone-based scaffolds: accelerated versus simulated physiological conditions composite electrospun scaffolds for engineering tubular bone grafts biomimetic tubular nanofiber mesh and platelet rich plasma-mediated delivery of bmp- for large bone defect regeneration differences between in vitro viability and differentiation and in vivo bone-forming efficacy of human mesenchymal stem cells cultured on pcl-tcp scaffolds combined marrow stromal cell-sheet techniques and high-strength biodegradable composite scaffolds for engineered functional bone grafts treatment of long bone defects and non-unions: from research to clinical practice the challenge of establishing preclinical models for segmental bone defect research custom-made composite scaffolds for segmental defect repair in long bones a tissue engineering solution for segmental defect regeneration in long bones establishment of a preclinical ovine model for tibial segmental bone defect repair by applying bone tissue engineering strategies autologous vs. allogenic mesenchymal progenitor cells for the reconstruction of critical sized segmental tibial bone defects in aged sheep prevalence of long-bone non-unions epidemiology of adult fractures: a review path analysis of factors for delayed healing and nonunion in operatively treated tibial shaft fractures animal models for preclinical assessment of bone morphogenetic proteins in the spine animal models for spinal fusion the use of recombinant human bone morphogenetic protein (rhbmp- ) to promote spinal fusion in a nonhuman primate anterior interbody fusion model. spine (phila pa osteogenic protein versus autologous interbody arthrodesis in the sheep thoracic spine. a comparative endoscopic study using the bagby and kuslich interbody fusion device. spine (phila pa lateral surgical approach to lumbar intervertebral discs in an ovine model scoliosis and its pathophysiology: do we understand it? spine (phila pa back pain and function years after brace treatment for adolescent idiopathic scoliosis: a casecontrol study-part i. spine (phila pa ) challenges to bone formation in spinal fusion bone graft substitutes for spinal fusion an update on bone substitutes for spinal fusion bone graft substitutes in spinal surgery spinal fusion surgery: animal models for tissue-engineered bone constructs bone morphogenetic protein in spinal fusion: overview and clinical update bone morphogenetic proteins for spinal fusion establishment and characterization of an open minithoracotomy surgical approach to an ovine thoracic spine fusion model cranioplasty after trephination using a novel biodegradable burr hole cover: technical case report patient-specific instrumentation for total knee arthroplasty: a review a review of rapid prototyped surgical guides for patient-specific total knee replacement new technologies in planning and performance of osteotonics: example cases in hand surgery computer-assisted prefabrication of individual craniofacial implants cad/cam dental systems in implant dentistry: update uncemented computer-assisted designcomputer-assisted manufacture femoral components in revision total hip replacement: a minimum follow-up of ten years uncemented custom computer-assisted design and manufacture of hydroxyapatite-coated femoral components: survival at to years ct lesion model-based structural allografts: custom fabrication and clinical experience rapid prototyping for biomedical engineering: current capabilities and challenges calvarial reconstruction by customized bioactive implant can bone tissue engineering contribute to therapy concepts after resection of musculoskeletal sarcoma? sarcoma key: cord- - vkhptas authors: wu, tong; perrings, charles title: the live poultry trade and the spread of highly pathogenic avian influenza: regional differences between europe, west africa, and southeast asia date: - - journal: plos one doi: . /journal.pone. sha: doc_id: cord_uid: vkhptas in the past two decades, avian influenzas have posed an increasing international threat to human and livestock health. in particular, highly pathogenic avian influenza h n has spread across asia, africa, and europe, leading to the deaths of millions of poultry and hundreds of people. the two main means of international spread are through migratory birds and the live poultry trade. we focus on the role played by the live poultry trade in the spread of h n across three regions widely infected by the disease, which also correspond to three major trade blocs: the european union (eu), the economic community of west african states (ecowas), and the association of southeast asian nations (asean). across all three regions, we found per-capita gdp (a proxy for modernization, general biosecurity, and value-at-risk) to be risk reducing. a more specific biosecurity measure–general surveillance–was also found to be mitigating at the all-regions level. however, there were important inter-regional differences. for the eu and asean, intra-bloc live poultry imports were risk reducing while extra-bloc imports were risk increasing; for ecowas the reverse was true. this is likely due to the fact that while the eu and asean have long-standing biosecurity standards and stringent enforcement (pursuant to the world trade organization’s agreement on the application of sanitary and phytosanitary measures), ecowas suffered from a lack of uniform standards and lax enforcement. highly pathogenic avian influenzas have become a major threat to human and livestock health in the last two decades. the h n panzootic ( ongoing) has been one the most geographically widespread and costly, resulting in the loss of hundreds of millions of poultry in countries [ ] and over human deaths worldwide-a mortality rate of percent [ , ] . for h n , and other h subtypes, most countries reporting poultry outbreaks also report evidence of the disease in wild bird populations, and the mechanisms for the spread of h n have been identified as a combination of wild bird transmission and the live poultry trade [ , ] . plos risk factors. our primary interest is in the role of live poultry imports as a source of traderelated avian influenza risk at the regional level. we note that other poultry products, such as packaged meat and eggs, do pose a risk, but it is significantly lower. although avian influenza can persist in frozen meat, contact with that meat is unlikely to cause infection [ ] . furthermore, since hpais are lethal to egg embryos, eggs are not a potential source of transmission [ ] . the data comprise an unbalanced panel covering countries over years; the lack of balance is due to the fact that membership of the eu changed over the timeframe. the response variable in all models estimated was a log transformation of the number of h n poultry outbreaks in a given country in a given year, obtained from the emergency prevention system for animal health (empres), a joint project of the fao and oie [ ] . the log transformation was applied to account for the wide disparities in the numbers of the outbreaks across countries. in , for example, indonesia recorded outbreaks while romania, the only eu country to be infected that year, had only . in addition to reflecting the differing directions and intensities of risk factors, this also reflects differences in reporting conventions for h n at the international level [ ] . a series of outbreaks may be reported separately in one country, but be treated as a single event in another. data on trade in live poultry were obtained from the united nations' comtrade database (comtrade.un.org) and resourcetrade.earth, a project of the royal institute of international affairs (www.chathamhouse.org). these report the total imports of live poultry into a given country in a given year by weight (kg). the data on trade in live poultry did not distinguish between different types of domestic birds, such as chickens, duck, and geese, but grouped them under a single commodity category of "live poultry." with respect to wild bird migration as a pathway for h n spread, we used the density of wild bird habitat as a proxy for the presence and scale of migratory bird populations, and the likelihood that wild and domestic birds will mix. lakes, wetlands, and (irrigated) agricultural areas have been consistently identified as wintering and breeding grounds for migratory birds, and as places where wild birds may come into contact with free-ranging poultry [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] . the indicator for wild bird habitat used in this study was the set of "important bird and biodiversity areas" (ibas) for "migratory and congregatory waterbirds" identified by birdlife the live poultry trade poses different avian influenza risks in different regions of the world table . the distribution of h n poultry outbreaks between - across the member states of asean, ecowas, and eu. "-" signifies that the country was not a member of its associated trade bloc in that given year . brunei cambodia indonesia laos international (datazone.birdlife.org). in their analysis of h n spread, kilpatrick, chmura ( ) also identified ibas as a proxy for migratory birds and the infection risks they pose. country-level statistics on socioeconomic and agro-ecological conditions were taken from the united nations' food and agriculture organization (www.fao.org/faostat/en/) and the world bank (data.worldbank.org). agricultural land cover was reported as a percentage of total land area of the country. per-capita gdp was reported in purchasing power parity terms as current international dollars. data for for these two variables were missing for certain countries. in these cases, the gaps were filled by extrapolating the missing data as a linear trend of the preceding years. we assume that agricultural land-where free-ranging chickens, ducks, and geese are commonly raised in all three regions-also acts as a relevant proxy for susceptible poultry. data on the biosecurity measures targeting avian influenza undertaken by each country were obtained from the world organisation for animal health (oie) (www.oie.int). these report a standardized series of biosecurity controls targeting wildlife and livestock diseases, including those related to surveillance, vaccination, border checks, and management of wild disease reservoirs, and whether or not a given country undertook them in a given year. we chose a subset of these biosecurity measures we considered most relevant to h n avian influenza risks for inclusion in our model. additionally, in any given year, there were to countries that did not provide a report of biosecurity measures to the oie; we assumed that this indicates an absence of action, and the dataset records these cases as zeroes. our modeling approach relied on generalized linear models (glm) to analyze a panel of data on disease outbreaks and associated risk factors. in this we follow others who have sought to predict the spread of h n at both national and international levels [ ] [ ] [ ] or h n [ ] [ ] [ ] . glms are well suited to epidemiological studies because of their flexibility regarding data type and the distribution of response variables, their simplicity of application, and their frequency of use [ ] . our identification strategy involved the selection of three specifications for each of two estimators. we adopted both random and fixed effects estimators. hausman tests conducted at the all-regions level favored a random effects estimator, as the p-value exceeded the % threshold below which fixed-effects regression is conventionally considered necessary. some factors that influence the likelihood and number of outbreaks in a given country or region are not likely to change significantly over the course of several years, or even a decade. in our dataset, netherlands for example, the amount of land covered by wild bird habitat is time-variant, while agricultural land and even per-capita gdp for many countries experienced relatively modest variations over the timeframe of the study. in this case, and as the hausman diagnostics indicate, a random effects estimator is more appropriate. nevertheless, since we wished to control for timeinvariant characteristics of regions and countries we also implemented fixed effects estimators at both the aggregate and trading bloc levels, implicitly assuming no changes in the trade or biosecurity environment at the bloc level that we are unable to control for. our first specification (model ) included a number of factors related to disease risk but excluded both live poultry imports and biosecurity measures. included predictors were land area, human population, per-capita gdp in purchasing power terms, agricultural area, wild bird habitat area, and the live chicken population. our second specification (model ) added intra-regional trade bloc and extra-bloc imports of live poultry. our third specification (model ) added four main biosecurity measures: border precautions, general surveillance, vaccination prohibition, and wild disease reservoir management. all are categories of oie-reported biosecurity measures taken against avian influenza. the general forms of the estimated random and fixed effects models were: where y it denotes the number of poultry outbreaks in country i in year t, x includes the predictors for model , z includes the additional predictors for model , u includes the additional predictors for model , ecowas and asean are dummy variables for the two titular regional trade blocs (the eu is the reference group), and u it and ε it are the "between" and "within" errors respectively. to account for heteroskedasticity, we used robust standard errors. finally, since the data used in this analysis are reported annually, and h n has been a conspicuous and fast-moving epidemic (meaning the effects of an outbreak are unlikely to persist over a long period of time) among poultry, we did not use a lag structure in our statistical analysis. therefore, we assumed that the factors driving an outbreak in a given year are contemporaneous with it (e.g., an outbreak that occurred in were modelled using trade volumes from ). we were also constrained by data availability in our use of annual increments: although monthly data exist for outbreaks, they do not for important predictor variables such as percapita gdp, human and poultry populations, the volume of live poultry traded, and biosecurity. regressions results from all models, including both random and fixed effects, are reported in tables - . at the all-regions level, the results for the random-and fixed-effects models were very similar, with the same set of predictor variables being statistically significant (i.e., p-values below the % or %) and the same direction of impact on the response variable. this set of predictors was human population (positive direction), per-capita gdp (negative direction), intra-trade bloc live poultry imports (negative direction), extra-trade bloc live poultry imports (positive direction), and the biosecurity measure of surveillance (negative direction). additionally, although the coefficient values for the same predictor differed between the two estimators, all pairs were within the same order of magnitude. the only exception to this was migratory waterbird habitat variable-the percent of land area covered by ibas for migratory and congregatory waterbirds. this was statistically significant and negative (i.e., had a mitigating impact on h n poultry outbreaks) for the fixed-effects model but was not significant for the random-effects model. the overall r-squared for the random-effects model was significantly higher than that for the fixed-effects model ( . vs. . ). the "between rsquared" value was particularly high ( . ) in the random effects model, signaling the importance of variation among countries (as opposed to "within r-squared," which measures the variation within countries over time). as we had expected, we found significant differences across trade regions. in the randomeffects model, ecowas diverged from all-regions conditions and from asean with respect to per-capita gdp and extra-bloc imports: while the two predictors were, respectively, riskdecreasing and risk-increasing at the all-regions level and in asean, they had the opposite impacts in ecowas. furthermore, ecowas differed from the all-regions level and from the eu in terms of intra-bloc imports: while this was risk-decreasing for the former two, it was risk-increasing for ecowas. finally, there were predictors that were statistically insignificant at the all-regions level but had a significant effect within different regions. for asean, agricultural land cover was a mitigating factor for outbreaks while wild disease reservoir management showed a strong positive relation with outbreaks. for ecowas, wild waterbird habitats and border precautions had a mitigating effect on outbreaks while vaccination prohibition and wild reservoir management had a positive effect. in the eu, the population of live chickens had a strong negative relation table . results from the regression models of h n outbreak risk factors for member states in all three regions; regressor coefficients are reported and statistically-significant factors are marked by asterisks. a blank space signifies that the variable was not included in the given model. units model model model with outbreaks, while vaccination prohibition, similar to the case with ecowas, was positively related. following liang, xu ( ), there is a perception that the long distance transmission of highly pathogenic avian influenza h n was largely due to wild bird migration, with the live poultry trade playing a minor and more localized role in some cases. our concern here has been to identify the nature of the risk posed by the live poultry trade in different regions of the world, and the conditions affecting that risk. our measure of development status, per-capita gdp, is simultaneously a proxy for modernization, biosecurity, consumption, and value-at-risk. as a proxy for modernization, it reflects risk-reducing differences in production methods. industrial livestock production methods typically include on-farm biosecurity measures that protect poultry from contact with disease-carrying wild birds. unlike traditional methods of free-range or "backyard" husbandry, factory production minimizes the likelihood of poultry intermingling with wild birds or being exposed to environmental pathogen pollution. for all its epidemiological, ecological, and ethical problems, industrial livestock production allows for more timely and widespread disease surveillance and vaccination, and for greater compliance with animal health regulations [ ] . at the same time, per-capita gdp growth is also associated with risk-increasing changes in meat consumption, and hence poultry production. indeed, the highest income elasticity of demand for meat and fish has been found in the poorest households and the poorest countries [ ] . in developing countries, % of the additions to meat consumption are from pork and poultry, with poultry dominating pork [ ] . absent changes in on-farm biosecurity, increased table . results from the regression models of h n poultry outbreak risk factors for the association of southeast asian nations (asean); regressor coefficients are reported and statistically-significant factors are marked by. a blank space signifies that the variable was not included in the given model. units model model model production implies increased risk. across all regions, the net effect of income growth is to reduce risk, dominating risk-increasing changes. in the ecowas region-the lowest income region-the effect is the opposite. the risk-increasing effects of income growth dominate the risk reducing effects (table ) . amongst the landscape variables-land area, the proportion in agriculture, and the proportion in ibas-our results reveal no uniform relation to h n outbreaks. at the all-regions level we found a weakly negative relation between outbreaks and the proportion of the land area in ibas (table ). this was driven by the european union, which includes the highest proportion of land area in ibas, but also the most industrialized forms of poultry production. the degree to which poultry production is industrialized also shows up in the coefficients on poultry numbers, which are negative and significant only for the eu (table ). while spatial heterogeneity at the landscape scale is important in terms of avian ecology, we were unable to take explicit account of these more detailed considerations in a country-scale analysis. the impacts of regional differences in biophysical conditions that are not directly controlled for are, however, included in bloc-level fixed effects. our primary concern is with the role of the live poultry trade, and how that differs between regions. across all regions we find that live poultry imports into a trade bloc are risk increasing. this is consistent with past studies that have shown that extra-bloc live poultry imports may be a significant source of additional avian influenza risk where they do not meet bloc sanitary and phytosanitary standards. the eu's common market and the asean free trade regime in particular have long-standing and standardized protocols, in accordance with the world trade organization's agreement on the application of sanitary and phytosanitary measures. but the two blocs have quite different exposures to external risk. a study of highly pathogenic avian influenza introductions to vietnam, for example, found that extra-asean imports of table . results from the regression models of h n poultry outbreak risk factors for the economic community of west african states (ecowas); regressor coefficients are reported and statistically-significant factors are marked by asterisks. a blank space signifies that the variable was not included in the given model. units model model model population # people . x - . x - �� . x - . x - �� . live poultry increased the risk of introduction [ ] . this is also what our study finds for the asean region (table ) . we do not see an equivalent effect for the eu (table ), reflecting differences in both import volumes and the biosecurity measures applied to imports. the eu imports less and applies stricter biosecurity measures to those imports. the ecowas story is different. extra-bloc live poultry imports are risk reducing, not risk increasing (table ). it is likely that imports from outside the bloc reduce avian influenza risk in the region in part because they meet biosecurity standards that are more stringent than the standards applied in the region. the effects of intra-bloc trade in live poultry mirror the effects of extra-bloc trade. in the eu and asean, intra-bloc trade is risk reducing (tables and ). this may reflect a "substitution effect" in which imports of safer intra-bloc poultry crowds out riskier extra-bloc imports. other studies have come to similar conclusions. eu-derived live poultry imports to spain, for example, were found to pose no threat of avian influenza introduction [ ] . once again, eco-was is the exception. extra-ecowas imports of live poultry are risk reducing while intrabloc imports are risk increasing (table ). this is likely due to poor internal biosecurity, such as lax standards and inconsistent execution of inspections. regulatory standards within the ecowas trade bloc have been weak for the whole of the study period [ ] . while harmonized sanitary and phytosanitary standards for the member states of ecowas were in principle adopted in , most ecowas states had yet to submit legislation for international certification by [ ] . failure to adopt and enforce unified standards may be partly due to income constraints in ecowas countries. in ppp terms, the bloc's per-capita gdp in was less than half that of asean and approximately / th that of the eu, meaning it had less resources available for biosecurity policies and institutions. political instability may be another important obstacle: a table . results from the regression models of h n poultry outbreak risk factors for the european union (eu); regressor coefficients are reported and statistically-significant factors are marked by asterisks. a blank space signifies that the variable was not included in the given model. units model model model the live poultry trade poses different avian influenza risks in different regions of the world number of ecowas member states, including nigeria, niger, sierra leone, mali, liberia, and cote d'ivoire have suffered from civil wars and armed insurgencies over the past two decades. such fraught geopolitical conditions are not conducive to the establishment and enforcement of cross-border regulations. it goes without saying, though, that certification of sanitary and phytosanitary legislation in ecowas states, and the establishment of enforcement agencies to bring states into compliance with the sps agreement and codex alimentarius is a necessary condition of improving regional trade-related biosecurity. in terms of biosecurity measures more specifically, we did not have direct measures of onfarm biosecurity (but conjecture that biosecurity is increasing in per-capita gdp), but we did have measures of four biosecurity policies at the national level. these include: ( ) border precautions (measures applied at airports, ports, railway stations or road check-points open to international movement of animal, animal products and other related commodities, where import inspections are performed to prevent the introduction of the disease, infection or infestation); ( ) general surveillance (surveillance not targeted at a specific disease, infection or infestation); ( ) prohibition of vaccination (prohibition of the use of a vaccine to control or prevent the infection or infestation); and ( ) management of wildlife reservoirs (measures to reduce the potential for wildlife to transmit the disease to domestic animals and human beings). the management of wild disease reservoirs differs widely across countries, but techniques include vaccination, treatment of infections with drugs, isolation of infected populations, population translocation, reproduction reduction, culling, and control (draining, flooding, or burning) of wild disease reservoir habitat [ ] . of these measures, only general surveillance was significant at the all-regions level, while at the bloc level the effects of the different measures were frequently ambiguous. in the eu, for example, only the prohibition of vaccination was significant, and then in positive relation to outbreaks. for poultry, vaccination may be prohibited because the practice makes it difficult to distinguish infected from vaccinated flocks. this makes it a concomitant of policies centered on livestock culling as the primary response to outbreak risk [ ] . no other biosecurity policy was found to have a statistically significant relation to outbreaks in the region. the same set of policies had opposite effects in asean and ecowas. the prohibition of vaccination and the management of wild reservoirs were positively related to outbreaks in ecowas but negatively related to outbreaks in asean, while border protection measures were negatively related to outbreaks in ecowas but positively related to outbreaks in asean. this may reflect regional disparities in the quality of implementation not captured in the data. but it may also reflect the greater importance of trade in the transmission of the disease in ecowas. in their survey of the international spread of h n in the early years of the global epidemic, kilpatrick, chmura ( ) found that transmission into europe was by wild birds, that transmission into southeast asia was by the poultry trade, and transmission into africa by a balance of both. our results suggest that after introduction, inter-country spread had differing dynamics in each region. while intra-bloc trade facilitated h n spread among west african countries, it did not in either europe or southeast asia. in these areas, greater risk was posed by out-ofregion live poultry imports. in recent decades, avian influenzas have emerged as a major threat to human and animal health across the world. in particular, hpai h n , which was first isolated in , has been the most widespread and among the most devastating in terms of livestock and human mortality. it has inflicted severe losses to poultry stocks and caused hundreds of human deaths. even today, as other avian influenzas have become epidemic, h n remains in circulation among wildlife and livestock. identifying and quantifying the mechanisms of its international spread can help lay the groundwork for prediction and mitigation. it may also provide an instructive framework for the management of other avian influenzas. in this study, we considered the risk posed by the international trade in live poultry and the effects of associated biosecurity measures. differing agro-ecological and socioeconomic conditions across the trade regions were shown to influence epidemic dynamics in different ways, with certain factors being risk-enhancing or risk-decreasing in one region but having the opposite effect, or no significant effect, in another. in policy terms, there is no one-size-fits-all solution to mitigating avian influenza spread. the particular conditions, including those related to the trade agreements and associated regulatory standards, of a given region need to be carefully considered. but overall, biosecurity measures are potentially effective at controlling h n risks, and should be undertaken as a means to forestall spread-in general, mitigation of epidemics is significantly more cost-efficient than suppression [ ] . on-farm and other forms of domestic biosecurity may be more important than trade-related measures, but where the protection of trade pathways is weak, the risk of avian influenza spread is clearly higher. supporting information s file. detailed information on data sources. the public sources of the data used in this study, and how they were acquired, are described. (docx) world organization for animal health. oie situation report for highly pathogenic avian influenza: update: / world health organization. cumulative number of confirmed human cases for avian influenza a(h n ) reported to who food and agriculture organization. h n situation update predicting the global spread of h n avian influenza combining spatial-temporal and phylogenetic analysis approaches for improved understanding on global h n transmission exotic effects of capital accumulation options for managing the infectious animal and plant disease risks of international trade forecasting biological invasions with increasing international trade risk of importing zoonotic diseases through wildlife trade, united states globalization and livestock biosecurity global traffic and disease vector dispersal influences on the transport and establishment of exotic bird species: an analysis of the parrots (psittaciformes) of the world economic factors affecting vulnerability to biological invasions. the economics of biological invasions structural change in the international horticultural industry: some implications for plant health the emergence and evolution of swine viral diseases: to what extent have husbandry systems and global trade contributed to their distribution and diversity? animal movements and the spread of infectious diseases wildlife trade and global disease emergence. emerging infectious diseases ecology of zoonoses: natural and unnatural histories bats are natural reservoirs of sars-like coronaviruses global perspective for foot and mouth disease control a hotspot of non-native marine fishes: evidence for the aquarium trade as an invasion pathway reducing the risks of the wildlife trade the worldwide airline network and the dispersal of exotic species land-use and socio-economic correlates of plant invasions in european and north african countries global transport networks and infectious disease spread epidemiologic clues to sars origin in china. emerging infectious diseases influenza a h n immigration is filtered out at some international borders a statistical phylogeography of influenza a h n . proceedings of the national academy of transboundary animal diseases: assessment of socio-economic impacts and institutional responses the spread of pathogens through trade in poultry meat: overview and recent developments the spread of pathogens through trade in poultry hatching eggs: overview and recent developments. revue scientifique et technique de l'office international des epizooties emergency prevention system (empres) for transboundary animal and plant pests and diseases. the empres-livestock: an fao initiative persistence of highly pathogenic avian influenza h n virus defined by agro-ecological niche mapping h n highly pathogenic avian influenza risk in southeast asia first introduction of highly pathogenic h n avian influenza a viruses in wild and domestic birds in denmark, northern europe environmental factors influencing the spread of the highly pathogenic avian influenza h n virus in wild birds in dynamic patterns of avian and human influenza in east and southeast asia avian influenza viruses in water birds agro-ecological features of the introduction and spread of the highly pathogenic avian influenza (hpai) h n in northern nigeria multiple introductions of h n in nigeria avian influenza h n viral and bird migration networks in asia environmental factors contributing to the spread of h n avian influenza in mainland china spatial distribution and risk factors of highly pathogenic avian influenza (hpai) h n in china different environmental drivers of highly pathogenic avian influenza h n outbreaks in poultry and wild birds mapping spread and risk of avian influenza a (h n ) in china predicting the risk of avian influenza a h n infection in live-poultry markets across asia potential geographic distribution of the novel avian-origin influenza a (h n ) virus animal disease and the industrialization of agriculture assessing current and future meat and fish consumption in sub-sahara africa: learnings from fao food balance sheets and lsms household survey data. global food security rising consumption of meat and milk in developing countries has created a new food revolution. the journal of nutrition risk of introduction in northern vietnam of hpai viruses from china: description, patterns and drivers of illegal poultry trade. transboundary and emerging diseases a quantitative assessment of the risk for highly pathogenic avian influenza introduction into spain via legal trade of live poultry mycotoxins: detection methods, management, public health and agricultural trade: cabi regulation status of quarantine pests of rice seeds in the economic community of west african states (ecowas) training manual on wildlife diseases and surveillance. paris: world organization for animal health emerging biological threats: a reference guide: a reference guide: abc-clio economic optimization of a global strategy to address the pandemic threat we would like to thank ann kinzig, jim collins, ben minteer, and peter daszak for their insightful comments and discussions on the research presented here. conceptualization: tong wu, charles perrings. key: cord- -u apzw authors: michael, edwin; sharma, swarnali; smith, morgan e.; touloupou, panayiota; giardina, federica; prada, joaquin m.; stolk, wilma a.; hollingsworth, deirdre; de vlas, sake j. title: quantifying the value of surveillance data for improving model predictions of lymphatic filariasis elimination date: - - journal: plos negl trop dis doi: . /journal.pntd. sha: doc_id: cord_uid: u apzw background: mathematical models are increasingly being used to evaluate strategies aiming to achieve the control or elimination of parasitic diseases. recently, owing to growing realization that process-oriented models are useful for ecological forecasts only if the biological processes are well defined, attention has focused on data assimilation as a means to improve the predictive performance of these models. methodology and principal findings: we report on the development of an analytical framework to quantify the relative values of various longitudinal infection surveillance data collected in field sites undergoing mass drug administrations (mdas) for calibrating three lymphatic filariasis (lf) models (epifil, lymfasim, and transfil), and for improving their predictions of the required durations of drug interventions to achieve parasite elimination in endemic populations. the relative information contribution of site-specific data collected at the time points proposed by the who monitoring framework was evaluated using model-data updating procedures, and via calculations of the shannon information index and weighted variances from the probability distributions of the estimated timelines to parasite extinction made by each model. results show that data-informed models provided more precise forecasts of elimination timelines in each site compared to model-only simulations. data streams that included year post-mda microfilariae (mf) survey data, however, reduced each model’s uncertainty most compared to data streams containing only baseline and/or post-mda or longer-term mf survey data irrespective of mda coverage, suggesting that data up to this monitoring point may be optimal for informing the present lf models. we show that the improvements observed in the predictive performance of the best data-informed models may be a function of temporal changes in inter-parameter interactions. such best data-informed models may also produce more accurate predictions of the durations of drug interventions required to achieve parasite elimination. significance: knowledge of relative information contributions of model only versus data-informed models is valuable for improving the usefulness of lf model predictions in management decision making, learning system dynamics, and for supporting the design of parasite monitoring programmes. the present results further pinpoint the crucial need for longitudinal infection surveillance data for enhancing the precision and accuracy of model predictions of the intervention durations required to achieve parasite elimination in an endemic location. we report on the development of an analytical framework to quantify the relative values of various longitudinal infection surveillance data collected in field sites undergoing mass drug administrations (mdas) for calibrating three lymphatic filariasis (lf) models (epifil, lym-fasim, and transfil), and for improving their predictions of the required durations of drug interventions to achieve parasite elimination in endemic populations. the relative information contribution of site-specific data collected at the time points proposed by the who monitoring framework was evaluated using model-data updating procedures, and via calculations of the shannon information index and weighted variances from the probability distributions of the estimated timelines to parasite extinction made by each model. results show that data-informed models provided more precise forecasts of elimination timelines in each site compared to model-only simulations. data streams that included year post-mda microfilariae (mf) survey data, however, reduced each model's uncertainty most compared to data streams containing only baseline and/or post-mda or longer-term mf survey data irrespective of mda coverage, suggesting that data up to this monitoring point may be optimal for informing the present lf models. we show that the improvements observed in the predictive performance of the best data-informed models may be a function of temporal changes in inter-parameter interactions. such best data-informed models may also produce plos mathematical models of parasite transmission, via their capacity for producing dynamical forecasts or predictions of the likely future states of an infection system, offer an important tool for guiding the development and evaluation of strategies aiming to control or eliminate infectious diseases [ ] [ ] [ ] [ ] [ ] [ ] [ ] . the power of these numerical simulation tools is based uniquely on their ability to appropriately incorporate the underlying nonlinear and multivariate processes of pathogen transmission in order to facilitate plausible predictions outside the range of conditions at which these processes are either directly observed or quantified [ ] [ ] [ ] [ ] . the value of these tools for guiding policy and management decisions by providing comparative predictions of the outcomes of various strategies for achieving the control or elimination of the major neglected tropical diseases (ntds) has been highlighted in a series of recent publications [ , , ] , demonstrating the crucial role these quantitative tools are beginning to play in advancing policy options for these diseases. while these developments underscore the utility of transmission models for supporting policy development in parasite control, a growing realization is that these models can be useful for this purpose only if the biological processes are well defined and demographic and environmental stochasticity are either well-characterized or unimportant for meeting the goal of the policy modelling exercise [ ] [ ] [ ] [ ] [ ] [ ] [ ] . this is because the realized predictability of any model for a system depends on the initial conditions, parameterizations and process equations that are utilized in its simulation such that model outcomes are strongly sensitive to the choice of values used for these variables [ ] . any misspecification of these system attributes will lead to failure in accurately forecasting the future behaviour of a system, with predictions of actual future states becoming highly uncertain even when the exact representation of the underlying deterministic process is well established but precise specification of initial conditions or forcing and/or parameter values is difficult to achieve [ , ] . this problem becomes even more intractable when theoretical models depend on parameter estimates taken from other studies [ , , ] . both these challenges, viz. sensitivity to forcing conditions and use of parameter estimates from settings that are different from the dynamical environment in which a model will be used for simulation, imply that strong limits will be imposed on the realized predictability of any given model for an application [ , , ] . as we have shown recently, if such uncertainties are ignored, the ability of parasite transmission models to form the scientific basis for management decisions can be severely undermined, especially when predictions are required over long time frames and across heterogeneous geographic locations [ , , ] . these inherent difficulties with using an idealized model for producing predictions to guide management have led to consideration of data-driven modelling procedures that allow the use of information contained within observations to improve specification and hence the predictive performance of process-based models [ , , , [ ] [ ] [ ] . such approaches, termed model-data fusion or data assimilation methods, act by combining models with various data streams (including observations made at different spatial or temporal scales) in a statistically rigorous way to inform initial conditions, constrain model parameters and system states, and quantify model errors. the result is the discovery of models that can more adequately capture the prevailing system dynamics in a site, an outcome which in turn has been shown to result in the making of significantly improved predictions for management decision making [ , , , ] . initially used in geophysics and weather forecasting, these methods are also beginning to be applied in ecological modelling, including more recently in the case of infectious disease modelling [ , ] . in the latter case, the approach has shown that it can reliably constrain a disease transmission model during simulation to yield results that approximate epidemiological reality as closely as possible, and as a consequence improve the accuracy of forecasts of the response of a pathogen system exposed to various control efforts [ - , , - ] . more recently, attention has also focused on the notion that a model essentially represents a conditional proposition, i.e. that running a model in a predictive mode presupposes that the driving forces of the system will remain within the bounds of the model conceptualization or specification [ ] . if these driving forces were to change, then it follows that even a model well-calibrated to a given historical dataset will fail. new developments in longitudinal data assimilation can mitigate this problem of potential time variation of parameters via the recursive adjustment of the model by assimilation of data obtained through time [ , , ] . apart from allowing assessment of whether stasis bias may occur in model predictions, such sequential model calibration with time-varying data can also be useful for quantifying the utility of the next measurement in maximizing the information gained from all measurements together [ ] . carrying out such longitudinal model-data analysis has thus the potential for providing information to improve the efficiency and cost-effectiveness of data monitoring campaigns [ , [ ] [ ] [ ] , along with facilitating more reliable model forecasts. a key question, however, is evaluating which longitudinal data streams provide the most information to improve model performance [ ] . indeed, it is possible that from a modelling perspective using more data may not always lead to a better-constrained model [ ] . this suggests that addressing this question is not only relevant to model developers, who need observational data to improve, constrain, and test models, but also for disease managers working on the design of disease surveillance plans. at a more philosophical level, we contend that these questions have implications for how current longitudinal monitoring data from parasite control programmes can best be exploited both scientifically and in management [ ] . specifically, we suggest that these surveillance data need to be analysed using models in a manner that allows the extraction of maximal information about the monitored dynamical systems so that this can be used to better guide both the collection of such data as well as the provision of more precise estimates of the system state for use in making state-dependent decisions [ , [ ] [ ] [ ] . currently, parasite control programmes use infection monitoring data largely from sentinel sites primarily to determine if an often arbitrarily set target is met [ ] . little consideration is given to whether these data could also be used to learn about the underlying transmission dynamics of the parasitic system, or how such learning can be effectively used by management to make better decisions regarding the interventions required in a setting to meet stated goals [ , ] . here, we develop an analytical framework to investigate the value of using longitudinal lf infection data for improving predictions of the durations of drug interventions required for achieving lf elimination by coupling data collected during mass drug interventions (mdas) carried out in three example field sites to three existing state-of-the-art lymphatic filariasis (lf) models [ , , , [ ] [ ] [ ] [ ] [ ] [ ] . to be managerially relevant to current who-specified lf intervention surveillance efforts, we evaluated the usefulness of infection data collected in these sites at the time points proposed by the who monitoring framework in carrying out the present assessment [ ] . this was specifically performed by ranking these different infection surveillance data streams according to the incremental information gain that each stream provided for reducing the prediction uncertainty of each model. longitudinal pre-and post-infection and mda data from representative sites located in each of the three major regions endemic for lf (africa, india, and papua new guinea (png)) were assembled from the published literature for use in constraining the lf models employed in this study. the three sites (kirare, tanzania, alagramam, india, and peneng, png) were selected on the basis that each represents the average endemic transmission conditions (average level of infection, transmitting mosquito genus) of each of these three major extant lf regions, while providing details on the required model inputs and data for conducting this study. these data inputs encompassed information on the annual biting rate (abr) and dominant mosquito genus, as well as mda intervention details, including the relevant drug regimen, time and population coverage of mda, and times and results of the conducted microfilaria (mf) prevalence surveys (table ) . note each site also provided these infection and mda data at the time points pertinent to the existing who guidelines for conducting lf monitoring surveys during a mda programme [ ] , which additionally, as pointed out above, allowed the assessment of the value of such infection data both for supporting effective model calibration and for producing more reliable intervention forecasts. the three existing lf models employed for this study included epifil, a deterministic monte carlo population-based model, and lymfasim and transfil, which are both stochastic, individual-based models. all three models simulate lf transmission in a population by accounting for key biological and intervention processes such as impacts of vector density, the life cycle of the parasite, age-dependent exposure, density-dependent transmission processes, infection aggregation, and the effects of drug treatments as well as vector control [ , , - , , , ] . although the three models structurally follow a basic coupled immigration-death model formulation, they differ in implementation (e.g. from individual to population-based), the total number of parameters included, and the way biological and intervention processes are mathematically incorporated and parameterized. the three models have been compared in recent work [ , ] , with full details of the implementation and simulation procedures for each individual model also described [ , , , , , , , , ] . individual model parameters and fitting procedures specific to this work are given in detail in s supplementary information. we used longitudinal data assimilation methods to sequentially calibrate the three lf models with the investigated surveillance data such that parameter estimates and model predictions reflect not only the information contained in the baseline but also follow-up data points. the available mf prevalence data from each site were arranged into four different temporal data streams to imitate the current who guidelines regarding the time points for conducting monitoring surveys during an mda programme. this protocol proposes that infection data be collected in sentinel sites before the first round of mda to establish baseline conditions, no sooner than months following the third round of mda, and no sooner than months following the fifth mda to assess whether transmission has been interrupted (defined as reduction of mf prevalence to below % in a population) [ , ] . thus, the four data streams considered for investigating the value of information gained from each survey were respectively: scenario -baseline mf prevalence data only, scenario -baseline and post-mda mf prevalence data, scenario -baseline, post-mda , and post-mda mf prevalence data, and scenario -baseline and post-mda mf prevalence data. in addition to these four data streams, a fifth model-only scenario (scenario ) was also considered where no site-specific data was introduced. in this case, simulations of interventions were performed using only model-specific parameter and abr priors estimated for each region. the first step for all models during the data assimilation exercises reported here was to initially simulate the baseline infection conditions in each site using a large number of samples ( , for epifil and transfil, and , - , for lymfasim) randomly selected from the parameter priors deployed by each model. the number of parameters which were left free to be fitted to these data by each model range from (lymfasim and transfil) to (epifil). the abr, a key transmission parameter in all three models, was also left as a free parameter whose distribution was influenced by the observed abr (table ) and/or by fits to previous region-specific datasets (see s supplementary information for model-specific implementations). the subsequent steps used to incorporate longitudinal infection data into the model calibration procedure varied among the models, but in all cases the goodness-of-fit of the model outputs for the site-specific mf prevalence data was assessed using the chi-square metric (α = . ) [ ] . epifil used a sequential model updating procedure to iteratively modify the parameters with the introduction of each subsequent follow up data point through time [ ] . this process uses parameter estimates from model fits to previous data as priors for the simulation of the next data which are successively updated with the introduction of each new observation, thus providing a flexible framework by which to constrain a model using newly available data. fig summarizes the iterative algorithm used for conducting this sequential model-data assimilation exercise [ ] . lymfasim and transfil, by contrast, included all the data in each investigated stream together for selecting the best-fitting models for each time series-i.e. model selection for each data series was based on using all relevant observations simultaneously in the fitting process [ , , ] . although a limitation of this batch estimation approach is that the posterior probability of each model is fixed for the whole simulation period, unlike the case in sequential data assimilation where a restricted set of parameters is exposed to each observation (as a result of parameter constraining by data used in the previous time step)which thereby yields models that give better predictions for different portions of the underlying temporal process-here we use both methods to include and assess the impact that this implementation difference may have on the results presented below. for all models, the final updated parameter estimates from each data stream were used to simulate the impact of observed mda rounds and for predicting the impact of continued mda to estimate how many years were required to achieve % mf prevalence. interventions were modelled by using the updated parameter vectors or models selected from each scenario for simulating the impact of the reported as well as hypothetical future mda rounds on the number of years required to reduce the observed baseline lf prevalence in each site to below the who transmission threshold of % mf prevalence [ ] . when simulating these interventions, the observed mda times, regimens, and coverages followed in each site were used (table ) , while mda was assumed to target all residents aged years and above. for making mf prevalence forecasts beyond the observations made in each site, mda simulations were extended for a total of annual rounds in each site at an assumed coverage of %. while the drug-induced mf kill rate and the duration of adult worm sterilization were fixed among the models (table ) , the worm kill rate was left as a free parameter to be estimated from post-intervention data to account for the uncertainty in this drug efficacy parameter [ , , ] . the number of years of mda required to achieve the threshold of % mf prevalence was calculated from model forecasts of changes in mf prevalence due to mda for each modeldata fusion scenario. the predictions from each model regarding timelines to achieve % mf for each fitting scenario were used to determine the information gained from each data stream compared to the in all scenarios, the initial epifil models were initialized with parameter priors and a chi-square fitting criterion was applied to select those models which represent the baseline mf prevalence data sufficiently well (α = . ). the accepted models were then used to simulate the impact of interventions on mf prevalence. the chi-square fitting criterion was sequentially applied to refine the selection of models according to the post-mda mf prevalence data included in the fitting scenario. the fitted parameters from selection of acceptable models at each data point were used to predict timelines to achieve % mf prevalence. the scenarios noted in the blue boxes indicate the final relevant updating step before using the fitted parameters to predict timelines to achieve % mf in that data fitting scenario. information attributable to the model itself [ , , ] . the relative information gained from a particular data stream was calculated as i d = h m -h md where h measures the entropy or uncertainty associated with a random variable, h m denotes predictions from the model-only scenario (scenario ) which essentially represents the impact of prior knowledge of the system, and h md signifies predictions from each of the four model-data scenarios (i.e. scenarios [ ] [ ] [ ] [ ] . the values of i d for each data scenario or stream were compared in a site to infer which survey data are most useful for reducing model uncertainty. the shannon information index was used to measure entropy, h, as follows: is the discrete probability density function (pdf) of the number of years of mda predicted by each fitted model to reach % mf, and is estimated from a histogram of the respective model predictions for m bins (of equal width in the range between the minimum and maximum values of the pdfs) [ , ] . to statistically compare two entropy values, a permutation test using the differential shannon entropy (dse) was performed [ ] . dse is defined as |h -h | where h was calculated from the distribution of timelines to achieve % mf for a given scenario, y , and h was calculated from the distribution of timelines to achieve % mf for a different scenario, y . the list of elements in y and y were combined into a single list of size y + y and the list was permuted , times. dse was then recalculated each time by calculating a new h from the first y elements and a new h from the last y elements from each permutation, from which p-values may be quantified as the proportion of all recalculated dses that were greater than the original dse. model predictions of the mean and variance in timelines to lf elimination were weighted according to the frequencies by which predictions occurred in a group of simulations. in general, if d , d ,. . .,d n are data points (model predictions in the present case) that occur in an ensemble of simulations with different weights or frequencies w ,w ,. . .,w n , then the weighted mean, here, n is the number of data points and n is the number of non-zero weights. in this study, the weighted variance of the distributions of predicted timelines to achieve % mf prevalence was calculated to provide a measure of the precision of model predictions in addition to the entropy measure, h. a similar weighting scheme was also used to pool the timeline predictions of all three models. here, predictions made by each of the three models for each data scenario were weighted as above, and a composite weighted % percentile interval for the pooled predictions was calculated for each data stream. this was done by first computing the weighted percentiles for the combined model simulations from which the pooled . th and . th percentile values were quantified. the matlab function, wprctile, was used to carry out this calculation. the extent by which parameter constraints are achieved through the coupling of models with data was evaluated to determine if improvements in such constraints by the use of additional data may lead to reduced model prediction uncertainty [ ] . parameter constraint was calculated as the ratio of the mean standard deviation of all fitted parameter distributions to the mean standard deviation of all prior parameter distributions. a ratio of less than one indicates the fitted parameter space is more constrained than the prior parameter space [ ] . this assessment was carried out using the epifil model only. in addition, pairwise parameter correlations were also evaluated to assess whether the sign, magnitude, and significance of these correlations changed by scenario to determine if using additional data might alter these interactions to better constrain a model. for this assessment, spearman's correlation coefficients and p-values testing the hypothesis of no correlation against the alternative of correlation were calculated, and the exercise was run using the estimated parameters from the epifil model. epifil was used to conduct a sensitivity analysis investigating whether the trend in relative information gained by coupling the model with longitudinal data was dependent on the interventions simulated. the same series of simulations (for three lf endemic sites and five fitting scenarios) were completed with the extended mda coverage beyond the observations given in table set here at % instead of % to represent an optimal control strategy. as before, the timelines to reach % mf prevalence in each fitting scenario were calculated and used to determine which data stream provided the model with the greatest gain of information. the results were compared to the original series of simulations to assess whether the trends are robust to changes in the intervention coverages simulated. epifil was also used to perform another sensitivity analysis expanding the number of data streams to investigate if the who monitoring scheme is adequate for informing the making of reliable model-based predictions of timelines for achieving lf elimination. to perform this sensitivity analysis, pre-and post-mda data from villupuram district, india that provide extended data points (viz. scenario - as previously defined, plus scenario -baseline, post-mda , post-mda , and post-mda mf prevalence data, and scenario -baseline, post-mda , post-mda , post-mda , and post-mda mf prevalence) were assembled from the published literature [ , ] . the timelines to reach % mf prevalence and the entropy for each of these additional scenarios were calculated to determine whether additional data streams over those recommended by who are required for achieving more reliable model constraints, which among these data might be considered as compulsory, and which might be optional for supporting predictions of elimination. differences in predicted medians, weighted variances and entropy values between data scenarios, models and sites were statistically evaluated using kruskall-wallis tests for equal medians, f-tests for equality of variance, and dse permutation tests, respectively. p-values for assessing significance for all pairwise tests were obtained using the benjamini-hochberg procedure for controlling the false discovery rate, i.e. for protecting against the likelihood of obtaining false positive results when carrying out multiple testing [ ] . here, our goal was twofold. first, to determine if data are required to improve the predictability of intervention forecasts by the present lf models in comparison with the use of theoretical models only, and second, to evaluate the benefit of using different longitudinal streams of mf survey data for calibrating the three models in order to determine which data stream was most informative for reducing the uncertainty in model predictions in a site. table summarises the key results from our investigation of these questions: these are the number of accepted best-fitting models for each data stream or scenario in the three study sites (table ) , the predicted median and range ( . th - . th percentiles) in years to achieve the mf threshold of % mf prevalence, the weighted variance and entropy values based on these predictions, and the relative information gained (in terms of reduced prediction uncertainty) by the use of longitudinal data for constraining the projections of each of the three lf models investigated. even though the number of selected best-fit models based on the chi-square criterion (see methods) differed for each site and model, these results indicate unequivocally that models constrained by data provided significantly more precise intervention predictions compared to model-only predictions ( table ). note that this was also irrespective of the two types of longitudinal data assimilation procedures (sequential vs. simultaneous) used by the different models in this study. thus, for all models and sites, model-only predictions made in the absence of data (scenario ) showed the highest prediction uncertainty, highlighting the need for data to improve the predictive performance of the present models. the relative information gained by using each data stream in comparison to model-only predictions further support this finding, with the best gains in reducing model prediction uncertainty provided by those data constraining scenarios that gave the lowest weighted variance and entropy values; as much as % to % reductions in prediction variance were achieved by these scenarios in comparison to modelonly predictions between the three models ( table ). the results also show, however, that data streams including post-mda mf survey data (scenarios and ) reduced model uncertainty (based on both the variance and entropy measures) most compared to data streams containing only baseline and/or post-mda mf survey data (scenarios and ) ( table ) . although there were differences between the three models (due to implementation differences either in how the models are run (monte carlo deterministic vs. individual-based) or in relation to how the present data were assimilated (see above)), overall, scenario , which includes baseline, post-mda , and post-mda data, was shown to most often reduce model uncertainty the greatest. additionally, there was no statistical difference between the performances of scenarios and in those cases where scenario resulted in the greatest gain of information (table ) . it is also noticeable that the best constraining data stream for each combination of site and model also produced as expected the lowest range in predictions of the numbers of years of annual mda required to achieve the % mf prevalence in each site, with the widest ranges estimated for model-only predictions (scenario ) and the shorter data streams (scenario ). in general, this constriction in predictions also led to lower estimates of the median times to achieve lf elimination, although this varied between models and sites ( table ) . the change in the distributions of predicted timelines to lf elimination without and with model constraining by the different longitudinal data streams is illustrated in fig for the kirare site (see s supplementary information for results obtained for the other two study villages investigated here). the results illustrate that both the location and length of the tail of the prediction distributions can change as models are constrained with increasing lengths of longitudinal data, with inclusion of post-mda mf survey data consistently producing a narrower or sharper range of predictions compared to when this survey point is excluded. fig compares the uncertainty in predictions of timelines to achieve elimination made by each of the three models without (scenario ) and via their constraining by the data streams providing the lowest prediction entropy for each of the models per site. note that variations in scenario predictions among the three models directly reflect the different model structures, parameterizations, and the presence (or absence) of stochastic elements. the boxplots in the figure, however, show that for all three sites and models, calibration of each model by data the lowest entropy scenario for each site is bolded and shaded grey. additional scenarios shaded grey are not significantly different from the lowest entropy scenario. data assimilation in filarial model predictions greatly reduces the uncertainty in predictions of the years of annual mda required to eliminate lf compared to model-only predictions, with the data streams producing the lowest entropy for simulations in each site significantly improving the precision of these predictions ( table ). this gain in precision, and thus the information gained using these data streams, is, as expected, greater for the stochastic lymfasim and transfil models compared to the deterministic epifil model. note also that even though the ranges in predictions of the annual mda years required to eliminate lf by the data streams providing the lowest prediction entropy differed statistically between the three models, the values overlapped markedly (e.g. for kirare the ranges are - , - , - for epifil, lymfasim and transfil data assimilation in filarial model predictions respectively), suggesting the occurrence of a similar constraining of predictive behaviour among the three models. to investigate this potential for a differential model effect, we further pooled the predictions from all three models for all the data scenarios and evaluated the value of each investigated data stream for improving their combined predictive ability. the weighted % percentile intervals from the pooled predictions were used for carrying out this assessment. the results are depicted in fig and indicate that, as for the individual model predictions, uncertainty in the collective predictions by the three lf models for the required number of years to eliminate lf using annual mda in each site may be reduced by model calibration to data, with the longitudinal mf prevalence data collected during the later monitoring periods (scenarios and ) contributing most to improving the multi-model predictions for each site. the boxplots show that by calibrating the models to data streams, more precise predictions are able to be made regarding timelines to achieve % mf prevalence across all models and sites. the results of pairwise f-tests for variance, performed to compare the weighted variance in timelines to achieve % mf prevalence between model-only simulations (scenario ) and the lowest entropy simulations (best scenario) (see table ), show that the predictions for the best scenarios are significantly different from the predictions for the model-only simulations. significance was determined using the benjamini-hochberg procedure for controlling the false discovery rate (q = . ). for epifil, lymfasim and transfil, the best scenarios are scenarios , , and for kirare, scenarios , , and for alagramam, and scenarios , , and for peneng, respectively. we attempted to investigate if model uncertainty in predictions by the use of longitudinal data was a direct function of parameter constraining by the addition of data. given the similarity in outcomes of each model, we remark on the results from the fits of the technically easier to run epifil model to evaluate this possibility here. the assessment of the parameter space constraint achieved through the inclusion of data was made by determining if the fitted parameter distributions for the model became reduced in comparison with priors as data streams were added to the system [ ] . the exercise showed that the size of the estimated parameter distributions reduced with addition of data, with even scenario data producing reductions for kirare and peneng (fig ) . in the case of alagramam, however, there was very little, if any, constraint in the fitted parameter space compared to the prior parameter space. this result, together with the fact that even using all the data in kirare and peneng produced up to only between . to % reductions in fitted parameter distributions when compared to the priors, indicate that the observed model prediction uncertainty in this study may be due to other complex factors connected with model parameterization. table provides the results of an analysis of pairwise parameter correlations of the selected best-fitting models for data scenario compared to those selected by the data stream that gave the best reduction in epifil prediction uncertainty for alagramam (scenario ). these results show that while the parameter space was not being constrained with the addition of more data, the pattern of parameter correlations changed in a complex manner between the two constraining data sets. for example, although the number of significantly correlated parameters did not differ, the magnitude and direction of parameter correlations were shown to change between the two data scenarios ( table ) . the corresponding results for kirare and peneng are shown in s supplementary information , and indicate that a broadly similar pattern of changes in parameter associations also occurred as a result of model calibration to the sequential data measured from those sites. this suggests that this outcome may constitute a general phenomenon at least with regards to the sequential constraining of epifil using longitudinal mf prevalence data. an intriguing finding (from all three data settings) is that the most sensitive parameters in this regard, i.e. with respect to altered strengths in pairwise parameter correlations, may be those representing the relationship of various components of host immunity with different transmission processes, including with adult worm mortality, rates of production and survival of mf, larval development rates in the mosquito vector and infection aggregation (table ) . this suggests that, as more constraining data are added, changes in the multidimensional parameter relationships related to host immunity could contribute to the sequential reductions in the lf model predictive uncertainty observed in this study. the lf elimination timeline predictions used above were based on modelling the impacts of annual mda given the reported coverages in each site followed by an assumed standard coverage for making longer term predictions (see methods). this raises the question as to whether the differences detected in the case of the best constraining data stream between the present study sites and between models ( table ) could be a function of the simulated mda coverages in each site. to investigate this possibility, we used epifil to model the outcome of changing the assumed mda coverage in each site on the corresponding entropy and information gain trends in elimination predictions made from the models calibrated to each of the site-specific data scenarios/streams investigated here. the results of increasing the assumed coverage of mda to % for each site are shown in fig and indicate that the choice of mda coverage in this study are unlikely to have significantly influenced the conclusion made above that the best performing data streams for reducing model uncertainty for predicting lf elimination pertains to data scenarios and . however, while the model-predicted timelines to achieve the % mf prevalence threshold using the observed mda coverage followed by % mda coverage showed that the data stream which most reduced uncertainty did not change from the impact of using the observed mda coverage followed by % mda coverage modelled for kirare and peneng (table , fig ) , this was not the case for alagramam, where data from scenario with a % coverage resulted in the greatest reduction in entropy compared to the original results using % coverage which indicated that scenario data performed best (table , fig ) . notably, though, the entropy values of predictions using the data scenario and constraints were not statistically different for this site (p-value < . ) (fig ) . epifil was also used to expand the number of calibration scenarios using a dataset with longer term post-mda data from villupuram district, india. this dataset contained two addition data streams: scenario which included baseline, post-mda , post-mda , and post-mda mf data, and scenario , which included baseline, post-mda , post-mda , post-mda , and post-mda mf data. scenario thus contained the most post-mda data and was demonstrated to be the most effective for reducing model uncertainty, but this effect was not statistically significantly different from the reductions produced by assimilating data contained in table . spearman parameter correlations for scenarios (lower left triangle) and (upper right triangle) for alagramam, india. data assimilation in filarial model predictions scenarios and ( table ). the inclusion of more data than are considered in scenario therefore did not result in any significant additional reduction in model uncertainty. epifil was used to evaluate the accuracy of the data-driven predictions of the timelines required to meet the goal of lf elimination based on breaching the who-set target of % mf for all sites, either scenario or had the lowest entropies, and scenario was not significantly different from scenario for kirare and alagramam. these results were not statistically different from the results given % coverage (see table ), suggesting that the data stream associated with the lowest entropy is robust to changes in the interventions simulated. scenarios where the weighted variance or entropy were not significantly different from the lowest entropy scenario are noted with the abbreviation ns. significance was determined using the benjamini-hochberg procedure for controlling the false discovery rate (q = . ). https://doi.org/ . /journal.pntd. .g table . predictions of timelines to achieve % mf in villupuram district, india, considering extended post-mda data. reporting those scenarios which are statistically significantly different from each other by numbers ( - ) as superscripts. for example, the weighted variance for scenario has the superscript numbers ( ) ( ) ( ) ( ) ( ) ( ) to indicate that the weighted variance for scenario is significantly different from the weighted variance for scenarios - . significance was determined using the benjamini-hochberg procedure for controlling the false discovery rate (q = . ) in all pairwise statistical tests. + information gained by each data stream (scenario - ) are presented in comparison to the information contained in the model-only simulation (scenario ) https://doi.org/ . /journal.pntd. .t data assimilation in filarial model predictions prevalence. this analysis was performed by using the longitudinal pre and post-infection and mda data reported for the nigerian site, dokan tofa, where elimination was achieved according to who recommended criteria after seven rounds of mda (table ). the data from this site comprised information on the abr and dominant mosquito genus, as well as details of the mda intervention carried out, including the relevant drug regimen applied, time and population coverage of mda, and outcomes from the mf prevalence surveys conducted at baseline and at multiple time points during mda [ ] . the results of model predictions of the timelines to reach below % mf prevalence as a result of sequential fitting to the mf prevalence data from this site pertaining to scenarios - (as defined above) are shown in table . note that in the post mda , and surveys, as no lf positive individuals were detected among the sample populations, we used a one-sided % clopper-pearson interval to determine the expected upper one-sided % confidence limits for these sequentially observed zero infection values data assimilation in filarial model predictions using the "rule of three" approximation after k empty samples formula [ ] . the results show that model constraining by scenario , which includes baseline and post-mda data, and scenario , which includes baseline, post-mda , and post-mda data, resulted in both the least entropy values and the shortest predicted times, i.e., from as low as to as high as years, required for achieving lf elimination in this site ( table ). the data in table show that the first instance the calculated one-sided upper % confidence limit in this setting fell below % mf prevalence also occurred post mda (i.e after years of mda). this is a significant result, and indicates that apart from being able to reduce prediction uncertainty, the best data-constrained models are also able to more accurately predict the maximal time ( years) by which lf elimination occurred in this site. our major goal in this study was to compare the reliability of forecasts of timelines required for achieving parasite elimination made by generic lf models versus models constrained by sequential mf prevalence surveillance data obtained from field sites undergoing mda. a secondary aim was to evaluate the relative value of data obtained at each of the sampling time points proposed by the who for monitoring the effects of lf interventions in informing these model predictions. this assessment allowed us to investigate the role of these data for learning system dynamics and measure their value for guiding the design of surveillance programmes in order to support better predictions of the outcomes of applied interventions. fundamentally, however, this work addresses the question of how best to use predictive parasite transmission models for guiding management decision making, i.e. whether this should be based on the use of ideal models which incorporate generalized parameter values or on models with parameters informed by local data [ ] . if we find that data-informed models can reduce prediction uncertainty significantly compared to the use of theoretical models unconstrained by data, then it is clear that to be useful for management decision making we require the application of model-data assimilation frameworks that can effectively incorporate information from appropriate data into models for producing reliable intervention projections. antithetically, such a finding implies that using unconstrained ideal models in these circumstances will provide only approximate predictions characterized by a degree of uncertainty that might be too large to be useful for reliable decision making [ , , ] . here, we have used three state-of-the-art lf models calibrated to longitudinal human mf prevalence data obtained from three representative lf study sites to carry out a systematic analysis of these questions in parasite intervention modelling (see also walker et al [ ] for a recent study highlighting the importance of using longitudinal sentinel site data for improving the prediction performances of the closely-related onchocerciasis models). further, by iteratively testing the reduction in the uncertainty of the projections of timelines required to achieve lf elimination in a site made by the models matching each observed data point, we have also quantified the relative values of temporal data streams, including assessing optimal record lengths, for informing the current lf models. our results provide important insights as to how best to use process models for understanding and generating predictions of parasite dynamics. they also highlight how site-specific longitudinal surveillance data coupled with models can be useful for providing information about system dynamics and hence for improving predictions of relevance to management decision-making. the first result of major importance from our work is that models informed by data can significantly reduce predictive uncertainty and hence improve performance of the present lf models for guiding policy and management decision-making. our results show that these improvements in predictive precision were consistent between the three models and across all three of our study sites, and can be very substantive with up to as much as % to % reductions in prediction variance obtained by the best data-constrained models in a site compared to the use of model-only predictions ( table ). the practical policy implications of this finding can also be gleaned from appraising the actual numerical ranges in the predictions made by each individual model for each of the modelling scenarios investigated here. in the case of epi-fil, the best data-informed model (scenario in peneng) gave an elimination prediction range of - years, while the corresponding model-only predictions for this site indicated a need for between - years of annual mda (table ). these gains in information from using data to inform model parameters and hence predictions were even larger for the two stochastic models investigated here, viz. lymfasim and transfil, where ranges as wide as - years predicted by model-only scenarios were reduced to - years for the best data-informed models in the case of lymfasim for kirare village, and from as broad as - years to - years respectively in the case of transfil for peneng (table ). these results unequivocally indicate that if parasite transmission models are used unconstrained by data, i.e. based on general parameter values uninformed by local data, it would lead to the making of predictions that would be marked by uncertainties that are likely to be far too large to be meaningful for practical policy making. if managers are risk averse, this outcome will also mean their need to plan interventions for substantially much longer than necessary, with major implications for the ultimate cost of the programme. note also that although statistically significant changes in the median years of mda required to achieve lf elimination were observed for the best datainformed models for all the three lf model types in each site, these were relatively small compared to the large reductions seen in each model's predictive uncertainly (table , fig ) . this result highlights that the major gains from constraining the present models by data lies in improving their predictive certainty rather than in advancing their average behaviour. however, our preliminary analysis of model predictive accuracy suggests that the best data-constrained models may also be able to generate more accurate predictions of the impact of control ( table ), indicating that, apart from simply reducing predictive uncertainty, such models could additionally have improved capability for producing more reliable predictions of the outcomes of interventions carried out in a setting. the iterative testing of the reduction in forecast uncertainty using mf surveillance data measured at time points proposed by the who (to support assessment of whether the threshold of % mf prevalence has been reached before implementation units can move to post-treatment surveillance [ ]) has provided further insights into the relative value of these data for improving the predictive performance of each of the present lf models. our critical finding here is that parameter uncertainty in all three lf models was similarly reduced by the assimilation of a few additional longitudinal data records (table ). in particular, we show that data streams comprising baseline + post-mda + post-mda (scenario ) and those comprising baseline + post-mda data (scenario ) best reduced parameter-based uncertainty in model projections of the impact of mdas carried out in each study site irrespective of the models used. although preliminary, a potential key finding is that the use of longer-term data additional to the data measured at the who proposed monitoring time points did not lead to a significant further reduction in parameter uncertainty (table ) . also, the finding that the who data scenarios and were adequate for constraining the present lf models appears not to be an artefact of variations in the mda coverages observed between the three study sites (fig ) . these results suggest that up to years of post-mda mf prevalence data are sufficient to constrain model predictions of the impact of lf interventions at a time scale that can go up to as high as to years depending on the site and model, and that precision may not improve any further if more new data are added ( table , table ). given that the who post-mda lf infection monitoring protocol was developed for the purpose solely focussed on supporting the meeting of set targets (e.g. the % mf prevalence threshold) and not on a priori hypotheses regarding how surveillance data could be used also to understand the evolution and hence prediction of the dynamical parasitic system in response to management action, our results are entirely fortuitous with respect to the value of the current lf monitoring data for learning about the lf system and its extinction dynamics in different settings [ ] . they do, nonetheless, hint at the value that coupling models to data may offer to inform general theory for guiding the collection and use of monitoring data in parasite surveillance programmes in a manner that could help extract maximal information about the underlying parasite system of interest. our assessment of whether the incremental increase in model predictive performance observed as a result of assimilating longitudinal data may be due to parameter constraining by the addition of data has shed intriguing new light on the impact that qualitative changes in dynamical system behaviour may have on parameter estimates and structure, and hence on the nature of the future projections of system change we can make from models. our major finding in this regard is that even though the parameter space itself may not be overly constrained by the best data stream (scenario in this case for alagramam village), the magnitude and direction of parameter correlations, particularly those representing the relationship of different components of host immunity with various transmission processes, changed markedly between the shorter (scenario ) and seemingly optimal data streams (scenario ). this qualitative change in system behaviour induced by alteration in parameter interactions in response to perturbations has been shown to represent a characteristic feature of complex adaptive ecological systems, particularly when these systems approach a critical boundary [ ] [ ] [ ] . this underscores yet another important reason to incorporate parameter information from data for generating sound system forecasts [ ] . the finding that additional data beyond years post-mda did not appear to significantly improve model predictive performance in this regard suggests that pronounced change in lf parameter interactions in response to mda interventions may occur generally around this time point for this parasitic disease, and that once in this parameter regime further change appears to be unlikely. this is an interesting finding, which not only indicates that coupling models to at least years post-mda will allow detection of the boundaries delimiting the primary lf parameter regions with different qualitative behaviour, but also that the current who monitoring protocol might be sufficient to allow this discovery of system change. although our principal focus in this study was in investigating the value of longitudinal data for informing the predictive performance of the current lf models, the results presented here have also underscored the existence of significant spatial heterogeneity in the dynamics of parasite extinction between the present sites ( table , fig ) . in line with our previous findings, this observed conditional dependency of systems dynamics on local transmission conditions means that timelines or durations of interventions required to break lf transmission (as depicted in table ) will also vary from site to site even under similar control conditions [ ] [ ] [ ] ] . as we indicated before, this outcome implies that we vitally require the application of models to detailed spatio-temporal infection surveillance data, such as that exemplified by the data collected by countries in sentinel sites as part of their who-directed monitoring and evaluation activities, if we are to use the present models to make more reliable intervention predictions to drive policy and management decisions (particularly with respect to the durations of interventions required, need for switching to more intensified or new mda regimens, and need for enhanced supplementary vector control) in a given endemic setting [ ] . as we have previously pointed out, the development of such spatially adaptive intervention plans will require the development and use of spatially-explicit data assimilation modelling platforms that can couple geostatistical interpolation of model inputs (eg. abr and/or sentinel site mf/ antigen prevalence data) with discovery of localized models from such data in order to produce the required regional or national intervention forecasts [ ] . the estimated parameter and prediction uncertainties presented here are clearly dependent on the model-data fusion methodology and its implementation, and the cost function used to discover the appropriate models for a data stream [ ] . while we have attempted to evaluate differences in individual model structures, their computer implementation, and the data assimilation procedures followed (e.g. sequential vs. simultaneous data assimilation), via comparing the collective predictions of the three models versus the predictions provided by each model singly, and show that these factors are unlikely to play a major role in influencing the current results, we indicate that future work must address these issues adequately to improve the initial methods we have employed in this work. currently, we are examining the development of sequential bayesian-based multi-model ensemble approaches that will allow better integration of each model's behaviour as well as better calculation of each model's transient parameter space at each time a new observation becomes available [ ] . this work also involves the development of a method to fuse information from several indicators of infection (e.g. mf, antigenemia, antibody responses [ ] ) together to achieve a more robust constraining of the present models. as different types of data can act as mutual constraints on a model, we also expect that such multiindicator model-data fusion methods will additionally address the problem of equifinality, which is known to complicate the parameterization of complex dynamical models [ , ] . of course, the ultimate test of the results reported here, viz. that lf models constrained by coupling to year post-mda data can provide the best predictions of timelines for meeting the % mf prevalence threshold in a site, is by carrying out the direct validation of our results against independent observations (as demonstrated by the preliminary validation study carried out here using the dokan tofa data (tables and )). we expect that data useful for performing these studies at scale may be available at the sentinel site level in the countries carrying out the current who-led monitoring programme. the present results indicate that access to such data, and to post-treatment surveillance data which are beginning to be assembled by many countries, is now a major need if the present lf models are to provide maximal information about parasite system responses to management and thus generate better predictions of system states for use in policy making and in judging management effectiveness in different spatiotemporal settings [ , ] . given that previous modelling work has indicated that if the globally fixed who-proposed % mf prevalence threshold is insufficient to break lf transmission in every setting (and thus conversely leading to significant infection recrudescence [ ] ), the modelling of such spatio-temporal surveillance data will additionally allow testing if meeting this recommended threshold will indeed result in successfully achieving the interruption of lf transmission everywhere. the epidemiology of filariasis control. the filaria epidemiological modelling for monitoring and evaluation of lymphatic filariasis control mathematical modelling and the control of lymphatic filariasis. the lancet infectious diseases heterogeneous dynamics, robustness/fragility trade-offs, and the eradication of the macroparasitic disease, lymphatic filariasis continental-scale, data-driven predictive assessment of eliminating the vector-borne disease, lymphatic filariasis, in sub-saharan africa by sequential modelling of the effects of mass drug treatments on anopheline-mediated lymphatic filariasis infection in papua new guinea assessing endgame strategies for the elimination of lymphatic filariasis: a model-based evaluation of the impact of dec-medicated salt predicting lymphatic filariasis transmission and elimination dynamics using a multi-model ensemble framework data-model fusion to better understand emerging pathogens and improve infectious disease forecasting ecological forecasting and data assimilation in a data-rich era the role of data assimilation in predictive ecology effectiveness of a triple-drug regimen for global elimination of lymphatic filariasis: a modelling study. the lancet infectious diseases epidemic modelling: aspects where stochasticity matters. mathematical biosciences relative information contributions of model vs. data to short-and long-term forecasts of forest carbon dynamics inference in disease transmission experiments by using stochastic epidemic models plug-and-play inference for disease dynamics: measles in large and small populations as a case study the limits to prediction in ecological systems big data need big theory too hierarchical modelling for the environmental sciences: statistical methods and applications parameter and prediction uncertainty in an optimized terrestrial carbon cycle model: effects of constraining variables and data record length bayesian calibration of simulation models for supporting management of the elimination of the macroparasitic disease, lymphatic filariasis an improved state-parameter analysis of ecosystem models using data assimilation. ecological modelling bayesian calibration of mechanistic aquatic biogeochemical models and benefits for environmental management the model-data fusion pitfall: assuming certainty in an uncertain world transmission dynamics and control of severe acute respiratory syndrome curtailing transmission of severe acute respiratory syndrome within a community and its hospital transmission dynamics of the etiological agent of sars in hong kong: impact of public health interventions philosophical issues in model assessment. model validation: perspectives in hydrological science a sequential monte carlo approach for marine ecological prediction a sequential bayesian approach for hydrologic model selection and prediction inferences about coupling from ecological surveillance monitoring: approaches based on nonlinear dynamics and information theory. towards an information theory of complex networks using model-data fusion to interpret past trends, and quantify uncertainties in future projections, of terrestrial ecosystem carbon cycling rate my data: quantifying the value of ecological data for the development of models of the terrestrial carbon cycle estimating parameters of a forest ecosystem c model with measurements of stocks and fluxes as joint constraints. oecologia global mapping of lymphatic filariasis adaptive management and the value of information: learning via intervention in epidemiology general rules for managing and surveying networks of pests, diseases, and endangered species geographic and ecologic heterogeneity in elimination thresholds for the major vector-borne helminthic disease, lymphatic filariasis modelling strategies to break transmission of lymphatic filariasis-aggregation, adherence and vector competence greatly alter elimination mathematical modelling of lymphatic filariasis elimination programmes in india: required duration of mass drug administration and post-treatment level of infection indicators mathematical models for lymphatic filariasis transmission and control: challenges and prospects. parasite vector the lymfasim simulation program for modeling lymphatic filariasis and its control the dynamics of wuchereria bancrofti infection: a model-based analysis of longitudinal data from pondicherry parameter sampling capabilities of sequential and simultaneous data assimilation: ii. statistical analysis of numerical results an evaluation of models for partitioning eddy covariance-measured net ecosystem exchange into photosynthesis and respiration normalized measures of entropy entropyexplorer: an r package for computing and comparing differential shannon entropy, differential coefficient of variation and differential expression impact of years of diethylcarbamazine and ivermectin mass administration on infection and transmission of lymphatic filariasis controlling the false discovery rate-a practical and powerful approach to multiple testing epidemiological and entomological evaluations after six years or more of mass drug administration for lymphatic filariasis elimination in nigeria the study of plant disease epidemics bayesian data assimilation provides rapid decision support for vector-borne diseases modelling the elimination of river blindness using long-term epidemiological and programmatic data from mali and senegal. epidemics socio-ecological dynamics and challenges to the governance of neglected tropical disease control. infectious diseases of poverty early-warning signals for critical transitions early warning signals of extinction in deteriorating environments practical limits for reverse engineering of dynamical systems: a statistical analysis of sensitivity and parameter inferability in systems biology models multi-sensor model-data fusion for estimation of hydrologic and energy flux parameters. remote sensing of environment key: cord- -qjdg o authors: simoes, joana margarida title: spatial epidemic modelling in social networks date: - - journal: aip conf proc doi: . / . sha: doc_id: cord_uid: qjdg o the spread of infectious diseases is highly influenced by the structure of the underlying social network. the target of this study is not the network of acquaintances, but the social mobility network: the daily movement of people between locations, in regions. it was already shown that this kind of network exhibits small world characteristics. the model developed is agent based (abm) and comprehends a movement model and a infection model. in the movement model, some assumptions are made about its structure and the daily movement is decomposed into four types: neighborhood, intra region, inter region and random. the model is geographical information systems (gis) based, and uses real data to define its geometry. because it is a vector model, some optimization techniques were used to increase its efficiency. since a long time, human epidemics have interested scientists of several areas. once diseases spread amongst people, it is impossible to ignore the role of social networks in embedding this phenomenon. therefore, the architecture and general topological features of the social network should be considered in the model. in several studies it has been considered the network of acquaintances [ ] , which is important in diseases that require a close, prolonged contact. however, in highly contagious diseases, the infection may be passed by a short physical contact, and in this case it is more important to track the movement of individuals. in this model, it was considered the social mobility network: the daily movement of individuals, which has been already referred in the literature as a complex network with a small world behaviour [ ] . in complex systems, the interaction among the constituents of the system, and the interaction between the system and its environment, are of such a nature that the system cannot be fully understood by simply analysing its components [ ] . in this paper, it is described a simulation system using artificial agents integrated with geographical information systems (gis) that helps to understand the spatial and temporal behaviour of a epidemic phenomena. the utility of spatially explicit agent oriented simulations is demonstrated by simulating alternative scenarios to compare differences in the spatial structure of the mobility network and the geographical distribution of individuals. agent based models (abm) offer a great potential for studying human complex behaviour, interacting within a spatial framework [ ] . unlike what happens with cellular automata (ca), in abm space is continuous and location is explicit, which means that individuals can be simulated, independent of the environment. this allows to specify rules focused on the individuals, and not on space. the present model is inspired by a site exchange cellular automata [ ] , which considers two phases for each time step: movement and infection, assuming there is no virus transmission while the individual is moving. the movement rules and infection rules are determined by a movement model, and a infection model that are going to be described on the next paragraphs. the domain of the model is divided into small subunits with a geographical relevance: the regions. this regions can have several definitions depending on the scale, but in this case it were consider the concelhos, according to the ine definition (administrative division code/ revision approved by the deliberation n of / / ). this choice was motivated by data availability and also for keeping the simulations reasonably fast. however, the model should be run with smaller regions, once conceptually the region definition is closer to the city definition. the movement rules try to emulate the daily movement of individuals according to the diversity of its activities (working, shopping, etc.). based on these regions, four ranges of movement were considered: neighbourhood, intra region, inter region and small world. neighborhood is the random movement of a individual within its immediate neighborhood (in this case it was considered a radius, defined as a parameter). this stands for the motion of an individual within its street or neighborhood; it is considered that it is where it spends most of its time. intra region is the random movement inside a region. this stands for the movement of an individual inside a city: for instance for working or going to the cinema. inter region is the random movement of an individual on the neighbor regions. this represents the travel to nearby cities, for instance for shopping or visiting friends and family. finally, it was considered a total random movement named small world (sw). this movement is a very tiny fraction of the total, and represents the situation that a small number of individuals have large range movements, which produce the shortcuts in the network, responsible for the sw phenomena. probabilities were attributed to each kind of movement, based on common sense of what are the most probable activities. this weights or probabilities should be based on real mobility data that were not available for this study, and therefore the results should be seen with caution. on fig. , it is shown a graphic with the probabilities attributed to each kind of movement. the update of the movement model is synchronous. the infection model considered was a slir (susceptible-latent-infected-removed). the contagious (state change from susceptible to latent) occurs every time a susceptible meets a infectious within a certain radius. the state change from latent to infectious, and infectious to removed is only determined by time. the contagious model is not specific for a disease, but it is flexible enough to be adjusted and fit the characteristics of a given disease. the update of the movement model is synchronous . the model is vector based, considering points moving continuously over a polygon layer (fig. ) . the polygon layer reads the geographic information from a shapefile (the esri gis format), using the shapelib api. the shapelib provides the programmer a structure with all the information contained in the shapefile, and in this way it is not necessary to program the low level access to the datafile (fig. ) . the advantage of reading these files is the possibility of displaying different geographical configurations. several gis functionalities were implemented (like buffers and zooms) and the algorithms were optimized taking into consideration the speed of the simulations. once the operation of searching for features was intensively used, one of the efficiency algorithms developed was a search method: the quadtree. a quadtree is a tree structure used to encode two-dimensional spaces. the image is recursively subdivided into sub quadrants. the quadtree is used for raster spaces, so in this case it was used a adaptation of the quadtree for vectorial spaces. structure of the vectorial quadtree implemented in this model. the sensitivity analysis of the movement model was performed by running simulations for exclusively one type of movement. the domain of the model for these simulations was portugal, with concelhos as regions. the initial conditions are the population distribution of census (ine) and the reported cases of mumps in , which were in the origin of a small epidemic in this country. although the simulations are agent based, the results are shown at the level of region, in order to be more perceptible. darker shadings correspond to a greater amount of infected and removed individuals in the region. in the neighbourhood movement simulation (fig. ) , the infection is much more restricted, in magnitude and in spatial extension, than in all other simulations. this is obviously due to the tighter movement range. it is also important to remark that the stability of the epidemic, occurs earlier in the neighbourhood simulations than in all others. the small world movement simulation (fig. ) presents a totally different distribution of population. as the individuals reach every part of the domain, so does the epidemic. however, for being so contagious its inefficiency is reduced because many individuals die before they transmitted the disease, and so the stabilization of the epidemic is reached later than in the other simulations. the impact of this component, even if present in a small amount, can be seen by running a simulation with no small world movement (fig. ) and another simulation with a small world movement probability of . (fig. ) . in the case when the small world movement is included, the epidemic reaches a greater number of people, and reaches a greater part of the country. this analysis calls attention to the importance of the mobility network, embedded in the epidemic model as it has a determinant impact in the evolution of the epidemic. in this network, it was shown how the small component of random movement (characteristic of small world networks) has an effective influence on the results, which enforces the belief that it should not be ignored when modelling social networks. however, by now this study still lacks of a network analysis, observing measures such as the average path length and clustering coefficient, that will allow to evaluate if a small world network has effectively emerged. as it was demonstrated on the previous chapter, the structure of the mobility network is determinant in the spatial pattern and on the magnitude of the epidemic. the movement should always be considered in human epidemics models. another conclusion from this work is that the agent based approach is very well suited for epidemic modelling and that vector based modelling , with a appropriated programming, is quite efficient and provides a realistic representation of reality. however there is still a lot of work to be done: the movement model needs to be analysed in terms of networks, and the infection model needs to be tested, to evaluate the importance of the different parameters and, in the future, fit the characteristics of specific diseases. there is already one dataset introduced in the model and some data analysis needs to be conducted to evaluate the model efficacy. although, due to several issues, it is always is difficult to match the model results with real data, this would provide a way of validate it. finally, one of the most useful applications of a spatial model like this, will be the introduction of vaccination barriers, that will allow to study different vaccination strategies. epidemics in hierarchical social networks scaling laws for the movement of people between locations in a large city modelling the spatial dynamics and social interaction of human recreators using gis and intelligent agents individual-based lattice model for spatial spread of epidemics " in discrete dynamics in deterministic site exchange cellular automata model for the spread of diseases in human settlements epidemics and percolation in small world-networks i would like to thank my supervisor, michael batty (casa) for reviewing the presentation that originated this paper and carmo gomes (fcul) for providing me the dataset i use on these simulations. key: cord- -zyjd rmp authors: peixoto, tiago p. title: network reconstruction and community detection from dynamics date: - - journal: nan doi: . /physrevlett. . sha: doc_id: cord_uid: zyjd rmp we present a scalable nonparametric bayesian method to perform network reconstruction from observed functional behavior that at the same time infers the communities present in the network. we show that the joint reconstruction with community detection has a synergistic effect, where the edge correlations used to inform the existence of communities are also inherently used to improve the accuracy of the reconstruction which, in turn, can better inform the uncovering of communities. we illustrate the use of our method with observations arising from epidemic models and the ising model, both on synthetic and empirical networks, as well as on data containing only functional information. the observed functional behavior of a wide variety largescale system is often the result of a network of pairwise interactions. however, in many cases, these interactions are hidden from us, either because they are impossible to measure directly, or because their measurement can be done only at significant experimental cost. examples include the mechanisms of gene and metabolic regulation [ ] , brain connectivity [ ] , the spread of epidemics [ ] , systemic risk in financial institutions [ ] , and influence in social media [ ] . in such situations, we are required to infer the network of interactions from the observed functional behavior. researchers have approached this reconstruction task from a variety of angles, resulting in many different methods, including thresholding the correlation between time series [ ] , inversion of deterministic dynamics [ ] [ ] [ ] , statistical inference of graphical models [ ] [ ] [ ] [ ] [ ] and of models of epidemic spreading [ ] [ ] [ ] [ ] [ ] [ ] , as well as approaches that avoid explicit modeling, such as those based on transfer entropy [ ] , granger causality [ ] , compressed sensing [ ] [ ] [ ] , generalized linearization [ ] , and matching of pairwise correlations [ , ] . in this letter, we approach the problem of network reconstruction in a manner that is different from the aforementioned methods in two important ways. first, we employ a nonparametric bayesian formulation of the problem, which yields a full posterior distribution of possible networks that are compatible with the observed dynamical behavior. second, we perform network reconstruction jointly with community detection [ ] , where, at the same time as we infer the edges of the underlying network, we also infer its modular structure [ ] . as we will show, while network reconstruction and community detection are desirable goals on their own, joining these two tasks has a synergistic effect, whereby the detection of communities significantly increases the accuracy of the reconstruction, which in turn improves the discovery of the communities, when compared to performing these tasks in isolation. some other approaches combine community detection with functional observation. berthet et al. [ ] derived necessary conditions for the exact recovery of group assignments for dense weighted networks generated with community structure given observed microstates of an ising model. hoffmann et al. [ ] proposed a method to infer community structure from time-series data that bypasses network reconstruction by employing a direct modeling of the dynamics given the group assignments, instead. however, neither of these approaches attempt to perform network reconstruction together with community detection. furthermore, they are tied down to one particular inverse problem, and as we will show, our general approach can be easily extended to an open-ended variety of functional models. bayesian network reconstruction.-we approach the network reconstruction task similarly to the situation where the network edges are measured directly, but via an uncertain process [ , ] : if d is the measurement of some process that takes place on a network, we can define a posterior distribution for the underlying adjacency matrix a via bayes' rule where pðdjaÞ is an arbitrary forward model for the dynamics given the network, pðaÞ is the prior information on the network structure, and pðdÞ ¼ p a pðdjaÞpðaÞ is a normalization constant comprising the total evidence for the data d. we can unite reconstruction with community detection via an, at first, seemingly minor, but ultimately consequential modification of the above equation where we introduce a structured prior pðajbÞ where b represents the partition of the network in communities, i.e., b ¼ fb i g, where b i ∈ f ; …; bg is group membership of node i. this partition is unknown, and is inferred together with the network itself, via the joint posterior distribution the prior pðajbÞ is an assumed generative model for the network structure. in our work, we will use the degreecorrected stochastic block model (dc-sbm) [ ] , which assumes that, besides differences in degree, nodes belonging to the same group have statistically equivalent connection patterns, according to the joint probability with λ rs determining the average number of edges between groups r and s and κ i the average degree of node i. the marginal prior is obtained by integrating over all remaining parameters weighted by their respective prior distributions, which can be computed exactly for standard prior choices, although it can be modified to include hierarchical priors that have an improved explanatory power [ ] (see supplemental material [ ] for a concise summary.). the use of the dc-sbm as a prior probability in eq. ( ) is motivated by its ability to inform link prediction in networks where some fraction of edges have not been observed or have been observed erroneously [ , ] . the latent conditional probabilities of edges existing between groups of nodes is learned by the collective observation of many similar edges, and these correlations are leveraged to extrapolate the existence of missing or spurious ones. the same mechanism is expected to aid the reconstruction task, where edges are not observed directly, but the observed functional behavior yields a posterior distribution on them, allowing the same kind of correlations to be used as an additional source of evidence for the reconstruction, going beyond what the dynamics alone says. our reconstruction approach is finalized by defining an appropriate model for the functional behavior, determining pðdjaÞ. here, we will consider two kinds of indirect data. the first comes from a susceptible-infected-susceptible (sis) epidemic spreading model [ ] , where σ i ðtÞ ¼ means node i is infected at time t, , otherwise. the likelihood for this model is where is the transition probability for node i at time t, with fðp; σÞ ¼ ð − pÞ σ p −σ , and where m i ðtÞ ¼ p j a ij lnð − τ ij Þσ j ðtÞ is the contribution from all neighbors of node i to its infection probability at time t. in the equations above, the value τ ij is the probability of an infection via an existing edge ði; jÞ, and γ is the → recovery probability. with these additional parameters, the full posterior distribution for the reconstruction becomes since τ ij ∈ ½ ; , we use the uniform prior pðτÞ ¼ . note, also, that the recovery probability γ plays no role on the reconstruction algorithm, since its term in the likelihood does not involve a [and, hence, gets cancelled out in the denominator pðσjγÞ ¼ pðγjσÞpðσÞ=pðγÞ]. this means that the above posterior only depends on the infection events → and, thus, is also valid without any modifications to all epidemic variants susceptible-infected (si), susceptibleinfected-recovered (sir), susceptible-exposed-infectedrecovered (seir), etc., [ ] , since the infection events occur with the same probability for all these models. the second functional model we consider is the ising model, where spin variables on the nodes s ∈ f− ; g n are sampled according to the joint distribution where β is the inverse temperature, j ij is the coupling on edge ði; jÞ, h i is a local field on node i, and zða; β; j; hÞ ¼ p s expðβ p i c à ; ; otherwiseg. the value of c à was chosen to maximize the posterior similarity, which represents the best possible reconstruction achievable with this method. nevertheless, the network thus obtained is severely distorted. the inverse correlation method comes much closer to the true network, but is superseded by the joint inference with community detection. empirical dynamics.-we turn to the reconstruction from observed empirical dynamics with unknown underlying interactions. the first example is the sequence of m ¼ votes of n ¼ deputies in the to session of the lower chamber of the brazilian congress. each deputy voted yes, no, or abstained for each legislation, which we represent as f ; − ; g, respectively. since the temporal ordering of the voting sessions is likely to be of secondary importance to the voting outcomes, we assume the votes are sampled from an ising model [the addition of zero-valued spins changes eq. ( ) only slightly by replacing coshðxÞ → þ coshðxÞ]. figure shows the result of the reconstruction, where the division of the nodes uncovers a cohesive government and a split opposition, as well as a marginal center group, which correlates very well with the known party memberships and can be used to predict unseen voting behavior (see supplemental material [ ] for more details). in fig. , we show the result of the reconstruction of the directed network of influence between n ¼ twitter users from retweets [ ] using a si epidemic model (the act of "retweeting" is modeled as an infection event, using eqs. ( ) and ( ) with γ ¼ ) and the nested dc-sbm. the reconstruction uncovers isolated groups with varying propensities to retweet, as well as groups that tend to influence a large fraction of users. by inspecting the geolocation metadata on the users, we see that the inferred groups amount, to a large extent, to different countries, although clear subdivisions indicate that this is not the only factor governing the influence among users (see supplemental material [ ] for more details). conclusion.-we have presented a scalable bayesian method to reconstruct networks from functional observations that uses the sbm as a structured prior and, hence, performs community detection together with reconstruction. the method is nonparametric and, hence, requires no prior stipulation of aspects of the network and size of the model, such as number of groups. by leveraging inferred correlations between edges, the sbm includes an additional source of evidence and, thereby, improves the reconstruction accuracy, which in turn also increases the accuracy of the inferred communities. the overall approach is general, requiring only appropriate functional model specifications, and can be coupled with an open ended variety of such models other than those considered here. [ , ] for details on the layout algorithm), and the edge colors indicate the infection probabilities τ ij as shown in the legend. the text labels show the dominating country membership for the users in each group. inferring gene regulatory networks from multiple microarray datasets dynamic models of large-scale brain activity estimating spatial coupling in epidemiological systems: a mechanistic approach bootstrapping topological properties and systemic risk of complex networks using the fitness model the role of social networks in information diffusion network inference with confidence from multivariate time series revealing network connectivity from response dynamics inferring network topology from complex dynamics revealing physical interaction networks from statistics of collective dynamics learning factor graphs in polynomial time and sample complexity reconstruction of markov random fields from samples: some observations and algorithms, in approximation, randomization and combinatorial optimization. algorithms and techniques which graphical models are difficult to learn estimation of sparse binary pairwise markov networks using pseudo-likelihoods inverse statistical problems: from the inverse ising problem to data science inferring networks of diffusion and influence on the convexity of latent social network inference learning the graph of epidemic cascades statistical inference approach to structural reconstruction of complex networks from binary time series maximum-likelihood network reconstruction for sis processes is np-hard network reconstruction from infection cascades escaping the curse of dimensionality in estimating multivariate transfer entropy causal network inference by optimal causation entropy reconstructing propagation networks with natural diversity and identifying hidden sources efficient reconstruction of heterogeneous networks from time series via compressed sensing robust reconstruction of complex networks from sparse data universal data-based method for reconstructing complex networks with binary-state dynamics reconstructing weighted networks from dynamics reconstructing network topology and coupling strengths in directed networks of discrete-time dynamics community detection in networks: a user guide bayesian stochastic blockmodeling exact recovery in the ising blockmodel community detection in networks with unobserved edges network structure from rich but noisy data reconstructing networks with unknown and heterogeneous errors stochastic blockmodels and community structure in networks nonparametric bayesian inference of the microcanonical stochastic block model for summary of the full generative model used, details of the inference algorithm and more information on the analysis of empirical data efficient monte carlo and greedy heuristic for the inference of stochastic block models missing and spurious interactions and the reconstruction of complex networks epidemic processes in complex networks spatial interaction and the statistical analysis of lattice systems equation of state calculations by fast computing machines monte carlo sampling methods using markov chains and their applications asymptotic analysis of the stochastic block model for modular networks and its algorithmic applications artifacts or attributes? effects of resolution on the little rock lake food web note that, in this case, our method also exploits the heterogeneous degrees in the network via the dc-sbm, which can refinements of this approach including thouless-anderson-palmer (tap) and bethe-peierls (bp) corrections [ ] yield the same performance for this example pseudolikelihood decimation algorithm improving the inference of the interaction network in a general class of ising models the simple rules of social contagion hierarchical block structures and high-resolution model selection in large networks hierarchical edge bundles: visualization of adjacency relations in hierarchical data key: cord- -b pyg b authors: cai, li-ming; li, zhaoqing; song, xinyu title: global analysis of an epidemic model with vaccination date: - - journal: j appl math comput doi: . /s - - - sha: doc_id: cord_uid: b pyg b in this paper, an epidemic dynamical model with vaccination is proposed. vaccination of both newborn and susceptible is included in the present model. the impact of the vaccination strategy with the vaccine efficacy is explored. in particular, the model exhibits backward bifurcations under the vaccination level, and bistability occurrence can be observed. mathematically, a bifurcation analysis is performed, and the conditions ensuring that the system exhibits backward bifurcation are provided. the global dynamics of the equilibrium in the model are also investigated. numerical simulations are also conducted to confirm and extend the analytic results. mathematical models have become important tools in analyzing the spread and control of infectious diseases [ ] . based on the theory of kermack and mckendrick [ ] , the spread of infectious diseases usually can be described mathematically by compart-mental models such as sir, sirs, seir, seirs models (where s represents the class of the susceptible population, e is the exposed class in the latent period, i is infectious class, r is the removed class, which has recovered with temporary or permanent immunity). in recent years, a variety of compartmental models have been formulated, and the mathematical analysis of epidemiology models has advanced rapidly, and the analyzed results are applied to infectious diseases [ , , ] . vaccination campaigns have been critical in attacking the spread of infectious diseases, e.g., pertussis, measles, and influenza. the eradication of smallpox has been considered as the most spectacular success of vaccination [ ] . although vaccination has been an effective strategy against infectious diseases, current preventive vaccine consisting of inactivated viruses do not protect all vaccine recipients equally. the vaccine-based protection is dependent on the immune status of the recipient [ , ] . for example, influenza vaccines protect - % of the recipients among healthy young adults and as low as - % of the elderly and others with weakened immune systems (such as hiv-infected or immuno-suppressed transplant patients) (see, [ , , ] ). since vaccination is the process of administering weakened or dead pathogens to a healthy person or animal, with the intent of conferring immunity against a targeted form of a related disease agent, the individuals having the vaccine-induced immunity can be distinguished from the recovered individuals by natural immunity. thus, vaccination can also be considered by adding some compartment naturally into the basic epidemic models. over the past few decades, a large number of simple compartmental mathematical models with vaccinated population have been used in the literature to assess the impact or potential impact of imperfect vaccines for combatting the transmission diseases [ , , , , , , , , , ] . in some of these studies (e.g., papers [ , , ] ), authors have shown that the dynamics of the model are determined by the disease's basic reproduction number . if < the disease can be eliminated from the community; whereas an endemic occurs if > . therefore, if an efficient vaccination campaign acts to reduce the disease's basic reproduction number below the critical level of , then the disease can be eradicated. while in other studies, such as alexander et al. [ ] and arino et al. [ ] , they have shown that the criterion for < is not always sufficient to control the spread of a disease. a phenomenon known as a backward bifurcation is observed. mathematically speaking, when a backward bifurcation occurs, there are at least three equilibria for < in the model: the stable disease-free equilibrium, a large stable endemic equilibrium, and a small unstable endemic equilibrium which acts as a boundary between the basins of attraction for the two stable equilibria. in some cases, a backward bifurcation leading to bistability can occur. thus, it is possible for the disease itself to become endemic in a population, given a sufficiently large initial outbreak. these phenomena have important epidemiological consequences for disease management. in recent years, backward bifurcation, which leads to multiple and subthreshold equilibria, has been attracting much attention (see, [ , , , , , , , , , , , , , , ] ). several mechanisms with vaccination have been identified to cause the occurrence of backward bifurcation in paper [ ] . in this paper, we shall investigate the effects of a vaccination campaign with an imperfect vaccine upon the spread of a non-fatal disease, such as hepatitis a, hepatitis b, tuberculosis and influenza, which features both exposed and infective stages. in particular, we focus on the vaccination parameters how to change the qualitative behavior of the model, which may lead to subthreshold endemic states via backward bifurcation. global stability results for equilibria are obtained. the model constructed in this paper is an extension of the model in paper [ ] , including a new compartment for the latent class (an important feature for the infectious diseases eg. hepatitis a, hepatitis b, tuberculosis and influenza) and the disease cycle. it is one of the aims of this paper to strengthen the disease cycle to cause multiple endemic equilibria. the paper is organized as follows. an epidemic model with vaccination of an imperfect vaccine is formulated in sect. , and the basic reproduction number, and the existence of backward bifurcation and forward bifurcation are analyzed in sect. . the global stability of the endemic equilibrium is established in sect. . the paper is concluded with a discussion. in order to derive the equations of the mathematical model, we divide the total population n in a community into five compartments: susceptible, exposed (not yet infectious), infective, recovered, and vaccinated; the numbers in these states are denoted by s(t), the flow diagram of the disease spread is depicted in fig. . all newborns are assumed to be susceptible. of these newborns, a fraction α of individuals are vaccinated, where α ∈ ( , ]. susceptible individuals are vaccinated at rate constant ψ. the parameter γ is the rate constant at which the exposed individuals become infectious, and γ is the rate constant that the infectious individuals become recovered and acquire temporary immunity. finally, since the immunity acquired by infection wanes with time, the recovered individuals have the possibility γ of becoming susceptible again. β is the transmission coefficient (rate of effective contacts between susceptible and infective individuals per unit time; this coefficient includes rate of contacts and effectiveness of transmission). since the vaccine does not confer immunity to all vaccine recipients, vaccinated individuals may become infected but at a lower rate than unvaccinated (those in class s). thus in this case, the effective contact rate β is multiplied by a scaling factor σ ( ≤ σ ≤ describes the vaccine efficacy, σ = represents vaccine that offers % protection against infection, while σ = models a vaccine that offers no protection at all). it is assumed that the natural death rate and birth rate are μ and the disease-induced death rate is ignored. thus the total population n is constant. since the model consider the dynamics of the human populations, it is assumed that all the model parameters are nonnegative. thus, the following model of differential equations is formulated based on the above assumptions and fig. , with nonnegative initial conditions and n ( ) > . system ( . ) is well posed: solutions remain nonnegative for nonnegative initial conditions. we illustrate here that there are limiting cases in system ( . ): if σ = , the vaccine is perfectly effective, and α = ψ = , there is no vaccination, system ( . ) will be reduced to the standard seirs model in [ ] ; if γ = and the limit γ → ∞, system ( . ) will be equivalent to an svir model in [ ] . if we let α = and γ = , system ( . ) can be reduced to an sveir epidemic model in [ ] , where authors aim to assess the potential impact of a sars vaccine via mathematical modelling. to explore the effect of the vaccination period and the latent period on disease dynamics, an sveir epidemic model with ages of vaccination and latency are formulated in paper [ ] . in papers [ , , , ] , authors have shown that the dynamics of the model are determined by the disease's basic reproduction number . that is, the disease free equilibrium is globally asymptotically stable for ≤ ; and there is a unique endemic equilibrium which is globally asymptotically stable if > . if ψ = and limit γ → ∞, system ( . ) will be reduced into an siv epidemic model in [ ] , where authors investigate the effect of imperfect vaccines on the disease's transmission dynamics. in [ ] , it is shown that reducing the basic reproduction number to values less than one no longer guarantees disease eradication. in this paper, we show that if a vaccination campaign with an imperfect vaccine and the disease cycle is considered, a more complicated dynamic behavior is observed in system ( . ). for example, the backward bifurcation occurs in system ( . ). in the following, first, it is easy to obtain that the total population n in system ( . ) is constant. to simplify our notation, we define the occupation variable of compartments s, e, i, v, and r as the respective fractions of a constant population n that belong to each of the corresponding compartments. we still write the occupation variable of compartments as s, e, i, v and r, respectively. thus, it is easy to verify that is positively invariant and globally attracting in r + . it suffices to study the dynamics of ( . ) on d. thus, system ( . ) can be rewritten as the following system: in the case σ = , system ( . ) reduces to an seirs model without vaccination [ ] , , is considered as the basic reproduction number of the model. the classical basic reproduction number is defined as the number of secondary infections produced by a single infectious individual during his or her entire infectious period. mathematically, the reproduction number is defined as a spectral radius r (which is a threshold quantity for disease control) that defines the number of new infectious generated by a single infected individual in a fully susceptible population [ ] . in the following, we shall use this approach to determine the reproduction number of system ( . ). it is easy to see that system ( . ) has always a disease-free equilibrium, ) can be rewritten as the jacobian matrices of f(x) and v(x) at the disease-free equilibrium p are, respectively, f v − is the next generation matrix of system ( . ). it follows that the spectral radius according to theorem in [ ] , the basic reproduction number of system ( . ) is the basic reproduction number r vac can be interpreted as follows: a proportion of γ μ+γ of exposed individuals progress to the infective stage before dying; μ+γ represents the number of the secondary infection generated by an infective individual when he or she is in the infectious stage. those newborns vaccinated individuals have generated the number μ+ψ of the secondary infection. average vaccinated individuals with vaccination rate ψ have generated the fraction of the secondary infection. now we investigate the conditions for the existence of endemic equilibria of system ( . ). any equilibrium (s, v, e, i, r) of system ( . ) satisfies the following equations: from the second and third equation of ( . ), we have hence, there exists no endemic equilibrium for r ≤ . for r > , the existence of endemic equilibria is determined by the presence in ( , ] of positive real solutions of the quadratic equation where, . ) and ( . ), we can see that the number of endemic equilibria of system ( . ) is zero, one, or two, depending on parameter values. for σ = (the vaccine is totally effective), it is obviously that there is at most one endemic equilibrium (p * (s * , e * , i * , r * , v * )) in system. from now on we make the realistic assumption that the vaccine is not totally effective, and thus < σ < . we notice that if r vac = , then we have since all the model parameters are positive, it follows from thus, r vac is a continuous decreasing function of ψ for ψ > , and if ψ < ψ crit , then r vac > and c < . therefore, it follows that p(i ) of eq. ( . ) has a unique positive root for r vac > . now we consider the case for r vac < . in this case, c > , and ψ ≥ ψ crit . from ( . ), it is easy to see that b(ψ) is an increasing function of ψ. thus, if b(ψ crit ) ≥ , then b(ψ) > for ψ > ψ crit . thus, p(i ) has no positive real root which implies system have no endemic equilibrium in this case. thus, let us consider the case is an linear increasing function of ψ. thus, there is a uniqueψ > ψ crit such that b(ψ) = , and thus (ψ) < . since (ψ) is a quadratic function of ψ with positive coefficient for ψ , (ψ) has a unique rootψ in [ψ crit ,ψ]. thus, for r vac < we have b(ψ) < , a > , c ≥ , and (ψ) > for ψ ∈ (ψ crit ,ψ). therefore, p(i ) has two possible roots and system from the above discussion, we have b(ψ) > for ψ >ψ, and (ψ) < for ψ ∈ (ψ,ψ). therefore, it follows that system ( . ) has no endemic equilibria for ψ >ψ. if r vac = , we have c = . in this case, system has a unique endemic equilibrium for b(ψ) < and no endemic equilibrium for b(ψ) > . summarizing the discussion above, we have the following theorem: for ψ crit < ψ <ψ and has no endemic equilibria for ψ >ψ; according to theorem of van den driesche and watmough [ ] , we have the following result. the disease-free equilibrium p is locally asymptotically stable when r vac < and unstable when r vac > . in the following, we first give a global result of the disease-free equilibrium of system ( . ) under some conditions. by directly calculating the derivative of l along system ( . ) and notice that s +σ v < , thus, we have it is easy to verify that the maximal compact invariant set in {(s, e, i, r, v ) ∈ : the global stability of p follows from the lasalle invariance principle [ ] . from the above discussion, we know that system ( . ) may undergo a bifurcation at the disease-free equilibrium when r vac = . now we establish the conditions on the parameter values that cause a forward or backward bifurcation to occur. to do so, we shall use the following theorem whose proof is found in castillo-chavez and song [ ] , which based on the use of the center manifold theory [ ] . for the following general system with a parameter φ. without loss of generality, it is assumed that x = is an equilibrium for system ( . ) for all values of the parameters φ, that is, f ( , φ) = for all φ. then the local dynamics of system ( . ) around x = are totally determined by a and b. (i) a > , b > . when φ < with |φ| , x = is locally asymptotically stable and there exists a positive unstable equilibrium; when < φ , x = is unstable and there exists a negative and locally asymptotically equilibrium; (ii) a < , b < . when φ < , with |φ| , x = is unstable; when < φ , x = is locally asymptotically stable and there exists a negative unstable equilibrium; (iii) a > , b < .when φ < , with |φ| , x = is unstable and there exists a locally asymptotically stable negative equilibrium; when < φ , x = is stable and a positive unstable equilibrium appears; (iv) a < , b > . when φ changes from negative to positive, x = changes its stability from stable to unstable. correspondingly, a negative unstable equilibrium becomes positive and locally asymptotically stable. now by applying theorem . , we shall show system ( . ) may exhibit a forward or a backward bifurcation when r vac = . consider the disease-free equilibrium p = (s , , , ) and choose β as a bifurcation parameter. solving r vac = gives let j denote the jacobian of the system ( . ) evaluated at the dfe p with β = β * . by directly computing, we have it is easy to obtain that j (p , β * ) has eigenvalues given by thus, λ = is a simple zero eigenvalue of the matrix j (p , β * ) and the other eigenvalues are real and negative. hence, when β = β * , the disease free equilibrium p is a non-hyperbolic equilibrium. thus, assumptions (a ) of theorem . is verified. now, we denote with ω = (ω , ω , ω , ω ), a right eigenvector associated with the zero eigenvalue λ = . thus, thus, we have from the above, we obtain that v = , γ μ + γ , , . let a and b be the coefficients defined as in theorem . . computation of a, b. for system ( . ), the associated non-zero partial derivatives of f (evaluated at the dfe (p ), x = s, x = i, x = e, x = r.) are given by ( . ) since the coefficient b is always positive, according to theorem . , it is the sign of the coefficient a, which decides the local dynamics around the disease-free equilibrium p for β = β * . if the coefficient a is positive, the direction of the bifurcation of system ( . ) at β = β * is backward; otherwise, it is forward. thus, we formulate a condition, which is denoted by (h ) : thus, if (h ) holds, we have a > , otherwise, a < . summarizing the above results, we have the following theorem. fig. the forward bifurcation diagram for model ( . ) value of ψ say ψ * such that backward bifurcation occurs of ψ < ψ * and forward bifurcation occurs if ψ > ψ * . both of these bifurcation diagrams are obtained by considering β as bifurcation parameter and later it is plotted with respect to r vac . in this section, we shall investigate the global stability of the unique endemic equilibrium for r vac > . here we shall apply the geometric approach [ , , ] to establish the global stability of the unique endemic equilibrium. in recent years, many authors [ , , , , ] have applied this method to show global stability of the positive equilibria in system. here, we follow the techniques and approaches in paper [ , ] to investigate global stability of the endemic equilibrium in system ( . ). here, we omit the introduction of the general mathematical framework of these theorems and only focus on their applications. in the previous section, we have showed that if r vac > , system ( . ) has a unique endemic equilibrium in d. furthermore, r vac > implies that the disease-free equilibrium p is unstable (theorem . ) . the instability of p , together with p ∈ ∂d, implies the uniform persistence of the state variables. this result can be also showed by using the same arguments from proposition . in [ ] and proposition . in [ ] . thus, we first give the following result: to prove our conclusion, we set the following differential equatioṅ where f : d(⊂ r n ) → r n , d is open set and simply connected and f ∈ c (r n ). letμ where, p(x) be a nonsingular ( ) matrix, the second additive compound of the jacobian matrix ∂ f /∂ x. μ is the lozinskiȋ measure with respect to a vector norm | · |. the following result comes from corollary . in paper [ ] . from proposition . , it is easy to verify that the condition (i) in theorem . holds. therefore, to prove our conclusion, we only verify that (ii) in theorem . holds. according to paper [ ] , the lozinskiȋ measure in theorem . can be evaluated as follow:μ where d + is the right-hand derivative. now we state our main result in this section. then the unique equilibrium p * in system ( . ) is globally asymptotically stable for r vac > . then, the jacobian matrix of system ( . ) can be written as the second additive compound [ ] (see, "appendix") of jacobian matrix is the × matrix given by where p f is the derivative of p in the direction of the vector field f . thus, we have p f p − = −diag(Ė/e,Ė/e,Ė/e,İ /i,İ /i,İ /i ). thus, we obtain that as in [ , ] , we define the following norm on r : where z ∈ r , with components z i , i = , . . . , and and let now we demonstrate the existence of some κ > such that by linearity, if this inequality is true for some z, then it is also true for −z. similar to analyzing methods in paper [ , ] , our proof is subdivided into eight separate cases, based on the different octants and the definition of the norm ( . ). to facilitate our analysis, we use the following inequalities: u (t) ≥ |z |, |z |, |z |, |z + z |, |z + z + z |, case . let u (z) > u (z),z , z , z > and |z | > |z | + |z |. then we have ||z|| = z and u (z) < z . taking the right derivative of ||z||, we have ( . ) case . by linearity, eq. ( . ) also holds for u > u and z , z , z < when |z | < |z | + |z |. thus, if we require that γ < γ + μ holds, then the inequality ( . ) holds for case and case . and |z | > |z |. thus, we have ||z|| = |z | + |z |, and u (z) < |z | + |z |. by directly calculating, we obtain that using the inequalities , |z |, |z + z + z | ≤ u (z) < |z | + |z |, and |z | ≤ |z |, we have ( . ) case . by linearity, eq. ( . ) also holds for u > u and z , thus, if we require that γ < γ + μ holds, then the inequality ( . ) holds for case and case . therefore, from the discussion above, we know that if inequalities ( . )hold, then there exists κ > such that d + ||z|| ≤ −κ||z|| for all z ∈ r and all nonnegative s, v, e and i . all conditions in theorem . can be satisfied when inequalities ( . ) hold. therefore, by theorem . , we can determine that if inequalities ( . ) hold, then the unique endemic equilibrium of system ( . ) is globally stable in d for r vac > . remark in sect. , we have shown that system ( . ) exhibit a backward bifurcation for r vac ≤ . as stressed in [ ] , for cases in which the model exhibits bistability, the compact absorbing set required in theorem . does not exist. by applying similar methods in [ ] , a sequence of surfaces that exists for time > and minimizes the functional measuring surface area may be obtained. therefore, the global dynamics of system ( . ) in the bistability region can be further investigated as it has been done in paper [ ] . in this paper, an epidemic model with vaccination has been investigated. by analysis, it is showed that the proposed model exhibits a more complicated dynamic behavior. backward bifurcation under the vaccination level conditions, and bistability phenomena can be observed. the global stability of the unique endemic equilibrium in the model is demonstrated for r vac > . note that the model ( . ) can be solved in an efficient way by means of the multistage adomian decomposition method (madm) as a relatively new method [ , , , ] . the madm has some superiority over the conventional solvers such as the r-k family. to illustrate the various theoretical results contained in this paper, the effect of some important parameter values on the dynamical behavior of system ( . ) is investigated in the following. now we consider first the role of the disease cycle on the backward bifurcation. if γ = , [i.e., the disease cycle-free in model ( . )], then the expression for the bifurcation coefficient, a, given in eq. ( . ) reduces to thus, the backward bifurcation phenomenon of system ( . ) will not occur if γ = . this is in line with results in papers [ , ] , where the disease cycle-free model ( . ) has a globally asymptotically stable disease-free equilibrium if the basic reproduction number is less than one. differentiating a, given in eq. ( . ) , with respect to γ gives hence, the bifurcation coefficient, a is an increasing functions of γ . thus, the feasibility of backward bifurcation occurring increases with disease cycle. now we consider the role of vaccination on the backward bifurcation. let α = ψ = σ = , then the expression for the bifurcation coefficient, a, given in eq. ( . ) , is reduces to thus, the backward bifurcation phenomenon of system ( . ) will not occur if α = ψ = σ = (i.e., the model ( . ) will not undergo backward bifurcation in the absence of vaccination). this is also in line with results in paper [ ] , where the vaccination-free model ( . ) has a globally asymptotically stable equilibrium if the basic reproduction number r is less than one. furthermore, the impact of the vaccine-related parameters (ψ, σ ) on the backward bifurcation is assessed by carrying out an analysis on the bifurcation coefficient a as follows. differentiating a, given in eq. ( . ), partially with respect to ψ, gives thus, the backward bifurcation coefficient, a is a decreasing function of the vaccination rate ψ. hence, the possibility of backward occurring decreases with increasing vaccination rate ( i.e., vaccinating more susceptible individuals decrease the likelihood of the occurrence of backward bifurcation). differentiating the bifurcation coefficient a, given in eq. ( . ), partially with respect to σ gives ∂a ∂σ = β * γ μ + γ m , with m = − γ γ (μ + γ )(μ + ψ) + μ( − α)(μ + γ )(μ + γ )(μ + ψ + σ (μα + ψ)) γ (μ + ψ)[μ( − α) + σ (μα + ψ)] + μ + γ γ + γ μ + γ + . thus, the bifurcation coefficient, a is a decreasing function with respect to σ . that is, the likelihood of backward bifurcation occurring decreases with increasing vaccine efficacy. let α = . , μ = . , β = . , ψ = . , γ = . , γ = . , γ = . . by direct calculating, it is easy to verify that m is negative and also condition (h ) is satisfied. figure depicts the backward bifurcation occurring phenomena with lower vaccine efficacy with σ = . ; fig. depicts the likelihood of backward bifurcation occurring with higher vaccine efficacy σ = . . in addition, it is obvious that our expression for the basic reproduction number in system ( . ), i.e., r vac = βγ (μ + γ )(μ + γ ) μ( − α) + σ (μα + ψ) μ + ψ is independent of the loss rate of immunity γ . from the above analysis, we have found that the dynamics of the model are not determined by the basic reproduction number, and the phenomena of the backward bifurcation in system may occur. moreover, it is found that the occurrence feasibility is increasing with the loss rate of immunity γ . from the following expression, it is easy to see that the policy of vaccinations with imperfect vaccines can decrease the the basic reproduction number r vac . thus, the imperfect vaccine may be beneficial to the community. this is also a positive point, sice it is know that the use of some imperfect vaccine can sometime result in detrimental consequences to the community [ , ] . at last, we must point out that although the system ( . ) with ( . ) is well posed mathematically, we acknowledge the biological reality that the fraction of the constant total population which occupies a compartment can only be within the subset q of rational values within r + , and furthermore only within a sub-subset of values within q belonging to n/n where n belongs to the integers z ∈ [ , n ]. in addition, we also point out that the analysis of the model ( . ) may become somewhat different if disease fatalities and more complex vital dynamics are included, in particular, if the population size is no longer constant. in the future, we may investigate many various modeling possibilities to simulate a real world biological process based on model ( . ). on the other hand, we note that the population in our model ( . ) is assumed to be homogeneously mixed. in fact, different individual may have different number of contacts. thus, a complex network-based approach on diseases transmission may be closer to a realistic situation [ , , ] . in the future, we shall investigate dynamics of the proposed model based on a complex network. a vaccination model for transmission dynamics of influenza infectious diseases of humans global results for an epidemic model with vaccination that exhibits backward bifurcation on the dynamics of an seir epidemic model with a convex incidence rate dynamical models of tuberculosis and their application dynamical behavior of an epidemic model for a vector-borne disease with direct transmission global stability of an epidemic model with carrier state in heterogeneous networks a new modified adomian decomposition method and its multistage form for solving nonlinear boundary value problems with robin boundary conditions a reliable algorithm for positive solutions of nonlinear boundary value problems by the multistage adomian decomposition method global stability of an sveir epidemic model with ages of vaccination and latency theoretical assessment of public health impact of imperfect prophylactic hiv- vaccines with with therapeutic benefits analytical approximate solutions for a general nonlinear resistor-nonlinear capacitor circuit model chaos control in the cerium-catalyzed belousov-zhabotinsky reaction using recurrence quantification analysis measures influenza vaccine efficacy in young, healthy adults nonlinear oscilations, dynamical systems, and bifurcations of vector fields an sveir model for assessing potential impact of an imperfect anti-sars vaccine backward bifurcation in epidemic control the mathematics of infectious diseases a contribution to the mathematical theory of epidemics. part i vaccination strategies and backward bifurcation in an age-sinceinfection structured model a simple vaccination model with multiple endemic states the stability of dynamical systems, regional conference series in applied mathematics global dynamics of vector-borne diseases with horizontal transmission in host population global anlysis of sis epidmeic model with a simple vaccination and multiple endemic equilibria on r.a. smiths autonomous convergence theorem. rocky mount global stability for the seir model in epidemiology a geometric approach to global-stability problems global stability of seirs models in epidemiology global dynamics of an seir epidemic model with vertical transmission new vaccine against tuberculosis: current developments and future challenges svir epidemic models with vaccination strategies modeling and dynamics of infectious diseases on the mechanism of strain replacement in epidemic models with vaccination, in current developments in mathematical biology progression age enhanced backward bifurcation in an epidemic model with super-infection logarithmic norms and projections applied to linear differential systems modelling the effect of imperfect vaccines on disease epidemiology global results for an sirs model with vaccination and isolation some application of hausdorff dimension inequalities for ordinary differential equations reproduction numbers and sub-threshold endemic equilibria for compartmental models of disease transmission backward bifurcation of an epidemic model with treatment global analysis of an sis model with an infective vector on complex networks global dynamics of a network epidemic model for waterborne diseases spread global attractivity and permanence of a sveir epidemic model with pulse vaccination and time delay who advisory committee on variola virus research report of the fourteenth meeting the periodic solution of a class of epidemic acknowledgements we would like to thank dr chin-hong park( editor-in-chief) and the four reviewers for their constructive comments and suggestions that have helped us to improve the manuscript significantly. the second additive compound matrix a [ ] for a × matrix a = (a i j ) is key: cord- - v ltunv authors: dungan, r. s. title: board-invited review: fate and transport of bioaerosols associated with livestock operations and manures date: - - journal: j anim sci doi: . /jas. - sha: doc_id: cord_uid: v ltunv airborne microorganisms and microbial by-products from intensive livestock and manure management systems are a potential health risk to workers and individuals in nearby communities. this report presents information on zoonotic pathogens in animal wastes and the generation, fate, and transport of bioaerosols associated with animal feeding operations and land applied manures. though many bioaerosol studies have been conducted at animal production facilities, few have investigated the transport of bioaerosols during the land application of animal manures. as communities in rural areas converge with land application sites, concerns over bioaerosol exposure will certainly increase. although most studies at animal operations and wastewater spray irrigation sites suggest a decreased risk of bioaerosol exposure with increasing distance from the source, many challenges remain in evaluating the health effects of aerosolized pathogens and allergens in outdoor environments. to improve our ability to understand the off-site transport and diffusion of human and livestock diseases, various dispersion models have been utilized. most studies investigating the transport of bioaerosols during land application events have used a modified gaussian plume model. because of the disparity among collection and analytical techniques utilized in outdoor studies, it is often difficult to evaluate health effects associated with aerosolized pathogens and allergens. invaluable improvements in assessing the health effects from intensive livestock practices could be made if standardized bioaerosol collection and analytical techniques, as well as the use of specific target microorganisms, were adopted. animal feeding operations (afo) generate vast quantities of manure (feces and urine) and wastewater that must be treated, stockpiled, or beneficially used. in the united states there are approximately , afo producing an estimated million wet tons of manure annually. of particular concern is the intensification of animal production, which has led to the creation of concentrated afo (cafo) that make up about % of all afo. the major producers of manure are cattle (beef and dairy), poultry (chicken and turkey), and swine operations (wright et al., ) . depending upon the animal production facility, the solid and liquid manures are typically stored in piles or holding ponds, mechanically dewatered, composted, anaerobically digested for biogas production, or a combination of the above. animal manures applied as solids, semi-solids, and liquids have traditionally been used as soil conditioners and as a source of nutrients for crop production (power and dick, ; risse et al., ) . when improperly managed, however, manures can pollute surface and ground waters with nutrients and pathogenic microorganisms (ritter, ) . because commercial livestock carry an increased microbial load in their gastrointestinal system, they are often reservoirs of zoonotic pathogens (temporarily or permanently), which can be transmitted to the environment in untreated manures (gerba and smith, ; venglovsky et al., ). an area of growing interest is airborne pathogens and microbial by-products generated at afo and during the land application of manures (chang et al., b; wilson et al., ; cole et al., ; chinivasagam et al., ; dungan and leytem, a; millner, ) , which can potentially affect the health of livestock, farm workers, and individuals in nearby residences (heederik et al., ) . land application of untreated solid and semi-solid manures and use of pressurized irrigation systems to apply liquid manures and wastewaters increase the chances that microorganisms will become aerosolized (teltsch et al., a; brooks et al., ; hardy et al., ; peccia and paez-rubio, ) . despite the potential for bioaerosol formation during these activities, very few research papers have addressed the risk of human exposure to pathogens during the land application of animal wastes (boutin et al., ; murayama et al., ) . to date, much of the research in this area has been conducted with municipal wastewaters (us epa , tanner et al., ; peccia and paez-rubio, ) and biosolids (dowd et al., ; brooks et al., a,b; tanner et al., ) . considering the fact that the number of cafo continues to grow (usda national agricultural statistics service, ), along with a growing farm worker and encroaching civilian population, an increased understanding of the fate and transport of airborne microorganisms is required to ensure public health is not compromised. the purpose of this review is to highlight the current knowledge of bioaerosol fate and transport, with a specific focus on bioaerosols generated at afo and during the land application of animal manures. readers seeking more information on bioaerosol collection and analytical methodologies should refer to a recent review by dungan and leytem ( b) . additional emphasis is placed on dispersion models as a means to assess the transport of bioaerosols and subsequent risk of exposure to individuals in the downwind plume. domesticated livestock harbor a variety of bacterial, viral, and protozoal pathogens, some of which pose a risk to other animals and humans. infectious diseases that are transmissible from animals to humans and vice versa are known as zoonoses. these diseases can be transmitted to humans through direct contact (skin wounds, mucous membranes), fecal-oral route, ingestion of contaminated food and water, or aerogenic route (e.g., droplets, dust). tables , , and present a list of important bacterial, viral, and protozoal zoonotic pathogens associated with animals and their wastes, respectively. many of these pathogens are endemic in commercial livestock and, therefore, are difficult to eradicate from both the animals and production facilities. some well-recognized zoonotic pathogens are escherichia coli o :h , salmonella spp., campylobacter jejuni, apthovirus that causes foot-and-mouth disease (fmd), and protozoal parasites such as cryptosporidium parvum and giardia lamblia. this section is not meant to be an exhaustive review of zoonotic pathogens; more detailed information on zoonoses can be found in krauss et al. ( ) and sobsey et al. ( ) . escherichia coli are native inhabitants of the gastrointestinal tract of mammals, but a subset of diarrhetic e. coli, known as enterohemorrhagic, enteropathogenic, (krauss et al., ) . salmonella occur in cattle, pigs, poultry, wild birds, pets, rodents, and other animals; however, only nontyphoidal salmonella (e.g., s. enterica serovar enteritidis) occurs in both humans and animals. human infection generally occurs through the ingestion of contaminated foodstuffs or excretions from sick or infected animals, resulting in acute gastroenteritis. campylobacter jejuni is among the most common causes of diarrheal disease in the united states, and this is attributed to the relatively low infectious dose (< organisms). the main reservoirs of c. jejuni are wild birds and poultry, although among farm animals pigs are important carriers. infection in humans occurs by ingestions of contaminated food (raw or undercooked poultry meat, pork, or milk) or water or by direct contact with contaminated feces. foot-and-mouth disease is a highly contagious and sometimes fatal viral disease of cloven-hoofed animals (domestic and wild). human infections with the fmd virus are rare and infections can usually be traced to direct handling of infected animals or contact during slaughter. cryptosporidium parvum is a protozoal parasite that is widespread in mammals and is increasingly recognized as a major cause of human diarrhea. in animals, clinical signs are most commonly observed in newborn calves. infected animals shed the organism in their feces, and human infection occurs though the ingestion of contaminated food and water. giardiasis, caused by various giardia spp. (e.g., g. lamblia), is considered one of the most prevalent parasitic infections in the world, especially in developing nations with poor sanitary practices. animal hosts of giardia spp. include cattle, sheep, pigs, cats, rodents, and other mammals, which are direct or indirect sources of human infection. transmission commonly occurs through the ingestion of food or water contaminated with feces. although the common route of transmission for many zoonotic pathogens is direct ingestion or contact, the inhalation of infectious particles should also be considered. it is well documented that communicable and noncommunicable human diseases are transmitted through airborne routes; however, the airborne transmission of some of the above-mentioned zoonotic pathogens is unknown and quite controversial. zoonotic pathogens, such as mycobacterium tuberculosis and hantavirus, are krauss et al. ( ) and sobsey et al. ( ) . known to be transmitted through aerogenic routes and are capable of causing severe disease in infected individuals (sobsey et al., ) . however, some enteric pathogens (e.g., salmonella spp.) are not typically associated with aerogenic routes of exposure, but based on studies with animals there is evidence suggesting that airborne transmission is possible (wathes et al., ; harbaugh et al., ; oliveira et al., ) . furthermore, there is much uncertainty associated with the dose-response of airborne pathogens and biological agents because many relationships have not been established to date (pillai and ricke, ; douwes et al., ; hermann et al., ). although the land application of manures is often utilized as a means to dispose of a waste by-product, rather than from a beneficial use perspective, manures are an excellent source of major plant nutrients such as nitrogen, phosphorus, and potassium, as well as some secondary nutrients. the application of manure not only improves soil nutrient status, but also has a significant effect on physical and biological properties (sommerfeldt and chang, ; khaleel et al., ; peacock et al., ) . manure applications increase the om content in soils, which in turn promotes the formation of water-stable soil aggregates and improves water infiltration, water-holding capacity, microbial activity, and overall productivity. to distribute the livestock manures and wastewaters to agricultural fields a variety of techniques are often utilized (pfost et al., ) . manures with a low moisture content, such as chicken litter or dewatered feces, can be land-applied using a manure slinger or spreader. wastes that have a very low solids content, such as wastewater from flush systems, holding ponds, or lagoons, can be land applied via furrow irrigation, directly injected (e.g., drag-hose), or sprayed using a tanker or pressurized irrigation systems (e.g., spray gun, center-pivot). application methods that launch liquid and solid manures into the air create a potentially hazardous situation as pathogens may become aerosolized and transported to downwind receptors (sorber and guter, ; brooks et al., ) . the aerosolized pathogens could potentially be directly inhaled or ingested after they land on fomites, water sources, or food crops. aerosolization is a process where fine droplets evaporate completely or to near dryness; thus, microorganisms in these droplets are transformed into solid or semi-solid particles (i.e., bioaerosols). during spray irrigation events of liquid manures and wastewaters, the water stream is broken up into droplets of various sizes. the size of the droplets is related to the sprinkler head configuration and operating pressure of the irrigation system. fine droplets, < μm in diameter, evapo-rate relatively quickly, whereas those > μm do not evaporate appreciably (hardy et al., ) . however, the evaporation rate of water droplets increases with decreasing humidity and increasing temperature. in a study conducted with low pressure sprinklers, total evaporation losses ranged from . to . % for smooth spray plate and . to . % for coarse serrated sprinklers (kohl et al., ) . in a us epa report ( ), the aerosolization efficiency (e) ranged from . to . %, with a median value of . % over spray irrigation events using rotating impact-sprinklers. aerosolization efficiency is the fraction of the total water sprayed that leaves the vicinity of the irrigation system as an aerosol, rather than as droplets. bioaerosols are viable and nonviable biological particles, such as bacteria, virus, fungal spores, and pollen grains and their fragments and by-products (e.g., endotoxins, mycotoxins), that are suspended in the air (grinshpun et al., ) . airborne microorganisms and their components are generated as a mixture of droplets or particles, having different aerodynamic diameters ranging from . to μm (lighthart, ; cox and wathes, ) . the generation of bioaerosols from water sources occurs during bubble bursting or splash, and wave action and microorganisms (single cells or groups) are usually surrounded by a thin layer of water . aside from natural activities, land spreading of slurries, pressurized spray irrigation events, and aeration basins at wastewater treatment plants are a few ways microorganisms become aerosolized. bioaerosols generated directly from relatively dry surfaces (e.g., feedlots, soils, plants) or during the land application of dry manures can be released as individual or groups of cells or associated with inorganic or organic particulate matter (cambra-lópez et al., ) . aerosol particles to μm in diameter are of the greatest concern because they are readily inhaled or swallowed, but the greatest retention in the lung alveoli occurs with the -to -μm particles (salem and gardner, ) . unlike microorganisms in soils, waters, and manures, aerosolized or airborne microorganisms are very susceptible to a variety of meteorological factors (cox and wathes, ) . the most significant factors that affect viability are relative humidity, temperature, and solar irradiance (table ). in general, laboratory and field studies have shown that microorganism viability decreases with decreases in relative humidity and increases in temperature and solar irradiance (poon, ; dimmock, ; ehrlich et al., b; goff et al., ; theunissen et al., ; lighthart and shaffer, ) . as relative humidity decreases, there is less water available to the microorganisms, which causes dehydration and subsequent inactivation of many microorganisms. however, because temperature influences relative humidity, it is often difficult to separate their effects (mohr, ) . targets of relative humidity-and temperature-induced inactivation of airborne microorganisms appear to be proteins and membrane phospholipids (cox and wathes, ) . viruses with structural lipids are stable at low relative humidities, whereas those without lipids are more stable at high relative humidities. oxygen concentration is also known to affect bacterial survival because it is involved in the inactivation of bioaerosols through the production of free radicals of oxygen (cox and baldwin, ; cox et al., ) . because bacteria are much more complex, biochemically and structurally, than viruses, viruses tend to be more resistant to the effects of oxygen and temperature-induced inactivation, except in the case of spore-forming bacteria such as clostridium spp. (mohr, ) . inactivation of bioaerosols by solar irradiance is highly dependent upon wavelength and is exacerbated by dehydration and oxygen (beebe, ; riley and kaufman, ; cox and wathes, ; ko et al., ) . short-wavelength ionizing radiation (e.g., x-rays, gamma rays, uv) induces free-radical-mediated reactions that cause damage to biopolymers, such as nucleic acids and proteins. another factor, known as the open-air factor, is based on the fact that the survival of many outdoor airborne microorganisms is generally poorer than in inside air under similar conditions (cox and wathes, ) . this effect was attributed to ozoneolefin reaction products in the outdoors. whereas the above-mentioned factors influence viability, microbial factors such as the type, genus, species, and strain of an organism also affect its airborne survival (songer, ; ehrlich et al., b) . microorganisms associated with droplets that evaporate to dryness or near-dryness before impacting the ground or vegetation are transported in air currents. when bioaerosols are released from a source, they can be transported short or long distances and are eventually deposited in terrestrial and aquatic environments (brown and hovmøller, ; jones and harrison, ; griffin, ) . the transport, behavior, and deposition of bioaerosols are affected by their physical properties (i.e., size, shape, and density) and meteorological factors they encounter while airborne. because most bioaerosols are not perfectly spherical, the most useful size definition is aerodynamic diameter, which is the major factor controlling their airborne behavior (kowalski, ) . aerodynamic diameter is defined as the diameter of a spherical particle of water (a unit density sphere) with which a bioaerosol or microorganism has the same settling velocity in air. meteorological factors such as wind velocity, relative humidity, temperature, and precipitation affect the transport of bioaerosols, with atmospheric stability being a major factor (lighthart and mohr, ; lighthart, ; jones and harrison, ). relative humidity not only affects microorganism viability as discussed above, but also affects settling velocity because it directly influences the density and aerodynamic diameter of the bioaerosol unit (ko et al., ; mohr, ) . the deposition of bio- aerosols occurs through gravitational settling, impaction, diffusion onto surfaces, and wash-out by raindrops (muilenberg, ) . for particles with an aerodynamic diameter > μm, gravitational settling and impaction are the leading causes of particle loss during transport (mohr, ) . for larger airborne particles (> μm), removal by raindrops is quite efficient. assessment of bioaerosol transport is generally accomplished by setting liquid impingement or solid impaction systems at an upwind location (background) and various downwind distances from the source (dungan and leytem, b) . in brief, the aerosol samplers are usually set at . m above the ground, which corresponds to the average breathing height for humans. air is then pulled through the samplers at a specified flow rate (e.g., . l·min − for glass impingers) for several minutes to hours using a vacuum pump. samples are then analyzed via culture-dependent or molecularbased (e.g., pcr) assays or microscopically to calculate a microorganism concentration per cubic meter of air. in the case of airborne endotoxins, samples are typically collected on filters, subsequently extracted using a weak tween solution, and analyzed using the kinetic limulus amebocyte lysate assay (schulze et al., ; dungan and leytem, c) . the most prevalent microorganisms identified in bioaerosol samples from afo are presented in table . with most bioaerosol studies, whether conducted at afo, composting facilities, wastewater treatment plants, biosolids application sites, or wastewater spray irrigations sites, the general trend observed is that the airborne microorganism concentrations decrease with distance from the source (goff et al., ; katzenelson and teltch, ; boutin et al., ; taha et al., ; green et al., ; low et al., ) . in a study at a swine operation, the average bacterial concentrations within the barns were . × cfu·m − , and although the outside air concentration decreased with distance from the facility, at m downwind the bacterial concentration was still . -fold greater ( cfu·m − ) than at the upwind location (green et al., ) . in a recent study by matković et al. ( ) , airborne concentrations of fungi inside a dairy barn were about × cfu·m − throughout the day (morning, noon, and night) and downwind concentrations approached background levels ( . to . × cfu·m − ) at distances as close as to m from the barn. at an open-lot dairy, the average endotoxin concentration at a background site was endotoxin units (eu)·m − , whereas at the edge of the lot and and , m further downwind, the average concentrations were , , and eu·m − , respectively (dungan and leytem, a) . table presents airborne concentrations for microorganisms and endotoxins within and downwind of various livestock operations. boutin et al. ( ) investigated bioaerosol emissions associated with the land application of swine and cattle slurries by way of tractor-pulled tanker and fixed high-pressure spray guns. near the source, total bacterial counts were about , cfu·m − , regardless of the land application method. the bacterial counts steadily decreased with distance from the application site and pathogenic bacteria such as salmonella, staphylococcus, and klebsiella pneumoniae were not detected. however, compared with tank spreading, which sprays closer to the ground, airborne bacterial concentrations were greater at greater distances from the spray guns, which is likely related to the upward discharge of slurry into the air that enhances droplet size reduction and drift. to our knowledge, the boutin et al. ( ) study is the only peer-reviewed report that addresses bioaerosol transport during spray irrigation of livestock manures, whereas most other reports address spray irrigation of industrial and municipal wastes (katzenelson and teltch, ; parker et al., ; camann et al., ; brooks et al., a; tanner et al., ) . in a preliminary pilot-scale field study conducted by kim et al. ( ) , swine manure was land-applied through a center pivot irrigation system and bioaerosol samples were collected upwind and , , and m downwind. total airborne coliform concentrations were found to decrease with distance, from about most probable number (mpn)·m − at m to near background concentrations at mpn·m − at m downwind. although the focus of this review is on bioaerosols associated with animal operations and manures, one could reasonably expect microorganisms in industrial and municipal wastewaters to behave similarly once aerosolized. differences in survivability may occur though, depending upon the concentration and type of om in the wastes because some organic substances are known to act as osmoprotectants (cox, ; marthi and lighthart, ) and may provide some degree of physical protection against uv radiation and drying (sobsey and meschke, ; aller et al., ) . parker et al. ( ) investigated the transport of aerosolized bacteria during the spray irrigation of potato processing wastewater. as with other similar studies, there was a decrease in the airborne microorganism concentration with distance from the irrigation system. these authors reported detection of coliforms at distances as far as . to . km from the source; however, there was no way to verify if they were above background concentrations because that information was not provided in the report. during the land application of liquid and dewatered domestic sewage sludge (biosolids) via spray tanker and spreader/slinger, respectively, indicator organisms (coliforms, clostridium perfringens, e. coli) were not detected at distances greater than m (brooks et al., b) . in most of the above-mentioned bioaerosol transport studies, fecal contamination indicator organisms were targeted. fecal indicator organisms are generally chosen because they are more abundant and easily identified in the aerosols (teltsch and katzenelson, ; bausum et al., ; brenner et al., ) , although they may behave differently from pathogens (dowd et al., ; carducci et al., ) . alternatively, to improve upon estimates of off-site transport of bio- aerosols, some researchers have used molecular-based approaches to track microorganisms from swine houses (duan et al., ) or during the land application of class b biosolids (low et al., ) and domestic wastewater (paez-rubio et al., ) . this approach is called microbial source tracking and has only recently been applied to aerosol samples. although emission rates for bioaerosols during the land application of livestock wastes are not currently available, emission rates have been calculated for the application of dewatered and liquid class b biosolids onto agricultural land. emission rate is a useful variable for understanding the impact of waste application, and similarities between application of municipal and livestock wastes can be made because the same spreading equipment is often used. during the land application of dewatered biosolids using a slinger, average emission rates for total bacteria, heterotrophic bacteria, total coliforms, sulfite-reducing clostridia, and endotoxin were reported to be . × cfu·s − , . × cfu·s − , . × cfu·s − , . × cfu·s − , and . × eu·s − , respectively . in a study conducted by tanner et al. ( ) , ground water seeded with e. coli was sprayed using a spray-tanker, and emission rates were reported to range from . to . × cfu·s − . interestingly, when studies were conducted using liquid biosolids, neither coliform bacteria nor coliphage were detected in air m downwind, although these microorganisms were detected in the biosolids. although no reason was given for the latter outcome, the direct measurement of bioaerosols does provide necessary information required for calculating emission rates. a bioaerosol emission rate is a required input variable for all aerosol fate and transport models that predict absolute concentration at a specified distance from the source . atmospheric dispersion modeling is a mathematical simulation used to predict the concentration of an air contaminant at various distances from a source. in an effort to assess the transport and diffusion of airborne microorganisms associated with human and livestock diseases, dispersion modeling has been utilized (sørensen et al., ; garten et al., ; pedersen and hansen, ) . in australia, atmospheric dispersion models have been developed as part of preparedness programs to manage potential outbreaks of foot-and-mouth disease (cannon and garner, ; garner et al., ) . in early bioaerosol transport studies, models were based upon a modified version of the inert particle dispersion model developed by pasquill ( ) . although some of the inert particle model assumptions will not be met at a typical afo, the model assumes ) gaussian distribution of particles in the crosswind and vertical planes; ) particles are emitted at a constant rate; ) diffusion in the direction of transport is negligible; ) particles are < μm in diameter (i.e., gravitational effects are negligible); ) particles are reflected from the ground (i.e., no deposition or reactions at surface); ) wind velocity and direction are constant; and ) terrain is flat. the original form of the inert particle dispersion model is where χ is the number of particles per cubic meter of air at a downwind location x, γ, and z (i.e., alongwind, crosswind, and vertical coordinates, respectively); q is the number of particles emitted per second; ū is the mean wind speed in meters per second; σ y and σ z are the sd of the crosswind and vertical displacements of particles at distance x downwind, respectively; and h is the height of the source including plume rise. if ground-level and centerline concentrations are to be determined, then z and γ are set to zero. for a groundlevel source h is also set to zero, the simplified equation then becomes because the pasquill dispersion model is based on inert particles, lighthart and frisch ( ) added a biological decay term as follows: χ (x,γ,z) bd = χ (x,γ,z) exp(-λt), [ ] where λ is the microbial death rate (per second) and t is approximated by x/ū. subsequent researchers utilized the biological decay term, along with the dispersion model, to assess bioaerosol transport from point sources (peterson and lighthart, ; teltsch et al., b; us epa, ; lighthart and mohr, ) . when only part of the material released into the atmosphere becomes an aerosol, as occurs during sprinkler irrigation, eq. [ ] becomes where e is the aerosolization efficiency factor (teltsch et al., b) . the microbial death and inactivation rates are generally derived from empirical laboratory data under static atmospheric conditions using pure cultures (hatch and dimmick, ) . therefore, it is imperative when developing microbial death rates to conduct the experiments with numerous microbial types and under varying environmental conditions (peterson and lighthart, ) . in laboratory studies, microbial death rates for sarcina lutea at °c were . × − and . × − s − at around and % relative humidity, whereas death rates for pasturella tularensis at °c were . × − and . × − s − at similar relative humidities, respectively (cox and goldberg, ; lighthart, ) . whereas these microbes are non-spore formers, one would expect spore-forming bacteria to survive longer under changing atmospheric conditions as a result of their ability to tolerate greater temperature and radiation (madigan and martinko, ) . as mentioned previously, the viability of airborne microorganisms will vary greatly depending upon growth media used and microbial genus and species being tested. in field trials conducted at pleasanton, ca, microbial death rates during the spray irrigation of municipal wastewater were determined under a variety of environmental conditions (us epa, ) . the median death rate constants for total coliform, fecal coliform, and coliphage were . , . , and . × − s − , respectively. death rate constants for e. coli, prepared in sterilized municipal wastewater, were reported to range from . × − s − in the morning to . × − s − in the afternoon (teltsch et al., b) . parker et al. ( ) modified pasquill's inert particle dispersion model to predict the transport of bioaerosols from an area source (i.e., sprinkler irrigation of potato processing wastewater). even though the model contained a biological decay term, the authors did not model decay or loss of viability of microorganisms due to a lack of experimental data. dowd et al. ( ) later used the same area-source model with microbial death rates from the literature to predict bioaerosol transport during the land application of dewatered domestic sewage sludge (biosolids). based upon model predictions at a high wind speed of m·s − , bacterial concentrations would be and . m − of air at and , m, respectively. to assess the risk of infection to workers and nearby populations, a beta-poisson model as described by haas ( ) was utilized. using dose-response data for salmonella typhimurium, the predicted risk of infection at m with a m·s − wind speed and h exposure period was %, whereas at , and , m it decreased to . and . %, respectively. risk of infection for coxsackievirus b was also determined; however, an incorrect dose-response value was used in the single-hit exponential model, and predicted risk of infection should have actually been about orders of magnitude less than their published values. overall, their model predictions suggest that bioaerosols from land-applied biosolids can increase the risk of viral and bacterial infection to onsite workers, but there was little or no risk to population centers > km from the application site under low-wind conditions (≤ m·s − ). the results from such studies should be used cautiously because the results were not empirically derived and, as outlined by pillai and ricke ( ) , there is uncertainty associated with the dose-response of different organisms and hosts. in a us epa report, microorganism concentrations in aerosols from spray irrigation events of municipal wastewater were predicted using an atmospher-ic diffusion model. the diffusion model consisted of principal components: where c d is the concentration of microorganisms per cubic meter of air; d d is the atmospheric diffusion factor at distance d from the source (s·m − ); q a is the aerosol source strength (microorganism s − ); m d is microorganism die-off factor (not to be confused with microbial death rate, λ) as described in eq. [ ] (i.e., number of organisms that are viable at distance d); and b is the background concentration (microorganisms m − ). d d is calculated using the inert particle dispersion model as shown in eq. [ ], but q was set to unity. for a wastewater irrigation event, the aerosol source strength was further defined as q a = w f e i, [ ] where w is the microorganism concentration in the wastewater (organisms l − ); f is the flow rate of the irrigation wastewater (l·s − ); e is the aerosolization efficiency factor ( < e ≤ ); and i is the microorganism impact factor (i.e., aggregate effect of all of factors affecting microorganism survivability; i > ). using input data from a us epa ( ) report, total coliform concentrations were determined m from the centerline of -m-long linear source under stable (summer night) and unstable (summer midday) atmospheric conditions. the wastewater flow rate during the irrigation event was set at l·s − , with a total coliform concentration of . × cfu·l − and respective night and midday wind speeds of and m·s − , e of . × − and . × − , i of . and . , λ of . and . s − , and aerosol age (a d ) of and s. the q a for total coliforms during night and midday was determined to be . × and . × cfu·s − , respectively. when background coliform concentrations were subtracted, the respective total airborne concentrations at m downwind were predicted to be only . and . × − cfu·m − . during midday conditions, fecal streptococci concentrations at m downwind were predicted to be -fold greater than total coliforms, even though the source concentration was -fold less. this is owing to the fact that fecal streptococci had a microorganism impact factor of . and death rate of zero. lighthart and mohr ( ) modified a version of the gaussian plume model used by peterson and lighthart ( ) to include an airborne microbial survival term that was a best-fit function of temperature, relative humidity, and solar radiation. the model included an algorithm using microbial source strength and local hourly mean weather data to drive the model through a typical summer or overcast and windy winter day. at high wind speeds or short travel times, the model predicted greater viable near-source concentrations because the microorganisms did not have time to become inactivated. as travel times were increased, due to slow wind speeds or longer distances, inactivation of microorganisms became more prevalent. lighthart and kim ( ) used a simulation model to describe the dispersion of individual droplets of water containing viable microbes. the droplet dispersion model was separated into submodels: ) aerosol generation, ) evaporation, ) dispersion, ) deposition, and ) microbial death. the position of each droplet, at each time step in the trajectory, was located in a -dimensional coordinate system. when the modeling process was repeated for many droplets, a simulation of a cloud of droplets then occurred. the effect of evaporation was determined to be an important factor when simulated in the model, as aerosols were carried further downwind. whereas the model takes into account the physical, chemical, and measured meteorological parameters for each water droplet, potential shortcomings revolved around the ability of the model to predict nearsource survival dynamics of airborne microorganisms (e.g., effect of microorganisms on water evaporation, critical water content of microbes). also, the droplet dispersion model does not take into account rapidly changing wind conditions (e.g., gusts) and, therefore, use of average wind velocities will lead to an oversimplification of meteorological conditions and microbial dispersion. when the model was compared with a release of pseudomonas syringae, deposition rates were found to be similar within m of the source. the simulation model was later used by ganio et al. ( ) to model a field spray event of bacillus subtilis var. niger spores. using the same meteorological conditions as the spray event, the model produced a bioaerosol deposition pattern somewhat similar to that obtained in the field (r = . ). a variety of short-and long-range dispersion models have been developed to understand and manage the airborne spread of epidemics such as foot-and-mouth disease (gloster et al., ; sørensen, ; cannon and garner, ; sørensen et al., ; rubel and fuchs, ; garner et al., ; mayer et al., ) . in a recent paper by gloster et al. ( ) , a historic outbreak of fmd in (hampshire, uk) was modeled using internationally recognized dispersion model systems. whereas one-half of the models [nuclear accident model (name), veterinary meteorological decision-support system (vetmet), plume dispersion emergency modeling system (pdems)] were run using observational data provided, the other one-half [australian integrated windspread model (aiwm), modéle lagrangien courte distance (mlcd), national atmospheric release advisory center (nrac)] used numerically derived meteorological data, and comparisons between outputs were made. using the same virus emission data, the models produced very similar h integrated concentrations along the major axis of the plume at , , , , and km. although there were differences between the estimates, as a result of model assumptions with respect to upward diffusion rates for surface material and choice of input weather data, most estimates were within one order of magnitude. these models also predicted similar directions for livestock at risk; however, additional model assumptions such as microbial fate and susceptibility to airborne infection can substantially modify the size and location of the downwind risk area. based on information presented in this review, it is evident that animal feeding operations and manure application practices contribute to the formation of bioaerosols at greater concentrations than found in background environments. as population centers grow and converge on such operations, there will be an increasing potential for exposure to airborne pathogens and microbial by-products that are transported off site. exposure to airborne bacteria, virus, fungi, and microbial by-products is not limited to inhalation routes because deposition on fomites, food crops, and water bodies and subsequent ingestion also represent transmission routes of concern. the ability to accurately quantify airborne microorganisms within and downwind from a source is important when evaluating health risks to exposed humans and animals. however, the actual risk of exposure from airborne pathogens has not been fully recognized for a variety of reasons including choice of bioaerosol collection technique, analytical methodology, target microorganism, and dispersion and infectivity model inputs. to date, most bioaerosol transport studies have targeted fecal indicator organisms because they are generally more abundant and easily detected. pathogens on the other hand are often at concentrations that are several orders of magnitude less than indicator organisms, making their detection difficult in highly diluted aerosol samples. because the survivability of aerosolized fecal indicator organisms is likely different from that of pathogens, a first step to improve future bioaerosol studies should include the selection of organisms that better represent targeted pathogens, along with standardized methods for their collection in outdoor environments. as molecular-based approaches improve with respect to sensitivity and rapidity, it may be appropriate to standardize and use such technologies to directly detect pathogens of interest in aerosol samples, avoiding the need for indicator organisms. standardization of target microorganisms and collection and analytical methodologies will improve the ability of researchers to compare results, refine dispersion models, and develop unified risk estimates. although animal operations and manure management practices are not currently regulated with respect to bioaerosol emissions, the possibility that control measures will someday be implemented is quite realistic. without standardized methodologies, regulatory agencies will have to base decisions on inconsistent data sets, and the effectiveness of mitigation strategies to control bioaerosol emissions will not be properly determined. because land application of manures will remain a viable nutrient utilization and disposal option into the foreseeable future, emphasis must be placed on research addressing the airborne transport of pathogens because there is a lack of information on this topic. furthermore, there is a surprising lack of information concerning the infectivity of aerosolized pathogens, especially enteric pathogens. clearly, a critical component of a risk determination is not only understanding bioaerosol dispersion and transport, but also the dose-response of zoonotic pathogens. to advance our understanding of risks associated with airborne pathogens from animal feeding operations, it will be necessary for a variety of scientists, including but not limited to aerobiologists, clinical microbiologists, epidemiologists, animal scientists, and risk modelers, to convene under a common setting to address these issues in more detail and work toward a common goal of standardizing of variety of bioaerosol collection and analytical methodologies. aerosol stability of infectious and potentially infectious reovirus particles volumetric assessment of airborne fungi in two sections of a rural indoor dairy cattle shed effect of temperature and relative humidity on the survival of airborne columbia sk group viruses airborne stability of simian virus the sea surface microlayer as a source of vira and bacterial enrichment in marine aerosols comparison of coliphage and bacterial aerosols at a wastewater spray irrigation site stability of disseminated aerosols of pasteurella tularensis subjected to simulated solar radiation and various humidities atmospheric bacterial contamination from landspreading of animal wastes: evaluation of the respiratory risk for people nearby animal viruses, coliphages, and bacteria in aerosols and wastewater at a spray irrigation site biological aerosol emission, fate, and transport from municipal and animal wastes estimation of bioaerosol risk of infection to residents adjacent to a land applied biosolids site using an empirically derived transport model a national study on the residential impact of biological aerosols from the land application of biosolids aerial dispersal of pathogens on the global and continental scales and its impact on plant disease microorganism levels in air near spray irrigation of municipal wastewater: the lubbock infection surveillance study airborne particulate matter from livestock production systems: a review of an air pollution problem. environ assessing the risk of windborne spread of foot-and-mouth disease in australia assessment of microbial parameters as indicators of viral contamination of aerosol from urban sewage treatment plants exposure assessment to airborne endotoxin, dust, ammonia, hydrogen sulfide and carbon dioxide in open style swine houses exposure of workers to airborne microorganisms in open-air swine houses investigation and application of methods for enumerating heterotrophs and escherichia coli in the air within piggery sheds mechanically ventilated broiler sheds: a possible source of aerosolized salmonella, campylobacter, and escherichia coli auditing and assessing air quality in concentrated feeding operation the survival of escherichia coli sprayed into air and into nitrogen from distilled water and from solutions protecting agents, as a function of relative humidity inactivation kinetics of some microorganisms subjected to a variety of stresses the toxic effect of oxygen upon the aerosol survival of escherichia coli b aerosol survival of serratia marcescens as a function of oxygen concentration, relative humidity, and time aerosol survival of pasteurella tularensis and the influence of relative humidity differences between the thermal inactivation of picornaviruses at "high" and "low" temperatures bioaerosol health effects and exposure assessment: progress and prospects bioaerosol transport modeling and risk assessment in relation to biosolid placement thermotolerant clostridia as an airborne pathogen indicator during land application of biosolids source identification of airborne escherichia coli of swine house surroundings using eric-pcr and rep-pcr airborne endotoxin concentrations at a large open lot dairy in southern idaho a concise review of methodologies used to collect and characterize bioaerosols and their application at concentrated animal feeding operations the effect of extraction, storage, and analysis techniques on the measurement of airborne endotoxins from a large dairy yearlong monitoring of airborne endotoxin at a concentrated dairy operation assessment of bioaerosols at a concentrated dairy operation survival of airborne pasteurella tularensis at different atmospheric temperatures effects of atmospheric humidity and temperature on the survival of airborne falvobacterium relationship between atmospheric temperature and survival of airborne bacteria direct detection of salmonella cells in the air of livestock stable by real-time pcr a comparison between computer modeled bioaerosol dispersion and a bioaerosol field spray event an integrated modeling approach to assess the risk of wind-borne spread of foot-and-mouth disease virus from infected premises modeling the transport and dispersion of airborne contaminants: a review of techniques and approaches sources of pathogenic microorganisms and their fate during land application of wastes airborne spread of foot-and-mouth disease: model intercomparison long distance transport of foot-and-mouth disease virus over sea emission of microbial aerosols from sewage treatment plants that use trickling filters bacterial plume emanating from the air surrounding swine confinement operations atmospheric movement of microorganisms in clouds of desert dust and implications for human health sampling for airborne microorganisms. page in manual for environmental microbiology estimation of risk due to low doses of microorganisms: a comparison of alternative methodologies rapid aerosol transmission of salmonella among turkeys in a simulated holding-shed environment technical background document: microbial risk assessment and fate and transport modeling of aerosolized microorganisms at wastewater land application facilities in idaho physiological responses of airborne bacteria to shifts in relative humidity health effects of airborne exposures from concentrated animal feeding operations a method to provide improved dose-response estimates for airborne pathogens in animals: an example using porcine reproductive and respiratory syndrome virus inactivation of airborne viruses by ultraviolet radiation the effects of meteorological factors on atmospheric bioaerosol concentrations-a review dispersion of enteric bacteria by spray irrigation changes in soil physical properties due to organic waste applications: a review computational fluid dynamics (cfd) modeling to predict bioaerosol transport behavior during center pivot wastewater irrigation influence of relative humidity on particle size and uv sensitivity of serratia marcescens and mycobacterium bovis bcg aerosols measurement of low pressure sprinkler evaporation loss aerobiological engineering handbook. page in aerosol science and particle dynamics zoonoses: infectious diseases transmissible from animals to humans survival of airborne bacteria in high urban concentration of carbon monoxide physics of microbial bioaerosols. page in atmospheric microbial aerosols: theory and mini-review of the concentration variations found in the alfresco atmospheric bacterial populations estimation of viable airborne microbes downwind from a point source simulation of airborne microbial droplet transport estimating downwind concentrations of viable airborne microorganisms in dynamic atmospheric conditions bacterial flux from chaparral into the atmosphere in mid-summer at a high desert location off-site exposure to respirable aerosols produced during the disk-incorporation of class b biosolids brock biology of microorganisms survival of bacteria during aerosolization effects of betaine on enumeration of airborne bacteria airborne fungi in a dairy barn with emphasis on microclimate and emissions a lagrangian particle model to predict the airborne spread of foot-and-mouth disease virus bioaerosols associated with animal production operations fate and transport of microorganisms in air. page in manual for environmental microbiology the outdoor aerosol. page in bioaerosols molecular identification of airborne bacteria associated with aerial spraying of bovine slurry waste employing s rrna gene pcr and gene sequencing techniques culture-independent characterization of archaeal biodiversity in swine confinement building aerosols culture-independent approach of the bacterial bioaerosol diversity in the standard swine confinement buildings, and assessment of the seasonal effect experimental airborne transmission of salmonella agona and salmonella typhimurium in weaned pigs emission rates and characterization of aerosols produced during the spreading of dewatered class b biosolids source bioaerosol concentration and rrna gene-based identification of microorganisms aerosolized at a flood irrigation wastewater reuse site microbial aerosols from food-processing waste spray fields the estimation of the dispersion of windborne material soil microbial community responses to dairy manure or ammonium nitrate applications quantification of airborne biological contaminants associated with land applied biosolids. water environment research foundation assessment tools in support of epidemiological investigations of airborne dispersion of pathogens estimation of downwind viable airborne microbes from a wet cooling tower-including settling land application equipment for livestock and poultry manure management bioaerosols from municipal and animal wastes: background and contemporary issues studies on the instantaneous death of airborne escherichia coli land application of agricultural, industrial, and municipal by-products assessment of bioaerosols in swine barns by filtration and impaction effect of relative humidity on the inactivation of airborne serratia marcescens by ultraviolet radiation land application of manure for beneficial reuse. page in animal agriculture and the environment, national center for manure & animal waste management white papers potential impact of land application of by-products on ground and surface water quality. page in land application of agricultural, industrial, and municipal by-products a decision-support system for realtime assessment of airborne spread of the foot-and-mouth disease virus health aspects of bioaerosols. page in atmospheric microbial aerosols: theory and applications. b. lighthart and a bioaerosol distribution patterns adjacent to two swine-growing-finishing housed confinement units in the american midwest endotoxin concentration in modern animal houses in southern bavaria ambient endotoxin level in an area with intensive livestock production concentrations and emissions of airborne endotoxins and microorganisms in livestock buildings in northern pathogen in animal wastes and the impacts of waste management practices on their survival, transport and fate. page in animal agriculture and the environment virus survival in the environment with special attention to survival in sewage droplets and other environmental media of fecal or respiratory origin. page . report for the world health organization changes in soil properties under annual applications of feedlot manure and different tillage practices influence of relative humidity on the survival of some airborne viruses health and hygiene aspects of spray irrigation sensitivity of the derma long-range gaussian dispersion model to meteorological input and diffusion parameters modelling the atmospheric dispersion of footand-mouth disease virus for emergency preparedness an integrated model to predict the atmospheric spread of fmd virus exposure to inhalable dust and endotoxins in agricultural industries introduction to aerobiology. page in manual for environmental microbiology estimating fugitive bioaerosol releases from static compost windrows: feasibility of a portable wind tunnel approach estimated occupational risk from bioaerosols generated during land application of class b biosolids bioaerosol emission rate and plume characteristics during land application of liquid class b biosolids airborne enteric bacteria and viruses from spray irrigation with wastewater isolation and identification of pathogenic microorganisms at wastewater-irrigated fields: ratios in air and wastewater die-away kinetics of aerosolized bacteria from sprinkler application of wastewater influence of temperature and relative humidity on the survival of chlamydia pneumoniae in aerosols the evaluation of microbiological aerosols associated with the application of wastewater to land estimating microorganism densities in aerosols from spray irrigation of wastewater pathogens and antibiotic residues in animal manures and hygienic and ecological risks related to subsequent land application effect of aerosolization on subsequent bacterial survival aerosol infection of calves and mice with salmonella typhimurium airborne microbial flora in a cattle feedlot agricultural uses of municipal, animal, and industrial byproducts concentrations of airborne endotoxin in cow and calf stables determination of the inflammatory potential of bioaerosols from a duck-fattening unit by using a limulus amebocyte lysate assay and human whole blood cytokine response airborne gramnegative bacterial flora in animal houses key: cord- - lxmq u authors: zhao, shi; musa, salihu s.; fu, hao; he, daihai; qin, jing title: large-scale lassa fever outbreaks in nigeria: quantifying the association between disease reproduction number and local rainfall date: - - journal: nan doi: . /s sha: doc_id: cord_uid: lxmq u lassa fever (lf) is increasingly recognised as an important rodent-borne viral haemorrhagic fever presenting a severe public health threat to sub-saharan west africa. in – , lf caused an unprecedented epidemic in nigeria and the situation was worsening in – . this work aims to study the epidemiological features of epidemics in different nigerian regions and quantify the association between reproduction number (r) and state rainfall. we quantify the infectivity of lf by the reproduction numbers estimated from four different growth models: the richards, three-parameter logistic, gompertz and weibull growth models. lf surveillance data are used to fit the growth models and estimate the rs and epidemic turning points (τ) in different regions at different time periods. cochran's q test is further applied to test the spatial heterogeneity of the lf epidemics. a linear random-effect regression model is adopted to quantify the association between r and state rainfall with various lag terms. our estimated rs for – ( . with % ci . – . ) was significantly higher than those for – ( . with % ci: ( . , . )) and – (ranged from . to . ). we report spatial heterogeneity in the rs for epidemics in different nigerian regions. we find that a one-unit (mm) increase in average monthly rainfall over the past months could cause a . % ( % ci . %– . %)) rise in r. there is significant spatial heterogeneity in the lf epidemics in different nigerian regions. we report clear evidence of rainfall impacts on lf epidemics in nigeria and quantify the impact. lassa fever (lf), caused by lassa virus (lasv), is increasingly recognised as an important rodent-borne viral haemorrhagic fever presenting a severe public health threat to some of the communities in sub-saharan west africa [ ] . discovered in [ ] , lf is endemic to much of rural nigeria and regions in the mano river union [ ] . lasv transmits from human to human, as well as via the zoonotic cycle [ , , ] . lf has a high case fatality rate ranging from % in the community to over % in hospital settings [ , , ] . the common reservoir of lasv is mastomys natalensis, one of the most widespread rodent species in sub-saharan africa [ , ] , which exhibits sensitive population dynamics to the water level, e.g. rainfall, flooded agricultural activities [ , ] . previous studies have recognised the ecological association between the population levels of rodents and rainfall [ ] [ ] [ ] . lf epidemics typically start in november and last until may of the following year, with the majority of cases occurring in the first quarter of the following year, in addition to sporadic cases reported throughout the year. the - epidemic in nigeria was an unprecedented lf epidemic in the country's history [ ] , which resulted in confirmed cases, including deaths, between january and march [ ] . the most recent epidemic in nigeria has already caused confirmed cases from january to march of , which included deaths [ ] . the five states of edo, ondo, ebonyi, bauchi and plateau are the only states that have been among the top hit hardest states in terms of number of lf cases in both the ( . % of total national cases) and ( . % of total national cases) epidemics. while there have been discussions about the association of rainfall level and lf incidence rate [ , ] , this association has not yet been demonstrated and quantified. this work aims to study the epidemiological features of epidemics in different nigerian regions between january and march . we estimate lf infectivity in terms of the reproduction number (r) and quantify the association between r and state rainfall. we explore the spatial heterogeneity of the lf epidemics and summarise the overall findings with model-average estimates. weekly lf surveillance data are obtained from the nigeria centre for disease control (ncdc), where the data are publicly available from the weekly situation reports released by ncdc [ ] . laboratory-confirmed case time series are used for analysis. we examine the major epidemics that occurred between january and march across the whole country and the aforementioned five states that were among the top hardest-hit states in both the and epidemics, i.e. edo, ondo, ebonyi, bauchi and plateau. the state rainfall records of each state were collected on monthly average basis from the historical records of the world weather online website [ ] . figure (a) and (b) shows the rainfall time series of the five states and the weekly reported lf cases across the entire nigeria. to test the credibility of the coincidence between rainfall and lf epidemic, we use a simple statistical regression model of 'case ∼ exp(α × rainfall) + θ', where α and θ are free parameters to be estimated. the 'rainfall' in the model represents the state rainfall time series with lag of - months. this lag term corresponds to the time interval between the rainfall and the development of rodent population [ ] . we check the least-square fitting outcomes of these regression models and select the model of lagged rainfall with the highest goodness-of-fit. the fitting significance is treated as the initiation of the quantitative association between state rainfall and the lf epidemic. four different nonlinear growth models are adopted to pinpoint the epidemiological features of each epidemic. the models are the richards, three-parameter logistic, gompertz and weibull growth models. these simple structured models are widely used to study s-shaped cumulative growth processes; e.g. the curve of a single-wave epidemic and have been extensively studied in previous work [ , ] . these models consider cumulative cases with saturation in the growth rate to reflect the progression of an epidemic due to reduction in susceptible pools or a decrease in the exposure to infectious rodent populations. the extrinsic growth rate increases to a maximum (i.e. saturation) before steadily declining to zero. the modelling and fitting via the growth models of the epidemic curve are illustrated in figure . we fit all models to the weekly reported lf cases in different regions and evaluate the fitting performance by the akaike information criterion (aic). we adopt the standard nonlinear least squares (nls) approach for model fitting and parameter estimation, following [ , ] . a p-value < . is regarded as statistically significant and the % confidence intervals (cis) are estimated for all unknown parameters. as we are using the cumulative number of the lf cases to conduct the model fitting, some fitting issues might occur, as per the studies in king et al. [ ] , due to the non-decreasing nature in the cumulative summation time series. the models are selected by comparing the aic to that of the baseline (or null) model. only the models with an aic lower than the aic of the baseline model are considered for further analysis. importantly, the baseline model adopted is expected to capture the trends of the time series. since the epidemic curves of an infectious disease commonly exhibit autocorrelations [ ] , we use autoregression (ar) models with a degree of , i.e. ar( ), as the baseline models for growth model selection. we also adopt the coefficient of determination (r-squared) and the coefficient of partial determination (partial r-squared) to evaluate goodness-of-fit and fitting improvement, respectively. for the calculation of partial r-squared, the ar( ) model is used as the baseline model. the growth models with a positive partial r-squared (indicating fitting improvement) against the baseline ar( ) model will be selected for further analyses. after the selection of models, we estimate the epidemiological features (parameters) of turning point (τ) and reproduction number (r) via the selected models. the turning point is defined as the time point of a sign change in the rate of case accumulation, i.e. from increasing to decreasing or vice versa [ , ] . the reproduction number, r, is the average number of secondary human cases caused by one primary human case via the 'human-to-rodent-to-human' transmission path [ , ] . when the population is totally (i.e. %) susceptible, the r will equate to the basic reproduction number, commonly denoted as r [ , ] . the reproduction number (r) is given in eqn ( ), here, γ is the intrinsic per capita growth rate from the nonlinear growth models and κ is the serial interval of the lasv infection. the serial interval (i.e. the generation interval) is the time between the infections of two successive cases in a chain of transmission [ , [ ] [ ] [ ] . the function h(·) represents the probability distribution of κ. hence, the function m(·) is the laplace transform of h(·) and specifically, m(·) is the moment generating function (mgf) of a probability distribution [ ] . according to previous work [ ] , we assume h(κ) to follow a gamma distribution with a mean of . days and a standard deviation (sd) of . days. therefore, r can be estimated with the values of γ from the fitted models [ , , , ] . the state rs were estimated from the γs of the fitted epidemic growth curves of each state. similarly, the national rs are estimated from the γs of the epidemic growth curves fitted to the national number of cases time series in different epidemic periods. we then summarise the κ and r estimates via the aic-weighted model averaging. the aic weights, w, of the selected models (with positive partial r-squared) are defined in eqn ( ), here, aic i is the aic of the i-th selected model and the aic min is the lowest aic among all selected models. thus, the i-th selected model has a weight of w i . the model-averaged estimator is the weighted average of the estimates in each selected model, which has been well studied in previous work [ , ] . for the aic-based model average of the r, there could be the situation that no growth model is selected according to the partial r-squared. in such cases, instead of the model average, we report the range of the r estimated from all growth models. testing the spatial heterogeneity of the lf epidemics after finding the model-averaged estimates, we apply cochran's q test to examine the spatial heterogeneity of the epidemics in different regions over the same period of time [ ] . for instance, we treat the model-averaged r estimates as the univariate meta-analytical response against different nigerian regions (states) and further check the heterogeneity by estimating the significance levels of the q statistics. a p-value < . is regarded as statistically significant. similar to the approach in the previous study [ ] , the association between the state rainfall level and lasv transmissibility are modelled by a linear mixed-effect regression (lmer) model in eqn ( ), here, e(·) represents the expectation function and j is the region index corresponding to different regions (states). term c j is the interception term of the j-th region to be estimated and it is variable from different regions, serving as the baseline scale of transmissibility in different states. the term t denotes the cumulative lag in the model and 〈rainfall j,t 〉 represents the average monthly rainfall of the previous t months from the turning point, τ, of the j-th region. the range of lag term, t, will be considered from to months, which is explained by the time interval between the peak of the rainfall and the peak of rodent population [ ] . as illustrated in figure , the reproduction numbers, r j s, are estimated for different epidemics from the selected growth models. the regression coefficient, β, is to be estimated. hence, the term (e β - ) × % is the percentage changing rate (of r), which can be interpreted as the percentage change in transmissibility due to a one-unit (mm) increase in the average of the monthly rainfall level over the past months. the framework of the regression is based on the exponential form of the predictor to model the expectation of transmissibility (e.g. r); this framework is inspired by previous work [ ] [ ] [ ] [ ] . to quantify the impacts of state rainfall, we calculate the percentage changing rate with different cumulative lags (t) from to months and estimate their significant levels. only the lag terms (t) with significant estimates are presented in this work. we present the analysis procedure in a flow diagram in figure . all analyses are conducted by using r (version . . [ ] ) and the r function 'nls' is employed for the nls estimation of model parameters. the rainfall time series of the five states and the weekly reported lf cases of the whole of nigeria are shown in figure (a) and (b). we observe that the major lf epidemics usually occur in nigeria between november and may of the following year. the cumulative lagged effects were observed via matching the peak timing of the rainfall and epidemic curves. in figure (c), we shift the rainfall time series of the five states by + months to match the trends of the national lf epidemic curve in nigeria. in figure (d) and (e), we find that the fit has a p-value < . , which indicates a statistically significant association between the lf cases and shifted rainfall curve. we fit four different growth models to the lf confirmed cases and estimate the model-average reproduction number (r) after model selection. we show the growth model fitting results in figure and the model estimation and selection results in table . most of the models have positive partial r-squared against the baseline ar( ) model. most of the regions exhibit an epidemic turning point (τ) ranging from the epidemiological week (ew) - , i.e. from the end of january to mid-march, in each year. out of four epidemics in the states of bauchi and plateau, there are three estimated τs after ew ( table ) many previous studies adopted the instantaneous reproduction number, commonly denoted by r t , which can be estimated by a renewable equation, to quantify the transmissibility of infectious diseases [ , , , , ] . the factors that affect the changing dynamics of r t include (i) the depletion of the susceptible population [ ] or decrease in the exposure to infectious sources, (ii) the change, usually it is the improvement, in the unmeasurable disease control efforts, e.g. contract tracing, travel restriction, school closure, etc., [ ] [ ] [ ] [ ] and local awareness of the epidemic [ ] , and (iii) the natural features of the pathogen, e.g. its original infectivity and other interepidemic factors [ , , ] . in this work, we choose to use the average reproduction number (r) rather than r t , as the measurement of the lasv transmissibility. the estimated r summarises the lasv transmissibility over the whole period of an epidemic. the reasons why we prefer r rather than r t are as follows. first, the temporal changes of the susceptible population or decrease in the exposure to infectious sources are removed from the r estimates due to the nature of the growth models. second, the changes of the susceptible population and/or disease awareness or control measures and the effect of the rainfall cannot be disentangled in the time-varying reproduction number, r t , the average reproduction number (r) adopted is a better proxy to explore the association between lf infectivity and rainfall. with respect to point (iii) and other heterogeneities of epidemics in different regions, we account for this issue by including the 'region' dummy variables in the lmer model in eqn ( ). these dummy variables serve as random effects to offset the regional heterogeneities of lf epidemics. therefore, we can then quantify a general effect, i.e. the β in eqn ( ), of the lagged rainfall on the lasv r estimate among different nigerian places. the association between total rainfall in a state and the lasv transmissibility (r) is modelled and quantified by the lmer model. in figure , we find a positive relation between rainfall and r. the estimated changing rate in r under a one-unit (mm) increase in the average monthly rainfall is summarised with different cumulative lag terms from to months (the t in eqn ( )). the range of lag in the rainfall from to months had previously been explained by the time interval between the peak of the rainfall and the peak of the rodent population [ ] . the estimates of the rainfall-associated changing rate in r with different lag terms were summarised in table . we report the most significant (i.e. with the lowest p-value) regression estimates that appear with a cumulative lag of months. the habitats of the lasv reservoir, i.e. rodents, include irrigated and flooded agricultural lands that are commonly found in and around african villages [ ] . the -month lag also coincides with the period between the dry and rainy seasons [ ] . the association between rodent population dynamics and rainfall levels has been demonstrated in a number of previous studies [ ] [ ] [ ] [ ] [ ] . hence, we consider the -month lagged estimation as our main results. namely, a one-unit (mm) increase in the average monthly rainfall over the past months is likely to cause a . % ( % ci . %- . %) rise in the r of the lf epidemic. we also remark that this 'one-unit (mm) increase in the average monthly rainfall over the past months' is equivalent to ' -unit (mm) increase in the total rainfall over the past months'. the present finding of the impact of lagged rainfall on lf epidemics suggests that the knowledge of such weather-driven epidemics could be gained by referring to past rainfall levels. for instance, if a relatively high amount of rainfall occurs, local measures, such as rodent population control, could be effective to reduce the lf risk. this speculation could also be verified by examining the rodent population data of the nigerian regions included in this work. the findings in this work are of public health interest and are helpful for policymakers in lf prevention and control. on the one hand, our findings suggest the existence of an association between rainfall and lasv transmissibility, which could be affected by the population dynamics of rodents [ ] . on the other hand, the positive relation between rainfall and r indicates that rainfall, particularly in states with a high lf risk, can be translated as a warning signal for lf epidemics. the modelling framework in this study should be easily extended to other infectious diseases. our work contains limitations. as in some african countries, the weather data are available only from a limited number of observatory stations and thus it is not sufficient to capture more accurate spatial variability. in this work, instead of exploring the spatial differences in the associations between rainfall and lf epidemic, we relaxed the setting and studied a general relationship. we qualified the general rainfall-associated changing rate of r in nigeria. for the transmissibility estimation, our growth modelling framework can provide the estimates of r, but not the basic reproduction number commonly denoted as r . however, according to the theoretical epidemiology [ , , , , ] , the r can be determined by r = r/s, where s denotes the population susceptibility. although s is not involved in our modelling framework, the information of s could be acquired from local serological surveillances. the existing literature reported . % seroprevalence among nigerian humans by the enzyme-linked immunosorbent assay (elisa) [ ] . hence, the r can be calculated as . by using s = - . % = . and the r = . as the average of the - lf epidemics. this was a data-driven modelling study, and we quantified the effect of rainfall as a weather-driven force of r based on previous ecological and epidemiological evidences [ , ] . since the transmission of lasv mainly relies on the rodent population, the factors including seasonality, agricultural land-using, subtropical or tropical forest coverage that could impact rodent ecology should be relevant and helpful in the analysis. however, due to availability of data, the agricultural land-using factors, e.g. pastureland, irrigated land, flooded agricultural land usage and forest coverage were absent in our analysis, which should be studied in the future if they become available. the lf epidemic reproduction numbers (r) of the whole of nigeria in - (r = . with % ci . - . ) and - (r ranged from . to . ) are significantly higher than in - (r = . with % ci . - . ). there is significant spatial heterogeneity in the lf epidemics of different nigerian regions. we report clear evidence of rainfall impacts on lf epidemics in nigeria and quantify this impact. a one-unit (mm) increase in the average monthly rainfall over the past months could cause a . % ( % ci . %- . %) rise in the r. the state rainfall information has potential to be utilised as a warning signal for lf epidemics. data. all data used for analysis are freely available via online public domains [ , ] . understanding the cryptic nature of lassa fever in west africa lassa fever epidemiological aspects of the epidemic lassa fever: epidemiology, clinical features, and social consequences the lassa fever fact sheet, the world health organization (who) lassa fever in post-conflict sierra leone population dynamics of the multimammate rat mastomys huberti in an annually flooded agricultural region of central mali stochastic seasonality and nonlinear density-dependent factors regulate population size in an african rodent the use of rainfall patterns in predicting population densities of multimammate rats the basis of reproductive seasonally in mastomys rats (rodentia: muridae) in tanzania forecasting rodent outbreaks in africa: an ecological basis for mastomys control in tanzania nigeria hit by unprecedented lassa fever outbreak the collection of the lassa fever outbreak situation reports, nigeria centre for disease control risk maps of lassa fever in west africa mapping the zoonotic niche of lassa fever in africa the historical weather records, the website of world weather online real-time parameter estimation of zika outbreaks using model averaging analysis of logistic growth models intervention measures, turning point, and reproduction number for dengue avoidable errors in the modelling of outbreaks of emerging pathogens, with special reference to ebola analysing increasing trends of guillain-barré syndrome (gbs) and dengue cases in hong kong using meteorological data how generation intervals shape the relationship between growth rates and reproductive numbers phase shifting of the transmissibility of macrolide sensitive and resistant mycoplasma pneumoniae epidemics in hong kong a new framework and software to estimate timevarying reproduction numbers during epidemics estimating individual and household reproduction numbers in an emerging epidemic associations between public awareness, local precipitation, and cholera in yemen in using modelling to disentangle the relative contributions of zoonotic and anthroponotic transmission: the case of lassa fever estimating initial epidemic growth rates simple framework for real-time forecast in a datalimited situation: the zika virus (zikv) outbreaks in brazil from to as an example model selection and multi-model inference, nd edn mortality burden of ambient fine particulate air pollution in six chinese cities: results from the pearl river delta study the long-term changing dynamics of dengue infectivity in guangdong ambient ozone and influenza transmissibility in hong kong cholera epidemic in yemen, - : an analysis of surveillance data a comparison study of zika virus outbreaks in french polynesia, colombia and the state of bahia in brazil modelling the large-scale yellow fever outbreak in luanda, angola, and the impact of vaccination r: a language and environment for statistical computing transmission dynamics of the influenza a (h n ) pandemic in india: the impact of holiday-related school closure different epidemic curves for severe acute respiratory syndrome reveal similar impacts of control measures strategic decision making about travel during disease outbreaks: a game theoretical approach mitigation of influenza b epidemic with school closures a simple model for complex dynamical transitions in epidemics community-based measures for mitigating the h n pandemic in china quantifying the seasonal drivers of transmission for lassa fever in nigeria mathematical epidemiology modeling infectious diseases in humans and animals viral hemorrhagic fever antibodies in nigerian populations acknowledgements. we appreciate the helpful comments from anonymous reviewers that improved this manuscript. the funding agencies had no role in the design and conduct of the study; collection, management, analysis, and interpretation of the data; preparation, review, or approval of the manuscript; or decision to submit the manuscript for publication.author contributions. sz conceived, carried out the study and drafted the first manuscript. sz and dh discussed the results. all authors revised the manuscript and gave final approval for publication.conflict of interest. the authors declare that they have no competing interests.ethical standards. since no personal data were collected, ethical approval and individual consent were not applicable. key: cord- -h bfdr authors: rasulev, bakhtiyor title: recent developments in d qsar and molecular docking studies of organic and nanostructures date: - - journal: handbook of computational chemistry doi: . / - - - - _ sha: doc_id: cord_uid: h bfdr the development of quantitative structure–activity relationship (qsar) methods is going very fast for the last decades. osar approach already plays an important role in lead structure optimization, and nowadays, with development of big data approaches and computer power, it can even handle a huge amount of data associated with combinatorial chemistry. one of the recent developments is a three-dimensional qsar, i.e., d qsar. for the last two decades, d-osar has already been successfully applied to many datasets, especially of enzyme and receptor ligands. moreover, quite often d qsar investigations are going together with protein–ligand docking studies and this combination works synergistically. in this review, we outline recent advances in development and applications of d qsar and protein–ligand docking approaches, as well as combined approaches for conventional organic compounds and for nanostructured materials, such as fullerenes and carbon nanotubes. the methodology of quantitative structure-activity relationship (qsar) is very well described in various publications (hansch et al. ; kubinyi a, b; eriksson et al. ) . in short, qsar is a method to find correlations and mathematical models for congeneric series of compounds, affinities of ligands to their binding sites, rate constants, inhibition constants, toxicological effect, and many other biological activities, based on structural features, as well as group and molecular properties, such as electronic properties, polarizability, or steric properties (klebe et al. ; hansch et al. ; karelson et al. ; kubinyi a, b; perkins et al. ; isayev et al. ; martin ; rasulev et al. ; puzyn et al. ) . thus, qsar approaches have been used for many types of biological activities to describe correlations for series of drugs and drug candidates (kubinyi a, b; veber et al. ) . in addition, in case of available crystallographic data on the proteins, the qsar models can be developed with the additional information from the three-dimensional ( d) structures of these proteins, interacting with drug candidates, by applying protein-ligand docking data for further qsar analysis, or, if there is no data on d structure of protein, then developing qsar based on three-dimensional features of investigated molecules (moro et al. ; ragno et al. ; gupta et al. ; hu et al. ; sun et al. ; araújo et al. ; ahmed et al. ). the second approach was named as d d qsar qsar approach (wise et al. ; cramer and bunce ; cramer et al. ; clark et al. ). there are also many other multidimensional approaches, including d qsar and d qsar, but all of them are just extension of qsar analysis to a number of conformations (orientations, tautomers, stereoisomers, or protonation states) per molecule, number of concentrations (dosages) per compound, etc (lill ) . in overall, when talking about d qsar, computational chemists usually assume that the qsar analysis takes into account a three-dimensional structure of the compound in minimal energy conformation and builds qsar model based on various d fields generated (kubinyi a, b) . a first similar to d qsar approach was developed by cramer in , which was the predecessor of d approaches called dynamic lattice-oriented molecular modeling system (dylomms) that involves the use of pca to extract vectors from the molecular interaction fields, which are then correlated with biological activities (wise et al. ) . later authors improved this approach and by combining the two existing techniques, grid and pls, has developed a powerful d qsar methodology, so-called comparative molecular field analysis (comfa) (cramer et al. ; clark et al. ). soon after, comfa has become a prototype of d qsar methods (kim et al. ; todeschini and gramatica ; podlogar and ferguson ) . comfa approach was then implemented in the sybyl software (tripos ) from tripos inc. as it was mentioned before, a good and fruitful approach is a combination of molecular docking and d qsar pharmacophore methods (patel et al. ; gupta et al. ; araújo et al. ; ahmed et al. ) . molecular docking and d qsar model are the two potent methods in drug discovery process. thus, virtual screening using d qsar approaches followed by docking has become one of the reputable methods for drug discovery and enhancing the efficiency in lead optimization (oprea and matter ) . the main advantage of this combined approach of d qsar and pharmacophore-based docking is to focus on specific key interaction for protein-ligand binding to improve drug candidates. ameliorate the selection of active compounds; it is optimal to use both methods like molecular docking and d qsar modeling (gopalakrishnan et al. ; klebe ; perola ; pajeva et al. ; yang ) . since the time of development of d qsar approach, a number of papers and methods' developments were published within d qsar methodology. let's briefly list and explain these methods here and then discuss recent developments and applications of these d qsars in the assessment of the properties of biologically active compounds and development of drugs and drug candidates. as can be seen from fig. , the number of publications related to d qsar approach is increasing every year, from to publications in the beginning of s to about publications per year in . it confirms the increasing importance of the method and successful application in many drug design projects. to give some view on a number of d qsar methods developed for the last three decades, we listed below the ligand-based d qsar methods and very short description per each of them. comfa -comparative molecular field analysis is the method which correlates the field values of the structure with biological activities. comfa generates an equation correlating the biological activity with the contribution of interaction energy fields at every grid point (cramer et al. ). the method was developed in the and still one of the most popular ones for d qsar modeling. comsia -comparative molecular similarity indices analysis (comsia) method, where the molecular similarity indices calculated from steric and electrostatic alignment (seal) similarity fields and applied as descriptors to encode steric, electrostatic, hydrophobic, and hydrogen bonding properties (klebe et al. ). this is a development of comfa method and also gets very popular in drug design. grid -this method and program was designed as an alternative to the original comfa approach. it is actually a force field which calculates the interaction energy fields in molecular-field analysis and determines the energetically favorable binding sites on molecules of known structure. the method to some extent is similar to comfa, and it computes explicit nonbonded (or non-covalent) interactions between a molecule of known d structure and a probe (i.e., a chemical group with certain user-defined properties). the probe is located at the sample positions on a lattice throughout and around the macromolecule. the method offers two distinct advantages, one of them is the use of a - potential function for calculating the interaction energies, which is smoother than the - form of the lennard-jones type in comfa, and another advantage is the availability of different types of probes (goodford ) . moreover, the program in addition of computing the regular steric and electrostatic potentials also calculates the hydrogen bonding potential using a hydrogen bond donor and acceptor, as well as the hydrophobic potential using a "dry probe." later on, a water probe was included to calculate hydrophobic interactions (kim et al. ; kim ) . msa -molecular shape analysis (msa) is a ligand-based d qsar method which attempts to merge conformational analysis with the classical hansch approach. the method deals with the quantitative characterization, representation, and manipulation of molecular shape in the construction of a qsar model (hopfinger ) . hasl -inverse grid-based approach represents the shapes of the molecules inside an active site as a collection of grid points (doweyko ) . the methodology of this approach begins with the intermediate conversion of the cartesian coordinates (x, y, z) for superposed set of molecules to a d grid consisting of the regularly spaced points that are ( ) arranged orthogonally to each other, ( ) separated by a particular distance termed as the resolution (which determines the number of grid points representing a molecule), and ( ) all sprawl within the van der waals radii of the atoms in the molecule. thus, the resulting set of points is referred to as the molecular lattice and represents the receptor active site map (like in comfa). quite important that the overall lattice dimensions are dependent on the size of the molecules and the resolution chosen. grind -this is the method that uses grid-independent descriptors (grind) which encode the spatial distribution of the molecular interaction fields (mif) of the studied compounds (pastor et al. ) . in the anchor-grind method (fontaine et al. ) , to compare the mif distribution of different compounds, the user defines a single common position in the structure of all the compounds in the series, so-called anchor point. this anchor point does not provide enough geometrical constrains to align the compounds studied; however, it is used by the method as a common reference point, making it possible to describe the geometry of the mif regions in a more precise way than grind does. the anchor point is particularly easy to assign in datasets having some chemical substituents well known as being crucial for the activity. in the anchor-grind approach, the r groups are described with two blocks of variables: the anchor-mif and the mifmif blocks (fig. ). the first one describes the geometrical distribution of the r mif relative to the anchor point, while the second one describes the geometrical distribution of the mif within each r group. these blocks are obtained after the following steps: (i) every r group is considered as attached to the scaffold, (ii) the anchor point is set automatically on an atom of the scaffold, (iii) a set of mif are calculated with the program grid (goodford ) as well as the shape field implemented in the program almond (cruciani et al. ) , and (iv), as the last step, the blocks of descriptors are computed from the anchor point and the filtered mif. authors also were able to incorporate a molecular shape into the grind descriptors . germ -genetically evolved receptor model (germ) is a method for d qsar and also for constructing d models of protein-binding sites in the absence of a crystallographically established or homology-modeled structure of the receptor (walters and hinds ) . as for many d qsar datasets, the primary requirement for germ is a structure-activity set for which a sensible alignment of realistic conformers has been determined. the methodology is the following: it encloses the superimposed set of molecules in a shell of atoms (analogous to the first layer of atoms in the active site) and allocates these atoms with explicit atom types (aliphatic h, polar h, etc., to match the types of atoms found in the investigated proteins). comma -comparative molecular moment analysis (comma) is one of the unique alignment-independent d qsar methods, which involves the computation of molecular similarity descriptors (similar to comsia) based on the spatial moments of molecular mass (i.e. shape) and charge distributions up to second-order as well as related quantities (silverman and platt ) . combine -comparative binding energy analysis (combine) method was developed to make use of the structural data from ligand-protein complexes, within a d qsar methodology. the method is based on the hypothesis where free energy of binding can be correlated with a subset of energy components calculated from the structures of receptors and ligands in bound and unbound forms (ortiz et al. ; lushington et al. ). comsa -comparative molecular surface analysis (comsa) is a non-grid d qsar method that utilizes the molecular surface to define the regions of the compounds which are required to be compared using the mean electrostatic potentials (meps) (polanski et al. (polanski et al. , . in overall, the methodology proceeds by subjecting the molecules in the dataset to geometry optimization and assigning them with partial atomic charges. afmoc -adaptation of fields for molecular comparison (afmoc) is a close to d qsar method that involves fields derived from the protein environments (i.e. not from the superimposed ligands as in comfa); therefore, it is also known as a "reverse" comfa (dafmoc) approach or protein-dependent d qsar (gohlke and klebe ) . the methodology is the following: a regularly spaced grid is placed into the receptor binding site, followed by mapping of the knowledgebased pair potentials between protein atoms and ligand atom probes onto the grid intersections resulting in the potential fields. thus, based on these potential fields, specific interaction fields are generated by multiplying distance-dependent atomtype properties of actual ligands docked into the active site with the neighboring grid values. in result, these atom-type-specific interaction fields are then correlated with the binding affinities using pls technique, which assigns individual weighting factors to each field value. coria -comparative residue interaction analysis (coria) is a d qsar approach which utilizes the descriptors that describe the thermodynamic events involved in ligand binding, to explore both the qualitative and the quantitative features of the ligand-receptor recognition process. the coria methodology is the following: initially it simply consisted of calculating the nonbonded (van der waals and coulombic) interaction energies between the ligand and the individual active site residues of the receptor that are involved in interaction with the ligand (datar et al. ; dhaked et al. ). by employing the genetic algorithmsupported pls technique (g-pls), these energies then correlated with the biological activities of molecules, along with the other physiochemical variables describing the thermodynamics of binding, such as surface area, lipophilicity, molar refractivity, molecular volume, strain energy, etc. somfa -self-organizing molecular-field analysis, where firstly the mean activity of training set is subtracted from the activity of each molecule to get their mean-centered activity values. the methodology is the following: • a d grid around the molecules with values at the grid points signifying the shape or electrostatic potential is generated. • the shape or electrostatic potential value at every grid point for each molecule is multiplied by its mean-centered activity. • the grid values for each molecule are summed up to give the master grids for each property. • then the so-called somfa property,i descriptors from the master grid values are calculated and correlated with the log-transformed molecular activities (robinson et al. ). knn-mfa -this relatively new method was developed and reported in by ajmani et al. ( ) . knn-mfa is a k-nearest neighbor molecular-field analysis. knn-mfa adopts a k-nearest neighbor principle for generating relationships of molecular fields with the experimentally reported activity. this method utilizes the active analogue principle that lies at the foundation of medicinal chemistry. like many d qsar methods, knn-mfa requires suitable alignment of a given set of molecules. this is followed by generation of a common rectangular grid around the molecules. the steric and electrostatic interaction energies are computed at the lattice points of the grid using a methyl probe of charge c . these interaction energy values are considered for relationship generation and utilized as descriptors to decide nearness between molecules. d-hovaifa -this method based on three-dimensional holographic vector of atomic interaction field analysis (zhou et al. ). initially the holographic vector for d qsar methods was developed by zhou et al. in (zhou et al. ). proceeding from two spatial invariants, namely, atom relative distance and atomic properties on the bases of three common nonbonded (electrostatic, van der waals, and hydrophobic) interactions which are directly associated with bioactivities, d-hovaif method derives multidimensional vectors to represent molecular steric structural characteristics. cmf -this is a recently introduced continuous molecular-field approach (baskin and zhokhova ) . this is a novel approach that consists in encapsulating continuous molecular fields into specially constructed kernels. it is based on the application of continuous functions for the description of molecular fields instead of finite sets of molecular descriptors (such as interaction energies computed at grid nodes) commonly used for this purpose. the feasibility of using molecular-field kernels in combination with the support vector regression (svr) machine learning method to build d qsar models has been demonstrated by the same authors earlier (zhokhova et al. ). authors claim that by combining different types of molecular fields and methods of their approximation, different types of kernels with different types of kernel-based machine learning methods, it is possible not only to present lots of existing methods in chemoinformatics and medicinal chemistry as particular cases within a single methodology but also to develop new approaches aimed at solving new problems (baskin and zhokhova ) . the example of application of this approach is described later in this chapter. phase -this is a flexible system (engine) (dixon et al. ) for common pharmacophore identification and assessment, d qsar model development, and d database creation and searching (within schrodinger suite, schrodinger, llc). it includes some subprograms, for example, ligprep, which attaches hydrogens, converts d structures to d, generates stereoisomers, and, optionally, neutralizes charged structures or determines the most probable ionization state at a user-defined ph. it also includes macromodel conformational search engine to generate a series of d structures that sample the thermally accessible conformational states. for purposes of d modeling and pharmacophore model development, each ligand structure is represented by a set of points in d space, which coincide with various chemical features that may facilitate non-covalent binding between the ligand and its target receptor. phase provides six built-in types of pharmacophore features: hydrogen bond acceptor (a), hydrogen bond donor (d), hydrophobic (h), negative ionizable (n), positive ionizable (p), and aromatic ring (r). in addition, users may define up to three custom feature types (x, y, z) to account for characteristics that don't fit clearly into any of the six built-in categories. to construct a d qsar model, a rectangular grid is defined to encompass the space occupied by the aligned training set molecules. this grid divides space into uniformly sized cubes, typically Å on each side, which are occupied by the atoms or pharmacophore sites that define each molecule. apf -in , totrov m (totrov introduced atomic property fields (apf) for d qsar analysis. apf concept is introduced as a continuous, multicomponent d potential that reflects preferences for various atomic properties at each point in space (fig. ) . the approach is extended to multiple flexible ligand alignments using an iterative procedure, self-consistent atomic property fields by optimization (scapfold). the application of atomic property fields and scapfold for virtual ligand screening and d qsar is tested on published benchmarks. the new method is shown to perform competitively in comparison to current state-of-the-art methods (comfa and comsia). thus, there are studies with comparative analysis of these two methods, phase and catalyst (hypogen). in , evans et al. ( ) provided a comparative study of phase and catalyst methods for their performance in determining d qsars and concluded that the performance of phase is better than or equal to that of catalyst hypogen, with the datasets and parameters used. authors found that within phase, the atom-based grid qsar model generally performed better than the pharmacophore-based grid, and by using overlays from catalyst to build phase grid qsar models, they found evidence that better performance of phase on these datasets was due to the use of the grid technique. in this part of the review, we discuss the new developments in the methods and in applications of d qsars for various chemicals, including nanostructured materials. for the last years, there was mainly improvement of the d qsar approaches which were developed before . as it was discussed above, all these methods as comfa, comsia, grid, somfa, etc., were developed in late s and early s. some of the recently introduced methods are just improvements of old approaches. however, if we take a look at applications, we can see many interesting publications and novel ligand developments which were designed by d qsar and docking methods. recently, a quite interesting study was performed by virsodia et al. ( ) on antitubercular activity of substituted n-phenyl- -methyl- -oxo- -phenyl- , , , -tetrahydropyrimidine- -carboxamides by application of d qsar using comfa and comsia methods. authors synthesized and assessed the antitubercular activity of all investigated compounds followed by comprehensive d qsar modeling. authors were able to get good models with r d . and . , with cross-validated q d . and . , respectively. authors stated that comfa and comsia contours helped them to design some new molecules with improved activity (virsodia et al. ) . another comfa and comsia study was performed by ravichandran et al. ( ) by analysis of anti-hiv activity of , , -thiazolidine derivatives. authors were able to get good models by comfa and comsia with r values . and . , respectively. the predictive model was evaluated using a test set comprising of molecules, and the predicted r values of comfa and comsia models were . and . , respectively. with the use of comsia method, kumar et al. ( ) were able to investigate novel bignelli dihydropyrimidines with potential anticancer activity. the developed model based on compounds showed a good statistical data -for training set q d . , while for the test set r d . . raparti et al. in (raparti et al. ) reported a study based on a novel knn-mfa d qsar which was discussed above, where authors synthesized, assessed for antimycobacterial activity, and investigated by d and d qsar approaches a series of ten compounds ( -(morpholin- -yl)-n -(arylidene)benzohydrazides). authors were able to get satisfactory for this size of dataset statistical results for d qsar model against m. tuberculosis (raparti et al. ), with r d . and q d . , respectively. another knn-mfa d qsar study was conducted and published by kishore jha et al. in (jha et al. . authors evaluated the antimicrobial activity of compounds by knn-mfa combined with various selection procedures. as selection methods, authors were using simulated annealing (sa), genetic algorithms (ga), and stepwise (sw) forward-backward methods. the developed model showed satisfactory results for this kind of studies, with q d . and r pred d . . authors concluded that the d qsar study has shown that less electronegative substituent would be favorable for the activity, and therefore the future molecules should be designed with less electronegative group to result in potentially active molecules. thus, recently, araújo et al. ( ) studied acetylcholine inhibitors (acheis) by application of combined approach, so-called receptor-dependent d qsar (rd d qsar) where they investigated a series of benzylpiperidine inhibitors of human acetylcholinesterase. they received two models with r d . , q d . and r d . , q d . , which were validated by a combined ga-pls approach. based on those models, authors have proposed four new benzylpiperidine derivatives and predicted the pic for each molecule. the good predicted potency of one of the benzylpiperidine derivatives indicated a promising potency for this candidate as a new huache inhibitor (araújo et al. ) . in another similar study, in gupta et al. ( conducted an interesting combined study with protein-ligand docking-based d qsar study of hiv- integrase inhibitors. they were using protein-ligand docking to identify a potential binding mode for inhibitors at hiv- in active site, and best docked conformation of certain molecule was used as a template for alignment. the docking was followed by comfa and comsia modeling, and authors developed very good models with r cv values of . and . , respectively, and non-cross-validated ones r ncv d . and . . this combined docking-based d qsar methodology showed really good predictive abilities and can be employed further in the development of better inhibitors for various proteins. one more study is worth to discuss where authors applied a combination of docking and d qsar to reveal the most important structural factors for the activity. here, hu et al. ( ) applied a receptor-and ligand-based d qsar study for a series of non-nucleoside hiv- reverse transcriptase inhibitors ( amino- -arylsulfonylbenzonitriles and their thio and sulfinyl congeners). authors were applying docking simulations to position the inhibitors into rt active site to determine the most probable binding mode and most reliable conformations. this complex receptor-based and ligand-based alignment procedure and different alignment modes allowed authors to obtain reliable and predictive comfa and comsia models with cross-validated q value of . and . , respectively. authors concluded that the comfa steric and comsia hydrophobic fields support the idea that bulkier and hydrophobic groups are favorable to bioactivity in the -and -positions of the b (benzene)-ring. at the same time, these groups are unfavorable in the -position. also, the comsia h-bond donor and acceptor fields suggest that the sulfide and sulfone inhibitors are more active than the sulfoxide ones due to h-bonding with protein residues. it is good to mention here another interesting study where combination of methods is used, including molecular docking and d qsar to develop a predictive qsar model. moro et al. ( ) suggested the use a combination of molecular electrostatic potential (mep) surface properties (autocorrelation vectors) with the conventional partial least-square (pls) analysis to produce a robust ligand-based d structure-activity relationship (automep/pls). they applied this approach to predict human a receptor antagonist activities. first of all, the approach was suggested as an efficient and alternative pharmacodynamic-driven filtering method for small-size virtual libraries. for this, authors generated a small-sized combinatorial library ( compounds) that was derived from the scaffold of the known human a antagonist pyrazolo-triazolo-pyrimidines (moro et al. ) . this is another successful example of combined approach of docking and d qsar to investigate and design active analogue compounds. authors were using multidock code that is part of moe suite (molecular operating environment (moe) ) to get a conformational sampling and then calculate interaction energies using mmff (halgren ) and use it for further steps. the meps were derived from a classical point charge model: the electrostatic potential for each molecule is obtained by moving a unit positive point charge across the van der waals surface, and it is calculated at various points j on this surface (moro et al. ) . authors were able to test the approach by synthesizing several predicted potent compounds, and they found that all the newly synthesized compounds are correctly predicted as potent human a antagonists (moro et al. ) . as a continuation of development of pharmacophore-and docking-based methods for qsar, the novel phase code was developed. this updated code then was used by amnerkar and bhusari ( ) to investigate by d qsar approach the anticonvulsant activity of some prop- -eneamido and -acetyl-pyrazolin derivatives compound aligned to the pharmacophore for which blue indicates nitrogen, yellow refers to sulfur, green indicates chlorine, gray indicates carbon, and white refers to hydrogen (reproduced with permission from reference (amnerkar and bhusari ). copyright elsevier, ) of aminobenzothiazole. they received a statistically significant d qsar model with r of . and q of . . the model was analyzed in order to understand the trends of investigated molecules for their anticonvulsant properties. authors found the influence of electron withdrawing, hydrogen bond donor, and negative/positive ionic and hydrophobic groups to anticonvulsant activity. authors believe that the derived d qsar as well as clues for possible structural modifications will be of interest and significance for the strategic design of more potent molecules in the benzothiazoles as anticonvulsant agents. the pharmacophore hypothesis generated from phase-based d qsar analysis can be seen in fig. . another phase application for d qsar study is published by pulla et al. ( ) . authors applied a d qsar approach to investigate silent mating-type information regulation homologue (sirt ) which is the homologous enzyme of silent information regulator- gene in yeast. sirt was believed to be overexpressed in many cancers (prostate, colon) and inflammatory disorders (rheumatoid arthritis); that is why it has good therapeutic importance. authors conducted both structurebased and ligand-based drug design strategies with utilizing high-throughput virtual screening of chemical databases. then an energy-based pharmacophore was generated using the crystal structure of sirt bound with a small molecule inhibitor and compared with a ligand-based pharmacophore model that showed four similar features. a d qsar model was developed and applied to generated structures. among the designed compounds, lead emerged as a promising sirt inhibitor with ic of . m and, at nanomolar concentration ( nm), attenuated the proliferation of prostate cancer cells (lncap) (pulla et al. ) . the d qsar model was developed using phase . module in maestro . software package developed by schrodinger, llc (dixon et al. ) . docking studies were executed using glide . module (halgren et al. ) . authors were validating the pharmacophore model by set composed of compounds, consisting of decoys and known inhibitors. the drug-like decoy set of compounds was obtained from the glide module (halgren et al. ) . a final d qsar model was developed based on dataset of molecules reported as sirt inhibitors in various literatures. to develop qsar model, phase module relied on pls regression applied to a large set of binary-valued variables. each independent variable in the model originated from the grid of cubic volume elements spanning the space occupied by the training set ligands. each training ligand in the training set was represented by binary code consisting of set of bit values ( or ) indicating the volume of elements occupied by van der waals model of the ligand. authors were able to get a very good d qsar model with r d . , q d . , and r ext d . . a validated d qsar model (for adhrr ) authors used to generate contour maps could help in understanding the importance of functional groups at specific positions toward biological activity. these insights could be known by comparing the contour maps of the most and least active compounds, as shown in fig. represented in pulla et al. ( ) . the blue and red cubes indicated the favorable and unfavorable regions, respectively, of the hydrogen bond donor effect, while light-red and yellow cubes indicated favorable and unfavorable regions, respectively, of the hydrophobic effect, and the cyan and orange cubes indicated favorable and unfavorable regions, respectively, of the electron-withdrawing effect. from fig. a , it can be seen that the blue favorable regions of the hydrogen bond donor effect were nearer to the donor feature (d ) of the active molecule; however, it could also be observed that blue boxes were also concentrated at the amide group beside thiophen, thus illustrating that additional donor groups at these regions (blue cubes) could increase biological activity. at the same time, in the inactive molecule, red unfavorable boxes were observed around the donor feature (d ), inferring the biological inactiveness of the molecule. in the case of the hydrophobic effect, the light-red color cubes were seen surrounding the hydrophobic feature (h , piperidine) of the active molecule, whereas the presence of few yellow unfavorable cubes indicated that these hydrophobic groups were not in the right position in the inactive molecule, illustrating the weak biological activity. next, in the case of the electron-withdrawing effect of the active molecule, the favorable cyan cubes were seen around the acceptor feature (a ), and cyan cubes were also seen near the pyrimidine ring. it inferred that the acceptor features near the pyrimidine ring could further increase the bioactivity of the molecule. however, in the case of the inactive molecule, mostly unfavorable orange cubes were observed around the acceptor feature (a ), illustrating the importance of the electron-withdrawing group in the activity of lead molecules. thus, another new combined docking-based d qsar study (sun et al. ) was published with the analysis of influenza neuraminidase inhibitors. the study was based on novel d-hovaifa method which is based on three-dimensional holographic vector of atomic interaction field analysis (zhou et al. ). as it was mentioned above, initially the holographic vector for d qsar was developed by (zhou et al. ). the method uses atomic relative distance and atomic properties on the bases of three common nonbonded (electrostatic, van der waals, and hydrophobic) interactions which are directly associated with bioactivities, and then d-hovaif method derives multidimensional vectors to represent molecular steric structural characteristics for further d qsar analysis. similarly to previous study, authors conducted a docking study to find the best docking pose and template for alignment. then authors were able to get good models for a large dataset of compounds and received the following correlation coefficients, r d . and r cv d . . authors claim that d-hovaifa can be applicable to molecular structural characterization and bioactivity prediction. in addition, it was showed that hovaifa and docking results are corresponding (sun et al. ) , which illustrates that hovaifa is an effective methodology for characterization of complex interactions of drug molecules. one more docking-based d qsar study was published in by sakkiah et al. ( ) , where authors conducted d qsar pharmacophore-based virtual screening and molecular docking for the identification of potential hsp inhibitors. authors were using hypo and hypogen (li et al. ) d-based pharmacophore models. based on the training set of compounds, they were able to develop a good model using pharmacophore generation module in discovery studio (accelrys) and then apply it for test set of compounds. for predicting activity, the correlation coefficients of the model for training and test sets were . and . , respectively. authors then applied the model to virtual screening of about , compounds (maybridge and scaffold databases) and finally selected compounds for docking studies. finally, selected compounds were reported that were showing high activity based on d qsar model and docking analysis. the developed hypogen pharmacophore model that was used for virtual screening of , compounds from the databases is represented in fig. . recently introduced and discussed previously the cmf approach for d qsar analysis was successfully applied for several datasets (baskin and zhokhova ) . authors applied cmf approach to build d qsar models for eight datasets through the use of five types of molecular fields (the electrostatic, steric, hydrophobic, hydrogen bond acceptor and donor ones). the d qsar models were developed for the following datasets: angiotensin converting enzyme (ace) inhibitors, acetylcholinesterase (ache) inhibitors, ligands for benzodiazepine receptors (bzr), cyclooxygenase- (cox- ) inhibitors, dihydrofolatereductase (dhfr) inhibitors, glycogen phosphorylase b (gpb) inhibitors, thermolysin (ther) inhibitors, and thrombine (thr) inhibitors. authors were able to get good models and then compare statistical characteristics of the developed models with the same characteristics built earlier for the same datasets using the most popular d qsar methods, comfa and comsia, based on the use of molecular fields. almost for all cases authors received better statistics than in original works. for example, for ace inhibitors, cmf approach showed q d . , while comfa and comsia showed . and . , respectively. d qsar for ache inhibitors showed for cmf q d . , while for comfa and comsia, it was . and . , respectively. d qsar for dhfr inhibitors showed for cmf q d . , while for comfa and comsia, it was only . and . , respectively. other datasets also showed better results. the only one dataset, for bzr receptors, showed not so high values (q d . ) but comparable with previous data received by comfa and comsia - . and . , respectively. as follows from the results presented in this paper, this particular implementation of the cmf approach provides an appealing alternative to the traditional lattice-based methodology. this method provides either comparable or enhanced predictive performance in comparison with state-of-theart d qsar methods, such as comfa and comsia. authors also discussed advantages and disadvantages of this approach. the potential advantages of this approach result from the ability to approximate electronic molecular structures with any desirable accuracy level, the ability to leverage the valuable information contained in partial derivatives of molecular fields (otherwise lost upon discretization) to analyze models and enhance their predictive performance, the ability to apply integral transforms to molecular fields and models, etc. the most attractive features of the cmf approach are its versatility and universality. at the same time, of the most serious limitations of the cmf approach, at least in its present form, comes from the mere nature of kernel-based machine learning methods. the amount of computational resources needed to calculate a kernel matrix scales as a square of the number of compounds in the training set. the mean amount of computational resources needed to calculate each element of a kernel matrix also scales as a square of the average number of atoms in molecules. as a result, it becomes impractical to build d qsar models using a training set with more than medium-sized compounds (baskin and zhokhova ) . another interesting combination approach, shih et al. ( ) in proposed a combination of d qsar methods in order to get better predictivity for the dataset. authors proposed for the first time a combination approach to integrate the pharmacophore (phmodel) (taha et al. ) , comfa, and comsia models for b-raf (raf family of serine/threonine kinases). the phmodel was implemented by the program accelrys discovery studio . . first, authors established ten phmodels and used them to align diverse inhibitor structures for generating the comfa and comsia models. then the partial least-square (pls) method was used and known b-raf inhibitors to validate the prediction ability of comfa and comsia models. finally, the goodness of hit (gh) test score was used as a benchmark for appraising the prediction ability of comfa and comsia models to screen a compound database. thus, ten phmodels were generated based on the training set inhibitors. each phmodel included four features: hydrogen bond acceptor (a), hydrogen bond donor (d), hydrophobic (hy), and ring aromatic (ra). the correlation coefficients for ten phmodels were very good and ranged from . to . . authors claim that this approach could be applied to screen inhibitor databases, optimize inhibitor structures, and identify novelty potent or specific inhibitors (shih et al. ) . one more combination study for raf inhibitors is worth to mention, was performed by yang et al. ( ) which was applied as a combination of docking, molecular dynamics (md), molecular mechanics poisson-boltzmann surface area (mm/pbsa) calculations, and d qsar analysis to investigate the detailed binding mode between b-raf kinases with the series of inhibitors and also to find the key structural features affecting the inhibiting activities. considering the difficulty in the accurate estimation of electrostatic interaction, the qm-polarized ligand docking and gbsa rescoring were applied to predict probable poses of these inhibitors bound into the active site of b-raf kinase. to obtain the rational conformation for developing d qsar models, authors applied the docking-based conformation selection strategy. moreover, the detailed interactions were analyzed on the basis of the results from md simulation and the free energy calculation for two inhibitors with much difference in their activity. authors investigated b-raf inhibitors and developed comfa and comsia models with r d . and . , respectively. in result, the structure-based d qsar models provided a further structural analysis and modifiable information for understanding the sars of these inhibitors. the important hydrophobic property of the -substitution of b-ring was required to be type inhibitors. the five substitutable positions of the c-ring could be further modified. authors concluded that the results obtained from the combined computational approach will be helpful for the rational design of novel type raf kinase inhibitors. recently, a group of computational scientists is proposed to apply a proteinprotein interaction (ppi) analysis to target small molecules. since currently in worlds life science, research is going on the booming of interactome studies, a lot of interactions can be measured in a high-throughput way, taking into account that large-scale datasets are already available. studies show that many different types of interactions can be potential drug targets. this boom of hts studies greatly broadens the drug target search space, which makes the drug target discovery difficult. in this case, computational methods are highly desired to efficiently provide candidates for further experiments and hold the promise to greatly accelerate the discovery of novel drug targets. thus, wang et al. ( ) published a study where they suggested a new method, where inhibition of protein-protein interaction (ppi) analysis offered as a promising source to improve the specificity of drugs with fewer adverse side effects. they proposed a machine learning method to predict ppi targets in a genomic-wide scale. authors developed a computational method, named as preppitar (wang et al. ) , to predict ppis as drug targets by uncovering the potential associations between drugs and ppis (fig. ) . authors investigated the databases and manually constructed a gold-standard positive dataset for drug and ppi interactions. their effort leads to a dataset with associations among ppis and fda-approved drugs and allowed them to build models and learn the association rules from the data. also, authors were able to characterize drugs by profiling in chemical structure, drug atc-code annotation, and side-effect space and represent ppi similarity by a symmetrical s-kernel based on protein amino acid sequence. at the end, a support vector machine (svm) is used to predict novel associations between drugs and ppis. the preppitar method was validated on the well-established gold-standard dataset. authors found that all chemical structures, drug atc code, and side-effect information are predictive for ppi target. authors claim that preppitar can serve as a useful tool for ppi target discovery and provide a general heterogeneous data-integrative framework. preppitar applies the kernel fusion method to integrate multiple information about drug, including chemical structure, atc code, and drug side effect to detect the interactions between drugs and ppis. (b) collecting known associations between drugs and ppis as gold-standard positives in a bipartite graph. (c) calculating drug-drug and ppi-ppi similarity metrics, where t i ; i d ; ; ; are the sequence similarity among proteins. (d) relating the similarity among drugs and similarity among ppis by kronecker product kernel and applying svm-based algorithm to predict the unknown associations between drugs and ppis (reproduced with permission from reference (wang et al. ). copyright oxford university press, nanomaterials are becoming an important component of the modern life and have been the subject of increasing number of investigations involving various areas of natural sciences and technology. however, theoretical modeling of physicochemical and biological activity of these species is still very scarce. the prediction of properties and activities of "classical" substances via correlating with molecular descriptors is a well-known procedure, by application qsar and d qsar methods. in spite of this, the application of qsar for the nanomaterials is a very complicated task, because of "nonclassical" structure of these materials. here, we would like to show first applications of the d qsar and docking methods for nanostructured materials, which are nevertheless possible and can be useful in predicting their various properties and activities (toxicity). thus, one of the first d qsar studies for nanostructured materials was provided in . durdagi et al. a have investigated novel fullerene analogues as potential hiv pr inhibitors. it was the first work where authors analyzed nanostructured compounds for anti-hiv activity using protein-ligand docking and d qsar approaches. moreover, authors conducted md simulations of ligand-free and the inhibitor bound hiv- pr systems to complement some previous studies and to provide proper input structure of hiv- pr in further docking simulations. then, five different combinations of stereoelectronic fields of d qsar/comsia models were obtained from the set of biologically evaluated and computationally designed fullerene derivatives (where training set d and test set d ) in order to predict novel compounds with improved inhibition effect. the best d qsar/comsia model yielded a cross-validated r value of . and a non-cross-validated r value of . . authors stated that the derived model indicated the importance of steric ( . %), electrostatic ( . %), h-bond donor ( . %), and h-bond acceptor ( . %) contributions (fig. ). in addition, the derived contour plots together with applied de novo drug design were then used as pilot models for proposing the novel analogues with enhanced binding affinities. interestingly, the investigated by authors the nanostructured compounds have triggered the interest of medicinal chemists to look for novel fullerene-type hiv- pr inhibitors possessing higher bioactivity. later this year, authors published a second study for the same type of fullerenebased nanomaterials (durdagi et al. b) . the same group published in another study for fullerene derivatives but functionalized by amino acids (durdagi et al. ). authors used in silico screening approach in order to propose potent fullerene analogues as anti-hiv drugs. two of the most promising derivatives showing significant binding scores were subjected to biological studies that confirmed the efficacy of the new compounds. the results showed that new leads can be discovered possessing higher bioactivity. authors used docking approach together with md simulations to get the best hits during the virtual screening. (durdagi et al. a ). copyright elsevier, in , the same group provided further analysis to design better anti-hiv fullerene-based inhibitors (tzoupis et al. ) . in this study authors employed a docking technique, two d qsar models, md simulations and the mm-pbsa method. in particular, authors investigated ( ) hydrogen bonding (h-bond) interactions between specific fullerene derivatives and the protease, ( ) the regions of hiv- pr that play a significant role in binding, ( ) protease changes upon binding, and ( ) various contributions to the binding free energy, in order to identify the most significant of them. the comfa and comsia methods were applied too, to build d qsar models, where good correlation coefficients were received, for both methods, r d . and . , respectively. authors claim that the computed binding free energies are in satisfactory agreement with the experimental results. another group published in a study that conducted a comprehensive investigation of fullerene analogues by combined computational approach including quantum chemical, molecular docking, and d descriptors-based qsar (ahmed et al. ) . authors stated that the protein-ligand docking studies and improved structure-activity models have been able both to predict binding affinities for the set of fullerene-c derivatives and to help in finding mechanisms of fullerene derivative interactions with human immunodeficiency virus type aspartic protease, hiv- pr. protein-ligand docking revealed several important molecular fragments that are responsible for the interaction with hiv- pr (fig. ). in addition, a density functional theory method has been utilized to identify the optimal geometries and predict physicochemical parameters of studied compounds. the five-variable ga-mlra-based model showed the best predictive ability (r train d . and r test d . ), with high internal and external correlation coefficients. calvaresi and zerbetto ( ) published a study where they investigated a fullerene binding with a set of proteins. authors investigated about proteins that are known to modify their activity upon interaction with c . the set was examined using patchdock (schneidman-duhovny et al. ) software with an algorithm that appraises quantitatively the interaction of c and the surface of each protein. the redundancy of the set allowed them to establish the predictive power of the approach that finds explicitly the most probable site where c docks on each protein. about % of the known fullerene-binding proteins fall in the top % of scorers. the close match between the model and experiments vouches for the accuracy of the model and validates its predictions. authors identified the sites of docking and discussed them in view of the existing experimental data available for protein c interaction. in addition, authors identified new proteins that can interact with c and discussed for possible future applications as drug targets and fullerene derivative bioconjugate materials. later, calvaresi and zerbetto ( ) published another study, where they investigated the binding of fullerene c with proteins. they one more time confirmed that hydrophobic pockets of certain proteins can accomodate a carbon cage either in full or in part. since the identification of proteins that are able to discriminate between different cages is an open issue, they were interested in investigating much larger library than in calvaresi and zerbetto ( ) . prediction of candidates is achieved with an inverse docking procedure that accurately accounts for (i) van der waals interactions between the cage and the protein surface, (ii) desolvation free energy, (iii) shape complementarity, and (iv) minimization of the number of steric clashes through conformational variations. a set of protein structures is divided into four categories that either select c or c (p-c or p-c ) and either accommodate the cages in the same pocket or in different pockets. thus, authors also confirmed the agreement with the experiments, where the kcsa potassium channel is predicted to have one of the best performances for both cages. recently, in xavier et al. (esposito et al. ) published a qsar study of decorated carbon nanotube investigation for toxicity using d fingerprints. in this study, authors proposed detailed mechanisms of action, relating to nanotoxicity, for a series of decorated (functionalized) carbon nanotube complexes based on previously reported qsar models. possible mechanisms of nanotoxicity for six endpoints (bovine serum albumin, carbonic anhydrase, chymotrypsin, hemoglobin along with cell viability, and nitrogen oxide production) have been extracted from the corresponding optimized qsar models. the molecular features relevant to each of the endpoint respective mechanism of action for the decorated nanotubes are also discussed in the study. based on the molecular information contained within the optimal qsar models for each nanotoxicity endpoint, either the decorator attached to the nanotube is directly responsible for the expression of a particular activity, irrespective of the decorator's d geometry and independent of the nanotube, or those decorators having structures that place the functional groups of the decorators as far as possible from the nanotube surface most strongly influence the biological activity. a docking study, together with comprehensive dft analysis was conducted by saikia et al. ( ) . authors made a simulation to analyze the interaction of nanomaterials with biomolecular systems, where they performed density functional calculations on the interaction of pyrazinamide (pza) drug with functionalized single-wall cnt (fswcnt) as a function of nanotube chirality and length, followed by docking simulation of fswcnt with pnca protein. the functionalization of pristine swcnt that facilitates in enhancing the reactivity of the nanotubes and formation of such type of nanotube-drug conjugate is thermodynamically feasible. docking studies predicted the plausible binding mechanism and suggested that pza loaded fswcnt facilitates in the target-specific binding of pza within the protein following a lock and key mechanism. authors noticed that no major structural deformation in the protein was observed after binding with cnt, and the interaction between ligand and receptor is mainly hydrophobic in nature. authors anticipate that these findings may provide new routes toward the drug delivery mechanism by cnts with long-term practical implications in tuberculosischemotherapy. in another study, turabekova et al. ( ) published a comprehensive study of carbon nanotube and pristine fullerene interactions with toll-like receptors (tlrs), which are responsible for immune response. having experimental data on hands and conducting comprehensive protein-ligand investigation, authors were able to show that cnt and fullerenes can bind to certain tlrs. authors suggested a hypothetical model providing the potential mechanistic explanation for immune and inflammatory responses observed upon exposure to carbon nanoparticles. specifically, authors performed a theoretical study to analyze cnt and c fullerene interactions with the available x-ray structures of tlr homo-and heterodimer extracellular domains. this assumption was based on the fact that similar to the known tlr ligands, both cnts and fullerenes induce, in cells, the secretion of certain inflammatory protein mediators, such as interleukins and chemokines. these proteins are observed within inflammation downstream processes resulting from the ligand molecule-dependent inhibition or activation of tlr-induced signal transduction. the computational studies have shown that the internal hydrophobic pockets of some tlrs might be capable of binding small-sized carbon nanostructures ( , armchair swcnts containing carbon atom layers and c fullerene). high binding scores and minor structural alterations induced in tlr ectodomains upon binding c and cnts further supported the proposed hypothesis (fig. ) . additionally, the proposed hypothesis is strengthened by the indirect experimental fig. , cnt-bound tlr /tlr ecds: (a) , cnt is bound to the tlr and tlr ecd interface dimerization area, (b) aligned structures of tlr ecds before (green carbon atoms) and after (blue carbon atoms) the impact opls refinement upon binding , cnts. the orientation of two parallel entrance loops and the side chains of hydrophobic phe , phe , and leu preventing the nanotube from intrusion are shown to be optimized (reproduced with permission from reference (turabekova et al. ) . copyright, royal society of chemistry, ) findings indicating that cnts and fullerenes induce an excessive expression of specific cytokines and chemokines (i.e., il- and mcp ). later, this kind of interaction was confirmed by md simulation provided by mozolewska et al. ( ) . in this study, authors made an attempt to determine if the nanotubes could interfere with the innate immune system by interacting with tlrs. for this purpose, authors used the following tlr structures downloaded from the rcsb protein data bank: tlr ( a c), tlr /md ( fxi), tlr ( v ), tlr ( a z), and the complexes of tlr /tlr ( z x) and tlr /tlr ( a ). the results of steered molecular dynamics (smd) simulations have shown that nanotubes interact very strongly with the binding pockets of some receptors (e.g., tlr ), which results in their binding to these sites without substantial use of the external force. in this chapter, we discussed d qsar and protein-ligand docking methods, recent applications of them for conventional organic compounds design and for nanostructured materials. despite of all pitfalls, the d qsar approach confirmed the importance and value in drug design and medicinal chemistry. moreover, the combination of d qsar approach with other techniques, including proteinligand docking study gives much better improvement in predictions of biologically active compounds and drug candidates. the development of methods for d qsar still continues, giving improved predictions for conventional organic compounds. thus, we believe that d qsar methods in the near future will be able to encode and model various organic and nanomaterials for important biological and physicochemical property improvement. receptor-and ligand-based study of fullerene analogues: comprehensive computational approach including quantum-chemical, qsar and molecular docking simulations three-dimensional qsar using the k-nearest neighbor method and its interpretation synthesis, anticonvulsant activity and d-qsar study of some prop- -eneamido and -acetyl-pyrazolin derivatives of aminobenzothiazole receptor-dependent (rd) d-qsar approach of a series of benzylpiperidine inhibitors of human acetylcholinesterase (huache) the continuous molecular fields approach to building d-qsar models baiting proteins with c fullerene sorting proteins comparative molecular field analysis (comfa). . toward its use with d-structural databases the dylomms method: initial results from a comparative study of approaches to d qsar comparative molecular field analysis (comfa). . effect of shape on binding of steroids to carrier proteins comparative residue interaction analysis (coria): a d-qsar approach to explore the binding contributions of active site residues with ligands exploring the binding of hiv- integrase inhibitors by comparative residue interaction analysis (coria) phase: a new engine for pharmacophore perception, d qsar model development, and d database screening. . methodology and preliminary results the hypothetical active site lattice. an approach to modelling active sites from data on inhibitor molecules computational design of novel fullerene analogues as potential hiv- pr inhibitors: analysis of the binding interactions between fullerene inhibitors and hiv- pr residues using d qsar, molecular docking and molecular dynamics simulations d qsar comfa/comsia, molecular docking and molecular dynamics studies of fullerene-based hiv- pr inhibitors in silico drug screening approach for the design of magic bullets: a successful example with anti-hiv fullerene derivatized amino acids methods for reliability and uncertainty assessment and for applicability evaluations of classification-and regression-based qsars exploring possible mechanisms of action for the nanotoxicity and protein binding of decorated nanotubes: interpretation of physicochemical properties from optimal qsar models d qsar methods: phase and catalyst compared incorporating molecular shape into the alignment-free grid-in dependent descriptors anchor-grind: filling the gap between standard d qsar and the grid-in dependent descriptors drugscore meets comfa: adaptation of fields for molecular comparison (afmoc) or how to tailor knowledge-based pair-potentials to a particular protein a computational procedure for determining energetically favorable binding sites on biologically important macromolecules a virtual screening approach for thymidine monophosphate kinase inhibitors as antitubercular agents based on docking and pharmacophore models docking-based d-qsar study of hiv- integrase inhibitors merck molecular force field. i. basis, form, scope, parameterization, and performance of mmff glide: a new approach for rapid, accurate docking and scoring. . enrichment factors in database screening exploring qsar a qsar investigation of dihydrofolate reductase inhibition by baker triazines based upon molecular shape analysis receptor-and ligand-based d-qsar study for a series of non-nucleoside hiv- reverse transcriptase inhibitors structure-toxicity relationships of nitroaromatic compounds design, synthesis and biological evaluation of , , -oxadiazole derivatives quantum-chemical descriptors in qsar/qspr studies thermodynamic aspects of hydrophobicity and biological qsar a critical review of recent comfa applications virtual ligand screening: strategies, perspectives and limitations. drug discovery today molecular similarity indices in a comparative analysis (comsia) of drug molecules to correlate and predict their biological activity qsar and d qsar in drug design part : methodology. drug discovery today qsar and d qsar in drug design part : applications and problems novel biginelli dihydropyrimidines with potential anticancer activity: a parallel synthesis and comsia study hypogen: an automated system for generating d predictive pharmacophore models multi-dimensional qsar in drug discovery whither combine? new opportunities for receptor-based qsar let's not forget tautomers chemical computing group inc. sherbooke st. west, suite # combined target-based and ligand-based drug design approach as a tool to define a novel d-pharmacophore model of human a adenosine receptor antagonists the application of a d-qsar (automep/pls) approach as an efficient pharmacodynamic-driven filtering method for small-sized virtual library: application to a lead optimization of a human a adenosine receptor antagonist preliminary studies of interaction between nanotubes and toll-like receptors integrating virtual screening in lead discovery prediction of drug binding affinities by comparative binding energy analysis combined pharmacophore modeling, docking, and d qsar studies of abcb and abcc transporter inhibitors grid-independent descriptors (grind): a novel class of alignment-independent three-dimensional molecular descriptors d qsar and molecular docking studies of benzimidazole derivatives as hepatitis c virus ns b polymerase inhibitors quantitative structure-activity relationship methods: perspectives on drug discovery and toxicology minimizing false positives in kinase virtual screens qsar and comfa: a perspective on the practical application to drug discovery the comparative molecular surface analysis (comsa)-a nongrid d qsar method by a coupled neural network and pls system: predicting p k a values of benzoic and alkanoic acids modeling robust qsar energy-based pharmacophore and three-dimensional quantitative structure-activity relationship ( d-qsar) modeling combined with virtual screening to identify novel small-molecule inhibitors of silent mating-type information regulation homologue (sirt ) using nano-qsar to predict the cytotoxicity of metal oxide nanoparticles docking and -d qsar studies on indolyl aryl sulfones. binding mode exploration at the hiv- reverse transcriptase non-nucleoside binding site and design of highly active n-( -hydroxyethyl) carboxamide and n-( -hydroxyethyl) carbohydrazide derivatives novel -(morpholin- -yl)-n -(arylidene) benzohydrazides: synthesis, antimycobacterial activity and qsar investigations qsar modeling of acute toxicity on mammals caused by aromatic compounds: the case study using oral ld for rats predicting anti-hiv activity of , , -thiazolidinone derivatives: d-qsar approach self-organizing molecular field analysis: a tool for structure-activity studies density functional and molecular docking studies towards investigating the role of single-wall carbon nanotubes as nanocarrier for loading and delivery of pyrazinamide antitubercular drug onto pnca protein d qsar pharmacophore based virtual screening and molecular docking for identification of potential hsp inhibitors patchdock and symmdock: servers for rigid and symmetric docking development of novel d-qsar combination approach for screening and optimizing b-raf inhibitors in silico comparative molecular moment analysis (comma): d-qsar without molecular superposition docking and d-qsar studies of influenza neuraminidase inhibitors using three-dimensional holographic vector of atomic interaction field analysis combining ligandbased pharmacophore modeling, quantitative structure-activity relationship analysis and in silico screening for the discovery of new potent hormone sensitive lipase inhibitors d qsar in drug design atomic property fields: generalized d pharmacophoric potential for automated ligand superposition, pharmacophore elucidation and d qsar immunotoxicity of nanoparticles: a computational study suggests that cnts and c fullerenes might be recognized as pathogens by toll-like receptors binding of novel fullerene inhibitors to hiv- protease: insight through molecular dynamics and molecular mechanics poisson-boltzmann surface area calculations molecular properties that influence the oral bioavailability of drug candidates synthesis, screening for antitubercular activity and d-qsar studies of substituted n-phenyl- -methyl- -oxo- -phenyl- , , , -tetrahydro-pyrimidine- -carboxamides genetically evolved receptor models: a computational approach to construction of receptor models computational probing protein-protein interactions targeting small molecules progress in three-dimensional drug design: the use of real-time colour graphics and computer postulation of bioactive molecules in dylomms pharmacophore modeling and applications in drug discovery: challenges and recent advances molecular dynamics simulation, free energy calculation and structure-based d-qsar studies of b-raf kinase inhibitors method of continuous molecular fields in the search for quantitative structure-activity relationships three dimensional holographic vector of atomic interaction field ( d-hovaif) key: cord- -s knxdne authors: perra, nicola; gonçalves, bruno title: modeling and predicting human infectious diseases date: - - journal: social phenomena doi: . / - - - - _ sha: doc_id: cord_uid: s knxdne the spreading of infectious diseases has dramatically shaped our history and society. the quest to understand and prevent their spreading dates more than two centuries. over the years, advances in medicine, biology, mathematics, physics, network science, computer science, and technology in general contributed to the development of modern epidemiology. in this chapter, we present a summary of different mathematical and computational approaches aimed at describing, modeling, and forecasting the diffusion of viruses. we start from the basic concepts and models in an unstructured population and gradually increase the realism by adding the effects of realistic contact structures within a population as well as the effects of human mobility coupling different subpopulations. building on these concepts we present two realistic data-driven epidemiological models able to forecast the spreading of infectious diseases at different geographical granularities. we conclude by introducing some recent developments in diseases modeling rooted in the big-data revolution. historically, the first quantitative attempt to understand and prevent infectious diseases dates back to when bernoulli studied the effectiveness of inoculation against smallpox [ ] . since then, and despite some initial lulls [ ] , an intense research activity has developed a rigorous formulation of pathogens' spreading. in this chapter, we present different approaches to model and predict the spreading of infectious diseases at different geographical resolutions and levels of detail. we focus on airborne illnesses transmitted from human to human. we are the carriers of such diseases. our contacts and mobility are the crucial ingredients to understand and model their spreading. interestingly, the access to large-scale data describing these human dynamics is a recent development in epidemiology. indeed, for many years only the biological roots of transmission were clearly understood, so it is not surprising that classical models in epidemiology neglect realistic human contact structures or mobility in favor of more mathematically tractable and simplified descriptions of unstructured populations. we start our chapter with these modeling approaches that offer us an intuitive way of introducing the basic quantities and concepts in epidemiology. advances in technology are resulting in increased data on human dynamics and behavior. consequently, modeling approaches in epidemiology are gradually becoming more detailed and starting to include realistic contact and mobility patterns. in sects. . and . we describe such developments and analyze the effects of heterogeneities in contact structures between individuals and between cities/subpopulations. with these ingredients in hand we then introduce state-of-the-art data-driven epidemiological models as examples of the modern capabilities in disease modeling and predictions. in particular, we consider gleam [ , ] , episims [ ] , and flute [ ] . the first model is based on the metapopulation framework, a paradigm where the inter-population dynamics is modeled using detailed mobility patterns, while the intra-population dynamics is described by coarse-grained techniques. the other tools are, instead, agent-based model (abm). this class of tools guarantees a very precise description of the unfolding of diseases, but need to be fed with extremely detailed data and are not computationally scalable. for these reasons their use so far has been limited to the study of disease spread within a limited numbers of countries. in comparison, metapopulation models include a reduced amount of data, while the approximated description of internal dynamics allows scaling the simulations to global scenarios. interestingly, the access to large-scale data on human activities has also started a new era in epidemiology. indeed, the big-data revolution naturally results in real time data on the health related behavior of individuals across the globe. such information can be obtained with tools that either require the active participation of individuals willing to share their health status or that is mined silently from individuals' health related data. epidemiology is becoming digital [ , ] . in sect. . we introduce the basic concepts, approaches, and results in this new field of epidemiology. in particular, we describe tools that, using search queries, microblogging, or other web-based data, are able to predict the incidence of a wide range of diseases two weeks ahead respect to traditional surveillance. epidemic models divide the progression of the disease into several states or compartments, with individuals transitioning compartments depending on their health status. the natural history of the disease is represented by the type of compartments and the transitions from one to another, and naturally varies from disease to disease. in some illnesses, susceptible individuals (s) become infected and infectious when coming in contact with one or more infectious (i) persons and remain so until their death. in this case the disease is described by the so-called si (susceptible-infected) model. in other diseases, as is the case for some sexual transmitted diseases, infected individuals recover becoming again susceptible to the disease. these diseases are described by the sis (susceptible-infected-susceptible) model. in the case of influenza like illnesses (ili), on the other hand, infected individuals recover becoming immune to future infections from the same pathogen. ilis are described by the sir (susceptible-infected-recovered) model. these basic compartments provide us with the fundamental description of the progression of an idealized infection in several general circumstances. further compartments can be added to accurately describe more realistic illnesses such as smallpox, chlamydia, meningitis, and ebola [ , , ] . keeping this important observation in mind, here we focus on the sir model. epidemic models are often represented using chart such as the one seen in fig. . . such illustrations are able to accurately represent the number of compartments and the disease's behavior in a concise and easily interpretable form. mathematically, models can also be accurately represented as reaction equations as we will see below. in general, epidemic models include two type of transitions, "interactive" and "spontaneous." interactive transitions require the contact between individuals in two different compartments, while spontaneous transitions occur naturally at a fixed rate per unit time. for example, in the transition between s to i, susceptible individuals become infected due to the interaction with infected individuals, i.e. sci ! i. the transition is mediated by individuals in the compartment i, see fig. but how can we model the infection process? intuitively we expect that the probability of single individual becoming infected must depend on ( ) the number of infected individuals in the population, ( ) the probability of infection given a contact with an infectious agent and, ( ) the number of such contacts. in this section we neglect the details of who is in contact with whom and consider instead individuals to be part of a homogeneously mixed population where everyone is assumed to be in contact with everyone else (we tackle heterogeneous contacts in sect. . ). in this limit, the per capita rate at which susceptible contract the disease, the force of infection , can be expressed in two forms depending on the type of population. in the first, often called mass-action law, the number of contacts per individual is independent of the total population size, and determined by the transmission rateǎ nd the probability of randomly contacting an infected individual, i.e. dˇi=n (where n is the population size). in the second case, often called pseudo massaction law, the number of contacts is assumed to scale with the population size, and the transmission rateˇ, i.e. dˇi. without loss of generality, in the following we focus on the first kind of contact. the sir framework is the crucial pillar to model ilis. think, for example, at the h n pandemic in , or the seasonal flu that every year spread across the globe. the progression of such diseases, from the first encounter to the recovery, happens in matters of days. for this reason, birth and death rates in the populations can be generally neglected, i.e. d t n Á for all times t. let us define the fraction of individuals in the susceptible, infected, and recovered compartments as s; i, and r. the sir model is then described by the following set of differential equations: where dˇi Áˇi n is the force of infection, and d t Á d dt . the first equation describes the infection process in a homogeneous mixed population. susceptible individuals become infected through random encounters with infected individuals. the second equation describes the balance between the in-flow (infection process, first term), and the out-flow (recovery process, second term) in compartment i. finally, the third equation accounts for the increase of the recovered population due to the recovery process. interestingly, the sir dynamical equations, although apparently very simple, due to their intrinsic non-linearity cannot be solved analytically. the description of the evolution of the disease can be obtained only through numerical integration of the system of differential equations. however, crucial analytic insight on the process can be obtained for early t t and late times t ! . under which conditions a disease starting from a small number, i , of individuals at time t is able to spread in the population? to answer this question let us consider the early stages of the spreading, i.e. t t . the equation for the infected compartment can be written as d t i d i.ˇs /, indicating an exponential behavior for early times. it then follows that if the initial fraction of susceptible individuals, s d s =n, is smaller than =ˇ, the exponent becomes negative and the disease dies out. we call this value the epidemic threshold [ ] of the sir model. the fraction of susceptibles in the population has to be larger than a certain value, that depends on the disease details, in order to observe an outbreak. typically, the initial cluster of infected individuals is small in comparison with the population size, i.e. s i , or s . in this case, the threshold condition can be re-written asˇ= > . the quantity: is called the basic reproductive number, and is a crucial quantity in epidemiology and provides a very simple interpretation of the epidemic threshold. indeed, the disease is able to spread if and only if each infected individual is able to infect, on average, more than one person before recovering. the meaning of r is then clear: it is simply the average number of infections generated by an initial infectious seed in a fully susceptible population [ ] . for any value of > , the sir dynamics will eventually reach a stationary, disease-free, state characterized by i d d t i d . indeed, infected individuals will keep recovering until they all reach the r compartment. what is the final number of recovered individuals? answering this apparently simple question is crucial to quantify the impact of the disease. we can tackle such conundrum dividing the first equation with the third equation in the system . . we obtain d r s d r s which in turn implies s t d s e r r t . unfortunately, this transcendent equation cannot be solved analytically. however, we can use it to gain some important insights on the sir dynamics. we note that for any r > , in the limit t ! , we must have s > . in other words, despite r , the disease-free equilibrium of an sir model is always characterized by some finite fraction of the population in the susceptible compartment, or, in other words, some individuals will always be able to avoid the infection. in the limit where r we can obtain an approximate solution for r (or equivalently for s d r ) by expanding s d s e r s at the second order around r . after a few basic algebraic manipulations we obtain in the previous sections we presented the basic concepts and models in epidemiology by considering a simple view of a population where individuals mix homogeneously. although such approximation allows a simple mathematical formulation, it is far from reality. individuals do not all have the same number of contacts, and more importantly, encounters are not completely random [ ] [ ] [ ] [ ] . some persons are more prone to social interactions than others, and contacts with family members, friends, and co-workers are much more likely than interactions with any other person in the population. over the last decade the network framework has been particularly effective in capturing the complex features and the heterogeneous nature of our contacts [ ] [ ] [ ] [ ] [ ] . in this approach, individuals are represented by nodes while links represent their interactions. as described in different chapters of the book (see chaps. , , and ), human contacts are not heterogeneous in both number and intensity [ ] [ ] [ ] [ ] ] but also change over time [ ] . this framework naturally introduces two timescales, the timescale at which the network connections evolve, g and the inherent timescale, p , of the process taking place over the network. although the dynamical nature of interactions might have crucial consequences on the disease spreading [ ] [ ] [ ] [ ] [ ] [ ] , the large majority of results in the literature deal with one of two limiting regimens [ , ] . when g p , the evolution of the network of contacts is much slower than the spreading of the disease and the network can be considered as static. on the other hand, when p g , the links are said to be annealed and changes in networks structure are much faster than the spreading of the pathogen. in both cases the two time-scales are well separated allowing for a simpler mathematical description. here we focus on the annealed approximation ( p g ) that provides a simple stage to model and understand the dynamical properties of epidemic processes. we refer the reader to chap. face-to-face interactions for recent approaches that relax this time-scale separation assumption. let us consider a network g .n; e/ characterized by n nodes connected by e edges. the number of contacts of each node is described by the degree k. the degree distribution p .k/ characterizes the probability of finding a node of degree k. empirical observations in many different domains show heavy-tailed degree distributions usually approximated as power-laws, i.e. p .k/ k ˛ [ , ] . furthermore, human contact networks are characterized by so-called assortative mixing, meaning a positive correlation between the degree of connected individuals. correlations are encoded in the conditional probability p .k jk/ that a node of degree k is connected with a node of degree k [ , ] . while including realistic correlations in epidemic models is crucial [ ] [ ] [ ] they introduce a wide set of mathematical challenges that are behind the scope of this chapter. in the following, we consider the simple case of uncorrelated networks in which the interdependence among degree classes is removed. how can we extend the sir model to include heterogeneous contact structures? here we must take a step further than simply treating all individuals the same. we start distinguishing nodes by degree while considering all vertices with the same degree as statistically equivalent. this is known as the degree block approximation and is exact for annealed networks. the quantities under study are now i k d i k n k ; s k d s k n k , and r k d r k n k , where the i k ; s k , and r k are the number of infected, susceptible, recovered individuals in the degree class k. n k instead describes the total number of nodes in the degree class k. the global averages are given by i d using this notation and heterogeneous mean field (hmf) theory [ ] , the system of differential equations ( . ) can now be written as: the contact structure introduces a force of infection function of the degree. in particular, k d k k where is the rate of infection per contact, i.e.ˇd k, and k describes the density of infected neighbors of nodes in the degree class k. intuitively, this density is a function of the conditional probability that a node k is connected to any node k and proportional to the number of infected nodes in each class in the simple case of uncorrelated networks the probability of finding a node of degree k in the neighborhood of a node in degree class k is independent of k. in this case k d d p k .k / p .k / i k =hki where the term k is due to the fact that at least one link of each infected node points to another infected vertex [ ] . in order to derive the epidemic threshold let us consider the early time limit of the epidemic process. as done in sect. . . . let us consider that at t t the population is formed mostly by susceptible individuals. in the present scenario this implies s k i k and r k k. the equation for the infected compartment then becomes d t i k d k i k . multiplying both sides for p .k/ and summing over all values of k we obtain d t i d hki i. in order to understand the behavior of i around t let us consider an equation built by multiplying both sides of the last equation by .k / p .k/ =hki and summing over all degree classes. we obtain d t d . hk i hki hki / . the fraction of infected individuals in each value of k will increase if and only if d t > . this condition is verified when [ ] : giving us the epidemic threshold of an sir process unfolding on an uncorrelated network. remarkably, due to their broad-tailed nature, real contact networks display fluctuations in the number of contacts (large hk i) that are significantly larger than the average degree hki resulting in very small thresholds. large degree nodes (hubs) facilitate an extremely efficient spreading of the infection by directly connecting many otherwise distant nodes. as soon as the hubs become infected diseases are able to reach a large fraction of the nodes in the network. real interaction networks are extremely fragile to disease spreading. while this finding is somehow worrisome, it suggests very efficient strategies to control and mitigate the outbreaks. indeed, hubs are central nodes and play a crucial role in the network connectivity [ ] and by vaccinating a small fraction of them one is able to quickly stop the spread of the disease and protect the rest of the population. it is important to mention that in realistic settings the knowledge of the networks' structure is often limited. hubs might not be easy to easily known and other indirect means must be employed. interestingly, the same feature of hubs that facilitates the spread of the disease also allows for their easy detection. since high degree nodes are connected to a large number of smaller degree nodes, one may simply randomly select a node, a, from the network and follow one of its links to reach another node, b. with high probability, node b has higher degree than a and is likely a hub. this effect became popularized as the friend paradox: on average your friends have more friends than you do [ ] . immunizing node b is then much more effective than immunizing node a. remarkably, as counter-intuitive as this methodology might seem, it works extremely well even in the case of quickly changing networks [ ] [ ] [ ] . the next step in the progression towards more realistic modeling approaches is to consider the internal structure of the nodes. if each node in the network represents a homogeneously mixed sub-population instead of a single individual and we consider the edges to represent interactions or mobility between the different subpopulations, then we are in the presence of what is known as meta-population. this concept was originally introduced by r. levins in [ ] for the study of geographically extended ecological populations where each node represents one of the ecological niches where a given population resides. the metapopulation framework was later extended for use in epidemic modeling by sattenspiel in . in a landmark paper [ ] sattenspiel considered two different types of interactions between individuals, local ones occurring within a given node, and social ones connecting individuals originating from different locations on the network. this idea was later expanded by sattenspiel and dietz to include the effects of mobility [ ] and thus laying the foundations for the development of epidemic models at the global scale. metapopulation epidemic models are extremely useful to describe particle reaction-diffusion models [ ] . in this type of model each node is allowed to have zero or more individuals that are free to diffuse among the nodes constituting the network. in our analysis, as done in the previous section, we follow the hmf approach and consider all nodes of degree k to be statistically equivalent and write all quantities in terms of the degree k. to start, let us define the average number of individuals in a node of degree k to be w k d where n k is the number of nodes with degree k and the sum is taken over all nodes i. the mean field dynamical equation describing the variation of the average number of individuals in a node of degree k is then: where p k and p kk represent, respectively, the rate at which particles diffuse out of a node of degree k and diffuse from a node of degree k to one of degree k . with these definitions, the meaning of each term of this equation becomes intuitively clear: the negative term represents individuals leaving the node, while the positive term accounts for individuals originating from other nodes arriving at this particular class of node. the conditional probability p .k jk/ encodes all the topological correlations of the network. by imposing that the total number of particles in the system remains constant, we obtain: that simply states that the number of particles arriving at nodes of degree k coming from nodes of degree k must be the same as the number of particles leaving nodes of degree k. the probabilities p k and p kk encode the details of the diffusion process [ ] . in the simplest case, the rate of movement of individuals is independent of the degree of their origin p k d p for all values of the degree. furthermore, if individuals that are moving simply select homogeneously among all of their connections, then we have p kk d p=k. in this case, the diffusion process will reach a stationary state when: where w d w=n, w is the total number of walkers in the system, and n the total number of nodes. the simple linear relation between w k and k serves as a strong reminder of the importance of network topology. nodes with higher degree will acquire larger populations of particles while nodes with smaller degrees will have proportionally smaller populations. however, even in the steady state, the diffusion process is ongoing, so individuals are continuously arriving and leaving any given node but are doing so in a way that maintains the total number of particles in each node constant. in more realistic settings, the traffic of individuals between two nodes is function of their degree [ ] : in this expression  modulates the strength of the diffusion flow between degree classes (empirical values are in the range : Ä Â Ä : [ ] ), where w is a constant and t k d w hk c i=hki is the proper normalization ensured by the condition in eq. ( . ). in these settings, the diffusion process reaches a stationary state when: note that for  d this solution coincides with the case of homogeneous diffusion [eq. ( . )]. combining this diffusion process with the (epidemic) reaction processes described above we finally obtain the full reaction-diffusion process. to do so we must simply write eq. ( . ) for each state of the disease (e.g., susceptible, infectious, and recovered for a simple sir model) and couple the resulting equations using the already familiar epidemic equations. the full significance of eq. ( . ) now becomes clear: nodes with higher degree have higher populations and are visited by more travelers, making them significantly more likely to also receive an infected individual that can act as the seed of a local epidemic. in a metapopulation epidemic context we must then consider two separate thresholds, the basic reproductive ratio, r , that determines whether or not a disease can spread within one population (node) and a critical diffusion rate, p c , that determines if individual mobility is sufficiently large to allow the disease to spread from one population to another. it is clear that if p d particles are completely unable to move from one population to another so the epidemic cannot spread across subpopulations and that if p d all individuals are in constant motion and the disease will inevitably spread to every subpopulation on the network with a transition occurring at some critical value p c . in general, the critical value p c cannot be calculated analytically using our approach as it depends non-trivially on the detailed structure of the network and the fluctuations of the diffusion rate of single individuals. however, in the case of uncorrelated networks a closed solution can be easily found for different mobility patterns. indeed, in the case where the mobility is regulated by eq. ( . ) we obtain: interestingly, the critical value of p is inversely proportional to the degree heterogeneity in the network, so that broad tailed networks have very low critical values. this simple fact explains why simply restricting travel between populations is a highly ineffective way to prevent the global spread of an epidemic. the mobility patterns considered so far are so-called markovian: individuals move without remembering where they have been nor they have a home where they return to after each trip. although this is a rough approximation of individuals behavior, markovian diffusion patterns are allowed to analytically describe the fundamental dynamical properties of many systems. recently, new analytic results have been proposed for non-markovian dynamics that include origin-destination matrices and realistic travel routes that follow shortest paths [ ] . in particular, the threshold within such mobility schemes reads as: the exponent Á, typically close to : in heterogeneous networks, emerges from the shortest paths routing patterns [ ] . interestingly, for values of Â Ä : , fixing Á d : , p c in the case of markovian mobility patterns is larger than the critical value in a system subject to non-markovian diffusion. the presence of origindestination matrices and shortest paths mobility lower the threshold facilitating the global spreading of the disease. instead, for values of  > : the contrary is true. in these models the internal contacts rate is considered constant across each subpopulation. interestingly, recent longitudinal studies on phone networks [ ] and twitter mention networks [ ] point to the evidence that contacts instead scale super-linearly with the subpopulation sizes. considering the heterogeneity in population sizes observed in real metapopulation networks, the scaling behavior entails deep consequence in the spreading dynamics. a recent study generalized the metapopulation framework considering such observations. interestingly, the critical mobility thresholds, in the case of mobility patterns described by eq. ( . ), changes significantly being lowered by such scaling features of human contacts [ ] . despite their simplicity, metapopulation models are extremely powerful tools in large scale study of epidemics. they easily lend themselves to large scale numerical stochastic simulations where the population and state of each node can be tracked and analyzed in great detail and multiple scenarios as well as interventions can be tested. the state of the art in the class of metapopulation approaches is currently defined by the global epidemic and mobility model (gleam) [ , ] . gleam integrates worldwide population estimates [ , ] with complete airline transportation and commuting databases to create a world wide description of mobility around the world that can then be used as the substrate on which the epidemic can spread. gleam divides the globe into transportation basins. each basin is defined empirically around an airport and the area of the basin is determined to be the region within which residents would likely use that airport for long distance travel. each basin represents a major metropolitan area such as new york, london, or paris. information about all civilian flights can be obtained from the international air transportation association (iata) [ ] and the official airline guide (oag) [ ] that are responsible for compiling up-to-date databases of flight information that airlines use to plan their operations. by connecting the population basins with the direct flight information from these databases we obtain the network that acts as a substrate for the reaction diffusion process. while most human mobility does not take place in the form of flights, the flight network provides the fundamental structure for long range travel that explains how diseases such as sars [ ] , smallpox [ ] , or ebola [ ] spread from country to country. to capture the finer details of within country mobility further information must be considered. gleam uses census information to create a commuting network at the basin level that connects neighboring metropolitan areas proportionally to the number of people who live in one are but work in the other. short-term short-distance mobility such as commuting is fundamentally different from medium-term long-distance airline travel. in one case, the typical timescale is work-day ( h) while in the other it is day. this timescale difference is taken into account in gleam in an effective, mean-field, manner instead of explicitly through a reaction process such as the one described above. this added layer is the final piece of the puzzle that brings the whole together and allows gleam to describe accurately the spread from one country to the next but also the spread happening within a given country [ ] . in fig. . we illustrate the progression in terms of detail that we have undergone since our initial description of simple homogeneously mixed epidemic models in a single population. with all these ingredients in place we have a fine grained description of mobility on a world wide scale on top of which we can finally build an epidemic model. within each basin, gleam still uses the homogeneous mixing approximation. this assumption is particularly suited for diseases that spread easily from person to person through airborne means such as ili. gleam describes influenza through an seir model as illustrated in fig. . . seir models are a modification of the sir model described above that includes a further compartment, exposed, to represent of the remaining symptomatic individuals, one half is sick enough to decide to not travel or commute while the remaining half continue to travel normally. despite their apparent complexity, large scale models such as gleam are controlled by just a small number of parameters and ultimately, it's the proper setting of these few parameters that is responsible for the proper calibration of the model and validity of the results obtained. most of the disease and mobility parameters are set directly from the literature or careful testing so that as little as possible remains unknown when it is time to apply it to a new outbreak. gleam was put to the test during the h n pandemic with great success. during the course of the epidemic, researchers were able to use official data as it was released by health authorities around the world. in the early days of the outbreak there was a great uncertainty about the correct value of the r for the /h n pdm strain in circulation so a methodology to determine it had to be conceived. one of the main advantages of epidemic metapopulation models is their computational tractability. it was this feature what proved invaluable when it came to determine the proper value of r . by plugging in a given set of parameters one is able to generate several hundreds or thousands of in silico outbreaks. each outbreak contains information not only about the number of cases in each city or country as a function of time but also information about the time when the first case occurs within a given country. in general, each outbreak will be different due to stochasticity and by combining all outbreaks generated for a certain parameter set we can calculate the probability distribution of the arrival times. the number of times that an outbreak generated the seeding of a country, say the uk, in the same day as it occurred in reality provides us with a measure of how likely the parameter values used are. by multiplying this probability for all countries with a known arrival time we can determine the overall likelihood of the simulation: where the product is taken over all countries c with known arrival time t c and the probability distribution of arrival times, p c .t/ is determined numerically for each set of input values. the set of parameters that maximizes this quantity is then the one whose values are the most likely to be correct. using this procedure the team behind gleam determined that the mostly likely value of the basic reproductive ratio was r d : [ ] , a value that was later confirmed by independent studies [ , ] . armed with an empirical estimate of the basic reproductive ratio for an ongoing pandemic, they then proceeded to use this value to estimate the future progression of the pandemic. their results predicting that the full peak of the pandemic would hit in october and november were published in early september [ ] . a comparison between these predictions and the official data published by the health authorities in each country would be published several years later [ ] clearly confirming the validity of gleam for epidemic forecasting in real time. indeed, the model predicted, months in advance, the correct peak week in % of countries in the north hemisphere for which real data was accessible. in the rest of cases the maximum error reported has been weeks. gleam can also be further extended to include age-structure [ ] , interventions and travel reductions. the next logical step in the hierarchy of large scale epidemic models is to take the description of the underlying population all the way down to the individual level with what are known as abm. the fundamental idea behind this class of model is a deceptively simple one: treat each individual in the population separately, assigning it properties such as age, gender, workplace, residence, family structure, etc: : : these added details give them a clear edge in terms of detail over metapopulation models but do so at the cost of much higher computational cost. the first step in building a model of this type is to generate a synthetic population that is statistically equivalent to the population we are interested in studying. typically this is in a hierarchical way, first generating individual households, aggregating households into neighborhoods, neighborhoods into communities, and communities into the census tracts that constitute the country. generating synthetic households in a way that reproduces the census data is far from a trivial task. the exact details vary depending on the end goal of the model and the level of details desired but the household size, age, and gender of household members are determined stochastically from the empirically observed distributions and conditional probabilities. one might start by determining the size of the household by extracting from the distribution of household size of the country of interest and selecting the age and gender of the head of the household proportionally to the number of heads of households for that household size that are in each age group. conditional on this synthetic individual we can then generate the remaining members, if any, of the household. the required conditional probability distributions and correlation tables can be easily generated [ ] from high quality census data that can be found for most countries in the world. this process is repeated until enough synthetic households have been generated. households are then aggregated into neighborhoods by selecting from the households according to the distribution of households in a specific neighborhood. neighborhoods are similarly aggregated into communities and communities into census tracts. each increasing level of aggregation (from household to country) represents a decrease in the level of social contact, with the most intimate contacts occurring at the household level and least intimate ones at the census tract or country level. the next step is to assign to each individual a profession and work place. workplaces are generated following a procedure similar to the generation of households and each employed individual is assigned a specific household. school age children are assigned a school. working individuals are assigned to work places in a different community or census tract in a way that reflects empirical commuting patterns. at this point, we have a fairly accurate description of where the entire population of a city or country lives and works. it is then not entirely surprising that this approach was first used to study in detail the demands imposed on the transportation system of a large metropolitan city. transims, the transportation analysis and simulation system [ ] , used an approach similar to the one described above to generate a synthetic population for the city of portland, in oregon (or) and coupled it with a route planner that would determine the actual route taken by each individual on her way to work or school as a way of modeling the daily toll on portland's transportation infrastructure and the effect that disruptions or modification might have in the daily lives of its population. episims [ ] was the logical extension of transims to the epidemic world. episims used the transims infrastructure to generate the contact network between individuals in portland, or. susceptible individuals are able to acquire the infection whenever they are in a location along with one or more infectious individuals. in this way the researchers are capable of observing as the disease spreads through the population and evaluate the effect that measures such as contact tracing and mass vaccination. more recent approaches have significantly simplified the mobility aspect of this kind of models and simply divide each h period into day time and nighttime. individuals are considered to be in contact with other members of their workplace during the day and with other household members during the night. in recent years, modelers have successfully expanded the large scale agent based approach to the country [ ] and even continent level [ ] . as the spatial scale of the models increased further modes of long-range transportation such as flights had to be considered. these are important to determine not only the seeding of the country under consideration through importation of cases from another country but also to connect distant regions in a more realistic way. flute [ ] is currently the most realistic large scale agent-based epidemic model of the continental united states. it considers that international seeding occurs at random in the locations that host the largest international airports in the us by, each day, randomly infecting in each location a number of individuals that is proportional to the international traffic of those airports. flute is a refinement of a previous model [ ] and it further refines the modeling of the infectious process by varying the infectiousness of an individual over time in the sir model that they consider. at the time of infection each individual is assigned one of six experimentally obtained viral load histories. each history prescribes the individuals viral load for each day of the infectious period and the infectiousness is considered to be proportional to the viral load. individuals may remain asymptotic for up to days after infection during which their infectiousness is reduced by % with respect to the symptomatic period. the total infectious period is set to days regardless of the length of the symptomatic period. given the complexity of the model the calibration of the disease parameters in order to obtain a given value of the basic reproductive ratio, r requires some finesse. chao et al. [ ] uses the definition of r to determine "experimentally" its value from the input parameters. it numerically simulates instances of the epidemic caused by a single individual within a person fully susceptible community for each possible age group of the seeding individual and use it to calculate the r a of each age group a. the final r is defined to the average of the various r a weighted by age dependent attack rate [ ] . the final result of this procedure is that the value of r is given by: where is the infection probability per unit contact and is given as input. flute was a pioneer in the way it completely released its source code, opening the doors of a new level of verifiability in this area. it has successfully used to study the spread of influenza viruses and analyze the effect of various interventions in the los angeles county [ ] and united states country level [ ] . the unprecedented amount of data on human dynamics made available by recent advances technology has allowed the development of realistic epidemic models able to capture and predict the unfolding of infectious disease at different geographical scales [ ] . in the previous sections, we described briefly some successful examples that have been made possible thanks to high resolution data on where we live, how we live, and how we move. data availability has started a second golden age in epidemic modeling [ ] . all models are judged against surveillance data collected by health departments. unfortunately, due to excessive costs, and other constraints their quality is far from ideal. for example, the influenza surveillance network in the usa, one of the most efficient systems in the world, is constituted of just providers that operate voluntarily. surveillance data is imprecise, incomplete, characterized by large backlogs, delays in reporting times, and the result of very small sample sizes. furthermore, the geographical coverage is not homogeneous across different regions, even within the same country. for these reasons the calibration and test of epidemic models with surveillance data induce strong limitations in the predictive capabilities of such tools. one of the most limiting issues is the geographical granularity of the data. in general, information are aggregated at the country or regional level. the lack of ground truth data at smaller scales does not allow a more precise selection and training of realistic epidemic models. how can we lift such limitations? data, data and more data is again the answer. at the end of almost billion of people had access to the internet while almost billion are phone subscribers, around % of which are actively using smartphones. the explosion of mobile usage boosted also the activity of social media platforms such as facebook, twitter, google+ etc. that now count several hundred million active users that are happy to share not just their thoughts, but also their gps coordinates. the incredible amount of information we create and access contain important epidemiologically relevant indicators. users complaining about catching a cold before the weekend on facebook or twitter, searching for symptoms of particular diseases on search engines, or wikipedia, canceling their dinner reservations on online platforms like opentable are just few examples. an intense research activity, across different disciplines, is clearly showing the potential, as well as the challenges and risks, of such digital traces for epidemiology [ ] . we are at the dawn of the digital revolution in epidemiology [ , ] . the new approach allows for the early detection of disease outbreaks [ ] , the real time monitoring of the evolution of a disease with an incredible geographical granularity [ ] [ ] [ ] , the access to health related behaviors, practices and sentiments at large scales [ , ] , inform data-driven epidemic models [ , ] , and development of statistical based models with prediction power [ , [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] . the search for epidemiological indicators in digital traces follows two methodologies: active and passive. in active data collection users are asked to share their health status using apps and web-based platforms [ ] . examples are influenzanet that is available in different european countries [ ] , and flu near you in the usa [ ] that engage tens of thousands of users that together provide the information necessary for the creation of interactive maps of ili in almost real time. in passive data collection, instead, information about individuals health status is mined from other available sources that do not require the active participation of users. news articles [ ] , queries on search engines [ ] , posts on online social networks [ , [ ] [ ] [ ] [ ] , page view counts on wikipedia [ , ] or other online/offline behaviors [ , ] are typical examples. in the following, we focus on the prototypical, and most famous, method of digital epidemiology, google flu trends (gft) [ ] , while considering also other approaches based on twitter and wikipedia data. gft is by far the most famous model in digital epidemiology. launched in november together with a nature paper [ ] describing its methodology, it has continuously made predictions on the course of seasonal influenza in countries around the world. the method used by gft is extremely simple. the percentage of ili visits, a typical indicator used by surveillance systems to monitor the unfolding of the seasonal flu, is estimated with a linear model based on search engine queries. this approach is general, and used in many different fields of science. a quantity of interest, in this case the percentage of ili visits p, is estimated using a correlated signal, in this case the ili related queries fraction q, that acts as surrogate. the fit allows the estimate of p as a function of the value of q: logit .p/ dˇ cˇ logit .q/ c ; ( . ) where logit .x/ d ln x x ,ˇ andˇ are fitting parameters, and is an error term. as clear from the expression, the gft is a simple linear fit, where the unknown parameters are determined considering historical data. the innovation of the system lies on the definition of q that is evaluated using hundreds of billions of searches on google. indeed, gft scans all the queries we submit to google, without using information about users' identity, in search of those that ili related. this is the paradigm of passive data collection in digital epidemiology. in the original model the authors measured the correlation of millions search queries with historic cdc data, finding that of them were enough to ensure the best correlation between the number of searches and the number of ili cases. the identity of such terms has been kept secret in order to avoid changes in users' behavior. however, the authors provided a list of topics associated with each one of them: were associated with influenza complications, to cold/flu remedies, to general terms for influenza, etc. although the search for the terms has been performed without prior information, none of the most representative terms were unrelated to the disease. in these settings gft showed a mean correlation of : with real data and was able to predict the surveillance value with - weeks ahead. gft is based on proprietary data that for many different constraints cannot be shared with the research community. other data sources, different in nature, are instead easily accessible. twitter and wikipedia are the two examples. indeed, both systems are available for download, with some limitations, through their respective apis. the models based on twitter are built within the same paradigm of gft [ , [ ] [ ] [ ] ] . tweets are mined in search of ili-related tweets, or other health conditions such as insomnia, obesity, and other chronic diseases [ , ] , that are used to inform regression models. such tweets are determined either as done in gft, or through more involved methods based on support vector machine (svm) or other machine learning methods that, provided an annotated corpus, find disease related tweets beyond simple keywords matches [ , [ ] [ ] [ ] ] . the presence of gps information or other self-reported geographical data allows the models to probe different granularities ranging from countries [ , , , ] to cities [ ] . while models based on twitter analyze users' posts, those based on wikipedia focus on pages views [ , ] . the basic intuition is that wikipedia is used to learn more about a diseases or a medication. plus, the website is so popular that is most likely one of the first results of search queries on most search engines. the methods proposed so far monitor a set of pages related to the disease under study. examples are influenza, cold, fever, dengue, etc. page views at the daily or weekly basis are then used a surrogates in linear fitting models. interestingly, the correlation with surveillance data ranges from : in the case of ebola to : in for ilis [ , ] , and allows accurate predictions up to weeks ahead. one important limitation of wikipedia based methods is the lack of geographical granularity. indeed, the view counts are reported irrespective of readers' location but the language of the page can be used as a rough proxy for location. such approximation might be extremely good for localized languages like italian but it poses strong limitations in the case of global languages like english. indeed, it is reported that % of pages views for english pages are done in the usa, % in the uk, and the rest in australia, canada and other countries [ ] . besides, without making further approximation such methods cannot provide indications at scales smaller than the country level. despite these impressive correlations, especially in the case of ilis, much still remains to be done. gft offers a particular clear example of the possible limitations of such tools. indeed, despite the initial success, it completely failed to forecast the h n pandemic [ , ] . the model was updated in september to increase the number of terms to , including the terms present in the original version. nevertheless, gft missed high out of weeks in the season - . in gft predicted a peak height more than double the actual value causing the underlying model to be modified again later that year. what are the reasons underlying the limitations of gft and other similar tools? by construction, gft relies just on simple correlations causing it to detect not only the flu but also things that correlate strongly with the flu such as winter patterns. this is likely one of the reasons why the model was not able to capture the unfolding of an off-season pandemic such as the h n pandemic. also, changes in the google search engine, that can inadvertently modify users' behavior, were not taken into account in gft. this factor alone possibly explains the large overestimation of the peak height in . plus, simple auto-regressive models using just cdc data can perform as well or better than gft [ ] . the parable of gft clearly shows both the potential and the risks of digital tools for epidemic predictions. the limitations of gft can possibly affect all similar approaches based on digital passive data collection. in particular, the use of simple correlations measures does not guarantee the ability of capturing the phenomena across different scales in space and time with respect to those used in the training. not to mention that correlations might be completely spurious. in a recent study for example, a linear model based on twitter simply informed with the timeline of the term zombie was shown to be a good predictor of the seasonal flu [ ] . despite such observations the potential of these models is invaluable to probe data that cannot be predicted by simple auto-regressive models. for example, flu activity at high geographical granularities, although very important, is measured with great difficulties by the surveillance systems. gft and other spatially resolved tools can effectively access to these local indicators, and provide precious estimates that can be used a complement for the surveillance and as input for generating epidemic models [ , ] . the field of epidemiology is currently undergoing a digital revolution due to the seemingly endless availability of data and computational power. data on human behavior is allowing for the development of new tools and models while the commoditization of computer resources once available only for world leading research institutions is making highly detailed large scale numerical approaches feasible at last. in this chapter, we present a brief review not only of the fundamental mathematical tools and concepts of epidemiology but also of some of the state-of-the-art and computational approaches aimed at describing, modeling, and forecasting the diffusion of viruses. our focus was on the developments occurring over the past decade that are sure to form the foundation for developments in decades to come. essai dune nouvelle analyse de la mortalité causée par la petite vérole et des advantages de l'inocoulation pur la prévenir. mémoires de mathematique physique de l infectious diseases in humans the role of the airline transportation network in the prediction and predictability of global epidemics modeling the spatial spread of infectious diseases: the global epidemic and mobility computational model modelling disease outbreaks in realistic urban social networks flute, a publicly available stochastic influenza epidemic simulation model digital disease detection: harnessing the web for public health surveillance digital epidemiology the mathematical theory of infectious diseases modeling infectious diseases in humans and animals a contribution to the mathematical theory of epidemics networks: an introduction scale-free networks complex networks: structure, robustness and function dynamical processes on complex networks computational social science linked: how everything is connected to everything else and what it means temporal networks telling tails explain the discrepancy in sexual partner reports simulated epidemics in an empirical spatiotemporal network of , sexual contacts what's in a crowd? analysis of face-to-face behavioral networks activity driven modeling of dynamic networks time varying networks and the weakness of strong ties epidemic spreading in non-markovian timevarying networks thresholds for epidemic spreading in networks modeling dynamical processes in complex socio-technical systems epidemic spreading in correlated complex networks spread of epidemic disease on networks correlations in weighted networks efficient immunization strategies for computer networks and populations using friends as sensors to detect global-scale contagious outbreaks controlling contagion processes in activity driven networks some demographic and genetic consequences of environmental heterogeneity for biological control population structure and the spread of disease a structured epidemic model incorporating geographic mobility among regions reaction-diffusion equations and their applications to biology epidemic modeling in metapopulation systems with heterogeneous coupling pattern: theory and simulations modeling human mobility responses to the large-scale spreading of infectious diseases the scaling of human interactions with city size the scaling of human contacts in reaction-diffusion processes on heterogeneous metapopulation networks the gridded population of the world version (gpwv ): population grids. palisades global rural-urban mapping project (grump), alpha version: population grids. palisades predictability and epidemic pathways in global outbreaks of infectious diseases: the sars case study human mobility and the worldwide impact of intentional localized highly pathogenic virus release assessing the international spreading risk associated with the west african ebola outbreak multiscale mobility networks and the spatial spreading of infectious diseases seasonal transmission potential and activity peaks of the new influenza a (h n ): a monte carlo likelihood analysis based on human mobility the transmissibility and control of pandemic influenza a (h n ) virus pandemic potential of a strain of influenza a (h n ): early findings real-time numerical forecast of global epidemic spreading: case study of a/h n pdm comparing large-scale computational approaches to epidemic modeling: agent-based versus structured metapopulation models creating synthetic baseline populations transims (transportation analysis simulation system) determinants of the spatiotemporal dynamics of the h n pandemic in europe: implications for real-time modelling mitigation strategies for pandemic influenza in the united states planning for the control of pandemic influenza a (h n ) in los angeles county and the united states predicting the behavior of techno-social systems epidemic processes in complex networks. arxiv preprint the parable of google flu: traps in big data analysis global capacity for emerging infectious disease detection web-based participatory surveillance of infectious diseases: the influenzanet participatory surveillance experience assessing vaccination sentiments with online social media: implications for infectious disease dynamics and control you are what you tweet: analyzing twitter for public health forecasting seasonal outbreaks of influenza forecasting seasonal influenza with stochastic microsimulations models assimilating digital surveillance data the use of twitter to track levels of disease activity and public concern in the us during the influenza a h n pandemic validating models for disease detection using twitter national and local influenza surveillance through twitter: an analysis of the - influenza epidemic towards detecting influenza epidemics by analyzing twitter messages detecting influenza epidemics using search engine query data detecting epidemics using wikipedia article views: a demonstration of feasibility with language as location proxy wikipedia usage estimates prevalence of influenzalike illness in the united states in near real-time guess who is not coming to dinner? evaluating online restaurant reservations for disease surveillance satellite imagery analysis: what can hospital parking lots tell us about a disease outbreak? public health for the people: participatory infectious disease surveillance in the digital age google flu trends using twitter to estimate h n influenza activity a content analysis of chronic diseases social groups on facebook and twitter. telemedicine and e-health reassessing google flu trends data for detection of seasonal and pandemic influenza: a comparative epidemiological study at three geographic scales predicting consumer behavior with web search acknowledgements bg was partially supported by the french anr project harms-flu (anr- -monu- ). key: cord- -voi gu l authors: xuan, huiyu; xu, lida; li, lu title: a ca-based epidemic model for hiv/aids transmission with heterogeneity date: - - journal: ann oper res doi: . /s - - - sha: doc_id: cord_uid: voi gu l the complex dynamics of hiv transmission and subsequent progression to aids make the mathematical analysis untraceable and problematic. in this paper, we develop an extended ca simulation model to study the dynamical behaviors of hiv/aids transmission. the model incorporates heterogeneity into agents’ behaviors. agents have various attributes such as infectivity and susceptibility, varying degrees of influence on their neighbors and different mobilities. additional, we divide the post-infection process of aids disease into several sub-stages in order to facilitate the study of the dynamics in different development stages of epidemics. these features make the dynamics more complicated. we find that the epidemic in our model can generally end up in one of the two states: extinction and persistence, which is consistent with other researchers’ work. higher population density, higher mobility, higher number of infection source, and greater neighborhood are more likely to result in high levels of infections and in persistence. finally, we show in four-class agent scenario, variation in susceptibility (or infectivity) and various fractions of four classes also complicates the dynamics, and some of the results are contradictory and needed for further research. focus on hiv/aids transmission among human groups in order to better understand its dynamical behavior. in epidemic modeling (see, e.g., bailey ; anderson and may ; murray ) , there are two frequently used methodologies: mathematical and simulation methods. for mathematical approaches, a cohort of people is often classified into susceptibles, infectives, and recovereds with (without) immunity (see, e.g., kermack and mckendrick ) . systems of differential equations are used to describe the linear (nonlinear) dynamics of epidemics. macroscopically, mathematical models can reveal the relationship among primary factors and describe their effects to epidemic spreading under certain assumptions. as to hiv/aids epidemic, many models have been proposed (may and anderson ; hyman et al. ; brauer and driessche ; wu and tan , etc.) . however, mathematical approaches have some serious drawbacks due to its intractability and the complexity of epidemics. moreover, the complicated nature of hiv/aids transmission makes it even harder to obtain analytical solutions and difficult to study them. in the early s, some researchers started to apply simulation approaches to this field. there is a large literature that addresses the computer simulation of epidemic dynamics (see, e.g., leslie and brunham ; atkinson ; rhodes and anderson ; rhodes and anderson ; ahmed and agiza ; benyoussef et al. ; tarwater and martin ) . particularly, cellular automata (ca) method (some literature refers to this as a lattice-based method) has been widely used in modeling complex adaptive systems. despite of its simple structure, ca is well suited to describing the propagation phenomena, such as rumor spreading, particle percolation, innovation propagation, and disease spreading. for instance, in epidemic modeling, fuentes and kuperman ( ) propose two ca models corresponding to the classical mathematical sis model and sis model respectively. ahmed and agiza ( ) develop a ca model that takes into consideration the latency and incubation period of epidemics and allow each individual (agent) to have distinctive susceptibility. gao et al. ( ) put forward a ca model for sars spreading which takes account of social influence. more recently, other methods such as agent-based modeling and system dynamics are introduced to this field (see, e.g., gordan ; bagni et al. ) . our paper contributes to this filed by developing an extended ca simulation model. we then use the new ca model to investigate some issues in hiv/aids epidemics. most models, including the foregoing ca models, have some limitations that fail to consider the peculiarities of hiv/aids epidemics and are thereby incapable of describing the epidemic accurately and completely (see frauenthal for more discussion). first, most of the models assume that there is no latent (or incubation) period. however, for some epidemics, especially aids, there are variously lasting periods of latency and incubation as well as behavior-varying infectivity (or susceptibility) during these periods. in fact, the development of aids involves a few stages in which an infected individual can exhibit different behaviors. those diversified behaviors, in turn, have some ignorable effects on the dynamics of hiv/aids. in light of this, we extend the conventional division of epidemic process (i.e., susceptible, infection, and removed) by dividing the infection period into three sub-stages, each corresponding to the clinical stage occurring in the course of aids development. due to the inability of classical ca approaches to accommodate those newly added state transitions events, we also borrow some ideas from discrete-event simulation techniques and make one agent's stage transitions being time-triggered instead of using some state-based transition rules. secondly, it is commonly assumed that individuals in the population are homogenous in the sense that they have equal infectivity and susceptibility, or they can exert the same influence on each other, etc. this assumption may be satisfied in commonly observed epidemics but not consistent with the hiv/aids epidemic. as we know, susceptibility and infectivity heavily depends on individuals' behavior. for examples, safe sex practices such as the use of condom could dramatically reduce the chance of infection. also, the way of hiv/aids transmission for one to another is various, depending on the interactions between people, and thus the probability of getting infected is determined in part by transmission routes and can be quite different between infected-male/susceptible-female and susceptible-male/infectedfemale interactions. under this assumption, the models that are confined to a single high-risk human group are not suitable in overall population cases. new models are needed to explicitly consider the complexity. therefore, we make an extension to the traditional ca model by introducing the extended definition of neighborhood and attaching some attributes to each agent such as infectivity and resistibility. we also define four types of agents that are characterized by different infectivity (and susceptibility) and various forms of neighborhood to represent four types of people in real life. in doing so, we will be able to investigate the dynamics of hiv/aids with heterogeneous groups in a realistic way. thirdly, classical ca models assume that agents in the grid are spatially fixed, that is, once an agent is placed in a cell, it does not move into another cell. this assumption is problematical because people in the real world are migratory. for instance, in china, millions of rural people leave their hometowns and seek jobs in the cities. the migration of population is a driving force for the spread of hiv/aids. ignoring the mobility of agents in epidemic models would jeopardize the creditability of the results obtained. considering this point, we incorporate agents' mobility into their behaviors. in our model, each agent is allowed to move randomly into one of its adjoined and unoccupied cells at random time intervals. recently, agent-based modeling is used in various fields to solve plenty of problems (see, e.g. zhang and bhattacharyya ; luo et al. ) . some reader might notice that our improved ca model have features that usually found in agent-based methodology. as a matter of fact, our method borrows much from agent-based simulation modeling. to make things simple, we prefer to view this model as being a ca models. this paper is organized as follows. in the next section, we present our extended ca simulation model. section gives a detailed description of simulation results and analyzes some influential factors that affect the dynamical behavior of the model. section concludes and points out some possible extensions and directions for future research. cellular automata have been extensively used as tools for modeling complex adaptive systems such as traffic flow, financial markets, chemical systems, biological groups, and other social systems (see e.g. gerhard and schuster ; gerhardt et al. ; weimar et al. ; karafyllidis and thanailakis ; karafyllidis ) . usually, a typical ca model consists of a regular two-dimension grid with a certain boundary condition and a swarm of agents living in the grid. the neighborhood of an agent is defined to some (or all) of the immediately adjacent cells and the agents who inhabit in the neighborhood are called neighbors. agents are restricted to local neighborhood interaction and hence are unable to communicate globally. there are several states agents can be in at each time and an agent's state at time t + is determined based on its neighbors' states at time t . the rules used in the determination of next-time states can be written as a mapping: where s is the set of states and t denotes simulation time. the mathematical properties of cellular automata have been studied in martin et al. ( ) . in our model, we consider a population of size n(t) at time t randomly distributed in a two-dimension w × w lattice. population growth rate r is fixed throughout the simulation. at each time, new agents are added to the model, and the dead removed. simulation time advances in a discrete way. the time interval (t, t + ) is specified to represent one week in real life. this assumption makes the simulations run reasonably fast (with respect to the whole progress of epidemics) without losing any time-specific clinical properties associated with hiv/aids. explicitly modeling the post-infection progression to aids is one feature of our model compared with conventional ca models. classical epidemic models divide the closed population into three subgroups: susceptible, infective, and recovered (removed). this simplified classification is not consistent with the epidemics in real life. particularly, it is well established that an individual, once infected with hiv, undergoes roughly three clinical phrases towards the full-blown aids: ( ) infected, not yet infectious, ( ) infectious, not yet asymptomatical, and, ( ) symptomatical (may and anderson ; may et al. ) . the lifetime of an individual should cover not only the process from health to infection, but also the sub-stages after infection. thus, we assume that each agent can go through the following states: • s : healthy state, initially, each agent is set to be in s state. healthy agents have no risk of being infected. when a healthy agent moves into the neighborhood of an infectious one or an infectious agent approaches him, the healthy agent's state will change from s to s because contacts with infectives incur the danger of infection. as for an agent in s state, it can transit in two directions: one direction is to change from s back to s , after all its infectious neighbors move away (or its dead neighbors are removed from the grid) or he leaves the neighborhoods of its infectious neighbors; the other direction is to change from s to s if he unluckily get infected. note that we assume infection is instantaneous, i.e., instantaneous transmission from an infected individual to a susceptible. a newly infected agent is unable to transmit hiv virus until seroconversion. the s state corresponds to the early stages of hiv infection. let t denote the duration of this period. empirical works have been done to estimate the parameter. , anderson and medley ( ) report t to lie between and days in transfusion-induced aids cases. in our model we assume that t is a random variable following a normal distribution with the mean μ and the variance σ . after t , the infected agent enters s state: infectious state. medically, the duration during which an infected is infectious but not yet symptomatic is called incubation period. we let t donate the period. empirical work suggests an average incubation period of around to around years (medley et al. . a weibull distribution are commonly used to describe this incubation period (see, e.g., anderson ; anderson and medley ) . furthermore, , anderson and medley ( ) estimated t with a weibull distribution (with a mean of . years and a median of . year) based on transfusioninduced aids cases. for simplicity, we take t as a real number drawn from a normal distribution with the mean μ and the variance σ rather than a weibull distribution. it should be pointed out that the simulation results generated with the normal distribution here are proved to have little, if any, difference compared with those generated when a weibull distribution is employed. during the t period, hiv viruses in the victim's body are constantly cloning themselves and eventually the immunity system collapses. at this point, the victim starts to show some symptoms and thus transit to s state: symptomatic stage. as usually, let t denote the duration of this period. rothenberg et al. ( ) report - year survival rates among idus (intravenous drug user) in new york city and find a median time of survival of days. chang et al. ( ) report a median survival time of . months. empirical work shows that almost all hiv infectives, excluding those who die from other causes, will inevitably develop aids and die of it (may and anderson ) . similarly, we assume t follows a normal distribution with the mean μ and the variance σ . eventually, the ill agent enters s state after t passes by agents in s state will be removed from the population at the beginning of the next time and all of their uninfected neighbors will be released from s state, back to s state. generally, these state transitions take place in the order of s , s , s , s , s , and s . it is impossible for an agent to return from s state to s or s state. this backward transition s to s , demonstrated by a dashed line in fig. , is due to the disappearance of threats posed by infectious agents. moreover, although the transitions among s , s , and s state are not relevant to the propagation process, this process is closely related to hiv/aids transmission. taking account of this procession is essential for a better understanding of hiv/aids transmission. in the model, all the events triggering transitions could be divided into two categories. one category is a rule-based, such as healthy-to-dangerous, dangerous-to-infected, and dangerous-to-healthy state-transition events. these events occur according to the ca transition rules: an agent's state at time t + is based not only on its own state but also on the states of its neighbors at time t . the other category is time-based, meaning that these events are scheduled at pre-specified times. for instance, an agent entering s state will be assigned a time indicating when to change to s state. after that amount of time elapse, the transition occurs spontaneously. despite the distinction between these two categories, the subtlety of implementing the two event-triggering mechanisms is very trivial and leaves no further elaboration necessary. actually, these six states can be divided into three "super" classes in terms of the taxonomy used in the classical mathematical models: s and s states correspond to the susceptible state; s , s , and s states belongs to the infection state, and s state is the removed state. obviously, s and s state could be treated as one single state without changing any results. the reason why we divide this into two sub-states is that this makes our model easily implemented and our logic decent and legible. hiv/aids epidemic differs from other epidemics in that its dynamics is heavily affected by individual's behavioral patterns and the interactions between them. for example, careful sex practices and sanitization measures in drug taking will make individuals less likely to be infected. behavioral patterns and interactions are mostly determined by individual's life styles, personalities, social networks, etc. however, the majority of models fail to take account of the heterogeneity in agents' behaviors. to capture this, we extend classical ca models by allowing each agent to have its own attributes such as mobility, infectivity, resistibility (susceptibility) and different extent of neighborhood. assume that each cell in the grid can be occupied by at most one agent at a time. at time t , agent i can move from one cell into one of its adjacent cells with probability p m i . here, p m i is a measurement of agent i's activity level. it is a fixed real number, drawn from a uniform distribution (p m min , p m max ) ( ≤ p m min ≤ p m max ≤ ). when p m min = p m max , the activity level across agents is equal and therefore agents have the same inclination to move around. one extreme case is p m min = p m max = , which corresponds to the situation in which agents stay in their initial places during simulation, whilst p m min = p m max = means that each agent will move into one empty neighboring cell at almost each time (he could get stocked and not move anywhere if it's neighborhood is occupied). it is easy to induce that the average time per move is calculated as /p m i . intuitively, high level of activity leads to speedy spreading. our simulation results verify this. besides the heterogeneity in agents' activity, another kind of heterogeneity is introduced when we assign various levels of infectivity and susceptibility to agents. let f i denote the infectivity level of agent i. f i is a real number drawn uniformly from the interval ( , ). it measures the possibility that agent i transmits hiv viruses to others when they meets. evidently, greater values of f i indicate higher infectiousness of agent i. suppose also that each agent has resistance to being infected. we denote this resistibility as r i for agent i. similarly, r i is also a real number drawn uniformly from the interval ( , ) and has the property that the greater the resistibility, the less is the chance of getting infected. note that the infectivity of an agent need not be a constant. an agent can have different level of infectivity, depending both on its state as well as on its behavior. it is widely believed that infectives experience two periods of high infectivity (see e.g. may and anderson ; may et al. ), one shortly after being infected and the other at the late stage of his illness. another example is that a patient might have high infectivity during the incubation period and low infectivity owing to good health care during the symptomatic period. although our model allow for various infectivity at different stages for a single agent, we adopt the fixed infectivity for each agent. in doing so, we can focus our attention on some significant issues. we leave various infectivity scenarios for future work. conventional ca models define two types of neighborhoods: moore neighborhood and von neumann neighborhood. in this paper, we extend the concept of ca neighborhood in order to better describe various situations encountered in agent-based modeling. figure illustrates the definition. as we can see in fig. b shows the classical moore neighborhood, and fig. d classical von neumann neighborhood. figure c represents an extended moore neighborhood with the order of × , and fig. e an extended × von neumann neighborhood. specially, fig. a can be simply viewed as an extended × moore (or von neumann) neighborhood. note that fig. b , f-i have the neighborhoods with one direction. this directional structure is able to capture the biases or preferences embedded in individuals' behavioral patterns and we can use different directions to represent variable ways of interactions. it is easy to see that the greater is the neighborhood, the larger extent to which an agent can exert its influence to its neighbors. given the above neighborhood definition, a concept of distance is naturally induced. let a pair of integer numbers (x, y) represent an agent's coordinates in the grid. the distance between agent i and j is thus given as ( ) next, we specify that the influence indicator m i,j of agent i and j satisfies the following condition: that is to say, the influence intensity is inversely proportional to the distance between them if the influence can be exerted, and zero otherwise. therefore, the infective impact i i,j of agent i on agent j can be expressed as and the probability of an agent infected by one of its neighbors is defined as: where p(·) is a function satisfying the conditions: ( ) p(·) is a real-valued function with the value between zero and one; ( ) p(·) is increased in i i,j and decreased in r i . in this model, we assume p(·) takes the form of the following equation: here, p(i i,j , r i ) is interpreted as the probability of agent i being infected by its neighbor j . denoted by b i the set of all its neighbors, agent i's overall probability of infection is thus given by it indicates that this overall probability is determined by the most influential neighbor. such specification makes sense in most cases. the model developed in sect. is implemented using java programming language with the repast software package. detailedly commented source code is available from the authors upon request. next, we begin our analysis by considering first a typical simulation run as a benchmark case. . benchmark case table lists the input parameters chosen for the benchmark case. in this case, the grid consists of × sites (w = ). the population size n is set to with the initial infected ratio α = . . all agents are homogeneous in terms of having the same infectivity f i = . , resistibility r i = . , and × moore neighborhood. they are uniformly distributed in the grid. the total simulation time t for each run is set to . figure shows a snapshot of the spatial distribution of the population at some time in a typical simulation. it is commonly believed that as the epidemic develops, its spread ends up with two typical situations: extinction or prevalence. figure depicts these two situations. in fig. a , the number of infections climbs early in the process. after about time t = , the infection level starts to drop slowly until it reaches zero at time t = , while in fig. b , the number of infections increases slowly and reaches an equilibrium level after t = . the intuition . the spatial distribution of the population at time t = behind is that in the first case, newly infected continuously enter the pool of infectives at a fairly low rate in early stages. after the lengthy incubation period, these infectives begin to develop aids and eventually die. the total number of them drops when the infection rate is very low with regard to the rate at which infectives leave the pool for some reason (dead in the model). while in the second case, healthy agents get infected at a relatively high rate in early stages. the infection level continues to increases because the number of removed agents is relatively small in later stages. thus high infection rate often leads to prevalence as demonstrated in fig. b . these two results can be found in real-world situations. notice that fig. the infections vs. time result in two simulations in the parameters setting, the growth rate r is almost zero (r = . ). in next subsection, we will explore the effects of various factors, such as population density, initial infection ratio, and infectivity, on the epidemic. now we keep other parameters constant as before and let the population density β vary to see how β (β = n/w ) affects the dynamics of hiv/aids transmission. tarwater and martin ( ) investigate this issue when studying the outbreaks of measles or measles-like infectious diseases. as one would expect, many common infectious diseases spread more rapidly at a high population density than at a low population density. figure illustrates the time series of the mean numbers of infectives for different population sizes n = , , , , and . we can see that when population density is relatively low (n = , ) , the infection levels are relatively low during the entire simulation and decline slowly in the later stages. this suggests that the epidemic died out eventually. in the and at last. for n = and , the infection numbers reach to a very high level and then drop rapidly. the collapse is because that so many infectives are removed from the model that the pool of infectives shrinks. for clarity, we also plot the fractions of infectives in the population vs. time in fig. . clearly, in late stages of the epidemic, the fractions are greater when β is great than when β is small. in summary, hiv/aids infection is more likely to persist at higher population densities. this is due to that with the population density increasing, the population contact rate rise, leading to increases in the probability of infection. early work (see, e.g., rhodes and anderson ) suggests that there is a threshold below which the epidemic would eventually dies out and above which it would persist. due to the limitation of ca methods, it is diffi- fig. mean number of infections vs. time for different initial infected ratios cult to pin down its exact value. however, with many simulation runs, we still can give an approximate interval in which the threshold lies. an interesting question one may ask is how epidemic spreading is affected by initial configurations of susceptibles and infectives, or whether the multiple infection sources will make the disease more likely to become endemic. with other parameters fixed as before, we run the simulations with α = . , . , . , and . , respectively. figure presents the result. clearly, as α increase, the infection level shifts upwards. in the case of α = . , the infection level reach at time t = , higher than in the case of α = . . by contrast, the level in the case of α = . climb to around at time t = and drops slightly to at time t = . the maximum infection is reached in the case of α = . at time t = , which is more than . spatially, more sources of infection imply greater chance of being infected within a certain area, letting hiv/aids epidemics to be more likely to spread out and persist. statistically speaking, an individual's probability of infection is generally proportional to the number of infectious sources. intuitively, the more migratory the population, the more likely that an epidemic is to spread. suppose an agents' activity can be measured by the number of contacts it makes with others within a unit of period of time. as a result, our model assumes that individuals' activity is measured by mobility. later stages. this is, in the case of p m max = . , the level is above and in the case p m max = . , the level is in the range ( , ). in the last case of p m max = . , the infection level fluctuates above , higher than those of other cases. so we conclude that mobility plays a significant role in the dynamics. it could explain why chinese government took rather strong measures to control the migratory people and quarantine the infectives or the suspects during the outbreak of sars in the spring of . as to our model, if agents are configured with higher mobilities, it is more likely that the hiv/aids infection can persist in the population, whilst if configured with lower mobilities, the infection would gradually diminish and eventually die out. next, we are to examine how neighborhood forms affect hiv/aids epidemic dynamics. the parameter sets are kept the same as in benchmark case, except for the adoption of different neighborhood forms. figure illustrates the simulation results generated in two cases: one with × von neumann neighborhood and the other with × moore neighborhood. in the case where the von neumann neighborhood is used, the level of infection goes up to about . it is clearly greater than that of the case with × moore neighborhood in which the level only reach about . such result suggests that with wider neighborhood, an agent is more likely to get influenced by its neighbors and therefore the likelihood of getting infected increases accordingly. it is easy to induce that the infection level rises with the order of neighborhood. this result also suggests that hiv/aids epidemic dynamics is significantly affected by strong interactions between agents. we now turn to examine heterogeneous mixing, i.e., different at-risk groups coexist. usually, heterogeneous mixing will make the dynamics more complicated and unpredictable. in the following, we assume the whole population is divided into four different groups as shown in table . as shown in table , class p has very low infectivity, low susceptibility, and × von neumann neighborhood. it can represents children and (or) elders in the population who hardly infect others and are easy to be infected. class pl refers to ordinary people who have relatively low infectivity and high resistibility (therefore low susceptibility). in our model, this class amounts to a large fraction of the whole population. class ph and class ph+ can represent the two high-risk groups observed in real life. agents of class ph have high infectivity and low resistibility duo to their high-risk behaviors like incautious sex without protection, needle sharing, unhygienic blood transfusion and so on. the biased × (or × ) von neumann neighborhood captures their potential oriented or biased behaviors. in contrast, agents of class ph+ with both higher infectivity and higher susceptibility represent those who are, although being in the minority, the most dangerous and malevolent group. such group does exist in real life. for instance, some crimes were reported in china in recent years that a few aids infectives intentionally have sex with innocent people or shot people with contaminated syringes in public places. they blame their infection on the society and the government for not being able to provide necessary health service and compensating too little. the × moore neighborhood indicates their intensive influence on others. we distinguish class ph+ from class ph in order to see whether such malevolent behaviors have significant impacts on the spread of hiv/aids and to what extent. while such rough classification may be incorrect or even erroneous, it surely is reasonable and well supported by our extended ca model. table gives the values of f i and r i used in the following simulations. as we will see later, the results generated with this classification are fairly consistent with those obtained through empirical work. figure gives a typical simulation result in the four-class scenario. the fractions of four classes here are: n p = . , n pl = . , n ph = . , and n ph+ = . . the infection curve in fig. is quite similar to that obtained in the single-class scenarios except that its level is fairly higher. in this section, we will investigate the effect of agents' susceptibility on the dynamics of hiv/aids. given n p = . , n pl = . , n ph = . , n ph+ = . and others as before, let r ph+ vary. figure gives the infection levels when r ph+ is set to be . , . , . , and . , respectively. as we can see, these equilibrium infections are almost at the same level, which is inconsistent with our expectation. the differences are so small that we cannot assure with confidence whether changes in susceptibility have impact on the epidemic dynamics. the possible reasons, we believe, are twofold: first, the role played by susceptibility may be not as decisive as the above factors. second, we may describe susceptibility in the wrong way and make it an inessential factor in our model. future work will reconsider this issue and find the better way to describe susceptibility. at last, we will examine whether changes in the fractions of some classes can affect epidemic behaviors. given n p = . , n pl = . and other parameters as before, we take n ph = . (n ph+ = . ), . ( . ), . ( . ), . ( . ), . ( . ), respectively. figure shows the infection levels in these five combinations. as observed from fig. , we obtained very similar results, compared with those in the foregoing analysis. with n ph+ decreasing, infection level declines. recall that ph+ has more influence on its neighbors than ph, thus leading to greater transmission to others and larger infection rate. this finding suggests the government should pay more attention to those who have high-risk life styles and revengeful behaviors. this makes it the essential issue of how highly infectious and malevolent individuals are restricted and controlled. the focus of the paper is on the modeling of the entire course of hiv/acid epidemics and heterogeneity in agents' behaviors. even though classical ca models are capable of describing the spread of common epidemics but fail to represent the complicated epidemics like hiv/aids disease. the ignorance of heterogeneity gives rise to unacceptable errors in the prediction of the development trends. in addition, components of a conventional ca system such as topological forms of grid, the definition of neighborhood, and state transition rules are simple and unchanged over time. this makes the modeling of complicated dynamics such as hiv/aids transmission difficult and uncontrollable. in this paper, we have developed an extended ca model to capture key epidemiological and clinical features of hiv/aids epidemic. first, we explicitly models and simulates the whole progression of hiv/aids disease (i.e., infected but not infectious, infectious but asymptomatic, symptomatic, and deceased). such improvement can give us a better understanding of the dynamics during the entire hiv/aids epidemics. in order to examine various degrees of influence between agents, we have introduced an extended definition of neighborhood to represent the intensity and bias of influence. this lets us gain insight into how various degrees of interactions affects the hiv/aids epidemics. another type of heterogeneity in disease-related attributes such as susceptibility, infectivity and durations of epidemic phrases is also taken into consideration. moreover, we also consider the effect of agents' mobility on epidemics dynamics. given all the improvements, we have obtained richer simulation results similar to those usually found in the mathematical models or other classical simulation models. we have identified some influential factors that greatly affect the hiv/aids epidemic dynamics. the main findings are that ) hiv/aids epidemic can end up in the two regimes: extinction and persistence; ) with these factors such as agents' mobility, population density, initial infection ratio, and the extent of neighborhood increasing, the infection level get higher. after crossing some critical point, the regime generated could change from dying-out to persistence at some point. this result is robust across many of the tested parameter combinations; ) in four-class scenarios, the great fraction of 'super' infectives (the ph+ class in our model) can also produce higher level of infection. however, our simulation study above is still preliminary. there are some issues needed to be addressed. first, as said before, we should redefine susceptibility in a better way to check its role in the dynamics of hiv/aids epidemic. second, most models posit that a virus carrier's infectivity is constant during the progress of a disease. however, this is not the case for hiv/aids epidemic. various infectivity at different stages could have substantial impact on the dynamics of hiv/aids transmission. this problem needs special attention. additional, further developments of our model, e.g. by adding age-related structure (griffiths et al. ) , different subgroup classification, and other heterogeneity, would greatly add to the appeal of this model. with these additions, a better understanding of hiv/aids and thorough empirical work are required. finally, a natural extension of the model is to include the assessment of various control policies and managerial strategies, and this will be a firm support for the decision-making in prevention programs against hiv/aids. on modeling epidemics. including latency, incubation and variable susceptibility the epidemiology of hiv infection: variable incubation plus infectious periods and heterogeneity in sexual activity infectious diseases of humans: dynamics and control possible demographic consequences of aids in developing countries epidemiology of hiv infection and aids: incubation and infectious periods, survival and vertical transmission a simulation model of the dynamics of hiv transmission in intravenous drug users a comparison of simulation models applied to epidemics the mathematical theory of infectious diseases and its applications dynamics of hiv infection on d cellular automata models for transmission of disease with immigration of infectives survival and mortality patterns of an acquired immunodeficiency syndrome (aids) cohort in new york state mathematical modeling in epidemiology cellular automata and epidemiological models with spatial dependence a heterogeneous cellular automata model for sars transmission a cellular automaton describing the formation of spatially ordered structures in chemical systems a cellular automaton model of excitable media a simple agent model of an epidemic an age-structured model for the aids epidemic the differential infectivity and staged progression models for the transmission of hiv a model for the influence of the greenhouse effect on insect and microorganism geographical distribution and population dynamics a model for predicting forest fire using cellular automata. ecological modelling a contribution to the mathematical theory of epidemics the dynamics of hiv spread: a computer simulation model flood decision support system on agent grid: method and implementation algebraic properties of cellular automata transmission dynamics of hiv infection the transmission dynamics of human immunodeficiency virus (hiv) [and discussion incubation period of aids in patients infected via blood transfusion the distribution of the incubation period for the acquired immunodeficiency syndrome (aids) mathematical biology i experiences creating three implementations of the repast agent modeling toolkit persistence and dynamics in lattice models of epidemic spread epidemic thresholds and vaccination in a lattice model of disease spread survival with the acquired immunodeficiency syndrome. experience with cases in new york city effects of population density on the spread of disease diffusion and wave propagation in cellular automaton models of excitable media modelling the hiv epidemic: a state-space approach effectiveness of q-learning as a tool for calibrating agent-based supply network models we would like to express our gratitude to the many people in xi'an jiaotong university who have participated in planning, data collection, and presenting the results. without their efforts, this simulation modeling would not have been possible. this work is supported by nsfc under contract . we are grateful to ting zhang and jie huang for excellent research assistance and the referees for many helpful comments that greatly improved the presentation. key: cord- -phepjf authors: hsieh, ying-hen; fisman, david n; wu, jianhong title: on epidemic modeling in real time: an application to the novel a (h n ) influenza outbreak in canada date: - - journal: bmc res notes doi: . / - - - sha: doc_id: cord_uid: phepjf background: management of emerging infectious diseases such as the influenza pandemic a (h n ) poses great challenges for real-time mathematical modeling of disease transmission due to limited information on disease natural history and epidemiology, stochastic variation in the course of epidemics, and changing case definitions and surveillance practices. findings: the richards model and its variants are used to fit the cumulative epidemic curve for laboratory-confirmed pandemic h n (ph n ) infections in canada, made available by the public health agency of canada (phac). the model is used to obtain estimates for turning points in the initial outbreak, the basic reproductive number (r( )), and for expected final outbreak size in the absence of interventions. confirmed case data were used to construct a best-fit -phase model with three turning points. r( )was estimated to be . ( % ci . - . ) for the first phase (april to may ) and . ( % ci . - . ) for the second phase (may to june ). hospitalization data were also used to fit a -phase model with r( )= . ( . - . ) and a single turning point of june . conclusions: application of the richards model to canadian ph n data shows that detection of turning points is affected by the quality of data available at the time of data usage. using a richards model, robust estimates of r( )were obtained approximately one month after the initial outbreak in the case of a (h n ) in canada. epidemics and outbreaks caused by emerging infectious diseases continue to challenge medical and public health authorities. outbreak and epidemic control requires swift action, but real-time identification and characterization of epidemics remains difficult [ ] . methods are needed to inform real-time decision making through rapid characterization of disease epidemiology, prediction of shortterm disease trends, and evaluation of the projected impacts of different intervention measures. real-time mathematical modeling and epidemiological analysis are important tools for such endeavors, but the limited public availability of information on outbreak epidemiology (particularly when the outbreak creates a crisis environment), and on the characteristics of any novel pathogen, present obstacles to the creation of reliable and credible models during a public health emergency. one needs to look no further than the sars outbreak, or ongoing concerns related to highly pathogenic avian influenza (h n ) or bioterrorism to be reminded of the need for and difficulty of real-time modeling. the emergence of a novel pandemic strain of influenza a (h n ) (ph n ) in spring highlighted these difficulties. early models of ph n transmission were subject to substantial uncertainties regarding all aspects of this outbreak, resulting in uncertainty in judging the pandemic potential of the virus and the implementation of reactive public health responses in individual countries (fraser et al. [ ] ). multiple introductions of a novel virus into the community early in the outbreak could further distort disease epidemiology by creating fluctuations in incidence that are misattributed to the behavior of a single chain of transmission. we sought to address three critical issues in real time disease modeling for newly emerged ph n : (i) to estimate the basic reproduction number; (ii) to identify the main turning points in the epidemic curve that distinguish different phases or waves of disease; and (iii) to predict the future course of events, including the final size of the outbreak in the absence of intervention. we make use of a simple mathematical model, namely the richards model, to illustrate the usefulness of near realtime modeling in extracting valuable information regarding the outbreak directly from publicly available epidemic curves. we also provide caveats regarding inherent limitations to modeling with incomplete epidemiological data. the accuracy of any modeling is highly dependent on the epidemiological characteristics of the outbreak considered, and most epidemic curves exhibit multiple turning points (peaks and valleys) during the early stage of an outbreak. while these may be due to stochastic ("random") variations in disease spread, and changes in either surveillance methods or case definitions, turning points may also represent time points where epidemics transition from exponential growth processes to processes that have declining rates of growth, and thus may identify effects of disease control programs, peaks of seasonal waves of infection, or natural slowing of growth due to infection of a critical fraction of susceptible individuals. for every epidemic, there is a suitable time point after which a given phase of an outbreak can be suitably modeled, and beyond which subsequent phases may be anticipated. detection of such "turning points" and identification of different phases or waves of an outbreak is of critical importance in designing and evaluating different intervention strategies. richards [ ] proposed the following model to study the growth of biological populations, where c(t) is the cumulative number of cases reported at time t (in weeks): here the prime "′" denotes the rate of change with respect to time. the model parameter k is the maximum case number (or final outbreak size) over a single phase of outbreak, r is the per capita growth rate of the infected population, and a is the exponent of deviation. the solution of the richards model can be explicitly given in terms of model parameters as using the richard model, we are able to directly fit empirical data from a cumulative epidemic curve to obtain estimates of epidemiological meaningful parameters, including the growth rate r. in such a model formulation, the basic reproduction number r is given by the formula r = exp(rt) where t is the disease generation time defined as the average time interval from infection of an individual to infection of his or her contacts. it has been shown mathematically [ ] that, given the growth rate r, the equation r = exp(rt) provides the upper bound of the basic reproduction number regardless of the distribution of the generation interval used, assuming there is little pre-existing immunity to the pathogen under consideration. additional technical details regarding the richards model can be found in [ ] [ ] [ ] . unlike the better-known deterministic compartmental models used to describe disease transmission dynamics, the richards model considers only the cumulative infected population size. this population size is assumed to have saturation in growth as the outbreak progresses, and this saturation can be caused by immunity, by implementation of control measures or other factors such as environmental or social changes (e.g., children departing from schools for summer holiday). the basic premise of the richards model is that the incidence curve of a single phase of a given epidemic consists of a single peak of high incidence, resulting in an s-shaped cumulative epidemic curve with a single turning point for the outbreak. the turning point or inflection point, defined as the time when the rate of case accumulation changes from increasing to decreasing (or vice versa) can be easily pinpointed as the point where the rate of change transitions from positive to negative; i.e., the moment at which the trajectory begins to decline. this time point has obvious epidemiologic importance, indicating either the beginning of a new epidemic phase or the peak of the current epidemic phase. for epidemics with two or more phases, a variation of the s-shaped richards model has been proposed [ ] . this multi-staged richards model distinguishes between two types of turning points: the initial s curve which signifies the first turning point that ends initial exponential growth; and a second type of turning point in the epidemic curve where the growth rate of the number of cumulative cases begins to increase again, signifying the beginning of the next epidemic phase. this variant of richards model provides a systematic method of determining whether an outbreak is single-or multi-phase in nature, and can be used to distinguish true turning points from peaks and valleys resulting from random variability in case counts. more details on application of the multi-staged richards model to sars can be found in [ , ] . readers are also referred to [ , ] for its applications to dengue. we fit both the single-and multi-phase richards models to canadian cumulative ph n cumulative case data, using publicly available disease onset dates obtained from the public health agency of canada (phac) website [ , ] . phac data represent a central repository for influenza case reports provided by each of canada's provinces and territories. onset dates represent best local estimates, and may be obtained differently in different jurisdictions. for example, the province of ontario, which comprises approximately / of the population of canada, and where most spring influenza activity was concentrated, replaces onset dates using a hierarchical schema, whereby missing onset dates may be replaced with dates of specimen collection (if known) or date of specimen receipt by the provincial laboratory system, if both dates of onset and specimen collection are missing. data were accessed at different time points during the course of the "spring wave (or herald wave)" of the epidemic in may-july of , whenever a new dataset is made available online by the phac. by sequentially considering successive s-shaped segments of the epidemic curve, we estimate the maximum case number (k) and locate turning points, thus generating estimates for cumulative case numbers during each phase of the outbreak. the phac cumulative case data is then fitted to the cumulative case function c(t) in the richards model with the initial time t = being the date when the first laboratory confirmed case was reported and the initial case number c = c( ) = , (the case number with onset of symptoms on that day). there were some differences between sequential epidemic curves in assigned case dates. for example, data posted by phac on may indicated an initial case date of april , but in the june data this had been changed to april , perhaps due to revision of the case date as a result of additional information. model parameter estimates based on the explicit solution given earlier can be obtained easily and efficiently using any standard software with a least-squares approximation tool, such as sas or matlab. daily incidence data by onset date were posted by phac until june , after which date only the daily number of laboratory-confirmed hospitalized cases in canada was posted. for the purpose of comparison, we also fit the hospitalization data to the richards model in order to evaluate temporal changes in the number of severe (hospitalized) cases, which are assumed to be approximately proportional to the total cases number. the case and hospitalization data used in this work are provided online as additional file . we fit the model to the daily datasets, acquired in real time, throughout the period under study. the leastsquared approximation of the model parameter estimation could converge for either the single-phased or the -phase richards models. for the sake of brevity, only four of these model fits are presented in table to demonstrate the difference in modeling results over time. the resulting parameter estimates with % confidence intervals (ci) (for turning point (t i ), growth rate (r), and maximum case number (k)), time period included in the model, and time period when the data set in question were accessed, is presented in table . note that all dates in the tables are given by month/day. we also note that the ci's for r reflect the uncertainty in t as well as in the estimates for r, and does not reflect the error due to the model itself, which is always difficult to measure. in order to compare the -phase and -phase models, we also calculate the akaike information criterion (aic) [ ] for the first, third, and fourth sets of data in table , where there is a model fit for the -phase model. the results, given in table , indicates that whenever there is a model fit for the -phase model, its aic value is always lower than that of the -phase model and hence compares favorably to the -phase model. parameter estimates fluctuate in early datasets, and the least-squared parameter estimations diverge within and between -phase and -phase models in a manner that seems likely to reflect artifact. in particular, for the earliest model fits, using data from april to may , the estimated reproductive number for the second phase is far larger than that obtained in the first phase, and that obtained using a single-phase model, and illustrating the pitfalls of model estimation using the limited data available early in an epidemic. estimates stabilize as the outbreak progresses, as can be seen with the final data sets (april to june and april to june ). for comparison, we plot the respective theoretical epidemic curves based on the richards model with the estimated parameters described in the table above in figure . as noted above, model can be used to estimate turning points (t i ) and basic reproductive numbers (r .), if the generation time t is know. we used t = . days ( % ci: . - . ), as obtained in [ ] by fitting an age stratified mathematical model to the first recognized influenza a (h n ) outbreak in la gloria, mexico. estimates are presented in table . we also conducted sensitivity analyses with r # calculated based on longer generation times (t = . ( . , . )) for seasonal influenza in [ ] (see last column in table ). excluding implausibly high estimates of r generated using initial outbreak data (april to may ), we obtain the estimates of r for the -phase model that range between . and . . inasmuch as richards model analyzes the general trends of an epidemic (e.g., turning point, reproductive number, etc.), it can be used to fit any epidemiological time series for a given disease process, as long as the rate of change in the recorded outcome is proportional to changes in the true number of cases. as such, for comparison, we fit our model using the time series for ph n hospitalizations in canada posted by phac on july [ ] (that last date these data were made available) ( table ) . this time series was easily fit to a one-phase model ( figure ) . further examples of using hospitalization or mortality data to fit the richards model can be found in [ ] . we used the richards model, which permits estimation of key epidemiological parameters based on cumulative case counts, to study the initial wave of influenza a (h n ) cases in canada. in most model fits, april - and may - were identified as early turning points for the outbreak, with a third and final turning point around june - in models based on longer time series. although this modeling approach was not able to detect turning points using some earlier data sets (e.g., those limited to the period from april to may ), in general the turning points identified were consistent across multiple models and time series. perhaps the most important divergence between models occurred with the detection of an april turning point in the case report time series, but not in the time series based on hospitalized cases. we believe this may be attributable to the small number of hospitalizations, relative to cases, that had occurred by that date, as well as the fact that hospitalization data only became available on april . the turning point can correspond to the point at which disease control activities take effect (such that the rate of change in epidemic growth begins to decline) or can represent the point at which an epidemic begins to wane naturally (for example, due to seasonal shifts or due to the epidemic having "exhausted" the supply of susceptibles such that the reproductive number of the epidemic declines below ). this quantity has direct policy relevance; for example, in the autumn ph n wave in canada, vaccination for ph n was initiated at or after the turning point of the autumn wave due to the time taken to produce vaccine; as the epidemic was in natural decline at that point, the impact of vaccination has subsequently been called into question. although the richards model is able to capture the temporal changes in epidemic dynamics over the course of an outbreak, it does not define their biological or epidemiological basis. as such, determining the nature of these turning points requires knowledge of "events on the ground" for correlation. we suspect that the last note that all dates in the tables are given by month/day. dates of posting are listed in parentheses. model duration indicates whether they fit a -phase or phase model. note that the maximum case number is rounded off to the nearest integer. r # is obtained using the generation interval of t = . ( . , . ) for seasonal influenza [ ] . table comparison of akaike information criterion (aic) values between -phase and -phase models for time periods with -phase model fit in table time table , last line) and a -phase model using hospitalization data (june ), this lag in turning points would actually be expected, due to the time from initial onset of symptoms until hospitalization, which was reported to have an interquartile range of - days in a recent study from canada [ ] . timelines for the -phase model for case data of / - / and the -phase model for hospitalization data are presented graphically in figure . in addition to identifying turning points, the richards model is useful for estimation of the basic reproductive number (r ) for an epidemic process, and our estimates derived using a richards model were consistent with estimates derived using other methods. for example, our r agrees almost perfectly with that of tuite et al., derived using a markov chain monte carlo simulation parameterized with individual-level data from ontario's public health surveillance system [ ] . our estimates of r is smaller than that derived by fraser et al. [ ] using mexican data, but such differences could relate in part to the different age distributions of these two countries [ ] , and may also reflect the fact that our estimate is obtained canadian data at a national level, while empirical mexican estimates were based on data from the town of la gloria with only residents. most epidemic curves in the early stage of a novel disease outbreak have multiple phases or waves due to simple stochastic ("random") variation, mechanisms of disease importing, initial transmission networks and individual/community behavior changes, improvements in the performance of surveillance systems, or changes in case definitions as the outbreak response evolves. however, changes in phase (signified by the presence of turning points identified using the richards model) may also pinpoint the timing of important changes in disease dynamics, such as effective control of the epidemic via vaccination or other control measures, depletion of disease-susceptible individuals (such that the effective reproductive number for the disease decreases to < ), or the peak of a "seasonal" wave of infection, as occurs with [ , , ] , some competing methods require more extensive and detailed data than are required to build a richards model, which requires only cumulative case data from an epidemic curve. as we also demonstrate here, the richards model produces fairly stable and credible estimates of reproductive numbers early in the outbreak, allowing these estimates to inform evolving disease table , derived using early case data accessed on may , closely approximate our final estimates (table , last row) . thus, while early estimation with the richards model failed to correctly detect turning points or accurately estimate the final outbreak size, it was nonetheless useful for rapid estimation of r within a month of first case occurrence in canada. as with any mathematical modeling technique, the approach presented here is subject to limitations, which include data quality associated with real-time modeling (as data are often subject to ongoing cleaning, correction, and reclassification of onset dates as further data become available), reporting delays, and problems related to missing data (which may be non-random). in our current study, the hierarchical approach used by canada's most populous province (ontario) for replacement of missing data could have had distorting effects on measured disease epidemiology: the replacement of missing onset dates with dates of specimen collection could have resulted in the artifactual appearance of early turning points identified by our model, due to limitations in weekend staffing early in the outbreak. if, as we believe to be the case, public health laboratories did not have sufficient emergency staffing to keep up with testing on weekends such that weekend specimen log-ins declined sharply, this would have created the appearance of epidemic "fade out" on weekends. other factors that might distort the apparent epidemiology of disease include changes in guidelines for laboratory testing of suspected cases, improved surveillance and public health alerts at later stages of the outbreak leading to increased case ascertainment or over-reporting of cases [ ] . however, the quality of the time series will tend to improve with the duration of the epidemic, both because stochastic variation is "smoothed out", and also because small variations become less important as the cumulative series becomes longer. we note that a further application of the richards model in the context of influenza would relate to comparison of the epidemiology of the influenza a h n epidemic to past canadian epidemics, though such an endeavor is beyond the scope of the present study. in summary, we believe that the richards model provides an important tool for rapid epidemic modeling in the face of a public health crisis. however, predictions based on the richards model (and all other mathematical models) should be interpreted with caution early in an epidemic, when one need to balance urgency with sound modeling. at their worst, hasty predictions are not only unhelpful, but can mislead public health officials, adversely influence public sentiments and responses, undermine the perceived credibility of future (more accurate) models, and become a hindrance to intervention and control efforts in general. additional file : electronic supplementary material. canada novel influenza a(h n ) daily laboratory-confirmed pandemic h n case and hospitalization data. epidemic science in real time pandemic potential of a strain of influenza a (h n ): early findings a flexible growth function for empirical use how generation intervals shape the relationship between growth rates and reproductive numbers sars epidemiology. emerging infectious diseases real-time forecast of multi-wave epidemic outbreaks. emerging infectious diseases richards model: a simple procedure for real-time prediction of outbreak severity intervention measures, turning point, and reproduction number for dengue turning points, reproduction number, and impact of climatological events on multi-wave dengue outbreaks public health agency of canada: cases of h n flu virus in canada public health agency of canada: cases of h n flu virus in canada a new look at the statistical model identification estimation of the serial interval of influenza pandemic influenza a (h n ) during winter influenza season in the southern hemisphere. influenza and other respiratory viruses critically ill patients with influenza a(h n ) infection in canada estimated epidemiologic parameters and morbidity associated with pandemic h n influenza age, influenza pandemics and disease dynamics comparative estimation of the reproduction number for pandemic influenza from daily case notification data the ideal reporting interval for an epidemic to objectively interpret the epidemiological time course initial human transmission dynamics of the pandemic (h n ) virus in north america. influenza and other respiratory viruses on epidemic modeling in real time: an application to the novel a (h n ) influenza outbreak in canada authors' contributions yhh conceived the study, carried out the analysis, and wrote the first draft. df interpreted the results and revised the manuscript. jw participated in the analysis, the interpretation of results, and the writing. all authors read and approved the final manuscript. the authors declare that they have no competing interests. key: cord- -osol wdp authors: ma, junling title: estimating epidemic exponential growth rate and basic reproduction number date: - - journal: infect dis model doi: . /j.idm. . . sha: doc_id: cord_uid: osol wdp the initial exponential growth rate of an epidemic is an important measure of the severeness of the epidemic, and is also closely related to the basic reproduction number. estimating the growth rate from the epidemic curve can be a challenge, because of its decays with time. for fast epidemics, the estimation is subject to over-fitting due to the limited number of data points available, which also limits our choice of models for the epidemic curve. we discuss the estimation of the growth rate using maximum likelihood method and simple models. this is a series of lecture notes for a summer school in shanxi university, china in . the contents are based on ma et al. (ma, dushoff, bolker, & earn, ) . we will study the initial exponential growth rate of an epidemic in section , the relationship between the exponential growth rate and the basic reproduction number in section , an introduction to the least square estimation and its limitations in section , an introduction to the maximum likelihood estimation in section , and the maximum likelihood estimation of the growth rate in section . epidemic curves are time series data of the number of cases per unit time. common choices for the time unit include a day, a week, a month, etc. it is an important indication for the severeness of an epidemic as a function of time. for example, fig. shows the cumulative number of ebola cases during the e ebola outbreak in western africa. the cumulative cases during the initial growth phase form an approximately linear relationship with time in log-linear scale. thus, in linear scale, the number of deaths increases exponentially with time. the mortality curve (the number of deaths per unit time) shows a similar pattern, as demonstrated by the daily influenza deaths in philadelphia during the influenza pandemic shown in fig. . in fact, most epidemics grow approximately exponentially during the initial phase of an epidemic. this can be illustrated by the following examples. where s is the fraction of susceptible individuals, i is the fraction of infectious individuals, and r is the fraction of recovered individuals; b is the transmission rate per infectious individual, and g is the recovery rate, i.e., the infectious period is exponentially distributed with a mean =g. linearize about the disease-free equilibrium (dfe) ð ; ; Þ, di dt zðb À gÞi: ( ) thus, if b À g > , then iðtÞ grows exponentially about the dfe. in addition, initially, sz , thus, the incidence rate (number of new cases per unit time) c ¼ bsi also increases exponentially. it is similar for an susceptible-exposed-infectious-recovered (seir) model, as illustrated by the following example. example . lets consider an seir model: where e is the fraction of latent individuals (infected but not infectious), s the rate that latent individuals leaving the class, i.e; , the mean latent period is exponentially distributed with mean =s; s, i, r, b and g are similarly defined as in example . again, ð ; ; ; Þ is a disease free equilibrium representing a completely susceptible population. linearize about this equilibrium, the equations for e and i are decoupled, and become de dt note that the jacobian matrix j ¼ Às b s Àg ! has two real eigenvalues, namely, l ¼ Àðs þ gÞ þ ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi ffi ðs À gÞ þ sb q ; l ¼ Àðs þ gÞ À ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi ffi ðs À gÞ þ sb q : thus, about the dfe, the solution of the model is asymptotically exponential with a rate l . similar to example , the incidence rate also grows exponentially initially. in general, suppose the infection states of an individual can be characterized by the following vector ð s ! ; i ! Þ, where s ! represents multiple susceptible states, and i ! represents multiple infectious (or latent) states. we also use s ! and i ! represent the number of individuals in each state. also assume that the epidemic can be modeled by the following generic system Þ is a dfe, and the initial number of infectious individuals i ! ð Þ is very small, then, initially, the dynamics of i is governed by the following linearized system if the def is unstable, then iðtÞ grows asymptotically exponentially. . the exponential growth rate and the basic reproduction number the exponential growth rate is, by itself, an important measure for the speed of spread of an infectious disease. it being zero is, like the basic reproduction number r ¼ , a disease threshold. the disease can invade a population if the growth rate is positive, and cannot invade (with a few initially infectious individuals) if it is negative. in fact, it can be used to infer r .there are two approaches to infer r from the exponential growth rate, a parametric one, and a non-parametric one. for the parametric approach, we need an underlying model that gives both the growth rate and r . example . consider the sir model ( ) in example . note that ð ; ; Þ is an disease free equilibrium, representing a completely susceptible population. as we discussed above, the exponential growth rate is l ¼ b À g. note that the basic reproduction number is r ¼ b=g . if, for example, g is estimated independently to l, then, lets look at a more complicated example. express b in terms of l and substitute it into r , then thus, if the mean infectious period =g and the mean latent period =s can be independently estimated on l, then r can be inferred from l. typically, for an epidemic model that contains a single transmission rate b, if all other parameters can be estimated independently to the exponential growth rate l, then l determines b, and thus determines r . models can be overly simplified for mathematical tractability. for example, both the sir model in example and the seir model in example assume exponentially distributed infectious period. however, the infectious period and the latent period are mostly likely not exponential. wallinga and lipsitch (wallinga & lipsitch, ) developed a non-parametric method to infer the basic reproduction number from the exponential growth rate without assuming a model. let hðaÞ be the probability that a random individual remain infectious a time units after being infected (i.e., a is the infection age); bðaÞ is the rate of transmission at the infection age a. then, tðaÞ ¼ hðaÞbðaÞ is the transmissibility of a random infectious individual at the infection age a, assuming that the whole population is susceptible. thus, in addition, we assume that the population is randomly mixed, i.e., every pair of individuals have identical rate of contact. let cðtÞdt be the number of new infections during the time interval ½t;t þ dt, that is, cðtÞ is the incidence rate, and sðtÞ be the average susceptibility of the population, i.e., the expected susceptibility of a randomly selected individual. in addition, new infections at time t is the sum of all infections caused by infectious individuals infected a time unit ago (i.e., at time t À a) if they remain infectious at time t (with an infectious age a) and their contact is susceptible. that is, and thus cðtÞ ¼ sðtÞ to compute r , we need to normalize tðaÞ as a probability density function, note that wðaÞda is the probability that a secondary infection occurs during the infection age interval ½a; a þ da. that is, wðaÞ is the probability density function of the generation time, i.e., the time from being infected to generate a secondary infection. this generation time is also called the serial interval. with the serial interval distribution wðtÞ, this means that the cðtÞ is only determined by r , wðtÞ and sðtÞ. at the beginning of an epidemic, where the epidemic grows exponentially (with an exponential growth rate l), sðtÞz and cðtÞ ¼ c e lt where c is the initial number of cases at where mðxÞ ¼ r ∞ e xa wðaÞda is the moment generating function of the serial time distribution wðaÞ. equation ( ) links the exponential growth rate to the basic reproduction number though the serial interval distribution only. that is, if we can estimate the serial interval distribution and the exponential growth rate independently, that we can infer the basic reproduction number. note that the serial interval distribution wðtÞ can be estimated independently to the exponential growth rate. for example, it can be estimated empirically using contact tracing. alternatively, one can also assume an epidemic model. here we discuss a few simple examples. example . consider an sir model. let fðaÞ be the cumulative distribution function of the infectious period, and a constant transmission rate b. the probability that an infected individual remains infectious a time units after being infected is and thus the transmissibility is tðaÞ ¼ b½ À fðaÞ; and the serial interval distribution is where m is the mean infectious period. for the special case that the infectious period is exponentially distributed with a rate g, i.e., fðaÞ ¼ À e Àga , this model becomes model ( ). then the density function of serial interval distribution is which is identical to the density function of infectious period distribution. the moment generating function is note that the exponential growth rate is l ¼ b À g, then lets consider a more complex example with multiple infected states. example . consider an seir model with a constant transmission rate b. let fðaÞ and gðaÞ be the cumulative distribution functions of the infectious period and the latent period, respectively. given the latent period t l ¼ [ a, the probability that an infectious individual is infectious a time units after being infected is À fða À [Þ:thus, hence, the serial interval distribution is for the special case that the latent period is exponentially distributed with a rate s (i.e., fðaÞ ¼ À e Àga ) and the latent period is exponentially distributed with a rate s (i.e., gðaÞ ¼ À e Àsa ), this model becomes model ( ), and wðaÞ ¼ gse Àga z a e ðgÀsÞs ds ¼ ðge Àga ÞÃðse Àsa Þ: that is, if both distributions are exponential, the serial interval distribution is the convolution of the latent period distribution and the infectious period distribution. in this case, the basic reproduction number is where m i ðxÞ and m l ðxÞ are the moment generating functions of the infectious period and latent period, respectively. in equation ( ), r ðtÞ ¼ r sðtÞ is the reproduction number, and thus this equation can be used to estimate the production number at any time t during the epidemic given the incidence curve cðtÞ, namely, this is similar to, but different from, the nonparametric method developed by wallingua and teunis (wallinga & teunis, ) . the least squares method is one of the most commonly used methods for parameter estimation in mathematical biology. this method is in fact a mathematical method. for a family of curves f ðt; q ! Þ, where q ! r m is a vector of parameters of the family, this method finds the curve f ðt; b qÞ in the family that minimizes the distance between the curve and a set of points , and x ! be the euclidean norm in r n , then the mathematical formulation of the least squares method is where argmin gives the parameter q ! that minimizes the objective function. for our purpose, the observations fðt i ; x i Þg nÀ i¼ is the epidemic curve, i.e., x is the number of initially observed cases, and x i is the number of new cases during the time interval ðt iÀ ; t . we aim to find an exponential function f ðt; c ; lÞ ¼ c e lt that minimizes its distance to the epidemic curve, i.e., the parameters q ¼ ðc ; lÞ. there are two commonly use methods to estimate the exponential growth rate l: . nonlinear least square to fit to f ðt; c ; lÞ ¼ c e lt directly; . linear least square to fit fðt i ; lnx i Þg to ln f ðt; c ; lÞ ¼ lnc þ lt. the nonlinear least squares method does not have an analytic solution. numerical optimization is needed to solve the minimization problem ( ). the linear least square method has an analytic solution: let [ ¼ lnc , then the least squares problem becomes the objective function is a quadratic function of [ and l, thus, the minimum is achieved at i¼ y i , which represents the average of any sequence fy i g n i¼ , then, and thus the best fit exponential growth rate ls b l ¼ do these two methods yield the same answer? to compare, we simulate an epidemic curve of the stochastic seir model in example , using the gillespie method (gillespie, ) . the simulated daily cases (number of individuals showing symptom on a day) are then aggregated into weekly cases. then, we use both methods to fit an exponential curve to the simulated epidemic curve. the simulated epidemic curve and the fitting results are shown in fig. . this exercise illustrates a challenge of fitting an exponential model to an epidemic curve: how to determine the time period to fit the exponential model. the exponential growth rate of an seir model decreases with time as the susceptible population decreases. in fig. , the epidemic curve peaks in week . we choose a sequence of nested fitting windows starting in the first week and ending in a week w for w ¼ ; ;…; . the seir model has an asymptotic exponential growth, so the fitted exponential growth rate is not monotonic near the beginning of the epidemic. for larger fitting windows, both methods give an exponential growth rate that decreases with the length of the fitting window. we need more data points to reduce the influence of the stochasticity. however, using more data points also risks of obtaining an estimate that deviates too much from the true exponential growth rate. there is no reliable method to choose a proper fitting window. fig. also shows that the linear and nonlinear least squares methods may not yield the same estimate. this is because of a major limitation of both least squares methods: they implicitly assume that the deviations jx i À f ðt i ; q ! Þj carry identical weights. with the nonlinear method, later data points (at larger times) deviate more from the exponential curve than the earlier data points, because the exponential growth slows down with time. thus, the method is more biased to the later data points. with the linear method, the deviations in lnx i are more even than in x i , and thus the linear method is less biased to the later data points than the nonlinear method does. the least squares method, as mentioned above, is a mathematical problem. it does not explicitly assume any error distributions, and thus cannot give us statistical information about the inference. for example, if we use two slightly different fitting windows and get two slightly different estimates, is the difference of the two estimates statistically significant? such a question cannot easily be answered by the least squares method. interestingly, the least squares methods make many implicit assumptions to the deviations. we have mentioned the implicit equal-weight assumption above. it also implicitly assumes that the order of the observations does not matter, and that positive and negative deviations are equivalent. thus, they implicitly assume that the deviations are independently identically and symmetrically distributed. in statistics, the least squares method is commonly used in linear and nonlinear regression with an addition assumption that the errors are independently and identically normally distributed. however, these assumption on the errors may not be appropriate. for example, the new cases at time t þ may be infected by those who are infected at time t. thus, the number of new cases at different times may not be independent. also, the number of cases is a counting variable, and thus its mean and variance may be closely related, meaning that the error may not be identically normally distributed. in the next section, we address some of these problems using the maximum likelihood method. the maximum likelihood method is a commonly used statistical method for parameter inference; see, e.g., [(bolker, to construct the likelihood function we need to make assumptions on the error distribution. there are two types of error: the process error and the observation error. the observation error is the error in the observation process. for example, most people with influenza do not go to see a doctor, and thus there is no record of these cases, resulting in an under-reporting of the number influenza cases. also, many influenza related deaths are caused by complications such as pneumonia, and influenza may not be recorded as the cause. typos, miscommunication, etc, can all result in observation errors. the process error originates from the stochasticity of the system that is independent to observation. for example, the disease dynamics is fig. . the simulated seir epidemic curve (upper) and the fitted exponential growth rate as a function of the end of the fitting window (lower). the epidemic curve is simulated stochastically from the seir model in example using the gillespie method (gillespie, ) with the parameters b ¼ : , s ¼ , g ¼ : , the rates have a time unit of a day. the daily cases are then aggregated by week. the data points are taken at times t i ¼ i, i ¼ ; ; ; … weeks. the theoretical exponential growth rate is l ¼ : per week. intrinsically stochastic. the time that an infectious individual recovers, and the time that a susceptible individual is infected, are all random variables that affects the number of new infections at any time, even if we eliminate all observation errors. these two types of errors have very different nature, and thus need very different assumptions. for example, it is reasonable to assume that observation errors are independent to each other, but process errors at a later time are commonly dependent on the process errors at earlier times. if observation errors are large and process errors are negligible, then we assume that the random variable x i corresponding to the observation x i is independently distributed with a probability mass function p i ðk; q ! Þ where k is the values that x i can take. then, the likelihood function is the maximization of this likelihood function rarely has an analytic solution, and commonly needs to be solved numerically. note that each factor (probability) can be very small, and thus the product may be very difficult to minimize numerically because of rounding errors (from the binary representation of real numbers in computers). it is a common practice to maximize the log-likelihood function for example, we assume that the number of cases xðt i Þ at time t i is independently poisson distributed with mean m i ¼ c e lti . then, the log-likelihood function note that the observed cases x i are constants, and thus the last term can be ignored for maximization. thus, this maximization problem can only be solved numerically. we choose poisson distribution because its simple form greatly simplifies the log-likelihood function. in addition, it does not introduce more parameters, which is valuable to avoid over-fitting when the number of data points available is small. if the process error is not completely negligible, then choosing an overly dispersed distribution, such as the negative binomial distribution may be desirable. a negative binomial distribution has two parameters, the success probability q ! and the shape parameter r > . for simplicity, we assume that the shape parameter r is the same at each time t i , and will; be estimated together with the model parameters q ! ; but q depend on t i . the probability mass function is and the log-likelihood function is again, the last term can be ignored for the optimization problem. in addition, there is a constraint r > . if process errors are large and observation errors are negligible, then we cannot assume that the observed values x iþ and x i are independent to each other. instead, for all i ¼ ; ; …; n À , we compute the probability mass function of x iþ given fx j ¼ x j g i j¼ , namely, q iþ ðk; q ! fx j g i j¼ Þ. then, the likelihood function is for simplicity, assume that x iþ is poisson distribution with mean m iþ ¼ x i e lðtiþ ÀtiÞ . note that, since we assumed no observation error, the initial condition c ¼ x is exact, and thus there is a single parameter l for the model. thus, and thus the log-likelihood function is x iÀ e lðt i Àt iÀ Þ þ x i lðt i À t iÀ Þ þ x i lnx i À lnx i !: again, the last two terms can be ignored in maximization because they are constants. thus, l ¼ argmax l x iÀ e lðt i Àt iÀ Þ þ ðt i À t iÀ Þx i l: it is much harder to formulate the likelihood function if process errors and observation errors must both be considered. we can simplify the problem by ignoring the process error and use an overly dispersed observation error distribution as a compensation. note that this simplification mainly affects the confidence intervals. the maximum likelihood method gives a point estimate, i.e., one set of parameter values that makes it mostly likely to observe the data. however, it is not clear how close the point estimates are to the real values. to answer this question we use an interval estimate, commonly known as a confidence interval. a confidence interval with a confidence level a is an interval that has a probability a that contains the true parameter value. a commonly used confidence level is %, which originates from a normal distribution. if a random variable x is normally distributed with a mean m and a standard deviation s, then the probability that x ½m À s; m þ s is %. the confidence interval can be estimated using the likelihood ratio test [ (bolker, ), p. ] . let c q !^b e the point estimate of the parameters. a value l is in the % confidence interval is equivalent to accepting with % probability that l is a possible growth rate. to determine this we fit a nested model by fixing the growth rate l ¼ l , suppose its point estimate is b q . we then compute the likelihood ratio the wilks' theorem (wilks, ) guarantees that, as the sample size becomes large, the statistics À lnl ¼ ½[ð b qÞ À[ð b q Þ is c distributed with a degree of freedom . we thus can compare À lnl with the % quantile of the c distribution and determine if l should be in the confidence interval or not. we can thus perform a linear search on both sides of the point estimate to determine the boundary of the confidence interval. we still have not addressed the problem of choosing a fitting window for an exponential model. recall that the challenge arises because the exponential growth rate of an epidemic decreases with time. instead of finding heuristic conditions for choosing the fitting window, we circumvent this problem by incorporating the decrease of the exponential growth rate into our model. we have two choices, using either a mechanistic model such as an sir or seir model, or a phenomenological model. naturally, if we know that a mechanistic model is a good description of the disease dynamics, fitting such a model to the epidemic curve is a good option (see, e.g., (chowell, ammon, hengartner, & hyman, ; pourabbas, d'onofrio, & rafanelli, ) ,). we use an sir model as an example. for simplicity, we assume that the process error is negligible, and the incidence rate is poisson distributed with a mean cðtÞ given by an sir model (cðtÞ ¼ bsin where n is the population size). to construct the log-likelihood function, we need to calculate cðtÞ, i.e., numerically solve the sir model. to do so, we need the transmission rate b. the recovery rate g, the initial fraction of infectious individuals ið Þ ¼ i (with the assumption that rð Þ ¼ , sð Þ ¼ À i , and thus i determines the initial conditions), in addition to the population size n. thus, the parameters of the model is q ! ¼ ðb; g; i ; nÞ. thus the log-likelihood function is (ignoring the constant terms) where the number of new cases cðt i Þ in the time interval ½t i ; t iþ is cðt i Þ ¼ sðt iþ Þ À sðt i Þ ; and sðt i Þ is solved numerically from the sir model. thus, [ implicitly depend on b, g and i through sðtÞ. one draw back using such a mechanistic model is its high computational cost, since each evaluation of the log-likelihood function requires solving the model numerically, and numerical optimization algorithms can be very hungry on function evaluations, especially if the algorithm depends on numerical differentiation. another draw back is that these mechanistic models can be overly simplified, and may not be a good approximation to the real disease dynamics. for example, for seasonal influenza, due to the fast evolution of the influenza virus, individuals have different history of infection, and thus have different susceptibility to a new strain. yet simple sir and seir models assume a population with a homogeneous susceptibility. thus using a simple sir to fit to an influenza epidemic may be an over simplification. however, realistic mechanistic models can be overly complicated, and involve too many parameters that are at best difficult to estimate. for example, a multi-group sir model depends on a contact matrix consisting of transmission rates between groups, which contains a large number of parameters if the model uses many groups. if all we need to estimate is the exponential growth rate, we only need a model that describes the exponential growth that gradually slows down. most cumulative epidemic curves grow exponentially initially, and then saturates at the final epidemic size. a simple phenomenological model can be used to describe the shape of the cumulative epidemic curve, but the model itself may not have realistic biological meaning. however, if simple mechanistic models cannot faithfully describe the epidemic process, using a simple phenomenological model with an analytical formula may be a better choice, at least numerically, because repetitively solving a system differential equations numerically, and differentiating the log-likelihood function numerically, can both be avoided with the analytical formula. here we discuss some examples for such models. the logistic model is the simplest model that shows an initial exponential growth followed a gradual slowing down and a saturation. the cumulative incidences cðtÞ (the total number of cases by time t) can be approximated by d dt cðtÞ ¼ rcðtÞ À cðtÞ k : where r is the exponential growth rate, and k ¼ lim t/∞ cðtÞ. let c ¼ cð Þ, its solution is the new cases cðt i Þ in a time period ½t i ; t iþ is thus the model parameters are q ! ¼ ðr; k; c Þ. note that it is less than the number of parameters of the simplest mechanistic model (i.e., the sir model). the logistic model has a fixed rate of slowing down of the exponential growth rate. to be more flexible, we can use the richards model (richards, ) for the cumulative incidence curve. the richards model, also called the power law logistic model, can be written as d dt cðtÞ where ais the parameter that controls the steepness of the curve. note that the logistic model is a special case with a ¼ . its solution is the new cases cðt i Þ in a time period ½t i ; t iþ is also given by ( ). the parameters are q ! ¼ ðr; k; c ; aÞ. to compare the performance of both the sir model and the phenomenological models, we fit these models to the stochastically simulated seir epidemic curve of weekly cases that we introduced in section (fig. ) . we assume that the process error is negligible, and the observations are poisson distributed about the mean that is given by the corresponding models. we use the maximum likelihood method. the results are shown in fig. . the predictions of the exponential model, as discussed before, quickly decreases as more data points are used. both the logistic model and the richards model give robust estimates with fitting windows ending up to the peak of the epidemic. the sir model gives a robust estimate for all fitting windows up to the whole epidemic curve. thus, the sir model is a good model to use to fit the exponential growth rate, even if it may not be the correct mechanistic model. (e.g., it ignores the latent period in this example). it requires more computational power, because the epidemic curve lacks an analytic formula, and needs to be numerically solved from a system of ordinary differential equations. the logistic model and the richards model can be used for all data points up to the peak of the epidemic. fig. also show that the sir model and the logistic model give the narrowest confidence intervals. however, narrower confidence intervals may not be desirable if it has a large chance that it does not contain the true value. due to errors, especially process errors, each realization of the underlying stochastic epidemic process yields a different epidemic curve. these epidemic curves may exhibit different exponential growth rates even if the underlying parameter values are the same. an observed epidemic curve is just a single realization of the epidemic process. does the estimated confidence intervals contain the theoretical exponential growth rate of the epidemic process? this question is answered by the "coverage probability", which is the probability that the confidence interval contains the true value. if the confidence interval properly considers all sources of stochasticity, then the coverage probability should be equal to its confidence level. to illustrate this, we numerically compute the coverage of the confidence intervals by simulating the seir model times and compute confident interval of the exponential growth rate for each realization, and compute the fraction of the confident intervals containing the theoretical value l ¼ : . the results is summarized in below: logistic model richards model coverage probability % % that is, even though the logistic model gives a narrow confidence interval, its coverage probability is low. the coverage probability of the confidence interval given by the richards model is also significantly lower than the confidence level. this is indeed caused by treating process errors as observation errors. if there is under reporting, that is, only a fraction p of the cases can be observed, then the observation error becomes larger as p decreases (i.e., more under reporting). the coverage will become larger as a result. for example, the case fatality ratio of the pandemic influenza is about % (frost, ) . thus, the mortality curve can be treated as the epidemic curve with a large under reporting ratio, and thus the observation error dominates. in this case ignoring the process error is appropriate. ecological models and data in r transmission dynamics of the great influenza pandemic of in geneva, switzerland: assessing the effects of hypothetical interventions statistics of influenza morbidity. with special reference to certain factors in case incidence and case-fatality a general method for numerically simulating the stochastic time evolution of coupled chemical reactions a method to estimate the incidence of communicable diseases under seasonal fluctuations with application to cholera a flexible growth function for empirical use how generation intervals shape the relationship between growth rates and reproductive numbers different epidemic curves for severe acute respiratory syndrome reveal similar impacts of control measures the comparison of the results of fitting the sir, exponential, logistic, and richards models to a simulated weekly incidence curve, as a function of the end point of the fitting window (upper). the epidemic curve (lower) is shown as a reference. the epidemic curve and the theoretical exponential this research is partially supported by a natural sciences and engineering research council canada discovery grant, and national natural science foundation of china (no. ). key: cord- -p mb r v authors: luo, yan; chalkou, konstantina; yamada, ryo; funada, satoshi; salanti, georgia; furukawa, toshi a. title: predicting the treatment response of certolizumab for individual adult patients with rheumatoid arthritis: protocol for an individual participant data meta-analysis date: - - journal: syst rev doi: . /s - - -x sha: doc_id: cord_uid: p mb r v background: a model that can predict treatment response for a patient with specific baseline characteristics would help decision-making in personalized medicine. the aim of the study is to develop such a model in the treatment of rheumatoid arthritis (ra) patients who receive certolizumab (ctz) plus methotrexate (mtx) therapy, using individual participant data meta-analysis (ipd-ma). methods: we will search cochrane central, pubmed, and scopus as well as clinical trial registries, drug regulatory agency reports, and the pharmaceutical company websites from their inception onwards to obtain randomized controlled trials (rcts) investigating ctz plus mtx compared with mtx alone in treating ra. we will request the individual-level data of these trials from an independent platform (http://vivli.org). the primary outcome is efficacy defined as achieving either remission (based on acr-eular boolean or index-based remission definition) or low disease activity (based on either of the validated composite disease activity measures). the secondary outcomes include acr ( % improvement based on acr core set variables) and adverse events. we will use a two-stage approach to develop the prediction model. first, we will construct a risk model for the outcomes via logistic regression to estimate the baseline risk scores. we will include baseline demographic, clinical, and biochemical features as covariates for this model. next, we will develop a meta-regression model for treatment effects, in which the stage risk score will be used both as a prognostic factor and as an effect modifier. we will calculate the probability of having the outcome for a new patient based on the model, which will allow estimation of the absolute and relative treatment effect. we will use r for our analyses, except for the second stage which will be performed in a bayesian setting using r jags. discussion: this is a study protocol for developing a model to predict treatment response for ra patients receiving ctz plus mtx in comparison with mtx alone, using a two-stage approach based on ipd-ma. the study will use a new modeling approach, which aims at retaining the statistical power. the model may help clinicians individualize treatment for particular patients. systematic review registration: prospero registration number pending (id# ). (continued from previous page) systematic review registration: prospero registration number pending (id# ). keywords: rheumatoid arthritis, certolizumab, individual participant data meta-analysis, prediction model, treatment response background rheumatoid arthritis (ra) is a chronic inflammatory disease, for which we cannot currently expect complete cure. the drugs that can delay disease progression are known as disease-modifying anti-rheumatic drugs (dmards). there are categories: conventional synthetic dmards (csdmards), biologic dmards (bdmards), and targeted synthetic dmards (tsdmards). bdmards can be further divided into several subtypes according to the target, among which the tumor necrosis factor (tnf) α inhibitors are the most classic and widely used. most ra patients undergo long-term treatment. according to the treat-to-target strategy proposed by the eular (european league against rheumatism) practice guideline [ ] , repeated assessment of disease activity should be performed every to months after a treatment is given, to evaluate the response and decide the next-step strategy: switching drugs, maintenance, tapering, or discontinuation. hence, the disease course of ra is composed of many short-term ( to months) intervention-response loops. for the purpose of improving long-term prognosis, such as delaying the progression of bone fusion or functional deficiency, short-term intervention-response loops need to have beneficial outcomes [ ] . to find the optimal treatment for a particular patient, it is necessary to personalize the treatment. it would be helpful if we could predict the probability of treatment response based on the patient's genetic, biologic, and clinical features. however, common evidence in the form of randomized controlled trials (rcts) or their meta-analyses (mas) at the aggregate level only reports average results. the drug that works for the average patients might not work or even be harmful for a particular patient. consequently, it is desirable to identify subgroups of patients associated with different treatment effects. individual participant data meta-analysis (ipd-ma) has been previously employed to develop prediction models for treatment effects [ ] [ ] [ ] [ ] . previous treatment response prediction models for ra were mainly based on observational studies [ ] [ ] [ ] [ ] [ ] . observational studies seem suited for predicting the absolute risk of an outcome, but it may be less satisfactory in estimating the relative risk between different drugs, because unknown confounders may persist even when we try to adjust for known confounders. on the other hand, though the population in rcts is highly restricted hence may be less representative, data from rcts are more rigorously collected and more likely to provide an unbiased estimate of the relative treatment effects [ ] . the synthesis of rct data via ipd-ma can increase the statistical power [ ] and have been used to predict treatment response [ , [ ] [ ] [ ] [ ] . to the best of the authors' knowledge, such an approach has not been taken to predict treatment response in ra to date. our aim is to develop a prediction model of treatment effects based on individual characteristics of ra patients through ipd-ma. since tnfα inhibitors are the most classic and widely used bdmards for ra, we will build a model for certolizumab (ctz), a tnfα inhibitor with sufficient ipd data, in this study. we will first estimate the pooled average effect sizes for the primary and secondary outcomes using one-stage bayesian hierarchical ipd-ma. the main objective of the study is to use a two-stage risk modeling approach to predict the individualized treatment effects interest [ ] . the first stage is to build a multivariable model aiming to predict the baseline risk for a particular patient blinded to treatment. in the second stage, this baseline risk score will be used as a prognostic factor and an effect modifier in an ipd meta-regression model to estimate the individualized treatment effects of ctz. we consider to validate and optimize the modeling approach in the present study, and plan eventually to expand it to an ipd network meta-analysis to compare several drug types (e.g., interleukin- inhibitors, anti-cd antibodies) as our future research perspective. the present protocol has been registered within the prospero database (provisional registration number id# ) and is being reported in accordance with the reporting guidance provided in the preferred reporting items for systematic reviews and meta-analyses protocols (prisma-p) statement [ ] (see the checklist in additional file ). the proposed ipd-ma will be reported in accordance with the reporting guidance provided in the preferred reporting items for systematic reviews and meta-analyses of individual participant data (prisma-ipd) statement [ ] . any amendments made to this protocol when conducting the study will be outlined and reported in the final manuscript. studies will be selected according to the following criteria: patients, interventions, outcomes, and study design. we will include adults ( years or older) who are diagnosed with either early ra ( american college of rheumatology (acr)/european league against rheumatism (eular) classification criteria) [ , ] or established ra ( classification criteria) [ ] . patients with inner organ involvement (such as interstitial lung diseases), vasculitis, or concomitant other systemic autoimmune diseases will be excluded. we will include both treatment-naïve patients and patients who have insufficient response to previous treatments. we will include patients with moderate to severe disease activity based on any validated composite disease activity measures. patients who have already achieved remission or at low disease activity at baseline will be excluded. patients who have used certolizumab (ctz) within months before randomization will be excluded. we will include rcts which compare certolizumab (ctz) plus methotrexate (mtx) with mtx monotherapy, regardless of doses. if a study compares ctz + any csdmards with any csdmards, we will only include patients on ctz + mtx or mtx from that study. trials investigating the tapering or discontinuation strategy of ctz will be excluded. our primary outcome is efficacy defined by disease states, which is achieving either remission (based on acr-eular boolean or index-based remission definition [ ] ) or low disease activity (based on either of the validated composite disease activity measures [ ] : das (disease activity score based on the evaluation of joints) ≤ . [ ] , cdai (clinical disease activity index) ≤ [ ] , sdai (simplified disease activity index) ≤ [ ] ) at months (allowance - months) after treatment, as a binary outcome. we choose it as our primary outcome because it is suggested as the indicator of the treatment target in both the practice guideline [ ] and the guideline for conducting clinical trials in ra [ ] , and it has been shown to provide more information for future joint damage and functional outcomes compared to relative response (change from baseline) [ ] . we have two secondary outcomes. one is efficacy defined by response (improvement from baseline), for which we will use the acr response criteria acr ( % improvement based on acr core set variables) [ ] . another is adverse events (aes). we will perform an ipd-ma separately for patients with all kinds of infectious aes within months since it is one of the most important aes for biologic agents. we will also describe other noticeable aes within months reported in the trials. we will not make predictions models for the secondary outcomes. we will include double-blind rcts only. if there are crossover rcts, only the data of the first phase will be used for analysis. cluster rcts, quasi-randomized trials, and observational studies will be excluded. we will conduct an electronic search of cochrane cen-tral, pubmed, and scopus from inception onwards, with the keywords: "rheumatoid arthritis," "certolizumab" or "cdp " "cimzia", "methotrexate" or "mtx," without language restrictions. a draft search strategy is provided in additional file . we will search who international clinical trials registry platform to find the registered studies. we will search the us food and drug administration (fda) reports to see if there are any unpublished reports from the pharmaceutical companies. for ipd, we will contact the company which markets certolizumab and request ipd through http://vivli.org. we will assess the representativeness of the ipd among all the eligible studies by investigating the potential differences between trials with ipd and those without ipd. two independent reviewers will screen the titles and abstracts retrieved from the electronic searches to assess for inclusion. if both reviewers agree that a trial does not meet eligibility criteria, it will be excluded. the full text of all the remaining articles will be obtained for further reading, and the same eligibility criteria will be applied to determine which to exclude. any disagreements will be resolved through discussion with a third member of the review team. two reviewers will independently extract the information for all the included studies at aggregate level. a detailed data extraction template will be developed and piloted on articles; after finalizing the items on the data extraction form, the articles will be re-extracted. the main information includes intervention/control details, trial implementation features (e.g., completion year, randomized numbers, dropouts, follow-up length), baseline demographic and disease-specific characteristics, and outcomes of interested. the above information will be used for: ( ) exploring the representativeness of the trials with ipd among all the eligible trials and ( ) confirming if the ipd is consistent with the reported results. when the ipd is ready to be used, we will identify the variables of interest before the analysis. the variables regarding intervention, control, and outcomes are defined as the above in the "eligibility criteria" section. with regard to patient or trial characteristics to be used as potential covariates in the prognostic model, based on the literature [ ] [ ] [ ] and our clinical practice, we propose the following factors as candidates of potential prognostic factors (pfs, baseline factors that may affect the response regardless of the treatment) (table ) , which will be used for baseline risk model development (see the "predicting treatment effect for patients with particular characteristics: a two-stage model" section below). we will try to collect all the information listed in table from the data, but only available factors that have been recorded in the trials will be added into the model. we will decide in which type (e.g., continuous, categorical, binary, etc.) a covariate will be put into the model according to the distribution of that covariate after we obtain the data. two independent reviewers will assess the risk of bias (rob) for each included rct according to "rob tool" proposed by the cochrane group [ ] . for the efficacy primary outcome, rcts will be graded as "low risk of bias," "high risk of bias," or "some concerns" in the following five domains: risk of bias arising from the randomization process, risk of bias due to deviations from the intended interventions, missing outcome data, risk of bias in measurement of the outcome, and risk of bias in selection of the reported result. the assessment will be adapted for ipd-ma, i.e., as per the obtained data and not the conducted and reported analyses in the original publications. finally, they will be summarized as an overall risk of bias according to the rob algorithm. since our primary aim is to develop a prediction model and not to get a precise estimation of the treatment effects, all the analyses will be based on ipd only. therefore, we will neither analyze aggregate data together nor investigate the robustness of the ipd-ma including aggregate data, for they are beyond the perspectives of the present study. we first synthesize the data using one-stage bayesian hierarchical ipd-ma [ ] . we will estimate the average relative treatment effect in terms of odds ratio (or) for efficacy. let y ij denote the dichotomous outcome of interest (y ij = for remission or low disease activity), for patient i where i = , , …, n j in trial j out of n trials, t ij be / for patient in control/intervention group, and p ij is the probability of having the outcome. where α j is the log odds of the outcome for the control group, in trial j, which is independent across trials; δ j is table potential candidates to be involved as prognostic factors in the prognostic model *factors that have been proved to be a prognostic factor for any treatments in previous studies # since genetic tests for ra are not routinely implemented in clinical practice, we anticipate that most studies will not report them. although genetics are often considered critical in precision medicine, we will consider it justifiable if no genetic information is included in our model, because there is no single one that has been proven to be strongly associated with the prognosis or treatment responses, and two studies have indicated that genetic information barely contribute in predicting treatment effects [ ] the treatment effect (log or), which we assume to be exchangeable across trials; δ is the summary estimate of the log-odds ratios for the intervention versus the control arm; and τ is the heterogeneity of δ across trials and normally distributed across trials. predicting treatment effect for patients with particular characteristics: a two-stage model data pre-processing within each study, the outcomes and the covariates will be evaluated for missing data, and we will further look at their distributional characteristics and correlations between the covariates (listed in the "at ipd level: for studies with ipd available" section). we will use multiple imputation methods for handling missing data [ ] . we will consider data transformation for continuous variables to resolve skewness and re-categorization for categorical variables if necessary. if two or more variables are highly correlated, we will only retain the variable that is most commonly reported across studies and in the literature or the variable that has the least missing values. stage : developing a baseline risk model in this step, we will build a multivariable model to predict the probability that a patient, given her or his baseline characteristics, is likely to achieve remission or low disease activity irrespective of treatment; we will refer to this model as the baseline risk model. the risk model can be built using the patients from the control group only, or from both intervention and control group. the former is more intuitive; however, a simulation study indicated that models based on the whole participants produced estimates with narrower distribution of bias and were less prone to overfitting [ ] . we will fit a multivariable logistic regression model: r ij is the probability of the outcome for patient i from trial j at the baseline. b j is the intercept, which is exchangeable across studies. pf ijk denotes the k prognostic factor (in total, there are p prognostic factors) in study j for patient i, and b kj is the regression coefficient for k prognostic factor in study j and is exchangeable across studies. in order to select the most appropriate model, we propose two approaches: ( ) use previously identified prognostic factors and through discussions with rheumatologists to decide the subset of the most clinically relevant factors and estimate the coefficients using penalized maximum likelihood estimation shrinkage method and ( ) use lasso penalization methods for variable selection and coefficient shrinkage [ ] . for each possible model, we will examine the sample size first, in order to assess the reliability of the model. we will calculate the events per variable (epv) for our model, using all the categories of categorical variables and the degrees of freedom of continuous outcomes [ ] . we will calculate efficient sample size for developing a logistic regression model [ ] . validation is essential in prediction model development. since no external data is available, we can only use internal validation. via resampling methods like bootstrap or cross-validation, we can estimate the calibration slope and c-statistic for each model, to indicate the ability of calibration and discrimination. stage : developing a meta-regression model for treatment effects we use the same notation system as that in the "average relative treatment effect: ipd-ma" section. the logit(r ij ) from stage will be used as a covariate in the meta-regression model, both as a prognostic factor and as an effect modifier. let logitðr ij Þ j denote the average of logit-risk for all the individuals in study j. the regression equation will be: γ a j is the log odds in the control group when a patient has a risk equal to the mean risk, which is assumed to be independent across trials. g j is the coefficient of the risk score, while g j is the treatment effect modification of the risk score for the intervention group versus the control group; both are assumed to be exchangeable cross trials and normally distributed about a summary estimate γ and γ respectively. predicting the probability of having the outcome for a new patient assume a new patient i who is not from any trial j has a baseline risk score g logitðr i Þ calculated from stage-one. in order to predict the absolute logitprobability to achieve the outcome, we use: we would have estimated δ, γ , and γ in the metaregression stage. we will estimate logitðrÞ as the mean of logit(r ij ) across all the individuals and studies. for a, we will estimate it by synthesizing all the control arms. then, we can calculate the individual probability of the outcome both for the control and the intervention and estimate the predicted absolute and relative treatment effect. to evaluate the performance of the two-stage prediction model, we will use internally validation methods via both the traditional measures, like c-statistic, and measures relevant to clinical usefulness. publication bias considering that we will probably not be able to include all the relevant research works, as some studies or their results were likely not published owing to non-significant results (study publication bias and outcome reporting bias) [ , ] , we will evaluate this issue by comparing the search and screening results (as we will try to retrieve possibly unpublished reports) with the ipd we can get. if necessary, we will address it by adding the study's variance as an extra covariate in the final ipd meta-regression model (see the section "predicting treatment effect for patients with particular characteristics: a two-stage model"-"stage : developing a meta-regression model for treatment effects"). statistical software we will use r for our analyses. stage will be performed in a bayesian setting using r jags. for the development of the baseline risk model, we will use the pmsampsize command to estimate if the available sample size is enough. we will examine the linear relationship between each one of the prognostic factors and the outcome via rcs and anova commands. the lasso model will be developed using the cv.glmnet command. we will use the lrm command for the predefined model based on prior knowledge, and then for the penalized maximum likelihood estimation, we will use the pentrace command. for the bootstrap internal validation (both for the baseline risk score and for the two-stage prediction model), we will use self-programmed r-routines. we have presented the study protocol for a prediction model of treatment effects for ra patients receiving ctz plus mtx, using a two-stage approach based on ipd-ma. though there are many optional drugs in treating ra, as treatment failure is relatively high, individualizing the treatment is imperative. many prognostic models for ra have been proposed, but no one is sufficiently satisfactory [ ] . we have discovered several problems. most previous models focused on long-term radiographic or functional prognosis. although they are certainly the critical outcomes that both clinicians and patients care about, the complex therapeutic changes during the long treatment process are extremely difficult to handle in developing prediction models. thus, it usually ends up with a simplified strategy, such as taking only the initial treatment into account, which compromises the clinical interpretation and relevancy of the model. on the other hand, a good short-term treatment response is always positively associated with good longterm prognosis [ , ] . predicting short-term treatment effect is instructive in clinical practice; however, research is lacking. a few established "short-term" diseaseactivity-oriented prediction models used an outcome measured at months or months. the problem is, unless in active-controlled studies, there would be considerable dropouts after - months; furthermore, due to ethnical issues, many trials would offer the nonresponders other active treatments after - months. under the itt principle, patients were commonly analyzed as originally allocated; when dropouts were not negligible, imputation methods were usually used, but mostly single imputation such as non-responder imputation or last observation carried forward (locf) [ ] . one may argue that these estimates were conservative to the intervention group though not precise. but in fact it is not always conservative for a relative effect estimate, while unbiased relative estimates are of critical interest in building personalized prediction models. as a result, in order to be methodologically rigorous, we choose the outcome measured at months, when the randomization is likely kept, and which is consistent with the assessment time recommended by the guideline [ ] . additionally, thanks to the ipd, we will be able to use multiple imputation to handle missing data, rather than the single imputations used in primary rcts. we will use a two-stage approach to construct the prediction model using ipd-ma. unlike the usual approach, which includes baseline features as prognostic factors and effect modifiers (through interaction terms) simultaneously, we first build a risk model for baseline factors, then treating the risk score as both a prognostic factor and an effect modifier. by doing so, overfitting problem caused by too many covariates and interaction terms can be alleviated. moreover, since penalization will only be used in common regression during risk modeling stage but not in meta-regression, the compromised penalization in meta-regression can be avoided. for stage , generally there are two types of risk models. one is an externally developed model, which is derived based on data independent from the data used at stage , such as established models from previous studies, or using some other studies. the other is an internally developed risk model, for which the same data will be used to build both the risk model and the treatment effects model [ ] . because there is no well-established risk model to predict the short-term disease activity for ra patients and also because we will very probably not have sufficient sample size to divide the entire data into two parts, we will use the internal risk model for our study. we acknowledge several limitations in our study. first, we handle effect modification at the level of risk scores, instead of particular covariates. that is, we will not try to identify specific effect modifiers. this may cause some problems in interpretation, as the concept of distinguishing prognostic factors and effect modifiers is well recognized. however, our approach assures the statistical power. second, due to the restricted sample size, only internal validation is planned while external validation is lacking. it needs to be validated on an external dataset in the future. third, we only focus on short-term treatment response for ra patients receiving two kinds of treatment, ctz and mtx. future studies may extend the scope to compare several kinds of therapies and treatment strategies and finally model for the long-term prognosis taking into consideration all the treatment processes. supplementary information accompanies this paper at https://doi.org/ . /s - - -x. additional file . prisma-p checklist. eular recommendations for the management of rheumatoid arthritis with synthetic and biological disease-modifying antirheumatic drugs: update committee for medicinal products for human use (chmp): guideline on clinical investigation of medicinal products for the treatment of rheumatoid arthritis a framework for developing, implementing, and evaluating clinical prediction models in an individual participant data meta-analysis developing and validating risk prediction models in an individual participant data meta-analysis cochrane ipdm-amg: individual participant data (ipd) meta-analyses of diagnostic and prognostic modeling studies: guidance on their use statistical approaches to identify subgroups in meta-analysis of individual participant data: a simulation study arthritis: which subgroup of patients with rheumatoid arthritis benefits from switching to rituximab versus alternative anti-tumour necrosis factor (tnf) agents after previous failure of an anti-tnf agent? prediction of response to methotrexate in rheumatoid arthritis prediction of response to targeted treatment in rheumatoid arthritis association of response to tnf inhibitors in rheumatoid arthritis with quantitative trait loci for cd and cd assessment of a deep learning model based on electronic health record data to forecast clinical outcomes in patients with rheumatoid arthritis personalized evidence based medicine: predictive approaches to heterogeneous treatment effects meta-analysis of individual participant data: rationale, conduct, and reporting getreal methods review g: get real in individual participant data (ipd) meta-analysis: a review of the methodology quantifying heterogeneity in individual participant data meta-analysis with binary outcomes a critical review of methods for the assessment of patient-level interactions in individual participant data meta-analysis of randomized trials, and guidance for practitioners cognitive-behavioral analysis system of psychotherapy, drug, or their combination for persistent depressive disorder: personalizing the treatment choice using individual participant data network metaregression preferred reporting items for systematic review and meta-analysis protocols (prisma-p) preferred reporting items for systematic review and meta-analyses of individual participant data: the prisma-ipd statement rheumatoid arthritis classification criteria: an american college of rheumatology/ european league against rheumatism collaborative initiative rheumatoid arthritis classification criteria: an american college of rheumatology/ european league against rheumatism collaborative initiative the american rheumatism association revised criteria for the classification of rheumatoid arthritis american college of rheumatology/european league against rheumatism provisional definition of remission in rheumatoid arthritis for clinical trials the definition and measurement of disease modification in inflammatory rheumatic diseases the disease activity score and the eular response criteria acute phase reactants add little to composite disease activity indices for rheumatoid arthritis: validation of a clinical activity score remission and active disease in rheumatoid arthritis: defining criteria for disease activity states the importance of reporting disease activity states in rheumatoid arthritis clinical trials american college of rheumatology. preliminary definition of improvement in rheumatoid arthritis remission-induction therapies for early rheumatoid arthritis: evidence to date and clinical implications assessing prognosis and prediction of treatment response in early rheumatoid arthritis: systematic reviews poor prognostic factors guiding treatment decisions in rheumatoid arthritis patients: a review of data from randomized clinical trials and cohort studies crowdsourced assessment of common genetic contribution to predicting anti-tnf treatment response in rheumatoid arthritis rob : a revised tool for assessing risk of bias in randomised trials meta-analysis using individual participant data: one-stage and two-stage approaches, and why they may differ missing data in randomised controlled trials: a practical guide using internally developed risk models to assess heterogeneity in treatment effects in clinical trials regression shrinkage and selection via the lasso risk prediction models: i. development, internal validation, and assessing the incremental value of a new (bio)marker minimum sample size for developing a multivariable prediction model: part ii -binary and time-to-event outcomes randomized controlled trials of rheumatoid arthritis registered at clinicaltrials.gov: what gets published and when dissemination and publication of research findings: an updated review of related biases rheumatoid arthritis treatment: the earlier the better to prevent joint damage evaluation of different methods used to assess disease activity in rheumatoid arthritis: analyses of abatacept clinical trial data a systematic review of randomised controlled trials in rheumatoid arthritis: the reporting and handling of missing data in composite outcomes publisher's note springer nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations authors' contributions yl and taf conceived the study. kc and gs designed the modeling strategy. ry and sf provided substantial contribution to the design of the study during its development. yl drafted the manuscript, and all the authors critically revised it. all authors gave final approval of the version to be published. this study was supported by the intramural support to the department of health promotion and human behavior, kyoto university graduate school of medicine/school of public health. the funder has no role in the study design, data collection, data analysis, data interpretation, writing of the report, or in the decision to submit for publication. the data that support the findings of this study are available from http://vivli. org but restrictions apply to the availability of these data, which were used under license for the current study, and so are not publicly available. data are however available from http://vivli.org upon reasonable request and application, after their permission. this study does not require institutional review board approval and participant consent.competing interests taf reports personal fees from mitsubishi-tanabe, msd, and shionogi and a grant from mitsubishi-tanabe, outside the submitted work; taf has a patent - . gs was invited to participate in two methodological meetings about the use of real-world data, organized by biogen and by merck. all the other authors report no competing interests to declare. key: cord- - qhgeirb authors: busby, j s; onggo, s title: managing the social amplification of risk: a simulation of interacting actors date: - - journal: j oper res soc doi: . /jors. . sha: doc_id: cord_uid: qhgeirb a central problem in managing risk is dealing with social processes that either exaggerate or understate it. a longstanding approach to understanding such processes has been the social amplification of risk framework. but this implies that some true level of risk becomes distorted in social actors’ perceptions. many risk events are characterised by such uncertainties, disagreements and changes in scientific knowledge that it becomes unreasonable to speak of a true level of risk. the most we can often say in such cases is that different groups believe each other to be either amplifying or attenuating a risk. this inherent subjectivity raises the question as to whether risk managers can expect any particular kinds of outcome to emerge. this question is the basis for a case study of zoonotic disease outbreaks using systems dynamics as a modelling medium. the model shows that processes suggested in the social amplification of risk framework produce polarised risk responses among different actors, but that the subjectivity magnifies this polarisation considerably. as this subjectivity takes more complex forms it leaves problematic residues at the end of a disease outbreak, such as an indefinite drop in economic activity and an indefinite increase in anxiety. recent events such as the outbreaks in the uk of highly pathogenic avian influenza illustrate the increasing importance of managing not just the physical development of a hazard but also the social response. the management of hazard becomes the management of 'issues', where public anxiety is regarded less as a peripheral nuisance and more as a legitimate and consequential element of the problem (leiss, ) . it therefore becomes as important to model the public perception of risk as it does to model the physical hazard-to understand the spread of concern as much as the spread of a disease, for example. in many cases the perception of risk becomes intimately combined with the physical development of a risk, as beliefs about what is risky behaviour come to influence levels of that behaviour and thereby levels of exposure. one of the main theoretical tools we have had to explain and predict public risk perception is the social amplification of risk framework due to kasperson et al ( ) . as we explain below, this framework claims that social processes often combine to either exaggerate or underplay the risk events experienced by a society. this results in unreasonable and disproportionate reactions to risks, not only among the lay public but also among legislators and others responsible for managing risk. but since its inception the idea of a 'real', objective process of social risk amplification has been questioned (rayner, ; rip, ) and, although work in risk studies and risk management continues to use the concept, it has remained problematic. the question is whether, if we lose the notion of some true risk being distorted by a social process, we lose all ability to anticipate and explain perplexing social responses to a risk event in a way that is informative to policymakers. we explore this question in the context of risks surrounding the outbreaks of zoonotic diseases-that is, diseases that cross the species barrier to humans from other animals. recent cases of zoonotic disease, such as bse, sars, west nile virus and highly pathogenic avian influenza (hpai), have been some of the most highly publicised and controversial risk issues encountered in recent times. many human diseases are zoonotic in origin but in cases such as bse and hpai the disease reservoirs remain in the animal population. this means that a public health risk is bound up with risk to animal welfare, and often risk to the agricultural economy, to food supply chains and to wildlife. this in turn produces difficult problems for risk managers and policymakers, who typically want to avoid a general public amplifying the risk and boycotting an industry and its products, but also want to avoid an industry underestimating a risk and failing to practice adequate biosecurity. the bse case in particular has been associated with ideas about risk amplification (eg, eldridge and reilly, ) and continues to appear in the literature (lewis and tyshenko, ) . other zoonoses, such as chronic wasting disease in deer herds, have also been seen as recent objects of risk amplification (heberlein and stedman, ) . in terms of the social reaction, not all zoonoses are alike. endemic zoonoses like e. coli do periodically receive public attention-for example following outbreaks at open farms and in food supply chains. but it is the more exotic zoonoses like bse and hpai that are more clearly associated with undue anxiety and ideas about social risk amplification. yet these cases also showed how uncertain the best, expertly assessed, supposedly objective risk level can be, and this makes it very problematic to retain the idea of an objective process of social risk amplification. such cases are therefore an important and promising setting for exploring the idea that amplification is only in the heads of social actors, and for exploring the notion that this might nonetheless produce observable, and potentially highly consequential, outcomes in a way that risk managers need to understand. our study involved two main elements, the second of which is the main subject of this article: . exploratory fieldwork to examine how various groups perceived risks and risk amplification in connection with zoonoses like the avian influenza outbreaks in ; . a systems dynamics simulation to work out what outcomes would emerge in a system of social actors who attributed amplification to other actors. in the remainder of the paper we first outline the fieldwork and its outcomes, and then describe the model and simulation. although the article concentrates on the latter, the two parts provide complementary elements of a process of theorising (kopainsky and luna-reyes, ) : the fieldwork, subjected to grounded analysis, produces a small number of propositions that are built into the systems dynamics model, and the model both operationalises these propositions and explores their consequences when operationalised in this way. the modelling is a basis for developing theory that is relevant to policy and decision making, rather than supporting a specific decision directly. a discussion and conclusion follow. traditionally, the most problematic aspect of public risk perception has been seen as its sometimes dramatic divergence from expert assessments-and the way in which this divergence has been seen as an obstacle both to managing risks specifically and to introducing new technology more generally. this has produced a longstanding interest in the individual perception of risk (eg, slovic, ) and in the way that culture selects particular risks for our attention (eg, douglas and wildavsky, ) . it has led to a strong interest in risk communication (eg, otway and wynne, ) . and it has been a central theme in the social amplification of risk framework (or sarf) that emerged in the late s (kasperson et al, ) . the notion behind social risk amplification, developed in a series of articles (kasperson et al, ; renn, ; burns et al, ; kasperson and kasperson, ) , is that a risk event produces signals that are processed and sometimes amplified by a succession of social actors behaving as communication 'stations'. they interact and observe each other's responses, sometimes producing considerable amplification of the original signal. a consequence is that there are often several secondary effects, such as product boycotts or losses of institutional trust, that compound the effect of the original risk event. a substantial amount of empirical work has been conducted on or around the idea of social amplification, for example showing that the largest influence on amplification is typically organisational misconduct (freudenberg, ) . it continues to be an important topic in the risk literature, not least in connection with zoonosis risks (eg, heberlein and stedman, ; lewis and tyshenko, ). there has always been a substantial critique of the basic idea of social risk amplification. its implication that there is some true or accurate level that becomes amplified is hard to accept in many controversial and contested cases where expertise is lacking or where there is no expert consensus (rayner, ) . the phenomenon of 'dueling experts' is common in conflicts over environmental health, for instance (nelkin, ) . more generally, the concept of risk amplification seems to suggest that there is a risk 'signal' that is outside the social system and is somehow amplified by it (rayner, ) . this seems misconceived when we take the view that ultimately risk itself is a social construction (hilgartner, ) or overlay on the world (jasanoff, ) . and it naturally leads to the view that contributors to the amplification, such as the media (bakir, ) , need to be managed more effectively, and that risk managers should concentrate on fixing the mistake in the public mind (rip, ) , when often it may be the expert assessment that is mistaken. it thus becomes hard to sustain the idea that there is a social process by which true levels of risk get distorted. and this appears to undermine the possibility that risk managers can have a way of anticipating very high or very low levels of social anxiety in any particular case. once risk amplification becomes no more than a subjective judgment by one group on another social group's risk responses, it is hard to see how risk issues can be dealt with on an analytical basis. however, subjective beliefs about risk can produce objective behaviours, and behaviours can interact to produce particular outcomes. and large discrepancies in risk beliefs between different groups are still of considerable interest, whether or not we can know which beliefs are going to turn out to be more correct. in the remainder of this article we therefore explore the consequences of the idea that social risk amplification is nothing more than an attribution, or judgment that one social actor makes of another, and try to see what implications this might have for risk managers based on a systems dynamics model. before this, however, we describe the fieldwork whose principal findings were meant to provide the main structural properties of the model. the aim of the fieldwork was to explore how social actors reason about the risks of recent zoonotic disease outbreaks, and in particular how they make judgments of other actors systematically amplifying or attenuating such risks. this involved a grounded, qualitative study of what a number of groups said in the course of a number of unstructured interviews and focus groups. it follows the general principle of using qualitative empirical work as a basis for systems dynamics modelling (luna-reyes and andersen, ) . focus groups were used where possible, for both lay and professional or expert actors; individual interviews were used where access could only be gained to relevant groups (such as journalists) as individuals. the participants were selected from a range of groups having a stake in zoonotic outbreaks such as avian influenza incidents and are listed in table . the focus groups followed a topic guide that was initially used in a pilot focus group and continually refined throughout the programme. they started with a short briefing on the specific topic of zoonotic diseases, with recent, well-publicised examples. the professional and expert groups were also asked to explain their roles in relation to the management of zoonotic diseases. participants were then invited to consider recent cases and other examples they knew of, discuss their reactions to the risks they presented, and discuss the way the risks had been, or were being, managed. their discussions were recorded and the recordings transcribed except in two cases where it was only feasible to record researcher notes. the individual interviews followed the same format. analysis of the transcripts followed a typical process of grounded theorising (glaser and strauss, ) , in which the aim was to find a way of categorising participants' responses that gave some theoretical insight into the principle of risk amplification as a subjective attribution. the categories were arrived at in a process of 'constant comparison' of the data and emerging, tentative categories until all responses have been satisfactorily categorised in relation to each other (glaser, ) . in glaser's words, 'validity is achieved, after much fitting of words, when the chosen one best represents the pattern. it is as valid as it is grounded'. our approach also drew on template analysis (king, ) in that we started with the basic categories of attributing risk amplification and risk attenuation, not a blank sheet. a fuller account of the analysis process and findings is given in a parallel publication (busby and duckett, ) . the first main theme to emerge from the data was the way in which actors privilege their own views, and construct reasons to hold on to them by finding explanations for other views as being systematically exaggerated or underplayed. it is surprising in a sense that this was relatively symmetrical. we expected expert groups to characterise lay groups as exaggerating or underplaying risk, but we also expected lay groups to use authoritative risk statements from expert groups and organisations of various kinds as ways of correcting their own initial and tentative beliefs. but there was no evidence for this kind of corrective process. the reasons that informants gave for why other actors systematically amplify or attenuate risk were categorised under five main headings: cognition, or the way they formed their beliefs; disposition, or their inherent natures; situation, or the particular circumstances; strategy, or deliberate, instrumental action; and structure, or basic patterns in the social or physical world. for example, one group saw the highly pathogenic avian influenza (hpai) outbreak at holton in the uk in as presenting a serious risk and explained the official advice that it presented only a very small risk as arising from a conspiracy between industry and government that the dispositions of the two naturally created. this second main theme was that some groups of informants often lacked specific and direct knowledge about relevant risks, and resorted to reasoning about other actors' responses to those risks. this reasoning involved moderating those observations with beliefs about whether other actors are inclined to amplify or attenuate risk. lay groups received information through the media but they had definite, and somewhat cliche´d, beliefs about the accuracy of risk portrayals in the media, for example. thus some informants saw the media treatment of hpai outbreaks as risk amplifying and portrayed the media as having an incentive to sensationalise coverage, but others (particularly virologists) saw media coverage as risk attenuating out of scientific ignorance. a third theme was that risk perceptions often came from the specific associations that arose in particular cases. for example, the holton hpai outbreak involved a large food processing firm that had earlier been involved in dietary and nutritional controversies. the firm employed intensive poultry rearing practices and was also importing partial products from a processor abroad. this particular case therefore bound together issues of intensive rearing, global sourcing, zoonotic outbreaks and lifestyle risks-incidental associations that enabled some informants to perceive high levels of risk and indignation, and portray others as attenuating this risk. the fourth theme was that some actors have specific reasons to overcome what they see as other actors' amplifications or attenuations. they do not just discount another actor's distortions but seek to change them. for example, staff in one government agency believed they had to correct farmers who were underplaying risk and not practicing sufficient bio-security, and also correct consumers who were exaggerating risk and boycotting important agricultural products. such actors do not simply observe other actors' expressed risk levels but try to communicate in such a way as to influence these expressed levels-for example through awareness-raising campaigns. the fieldwork therefore pointed to a model in which actors like members of the public based their risk evaluations on what they were told by others, corrected in some way for what they expected to be others' amplifications or attenuations; discrepancies between their current evaluations and those of others would be regarded as evidence of such amplifications, rather than being used to correct their own evaluations. the findings also indicated a model in which risk managers would communicate risk levels in a way that was intended to overcome the misconceptions of actors like the public. these are the underpinning elements of the models we describe below. systems dynamics was a natural choice for this modelling on several grounds. first, there is an inherent stress on endogeneity in the basic idea of social risk amplification, and in particular in the notion that it is an attribution. risk responses first and foremost reflect the way people think about risks and think about the responses of other people to those risks. second, the explicit and intuitive representation of feedback loops was important to show the reflective nature of social behaviour: how actors see the impact of their risk responses on other actors and modify their responses accordingly. third, memory plays an important part in this, since the idea that some actor is a risk amplifier will be based on remembering their past responses, and the accumulative capacity of stocks in systems dynamics provides an obvious way of representing social memory. developing a systems dynamics model on the grounded theory therefore followed naturally, and helped to add a deductive capability to the essentially inductive process of grounded theory (kopainsky and luna-reyes, ) . kopainsky and luna-reyes ( ) also point out that grounded theory can produce large and rich sets of evidence and overly complex theory, making it important to have a rigorous approach to concentrating on small numbers of variables and relationships. thus, in the modelling we describe in the next section, the aim was to try to represent risk amplification with as little elaboration as possible, so that it would be clear what the consequences of the basic structural commitments might be. this meant reduction to the simplest possible system of two actors, interacting repeatedly over time during the period of an otherwise static risk event (such as a zoonosis outbreak). applications of systems dynamics have been wide-ranging, addressing issues in domains ranging from business (morecroft and van der heijden, ) to military (minami and madnick, ) , from epidemiology (dangerfield et al, ) to diffusion models in marketing (morecroft, ) , from modelling physical state such as demography (meadows et al, ) to mental state such as trust martinez-moyano and samsa, ) . applications to issues of risk, particularly risk perception, are much more limited. there has been some application of system dynamics to the diffusion of fear and sarf, specifically (burns and slovic, ; sundrani, ) , but not to the idea of social amplification as an attribution. probably the closest examples to our work in the system dynamics literature deal with trust. luna-reyes et al ( ), for example, applied system dynamics to investigate the role of knowledge sharing in building trust in complex projects. to make modelling tractable, the authors make several simplifying assumptions including the aggregation of various government agencies as a single actor and various service providers as another actor. each actor accumulates the knowledge of the other actor's work, and the authors explore the dynamics that emerge from their interaction. greer et al ( ) modelled similar interactions-this time between client and contractor-each having its own, accumulated understandings of a common or global quantity (in this case the 'baseline' of work a project). martinez-moyano and samsa ( ) developed a system dynamics model to support a feedback theory of trust and confidence. this represented the mutual interaction between two actors (government and public) in a social system where each actor assesses the trustworthiness of the other actor over time, with both actors maintaining memories of the actions and outcomes of the other actor. our approach draws from all these studies, modelling a system in which actors interact on the basis of remembered, past interactions as they make assessments of some common object. the actors are in fact groups of individuals who are presumed to be acting in some concerted way. although this may seem questionable there are several justifications for doing so: ( ) the aim is not to represent the diversity of the social world but to explore the consequences of specific ideas about phenomena like social risk amplification; ( ) in some circumstances a 'risk manager' such as a private corporation or a government agency may act very much like a unit actor, especially when it is trying to coordinate its communications in the course of risk events; ( ) equally in some circumstances it may be quite realistic to see a 'public' as acting in a relatively consensual way whose net, aggregate or average response is of more interest than the variance of response. in the following sections we develop a model in three stages. in the first, we represent the conventional view of social risk amplification; in the second, we add our subjective, attributional approach in a basic form; and in the third we make the attributional elements more realistically complex. the aim is to explore the implications of the principal findings of the fieldwork, and our basic theoretical commitments to social risk amplification as an attribution, with as little further adornment as possible, while also incorporating elements shown in the literature to be important aspects of risk amplification. in the first model, shown in figure , we represent in a simple way the basic notion of social risk amplification. the fundamental idea is that risk responses are socially developed, not simply the sum of the isolated reactions of unconnected individuals. the model represents a population as being in one of two states of worry. this is simpler than the three-state model of burns and slovic ( ) particularly adds to the model. there is also no need for a recovering or removal state, as in sir (susceptible infectious recovered) models (sterman, , p ) , since there is no concept of immunity and it seems certain that people can be worried by the same thing all over again. the flow from an unworried state to a worried state is a function of how far the proportion in the worried state exceeds that normally expected in regard to a risk event such as a zoonotic disease outbreak. members of the public expect some of their number to become anxious in connection with any risk issue: when, through communication or observation, they realise this number exceeds expectation, this in itself becomes a reason for others to become anxious. this observation of fellow citizens is not medium-specific, so it is a combination of observation by word-of-mouth, social networks and broadcast media. in terms of how this influences perception, various processes are suggested in the literature. for example, there is a variety of 'social contagion' effects (levy and nail, ; scherer and cho, ) relevant to such situations. social learning (bandura, ) or 'learning by proxy' (gardner et al, ) may also well be important. we do not model specific mechanisms but only an aggregate process by which the observation of worry influences the flow into a state of being worried. the flow out of the worried state is a natural relaxation process. it is hard to stay worried about a specific issue for any length of time, and the atrophy of vigilance is reported in the literature (freudenberg, ) . there is also a base flow between the states, reflecting the way in which-in the context of any public risk event-there will be some small proportion of the population that becomes worried, irrespective of peers and public information. this base flow also has the function of dealing with the 'startup problem' in which zero flow is a potential equilibrium for the model (sterman, , p ) . the public risk perception in this model stands in relation to an expert, supposedly authoritative assessment of the risk. people worry when seeing others worry, but moderate this response when exposed to exogenous information-the expert or managerial risk assessment. what ultimately regulates worry is some combination of these two elements and it is this regulatory variable that we call a resultant 'risk perception'. unlike burns and slovic ( ) we do not represent this as a stock because it is not anyone's belief, and so need not have inertia. the fact that various members of the public are in different states of worry means that there is no belief that all share, as such. instead, risk perception is an emergent construct on which flows between unworried and worried states depend (and which also determines how demand for risky goods changes, as we explain below). in the simplest model we simply take this resultant risk perception as a weighted geometric mean of the risk implied by the proportion of the population worried and the publically known expert risk assessment. the expert assessment grows from zero toward a finite level, for a certain period, before decaying again to zero. this reflects a time profile for typical risk events-for example zoonotic outbreaks such as sars-where numbers of reported cases climb progressively and rapidly to a peak before declining (eg, leung et al, ) . the units for risk perception and the expert assessment are arbitrary, but for exposition are taken as probabilities of individual fatality during a specific risk event. numerical values of the exogenous risk-related variables are based on an outbreak in which the highest fatality probability is À . but risks in a modern society tend to vary over several orders of magnitude. typically, individual fatality probabilities of À are regarded as 'a very low level of risk', whereas risks of À are seen as very high and at the limit of tolerability for risks at work (hse, ) . because both assessed and perceived risks are likely to vary widely, discrepancies between risk levels are represented as ratios. the way in which the expert assessment is communicated to the public is via some homogenous channel we have simply referred to as the 'media'. in our basic model we represent in very crude terms the way in which this media might exaggerate the difference between expert assessment and public perception. but the sarf literature suggests there is no consistent relationship between media coverage and either levels of public concern or frequencies of fatalities (breakwell and barnett, ; finkel, ) , so the extent of this exaggeration is likely to be highly case specific. it is also possible that the media have an effect on responses by exaggerating to a given actor its own responses. the public, for example, could have an inflated idea of how worried they are because newspapers or blogs portray it to be so. but we do not represent this because it is so speculative and may be indeterminable empirically. finally, the base model also represents the way in which risk perception influences behaviour, in particular the consumption of the goods or services that expose people to the risk in question. the holton uk outbreak of hpai, for example, occurred at a turkey meat processing plant and affected demand for its products; the sars outbreak affected demand for travel, particularly aviation services. brahmbhatt and dutta ( ) even refer to the economic disruption caused by 'panicky' public responses as 'sars type' effects. there are many complications here, not least that reducing consumption of one amenity as a result of heightened risk perception may increase consumption of a riskier amenity. air travel in the us fell after / but travel by car increased and aggregate risk levels were said to have risen in consequence (gigerenzer, ) . a further complication is that in certain situations, such as bank runs (diamond and dybvig, ), risk perceptions are directly self-fulfilling rather than self-correcting. the most common effect is probably that heightened risk perceptions will lead to reduced demand for the amenity that causes exposure, leading to reductions in exposure and reductions in the expert risk assessment, but it is worth noting that the effect is case-specific. the expert risk assessment is therefore not exogenous, and there is a negative feedback loop that operates to counteract rising risk perceptions. as we show later from the simulation outcomes, the base model shows a public risk perception that can be considerably larger than the expert risk assessment. it therefore seems to show 'risk amplification'. but there is no variable that stands for risk in the model: there are only beliefs about risk (called either assessments or perceptions). the idea that social risk amplification is a subjective attribution, not an objective phenomenon, means that this divergence of risk perception and expert assessment does not amount to risk amplification. and it says that actors see others as being risk amplifiers, or attenuators, and develop their responses accordingly. this means that we need to add to sarf, and the basic model of the previous section, the processes by which actors observe, diagnose and deal with other actors' risk assessments or perceptions. what our fieldwork revealed was that the social system did not correct 'mistaken' risk perceptions in some simpleminded fashion. in other words, it was not the case that people formed risk perceptions, received information about expert assessment, and then corrected their perceptions in the correct direction. instead, as we explained earlier, they found reasons why expert assessments, and in fact the risk views of any other group, might be subject to systematic amplification or attenuation. they then corrected for that amplification. risk managers, on the other hand, had the task of overcoming what they saw as mistaken risk responses in other groups, not simply correcting for them. therefore in the second model, shown in figure , we now have a subsystem in which a risk manager (a government agency or an industrial undertaking in the case of zoonotic disease outbreaks) observes the public risk perception in relation to the expert risk assessment, and communicates a risk level that is designed to compensate for any discrepancy between the two. commercial risk managers will naturally want to counteract risk amplification that leads to revenue losses from product and service boycotts, and governmental risk managers will want to counteract the risk amplification that produces panic and disorder. as beck et al ( ) report, the uk bse inquiry found that risk managers' approach to communicating risk 'was shaped by a consuming fear of provoking an irrational public scare'. the effect is symmetrical to the extent that the public in turn observes discrepancies between managerial communications and its own risk perceptions, and attributes amplification or attenuation accordingly. attributions are based on simple memory of past observations. this historical memory of another actor's apparent distortions is sometimes mentioned in the sarf literature (kasperson et al, ; poumadere and mays, ) . this memory is represented as stocks of observed discrepancies, reaching a level m i (t)for actor i at time t. the managerial memory, for example, is r public ðtÞ r expert ðtÞ dt m i (t) implies that actor i sees the other actor as exaggerating risk, while m i (t)o implies perceived attenuation. the specific deposits in an actor's memory are not retrievable, and equal weight is given to every observation that contributes to it. the perceived scale of amplification is the time average of memory content, and the confidence the actor has in this perceived amplification is Àe À|m(t)| where confidence grows logarithmically towards unity as the magnitude of the memory increases. the managerial actor modifies the risk level it communicates by the perceived scale of public amplification raised to the power of its confidence, while the public adjusts the communicated risk level it takes account of by the perceived scale of managerial attenuation raised to the power of its confidence in this. in the third model, in figure , we add three elements found in the risk amplification literature that become especially relevant to the idea of risk amplification as a subjective attribution: confusion, distrust and differing perceptions about the significance of behavioural change. the confusion issue reflects the way an otherwise authoritative actor's view tends to be discounted if it shows evidence of confusion, uncertainty or inexplicable change. two articles in the recent literature on zoonosis risk (bergeron and sanchez, ; heberlein and stedman, ) specifically describe the risk amplifying effect of the authorities seeming confused or uncertain. the distrust issue reflects the observation that 'distrust acts to heighten risk perception . . . ' (kasperson et al, ) , and that it is 'associated with perceptions of deliberate distortion of information, being biased, and having been proven wrong in the past' (frewer, , p ) . a distinguishing aspect of trust and distrust is the basic asymmetry such that trust is quick to be lost and slow to be gained (slovic, ) . in figure , the confusion function is based on the rate of change of attributed amplification, not rate of change communication itself, since some change in communication might appear justified if correlated with a change in public perception: g ¼ À e Àg c g ðtÞ j j ; where c g (t) is the change in managerial amplification in unit time. the distrust function is based on the extent of remembered attributed amplification: f ¼ À e Àf m g ðtÞ j j ; where m g (t) is the memory of managerial risk amplification at time t and f is the distrust parameter. there is no obvious finding in the literature that would help us set the value of such a parameter. the combination of the confusion and distrust factors is a combination of an integrator and a differentiator. it is used to determine how much weight is given to managerial risk communications in the formation of the resultant risk perception. it is defined such that as distrust and confusion both approach unity, this weight w tends to zero: w ¼ w max ( Àg)( Àf). this weight was exogenous in the previous model, so the effect of introducing confusion and distrust is also to endogenise the way observation of worry is combined with authoritative risk communication. the third addition in this model is an important disproportionality effect. the previous models assume that risk managers base their view of the public risk perception on some kind of direct observation-for example, through clamour, media activity, surveys and so on. in practice, the managerial view is at least partly based on the public's consumption of the amenity that is risk, for example the consumption of beef during the bse crisis, or flight bookings and hotel reservations during the sars outbreak. the problem is that when a foodstuff like beef becomes a risk object it may be easy for many people to stop consuming it, and such a response from the consumer's perspective can be proportionate to even a mild risk assessment. reducing beef consumption is an easy precaution for most of the population to take (frewer, ) , so rational even when there is little empirical evidence that there is a risk at all (rip, ) . yet this easy response of boycotting beef may be disastrous for the beef industry, and therefore seem highly disproportionate to the industry, to related industries and to government agencies supporting the industry. unfortunately there is considerable difficulty in quantifying this effect in general terms. recent work (mehers, ) looking at the effect of heightened risk perceptions around the avian influenza outbreak at a meat processing plant suggests that the influence on the demand for the associated meat products was very mixed. different regions and different demographic groups showed quite different reactions, for example, and the effect was confounded by actions (particularly price changes) taken by manufacturer and retailers. our approach is to represent the disproportionality effect with a single exogenous factorthe relative substitutability of the amenity for similar amenities on the supply and demand side. the risk manager interprets any change in public demand for the amenity multiplied by this factor as being the change in public risk perception. if the change in this inferred public risk perception exceeds that observed directly (for example by opinion survey), then it becomes the determinant of how risk managers think the public are viewing the risk in question. this relative substitutability is entirely a function of the specific industry (and so risk manager) in question: there is no 'societal' value for such a parameter, and the effects of a given risk perception on amenity demand will always be case specific. for example, brahmbhatt and dutta ( ) reported that the sars outbreak led to revenue losses in beijing of % in tourist attractions, exhibitions and hotels, but of - % in travel agencies, airlines, railways and so on. the effects are substantial but a long way from being constant. in this section we briefly present the outcomes of simulation with two aims: first to show how the successive models produce differences in behaviour, if at all, and thereby to assess how much value there is in the models for policymakers; second to assess how much uncertainty in figure model of a more complex attributional view of risk amplification. outcomes such as public risk perception is produced by uncertainty in the exogenous parameters. figure shows the behaviour of the three successive models in terms of public risk perception and expert risk assessment. for the three models, the exogenous variables are set at their modal values and when variables are shared between models they have the same values. the expert risk assessment is thus very similar for each model, as shown in the figure, rising towards its target level, falling as public risk perception reduces exposure, and then ceasing as the crisis ends around day . in the base model, the public risk perception is eight times higher than the expert assessment at its peak, which occurs some days after that in the expert assessment. but once the attributional view of risk amplification is modelled, this disparity becomes much greater, and it occurs earlier. in the simple attributional system the peak discrepancy is over times, and in the complex attributional system nearly times, both occurring within days of the expert assessment peak. thus the effect of seeing risk amplification as the subjective judgment of one actor about another is, given the assumptions in our models, to polarise risk beliefs much more strongly and somewhat more rapidly. we can no longer call the outcome a 'risk amplification' since, by assumption, there is no longer an objective risk level exogenous to the social system. but there is evidently strong polarisation. there is some qualitative difference in the time profile of risk perception between the three models, as shown in the previous figure where the peak risk perception occurs earlier in the later models. there are also important qualitative differences in the time profiles of stock variables amenity demand and worried population, as shown in figure . when the attributional view is taken, both demand and worry take longer to recover to initial levels, and when the more complex attributional elements are modelled (the effects of mistrust, confusion and different perceptions of the meaning of changes in demand), the model indicates that little recovery takes place at all. the scale of the recovery depends on the value of the exogenous parameters, and some of these (as we discuss below) are case specific. but of primary importance is the way the weighting given to managerial communications or expert assessment is dragged down by public attributions. this result indicates the importance of a complex, attributional view of risk amplification. unlike the base model, in the attributional model it is much more likely there will be an indefinite residue from a crisis-even when the expert assessment of risk falls to near zero. figures and show the time development of risk perception in the third model in terms of the mean outcome with (a) % confidence intervals on the mean and (b) tolerance intervals for % confidence in % coverage over runs, with triangular distributions assigned to the exogenous parameters and plausible ranges based solely on the author's subjective estimates. the exogenous parameters fall into two main groups. the first group is of case-specific factors and would be expected to vary between risk events. this includes, for example, the relative substitutability of the amenity that is the carrier of the risk, and the latency before changes in demand for this amenity change the level of risk exposure. the remaining parameters are better seen as social constants, since there is no theoretical reason to think that they will vary from one risk event to another. these include factors like the natural vigilance period among the population, the normal flow of people into a state of worry, the latency before people become aware of a discrepancy between emergent risk perception and the proportion of the population that is in a state of worry. figure shows the confidence and tolerance intervals with the social constants varying within their plausible ranges and the case-specific factors fixed at their modal values, and figure vice versa. thus figure shows the effect of our uncertainty about the character of society, figure outcomes of the three models. whereas figure shows the effect of the variability we would expect among risk events. the substantial difference between the means in risk perception between the two figures reflects large differences between means and modes in the distributions attributed to the parameters, which arises because plausible ranges sometimes cover multiple orders of magnitude (eg, the confusion and distrust constants both range from to with modes of , and the memory constant from to with a mode of ). these figures do not give a complete understanding, not least because interactions between the two sets of parameters are possible, but they show a reasonably robust qualitative profile. figure shows the 'simple' correlation coefficients between resultant risk perception and the policy-relevant exogenous parameters over time, as recommended by ford and flynn ( ) as an indication of the relative importance of model inputs. at each day of the simulation, the sample correlation coefficient is calculated for each parameter over the runs. no attempt has been made to inspect whether the most important inputs are correlated, and to refine the model in the light of this. nonetheless the figure gives some indication of how influential are the most prominent parameters: the expert initial assessment level (ie, the original scale of the risk according to expert assessment), the expert assessment adjustment time (ie, the delay in the official estimate reflecting the latest information), the base flow (the flow of people between states of non-worry and worry in relation to a risk irrespective of the specific social influences being modelled) and the normal risk perception (the baseline against which the resultant risk perception is gauged, reflecting a level of risk that would be unsurprising and lead to no increase in the numbers of the worried). the first of these is case-specific, but the other three would evidently be worth empirical investigation given their influence in the model. it is extremely difficult to test such outcomes against empirical data because cases differ so widely and it is unusual to find data on simultaneous expert assessments and public perceptions over short-run risk events like disease outbreaks, particularly outbreaks of zoonotic disease. but a world bank paper of , on the economic effects of infectious disease outbreaks (primarily sars, a zoonotic disease), collected together data gathered on the sars outbreak, and some-primarily that of lau et al ( ) -showed the day-by-day development of risk perception alongside reported cases. figure is based on lau et al's data ( ) , and shows the number of reported cases of sars as a proportion of the hong kong population at the time, together with the percentage of people in a survey expressing a perception that they had a large or very large chance of infection from sars. the two lines can be regarded as reasonably good proxies for the risk perception and expert assessment outcomes in figure and they show a rough correspondence: a growth in both perception and expertly assessed or measured 'reality', followed by a decay, in which the perception appears strongly exaggerated from the standpoint of the expert assessment. the perceptual gap is about four orders of magnitude-greater than even the more complex attributional system in our modelling. moreover, the risk perception peak occurs early, and in fact leads the reported cases peak. it is our models and especially in which the perception peak occurs early (although it never leads the expert assessment peak). the implications of the work the social amplification of risk framework has always been presented as an 'integrative framework' (kasperson et al, ) , rather than a specific theory, so there has always been a need for more specific modelling to make its basic concepts precise enough to be properly explored. at the same time, as suggested earlier, its implication that there is some true level of risk that becomes distorted in social responses has been criticised for a long time. we therefore set out to explore whether it is possible to retain some concept of social risk amplification in cases where even expert opinion tends to be divided, the science is often very incomplete, and past expert assessment has been discredited. zoonotic disease outbreaks provide a context in which such conditions appear to hold. our fieldwork broadly pointed to a social system in which social actors of all kinds privilege their own risk views, in which they nonetheless have to rely on other actor's responses in the absence of direct knowledge or experience of the risks in question, in which they attribute risk amplification or attenuation to other actors, and in which they have reasons to correct for or overcome this amplification. to explore how we can model such processes has been the main purpose of the work we have described. and the resulting model provides specific indications of what policymakers need to deal with-a much greater polarisation of risk beliefs, and potentially a residue of worry and loss of demand after the end of a risk crisis. it also has the important implication that risk managers' perspectives should shift, from correcting a public's mistakes about risk to thinking about how their own responses and communications contribute to the public's views about a risk. our approach helps to endogenise the risk perception problem, recognising that it is not simply a flaw in the world 'out there'. it is thus an important step in becoming a more sophisticated risk manager or manager of risk issues (leiss, ) . it is instructive to compare this model with models like that of luna-reyes et al ( ) which essentially involve a convergent process arise from knowledge sharing, and the subsequent development of trust. we demonstrate a process in which there is knowledge sharing, but a sharing that is undermined by expectations of social risk amplification. observing discrepancies in risk beliefs leads not to correction and consensus but to self-confirmation and polarisation. our findings in some respects are similar to greer et al ( ) , who were concerned with discrepancies in the perceptions of workload in the eyes of two actors involved in a common project. such discrepancies arose not from exogenous causes but from unclear communication and delay inherent in the social system. all this reinforces the long-held view in the risk community, and of risk communication researchers in particular, that authentic risk communication should involve sustained relationships, and the open recognition of uncertainties and difficulties that would normally be regarded as threats to credibility (otway and wynne, ) . the reason is not just the moral requirement to avoid the perpetuation of powerful actors' views, and not just the efficiency requirement to maximise the knowledge base that contributes to managing a risk issue. the reason is also that the structure of interactions can be unstable, producing a polarisation of view that none of the actors intended. actors engaged with each other can realise this and overcome it. a basic limitation to the use of the models to support specific risk management decisions, rather than give more general insight into social phenomena, is that there are very few sources of plausible data for some important variables in the model, such as the relaxation delay defining how long people tend to stay worried about a specific risk event before fatigue, boredom or replacement by worry about a new crisis leads them to stop worrying. it is particularly difficult to see where values of the case-specific parameters are going to come from. other sd work on risk amplification at least partly avoids the calibration problem by using unit-less normalised scales and subjective judgments (burns and slovic, ) . and one of the benefits of this exploratory modelling is to suggest that such variables are worthwhile subjects for empirical research. but at present the modelling does not support prediction and does not help determine best courses of action at particular points in particular crises. in terms of its more structural limitations, the model is a small one that concentrates specifically on the risk amplification phenomenon to the exclusion of the many other processes that, in any real situation, risk amplification is connected with. as such, it barely forms a 'microworld' (morecroft, ) . it contrasts with related work such as that of martinez-moyano and samsa's ( ) modelling of trust in government, which similarly analyses a continuing interaction between two aggregate actors but draws extensively on cognitive science. however, incorporating a lot more empirical science does not avoid having to make many assumptions and selections that potentially stand in the way of seeing through to how a system produces its outcomes. the more elaborate the model the more there is to dispute and undermine the starkness of an interesting phenomenon. we have had to make few assumptions about the world, about psychology and about sociology before concluding that social risk amplification as little more than a subjective attribution has a strongly destabilising potential. this parsimony reflects towill's ( ) notion that we start the modelling process by looking for the boundary that 'encompasses the smallest number of components within which the dynamic behaviour under study is generated'. the model attempts to introduce nothing that is unnecessary to working out the consequences of risk amplification as an attribution. as ghaffarzadegan et al ( ) point out in their paper on small models applied to problems of public policy, and echoing forrester's ( ) argument for 'powerful small models', the point is to gain accessibility and insight. having only 'a few significant stocks and at most seven or eight major feedback loops', small models can convey the counterintuitive endogenous complexity of situations in a way that policymakers can still follow. they are small enough to show systems in aggregate, to stress the endogeneity of influences on the system's behaviour, and to clearly illustrate how policy resistance comes about (ghaffarzadegan et al, ) . as a result they are more promising as tools for developing correct intuitions, and for helping actors who may be trapped in a systemic interaction to overcome this and reach a certain degree of self-awareness (lane, ) . the intended contribution of this study has been to show how to model a long-established, qualitative framework for reasoning about risk perception and risk communication, and in the process deal with one of the main criticisms of this framework. the idea that in a society the perception of a risk becomes exaggerated to the point where it bears no relation to our best expert assessments of the risk is an attractive one for policymakers having to deal with what seem to be grossly inflated or grossly under-played public reactions to major events. but this idea has always been vulnerable to the criticism that we cannot know objectively if a risk is being exaggerated, and that expert assessments are as much a product of social processes as lay opinion. the question we posed at the start of the paper was whether, in dropping a commitment to the idea of an objective risk amplification, there is anything left to model and anything left to say to policymakers. our work suggests that there is, and that modelling risk amplification as something that one social actor thinks another is doing is a useful thing to do. there were some simple policy implications emerging from this modelling. for example, once you accept that there is no objective standard to indicate when risk amplification is occurring, actors are likely to correct for other actors' apparent risk amplifications and attenuation, instead of simple-mindedly correcting their own risk beliefs. this can have a strongly polarising effect on risk beliefs, and can produce residual worry and loss of demand for associated products and services after a crisis has passed. the limitations of the work point to further developments in several directions. first, there is a need to explore various aspects of how risk managers experience risk amplification. for example, the modelling, as it stands, concentrates on the interactions of actors in the context of a single event or issue-such as a specific zoonotic outbreak. in reality, actors generally have a long history of interaction around earlier events. we take account of history within an event, but not between events. a future step should therefore be to expand the timescale, moving from intra-event interaction to inter-event interaction. the superposition of a longer term process is likely to produce a model in which processes acting over different timescales interact and cannot simply be treated additively (forrester, ) . it also introduces the strong possibility of discontinuities, particularly when modelling organisational or institutional actors like governments whose doctrines can change radically following elections-rather like the discontinuities that have to be modelled to represent personnel changes and consequences like scapegoating (howick and eden, ) . another important direction of work would be a modelling of politics and power. it is a common observation in risk controversies that risk is a highly political construction-being used by different groups to gain resources and influence. as powell and coyle ( ) point out, the systems dynamics literature makes little reference to power, raising questions about the appropriateness of our modelling approach to a risk amplification subject-both in its lack of power as an object for modelling, and its inattention to issues of power surrounding the use of the model and its apparent implications. powell and coyle's ( ) politicised influence diagrams might provide a useful medium for representing issues of power, both within the model of risk amplification and in the understanding of the system in which the model might be influential. the notion, as currently expressed in our modelling, that it is always in one actor's interest to somehow correct another's amplification simply looks naı¨ve. greenpeace v. shell: media exploitation and the social amplification of risk framework (sarf) social learning theory public administration, science and risk assessment: a case study of the uk bovine spongiform encaphalopathy crisis media effects on students during sars outbreak world bank policy research working paper , the world bank east asia and pacific region chief economist's office social amplification of risk and the layering method the diffusion of fear: modeling community response to a terrorist strike incorporating structural models into research on the social amplification of risk: implications for theory construction and decision making social risk amplification as an attribution: the case of zoonotic disease outbreaks model-based scenarios for the epidemiology of hiv/aids: the consequences of highly active antiretroviral therapy bank runs, deposit insurance, and liquidity risk and culture: an essay on the selection of technological and environmental dangers risk and relativity: bse and the british media the social amplification of risk perceiving others' perceptions of risk: still a task for sisyphus statistical screening of system dynamics models nonlinearity in high-order models of social systems system dynamics-the next fifty years institutional failure and the organizational amplification of risk: the need for a closer look trust, transparency, and social context: implications for social amplification of risk workers' compensation and family and medical leave act claim contagion how small system dynamics models can help the public policy process out of the frying pan into the fire: behavioral reactions to terrorist attacks conceptualization: on theory and theorizing using grounded theory the discovery of grounded theory improving interorganizational baseline alignment in large space system development programs socially amplified risk: attitude and behavior change in response to cwd in wisconsin deer the social construction of risk objects: or, how to pry open networks of risk on the nature of discontinuities in system dynamics modelling of disrupted projects reducing risks, protecting people bridging the two cultures of risk analysis the social amplification and attenuation of risk the social amplification of risk: assessing fifteen years of research and theory the social amplification of risk the social amplification of risk: a conceptual framework qualitative methods and analysis in organizational research: a practical guide closing the loop: promoting synergies with other theory building approaches to improve systems dynamics practice social theory and systems dynamics practice monitoring community responses to the sars epidemic in hong kong: from day to day the chamber of risks: understanding risk controversies a tale of two cities: community psychobehavioral surveillance and related impact on outbreak control in hong kong and singapore during the severe acute respiratory syndrome epidemic contagion: a theoretical and empirical review and reconceptualization the impact of social amplification and attenuation of risk and the public reaction to mad cow disease in canada collecting and analyzing qualitative data for system dynamics: methods and models knowledge sharing and trust in collaborative requirements analysis a feedback theory of trust and confidence in government the limits to growth: the -year update on the quantitative analysis of food scares: an exploratory study into poultry consumers' responses to the h n avian influenza outbreaks in the uk food supply chain dynamic analysis of combat vehicle accidents strategy support models systems dynamics and microworlds for policymakers modelling the oil producers: capturing oil industry knowledge in a behavioural simulation model. modelling for learning, a special issue of the science controversies: the dynamics of public disputes in the united states risk communication: paradigm and paradox (guest editorial) the dynamics of risk amplification and attenuation in context: a french case study identifying strategic action in highly politicized contexts using agent-based qualitative system dynamics muddling through metaphors to maturity: a commentary on kasperson et al. 'the social amplification of risk' risk communication and the social amplification of risk should social amplification of risk be counteracted folk theories of nanotechnologists a social network contagion theory of risk perception perception of risk perceived risk, trust and democracy business dynamics: systems thinking and modelling for a complex world understanding social amplification of risk: possible impact of an avian flu pandemic. masters dissertation, sloan school of management and engineering systems division acknowledgements-many thanks are due to the participants in the fieldwork that underpinned the modelling, and to dominic duckett who carried out the fieldwork. we would also like to thank the anonymous reviewers of an earlier draft of this article for insights and suggestions that have considerably strengthened it. the work was partly funded by a grant from the uk epsrc. key: cord- - oqkdbq authors: bley, katja; schön, hendrik; strahringer, susanne title: overcoming the ivory tower: a meta model for staged maturity models date: - - journal: responsible design, implementation and use of information and communication technology doi: . / - - - - _ sha: doc_id: cord_uid: oqkdbq when it comes to the economic and strategic development of companies, maturity models are regarded as silver bullets. however, the existing discrepancy between the large amount of existing, differently developed models and their rare application remains astonishing. we focus on this phenomenon by analyzing the models’ interpretability and possible structural and conceptual inconsistencies. by analyzing existing, staged maturity models, we develop a meta model for staged maturity models so different maturity models may share common semantics and syntax. our meta model can therefore contribute to the conceptual rigor of existing and future maturity models in all domains and can be decisive for the success or failure of a maturity measurement in a company. economic development, the assumption of growth, and the ongoing transformation of processes increase the competitive pressure on enterprises of all sizes. hence, they search for tools that can help determine their current benchmarking position or assess their subsequent performance compared to a predefined best-practice performance [ ] . one famous example of these benchmarking tools is maturity models (mms). considering their components, two definitions result: 'maturity' as a "state of being complete, perfect or ready" [ ] and 'model' as "an abstraction of a (real or language based) system allowing predictions or inferences to be made" [ ] . thus, an mm can be regarded as an abstraction of a system that allows predictions or inferences about a complete or perfect state. its aim is a structured, systematic elaboration of best practices and processes that are related to the functioning and structure of an organization [ ] . the mm is divided into different levels, which are used as benchmarks for the overall maturity of an organization. by application, an object is able to assess its own position in a predefined best-practice approach. an awareness of the maturity level in a particular domain is necessary in order to recognize improvement potentials and stimulate a continuous improvement process [ ] . especially when it comes to emerging phenomena like digitalization or the industrial internet, mms became a favored instrument especially for smes in order to determine their own digitalization status and corresponding guidelines for improvement. but it is not only in this area that there has been an enormous increase in mm publications in recent years. due to their general applicability, the rather simple development, and the promising competitive advantages many mms have emerged in both science and practice over the last decade. they differ in the concepts of maturity, the considered domains, the development approach, or the target group [ , ] . although there exist several approaches to the rigorous development (i.e. [ ] [ ] [ ] ), and there are mms that already follow these guidelines and should be considered as complete and thorough, a plethora of developers still see a need for new and different kinds of mms. a possible reason for this phenomenon may be that current mms lack in consistency regarding their semantics, syntax, and concepts. this leads to a situation where many different mms introduce ontologies that are either conceptually the same but are used for different subjects and in different contexts so researchers do not understand or recognize similar models or they try to differentiate from existing approaches by using different terminology (e.g. level vs. stages, maturity vs. assessment, dimension vs. area). this in turn weakens the benefits of mms for companies, as they are unable to understand, compare, and apply the models due to different terminology or incomprehensible concepts and relationships. as a consequence, a solution is needed that proposes theoretical and practical solutions for the semantic and syntactical pitfalls and difficulties mentioned. we address this problem of conceptual divergence by introducing a uml-based meta model of mmsa formal construct that summarizes the various mm concepts, features, and interdependencies and relates them to each other. we introduce this meta model regarding the different mm concepts, where each mm can be an instance of it as it provides a conceptual template for the rigorous development of new and the evaluation of existing maturity models. we thereby focus on staged mms, as they have a holistic and cross-organizational structure that is dominant in mm research. based on the most popular mms as well as several newly developed mms, we summarize and discuss already existing mm concepts. we are able to show a valid instantiation of our maturity meta model by classifying different staged mms in their properties and concepts with our meta model. according to [ ] , mms typically consist of: (a) a number of levels (commonly - ), (b) a generic description of each level, (c) a number of dimensions or "process areas", (d) a number of elements or activities for each process area, and (e) a description of the performance of each activity or element as it might perform or might be performed at each maturity level. the purpose of mm assessment can vary. depending on its intended aim, the mm focuses on an assessment of the current state of development (descriptive), the identification of improvement potential (prescriptive), or its application as a benchmarking tool (comparative) for an object under study [ ] . mm development has faced an evolution of around years. the early approaches of these models reach back to the late s, when crosby and nolan developed the first preliminaries of today's mm: the five-staged quality management maturity grid and the stage theory on electronic data processing [ , ] . the concept of these models gained more and more attention, and to date, the capability maturity model integration (cmmi) [ ] represents one of the most well-known and applied mms. although there are different types of mms, we focus on the dominant form, the staged mm. its basic assumption is that an organization is constantly evolving, interand intra-organizationally, due to learning effects and improvements. this evolution process is represented in a simplified step-by-step approach using a certain number of degrees of maturity (usually - ), a combination of key factors that must be given in order to achieve a certain level of maturity. figure shows a common representation of a staged mm. when using such an mm, the presence of relevant factors is measured via indicators. after conducting this assessment, the results are summarized with corresponding evaluations and a representative maturity level (see fig. ). researchers who develop an mm have to decide on a research design for model development (i.e., empirical qualitative/quantitative, conceptual, design-oriented or other), specific research methods to be applied (i.e., case studies, surveys, literature reviews, interviews, conceptual developments or no method at all), and the research content that has to be elaborated and described (i.e., conceptual, descriptive, assessment, comparative, or others) [ ] . considering the long history of mms, multiple development guidelines evolved over time that focus on these decisive processes and can be regarded as supportive instruments for mm development. [ ] present the most established standard for the development of mms: a six-step generic phase model comprising scope, design, populate, test, deploy, and maintain. a different procedure model is postulated by [ ] , who strongly focus on a design science research approach. they suggest an eight-phase development approach, focusing on problem definition, comparison of existing mms, determination of the development strategy, iterative mm development, the conception of transfer and evaluation, implementation of transfer media, evaluation, and rejection of mm. [ ] set up general design principles (principles of form and function) for the development of useful mms. a framework for the characterization of existing mms is presented by [ ] . in order to create sound and widely accepted mms, he proposes two "maturity model cycles". based on a summarizing, many approaches can support researchers in creating mms. however, these guidelines are limited in their interpretability and validity, as they do not provide concrete terminology specifications or structural concept models. consequently, many models were built following these recommendations, but they still differ in the terminology used, their descriptive principles, the extent of provided information about factors, and the factors' influence on the degree of maturity [ ] . in spite of their salient popularity, there are also critical voices regarding the scientific implications and the practical applicability of mms. from a scientific point of view, [ ] criticize a lack of empirical research foundation during the development process. this in turn, often leads to a simple copy of existing model structures (i.e., five-staged-mms) without considering the conceptual suitability for the given context. furthermore, [ ] mention the lack of validation when selecting factors or dimensions for the model as well as the linear course of maturation. another aspect is the missing operationalization of maturity measurements which impedes replication of an evaluation. similarly, [ ] argue that a simple demonstration of a gap (as done in many mms) is not enough. only understanding cause and effect can help to guide action. otherwise these models might convey the impression of "falsified certainty" to their users [ ] . a common promise of mms is an increase in competitiveness, effectiveness and efficiency in a specific domain or a general improvement of constitution. however, mms lack a clear description and definition of these concepts in their conceptual grounding, making them theoretical constructs but not applicable models for users (i.e., smes). [ ] summarize shortcomings of mms from a practical perspective. they state that mms are often inflexible, although a high flexibility is required. similar to [ ] , they argue that mms focus on identifying gaps and raising awareness towards a specific topic, but enterprises are in need of concrete recommendations for action. furthermore, these models are disciplinary, impractical, and overwhelming for many enterprises, especially for smes where a corresponding executive management level is missing. however, not only the application and general structure of models is a topic of criticism. inconsistent terminology of the models' concepts plays a major role in mm research. when developing new mms, developers tend to invent new titles for common mm approaches in order to stand out from the plethora of already existing mm terms. for instance, designations like "maturity model", "assessment model", "roadmap", "maturity framework", "maturity matrix", "guide to maturity", and "excellence model" are synonyms for the concept of an mm. inconsistent terminology used within mms, indirectly results in an inflation of similar mms, as these models will not be identified as identical or at least similar approaches. this inconsistent terminology may cause, among other problems, two different semantic-syntax errors between mms: (a) developers use the same terminology for maturity concepts, but each concept has a different interpretation (e.g., homonyms), and (b) developers address the same concept but label it in a different way (e.g., synonyms). for instance, to describe the field of interest an mm is applied to, mms use different terms: "domain", "area", or "dimension". but not only the terminology, also the relationships between these concepts remain unclear. there rarely is a standardized definition framework of concepts and relationships. although there exists a meta model for project management competence models, it is only available in german and focuses on the broader application context, not on the maturity model itself [ ] . in conclusion, these inconsistencies represent a research gap as they lead to a situation where a comparison of existing mms is almost impossible due to a lack of understanding what is actually provided. a recent study revealed that almost all scientific and many consultancy mms (in the field of industrial internet) are still not being used or are even unknown in business practice, although potential applicants stated a need for and interest in these models in general [ ] . as this discrepancy comes as a surprise, we assume a weak point either in the development or application of mms (fig. ) ; otherwise, the reluctant use in praxis is not explicable. a possible explanation for this phenomenon may be located in the different interpretations of mms' concepts on both sides of stakeholders. developers may use concepts and relationships that are interpreted differently by model applicants and by other developers in the same field of research. as a result, possible applicants as well as researchers misinterpret an mm's structure and concepts. in other words, syntax and semantics vary across mms, which is why, on the one hand, companies (as applicants) do not use or apply the models, as they simply do not understand their structure and effects. on the other hand, researchers tend to develop even more and new mms, as they do not recognize existing models as similar or comparable mm approaches (i.e., from to , new mms had been developed in the domain of industry . [ ] ). what is needed is an overarching, holistic conceptual framework that can unify and standardize the individual objectives and goals. by focusing on the developer's perspective in this paper, the consistency of existing models can be validated, and future models can be developed in a rigorous way, which will then be the foundation for a sound assessment of models later on. a meta model, as "a model of models" [ ] , is an approach that is able to meet the above-mentioned requirements. originating from software engineering, it can be understood as a model specifying underlying models that are considered instantiations of the meta model. the meta model itself is instantiated from a meta model [ ] . in many cases, the meta meta model specifies a subset of the modeling language uml. a meta model typically represents the language (linguistic) and/or structure (ontological) of the underlying (that is, the subordinate) models [ ] . it is used as an abstract description of ( ) the unification of concepts that are part of the model, ( ) the specifications and definition of concepts, and ( ) the specifications and definition of concepts. thus, multiple valid models can be instances from the same meta model. the schematic meta-modeling approach for mms is shown in fig. . on the top layer (m ), the developed meta model is located. it specifies the underlying mm on m . every concept on m has to adhere to a type concept specified on m . please note that the mm on m is, in principle, generally valid for all enterprises. it specifies the ranges and used concepts for a later assessment. the developer creates an mm type on m (according to the meta model on m ), which is "instantiated" on m . on the m level, the "instantiated" and applied mm characterizes a concrete enterprise. this can be done by the user (the person who applies a constructed mm onto an enterprise). with a detailed meta model for mms, different mms can be regarded as instances of it, leading to common semantics and syntaxes between those models. only by providing a unified meta model ("specification" in fig. ) , that reveals concepts, relations and entities, will the developer be able to develop a rigorous and logical mm through instantiation ("structure" in fig. ). our research can therefore contribute fundamentally to the development of future staged mms in all areas, as its structure on the m level and its instantiation on the m level are directly interdependent and thus decisive for the success or failure of a maturity measurement in a company. however, compared to classical meta-modeling, where the m model concepts serve as types for m level instances, here, the instantiation of m from m must be interpreted differently. during assessment (the process that creates the m instances), only some of the concepts from the m level are instantiated with "values" that ultimately (typically by calculation) yield the resulting maturity level for the assessed enterprise. the development of the meta model for maturity models ( m) was based on a study of the most common and representative staged mms. in order to elaborate sufficient meta model elements that are valid for a broad class of staged mms, an analysis of different staged mms, their development and their structure was conducted to summarize and analyze existing concepts, their relationships as well as their multiplicities and instantiations. as a result, universally applicable meta concepts were found and related to each other. for that, we followed an opportunistic approach: we identified the necessary meta model concepts (and their relations) by reduction and selection via decision criteria based on the available content of the investigated mms. the resulting meta model in fig. and table show the m level structure as a conceptual and formal description of an mm, whose instances (on m level) reflect the actual concepts and interrelations within various staged mms investigated in our study (see also sect. ). the staged structure of mms is built on common assumptions, which influenced the meta model's design: . a maturity model is typically developed within a domain, as it represents a path of growth for a specific field of interest. . an mm consists of several maturity levels (i.e., l -l ), which have an ascending order and are intended to represent the improvement path of an object's ability. . the mm can (but does not have to) be divided into dimensions (i.e., d -d ). these dimensions divide the object of study into fields of interest, mostly depending on the domain in which the model is applied. this subdivision serves the possibility of an incremental measurement of maturity in a certain area. . factors are properties of the maturity level and (if available) are grouped in dimensions. there are two possible ways of maturity determination by factors: a. in the first approach, the respective maturity level is specified by checking the threshold of certain (related) factors that are applied. due to the fact that a factor could be used on several levels (however, with different values per level), the factor specification, which assigns the concrete expected value of a factor per maturity level, is used between both. thus, e.g., level l could use factor f with the factor specification f l while level l is still able to use f as well (then with its own factor specification f l ). this is used to express evolving enterprise maturity via growing requirements per level of the same factor. b. second, there is the possibility that the maturity level can be determined per dimension instead of "requirements per level". for this purpose, the factors need to be assigned to the respective dimensions and calculated by using a dimensionspecific aggregation formula (e.g., sum, average, median). in contrast to the approach (a), the dimensions do not use any threshold as intermediate value. thus, a factor can only be assigned to one dimension, and the dimension maturity is calculated by its set of factors (and their current value). a factor cannot be used twice, and the dimension maturity is always a representation of the current state of the enterprise (within this dimension). the mm can make use of both, several dimensions of maturity and an overall level maturity since maturity levels and dimensions are modeled orthogonally in the meta model. however, the specification of the maturity level is still mandatory. within an mm, its value is still abstract and not directly transferable into an object under study. this transfer is done by using indicators. indicators are measurable properties of one or more factors and are directly retrievable (measurable) for an object under study. as a factor can have many indicators, and each indicator can be relevant for multiple factors, the factors retrieve their values from their indicators by using an indicator aggregation. this is a pure technical construct to allow the shared usage of indicators by different factors. however, the aggregation functions (e.g., sum, average, median) are assigned to the factor to yield its value. . the indicator measurement is done on the basis of indicator types. this could be, e.g., likert, ordinal, or cardinal scales, counting or index values. the final meta model with its concepts can be used to develop the initial-staged mm. the defined elements (concepts) and their relationships are regarded as best practices in mm development and are practically needed concepts in order to build a functioning staged mm. although one could decide for several variations, the already discovered concepts in the meta model are the sum of (extracted) expert knowledge from previous mms. thus, the non-existence of any part of the meta model in an actual mm is neither considered as an issue in the meta model nor is the mm regarded as incorrect. rather, the interpretation is that the actual mm lacks a feature that is typically used by other mms, and it could potentially be improved with such. further, the relatively small amount of meta concepts is not necessarily a drawback, since the meta model only contains the most powerful concepts found. despite the few concepts, it provides a strong, defined, and universal framework for a whole class of staged mms. table summarizes an analysis of currently existing and applicable staged mms regarding the mapping between the developed meta model and the concepts used. the selected mms represent a sample from different domains, years, and development approaches. the bpmm and cmmi are representatives of the most famous mms developed [ , ] . leyh et al. [ ] , gökalp et al. [ ] , schumacher et al. [ ] , and luftman [ ] are representatives of scientific mm approaches from different years. property of the organization, which represents the object/area/process of investigation; used by one or more maturity level and can be related to a dimension within an organization factor specification technical and foundational requirement construct for determining the maturity level; acts as the maturity level's individual expression of a factor (needed due to multi-usage of factors) indicator measurable property of one or more factors within an organization; uses an indicator type for measurement indicator type measuring method for determining of the value of the indicator maturity level rank of the organizational maturity that results from factor evaluation by using the factor specification (see .a) or aggregation of a dimension's related factors (see .b); subdivides the maturity model and represents a relative degree of organizational ability/maturity from the analysis of existing mms, we conclude that very few models have indepth explanations of all the concepts and relationships used. however, almost every examined model fulfills common concepts like the mm's name, the domain in which it is located, as well as a description of its respective maturity levels (except one). the concepts' names "factor" and "indicator" are not used within the models. although [ ] use the term "indicators", the definition matches the m-concept of a factor. the m-concept of "indicator" is not further described in their model. in general, the lowlevel concepts (like, e.g., indicator) are rarely specified, possibly due to abstraction from the actual calculation of a maturity level. when it comes to the interpretation of the examined concepts and relationships of factors, indicators, and requirements, no mm can be regarded as an exact instantiation of our meta model, as not every concept is present. there is often a lack of information in the available publications about factors and indicators, which are subject to further calculations, or information on which requirements can be used to derive the maturity level. the mere mentioning of concepts allows conclusions to be drawn about the fact that the authors have basically dealt with the respective concepts, but the relationships between them remain undefined. possible, general reasons for this phenomenon could be the length of the publications in which the mms are presented. conference papers are often too short to describe a comprehensive presentation of all underlying assumptions. another reason could be that papers do not describe mms for application purposes, but rather focus on their contribution. however, we do not consider existing staged mms that do not match the meta model as incorrect. the meta model, as a conceptual orientation, could help other researchers or developers to understand and interpret the respective intention and improves the developer's cycle of mm development. we claim that this will help these models to overcome their ivory tower. the rather few but strong concepts defined in the meta model are a core skeleton of staged mms that may lead both the development as well as the understanding of mms. although we have built our meta model on a broad analysis of existing models, our research has limitations. thus, our meta model approach is only valid for staged models and therefore cannot explain other types of mms (continuous or focus area mms). in addition, the meta model initially only considers the development cycle and must be further developed for the application cycle. however, as [ ] proposed, both cycles should not be analyzed concurrently anyway, as they differ in their requirements. to this point, we have been able to show with the current approach that the consistency and concepts of existing mms are often not entirely described in the development cycle. it will be the focus of our future work to concentrate on the application cycle and to compare concepts of assessment with provided concepts in existing mms. in this paper, a meta model for mms was introduced, which can be used for a standardized and consistent development of mms. the related concepts, elements, and their relationships were explained and specified in detail. this was done by the analysis of relevant literature that introduces staged mms and by extracting their core concepts, including their syntax and semantics. further, the final m was evaluated against several mms from the research literature showing that the majority of mms lack in the exact specification of their elements and relationships. this uncertainty and divergence in mm specifications often lead to inconsistent applications and implications derived from their application. the presented m is therefore beneficial regarding consistent mm development and comparison. the m, together with its defined concepts and relationships, is a compact but powerful tool for mm developers for initial development and evaluation of their work, as well as to be consistent with related work and the staged mm semantics in general. however, the m only covers the structural part of an instantiated mm. consideration of the assessment part as an integrated aspect of the m would be valuable for the mm user. we therefore intend to develop the assessment aspect in the next research step. also, we do not claim completeness regarding the concepts in the m since we only introduced the concepts that are used by many different mms. additional concepts could be introduced; however, we strived for simplicity and usability instead of a complex and over specified meta model. the m is only applicable for staged mms; thus, a meta model for other mm types (e.g., continuous mms) has to be constructed separately. a role-based maturity model for digital relevance the oxford english dictionary. clarendon matters of (meta-) modeling the maturity of maturity model research: a systematic mapping study operational excellence driven by process maturity reviews: a case study of the abb corporation business process maturity models: a systematic literature review what makes a useful maturity model? a framework of general design principles for maturity models and its demonstration in business process management developing maturity models for it management understanding the main phases of developing a maturity assessment model the use of maturity models/grids as a tool in assessing product development capability prozessverbesserung mit reifegradmodellen quality is free: the art of making quality certain software engineering institute: cmmi® for development maturity assessment models: a design science research approach maturity models development in is research: a literature review knowing "what" to do is not enough: turning knowledge into action inductive design of maturity models: applying the rasch algorithm for design science research project management maturity models: the silver bullets of competitive advantage? kompetenz-und reifegradmodelle für das projektmanagement: grundlagen maturity models in the age of industry . -do the available models correspond to the needs of business practice? object management group: mda guide version object management group: uml infrastructure specification metamodellierung als instrument des methodenvergleichs: eine evaluierung am beispiel objektorientierter analysemethoden. shaker business process maturity model (bpmm) the application of the maturity model simmi . in selected enterprises development of an assessment model for a maturity model for assessing industry . readiness and maturity of manufacturing enterprises assessing business-it alignment maturity key: cord- -l b n ej authors: young, colin r.; welsh, c. jane title: animal models of multiple sclerosis date: journal: sourcebook of models for biomedical research doi: . / - - - - _ sha: doc_id: cord_uid: l b n ej to determine whether an immunological or pharmaceutical product has potential for therapy in treating multiple sclerosis (ms), detailed animal models are required. to date many animal models for human ms have been described in mice, rats, rabbits, guinea pigs, marmosets, and rhesus monkeys. the most comprehensive studies have involved murine experimental allergic (or autoimmune) encephalomyelitis (eae), semliki forest virus (sfv), mouse hepatitis virus (mhv), and theiler’s murine encephalomyelitis virus (tmev). here, we describe in detail multispecies animal models of human ms, namely eae, sfv, mhv, and tmev, in addition to chemically induced demyelination. the validity and applicability of each of these models are critically evaluated. multiple sclerosis (ms) affects about , people in the united states and is a major cause of nervous system disability in adults between the ages of and years. the symptoms are diverse, ranging from tremor, nystagmus, paralysis, and disturbances in speech and vision. extensive demyelination is seen in the neuronal lesions. the clinical heterogeneity of ms, as well as the finding of different pathological patterns, suggests that ms may be a spectrum of diseases that may represent different pathological processes. this has led to the development of many different animal models, including rodents and nonhuman primates, that reflect the pathological processes and could allow for the development of therapeutic approaches. at the present time, the exact etiological mechanism in humans is not clear; however, several animal models are available providing insight into disease processes. the relative inaccessibility and sensitivity of the central nervous system (cns) in humans preclude studies on disease pathogenesis, and so much of our understanding of infections and immune responses has been derived from experimental animal models. the experimental systems include theiler's virus, mouse hepatitis virus, and semliki forest virus infections of laboratory rodents. additional information has been obtained from studies of experimental infections of other animals that result in demyelination, notably maedi-visna virus in sheep and canine distemper virus in dogs. in humans and animals, most natural cases of demyelinating disease are rare complications of viral infections. one possible reason for the low incidence of demyelination following viral infections could be the low efficiency of neuroinvasion. however, a correlation between cns infection and clinical disease is difficult to determine. the role of genetics and environmental factors in ms is complex. factors such as geographical location, ethnic background, and clustering in temperate climates all contribute to susceptibility. individuals with a north european heritage are statistically more susceptible to ms than those from a more tropical environment and it is more common in women. epidemiological data indicate that ms is not a single-gene disorder and that additionally environmental factors contribute to the disease. data from genetic studies indicate that although mhc genes clearly contribute to disease susceptibility and/or resistance, it is probable that a combination of environmental factors may additionally contribute to disease development in genetically predisposed individuals. to understand the initiating factors and progression of ms, researchers have turned to experimental model systems. since this disease cannot be recreated in a tissue culture system, much effort has been directed to the use of laboratory animals. those animal models should mirror the clinical and pathological fi ndings observed in human ms. ideally, the animal model should be in a species that is easy to handle, inexpensive, can be kept in large numbers, and is easily bred in laboratory conditions. the most frequently used animals are laboratory rodents, including mice, rats, guinea pigs, and hamsters. one of the most useful aspects of laboratory rodents as animal models of disease is the vast array of inbred strains of the species available, most notably in experimental mice. additionally, very valuable information has been obtained from studies using larger animals including sheep, dogs, cats, and nonhuman primates. models of ms fall into two main groups: viral and nonviral. viral models are immensely relevant since epidemiological studies suggest an environmental factor, and almost all naturally occurring cns demyelinating diseases of humans and animals of known etiology are caused by a virus. these include in humans, subacute sclerosing panencephalitis (sspe)-caused by measles or rubella viruses, progressive multifocal leukoencephalopathy (pml)-caused by jc virus, and human t lymphotrophic virus- (htlv- )-associated myelopathy (ham)-caused by htlv- ; in animals, these include visna virus in sheep and canine distemper in dogs. however, no one virus has consistently been associated with human ms, although it is likely that more than one virus could trigger the disease. of the nonviral models of ms, experimental allergic encephalomyelitis (eae) is the most widely studied. eae is characterized by inflammatory infiltrates in the cns that can be associated with demyelinating lesions. in eae, the disease is initiated by the extraneural injection of cns material, or purified myelin components, emulsified in an adjuvant, the most commonly employed one being complete freund's adjuvant containing mycobacterium tuberculosis h ra. however, no naturally occurring autoimmune correlate of this experimental disease is known, although it is extensively researched as a model of ms, with the reasoning that ms may be such a disease. the most widely studied models of ms are the experimental infections of rodents resulting in an inflammatory demyelinating disease in the cns, such as theiler's virus, mouse hepatitis virus, and semliki forest virus. each of these infections gives rise to lesions of mononuclear cell inflammatory demyelination throughout the brain and spinal cord but not in the peripheral nervous system. as such, this histopathology correlates with human ms, although it does not preclude the fact that the viruses could gain access to the cns via the peripheral nervous system. these viral models demonstrate how a virus can easily reproduce cns disease, which is comparatively rare in humans, and how this can be influenced by many factors including both genetic and immunological. experimental studies in induced animal models have the advantage over studies in spontaneous models in that the onset and progression of the disease can be controlled. although it has been proposed that some autoimmune diseases may have a viral etiology, virus-induced autoimmunity is a controversial subject. epidemiological studies of ms provide strong evidence for the involvement of a viral etiology in the onset of disease. theiler's virus-induced demyelination, a model for human ms, bears several similarities to the human disease: an immune-mediated demyelination, involvement of cd + helper t cells and cd + cytotoxic t cells, delayed type hypersensitivity responses to viral antigens and autoantigens, and pathology. indeed this mouse model may provide a scenario that closely resembles chronic progressive ms. theiler's murine encephalomyelitis virus (tmev) is a picornavirus that causes an asymptomatic gastrointestinal infection, followed by occasional paralysis. there are two main strains of tmev, the virulent strains and persistent theiler's original (to) strains. the virulent gdvii strains of theiler's virus are highly neurovirulent and when injected intracranially, cause death by encephalitis within h. gdvii strains also cause differing forms of paralysis depending on the route of inoculation (see table - ). from these studies it appears that the gdvii virus may gain access to the cns by retrograde axonal transport rather than by a hematogeneous route. infection of susceptible strains of mice with the persistent to strains bean, da, ww, or yale results in a primary demyelinating disease that closely resembles human ms. infection of resistant strains of mice with bean does not result in demyelinating disease, since these mice are able to clear virus from the cns. susceptible mice fail to clear virus from the cns, possibly resulting from poor natural killer (nk) cell and cytotoxic t lymphocyte (ctl) responses. persistent viral infection of the cns is required for demyelination. following the intracranial injection of susceptible mice with bean, virus replicates both in the brain and spinal cord. one month postinfection, viral titers decrease, and high levels of neutralizing antibodies are detected (figure - ) . at this point in the disease, neurons may become infected with virus and mice may develop a nonprogressive flaccid paralysis of the forelimbs and/or hindlimbs. this is sometimes referred to as a polio-like disease, but this is confusing since flaccid paralysis in mice infected with poliovirus is progressive and normally results in death. in the late phase of the disease, astrocytes, oligodendrocytes, and macrophage/microglial cells become infected with virus. also in the demyelinating disease there is both b and t cell autoimmunity, directed against myelin and its antigenic components. genetics of persistent infection and demye-linating disease all inbred mouse strains inoculated intracerebrally with tmev show early encephalomyelitis, but not all strains remain persistently infected. resistant strains normally this trait is under multigenic control, with h- mhc class i genes being the most prominent. additionally, several non-h- quantitative trait loci (qtl) have been identified within the same h- haplotypes that control persistence. there is generally a good correlation in inbred strains between susceptibility to three phenotypes (viral load, pathology, and symptoms), suggesting that variations in both demyelination and clinical disease may result from how each mouse strain can control the viral load during the persistent infection. using b congeneic and recombinant strains of mice, susceptibility to disease has been mapped to the h- d region. furthermore, resistant haplotypes are dominant and the same locus controls viral load during persistence and demyelination. currently, non-h- susceptibility loci have been identified as having an effect on susceptibility to theiler's virus-induced disease (tvid) (see table - ). the mechanism(s) of tvid may be different for different mouse strains, but most of the information has come from studies of sjl/j mice infected with the da or bean strain of virus. the virus infects oligodendrocytes, and the resulting demyelinating disease could be due, in part, to the virus killing oligodendrocytes directly or by the virus-specific cd + ctls present in the lesions. a series of experiments has demonstrated that demyelination correlates with the presence of a cd + t cell-mediated response against viral epitopes. these cells secrete cytokines such as interferon (ifn)-γ that activate both microglial cells and invading monocytes, which subsequently secrete factors such as tumor necrosis factor (tnf)-α and thus can cause "bystander" demyelination. activated macrophages ingest and degrade damaged myelin. autoantibodies , and myelin-specific cd + t cells have been shown in sjl/j mice several months after intracranial inoculation. , epitope spreading in these mice commences with recognition of a proteolipid protein (plp) epitope, and then progresses to additional plp epitopes and then to myeloid basic protein (mbp) epitopes. a direct demonstration that disease can be maintained on a purely autoimmune footing, after infection has been eradicated, has not been shown. immunity and theiler's virus the first response to viral infection is the production of type i interferons, which are critical for viral clearance. ifn-α/β receptor knockout mice injected with tmev die of encephalomyelitis within days of infection. nk cells are activated early in infection with certain viruses. in tmev infection susceptible sjl mice have a % lower nk cell activity in comparison to the highly resistant c bl/ mice. this low nk activity in sjl mice is in part due to a defect in the thymus impairing the responsiveness of nk cells to stimulation by ifn-β. the pivotal role of nk cells in early tmev clearance is demonstrated by the finding that resistant mice depleted of nk cells by monoclonal antibodies to nk . develop severe signs of gray matter disease. in the early disease, both cd + and cd + t cells have been shown to be important in viral clearance. in early disease cd + t cells are required for b cells to produce antibodies for viral clearance. these cd + t cells secrete ifn-γ, which in vitro inhibits tmev replication and has a protective role in vivo. cd + t cells are also important in viral clearance, as demonstrated by the finding that cd + t cell-depleted mice fail to clear virus and develop a more severe demyelinating disease. cd + t cells also provide protection against tvid when adoptively transferred to a tvid-susceptible strain, balb/c.anncr. thus, cd + t cells are implicated in viral clearance and resistance to demyelination. higher ctl activity has been demonstrated in tvid-resistant c bl/ mice as compared to resistant sjl/j mice. these ctls may play an important role in viral disease since they may recognize viral determinants and/or they may inhibit delayed type hypersensitivity responses. the relative roles of th /th cells in tvid are very complex, and a simple picture of a th or th polarization during infection may not be apparent. a pathogenic role for th cells during late demyelinating disease is demonstrated by the finding that both tvid correlates with delayed type hypersensitivity responses to tmev and that the depletion of cd + t cells during late disease results in the amelioration of clinical signs. high levels of the proinfl ammatory th cytokines ifn-γ and tnf-α in late disease correlate well with maximal disease activity. evidence demonstrating the protective role of th in tvid has been shown in experiments in which skewing the immune response toward th immunity in tmev infection diminishes the later demyelinating disease. however, other studies have shown that the th /th balance did not explain the difference in susceptibility to tvid. th cytokines are generally pathogenic during late demyelinating disease, whereas th cytokines are protective. and is characterized by abnormally thin myelin sheaths in relation to axon diameter. to date, however, there are few reliable data on the frequency of remyelination in ms patients. stimulation of remyelination is a potential treatment for ms. the tmev model of ms can be used to study remyelination using remyelinationpromoting antibodies. in this remyelination model, sjl/j mice, aged - weeks, are injected with a µl volume containing , pfu of daniel strain intracerebrally. all animals develop mild encephalitis, which resolves within days after the injection. the infected mice then develop the chronic demyelinating disease that gradually progresses over several months. to study remyelination, mice that had been infected with tmev for months receive a single intraperitoneal injection of . mg (∼ . g/kg body weight) of a recombinant remyelinating antibody (rhigm ) in phosphate buffered saline (pbs). in one study, . % of lesions in animals treated with rhigm showed retraction of varying degrees, presumably the effect of remyelination in these lesions. the direct binding of rhigm to demyelinated lesions is consistent with the hypothesis that these antibodies work directly in the lesions, probably by binding to the cns glia to induce remyelination. thus, this murine tmev model can also be used as a model with which to examine different modes of remyelination. mouse hepatitis virus (mhv) is a member of the coronaviridae, a group of large positive sense enveloped rna viruses. depending on the strain of virus used, mhv causes a variety of diseases including enteritis, hepatitis, and demyelinating encephalomyelitis. infection of mice with the neutrotropic jhm strain of mhv causes encephalitis, followed by chronic demyelination. virus is not cleared from the cns, resulting in a persistent infection. after intracerebral or intranasal infection with mhv, virus enters the brain and causes encephalitis. intranasal infection with mhv-jhm or -a leads to viral spread through the olfactory bulb and along the olfactory tracts, as well as along the trigeminal nerve to the mesencephalic nucleus. up to days postinfection (pi) early viral spread is via specific neural pathways and neural connections. viral titers peak at about day pi in the brain and later in the spinal cord and virus is cleared by days - pi. however, viral antigen is still detectable up to day pi. additionally, viral rna is detectable in the brain as late at - months postinfection, although the amount of rna decreases with time. liver infection can occur after any route of infection (in, ic, ig, or ip), with viral titers peaking at day pi and hepatitis developing during the first - weeks. cns demyelination develops as active mhv infection resolves. the lesions observed are histologically very similar to those observed in ms patients. these mhv lesions are characterized by primary demyelination accompanied by naked axons, and are found scattered throughout the spinal cord. the peripheral nervous system is not affected. chronic lesions are associated with lipid-laden macrophages, scattered lymphocytes, and perivascular cuffing. these chronic lesions can persist as late as day pi, and demyelinating axons can be seen as late as months postinfection. chronic diseases in mhv-infected mice are associated with ataxia, hindlimb paresis, and paralysis, followed by a recovery. this animal recovery is mediated by cns remyelination, beginning anywhere from to days pi. c bl/ mice (h- b ) are susceptible to mhv infection. in this murine model adult mice (of weight - g) are anesthetized by inhalant anesthesia and receive an intracerebral injection of approximately pfu of a neurotropic mhv strain in a volume of approximately µl of pbs. this intracerebral injection of mhv results in a biphasic disease: an acute encephalomyelitis with myelin loss, followed - days later by an immunemediated demyelinating encephalomyelitis with progressive destruction of the cns. there is an - % survival rate of mice injected with this mhv, with animals usually succumbing during the first weeks of acute infection. animals surviving this acute stage show a % chance of survival. control animals injected intracerebrally with sterile pbs show no clinical signs or histological defects. electron micrographs of demyelinating lesions show that macrophage processes slip between layers in the myelin sheath, implying that macrophages could indeed be mediating demyelination. the appearance of macrophages within the cns also correlates with the development of lesions. additionally, they do not appear in large numbers in the absence of lymphocytes, so it is possible that myelin damage is caused by a nonmacrophagedependent mechanism and that macrophages may only clear up the damaged myelin. in contrast to other mouse models of demyelination, there does not appear to be a clear role for any single lymphocytic or monocytic subset mediating the demyelination. rather, it appears that a balance of immune components may be necessary for viral clearance and that various pathways, both immune and nonimmune, may cause the ensuing demyelinating events. recently, progress has been made in further identifying the immune cells required for demyelination. experimental infection of severe combined immunodeficiency (scid) mice, lacking t cells, results in fulminate encephalitis without demyelination. adoptive transfer of splenocytes from syngeneic immunocompetent mice into infected scid mice results in demyelination within - days posttransfer. additional experiments indicated that either cd + or cd + t cell subsets are capable of initiating this process. however, mice that receive splenocytes depleted of cd + t cells survive longer and develop more demyelination than mice receiving splenocytes depleted of cd + t cells. thus, experimental scid mice demonstrate that the roles of each t cell subset in demyelinating diseases are not equal. ifn-γ is a critical mediator of homeostasis and infl ammation in ms and many of its rodent models. bone marrow chimera mice have been used to address the role of ifn-γ in bystander demyelination mediated by cd + t cells. these chimeras as rodent models for jhm have addressed a hypothesis that ifn-γ produced by cd + t cells, and not from other sources, was the critical component in mediating bystander demyelination. this chimeric approach did not compromise ifn-γ production by cells such as nk cells and dendritic cells, thus preserving the innate immune response to the virus. the results demonstrated that ifn-γ produced by these innate cells was unable to initiate the demyelinating disease, even in the context of activated cd + t cells lacking only the ability of produce ifn-γ. these findings highlight the role that cd + t cells have in demyelination in jmh-infected mice. it has been demonstrated that ifn-γ is critical in other animal models of demyelination and in ms. semliki forest virus (sfv) is an alphavirus of the togaviridae. the virus has been isolated from mosquitoes, but the natural host is unknown. sfv is a single-stranded positive strand rna virus that has been cloned and sequenced. the most commonly studied strains used in adult mice are the virulent l strain and the avirulent a ( ) strain. both of these strains are avirulent in neonatal and suckling mice by all routes of infection. experimental infection of mice with sfv is widely used as a model to study the mechanism of virus-induced cns disease. sfv has the advantage of being neuroinvasive as well as neurotropic, thus allowing studies of viral entry into the cns and the integrity of the bloodbrain barrier (bbb). following intraperitoneal injection with pfu sfv in . ml pbs containing . % bovine serum albumin, all strains replicate in muscles and other tissues, resulting in a plasma viremia. virus then crosses the cerebral vascular endothelial cells, resulting in infection of neurons and oligodendrocytes. in neonatal or adult mice, infection with virulent strains results in widespread infection that is lethal within a few days. in contrast, infection of mice with the a ( ) strain results in a cns infection, and infectious virus is cleared from the brain by day . infiltrating mononuclear cells are observed days pi and peak at about day . focal lesions of demyelination throughout the cns are observed days pi and peak between and days pi. sfv-induced demyelinating diseases have been widely studied following intraperitoneal injection of adult mice with the a ( ) strain of the virus. following intraperitoneal injection, virus is detected in the brain by h. viral titers then rise, but rapidly decline following initiation of the immune response. interestingly, although infectious virus can be detected only up to day pi, realtime polymerase chain reaction (rt-pcr) studies detect viral rna up to day pi. thus, it is possible that there is persistence of viral antigen(s). disturbance of the bbb occurs between and days pi, which corresponds to the increase in inflammatory cell infi ltration and reduction in viral titer and which may be related to the influx of cells or cytokine-mediated effects. the presence of macrophages, activated microglia, and the proinfl ammatory cytokines tnf-α, ifn-γ, interleukin (il)- α, il- , il- , and granulocyte-macrophage colony-stimulating factor (gm-csf) during sfv-induced demyelination, in addition to enhancing the infl ammatory response, may also play a role in controlling viral infection since il- , ifn-γ, and tnf have direct antiviral activity. additionally, ifn-γ and tnf production peripherally coincides with sfv-induced encephalitis in sjl and b mice. interestingly, these same cytokines predominate in ms lesions. an intense inflammatory response characterized by perivascular cuffi ng is apparent histologically from days. demyelination, as demonstrated using luxol fast blue staining of sections, is apparent by days. however, small focal lesions of demyelination can be observed using electron microscopy by day . a striking feature of sfv infection appears in the optic nerve, where there are demyelinating lesions and changes in visually evoked responses and axonal transport. this optic neuritis also occurs in human ms. it appears that sfv-induced demyelination in this mouse model is accompanied by neurophysiologically demonstrable visual deficits very similar to those found in ms patients. thus, this may provide a very useful animal model for research into ms. the advantages of this model are that genetic and environmental factors can be readily controlled, while the low cost and fast reproductive rate make experimental design considerably easier. no demyelination is observed following sfv infection of scid mice or athymic mice. in the absence of specific immune responses, scid mice infected with sfv a ( ) have a persistent viremia, a persistent and restricted cns infection, and no lesions of demyelination. comparison of the infection to that in nu/nu and balb/c mice and studies on the transfer of immune sera show that immunoglobulin m (igm) antibodies clear the viremia but not the brain virus and that infections of brain virus can be reduced by igg antibodies. these igg antibodies can abolish infectivity titers in the brain but cannot remove all viral rna. adoptive cell transfer studies and administration of anti-cd antibodies demonstrate that demyelination following sfv infection is dependent on cd + t cells. this is consistent with the finding that the cns inflammatory infiltrate is dominated by cd + t cells. this fi nding is analogous to that in ms and is in contrast to that in eae, where cd + cells predominate. in the eae autoimmune model of ms, studies suggest that a th cytokine profile predominates. another point of difference between the eae model and the sfv model is shown in the th /th profiles. following infection with sfv, th and th cytokines were detected in the cns and both were present throughout the time course studied, indicating that there was no bias of th response in the cns, nor were changes apparent with time. the experimental disease eae has been investigated in many strains of animals including mice, rats, guinea pigs, rabbits, marmosets, and rhesus monkeys. eae is an autoimmune infl ammatory disease of the cns and is characterized by perivascular and subpial inflammatory infiltrates and demyelinating lesions. the disease is usually initiated by injection of autoantigens emulsifi ed in an adjuvant. the progression and pathology of lesions observed depend on the type of antigen used in the injection, the method of injection, and the strain of animal used. because of its very nature, eae as a model of ms does not address certain pertinent questions relating to ms, such as age-related onset of disease or epidemiology. a major difference between eae and viral models of ms is that in eae the inflammatory response is directed to autoantigens. a feature of the eae model is that the course of the disease can be relapsing and remitting. studies of eae have been used to identify antigenic determinants on components of myelin. using bioinformatic technology these determinants have been used to search available databases of viral and bacterial proteins. results indicate numerous viral and bacterial protein segments with probable sequence similarity to myelin basic protein determinants. experimental allergic encephalomyelitis in rabbits eae has been induced in rabbits by footpad inoculation with rabbit spinal cord homogenate, resulting in hindlimb paresis or paralysis. rabbits with -day paraplegia showed increased spinal cord incorporation of radioactive drugs administered in the epidural space. thus, this demyelinating disease process may expose the spinal cord to larger amounts of sub-stances administered neuraxially. it is therefore possible that this rabbit model could be used to investigate the incorporation of radioactive therapeutic drugs in the epidural space. experimental allergic encephalomyelitis in guinea pigs guinea pigs have also been investigated to determine whether they may serve as useful eae models of ms. the interest in guinea pigs stems from the fact that group cd glycoprotein homologues, which in humans present foreign and self lipid and glycolipid antigens to t cells, are not found in mice and rats but are present in guinea pigs. in this guinea pig model, animals have been sensitized for eae, and cd and mhc class ii expression has been measured in the cns. in normal guinea pigs low level mhc class ii occurred on meningeal macrophages and microglial cells, whereas immunoreactivity for cd was absent. in the eae cns, however, the majority of infiltrating cells were mhc ii + and microglia showed increased expression, whereas cd immunoreactivity was detected on astrocytes, b cells, and macrophages. minimal cd and mhc ii coexpression was detected on inflammatory cells or glia. thus, in this guinea pig eae model group cd molecules are upregulated in the cns on subsets of cells distinct from the majority of mhc iibearing cells. this expression of cd proteins in such eae lesions broadens the potential repertoire of antigens recognized at these sites and highlights the value of this guinea pig model of human ms. experimental allergic encephalomyelitis in rats rats were injected with spinal cord homogenate or the encephalitogen; myelin basic protein induced eae in genetically susceptible dark agouti (da) rats but not in albino oxford (ao) rats. here -to -week-old rats were immunized in either or both hind footpads with . ml antigenic emulsion containing µg rat spinal cord tissue in complete freund's adjuvant (cfa). rats are monitored from day after inoculation and the severity of disease was assessed by grading tail, hindlimb, and forelimb weakness, each on a sale of (no disease), (loss of tail tonicity), (hindlimb weakness), (hindlimb paralysis), to (moribund or dead). clinical disease in susceptible strains of rats is apparent in all animals, and the onset of disease occurs at day postinjection. at the peak of the clinical manifestation of eae there is a marked increase in the level of infiltration of cells accompanied by a lack of activation in susceptible da rats, whereas it remains elevated in resistant ao rats. at the peak of clinical disease da rat spinal cords contain high levels of cd + t cells. da rats also contained times as many live cd + t cells as ao rats. astrocytosis, as an indication of cns reaction to the presence of infl ammatory cells, was clearly observed in both rat species. microglial activation persists in resistant ao rats, whereas activation is downregulated in da rats. in this model it is speculated that at the peak of disease, infiltrating monocytes and macrophages are the main antigen-presenting and effector cells. rat eae may also be induced by the injection of xenogeneic myelin. for example, -to -week-old lewis rats injected in both hind footpads with an emulsion containing µg of guinea pig myelin basic protein and cfa develop acute eae. also chronic relapsing eae (cr-eae) may also be induced in this rat model using a regimen of intraperitoneal injections of mg/kg of cyclosporin a. pathology studies indicate that in acute and cr-eae, mcp- and its receptor ccr are significantly upregulated throughout the course of cr-eae and that a large number of macrophages infiltrated the cr-eae lesion. this suggests that macrophages recruited by mcp- and ccr -expressing cns cells are responsible for the development and relapse of eae. thus, in this rat model, in addition to t cells, macrophages are another target for immunotherapy studies for neurological autoimmune diseases. a more recent development of a rat eae model involves using human mbp as antigen. here eae was induced by the immunization of female wistar rats with human mbp. it was found that most of the rats developed tail tone loss and hindlimb paralysis together with demyelination, infiltrative lymphocyte foci, and "neurophagia" in the cortex of cerebra and in the white matter of the spinal cord. this study further demonstrated that this rat model of eae induced by human mbp resembles many features of human ms and may promise to be a better animal model for the study of ms. the use of cfa is not a prerequisite for the development of rat eae. for example, eae can be induced in -to -week-old da rats by a single hind footpad injection of an encephalitogenic emulsion consisting of rat or guinea pig spinal cord homogenate (sch) in pbs. the reason for not wanting to use cfa is that in itself it induces a strong inflammatory response and exerts numerous immunomodulatory properties. additionally, cfa induces a strong anti-purified protein derivative (ppd) response and may induce adjuvant arthritis, another autoimmune disease. the susceptibility of da rats to eae induction with sch depends upon the origin of the cns tissue, the homologous tissue being the more efficient encephalitogen. da rats that recovered from eae that had been induced with homologous sch without adjuvant and then immunized with the encephalitogenic emulsion containing cfa developed clinical signs of the disease. neurological signs in rechallenged rats were milder, but first signs appeared earlier. the earlier onset of eae observed in da rats after rechallenge has been attributed to the reactivation of memory cells. taken together, these experiments demonstrate that eae can be effi ciently and reproducibly induced in da rats without the use of cfa. this experimental model for understanding the basic mechanisms involved in autoimmunity within the cns, without the limitations and inherent problems imposed by the application of adjuvants, may represent one of the most reliable rodent models of ms. the rat as an experimental model could be used to evaluate new immunotherapies of eae. these include antigen-induced mucosal tolerance, treatment with cytokines, and dendritic cellbased immunotherapy. the ideal treatment of diseases with an autoimmune background such as ms should specifically eliminate autoreactive t cells without affecting the integrity of the immune system. one way to achieve this would be to induce immunological tolerance to autoantigens by the oral or nasal administration of autoantigen. several studies have shown that nasal administration of soluble antigens results in peripheral tolerance by immune deviation or the induction of other regulatory mechanisms. in the rat model this tolerance has been investigated using synthetic peptides of mbp, mbp - , - , and - . nasal administration of the encephalitogenic mbp - or - suppresses eae. mbp - and - given together had synergistic effects in suppressing eae and reversed ongoing eae. a problem, however, of antigen-specific therapy by the nasal route is that one antigen, or peptide, may be effective in inducing tolerance in one strain of animal but not in another. one way of treating ongoing eae may be the use of an altered peptide ligand with high tolerogenic efficacy when administered nasally. cytokines have been widely used in disease prevention and treatment. cytokine immunotherapy in ms could employ one or two basic strategies: first, to administer immune response downregulatory cytokines, or second, to administer inhibitors of proinfl ammatory cytokines. the nasal route of administering these cytokines has been studied in the rat eae model. nasal administration of low doses of antiinflammatory or regulatory cytokines such as il- , il- , or tumor growth factor (tgf)-β inhibits development of rat eae when given before or on the day of immunization, but by differing mechanisms. nasally administered il- reduced both peripheral immune responses and microglia activation in the cns, whereas nasal administration of il- or tgf-β triggered the activation of dendritic cells (dcs). however, nasal administration of cytokines alone fails to treat ongoing lewis rat eae. interestingly, nasal administration of mbp - + il- or mbp - + il- suppresses ongoing eae in lewis rats. the suppression of eae by mbp - + il- is associated with the induction of a broad lymphocyte hyporesponsiveness. although this combined administration of autoantigen plus cytokine may be effective in treating rat eae, the applicability of this to human ms is severely limited by the lack of knowledge of the pathologically relevant autoantigen(s) in ms. dcs not only activate lymphocytes, but also induce t cell tolerance to antigens. use of tolerogenic dcs is thus a possible immunotherapeutic strategy for treatment of eae, and indeed this has been studied in some detail. however, mbp - -pulsed dcs only prevented the development of eae and failed to treat ongoing eae in lewis rats. in an attempt to treat ongoing eae, splenic dcs have been isolated from healthy lewis rats and modifi ed in vitro with cytokines ifn-β, il- , il- , or tgf-β . upon subcutaneous injection into lewis rats on day pi with mbp - + fca, ifn-β or tgf-β -modifi ed dcs promoted immune protection from eae. the common marmoset callithrix jacchus is an outbred species characterized by a naturally occurring bone marrow chimerism. the marmoset is a primate phylogenetically close to humans, and has been studied as an animal model for ms. eae can be induced in the common marmoset by the injection of human brain white matter, dispersed in demineralized water to a concentration of mg/ml and emulsified with cfa containing . mg/ml of mycobacterium butyricum h a. monkeys are injected intracutaneously with µl of emulsion into the dorsal skin at several locations. clinical disease in this model is scored daily on a scale from to : = no clinical signs; . = apathy, loss of appetite, and an altered walking pattern without ataxia; = lethargy and/or anorexia; = ataxia; . = paraparesis or monoparesis and/or sensory loss and/or brainstem syndrome; = paraplegia or hemiplegia; = quadriplegia; and = spontaneous death attributable to eae. here the onset of disease, as measured by clinical scores, is variable among animals between and weeks postinoculation. additionally, the maximal clinical scores are variable among animals and range between and . on histopathological examination, large plaques of demyelination are observed in the white matter of cerebral hemispheres, mainly localized around the wall of lateral ventricles, in the hemispheric white matter, corpus callosum, optic nerves, and optic tracts. the demyelinated areas show a moderate or severe degree of inflammation characterized by perivascular cuffs of mononuclear cells. in the spinal cord, widespread demyelination is also observed. areas of demyelination involve the ventral, lateral, and dorsal columns of the spinal cord, especially in the outer part of the spinal tracts. thus, pathology in the marmoset model is characterized by infl ammation, demyelination, and astrogliosis. interestingly, this model demonstrates the presence of axonal damage in demyelinating plaques. indeed, axonal damage and loss are well-known events in ms. in ms, axonal damage appears to be an early event, related to an acute inflammation. in the marmoset eae, axonal damage also occurs in areas of acute and early inflammation and demyelination. this eae in c. jacchus is of special interest because of the resemblance of this model to the human disease, and the similarity between the immune systems of marmosets and humans. the type of clinical signs of eae in marmosets depends largely on the antigens used for disease induction. sensitization of marmosets to human myelin induces a relapsing-remitting, secondary-progressive disease course. lesions in this model represent all stages present in chronic ms. marmosets inoculated with mbp develop only mild inflammatory disease unless bordetella pertussis is used with the encephalitogen. cns demyelination critically depends on the presence of antibodies to myelin oligodendrocyte glycoprotein (mog), a minor cns component. marmosets sensitized to a chimeric protein of mbp and proteolipid protein (of myelin) develop clinical eae only after the autoimmune reaction has spread to mog. marmosets immunized with recombinant human mog - do not develop relapsing-remitting disease but only chronic-progressive disease. during the asymptomatic phase of this primary progressive-like disease, which can last from to weeks, brain lesions are detectable using magnetic resonance imaging (mri), but are not expressed clinically. the induction of eae with mbp or white matter tissue homogenate (wmh) has been well established in rhesus monkeys (macaca mulatta). the rhesus monkey was the first animal species in which eae was deliberately induced. that autoimmunity to brain antigens could induce paralytic disease was confirmed by studies in rhesus monkeys given repeated inoculations of brain homogenates. mog-induced eae has also been produced in this nonhuman primate species that is a highly relevant model for the human disease. the close similarity of the human and rhesus monkey immune system is illustrated by the high degree of similarity between the polymorphic mhc and t cell receptor genes between these two primates. to produce this mog-induced eae, monkeys are injected, under anesthesia, with a total of ml of : emulsion composed of µg mog in pbs and cfa at sites into the dorsal skin. overt clinical signs are scored daily according to the following criteria: ( ) no clinical signs; ( . ) loss of appetite, apathy, and altered walking; ( ) lethargy, anorexia, substantial reduction of the general condition, and loss of tail tonus; ( ) ataxia, tail biting, sensory loss, and/or blindness; ( . ) incomplete paralysis of one (hemiparesis) or two sides (paraparesis); ( ) complete paralysis of one (hemiplegia) or two sides (paraplegia); ( ) complete paralysis (quadriplegia); and ( ) death. the onset of clinical disease varies between animals, and occurs at days - after encephalitogenic challenge. all monkeys, however, develop clinical disease and all achieved a score of on clinical severity. the current available panel of nonhuman primate eae models may reflect the spectrum of inflammatory demyelinating diseases in the human population. these eae models can therefore be used to investigate pathogenic mechanisms and to develop more effective therapies. the most widely studied animal model of eae is that of the mouse. in common with other animal models of eae, disease induction varies depending on both the sex of the animals, the mouse strain used, as well as the origin of the spinal cord encephalitogen. in this model mice, aged - weeks, are immunized subcutaneously in four sites over the back with - µg of guinea pig mbp emulsified in equal volumes of cfa containing - µg heat-killed mycobacterium tuberculosis. mice also receive µg of pertussis toxin in . ml pbs intraperitoneally at the time of immunization and h later. mice are then scored daily for clinical signs of eae for at least days as follows: , no clinical signs; + , limp tail or waddling gait with tail tonicity; + , ataxia or waddling gait with tail limpness; + , partial hindlimb paralysis; + , total hindlimb paralysis; and , moribund/death. for each strain of mice there is variation in day of onset of disease, varying from day to day postinfection; incidence of disease, varying from % to % of animals; incidence of mortality, varying from % to % of animals; and mean clinical scores, varying from to . . many mouse strains have been employed in the study of eae, and while the sjl strain has been most frequently used to model gender differences in both disease onset and severity, the sjl model has some limitations due to its diminished cd + t cell repertoire. certain susceptible strains of mice, such as fvb mice, show a relapsing-remitting course of disease that bears some resemblance to ms. fvb mice therefore may serve as a mouse strain into which various transgenes may be introduced for the purpose of studying their influence on eae and for exploring new therapeutic approaches. since eae is a well-studied disease in mice, mimicking many clinical and pathological features of ms, including cns infl ammation and demyelination, it is of significance that it can also be used as an appropriate model to study ms-related pain. it has been clearly demonstrated in sjl that in both "active" and "passive" eae, there is an initial increase in tail withdrawal latency (hypoalgesia) that peaked several days prior to the peak in motor defi cits during the acute disease phase. during the chronic disease phase, tail withdrawal latencies decreased and were significantly faster than control latencies for up to days postimmunization. thus, it is possible to use both murine active and passive eae as models for ms-related pain. while specific immunotherapeutic strategies are effective in experimental model systems, translation to the human disease has genetically been poorly tolerated or has proved to be ineffective. this conflict may in part be due to the model systems used as well as the poor correlation of in vitro findings compared to those observed in vivo. in biozzi abh mice, which express the novel mhc class ii a, eae occurs following immunization with myelin proteins and peptide epitopes of these proteins; however, only plp peptide - , mog peptide - , or spinal cord homogenate reproducibly induces chronic relapsing eae (creae) with infl ammation and demyelination. creae provides a wellcharacterized reproducible system to develop therapeutic strate-gies during established relapsing autoimmune neurological disease and is pertinent to ms. in creae in abh mice, relapse and progression of disease are associated with emergence and broadening of the immune repertoire due to release of myelin antigens following myelin damage. thus, this creae model in biozzi abh mice is very well suited as a model with which to examine the effect of therapeutic strategies in a dynamic system. disease susceptibility in human ms is associated with three mhc class ii alleles in the hla-dr haplotype, drb * , drb * , and dqb * . an autoimmune pathogenesis has been hypothesized in which one or more of these mhc class ii molecules presents cns-derived self-antigens to autoaggressive cd + t cells, which infiltrate the cns initiating an infl ammatory response. however, the target autoantigens in ms are unknown. immunization of mice with myelin or other brainassociated proteins induces eae, a disease resembling ms both clinically and pathologically. the proteins, mbp, plp, and mog, components of the myelin sheath, are candidate antigens. indeed, t cells that are reactive to these antigens have been demonstrated in ms patients. mice expressing the human hla-dr (drb * ) molecule are capable of presenting peptides from all these three ms candidate autoantigens. it is possible in ms that while t cells responding to one of these antigens may initiate the disease, epitope spreading and the recruitment of t cells with additional specificities, as the disease progresses, could lead to infl ammatory responses to several proteins resulting in an escalation of the autoimmune response. transgenic mouse models of multiple sclerosis are now well established. the following are two examples of such transgenic models. first, ms is associated with hla class ii molecules hla-dr , -dr , and -dr . in humans it is difficult to analyze the individual roles of hla molecules in disease pathogenesis due to the heterogeneity of mhc genes, linkage disequilibrium, the influence of non-mhc genes, and the contribution of environmental factors. however, the specific roles of each of these class ii molecules can be addressed using transgenic models expressing these hla genes. this model could prove useful in deciphering the role of hla molecules and autoantigens in ms. second, while eae has been a valuable model for the immunopathogen-esis of ms, it has sometimes been difficult to reconcile the fi ndings and therapies in the rodent models and the cellular and molecular interactions that can be studied in human disease. humanized transgenic mice offer a means of achieving this, through the expression of disease-implicated hla class ii molecules, coexpressed with a cognate hla-class ii-restricted, myelin-specifi c t cell receptor derived from a human t cell clone implicated in disease. such transgenic mice could provide an excellent model for studying epitope spreading in a humanized immunogenetic environment and for testing of immunotherapies. the majority of the current therapies being planned for phase ii and iii trials in ms were first examined in eae. thus a particularly pertinent question is whether eae is a suitable relevant research tool for ms? some researchers believe that while eae is a useful model of acute human cns demyelination, its contribution to the understanding of ms is limited. eae is an acute monophasic illness, as compared to ms, which is a chronic relapsing disease, and may be more suited as a model of acute disseminated encephalomyelitis (adem). drawbacks of the eae model include the following: ( ) the nature of the inflammatory response in eae as compared to ms; ( ) th- -mediated disease in eae as compared to ms; ( ) differences in the pathology between eae and ms; and ( ) pitfalls in extending immunotherapies from eae to ms (see table - ). consequently it may be concluded that the clinical picture of eae presented depends not only on the animal species used, but also on the route of administration of the encephalitogen and the nature of the encephalitogen, mbp, plp, or mog. it is thus possible that these eae models are somewhat imprecise methods to study the pathogenesis of ms or to develop therapeutic strategies. the nonhuman primate eae models are of primary importance for the safety and efficacy of testing new therapeutics for ms that may not work sufficiently well in species distant from humans, such as rodents. questions concerning the immunogenicity of biological therapeutics have also been addressed in nonhuman primates. many biological therapeutics, such as anti-cd antibodies and altered peptide ligands, have been investigated in rodents. although some of these therapeutics have been effective in treating eae in rodents, they have proven to be partially effective, or in some cases detrimental, in ms patients. this ultimately raises the question of whether rodent models are the appropriate animal models for testing new therapeutic strategies for use in human ms. there are several examples, in humans and animals, of demyelinating diseases not associated with viral infections, such as demyelination associated with vitamin deficiency or toxins. many different animal models of eae have been studied using various mri techniques. the clinical features of such models depend greatly upon the route of inoculation of the encephalitogen as well as the species and strain of animal used. inoculation routes such as subcutaneous, footpad, or intraperitoneal are not helpful in determining the onset or location of the lesion in the brain or spinal cord. thus, to create demyelinating lesions of precisely known locations and time courses, stereotaxic techniques are used to inoculate animals with chemicals that induce demyelinating lesions in the brain. several chemicals, such as ethidium bromide, cuprizone, and lysophosphatidylcholine (lpc), when injected directly into nerves or into the cns, produce lesions of demyelination. for demyelination studies with lpc, male wistar rats are anesthetized with sodium pentathal and fixed in a rat head restraining stereotaxic surgical table, head shaved, a burr hole created, and . µl of a % lpc solution in isotonic saline injected using an injector cannula. then lpc is infused at the rate of . µl/min for the next - min. the cannula is then removed and the burr hole closed using bone wax. rats are then observed daily and histological studies are carried out from day to day after lpc injection to cover the entire process of disease evolution. using this lpc-induced demyelination it is possible to observe the complete pathological process of demyelination and remyelination in this animal model of ms. demyelination can be observed with the maximum value occurring on day . after day , remyelination starts with a reduction in edema. this model could be particularly useful for studying remyelination. one prominent feature of all chemically induced lesions is that the demyelinating lesions, and subsequent remyelination, can be studied without the interference of immune mechanisms. this has a tremendous advantage over virally induced models of ms: since no virus was inoculated none can remain to affect the remyelination. the use of animal models to investigate the pathogenesis of neuroinflammatory disorders of the central nervous system a multi-generational family with multiple sclerosis the geography of multiple sclerosis refl ects genetic susceptibility pathogenesis of virus-induced demyelination a comparison of the neurotropism of theiler's virus and poliovirus in cba mice theiler's virus infection in mice: an unusual biphasic disease leading to demyelination theiler's murine encephalomyelitis virus infection in mice: a persistent viral infection of the central nervous system which induces demyelination the effect of restraint stress on the neuropathogenesis of theiler's virus-induced demyelination, a murine model for multiple sclerosis the interaction of two groups of murine genes determines the persistence of theiler's virus in the central nervous system the genetics of the persistent infection and demyelinating disease caused by theiler's virus susceptibility to theiler's virus-induced demyelination. mapping of the gene within the h- d region the effect of l t t cell depletion on the pathogenesis of theiler's murine encephalomyelitis virus infection in cba mice characterization of b lymphocytes present in the demyelination lesions induced by theiler's virus investigation of the role of autoimmune responses to myelin in the pathogenesis of theiler's virus-induced demyelinating disease persistent infection with theiler's virus leads to cns autoimmunity via epitope spreading role of natural killer cells as immune effectors in encephalitis and demyelination induced by theiler's virus the role of cd + t cells in the acute and chronic phases of theiler's virus-induced disease in mice quantitative, not qualitative, differences in cd + t cell responses to theiler's murine encephalomyelitis virus between resistant c bl/ and susceptible sjl/j mice inhibition of theiler's virus mediated demyelination by peripheral immune tolerance induction in vivo evaluation of remyelination in rat brain by magnetization transfer imaging a human antibody that promotes remyelination enters the cns and decreases lesion load as detected by t -weighted spinal cord mri in a virus-induced murine model of ms persistent viral infections weiss sr the organ tropism of mouse hepatitis virus a in mice is dependent on dose and route of inoculation cd + and cd + t cells are not major effectors of mouse hepatitis virus a -induced demyelinating disease cellular reservoirs for coronavirus infection of the brain in beta -microglobulin knockout mice autoimmune and virus induced demyelinating diseases. a review dynamic regulation of alpha-and beta-chemokine expression in the central nervous system during mouse hepatitis virus-induced demyelinating disease remyelination, axonal sparing, and locomotor recovery following transplantation of glial-committed progenitor cells into the mhv model of multiple sclerosis oligodendrocytes and their myelin-plasma membrane connections in jhm mouse hepatitis virus encephalomyelitis macrophage infiltration, but not apoptosis, is correlated with immune-mediated demyelination, following murine infection with a neurotropic coronavirus cd and cd t cells have redundant but not identical roles in virus-induced demyelination cd + t cr+ and cd + tcr-cells in whole bone marrow facilitate the engraftment of hematopoietic stem cells across allogeneic barriers bystander cd t-cellmediated demyelination is interferon-γ-dependent in a coronavirus model of multiple sclerosis a role for α -integrin in the pathology following semliki forest virus infection replication of the a ( ) strain of semliki forest virus is restricted in neurons demyelination induced in mice by avirulent semliki forest virus. ii. an ultrastructural study of focal demyelination in the brain long-term effects of semliki forest virus infection in the mouse central nervous system a case for cytokines as effector molecules in the resolution of virus infection the adhesion molecule and cytokine profi le of multiple sclerosis lesions physiological deficits in the visual system of mice infected with semliki forest virus and their correlation with those seen in patients with demyelinating disease role of immune responses in protection and pathogenesis during semliki forest virus encephalitis isolation and characterization of cells infiltrating the spinal cord during the course of chronic relapsing experimental allergic encephalomyelitis in the biozzi ab/h mice characterization of the cellular and cytokine response in the central nervous system following semliki forest virus infection probable epitopes: relationships between myelin basic protein antigenic determinants and viral and bacterial proteins acute experimental allergic encephalomyelitis increases lumbar spinal cord incorporation of epidurally administered [( )h]-d-mannitol and [( )c]-carboxyl-inulin in rabbits upregulation of group cd antigen presenting models in guinea pigs with experimental autoimmune encephalomyelitis: an immunohistochemical study neurological response after induction of experimental allergic encephalomyelitis in susceptible and resistant rat strains upregulation of monocyte chemotactic protein- and cc chemokine receptor in the central nervous system is closely associated with relapse of autoimmune encephalomyelitis in lewis rats evaluation of a rat model of experimental autoimmune encephalomyelitis with human mbp as antigen induction of experimental autoimmune encephalomyelitis in dark agouti rats without adjuvant inhibition of experimental autoimmune encephalomyelitis by inhalation but not oral administration of the encephalitogenic peptide: influence of mhc binding affi nity dendritic cells and the control of immunity rat models as tools to develop new immunotherapies eae in the common marmoset callithrix jacchus demyelination and axonal damage in a non-human primate model of multiple sclerosis histopathological characterization of magnetic resonance imaging-detectable brain white matter lesions in a primate model of multiple sclerosis: a correlative study in the experimental autoimmune encephalomyelitis model in common marmosets (callithrix jacchus) experimental allergic encephalomyelitis in the new world monkey, callithrix jacchus observations on the attempts to produce acute disseminated allergic encephalomyelitis in primates rhesus monkeys are highly susceptible to experimental autoimmune encephalomyelitis induced by myelin oligodendrocyte glycoprotein: characterization of immunodominant t-and b-cell epitopes sex differences in experimental autoimmune encephalomyelitis in multiple murine strains eae susceptibility in fvb mice hyperalgesia in an animal model of multiple sclerosis encephalitogenic and tolerogenic potential of altered peptide ligands of mog and plp in biozzi abh mice native myelin oligodendrocyte glycoprotein promotes severe chronic neurological disease and demyelination in biozzi abh mice hla class ii associated genetic susceptibility in multiple sclerosis: a critical reevaluation progress in determining the causes and treatment of multiple sclerosis t cell reactivity to multiple myelin antigens in multiple sclerosis patients and healthy controls humanized animal models for autoimmune diseases identifi cation of t cell epitopes on human proteolipid protein and induction of experimental autoimmune encephalomyelitis in hla class ii-transgenic mice disease-related epitope spread in a humanized t cell receptor transgenic model of multiple sclerosis experimental allergic encephalomyelitis: a misleading model of multiple sclerosis the use of monoclonal antibodies for treatment of autoimmune disease an altered peptide ligand mediates immune deviation and prevents autoimmune encephalomyelitis the ups and downs of multiple sclerosis therapeutics gadolinium enhancement in acute and chronic progressive eae in the guinea pig sequential diffusion-weighted magnetic resonance image study of lysophosphatidyl choline-induced experimental demyelinating lesion: an animal model of multiple sclerosis the authors would like to thank dana parks for expert secretarial assistance with this manuscript. key: cord- -kuwy pbo authors: liu, jiajun; neely, michael; lipman, jeffrey; sime, fekade; roberts, jason a.; kiel, patrick j.; avedissian, sean n.; rhodes, nathaniel j.; scheetz, marc h. title: development of population and bayesian models for applied use in patients receiving cefepime date: - - journal: clin pharmacokinet doi: . /s - - - sha: doc_id: cord_uid: kuwy pbo background and objective: understanding pharmacokinetic disposition of cefepime, a β-lactam antibiotic, is crucial for developing regimens to achieve optimal exposure and improved clinical outcomes. this study sought to develop and evaluate a unified population pharmacokinetic model in both pediatric and adult patients receiving cefepime treatment. methods: multiple physiologically relevant models were fit to pediatric and adult subject data. to evaluate the final model performance, a withheld group of pediatric patients and two separate adult populations were assessed. results: seventy subjects with a total of cefepime concentrations were included in this study. all adults (n = ) on average weighed . kg and displayed a mean creatinine clearance of . ml/min. all pediatric subjects (n = ) had mean weight and creatinine clearance of . kg and . ml/min, respectively. a covariate-adjusted two-compartment model described the observed concentrations well (population model r( ), . %; bayesian model r( ), . %). in the evaluation subsets, the model performed similarly well (population r( ), . %; bayesian r( ), . %). conclusion: the identified model serves well for population dosing and as a bayesian prior for precision dosing. electronic supplementary material: the online version of this article ( . /s - - - ) contains supplementary material, which is available to authorized users. cefepime is a commonly utilized antibiotic for nosocomial infections. rising resistance, manifesting as increased cefepime minimum inhibitory concentrations (mics), has led to more frequent clinical failures [ , ] . to advise clinical outcomes according to mics, the clinical and laboratory standards institute updated the susceptibility breakpoints and then created a category of susceptibledose dependent for mics of and mg/l for enterobacteriaceae spp. [ ] . achieving goal pharmacokinetic exposures to effectively treat these higher mics can require a precision dosing approach. the online version of this article (https ://doi.org/ . /s - - - ) contains supplementary material, which is available to authorized users. cefepime, like other β-lactams, has pharmacodynamic activity governed by 'time-dependent' activity. the fraction of time that the unbound drug concentration exceeds the mic (ft >mic ) for the dosing interval is the pharmacokinetic/pharmacodynamic (pk/pd) efficacy target for cefepime [ ] , and a target of - % has been previously proposed [ ] [ ] [ ] [ ] . for the currently approved cefepime product and combination agents in the pipeline [ , ] , understanding cefepime disposition and variability is crucial to for optimal treatment of patients. as inter-and intra-patient pk variability can impact the achievement of pd goals, understanding the precision of population dosing is important. further, to fully realize precision dosing, individualized models (e.g., bayesian models) are needed. once developed, these models will form the basis for adaptive feedback and control strategies when paired with real-time drug assays. the purpose of this study was to: ( ) develop and evaluate a unified cefepime population pk model for adult and pediatric patients, and ( ) construct an individualized model that can be utilized to deliver precision cefepime dosing. data from four clinical cefepime pk studies representing unique groups of patients were compiled. subject demographics and study methodologies have been previously described [ ] [ ] [ ] [ ] . in brief, populations represented were febrile neutropenic adults with hematologic malignancies [ , ] , those with critical illness [ ] , and children with presumed or documented bacterial infections [ ] . for the two studies that evaluated adults with neutropenic fever, sime et al. prospectively enrolled patients receiving chemotherapy and/or stem cell transplant who subsequently developed febrile neutropenia and were administered maximum doses of cefepime [ ] . a total of cefepime plasma concentrations in presumably steady-state dosing intervals (third, sixth, and ninth) were analyzed for pk target attainment. whited et al. prospectively studied similar patients (n = ) who were admitted to hematology-oncology services and were receiving cefepime at a maximum dosage for febrile neutropenia [ ] . cefepime pk samples were obtained during steady state and analyzed for population parameters. critically ill adults were studied by roberts et al. as a prospective multinational pk study and included patients who received cefepime (only n = were included for model evaluation) [ ] . last, reed et al. characterized cefepime pharmacokinetics in hospitalized pediatric patients (above months of age) who received cefepime as monotherapy for bacterial infections [ ] . for our study, only those who received intravenous cefepime were included for model development. adult (n = ) and partial pediatric (n = ) datasets were utilized for pk model building ( fig. ) [ , ] . model evaluation was performed with other datasets consisting of independent adult (n = , n = ) and pediatric (n = ) patients [ ] [ ] [ ] ] . pediatric patients from reed et al. [ ] were randomized into the model building or the evaluation dataset. all clinical patient-level data included age, weight, and serum creatinine. an estimated creatinine clearance (crcl) was calculated for each patient [ ] . the cockcroft-gault formula (applied to all subjects) served as a standardized descriptor for the elimination rate constant. this study was exempted by the institutional review board at midwestern university chicago college of pharmacy. to construct the base pk models, the nonparametric adaptive grid (npag) algorithm [ , ] within the pmetrics (version . . ) package [ ] for r [ ] was utilized. multiple physiologically relevant, one-and two-compartmental pk models were built and assessed. the one-compartment structural model included an intravenous cefepime dose into and parameterized total cefepime elimination rate constant (k e ) from the central compartment. the two-compartment model included additional parameterizations of intercompartmental transfer constants between central and peripheral compartments (k cp and k pc ). in candidate models, total cefepime elimination was explored according to full renal and partial renal clearance (cl) models [i.e., nonrenal elimination (k e intcpt) and renal elimination descriptor (k e vectorized as a function of glomerular filtration estimates)] [ , ] . assay error was included into the model using a polynomial equation in the form of standard deviation (sd) a unified cefepime population pharmacokinetic model has been developed from adult and pediatric patients and evaluates well in independent populations. when paired with real-time β-lactam assays, a precision dosing approach will optimize drug exposure and improve clinical outcomes. as a function of each observed concentration, y (i.e., sd = c + c · y). observation weighting was performed using gamma (i.e., error = sd · gamma), a multiplicative variance model to account for extra process noise. gamma was initially set at with c and c equal to . and . , respectively. covariate relationships were assessed using the 'pmstep' function in pmetrics by applying stepwise linear regressions (forward selection and backwards elimination) of all covariates on pk parameters. additionally, a priori analyses examined the effect of covariates on cefepime k e , and both weight and crcl were variables considered a priori to have a high potential likelihood to impact cefepime pharmacokinetics ( [ , , ] . weight and crcl were standardized to kg and ml/min, respectively. further, an allometric scaler was applied to standardized weight (i.e., quotient of weight in kg divided by kg raised to the negative . th power) as a covariate adjustment to k e (esm). ultimate model retention was governed according to criteria described below. the best-fit pk and error model was identified by the change in objective function value (ofv) calculated as differences in − log-likelihood, with a reduction of . in ofv corresponding to p < . based on chi-square distribution and one degree of freedom. further, the best-fit model was selected based on the rule of parsimony and the lowest akaike's information criterion scores. goodness of fit of the competing models were evaluated by regression on observed vs. predicted plots, coefficients of determination, and visual predictive checks. predictive performance was assessed using bias and imprecision in both population and individual prediction models. bias was defined as mean weighted prediction error; imprecision was defined as bias-adjusted mean weighted squared prediction error. posterior-predicted cefepime concentrations for each study subject were calculated using individual median bayesian posterior parameter estimates. to evaluate the final adjusted model, the npag algorithm [ , ] was employed to assess the performance with separate data sets (fig. ). the population joint density from the best-fit covariate adjusted model was employed as a bayesian prior for the randomly withheld pediatric data adult data adult data adult data fig. schematic for data sources in model development and evaluation and separate adult data. in the evaluation process, structural model, model parameters, assay error, and observation weighting were unchanged. goodness of fit of the competing models were determined as described above. simulation was performed to examine the exposures predicted by the final model, employing all support points from the population parameter joint density in the final npag analysis [ , ] . each support point was treated as a mean vector surrounded by the population variance-covariance matrix (i.e., covariance equal to the population covariance divided by the total number of support points). for each subject, simulated profiles were created with predicted outputs at . -h intervals. covariate values for each simulated subject were fixed based on arithmetic means of observed weight and crcl for corresponding adult and pediatric populations. semi-parametric monte carlo sampling was performed from the multimodal multivariate distribution of parameters with the parameter space concordant with the npag population analysis results (i.e., best-fit model) [ table s of the esm] [ ] . maximum dosing regimens were simulated for adult and pediatric populations (total n = ): g every h infused over . h and mg/kg every h infused over . h, respectively. protein binding of % (i.e., % free fraction of total cefepime dose) was accounted for in predicting cefepime concentrations [ ] . the pk/pd target of ft >mic ≥ % was utilized across doubling mics of . - mg/l over the first h of cefepime therapy [ ] . estimates are provided from the first h of simulations as timely administration of effective antimicrobial agents is associated with increased survival [ ] . a total of clinically diverse subjects, contributing cefepime concentrations, were included in this study (n = subjects for model development; n = subjects for evaluation) (fig. a total of cefepime observations were available for model development. cefepime concentrations ranged from . to . μg/ml. the base one-and two-compartment models (without covariate adjustment) produced reasonable fits for observed and bayesian posterior-predicted cefepime concentrations (r = . % and . %, respectively), but population estimates were unsatisfactory (r = . % and . %, respectively) ( table ) . weight and crcl displayed relationships with the standard two-compartment model (i.e., base two-compartment model). volume of distribution was associated with weight (p < . ) and ke (total) was associated with crcl (p < . ). after standardizing weight (to kg) without an allometric scaler in the base two-compartment model, fits for both population and bayesian posterior estimates against the observed data improved (r = . % and . %, respectively; ofv change, ). bias and imprecision for bayesian posterior fits were − . and . , respectively. when covariates (i.e., weight to volume of distribution and k e ; crcl to k e ) and the allometric scaler were applied in the two-compartment model, bayesian posteriors fit well (r = . %; fig. right) with low bias and imprecision (− . and . , respectively), and the population pk model produced good fits of the observed cefepime concentrations (r = . %, bias = . , imprecision = . ; fig. left) . the ofv change from the weightadjusted two-compartment model to the final model was significant at − (p < . ) [ table ]. the final model also produced acceptable predicted checks (fig. ) . thus, a two-compartment model with weight and crcl as covariate adjustments and allometric scaling was selected as the final pk model. the population parameter values from the final pk model are summarized in table . structural model and differential equations that define the population pk are listed in the esm. the population parameter value covariance matrix can be found in table . additionally, weighted residual error plots for the best-fit model (fig. s ) and scatter plots for covariates for the base structural model (fig. s ) can be found included in the esm. for the evaluation subset, bayesian priors resulted in reasonably accurate and precise predictions (population r = . %, bayesian r = . %; fig. ). fig. goodness-of-fit plots for best-fit population cefepime pk model (model development) results of the probability of target attainment (pta) analysis are shown in table this study created a population and individual pk model for adult and pediatric patients and can serve as a bayesian prior for precision dosing. when paired with a real-time assay for cefepime, this model allows for precise and accurate predictions of cefepime disposition via adaptive feedback control. in the absence of real-time assays, these cefepime pk parameters facilitate more accurate population-based dosing table population parameter value covariance matrix for the best-fit model strategies. previous work by rhodes et al. has shown an absolute difference of approximately % in survival probability across the continuum of achieving - % ft >mic in adult patients with gram-negative bloodstream infections, thus understanding the dose and re-dosing interval necessary to achieve optimal pk exposures should greatly improve clinical outcomes for patients treated with cefepime [ ] . individualized dosing and therapeutic drug monitoring of β-lactam antibiotics (e.g., cefepime) are critically important to achieving optimal drug exposure (i.e., optimal ft >mic as the pk/pd target) and improving clinical outcomes [ , , ] . precision medicine has been named as a major focus for the national health institute with $ million invested [ ] , yet precision medicine has mostly focused on genomic differences [ , ] . precision dosing is an important facet of precision medicine, and renewed efforts in precision dosing in the real-world setting are being pursued [ ] . cefepime is a highly relevant example. while rigorous reviews and analyses are conducted during the development phase of an antibiotic, dose optimization is far less ideal for the types of patients who ultimately receive the drug. this is highlighted by the fact that although cefepime-associated neurotoxicity is rare, this serious and potentially life-threatening adverse event has been increasingly reported and few strategies exist for optimizing and delivering precision exposures [ , ] . lamoth et al. found that a cefepime trough concentration of ≥ mg/l has a % probability of predicting neurotoxicity [ ] . huwyler et al. identified a similar predictive threshold of > mg/dl (five-fold increased risk for neurologic events) [ ] . in contrast, rhodes et al. found the cut-off of mg/l to be suboptimal [ ] . furthermore, rhodes et al. performed simulations from literature cefepime data and observed a high intercorrelation amongst all pk parameters (i.e., area under the curve at steady state, maximum plasma concentration, and minimum plasma concentration), suggesting that more work is needed to establish the pharmacokinetic/toxicodynamic (pk/td) profile for cefepime. in addition to complications by these less-than-ideal pk/ td data, clinicians are left to treat patients with extreme age differences, organ dysfunction, and comorbid conditions affecting antibiotic pharmacokinetics/pharmacodynamics [ ] . further, a contemporary dose reduction strategy based on estimated renal function (e.g., estimated crcl using the cockcroft-gault formula) is also likely to be confounded in these patients by intrinsic pk variability, such as changes in volume of distribution, and the challenges of accurately estimating the glomerular filtration rate at any point in time, leading to more 'uncertainties' in balancing dose optimization and adverse events [ , ] . these 'real-world' patients are often under-represented, and thus not well understood, from a pk/pd and pk/td standpoint during the drug approval process. bridging to the more typical patients that are clinically treated is important and central to the mission of precision medicine. the findings of this study can be used to guide cefepime dosing in these 'real-world' patients. several other studies have reviewed population cefepime pharmacokinetics. sime [ ] . in our pediatric population, means of cl and elimination half-life were . l/h and . h, respectively. our simulation findings are similar to those of shoji et al. that the maximum pediatric cefepime dosing did not adequately achieve optimal exposure to target higher mics. while the cefepime package labeling recommends maximum dosages of g every h for adult patients with neutropenic fever and mg/kg every h for pediatric patients with pneumonia and/or neutropenic fever, there may be a need to extend these dosing regimens to other populations (in the absence of aforementioned indications) to achieve the best clinical outcomes by optimizing the pk/pd attainment goals [ ] . other studies also performed a simulation for pta with different cefepime regimens and renal functions. tam et al. found that with a pd target of % f t>mic, g every h ( -minute infusion) achieved approximately % pta for mic of mg/l in patients with crcl of ml/min while g every h achieved barely above % pta for an mic of mg/l in the same population [ ] . nicasio et al. also conducted a simulation using a pd target of % f t>mic in the critically ill with varying renal function. the maximum recommended dosage ( g every h) in patients with crcl between and ml/min achieved a pta of . % at an mic of mg/l; however, when the same regimen was infused over . h, the pta achieved was significantly lower [ ] . collectively, these findings suggest that cefepime exposure is highly variable and may be clinically suboptimal in a large number of patients commonly treated with cefepime. these findings support the need for precision dosing and therapeutic drug monitoring for β-lactam antibiotics to reach optimal pk/pd targets given the high variability in drug exposures. our study is not without limitations. although a relatively large and diverse cohort was included in model development and evaluation, we did not specifically assess certain subgroups such as patients with morbid obesity and severe renal dysfunction. these conditions may require patient-specific models. second, many studies to date included 'real-world' patients with various disease sates (e.g., neutropenic fever, renal failure, sepsis); however, all studies were conducted under the research protocol where doses, and administration times were all carefully confirmed. additional efforts will be needed to evaluate model performance in clinical contexts. a unified population model for cefepime in adult and pediatric populations was developed and demonstrated excellent performance on evaluation. current cefepime dosages are often suboptimal, and population variability is high. precision dosing approaches and real-time assays are needed for cefepime to optimize drug exposure and improve clinical outcomes. failure of current cefepime breakpoints to predict clinical outcomes of bacteremia caused by gram-negative organisms evaluation of clinical outcomes in patients with gram-negative bloodstream infections according to cefepime mic performance standards for antimicrobial susceptibility testing pharmacokinetic/pharmacodynamic parameters: rationale for antibacterial dosing of mice and men defining clinical exposures of cefepime for gram-negative bloodstream infections that are associated with improved survival clinical pharmacodynamics of cefepime in patients infected with pseudomonas aeruginosa relationship between pk/pd of cefepime and clinical outcome in febrile neutropenic patients with normal renal function interrelationship between pharmacokinetics and pharmacodynamics in determining dosage regimens for broad-spectrum cephalosporins cefepime/vnrx- broad-spectrum activity is maintained against emerging kpc-and pdc-variants in multidrug-resistant k. pneumoniae and p. aeruginosa pharmacokinetics of intravenously and intramuscularly administered cefepime in infants and children dali: defining antibiotic levels in intensive care unit patients: are current beta-lactam antibiotic doses sufficient for critically ill patients? adequacy of high-dose cefepime regimen in febrile neutropenic patients with hematological malignancies pharmacokinetics of cefepime in patients with cancer and febrile neutropenia in the setting of hematologic malignancies or hematopoeitic cell transplantation. pharmacotherapy guidance document: population pharmacokinetics guidance for industry prediction of creatinine clearance from serum creatinine an adaptive grid non-parametric approach to pharmacokinetic and dynamic (pk/pd) population models accurate detection of outliers and subpopulations with pmetrics, a nonparametric and parametric pharmacometric modeling and simulation package for r r: a language and environment for statistical computing cefepime clinical pharmacokinetics pharmacokinetics of cefepime in patients with respiratory tract infections cefepime in intensive care unit patients: validation of a population pharmacokinetic approach and influence of covariables population modeling and monte carlo simulation study of the pharmacokinetics and antituberculosis pharmacodynamics of rifampin in lungs duration of hypotension before initiation of effective antimicrobial therapy is the critical determinant of survival in human septic shock pharmacokinetics-pharmacodynamics of antimicrobial therapy: it's not just for mice anymore therapeutic drug monitoring of the beta-lactam antibiotics: what is the evidence and which patients should we be using it for? the white house. fact sheet: president obama's precision medicine initiative precision medicine: from science to value. health aff (millwood) precision medicine: changing the way we think about healthcare precision dosing: defining the need and approaches to deliver individualized drug dosing in the real-world setting cefepime and risk of seizure in patients not receiving dosage adjustments for kidney impairment characterizing cefepime neurotoxicity: a systematic review high cefepime plasma concentrations and neurological toxicity in febrile neutropenic patients with mild impairment of renal function cefepime plasma concentrations and clinical toxicity: a retrospective cohort study an exploratory analysis of the ability of a cefepime trough concentration greater than mg/l to predict neurotoxicity performance of the cockcroft-gault, mdrd, and new ckd-epi formulas in relation to gfr, age, and body size population pharmacokinetics of high-dose, prolonged-infusion cefepime in adult critically ill patients with ventilator-associated pneumonia population pharmacokinetic assessment and pharmacodynamic implications of pediatric cefepime dosing for susceptible-dose-dependent organisms pharmacokinetics and pharmacodynamics of cefepime in patients with various degrees of renal function acknowledgements j.a. roberts would like to acknowledge funding key: cord- - kwlteyr authors: wu, nicholas c; dai, lei; olson, c anders; lloyd-smith, james o; sun, ren title: adaptation in protein fitness landscapes is facilitated by indirect paths date: - - journal: nan doi: . /elife. sha: doc_id: cord_uid: kwlteyr the structure of fitness landscapes is critical for understanding adaptive protein evolution. previous empirical studies on fitness landscapes were confined to either the neighborhood around the wild type sequence, involving mostly single and double mutants, or a combinatorially complete subgraph involving only two amino acids at each site. in reality, the dimensionality of protein sequence space is higher ( (l)) and there may be higher-order interactions among more than two sites. here we experimentally characterized the fitness landscape of four sites in protein gb , containing ( ) = , variants. we found that while reciprocal sign epistasis blocked many direct paths of adaptation, such evolutionary traps could be circumvented by indirect paths through genotype space involving gain and subsequent loss of mutations. these indirect paths alleviate the constraint on adaptive protein evolution, suggesting that the heretofore neglected dimensions of sequence space may change our views on how proteins evolve. doi: http://dx.doi.org/ . /elife. . the fitness landscape is a fundamental concept in evolutionary biology (kauffman and levin, ; poelwijk et al., ; romero and arnold, ; hartl, ; kondrashov and kondrashov, ; de visser and krug, ) . large-scale datasets combined with quantitative analysis have successfully unraveled important features of empirical fitness landscapes (kouyos et al., ; barton et al., ; szendro et al., ) . nevertheless, there is a huge gap between the limited throughput of fitness measurements (usually on the order of variants) and the vast size of sequence space. recently, the bottleneck in experimental throughput has been improved substantially by coupling saturation mutagenesis with deep sequencing (fowler et al., ; hietpas et al., ; jacquier et al., ; wu et al., ; thyagarajan and bloom, ; qi et al., ; stiffler et al., ) , which opens up unprecedented opportunities to understand the structure of high-dimensional fitness landscapes (jiménez et al., ; pitt and ferré-d'amaré, ; payne and wagner, ) . previous empirical studies on combinatorially complete fitness landscapes have been limited to subgraphs of the sequence space consisting of only two amino acids at each site ( l genotypes) (weinreich et al., ; lunzer et al., ; o'maille et al., ; lozovsky et al., ; franke et al., ; tan et al., ) . most studies of adaptive walks in these diallelic sequence spaces focused on "direct paths" where each mutational step reduces the hamming distance from the starting point to the destination. however, it has also been shown that mutational reversions can occur during adaptive walks in diallelic sequence spaces such that adaptation proceeds via "indirect paths" (depristo et al., ; berestycki et al., ; martinsson, ; li, ; palmer et al., ) . in sequence space with higher dimensionality ( l , for a protein sequence with l amino acid residues), the extra dimensions may further provide additional routes for adaptation (gavrilets, ; cariani, ) . although the existence of indirect paths has been implied in different contexts, it has not been studied systematically and its influence on protein adaptation remains unclear. another underappreciated property of fitness landscapes is the influence of higher-order interactions. empirical evidence suggests that pairwise epistasis is prevalent in fitness landscapes (kvitek and sherlock, ; kouyos et al., ; o'maille et al., ; lozovsky et al., ) . specifically, sign epistasis between two loci is known to constrain adaptation by limiting the number of selectively accessible paths (weinreich et al., ) . higher-order epistasis (i.e. interactions among more than two loci) has received much less attention and its role in adaptation is yet to be elucidated (weinreich et al., ; palmer et al., ) . in this study, we investigated the fitness landscape of all variants ( = , ) at four amino acid sites (v , d , g and v ) in an epistatic region of protein g domain b (gb , amino acids in total) (figure -figure supplement ) , an immunoglobulin-binding protein expressed in streptococcal bacteria (sjö bring et al., ; sauer-eriksson et al., ) . the four chosen sites contain of the top positively epistatic interactions among all pairwise interactions in protein gb , as we previously characterized (figure -figure supplement ) . thus the sequence elife digest proteins can evolve over time by changing their component parts, which are called amino acids. these changes usually happen one at a time and natural selection tends to preserve those changes that make the protein more efficient at its specific tasks, while discarding those that impair the protein's activity. however the effect of each change depends on the protein as a whole, and so two changes that separately make the protein worse can make it much better if they occur together. this phenomenon is called epistasis and in some cases it can trap proteins in a suboptimal form and prevent them from improving further. proteins are made from twenty different kinds of amino acid, and there are millions of different combinations of amino acids that could, in theory, make a protein of a given length. studying protein evolution involves making variants of the same protein, each with just a few changes, and comparing how efficient, or "fit", they are. previous studies only measured the fitness of a few variants and showed that epistasis could block protein evolution by requiring the protein to lose some fitness before it could improve further. however, new techniques have now made it easier to study protein evolution by testing many more protein variants. wu, dai et al. focused on four amino acids in part of a protein called gb and tested the efficiency of every possible combination of these four amino acids, a total of , ( ) variants. contrary to expectations, the results suggested that the protein could evolve quickly to maximise fitness despite there being epistasis between the four amino acids. overcoming epistasis typically involved making a change to one amino acid that paved the way for further changes while avoiding the need to lose fitness. the original change could then be reversed once the epistasis was overcome. the complexity of this solution means it can only be seen by studying a large number of protein variants that represent many alternative sequences of protein changes. wu, dai et al. conclude that proteins are able to achieve a higher level of fitness through evolution by exploring a large number of changes. there are many possible changes for each protein and it is this variety that, despite epistasis, allows proteins to become naturally optimised for the tasks that they perform. while the full complexity of protein evolution cannot be explored at the moment, as technology advances it will become possible to study more protein variants. such advances would therefore hopefully allow researchers to discover even more about the natural mechanisms of protein evolution. space is expected to cover highly beneficial variants, which presents an ideal scenario for studying adaptive evolution. moreover, this empirical fitness landscape is expected to provide us insights on how high dimensionality and epistasis would influence evolutionary accessibility. briefly, a mutant library containing all amino acid combinations at these four sites was generated by codon randomization. the "fitness" of protein gb variants, as determined by both stability (i.e. the fraction of folded proteins) and function (i.e. binding affinity to igg-fc), was measured in a high-throughput manner by coupling mrna display with illumina sequencing (see materials and methods, figure figure supplement ) (roberts and szostak, ; olson et al., ) . the relative frequency of mutant sequences before and after selection allowed us to compute the fitness of each variant relative to the wild type protein (wt). while most mutants had a lower fitness compared to wt (fitness < ), . % of mutants were beneficial (fitness > ). (figure -figure supplement ) . we note that this study does not aim to extrapolate protein fitness to organismal fitness. although there are examples showing that protein fitness in vitro correlates with organismal fitness in vivo (natarajan et al., ; wu et al., ) , this relation may not be linear and is likely to be systemspecific due to the difference in selection pressures in vitro and in vivo (pál et al., ; hingorani and gierasch, ) . to understand the impact of epistasis on protein adaptation, we first analyzed subgraphs of sequence space including only two amino acids at each site ( figure a) . each subgraph represented a classical adaptive landscape connecting wt to a beneficial quadruple mutant, analogous to previously studied protein fitness landscapes (weinreich et al., ; szendro et al., ) . each variant is denoted by the single letter code of amino acids across sites , , and (for example, wt sequence is vdgv). each subgraph is combinatorially complete with = variants, including wt, the quadruple mutant, and all intermediate variants. we identified a total of subgraphs in which the quadruple mutant was the only fitness peak. by focusing on these subgraphs, we essentially limited the analysis to direct paths of adaptation, where each step would reduce the hamming distance from the starting point (wt) to the destination (quadruple mutant). out of possible direct paths, the number of selectively accessible paths (i.e. with monotonically increasing fitness) varied from to among the subgraphs ( figure b ). in the most extreme case, only one path was accessible from wt to the quadruple mutant wlfa ( figure a) . we also observed a substantial skew in the computed probability of realization among accessible direct paths (figure -figure supplement ), suggesting that most of the realizations in adaptation were captured by a small fraction of possible trajectories (weinreich et al., ) . these results indicated the existence of sign epistasis and reciprocal sign epistasis, both of which may constrain the accessibility of direct paths (weinreich et al., ; tufts et al., ) . indeed, we found that these two types of epistasis were prevalent in our fitness landscape ( figure c ). furthermore, we classified the types of all pairwise epistasis in each subgraph and computed the level of ruggedness as f sign þ f reciprocal , where f type was the fraction of each type of pairwise epistasis. as expected, the number of selectively inaccessible direct paths, i.e. paths that involve fitness declines, was found to be positively correlated with the ruggedness induced by pairwise epistasis (figure -figure supplement , pearson correlation = . , p= .  À ) (poelwijk et al., ) . our findings support the view that direct paths of protein adaptation are often constrained by pairwise epistasis on a rugged fitness landscape (weinreich et al., ; kondrashov and kondrashov, ) . in particular, adaptation can be trapped when direct paths are blocked by reciprocal sign epistasis. however, crucially, this analysis was limited to mutational trajectories within a subgraph of the sequence space. in reality, the dimensionality of protein sequence space is higher. intuitively, when an extra dimension is introduced, a local maximum may become a saddle point and allow for further adaptation -a phenomenon that is also known as "extra-dimensional bypass" (gavrilets, ; cariani, ; gutiérrez and maere, ) . with our experimental data, we observed two distinct mechanisms of bypass, either using an extra amino acid at the same site or using an additional site, that allow proteins to continue adaptation when no direct paths were accessible due to reciprocal sign epistasis ( figure ). the first mechanism of bypass, which we termed "conversion we identified a total of subgraphs in which the quadruple mutant was the only fitness peak. the number of accessible direct paths from wt to the quadruple mutant is shown for each subgraph. the maximum number of direct paths is . (c) the fraction of three types of pairwise epistasis around wt ( out of ), randomly sampled from the entire sequence space ( in total), or in the neighborhood of the top fitness variants and lethal variants. we note that this analysis is different from previous studies on how epistasis changes along adaptive walks, where the quadruples are chosen such that the fitness values of genotype , and are in increasing order (greene and crona, ) . sign epistasis and reciprocal sign epistasis, both of which can block adaptive paths, are prevalent in the fitness landscape. classification scheme of epistasis is shown at the top. each node represents a genotype, which is within a sequence space of two loci and two alleles. green arrows represent the accessible paths from genotype " " to a beneficial double mutant " " (colored in red). doi: . /elife. . the following figure supplements are available for figure : may open up potential paths that circumvent the reciprocal sign epistasis. the starting point is and the destination is (in red). green arrows indicate the accessible path. a successful bypass would require a "conversion" step that substitutes one of the two interacting sites with an extra amino acid ( ! ), followed by the loss of this mutation later ( ! ). the original reciprocal sign epistasis is changed to sign epistasis on the new genetic background after conversion. (b) among~ , randomly sampled reciprocal sign epistasis, > % of them can be circumvented by at least one conversion bypass (i.e. success, inset). the number of available bypass for the success cases is shown as histogram. (c) the second mechanism of bypass involves an additional site. in this case, adaptation involves a "detour" step to gain mutation at the third site ( ! ), followed by the loss of this mutation ( ! ). the original reciprocal sign epistasis is changed to either magnitude epistasis or sign epistasis on the new genetic background after detour ( bypass", works by converting to an extra amino acid at one of the interacting sites (palmer et al., ) . consider a simple scenario with only two interacting sites. if the sequence space is limited to amino acids at each site, as in past analyses of adaptive trajectories, the number of neighbors is ; however, if all possible amino acids were considered, the total number of neighbors would be . some of these extra neighbors may lead to potential routes that circumvent the reciprocal sign epistasis ( figure a ). in this case, a successful bypass would require a conversion step that substitutes one of the two interacting sites with an extra amino acid ( ! ), followed by the loss of this mutation ( ! ). this bypass is feasible only if the original reciprocal sign epistasis is changed to sign epistasis after the conversion. to test whether such bypasses were present in our system, we randomly sampled pairwise interactions from the sequence space and analyzed the~ , reciprocal sign epistasis among them (see materials and methods). more than % of the time there was at least one successful conversion bypass and in many cases multiple bypasses were available ( figure b ). the second mechanism of bypass, which we termed "detour bypass", involves an additional site ( figure c ). in this case, adaptation can proceed by taking a detour step to gain a mutation at the third site ( ! ), followed by the later loss of this mutation ( ! ) (depristo et al., ; palmer et al., ) . detour bypass was observed in our system ( figure d ), but was not as prevalent and had a lower probability of success than conversion bypass. out of possible detour bypasses for a chosen reciprocal sign epistasis, we found that there were on average . conversion bypasses and . detour bypasses available. we note, however, that the lower prevalence of detour bypass in our fitness landscape (l= ) does not necessarily mean that it should be expected to be less frequent than conversion bypass in other systems. while the maximum number of possible conversion bypasses is always fixed (  À ¼ ), the maximum number of possible detour bypasses (  (l À )) is proportional to the sequence length l of the entire protein (whereas our study uses a subset l = ). the pervasiveness of extra-dimensional bypasses in our system contrasts with the prevailing view that adaptive evolution is often blocked by reciprocal sign epistasis, when only direct paths of adaptation are considered. the two distinct mechanisms of bypass both require the use of indirect paths, where the hamming distance to the destination is either unchanged (conversion) or increased (detour). in order to circumvent the inaccessible direct paths via extra dimensions, reciprocal sign epistasis must be changed into other types of pairwise epistasis. for detour bypass, this means that the original reciprocal sign epistasis is changed to either magnitude epistasis or sign epistasis in the presence of a third mutation ( we proved that higher-order epistasis is necessary for the scenario that reciprocal sign epistasis is changed to magnitude epistasis, as well as for one of the two scenarios that reciprocal sign epistasis is changed to sign epistasis (see materials and methods). this suggests a critical role of higher-order epistasis in mediating detour bypass. to confirm the presence of higher-order epistasis, we decomposed the fitness landscape by fourier analysis (see materials and methods, figure -figure supplement ) weinreich et al., ; neidhart et al., ) . the fourier coefficients can be interpreted as epistatic interactions of different orders (weinreich et al., ; de visser and krug, ) , including the main effects of single mutations (the first order), pairwise epistasis (the second order), and higher-order epistasis (the third and the fourth order). the fitness of variants can be reconstructed by expansion of fourier coefficients up to a certain order (figure -figure supplement ). in our system with four sites, the fourth order fourier expansion will always reproduce the measured fitness (i.e. the fraction of variance in fitness explained equals ). when the second order fourier expansion does not reproduce the measured fitness, it indicates the presence of higher-order epistasis. in this way, we identified the . % of subgraphs with greatest fitness contribution from higher-order epistasis ( figure a , red lines) and visualized the corresponding quadruple mutants by the sequence logo plot ( figure b ). the skewed composition of amino acids in these subgraphs indicates that higherorder interactions are enriched among specific amino acid combinations of site , and . this interaction among sites is consistent with our knowledge of the protein structure, where the side chains of sites , , and can physically interact with each other at the core (figure -figure supplement a) and destabilize the protein due to steric effects ( figure -figure supplement ) . in the presence of higher-order epistasis, epistasis between any two sites would vary across different genetic backgrounds. we computed the magnitude of pairwise epistasis (") between each pair of amino acid substitutions (see materials and methods) (khan et al., ) , and observed numerous instances where the sign of pairwise epistasis depended on genetic background. for example, g l and v h were positively epistatic when site was isoleucine [i], but the interaction changed to negative epistasis when site carried a tyrosine [y] or a tryptophan [w] ( figure c-d) . similar patterns were observed in other pairwise interactions among site , and , such as g f/v a and v w/v h (figure -figure supplement ) . the observed pattern of higher-order epistasis was consistent with the results of the fourier analysis ( figure b ). for example, site was mostly excluded from higher-order epistasis; tyrosine [y] or tryptophan [w] at site were involved in the most significant higher-order interactions, as they often changed the sign of pairwise epistasis. higher-order epistasis can also switch the type of pairwise epistasis, such as shifting from reciprocal sign epistasis to magnitude or sign epistasis (figure -figure supplement ), which in turn is important for the existence of detour bypass. our analysis on circumventing reciprocal sign epistasis revealed how indirect paths could open up new avenues of adaptation. to study the impact of indirect paths at a global scale, we performed simulated adaptation in the entire sequence space of , variants. the fitness landscape was completed by imputing fitness values of the , missing variants (i.e. . % of the sequence space) that had fewer than sequencing read counts in the input library. our model of protein fitness incorporated main effects of single mutations, pairwise interactions, and three-way interactions among site , and (see materials and methods, figure -figure supplement ). we used predictor selection based on biological knowledge, followed by regularized regression, which has been demonstrated to ameliorate possible bias in the inferred fitness landscape (otwinowski and plotkin, ) . in the complete sequence space, we identified a total of fitness peaks (i.e. local maxima); among them peaks had fitness larger than wt and their combined basins of attraction covered % of the sequence space ( figure a) . we then simulated adaptation on the fitness landscape using three different models of adaptive walks (see materials and methods), namely the greedy model (de visser and krug, ), correlated fixation model (gillespie, ) , and equal fixation model (weinreich et al., ) . in the greedy model, adaptation proceeds by sequential fixation of mutations that render the largest fitness gain at each step. the other two models assign a nonzero fixation probability to all beneficial mutations, either weighted by (correlated fixation model) or independent of (equal fixation model) the relative fitness gain. the greedy model represents adaptive evolution of a large population with pervasive clonal interference (de visser and krug, ). the correlated fixation model represents adaptive evolution of a population under the scheme of strong-selection/weak-mutation (sswm) (gillespie, ) , which assumes that the time to fixation is much shorter than the time between mutations and the fixation probability of a given mutation is proportional to the improvement in fitness. the equal fixation model represents a simplified scenario of adaptation where all beneficial mutations fix with equal probability (weinreich et al., ) . among all the possible adaptive paths to fitness peaks, many of them involved indirect paths, i.e. they employed mechanisms of extra-dimensional bypass ( figure b, figure -figure supplement ). we classified each step on the adaptive paths into three categories based on the change of hamming distance to the destination (a fitness peak, in this case): "towards (- )", "conversion ( )", and "detour (+ )" ( figure c ). conversion was found to be pervasive during adaptation in our fitness landscape ( % of mutational steps for greedy model, % for correlated fixation model, % for equal fixation model). the use of detour was less frequent ( . % of mutational steps for greedy model, . % for correlated fixation model, . % for equal fixation model), in accordance with the previous observation that detour bypass was less available than conversion bypass in our fitness landscape with l ¼ . a conversion step would increase the length of an adaptive path by , while a detour step would increase the length by . as a result, an indirect path can be substantially longer than a direct path consisting of only "towards" steps. we found that many of the adaptive paths required more than steps, which was the maximal length of a direct path between any variants in this landscape ( figure d ). interestingly, because indirect adaptive paths involved more variants of intermediate fitness, the use of conversion and detour steps depended on the strength of selection. consistent with previous studies (orr, (orr, , , when mutations conferring larger fitness gains were more likely to fix (e.g. greedy model and correlated fixation model), adaptation favored direct moves toward the destination, thus leading to a shorter adaptive paths ( figure c-d) . this suggests that the strength of selection interacts with the topological structure of fitness landscapes to determine the length and directness of evolutionary trajectories. given that extra-dimensional bypasses can help proteins avoid evolutionary traps, we expect that their existence would facilitate adaptation in rugged fitness landscapes. indeed, we found that indirect paths increased the number of genotypes with access to each fitness peak ( figure e ). in addition, the fraction of genotypes with accessible paths to all fitness peaks increased from from % to % when indirect adaptive paths were allowed (figure -figure supplement c) . we also found that a substantial fraction of beneficial variants (fitness > ) in the sequence space were accessible from wt only if indirect paths were used ( figure f) . we repeated the analysis in figure f with the consideration of the constraints imposed by the standard genetic code (figure -figure supplement a). the constraints from the genetic code decreased the number of accessible variants due to the reduction in connectivity. however, this reduction in connectivity did not alter our core finding that indirect paths substantially increase evolutionary accessibility (figure -figure supplement b). taken together, these results suggest that indirect paths promote evolutionary accessibility in rugged fitness landscapes. this enhanced accessibility would allow proteins to explore more sequence space and lead to delayed commitment to evolutionary fates (i.e. fitness peaks) (palmer et al., ) . consistent with this expectation, our simulations showed that many mutational trajectories involving extra-dimensional bypass did not fully commit to a fitness peak until the last two steps (figure -figure supplement ). in our analysis, we have limited adaptation to the regime where fitness is monotonically increasing via sequential fixation of one-step beneficial mutants. when this assumption is relaxed, adaptation can sometimes proceed by crossing fitness valleys, such as via genetic drift or recombination (de visser and krug, ; weissman et al., ; ostman et al., ; poelwijk et al., ; weissman et al., ) . another simplification in most of our analyses is to treat all sequences in a "protein space" (smith, ) , where two sequences are considered as neighbors if they differ by a single amino-acid substitution. in practice, amino acid substitutions occurring via a single nucleotide mutation are limited by the genetic code, so the total number of one-step neighbors would be (figure -figure supplement ) . we also expect fitness landscapes of different systems to have different topological structure. even in our system (with > % coverage of the genotype space), the global structure of the fitness landscape is influenced by the imputed fitness values of missing variants, which can vary when different fitness models or fitting methods are used. our analysis also ignored measurement errors, but the measurement errors are expected to be very small due to the high reproducibility in the data (figure -figure supplement b). both imputation of missing variants and measurement errors can lead to slight mis-specification of the topological structure of the fitness landscape. finally, we note that the four amino acids chosen in our study are in physical proximity and have strong epistatic interactions. while the availability of conversion bypass only depends on the dimensionality at each site, the degree of higher-order epistasis and the availability of detour bypasses can be quite different in other fitness landscapes. although the details of a particular fitness landscape can influence the quantitative role of different bypass mechanisms, this does not undermine the generality of our conceptual findings on extra-dimensional bypass, higher-order epistasis, and their roles in protein evolution. higher-order epistasis has been reported in a few biological systems (wang et al., ; pettersson et al., ; palmer et al., ) , and is likely to be common in nature (weinreich et al., ) . in this study, we observed the presence of higher-order epistasis and systematically quantified its contribution to protein fitness. our results suggest that higher-order epistasis can either increase or decrease the ruggedness induced by pairwise epistasis, which in turn determines the accessibility of direct paths in a rugged fitness landscape (figure -figure supplement ). we also revealed the important role of higher-order epistasis in mediating detour bypass, which could promote evolutionary accessibility via indirect paths. our work demonstrates that even in the most rugged regions of a protein fitness landscape, most of the sequence space can remain highly accessible owing to the indirect paths opened up by high dimensionality. the enhanced accessibility mediated by indirect paths may provide a partial explanation for some observations in viral evolution. for example, throughout the course of infection hiv always seems to find a way to revert to the global consensus sequence, a putatively "optimal" hiv- sequence after immune invasion (zanini et al., ) . as we pointed out, the possible number of detour bypasses scales up with sequence length, so it will be interesting to study how extra-dimensional bypass influences adaptation in sequence space of even higher dimensionality. for example, it is plausible that the sequence of a large protein may never be trapped in adaptation (gavrilets, ) , so that adaptive accessibility becomes a quantitative rather than qualitative problem. given the continuing development of sequencing technology, we anticipate that the scale of experimentally determined fitness landscapes will further increase, yet the full protein sequence space is too huge to be mapped exhaustively. does this mean that we will never be able to understand the full complexity of fitness landscapes? or perhaps big data from high-throughput measurements will guide us to find general rules? by coupling state-of-the-art experimental techniques with novel quantitative analysis of fitness landscapes, this work takes the optimistic view that we can push the boundary further and discover new mechanisms underlying evolution (fisher et al., ; desai, ; szendro et al., ] ). two oligonucleotides (integrated dna technologies, coralville, ia), '-agt cta gta tcc aac ggc nns nns nnk gaa tgg acc tac gac gac gct acc aaa acc tt- ' and '-ttg taa tcg gat cct ccg gat tcg gtm nnc gtg aag gtt ttg gta gcg tcg tcg t- ' were annealed by heating to ˚c for min and cooling to room temperature over hr. the annealed nucleotide was extended in a reaction containing . mm of each oligonucleotide, mm nacl, mm tris-hcl ph . , mm mgcl , mm dtt, mm each dntp, and units klenow exo-(new england biolabs, ipswich, ma) for mins at ˚c. the product (cassette i) was purified by purelink pcr purification kit (life technologies, carlsbad, ca) according to manufacturer's instructions. a constant region was generated by pcr amplification using kod dna polymerase (emd millipore, billerica, ma) with . mm mgso , . mm of each dntp (datp, dctp, dgtp, and dttp), . ng protein gb wild type (wt) template, and . mm each of '-ttc taa tac gac tca cta tag gga caa tta cta ttt aca tat cca cca tg- ' and '-agt cta gta tcc tcg acg ccg ttg tcg tta gcg tac tgc- '. the sequence of the wt template consisted of a t promoter, ' utr, the coding sequence of protein gb , ' poly-gs linkers, and a flag-tag (figure figure supplement b) ( . the thermocycler was set as follows: min at ˚c, then three-step cycles of s at ˚c, s at ˚c, and s at ˚c, and min final extension at ˚c. the product (constant region) was purified by purelink pcr purification kit (life technologies) according to manufacturer's instructions. both the purified constant region and cassette i were digested with bcivi (new england biolabs) and purified by purelink pcr purification kit (life technologies) according to manufacturer's instructions. ligation between the constant region and cassette i (molar ratio of : ) was performed using t dna ligase (new england biolabs). agarose gel electrophoresis was performed to separate the ligated product from the reactants. the ligated product was purified from the agarose gel using zymoclean gel dna recovery kit (zymo research, irvine, ca) according to manufacturer's instructions. pcr amplification was then performed using kod dna polymerase (emd millipore) with . mm mgso , . mm of each dntp (datp, dctp, dgtp, and dttp), ng of the ligated product, and . mm each of '-ttc taa tac gac tca cta tag gga caa tta cta ttt aca tat cca cca tg- ' and '-gga gcc gct acc ctt atc gtc gtc atc ctt gta atc gga tcc tcc gga ttc- '. the thermocycler was set as follows: min at ˚c, then three-step cycles of s at ˚c, s at ˚c, and s at ˚c, and min final extension at ˚c. the product, which is referred as "dna library", was purified by purelink pcr purification kit (life technologies) according to manufacturer's instructions. affinity selection by mrna display (roberts and szostak, ; olson et al., ) was performed as described (figure -figure supplement a ) . briefly, the dna library was transcribed by t rna polymerase (life technologies) according to manufacturer's instructions. ligation was performed using nmol of mrna, . nmol of '-ttt ttt ttt ttt gga gcc gct acc- ', and . nmol of '-/ phos/-d(a) -(c ) -d(acc)-puromycin by t dna ligase (new england biolabs) in a ml reaction. the ligated product was purified by urea page and translated in a ml reaction volume using retic lysate ivt kit (life technologies) according to manufacturer's instructions followed by incubation with mm final concentration of kcl and mm final concentration of mgcl for at least min at room temperature to increase the efficiency for fusion formation (liu et al., ) . the mrna-protein fusion was then purified using anti-flag m affinity gel (sigma-aldrich, st. louis, mo). elution was performed using x flag peptide (sigma-aldrich). the purified mrna-protein fusion was reverse transcribed using superscript iii reverse transcriptase (life technologies). this reverse transcribed product, which was referred as "input library", was incubated with pierce streptavidin agarose (sa) beads (life technologies) that were conjugated with biotinylated human igg-fc (rockland immunochemicals, limerick, pa). after washing, the immobilized mrna-protein fusion was eluted by heating to ˚c. the eluted sample was referred as "selected library". pcr amplification was performed using kod dna polymerase (emd millipore) with . mm mgso , . mm of each dntp (datp, dctp, dgtp, and dttp), the selected library, and . mm each of '-cta cac gac gct ctt ccg atc tnn nag cag tac gct aac gac aac g- ' and '-tgc tga acc gct ctt ccg atc tnn nta atc gga tcc tcc gga ttc g- '. the underlined "nnn" indicated the position of the multiplex identifier, gtg for input library and tgt for post-selection library. the thermocycler was set as follows: min at ˚c, then to three-step cycles of s at ˚c, s at ˚c, and s at ˚c, and min final extension at ˚c. the product was then pcr amplified again using kod dna polymerase (emd millipore) with . mm mgso , . mm of each dntp (datp, dctp, dgtp, and dttp), the eluted product from mrna display, and . mm each of '-aat gat acg gcg acc acc gag atc ta cac tct ttc cct aca cga cgc tct tcc g- ' and '-caa gca gaa gac ggc ata cga gat cgg tct cgg cat tcc tgc tga acc gct ctt ccg- '. the thermocycler was set as follows: min at ˚c, then to three-step cycles of s at ˚c, s at ˚c, and s at ˚c, and min final extension at ˚c. the pcr product was then subjected to  bp paired-end sequencing using illumina hiseq platform. we aimed to obtain at least million paired-end reads for each input library and post-selection library such that the average coverage for each variant would be more than paired-end reads. there were , , paired-end reads obtained for the input library and , , paired-end reads obtained for the post-selection library. raw sequencing data have been submitted to the nih short read archive under accession number: bioproject prjna . we were able to compute the fitness for . % of all variants from the sequencing data. the fitness measurements in this study were highly consistent with our previous study on the fitness of single and double mutants in protein gb (pearson correlation = . , figure -figure supplement b) . the first three nucleotides of both forward read and reverse read were used for demultiplexing. if the first three nucleotides of the forward read were different from that of the reverse read, the given paired-end read would be discarded. for both forward read and reverse read, the nucleotides that were corresponding to the codons of protein gb sites , , , and were extracted. if coding sequence of sites , , , and in the forward read and that in the reverse read did not reversecomplement each other, the paired-end read would be discarded. subsequently, the occurrence of individual variants at the amino acid level for site , , , and in both input library and selected library were counted, with each paired-end read represented count. custom python scripts and bash scripts were used for sequencing data processing. all scripts have been deposited to https:// github.com/wchnicholas/proteingfourmutants. the fitness (w) for a given variant i was computed as: where count i;selected represented the count of variant i in the selected library, count i;input represented the count of variant i in the input library, count wt;selected represented the count of wt (vdgv) in the selected library, and count wt;input represented the count of wt (vdgv) in the input library. therefore, the fitness of each variant, w i , could be viewed as the fitness relative to wt (vdgv), such that = . variants with count input < were filtered to reduce noise. the fraction of all possible variants that passed this filter was . % ( , out of , all possible variants). the fitness of each single substitution variant was referenced to our previous study , because the sequencing coverage of single substitution variants in our previous study was much higher than in this study (~ fold higher). hence, our confidence in computing fitness for a single substitution variant should also be much higher in our previous study than this study. subsequently, the fitness of each single substitution in this study was calculated by multiplying a factor of . by the fitness of that single substitution computed from our previous study . this is based on the linear regression analysis between the single substitution fitness as measured in our previous study and in this study, which had a slope of . and a y-intercept of~ . the fitness of each profiled variant is shown in supplementary file . the three types of pairwise epistasis (magnitude, sign and reciprocal sign) were classified by ranking the fitness of the four variants involved (greene and crona, ) . to quantify the magnitude of epistasis (") between substitutions a and b on a given background variant bg, the relative epistasis model (khan et al., ) was employed as follows: where w ab represents the fitness of the double substitution, ln(w a ) and ln(w b ) represents the fitness of each of the single substitution respectively, and w bg represents the fitness of the background variant. as described previously , there exists a limitation in determining the exact fitness for very low-fitness variants in this system. to account for this limitation, several rules were adapted from our previous study to minimize potential artifacts in determining " . we previously determined that the detection limit of fitness (w) in this system is~ . . rule ) if max( wab wbg , wa wbg , wb wbg ) < . , " ab;bg;adjusted = rule ) if min(w a , w b , wa wbg , wb wbg ) < . , " ab;bg;adjusted = max( , " ab;bg ) rule ) if min(w ab , wab wbg ) < . , " ab;bg;adjusted = min( , " ab;bg ) rule prevents epistasis being artificially estimated from low-fitness variants. rule prevents overestimation of epistasis due to low fitness of one of the two single substitutions. rule prevents underestimation of epistasis due to low fitness of the double substitution. of note, " ab;bg;adjusted was set to if both rule and rule were satisfied. to compute the epistasis between two substitutions, a and b, on a given background variant bg, " ab;bg;adjusted would be used if any one of the above three rules was satisfied. otherwise, " ab;bg would be used. fitness decomposition was performed on all subgraphs without missing variants ( , subgraphs in total). we decomposed the fitness landscape into epistatic interactions of different orders by fourier analysis (stadler, ; szendro et al., ; weinreich et al., ; neidhart et al., ) . the fourier coefficients given by the transform can be interpreted as epistasis of different orders (weinreich et al., ; de visser and krug, ) . for a binary sequencez with dimension l (z i equals if mutation is present at position i, or otherwise), the fourier decomposition theorem states that the fitness function f ðzÞ can be expressed as (weinberger, ) : the formula for the fourier coefficientsfk is then: for example, we can expand the fitness landscape up to the second order, i.e. with linear and quadratic terms where s i ðÀ Þ zi fþ ; À g, andẽ i is a unit vector along the i th direction. in our analysis of subgraphs, there are a total of ¼ terms in the fourier decomposition, with i À Á terms for the i th order (i ¼ ; ; ; ; ). we can expand the fitness landscape up to a given order by ignoring all higherorder terms in equation . in this paper, we refer to higher-order epistasis as non-zero contribution to fitness from the third order terms and beyond. the fitness values for , variants ( . % of the entire sequence space) were not directly measured (read count in the input pool = ) or were filtered out because of low read counts in the input pool (see section "calculation of fitness"). to impute the fitness of these missing variants, we performed regularized regression on fitness values of observed variants using the following model (hinkley et al., ; otwinowski and plotkin, ) : here, f is the protein fitness. a is the intercept that represents the log fitness of wt; b i represents the main effect of a single mutation, i; m i is a dummy variable that equals if the single mutation i is present in the sequence, or if the single mutation is absent; and n m ¼  À Á ¼ is the total number of single mutations. similarly, g j represents the effect of interaction between a pair of mutations; p j is the dummy variable that equals either or depending on the presence of that those two mutations; and n p ¼  À Á ¼ is the total number of possible pairwise interactions. in addition to the main effects of single mutations and pairwise interactions, the three-way interactions among sites , and are included in the model, based on our knowledge of higher-order epistasis (figure ) . d k represents the effect of three-way interactions among sites , and ; t k is the dummy variable that equals either or depending on the presence of that three-way interaction; and n t ¼ ¼ is the total number of three-way interactions. thus, the total number of coefficients in this model is , including main effects of each site (i.e. additive effects), interactions between pairs of sites (i.e. pairwise epistasis), and a subset of three-way interactions (i.e. higher-order epistasis). out of the , variants with experimentally measured fitness values, , variants were non-lethal (f > ) and were used to fit the model coefficients using lasso regression (matlab r b). lasso regression adds a penalty term l p j j ( stands for any coefficient in the model) when minimizing the least squares, thus it favors sparse solutions of coefficients (figure -figure supplement b) . we calculated the -fold cross-validation mse (mean squared errors) of the lasso regression for a wide range of penalty parameter l (figure -figure supplement a) . l ¼ À is chosen. for measured variants, the model-predicted fitness values were highly correlated with the actual fitness values (pearson correlation= . , figure -figure supplement c). we then used the fitted model to impute the fitness of the , missing variants and complete the entire fitness landscape. imputed fitness values for missing variants are listed in supplementary file . simulating adaptation using three models for fixation python package "networkx" was employed to construct a directed graph that represented the entire fitness landscape for sites , , , and . a total of = , nodes were present in the directed graph, where each node represented a -site variant. for all pairs of variants separated by a hamming distance of , a directed edge was generated from the variant with a lower fitness to the variant with a higher fitness. therefore, all successors of a given node had a higher fitness than the given node. a fitness peak was defined as a node that had out-degree. three models, namely the greedy model (de visser and krug, ), the correlated fixation model (gillespie, ) , and the equal fixation model (weinreich et al., ) , were employed in this study to simulate the mutational steps in adaptive trajectories. under all three models, the probability of fixation of a deleterious or neutral mutation is . considering a mutational trajectory initiated at a node, n i with a fitness value of w i , where n i has m successors, (n , n , . . . n m ) with fitness values of (w , w , . . . w m ). then the probability that the next mutational step is from n i to n k , where k ( , , . . . m), is denoted p ifik and called the probability of fixation, and can be computed for each model as follows. for the greedy model (deterministic model), if w k ¼ maxðw ; w ; :::w m Þ; p i!k ¼ ( ) otherwise; p i!k ¼ for the correlated fixation model (non-deterministic model), for the equal fixation model (non-deterministic model), to compute the shortest path from a given variant to all reachable variants, the function "singlesource shortest path" in "networkx" was used. if the shortest path between a low-fitness variant and a high-fitness variant does not exist, it means that the high-fitness variant is inaccessible. if the length of the shortest path is larger than the hamming distance between two variants, it means that adaptation requires indirect paths. under constraints imposed by the standard genetic code, the connectivity of the directed graph that represented the fitness landscape was restricted according to the matrix shown in figure figure supplement a. the genetic distance between two variants was calculated according to the matrix shown in figure -figure supplement a. if the length of the shortest path is larger than the genetic distance between two variants, it means that adaptation requires indirect paths. in the subgraph analysis shown in figure -figure supplement , the fitness landscape was restricted to amino acids at each of the sites (the wt and adapted alleles). there was a total of variants, hence nodes, in a given subgraph. only those subgraphs where the fitness of all variants was measured directly were used (i.e. any subgraph with missing variants was excluded from this analysis). mutational trajectories were generated in the same manner as in the analysis of the entire fitness landscape (see subsection "simulating adaptation using three models for fixation"). in a subgraph with only one fitness peak, the probability of a mutational trajectory from node i to node j via intermediate a, b, and c was as follows: to compute the gini index for a given set of mutational trajectories from node i to node j, the probabilities of all possible mutational trajectories were sorted from large to small. inaccessible trajectories were also included in this sorted list with a probability of . this sorted list with t trajectories was denoted as (p i!j; , p i!j; , . . . p i!j;t ), where p i!j; was the largest and p i!j;t was the smallest. this sorted list was converted into a list of cumulative probabilities, which is denoted as (a i!j; , a i!j; , . . . a i!j;t ), where a i!j;t = p t n¼ p i!j;t . the gini index for the given subgraph was then computed as follows: gini index ¼  p tÀ n¼ ða i!j;n Þ þ a i!j;t À t t À ( ) sequence logo was generated by weblogo (http://weblogo.berkeley.edu/logo.cgi) (crooks et al., ) . the visualization of basins of attraction ( figure a ) was generated using graphviz with "fdp" as the option for layout. the ddg prediction was performed by the ddg monomer application in rosetta software (das and baker, ) using the parameters from row of table i in kellogg et al. (kellogg et al., ) . here we prove that higher-order epistasis is required for two possible scenarios of extra-dimensional bypass via an additional site (figure -figure supplement ) . for a fitness landscape defined on a boolean hypercube, we can expand the fitness as taylor series (weinberger, ) . f ¼ a þ a þ a þ a f ¼ a þ a þ a þ a þ a þ a þ a þ a to prove that higher-order epistasis is present is equivalent to prove that a ¼ . the fitness difference between neighbors is visualized by the directed edges that go from low-fitness variant to high-fitness variant, thus each edge represents an inequality. no cyclic paths are allowed in this directed graph. the reciprocal sign epistasis ( figure -figure supplement a) gives, : a < ( ) : a < ( ) ! : a þ a > ( ) ! : a þ a > the detour step ( ! ) and the loss step ( ! ) are required for extra-dimensional bypass, ! : a > ( ) : a þ a þ a þ a < combining inequality ( ) and ( ) gives combining inequality ( ) and ( ) gives combining the above two inequalities with ( ) and ( ), we arrive at a < for the scenario in (c), the proof of higher-order epistasis is similar. we have (the yellow edge) ! : a þ a > ( ) combining the above inequality with ( ), ( ) and ( ), we arrive at for the scenario in (d), when a þ a < , all the inequalities can be satisfied with a ¼ . so higher-order epistasis is not necessary in this case. scaling laws describe memories of host-pathogen riposte in the hiv population accessibility percolation with backsteps extradimensional bypass weblogo: a sequence logo generator macromolecular modeling with rosetta empirical fitness landscapes and the predictability of evolution mutational reversions during adaptive protein evolution statistical questions in experimental evolution evolutionary dynamics and statistical physics high-resolution mapping of protein sequence-function relationships evolutionary accessibility of mutational pathways two crystal structures of the b immunoglobulin-binding domain of streptococcal protein g and comparison with nmr evolution and speciation on holey adaptive landscapes molecular evolution over the mutational landscape the changing geometry of a fitness landscape along an adaptive walk modeling the evolution of molecular systems from a mechanistic perspective what can we learn from fitness landscapes? experimental illumination of a fitness landscape comparing protein folding in vitro and in vivo: foldability meets the fitness challenge a systems analysis of mutational effects in hiv- protease and reverse transcriptase capturing the mutational landscape of the beta-lactamase tem- comprehensive experimental fitness landscape and evolutionary network for small rna towards a general theory of adaptive walks on rugged landscapes role of conformational sampling in computing mutation-induced changes in protein structure and stability negative epistasis between beneficial mutations in an evolving bacterial population topological features of rugged fitness landscapes in sequence space exploring the complexity of the hiv- fitness landscape reciprocal sign epistasis between frequently experimentally evolved adaptive mutations causes a rugged fitness landscape phase transition for accessibility percolation on hypercubes optimized synthesis of rna-protein fusions for in vitro protein selection stepwise acquisition of pyrimethamine resistance in the malaria parasite the biochemical architecture of an ancient adaptive landscape accessibility percolation and first-passage site percolation on the unoriented binary hypercube epistasis among adaptive mutations in deer mouse hemoglobin exact results for amplitude spectra of fitness landscapes single-round, multiplexed antibody mimetic design through mrna display a comprehensive biophysical description of pairwise epistasis throughout an entire protein domain the population genetics of adaptation: the adaptation of dna sequences a minimum on the mean number of steps taken in adaptive walks impact of epistasis and pleiotropy on evolutionary adaptation inferring fitness landscapes by regression produces biased estimates of epistasis quantitative exploration of the catalytic landscape separating divergent plant sesquiterpene synthases comprehensive and quantitative mapping of energy landscapes for protein-protein interactions by rapid combinatorial scanning delayed commitment to evolutionary fate in antibiotic resistance fitness landscapes the robustness and evolvability of transcription factor binding sites replication and explorations of high-order epistasis using a large advanced intercross line pedigree rapid construction of empirical rna fitness landscapes empirical fitness landscapes reveal accessible evolutionary paths a quantitative highresolution genetic profile rapidly identifies sequence determinants of hepatitis c viral fitness and drug sensitivity rna-peptide fusions for the in vitro selection of peptides and proteins exploring protein fitness landscapes by directed evolution crystal structure of the c fragment of streptococcal protein g in complex with the fc domain of human igg streptococcal protein g. gene structure and protein binding properties natural selection and the concept of a protein space landscapes and their correlation functions evolvability as a function of purifying selection in tem- blactamase quantitative analyses of empirical fitness landscapes hidden randomness between fitness landscapes limits reverse evolution the inherent mutational tolerance and antigenic evolvability of influenza hemagglutinin epistasis constrains mutational pathways of hemoglobin adaptation in high-altitude pikas genetic background affects epistatic interactions between two beneficial mutations fourier and taylor series on fitness landscapes darwinian evolution can follow only very few mutational paths to fitter proteins should evolutionary geneticists worry about higher-order epistasis? perspective: sign epistasis and genetic costraint on evolutionary trajectories the rate at which asexual populations cross fitness valleys the rate of fitness-valley crossing in sexual populations mechanisms of host receptor adaptation by severe acute respiratory syndrome coronavirus high-throughput profiling of influenza a virus hemagglutinin gene at single-nucleotide resolution population genomics of intrapatient hiv- evolution we would like to thank jesse bloom and joshua plotkin for helpful comments on early versions of the manuscript. ncw was supported by philip whitcome pre-doctoral fellowship, audree fowler fellowship in protein science, and ucla dissertation year fellowship. ld was supported by hhmi postdoctoral fellowship from jane coffin childs memorial fund for medical research. rs was supported by nih r de . the funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. the funders had no role in study design, data collection and interpretation, or the decision to submit the work for publication. author contributions ncw, conception and design, acquisition of data, analysis and interpretation of data, drafting or revising the article; ld, jol-s, analysis and interpretation of data, drafting or revising the article; cao, conception and design, acquisition of data; rs, conception and design, drafting or revising the article key: cord- - tq cr o authors: vertrees, roger a.; goodwin, thomas; jordan, jeffrey m.; zwischenberger, joseph b. title: tissue culture models date: - - journal: molecular pathology of lung diseases doi: . / - - - - _ sha: doc_id: cord_uid: tq cr o the use of tissue cultures as a research tool to investigate the pathophysiologic bases of diseases has become essential in the current age of molecular biomedical research. although it will always be necessary to translate and validate the observations seen in vitro to the patient or animal, the ability to investigate the role(s) of individual variables free from confounders is paramount toward increasing our understanding of the physiology of the lung and the role of its cellular components in disease. additionally, it is not feasible to conduct certain research in humans because of ethical constraints, yet investigators may still be interested in the physiologic response in human tissues; in vitro characterization of human tissue is an acceptable choice. the use of tissue cultures as a research tool to investigate the pathophysiologic bases of diseases has become essential in the current age of molecular biomedical research. although it will always be necessary to translate and validate the observations seen in vitro to the patient or animal, the ability to investigate the role(s) of individual variables free from confounders is paramount toward increasing our understanding of the physiology of the lung and the role of its cellular components in disease. additionally, it is not feasible to conduct certain research in humans because of ethical constraints, yet investigators may still be interested in the physiologic response in human tissues; in vitro characterization of human tissue is an acceptable choice. tissue culture techniques have been utilized extensively to investigate questions pertaining to lung physiology and disease. the isolation and propagation of human bronchial epithelial cells has allowed investigators to begin to characterize the interactions and reactions that occur in response to various stimuli. moreover, the culture of human airway smooth muscle has allowed researchers to investigate a pathologic cascade that occurs in asthma as well as other physiologic responses in the smooth muscle of the lung. numerous lung cancer cell lines have been established to investigate their responses to chemotherapy and determine their biologic properties. overall, the use of cultured human lung tissue has provided a windfall of information on the pathogenesis of diseases that affect the lung and on the basic physiology and development of the lung in general. despite this wealth of information in the literature, this chapter is the fi rst to discuss the use of tissue culture models to examine the physiology and pathologic basis of lung diseases. in light of this, we briefl y discuss the history and principles behind the utilization of tissue culture. we then discuss the current use of tissue culture to examine many of the from henrietta lacks were cultivated into the fi rst immortal cell line-"hela." hela cells are still one of the most widely used cell lines today. since the s, tissue culture has become fi rmly established as a mechanism to answer many questions in biomedical research. today, tissue culture is widely used to investigate diseases that affect the lung, and through this work we have been able to increase our understanding of the pathologic cascades that occur in lung diseases, as well as the normal physiologies of the lung. tissue culture is a commonly used generic term for the in vitro cultivation of cells, attributed to the early cultures that generally consisted of heterogeneous cultures of crudely disaggregated tissues. currently, many terms are used that can be encompassed by the term: organ culture, cell culture, primary explants, and ex vivo propagation all deal with the in vitro cultivation of cells or tissues. cell culture in general can be applied either to primary cells (e.g., those with a fi nite life span) or to cell lines (e.g., hela cells). additionally, these cultures can be either a homogenous or a heterogenous group of cells. primary cell culture involves the isolation of cells from a tissue by disaggregation. single cell suspensions from tissues can be completed through either enzymatic digestion of extracellular matrix surrounding the cells-such as with ethylenediaminetetraacetic acid, trypsin, or collagenase-or mechanical disaggregation. these disaggregation procedures have the disadvantage of possibly injuring cells. if the cells of interest are adherent viable cells, they will be separated from nonviable cells when the medium is changed. alternatively, viable cells can be separated from nonviable cells prior to culture by subjecting the single cell suspension to density gradient centrifugation (e.g., hypaque). primary cells have an advantage of possessing many of the biologic properties that they possessed in vivo because they are not transformed. primary cells, unlike cell lines, are not immortal and have only a fi nite survival time in culture before becoming senescent. variant cells, however, as well as those obtained from neoplastic tissue, may proliferate infi nitely, thus becoming immortal in vitro. this will eventually allow the immortal cell to take over the culture and can be thought of as a cell line. in general, primary human cultures will survive for - passages in vitro, although this number is dependent on cell type, conditions, and possibly other unknown factors. primary cells are widely used to examine the effects of toxins, infectious agents, or other cellular interactions that would not be feasible in vivo. primary cells have a disadvantage of being a heterogeneous mixture of cells upon primary isolation, with the type of cell obtained generally a component of the disag-gregation method used. the most common contaminant seen following isolation of primary cells is cells of mesenchymal origin (e.g., fi broblasts). however, advances have been made that allow the culture of homogenous populations of cells. for instance, cell surface molecules specifi c for the cells of interest may be tagged with monoclonal antibodies. techniques such as fl uorescenceactivated cell sorting or the use of magnetic beads can be utilized to enrich the single cell suspension for the cell type of interest. additionally, some investigators have recently exploited unique characteristics of certain cells, such as the presence of p-glycoprotein or multidrug resistance-associated proteins expressed on endothelial cells, to poison other contaminating cells in culture. another type of primary cell culture is "primary explants." this type of culture is not subjected to a disaggregation procedure like the primary cell technique described earlier. therefore, single cell suspensions do not occur. briefl y, tissue samples are dissected and fi nely minced. these tissue pieces are then placed onto the surface of a tissue culture plate. following plating of tissue pieces, cells have been shown to migrate out of the tissue and onto the tissue culture surface. this technique is useful when cells of interest may become damaged or lost in the disaggregation technique described earlier and is often used to culture human bronchial epithelial cells. cell lines are another useful source of cells to investigate questions in biomedical research. these cells have the advantage of being immortal as opposed to the fi nite life spans that primary cells possess. additionally, they are generally well studied and characterized, leaving few experimental variables to worry about. these cells however, are prone to dedifferentiation-a process by which they lose the phenotypic characteristics of the cell from which they began. many of the early cell lines were established from tumor tissue and as such possess abnormal growth characteristics. newer cell lines have been established by molecular techniques such as inserting a telomerase gene into a cell to allow it to replicate infinitely. because of the phenotypic changes that allow cell lines to replicate infi nitely in culture, they are often a fi rst choice for experiments; however, they are also highly criticized in light of their nonnatural phenotype. organ culture, as the name implies, involves ex vivo culture of the whole or signifi cant portion of the organ. the main advantage to this type of culture is the retention and preservation of the original cell-cell interaction and extracellular architecture. this type of culture may be particularly important when experimental design necessitates the use of an ex vivo system, but researchers still need to retain the original organ architecture to answer questions posed. these types of cultures do not grow rapidly, however, and are therefore not suitable for experiments needing large numbers of a particular cell type. advantages and limitations of tissue culture tissue culture has become the penultimate tool of the reductionist biologist. the utilization of tissue culture as a research methodology has allowed investigators to study isolated interactions in its near-normal environment. these experiments by their very nature introduce artifacts; however, they do minimize the number of confounding variables that may affect a particular experiment. for instance, tissue culture allows investigators to determine the effects of one particular treatment on a particular cell type, which would not be feasible in vivo. additionally, tissue culture models of disease allow investigators to obtain samples and make observations more readily than those done in vivo. however, it is the relative simplicity of experiments done in vitro that allows models of disease or physiology to come under frequent and warranted criticism. these models do not take into consideration the complexity of biologic systems. diminishing possible confounding variables by culturing cells in vitro brings up the constant criticism of how applicable results are because of alterations of the normal cellular environment in vivo. for example, cell-cell interactions in vitro are reduced and unnatural. moreover, the culture does not contain the normal heterogeneity and threedimensional architecture that is seen in vivo. this said, however, tissue culture biology has proved to be successful in many ways. we have briefl y discussed the advantages that experimental systems using tissue culture affords researchers studying physiology and pathogenesis. because of its ability to isolate individual variables and determine their role(s) in physiology, cell culture has become an integral tool in deciphering the pathologic cascades that occur in human disease. diseases that affect lung are no exception. many diseases that affect the lung, and humans in general, are multifactorial. this begs the question how can cell culture, because of its reductionist nature only dealing with a minimal number of variables, help to solve the unknown questions and decipher the components involved in disease? often, clinical observations, and the questions arising therein, have been the launching pad for investigation. for instance, observations of massive infl ammation in the bronchoalveolar lavage samples of patients with acute respiratory disease syndrome (ards), consistent with damage seen in histologic samples, prompted investiga-tors to determine the role(s) of infl ammation in the etiology of ards. through the use of cell culture, investigators were able to determine individual interactions that occurred in the disease process. investigators have utilized culture models employing microcapillary endothelial cells under fl ow conditions to understand the role of proinfl ammatory cytokines in the cytokinesis and emigration of neutrophils in disease. using a model of pulmonary endothelium under fl ow conditions allowed investigators to demonstrate the importance of certain proinfl ammatory cytokines in ards. the role of inhaled toxicants in lung injury, and the mechanism(s) by which they cause disease, is another area of investigation that has utilized cell culture. scientists have developed diverse and unique tissue culture systems that contain air-liquid barriers of lung epithelium and subjected these cells to various gaseous toxicants to determine what occurs following inhalation of various chemicals. utilizing these types of systems, investigators are able to control the exposure time and other variables that may be diffi cult when determining inhaled toxicant effects in vivo. moreover, the use of tissue culture, as opposed to an animal model, allows investigators to observe effects kinetically, without undue changes (e.g., sacrifi ce) and expense in the experimental model. a tissue culture model also permits an investigator to observe multiple changes in real time, such as cellular integrity, cell signaling and intracellular traffi cking, protein expression changes, oxidant-induced cellular damage, and more. deciphering each of these changes in an animal model would be extremely diffi cult; through employing a tissue culture model, researchers are able to tightly control the experimental system while isolating the events of interest. further examples of how tissue culture models are currently being used to elucidate questions in lung physiology and disease are discussed later in the section on lung tissue cell lines. maintaining cells in vitro was initially a very diffi cult task. many characteristics need to be fulfi lled before a successful cell culture occurs. some of these characteristics are dependent on the type of tissue being studied; others may depend on specifi c requirements of the individual cells. various chemically defi ned media are now available commercially to support the growth and differentiation of numerous cell types. the creation of defi ned media has allowed investigators to culture a multitude of cell types while controlling the local environment to answer pertinent questions. for example, glucose can be removed from a culture medium in order to study its effects on cellular metabolism, relative position in the cell cycle, and many other effects. each chemical component is known in these media. additionally, investigators can add growth factors to nourish their cell cultures. the medium chosen when culturing cells in tissue culture must fi t two main requirements: ( ) it must allow cells to continue to proliferate in vitro, and ( ) it must allow the preservation of the certain specialized functions of interest. the most common medium formulations used currently in lung research are dulbecco's modifi ed eagle's medium, minimum essential medium, rpmi , and ham's f- . occasionally, investigators develop new medium types to attain a formulation that optimizes their own experimental conditions. fetal bovine serum is a common additive to most tissue culture media, although some investigators choose to forgo this additive for more defi ned supplementation. additionally, others may choose sera from other sources such as human serum when culturing cells of human origin. inactivation of complement by heat treating serum for hr at °c was initially very popular in tissue culture. however, it has become clear that this treatment may in fact damage some of the proteinaceous growth factors present in the medium, rendering it less effective. currently, many experts recommend heat inactivation only if the cell type of interest is particularly sensitive to complement. more specifi c examples of medium utilized in lung tissue culture models are given later in the section on lung tissue cell lines. when deciphering if the current culture conditions are suffi cient for the experimental design, the investigator must determine which cellular characteristics are important. not only are the general characteristics, such as adhesion, multiplication, and immortalization of cell types important, but so are tissue-specifi c characteristics. of importance to pulmonary research, the lung is a unique environment to simulate in vitro because of the air-liquid interface. recently, investigators have made use of culture insert wells (e.g., transwells, corning) in order to study this interaction. cell adhesion nearly all normal or neoplastic human epithelial cells will attach with relative ease to tissue culture surfaces. most tissue culture models utilizing tissue of lung origin fi t this description, with the notable exception of small cell lung carcinoma cell lines. however, for culture cells that may loosely adhere, or may not adhere at all, scientists coat tissue culture surfaces with extracellular matrix proteins. incubating tissue culture surfaces with serum, as well as laminin, fi bronectin, or collagen, prior to culture has been shown to improve attachment of fi nicky cells. these treatments also help in replicating the normal attachment of cells to extracellular matrix proteins in vivo. the development of continuous cell lines may be serendipitous, as was the development of early cell lines. in brief, many investigators would continue splitting primary cell cultures until one or more cell clones became immortal. unfortunately, the changes that generally occurred in culture led to cells with abnormal phenotypes that had undergone dedifferentiation. today, many investigators choose to use molecular biology techniques, exploiting our current knowledge of oncogenic viruses and enzymatic processes of cellular aging to transform primary cells in vitro to an immortal phenotype. it is known that the large t antigen present in the sv (simian virus) virus is capable of transforming cells to an abnormal phenotype. , , moreover, transfection of primary cells with a transposase enzyme has also been shown to induce an immortal phenotypic change while preserving most normal cellular functions and phenotypes. dedifferentiation a commonly encountered problem in tissue culture is dedifferentiation. this loss of phenotype may be insignificant to the research at hand or it may be critical, and it must be dealt with on a case by case basis. when a cell culture undergoes dedifferentiation it is often unclear whether undifferentiated cells took over the culture of terminally differentiated cells or whether a primary cell of interest became immortal under the culture conditions. the functional environment in which cells are cultured is critical when correlating experimental results to those seen in vivo. we previously alluded to the importance of the environment in which cells are cultured when discussing the advantages and limitations of tissue culture. investigators have frequently strived to replicate integral in vivo environments in vitro in order to increase the signifi cance of their experimental results. the development of cell culture insert wells (e.g., transwells, corning) has allowed investigators to culture bronchial or alveolar epithelial cells at an air-liquid interface. this ability allows investigators to begin to replicate a signifi cant aspect of these cells' functional environment in vitro, thereby increasing their understanding of the effects of gaseous particles on pulmonary epithelial cells. alternatively, scientists have also cultured epithelial cells on a roller bottle apparatus. this method allows investigators to determine the amount of time the apical epithelial cell surface is in contact with the air. capillary cell cultures have also come under frequent criticism when cultured in a monolayer in a tissue culture plate. investigators have been able to utilize gel matrices in which capillary cells form tubule-like structures, more closely replicating the architecture these cells maintain in vivo. additionally, endothelial cells are constantly under fl ow conditions in vivo. addressing this condition in vitro has allowed investigators to look at the role of endothelial cells during infl ammation-helping to increase the understanding of the role endothelium plays in acute lung injury. at times, researchers may also choose to determine the effects of soluble factors (e.g., cytokines, hormones, neurotransmitters) from acute patients or animal models in a cell culture model. the milieu of soluble factors present in the serum that may play a role in a disease state is considerable. moreover, these factors may have actions alone that are different when combined with other soluble factors. reconstituting every factor presents a diffi culty in vitro and leaves the possibility that an unknown factor may be missing. to address this, investigators have harvested sera from patients or animal models and used these samples as additives in their media formulations. for instance, through the use of serum samples from an animal model of smoke/burn injury-induced acute lung injury, investigators have demonstrated that use of arteriovenous co removal in acute lung injury signifi cantly reduces apoptotic cell death in epithelial cells. lung tissue cell lines: establishment and signifi cance the diversity of research fi elds utilizing tissue culture models of lung diseases is extensive. in this section, we will give a brief overview of the main lung cell types that are being utilized in research today to answer pressing questions about lung physiology and the pathophysiology of pulmonary disease. included in this discussion is also an overview of cell isolation and culture. the use of normal human bronchial epithelial (hbe) cells is extensively reported in the literature. based on a method pioneered by lechner et al., bronchial fragments obtained from surgery, autopsy, or biopsy specimens may be used as explants. the outgrowth of bronchial epithelial cells occurs readily from these explants when grown in medium supplemented with bovine pituitary extract and epidermal growth factor. alternatively, these cells have also been demonstrated to grow in basal keratinocyte serum-free medium without supplementation; however, they demonstrate a slower growth rate and earlier senescence. cultures of hbe cells are valuable for determining the responses to toxic inhaled pollutants. in vitro exposure systems based on these methods have several advantages. first, in vitro exposure systems can be stringently controlled and reproduced much better than in animal systems; second, individual determination of the cell types' responses to pollutants allows for a better characterization of the individual involvement of the cell type to a biologic response. finally, in vitro determination of the responses to toxic agents allows investigators to observe the reactions of human cells when testing in humans is not feasible because of ethical restraints. in vitro study of the responses of bronchial cells to gaseous pollutants is not without its diffi culties. wallaert et al. have described these constraints well. briefl y, because of the gaseous nature of the pollutants, culture systems should be designed that allow signifi cant exposure times to pollutants while also taking care to inhibit cells from drying out when exposed to air. to facilitate these experiments, roller bottle cultures have been developed that allow cells direct contact with the ambient air. alternatively, cells have been grown on a membrane fi lter and cultured at an air-liquid interface, which allows constant exposure to the experimental treatment. the same type of experiments that are used to determine the responses of cells to inhaled toxicants have also been used to characterize responses to inhaled pharmaceuticals. in addition to the characterization of responses to inhaled agents, epithelial cell cultures, notably alveolar epithelium obtained from fetal lung tissue, have allowed investigators to characterize the liquid transport phenotype that occurs in the developing lung. characterization of the cl − ion secretion system, which occurs in the distal lung epithelium throughout gestation, has been shown to be integral in the stimulation of growth of the developing lung by regulating liquid secretion. likewise, a phenotypic switch of na + absorptive capacity has been described toward the end of gestation, which is important for preparation of the lung to function postpartum and beyond. these culture systems have elucidated important physiologic changes that occur in the developing lung. similar experiments have demonstrated that while ion transport plays a crucial role in this process other hormones and neurotransmitters are also important. pulmonary endothelial cells represent a unique type of endothelium because of their paradoxical responses to hypoxia. this uniqueness highlights the need to utilize cell culture models of pulmonary endothelium as opposed to other endothelia when interested in investigating their role(s) in pulmonary physiology. several investigators have described the isolation and culture of pulmonary endothelial cells. persistent pulmonary hypertension of the newborn, also known as neonatal pulmonary hyper-tension, is caused by a disorder of the pulmonary vasculature from fetal to neonatal circulation, culminating in hypoxemic respiratory failure and death. the inciting events that culminate in neonatal pulmonary hypertension are multifactorial. despite this, decreased production of vasodilator molecules such as nitric oxide and prostaglandin i in the pulmonary endothelium has been shown to be a critical component of disease progression. primary cell cultures of human airway smooth muscle tissue can be obtained utilizing a method described by halayko et al. in which they isolated and characterized airway smooth muscle cells obtained from canine tracheal tissue. briefl y, airway smooth muscle cells were obtained by fi nely mincing tissue and subjecting it to an enzymatic disaggregation solution containing collagenase, type iv elastase, and type xxvii nagarse protease. following generation of a single cell suspension, cells may be grown in dulbecco's modifi ed eagle's medium supplemented with % fetal bovine serum. halayko et al. obtained approximately . × smooth muscle cells per gram of tissue using this method. although halayko et al. pioneered this technique using trachealis tissue, many other investigators have obtained airway smooth muscle cells from a variety of biopsy specimens. airway smooth muscle hyperreactivity and hypertrophy has been known for nearly years to be an important end response of asthma. the use of airway smooth muscle in vitro has been vital toward delineating the pathologic steps that occur in asthma, as well as testing of potential therapeutics that may help to decrease the morbidity and mortality of asthma. additionally, the relative paucity of in vivo models of asthma further illustrates the value of isolation and characterization of smooth muscle cells from asthmatic patients in vitro. using airway smooth muscle cell culture, investigators have characterized both the hypertrophic and hyperplastic growth of smooth muscle in individuals. investigation of the potential stimuli that lead to airway smooth muscle proliferation and hypertrophy have led researchers to implicate the mitogen-activated protein kinase family members, extracellular signal-regulated kinase- and - , and the phosphoinositol- kinase pathways in pathogenesis. additionally, mediators directing smooth muscle migration have also been observed in vitro and may play a role in the progression of asthma. platelet-derived growth factor, fi broblast growth factor- , and transforming growth factor-β (tgf-β) have all been shown to play a role in the migratory response of smooth muscle cells seen in asthma. additionally, contractile agonists such as leukotriene e have been shown to potentiate the migratory responses seen with platelet-derived growth factor treatment. human airway smooth muscle cell culture has also been utilized to investigate possible pharmacologic interventions for the treatment of asthma. β -agonists have been shown to decrease the rate of dna synthesis and likewise decrease the hyperplasia seen in airway smooth muscle cells in response to mitogenic stimuli through an increase in cyclic adenosine monophosphate. like β agonists, glucocorticoids have similar antiproliferative activities. lung cancer tissue and the development of novel therapeutics culture of neoplastic cells from human tumors has allowed investigators to harvest a wealth of knowledge into the biology of lung cancers; moreover, these cultures have provided potential models to test potential therapeutics. the propagation of lung cancer cells in vitro has been covered in great depth previously. in contrast to primary cell cultures, cultures of neoplastic cells are immortal, allowing their easy growth in culture with less chance of being overgrown by mesenchymal cells such as fi broblasts. the relative ease of growth in culture has led to many cell lines of lung cancer tissue. the national cancer institute, recognizing the need for a variety of lung cancer cell lines (both small cell and non-small cell), helped establish over cell lines. these lines are a wonderful resource for investigators given that they are extensively characterized, and many have full clinical data available. moreover, many of these cell lines are now easily available through the american type culture collection for a modest handling fee. additionally, if investigators do not wish to use currently established lung cancer cell lines, obtaining clinical samples for use in tissue culture models is relatively easy. the same methods used to obtain biopsy specimens for clinical staging can also be used to begin cell cultures. following culture and initial characterization of lung cancer cell lines, many investigators have demonstrated that lung cancer cell lines maintain a similar phenotype after establishment. specifi cally, it has been verifi ed that injection of lung cancer cell lines into nude mice exhibit similar histopathology to the original tumor, indicating minimal change occurred following establishment of the cell lines. small cell lung carcinoma (sclc) cell lines have been established from a multitude of biopsy specimens, including bone marrow, lymph nodes, and pleural effusions. , once viable cells have been obtained from clinical samples, cells are easily maintained in a basal cell culture medium such as rpmi in a humidifi ed incubator at °c and % co , although the initial isolations of sclc lines utilized hites and acl- media. most established sclc cell lines maintain a neuroendocrine phenotype in culture; however, baillie-johnson et al. noticed considerable heterogeneity in the cell lines they established, highlighting the signifi cance that establishing a cell line from the clinical sample of interest may provide investigators with a line that possesses the exact phenotypic properties of interest. small cell carcinoma poses many diffi culties to surgical treatment, owing to its early and widespread metastasis. therefore, combination chemotherapy is generally utilized in treatment. unfortunately, despite initial sensitivities, sclc tumors become resistant to further treatment. utilizing in vitro cultures of sclc cell lines, sethi et al. began to describe how extracellular matrix proteins can protect sclc against apoptosis-inducing chemotherapeutics through β -integrin-mediated survival signals. these data indicate that extracellular matrix proteins surrounding sclc may play a role in the local recurrence seen in patients following chemotherapy in vivo and suggest novel therapeutics aimed at blocking these survival signals. non-small cell lung carcinoma (nsclc) cell lines including squamous cell carcinoma, adenocarcinoma, and large cell carcinoma have all been established. despite the fact that nsclc cells comprise three distinct histologic cell types, all cell types can be established relatively easily. the primary treatment protocol for patients affl icted by nsclc is generally surgical resection of the tumor; therefore, tumor cells for culture are readily available. these cell types can be grown under conditions similar to those described for sclc. infectious diseases play a unique role in lung pathology in light of their roles as either important contributors or consequences of many lung diseases. for instance, certain lung diseases may predispose patients to infection: patients affl icted with obstructive lung diseases, as well as cystic fi brosis patients, commonly suffer from severe and recurrent bacterial infections. additionally, patients may become superinfected following a viral respiratory infection. systemic infections, such as gram-negative bacterial sepsis, may lead to lung diseases such as ards. human type ii alveolar pneumocytes and acute lung injury/acute respiratory distress syndrome pulmonary alveolar type ii cells are a unique cell subset that carries out highly specialized functions that include synthesis and secretion of surfactant, a unique composition of lipoproteins that act to reduce surface tension at the alveolar air-liquid interface. defi ning the molecular mechanisms leading to production of surfactant by type ii pneumocytes is important in many disease processes. the pathogenic sequence that results in ards, the most severe manifestation of alveolar lung injury, is generally thought to be initiated by a systemic infl ammatory response. despite this knowledge, there still exist many questions about the initial triggers and pathologic steps that occur in ards. greater understanding of these steps may help to develop new treatment regimes. currently, treatment of ards consists of mechanical ventilation, which helps to stabilize blood gases. however, mechanical ventilation itself may provoke further infl ammation in the alveoli, thereby decreasing compliance and gas exchange in the alveoli. the cell type of particular interest in ards and diffuse alveolar damage is the type ii pneumocytes. [ ] [ ] [ ] [ ] [ ] until recently, studies trying to decipher the pathologic sequence in acute lung injury have had to rely on standard lung epithelial cell lines. recently, however, human type ii alveolar epithelial cells (pneumocytes) have been successfully isolated from fetal human lung tissue by collagenase digestion. briefl y, fetal lung tissues were minced and incubated in a serum-free medium containing dibutyryl cyclic adenosine monophosphate for days. the tissue explants were then treated with collagenase and incubated with deae-dextran to eliminate contaminating fi broblasts. cells were then plated onto tissue culture dishes treated with extracellular matrix derived from mdck cells and cultured overnight in waymouth's medium containing % serum. these steps resulted in relatively pure populations of human type ii pneumocytes that were then cultured at an air-liquid interface. using these methods, alcorn et al. were able to maintain a primary culture that retained the morphologic and biochemical characteristics of type ii pneumocytes for up to weeks. conventional bioreactors and three-dimensionality: the origins of three-dimensional culture carrel postulated that tissue development was linked to access to nutrient supply, noting that peripheral cells grew readily, and internal cells became necrotic presumably based on their distance from the nutrient source. to circumvent this issue, carrel implemented cultures on silk veils, preventing the plasma clots of the growth media from deforming or becoming spherical, thus facilitating the internal cell's ability to obtain nutrient replenishment. many attempts were made in standard culture systems (bioreactors) and other culture apparatuses to escape the constraints of two-dimensional cell culture, with the intent of yielding high-fi delity human and mammalian tissues, and thus emphasizing the need for development of three-dimensional biology. another famous researcher, leighton, improved on carrel's techniques in the s and s. leighton's major contribution to three-dimensional culture technology was the introduction of the idea of a sponge matrix as a substrate on which to culture tissues. , leighton fi rst experimented on cellulose sponges surrounded by plasma clots resident within glass tubes. he devised a system to grow -to -mm tissue explants on sponges, using small amounts of chick plasma and embryo extract. after the mixture solidifi ed on the sponge leighton added the nutrient media and inserted the "histoculture" in a roller apparatus to facilitate nutrient mass transfer. he experimented with many sponge combinations, discovering that collagen-impregnated cellulose sponges were optimal for sustaining the growth of native tissue architecture. , leighton was successful in growing many different tissue types on the sponge-matrix cultures. , leighton also found that c hba mouse mammary adenocarcinoma cells, when grown on sponge-matrix histoculture, aggregated "much like the original tumor, forming distinct structures within the tumors such as lumina and stromal elements, and glandular structures." an extremely important difference of this threedimensional histoculture from the standard two-dimensional culture is the apparent quiescence of the stromal component and the balanced growth of these cells with regard to the overall culture. leighton further advanced the concept of three-dimensional histoculture to histophysiologic gradient cultures. these cultures are conducted in chambers that allow metabolic exchange between "the pool of medium and the culture chamber by diffusion across a membrane." histophysiologic gradient cultures mimic, to some degree, diffusion in tissues. from the pioneering work of carrel and leighton, other methods of emulating three-dimensional cultures have been developed, such as embedding cells and tissues in collagenous gels of rat tail as per the techniques of nandi and colleagues. many of the advantages of threedimensional cultures seen by leighton, nandi, and others may be attributed to permitting the cells to retain their normal shape and special associations. this global concept will be important as we begin to understand and recall the physical and environmental characteristics of the rotating-wall vessel systems. other methods of three-dimensional culture encompass a technique known as organ culture or culture on a fi lter, a strategy developed by strangeways and fell and robinson. tissue explants were grown on lens paper in a watch glass containing liquid culture medium. browning and trier found "that for some tissues, it is critical to keep the cultures at the air-liquid interface," thus allowing the tissues to experience conditions similar to the in vivo environment. another strategy is the use of three-dimensional cultures known as proto-tissues, or aggregates of cells, used to form spheroids. this technique was popularized by sutherland and colleagues more than years ago when they manipulated aggregates of cells into a spherical confi guration by spinning agitation of the cells in spinner fl asks. this technique produced pseudo-tissue-like organoids useful for research evaluations. each of these methodologies will be of benefi t as we continue to examine strategies for achieving three-dimensional lung tissue constructs. , finally, membrane bioreactors are capable of retaining enzymes, organelles, and microbial, animal, and plant cells behind a membrane barrier, trapped in a matrix or adherent to the membrane surface. in , gallup and gerhardt fi rst used the membrane bioreactor for dialysis culture of serratia marcescens. immobilized enzyme microencapsulation was pioneered by chang, but butterworth et al. fi rst developed the enzyme membrane reactor to successfully accomplish starch hydrolysis with α-amylase. likewise, for animal cell culturing, knazek et al. fi rst cultured human choriocarcinoma cells on compacted bundles of amicon fi bers. many reviews on the particular applications of hollow fi ber and immobilized bioreactant bioreactors for enzyme catalysts, microbial cells, and animal cell culture are available. [ ] [ ] [ ] [ ] [ ] [ ] as presented previously, tissue-engineering applications of three-dimensional function and structure are well known in medical science research. in microgravity three-dimensional aggregates form, facilitating the expression of differentiated organotypic assemblies. investigations to determine the effect of composite matrices, spiked with esterifi ed hyaluronic acid and gelatin, to augment osteochondral differentiation of cultured, bone marrow-derived mesenchymal progenitor cells and the effects of the matrix on cellular differentiation have been examined in vitro and in vivo. briefl y, empty and populated matrices cultured for days, with and without tgf-β demonstrated the following results. cells implanted in the matrix produced a robust type ii collagen extracellular matrix in vitro. matrices placed in immunodefi cient mice yielded no differentiation in empty constructs, osteochondral differentiation in loaded implants, and an enhanced level of differentiation in preimplantation in vitro-cultured matrices containing tgf-β . these results demonstrate the utility of three-dimensional matrix for presentation of bone mesenchymal progenitor cells in vivo for repair of cartilage and bone defects as well as indicate the efficacy for in vitro tissue engineering regimes. these techniques lend themselves to microgravity and ground-based research tissue cultures alike. many earth-based laboratories are researching and developing hemopoietic bone marrow cultures of stem cell origin, and three-dimensional confi gurations are providing promising results as illustrated by schoeters and coworkers. they report that murine bone marrow cells, cultured under long-term hemopoietic conditions, produce mineralized tissue and bone matrix proteins in vitro but only when precipitated by the presence of adherent bone stroma cells in three-dimensional collagen matrices. at a concentration of × l stromal cells, mineralization occurs in days. in contrast, twodimensionally oriented marrow fragments at × cells require requires more than days before mineralization can similarly be detected. two-dimensional long-term marrow culture facilitates and enhances expansion of the stromal component and rudimentary differentiation of osteogenic-like cells in the adherent stromal layer as verifi ed by type i collagen or cells positive for alkaline phosphatase. production of osteonectin and osteocalcin, a bone-specifi c protein, combined with calcifi cation is observed only in threedimensional cultures. these studies demonstrate the need for and benefi t of three-dimensionality and the application to the microgravity environment. as we can see, this further reinforces the quest for threedimensionality and the potential of modeling the microgravity environment. investigations clearly show the need for the application of three-dimensional study techniques in lung pathophysiologic studies. interestingly, three-dimensional biology has facilitated full-scale investigations into most areas of tissue engineering, cell biology and physiology, immunology, and cancer research. anchorage-dependent cells are widely cultured on microcarriers. studies show that for the purposes of improved surface-to-volume ratio and scale up, the microcarrier suspension culture provides excellent potential for high-density cell growth. in addition, microcarriers serve well as structural supports for three-dimensional assembly, the composite of which is the basis for threedimensional tissue growth. conventional culture systems for microcarrier cultures (i.e., bioreactors) use mechanical agitation to suspend microcarriers and thus induce impeller strikes as well as fl uid shear and turbulence at the boundary layer between the wall and the fl uid. investigators have attempted to make a complete study of the most effi cient bioreactor designs and agitation regimens. they concluded that virtually all stirred-tank bioreactors operate in the turbulent regimen. it has been demonstrated that bead-to-bead bridging of cells is enhanced signifi cantly at lower agitation rates in a stirred reactor. excessive agitation from either stirring or gas bubble sparging has been documented as a cause of cell damage in microcarrier cell cultures. , to overcome the problems induced by these mechanisms, investigators developed alternative culture techniques such as porous microcarriers to entrap cells, increased viscosity of culture medium, bubble-free oxygenation, and improved methods for quiescent inoculation. , these steps decreased the damage attributed to turbulence and shear forces but failed to signifi cantly rectify the problems. reactor systems of substantially increased volume exhibit less agitation-related cell damage. this is presumably because of the decreased frequency of cell-microcarrier contact with the agitation devices in the systems. research-scale investigations do not afford the luxury of experimenting with large-scale production systems. therefore, if a large-volume system is indeed more quiescent, an improved bioreactor system should emulate the fl uid dynamics present in the upper regions of large-scale reactors in which cells and microcarriers reside with minimal agitation. the problem, then, is to suspend microcarriers and cells without inducing turbulence or shear while providing adequate oxygenation and nutritional replenishment. the term rotating-wall vessel comprises a family of vessels, batch fed and perfused, that embody the same fl uid dynamic operating principles. these principles are ( ) solid body rotation about a horizontal axis that is characterized by (a) colocation of particles of different sedimentation rates, (b) extremely low fl uid shear stress and turbulence, and (c) three dimensional spatial freedom; and ( ) oxygenation by active or passive diffusion to the exclusion of all but dissolved gasses from the reactor chamber, yielding a vessel devoid of gas bubbles and gas-fl uid interface (zero head space). , three-dimensional models of lung disease current cell culture models have shortcomings resulting in unreliable tumor growth, uncharacteristic tumor development, nonhuman tumors, and inadequate methods of detection. cells propagated under traditional culture conditions differ widely in their expression of differentiated markers, adhesion receptors, and growth factor receptors compared with cells in situ or those grown as tissue-like structures. , this is of concern because the phenotypic changes leading to malignant transformation often stem from alterations in the balanced and multifaceted roles of growth factors, receptors, and cytokines (reviewed by herlyn et al. ). with increasing evidence of the importance of adhesive contacts, paracrine cross-talk between different cell types, and signaling cascades that link the cell with a complex substratum, there is now recognition that models must be developed that better simulate these complexities. there is still much to learn about the dynamic relationships among the different phenotypes found in the normal lung and in lung cancers. until a cell culture system is developed that allows differentiation to occur, it is diffi cult to make any fi rm statement about relating effects in cell culture to clinical practice. tissue engineering is very embryonic in development and currently nearly universally focused on building replacement tissues. a new technology developed at the nasa johnson space center used to study colon cancer has been adapted to three-dimensional in vitro lung tissue culture models but has not been reported on to date. rotating-wall vessels are horizontally rotating cylindrical tissue culture vessels that provide controlled supplies of oxygen and nutrients with minimal turbulence and extremely low shear. these vessels suspend cells and microcarriers homogeneously in a nutrient-rich environment, which allows the three-dimensional assembly of cells to tissue. prior to seeding rotating-wall vessels (synthecon, inc, houston, tx), cells were cultured in standard t fl asks (corning, corning, ny) in gtsf- medium ( psebm) in a humidifi ed °c, % co incubator. the rotating-wall vessels were seeded with - mg/ml cultispher-gl microcarriers (hyclone laboratories, inc., logan, ut) followed by beas -b or bzr-t cells (atcc, baltimore, md) at a density of × cells/ml. cultures were grown in the rotating-wall vessels for - days for formation of -to -mm diameter tumor masses. rotating-wall vessel rotation was initiated at rpm and increased as aggregate size became larger. stationary control cultures were initiated under the same conditions using fep tefl on bags (american fluoroseal, columbia, md). at -hour intervals ph, dissolved co , and dissolved o were determined using a corning model clinical blood gas analyzer. glucose concentration was determined using a beckman model clinical glucose analyzer (beckman, fullerton, ca). cell samples were harvested every hr and fi xed with omnifi x (xenetics, tustin, ca) for immunohistochemistry or fi xed with % glutaraldehyde/ % paraformaldehyde in . m cacodylic buffer (electron microscopy sciences, fort washington, pa) for scanning electron microscopy. cancer models already developed by nasa investigators include growth and differentiation of an ovarian tumor cell line, - growth of colon carcinoma lines, and three-dimensional aggregate and microvillus formation in a human bladder carcinoma cell line. in support as an appropriate model for cancer, even the most rudimentary three-dimensional cellular structures exhibit different phenotypes than cell lines cultured under twodimensional conditions. properties such as responses to tgf-β, drug resistance to cisplatin or cyclophosphamide, and resistance to apoptosis are all altered in various types of cell aggregates. many investigations sustain consistent evidence that cells growing in three-dimensional arrays appear more resistant to cytotoxic chemoagents than cells in monolayer culture. li et al. found that spheroids were more resistant to cytosine arabinoside by -fold and methotrexate by -fold when compared with single cell suspensions. further monolayer cultures of colon carcinoma cells were sensitive to piericidin c in contrast to responses within in vivo colon tumors or three-dimensional slices of tumors grown in vitro. numerous other investigations have revealed increased levels of drug resistance of spheroids compared with single cell monolayers. , questions of poor diffusion and insuffi cient drug absorption within spheroids and a relatively frequent high proportion of resting cells have clouded differences in drug resistance, which could be the result of nutrient deprivation and hypoxia. heppner and colleagues executed precise experiments that confi rmed threedimensional structure and function as the causative agent and was responsible for drug resistance rather than simple inaccessibility to nutrients or the drug concentration. heppner embedded tumor specimens or cell aggregates in collagen gels, exposed the culture to various cytotoxic drugs, and compared the drug responses of the same cells in monolayers. these experiments revealed an increased resistance in the three-dimensional tumor arrays of a remarkable , -fold greater than in monolayer cultures, and a similar result was seen in three-dimensional histocultures in collagen. the tumor cells grew in the presence of drug concentrations that rendered monolayers to a viability less than . % of control cultures. amazingly, heppner observed that the cells became sensitive again when replated as monolayers and fi nally showed that even when exposed to melphalan and -fl uorouracil in monolayer cells transferred to collagen gels were again resistant based on three-dimensional architecture. thus, the cells were exposed to the drugs as monolayers, facilitating access to the drugs, and, once the cells were transferred after drug exposure to a threedimensional structure, high resistance to the drugs was sustained. , [ ] [ ] [ ] [ ] based on the caliber of data referenced above, teicher et al. serially passaged through multiple ( ) transfers emt- tumors in mice that were treated with thiotepa, cisplatin, and cyclophosphamide over a prolonged month period, thus producing extremely drug-resistant tumors in vivo. when these tumors were grown as monolayer cultures, they were as drug sensitive as the parental cells. kobayashi and colleagues grew the same in vivo drug-resistant tumor cell lines as spheroids in threedimensional arrays, and resistance was almost , times that of the parent line with selected drugs, an example being the active form of cyclophosphamide used in vitro. similarly extreme resistance was also observed to cisplatin and thiotepa. this resistance was not seen in monolayer cultures, even when the monolayers were cultured on traditional extracellular matrix substrates. these experiments reconfi rmed that cells in a three-dimensional array are more drug resistant than monolayer cells in vitro and demonstrated that three-dimensional cellular confi gurations can and do become resistant to super pharmacologic doses of drugs by forming compact structures. rotating-wall vessel tumor models several important human tumor models have been created in rotating-wall vessel cultures, specifi cally, lung, prostate, colon, and ovarian. , , , many of these models involve cancers that are leading killers in our society. we present two such examples in this section, colon and prostate carcinoma. as previously reviewed, the literature indicates the remarkable difference between chemotherapeutic cytotoxicity in two-dimensional and three-dimensional cellular constructs, which may be predicated on a number of criteria. therefore, a threedimensional tumor model that emulates differentiated in vivo-like characteristics would provide unique insights into tumor biology. goodwin et al. detail the fi rst construction of a complex three-dimensional ex vivo tumor in rotatingwall vessel culture composed of a normal mesenchymal base layer (as would be seen in vivo) and either of two established human colon adenocarcinoma cell lines, ht- , an undifferentiated line, and ht- km a stable, moderately differentiated subline of ht- . each of these engineered tumor tissues produced tissue-like aggregates (tlas) with glandular structures, apical and internal glandular microvilli, tight intercellular junctions, desmosomes, cellular polarity, sinusoid development, internalized mucin, and structural organization akin to normal colon crypt development. necrosis was minimal throughout the tissue masses up to days of culture while achieving > . cm in diameter. other notable results included enhanced growth of neoplastic colonic epithelium in the presence of mixed normal human colonic mesenchyme. these results mimic the cellular differentiation seen in vivo and are similar to results obtained with other tumor types. prostate carcinoma has also been modeled in the rotating-wall vessel system by several investigators. [ ] [ ] [ ] one of the most comprehensive descriptions of these engineered tissues is detailed by wang et al. in that review, the authors describe the ability of the rotatingwall vessel system to recapitulate human prostate carcinoma (lncap) and bone stroma (mg ) to illuminate the evolution of prostate tumorigenesis to the metastatic condition. in particular, the lncap and arcap models represented in the review are known to be lethal in the human, being androgen independent and metastatic. rotating-wall vessel tla engineering also allowed indepth study of epithelial and stromal interactions, which are the facilitating elements of the continuance of lncap prostate-specifi c antigen production in vitro. when lncap was cultured in three dimensions without stroma, production of prostate-specifi c antigen ceased and metastatic markers were not observed. the authors outline the process of malignant transformation, demonstrating that these metastatic models are only possible in threedimensional tlas and are achieved by specifi c geometric relationships in three-dimensional confi guration. furthermore, they show through direct comparison with other culture systems the advantages of the rotating-wall vessel system to allow synergistic relationships to study this disease state. unlike two-dimensional models, these rotating-wall vessel tumor tissues were devoid of metabolic and nutrient defi ciencies and demonstrated in vivo-like architecture. these data suggest that the rotating-wall vessel affords a new model for investigation and isolation of growth, regulatory, and structural processes within neoplastic and normal tissues. in this section, we explore the utility of rotating-wall vessel tlas as targets for microbial infection and disease. several studies have been conducted recently that indicate that three-dimensional tissues respond to infective agents with greater fi delity and with a more in vivo-like response than traditional two-dimensional cultures. nickerson et al. describe the development of a threedimensional tla engineered from int- cells of the human small intestine, which were used as targets for the study of salmonella typhimurium. in this study, threedimensional tlas were used to study the attachment, invasion, and infectivity of salmonella into human intestinal epithelium. immunocytochemical characterization and scanning and transmission electron microscopic analyses of the three-dimensional tlas revealed that the tlas more accurately modeled human in vivo differentiated tissues than did two-dimensional cultures. the level of differentiation in the int- tlas was analogous to that found in previously discussed small intestine tlas and from other organ tissues reconstructed in rotatingwall vessels. analysis of the infectivity studies revealed salmonella attached and infected in a manner significantly different from that in control two-dimensional cultures. during an identical exposure period of infection with salmonella, the three-dimensional tlas displayed a minor loss of structural integrity when compared with the two-dimensional int- cultures. furthermore, salmonella demonstrated a greatly reduced ability to adhere, invade, and induce the apoptotic event in these int- three-dimensional tlas than in twodimensional cultures. this result is not unlike the in vivo human response. two-dimensional cultures were significantly damaged within several hours of contact with the bacteria; conversely, although "pot marks" could be seen on the surfaces of the three-dimensional tlas, they remained structurally sound. cytokine analysis and expression postinfection of three-dimensional tlas and two-dimensional cultures with salmonella exhibited remarkable differences in expressed levels of interleukin (il)- α, il- β, il- , il- ra, and tumor necrosis factor-α mrnas. additionally, noninfected three-dimensional tlas constitutively demonstrated elevated levels of tgf-β mrna and prostaglandin e compared with noninfected two-dimensional cultures of int- . as previously stated, traditional two-dimensional cell monolayers lack adequate fi delity to emulate the infection dynamics of in vivo microbial adhesion and invasion. the respiratory epithelium is of critical importance in protecting humans from disease. exposed to the environment, the respiratory epithelium acts as a barrier to invading microbes present in the air, defending the host through a multilayered complex system. the three major layers of the human respiratory epithelium are pseudostratifi ed epithelial cells, a basement membrane, and underlying mesenchymal cells. ciliated, secretory, and basal epithelial cells are connected by intercellular junctions and anchored to the basement membrane through desmosomal interactions. together with tight junctions and the mucociliary layer, the basement membrane maintains the polarity of the epithelium and provides a physical barrier between the mesenchymal layer and the airway. , infi ltrating infl ammatory and immune cells move freely between the epithelial and subepithelial compartments. airway epithelial cells play a vital role in host defense by blocking paracellular permeability and modulating airway function through cellular interactions. ciliated epithelial cells block invasion of countless inhaled microorganisms by transporting them away from the airways. as regulators of the innate immune response, epithelial cells induce potent immunomodulatory and infl ammatory mediators such as cytokines and chemokines that recruit phagocytic and infl ammatory cells that remove microbes and enhance protection. , , , ideally, cell-based models should reproduce the structural organization, multicellular complexity, differentiation state, and function of the human respiratory epithelium. immortalized human epithelial cell lines, such as beas- b, primary normal human bronchial epithelial cells, and air-liquid interface cultures, are used to study respiratory virus infections in vitro. traditional monolayer cultures (two-dimensional) of immortalized human bronchoepithelial cells represent homogenous lineages. although growing cells in monolayers is convenient and proliferation rates are high, such models lack the morphology and cell-cell and cell-matrix interactions characteristic of human respiratory epithelia. thus, their state of differentiation and intracellular signaling pathways most likely differ from those of epithelial cells in vivo. primary cell lines of human bronchoepithelial cells provide a differentiated model similar to the structure and function of epithelial cells in vivo; however, this state is short lived in vitro. , air-liquid interface cultures of primary human bronchoepithelial cells (or submerged cultures of human adenoid epithelial cells) are grown on collagen-coated fi lters in wells on top of a permeable fi lter. these cells receive nutrients basolaterally, and their apical side is exposed to humidifi ed air. the result is a culture of well-differentiated heterogeneous (ciliated, secretory, basal) epithelial cells essentially identical to airway epithelium in situ. , although this model shows fi delity to the human respiratory epithelium in structure and function, maintenance of consistent cultures is not only diffi cult and time consuming but also limited to small-scale production and thus limits industrial research capability. true cellular differentiation involves sustained complex cellular interactions [ ] [ ] [ ] in which cell membrane junctions, extracellular matrices (e.g., basement membrane and ground substances), and soluble signals (endocrine, autocrine, and paracrine) play important roles. [ ] [ ] [ ] [ ] this process is also infl uenced by the spatial relationships of cells to each other. each epithelial cell has three membrane surfaces: a free apical surface, a lateral surface that connects neighboring cells, and a basal surface that interacts with mesenchymal cells. recently viral studies by goodwin et al. and suderman et al. were conducted with rotating-well vessel-engineered tla models of normal human lung. this model is composed of a coculture of in vitro threedimensional human bronchoepithelial tlas engineered using a rotating-wall vessel to mimic the characteristics of in vivo tissue and to provide a tool to study human respiratory viruses and host-pathogen cell interactions. the tlas were bioengineered onto collagen-coated cyclodextran beads using primary human mesenchymal bronchial-tracheal cells as the foundation matrix and an adult human bronchial epithelial immortalized cell line (beas- b) as the overlying component. the resulting tlas share signifi cant characteristics with in vivo human respiratory epithelium, including polarization, tight junctions, desmosomes, and microvilli. the presence of tissuelike differentiation markers, including villin, keratins, and specifi c lung epithelium markers, as well as the production of tissue mucin, further confi rm these tlas differentiated into tissues functionally similar to in vivo tissues. increasing virus titers for human respiratory syncytial virus (wtrsva ) and parainfl uenza virus type (wtpiv js) and the detection of membrane-bound glycoproteins (f and g) over time confi rm productive infections with both viruses. viral growth kinetics up to day pi with wtrsva and wtpiv js were as follows: wtpiv js replicated more effi ciently than wtrsva in tlas. peak replication was on day for wtpiv js (approximately log particle forming units [pfu] per milliliter) and on day for wtrsva (approximately log pfu/ml). viral proliferation remained high through day when the experiments were terminated. viral titers for severe acute respiratory syndrome-coronavirus were approximately log pfu/ml at day pi. human lung tlas mimic aspects of the human respiratory epithelium well and provide a unique opportunity to study the host-pathogen interaction of respiratory viruses and their primary human target tissue independent of the host's immune system, as there can be no secondary response without the necessary immune cells. these rotating-wall vessel-engineered tissues represent a valuable tool in the quest to develop models that allow analysis and investigation of cancers and infectious disease in models engineered with human cells alone. we have explored the creation of three-dimensional tlas for normal and neoplastic studies and fi nally as targets for microbial infections. perhaps carrel and leighton would be fascinated to know that from their early experiments in three-dimensional modeling and the contributions they made has sprung the inventive spirit to discover a truly space age method for cellular recapitulation. the biochemical basis of pulmonary function the pathology of bronchial asthma three-dimensional histoculture: origins and applications in cancer research histoculture of human breast cancers studies on the propagation in vitro of poliomyelitis viruses. iv. viral multiplication in a stable strain of human malignant epithelial cells (strain hela) derived from an epidermoid carcinoma of the cervix effects of paramyxoviral infection on airway epithelial cell foxj expression, ciliogenesis, and mucociliary function culture of animal cells: a manual of basic technique preclinical models of lung cancer: cultured cells and organ culture human papillomavirus and the development of cervical cancer: concept of carcinogenesis relationship of alveolar epithelial injury and repair to the induction of pulmonary fi brosis generation of human pulmonary microvascular endothelial cell lines promotion of mitochondrial membrane complex assembly by a proteolytically inactive yeast lon culture and transformation of human airway epithelial cells a mechanism of hyperthermia-induced apoptosis in ras-transformed lung cells smoke/burn injury-induced respiratory failure elicits apoptosis in ovine lungs and cultured lung cells, ameliorated with arteriovenous co removal clonal growth of epithelial cells from normal adult human bronchus experimental systems for mechanistic studies of toxicant induced lung infl ammation regulation of vasodilator synthesis during lung development markers of airway smooth muscle cell phenotype airway smooth muscle cell proliferation: characterization of subpopulations by sensitivity to heparin inhibition divergent differentiation paths in airway smooth muscle culture: induction of functionally contractile myocytes airway wall remodelling and hyperresponsiveness: modelling remodelling in vitro and in vivo nci series of cell lines: an historical perspective establishment and characterisation of cell lines from patients with lung cancer (predominantly small cell carcinoma) cell culture methods for the establishment of the nci series of lung cancer cell lines extracellular matrix proteins protect small cell lung cancer cells against apoptosis: a mechanism for small cell lung cancer growth and drug resistance in vivo radioimmunoassay of pulmonary surface-active material in the tracheal fl uid of the fetal lamb selected anatomic burn pathology review for clinicians and pathologists overview of ventilatorinduced lung injury mechanisms surfactant protein-a levels in patients with acute respiratory distress syndrome positive end-expiratory pressure modulates local and systemic infl ammatory responses in a sepsis-induced lung injury model injurious mechanical ventilation and end-organ epithelial cell apoptosis and organ dysfunction in an experimental model of acute respiratory distress syndrome alveolar infl ation during generation of a quasi-static pressure/volume curve in the acutely injured lung evaluation of mechanical ventilation for patients with acute respiratory distress syndrome as a result of interstitial pneumonia after renal transplantation primary cell culture of human type ii pneumonocytes: maintenance of a differentiated phenotype and transfection with recombinant adenoviruses collagen-coated cellulose sponge: three dimensional matrix for tissue culture of walker tumor histophysiologic gradient culture of stratifi ed epithelium the three-dimensional question: can clinically relevant tumor drug resistance be measured in vitro? structural biology of epithelial tissue in histophysiologic gradient culture tissue culture in relation to growth and differentiation the growth, development and phosphatase activity of embryonic avian femora and end-buds cultivated in vitro organ culture of mucosal biopsies of human small intestine growth of nodular carcinomas in rodents compared with multicell spheroids in tissue culture dialysis fermentor systems for concentrated culture of microorganisms semipermeable microcapsules application of ultrafi ltration for enzyme retention during continuous enzymatic reaction cell culture on artifi cial capillaries: an approach to tissue growth in vitro membranes and bioreactors: a technical challenge in biotechnology membrane bioreactors: engineering aspects membrane bioreactors: present and prospects artifi cial membrane as carrier for the immobilization of biocatalysts hollow fi ber cell culture in industry hollow fi ber enzyme reactors engineering of osteochondral tissue with bone marrow mesenchymal progenitor cells in a derivatized hyaluronan-gelatin composite sponge haemopoietic long-term bone marrow cultures from adult mice show osteogenic capacity in vitro on -dimensional collagen sponges microcarrier/cultures of animal cells mammalian cell culture: engineering principles and scale-up morphologic differentiation of colon carcinoma cell lines ht- and ht- km in rotating-wall vessels hydrodynamic effects on animal cells in microcarrier cultures physical mechanisms of cell damage in microcarrier cell culture bioreactors growth and death in over agitated microcarrier cell cultures cell death in the thin fi lms of bursting bubbles growth of anchoragedependent cells on macroporous microcarriers viscous reduction of turbulent damage in animal cell culture biological experiences in bubble-free aeration system critical parameters in the microcarrier culture of animal cells the large-scale cultivation of mammalian cells analysis of gravity-induced particle motion and fl uid perfusion fl ow in the nasa-designed rotating zero-head-space tissue culture vessel cell culture for threedimensional modeling in rotating-wall vessels: an application of simulated microgravity autocrine and paracrine roles for growth factors in melanoma growth-regulatory factors for normal, premalignant, and malignant human cells in vitro reduced shear stress: a major component in the ability of mammalian tissues to form three-dimensional assemblies in simulated microgravity threedimensional culture of a mixed mullerian tumor of the ovary: expression of in vivo characteristics threedimensional modeling of t- human bladder carcinoma cell line: a new simulated microgravity vessel multicellular resistance: a new paradigm to explain aspects of acquired drug resistance of solid tumors comparison of cytotoxicity of agents on monolayer and spheroid systems modifi ed -tumour (l , colon ) assay to screen for solid tumor selective agents signifi cance of three-dimensional growth patterns of mammary tissues in collagen gels assessing tumor drug sensitivity by a new in vitro assay which preserves tumor heterogeneity and subpopulation interactions factors affecting growth and drug sensitivity of mouse mammary tumor lines in collagen gel cultures origin and deposition of basement membrane heparan sulfate proteoglycan in the developing intestine tumor resistance to alkylating agents conferred by mechanisms operative only in vivo acquired multicellular-mediated resistance to alkylating agents in cancer heat sterilisation to inactivate aids virus in lyophilised factor viii threedimensional growth patterns of various human tumor cell lines in simulated microgravity of a nasa bioreactor long term organ culture of human prostate tissue in a nasa-designed rotating wall bioreactor establishment of a three-dimensional human prostate organoid coculture under microgravity-simulated conditions: evaluation of androgen-induced growth and psa expression three-dimensional coculture models to study prostate cancer growth, progression, and metastasis to bone threedimensional tissue assemblies: novel models for the study of salmonella enterica serovar typhimurium pathogenesis series introduction: innate host defense of the respiratory epithelium the airway epithelium: structural and functional properties in health and disease apicobasal polarization: epithelial form and function robbins infectious diseases, th ed. philadelphia: wb saunders epithelial regulation of innate immunity to respiratory syncytial virus epithelia cells as regulators of airway infl ammation human bronchial epithelial cells with integrated sv virus t antigen genes retain the ability to undergo squamous differentiation identifi cation and culture of human bronchial epithelial cells developing differentiated epithelial cell cultures: airway epithelial cells evaluation of anchorageindependent proliferation in tumorigenic cells using the redox dye alamar blue airway epithelium and mucus: intracellular signaling pathways for gene expression and secretion disorganization of stroma alters epithelial differentiation of the glandular stomach in adult mice expression and function of cell surface extracellular matrix receptors in mouse blastocyst attachment and outgrowth milk protein expression and ductal morphogenesis in the mammary gland in vitro: hormone-dependent and -independent phases of adipocyte-mammary epithelial cell interaction cell replication of mesenchymal elements in adult tissues. i. the replication and migration of mesenchymal cells in the adult rabbit dermis defi ning conditions to promote the attachment of adult human colonic epithelial cells laminin expression in colorectal carcinomas varying in degree of differentiation infl uence of mammary cell differentiation on the expression of proteins encoded by endogenous balb/c mouse mammary tumor virus genes opinion: building epithelial architecture: insights from three-dimensional culture models threedimensional engineered high fi delity normal human lung tissue-like assemblies (tla) as targets for human respiratory virus infection severe acute respiratory syndrome (sars)-cov infection in a threedimensional human bronchial-tracheal (hbte) tissue-like assembly key: cord- - wyebaxb authors: kurahashi, setsuya title: an agent-based infectious disease model of rubella outbreaks date: - - journal: agents and multi-agent systems: technologies and applications doi: . / - - - - _ sha: doc_id: cord_uid: wyebaxb this study proposes a simulation model of rubella. sir (susceptible, infected, recovered) model has been widely used to analyse infectious diseases such as influenza, smallpox, bioterrorism, to name a few. on the other hand, agent-based model begins to spread in recent years. the model enables to represent the behaviour of each person on the computer. it also reveals the spread of infection by simulation of the contact process among people in the model. the study designs a model based on smallpox and ebola fever model in which several health policies are decided such as vaccination, the gender-specific workplace and so on. the infectious simulation of rubella, which has not yet vaccinated completely for men in japan, is implemented in the model. as results of experiments using the model, it has been found that preventive vaccine to all the men is crucial factors to prevent the spread in women. infectious diseases have been serious risk factors in human societies for centuries. smallpox has been recorded in human history since more than b.c . people have also been suffering from many other infectious diseases such as malaria, cholera, tuberculosis, typhus, aids, influenza, etc. although people have tried to prevent and hopefully eradicate them, a risk of unknown infectious diseases including sars, a new type of infectious diseases, as well as ebola haemorrhagic fever and zika fever have appeared on the scene. a model of infectious disease has been studied for years. sir (susceptible, infected, recovered) model has been widely used to analyse such diseases based on a mathematical model. after an outbreak of sars, the first sir model of sars was published and many researchers studied the epidemic of the disease using this model. when an outbreak of a new type of influenza is first reported, the u.s. government immediately starts an emergency action plan to estimate parameters of its sir model. nevertheless, the sir model has difficulty to analyse which measures are effective because the model has only one parameter to represent infectiveness. for example, it is difficult for the sir model to evaluate the effect of temporary closing of classes because of the influenza epidemic. the agent-based approach or the individual-based approach has been adopted to conquer these problems in recent years [ ] [ ] [ ] [ ] . the model enables to represent behaviour of each person. it also reveals the spread of an infection by simulation of the contact process among people in the model. in this study, we developed a model to simulate rubella based on the infectious disease studies using agent-based modelling. what we want to know is how to prevent an epidemic of infectious diseases not only using mechanisms of the epidemic but also decision-making of health policy [ ] . we aim to study the relationship between antibody holding rate of men and the spread of infection by constructing infection of rubella virus with the agent-based model and repeating simulation experiment on a computer. although our previous study described the infectious disease model of smallpox and ebola [ ] , this paper proposes a new model of rubella which has caused crucial problems for pregnant women in recent years. in sect. , as examples of infections that occurred in the past, we will explain smallpox, ebola hemorrhagic fever, zika fever and rubella. section describes related research on infectious disease models. section explains the basic model and describes the rubella model. section explains the experimental results and sect. discusses them. finally, we summarize the whole in sect. . the smallpox virus affects the throat where it invades into the blood and hides in the body for about days. patients developed a high fever after that, but rashes do not appear until about days after the infection. while not developing rashes, smallpox virus is able to infect others. after days, red rashes break out on the face, arms and legs, and subsequently they spread over the entire body. when all rashes generate pus, patients suffer great pains; finally, % of patients succumb to the disease. for thousands of years, smallpox was a deadly disease that resulted in thousands of deaths. a source of ebola infection is allegedly by eating a bat or a monkey, but it is unknown whether eating these animals is a source of the infection. due to the recent epidemic, which began in guinea in december , , deaths have been confirmed. the authorities of guinea, liberia and sierra leone have each launched a state committee of emergency and have taken measures to cope with the situation. the prohibition of entry over the boundary of guinea is included in these measures. zika fever is an illness caused by zika virus via the bite of mosquitoes. it can also be potentially spread by sex according to recent report [ , ] . most cases have no symptoms and present are usually mild including fever, red eyes, joint pain and a rash [ ] , but it is believed that the zika fever may cause microcephaly which severely affects babies by a small head circumference. rubella is a type of viral infection caused by the rubella virus [ , ] . in japan, there were epidemics ( , , , ) once every years, but after a male and female infant was the subject of periodic vaccination, no big epidemic occurred. however, in , , people outbreaks were estimated and ten congenital rubella syndromes were reported. a large epidemic occurred in asia in , and from to , an epidemic exceeding , cases occurred mainly in adult males who did not take the vaccine [ ] . the epidemic recurred in , as of october the number of annual infections of rubella was about people, the national institute of infectious diseases announced emergency information on rubella epidemics. the centers for disease control and prevention (u.s.) raised the rubella alert level in japan to the second 'recommendation' among the three levels [ ] . they recommended that pregnant women who are not protected against rubella through either vaccination or previous rubella infection should not travel to japan during this outbreak. epstein [ , ] made a smallpox model based on epidemics in europe from to . in the model, families from two towns were surveyed. the family includes two parents and two children, thus the population is each from each town. all parents go to work in their town during the day except % of adults who go to another town. all children attend school. there is a communal hospital serving the two towns in which each five people from each town work. this model was designed as an agent-based model, and then simulation of infectious disease was conducted using the model. as results of experiments showed that ( ) in a base model in which any infectious disease measures were not taken, the epidemic spread within days and % of people died, ( ) a trace vaccination measure was effective but it was difficult to trace all contacts to patients in an underground railway or an airport, ( ) a mass vaccination measure was effective, but the number of vaccinations would be huge so it was not realistic and ( ) epidemic quenching was also effective, and reactive household trace vaccination along with pre-emptive vaccination of hospital workers showed a dramatic effect. ohkusa [ ] evaluated smallpox measures using an individual-based model of infectious diseases. the model supposed a town including , habitats and a public health centre. in the model, one person was infected with smallpox virus at a shopping mall. they compared between a trace vaccination measure and a mass vaccination measure. as a result of simulation, it was found that the effect of trace vaccination dropped if the early stage of infection was high and the number of medical staff is small, while the effect of mass vaccination was stable. therefore, timely and concentrate mass vaccination is required when virus starts spreading. the estimation about the number, place and time of infection is needed quickly and the preparation of an emergency medical treatment and estimation system is required for such occasions. regarding measles epidemics, agent-based simulation models of measles transmission have been developed using the framework for reconstructing epidemiological dynamics, a data-driven agent-based model to simulate the spread of an airborne infectious disease in an irish town, and so on [ , ] . from these studies, the effectiveness of an agent-based model has been revealed, yet these are not sufficient models to consider a relationship between antibody holding rate of men and women, and commutation routes and the gender-specific workplace. in addition, authorities need to make a decision regarding measles-rubella mixed (mr) vaccine to men. this study takes into account these extensions. we designed a health policy simulation model of infectious disease based on epstein's smallpox model. the model includes smallpox, ebola haemorrhagic fever and rubella. we assume all individuals to be susceptible which means no background of immunity. families live in two towns. the family includes two parents and two children. therefore, the population is each in each town. all parents go to work in their town during the day except % of adults commute to another town. all children attend school. there is a communal hospital serving two towns in which five people from each town work. each round consists of an interaction through the entire agent population. the call order is randomized each round and agents are processed or activated, serially. on each round, when an agent is activated, she identifies her immediate neighbours for interaction. each interaction results in a contact. in turn, that contact results in a transmission of the infection from the contacted agent to the active agent with probability. the probability of contact at an interaction is . at a workplace and a school, while . at a home and a hospital. the probability of infection at a contact is . at a workplace and a school, while . at a home and a hospital. in the event the active agent contracts the disease, she turns blue to green and her own internal clock of disease progression begins. after days, she will turn yellow and begins infecting others. length of noncontagious period is days, and early rash contagious period is days. unless the infected individual is vaccinated within days of exposure, the vaccine is ineffective. at the end of day , smallpox rash is finally evident. next day, individuals are assumed to hospitalize. after more days, during which they have a cumulative % probability of mortality, surviving individuals recover and return to circulation permanently immune to further infection. dead individuals are coloured black and placed in the morgue. immune individuals are coloured white. individuals are assumed to be twice as infectious during days - as during days - . in the event the active agent contracts the disease, she turns blue to green and her own internal clock of disease progression begins. after days, she will turn yellow and begins infecting others. however, her disease is not specified in this stage. after days, she begins to have vomiting and diarrhoea and the disease is specified as ebola. unless the infected individual is dosed with antiviral medicine within days of exposure, the medicine is ineffective. this is an imaginary medicine to play the policy game. at the end of day , individuals are assumed to hospitalize. after more days, during which they have a cumulative % probability of mortality, surviving individuals recover and return to circulation permanently immune to further infection. dead individuals are coloured black and placed in the morgue. immune individuals are coloured white. other settings are the same as smallpox. rubella is a viral infectious disease characterized by fever and rash. since symptoms range from subclinical to complications, it is difficult to judge as rubella only with symptoms. if a pregnant woman until about weeks of pregnancy infects rubella with a virus, there is a possibility that the baby will develop congenital rubella syndrome (crs). in consequence, congenital heart disease, hearing loss, cataract, pigmentary retinopathy and the like occur as congenital abnormalities. for this reason, the pre-inoculation of the vaccine is extremely important. in japan, however, only junior high school girls were eligible for regular vaccination of rubella vaccine from to . in the past, vaccination was recommended for children under the age of , but due to the frequent occurrence of meningitis caused by the vaccine strains, the use was discontinued after that. thereafter, the national epidemic of measles occurred mainly in the - generations in . 'prevention guideline on specific infectious diseases related to measles' was announced by the ministry of health, labour and welfare, and rubella was also designated as a disease to take measures. and during the years from to , as a periodical inoculation at the first period ( -year-old child), the second period (before elementary school entrance), the third period (first grader of junior high school) and the fourth period (third grader of high school), mr vaccine was to be inoculated. from fiscal , as a rule, measles-rubella mixed (mr) vaccine is inoculated in infants of the first period and children before elementary school entrance of the second period. according to a survey of , females possess about % of antibodies in all ages, while males only possess about % of - years old, and - % of - years old. the antibodies holding rate in middle-aged males stays low. regarding the infection process, fever, rashes and lymphadenopathy (especially the posterior portion of the auricle, the back of the auricle, the neck) appear after a latency period of - days from infection. fever is observed in about half of rubella patients. in addition, inapparent infection is about - %. the process of infection in a rubella model is plotted in fig. . the model employs with the basic parameters to the disease. an orange line and a blue line indicate the number of infected and recovered people, respectively. when a player adopts a basic model, it takes approximately days until convergence of the outbreak and more than people have infected. however, the result of executing the experiment many times is greatly different. figure shows the histogram of the results of runs. the horizontal axis shows the number of infected people, the vertical axis shows occurrence frequency, the blue next, the results when men and women are working separately in the workplace are shown in fig. . as the results show, the frequency of infection by more than men has increased to over % of the total. this result is thought to be caused by a low antibody holding rate of males. however, the total number of women infected has not increased. next, we conducted an experiment with a model that introduced the railway to commute. in this model, adults commute by railway. as the experimental result in fig. shows, the number of infected men has not only increased dramatically but also the number of women infected has increased. in the basic model without the railway, the frequency of infection of one or more women became %, but in the model using the railway, it has increased to %. especially, the case where more than ten women were infected was % or more, which was a severe result. figure shows the experimental results when men and women work separately in a railway model. the result is more serious. the frequency fig. the experimental result of rubella model of the workplace separated the sexes with a railway of infection of one or more women increases to %, and the case where more than ten women were infected was %. it was estimated by the gender-specific workplace model that the low infection rate of middle-aged men as low as % is the cause of infection spread. based on this, we experimented in the case of strengthening medical policy promoting vaccination for males and raising the antibody holding rate to %. figure shows experimental results at the gender-specific workplace, railway use and male antibody holding rate of %. according to this, both males and females, the proportion of infected people who expanded to more than one person has drastically decreased to less than %. by combining other infection prevention measures, it is possible to control the spread of infection of rubella. from these experimental results, the rubella virus infection is not represented by a simple statistical distribution. it is considered that the infection process is a complicated system in which positive feedback works by accidentality of infection route fig. the experimental result of rubella model of the workplace separated the sexes with a railway and % antibody holding rate and interaction between infected people. it is also speculated that the spread of infection will start in a workplace where many men do not possess antibodies. one of the major factors of infection spread is commuters with many opportunities to come into close contact with the unspecified majority in a train. in addition, it became clear that raising the antibody holding rate of males is an important measure to prevent the spread of the whole infection. this study proposes a simulation model of rubella. it also evaluates health policies to prevent an epidemic. as health policies, vaccination to men, avoidance of separated workplace by sexes and avoidance crowds in a train were implemented in the model. as a result of experiments, it has been found that vaccination to middle-aged men and avoidance crowds in a train are crucial factors to prevent the spread of rubella. using a public transportation to commute, however, is inevitable in a modern society. even if % of people including men and women were vaccinated, it would not prevent the epidemic completely. by combining other infection prevention measures, it is possible to control the spread of infection of rubella. individual-based computational modeling of smallpox epidemic control strategies containing a large bioterrorist smallpox attack: a computer simulation approach agent-based models networks, crowds, and markets: reasoning about a highly connected world risk and benefit of immunisation: infectious disease prevention with immunization a health policy simulation model of smallpox and ebola haemorrhagic fever zika virus: rapid spread in the western hemisphere zika virus spreads to new areas -region of the americas who: zika virus, world health organization, media centre, fact sheets national institute of infectious diseases: fiscal year rubella immunization status and status of antibody retention -survey on infectious disease epidemic survey in (provisional result) rubella in japan toward a containment strategy for smallpox bioterror: an individual-based computational approach generative social science: studies in agent-based computational modeling an evaluation of counter measures for smallpox outbreak using an individual based model and taking into consideration the limitation of human resources of public health workers the role of vaccination coverage, individual behaviors, and the public health response in the control of measles epidemics: an agent-based simulation for california an open-data-driven agent-based model to simulate infectious disease outbreaks key: cord- - farxcc authors: koibuchi, yukio; sato, shinji title: numerical simulation of urban coastal zones date: journal: advanced monitoring and numerical analysis of coastal water and urban air environment doi: . / - - - - _ sha: doc_id: cord_uid: farxcc nan water quality is directly and indirectly influenced by a variety of flows. for example, since phytoplankton drifts passively according to water flows, it is strongly influenced by the distribution of the flow of the bay (lucas et al. ) . bay water motions also exert a strong influence on other water quality parameters. these kinds of influences are direct and obvious. the degree of influence naturally increases with the strength of the currents. in contrast, even if the flow is very small, it has some spatial patterns in common with outer bay directions. in this case, a nutrient load that is discharged at the head of the bay is transported to distant locations and, probably, the water exchange rate of the bay also increases. the nutrient loading that is permissible -i.e. which the bay can accept -will then increase due to the increase in the water exchange rate. for water retention rates, the strength of currents is not as important as spatial patterns. for example, tidal currents are dominant in urban coastal areas. however they are oscillatory. water particles in tidal currents move to the head of the bay during flood tides, but move back to the mouth of the bay during ebb tides. as a result, tidal currents do not substantially transport water particles. in contrast, density currents are clearly weaker than tidal currents, but they flow in one direction continuously and transport water particles more efficiently. as a result, they substantially effect water retention times and ecosystem characteristics. a water quality model for bays must therefore consist of a threedimensional circulation model and an ecosystem model that describes pelagic and benthic aspect of nutrients cycling. this section focuses on the physical modeling and the next section deals with water quality modeling. the final section discusses the application of these models to tokyo bay. many three-dimensional hydrodynamic models have been developed in the last decades, including pom (princeton ocean model; blumberg and mellor ) , ch d (curvilinear hydrodynamics in dimensions; johnson et al. ) and roms (regional ocean modeling system; maccready et al. ; li et al. ) . these models solve the navier-stokes equation with the forcing (the wind stress, coriolis force and buoyancy force) under adequate approximations that are called the hydrostatic and boussinesq approximations. hydrostatic approximation assumes that there is a perfect balance between pressure gradients and gravity: in other words, no acceleration occurs in a vertical direction. this is justified because the aspect ratio of urban coastal areas is extremely small, and hence the vertical motions are considered to be small and also to be further inhibited by gravitational forces under stable density stratification. this means that vertical acceleration is negligible and the fluid behaves as though it were under static equilibrium as far as vertical motion is concerned (prudman ) . density variations in urban coastal areas are also small (less than % or so), and so density can be considered to be constant, except when body forces resulting from the motion of a density stratified fluid in a gravitational field are concerned. this approximation is called the boussinesq approximation: in other words, changes in the mass or inertia of a fluid body due to the changes in its density are negligible, while the same changes in density are consequential when the gravitational field is present ( kantha and clayson ) . therefore, following boussinesq ( ) , this approximation justifies replacing r by a constant reference density r everywhere except in terms involving gravitational acceleration constant g. under such approximations, the governing equations are transformed as follows: here, t stands for time. u , v, and w are the velocity components in x , y , and z directions. the symbol r′ is the reference density and it is defined as r = r + r′ . the symbols a x , a y and a z are the eddy viscosities in x , y , and z directions. the symbol g is the acceleration due to gravity. all currents including tides, wind-driven currents, and density currents in urban coastal zones are strongly influenced by geometry and bathymetry, whereas these areas are rarely regular in shape. in particular, a coastline near an urban coastal area is more complex, due to reclamations and the constructions of harbors, than a natural one. in addition, the uniformity of the bathymetry further lessened as a result of dredging for vessel transport. a computational grid is required to accurately represent such complex geometry and bathymetry. for this reason, the selection of which grid system to use has varied along with the progress of modeling, although the governing equations are not basically different. for vertical coordinate systems (shown in fig. - ), cartesian ( z -coordinate vertical grid) and sigma-coordinate grids have been widely used. a cartesian grid is easy to understand, and shows the correspondence between program codes and governing equations. it is sometimes more accurate than sigmacoordinate grids, especially if the bathymetry of the bay is simple and mild. the sigma-coordinate system tends to have an error featuring the presence of steep-bottom topography. however, unless an excessively large number of vertical levels are employed, the cartesian grid fails to represent the bottom topography with satisfied accuracy. the sigma-coordinate system is convenient in the sense that it can essentially introduce a "flattening out" mechanism for variable bottoms at z = −h(x, y). the flow near the seabed is also calculated well. moreover, the sigmacoordinate system is easy to program since the number of vertical grids can be the same, and the setting of boundary conditions will be simple. the sigma-coordinate system has long been widely used in both meteorology and oceanography (phillips ; freeman et al. ) . after the incorporation of approximations, the governing equations in the sigma coordinate system are as follows: where, t stands for time. u , v, and s ⋅ are the velocity components in x , y, and z directions in the s ⋅ coordinate system. h is change in surface elevation, h is initial water depth, and h is the total water depth ( h=h + h ). f is the coriolis coefficient and p stands for pressure. r and r′ are the constant reference density and the deviation from it, and r = r + r′ . a h and a v are the horizontal and vertical eddy viscosity coefficients, respectively. g is acceleration due to gravity. recently, the stretched grid system (s-grid system) has also been popular. this grid system is an extension of sigma-coordinate system. generally, a sigma-coordinate grid divides the vertical coordinate into an equal number of points. the s-grid system has a higher resolution near the surface and bottom (haidvogel et al. ) . wind-stress and bottom friction are considered at surface and bottom boundary conditions, respectively. settings in boundary conditions are easy in sigma-coordinate systems since they use the same number of vertical grids. at lateral boundaries, normal velocities are set at zero, and a free slip condition is applied to the friction terms. at open boundary, the velocity gradient is set at zero. horizontal computational grids are also modified to fit topography. the simplest horizontal computational grid is the rectangular grid with fixed spacing. the rectangular grid is equivalent to the cartesian vertical grid. recently, curvilinear coordinate systems have been widely used. these systems allow greater flexibility than rectangular grid systems. fig. - is (from ming et al. ) an example of a horizontal curvilinear coordinate system. the chesapeake bay, like other bays, has a typical complex geometry, and thus a horizontal curvilinear coordinate system is advantageous (li et al. ) . this is extended to a nested grid system in which finer grids are used in regions to yield detailed information. in urban coastal areas, density difference plays an essential role for water quality and currents. one of the most important phenomena induced by the density effect is stratification. once stratification occurs in a coastal zone, surface water and bottom water are isolated. this process is very important when we discuss the distributions of pollutants from the land. stratification also enhances the increase of phytoplankton in the surface layer and oxygen depletion near the seabed. moreover, estuarine circulation is induced by density differences in the salty sea water and the river flow. fig. - shows a schematic diagram of estuarine circulation. river water runs through the urban area and runs off from the river mouth, spreading over the sea surface like a veil since river water has low density compared with saline sea water. to cancel the density difference between the river water and the saline water, a great deal of sea water is entrained into the river water flow. such a mixing process continues until the river water reaches the same density as the surrounding sea water, resulting in vertical circulation in the bays that is is several to ten times greater than the river flux (unoki ) . thus, estuarine circulation induces seaward currents on the surface and landward currents near the bottom. the speed of the currents is slow compared with the tidal currents, as explained previously. since estuarine circulations are in a fixed direction, its material transport is very effective over a long time scale in spite of the small magnitude of its velocity. estuarine circulation also plays an important role in the nutrient cycles of stratified bays. organic matters are deposited on the seabed after phytoplankton blooms or river runoffs. they are decomposed by bacteria in the seabed. these nutrients are supplied from the seabed under anoxic condition in summer. in order to include density effects in the numerical model, conservation equations for temperature and salinity are also included. then, to obtain a realistic prediction for vertical stratification, a turbulent closure model is employed (mellor and yamada ) . consequently, flows driven by various mechanisms -e.g. the gravitational, wind-driven, and topographically induced flows -can be reproduced within physical numerical models. diffusion equations for temperature and salinity in the sigma coordinate system are as follows: here, t and s stand for temperature and salinity, respectively. heat balance and moisture balance at surface are considered as surface boundary condition for temperature and salinity, respectively. c p and q stands for specific heat coefficient and net surface heat flux at the surface, respectively. r stands for river discharge. k h and k v are the horizontal and vertical eddy diffusion coefficients, respectively. the vertical mixing coefficients, v a and v k are obtained by appealing to the second order turbulent closure scheme of mellor and yamada ( ) , which characterizes turbulence by equations for the kinetic energy of turbulence, q , and turbulence macro scale l , according to: r r ( . ) wall proximity functions w is defined as follows: . ) mixing coefficients are given as: the stability functions s m , s h , and s q are analytically derived from algebraic relations. they are functionally dependent on , q and l . these relations are derived from closure hypotheses described by mellor ( ) and later summarized by mellor and yamada ( ) . a semi-implicit finite difference scheme has been adopted where equations are discretized explicitly in the horizontal direction and implicitly in the vertical direction. an arakawa c staggered grid has been used with the first order upwind scheme. the tri-diagonal formation of the momentum equation is utilized, and in combination with the mass conservation equation, an algebraic equation is obtained where the only unknown variable is the surface elevation h in implicit form. this algebraic equation is solved through the successive over relaxation (sor) method. the phenomena in urban coastal zone are not only physical but also biological or chemical, each of which relates to the other. the ecosystems and water quality of urban coastal zones are highly complicated. to deal with these complex systems, a water quality model is composed of a three-dimensional physical circulation model and an ecosystem model that describes pelagic and benthic aspect of nutrients cycling. the pelagic and benthic systems also have interactions with each other. in ecosystem models, each water quality variable is often called a compartment. various kinds of models are also proposed for ecosystem models (kremer and nixon ; fasham et al. ; chai et al. ; kishi et al. ) . some models, such as ce-qual-icm (cerco and cole ) each model is developed with basic aquatic compartments such as phytoplankton, zooplankton, and nutrients. some differences in the modeling of sediment, detritus, and the detailed modeling of phytoplankton exist, depending on the target ecosystems and the objectives of the study. as a result, no fully adaptive model applicable for all water bodies exists. if we were to make a model that could be adapted for all areas, its results would be too complex to discuss. it would not be so different from observing the real world. for example, ocean ecosystem models tend to focus only on pelagic systems. they tend to ignore benthic modeling, since the open ocean is deep enough to prevent the return of detritus to the seabed. meanwhile, ocean ecosystem models generally deal with some metals in order to represent the limiting factor of phytoplankton. these metals are fully abundant in urban coastal zones. however, they are often depleted during phytoplankton growth in the open ocean. on the other hand, the concentration of phytoplankton in coastal areas is highly variable both spatially and temporally as compared to the open sea. subsequent sedimentation of this bloom also constitutes a major input to benthic ecology (waite et al. ; matsukawa ; yamaguchi et al. ) . to represent these phenomena, ecosystem models of coastal zones usually cover benthic systems. ecosystem models solve conservation equations for relevant components with appropriate source and sink terms. this is the same as temperature and salinity modeling in physical models, as explained in sect. . . . for the sigma coordinate system, a mathematical formulation of the conservation of mass is written by: where c denotes concentration of the water quality variable and t is time. fluxes into and out of the target control volume are calculated by using physical model results. s(x,y,s,t) represents sources or sinks of the water quality variable due to internal production and the removal of the biogeochemical effect. it also represents the kinetic interactions of each compartment. w(x,y,s,t) represents the external inputs of the variable c . for example, phytoplankton constitutes the first level in the food chain of the pelagic ecosystem of bays. phytoplankton photosynthesizes by using sunlight and increases. at this time, the source term s of phytoplankton is increased depending on the amount of photosynthesis that takes place. phytoplankton is then decreased by the grazing of zooplankton. the source term s of zooplankton is increased along with this grazing, and the source term of phytoplankton is decreased. ecosystem models basically express the relationship of each compartment through mathematical expressions. these models provide a quantitative description of the influences of physical circulation on the biological and chemical processes of urban coastal zones. the ecosystem model introduced here was developed to simulate the nutrient budget of an urban coastal zone. it includes the temporal and spatial variations of phytoplankton, nutrients, detritus, and dissolved oxygen (do). in urban coastal zones, nutrients emitted from urban areas are not a limiting factor for phytoplankton growth. however, quantifying the nutrient budget is essential for analyzing and restoring the ecosystems of urban coastal zones. fig. . shows schematic interactions of a lower trophic ecosystem model which is used for tokyo bay (koibuchi et al. ) this model has state variables: phytoplankton (phy), zooplankton (zoo), nutrients (nh , no , po and si), labile detritus (ldon, ldop, ldosi) and refractory detritus (rdon, rdop, rdosi) for each nutrient, labile detritus carbon (ldoc), refractory detritus carbon (rdoc), dissolved organic carbon (doc), and dissolved oxygen (do), as well as sedimentation processed from particulate organic material. since the basic structure of the model follows the widely applied ce-qual-icm cole , ) , this section mainly focuses on our modifications of the ce-qual-icm model in the following description. this model deals with four phytoplankton groups. phyd is based on skeletonema costatum, which is a dominant phytoplankton species in tokyo bay. phyd represent a winter diatom group (such as eucampia ). phyr is a mixed summer assemblage consisting primarily of heterosigma akashiwo and thalassosira . phyz denotes the dinoflagellates. these four phytoplankton assemblages have different optimal levels of light for photosynthesis, maximum growth rates, optimal temperatures for growth, and half saturation constants for nutrient uptake. diatoms only use silica during growth. the time rate of change of phytoplankton due to biological activity and sink is given by: ( . ) where x = d , d , r, z , denote each phytoplankton assemblage. the phytoplankton growth rate m depends on temperature t , on photosynthetically available radiation i , and on the nutrient concentration of nitrogen, phosphorus, and silica : idealized nutrient cycling in tokyo bay's ecosystem according to the model of koibuchi et al. ( ) . cycling between the state variables: phytoplankton, zooplankton, nutrients (nitrogen, phosphorus, and silicate), labile detritus and refractory detritus for each nutrient, and dissolved oxygen, as well as sedimentation processed from particulate organic materials where m max (t) is the growth rate at ambient temperature, which relates m max , the maximum growth rate m max = m · . t (eppey ) , t opt , the optimal temperature of each plankton assemblage, and b and b are shaping coefficients, k no , k nh , k po and k si , which is the michaelis of the halfsaturation constant for each nutrient. i is exponentially decreasing with water depth z according to: . ) where i is shortwave radiation, and par is the fraction of light that is available for photosynthesis. k w , k chl , and k sal are the light attenuation coefficients for water, chlorophyll, and depth average salinity, respectively. suspended sediment reduces underwater light intensity and affects the growth of phytoplankton. the effect of suspended sediment concentration on light intensity should be simulated as its own compartment. however, the re-suspension rate of mixed mud and the available data on the concentration of sediment suspended in river water and on the seabed are very limited. therefore, we used observation results of salinity based on field observation data from and , as shown in fig. - . the function l(i ) represents the photosynthesis-light relationship (evans and parslow ) , the rate of phytoplankton grazing, g, which is a function of an ambient temperature: where k grz is the predation rate at °c. other phytoplankton loss terms are mortality, represented by the linear rate m p , where w px is the constant vertical sinking velocity for each phytoplankton. the growth rates of zooplankton are expressed as follows: here b is the assimilation efficiency of phytoplankton by zooplankton, and l bm and l e denote excretion due to basal metabolism and ingestion, while the remaining fraction is transferred to the detritus. m z is the loss coefficient of zooplankton mortality. the nutrient compartments have four principal forms for each nutrient (nitrogen, phosphorus, and silica): dissolved organic nutrients, labile and refractory particulate organic nutrients (lpon and rpon, respectively) , and dissolved inorganic nutrients. only the dissolved inorganic nutrients are utilized by phytoplankton for growth. nutrients are changed to these various organic and inorganic forms via respiration and predation. fig. . shows an example of nutrient cycles using phosphorus. dop, lpop, and rpop work as a pool of phosphorus. for example, certain labile compounds that are rapidly degraded, such as the sugars and amino acids in the particulate organic matter deposited on the sediment surface, decompose readily; others, such as cellulose, are more refractory, or resistant to decomposition. table . shows the distributions of each detritus form by each event, based on pett ( ) . ammonia and nitrate are utilized by phytoplankton for growth. ammonia is the preferred form of inorganic nitrogen for algal growth, but phytoplankton utilize nitrate when ammonia concentrations become depleted. nitrogen is returned from algal biomass to the various dissolved and particulate organic nitrogen pools through respiration and predatory grazing. the time rates for variations due to the biological processes of nitrate and ammonium are as follows. denitrification does not occur in the pelagic water systems in this model, but rather occurs in the anoxic sediment layer. as a result, if denitrification occurs in the sediment, nitrate is transferred by diffusion effect into the sediment. phosphorus kinetics is basically similar to nitrogen kinetics except for the denitrification and the alkaline phosphatase effects of the dop degradation processes. many phytoplankton can enhance alkaline phosphatase activity. this effect makes it possible for them to use phosphate from dop pools (fitzgerald and nelson ) . this effect is formulated in the following model: ( . ) where, r dop is the decomposition rate for dop, r dop-min is the minimum constant of dop decomposition(day − ), k po is a half saturation constant of phosphate uptake, and r dop-di is the acceleration effect of dop decomposition by diatoms. the kinetics of the silica is fundamentally the same as the kinetics of the phosphorus. only diatoms utilize silica during growth. silica is returned to the unavailable silica pool during respiration and predation. the sediment system is zoned in two layers (see fig. . ), an aerobic and an anoxic layer. organic carbon concentrations in the sediment are controlled by detritus burial velocity, the speed of labile and refrigerate organic carbon decomposition, and the rate constant for the diagenesis of particulate organic carbon. the thickness of the aerobic layer is calculated by oxygen diffusion when the amount of oxygen at the bottom layer of the pelagic system isn't zero. the nutrient model, which is a simplified version of the model, treats the nutrients ammonium, nitrate, phosphate, and silica and their exchanges with the pelagic system. silicate-dependent diatoms and non-silicate-dependent algae are distinguished. dissolved oxygen is an essential index for the water quality of an urban coastal zone. sources of do included in the model are reaeration at the sea surface, photosynthesis of phytoplankton, and do in inflows. the sink of do includes respiration of phytoplankton and zooplankton, oxidation of detritrial carbon (ldoc and rdoc), nitrification, and sediment oxygen demand. the time variation of do is formulated as follows: ( ) where, k oc is the oxygen to carbon ratio. k on is the oxygen to nitrogen ratio. k a and q a t- denote the reaeration rate at °c and the temperature coefficient for reaeration at the sea surface, respectively. k nh and q nh t- are the ammonia oxidation rate at °c and the temperature coefficient. k nit is the half saturation constant of ammonia oxidation. k rdoc and q rdoc t- are the rdoc mineralization rate at °c and the temperature coefficient for rdoc mineralization. k ldoc and q ldoc t- mark the ldoc mineralization rate at °c and the temperature coefficient for ldoc mineralization. k mldoc is the half saturation constant for ldoc mineralization. the concentration of do saturation is proportional to temperature and salinity. oxygen saturation value is calculated by using the following equation: where t is temperature and s is salinity. tokyo bay is located at the central part of the main island of japan. the inner bay, the north of which is km in length in its narrowest channel (fig. . ) along the main axis of the bay, connects to the pacific ocean. its average depth and width are m and km, respectively. tokyo bay is one of the most eutrophicated bays in japan. phytoplankton increase in the surface layer from late spring to early fall, and oxygen depletion and the formation of hydrogen sulfide occur on the sea bed. the sea-water color at the head of the bay sometimes becomes milky blue-green in late summer after a continuous north wind (koibuchi et al. ) . this phenomenon is called a blue tide. in the last decade, a variety of water quality observation equipment has been developed. this has made it easier to measure water quality than in the past. however, even with advanced technology, measuring the flux of nutrients is not easy. to quantify the nutrients budget, we applied our numerical model to tokyo bay. the computational domain was divided into km horizontal grids with vertical layers. computation was carried out from april , to march , , with time increments of s provided by the japan meteorological agency giving hourly meteorological data that included surface wind stress, precipitation, and solar radiation. at the open boundary, an observed tide level was obtained which can be downloaded from the japan oceanographic data center (jodc) of the japan coastal guard. don and dop was obtained at % of tn, tp based on the observation results of suzumura and ogawa ( ) at the open boundary. fig. - shows a temporal variation of the computed density at s . the simulation of water column density over the whole period (april -october ) agreed well with measured density. variations between simulated and observed values were generally less than . through the water column. time variation of density effectively reproduced observed results, including shortterm wind-induced variation, formation of stratification during summer, and mixing after continuous strong wind in the middle of october. calculation results also reproduced an upwelling event during the summer season. total chlorophyll-a concentrations in the surface water were reproduced relatively well by model simulations. the model captured the temporal increase in chlorophyll-a that were seen in the observation results, as denoted by arrows in fig. - . during these periods, phytoplankton increased more than m g/l. in tokyo bay, a red tide is defined as a chlorophyll a concentration of greater than m g/l. four different types of plankton assemblages do showed high variability compared with field measurements at the bottom, and was relatively higher than the field data from late september through october. oxygen-depleted water was made on the seabed, representing a basic trend for do variations (fig. - ) . the simulation of phosphate captured not only the observed increase in the surface layer during the summer season, but also inter-annual variability observed over the study period (fig. - ) . for example, phosphate concentrations increased from june to july due to phosphate release from sediment. simulation results represented this kind of trend based on the oxygen-depleted water. fig. - shows nitrate concentration. nitrate concentration in the surface layer fluctuated considerably during this period. nitrate levels doubled or tripled occasionally at the surface. the timing of the high nitrate concentrations and the low density in the surface layer coincided with increases in the river discharge. concentrations of nitrate were underestimated in bottom waters during summer. further study is needed to simulate denitrification processes in the sediment layer. fig. - shows the calculation results of an annual budget of nitrogen and phosphorus in tokyo bay. the annual budget is useful in understanding nutrient cycles. nitrogen is supplied to a considerable degree from rivers, since atmospheric nitrogen input is significant around urban areas. phytoplankton uptake the nitrogen and sink to bottom waters, where they are decomposed by heterotrophic processes which consume oxygen. at the head of the bay (between the line and line in fig. - ) , about % of nitrogen is sunk as detritus and % of it is lost into the atmosphere by denitrification. ammonia released from sediment reaches %. about % of the nitrogen load flows out from the bay. in contrast, atmospheric phosphorus input to the bay is negligible compared to the contribution of phosphorus from other sources. phosphate is released from sediment in the same amount as that discharged from the rivers, and it is transported to the head of the bay by estuarine circulation. as a result, the amount of phosphate in the inner bay remains high. in contrast, nitrogen is mainly supplied from the river mouth and transported quickly out of the bay. in conclusion, nitrogen and phosphorus showed important differences in the mechanisms by which they cycle in fig. - . summary of fluxes and process rates calculated in tokyo bay from january to january . units are given in ton/year for each element the bay. the regeneration of nutrients and their release from the sediment is an important source for phytoplankton growth and is equal to the contributions from the rivers. phosphorus in particular is largely retained within the system through recycling between sediment and water. these results denote the difficulty of improving the eutrophication of bays through the construction of sewage treatment plants alone. big cities have long been developed near waterfronts. even now, naval transport remains one of the most important transportation systems, especially for heavy industries and agriculture. today, many of the world's largest cities are located on coastal zones, and therefore vast quantities of human waste are discharged into near-shore zones (walker ) . fifty percent of the world's populations live within km of the sea. many people visit urban coastal zones for recreation and leisure, and we also consume seafood harvested from this area. as a result, effluents released into the water pose a risk of pathogen contamination and human disease. this risk is particularly heightened for waters that receive combined sewer overflows (csos) from urban cities where both sanitary and storm waters are conveyed in the same sewer system. to decrease the risk from introduced pathogens, monitoring that is both well designed and routine is essential. however, even though a surprising number of pathogens have been reported in the sea, measuring these pathogens is difficult and time consuming -not least because such pathogens typically exist in a "viable but non-culturable" (vbnc) state. in addition, the physical environments of urban coastal zones vary widely depending on time and location. their complicated geographical features border both inland and outer oceans, and so both inland and outer oceans affect them. for example, tidal currents, which are a dominant phenomenon in this area, oscillate according to diurnal periods. even if the emitted levels of pathogens were constant and we could monitor the levels of pathogen indicator organisms at the same place, they would fluctuate according to tidal periods. density stratification also changes with the tides. consequently, the frequent measurement of pathogens is needed to discuss the risk pathogens pose in urban coastal zones. however, this kind of frequent monitoring appears to be impossible. to solve this conundrum and achieve an assessment of pathogen risk, we developed a set of numerical models that expand upon the models developed in sect. . and that include a pathogens model coupled with a three-dimensional hydrodynamic model. section . . deals with the distributions of pathogens in urban coastal zones. the pathogens model is explained in greater detail in sect. . . . section . . deals with numerical experiments that help to understand the effects of appropriate countermeasures. figure - shows some typical density distributions patterns in urban coastal zones. the changing balance between tidal amplitudes and river discharge is responsible for the differences among these patterns. as tidal currents increase, the production of turbulent kinetic energy grows and can become the largest source of mixing in the shallow coastal waters. on the other hand, river-discharged water has a low density, creating a density difference between sea water and land-input water. in salt-wedge estuaries ( fig. - , top) , river water is discharged into a small tidal-range sea. the strength of the tidal currents decreases relative to the river flow. this creates a vertical stratification of density. as a result, river water distributes like a veil over the sea's surface and moves seaward. fig. - . cross-sectional view of the mixing patterns in urban coastal zones in contrast, bottom water moves to the river mouth and mixes with the river water. under such conditions, pathogens move on the surface of the sea, and further mixing with low-density fresh water is restricted by stratification. in partially mixed estuaries ( fig. - , middle) , the tidal force becomes a more effective mixing mechanism. fresh water and sea water are mixed by turbulent energy. as a result, pathogens that are emitted from sewer treatment plants are more mixed than those in the static salt-wedge estuaries. in well-mixed estuaries ( fig. - , bottom) , the mixing of salt and river waters becomes more complete due to the increased strength of tidal currents relative to river flow. here, the density difference is developed in a horizontal direction. as a result, pathogens are mixed in the water column and settle down on the sea bed, in turn contaminating estuarine waters during the spring tide or contaminating rainfall through re-suspension (pommepuy et al. ) . figure - shows distributions of pathogens under coastal environments. these pathogens encounter a wide range of stresses including uv rays (sinton et al. ) , temperature differences (matsumoto and omura ) , ph (solić and krstulović ) , salinity (omura et al. ) , and lack of nutrients. the pathogens are transported by currents and continue to become part of sedimentation and to be re-suspended in urban coastal zones (pommepuy et al. ) . the modeling of major pathogens of concern (including adenovirus, enterovirus, rotavirus, norovirus, and coronavirus) is not usually conducted owing to the difficulty of modeling and the lack of observational data in coastal environments. we modeled escherichia coliform (e. coli) by using experimental data in coastal sea water. this model consists of a three-dimensional hydrodynamic model and an e. coli model (onozawa et al. ) . the mathematical framework employed in the e. coli model takes the same approach that was explained in sect. . . . the mass balance of e. coli is expressed as follows: . ) where coli denotes concentrations of e. coli (cfu/ ml), and t is time. sink represents the sinking speed of e. coli . u i denotes flow speed for the calculation of the advection term. e i denotes the diffusion coefficients. sal denotes the salinity-dependent die-off rate ( ppt/day). sunlight is generally recognized to be one source by which bacteria are inactivated, due to uv damage to the bacterial cell (sinton et al. ) . however, this particular target area has high turbidity that rapidly absorbs uv rays at the sea's surface. as a result, this process has been ignored in this model. in this model, we can see the numerical simulations performed with two nested domains to fit the complex geography feature around the odaiba area . these nested grids make possible a representation of the stratification effect. a detailed configuration of the model is summarized in table - . the two computational domains cover the whole fig. - ) that includes both the salinity simulation and observation results. the simulation results show stratification, mixing, and an upwelling phenomenon, and include levels and timing. fig. - shows a comparison between observation results and a calculation for temperature and salinity in a fine grid scale (domain in fig. - ) . variations between the simulated and observed values were generally less than . °c and psu through the water column. the timing and periods of upwelling events were captured accurately. after rain fall, river discharge was increased remarkably. model results adequately represent precipitation variation events and their effects. figure - shows a comparison between modeled and measured e. coli at stn. . the current standard for acceptably safe beaches for swimming set by of the ministry of the environment of japan is a fecal coliform rate of coliforms unit per ml (cfu/ ml). this index of fecal coliform includes not only e. coli but also others. however, it is well known that the majority of the fecal coliform in this area comes from e. coli . therefore, we use a value of cfu/ ml e. coli as the standard for the safety of swimming in the sea. from the calculation results, we can see that durations when the standards are exceeded are very limited, and that most of the summer period falls below the standard for swimming. this result also denotes that the increasing rates of e. coli do not agree with levels of precipitation. even in small precipitations, e. coli significantly increased. understanding the effects of physical factors is important to understanding the fate and distributions of pathogens. such an understanding is in turn variations in levels of e. coli are directly correlated with the discharge from pumping stations, tidal currents, river discharges, and density distributions, as explained in sect. . . . as a result, the distributions of cso differ according to timing, even when the level of discharge is the same. we performed numerical experiments in order to evaluate the contributions of these different discharges and phenomena to cso distributions. the first numerical experiment was a nowcast simulation that calculated e. coli distributions under realistic conditions. the second experiment was a numerical experiment to estimate the effects of a waste-reservoir that was being constructed near the shibaura area. numerical experiments were also applied odaiba area, which is used as a bathing area. figure - shows temporal variations of precipitation and river discharges (top), as well as tide levels and e. coli concentrations discharged from three different areas. shibaura and sunamachi area are located at the upper bay location from the odaiba area. morigasaki has the largest area, but is located in the lower bay location from the odaiba area (see fig. in this spring tide period, tidal ranges can reach m. small precipitations were measured from august th to th. river discharge increased with precipitation, reaching m /s. the levels of e. coli increased rapidly after the rainfall event, due mainly to discharges from the sunamachi and chibaura areas. near the end of this period of increase, effluent from morigasaki also reached the odaiba area. fig. - shows the spatial distributions of e. coli from three different times. from this fig., the e. coli emitted from the morigasaki area can be seen to have been transported from the lower region of the bay to the odaiba area. this is because the small amount of river discharge resulted in a thin layer of low-density, highly concentrated e. coli on the surface of the sea, and tended to isolate the e. coli by preventing it from mixing with the water column. in contrast, fig. - shows a large precipitation case under the neap tide period. large amounts of precipitation produced a large river discharge that reached m /s. in this period, only the upper bay's csos arrived at the odaiba area. no contributions from the morigasaki area took place. in conclusion, the concentrations of e. coli vary widely according to space and time. the density distributions produced by the balance of fig. - . spatial distributions of e. coli at odaiba area tides and river discharges have very complex effects. e. coli concentrations reached maximum levels after small precipitation events, but did not increase so much under large precipitation events due to mixing. these kinds of results would be impossible to understand only from observation. the model successfully captured complex distributions of e. coli and helped our understanding of pathogens contaminations. to mitigate cso pollutions, the construction of storage tanks at three sites in tokyo has been planned by the tokyo metropolitan government. shibaura is the target area of this plan around the odaiba area. numerical numerical simulation was performed with and without the proposed storage tank, which has a capacity of , m . this storage tank can store csos after rainfall. to include the effect of continuous rain, we assumed that the csos stored in the tank could be purified within day. table . shows the calculation results for the mitigation effect of the storage tank. these numbers denote the dates when the standards for bathing in the sea (over cfu/ ml) were exceeded. fig. . shows the observation stations. from this table, stn. shows the largest decrease in csos among these five stations. this is because stn. is located closest to the pumping stations, and therefore would be most sensitive to the csos. before the construction of the cso storage tank, minimum bathing standards for cso levels were exceeded on days. after the construction of the storage tank, the duration of cso levels that exceeded safety standards decreased to days. there was an improvement of days. on the other hand, other stations only days, or in some cases, less than day. such differences could not be observed, especially in those stations that are located inside the odaiba area due to the enclosed feature of bathymetry. for example, over , storage tanks have been built in germany alone, and another , were planned during the s in germany. our plans for dealing with csos are not enough to mitigate the effects of csos completely. at the same time, these results show us the complexity of pathogens distributions and the importance of numerical modeling for this problem. numerical simulation is one of the most important tools for the management of water quality and ecosystems in urban coastal zones. we have developed a water quality model to simulate both nutrient cycles and pathogens distributions, and coupled it with a three-dimensional hydrodynamic model of urban coastal areas. to quantify the nutrients budget, a numerical model should include material cycles with phytoplankton, zooplankton, carbons, nutrients, and oxygen. we applied this model to the tokyo bay and simulated water column temperatures, salinity, and nutrient concentrations that were closely linked with field observations. this model successfully captured periods of timing, stratification events, and subsequent changes in bottom water oxygen and nutrients. our model results also indicated that there were clear differences between the material cycles of nitrogen and phosphorus inside the bay. the regeneration of nutrients and its release from sediment was found to be a source of phytoplankton growth on the same order of importance as contributions from rivers. in particular, phosphorus was found to have been largely retained within the system through recycling between sediment and water. we also developed a pathogen model that includes e. coli and is applied to the simulation of cso influences in urban coastal zones. these results indicate that, because of stratification, concentrations of e. coli significantly increase after even small precipitation events. from this study, the balance between tidal mixing and river waters can be seen to be significant. however, these are only two case studies; it remains necessary to simulate the structure and characteristics of cso distributions and their impact on urban coastal zone pollution. such simulations remain as future works to be undertaken. ministry of the environment a description of a three-dimensional coastal ocean circulation model phytoplankton kinetics in a subtrophical estuary: eutrophication three-dimensional eutrophication model of chesapeake bay user's guide to the ce-qual-icm: three-dimensional eutrophication model one dimensional ecosystem model of the equatorial pacific upwelling system part i: model development and silicon and nitrogen cycle temperature and phytoplankton growth in the sea a model of annual plankton cycles a nitrogen-based model of phytoplankton dynamics in the oceanic mixed layer extractive and enzymatic analysis for limiting or surplus phosphorus in algae a modified sigma equations; approach to the numerical modeling of great lake hydrodynamics model evaluation experiments in the north atlantic basin: simulations in nonlinear terrain-following coordinates verification of a three-dimensional hydrodynamic model of chesapeake bay small scale processes in geophysical fluid flows nemuro -a lower trophic level model for the north pacific marine ecosystem blue tide occurred in the west of tokyo bay in summer of study on budget and circulation of nitrogen and phosphorus in tokyo bay a coastal marine ecosystem: simulation and analysis simulations of chesapeake bay estuary: sensitivity to turbulence mixing parameterizations and comparison with observations processes governing phytoplankton blooms in estuaries. ii. the role of transport in global dynamics long-term isohaline salt balance in an estuary nitrogen budget in tokyo bay with special reference to the low sedimentation to supply ratio some factors affecting the survival of fecal indicator bacteria in sea water analytic prediction of the properties of stratified planetary surface layers development of a turbulence closure model for geophysical fluid problems viability and adaptability of e-coli. and enterococcus group to salt water with high concentration of sodium chloride numerical calculation of combined sewer overflow(cso) due to heavy rain around daiba in the head of tokyo bay kenetics of microbial mineralization of organic carbon from detrital skeletonema costatum cells a coordinate system having some special advantages for numerical forecasting enteric bacterial survival factors sunlight inactication of fecal indicator bacteria and bacteriophages from waste stabilization pond effluent in fresh and saline waters separate and combined effects of solar radiation, temperature, salinity, and ph on the survival of feacal coliforms in seawater characterization of dissolved organic phosphrous in coastal seawater using ultrafiltration and phosphoydrolytic enzymes relation between the transport of gravitational circulation and the river discharge in bays spring bloom sedimentation in a subarctic ecosystem the coastal zone seasonal changes of organic carbon and nitrogen production by phytoplankton in the estuary of river tamagawa key: cord- -v rllyyu authors: puzyn, tomasz; gajewicz, agnieszka; leszczynska, danuta; leszczynski, jerzy title: nanomaterials – the next great challenge for qsar modelers date: - - journal: recent advances in qsar studies doi: . / - - - - _ sha: doc_id: cord_uid: v rllyyu in this final chapter a new perspective for the application of qsar in the nanosciences is discussed. the role of nanomaterials is rapidly increasing in many aspects of everyday life. this is promoting a wide range of research needs related to both the design of new materials with required properties and performing a comprehensive risk assessment of the manufactured nanoparticles. the development of nanoscience also opens new areas for qsar modelers. we have begun this contribution with a detailed discussion on the remarkable physical–chemical properties of nanomaterials and their specific toxicities. both these factors should be considered as potential endpoints for further nano-qsar studies. then, we have highlighted the status and research needs in the area of molecular descriptors applicable to nanomaterials. finally, we have put together currently available nano-qsar models related to the physico-chemical endpoints of nanoparticles and their activity. although we have observed many problems (i.e., a lack of experimental data, insufficient and inadequate descriptors), we do believe that application of qsar methodology will significantly support nanoscience in the near future. development of reliable nano-qsars can be considered as the next challenging task for the qsar community. one diameter of nm or less. when nanoparticles are intentionally synthesized to be used in consumer goods, they are called "nanomaterials" [ ] . nowadays, years after feyman's lecture, nanotechnology has emerged at the forefront of science and technology developments and nanomaterials have found a wide range of applications in different aspects of human life. for example, nanoparticles of such inorganic compounds as tio and zno oxides are used in cosmetics [ ] , sunscreens [ ] , solar-driven self-cleaning coatings [ ] , and textiles [ ] . nanosized cuo has replaced noble metals in newer catalytic converters for the car industry [ ] . nanopowders of metals can be used as antibacterial substrates (e.g., the combination of the pure nanosilver ion with fiber to create antiodor socks) [ ] . finally, metal salts (i.e., cdse quantum dots) have found many applications in electronics and biomedical imaging techniques [ , ] . the discoveries of fullerene (c ) in by kroto et al. [ ] and carbon nanotubes in by iijima [ ] opened a new area of the tailored design of carbonbased nanomaterials. carbon-based nanomaterials are currently used, among other applications, for synthesis of polymers characterized by enhanced solubility and processability [ ] and for manufacturing of biosensors [ ] . they also contribute to a broad range of environmental technologies including sorbents, high-flux membranes, depth filters, antimicrobial agents, and renewable energy supplies [ ] . according to current analysis [ ] , about different products containing nanomaterials were officially on the market in . most of them ( ) have been manufactured in the usa, in east asia (china, taiwan, korea, japan), in europe, and only in other countries. it is interesting that the number ( ) is two times higher than the number of nanoproducts in the previous year. investments in nanotechnology industry grew from $ billion in to $ billion in andif one can believe the forecast -will reach $ . trillion in . without doubt, nothing is able to stop such a rapidly developing branch of technology and we should be prepared for (better or worse) living day by day in symbiosis with nanomaterials. the astonishing physical and chemical properties of engineered nanoparticles are attributable to their small size. in the nanometer-scale, finite size effects such as surface area and size distribution can cause nanoparticles to have significantly different properties as compared to the bulk material [ ] . for instance, by decreasing the size of gold samples one induces color changes from bright yellow through reddish-purple up to blue. however, from the physico-chemical viewpoint, the novel properties of nanoparticles can also be determined by their chemical composition, surface structure, solubility, shape, ratio of particles in relation to agglomerates, and surface area to volume ratio. all these factors may give rise to unique electronic, magnetic, optical, and structural properties and, therefore, lead to opportunities for using nanomaterials in novel applications and devices [ ] . new, characteristic properties of nanomaterials include greater hardness, rigidity, high thermal stability, higher yield strength, flexibility, ductility, and high refractive index. the band gap of nanometer-scale semiconductor structures decreases as the size of the nanostructure decreases, raising expectations for many possible optical and photonic applications [ ] . with respect to the size of the grains, it has been suggested that nanomaterials would exhibit increased (typically - times) strength and hardness as compared to their microcrystalline counterparts. for example, the strength of nanocrystalline nickel is five orders of magnitude higher than that of the corresponding microcrystalline nickel [ ] . interestingly, the observed strength of crystalline nanomaterials is accompanied by a loss of ductility, which can result in a limitation of their utility [ ] . however, some of the nanocrystalline materials have the ability to undergo considerable elongation and plastic deformation without failing (even up to - %). such machinability and superplasticity properties have been observed for ceramics (including monoliths and composites), metals (including aluminum, magnesium, iron, titanium), intermetallic elements (including iron, nickel, and titanium base), and laminates [ ] . although the atomic weight of carbon nanotubes is about one-sixth of the weight of steel, their young's modulus and tensile strength are, respectively, five and times higher than those of steel [ ] . in addition, nanoparticles, because of their very small sizes and surface/interface effects such as the fundamental change in coordination, symmetry, and confinement, they may exhibit high magnetic susceptibility. a variety of nanoparticles reveal anomalous magnetic properties such as superparamagnetism. this opens new areas of potential application for them, such as data storage and ferrofluid technology [ ] . according to recent studies, nanoparticles may have also great potential in medical application, mostly due to their good biocompatibility that allows them to promote electron transfer between electrodes and biological molecules. for instance, the high biocompatibility of magnetite nanocrystals (fe o ) makes them potentially useful as the magnetic resonance imaging contrast agents [ ] . one of the unique aspects of nanoparticles is their high wettability, termed by fujishima [ ] as superhydrophilicity. depending upon the chemical composition, the surface can exhibit superhydrophilic characteristics. for example, titanium dioxide (tio ), at sizes below a few nm, can decrease the water contact angle to ± • [ ] . nano-sized composites, due to the chemical composition and viscosity of the intercrystalline phase, may provide a significant increase in creep resistance. it has been demonstrated that alumina/silicon carbide composites are characterized by a minimum creep rate, three times lower than the corresponding monolith [ ] . as mentioned in section . , different types of nanomaterials are increasingly being developed and used by industry. however, little is known about their toxicity, including possible mutagenic and/or carcinogenic effects [ ] . some recent contributions report evident toxicity and/or ecotoxicity of selected nanoparticles and highlight the potential risk related to the development of nanoengineering. evidently, there is insufficient knowledge regarding the harmful interactions of nanoparticles with biological systems as well as with the environment. it is well known that the most important parameters with respect to the induction of adverse effects by a xenobiotic compound are its dose, dimension, and durability. conversely, it is well established that nano-sized particles, due to their unique physical and chemical properties discussed above, behave differently from their larger counterparts of the same chemical composition [ ] [ ] [ ] [ ] [ ] [ ] . because of the difference between nanoparticles and bulk chemicals, the risk characterization of bulk materials cannot be directly extrapolated to nanomaterials. the biological activity of nanoparticles and their unique properties causing harmful effects are highly dependent on their size. nanoparticles, because of their small size, may pass organ barriers such as skin, olfactory mucosa, and the blood-brain barrier [ ] [ ] [ ] , readily travel within the circulatory system of a host, and deposit in target organs. this is not possible with the same material in a larger form [ ] . indeed, reduction of the particle's size to the nanoscale level results in a steady increase of the surface to volume ratio. as a consequence, a larger number of potentially active groups per mass unit is "available" on the surface and might interact with biological systems [ ] . this is one possible explanation why nano-sized particles of a given compound are generally more toxic than the same compound in its larger form [ ] . however, oberdörster et al. [ ] suggested that the particle size is not the only possible factor influencing toxicity of nanomaterials. the following features should be also considered: • size distribution, • agglomeration state, • shape, • porosity, • surface area, • chemical composition, • structure-dependent electronic configuration, • surface chemistry, • surface charge, and • crystal structure. natural and anthropogenic nanoparticles gain access into the human body through the main ports of entry including the lungs, the skin, or the gastrointestinal tract. the unique properties of nanoparticles allow them not only to pnetrate physiological barriers but also to travel throughout the body and interact with subcellular structures. toxicological studies show that nanoparticles can be found in various cells such as mitochondria [ , ] , lipid vesicles [ ] , fibroblasts [ ] , nuclei [ ] , and macrophages [ ] . depending on their localization inside the cell, nanoparticles can induce formation of reactive oxygen species (ros), for instance, superoxide radicals, hydroxyl radicals reactive nitrogen [ ] , sulfur [ ] , and other species stressing the body in a similar manner to the effect of ros [ ] . this results in oxidative stress and inflammation, leading to the impacts on lung and cardiovascular health [ ] . it is worth noting that normally, due to the presence of antioxidant molecules (i.e., vitamin c and glutathione), the body's cells are able to defend themselves against ros and free radicals damage. however, when a large dose of strongly electrophilic nanoparticles enter the body, the balance between reduced glutathione (gsh) and its oxidized form (gssg) is destroyed [ ] and the unscavenged oxidants cause cell injuries by attacking dna, proteins, and membranes [ ] . at the cellular level, oxidative stress is currently the best developed paradigm depicting the harmful effects of nano-sized particles [ , , ] . the mechanism of oxidative stress occurring at the molecular level is mainly responsible for observed cytotoxic and genotoxic effects induced by nanoparticles. cytotoxicity of selected nanospecies has been confirmed by many researchers. for example, fullerene (c ) particles suspended in water are characterized by antibacterial activity against escherichia coli and bacillus subtilis [ ] and by cytotoxicity to human cell lines [ ] . single multiwalled carbon nanotubes (cwcnts and mwcnts) are also toxic to human cells [ , ] . nano-sized silicon oxide (sio ), anatase (tio ), and zinc oxide (zno) can induce pulmonary inflammation in rodents and humans [ ] [ ] [ ] . epidemiological studies have shown that nanoparticles might be genotoxic to humans [ ] . irreversible dna modifications resulting from the activity of ros may lead to heritable mutations, involving a single gene, a block of genes, or even whole chromosomes. dna damage may also disrupt various normal intracellular processes, such as dna replication and modulate gene transcription, causing abnormal function or cell death [ , , ] . until now, more than different oxidative dna lesions have been found. the most investigated oh-related dna lesions is -hydroxydeoxyguanosine ( -ohdg) [ ] , which may be induced by several particles such as asbestos, crystalline silica, coal fly ashes. oxygen free radicals may overwhelm the antioxidant defense system by mediating formation of base adducts, such as -hydroxydeoxyguanosine, and therefore play a key role in initiation of carcinogenesis [ ] . data on neurotoxic effects of engineered nanoparticles are very limited, but it has been reported that inhaled nanoparticles, depending on their size, may be distributed to organs and surrounding tissues, including the olfactory mucosa or bronchial epithelium and then can be translocated via the olfactory nerves to the central nervous system [ ] . there is also some evidence that nano-sized particles can penetrate and pass along nerve axons and dendrites of neurons into the brain [ ] . recent studies confirm the translocation of nanoparticles from the respiratory tract into the central nervous system; for example, inhalation with nm magnesium oxide in rats showed that manganese can be taken up into olfactory neurons and accumulated in the olfactory bulb [ ] . the particles at the nanoscale may also gain access to the brain across the bloodbrain barrier [ ] . there is experimental evidence that oxidative stress also plays an important role in neurodegenerative diseases and brain pathology, for instance, hallervorden-spatz syndrome, pick's disease, alzheimer's disease, or parkinson's disease [ ] . the effects of nanoparticles on the immune system are still unclear. although the reticuloendothelial system (res) is able to eliminate nanoparticles, several toxicological studies have suggested that nanoscale particles' interaction with the defense activities of immune cells can change their antigenicity and stimulate and/or suppress immune responses. direct experiments showed that dendritic cells and macrophages uptake of nanoparticle-protein complexes may change the formation of the antigen and initiate an autoimmune response [ ] . several studies have also reported that nanoparticles may induce damage to red blood cells (erythrocytes). bosi et al. [ ] have studied the hemolytic effect of different water-soluble c fullerenes. preliminary results indicate that hemolytic activity depends on the number and position of the cationic surface groups. however, no clinically relevant toxicity has yet been demonstrated [ ] . nano-sized particles such as volcanic ash, dust storms, or smoke from natural fires have always been present in the environment. however, the recent progress of industry has increased engineered nanoparticle pollution. the unique size-specific behavior and specific physical-chemical properties, in combination with toxicity to particular living organisms, may also result in harmful effects on the level of whole environmental ecosystems [ ] . in the pioneering report on the non-human toxicity of fullerene, eva oberdörster [ ] observed that manufactured nanomaterials can have negative impacts on aquatic organisms. water-soluble c fullerenes cause oxidative damage (lipid peroxydation in the brain) and depletion of glutathione in the gill of juvenile largemouth bass (micropterus salmoides) at a concentration of . ppm. however, these results might be disputable, because the authors used the organic solvent tetrahydrofuran (thf) to disaggregate c fullerenes, thf is classified as a neurotoxin [ ] . subsequently, lover and klaper [ ] observed the toxicological impact of nanoparticles of fullerenes (c ) and titanium dioxide (tio ) to daphnia magna: c and tio caused mortality with a lc value of . ppm for tio and a lc value of ppb for the fullerene. in this case the authors also used thf for solubilization of hydrophobic c , thus the results are also of lower credibility. interestingly, in similar experiments by andrievsky et al. [ ] with "fullerene water solutions" (hydrated fullerenes, c · nh o), no mortality was observed. in a later study, adams et al. [ ] confirmed the acute toxicity of selected nanosized metal oxides against d. magna. he observed that sio particles were the least toxic and that toxicity increased from sio to tio to zno. a further study by the authors [ ] showed that these three photosensitive nanoscale metal oxides in water suspensions have similar antibacterial activity to gram-positive (b. subtilis) and gram-negative (e. coli) bacteria (sio < tio < zno). all the metal oxides nanoparticles tested inhibited the growth of both gram-positive and gram-negative bacteria; however, b. subtilis was more sensitive than e. coli. similar results have been observed for a bath of zno, tio , and cuo against bacterium vibrio fischeri and crustaceans d. magna and thamnocephalus platyurus [ ] . the antibacterial effects of nano-sized metal oxides to v. fischeri were similar to the rank of toxicity to d. magna and t. platyurus; they increased from tio to cuo and zno. it is also very important to recognize that titanium dioxide was not toxic even at the g/l level, which means that not all nanoparticles of metal oxides induce toxicity. smith et al. [ ] investigated the ecotoxicological potential of single-walled carbon nanotubes (swcnt) to rainbow trout (oncorhynchus mykiss) showing that the exposure to dispersed swcnt causes respiratory toxicity -an increase of the ventilation rate, gill pathologies, and mucus secretion. additionally, the authors observed histological changes in the liver, brain pathology, and cellular pathologies, such as individual necrotic or apoptotic bodies, in rainbow trout exposed to . mg/l swcnt. mouchet et al. [ ] analyzed the acute toxicity and genotoxicity of double-walled carbon nanotubes (dwnts) to amphibian larvae (xenopus laevis). the authors did not observe any effects at concentrations between and mg/l. however, at the highest concentrations ( mg/l) % of mortality was measured, while at the lowest concentrations ( mg/l) reduced size and/or a cessation of growth of the larvae were observed. summarizing this section, there is strong evidence that chemicals, when synthesized at the nanoscale, can induce a wide range of specific toxic and ecotoxic effects. moreover, even similar compounds from the same class can differ in toxicity. the available data on toxicity are still lacking; thus, more comprehensive and systematic studies in this area are necessary and very important. as demonstrated in this book, quantitative structure-activity relationship (qsar) methods can play an important role in both designing new products and predicting their risk to human health and the environment. however, taking into account the specific properties of nanomaterials and their still unknown modes of toxic action, this class of compounds seems to be much more problematic for qsar modelers than the "classic" (small, drug-like) chemicals. until now, more than different descriptors have been developed and used for the characterization of molecular structure (chapter ). in general, the descriptors can be classified according to their dimensionality. constitutional descriptors, socalled "zero-dimensional," are derived directly from the formula (e.g., the number of oxygen atoms). descriptors of bulk properties, such as n-octanol/water partition coefficient or water solubility, are classified as "one-dimensional" descriptors. topological descriptors based on the molecular graph theory are called "twodimensional" descriptors and characterize connections between individual atoms in the molecule. "three-dimensional" descriptors reflect properties derived from the three-dimensional structure of a molecule optimized at the appropriate level of quantum-mechanical theory. "four-dimensional" descriptors are defined by molecular properties arising from interactions of the molecule with probes characterizing the surrounding space or by stereodynamic representation of a molecule, including flexibility of bonds, conformational behavior, etc. [ ] [ ] [ ] [ ] [ ] . only a little is known about applicability of those "traditional" descriptors for the characterization of nanostructures. some authors [ ] [ ] [ ] postulate that the existing descriptors are insufficient to express the specific physical and chemical properties of nanoparticles. thus, novel and more appropriate types of the descriptors must be developed. a group of nanoparticles is structurally diversified. in fact, this group has been defined arbitrarily in some way, taking into account size as the only criterion of the particles' membership. therefore, structures as various as nanotubes, fullerenes, crystals, and atom clusters as well as chemical species of such different properties as metals, non-metals, organic compounds, inorganic compounds, conductors, semi-conductors, and isolators were put together into one single group. since nanoparticles are not structurally homogenous, a common mechanism of toxicity cannot be expected for all of them. in consequence, toxicity and other properties should be studied within the most appropriately chosen sub-classes of structural and physico-chemical similarity. what is the best way to define the sub-classes? the answer might be given based on a stepwise procedure recommended by the oecd guidance document on the grouping of chemicals [ ] (see also chapter ) . along with the guidelines, the following eight steps should be performed: . development of the category hypothesis, definition, and identification of the category members. the category can be defined based on chemical similarity, physico-chemical properties, toxicological endpoint, and/or mechanism of action, as well as in terms of a metabolic pathway. . gathering of data for each category members. all existing data should be collected for each member of the category. . evaluation of available data for adequacy. the data should be carefully evaluated at this stage according to the commonly accepted protocols (i.e., according to the appropriate oecd guidance). . construction of a matrix of data availability (category endpoints vs. members). the matrix is to indicate whether data are available or not. performing of a preliminary evaluation of the category and filling data gaps. the preliminary evaluation should indicate if (i) the category rationale is supported and (ii) the category is sufficiently robust for the assessment purpose (contains sufficient, relevant and reliable information). . performing of additional testing (experiments). based on the preliminary evaluation (especially evaluation of the robustness), additional experiments and group members for testing can be proposed. . performing of a further assessment of the category. if new data from the additional testing are generated, the category should be revised according to the criteria from step . . documenting of the finalized category. finally, the category should be documented in the form of a suitable reporting format proposed by the guidance. the currently proposed [ ] working classification scheme for nanostructured particles includes nine categories: . spherical or compact particles; . high aspect ratio particles; . complex non-spherical particles; . compositionally heterogeneous particles -core surface variation; . compositionally heterogeneous particles -distributed variation; . homogeneous agglomerates; . heterogeneous agglomerates; . active particles; . multifunctional particles. this classification has been adapted from the original work of maynard and aitken [ ] . what types of structural properties should be described within the groups? as previously discussed in section . , the diameter of a nanoparticle is important, but it is not the only one possible factor influencing the mode of toxic action. the additional structural characteristics which must also be appropriately expressed are size distribution, agglomeration, shape, porosity, surface area, chemical composition, electronic configuration, surface chemistry, surface charge, and crystal structure. in contrast to the classic qsar scheme, an entire characterization of a nanostructure may be impossible only when computational methods are employed. novel descriptors reflecting not only molecular structure, but also supra-molecular pattern (size, shape of the nanoparticles, etc.) should be derived from both computational and experimental techniques. the fastest and relatively easy step of characterizing the structure is the calculation of constitutional and topological descriptors. an interesting and very practical idea in this field is to replace a series of simple descriptors by one, so-called "technological attributes code" or "smiles-like code" [ ] [ ] [ ] [ ] . for instance, a nanoparticle of ceramic zirconium oxide, existing in bulk form and synthesized at a temperature of • c can be expressed by the code "zr,o,o,cer,%e" [ ] . similar to the simplified molecular input line entry system (smiles), the international chemical identifier (inchi) might also be used directly as a descriptor of chemical composition [ ] . another possibility is to apply descriptors derived from either molecular graph (mg) or the graphs of atomic orbitals (gao) theory [ ] [ ] [ ] . in the first case, vertexes in the graph represent atoms, while edges represent covalent bonds. in the second method, vertexes refer to particular atomic orbitals ( s, s, p, etc.), while edges connect the orbitals belonging to different atoms (figure - ) . based on the molecular graphs, faulon and coworkers [ ] [ ] [ ] [ ] have developed the signature molecular descriptor approach for the characterization of fullerenes and nanotubes. the signature is a vector including extended valences of atoms derived from a set of subgraphs, following the five-step algorithm: . constructing of a subgraph containing all atoms and bonds that are at a distance no greater than the given signature height; . labeling the vertices in a canonical order; . constructing a tree spanning all the edges; . removing of all canonical labels that appear only one time; . writing the signature by reading the tree in a depth-first order. the signature descriptor can be utilized not only for direct qsar modeling, but also for calculating a range of topological indices (i.e., the wiener index). [ ] [ ] [ ] ] without doubt, simplicity of calculation is the most significant advantage of the topological descriptors. however, in many cases these two-dimensional characteristics are insufficient to investigate more complex phenomena. in such a situation, a more sophisticated approach must be employed to describe the structure appropriately. as mentioned previously, quantum-mechanical calculations can deliver useful information on the three-dimensional features (see chapter ). among others, they include: molecular geometry (bond lengths, valence, and torsion angles), electron distribution, ionization potential, electron affinity, surface reactivity, and band gap. when performing quantum-mechanical calculations, there are always two important assumptions to be introduced. first one is an appropriate molecular model; the second one is the appropriate level of the theory. both assumptions are closely related: when the model (system) is too large, the calculations at the highest levels of the theory are impossible, because of large computational time and other technical resources to be required [ ] . small fullerenes and carbon nanotubes can be treated as whole systems and modelled directly with quantum-mechanical methods. among the theory levels, the density functional theory (dft) recently seems to have been accepted as the most appropriate and practical choice for such calculations. indeed, dft methods can serve as a good alternative for conventional ab initio calculations, when a step beyond the means field approximation is crucial and the information on the electron correlation significantly improves the results (e.g., hartree-fock -hf method in conjunction with møller-pleset the second-order correction -mp ). unfortunately, even "small" fullerenes and carbon nanotubes (containing between and carbon atoms) are, in fact, large from quantum-mechanical point of view. therefore, the "classic" ab initio calculations might be impractical because of the reasons mentioned in the previous paragraph, whereas dft can be performed in reasonable time. the functional commonly utilized for dft is abbreviated with the b lyp symbol. in b lyp calculations (eq. - ) the exchange-correlation energy e xc is expressed as a combination (a , a x , and a c are the parameters) of four elements: (i) the exchange-correlation energy from the local spin density approximation (lsda, e lsda xc ), (ii) the difference between the exchange energy from hartree-fock (e hf x ) and lsda (e lsda x ), (iii) becke's exchange energy with gradient correction (e b x ), and (iv) the correlation energy with lee-yang-parr correction (e lyp c ) [ , ] : sometimes, when a system is too large from the quantum-mechanical point of view, the calculations are practically impossible. the situation is very common for larger crystalline nanoparticles (i.e., nanoparticles of metal oxides: tio , al o , sno , zno, etc.) and, in such cases, a simplified model of the whole structure must first be appropriately selected. in general, there are two strategies for modeling of crystalline solids: (i) an application of the periodic boundary conditions (pbss) and (ii) calculations based on the molecular clusters. in the first approach, calculations for a single unit cell are expanded in the three dimensions with respect to the translational symmetry by employing appropriate boundary conditions (i.e., the unit cell should be neutral and should have no dipole moment). in doing so, the model includes information on the long-range forces occurring in the crystal. however, the cell size should be large enough to also be able to model defects in the surface and to eliminate the spurious interactions between periodically repeated fragments of the lattice [ ] [ ] [ ] . in the second approach, a small fragment or so-called "cluster," is cut off from the crystal structure and then used as a simplified model for calculations. the only problem is how to choose the diameter of the cluster correctly? this must be performed by reaching a compromise between the number of atoms (and thus the required time of computations) and the expected accuracy (and hence level of the theory to be employed). it is worth mentioning that the molecular properties can be divided into two groups depending on how they change with increasing size of the cluster (going from molecular clusters to the bulk form). they are (i) scalable properties, varying smoothly until reaching the bulk limit and (ii) non-scalable properties, when the variation related to increasing size of the cluster is not monotonic. although the cluster models usually avoid the long-range forces, they have found many applications in modeling of local phenomena and interactions on the crystal surface [ ] . as previously mentioned, in addition to calculated properties, experimentally derived properties may also serve as descriptors for developing nano-qsars (table - ). the experimental descriptors seem to be especially useful for expressing size distribution, agglomeration state, shape, porosity, and irregularity of the surface area. interestingly, the experimental results can be combined with numerical methods to define new descriptors. for example, images taken by scanning electron microscopy (sem), transmission electron microscopy (tem), or atomic force (figure - ) might be processed with use of novel chemometric techniques of image analysis. namely, a series of images for different particles of a given nanostructure should first be taken. then, the pictures must be numerically averaged and converted into a matrix containing numerical values that correspond to intensity of each pixel in the gray scale or color value in the rgb scale. new descriptors can be defined based on the matrix (i.e., a shape descriptor can be calculated as a sum of non-zero elements in the matrix; porosity -as a sum of relative differences between each pixel and its "neighbors," etc.) [ ] . without doubt, an appropriate characterization of the nanoparticles' structure is currently one of the most challenging tasks in nano-qsar. although more than qsar descriptors have been defined so far, they may be inadequate to express the supramolecular phenomena governing the unusual activity and properties of nanomaterials. as a result, much more effort in this area is required. an important step related to the numerical description of chemical structure and qsar modeling involves establishing a qualitative relationship between the structure of a nanoparticle and its various electronic properties. the b lyp functional and the standard - g(d) polple's style basis set were applied by shukla and leszczynski [ ] to investigate the relationships between the shape, size, and electronic properties of small carbon fullerenes, nanodisks, nanocapsules, and nanobowls. they found out that the ionization potentials decrease, while the electron affinities increase in going from the c fullerenes to the closed nanodisks, capsules, and open bowl-shaped nanocarbon clusters. in similar studies performed for capped and uncapped carbon nanotubes at the b lyp/ - g(d) level of theory by yumura et al. [ , ] , the authors demonstrated that the tube lengths, edge structures, and end caps play an important role in determining the band gap expressed as a difference between the energies of the highest occupied and lowest unoccupied molecular orbitals (homo-lumo) and vibrational frequencies. wang and mezey [ ] characterized electronic structures of open-ended and capped carbon nanoneedles (cnns) at the same theory level (b lyp/ - g(d)) concluding that conductivity of the studied species is strictly correlated to their size. only very long cnns structures have band gaps sufficiently narrow to be semiconductors, while the band gaps of very short and thin structures are too large to conduct electrons. similarly, poater et al. [ , ] observed that the parr electrophilicity and electronic movement described by the chemical potential increase with increasing length of the carbon nanoneedles and very "short" structures (containing four layers and less) have a homo-lumo gap too large to allow conductivity. moreover, simeon et al. [ ] , by performing b lyp calculations, demonstrated that a replacement of the fullerene carbon atom with a heteroatom results in a significant change of electronic and catalytic properties of the fullerene molecule. similar studies have been performed for crystalline metal semi-conductors with the use of the cluster calculations. as mentioned in section . . , some electronic properties are scalable. they change with the changing size of the cluster until the bulk limit is reached. known examples of such properties are the homo-lumo gap (band gap) and the adiabatic electron detachment energy. for instance, the band gap of zno nanoparticles decreases with increasing diameter of the particle up to the bulk value observed for about nm [ ] . similarly, the bulk limits of the homo-lumo gap and the detachment energy for titanium oxide anion clusters of increasing size (increasing n) were reached already for n= [ , ] . in the classic formalization of qsars, electronic properties (e.g., homo, lumo, ionization potential) have been utilized as "ordinary" molecular descriptors. as discussed above, this approach should be revised for nanoparticles, for which the properties vary with size of a particle and this variation cannot be simply described by a linear function. it is not out of the question that similar phenomena might be observed also for other types of the "traditional" descriptors and further studies in this area are required and strongly justified. regarding the five oecd principles for the validation of a (q)sar as discussed in chapters and , an ideal qsar model, applicable for regulatory purpose, should be associated with (i) a well-defined endpoint; (ii) an unambiguous algorithm; (iii) a defined domain of applicability; (iv) appropriate measures of goodness-offit, robustness, and predictivity; and (v) a mechanistic interpretation, if possible. unfortunately, it is extremely difficult to fulfill all of these principles for (q)sars applicable to nanomaterials. there are two main difficulties related to the development of nano-qsars. the first one is lack of sufficiently numerous and systematic experimental data, while the second one is very limited knowledge on mechanisms of toxic action. as we mentioned many times, regarding their structure, the class of nanomaterials is not homogenous, combining a range of physico-chemical properties, as well as possible mechanisms of metabolism and toxicity. thus, it is impossible to assume one common applicability domain for all nanomaterials. each mode of toxicity and each class of nanomaterials should be studied separately. analyzing the literature data (section . ) it must be concluded that even if a class of structurally similar nanoparticles is tested with the same laboratory protocol, the number of tested compounds is often insufficient to perform comprehensive internal and external validation of a model and to calculate the appropriate measures of robustness and predictivity in qsar. for instance, limbach et al. [ ] have proposed two rankings of cytotoxicity of seven oxide nanoparticles based on the in vitro study of human and rodent cells. the rankings were as follows: (i) fe o ≈asbestos > zno > ceo ≈zro ≈tio ≈ca (po ) and (ii) zno > asbestos≈zro > ca (po ) ≈fe o ≈ceo ≈tio , respectively, for human (mesothelinoma) and rodent cells. in another paper by the same research group, the authors have found that for four metal nanoparticles -namely, tio , fe o , mn o , and co o -the chemical composition was the main factor determining the formation of reactive oxygen responsible for toxicity toward human lung epithelial cells [ ] . obviously, the results cannot be combined together and a data set containing five or six compounds is too small to build an appropriately validated qsar model. do these restrictions and problems mean qsar modelers are not able to provide useful and reliable information for nanoparticles? we do not believe this to be true. the amount of data will increase along with increasing number of nanotoxicological studies. however, no one can expect the accumulation in the next few years of such extensive data for nanomaterials, as it is now available for some environmental pollutants, pharmaceuticals, and "classical" industrial chemicals [ , ] . despite the limitations, there are some very promising results of preliminary nano-qsar studies which are reviewed below. toropov et al. [ ] have developed two models defining the relationships between basic physico-chemical properties (namely, water solubility, log s, and n-octanol/water partition coefficient, log p) of carbon nanotubes and their chiral vectors (as structural descriptors). the two-element chiral vector (n, m) contains information about the process of rolling up the graphite layer when a nanotube is formed. it had been previously known [ ] that the elements of the chiral vector are related to conductivity. at this point, toropov et al. confirmed, using the qsprbased research, that the vector is also strictly related to other properties. the models developed were defined by the following two equations (eqs. - and - ): log s = − . − . n − . m r = . , s = . , f = log p = − . + . n − . m r = . , s = . , f = . ( - ) the study was based on experimental data being available for only types of carbon nanotube. to perform an external validation, the authors divided the compounds into a training set (n= ) and a test set (n test = ). statistics of the validation were r test = . , s test = . , and f test = . and r test = . , s test = . , and f test = . , respectively, for the models for water solubility and n-octanol/water partition coefficient. without doubt, these were the first such qspr models developed for nanoparticles. however, the ratio of descriptors to compounds (the topliss ratio) was low, thus the model might be unstable (see discussion in chapter for more detail). another contribution by toropov and leszczynski [ ] presents a model predicting young's modulus (ym) for a set of inorganic nanostructures (eq. - ). martin et al. [ ] have proposed two qsar models predicting the solubility of buckminsterfullerene (c ), respectively, in n-heptane (log s heptane ) and n-octanol (log s octanol ) (eqs. - and - ): the symbols r and s refer to leave- %-out cross-validation. the authors applied codessa descriptors, namely, rncg -relative negative charge (zefirov's pc); asic -average structural information content of the second order; e min ee (cc) -minimum exchange energy for a c-c bond; ic -first-order information content; and rpcs -relative positive charged surface area. interestingly, the models were calibrated on compounds including polycyclic aromatic hydrocarbons (pahs) containing between two and six aromatic rings and the fullerene. although values of solubility predicted for the fullerene seem to be reasonable, the authors did not validate the applicability domain of the models. indeed, the structural difference between hydrocarbons and the fullerene is probably too large to make reliable predictions for c (the polycyclic hydrocarbons are planar, but the fullerene is spherical). in addition, the experimental values of log s for pahs ranged from - . to . in heptane and from − . to − . in octanol, while the experimental values for the fullerene were − . and . in heptane and octanol, respectively. an interesting area of nano-qsar applications is estimating solubility of a given nanoparticle in a set of various solvents. in that case, the main purpose of molecular descriptors is to correctly characterize the variation in interactions between the particle and the molecules of different solvents [ ] . in fact, it means that the descriptors are related to the structure of solvents rather than to the nanoparticle structure. murray et al. [ ] have developed a model characterizing the solubility of c in organic solvents by employing three following descriptors: two quantities, σ tot and υ reflecting variability and degree of balance of electrostatic potential on the solvent surface and the surface area, sa (eq. - ). , s= . ) , nothing is known about its predictive ability, because the model has not been validated. a set of linear models built separately for individual structural domains, namely alkanes (n= ), alkyl halides (n= ), alcohols (n= ), cycloalkanes (n= ), alkylbenzenes (n= ), and aryl halides (n= ), was published by sivaraman et al. [ ] . the models were based on connectivity indices, numbers of atoms, polarizability, and variables indicating the substitution pattern as molecular descriptors for the solvents. the values of r for particular models ranged between . (alkyl halides) and . (cycloalkanes) with the corresponding values of s from . (alkyl halides) to . (cycloalkanes). the authors concluded that it was impossible to obtain a unified model that included all solvents. however, when the first three classes of solvents (i.e., alkanes, alkyl halides, and alcohols) were combined together into one model, the results of an external validation performed were satisfactory. as well as linear approaches, non-linear models have been constructed. for instance, kiss et al. [ ] applied an artificial neural network utilizing molar volume, polarizability parameter, lumo, saturated surface, and average polarizability as structural descriptors of solvents. they observed that for most of the solvents studied (n= ) solubility decreases with increasing molar volume and increases with polarizability and the saturated surface areas of the solvents. the reported value of s in that case was . of log units. the values of r and f were . and , respectively. in another study [ ] the authors proposed modeling with both multiple linear regression with heuristic selection of variables (hm-mlr) and a least-squares support vector machine (svm). then they compared both models with each other. both models were developed with codessa descriptors [ ] . interestingly, the results were very similar (the model using svm had slightly better characteristics). the values of r for the linear and non-linear model were, respectively, . and . , while the values of f were and . the reported root mean square errors were . for the linear model (hm-mlr) and . for the model employing svm. when analyzing all the results it might be concluded that the main factor responsible for differences in the model error is related to the type of the descriptors rather than to the mathematical method of modeling. recently, toropov et al. [ ] developed an externally validated one-variable model for c solubility using additive optimal descriptors calculated from the international chemical identifier (inchi) code (eq. - ): log s = − . (± . ) + . (± . ) dcw(inchi) n = , the descriptor dcw(inchi) is defined as the sum of the correlation weights cw(i k ) for individual ichi attributes i k characterizing the solvent molecules. the example of the dcw(inchi) calculation is presented in table - . the values of cw(i k ) were optimized by the monte carlo method. all of the above models refer to physico-chemical properties as the endpoints, thus they are also termed quantitative structure-property relationships (qsprs). currently, there are only a small number of qsars related directly to nanomaterials' activity. in tsakovska [ ] proposed the application of qsar methodology to predict protein-nanoparticle interactions. in durdagi et al. published two papers [ , ] presenting qsar-based design of novel inhibitors of human immunodeficiency virus type aspartic protease (hiv- pr). in the first work [ ] the authors developed a three-dimensional qsar model with comparative molecular similarity indices analysis (comsia) method for derivatives of fullerene c . the values of r and q for the training set (n= ) were . and . , respectively. the absolute values of residuals in the validation set (n= ) ranged from . to . logarithmic units of ec (μm). the second model [ ] were characterized by lower values of the statistics (n= , r = . and q = . ). however, in that case the predictions for an external set of compounds (n test = ) were possible with an acceptable level of error. in addition, the authors proposed nine novel structures indicating possible inhibitor activity based on the model obtained. they concluded that steric effects play the most important role in the inhibition mechanism as well as electrostatic and h-donor/acceptor properties. however, the last two types of interactions are of lower importance. similarly, smiles-based optimal descriptors have been successfully applied for modeling hiv- pr fullerene-based inhibitors [ ] . the model reported by toropov et al. [ ] was described by the following equation and parameters: the dcw descriptor in this case is defined as the following (eq. - ): cw(sa k ) where the sa k is a smiles attribute, i.e., one symbol (e.g., "o," "=," "v") or two symbols (e.g., "al," "bi," "cu") in the smiles notation. numbers of double bonds have been used as global smiles attributes. they are denoted as "= " and "= ." "= " is the indicator of one double bond and "= " is the indicator of two double bonds. although we strongly believe in the usefulness and appropriateness of qsar methodology for nanomaterial studies, the number of available models related to activity and toxicity is still very limited. when analyzing the situation, it seems that the main limitation is insufficient amount of existing experimental data. in many cases, lack of data precludes an appropriate implementation of statistical methods, including necessary external validation of the model. the problem of the paucity of data will be solved only when a strict collaboration between the experimentalists and qsar modelers is established. the role of the modelers in such studies should not be restricted only to rationalization of the data after completing the experimental part, but also they must be involved in the planning of the experimentation. since the experiments on nanomaterials are usually expensive, a kind of compromise between the highest possible number of compounds for testing and the lowest number of compounds necessary for developing a reliable qsar model should be reached. regarding the limited amount of data and high costs of the experiments, the idea of applying novel read-across techniques enabling preliminary estimation of data (chapter ) [ , ] is very promising. however, no one has yet tried to implement this technique to nanomaterials. without doubt, a large and increasing aspect of the near future of chemistry and technology will be related to the development of nanomaterials. on one hand, due to their extraordinary properties, nanomaterials are becoming a chance for medicine and industry. but, on the other hand, the same properties might result in new pathways and mechanisms of toxic action. in effect, the work with nanomaterials is challenging for both "types" of chemists: those who are searching for and synthesizing new chemicals and those who are working on risk assessment and protection of humans from the effects of these chemicals. when analyzing the current status of nano-qsar, the four noteworthy suggestions for further work can be made: . there is a strong need to supplement the existing set of molecular descriptors by novel "nanodescriptors" that can represent size-dependent properties of nanomaterials. . a stronger than usual collaboration between the experimentalists and nano-qsar modelers seems to be crucial. on one hand, it is necessary to produce data of higher usefulness for qsar modelers (more compounds, more systematic experimental studies within groups of structural similarity, etc.). on the other hand, a proper characterization of the nanomaterials structure is not possible only at the theoretical (computational) level. in such situation, experiment-based structural descriptors for nano-qsar might be required. . it is possible that the current criteria of the models'quality (the five oecd rules) will have to be re-evaluated and adapted to nanomaterials. this is due to the specific properties of chemicals occurring at the "nano" level (i.e., electronic properties change with changing size) and the very limited number of data (problems with the "classic" method of validation which is biased to small, low molecular weight molecules). . greater effort is required in the areas of grouping nanomaterials and nano-readacross. this technique might be useful especially at the initial stage of nano-qsar studies, when the experimental data are scarce. in summary, the development of reliable nano-qsar is a serious challenge that offers an exciting new direction for qsar modelers. this task will have to be completed before the massive production of nanomaterials in order to prevent potentially hazardous molecules from being released into the environment. in the long term, prevention is always more efficient and cheaper than clean-up. there's plenty of room at the bottom. an invitation to enter a new field of physics the potential risks of nanomaterials: a review carried out for ecetoc inorganic and organic uv filters: their role and efficiency in sunscreens and suncare products solar-driven self-cleaning coating for a painted surface synthesis activity and characterization of textiles showing self-cleaning activity under daylight irradiation synthesis, characterization and catalytic properties of cuo nanocrystals with various shapes consumer products inventory of nanotechnology products ( ) the project on emerging nanotechnologies the use of nanocrystals in biological detection fluorescent cdse/zns nanocrystal-peptide conjugates for long-term, nontoxic imaging and nuclear targeting in living cells helical microtubules of graphitic carbon fullerene containing polymers: a review on their synthesis and supramolecular behavior in solution role of carbon nanotubes in electroanalytical chemistry: a review environmental applications of carbon-based nanomaterials the challenge of regulating nanomaterials toxic potential of materials at the nanolevel size effects on the band-gap of semiconductor compounds room temperature creep behavior of nanocrystalline nickel produced by an electrodeposition technique high tensile ductility in a nanostructured metal recent development in the mechanics of superplasticity and its applications recent advances in nanotechnology magnetic properties of nanoparticle assemblies one-pot reaction to synthesize biocompatible magnetite nanoparticles titanium dioxide photocatalysis tensile creep behavior of alumina/silicon carbide nanocomposite health and environmental impact of nanotechnology: toxicological assessment of manufactured nanoparticles significance of particle parameters in the evaluation of exposure-doseresponse relationships of inhaled particles particle toxicology: from coal mining to nanotechnology current hypotheses on the mechanisms of toxicity of ultrafine particles dosimetry and toxicology of ultrafine particles nanotoxicology: an emerging discipline evolving from studies of ultrafine particles long-term clearance kinetics of inhaled ultrafine insoluble iridium particles from the rat lung, including transient translocation into secondary organs translocation of inhaled ultrafine particles to the brain translocation of inhaled ultrafine manganese oxide particles to the central nervous system modified opinion on the appropriateness of the risk assessment methodology in accordance with the technical guidance documents for new and existing substances for assessing the risks of nanomaterials. scientific committee on emerging and newly identified health risks principles for characterizing the potential human health effects from exposure to nanomaterials: elements of a screening strategy ultrafine particulate pollutants induce oxidative stress and mitochondrial damage comparison of the abilities of ambient and manufactured nanoparticles to induce cellular toxicity according to an oxidative stress paradigm combustion-derived ultrafine particles transport organic toxicants to target respiratory cells cytotoxicity of single-wall carbon nanotubes on human fibroblasts formation of nucleoplasmic protein aggregates impairs nuclear function in response to sio nanoparticles biological behavior of hat-stacked carbon nanofibers in the subcutaneous tissue in rats oxidative stress-induced dna damage by particulate air pollution reactive sulfur species: an emerging concept in oxidative stress ed) nanonotoxicology: characterization, dosing and health effects oxidative stress: oxidants and antioxidants size-dependent proinflammatory effects of ultrafine polystyrene particles: a role for surface area and oxidative stress in the enhanced activity of ultrafines unusual inflammatory and fibrogenic pulmonary responses to single-walled carbon nanotubes in mice ultrafine particles bacterial cell association and antimicrobial activity of a c water suspension the differential cytotoxicity of water-soluble fullerenes cellular toxicity of carbon-based nanomaterials pulmonary effects of inhaled zinc oxide in human subjects, guinea pigs, rats, and rabbits investigations on the inflammatory and genotoxic lung effects of two types of titanium dioxide: untreated and surface treated comparing study of the effect of nanosized silicon dioxide and microsized silicon dioxide on fibrogenesis in rats combustion-derived nanoparticles: a review of their toxicology following inhalation exposure damage to dna by reactive oxygen and nitrogen species: role in inflammatory disease and progression to cancer oxyradicals and dna damage role of oxygen free radicals in carcinogenesis and brain ischemia neurotoxicity of low-dose repeatedly intranasal instillation of nano-and submicron-sized ferric oxide particles in mice hallervorden-spatz syndrome and brain iron metabolism hemolytic effects of water-soluble fullerene derivatives preclinical studies to understand nanoparticle interaction with the immune system and its potential effects on nanoparticle biodistribution effect-oriented physicochemical characterization of nanomaterials manufactured nanomaterials (fullerenes, c ) induce oxidative stress in the brain of juvenile largemouth bass the myth about toxicity of pure fullerenes is irreversibly destroyed. eighth biennial workshop "fullerenes and atomic clusters" iwfac' daphnia magna mortality when exposed to titanium dioxide and fullerene (c ) nanoparticles is the c fullerene molecule toxic?! fuller nanotub carbon nanostruct comparative toxicity of nano-scale tio , sio and zno water suspensions comparative eco-toxicity of nanoscale tio , sio , and zno water suspensions toxicity of nanosized and bulk zno, cuo and tio to bacteria vibrio fischeri and crustaceans daphnia magna and thamnocephalus platyurus toxicity of single walled carbon nanotubes to rainbow trout, (oncorhynchus mykiss): respiratory toxicity, organ pathologies, and other physiological effects characterisation and in vivo ecotoxicity evaluation of double-wall carbon nanotubes in larvae of the amphibian xenopus laevis handbook of molecular descriptors d-qsar of human immunodeficiency virus (i) protease inhibitors. iii. interpretation of comfa results d-qsar analysis of a set of ecdysteroids and a comparison to comfa modeling in silico modelling of hazard endpoints: current problems and perspectives hierarchic system of qsar models ( d- d) on the base of simplex representation of molecular structure a new approach to the characterization of nanomaterials: predicting young's modulus by correlation weighting of nanomaterials codes predicting water solubility and octanol water partition coefficient for carbon nanotubes based on the chiral vector computational nanotoxicology -towards a structure-activity based paradigm for investigation of the activity of nanoparticles guidance document on the grouping of chemicals. organisation of economic cooperation and development assessing exposure to airborne nanomaterials: current abilities and future requirements smiles as an alternative to the graph in qsar modelling of bee toxicity additive smiles-based optimal descriptors in qsar modelling bee toxicity: using rare smiles attributes to define the applicability domain optimisation of correlation weights of smiles invariants for modelling oral quail toxicity smiles in qspr/qsar modeling: results and perspectives additive in chi-based optimal descriptors: qspr modeling of fullerene c solubility in organic solvents qspr modeling mineral crystal lattice energy by optimal descriptors of the graph of atomic orbitals the graph of atomic orbitals and its basic properties. . wiener index the graph of atomic orbitals and its basic properties. . zagreb indices the signature molecular descriptor. . inversequantitative structure-activity relationship of icam- inhibitory peptides the signature molecular descriptor. . enumerating molecules from their extended valence sequences the signature molecular descriptor. . canonizing molecules using extended valence sequences the signature molecular descriptor. . using extended valence sequences in qsar and qspr studies introduction to computational chemistry development of the colle-salvetti correlation energy formula into a functional of the electron density density-functional thermochemistry. iii the role of exact exchange periodic boundary conditions in ab initio calculations electronic and structural properties of the ( ) and ( ) zno surfaces density functional theory study on the structural and electronic properties of low index rutile surfaces for tio /sno /tio and sno /tio /sno composite systems clusters: a bridge across the disciplines of physics and chemistry a new concept of molecular nanodescriptors for qsar/qspr studies nanoparticle analysis and characterization methodologies in environmental risk assessment of engineered nanoparticles a density functional theory study on the effect of shape and size on the ionization potential and electron affinity of different carbon nanostructures quantum-size effects in capped and uncapped carbon nanotubes end-cap effects on vibrational structures of finitelength carbon nanotubes the electronic structures and properties of open-ended and capped carbon nanoneedles modelling nanoneedles: a journey towards nanomedicine modeling the structure-property relationships of nanoneedles: a journey toward nanomedicine ab initio quantum chemical studies of fullerene molecules with substituents c x x=si variable band gap zno nanostructures grown by pulsed laser deposition theoretical study of the electronic structure and stability of titanium dioxide clusters (tio ) n with n= - probing the electronic structure and band gap evolution of titanium oxide clusters (tio ) n -(n= − ) using photoelectron spectroscopy oxide nanoparticle uptake in human lung fibroblasts: effects of particle size, agglomeration, and diffusion at low concentrations in vitro cytotoxicity of oxide nanoparticles: comparison to asbestos, silica, and the effect of particle solubility handbook of physical-chemical properties and environmental fate for organic chemicals clar valence bond representation of pi-bonding in carbon nanotubes qspr modeling of solubility of polyaromatic hydrocarbons and fullerene in -octanol and n-heptane solubility of fullerene representation of c solubilities in terms of computed molecular surface electrostatic potentials and areas qspr modeling for solubility of fullerene (c ) in organic solvents artificial neural network approach to predict the solubility of c in various solvents accurate quantitative structure-property relationship model to predict the solubility of c in various solvents based on a novel approach using a least-squares support vector machine codessa pro. comprehensive descriptors for structural and statistical analysis computational modelling of nanoparticles d qsar comfa/comsia, molecular docking and molecular dynamics studies of fullerene-based hiv- pr inhibitors computational design of novel fullerene analogues as potential hiv- pr inhibitors: analysis of the binding interactions between fullerene inhibitors and hiv- pr residues using d qsar, molecular docking and molecular dynamics simulations smiles-based optimal descriptors: qsar analysis of fullerene-based hiv- pr inhibitors by means of balance of correlations an application of graphs of atomic orbitals for qsar modeling of toxicity of metal oxides. th annual federation of analytical chemistry and spectroscopy societies toward in silico approaches for investigating the activity of nanoparticles in therapeutic development key: cord- - esrper authors: lin, cheng-yung; chiang, cheng-yi; tsai, huai-jen title: zebrafish and medaka: new model organisms for modern biomedical research date: - - journal: j biomed sci doi: . /s - - - sha: doc_id: cord_uid: esrper although they are primitive vertebrates, zebrafish (danio rerio) and medaka (oryzias latipes) have surpassed other animals as the most used model organisms based on their many advantages. studies on gene expression patterns, regulatory cis-elements identification, and gene functions can be facilitated by using zebrafish embryos via a number of techniques, including transgenesis, in vivo transient assay, overexpression by injection of mrnas, knockdown by injection of morpholino oligonucleotides, knockout and gene editing by crispr/cas system and mutagenesis. in addition, transgenic lines of model fish harboring a tissue-specific reporter have become a powerful tool for the study of biological sciences, since it is possible to visualize the dynamic expression of a specific gene in the transparent embryos. in particular, some transgenic fish lines and mutants display defective phenotypes similar to those of human diseases. therefore, a wide variety of fish model not only sheds light on the molecular mechanisms underlying disease pathogenesis in vivo but also provides a living platform for high-throughput screening of drug candidates. interestingly, transgenic model fish lines can also be applied as biosensors to detect environmental pollutants, and even as pet fish to display beautiful fluorescent colors. therefore, transgenic model fish possess a broad spectrum of applications in modern biomedical research, as exampled in the following review. although zebrafish (danio rerio) and medaka (oryzias latipes) are primitive vertebrates, they have several advantages over other model animals. for example, they are fecund and light can control their ovulation. spawning takes place frequently and no limitation in their spawning season. microinjection of fertilized eggs is easily accessible and relatively cheap. their embryos are transparent, making it easy to monitor the dynamic gene expression in various tissues and organs in vivo without the need to sacrifice the experimental subjects. their genome sizes are approximately to % of the mammalian genome, making them the only vertebrates available for large-scale mutagenesis. their maturation time takes only ~ months, which is relatively less laborious and time-saving for generating transgenic lines. in addition, many routine techniques of molecular biology and genetics, including knock-in, knockdown and knockout, are well developed in the model fish. therefore, zebrafish and medaka are new excellent animal systems for the study of vertebrate-specific biology in vivo. the f transgenic line can be established once the exogenous gene can be successfully transferred to the embryos, followed by stable germline transmission of the transgene to the f generation. generally, around - % of treated embryos have a chance to achieve germline transmission [ ] . it has been reported that a foreign gene flanked with inverted terminal repeats of adeno-associated virus can be used to enhance the ubiquitous expression and stable transmission of transgene in model fish [ ] . meanwhile, transgenesis can be facilitated by using the tol transposon derived from medaka [ ] . transposase catalyzes transposition of a transgene flanked with the tol sequence [ ] . the efficiency of tol- -mediated germline transmission could range from to % of injected embryos [ , ] . a cutting-edge technique has taken the study of fish gene transfer to new horizons, such as knockout zebrafish by the transcription activator-like effector nuclease (talen) system and the clustered regularly interspaced short palindromic repeats (crispr) combined with crispr-associated proteins (cas ) [ , ] . the talen system involves the dna recognition domain of transcription activator-like effectors (tales) and a nuclease domain for generation of nicks on dna sequences. the crispr/cas system directed by a synthetic single guide rna can induce targeted genetic knockout in zebrafish. the main difference between these two systems is based on their recognition mechanisms. unlike the tales applied in the talen system, the crispr/cas system recognizes its target dna fragment by the complementary non-coding rna. the development of the talen and crispr/cas systems provides new genomic editing approaches for establishing genetic knockout fish lines [ ] . the fluorescence protein gene (fpg) has been widely applied as a reporter gene in studies of the transgene expression by direct visualization under fluorescent microscopy in vivo [ ] . many transgenic model fish lines harbor an fpg driven by various tissue-specific promoters, including the erythroid-specific gata promoter [ ] , muscle-specific α-actin promoter [ ] , rod-specific rhodopsin promoter [ ] , neuron-specific isl- promoter [ ] , pancreas-specific pdx- and insulin promoters [ ] , myocardium-specific cmlc promoter [ ] , liver-specific l-fabp promoter [ ] , bone-specific col a promoter [ ] , macrophage-specific mfap promoter [ ] , and germ cell-specific vasa promoter [ ] . using medaka β-actin promoter, tsai's lab generated a transgenic line of medaka displaying green fp ubiquitously around the whole fish from f through f generations in a mendelian inheritance manner [ ] . this is known as the first transgenic line of glowing pet fish, which was reported by science [ ] and far eastern economic review [ ] and honored to be selected as among "the coolest inventions of " by time [ ] . the dna sequences of the aforementioned promoters ranging from . to . kb are sufficient to drive the fpg reporter to mimic the tissue-specific expression of endogenous gene. however, some genes require a longer regulatory dna sequence, such as more than kb, to fully recapitulate the characteristic expression profiles of endogenous genes. in that case, bacterial artificial chromosome (bac) and phage p -dereived artificial chromosome (pac) have been commonly used for this purpose [ ] . for example, the zebrafish rag gene, flanked with pac dna containing kb at the ′ upstream and kb at the ′ downstream, can be expressed specifically in lymphoid cells. instead of using the tedious chi-site dependent approach, jessen et al. reported a two-step method to construct a bac clone [ ] . employing this protocol, chen et al. constructed a bac clone containing the upstream kb range of zebrafish myf and generated a transgenic line tg (myf :gfp) [ ] . this transgenic line is able to recapitulate the somite-specific and stagedependent expression of the endogenous myf at an early developmental stage. in summary, all the above transgenic lines should be very useful materials for studying both gene regulation and cell development. zebrafish is particularly useful for studying heart development for the following reasons: (a) zebrafish have a primitive form of the heart, which is completely developed within h post-fertilization (hpf). (b) the cardiac development can be easily observed in the transgenic line possessing a fp-tagged heart. (c) the zebrafish embryos with a defective cardiovascular system can still keep on growing by acquiring oxygen diffused from water. (d) discovery of genes involved in heart development can be facilitated by a simple haploid mutation method [ ] . for example, using the zebrafish jekyll mutant, which has defective heart valves, walsh and stainier discovered that udpglucose dehydrogenase is required for zebrafish embryos to develop normal cardiac valves [ ] . tsai's lab is the first group to generate a transgenic zebrafish line that possesses a gfp-tagged heart [ ] . this line was established from zebrafish embryos introduced with an expression construct, in which the gfp reporter is driven by an upstream control region of zebrafish cardiac myosin light chain gene (cmlc ). using this transgenic line, raya et al. found that the notch signaling pathway is activated is during the regenerative response [ ] . shu et al. reported that na, k-atpase α b and α isoforms have distinct roles in the patterning of zebrafish heart [ ] . this transgenic line should also be useful for studying the dynamic movement and cell fate of cardiac primordial cells. for example, forouhar et al. proposed a hydro-impedance pump model for the embryonic heart tubes of zebrafish [ ] . a d dynamic image of cardiac development has been developed [ ] . furthermore, hami et al. reported that a second heart field is required during cardiac development [ ] . thus, recently, nevis et al. stated that tbx plays a function for proliferation of the second heart field, and the zebrafish tbx -null mutant resemble the heart defects in digeorge syndrome [ ] . thus, the expression pattern of heart-specific genes could be analyzed based on heart progenitor cells collected in this transgenic line. the analysis of gene or protein expression dynamics at different developmental stages could also be conducted. furthermore, this transgenic fish is a potential platform for detecting chemicals, drugs and environmental pollutants affecting heart development, as detailed in following section. in vivo transient assay of the injected dna fragments in model fish embryos is a simple yet effective way to analyze the function of regulatory cis-elements. for example, myf , one of myogenesis regulatory factors (mrf), plays key roles in the specification and differentiation of muscle primordial cells during myogenesis. the expression of myf is somite-specific and stage-dependent, and its activation and repression are delicately orchestrated. using in vivo transient assay, chen et al. found that a novel ciselement located at − /- is essential for somite-specific expression of myf [ ] . lee et al. revealed that this − /- cis-element is specifically bound by forkhead box d , and proposed that somite development is regulated by the pax -foxd -myf axis [ ] . besides foxd , foxd , another protein in the forkhead box family, is necessary for maintaining the anterior-posterior polarity of somite cells in mesenchymal-epithelial transition [ ] . the expression of foxd is regulated by fgf signaling in anterior presomitic mesoderm (psm), which indicates that fgf-foxd -mesp signaling takes place in somitogenesis [ ] . furthermore, analysis of the loci of adjacent mrf and myf revealed the complicated regulation mechanism of the mrf genes. it was also found that the biological function of mrf is related to myofibril alignment, motor axon growth, and organization of axonal membrane [ ] . the molecular mechanism that underlies the repression of myf has also been reported. for example (fig. a) , a strong repressive element of zebrafish myf was found within intron i (+ /+ ) [ ] . this repressive element is modulated by a novel intronic microrna, termed mir-in or mir- [ ] . when myf transcripts reach the highest level after specification, the accumulated mir- starts to reduce the transcription of myf through silencing the positive factor dickkopf-related protein (dkk r or dkk a) for the myf promoter [ ] . itgα b is a receptor of secretory dkk a and that interaction between itgα b and dkk a is required to drive the downstream signal transduction which regulates myf promoter activity in somite during embryogenesis of zebrafish [ ] . dkk a regulates p a phosphorylation to maintain smad stability, which in turn enables the formation of the smad / a/ complex required for the activation of the myf promoter [ ] . however, when myf transcripts are reduced at the later differentiation stage, mir- is able to be transcribed by its own promoter [ ] (fig. b) . furthermore, increased expression of mir- interacts with its receptor itgα b, resulting in the phosphorylation of p a and the formation of the smad / a/ complex, which in turn, activates the myf promoter activity. when myf is highly transcribed, the intronic mir- suppresses the transcription of myf through silencing the dkk a [ , [ ] [ ] [ ] . b at the late muscle development, mir- starts transcription at its own promoter and switches to silence homer b to control the homeostasis of intracellular calcium concentration ([ca + ] i ) in fast muscle cells [ ] . either mir- -knockdown or homer- b-overexpression causes the increase of homer- b protein, resulting in an enhanced level of [ca + ] i , which in turn, disrupts sarcomeric actin filament organization. in contrast, either mir- -overexpression or homer- b-knockdown causes the decrease of homer- b, resulting in a reduced [ca + ] i and thus a defective muscle phenotype [ ] controls the intracellular concentration of ca + ([ca + ] i ) in fast muscle cells through subtly reducing homer- b expression. the homeostasis of [ca + ] i is required during differentiation to help maintain normal muscle development [ ] . nevertheless, it remains to be investigated how mir- switches its target gene at different developmental stages. apart from the regulation of somitogenesis, myf is also involved in craniofacial muscle development. the functions of myf in cranial muscles and cartilage development are independent of myod, suggesting that myf and myod are not redundant. thus, three possible pathways could be associated with the molecular regulation between myf and myod: (i) myf alone is capable of initiating myogenesis, (ii) myod initiates muscle primordia, which is subdivided from the myf -positive core, and (iii) myod alone, but not myf , modulates the development of muscle primordia [ ] . furthermore, the six a gene was found to play an important role in the interaction between myf and myod [ ] . in cartilage development, myf is expressed in the paraxial mesoderm at the gastrulation stage. myf plays a role in mesoderm fate determination by maintaining the expression of fgf / , which in turn, promotes differentiation from neural crest cells to craniofacial cartilage [ ] . this research on myf not only reveals that it has different functions between craniofacial muscle development and somitogenesis, but it also opens up a new study field for understanding craniofacial muscle development. hinits et al. reported no phenotype both in myf knockdown embryos and myf -null mutant, suggesting that myf is rather redundant in somitogenesis of zebrafish [ ] . however, it is hard to reasonably explain why these embryos and mutant are all lethal and can't grow to adulthood. on the other hand, lin et al. reported an observable defective phenotype in myf -knockdown embryos, and claimed that the concentration of myf -mo they used can inhibit maternal myf mrna translation [ ] . this discrepancy might attribute the effectiveness of mo used or the different phenotypes between the knockdown embryos and knockout mutant in this case. the retina-specific expression of the carp rhodopsin gene is controlled by two upstream regulatory dna ciselements [ ] . one is located at − to − , which is the carp neural retina leucine zipper response-like element; the other is located at − to − , which is a carpspecific element crucial to reporter gene expression in medaka retinae. intriguingly, immediate activation of early growth response transcriptional regulator egr could result in the incomplete differentiation of retina and lens, leading to microphthalmos [ ] . another important factor for ocular development is the adpribosylation factor-like interacting protein (arl ip ). loss of arl ip function leads to the absence of retinal neurons, disorganized retinal layers and smaller optic cups [ ] . upon losing arl ip , retinal progenitors continued to express cyclin d , but not shh or p kip , suggesting that eye progenitor cells remained at the early progenitor stage, and could not exit the cell cycle to undergo differentiation [ ] . additionally, it has been reported that arl ip is essential for specification of neural crest derivatives, but not neural crest induction. tu et al. found that arl ip mutation causes abnormal neural crest derivative tissues as well as reduced expression of neural crest specifier genes, such as foxd , snail b and sox , indicating that arl ip is involved in specification, but not induction, of neural crest cells [ ] . furthermore, they found that arl ip could play an important role in the migration of neural crest cells because in the arl ip -knockdown embryos, crestinand sox -expressing neural crest cells failed to migrate ventrally from neural tube into trunk. more recently, lin et al. found that ras-related nuclear (ran) protein is conjugated with arl ip , and proposed that ran protein associates with arl ip to regulate the development of retinae [ ] . to date, no in vivo model system has been established to identify cells in the cns that can specifically respond with regeneration after stresses, and, even if identified, no method is in place to trace these responsive cells and further identify their cell fates during hypoxic regeneration. to address these issues, lee et al. generated a transgenic zebrafish line huorfz, which harbors the upstream open reading frame (uorf) from human ccaat/enhancer-binding protein homologous protein gene (chop), fused with the gfp reporter and driven by a cytomegalovirus promoter [ ] . after huorfz embryos were treated with heat-shock or under hypoxia, the gfp signal was exclusively expressed in the cns, resulting from impeding the huorf chop -mediated translation inhibition [ ] . interestingly, zeng et al. found that gfp-(+) cells in spinal cord respond to stress, survive after stress and differentiate into neurons during regeneration (chih-wei zeng, yasuhiro kamei and huai-jen tsai, unpublished data). micrornas (mirnas) are endogenous single-stranded rna molecules of - nucleotides (nt) that repress or activate the translation of their target genes through canonical seed-and non-canonical centered mirna binding sites. the known mechanisms involved in mirnas-mediated gene silencing are decay of mrnas and blockage of translation [ ] [ ] [ ] . probably the expression of ~ % of human genes is regulated by mirnas [ , ] . therefore, to understand gene and function in cells or embryos, it is important to exactly know the target gene(s) of a specific mirna at different phase of cells or at particular stages of developing embryos. instead of using a bioinformatic approach, the tsai's lab developed the labeled mirna pull-down (lamp) assay system, which is a simple but effective method to search for the candidate target gene(s) of a specific mirna under investigation [ ] . lamp assay system yields fewer falsepositive results than a bioinformatic approach. taking advantage of lamp, scientists discovered that mir- silences different target genes at different developmental stages, e.g., at early stage, mir- targets dkk a [ ] , while at late stage, it targets homer- b [ ] (fig. ) . in another example (fig. ) , mir- and mir- are two muscle-specific micrornas sharing the same seed sequences. they are able to modulate the expression of vascular endothelial growth factor aa (vegfaa) and serve as cross-tissue signaling regulators between muscle and vessels. since mir- and mir- share identical seed sequences, stahlhut et al. demonstrated that they can silence the same target gene, such as vegfaa, and considered them as a single cross-tissue regulator termed as mir- / [ ] . mir- / reduces the level of vegfaa, resulting in the inhibition of the angiogenic signaling [ ] . surprisingly, using the lamp assay system, lin et al. reported that the target genes for mir- and mir- are different [ ] . while mir- targets vegfaa, mir- targets seryl-trna synthetase gene (sars). sars is a negative regulator of vegfaa. although both mir- and mir- have identical seed sequences, the sars- ′utrs of zebrafish, human and mouse origins can be recognized only by mir- in zebrafish embryos and mammalian cell lines (hek- t and c c ), but not by mir- [ ] . conversely, the vegfaa- ′utr is targeted by mir- , but not by mir- . therefore, lin et al. concluded that mir- and mir- are actually two distinct regulators and play opposing roles in zebrafish angiogenesis. the mir- /sars/vegfaa pathway promotes embryonic angiogenesis by indirectly controlling vegfaa, while mir- /vegfaa pathway plays an anti-angiogenic role by directly reducing vegfaa. interestingly, they also found that the mir- /sars/vegfaa pathway increasingly affects embryonic angiogenesis at late developmental stages in somitic cells [ ] . it remains to be studied how mir- increases in abundance at late stage. different from mammals, zebrafish have the ability to regenerate injured parts in the cns. many mirnas have been found in the cns. since mirnas are involved in many aspects of development and homeostatic pathways, they usually play important roles in regeneration [ ] . it has been shown that several mirnas have prominent fig. mir- and mir- silence different target genes and play opposing roles in zebrafish angiogenesis. both mir- and mir- are musclespecific micrornas and share identical seed sequences. however, they silence different target genes to affect the secreted vegfaa level through different pathways [ ] . the mir- /sars/vegfaa pathway plays a positive role in angiogenesis since sars, a negative factor for vegfaa promoter transcription, is silenced by mir- , resulting in the increase of vegfaa. however, the mir- /vegfaa pathway plays a negative role since vegfaa is silenced directly by mir- . dynamic changes of mir- and mir- levels are also observed [ ] . the mir- level gradually increases between and hpf and significantly increases further between and hpf, while the mir- level is only slightly changed during this same period. consequently, vegfaa increases greatly from to hpf, which might be responsible for the continuous increase of mir- /sars/vegfaa pathway, but not mir- /vegfaa pathway. therefore, temporal regulation of the expression of mir- and mir- with different target genes occur during embryonic angiogenesis in somitic cells of zebrafish functions in regulating the regeneration process. for example, mir- promotes spinal cord repair by enhancing angiogenesis [ ] , and the mir- family represses proliferation in the adult mouse heart [ ] . furthermore, mir-nas mir- b and mir- are identified following optic nerve crush. by gene ontology analysis, mir- b and mir- are found to regulate genes, including eva a, layna, nefmb, ina, si:ch - a . , smoc , and sb:cb . these genes are involved in cell survival or apoptosis, indicating that these two mirnas are potential regulators of optic nerve regeneration [ ] . although the main hematopoietic sites in zebrafish differ from those in mammals, both zebrafish and mammals share all major blood cell types that arise from common hematopoietic lineages [ ] . moreover, many genes and signaling pathways involved in hematopoiesis are conserved among mammals and zebrafish. for example, scl, one of the first transcription factors expressed in early hematopoietic cells, is evolutionarily conserved. during definitive hematopoiesis, runx marks hematopoietic stem cells (hscs) in both mouse and fish. additionally, in differentiated populations, gata , the erythroid lineage regulator, pu. and c/ebp, the myeloid lineage regulators, and ikaros, a mark of the lymphoid population, are in accordance with the hematopoietic hierarchy in zebrafish and mammals [ ] . thus, the findings with respect to zebrafish blood development could be applied to mammalian system. genetic screening in zebrafish has generated many blood-related mutants that help researchers understand hematopoietic genes and their functions [ ] . for example, the spadetail mutant carrying a mutated tbx exhibits defective mesoderm-derived tissues, including blood. this mutant displays the decrease levels tal , lmo , gata , fli and gata in the posterior lateral mesoderm, indicating the important role of tbx during hemangioblast regulation [ ] . chemical screening in zebrafish using biologically active compounds is also a powerful approach to identify factors that regulate hscs. for example, it is well known that prostaglandin (pg) e increases the induction of stem cells in the aorta-gonad-mesonephros region of zebrafish, as demonstrated by increasing expressions of runx and cmyb, which, in turn, increases engraftment of murine marrow in experimental transplantation [ ] . in human clinical trials, the treatment of cord blood cells with dimethyl pge caused an increase in long-term engraftment [ ] , suggesting that a compound identified in zebrafish could have clinical application in humans. model fish are excellent materials for the study of human diseases due to some mutants display similar phenotypes of human diseases [ ] . in addition, essential genes and thereof regulation to control the development of tissues or organs are highly conserved [ ] . for example, tbx is a t-box transcription factor responsible for cell-type specification and morphogenesis. the phenotypes of tbx mutant are highly similar among mammals and zebrafish. thus, transgenic fish with heart-specific fluorescence could provide a high-through screening platform for drugs for cardiovascular disease. for example, the tsai's lab established a transgenic line which could be induced to knock down the expression level of cardiac troponin c at any developmental stage, including embryos, larva or adult fish. the reduction of troponin c resulted in mimicry of dilated cardiomyopathy, and the incomplete atrioventricular blocking disease in humans. therefore, this transgenic line is expected to make a significant contribution to drug screening and the elucidation of the molecular mechanisms underlying cardiovascular diseases. next, the effect of drugs on embryonic development was also studied. amiodarone, which is a class iii antiarrhythmic agent, is being used for the treatment of tachyarrhythmia in humans. however, amiodarone-treated zebrafish embryos were found to exhibit backflow of blood in the heart [ ] . subsequent research showed that amiodarone caused failure of cardiac valve formation [ ] . specifically, amiodarone induces ectopic expression of similar to versican b (s-vcanb), resulting in repression of egfr/gsk β/snail signaling, which in turn, upregulates cdh at the heart field, and causes defective cardiac valves [ ] . moreover, amiodarone was found to repress metastasis of breast cancer cells by inhibiting the egfr/erk/snail pathway [ ] , a phenomenon analogous to the inhibitory effects of amiodarone on emt transition observed in the heart. last but not least, although zebrafish has a twochambered heart, relative to mouse, rat, and rabbit, its heart rate, action potential duration (apd) and electrocardiogram (ecg) morphology are similar to those of humans. [ , ] . additionally, tsai et al. demonstrated that the in vitro ecg recording of zebrafish heart is a simple, efficient and high throughput assay [ ] . thus, zebrafish can serve as a platform for direct testing of drug effect on apd prolongation and prolonged qt interval, which is required by the fda as a precondition for drug approval. zebrafish become a popular experimental animal for the studies of human cancer [ ] , in part because the fish homologs of human oncogenes and tumor suppressor genes have been identified, and in part because signaling pathways regulating cancer development are conserved [ ] [ ] [ ] . amatruda et al. reported that many zebrafish tumors are similar to those of human cancer in the histological examination [ ] . the zebrafish transgenic line with skin-specific red fluorescence could be applied for skin tumor detection [ ] . when the embryos of this line were treated with solutions containing arsenic, the tumors induced on the skin could be easily identified by naked eye under fluorescent microscope. therefore, this transgenic line can be potentially used for the study of skin diseases. for example, the common skin cancer melanoma may be screened by the red fluorescence expression in this transgenic line. zebrafish transgenic line could also be applied to establish models simulating melanoma development. the human oncogenic braf v e was expressed under the control of the zebrafish melanocyte mitfa promoter to establish a melanoma model [ ] . combining skinspecific red fluorescence with mitfa-driven oncogene expression, the melanoma could be easily traced. therefore, transgenic lines and mutants of model fish could provide abundant resources for mechanistic studies and therapeutic research in human diseases. metastasis involves processes of sequential, interlinked and selective steps, including invasion, intravasation, arrest in distant capillaries, extravasation, and colonization [ ] . zebrafish is again an alternative organism for in vivo cancer biology studies. in particular, xenotransplantation of human cancer cells into zebrafish embryos serves as an alternative approach for evaluating cancer progression and drug screening [ ] . for example, human primary tumor cells labeled with fluorescence have already been implanted in zebrafish liver, and the invasiveness and metastasis of these cells were directly observable and easily traceable [ ] . to investigate the mechanism of local cancer cell invasion, human glioblastoma cells labeled with fluorescence were infiltrated into the brain of zebrafish embryos. it was observed that the injected cells aligned along the abluminal surface of brain blood vessels [ ] . by grafting a small amount of highly metastatic human breast carcinoma cells onto the pericardial membrane of zebrafish embryos at hpf, tumor cells were observed to move longitudinally along the aorta [ ] . similarly, highly metastatic human cancer cells labeled with fluorescence were injected into the pericardium of -hpf embryos. afterwards, it is possible to visualize how cancer cells entered the blood circulation and arrested in small vessels in head and tail [ ] . in another example, zebrafish embryos were injected with tumorigenic human glioma stem cells at different stages of metastasis, including beginning, approaching, clustering, invading, migrating, and transmigrating [ ] . thus, grafting a small number of labeled tumor cells into transparent zebrafish embryos allows us to dynamically monitor the cancer cells without the interference of immune suppression. apart from its utility in analyzing the mechanisms of tumor dissemination and metastasis, the zebrafish model can also be applied to screen potential anticancer compounds or drugs. in addition, zebrafish feature such advantages as easy gene manipulation, short generation cycle, high reproducibility, low maintenance cost, and efficient plating of embryos [ , ] . therefore, this small fish is second only to scid and nude mice as xenograft recipients of cancer cells. leukemia is a cancer related to hematopoiesis. most often, leukemia results from the abnormal increase of white blood cells. however, some human cancers of bone marrow and blood origins have their parental cells from other blood cell types. the search for efficacious therapies for leukemia is ongoing. interestingly, the developmental processes and genes related to hematopoiesis are similar between zebrafish and humans, making zebrafish a feasible model for the study of leukemia. in addition, gene expression in zebrafish could be conveniently modified by several approaches, e.g., mo-induced gene knockdown, talens and crispr/cas gene knockout, and dna/ rna introduced overexpression [ , ] . in the study of yeh et al. [ ] , the zebrafish model was applied to screen for chemical modifiers of aml -eto, an oncogenic fusion protein prevalent in acute myeloid leukemia (aml). treatment of zebrafish with chemical modifiers of aml -eto resulted in hematopoietic dysregulation and elicited a malignant phenotype similar to human aml. cyclooxygenase- (cox ) is an enzyme causing inflammation and pain. nimesulide is an inhibitor of cox and an antagonist to aml -eto in hematopoietic differentiation. fms-like tyrosine kinase (flt ) is a class iii receptor tyrosine kinase which is normally expressed in human hematopoietic stem and progenitor cells (hspcs) [ ] . internal tandem duplication (itd), which may occur at either the juxtamembrane domain (jmd) or the tyrosine kinase domain (tkds) of flt , is observed in one-third of human aml. zebrafish flt shares an overall , , and % sequence identity with that of human, mouse, and rat, respectively. however, the jmd and the activation loops of tkd are highly conserved, implicating that the functions of flt signaling are evolutionally conserved. overexpression of human flt -itd in zebrafish embryos induces the ectopic expansion of flt -itd positive myeloid cells. if those embryos are treated with ac , a potent and relatively selective inhibitor of flt , flt -itd myeloid expansion is effectively ameliorated [ ] . in another example, isocitrate dehydrogenase (idh) and are involved in citric acid cycle in intermediary metabolism. idh mutations are found in approximately % of cytogenetically abnormal aml, suggesting a pathogenetic link in leukemia initiation [ , ] . injection of either human idh -r h or zebrafish idh -r h, a mutant corresponding to human idh -r h, resulted in increased -hydroxyglutarate, which in turn induced the expansion of primitive myelopoiesis [ ] . taken together, these reports suggest that the molecular pathways involved in leukemia are conserved between humans and zebrafish. based on the aforementioned experimental evidence, zebrafish can be an exceptional platform for mimicking human myelodysplastic syndromes and establishing an in vivo vertebrate model for drug screening. several liver tumor models have been reported by liverspecific expression of transgenic oncogenes such as kras, xmrk and myc. these transgenic lines of zebrafish usually generate liver tumors with various severity from hepatocellular adenoma (hca) to hepatocellular carcinoma (hcc) [ ] [ ] [ ] . these three transgenic liver cancer models have been used to identify differentially expressed genes through rna-sage sequencing. for example, researchers have searched genes either up-or downregulated among the three tumor models and analyzed the possible signaling pathways. then, correlation between zebrafish liver tumor signatures and the different stages of human hepatocarcinogenesis was determined [ ] . high tumor incidence and convenient chemical treatment make this inducible transgenic zebrafish a plausible platform for studying on liver tumor progression, regression, and anticancer drug screening. interestingly, zebrafish become a modern organism for studying on depressive disorders [ ] [ ] [ ] . because the physiological (neuroanatomical, neuroendocrine, neurochemical) and genetic characteristics of zebrafish are similar to mammals, zebrafish are ideal for high-throughput genetic and chemical genetic screening. furthermore, since behavioral test of zebrafish for cognitive, approach-avoidance, and social paradigms are available, the identification of depression-like indices in response to physiological, genetic, environmental, and/ or psychopharmacological alterations is feasible [ ] . actually, zebrafish display highly robust phenotypes of neurobehavioral disorders such as anxiety-like and approach-avoidance behaviors. furthermore, novel information of behavioral indices can be exposed, including geotaxis via top-bottom vertical movement [ ] . zebrafish behavior can also be monitored using automated behavioral tracking software, which enhances efficiency and reduces interrater variance [ ] . additionally, zebrafish offer a potential insight into the social aspects of depression [ ] and may be suitable for studying the cognitive deficits of depression [ ] and its putative etiological pathways [ ] . last but not least, zebrafish are highly sensitive to psychotropic drugs, such as antidepressants, anxiolytics, mood stabilizers, and antipsychotics [ ] [ ] [ ] , serving as an important tool for drug discovery. aromatic hydrocarbons, heavy metals and environmental estrogens are currently being used to test the impact of environmental pollutants on animals [ ] . these studies mainly focused on mortality and abnormality rates. however, the developing embryos may have already been damaged in a subtle way that would have precluded direct observation of morphology and detection of mortality. to overcome this drawback, transgenic fish can be used because they are designed to study (a) whether toxicants cause defective genes during embryogenesis; (b) whether pollutants affect the expression of tissue-specific gene; and (c) whether the impact of pollutants on embryonic development is dosage dependent. pollutants can be directly detected by simply observing the coloration change of cells before or after the pollutants can cause morphological damage. therefore, transgenic model fish are promising organisms for use as bioindicators to environmental toxicants and mutagens [ , ] . in addition, chen and lu reported that the environmental xenobiotics can be detected by a transgenic line of medaka carrying a gfp reporter driven by cytochrome p a promoter (cyp a-gfp) [ ] . furthermore, the environmental xenoestrogenic compounds can be specifically detected by a hybrid transgenic line derived from crossing between line cyp a-gfp and line vg-lux whose lux reporter activity is driven by a vitellogenin promoter [ ] . lee et al. reported another zebrafish transgenic line, termed huorfz [ ] , as it has been described in pervious section . at normal condition, the translation of the transferred huorf chop -gfp mrna in huorfz embryos is completely suppressed by an inhibitory uorf of human chop mrna (huorf chop ). however, when the huorfz embryos were under er stress, such as heat shock, cold shock, hypoxia, metals, alcohol, toxicants or drugs, the downstream gfp became apparent due to the blockage of huorf chop -mediated translation inhibition. therefore, huorfz embryos can be used to study the mechanism of translational inhibition. additionally, huorfz embryos can serve a living material to monitor the contamination of hazardous pollutants [ ] . besides the universal huorfz system, zebrafish could also be indicators for specific pollutants. for example, xu et al. reported a transgenic zebrafish tg (cyp a:gfp) which can serve as an in vivo assay for screening xenobiotic compounds, since cyp a is involved in the aryl hydrocarbon receptor pathway, and can be induced in the presence of dioxins/dioxin-like compounds and polycyclic aromatic hydrocarbons [ ] . additional advantages of zebrafish include the small size, abundant number, rapid development and transparent eggs. these features make this model fish more accessible for the studies of molecular toxicology. it is increasingly clear that the transgenic fish model is a powerful biomaterial for the studies of multiple disciplines, including molecular biology, developmental biology, neurobiology, cancer biology and regenerative medicine. it provides a simple, yet effective, in vivo approach to identify regulatory dna sequences, as well as determine gene function and molecular pathways. more importantly, an increasing number of papers have reported that (a) the defective phenotype of mutants of model fish can photocopy with known human disorders; and (b) drugs have similar effects on zebrafish and mammalian systems. therefore, the transgenic fish model offers a useful platform for high-throughput drug screening in biomedical sciences. additionally, it can serve as an environmental indicator for detecting pollutants in our daily lives. nevertheless, there are several limitations and caveats of this fish model. first, unlike mammals, fish lack the heart septation, lung, mammary gland, prostate gland and limbs, which make the fish model impossible for studies of these tissues and organs. additionally, fish are absent of placenta so that fish embryos are directly exposed to the environment (e.g., drugs or pollutants) without involving the placenta. second, fish are poikilothermic and usually maintained below °c, which may not be optimal for those mammalian agents adapted for °c in evolution. last, since the zebrafish genome is tetraploid, it is less straight forward to conduct loss-offunction studies for certain genes. the molecular biology of transgenic fish enhanced expression and stable transmission of transgenes flanked by inverted terminal repeats from adeno-associated virus in zebrafish identification of the tol transposase of the medaka fish oryzias latipes that catalyzes excision of a nonautonomous tol element in zebrafish danio rerio functional dissection of the tol transposable element identified the minimal cis-sequence and a highly repetitive sequence in the subterminal region essential for transposition a transposon-mediated gene trap approach identifies developmentally regulated genes in zebrafish heritable gene targeting in zebrafish using customized talens efficient genome editing in zebrafish using a crispr-cas system crispr/cas and talen-mediated knock-in approaches in zebrafish the aequorea victoria green fluorescent protein can be used as a reporter in live zebrafish embryos gata- expression pattern can be recapitulated in living transgenic zebrafish using gfp reporter gene high-frequency generation of transgenic zebrafish which reliably express gfp in whole muscles or the whole body by using promoters of zebrafish origin isolation of a zebrafish rod opsin promoter to generate a transgenic zebrafish line expressing enhanced green fluorescent protein in rod photoreceptors visualization of cranial motor neurons in live transgenic zebrafish expressing green fluorescent protein under the control of the islet- promoter/enhancer analysis of pancreatic development in living transgenic zebrafish embryos germ-line transmission of a myocardium-specific gfp transgene reveals critical regulatory elements in the cardiac myosin light chain promoter of zebrafish -bp liver regulatory sequence in the liver fatty acid binding protein (l-fabp) gene is sufficient to modulate liver regional expression in transgenic zebrafish establishment of a bone-specific col a : gfp transgenic zebrafish the macrophage-specific promoter mfap allows live, long-term analysis of macrophage behavior during mycobacterial infection in zebrafish expression of a vas:: egfp transgene in primordial germ cells of the zebrafish uniform gfp-expression in transgenic medaka (oryzias latipes) at the f generation random samples: that special glow genetics: fish that glow in taiwan coolest inventions : light and dark-red fish, blue fish and glow-in-dark fish modification of bacterial artificial chromosomes through chi-stimulated homologous recombination and its application in zebrafish transgenesis artificial chromosome transgenesis reveals longdistance negative regulation of rag in zebrafish multiple upstream modules regulate zebrafish myf expression use of the gal -uas technique for targeted gene expression in zebrafish udp-glucose dehydrogenase required for cardiac valve formation in zebrafish activation of notch signaling pathway precedes heart regeneration in zebrafish k-atpase is essential for embryonic heart development in the zebrafish the embryonic vertebrate heart tube is a dynamic suction pump four-dimensional cardiac imaging in living embryos via postacquisition synchronization of nongated slice sequences zebrafish cardiac development requires a conserved secondary heart field tbx is required for second heart field proliferation in zebrafish novel regulatory sequence − /- functions as a key element to drive the somite-specificity of zebrafish myf- foxd mediates zebrafish myf expression during early somitogenesis foxd mediates anterior-posterior polarity through upstream modulator fgf signaling during zebrafish somitogenesis inactivation of zebrafish mrf leads to myofibril misalignment and motor axon growth disorganization novel cis-element in intron represses somite expression of zebrafish myf- microrna- regulates fast muscle differentiation through modulating the target gene homer- b in zebrafish embryos novel intronic microrna represses zebrafish myf promoter activity through silencing dickkopf- gene zebrafish dkk a protein regulates the activity of myf promoter through interaction with membrane receptor integrin α b dickkopf- -related gene regulates the expression of zebrafish myf gene through phosphorylated p a-dependent smad activity myogenic regulatory factors myf and myod function distinctly during craniofacial myogenesis of zebrafish the transcription factor six a plays an essential role in the craniofacial myogenesis of zebrafish normal function of myf during gastrulation is required for pharyngeal arch cartilage development in zebrafish embryos differential requirements for myogenic regulatory factors distinguish medial and lateral somitic, cranial and fin muscle fibre populations retina-specific ciselements and binding nuclear proteins of carp rhodopsin gene egr gene knockdown affects embryonic ocular development in zebrafish the embryonic expression patterns and the knockdown phenotypes of zebrafish adp-ribosylation factor-like interacting protein gene arl ip plays a role in proliferation during zebrafish retinogenesis zebrafish arl ip is required for neural crest development during embryogenesis ras-related nuclear protein is required for late developmental stages of retinal cells in zebrafish eyes transgenic zebrafish model to study translational control mediated by upstream open reading frame of human chop gene a parsimonious model for gene regulation by mirnas gene silencing by micrornas: contributions of translational repression and mrna decay regulation of mrna translation and stability by micrornas micrornas: target recognition and regulatory functions microrna target predictions in animals labeled microrna pull-down assay system: an experimental approach for high-throughput identification of micrornatarget mrnas mir- and mir- regulate angiogenesis by modulating vegfa expression in zebrafish mir- and mir- target different genes to have opposing roles during angiogenesis in zebrafish embryos concise review: new frontiers in microrna-based tissue regeneration administration of microrna- promotes spinal cord regeneration in mice regulation of neonatal and adult mammalian heart regeneration by the mir- family integrated analyses of zebrafish mirna and mrna expression profiles identify mir- b and mir- as potential regulators of optic nerve regeneration transplantation and in vivo imaging of multilineage engraftment in zebrafish bloodless mutants hematopoiesis: an evolving paradigm for stem cell biology transcriptional regulation of hematopoietic stem cell development in zebrafish mutantspecific gene programs in the zebrafish prostaglandin e regulates vertebrate haematopoietic stem cell homeostasis prostaglandin e enhances human cord blood stem cell xenotransplants and shows long-term safety in preclinical nonhuman primate transplant models from zebrafish to human: modular medical models the heartstrings mutation in zebrafish causes heart/fin tbx deficiency syndrome the toxic effect of amiodarone on valve formation in the developing heart of zebrafish embryos amiodarone induces overexpression of similar to versican b to repress the egfr/gsk b/snail signaling axis during cardiac valve formation of zebrafish embryos cancer metastasis and egfr signaling is suppressed by amiodarone-induced versican v in vivo recording of adult zebrafish electrocardiogram and assessment of drug-induced qt prolongation zebrafish model for human long qt syndrome in-vitro recording of adult zebrafish heart electrocardiogram -a platform for pharmacological testing liver development and cancer formation in zebrafish zebrafish as a cancer model zebrafish modelling of leukaemias catch of the day: zebrafish as a human cancer model zebrafish as a cancer model system a keratin transgenic zebrafish tg(k ( . ):rfp) treated with inorganic arsenite reveals visible overproliferation of epithelial cells braf mutations are sufficient to promote nevi formation and cooperate with p in the genesis of melanoma the pathogenesis of cancer metastasis: the 'seed and soil' hypothesis revisited zebrafish xenotransplantation as a tool for in vivo cancer study metastatic behaviour of primary human tumours in a zebrafish xenotransplantation model calpain is required for the invasion of glioblastoma cells in the zebrafish brain microenvironment distinct contributions of angiogenesis and vascular co-option during the initiation of primary microtumors and micrometastases visualizing extravasation dynamics of metastatic tumor cells a novel zebrafish xenotransplantation model for study of glioma stem cell invasion quantitative phenotyping-based in vivo chemical screening in a zebrafish model of leukemia stem cell xenotransplantation zebrafish-based systems pharmacology of cancer metastasis discovering chemical modifiers of oncogene-regulated hematopoietic differentiation stk- , the human homolog of flk- /flt- , is selectively expressed in cd + human bone marrow cells and is involved in the proliferation of early progenitor/ stem cells functions of flt in zebrafish hematopoiesis and its relevance to human acute myeloid leukemia cancer-associated metabolite -hydroxyglutarate accumulates in acute myelogenous leukemia with isocitrate dehydrogenase and mutations regulation of cancer cell metabolism functions of idh and its mutation in the regulation of developmental hematopoiesis in zebrafish inducible and repressable oncogene-addicted hepatocellular carcinoma in tet-on xmrk transgenic zebrafish an inducible kras (v ) transgenic zebrafish model for liver tumorigenesis and chemical drug screening a transgenic zebrafish liver tumor model with inducible myc expression reveals conserved myc signatures with mammalian liver tumors xmrk, kras and myc transgenic zebrafish liver cancer models share molecular signatures with subsets of human hepatocellular carcinoma gaining translational momentum: more zebrafish models for neuroscience research zebrafish as an emerging model for studying complex brain disorders zebrafish models for translational neuroscience research: from tank to bedside zebrafish models of major depressive disorders threedimensional neurophenotyping of adult zebrafish behavior aquatic blues: modeling depression and antidepressant action in zebrafish social modulation of brain monoamine levels in zebrafish can zebrafish learn spatial tasks? an empirical analysis of place and single cs-us associative learning cognitive dysfunction in depression-pathophysiology and novel targets a larval zebrafish model of bipolar disorder as a screening platform for neuro-therapeutics role of serotonin in zebrafish (danio rerio) anxiety: relationship with serotonin levels and effect of buspirone, way , sb , fluoxetine and para-chlorophenylalanine (pcpa) in two behavioral models an affective disorder in zebrafish with mutation of the glucocorticoid receptor global water pollution and human health transgenic zebrafish for detecting mutations caused by compounds in aquatic environments mutational spectra of benzo [a] pyrene and meiqx in rpsl transgenic zebrafish embryos transgenic fish technology: basic principles and their application in basic and applied research gfp transgenic medaka (oryzias latipes) under the inducible cyp a promoter provide a sensitive and convenient biological indicator for the presence of tcdd and other persistent organic chemicals zebrafish transgenic line huorfz is an effective living bioindicator for detecting environmental toxicants generation of tg (cyp a:gfp) transgenic zebrafish for development of a convenient and sensitive in vivo assay for aryl hydrocarbon receptor activity the authors declare that they have no competing interests.authors' contributions hjt conceptualized, organized, charged and revised the content, and hjt, cyl and cyc wrote the manuscript together. all authors read and approved the final manuscript.• we accept pre-submission inquiries • our selector tool helps you to find the most relevant journal submit your next manuscript to biomed central and we will help you at every step: key: cord- - s lghj authors: buonomo, bruno title: effects of information-dependent vaccination behavior on coronavirus outbreak: insights from a siri model date: - - journal: nan doi: . /s - - - sha: doc_id: cord_uid: s lghj a mathematical model is proposed to assess the effects of a vaccine on the time evolution of a coronavirus outbreak. the model has the basic structure of siri compartments (susceptible–infectious–recovered–infectious) and is implemented by taking into account of the behavioral changes of individuals in response to the available information on the status of the disease in the community. we found that the cumulative incidence may be significantly reduced when the information coverage is high enough and/or the information delay is short, especially when the reinfection rate is high enough to sustain the presence of the disease in the community. this analysis is inspired by the ongoing outbreak of a respiratory illness caused by the novel coronavirus covid- . aviruses can infect people and then spreading from person-to-person. however, this may happen with serious consequences: well known cases are that of severe acute respiratory syndrome (sars) which killed people worldwide during - outbreak [ ] , and the more recent case of middle east respiratory syndrome coronavirus (mers), where a total of confirmed cases including associated deaths were reported, the majority from saudi arabia (at the end of november , [ ] ). therefore, coronavirus may represent a serious public health threat. the emergency related to the novel outbreak in china is still ongoing at time of writing this article and it is unclear how the situation worldwide will unfold. the news released by media create great concern and behavioral changes can be observed in the everyday life of individuals, even in europe where at the moment only few cases have been reported. for example, the fear of coronavirus has driven rapidly to sold out of protective face masks in pharmacies in italy long before the first case in the country was reported [ ] . a specific aspects of diseases caused by coronavirus is that humans can be reinfected with respiratory coronaviruses throughout life [ ] . the duration of immunity for sars, for example, was estimated to be greater than years [ ] . moreover, investigations on human coronavirus with infected volunteers has shown that even though the immune system react after the infection (serum-specific immunoglobulin and igc antibody levels peak - days after infection) at one year following experimental infection there is only partial protection against re-infection with the homologous strain [ ] . predictions or insight concerning the time-evolution of epidemics, especially when a new emerging infectious disease is under investigation, can be obtained by using mathematical models. in mathematical epidemiology, a large amount of literature is devoted to the use of the so called compartmental epidemic models, where the individuals of the community affected by the infectious disease are divided in mutually exclusive groups (the compartments) according to their status with respect to the disease [ , , , , ] . compartmental epidemic models are providing to be the first mathematical approach for estimating the epidemiological parameter values of covid- in its early stage and for anticipating future trends [ , , ] . when the disease under interest confer permanent immunity from reinfection after being recovered, the celebrated sir model (susceptible-infectious-recovered) and its many variants are most often adopted. however, where reinfection cannot be neglected the sirs model (susceptible-infectious-recovered, and again susceptible) and its variants may be used, under the assumption that infection does not change the host susceptibility [ , , , , ] . since the disease of our interest has both reinfection and partial immunity after infection, we consider as starting point the so-called siri model (susceptibleinfectious-recovered-infectious) which takes into account of both these features (see [ ] and the references contained therein for further information on siri model). when the epidemic process may be decoupled from the longer time-scale demographic dynamics, i. e. when birth and natural death terms may be neglected, one gets a simpler model with an interesting property. in fact, according to the values of three relevant parameters (the transmission rate, the recovery rate and the reinfection rate), the model exhibits three different dynamics [ , ] : (i) no epidemic will occur, in the sense that the fraction of infectious will decrease from the initial value to zero; (ii) an epidemic outbreak occurs, in the sense that the fraction of infectious will initially increase till a maximum value is reached and then it decreases to zero; (iii) an epidemic outbreak occurs and the disease will permanently remain within the population. at time of writing this paper, scholars are racing to make a vaccine for the novel covid- coronavirus available. as of february , , it was announced that 'the first vaccine could be ready in months' [ ] . therefore, it becomes an intriguing problem to qualitatively assess how the administration of a vaccine could affect the outbreak, taking into account of the behavioral changes of individuals in response to the information available on the status of the disease in the community. this is the main aim of this paper. the scenario depicted here is that of a community where a relatively small quantity of infectious is present at time of delivering the vaccine. the vaccination is assumed to be fully voluntary and the choice to get vaccinated or not is assumed to depend in part on the available information and rumors concerning the spread of the disease in the community. the behavioral change of individuals is introduced by employing the method of information-dependent models [ , , ] which is based on the introduction of a suitable information index. such an approach has been applied to general infectious diseases [ , , , , ] as well as specific ones, including childhood diseases like measles, mumps and rubella [ , ] and is currently under development (for very recent papers see [ , , ] ). therefore, another goal of this manuscript is to provide an application of the information index to a simple model containing relevant features of a coronavirus disease. specifically, we use epidemiological parameter values based on early estimation of novel coronavirus covid- [ ] . the rest of the paper is organized as follows: in sect. we introduce the basic siri model and recall its main properties. in sect. we implement the siri model by introducing the information-dependent vaccination. the epidemic and the reinfection thresholds are discussed in sect. . section is devoted to numerical investigations: the effects of the information parameters on the time evolution of the outbreak are discussed. conclusions and future perspective are given in sect. . since the disease of our interest has both reinfection and partial immunity after infection, we first consider the siri model, which is given by the following nonlinear ordinary differential equations (the upper dot denotes the time derivative) [ ] : here s, i and r denote, respectively, the fractions of susceptible, infectious (and also infected) and recovered individuals, at a time t (the dependence on t is omitted); β is the transmission rate; γ is the recovery rate; μ is the birth/death rate; σ ∈ ( , ) is the reduction in susceptibility due to previous infection. model ( ) assumes that the time-scale under consideration is such that demographic dynamics must be considered. however, epidemics caused by coronavirus often occurs quickly enough to neglect the demographic processes (as in the case of sars in [ ] [ ] . when the epidemic process is decoupled from demography, i.e. when μ = , one obviously gets the reduced model:Ṡ this very simple model has interesting properties. indeed, introduce the basic reproduction number r = β/γ . it has been shown that the solutions have the following behavior [ ] : if r ≤ , then no epidemic will occur, in the sense that the state variable i (t) denoting the fraction of infectious will decrease from the initial value to zero; if r ∈ ( , /σ ), then an epidemic outbreak will follow, in the sense that the state variable i (t) will initially increase till a maximum value is reached and then it decreases to zero; if r > /σ , then an epidemic outbreak will follow and the disease will permanently remain within the population, in the sense that the state variable i (t) will approach (after a possibly non monotone transient) an endemic equilibrium e, given by: where: the equilibrium e is globally asymptotically stable [ ] and it is interesting to note that, since the demography has been neglected, the disease will persist in the population due to the reservoir of partially susceptible individuals in the compartment r. from a mathematical point of view, the threshold r = r σ , where r σ = /σ , is a bifurcation value for model ( ) . this does not happen for model ( ) . in fact, when demography is included in the model, the endemic equilibrium exists for r > , where r = β/(μ + γ ) and therefore both below and above the reinfection threshold. model ( ) (as well as ( )) is a simple model which is able to describe the timeevolution of the epidemic spread on a short time-scale. however, it does not takes into account of possible control measure. the simplest one to consider is vaccination. we consider the scenario where the vaccination is assumed to be fully voluntary. in order to emphasize the role of reinfection, we assume that only susceptible individuals (i.e. individuals that did not experience the infection) consider this protective option. when the vaccine is perfect (i.e. it is an ideal vaccine which confer percent life-long immunity) one gets the following model: where v denotes the fraction of vaccinated individuals and ϕ is the vaccination rate. in the next section we will modify the siri model ( ) to assess how an hypothetical vaccine could control the outbreak, taking into account of the behavioral changes of individuals produced by the information available on the status of the disease in the community. we modify the siri model by employing the idea of the information-dependent epidemic models [ , ] . we assume that the vaccination is fully voluntary and information-dependent, in the sense that the choice to get vaccinated or not depends on the available information and rumors concerning the spread of the disease in the community. the information is mathematically represented by an information index m(t), which summarizes the information about the current and past values of the disease and is given by the following distributed delay [ ] [ ] [ ] ] : here, the functiong describes the information that individuals consider to be relevant for making their choice to vaccinate or not to vaccinate. it is often assumed thatg depends only on prevalence [ , , , ] where g is a continuous, differentiable, increasing function such that g( ) = . in particular, we assume that: in ( ) the parameter k is the information coverage and may be seen as a 'summary' of two opposite phenomena, the disease under-reporting and the level of media coverage of the status of the disease, which tends to amplify the social alarm. the range of variability of k may be restricted to the interval ( , ) (see [ ] ). the delay kernel k (t) in ( ) is a positive function such that +∞ k (t)dt = and represents the weight given to past history of the disease. we assume that the kernel is given by the first element erl ,a (t) of the erlangian family, called weak kernel or exponentially fading memory. this means that the maximum weight is assigned to the current information and the delay is centered at the average /a. therefore, the parameter a takes the meaning of inverse of the average time delay of the collected information on the disease. with this choice, by applying the linear chain trick [ ] , the dynamics of m is ruled by the equation: we couple this equation with model ( ). the coupling is realized through the following information-dependent vaccination rate: where the constant ϕ ∈ ( , ) represents the fraction of the population that chooses to get vaccinate regardless of rumors and information about the status of the disease in the population, and ϕ (m(t)) represents the fraction of the population whose vaccination choice is influenced by the information. generally speaking, we require that ϕ ( ) = and ϕ is a continuous, differentiable and increasing function. however, as done in [ , ] , we take: where ε > . this parametrization leads to an overall coverage of −ε (asymptotically for m → ∞). here we take ε = . , which means a roof of % in vaccine uptakes under circumstances of high perceived risk. we also take d = [ ] . note that this choice of parameter values implies that a . % vaccination coverage is obtained in correspondence of an information index m = . (see fig. ). finally we assume that the vaccine is not perfect, which is a more realistic hypothesis, so that the vaccinated individuals may be infected but with a reduced susceptibility ψ. the siri epidemic model with information-dependent vaccination that we consider is therefore given by the meaning of the state variables, the parameters and their baseline values are given in table . note that ( ) to get: let us introduce the quantity which is the basic reproduction number of model ( ) [ ] . from the second equation of ( ) it easily follows thatİ where: it immediately follows that, if i ( ) > , then: assuming that i ( ) > and r( ) = v ( ) = (and therefore s( ) < ) it follows that: if p < /s( ), then the epidemic curve initially decays. if p > /s( ) the epidemic takes place since the infectious curve initially grows. from the first equation in ( ) it can be seen that at equilibrium it must bes = . therefore, all the possible equilibria are susceptible-free. since the solutions are clearly bounded, this means that for large time any individual who was initially susceptible has experienced the disease or has been vaccinated. looking for equilibria in the form e = Ĩ ,r,Ṽ ,m , from ( ) we get: disease-free equilibria: ifĨ = . it can be easily seen from ( ) that therefore there are infinitely many disease-free equilibria of the form endemic equilibrium: we begin by looking for equilibria such that this implies that: therefore:Ṽ = andr it follows that an unique susceptibles-free endemic equilibrium exists, which is given by: where which exists only if the quantity is the reinfection threshold. when σ > σ c the disease may spread and persist inside the community where the individuals live. note that in classical sir models the presence of an endemic state is due to the replenishment of susceptibles ensured by demography [ ] , which is not the case here. the local stability analysis of e requires the jacobian matrix of system ( ): taking into account of ( ), ( ) and that v = , it follows the eigenvalues are: and the eigenvalues of the submatrix: the trace is negative and the determinant is so that e is locally asymptotically stable. (i) the stable endemic state e can be realized thanks to the imperfection of the vaccine, in the sense that when ψ = in ( ) the variable v is always increasing. (ii) the information index, in the form described in sect. , may be responsible of the onset of sustained oscillations in epidemic models both in the case of delayed information (see e.g. [ , , , ] ) and instantaneous information (as it happens when the latency time is included in the model [ ] ). in all these mentioned cases, the epidemic spread is considered on a long time-scale and demography is taken into account. the analysis in this section clearly shows that sustained oscillations are not possible for the short time-scale siri model with information. table we use epidemiological parameter values based on early estimation of novel coronavirus covid- provided in [ ] . the estimation, based on the use of a seir metapopulation model of infection within chinese cities, revealed that the transmission rate within the city of wuhan, the epicenter of the outbreak, was . day − , and the infectious period was . days (so that γ = . day − ). therefore the brn given in ( ) is p = . (of course, in agreement with the estimate in [ ] ), and the value σ c := /p = . is the threshold for the infection rate. for vaccinated individuals, the relative susceptibility (compared to an unvaccinated individuals) is set ψ = . , which means that vaccine administration reduces the transmission rate by % (vaccine efficacy = . ). this value falls within the estimates for the most common vaccine used in the usa, where vaccine efficacy ranges between . and . (see table . , p. , in [ ] ). as for the relative susceptibility of recovered individuals, we consider two relevant baseline cases: (i) case i: σ = . . this value is representative of a reinfection value below the reinfection threshold σ c ; (i) case ii: σ = . . this value is representative of a reinfection value above the reinfection threshold σ c . the information parameter values are mainly guessed or taken from papers where the information dependent vaccination is used [ , ] . the information coverage k ranges from a minimum of . (i.e. the public is aware of % of the prevalence) to . the average time delay of information ranges from the hypothetical case of immediate information (t = ) to a delay of days. the description and baseline values of the parameters are presented in table . the initial data reflect a scenario in which a small portion of infectious is present in the community at time of administrating the vaccine. furthermore, coherently with the initial data mentioned in sect. , we assume that: and, clearly, s( ) = − i ( ). according to the analysis made in sect. , values of σ below the threshold σ c implies that the epidemic will eventually die out. when σ is above σ c , then the disease is sustained endemically by reinfection. this behavior is illustrated in fig. , where it is considered the worst possible scenario, where k = . and t = days. in fig. , left panel, the continuous line is obtained for σ = . . vaccination is not able to influence the outbreak, due to the large delay. however, even though an epidemic peak occurs after three weeks, thereafter the disease dies out due to the low level of reinfection. the case σ = . is represented by the dotted line. as expected, the reinfection is able to 'restart' the epidemic. the trend (here captured for one year) would be to asymptotically converge to the endemic equilibrium e . the corresponding time evolution of the information index m is shown in fig. , right panel. in particular, in the elimination case (σ = . ), the information index reaches a maximum of . (approx.) which corresponds to a vaccination rate of . % (see fig. ). after that, it declines but, due to memory of past events, the information index is still positive months after the elimination of the disease. the 'social alarm' produced in the case σ = . is somehow represented by the increasing continuous curve in fig. , right panel. at the end of the time frame it is m ≈ . which corresponds to a vaccination rate of %. in summary, a large reinfection rate may produce a large epidemic. however, even in this worst scenario, the feedback produced by behavioral changes due to information may largely affect the outbreak evolution. in fig. table model ( ), is given by the quantity: more informed people react and vaccinate and this, in turn, contribute to the elimination of the disease. therefore, a threshold value k c exists above which the disease can be eliminated. an insight on the overall effect of parameter k on the epidemic may be determined by evaluating how it affects the cumulative incidence (ci), i.e. total number of new cases in the time frame [ , t f ]. we also introduce the following index which measures the relative change of cumulative incidence for two different values, say p and p , of a given parameter p over the simulated time frame (in other words, the percentage variation of the cumulative incidence varying p from p ). in fig. (first plot from the left) it is shown the case of a reinfection value σ = . , that is under the reinfection threshold. it can be seen how ci is declining with increasing k. in fig. (second plot from the left) a comparison with the case of low information coverage, k = . , is given: a reduction till % of ci may be reached by increasing the value of k till k = . . when the reinfection value is σ = . (fig. , third and fourth plot), that is above the reinfection threshold, the 'catastrofic case' is represented in correspondence of k = . . this case is quickly recovered by increasing k, as we already know from fig. , because of the threshold value k c , between . and . , which allows to pass from the endemic to no-endemic asymptotic state. then, again table ci is declining with increasing k. this means that when reinfection is high, the effect of information coverage is even more important. in fact, in this case the prevalence is high and a high value of k result in a greater behavioral response by the population. in fig. it is shown the influence of the information delay t on ci. in the case σ = . ci grows concavely with t (first plot from the left). in fig. (second plot) a comparison with the case of maximum information delay, t = days, is given: a reduction till % of ci may be reached by reducing the value of t till to very few days. when the reinfection value is σ = . (fig. , third and fourth plot), that is above the reinfection threshold, ci increases convexly with t . a stronger decreasing effect on ci can be seen by reducing the delay from t = days to t ≈ , and a reduction till % of ci may be reached by reducing the value of t till to very few days. in this paper we have investigated how a hypothetical vaccine could affect a coronavirus epidemic, taking into account of the behavioral changes of individuals in response to the information about the disease prevalence. we have first considered a basic siri model. such a model contains the specific feature of reinfection, which is typical of coronaviruses. reinfection may allow the disease to persist even when the time-scale of the outbreak is small enough to neglect the demography (births and natural death). then, we have implemented the siri model to take into account of: (i) an available vaccine to be administrated on voluntary basis to susceptibles; (ii) the change in the behavioral vaccination in response to information on the status of the disease. we have seen that the disease burden, expressed through the cumulative incidence, may be significantly reduced when the information coverage is high enough and/or the information delay is short. when the reinfection rate is above the critical value, a relevant role is played by recovered individuals. this compartment offers a reservoir of susceptibles (although with a reduced level of susceptibility) and if not vaccinate may contribute to the re-emergence of the disease. on the other hand, in this case a correct and quick information may play an even more important role since the social alarm produced by high level of prevalence results, in turn, in high level of vaccination rate and eventually in the reduction or elimination of the disease. the model on which this investigation is based is intriguing since partial immunity coupled to short-time epidemic behavior may lead to not trivial epidemic dynamics (see the 'delayed epidemic' case, where an epidemics initially may decrease to take off later [ ] ). however, it has many limitations in representing the covid- propagation. for example, the model represents the epidemics in a closed community over a relatively short time-interval and therefore it is unable to capture the complexity of global mobility, which is one of the main concerns related to covid- propagation. another limitation, which is again related to the global aspects of epidemics like sars and covid- , is that we assume that individuals are influenced by information on the status of the prevalence within the community where they live (i.e. the fraction i is part of the total population) whereas local communities may be strongly influenced also by information regarding far away communities, which are perceived as potential threats because of global mobility. moreover, in absence of treatment and vaccine, local authorities face with coronavirus outbreak using social distancing measures, that are not considered here: individuals are forced to be quarantined or hospitalized. nevertheless, contact pattern may be reduced also as response to information on the status of the disease. in this case the model could be modified to include an information-dependent contact rate, as in [ , ] . finally, the model does not include the latency time and the diseaseinduced mortality is also neglected (at the moment, the estimate for covid- is at around %). these aspects will be part of future investigations. mascherine sold out in farmacie roma data-based analysis, modelling and forecasting of the novel coronavirus ( -ncov) outbreak. medrxiv preprint infectious diseases of humans. dynamics and control mathematical epidemiology oscillations and hysteresis in an epidemic model with informationdependent imperfect vaccination global stability of an sir epidemic model with information dependent vaccination globally stable endemicity for infectious diseases with information-related changes in contact patterns modeling of pseudo-rational exemption to vaccination for seir diseases the time course of the immune response to experimental coronavirus infection of man mathematical structures of epidemic systems a new transmission route for the propagation of the sars-cov- coronavirus information-related changes in contact patterns may trigger oscillations in the endemic prevalence of infectious diseases bistable endemic states in a susceptible-infectious-susceptible model with behavior-dependent vaccination vaccinating behaviour, information, and the dynamics of sir vaccine preventable diseases. theor bifurcation thresholds in an sir model with informationdependent vaccination fatal sir diseases and rational exemption to vaccination the impact of vaccine side effects on the natural history of immunization programmes: an imitation-game approach infection, reinfection, and vaccination under suboptimal immune protection: epidemiological perspectives epidemiology of coronavirus respiratory infections epidemics with partial immunity to reinfection modeling infectious diseases in humans and animals nonlinear dynamics of infectious diseases via informationinduced vaccination and saturated treatment modeling the interplay between human behavior and the spread of infectious diseases an introduction to mathematical epidemiology bistability of evolutionary stable vaccination strategies in the reinfection siri model biological delay systems: linear stability theory the change of susceptibility following infection can induce failure to predict outbreak potential for r novel coronavirus -ncov: early estimation of epidemiological parameters and epidemic predictions world health organization: cumulative number of reported probable cases of sars middle east respiratory syndrome coronavirus world health organization: novel coronavirus world health organization: director-general's remarks at the media briefing on -ncov on statistical physics of vaccination duration of antibody responses after severe acute respiratory syndrome stability and bifurcation analysis on a delayed epidemic model with information-dependent vaccination publisher's note springer nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations conflict of interest the author states that there is no conflict of interest. key: cord- -cxua o t authors: wang, rui; jin, yongsheng; li, feng title: a review of microblogging marketing based on the complex network theory date: - - journal: international conference in electrics, communication and automatic control proceedings doi: . / - - - - _ sha: doc_id: cord_uid: cxua o t microblogging marketing which is based on the online social network with both small-world and scale-free properties can be explained by the complex network theory. through systematically looking back at the complex network theory in different development stages, this chapter reviews literature from the microblogging marketing angle, then, extracts the analytical method and operational guide of microblogging marketing, finds the differences between microblog and other social network, and points out what the complex network theory cannot explain. in short, it provides a theoretical basis to effectively analyze microblogging marketing by the complex network theory. as a newly emerging marketing model, microblogging marketing has drawn the domestic academic interests in the recent years, but the relevant papers are scattered and inconvenient for a deep research. on the microblog, every id can be seen as a node, and the connection between the different nodes can be seen as an edge. these nodes, edges, and relationships inside form the social network on microblog which belongs to a typical complex network category. therefore, reviewing the literature from the microblogging marketing angle by the complex network theory can provide a systematic idea to the microblogging marketing research. in short, it provides a theoretical basis to effectively analyze microblogging marketing by the complex network theory. the start of the complex network theory dates from the birth of small-world and scale-free network model. these two models provide the network analysis tools and information dissemination interpretation to the microblogging marketing. "six degrees of separation" found by stanley milgram and other empirical studies show that the real network has a network structure of high clustering coefficient and short average path length [ ] . watts and strogatz creatively built the smallworld network model with this network structure (short for ws model), reflecting human interpersonal circle focus on acquaintances to form the high clustering coefficient, but little exchange with strangers to form the short average path length [ ] . every id in microblog has strong ties with acquaintance and weak ties with strangers, which matches the ws model, but individuals can have a large numbers of weak ties in the internet so that the online microblog has diversity with the real network. barabàsi and albert built a model by growth mechanism and preferential connection mechanism to reflect that the real network has degree distribution following the exponential distribution and power-law. because power-law has no degree distribution of the characteristic scale, this model is called the scale-free network model (short for ba model) [ ] . exponential distribution exposes that most nodes have low degree and weak impact while a few nodes have high degree and strong impact, confirming "matthew effect" in sociology and satisfying the microblog structure that celebrities have much greater influence than grassroots, which the small-world model cannot describe. in brief, the complex network theory pioneered by the small-world and scalefree network model overcomes the constraints of the network size and structure of regular network and random network, describes the basic structural features of high clustering coefficient, short average path length, power-law degree distribution, and scale-free characteristics. the existing literature analyzing microblogging marketing by the complex network theory is less, which is worth further study. the complex network theory had been evoluted from the small-world scale-free model to some major models such as the epidemic model and game model. the diffusion behavior study on these evolutionary complex network models is valuable and can reveal the spread of microblogging marketing concept in depth. epidemic model divides the crowd into three basic types: susceptible (s), infected (i), and removed (r), and build models according to the relationship among different types during the disease spread in order to analyze the disease transmission rate, infection level, and infection threshold to control the disease. typical epidemic models are the sir model and the sis model. differences lie in that the infected (i) in the sir model becomes the removed (r) after recovery, so the sir model is used for immunizable diseases while the infected (i) in the sis model has no immunity and only becomes the susceptible (s) after recovery. therefore, the sis model is used for unimmunizable diseases. these two models developed other epidemic model: sir model changes to sirs model when the removed (r) has been the susceptible (s); sis model changes to si model presenting the disease outbreaks in a short time when the infected (i) is incurable. epidemic model can be widely seen in the complex network, such as the dissemination of computer virus [ ] , information [ ] , knowledge [ ] . guimerà et al. finds the hierarchical and community structure in the social network [ ] . due to the hierarchical structure, barthélemy et al. indicate that the disease outbreak followed hierarchical dissemination from the large-node degree group to the small-node degree group [ ] . due to the community structure, liu et al. indicate the community structure has a lower threshold and greater steady-state density of infection, and is in favor of the infection [ ] ; fu finds that the real interpersonal social network has a positive correlation of the node degree distribution, but the real interpersonal social network has negative [ ] . the former expresses circles can be formed in celebrities except grassroots, but the latter expresses contacts can be formed in celebrities and grassroots on the microblog. the game theory combined with the complex network theory can explain the interpersonal microlevel interaction such as tweet release, reply, and retweet because it can analyze the complex dynamic process between individuals such as the game learning model, dynamic evolutionary game model, local interaction model, etc.( ) game learning model: individuals make the best decision by learning from others in the network. learning is a critical point to decision-making and game behavior, and equilibrium is the long-term process of seeking the optimal results by irrational individuals [ ] . bala and goyal draw the "neighbor effect" showing the optimal decision-making process based on the historical information from individuals and neighbors [ ] . ( ) dynamic evolutionary game model: the formation of the social network seems to be a dynamic outcome due to the strategic choice behavior between edge-breaking and edge-connecting based on the individual evolutionary game [ ] . fu et al. add reputation to the dynamic evolutionary game model and find individuals are more inclined to cooperate with reputable individuals in order to form a stable reputation-based network [ ] . ( ) local interaction model: local network information dissemination model based on the strong interactivity in local community is more practical to community microblogging marketing. li et al. restrain preferential connection mechanism in a local world and propose the local world evolutionary network model [ ] . burke et al. construct a local interaction model and find individual behavior presents the coexistence of local consistency and global decentrality [ ] . generally speaking, microblog has characteristics of the small-world, scale-free, high clustering coefficient, short average path length, hierarchical structure, community structure, and node degree distribution of positive and negative correlation. on one hand, the epidemic model offers the viral marketing principles to microblogging marketing, such as the sirs model can be used for the long-term brand strategy and the si model can be used for the short-term promotional activity; on the other hand, the game model tells microblogging marketing how to find opinion leaders in different social circles to develop strategies for the specific community to realize neighbor effect and local learning to form global microblog coordination interaction. rationally making use of these characteristics can preset effective strategies and solutions for microblogging marketing. the complex network theory is applied to biological, technological, economic, management, social, and many other fields by domestic scholars. zhou hui proves the spread of sars rumors has a typical small-world network features [ ] . duan wenqi studies new products synergy diffusion in the internet economy by the complex network theory to promote innovation diffusion [ ] . wanyangsong ( ) analyzes the dynamic network of banking crisis spread and proposes the interbank network immunization and optimization strategy [ ] . although papers explaining microblogging marketing by the complex network theory have not been found, these studies have provided the heuristic method, such as the study about the online community. based on fu's study on xiao nei sns network [ ] , hu haibo et al. carry out a case study on ruo lin sns network and conclude that the online interpersonal social network not only has almost the same network characteristics as the real interpersonal social network, but also has a negative correlation of the node degree distribution while the real interpersonal social network has positive. this is because the online interpersonal social network is more easier for strangers to establish relationships so that that small influence people can reach the big influence people and make weak ties in plenty through breaking the limited range of real world [ ] . these studies can be used to effectively develop marketing strategies and control the scope and effectiveness of microblogging marketing. there will be a great potential to research on the emerging microblog network platform by the complex network theory. the complex network theory describes micro and macro models analyzing the marketing process to microblogging marketing. the complex network characteristics of the small-world, scale-free, high clustering coefficient, short average path length, hierarchical structure, community structure, node degree distribution of positive and negative correlation and its application in various industries provide theoretical and practical methods to conduct and implement microblogging marketing. the basic research idea is: extract the network topology of microblog by the complex network theory; then, analyze the marketing processes and dissemination mechanism by the epidemic model, game model, or other models while taking into account the impact of macro and micro factors; finally, find out measures for improving or limiting the marketing effect in order to promote the beneficial activities and control the impedimental activities for enterprizes' microblogging marketing. because the macro and micro complexity and uncertainty of online interpersonal social network, the previous static and dynamic marketing theory cannot give a reasonable explanation. based on the strong ties and weak ties that lie in individuals of the complex network, goldenberg et al. find: ( ) after the external short-term promotion activity, strong ties and weak ties turn into the main force driving product diffusion; ( ) strong ties have strong local impact and weak transmission ability, while weak ties have strong transmission ability and weak local impact [ ] . therefore, the strong local impact of strong ties and strong transmission ability of weak ties are required to be rationally used for microblogging marketing. through system simulation and data mining, the complex network theory can provide explanation framework and mathematical tools to microblogging marketing as an operational guide. microblogging marketing is based on online interpersonal social network, having difference with the nonpersonal social network and real interpersonal social network. therefore, the corresponding study results cannot be simply mixed if involved with human factors. pastor-satorras et al. propose the target immunization solution to give protection priority to larger degree node according to sis scale-free network model [ ] . this suggests the importance of cooperation with the large influential ids as opinion leaders in microblogging marketing. remarkably, the large influential ids are usually considered as large followers' ids on the microblog platform that can be seen from the microblog database. the trouble is, as scarce resources, the large influential ids have a higher cooperative cost, but the large followers' ids are not all large influential ids due to the online public relations behaviors such as follower purchasing and watering. this problem is more complicated than simply the epidemic model. the complex network theory can be applied in behavior dynamics, risk control, organizational behavior, financial markets, information management, etc.. microblogging marketing can learn the analytical method and operational guide from these applications, but the complex network theory cannot solve all the problems of microblogging marketing, mainly: . the complexity and diversity of microblogging marketing process cannot completely be explained by the complex network theory. unlike the natural life-like virus, individuals on microblog are bounded rational, therefore, the decisionmaking processes are impacted by not only the neighbor effect and external environment but also by individuals' own values, social experience, and other subjective factors. this creates a unique automatic filtering mechanism of microblogging information dissemination: information recipients reply and retweet the tweet or establish and cancel contact only dependent on their interests, leading to the complexity and diversity. therefore, interaction-worthy topics are needed in microblogging marketing, and the effective followers' number and not the total followers' number of id is valuable. this cannot be seen in disease infection. . there are differences in network characteristics between microblog network and the real interpersonal social network. on one hand, the interpersonal social network is different from the natural social network in six points: ( ) social network has smaller network diameter and average path length; ( ) social network has higher clustering coefficient than the same-scale er random network; ( ) the degree distribution of social network has scale-free feature and follows power-law; ( ) interpersonal social network has positive correlation of node degree distribution but natural social network has negative; ( ) local clustering coefficient of the given node has negative correlation of the node degree in social network; ( ) social network often has clear community structure [ ] . therefore, the results of the natural social network are not all fit for the interpersonal social network. on the other hand, as the online interpersonal social network, microblog has negative correlation of the node degree distribution which is opposite to the real interpersonal social network. this means the results of the real interpersonal social network are not all fit for microblogging marketing. . there is still a conversion process from information dissemination to sales achievement in microblogging marketing. information dissemination on microblog can be explained by the complex network models such as the epidemic model, but the conversion process from information dissemination to sales achievement cannot be simply explained by the complex network theory, due to not only individual's external environment and neighborhood effect, but also consumer's psychology and willingness, payment capacity and convenience, etc.. according to the operational experience, conversion rate, retention rates, residence time, marketing topic design, target group selection, staged operation program, and other factors are needed to be analyzed by other theories. above all, microblogging marketing which attracts the booming social attention cannot be analyzed by regular research theories. however, the complex network theory can provide the analytical method and operational guide to microblogging marketing. it is believed that microblogging marketing on the complex network theory has a good study potential and prospect from both theoretical and practical point of view. the small world problem collective dynamics of 'small-world' networks emergence of scaling in random networks how viruses spread among computers and people information exchange and the robustness of organizational networks network structure and the diffusion of knowledge team assembly mechanisms determine collaboration network structure and team performance romualdo pastor-satorras, alessandro vespignani: velocity and hierarchical spread of epidemic outbreaks in scale-free networks epidemic spreading in community networks social dilemmas in an online social network: the structure and evolution of cooperation the theory of learning in games learning from neighbors a strategic model of social and economic networks reputation-based partner choice promotes cooperation in social networks a local-world evolving network model the emergence of local norms in networks research of the small-world character during rumor's propagation study on coordinated diffusion of new products in internet market doctoral dissertation of shanghai jiaotong university structural analysis of large online social network talk of the network: a complex systems look at the underlying process of word-of-mouth immunization of complex networks meeting strangers and friends of friends: how random are socially generated networks key: cord- - v dcmt authors: papariello, luca; bampoulidis, alexandros; lupu, mihai title: on the replicability of combining word embeddings and retrieval models date: - - journal: advances in information retrieval doi: . / - - - - _ sha: doc_id: cord_uid: v dcmt we replicate recent experiments attempting to demonstrate an attractive hypothesis about the use of the fisher kernel framework and mixture models for aggregating word embeddings towards document representations and the use of these representations in document classification, clustering, and retrieval. specifically, the hypothesis was that the use of a mixture model of von mises-fisher (vmf) distributions instead of gaussian distributions would be beneficial because of the focus on cosine distances of both vmf and the vector space model traditionally used in information retrieval. previous experiments had validated this hypothesis. our replication was not able to validate it, despite a large parameter scan space. the last years have seen proof that neural network-based word embedding models provide term representations that are a useful information source for a variety of tasks in natural language processing. in information retrieval (ir), "traditional" models remain a high baseline to beat, particularly when considering efficiency in addition to effectiveness [ ] . combining the word embedding models with the traditional ir models is therefore very attractive and several papers have attempted to improve the baseline by adding in, in a more or less ad-hoc fashion, word-embedding information. onal et al. [ ] summarized the various developments of the last half-decade in the field of neural ir and group the methods in two categories: aggregate and learn. the first one, also known as compositional distributional semantics, starts from term representations and uses some function to combine them into a document representation (a simple example is a weighted sum). the second method uses the word embedding as a first layer of another neural network to output a document representation. the advantage of the first type of methods is that they often distill down to a linear combination (perhaps via a kernel), from which an explanation about the representation of the document is easier to induce than from the neural network layers built on top of a word embedding. recently, the issue of explainability in ir and recommendation is generating a renewed interest [ ] . in this sense, zhang et al. [ ] introduced a new model for combining highdimensional vectors, using a mixture model of von mises-fisher (vmf) instead of gaussian distributions previously suggested by clinchant and perronnin [ ] . this is an attractive hypothesis because the gaussian mixture model (gmm) works on euclidean distance, while the mixture of von mises-fisher (movmf) model works on cosine distances-the typical distance function in ir. in the following sections, we set up to replicate the experiments described by zhang et al. [ ] . they are grouped in three sets: classification, clustering, and information retrieval, and compare "standard" embedding methods with the novel movmf representation. in general, we follow the experimental setup of the original paper and, for lack of space, we do not repeat here many details, if they are clearly explained there. all experiments are conducted on publicly available datasets and are briefly described here below. classification. two subsets of the movie review dataset: (i) the subjectivity dataset (subj) [ ] ; and (ii) the sentence polarity dataset (sent) [ ] . clustering. the newsgroups dataset was used in the original paper, but the concrete version was not specified. we selected the "bydate" version, because it is, according to its creators, the most commonly used in the literature. it is also the version directly load-able in scikit-learn , making it therefore more likely that the authors had used this version. retrieval. the trec robust collection [ ] . the methods used to generate vectors for terms and documents are: tf-idf. the basic term frequency -inverse document frequency method [ ] . implemented in the scikit-learn library . [ ] . lda. latent dirichlet allocation [ ] . cbow. word vec [ ] in the continuous bag-of-word (cbow) architecture. pv-dbow/dm. paragraph vector (pv) is a document embedding algorithm that builds on word vec. we use here both its implementations: distributed bag-of-words (pv-dbow) and distributed memory (pv-dm) [ ] . the lsi, lda, cbow, and pv implementations are available in the gensim library . the fk framework offers the option to aggregate word embeddings to obtain fixed-length representations of documents. we use fisher vectors (fv) based on (i) a gaussian mixture model (fv-gmm) and (ii) a mixture of von mises-fisher distributions (fv-movmf) [ ] . we first fit (i) a gmm and (ii) a movmf model on previously learnt continuous word embeddings. the fixed-length representation of a document x containing t words w i -expressed as where k is the number of mixture components. the vectors g x i , having the dimension (d) of the word vectors e wi , are explicitly given by [ , ] : where ω i are the mixture weights, γ t (i) = p(i|x t ) is the soft assignment of x t to (i) gaussian and (ii) vmf distribution i, and σ i = diag(Σ i ), with Σ i the covariance matrix of gaussian i. in (i), σ i refers to the mean vector; in (ii) it indicates the mean direction and κ i is the concentration parameter. we implement the fk-based algorithms by ourselves, with the help of the scikit-learn library for fitting a mixture of gaussian models and of the spherecluster package for fitting a mixture of von mises-fisher distributions to our data. the implementation details of each algorithm are described in what follows. each of the following experiments is conceptually divided in three phases. first, text processing (e.g. tokenisation); second, creating a fixed-length vector representation for every document; finally, the third phase is determined by the goal to be achieved, i.e. classification, clustering, and retrieval. for the first phase the same pre-processing is applied to all datasets. in the original paper, this phase was only briefly described as tokenisation and stopword removal. it is not given what tokeniser, linguistic filters (stemming, lemmatisation, etc.), or stop word list were used. knowing that the gensim library was used, we took all standard parameters (see provided code ). gensim however does not come with a pre-defined stopword list, and therefore, based on our own experience, we used the one provided in the nltk library for english. for the second phase, transforming terms and documents to vectors, zhang et al. [ ] specify that all trained models are dimensional. we have additionally experimented with dimensionality (used by clinchant and perronnin [ ] for clustering) and , as we hypothesized that might be too low. the tf-idf model is dimensional (i.e. only the top terms based on their tf-idf value are used), while the fischer-kernel models are × d dimensional, where d = { , , }, as just explained. in what follows, d refers to the dimensionality of lsi, lda, cbow, and pv models. the cbow and pv models are trained using a default window size of , keeping both low and high-frequency terms, again following the setup of the original experiment. the lda model is trained using a chunk size of documents and for a number of iterations over the corpus ranging from to . for the fk methods, both fitting procedures (gmm and movmf) are independently initialised times and the best fitting model is kept. for the third phase, parameters are explained in the following sections. logistic regression is used for classification in zhang et al., and therefore also used here. the results of our experiments, for d = and -dimensional feature vectors, are summarised in table . for all the methods, we perform a parameter scan of the (inverse) regularisation strength of the logistic regression classifier, as shown in fig. (a) and (b) . additionally, the learning algorithms are trained for a different number of epochs and the resulting classification accuracy assessed, cf. fig. (c) and (d). figure (a) indicates that cbow, fv-gmm, fv-movmf, and the simple tf-idf, when properly tuned, exhibit a very similar accuracy on subj -the given confidence intervals do not indeed allow us to identify a single, best model. surprisingly, tf-idf outperforms all the others on the sent dataset ( fig. (b) ). increasing the dimensionality of the feature vectors, from d = to , has the effect of reducing the gap between tf-idf and the rest of the models on the sent dataset (see table ). for clustering experiments, the obtained feature vectors are passed to the kmeans algorithm. the results of our experiments, measured in terms of adjusted rand index (ari) and normalized mutual information (nmi), are summarised in table . we used both d = and -dimensional feature vectors. note that the evaluation of the clustering algorithms is based on the knowledge of the ground truth class assignments, available in the newsgroups dataset. as opposed to classification, clustering experiments show a generous imbalance in performance and firmly speak in favour of pv-dbow. interestingly, tf-idf, fv-gmm, and fv-movmf, all providing high-dimensional document representations, have a low clustering effectiveness. lsi and lda achieve low accuracy (see table ) and are omitted here for visibility. the left panels [(a) and (b)] show the effect of (inverse) regularisation of the logistic regression classifier on the accuracy, while the right panels [(c) and (d)] display the effect of training for the learning algorithms. the two symbols on the right axis in panels (a) and (b) indicate the best (fv-movmf) results reported in [ ] . for these experiments, we extracted from every document of the test collection all the raw text, and preprocessed it as described in the beginning of this section. the documents were indexed and retrieved for bm with the lucene . search engine. we experimented with three topic processing ways: ( ) title only, ( ) description only, and ( ) title and description. the third way produces the best results and closest to the ones reported by zhang et al. [ ] , and hence are the only ones reported here. an important aspect of bm is the fact that the variation of its parameters k and b could bring significant improvement in performance, as reported by lipani et al. [ ] . therefore, we performed a parameter scan for k ∈ [ , ] and b ∈ [ , ] with a . step size for both parameters. for every trec topic, the scores of the top documents retrieved from bm were normalised to [ , ] with the min-max normalisation method, and were used in calculating the scores of the documents for the combined models [ ] . the original results, those of our replication experiments with standard (k = . and b = . ) and best bm parameter values-measured in terms of mean average precision (map) and precision at (p@ )-are outlined in table . we replicated previously reported experiments that presented evidence that a new mixture model, based on von mises-fisher distributions, outperformed a series of other models in three tasks (classification, clustering, and retrievalwhen combined with standard retrieval models). since the source code was not released in the original paper, important implementation and formulation details were omitted, and the authors never replied to our request for information, a significant effort has been devoted to reverse engineer the experiments. in general, for none of the tasks were we able to confirm the conclusions of the previous experiments: we do not have enough evidence to conclude that fv-movmf outperforms the other methods. the situation is rather different when considering the effectiveness of these document representations for clustering purposes: we find indeed that the fv-movmf significantly underperforms, contradicting previous conclusions. in the case of retrieval, although zhang et al.'s proposed method (fv-movmf) indeed boosts bm , it does not outperform most of the other models it was compared to. clustering on the unit hypersphere using von mises-fisher distributions latent dirichlet allocation aggregating continuous word embeddings for information retrieval indexing by latent semantic analysis distributional structure. word let's measure run time! extending the ir replicability infrastructure to include performance aspects distributed representations of sentences and documents verboseness fission for bm document length normalization efficient estimation of word representations in vector space neural information retrieval: at the end of the early years a sentimental education: sentiment analysis using subjectivity summarization based on minimum cuts seeing stars: exploiting class relationships for sentiment categorization with respect to rating scales the trec robust retrieval track. sigir forum aggregating neural word embeddings for document representation ears : the nd international workshop on explainable recommendation and search authors are partially supported by the h safe-deed project (ga ). key: cord- -z a eoyo authors: brockmann, dirk title: human mobility, networks and disease dynamics on a global scale date: - - journal: diffusive spreading in nature, technology and society doi: . / - - - - _ sha: doc_id: cord_uid: z a eoyo disease dynamics is a complex phenomenon and in order to address these questions expertises from many disciplines need to be integrated. one method that has become particularly important during the past few years is the development of computational models and computer simulations that help addressing these questions. in the focus of this chapter are emergent infectious diseases that bear the potential of spreading across the globe, exemplifying how connectivity in a globalized world has changed the way human-mediated processes evolve in the st century. the examples of most successful predictions of disease dynamics given in the chapter illustrate that just feeding better and faster computers with more and more data may not necessarily help understanding the relevant phenomena. it might rather be much more useful to change the conventional way of looking at the patterns and to assume a correspondingly modified viewpoint—as most impressively shown with the examples given in this chapter. in early , news accumulated in major media outlets about a novel strain of influenza circulating in major cities in mexico [ ] . this novel h n strain was quickly termed "swine flu", in reference to its alleged origin in pig populations before jumping the species border to humans. very quickly public health institutions were alerted and saw the risk of this local influenza epidemic becoming a major public health problem globally. the concerns were serious because this influenza strain was of the h n subtype, the same virus family that caused one of the biggest pandemics in history, the spanish flu that killed up to million people in the beginning of the th century [ ] . the swine flu epidemic did indeed develop into a pandemic, spreading across the globe in matters of months. luckily, the strain turned out to be comparatively mild in terms of symptoms and as a health hazard. nevertheless, the concept of emergent infectious diseases, novel diseases that may have dramatic public health, societal and economic consequences reached a new level of public awareness. even hollywood picked up the topic in a number of blockbuster movies in the following years [ ] . only a few years later, mers hit the news, the middle east respiratory syndrome, a new type of virus that infected people in the middle east [ ] . mers was caused by a new species of corona virus of the same family of viruses that the sars virus belonged to. and finally, the ebola crisis in west african countries liberia, sierra leone and guinea that although it did not develop into a global crisis killed more than people in west africa [ ] . emergent infectious diseases have always been part of human societies, and also animal populations for that matter [ ] . humanity, however, underwent major changes along many dimensions during the last century. the world population has increased from approx. . billion in to . billion in [ ] . the majority of people now live in so-called mega-cities, large scale urban conglomerations of more than million inhabitants that live in high population densities [ ] often in close contact with animals, pigs and fowl in particular, especially in asia. these conditions amplify not only the transmission of novel pathogens from animal populations to human, high frequency human-to-human contacts yield a potential for rapid outbreaks of new pathogens. population density is only one side of the coin. in addition to increasing faceto-face contacts within populations we also witness a change of global connectivity [ ] . most large cities are connected by means of an intricate, multi-scale web of transportation links, see fig. . . on a global scale worldwide air-transportation dominates this connectivity. approx. , airports and , direct connections span the globe. more than three billion passengers travel on this network each year. every day the passengers that travel this network accumulate a total of more than billion kilometers, which is three times the radius of our solar system [ , ] . clearly this amount of global traffic shapes the way emergent infectious diseases can spread across the globe. one of the key challenges in epidemiology is preparing for eventual outbreaks and designing effective control measures. evidence based control measures, however, require a good understanding of the fundamental features and characteristics of spreading behavior that all emergent infectious diseases share. in this context this means addressing questions such as: if there is an outbreak at fig. . the global air-transportation network. each node represents one of approx. airports, each link one of approx. direct connections between airports. more than billion passengers travel on this network each year. all in all every day more than billion km are traversed on this network, three times the radius of our solar system location x when should one expect the first case at a distant location y? how many cases should one expect there? given a local outbreak, what is the risk that a case will be imported in some distant country. how does this risk change over time? also, emergent infectious diseases often spread in a covert fashion during the onset of an epidemic. only after a certain number of cases are reported, public health scientists, epidemiologist and other professionals are confronted with cases that are scattered across a map and it is difficult to determine the actual outbreak origin. therefore, a key question is also: where is the geographic epicenter of an ongoing epidemic? disease dynamics is a complex phenomenon and in order to address these questions expertises from many disciplines need to be integrated, such as epidemiolgy, spatial statistics, mobility and medical research in this context. one method that has become particularly important during the past few years is the development of computational models and computer simulations that help address these questions. these are often derived and developed using techniques from theoretical physics and more recently complex network science. modeling the dynamics of diseases using methods from mathematics and dynamical systems theory has a long history. in kermack and mckenrick [ ] introduced and analyzed the "suceptible-infected-recovered" (sir) model, a parsimoneous model for the description of a large class of infectious diseases that is also still in use today [ ] . the sir model considers a host population in which individuals can be susceptible (s), infectious (i) or recovered (r). susceptible individuals can aquire a disease and become infectious themselves and transmit the disease to other susceptible individuals. after an infectious period individuals recover, acquire immunity, and no longer infect others. the sir model is an abstract model that reduces a real world situation to the basic dynamic ingredients that are believed to shape the time course of a typical epidemic. structurally, the sir model treats individuals in a population in much the same way as chemicals that react in a wellmixed container. chemical reactions between reactants occur at rates that depend on what chemicals are involved. it is assumed that all individuals can be represented only by their infectious state and are otherwise identical. each pair of individuals has the same likelihood of interacting. schematically, the sir model is described by the following reactions where and are transmission and recovery rates per individual, respectively. the expected duration of being infected, the infectious period is given by t = − which can range from a few days to a few weeks for generic diseases. the ratio of rates r = ∕ is known as the basic reproduction ratio, i.e. the expected number of secondary infections caused by a single infected individual in a fully susceptible population. r is the most important epidemiological parameter because the value of r determines whether an infectious disease has the potential for causing an epidemic or not. when r > a small fraction of infected individuals in a susceptible population will cause an exponential growth of the number of infections. this epidemic rise will continue until the supply of susceptibles decreases to a level at which the epidemic can no longer be sustained. the increase in recovered and thus immune individuals dilutes the population and the epidemic dies out. mathematically, one can translate the reaction scheme ( . ) into a set of ordinary differential equations. say the population has n ≫ individuals. for a small time interval Δt and a chosen susceptible individual the probability of that individual interacting with an infected is proportional to the fraction i∕n of infected individuals. because we have s susceptibles the expected change of the number susceptibles due to infection is where the rate is the same as in ( . ) and the negative sign accounts for the fact that the number of susceptibles decreases. likewise the number of infected individuals is increased by the same amount Δi = +Δt × × s × i∕n. the number of infecteds can also decrease due to the second reaction in ( . ). because each infected can spontaneouly recover the expected change due to recovery is based on these assumptions eqs. ( . ) and ( . ) become a set of differential equations that describe the dynamics of the sir model in the limit Δt → : depending on the magnitude of n a model in which reactions occur randomly at rates and a stochastic system generally exhibits solutions that fluctuate around the solutions to the deterministic system of eq. ( . ) . both, the deterministic sir model and the more general particle kinetic stochastic model are designed to model disease dynamics in a single population, spatial dynamics or movement patterns of the host population are not accounted for. these systems are thus known as well-mixed systems in which the analogy is one of chemical reactants that are well-stirred in a chemical reaction container as mentioned above. when a spatial component is expected to be important in natural scenario, several methodological approaches exist to account for space. essentially the inclusion of a spatial component is required when the host is mobile and can transport the state of infection from one location to another. the combination of local proliferation of an infection and the disperal of infected host individuals then yields a spread along the spatial dimension [ , ] . one of the most basic ways of incorporating a spatial dimension and host dispersal is by assuming that all quantities in the sir model are also functions of a location , so the state of the system is defined by s( , t), j( , t) and r( , t). most frequently two spatial dimensions are considered. the simplest way of incorporating dispersal is by an ansatz following eq. ( . ) in chap. which assumes that individuals move diffusively in space which yields the reaction-diffusion dynamical system where e.g. in a two-dimensional system with = (x, y) the laplacian is ∇ = ∕ x + ∕ y and the parameter d is the diffusion coefficient. the reasoning behind this approach is that the net flux of individuals of one type from one location to a neighboring location is proportional to the gradient or the difference in concentration of that type of individuals between neighboring locations. the key feature of diffusive dispersal is that it is local, in a discretized version the laplacian permits movements only within a limited distance. in reaction diffusion systems of this type the combination of initial exponential growth (if r = ∕ > ) and diffusion (d > ) yields the emergence of an epidemic wavefront that progresses at a constant speed if initially the system is seeded with a small patch of infected individuals [ ] . the advantage of parsimoneous models like the one defined by eq. ( . ) is that properties of the emergent epidemic wavefront can be computed analytically, e.g. the speed of the wave in the above system is related to the basic reproduction number and diffusion coefficient by in which we recognize the relation of eq. ( . ). another class of models considers the reaction of eq. ( . ) to occur on two-dimensional (mostly square) lattices. in these models each lattice site is in one of the states s, i or r and reactions occur only with nearest neighbors on the lattice. these models account for stochasticity and spatial extent. given a state of the system, defined by the state of each lattice site, and a small time interval Δt, infected sites can transmit the disease to neighboring sites that are susceptible with a probability rate . infected sites also recover to the the system is identical to the system depicted in (a). however, in addition to the generic next neighbor transmission, with a small but significant probability a transmission to a distant site can occur. this probability also decreases with distance as an inverse power-law, e.g. where the exponent is in the range < < . because the rare but significant occurance of long-range transmissions, a more complex pattern emerges, the concentric nature observed in system a is gone. instead, a fractal, multiscale pattern emerges r state and become immune with probability Δt. figure . a illustrates the time course of the lattice-sir model. seeded with a localized patch of infected sites, the system exhibits an asymptotic concentric wave front that progresses at an overall constant speed if the ratio of transmission and recovery rate is sufficiently large. without the stochastic effects that yield the irregular interface at the infection front, this system exhibits similar properties to the reaction diffusion system of eq. ( . ). in both systems transmission of the disease in space is spatially restricted per unit time. the stochastic lattice model is particularly useful for investigating the impact of permitting long-distance transmissions. figure . b depicts temporal snapshots of a simulation that is identical to the system of fig. . a apart from a small but significant difference. in addition for infected sites to transmit the disease to neighboring susceptible lattice sites, every now and then (with a probability of %) they can also fig. . ) geographic distance to the initial outbreak location is no longer a good predictor of arrival time, unlike in systems with local or spatially limited host mobility infect randomly chosen lattice sites anywhere in the system. the propensity of infecting a lattice site at distance r decreases as an inverse power-law as explained in the caption to fig. . . the possibility of transmitting to distant locations yields new epidemic seeds far away that subsequently turn into new outbreak waves and that in turn seed second, third, etc. generation outbreaks, even if the overall rate at which long-distance transmission occur is very small. the consequence of this is that the spatially coherent, concerntric pattern observed in the reaction diffusion system is lost, and a complex spatially incoherent, fractal pattern emerges [ ] [ ] [ ] . practically, this implies that the distance from an initial outbreak location can no longer be used as a measure for estimating or computing the time that it takes for an epidemic to arrive at a certain location. also, given a snapshot of a spreading pattern, it is much more difficult to reconstruct the outbreak location from the geometry of the pattern alone, unlike in the concentric system where the outbreak location is typically near the center of mass of the pattern. a visual inspection of the air-transportation system depicted in fig. . is sufficiently convincing that the significant fraction of long-range connections in global mobility will not only increase the speed at which infectious diseases spread but, more importantly, also cause the patterns of spread to exhibit high spatial incoherence and complexity caused by the intricate connectivity of the air-transportation network. as a consequence we can no longer use geographic distance to an emergent epidemic epicenter as an indicator or measure of "how far away" that epicenter is and how long it will take to travel to a given location on the globe. this type of decorrelation is shown in fig. . for two examples: the sars epidemic and the influenza h n pandemic. on a spatial resolution of countries, the figure depicts scatter plots of the epidemic arrival time as a function of geodesic (shortest distance on the surface of the earth) distance from the initial outbreak location. as expected, the correlation between distance and arrival time is weak. given that models based on local or spatially limited mobility are inadequate, improved models must be developed that account for both, the strong heterogeneity in population density, e.g. that human populations accumulate in cities that vary substantially in size, and the connectivity structure between them that is provided by data on air traffic. in a sense one needs to establish a model that captures that the entire population is a so-called meta-population, a system of m = , … , m subpopulation, each of size n m and traffic between them, e.g. specifying a matrix f nm that quantifies the amount of host individuals that travel from population m to population n in a given unit of time [ , ] . for example n n could correspond to the size of city n and f nm the amount of passengers the travel by air from m to n. one of earliest and most employed models for disease dynamics using the meta-population approach is a generalization of eq. ( . ) in which each population's dynamics is governed by the ordinary sir model, e.g. ds n ∕dt = − s n i n ∕n n ( . ) di n ∕dt = s n i n ∕n n − i n dr n ∕dt = i n where the size n n = r n + i n + s n of population n is a parameter. in addition to this, the exchange of individuals between populations is modeled in such a way that hosts of each class move from location m to location n with a probability rate nm which yields which is a generic metapopulation sir model. in principle one is required to fix the infection-related parameters and and the population sizes n m as well as the mobility rates nm , i.e. the number of transitions from m to n per unit time. however, based on very plausible assumptions [ ] , the system can be simplified in such a way that all parameters can be gauged against data that is readily available, e.g. the actual passenger flux f nm (the amount of passengers that travel from m to n per day) that defines the air-transportation network, without having to specify the absolute population sizes n n . first the general rates nm have to fulfill the condition nm n m = mn n n if we assume that the n n remain constant. if we assume, additionally, that the total air traffic flowing out of a population n obeys where the dynamic variables are, again, fractions of the population in each class: s n = s n ∕n n , j n = i n ∕n n , and r n = r n ∕n n . in this system the new matrix p mn and the new rate parameter can be directly computed from the traffic matrix f nm and the total population involved n = ∑ m n m according to where f = ∑ n,m f mn is the total traffic in the network. the matrix p nm is therefore the fraction of passengers that are leaving node m with destination n. because passengers must arrive somewhere we have ∑ n p nm = . an important first question is concerning the different time scales, i.e. the parameters , and that appear in system ( . ) . the inverse − = t is the infectious period, that is the time individuals remain infectious. if we assume t ≈ - days and r = ∕ ≈ both rates are of the same order of magnitude. how about ? the total number of passengers f is approximately × per day. if we assume that n ≈ × people we find that it is instructive to consider the inverse t travel = − ≈ days. on average a typical person boards a plane every - years or so. keep in mind though that this is an average that accounts for both a small fraction of the population with a high frequency of flying and a large fraction that almost never boards a plane. the overall mobility rate is thus a few orders of magnitude smaller than those rates related to transmissions and recoveries. this has important consequences for being able to replace the full dynamic model by a simpler model discussed below. figure . depicts a numerical solution to the model defined by eq. ( . ) for a set of initial outbreak locations. at each location a small seed of infected individuals initializes the epidemic. global aspects of an epidemic can be assessed by the total fraction of infected individuals j g (t) = ∑ n c n j n (t) where c n is the relative size population n with respect to the entire population size n . as expected the time course of a global epidemic in terms of the epicurve and duration depends substantially on the initial outbreak location. a more important aspect is the spatiotemporal pattern generated by the model. figure . depicts temporal snapshots of simulations initialized in london and chicago, respectively. analogous to the qualitative patterns observed in fig. . b, we see that the presence of long-range connections in the worldwide air-transportation network yields incoherent spatial patterns much unlike the regular, concentric wavefronts observed in systems without long-range mobility. figure . shows that also the model epidemic depicts only a weak correlation between geographic distance to the outbreak location and arrival time. for a fixed geographic distance arrival times at different airports can vary substantially and thus the traditional geographic distance is useless as a predictor. the system defined by eq. ( . ) is one of the most parsimoneous models that accounts for strongly heterogeneous population distributions that are coupled by traffic flux between them and that can be gauged against actual population size distributions and traffic data. surprisingly, despite its structural simplicity this type of model has been quite successful in accounting for actual spatial spreads of past epi-and pandemics [ ] . based on early models of this type and aided by the exponential increase of computational power, very sophisticated models have been developed that account for factors that are ignored by the deterministic metapopulation sir model. in the most sophisticated approaches, e.g. gleam [ ] , the global epidemic and mobility computational tool, not only traffic by air but other means of transportation are considered, more complex infectious dynamics is considered and in hybrid dynamical systems stochastic effects caused by random reactions and mobility events are taken into account. household structure, available hospital beds, seasonality have been incorporated as well as disease specific features, all in order to make predictions more and more precise. the philosophy of this type of research line heavily relies on the increasing advancement of both computational power as well as more accurate and pervasive data often collected in natural experiments and webbased techniques [ ] [ ] [ ] [ ] [ ] . despite the success of these quantitative approaches, this strategy bears a number of problems some of which are fundamental. first, with increasing computational methods it has become possible to implement extremely complex dynamical systems with decreasing effort and also without substantial knowledge of the dynamical properties that often nonlinear dynamical systems can possess. implementing a lot of dynamical detail, it is difficult to identify which factors are essential for an observed phenomenon and which factors are marginal. because of the complexity that is often incorporated even at the beginning of the design of a sophisticated model in combination with the lack of data modelers often have to make assumptions about the numerical values of parameters that are required for running a computer simulation [ ] . generically many dozens of unknown parameters exist for which plausible and often not evidence-based values have to be assumed. because complex computational models, especially those that account for stochasticity, have to be run multiple times in order to make statistical assessments, systematic parameter scans are impossible even with the most sophisticated supercomputers. finally, all dynamical models, irrespective of their complexity, require two ingredients to be numerically integrated: ( ) fixed values for parameters and ( ) initial conditions. although some computational models have been quite successful in describing and reproducing the spreading behavior of past epidemics and in situations where disease specific parameters and outbreak locations have been assessed, they are difficult to apply in situations when novel pathogens emerge. in these situations, when computational models from a practical point of view are needed most, little is known about these parameters and running even the most sophisticated models "in the dark" is problematic. the same is true for fixing the right initial con-ditions. in many cases, an emergent infectious disease initially spreads unnoticed and the public becomes aware of a new event after numerous cases occur in clusters at different locations. reconstructing the correct initial condition often takes time, more time than is usually available for making accurate and valueable predictions that can be used by public health workers and policy makers to devise containment strategies. given the issues discussed above one can ask if alternative approaches exist that can inform about the spread without having to rely on the most sophisticated highly detailed computer models. in this context one may ask whether the complexity of the observed patterns that are solutions to models like the sir metapopulation model of eq. ( . ) are genuinely complex because of the underlying complexity of the mobility network that intricately spans the globe, or whether a simple pattern is really underlying the dynamics that is masked by this complexity and our traditional ways of using conventional maps for displaying dynamical features and our traditional ways of thinking in terms of geographic distances. in a recent approach brockmann and helbing [ ] developed the idea of replacing the traditional geographic distance by the notion of an effective distance derived from the topological structure of the global air-transportation network. in essence the idea is very simple: if two locations in the air-transportation network exchange a large number of passengers they should be effectively close because a larger number of passengers implies that the probability of an infectious disease to be transmitted from a to b is comparatively larger than if these two locations were coupled only by a small number of traveling passengers. effective distance should therefore decrease with traffic flux. what is the appropriate mathematical relation and a plausible ansatz to relate traffic flux to effective distance? to answer this question one can go back to the metapopulation sir model, i.e. eq. ( . ) . dispersal in this equation is governed by the flux fraction p nm . recall that this quantity is the fraction of all passengers that leave node m and arrive at node n. therefore p nm can be operationally defined as the probability of a randomly chosen passenger departing node m arriving at node n. if, in a thought experiment, we assume that the randomly selected person is infectious, p nm is proportional to the probability of transmitting a disease from airport m to airport n. we can now make the following ansatz for the effective distance: ( . ) where d ≥ is a non-negative constant to be specified later. this definition of effective distance implies that if all traffic from m arrives at n and thus p nm = the effective distance is d nm = d which is the smallest possible value. if, on the other hand p nm becomes very small, d nm becomes larger as required. the definition ( . ) applies to nodes m and n that are connected by a link in the network. what about pairs of nodes that are not directly connected but only by paths that require intermediate steps? given two arbitrary nodes, an origin m and a destination n, an infinite amount of paths (sequence of steps) exist that connect the two nodes. we can define the shortest effective route as the one for which the accumulation of effective distances along the legs is minimal. so for any path we sum the effective distance along the legs according to eq. ( . ) adding up to an effective distance d nm . this approach also explains the use of the logarithm in the definition of effective distance. adding effective distances along a route implies the multiplication of the probabilities p nm along the involved steps. therefore the shortest effective distance d nm is equivalent to the most probable path that connect origin and destination. the parameter d is a free parameter in the definition and quantifies the influence of the number of steps involved in a path. typically it is chosen to be either or depending on the application. one important property of effective distance is its asymetry. generally we have this may seem surprising at first sight, yet it is plausible. consider for example two airports a and b. let's assume a is a large hub that is strongly connected to many other airports in the network, including b. airport b, however, is only a small airport with only as a single connection leading to a. the effective distance b → a is much smaller (equal to d ) than the effective distance from the hub a to the small airport b. this accounts for the fact that if, again in a thought experiment, a randomly chosen passenger at airport b is most definitely going to a whereas a randomly chosen passenger at the hub a is arriving at b only with a small probability. given the definition of effective distance one can compute the shortest effective paths to every other node from a chosen and fixed reference location. each airport m thus has a set of shortest paths p m that connect m to all other airports. this set forms the shortest path tree t m of airport m. together with the effective distance matrix d nm the tree defines the perspective of node m. this is illustrated qualitatively in the fig. . that depicts a planar random triangular weighted network. one can now employ these principles and compute the shortest path trees and effective distances from the perspective of actual airports in the worldwide airtransportation network based on actual traffic data, i.e. the flux matrix f nm . figure . depicts the shortest path tree of one of the berlin airports (tegel, txl). the radial distance of all the other airports in the network is proportional to their effective distance from txl. one can see that large european hubs are effectively close to txl as expected. however, also large asian and american airports are effectively close to txl. for example the airports of chicago (ord), beijing (pek), miami (mia) and new york (jfk) are comparatively close to txl. we can also see that from the perspective of txl, germany's largest airport fra serves as a gateway to a considerable fraction of the rest of the world. because the shortest path tree also represents the most probable spreading routes one can use this method to identify airports that are particularly important in terms of distributing an infectious disease throughout the network. the shortest path trees are also those paths that correspond to the most probable paths of a random walker that starts at the reference location and terminates at the respective target node the use of effective distance and representing the air-transportation network from the perspective of chosen reference nodes and making use of the more plausible notion of distance that better reflects how strongly different locations are coupled in a networked system is helpful for "looking at" the world. yet, this representation is more than a mere intuitive and plausible spatial representation. what are the dynamic consequences of effective distance? the true advantage of effective distance is illustrated in fig. . . this figure depicts the identical computer-simulated hypothetical pandemic diseases as fig. . . unlike the latter, that is based on the traditional geographic representation, fig. . employs the effective distance and shortest path tree representation from the perspective of the outbreak location as discussed above. using this method, the spatially incoherent patterns in the traditional representation are transformed into concentric spreading patterns, similar to those expected for simple reaction diffusion systems. this shows that the complexity of observed spreading patterns is actually equivalent to simple spreading patterns that are just convoluted and masked by the underlying network's complexity. this has important consequences. because only the topological features of the network are used for computing the effective distance and no dynamic features are required, the concentricy of the emergent patterns are a generic feature and independent of dynamical properties of the underlying model. it also means that in effective distance, contagion processes spread at a constant speed, and just like in the simple reaction diffusion model one can much better predict the arrival time of an epidemic wavefront, knowing the speed and effective distance. for example if shortly after an epidemic outbreak the spreading commences and the initial spreading speed is assessed, one can forecast arrival times without having to run computationally expensive simulations. even if the spreading speed is unknown, the shortest path trees and effective distance from the perspective of airport tegel (txl) in berlin. txl is the central node. radial distance in the tree quantifies the effective distance to the reference node txl. as expected large european hubs like frankfurt (fra), munich (muc) and london heathrow (lhr) are effective close to txl. however, also hubs that are geographically distant such as chicago (ord) and beijing (pek) are effectively closer than smaller european airports. note also that the tree structure indicates that fra is a gateway to a large fraction of other airports as reflected by the size of the tree branch at fra. the illustration is a screenshot of an interactive effective distance tool available online [ ] effective distance which is independent of dynamics can inform about the sequence of arrival times, or relative arrival times. the benefit of the effective distance approach can also be seen in fig. . in which arrival times of the sars epidemic and the h n pandemic in affected countries are shown as a function of effective distance to the outbreak origin. comparing this figure to fig. . we see that effective distance is a much better predictor of arrival time, a clear linear relationship exists between effective distance ord lhr fig. . simulations and effective distance. the panels depict the same temporal snapshots of computer simulated hypothetical pandemic scenarios as in fig. . . the top row corresponds to a pandemic initially seeded at lhr (london) the bottom row at ord (chicago). the networks depict the shortest path tree effective distance representation of the corresponding seed airports as in fig. . . the simulated pandemics that exhibit spatially incoherent complex patterns in the traditional representation (fig. . ) are equivalent to concentric wave fronts that progress at constant speeds in effective distance space. this method thus substantially simplifies the complexity seen in conventional approaches and improves quantitative predictions and epidemic arrival. thus, effective distance is a promising tool and concept for application in realistic scenarios, being able to provide a first quantitative assessment of an epidemic outbreak and its potential consequences on a global scale. in a number of situation epidemiologists are confronted with the task of reconstructing the outbreak origin of an epidemic. when a novel pathogen emerges in some cases the infection spreads covertly until a substantial case count attracts attention and public health officials and experts become aware of the situation. quite often cases occur much like the patterns depicted in fig. . b in a spatially incoherent way because of the complexity of underlying human mobility networks. when cases emerge at apparently randomly distributed locations it is a difficult task to assess where the event initially started. the computational method based on effective distance can also be employed in these situations provided that one knows the underlying mobility network. this is because the concentric pattern depicted in fig. . is . compared to the conventional use of geographic distance effective distance is a much better predictor of epidemic arrival time as is reflected by the linear relationship between arrival time and effective distance, e.g. compare to fig. . . right: the same analysis for the sars epidemic. also in this case effective distance is much more strongly correlated with arrival time than geographic distance only observed if and only if the actual outbreak location is chosen as the center perspective node. in other words, if the temporal snapshots are depicted using a different reference node the concentric pattern is scrambled and irregular. therefore, one can use the effective distance method to identify the outbreak location of a spreading process based on a single temporal snapshot. this method is illustrated in a proofof-concept example depicted in fig. . . assume that we are given a temporal snapshot of a spreading process as depicted in fig. . a and the goal is to reconstruct the outbreak origin from the data. conventional geometric considerations are not sucessful because the network-driven processes generically do not yields simple geometric patterns. using effective distance, we can now investigate the pattern from the perspective of every single potential outbreak location. we could for example pick a set of candidate outbreak locations (panel (b) in the figure). if this is done we will find that only for one candidate outbreak location the temporal snapshot has the shape of a concentric circle. this must be the original outbreak location. this process, qualitatively depicted in the figure, can be applied in a quantitative way and has been applied to actual epidemic data such as the ehec outbreak in germany [ ] . outbreak reconstruction using effective distance. a the panel depicts a temporal snapshot of a computer simulated hypothetical pandemic, red dots denote airports with a high prevalence of cases. from the snapshot alone it is difficult to assess the outbreak origin which in this case is ord (chicago). b a choice of potential outbreak locations as candidates. c for these candidate locations the pattern is depicted in the effective distance perspective. only for the correct outbreak location the pattern is concentric. this method can be used quantitatively to identify outbreaks of epidemics that initially spread in a covert way emergent infectious diseases that bear the potential of spreading across the globe are an illustrative example of how connectivity in a globalized world has changed the way human mediated processes evolve in the st century. we are connected by complex networks of interaction, mobility being only one of them. with the onset of social media, the internet and mobile devices we share information that proliferates and spreads on information networks in much the same way (see also chap. ) . in all of these systems the scientific challenge is understanding what topological and statistical features of the underlying network shape particular dynamic features observed in natural systems. the examples addressed above focus on a particular scale, defined by a single mobility network, the air-transportation network that is relevant for this scale. as more and more data accumulates, computational models developed in the future will be able to integrate mobility patterns at an individual resolution, potentially making use of pervasive data collected on mobile devices and paving the way towards predictive models that can account very accurately for observed contagion patterns. the examples above also illustrate that just feeding better and faster computers with more and more data may not necessarily help understanding the fundamental processes and properties of the processes that underly a specific dynamic phenomenon. sometimes we only need to change the conventional and traditional ways of looking at patterns and adapt our viewpoint appropriately. note , examples are contagion, a surprisingly accurate depiction of the consequences of a severe pandemic, and rise of the planet of the apes that concludes with ficticious explanation for the extinction of mankind due to a man made virus in the future the fates of human societies proc. r. soc. lond infectious diseases of humans: dynamics and control modeling infectious diseases in humans and animals human mobility and spatial disease dynamics proc. natl. acad. sci. usa key: cord- -yazo lga authors: brauer, fred title: compartmental models in epidemiology date: journal: mathematical epidemiology doi: . / - - - - _ sha: doc_id: cord_uid: yazo lga we describe and analyze compartmental models for disease transmission. we begin with models for epidemics, showing how to calculate the basic reproduction number and the final size of the epidemic. we also study models with multiple compartments, including treatment or isolation of infectives. we then consider models including births and deaths in which there may be an endemic equilibrium and study the asymptotic stability of equilibria. we conclude by studying age of infection models which give a unifying framework for more complicated compartmental models. we will be concerned both with epidemics which are sudden outbreaks of a disease, and endemic situations, in which a disease is always present. epidemics such as the outbreak of sars, the ebola virus and avian flu outbreaks are events of concern and interest to many people. the spanish flu epidemic caused millions of deaths, and a recurrence of a major influenza epidemic is a dangerous possibility. an introduction of smallpox is of considerable concern to government officials dealing with terrorism threats. an endemic situation is one in which a disease is always present. the prevalence and effects of many diseases in less developed countries are probably less well-known but may be of even more importance. every year millions of people die of measles, respiratory infections, diarrhea and other diseases that are easily treated and not considered dangerous in the western world. diseases such as malaria, typhus, cholera, schistosomiasis, and sleeping sickness are endemic in many parts of the world. the effects of high disease mortality on mean life span and of disease debilitation and mortality on the economy in afflicted countries are considerable. our goal is to provide an introduction to mathematical epidemiology, including the development of mathematical models for the spread of disease as well as tools for their analysis. scientific experiments usually are designed to obtain information and to test hypotheses. experiments in epidemiology with controls are often difficult or impossible to design and even if it is possible to arrange an experiment there are serious ethical questions involved in withholding treatment from a control group. sometimes data may be obtained after the fact from reports of epidemics or of endemic disease levels, but the data may be incomplete or inaccurate. in addition, data may contain enough irregularities to raise serious questions of interpretation, such as whether there is evidence of chaotic behaviour [ ] . hence, parameter estimation and model fitting are very difficult. these issues raise the question of whether mathematical modeling in epidemiology is of value. our response is that mathematical modeling in epidemiology provides understanding of the underlying mechanisms that influence the spread of disease and, in the process, it suggests control strategies. in fact, models often identify behaviours that are unclear in experimental data -often because data are non-reproducible and the number of data points is limited and subject to errors in measurement. for example, one of the fundamental results in mathematical epidemiology is that most mathematical epidemic models, including those that include a high degree of heterogeneity, usually exhibit "threshold" behaviour. in epidemiological terms this can be stated as follows: if the average number of secondary infections caused by an average infective, called the basic reproduction number, is less than one a disease will die out, while if it exceeds one there will be an epidemic. this broad principle, consistent with observations and quantified via epidemiological models, has been consistently used to estimate the effectiveness of vaccination policies and the likelihood that a disease may be eliminated or eradicated. hence, even if it is not possible to verify hypotheses accurately, agreement with hypotheses of a qualitative nature is often valuable. expressions for the basic reproduction number for hiv in various populations have been used to test the possible effectiveness of vaccines that may provide temporary protection by reducing either hiv-infectiousness or susceptibility to hiv. models are used to estimate how widespread a vaccination plan must be to prevent or reduce the spread of hiv. in the mathematical modeling of disease transmission, as in most other areas of mathematical modeling, there is always a trade-off between simple models, which omit most details and are designed only to highlight general qualitative behaviour, and detailed models, usually designed for specific situations including short-term quantitative predictions. detailed models are generally difficult or impossible to solve analytically and hence their usefulness for theoretical purposes is limited, although their strategic value may be high. in these notes we describe simple models in order to establish broad principles. furthermore, these simple models have additional value as they are the building blocks of models that include more detailed structure. many of the early developments in the mathematical modeling of communicable diseases are due to public health physicians. the first known result in mathematical epidemiology is a defense of the practice of inoculation against smallpox in by daniel bernoulli, a member of a famous family of mathematicians (eight spread over three generations) who had been trained as a physician. the first contributions to modern mathematical epidemiology are due to p.d. en'ko between and [ ] , and the foundations of the entire approach to epidemiology based on compartmental models were laid by public health physicians such as sir ross, r.a., w.h. hamer, a.g. mckendrick and w.o. kermack between and , along with important contributions from a statistical perspective by j. brownlee. a particularly instructive example is the work of ross on malaria. dr. ross was awarded the second nobel prize in medicine for his demonstration of the dynamics of the transmission of malaria between mosquitoes and humans. although his work received immediate acceptance in the medical community, his conclusion that malaria could be controlled by controlling mosquitoes was dismissed on the grounds that it would be impossible to rid a region of mosquitoes completely and that in any case mosquitoes would soon reinvade the region. after ross formulated a mathematical model that predicted that malaria outbreaks could be avoided if the mosquito population could be reduced below a critical threshold level, field trials supported his conclusions and led to sometimes brilliant successes in malaria control. however, the garki project provides a dramatic counterexample. this project worked to eradicate malaria from a region temporarily. however, people who have recovered from an attack of malaria have a temporary immunity against reinfection. thus elimination of malaria from a region leaves the inhabitants of this region without immunity when the campaign ends, and the result can be a serious outbreak of malaria. we will begin with an introduction to epidemic models. next, we will incorporate demographic effects into the models to explore endemic states, and finally we will describe models with infectivity depending on the age of infection. our approach will be qualitative. by this we mean that rather than attempting to find explicit solutions of the systems of differential equations which will form our models we will be concerned with the asymptotic behaviour, that is, the behaviour as t → ∞ of solutions. this material is meant to be an introduction to the study of compartmental models in mathematical epidemiology. more advanced material may be found in many other sources, including chaps. - of this volume, the case studies in chaps. - , and [ , - , , , , ]. an epidemic, which acts on a short temporal scale, may be described as a sudden outbreak of a disease that infects a substantial portion of the population in a region before it disappears. epidemics usually leave many members untouched. often these attacks recur with intervals of several years between outbreaks, possibly diminishing in severity as populations develop some immunity. this is an important aspect of the connection between epidemics and disease evolution. the book of exodus describes the plagues that moses brought down upon egypt, and there are several other biblical descriptions of epidemic outbreaks. descriptions of epidemics in ancient and medieval times frequently used the term "plague" because of a general belief that epidemics represented divine retribution for sinful living. more recently some have described aids as punishment for sinful activities. such views have often hampered or delayed attempts to control this modern epidemic . there are many biblical references to diseases as historical influences, such as the decision of sennacherib, the king of assyria, to abandon his attempt to capture jerusalem about bc because of the illness of his soldiers (isaiah , [ ] [ ] [ ] . the fall of empires has been attributed directly or indirectly to epidemic diseases. in the second century ad the so-called antonine plagues (possibly measles and smallpox) invaded the roman empire, causing drastic population reductions and economic hardships. these led to disintegration of the empire because of disorganization, which facilitated invasions of barbarians. the han empire in china collapsed in the third century ad after a very similar sequence of events. the defeat of a population of millions of aztecs by cortez and his followers can be explained in part by a smallpox epidemic that devastated the aztecs but had almost no effect on the invading spaniards thanks to their built-in immunities. the aztecs were not only weakened by disease but also confounded by what they interpreted as a divine force favoring the invaders. smallpox then spread southward to the incas in peru and was an important factor in the success of pizarro's invasion a few years later. smallpox was followed by other diseases such as measles and diphtheria imported from europe to north america. in some regions, the indigenous populations were reduced to one tenth of their previous levels by these diseases. between and the indian population of mexico was reduced from million to million. the black death spread from asia throughout europe in several waves during the fourteenth century, beginning in , and is estimated to have caused the death of as much as one third of the population of europe between and . the disease recurred regularly in various parts of europe for more than years, notably as the great plague of london of - . it then gradually withdrew from europe. as the plague struck some regions harshly while avoiding others, it had a profound effect on political and economic developments in medieval times. in the last bubonic plague epidemic in france ( - ), half the population of marseilles, % of the population in nearby toulon, % of the population of arles and % of the population of aix and avignon died, but the epidemic did not spread beyond provence. the historian w.h. mcneill argues, especially in his book [ ] , that the spread of communicable diseases has frequently been an important influence in history. for example, there was a sharp population increase throughout the world in the eighteenth century; the population of china increased from million in to million in and the population of europe increased from million in to million in . there were many factors involved in this increase, including changes in marriage age and technological improvements leading to increased food supplies, but these factors are not sufficient to explain the increase. demographic studies indicate that a satisfactory explanation requires recognition of a decrease in the mortality caused by periodic epidemic infections. this decrease came about partly through improvements in medicine, but a more important influence was probably the fact that more people developed immunities against infection as increased travel intensified the circulation and co-circulation of diseases. perhaps the first epidemic to be examined from a modeling point of view was the great plague in london ( - ). the plague was one of a sequence of attacks beginning in the year of what came to be known as the black death. it is now identified as the bubonic plague, which had actually invaded europe as early as the sixth century during the reign of the emperor justinian of the roman empire and continued for more than three centuries after the black death. the great plague killed about one sixth of the population of london. one of the few "benefits" of the plague was that it caused cambridge university to be closed for two years. isaac newton, who was a student at cambridge at the time, was sent to his home and while "in exile" he had one of the most productive scientific periods of any human in history. he discovered his law of gravitation, among other things, during this period. the characteristic features of the great plague were that it appeared quite suddenly, grew in intensity, and then disappeared, leaving part of the population untouched. the same features have been observed in many other epidemics, both of fatal diseases and of diseases whose victims recovered with immunity against reinfection. in the nineteenth century recurrent invasions of cholera killed millions in india. the influenza epidemic of - killed more than million people overall, more than half a million in the united states. one of the questions that first attracted the attention of scientists interested in the study of the spread of communicable diseases was why diseases would suddenly develop in a community and then disappear just as suddenly without infecting the entire community. one of the early triumphs of mathematical epidemiology [ ] was the formulation of a simple model that predicted behaviour very similar to the behaviour observed in countless epidemics. the kermack-mckendrick model is a compartmental model based on relatively simple assumptions on the rates of flow between different classes of members of the population. there are many questions of interest to public health physicians confronted with a possible epidemic. for example, how severe will an epidemic be? this question may be interpreted in a variety of ways. for example, how many individuals will be affected altogether and thus require treatment? what is the maximum number of people needing care at any particular time? how long will the epidemic last? how much good would quarantine or isolation of victims do in reducing the severity of the epidemic? these are some of the questions we would like to study with the aid of models. we formulate our descriptions as compartmental models, with the population under study being divided into compartments and with assumptions about the nature and time rate of transfer from one compartment to another. diseases that confer immunity have a different compartmental structure from diseases without immunity. we will use the terminology sir to describe a disease which confers immunity against re-infection, to indicate that the passage of individuals is from the susceptible class s to the infective class i to the removed class r. on the other hand, we will use the terminology sis to describe a disease with no immunity against re-infection, to indicate that the passage of individuals is from the susceptible class to the infective class and then back to the susceptible class. other possibilities include seir and seis models, with an exposed period between being infected and becoming infective, and sirs models, with temporary immunity on recovery from infection. the independent variable in our compartmental models is the time t and the rates of transfer between compartments are expressed mathematically as derivatives with respect to time of the sizes of the compartments, and as a result our models are formulated initially as differential equations. possible generalizations, which we shall not explore in these notes, include models in which the rates of transfer depend on the sizes of compartments over the past as well as at the instant of transfer, leading to more general types of functional equations, such as differential-difference equations, integral equations, or integro-differential equations. in order to model such an epidemic we divide the population being studied into three classes labeled s, i, and r. we let s(t) denote the number of individuals who are susceptible to the disease, that is, who are not (yet) infected at time t. i(t) denotes the number of infected individuals, assumed infectious and able to spread the disease by contact with susceptibles. r(t) denotes the number of individuals who have been infected and then removed from the possibility of being infected again or of spreading infection. removal is carried out either through isolation from the rest of the population or through immunization against infection or through recovery from the disease with full immunity against reinfection or through death caused by the disease. these characterizations of removed members are different from an epidemiological perspective but are often equivalent from a modeling point of view which takes into account only the state of an individual with respect to the disease. in formulating models in terms of the derivatives of the sizes of each compartment we are assuming that the number of members in a compartment is a differentiable function of time. this may be a reasonable approximation if there are many members in a compartment, but it is certainly suspect otherwise. in formulating models as differential equations, we are assuming that the epidemic process is deterministic, that is, that the behaviour of a population is determined completely by its history and by the rules which describe the model. in other chapters of this volume linda allen and ping yan describe the study of stochastic models in which probabilistic concepts are used and in which there is a distribution of possible behaviours. the developing study of network science, introduced in chap. of this volume and described in [ , , ] , is another approach. the basic compartmental models to describe the transmission of communicable diseases are contained in a sequence of three papers by w.o. kermack and a.g. mckendrick in mckendrick in , mckendrick in , and . the first of these papers described epidemic models. what is often called the kermack-mckendrick epidemic model is actually a special case of the general model introduced in this paper. the general model included dependence on age of infection, that is, the time since becoming infected. curiously, kermack and mckendrick did not explore this situation further in their later models which included demographic effects. age of infection models have become important in the study of hiv/aids, and we will return to them in the last section of this chapter. the special case of the model proposed by kermack and mckendrick in which is the starting point for our study of epidemic models is a flow chart is shown in fig. ( ) an average member of the population makes contact sufficient to transmit infection with βn others per unit time, where n represents total population size (mass action incidence) . ( ) infectives leave the infective class at rate αi per unit time. ( ) there is no entry into or departure from the population, except possibly through death from the disease. according to ( ) , since the probability that a random contact by an infective is with a susceptible, who can then transmit infection, is s/n , the number of new infections in unit time per infective is (βn )(s/n ), giving a rate of new infections (βn )(s/n )i = βsi. alternately, we may argue that for a contact by a susceptible the probability that this contact is with an infective is i/n and thus the rate of new infections per susceptible is (βn )(i/n ), giving a rate of new infections (βn )(i/n )s = βsi. note that both approaches give the same rate of new infections; there are situations which we shall encounter where one is more appropriate than the other. we need not give an algebraic expression for n since it cancels out of the final model, but we should note that for a disease that is fatal to all who are infected n = s +i; while, for a disease from which all infected members recover with immunity, n = s + i + r. later, we will allow the possibility that some infectives recover while others die of the disease. the hypothesis ( ) really says that the time scale of the disease is much faster than the time scale of births and deaths so that demographic effects on the population may be ignored. an alternative view is that we are only interested in studying the dynamics of a single epidemic outbreak. in later sections we shall consider models that are the same as those considered in this first section except for the incorporation of demographic effects (births and deaths) along with the corresponding epidemiological assumptions. the assumption ( ) requires a fuller mathematical explanation, since the assumption of a recovery rate proportional to the number of infectives has no clear epidemiological meaning. we consider the "cohort" of members who were all infected at one time and let u(s) denote the number of these who are still infective s time units after having been infected. if a fraction α of these leave the infective class in unit time then u = −αu , and the solution of this elementary differential equation is thus, the fraction of infectives remaining infective s time units after having become infective is e −αs , so that the length of the infective period is distributed exponentially with mean ∞ e −αs ds = /α, and this is what ( ) really assumes. the assumptions of a rate of contacts proportional to population size n with constant of proportionality β, and of an exponentially distributed recovery rate are unrealistically simple. more general models can be constructed and analyzed, but our goal here is to show what may be deduced from extremely simple models. it will turn out that many more realistic models exhibit very similar qualitative behaviours. in our model r is determined once s and i are known, and we can drop the r equation from our model, leaving the system of two equations we are unable to solve this system analytically but we learn a great deal about the behaviour of its solutions by the following qualitative approach. to begin, we remark that the model makes sense only so long as s(t) and i(t) remain non-negative. thus if either s(t) or i(t) reaches zero we consider the system to have terminated. we observe that s < for all t and i > if and only if s > α/β. thus i increases so long as s > α/β but since s decreases for all t, i ultimately decreases and approaches zero. if s( ) < α/β, i decreases to zero (no epidemic), while if s( ) > α/β, i first increases to a maximum attained when s = α/β and then decreases to zero (epidemic). we think of introducing a small number of infectives into a population of susceptibles and ask whether there will be an epidemic. the quantity βs( )/α is a threshold quantity, called the basic reproduction number and denoted by r , which determines whether there is an epidemic or not. if r < the infection dies out, while if r > there is an epidemic. the definition of the basic reproduction number r is that the basic reproduction number is the number of secondary infections caused by a single infective introduced into a wholly susceptible population of size k ≈ s( ) over the course of the infection of this single infective. in this situation, an infective makes βk contacts in unit time, all of which are with susceptibles and thus produce new infections, and the mean infective period is /α; thus the basic reproduction number is actually βk/α rather than βs( )/α. instead of trying to solve for s and i as functions of t, we divide the two equations of the model to give and integrate to find the orbits (curves in the (s, i)-plane, or phase plane) with c an arbitrary constant of integration. here, we are using log to denote the natural logarithm. another way to describe the orbits is to define the note that the maximum value of i on each of these orbits is attained when s = α/β. note also that since none of these orbits reaches the i -axis, s > for all times. in particular, s ∞ = lim t→∞ s(t) > , which implies that part of the population escapes infection. let us think of a population of size k into which a small number of infectives is introduced, so that s ≈ k, i ≈ , and r = βk/α. if we use the fact that lim t→∞ i(t) = , and let s ∞ = lim t→∞ s(t), then the relation from which we obtain an expression for β/α in terms of the measurable quantities s and s ∞ , namely we may rewrite this in terms of r as the final size relation in particular, since the right side of ( . ) is finite, the left side is also finite, and this shows that s ∞ > . it is generally difficult to estimate the contact rate β which depends on the particular disease being studied but may also depend on social and behavioural factors. the quantities s and s ∞ may be estimated by serological studies (measurements of immune responses in blood samples) before and after an epidemic, and from these data the basic reproduction number r may be estimated by using ( . ) . this estimate, however, is a retrospective one which can be determined only after the epidemic has run its course. initially, the number of infectives grows exponentially because the equation for i may be approximated by i = (βk − α)i and the initial growth rate is this initial growth rate r may be determined experimentally when an epidemic begins. then since k and α may be measured β may be calculated as however, because of incomplete data and under-reporting of cases this estimate may not be very accurate. this inaccuracy is even more pronounced for an outbreak of a previously unknown disease, where early cases are likely to be misdiagnosed. the maximum number of infectives at any time is the number of infectives when the derivative of i is zero, that is, when s = α/β. this maximum is given by obtained by substituting s = α/β, i = i max into ( . ). the village of eyam near sheffield, england suffered an outbreak of bubonic plague in - the source of which is generally believed to be the great plague of london. the eyam plague was survived by only of an initial population of persons. as detailed records were preserved and as the community was persuaded to quarantine itself to try to prevent the spread of disease to other communities, the disease in eyam has been used as a case study for modeling [ ] . detailed examination of the data indicates that there were actually two outbreaks of which the first was relatively mild. thus we shall try to fit the model ( . ) over the period from mid-may to mid-october , measuring time in months with an initial population of seven infectives and susceptibles, and a final population of . values of susceptibles and infectives in eyam are given in [ ] for various dates, beginning with s( ) = , i( ) = , shown in table . . the relation ( the actual data for the eyam epidemic are remarkably close to the predictions of this very simple model. however, the model is really too good to be true. our model assumes that infection is transmitted directly between people. while this is possible, bubonic plague is transmitted mainly by rat fleas. when an infected rat is bitten by a flea, the flea becomes extremely hungry and bites the host rat repeatedly, spreading the infection in the rat. when the host rat dies its fleas move on to other rats, spreading the disease further. as the number of available rats decreases the fleas move to human hosts, and this is how plague starts in a human population (although the second phase of the epidemic may have been the pneumonic form of bubonic plague, which can be spread from person to person). one of the main reasons for the spread of plague from asia into europe was the passage of many trading ships; in medieval times ships were invariably infested with rats. an accurate model of plague transmission would have to include flea and rat populations, as well as movement in space. such a model would be extremely complicated and its predictions might well not be any closer to observations than our simple unrealistic model. in [ ] a stochastic model was also used to fit the data, but the fit was rather poorer than the fit for the simple deterministic model ( . ). in the village of eyam the rector persuaded the entire community to quarantine itself to prevent the spread of disease to other communities. this policy actually increased the infection rate in the village by keeping fleas, rats, and people in close contact with one another, and the mortality rate from bubonic plague was much higher in eyam than in london. further, the quarantine could do nothing to prevent the travel of rats and thus did little to prevent the spread of disease to other communities. one message this suggests to mathematical modelers is that control strategies based on false models may be harmful, and it is essential to distinguish between assumptions that simplify but do not alter the predicted effects substantially, and wrong assumptions which make an important difference. the assumption in the model ( . ) of a rate of contacts per infective which is proportional to population size n , called mass action incidence or bilinear incidence, was used in all the early epidemic models. however, it is quite unrealistic, except possibly in the early stages of an epidemic in a population of moderate size. it is more realistic to assume a contact rate which is a non-increasing function of total population size. for example, a situation in which the number of contacts per infective in unit time is constant, called standard incidence, is a more accurate description for sexually transmitted diseases. we generalize the model ( . ) by replacing the assumption ( ) by the assumption that an average member of the population makes c(n ) contacts in unit time with c (n ) ≥ [ , ] , and we define it is reasonable to assume β (n ) ≤ to express the idea of saturation in the number of contacts. then mass action incidence corresponds to the choice c(n ) = βn , and standard incidence corresponds to the choice c(n ) = λ. some epidemic models [ ] have used a michaelis-menten type of interaction of the form another form based on a mechanistic derivation for pair formation [ ] leads to an expression of the form data for diseases transmitted by contact in cities of moderate size [ ] suggests that data fits the assumption of a form with a = . quite well. all of these forms satisfy the conditions c (n ) ≥ , β (n ) ≤ . because the total population size is now present in the model we must include an equation for total population size in the model. this forces us to make a distinction between members of the population who die of the disease and members of the population who recover with immunity against reinfection. we assume that a fraction f of the αi members leaving the infective class at time t recover and the remaining fraction ( − f ) die of disease. we use s, i, and n as variables, with n = s + i + r. we now obtain a three-dimensional model we also have the equation r = −fαi, but we need not include it in the model since r is determined when s, i, and n are known. we should note that if f = the total population size remains equal to the constant k, and the model ( . ) reduces to the simpler model ( . ) with β replaced by the constant β(k). we wish to show that the model ( . ) has the same qualitative behaviour as the model ( . ), namely that there is a basic reproduction number which distinguishes between disappearance of the disease and an epidemic outbreak, and that some members of the population are left untouched when the epidemic passes. these two properties are the central features of all epidemic models. for the model ( . ) the basic reproduction number is given by because a single infective introduced into a wholly susceptible population makes c(k) = kβ(k) contacts in unit time, all of which are with susceptibles and thus produce new infections, and the mean infective period is /α. in addition to the basic reproduction number r there is also a timedependent running reproduction number which we call r * , representing the number of secondary infections caused by a single individual in the population who becomes infective at time t. in this situation, an infective makes c(n ) = nβ(n ) contacts in unit time and a fraction s/n are with susceptibles and thus produce new infections. thus it is easy to see that for the model ( . ) the running reproduction number is given by if r * < for all large t, the epidemic will pass. we may calculate the rate of change of the running reproduction number with respect to time, using ( . ) and ( . ) to find that and an epidemic begins. however, r * decreases until it is less than and then remains less than . thus the epidemic will pass. if r < then and there is no epidemic. from ( . ) we obtain integration of these equations from to t gives when we combine these two equations, eliminating the integral expression, and use if we let t → ∞, s(t) and n (t) decrease monotonically to limits s ∞ and n ∞ respectively and i(t) → . this gives the relation in this equation, k−n ∞ is the change in population size, which is the number of disease deaths over the course of the epidemic, while k − s ∞ is the change in the number of susceptibles, which is the number of disease cases over the course of the epidemic. in this model, ( . ) is obvious, but we shall see in a more general setting how to derive an analogous equation from which we can calculate an average disease mortality. equation ( . ) generalizes to the infection age epidemic model of kermack and mckendrick. if we use the same approach as was used for ( . ) to show that s ∞ > , we obtain and we are unable to proceed because of the dependence on n . however, we may use a different approach to obtain the desired result. we assume that β( ) is finite, thus ruling out standard incidence. if we let t → ∞ in the second equation of ( . ) we obtain the first equation of ( . ) may be written as since since the right side of this inequality is finite, the left side is also finite and this establishes that s ∞ > . in addition, if we use the same integration together with the inequality we obtain a final size inequality if β(n ) → ∞ as n → we must use a different approach to analyze the limiting behaviour. it is possible to show that s ∞ = is possible only if n → and k β(n )dn diverges, and this is possible only if f = , that is, only if all infectives ide of disease. the assumption that β(n ) is unbounded as n → is biologically unreasonable. in particular, standard incidence is not realistic for small population sizes. a more realistic assumption would be that the number of contacts per infective in unit time is linear for small population size and saturates for larger population sizes, which rules out the possibility that the epidemic sweeps through the entire population. in many infectious diseases there is an exposed period after the transmission of infection from susceptibles to potentially infective members but before these potential infectives can transmit infection. if the exposed period is short it is often neglected in modeling. a longer exposed period could perhaps lead to significantly different model predictions, and we need to show that this is not the case. to incorporate an exponentially distributed exposed period with mean exposed period /κ we add an exposed class e and use compartments s, e, i, r and total population size n = s +e +i +r to give a generalization of the epidemic model ( . ). we also have the equation r = −fαi, but we need not include it in the model since r is determined when s, i, and n are known. a flow chart is shown in fig. . . the analysis of this model is the same as the analysis of ( . ), but with i replaced by e + i. that is, instead of using the number of infectives as one of the variables we use the total number of infected members, whether or not they are capable of transmitting infection. some diseases have an asymptomatic stage in which there is some infectivity rather than an exposed period. this may be modeled by assuming infectivity reduced by a factor ε e during an exposed period. a calculation of the rate of new infections per susceptible leads to a model there is a final size relation like ( . ) for the model ( . ). integration of the sum of the first two equations of ( . ) from to ∞ gives and division of the first equation of ( . ) by s followed by integration from to ∞ gives the same integration using β(n ) ≤ β( ) < ∞ shows as in the previous section that s ∞ > . one form of treatment that is possible for some diseases is vaccination to protect against infection before the beginning of an epidemic. for example, this approach is commonly used for protection against annual influenza outbreaks. a simple way to model this would be to reduce the total population size by the fraction of the population protected against infection. however, in reality such inoculations are only partly effective, decreasing the rate of infection and also decreasing infectivity if a vaccinated person does become infected. to model this, it would be necessary to divide the population into two groups with different model parameters and to make some assumptions about the mixing between the two groups. we will not explore such more complicated models here. if there is a treatment for infection once a person has been infected, we model this by supposing that a fraction γ per unit time of infectives is selected for treatment, and that treatment reduces infectivity by a fraction δ. suppose that the rate of removal from the treated class is η. the sit r model, where t is the treatment class, is given by a flow chart is shown in fig. . . it is not difficult to prove, much as was done for the model ( . ) that in order to calculate the basic reproduction number, we may argue that an infective in a totally susceptible population causes βk new infections in unit time, and the mean time spent in the infective compartment is /(α + γ). in addition, a fraction γ/(α+γ) of infectives are treated. while in the treatment stage the number of new infections caused in unit time is δβk, and the mean time in the treatment class is /η. thus r is it is also possible to establish the final size relation ( . ) by means similar to those used for the simple model ( . ). we integrate the first equation of ( . ) to obtain log integration of the third equation of ( . ) gives integration of the sum of the first two equations of ( . ) gives combination of these three equations and ( . ) gives if β is constant, this relation is an equality, and is the same as ( . ). an actual epidemic differs considerably from the idealized models ( . ) or ( . ), as was shown by the sars epidemic of - . some notable differences are: . as we have seen in the preceding section, at the beginning of an epidemic the number of infectives is small and a deterministic model, which presupposes enough infectives to allow homogeneous mixing, is inappropriate. . when it is realized that an epidemic has begun, individuals are likely to modify their behaviour by avoiding crowds to reduce their contacts and by being more careful about hygiene to reduce the risk that a contact will produce infection. . if a vaccine is available for the disease which has broken out, public health measures will include vaccination of part of the population. various vaccination strategies are possible, including vaccination of health care workers and other first line responders to the epidemic, vaccination of members of the population who have been in contact with diagnosed infectives, or vaccination of members of the population who live in close proximity to diagnosed infectives. . diagnosed infectives may be hospitalized, both for treatment and to isolate them from the rest of the population. . contact tracing of diagnosed infectives may identify people at risk of becoming infective, who may be quarantined (instructed to remain at home and avoid contacts) and monitored so that they may be isolated immediately if and when they become infective. . in some diseases, exposed members who have not yet developed symptoms may already be infective, and this would require inclusion in the model of new infections caused by contacts between susceptibles and asymptomatic infectives from the exposed class. . isolation may be imperfect; in-hospital transmission of infection was a major problem in the sars epidemic. in the sars epidemic of - in-hospital transmission of disease from patients to health care workers or visitors because of imperfect isolation accounted for many of the cases. this points to an essential heterogeneity in disease transmission which must be included whenever there is any risk of such transmission. all these generalizations have been considered in studies of the sars epidemic of - . while the ideas were suggested in sars modelling, they are in fact relevant to any epidemic. one beneficial effect of the sars epidemic has been to draw attention to epidemic modelling which may be of great value in coping with future epidemics. if a vaccine is available for a disease which threatens an epidemic outbreak, a vaccinated class which is protected at least partially against infection should be included in a model. while this is not relevant for an outbreak of a new disease, it would be an important aspect to be considered in modelling an influenza epidemic or a bioterrorist outbreak of smallpox. for an outbreak of a new disease, where no vaccine is available, isolation and quarantine are the only control measures available. let us formulate a model for an epidemic once control measures have been started. thus, we assume that an epidemic has started, but that the number of infectives is small and almost all members of the population are still susceptible. we formulate a model to describe the course of an epidemic when control measures are begun under the assumptions: . exposed members may be infective with infectivity reduced by a factor ε e , ≤ ε e < . . exposed members who are not isolated become infective at rate κ . . we introduce a class q of quarantined members and a class j of isolated members. . exposed members are quarantined at a proportional rate γ in unit time (in practice, a quarantine will also be applied to many susceptibles, but we ignore this in the model). quarantine is not perfect, but reduces the contact rate by a factor ε q . the effect of this assumption is that some susceptibles make fewer contacts than the model assumes. . there may be transmission of disease by isolated members, with an infectivity factor of ε j . . infectives are diagnosed at a proportional rate γ per unit time and isolated. in addition, quarantined members are monitored and when they develop symptoms at rate κ they are isolated immediately. . infectives leave the infective class at rate α and a fraction f of these recover, and isolated members leave the isolated class at rate α with a fraction f recovering. these assumptions lead to the seqijr model [ ] here, we have used an equation for n to replace the equation the model before control measures are begun is the special case of ( . ). it is the same as ( . ). we define the control reproduction number r c to be the number of secondary infections caused by a single infective in a population consisting essentially only of susceptibles with the control measures in place. it is analogous to the basic reproduction number but instead of describing the very beginning of the disease outbreak it describes the beginning of the recognition of the epidemic. the basic reproduction number is the value of the control reproduction number with in addition, there is a time-dependent effective reproduction number r * which continues to track the number of secondary infections caused by a single infective as the epidemic continues with control measures (quarantine of asymptomatics and isolation of symptomatics) in place. it is not difficult to show that if the inflow into the population from travellers and new births is small (i.e., if the epidemiological time scale is much faster than the demographic time scale), our model implies that r * will become and remain less than unity, so that the epidemic will always pass. even if r c > , the epidemic will abate eventually when the effective reproduction number becomes less than unity. the effective reproduction number r * is essentially r c multiplied by a factor s/n , but allows time-dependent parameter values as well. however, it should be remembered that if the epidemic takes so long to pass that there are enough new births and travellers to keep r * > , there will be an endemic equilibrium meaning that the disease will establish itself and remain in the population. we have already calculated r for ( . ) and we may calculate r c in the same way but using the full model with quarantined and isolated classes. we obtain each term of r c has an epidemiological interpretation. the mean duration in e is /d with contact rate ε e β, giving a contribution to r c of ε e kβ(k)/d . a fraction κ /d goes from e to i, with contact rate β and mean duration /d , giving a contribution of kβ(k)κ /d d . a fraction γ /d goes from e to q, with contact rate ε e ε q β and mean duration /κ , giving a contribution of ε e ε q kβ(k)γ /d κ . a fraction κ γ /d d goes from e to i to j, with a contact rate of ε j β and a mean duration of /α , giving a contribution of ε j kβ(k)κ γ /α d d . finally, a fraction γ /d goes from e to q to j with a contact rate of ε j β and a mean duration of /α giving a contribution of ε j kβ(k)γ /d α . the sum of these individual contributions gives r c . in the model ( . ) the parameters γ and γ are control parameters which may be varied in the attempt to manage the epidemic. the parameters q and j depend on the strictness of the quarantine and isolation processes and are thus also control measures in a sense. the other parameters of the model are specific to the disease being studied. while they are not variable, their measurements are subject to experimental error. the linearization of ( . ) at the disease-free equilibrium (k, , , , , k) the corresponding characteristic equation is a fourth degree polynomial equation whose leading coefficient is and whose constant term is a positive constant multiple of − r c , thus positive if r c < and negative if r c > . if r c > there is a positive eigenvalue, corresponding to an initial exponential growth rate of solutions of ( . ). if r c < it is possible to show that all eigenvalues of the coefficient matrix have negative real part, and thus solutions of ( . ) die out exponentially [ ] . next, we wish to show that analogues of the relation ( . ) and s ∞ > derived for the model ( . ) are valid for the management model ( . ). we begin by integrating the equations for s + e, q, i, j, and n of ( . ) with respect to t from t = to t = ∞, using the initial conditions we obtain, since e, q, i, and j all approach zero at t → ∞, in order to relate (k − s ∞ ) to (k − n ∞ ), we need to express thus we have this has the form, analogous to ( . ), with c, the disease death rate, given by the mean disease death rate may be measured and this expression gives information about some of the parameters in the model which can not be measured directly. it is easy to see that an argument similar to the one used for ( . ) but technically more complicated may be used to show that s ∞ > for the treatment model ( . ). thus the asymptotic behaviour of the management model ( . ) is the same as that of the simpler model ( . ). if the control reproduction number r c is less than the disease dies out and if r c > there is an epidemic which will pass leaving some members of the population untouched. the underlying assumptions of the models of kermack-mckendrick type studied in this chapter are that the sizes of the compartments are large enough that the mixing of members is homogeneous. while these assumptions are probably reasonable once an epidemic is well underway, at the beginning of a disease outbreak the situation may be quite different. at the beginning of an epidemic most members of the population are susceptible, that is, not (yet) infected, and the number of infectives (members of the population who are infected and may transmit infection) is small. the transmission of infection depends strongly on the pattern of contacts between members of the population, and a description should involve this pattern. since the number of infectives is small a description involving an assumption of mass action should be replaced by a model which incorporates stochastic effects. one approach would be a complete description of stochastic epidemic models, for which we refer the reader to the chapter on stochastic models in this volume by linda allen. another approach would be to consider a stochastic model for an outbreak of a communicable disease to be applied so long as the number of infectives remains small, distinguishing a (minor) disease outbreak confined to this initial stage from a (major) epidemic which occurs if the number of infectives begins to grow at an exponential rate. once an epidemic has started we may switch to a deterministic compartmental model. this approach is described in chap. on network models in this volume. there is an important difference between the behaviour of network models and the behaviour of models of kermack-mckendrick type, namely that for a stochastic disease outbreak model if r < the probability that the infection will die out is , while if r > there is a positive probability that the infection will persist, and will lead to an epidemic and a positive probability that the infection will increase initially but will produce only a minor outbreak and will die out before triggering a major epidemic. epidemics which sweep through a population attract much attention and arouse a great deal of concern. as we have mentioned in the introduction, the prevalence and effects of many diseases in less developed countries are probably less well-known but may be of even more importance. there are diseases which are endemic in many parts of the world and which cause millions of deaths each year. we have omitted births and deaths in our description of models because the time scale of an epidemic is generally much shorter than the demographic time scale. in effect, we have used a time scale on which the number of births and deaths in unit time is negligible. to model a disease which may be endemic we need to think on a longer time scale and include births and deaths. for diseases that are endemic in some region public health physicians need to be able to estimate the number of infectives at a given time as well as the rate at which new infections arise. the effects of quarantine or vaccine in reducing the number of victims are of importance, just as in the treatment of epidemics. in addition, the possibility of defeating the endemic nature of the disease and thus controlling or even eradicating the disease in a population is worthy of study. measles is a disease for which endemic equilibria have been observed in many places, frequently with sustained oscillations about the equilibrium. the epidemic model of the first section assumes that the epidemic time scale is so short relative to the demographic time scale that demographic effects may be ignored. for measles, however, the reason for the endemic nature of the disease is that there is a flow of new susceptible members into the population, and in order to try to model this we must include births and deaths in the model. the simplest way to incorporate births and deaths in an infectious disease model is to assume a constant number of births and an equal number of deaths per unit time so that the total population size remains constant. this is, of course, feasible only if there are no deaths due to the disease. in developed countries such an assumption is plausible because there are few deaths from measles. in less developed countries there is often a very high mortality rate for measles and therefore other assumptions are necessary. the first attempt to formulate an sir model with births and deaths to describe measles was given in by h.e. soper [ ] , who assumed a constant birth rate µk in the susceptible class and a constant death rate µk in the removed class. his model is this model is unsatisfactory biologically because the linkage of births of susceptibles to deaths of removed members is unreasonable. it is also an improper model mathematically because if r( ) and i( ) are sufficiently small then r(t) will become negative. for any disease model to be plausible it is essential that the problem be properly posed in the sense that the number of members in each class must remain non-negative. a model that does not satisfy this requirement cannot be a proper description of a disease model and therefore must contain some assumption that is biologically unreasonable. a full analysis of a model should include verification of this property. a model of kermack and mckendrick [ ] includes births in the susceptible class proportional to total population size and a death rate in each class proportional to the number of members in the class. this model allows the total population size to grow exponentially or die out exponentially if the birth and death rates are unequal. it is applicable to such questions as whether a disease will control the size of a population that would otherwise grow exponentially. we shall return to this topic, which is important in the study of many diseases in less developed countries with high birth rates. to formulate a model in which total population size remains bounded we could follow the approach suggested by [ ] in which the total population size is held constant by making birth and death rates equal. such a model is because s + i + r = k, we can view r as determined when s and i are known and consider the two-dimensional system we shall examine a slightly more general sir model with births and deaths for a disease that may be fatal to some infectives. for such a disease the class r of removed members should contain only recovered members, not members removed by death from the disease. it is not possible to assume that the total population size remain constant if there are deaths due to disease; a plausible model for a disease that may be fatal to some infectives must allow the total population to vary in time. the simplest assumption to allow this is a constant birth rate Λ, but in fact the analysis is quite similar if the birth rate is a function Λ(n ) of total population size n . let us analyze the model where n = s + i + r, with a mass action contact rate, a constant number of births Λ per unit time, a proportional natural death rate µ in each class, and a rate of recovery or disease death α of infectives with a fraction f of infectives recovering with immunity against reinfection. in this model if f = the total population size approaches a limit k = Λ/µ. then k is the carrying capacity of the population. if f < the total population size is not constant and k represents a carrying capacity or maximum possible population size, rather than a population size. we view the first two equations as determining s and i, and then consider the third equation as determining n once s and i are known. this is possible because n does not enter into the first two equations. instead of using n as the third variable in this model we could have used r, and the same reduction would have been possible. if the birth or recruitment rate Λ(n ) is a function of total population size then in the absence of disease the total population size n satisfies the differential equation the carrying capacity of population size is the limiting population size k, satisfying the condition Λ (k) < µ assures the asymptotic stability of the equilibrium population size k. it is reasonable to assume that k is the only positive equilibrium, so that Λ(n ) > µn for ≤ n ≤ k. for most population models, however, if Λ(n ) represents recruitment into a behavioural class, as would be natural for models of sexually transmitted diseases, it would be plausible to have Λ( ) > , or even to consider Λ(n ) to be a constant function. if Λ( ) = , we require Λ ( ) > µ because if this requirement is not satisfied there is no positive equilibrium and the population would die out even in the absence of disease. we have used a mass action contact rate for simplicity, even though a more general contact rate would give a more accurate model, just as in the epidemics considered in the preceding section. with a general contact rate and a density-dependent birth rate we would have a model if f = , so that there are no disease deaths, the equation for n is so that n (t) approaches a limiting population size k. the theory of asymptotically autonomous systems [ , , , ] implies that if n has a constant limit then the system is equivalent to the system in which n is replaced by this limit. then the system ( . ) is the same as the system ( . ) with β replaced by the constant β(k) and n by k, and Λ(n ) replaced by the constant Λ(k) = µk. we shall analyze the model ( . ) qualitatively. in view of the remark above, our analysis will also apply to the more general model ( . ) if there are no disease deaths. analysis of the system ( . ) with f < is much more difficult. we will confine our study of ( . ) to a description without details. the first stage of the analysis is to note that the model ( . ) is a properly posed problem. that is, since s ≥ if s = and i ≥ if i = , we have s ≥ , i ≥ for t ≥ and since n ≤ if n = k we have n ≤ k for t ≥ . thus the solution always remains in the biologically realistic region s ≥ , i ≥ , ≤ n ≤ k if it starts in this region. by rights, we should verify such conditions whenever we analyze a mathematical model, but in practice this step is frequently overlooked. our approach will be to identify equilibria (constant solutions) and then to determine the asymptotic stability of each equilibrium. asymptotic stability of an equilibrium means that a solution starting sufficiently close to the equilibrium remains close to the equilibrium and approaches the equilibrium as t → ∞ >, while instability of the equilibrium means that there are solutions starting arbitrarily close to the equilibrium which do not approach it. to find equilibria (s ∞ , i ∞ ) we set the right side of each of the two equations equal to zero. the second of the resulting algebraic equations factors, giving two alternatives. the first alternative is i ∞ = , which will give a disease-free equilibrium, and the second alternative is βs ∞ = µ + α, which will give an endemic equilibrium, provided βs ∞ = µ+α < βk. if i ∞ = the other equation gives s ∞ = k = Λ/µ. for the endemic equilibrium the first equation gives we linearize about an equilibrium (s ∞ , i ∞ ) by letting y = s−s ∞ , z = i−i ∞ , writing the system in terms of the new variables y and z and retaining only the linear terms in a taylor expansion. we obtain a system of two linear differential equations, the coefficient matrix of this linear system is we then look for solutions whose components are constant multiples of e λt ; this means that λ must be an eigenvalue of the coefficient matrix. the condition that all solutions of the linearization at an equilibrium tend to zero as t → ∞ is that the real part of every eigenvalue of this coefficient matrix is negative. at the disease-free equilibrium the matrix is which has eigenvalues −µ and βk − µ − α. thus, the disease-free equilibrium is asymptotically stable if βk < µ + α and unstable if βk > µ + α. note that this condition for instability of the disease-free equilibrium is the same as the condition for the existence of an endemic equilibrium. in general, the condition that the eigenvalues of a × matrix have negative real part is that the determinant be positive and the trace (the sum of the diagonal elements) be negative. since βs ∞ = µ+α at an endemic equilibrium, the matrix of the linearization at an endemic equilibrium is and this matrix has positive determinant and negative trace. thus, the endemic equilibrium, if there is one, is always asymptotically stable. if the quantity is less than one, the system has only the disease-free equilibrium and this equilibrium is asymptotically stable. in fact, it is not difficult to prove that this asymptotic stability is global, that is, that every solution approaches the disease-free equilibrium. if the quantity r is greater than one then the disease-free equilibrium is unstable, but there is an endemic equilibrium that is asymptotically stable. again, the quantity r is the basic reproduction number. it depends on the particular disease (determining the parameter α) and on the rate of contacts, which may depend on the population density in the community being studied. the disease model exhibits a threshold behaviour: if the basic reproduction number is less than one the disease will die out, but if the basic reproduction number is greater than one the disease will be endemic. just as for the epidemic models of the preceding section, the basic reproduction number is the number of secondary infections caused by a single infective introduced into a wholly susceptible population because the number of contacts per infective in unit time is βk, and the mean infective period (corrected for natural mortality) is /(µ + α). there are two aspects of the analysis of the model ( . ) which are more complicated than the analysis of ( . ). the first is in the study of equilibria. because of the dependence of Λ(n ) and β(n ) on n , it is necessary to use two of the equilibrium conditions to solve for s and i in terms of n and then substitute into the third condition to obtain an equation for n . then by comparing the two sides of this equation for n = and n = k it is possible to show that there must be an endemic equilibrium value of n between and k. the second complication is in the stability analysis. since ( . ) is a threedimensional system which can not be reduced to a two-dimensional system, the coefficient matrix of its linearization at an equilibrium is a × matrix and the resulting characteristic equation is a cubic polynomial equation of the form λ + a λ + a λ + a = . the routh-hurwitz conditions a > , a a > a > are necessary and sufficient conditions for all roots of the characteristic equation to have negative real part. a technically complicated calculation is needed to verify that these conditions are satisfied at an endemic equilibrium for the model ( . ). the asymptotic stability of the endemic equilibrium means that the compartment sizes approach a steady state. if the equilibrium had been unstable, there would have been a possibility of sustained oscillations. oscillations in a disease model mean fluctuations in the number of cases to be expected, and if the oscillations have long period could also mean that experimental data for a short period would be quite unreliable as a predictor of the future. epidemiological models which incorporate additional factors may exhibit oscillations. a variety of such situations is described in [ , ] . the epidemic models of the first section also exhibited a threshold behaviour but of a slightly different kind. for these models, which were sir models without births or natural deaths, the threshold distinguished between a dying out of the disease and an epidemic, or short term spread of disease. from the third equation of ( . ) we obtain where n = s + i + r. from this we see that at the endemic equilibrium n = k − ( − f )αi/µ, and the reduction in the population size from the carrying capacity k is the parameter α in the sir model may be considered as describing the pathogenicity of the disease. if α is large it is less likely that r > . if α is small then the total population size at the endemic equilibrium is close to the carrying capacity k of the population. thus, the maximum population decrease caused by disease will be for diseases of intermediate pathogenicity. in order to describe a model for a disease from which infectives recover with immunity against reinfection and that includes births and deaths as in the model ( . ), we may modify the model ( . ) by removing the equation for r and moving the term fαi describing the rate of recovery from infection to the equation for s. this gives the model describing a population with a density-dependent birth rate Λ(n ) per unit time, a proportional death rate µ in each class, and with a rate α of departure from the infective class through recovery or disease death and with a fraction f of infectives recovering with no immunity against reinfection. in this model, if f < the total population size is not constant and k represents a carrying capacity, or maximum possible population size, rather than a constant population size. it is easy to verify that if we add the two equations of ( . ), and use n = s + i we obtain for the sis model we are able to carry out the analysis with a general contact rate. if f = the equation for n is and n approaches the limit k. the system ( . ) is asymptotically autonomous and its asymptotic behaviour is the same as that of the single differential equation where s has been replaced by k − i. but ( . ) is a logistic equation which is easily solved analytically by separation of variables or qualitatively by an equilibrium analysis. we find that i → if kβ(k) < (µ + α), or r < and i → i ∞ > with to analyze the sis model if f < , it is convenient to use i and n as variables instead of s and i, with s replaced by n − i. this gives the model equilibria are found by setting the right sides of the two differential equations equal to zero. the first of the resulting algebraic equations factors, giving two alternatives. the first alternative is i = , which will give a disease-free equilibrium i = , n = k, and the second alternative is β(n )(n − i) = µ + α, which may give an endemic equilibrium. for an endemic equilibrium (i ∞ , n ∞ ) the first equation gives substitution into the other equilibrium condition gives which can be simplified to while the right side of ( . ) is since if r > the left side of ( . ) is less than the right side of ( . ), and this implies that ( . ) has a solution for n, < n < k. thus there is an endemic equilibrium if r > . if r < this reasoning may be used to show that there is no endemic equilibrium. the linearization of ( . ) at an equilibrium (i ∞ , n ∞ ) has coefficient matrix at the disease-free equilibrium the matrix is which has eigenvalues Λ (k) − µ and kβk − (µ + α). thus, the diseasefree equilibrium is asymptotically stable if kβ(k) < µ + α, or r < , and unstable if kβ(k) > µ+α, or r > . note that the condition for instability of the disease-free equilibrium is the same as the condition for the existence of an endemic equilibrium. at an endemic equilibrium, since β( it is clear that the coefficient matrix has negative trace and positive determinant if Λ (n ) < µ and this implies that the endemic equilibrium is asymptotically stable. thus, the endemic equilibrium, which exists if r > , is always asymptotically stable. if r < the system has only the disease-free equilibrium and this equilibrium is asymptotically stable. in the case f = the verification of these properties remains valid if there are no births and deaths. this suggests that a requirement for the existence of an endemic equilibrium is a flow of new susceptibles either through births, as in the sir model or through recovery without immunity against reinfection, as in the sis model with or without births and deaths. if the epidemiological and demographic time scales are very different, for the sir model we observed that the approach to endemic equilibrium is like a rapid and severe epidemic. the same happens in the sis model, especially if there is a significant number of deaths due to disease. if there are few disease deaths the number of infectives at endemic equilibrium may be substantial, and there may be damped oscillations of large amplitude about the endemic equilibrium. for both the sir and sis models we may write the differential equation for i as which implies that whenever s exceeds its endemic equilibrium value s ∞ , i is increasing and epidemic-like behaviour is possible. if r < and s < k it follows that i < , and thus i is decreasing. thus, if r < , i cannot increase and no epidemic can occur. next, we will turn to some applications of sir and sis models, taken mainly from [ ] . in order to prevent a disease from becoming endemic it is necessary to reduce the basic reproduction number r below one. this may sometimes be achieved by immunization. if a fraction p of the Λ newborn members per unit time of the population is successfully immunized, the effect is to replace k by k ( − p) , and thus to reduce the basic reproduction number to r ( − p) . a population is said to have herd immunity if a large enough fraction has been immunized to assure that the disease cannot become endemic. the only disease for which this has actually been achieved worldwide is smallpox for which r is approximately , so that % immunization does provide herd immunity. for measles, epidemiological data in the united states indicate that r for rural populations ranges from . to . , requiring vaccination of . - . % of the population. in urban areas r ranges from . to . , requiring vaccination of . - . % of the population. in great britain, r ranges from . to . , requiring vaccination of - % of the population. the measles vaccine is not always effective, and vaccination campaigns are never able to reach everyone. as a result, herd immunity against measles has not been achieved (and probably never can be). since smallpox is viewed as more serious and requires a lower percentage of the population be immunized, herd immunity was attainable for smallpox. in fact, smallpox has been eliminated; the last known case was in somalia in , and the virus is maintained now only in laboratories (although there is currently some concern that it may be reintroduced as a bioterrorism attack). the eradication of smallpox was actually more difficult than expected because high vaccination rates were achieved in some countries but not everywhere, and the disease persisted in some countries. the eradication of smallpox was possible only after an intensive campaign for worldwide vaccination [ ] . in order to calculate the basic reproduction number r for a disease, we need to know the values of the contact rate β and the parameters µ, k, and α. the parameters µ, k, and α can usually be measured experimentally but the contact rate β is difficult to determine directly. there is an indirect means of estimating r in terms of the life expectancy and the mean age at infection which enables us to avoid having to estimate the contact rate. in this calculation, we will assume that β is constant, but we will also indicate the modifications needed when β is a function of total population size n . the calculation assumes exponentially distributed life spans and infective periods. in fact, the result is valid so long as the life span is exponentially distributed. consider the "age cohort" of members of a population born at some time t and let a be the age of members of this cohort. if y(a) represents the fraction of members of the cohort who survive to age (at least) a, then the assumption that a fraction µ of the population dies per unit time means that y (a) = −µy(a). since y( ) = , we may solve this first order initial value problem to obtain y(a) = e −µa . the fraction dying at (exactly) age a is −y (a) = µy(a). the mean life span is the average age at death, which is since y(a) = e −µa , this reduces to /µ. the life expectancy is often denoted by l, so that we may write the rate at which surviving susceptible members of the population become infected at age a and time t + a, is βi(t + a). thus, if z(a) is the fraction of the age cohort alive and still susceptible at age a, z (a) = −[µ+βi(t +a)]z(a). solution of this first linear order differential equation gives the mean length of time in the susceptible class for members who may become infected, as opposed to dying while still susceptible, is and this is the mean age at which members become infected. if the system is at an equilibrium i ∞ , this integral may be evaluated, and the mean age at infection, denoted by a, is given by for our model the endemic equilibrium is and this implies this relation is very useful in estimating basic reproduction numbers. for example, in some urban communities in england and wales between and the average age of contracting measles was . years. if life expectancy is assumed to be years, this indicates r = . . if β is a function β(n ) of total population size the relation ( . ) becomes if disease mortality does not have a large effect on total population size, in particular if there is no disease mortality, this relation is very close to ( . ). the relation between age at infection and basic reproduction number indicates that measures such as inoculations, which reduce r , will increase the average age at infection. for diseases such as rubella (german measles), whose effects may be much more serious in adults than in children, this indicates a danger that must be taken into account: while inoculation of children will decrease the number of cases of illness, it will tend to increase the danger to those who are not inoculated or for whom the inoculation is not successful. nevertheless, the number of infections in older people will be reduced, although the fraction of cases which are in older people will increase. many common childhood diseases, such as measles, whooping cough, chicken pox, diphtheria, and rubella, exhibit variations from year to year in the number of cases. these fluctuations are frequently regular oscillations, suggesting that the solutions of a model might be periodic. this does not agree with the predictions of the model we have been using here; however, it would not be inconsistent with solutions of the characteristic equation, which are complex conjugate with small negative real part corresponding to lightly damped oscillations approaching the endemic equilibrium. such behaviour would look like recurring epidemics. if the eigenvalues of the matrix of the linearization at an endemic equilibrium are −u ± iv, where i = − , then the solutions of the linearization are of the form be −ut cos(vt + c), with decreasing "amplitude" be −ut and "period" π v . for the model ( . ) we recall from ( . ) that at the endemic equilibrium we have βi ∞ + µ = µr , βs ∞ = µ + α and from ( . ) the matrix of the linearization is the eigenvalues are the roots of the quadratic equation if the mean infective period /α is much shorter than the mean life span /µ, we may neglect the terms that are quadratic in µ. thus, the eigenvalues are approximately and these are complex with imaginary part µ(r − )α. this indicates oscillations with period approximately π µ(r − )α . we use the relation µ(r − ) = µl/a and the mean infective period τ = /α to see that the interepidemic period t is approximately π √ aτ . thus, for example, for recurring outbreaks of measles with an infective period of weeks or / year in a population with a life expectancy of years with r estimated as , we would expect outbreaks spaced . years apart. also, as the "amplitude" at time t is e −µr t/ , the maximum displacement from equilibrium is multiplied by a factor e −( )( . )/ = . over each cycle. in fact, many observations of measles outbreaks indicate less damping of the oscillations, suggesting that there may be additional influences that are not included in our simple model. to explain oscillations about the endemic equilibrium a more complicated model is needed. one possible generalization would be to assume seasonal variations in the contact rate. this is a reasonable supposition for a childhood disease most commonly transmitted through school contacts, especially in winter in cold climates. note, however, that data from observations are never as smooth as model predictions and models are inevitably gross simplifications of reality which cannot account for random variations in the variables. it may be difficult to judge from experimental data whether an oscillation is damped or persistent. in the model ( . ) the demographic time scale described by the birth and natural death rates Λ and µ and the epidemiological time scale described by the rate α of departure from the infective class may differ substantially. think, for example, of a natural death rate µ = / , corresponding to a human life expectancy of years, and epidemiological parameters α = , f = , describing a disease from which all infectives recover after a mean infective period of / year, or two weeks. suppose we consider a carrying capacity k = , and take β = . , indicating that an average infective makes ( . )( , ) = contacts per year. then r = . , and at the endemic equilibrium we have s ∞ = . , i ∞ = . , r ∞ = . . this equilibrium is globally asymptotically stable and is approached from every initial state. however, if we take s( ) = , i( ) = , r( ) = , simulating the introduction of a single infective into a susceptible population and solve the system numerically we find that the number of infectives rises sharply to a maximum of and then decreases to almost zero in a period of . year, or about months. in this time interval the susceptible population decreases to and then begins to increase, while the removed (recovered and immune against reinfection) population increases to almost , and then begins a gradual decrease. the size of this initial "epidemic" could not have been predicted from our qualitative analysis of the system ( . ). on the other hand, since µ is so small compared to the other parameters of the model, we might consider neglecting µ, replacing it by zero in the model. if we do this, the model reduces to the simple kermack-mckendrick epidemic model (without births and deaths) of the first section. if we follow the model ( . ) over a longer time interval we find that the susceptible population grows to after years, then drops to during a small epidemic with a maximum of infectives, and exhibits widely spaced epidemics decreasing in size. it takes a very long time before the system comes close to the endemic equilibrium and remains close to it. the large initial epidemic conforms to what has often been observed in practice when an infection is introduced into a population with no immunity, such as the smallpox inflicted on the aztecs by the invasion of cortez. if we use the model ( . ) with the same values of β, k and µ, but take α = , f = to describe a disease fatal to all infectives, we obtain very similar results. now the total population is s + i, which decreases from an initial size of , to a minimum of and then gradually increases and eventually approaches its equilibrium size of . . thus, the disease reduces the total population size to one-fourth of its original value, suggesting that infectious diseases may have large effects on population size. this is true even for populations which would grow rapidly in the absence of infection, as we shall see later. many parts of the world experienced very rapid population growth in the eighteenth century. the population of europe increased from million in to million in . in the same time period the population of great britain increased from . million to . million, and the population of china increased from million to million [ ] . the population of english colonies in north america grew much more rapidly than this, aided by substantial immigration from england, but the native population, which had been reduced to one tenth of their previous size by disease following the early encounters with europeans and european diseases, grew even more rapidly. while some of these population increases may be explained by improvements in agriculture and food production, it appears that an even more important factor was the decrease in the death rate due to diseases. disease death rates dropped sharply in the eighteenth century, partly from better understanding of the links between illness and sanitation and partly because the recurring invasions of bubonic plague subsided, perhaps due to reduced susceptibility. one plausible explanation for these population increases is that the bubonic plague invasions served to control the population size, and when this control was removed the population size increased rapidly. in developing countries it is quite common to have high birth rates and high disease death rates. in fact, when disease death rates are reduced by improvements in health care and sanitation it is common for birth rates to decline as well, as families no longer need to have as many children to ensure that enough children survive to take care of the older generations. again, it is plausible to assume that population size would grow exponentially in the absence of disease but is controlled by disease mortality. the sir model with births and deaths of kermack and mckendrick [ ] includes births in the susceptible class proportional to population size and a natural death rate in each class proportional to the size of the class. let us analyze a model of this type with birth rate r and a natural death rate µ < r. for simplicity we assume the disease is fatal to all infectives with disease death rate α, so that there is no removed class and the total population size is n = s + i. our model is from the second equation we see that equilibria are given by either i = or βs = µ + α. if i = the first equilibrium equation is rs = µs, which implies s = since r > µ. it is easy to see that the equilibrium ( , ) is unstable. what actually would happen if i = is that the susceptible population would grow exponentially with exponent r − µ > . if βs = µ + α the first equilibrium condition gives which leads to thus, there is an endemic equilibrium provided r < α + µ, and it is possible to show by linearizing about this equilibrium that it is asymptotically stable. on the other hand, if r > α + µ there is no positive equilibrium value for i. in this case we may add the two differential equations of the model to give and from this we may deduce that n grows exponentially. for this model either we have an asymptotically stable endemic equilibrium or population size grows exponentially. in the case of exponential population growth we may have either vanishing of the infection or an exponentially growing number of infectives. if only susceptibles contribute to the birth rate, as may be expected if the disease is sufficiently debilitating, the behaviour of the model is quite different. let us consider the model which has the same form as the celebrated lotka-volterra predator-prey model of population dynamics. this system has two equilibria, obtained by setting the right sides of each of the equations equal to zero, namely ( , ) and an endemic equilibrium ((µ + α)/β, (r − µ)/β). it turns out that the qualitative analysis approach we have been using is not helpful as the equilibrium ( , ) is unstable and the eigenvalues of the coefficient matrix at the endemic equilibrium have real part zero. in this case the behaviour of the linearization does not necessarily carry over to the full system. however, we can obtain information about the behaviour of the system by a method that begins with the elementary approach of separation of variables for first order differential equations. we begin by taking the quotient of the two differential equations and using the relation to obtain the separable first order differential equation integration gives the relation where c is a constant of integration. this relation shows that the quantity is constant on each orbit (path of a solution in the (s, i− plane). each of these orbits is a closed curve corresponding to a periodic solution. this model is the same as the simple epidemic model of the first section except for the birth and death terms, and in many examples the time scale of the disease is much faster than the time scale of the demographic process. we may view the model as describing an epidemic initially, leaving a susceptible population small enough that infection cannot establish itself. then there is a steady population growth until the number of susceptibles is large enough for an epidemic to recur. during this growth stage the infective population is very small and random effects may wipe out the infection, but the immigration of a small number of infectives will eventually restart the process. as a result, we would expect recurrent epidemics. in fact, bubonic plague epidemics did recur in europe for several hundred years. if we modify the demographic part of the model to assume limited population growth rather than exponential growth in the absence of disease, the effect would be to give behaviour like that of the model studied in the previous section, with an endemic equilibrium that is approached slowly in an oscillatory manner if r > . example. (fox rabies) rabies is a viral infection to which many animals, especially foxes, coyotes, wolves, and rats, are highly susceptible. while dogs are only moderately susceptible, they are the main source of rabies in humans. although deaths of humans from rabies are few, the disease is still of concern because it is invariably fatal. however, the disease is endemic in animals in many parts of the world. a european epidemic of fox rabies thought to have begun in poland in and spread through much of europe has been modeled. we present here a simplified version of a model due to r.m. anderson and coworkers [ ] . we begin with the demographic assumptions that foxes have a birth rate proportional to population size but that infected foxes do not produce offspring (because the disease is highly debilitating), and that there is a natural death rate proportional to population size. experimental data indicate a birth rate of approximately per capita per year and a death rate of approximately . per capita per year, corresponding to a life expectancy of years. the fox population is divided into susceptibles and infectives, and the epidemiological assumptions are that the rate of acquisition of infection is proportional to the number of encounters between susceptibles and infectives. we will assume a contact parameter β = , in rough agreement with observations of frequency of contact in regions where the fox density is approximately fox/km , and we assume that all infected foxes die with a mean infective period of approximately days or / year. these assumptions lead to the model with β = , r = . , µ = . , α = . as this is of the form ( . ), we know that the orbits are closed curves in the (s, i) plane, and that both s and i are periodic functions of t. we illustrate with some simulations obtained using maple (figs. . , . , and . ). it should be noted from the graphs of i in terms of t that the period of the oscillation depends on the amplitude, and thus on the initial conditions, with larger amplitudes corresponding to longer periods. a warning is in order here. the model predicts that for long time intervals the number of infected foxes is extremely small. with such small numbers, the continuous deterministic models we have been using (which assume that population sizes are differentiable functions) are quite inappropriate. if the density of foxes is extremely small an encounter between foxes is a random event, and the number of contacts cannot be described properly by a function of population densities. to describe disease transmission properly when population sizes are very small we would need to use a stochastic model. now let us modify the demographic assumptions by assuming that the birth rate decreases as population size increases. we replace the birth rate of r per susceptible per year by a birth rate of re −as per susceptible per year, with a a positive constant. then, in the absence of infection, the fox population is given by the first order differential equation we omit the verification that the equilibrium n = is unstable while the positive equilibrium n = ( /a) log(r/µ) is asymptotically stable. thus, the population has a carrying capacity given by the model now becomes if βs = µ + α there is an endemic equilibrium with βi + µ = re −as . a straightforward computation, which we shall not carry out here shows, that the disease-free equilibrium is asymptotically stable if r = βk/(µ + α) < and unstable if r > , while the endemic equilibrium, which exists if and only if r > , is always asymptotically stable. another way to express the condition for an endemic equilibrium is to say that the fox population density must exceed a threshold level k t given by with the parameter values we have been using, this gives a threshold fox density of . fox/km . if the fox density is below this threshold value, the fox population will approach its carrying capacity and the disease will die out. above the threshold density, rabies will persist and will regulate the fox population to a level below its carrying capacity. this level may be approached in an oscillatory manner for large r . the epidemic model of kermack we will describe a general age of infection model and carry out a partial analysis; there are many unsolved problems in the analysis. we continue to let s(t) denote the number of susceptibles at time t and r(t) the number of members recovered with immunity, but now we let i * (t) denote the number of infected (but not necessarily infective) members. we make the following assumptions: . the population has a birth rate Λ(n ), and a natural death rate µ giving a carrying capacity k such that Λ(k) = µk, Λ (k) < µ. . an average infected member makes c(n ) contacts in unit time of which s/n are with susceptibles. we define β(n ) = c(n )/n and it is reasonable to assume that β (n ) ≤ , c (n ) ≥ . . b(τ ) is the fraction of infecteds remaining infective if alive when infection age is τ and b µ (τ ) = e −µτ b(τ ) is the fraction of infecteds remaining alive and infected when infection age is τ . letb µ ( ) = ∞ b µ (τ )dτ. ( in previous sections we have used b(τ ) = e −ατ , which would give b µ (τ ) = e −(µ+α)τ . we let i (t) be the number of new infecteds at time t, i(t, τ ) be the number of infecteds at time t with infection age τ , and let φ(t) be the total infectivity at time t. then differentiation of the equation for i * gives three terms, including the rate of new infections and the rate of natural deaths. the third term gives the rate of recovery plus the rate of disease death as thus the si * r model is since i * is determined when s, φ, n are known we have dropped the equation for i * from the model, but it will be convenient to recall if f = then n (t) approaches the limit k, the model is asymptotically autonomous and its dimension may be reduced to two, replacing n by the constant k. we note, for future use, that we define m = ( − f )( − µb µ ( ), and ≤ m ≤ . we note, however, that if f = then m = . we also have, using integration by parts, if a single infective is introduced into a wholly susceptible population, making kβ(k) contacts in unit time, the fraction still infective at infection age τ is b µ (τ ) and the infectivity at infection age τ is a µ (τ ). thus r , the total number of secondary infections caused, is example. (exposed periods) one common example of an age of infection model is a model with an exposed period, during which individuals have been infected but are not yet infected. thus we may think of infected susceptibles going into an exposed class (e), proceeding from the exposed class to the infective class (i) at rate κe and out of the infective class at rate αi. exposed members have infectivity and infective members have infectivity . thus i * = e + i and φ = i. we let u(τ ) be the fraction of infected members with infection age τ who are not yet infective if alive and v(τ ) the fraction of infected members who are infective if alive. then the fraction becoming infective at infection age τ if alive is κu(τ ), and we have the solution of the first of the equations of ( . ) is when we multiply this equation by the integrating factor e ατ and integrate, we obtain the solution and this is the term a µ (τ ) in the general model. the term b(τ ) is u(τ )+v(τ ). thus we have with these choices and the identifications we may verify that the system ( . ) reduces to which is a standard seir model. for some diseases there is an asymptomatic period during which individuals have some infectivity rather than an exposed period. if the infectivity during this period is reduced by a factor ε, then the model can be described by the system this may be considered as an age of infection model with the same identifications of the variables and the same choice of u(τ ), v(τ ) but with a(τ ) = εu(τ ) + v(τ ). there is a disease-free equilibrium s = n = k, φ = of ( . ). endemic equilibria (s, n, φ) are given by if f = the third condition gives Λ(n ) = µn , which implies n = k. then the second condition may be solved for s, after which the first condition may be solved for φ. thus, there is always an endemic equilibrium. if f < the second of the equilibrium conditions gives now substitution of the first two equilibrium conditions into the third gives an equilibrium condition for n , namely then we must have Λ(n ) < µn. however, this would contradict the demographic condition Λ(n ) > µn, < n < k imposed earlier. this shows that if r < there is no endemic equilibrium. if r > for n = the left side of ( . ) is non-negative while the right side is negative. for n = k the left side of ( . ) is µk( − m ) while the right side is this shows that there is an endemic equilibrium solution for n . the linearization of ( . ) at an equilibrium (s, n, φ) is the condition that this linearization has solutions which are constant multiples of e −λτ is that λ satisfies a characteristic equation. the characteristic equation at an equilibrium (s, n, φ) is here, the choice of q(λ) is motivated by the integration by parts formula the characteristic equation then reduces to where p = β(n ) + sβ (n ) ≥ . the characteristic equation for a model consisting of a system of ordinary differential equations is a polynomial equation. now we have a transcendental characteristic equation, but there is a basic theorem that if all roots of the characteristic equation at an equilibrium have negative real part then the equilibrium is asymptotically stable [ , chap. ] . at the disease-free equilibrium s = n = k, φ = the characteristic equation is since the absolute value of the left side of this equation is no greater than kβ(k)Â µ ( ) if λ ≥ the disease-free equilibrium is asymptotically stable if and only if r = kβ(k)Â µ ( ) < . in the analysis of the characteristic equation ( . ) it is helpful to make use of the following elementary result: if |p (λ)| ≤ , g(λ) > for λ ≥ , then all roots of the characteristic equation p (λ) = + g(λ) satisfy λ < . to prove this result, we observe that if λ ≥ the left side of the characteristic equation has absolute value at most while the right side has absolute value greater than . if f = , the characteristic equation reduces to we have |sβ(n )Â µ (λ)| ≤ s β (n )Â µ ( ) = the term φβ(n ) λ + µ in ( . ) has positive real part if λ ≥ . it follows from the above elementary result that all roots satisfy λ < , so that the endemic equilibrium is asymptotically stable. thus all roots of the characteristic equation ( . ) have negative real part if f = . the analysis if f < is more difficult. the roots of the characteristic equation depend continuously on the parameters of the equation. in order to have a root with λ ≥ there must be parameter values for which either there is a root at "infinity", or there is a root λ = or there is a pair of pure imaginary roots λ = ±iy, y > . since the left side of ( . ) approaches while the right side approaches as λ → ∞, λ ≥ , it is not possible for a root to appear at "infinity". for λ = , since sβ(n )Â µ ( ) = and β (n ) ≤ the left side of ( . ) is less than at λ = , while the right side is greater than since − Λ (n )b µ ( ) > − Λ (n )/µ > if Λ (n ) < µ. this shows that λ = is not a root of ( . ), and therefore that all roots satisfy λ < unless there is a pair of roots λ = ±iy, y > . according to the hopf bifurcation theorem [ ] a pair of roots λ = ±iy, y > indicates that the system ( . ) has an asymptotically stable periodic solution and there are sustained oscillations of the system. a somewhat complicated calculation using the fact that since b µ (τ ) is monotone non-increasing, in ( . ) has negative real part for some y > . this is not possible with mass action incidence, since β (n ) = ; thus with mass action incidence the endemic equilibrium of ( . ) is always asymptotically stable. since β (n ) ≤ , instability requires b µ (iy) = there are certainly less restrictive conditions which guarantee asymptotic stability. however, examples have been given [ , ] of instability, even with f = , Λ (n ) = , where constant infectivity would have produced asymptotic stability. their results indicate that concentration of infectivity early in the infected period is conducive to such instability. in these examples, the instability arises because a root of the characteristic equation crosses the imaginary axis as parameters of the model change, giving a pure imaginary root of the characteristic equation. this translates into oscillatory solutions of the model. thus infectivity which depends on infection age can cause instability and sustained oscillations. in order to formulate an si * s age of infection model we need only take the si * r age of infection model ( . ) and move the recovery term from the equation for r (which was not listed explicitly in the model) to the equation for s. we obtain the model although we will not carry out any analysis of this model, it may be attacked using the same approach as that used for ( . ) . it may be shown that if r = kβ(k)Â µ ( ) < the disease-free equilibrium is asymptotically stable. if r > there is an endemic equilibrium and the characteristic equation at this equilibrium is where p = β(n ) + sβ (n ) ≥ . many diseases, including most strains of influenza, impart only temporary immunity against reinfection on recovery. such disease may be described by sis age of infection models, thinking of the infected class i * as comprised of the infective class i together with the recovered and immune class r. in this way, members of r neither spread or acquire infection. we assume that immunity is lost at a proportional rate κ. we let u(τ ) be the fraction of infected members with infection age τ who are infective if alive and v(τ ) the fraction of infected members who are not recovered and still immune if alive. then the fraction becoming immune at infection age τ if alive is αu(τ ), and we have these equations are the same as ( . ) obtained in formulating the seir model with α and κ interchanged. thus we may solve to obtain we take b(τ ) = u(τ )+v(τ ), a(τ ) = u(τ ). then if we define i = φ, r = i * −φ, the model ( . ) is equivalent to the system which is a standard sirs model. if we assume that, instead of an exponentially distributed immune period, that there is an immune period of fixed length ω we would again obtain u(τ ) = e −ατ , but now we may calculate that v(τ ) = − e −ατ , (τ ≤ ω), v(τ ) = e −ατ (e αω − ), (τ > ω) . to obtain this, we note that v (τ ) = αu(τ ), (τ ≤ ω), v (τ ) = αu(τ ) − αu(τ − ω), (τ > ω) . from these we may calculate a(τ ), b(τ ) for an si * s model. since it is known that the endemic equilibrium for an sirs model with a fixed removed period can be unstable [ ] , this shows that ( . ) may have roots with non-negative real part and the endemic equilibrium of an si * s age of infection model is not necessarily asymptotically stable. the si * r age of infection model is actually a special case of the si * s age of infection model. we could view the class r as still infected but having no infectivity, so that v(τ ) = . the underlying idea is that in infection age models we divide the population into members who may become infected and members who can not become infected, either because they are already infected or because they are immune. we conclude by returning to the beginning, namely an infection age epidemic model closely related to the original kermack-mckendrick epidemic model [ ] . we simply remove the birth and natural death terms from the si * r model ( . ) . the result is and this shows that s ∞ > . we recall that we are assuming here that β( ) is finite; in other words we are ruling out standard incidence. it is possible to show that s ∞ can be zero only if n → and k β(n )dn diverges. however, from ( . ) we see that this is possible only if f = . if there are no disease deaths, so that the total population size n is constant, or if β is constant (mass action incidence), the above integration gives the final size relation log we may view the epidemic management model ( . ) as an age of infection model. we define i * = e + q + i + j, and we need only calculate the kernels a(τ ), b(τ ). we let u(τ ) denote the number of members of infection age τ in e, v(τ ) the number of members of infection age τ in q, w(τ ) the number of members of infection age τ in i, and z(τ ) the number of members of infection age τ in j. then (u, v, w, z) satisfies the linear homogeneous system with constant coefficient with initial conditions u( ) = , v( ) = , w( ) = , z( ) = . this system is easily solved recursively, and then the system ( . ) is an age of infection epidemic model with a(τ ) = ε e u(τ )+ε e ε q v(τ )+w(τ )+ε j z(τ ), b(τ ) = u(τ )+v(τ )+w(τ )+z(τ ) . in particular, it now follows from the argument carried out just above that s ∞ > for the model ( . ). the proof is less complicated technically than the proof obtained for the specific model ( . ). the generalization to age of infection models both unifies the theory and makes some calculations less complicated. population dynamics of fox rabies in europe infectious diseases of humans. oxford science publications mathematical models in population biology and epidemiology vertically transmitted diseases: models and dynamics. biomathematics mathematical approaches for emerging and reemerging infectious diseases: an introduction mathematical approaches for emerging and reemerging infectious diseases: models, methods and theory the role of long incubation periods in the dynamics of hiv/aids. part : single populations models asymptotically autonomous epidemic models epidemic modelling: an introduction overall patterns in the transmission cycle of infectious disease agents the first epidemic model: a historical note on p.d. en'ko detecting nonlinearity and chaos in epidemic data singapore and beijing experience the saturating contact rate in marriage and epidemic models qualitative analysis for communicable disease models an immunization model for a hetereogeneous population the mathematics of infectious diseases periodicity in epidemic models periodicity and stability in epidemic models: a survey abzweigung einer periodischen lösungen von einer stationaren lösung eines differentialsystems mckendrick: a contribution to the mathematical theory of epidemics mckendrick: contributions to the mathematical theory of epidemics, part. ii mckendrick: contributions to the mathematical theory of epidemics, part. iii asymptotically autonomous differential systems dynamic models of infectious diseases as regulators of population size plagues and peoples the global condition network theory and sars: predicting outbreak diversity epidemic models: their structure and relation to data the structure and function of complex networks raggett: modeling the eyam plague interpretation of periodicity in disease prevalence exploring complex networks asymptotically autonomous differential equations in the plane mathematics in population biology mathematical and statistical approaches to aids epidemiology how may infection-age dependent infectivity affect the dynamics of hiv/aids? reproduction numbers and sub-threshold endemic equilibria for compartmental models of disease transmission theory of nonlinear age-dependent population dynamics. marcel dekker key: cord- -i ecxgus authors: nan title: abstracts of publications related to qasr date: - - journal: nan doi: . /qsar. sha: doc_id: cord_uid: i ecxgus nan tive mechanisms p. - . edited by magee, p.s., henry, d.r., block, j.h., american chemical society, washington, . results: an overview is given on the approaches for the discovery and design concepts of bioactive molecules: a) natural products derived from plant extracts and their chemically modified derivatives (cardiac glycosides, atropine, cocaine, penicillins, cephalosporins, tetracyclines and actinomycins, pyrethrins and cyclosporin; b) biochemically active molecules and their synthetic derivatives: acetylcholine, histamine, cortisonelhydrocortisone, indole- -acetic acid (phenoxyacetic acid herbicides); c) principles of selective toxicity is discussed exemplified by trimethoprimlmethotrexate, tetracyclines, acylovir, azidothymidine, antifungal agents; d) metabolism of xenobiotics; e) exploitation of secondary effects (serendipity); f) receptor mapping; g) quantitative structure-activity relationship studies; h) empirical screening (shotgun approach). results: past and present of qsar is overviewed: a) historical roots; b) the role of qsar models in rational drug design, together with a simplified diagram of the steps involved in drug development, including the place of qsar investigations; c) classification of qsar models: structure-cryptic (property-activity) models, structure-implicit (quantum chemical) models, structure-explicit (structure-activity) and structure-graphics (computer graphics) models; d) a non-empirical qsar model based on quantities introduced for identification of chemical structures, using szymansk's and randic's identification (id) numbers, including applications for alkyltriazines. bioessays, , ( ) , - . results: a review is given on recent observations about receptor structure and the dynamic nature of drug receptors and the significance of receptor dynamics for drug design: a) receptors are classified according to structure and function (i) ion channels (nicotinic acetylcholine, gaba, glycine); (ii) g protein linked [adrenergic ( c x ,~) , muscarinic acetylcholine, angiotensin, substance k, rhodopsin]; (iii) tyrosine kinase (insulin, igf, egf, pdgf); (iv) guanylate cyclase (atrial natriuretic peptide, speractin); b) protein conformational changes can be best studied on allosteric proteins whose crystal structure is available (e.g. hemoglobin, aspartate transcarbamylase, tryptophan repressor) (no high resolution of a receptor structure is known); c) receptor conformational changes can be studied by several indirect approaches (i) spectral properties of covalent or reversibly bound fluorescent reporter groups; (ii) the sensitivity of the receptor to various enzymes; (iii) the sedimentation of chromatographic properties of the receptor; the affinity of binding of radioligands; (iv) the functional state of the receptor; d) there are many unanswered questions: e.g. (i) are there relatively few conformational states for receptors with fluctuations around them or many stable conformational states; (ii) how can static structural information be used in drug design when multiple receptor conformations exist. title: designing molecules and crystals by computer. (review) author: koide, a. ibm japan limited, tokyo scientific center, tokyo research laboratory - sanban-cho, chiyoda-ku, tokyo , japan. source: ibm systems journal , ( ), - . results: an overview is given on three computer aided design (cad) systems developed by ibm tokyo scientific center: a) molecular design support system providing a strategic combination of simulation programs for industrial research and development optimizing computational time involved and the depth of the resulting information; b) molworld on ibm personal systems intended to create an intelligent visual environment for rapidly building energetically stable d molecular geometries for further simulation study; c) molecular orbital graphics system designed to run on ibm mainframe computers offering highly interactive visualization environment for molecular electronic structures; d) the systems allow interactive data communication among the simulation programs for their strategically combined use; e) the structure and functions of molworld is illustrated on modeling the alanine molecule: (i) data model of molecular structures; (ii) chemical formula input; (iii) generation of d molecular structure; (iv) formulation of bonding model; (v) interactive molecular orbital graphics; (vi) methods of visualizing electronic structures; (vii) use of molecular orbital graphics for chemical reactions. title: interfacing statistics, quantum chemistry, and molecular modeling. (review) author: magee, p.s. biosar research project vallejo ca , usa. source: acs symposium series , no. . in: probing bioactive mechanisms p. - . edited by magee, p.s., henry, d.r., block, j.h., american chemical society, washington, . results: a review is given on the application and overlap of quantum chemical, classical modeling and statistical approaches for the quant. struct.-act. relat. , - ( ) abstr. - understanding of binding events at the molecular level. a new com-a) qsar of cns drugs has been systematically discussed according plementary method called statistical docking experiment is also to the following classes: (i) general (nonspecific) cns depressants: presented: general anesthetics; hypnotics and sedatives; (ii) general insights obtained using energy-minimized structures; activation in the bound state, types and energies of interactions at the receptor site and in crystal; four successful examples (significant regression equations) are given for the modeling of binding events using physico-chemical descriptors and correlation analysis: (i) binding of a diverse set of pyridines to silica gel during thin-layer chromatography; (ii) binding of meta-substituted n-methyl-arylcarbamates to bovine erythrocyte ache; (iii) binding of meta-substituted n-methyl-arylcarbamates to ache obtained from susceptible and resistant green rice leafhoppers; (iv) activity of phenols inhibiting oxidative phosphorylation of adp to atp in yeast; a new statistical method for mapping of binding sites has been developed based on the hypermolecule approach, identifying key positions of binding and nature of the energy exchange between the hypermolecule atoms and the receptor site; two examples are given on the successful application of statistical modeling (statistical docking experiment) based on the hypermolecule approach: (i) inhibition of housefly head ache by metasubstituted n-methyl-arylcarbamates (n = , r = . , s = . , f = . ); (ii) inhibition of housefly head ache by orthosubstituted n-methyl-arylcarbamates (n = , r = . , s = . , f = . ). (nonspecific) cns stimulants; (iii) selective modifiers of cns functions: anticonvulsants, antiparkinsonism drugs, analgetics and psychopharmacological agents; (iv) miscellaneous: drugs interacting with central a-adrenoreceptors, drugs interacting with histamine receptors, cholinergic and anticholinergic drugs; b) the review indicates that the fundamental property of the molecules which mostly influence the activity of cns drugs is hydrophobicity (they have to pass the cell membrane and the bloodbrain barrier); c) electronic parameters, indicative of dipole-dipole or charge-dipole interactions, charge-transfer phenomena, hydrogen-bond formation, are another important factor governing the activity of most cns agents; d) topographical, lipophylic and electronic structures of cns pharmacophores are reviewed; e) qsar equations, tables and figures from references are shown and discussed. the relevant template for each atom in the molecule is mapped into a bit array and the appropriate atomic position is marked; volume comparisons (e.g. common volume or excluded volume) are made by bit-wise boolean operations; the algorithm for the visualization of the molecular surface comprising the calculated van der walls volume is given; comparisons of cpu times required for the calculation of the van der waals molecular volumes of various compounds using the methods of stouch and jurs, pearlman, gavazotti and the new method showed that similar or better results can be achieved using the new algorithm with vax-class computers on molecules containing up to several hundred atoms. abtstr. - quant. struct.-act. relat. , - ( ) one of the important goal of protein engineering is the design of isosteric analogues of proteins; major software packages are available for molecular modeling are among others developed by (i) biodesign, inc., pasadena, california; (ii) biosym technologies, san diego, california; (iii) tripos, st. louis, missouri; (iv) polygen, waltham, massachusetts; (v) chemical design ltd. oxford; the molecular modelling packages use three basic parameters: (i) descriptive energy field; (ii)algorithm for performing molecular mechanics calculations; (iii) algorithm for performing molecular dynamics calculations; modelling study of the binding events occurring between the envelop protein (gp ) of the aids (hiv) virus and its cellular receptor (cd ) protein supported the hypothesis that this domain was directly involved in binding the gp envelop protein leading to the design of conformationally restricted synthetic peptides binding to cd . title: finding washington, . results: a new technique called "homology graphing" has been developed for the analysis of sequence-function relationships in proteins which can be used for sequence based drug design and the search for lead structures: a) as target protein is inhibited by the ligands of other proteins having sequence similarity, computer programs have been developed for the search of the similarity of proteins; b) proteins are organized into hierarchical groups of families and superfamilies based on their global sequence similarities; c) global sequence similarities were used to find inhibitors of acetolactate synthase (als) and resulted in a quinone derivative as a lead structure of new als inhibitors; d) local sequence similarities of bacterial and mammal glutathione synthase (gsh) were used to find inhibitors of gsh; e) it was shown that the sequence segment of gsh was similar to dihydrofolate reductase (dhfr) is part of the atp-binding site; f) biological bases of local similarity between sequences of different proteins were indicated: molecular evolution of proteins and functionally important local regions; g) homology graph, as a measure of sequence similarity was defined; h) sequence-chemical structure relationship based on homology graph and the procedure to find lead structures was illustrated by an example resulting in a list of potential inhibitors selected by the procedure based on the sequence segment from residue to of the sequence of tobacco als. source: acs symposium series , no. . in: probing bioactive mechanisms p. - . edited by magee, p.s.. henry, d.r., block, j.h., american chemical society, washington. . results: a review is given on the molecular design of the following major types of antifungal compound in relation to biochemistry, molecular modeling and target site fit: a) squalene epoxidase inhibitors (allilamines and thiocarbanilates) blocking conversion of squalene , -oxidosqualene; b) inhibitors of sterol c- demethylation by cytochrome p- (piperazines pyridines, pyrimidines, imidazoles and triazoles); c) inhibitors of sterol a' -t a' isornerization andlor sterol reductase inhibitors (morpholines); d) benzimidazoles specifically interfering with the formation of microtubules and the activity phenylcarbamates on benzimidazole resistant strains; e) carboxamides specifically blocking the membrane bound succinate ubiquinone oxidoreductase activity in the mitochondria electron transport chain in basidiomycetes; f) melanin biosynthesis inhibitors selectively interfering with the polyketide pathway to melanin in pyricularia oryzae by blocking nadph dependent reductase reactions of the pathway (fthalide, pcba, chlobentiazone, tricyclazole, pyroquilon, pp ). title: quantitative modeling of soil sorption for xenobiotic chemicals. (review) author: sabljic, a. theoretical chemistry group, department of physical chemistry, institute rudjer boskovic hpob , yu- zagreb, croatia, yugoslavia. source: environ. health perspect. , ( ), - . results: the environmental fate of organic pollutants depends strongly on their distribution between different environmental compartments. a review is given on modeling the soil sorption behavior of xenobiotic chemicals: a) distribution of xenobiotic chemicals in the environment and principles of its statistical modeling; b) quantitative structure-activity relationship (qsar) models relating chemical, biological or environmental activity of the pollutants to their structural descriptors or physico-chemical properties such as logp values and water solubilities; c) analysis of the qsar existing models showed (i) low precision of water solubility and logp data; (ii) violations of some basic statistical laws; d) molecular connectivity model has proved to be the most successful structural parameter modeling soil sorption; e) highly significant linear regression equations are cited between k , values and the first order molecular connectivity index ( ' x ) of a wide range of organic pollutants such as polycyclic aromatic hydrocarbons (pahs) and pesticides (organic phosphates, triazines, acetanilides, uracils, carbamates, etc.) with r values ranging from . to . and s values ranging from . to . ; f) the molecular connectivity model was extended by the addition of a single semiempirical variable (polarity correction factor) resulting in a highly significant linear regression equations between the calculated and measured ko, values of the total set of compounds (n = , r = . , s = . , f = ); g) molecular surface areas and the polarity of the compounds were found to be responsible for the majority of the variance in the soil sorption data of a set of structurally diverse compounds. title: strategies for the use of computational sar methods in assessing genotoxicity. (review) results: a review is given on the overall strategy and computational sar methods for the evaluation of the potential health effects of chemicals. the main features of this strategy are discussed as follows: a) generalized sar model outlining the strategy of developing information for the structure-activity assessment of the potential biological effects of a chemical or a class of chemicals; b) models for predicting health effects taking into account a multitude of possible mechanisms: c) theoretical models for the mechanism of the key steps of differential activity at the molecular level; d) sar strategies using linear-free energy methods such as the hansch approach; e) correlative sar methods using multivariate techniques for descriptor generation and an empirical analysis of data sets with large number of variables (simca, adapt, topkat, case, etc.); f) data base considerations describing three major peer-reviewed genetic toxicology data bases (i) national toxicology program (ntp) containing short term in vitro and in vivo genetic tests; (ii) data base developed by the epa gene-tox program containing different short term bioassays for more than compounds, used in conjunction with adapt, case and topkat; (iii) genetic activity profile (gap) in form of bar graphs displaying information on various tests using a given chemical. title: quantitative structure-activity relationships. principles, and authors: benigni,, r.; andreoli, c.; giuliani, a. applications to mutagenicity and carcinogenicity. (review) laboratory of toxicology and ecotoxicology, istituto superiore di sanita rome, italy. source: mutat. res. , ( ), - . results: methods developed for the investigation for the relationships between structure and toxic effects of compounds are summarized: a) the extra-thermodynamic approach: the hansch paradigm, physical chemical properties that influence biological activity and their parametrization, originality of the hansch approach, receptors and pharmacophores: the natural content of the hansch approach, predictive value of qsars, a statistifa tool: multiple linear regression analysis, the problem of correlations among molecular descriptors, other mathematical utilizations of extrathermodynamic parameters; b) the substructural approach: when topological (substructural) descriptors are needed, how to use topological decriptors; c) qsar in mutagenicity and carcinogenicity: general problems, specific versions of the substructural approach used for mutagenicity and carcinogenicity, applications to mutagenicity and carcinogenicity. title: linking structure and data. (review) author: bawden, d. source: chem. britain , (nov) , i - . address not given. results: the integration of information from different sources, particularly linking structural with non-structural information is an important consideration in chemical information technology. a review is given on integrated systems: a) socrates chemicallbiological data system for chemical structure and substructure searching combined with the retrieval of biological and physicochemical data, compound availability, testing history, etc.; b) psidom suite of pc based structure handling routines combining chemical structure with the retrieval of text and data; c) cambridge crystal structure databank on x-ray data of organic compounds integrating information on chemical structure, crystal conformation, numerical information on structure determination, bibliographic reference and keywording; d) computer aided organic synthesis for structure and substructure search, reaction retrieval, synthetic analysis and planning, stereochemical analysis, product prediction and thermal hazard analysis. title: determination of three-dimensional structures of proteins and nucleic acids in solution by nuclear magnetic resonance spectroscopy. source: critical rev. biochem. mol. biol. , ( ) , - . results: a comprehensive review is given on the use of nmr spectroscopy for the determination of d structures of proteins and nucleic acids in solution discussing the following subjects: a) theoretical basis of two-dimensional ( d) nmr and the nuclear overhauser effect (noe) measurements for the determination of d structures is given; b) sequential resonance assignment for identifying spin systems of protein nmr spectra and nucleic acid spectra, selective isotope labeling for extension to larger systems and the use of site specific mutagenesis; c) measurement and calculation of structural restraints of the molecules (i) interproton distances; (ii) torsion angle restrains; (iii) backbone torsion angle restraints; (iv) side chain torsion angle restraints; (v) stereospecific assignments; (vi) dihedral angle restraints in nucleic acids; d) determination of secondary structure in proteins; e) determination of tertiary structure in proteins using (i) metric matrix distance geometry (ii) minimization in torsion angle space; (iii) restrained molecular dynamics; (iv) dynamical simulated annealing; (v) folding an extended strand by dynamical simulated annealing; (vi) hybrid metric matrix distance geometry-dynamical simulated annealing method; (vii) dynamical simulated annealing starting from a random array of atoms; f) evaluation of the quality of structures generated from nmr data illustrated by studies for the structure determination of proteins and oligonucleotides using various algorithms and computer programs; g) comparisons of solution and x-ray structures of (i) globular proteins; (ii) related proteins; (iii) nonglobular proteins and polypeptides; h) evaluation of attainable precision of the determination of solution structures of proteins for which no x-ray structures exist (i) bds-i (small -residue protein from the sea anemone sulcata; (ii) hirudin (small -residue protein from leech which is a potent natural inhibitor of coagulation); i) structure determination by nmr is the starting point for the investigation of the dynamics of conformational changes upon ligand abtstr. - quant. struct.-act. relat. , - ( ) binding, unfolding kinetics, conformational equilibria between different conformational states, fast and slow internal dynamics and other phenomena. title: aladdin. an integrated tool for computer-assisted molecular design and pharmacophore recognition from geometric, steric, and substructure searching of three-dimensional molecular structures. ( aladdin has the ability to (i) objectively describe receptor map hypothesis; (ii) scan a database to retrieve untested compounds which is predicted to be active by a receptor map hypothesis; (iii) quantitatively compare receptor map hypotheses for the same biological activity; (iv) design compounds that probe the bioactive conformation of a flexible ligand; (v) design new compounds that a receptor map hypothesis predicts to be active; (vi) design compounds based on structures from protein x-ray crystallography; a search made by aladdin in a database for molecules that should have d dopaminergic activity recognized unexpected d dopamine agonist activity of existing molecules; a comparison of two superposition rules for d agonists was performed by aladdin resulted in a clear discrimination between active and inactive compounds; a compound set was designed that match each of the three lowenergy conformations of dopamine resulting in novel active analogues of known compounds; mimics of some sidc ~ . . p.piide beta turns were designed, in order to demonstrate that aladdin can find small molecules that match a portion of a peptide chain and/or backbone; results: lately a number of chemical information systems based on three-dimensional ( -d) molecular structures have been developed and used in many laboratories: a) concord uses empirical rules and simplified energy minimization to rapidly generate approximate but usually highly accurate -d molecular structures from chemical notation or molecular connection table input; b) chemical abstracts service (cas) has added -d coordinates for some million organic substances to the cas registry file; c) cambridge structural database system contains x-ray and neutron diffraction crystal structures for tens of thousands of compounds; d) maccs d developed by molecular design ltd., contains the standard maccs-i structures to which additional -d data, such as cartesian coordinates, partial atomic charges and molecular mechanics energy are added; maccs d allows exact match, geometric, submodel and substructure searching of -d models with geometric constrains specified to certain degree of tolerance; two -d databases are also available from molecular design that can be searched using maccs d [drug data report ( , models) and fine chemicals directory ( , models)]; e) aladdin (daylight chemical information systems) is also searches databases of -d structures to find compounds that meet biological, substructural and geometric criteria such as ranges of distances, angles defined by three points (dihedral angles) and plane angles that the geometric object must match. aladdin is one of a number of menus working within the framework provided by daylight's chemical information system. title: improved access to supercomputers boosts chemical applica-author: borman, s. c&en sixteenth st., n.w., washington dc , usa. source: c&en , ( ) , - . results: supercomputers have been much more accessible by scientists and engineers in the past few years in part as a result of the establishment of national science foundation (nsf) supercomputer centers. the most powerful class of supercomputers have program execution rates of million to billion floating-point operations per second, memory storage capacities of some ten million to miltion computer words and a standard digital word size of bits, the equivalent of about decimal digits. the following examples are given for the use of supercomputer resources for chemical calculations and modeling: a) modeling of key chromophores in the photosynthetic reaction center of rhodopseudomonas viridis showing the heme group, the iron atom and the chlorophyll which absorbs light and causes rapid transfer of electron to pheophtin and then to the quinone; modeling includes a significant part of the protein having about atoms out of a total of some , ; quant. struct.-act. relat. , - ( ) abstr. - b) modeling of transition state of reaction between chloride and methyl chloride including electron clouds and water molecules surrounding the reaction site; c) analysis of nucleic acid and protein sequences to evaluate the secondary structure of these biopolymers; d) construction of a graphical image of hexafluoropropylene oxide dimer, a model for dupont krytox high performance lubricant; e) calculation of the heats of formation of diaminobenzene isomers indicated that the target para isomer was kcal/mol less stable then the meta isomer byproduct therefore the development for its large scale catalytic synthesis was not undertaken (saving was estimated to be $ to $ million). , b) comparison of the newly defined eo, parameter with the taft-kutter-hansch e, (tkh e,) parameter showed characteristic steric effects of ortho-alkoxy and n-bonded planar type substituents (e.g. no,, ph); c) in various correlation analyses using retrospective data eo, satisfactorily represented the steric effects of ortho-substituents on reactivity and biological activity of various organic compounds; d) semi-empirical am calculations using a hydrocarbon model to study the steric effects of a number of ortho-substituents resulted in the calculation of the es value (difference in the heat of formation between ortho-substituted toluene and t-butylbenzene) which linearly correlated with the eo, and the tkh e, parameters; e) effects of di-ortho substitution on lipophilicity could be mostly expressed by the summed effect of the -and -position substituents; t) highly significant regression equations were calculated for the pk, values of di-ortho-substituted benzoic acids using various substituent parameters; g) quantitative analysis of the effect of ortho-substitution is difficult because it is a result of overlapping steric and electronic effects. title: calculation of partition coefficient of n-bridgehead com- ( i i ) is more lipophilic than propanolol- -sulphate (iv)]. fig. shows the relationship between lipophilicity and ph for the compounds (circle represents (i), triangle (ii), rhomboid (m) and square gv - ( ) abstr. - f (rekker's constant, characterizing hydrophobicity). results: a good agreement was found between the observed and calculated logp values of i ( . and . , respectively) and for iii. the hydrophobicity of i was found to be significantly lower than that of i ( . and . , respectively) . the large deviation was attributed to the surface reduction as a result of condensed ring formation in i. since interesting pharmacological activities have been reported for several derivatives of this type of compounds, the hydrophobicity of the unsubstituted lh-indolo[ , -c]quinoline has been calculated to be . : ( ) [interaction energy between a molecule and the binding site model was assumed to be the sum of its atomic contributions according to the expres-e , . , , ,~~(~) was the interaction energy parameter between the site region rand the atom-type of atom a and ag(b) was the total interaction energy for the binding mode b (binding mode was regarded as feasible when the molecule was in its energetically most favorable conformation)]. sion ag(b) = erelion reatomi a in r er,typc(a). where results: for development of the binding site model, first a simple geometry was proposed and agm- agm,calc i agm+) was calculated for the whole set of compounds. if the calculated binding energy of any of the compounds was outside of the above boundary, the proposed site geometry was rejected and a more complex one was considered. this procedure had been repeated until all molecules in the set could be fitted within the experimental data range. as a result a d, five-region voronoi binding site model has been developed for the pahs containing a trigonal pyramid (rl) in the center and portions r rs having infinite volumes and indicated by boundary planes. region r, represented access to the solvent and regions r rs were blocked for binding ( fig. ) : pyrene is shown in its optimal binding mode with its atom barely touching the boundary surfaces and edges: calculations showed that benzene and other monoaromatic ring compounds should be very weak competitors for the b[a]p site. the model correctly predicted the binding energy of nine competitors outside of the training set. '% (wiener index calculated as the sum of all unique shortest distances between atoms in the hydrogen suppressed graph of the compound); (wiener index calculated as the sum of all geometric distances between atoms in the hydrogen suppressed molecule of the compound). results: the traditional d wiener number is defined as the sum of the lengths of all possible routes in the molecular graph. here the length is proposed to be calculated as the real three-dimensional length between atoms: this is the d wiener number. this number has many of the advantageous features of the related and very much studied d wiener number. additionally, it is highly discriminative and its use in quantitative structure-property relation studies (qspr) appears to be encouraging, according to the preliminary calculations. of these the most convincing is the set of statistical parameters for the linear correlation between the experimental and calculated enthalpy functions of dw the lower alkanes not shown here. three different models have been tried and in all cases the d wiener number seemed to be superior to the d one as it is reflected in (eqs. - ). a) gaba receptors in human mouse, rat and bovine brain tissues, membrane preparations and cellular uptake systems; b) gaba receptors in cat and rat spinal cord preparations; c) cultured astrocytes. as in the equations nearly all indicator variables had negative regression coefficients it was concluded that instead of searching for better analogs, the research should be directed toward degradable pro-gaba or pro-muscimol derivatives that are efficiently taken up into the central nervous system (cns). . (+ . ) irng + . ( ) title: synthesis and qsar of -aryl- -(~- -quinolyi/l-isoqui-noly ethyl)piperazines and some related compounds as hypotensive agents. authors ( ) based on eq. , an optimal logp is predicted (logpo = . ). the highest activity was produced by the -( -methylphenyl)- -(~- -qui-data determined: chemical descriptors: abtstr. - quant. struct.-act. relat. , - ( ) nolylethyl) piperazine, its logp value being near to the optimal value ( . ). l.og(bph) values calculated by eq. agree well with the observed ones. source: toxicology , ( ), - . compounds: , -dimethoxyphenol, -chlorophenol, . -dichlorophenol, -methyl- -nitropheno , , dichlorophenol, , , -trichlorophenol, , , , -tetrachlorophenol, , , -triiodophenol, pentachlorophenol. biological material: chinese hamster ovary (cho) cells. data taken from the literature: ezoc; ecsoc; eczoa; ec~oa [concentration (mmol l) of the compound leading to a or % inhibition of the cell growth or adenosine uptake, respectively]. data determined: ego; ecso [concentration (mmol l) of the compound leading to a or % inhibition of the na+/k+-atpase activity, respectively]. chemical descriptors: logp (logarithm of the partition coefficient in i-octanollwater); u (hammett's constant, characterizing the electron-withdrawing power of the substituent); e, (taft's constant, characterizing steric effects of the substituent); x (molecular connectivity index, calculated by koch's method). results: highly significant linear relationships were calculated between log (eczo) and logp (r = - . ). the relationship between log(ec,,) and u being less good (r = - . ). combining the two parameters the relationship has improved (eq. i): ( ) (logarithm of the partition coefficient in i-octanollwater); (hansch-fujita's substituent constant characterizing hydrophobicity); (hammett's constant, characterizing the electron-withdrawing power of the substituent); (sterimol steric parameter, characterizing the steric effect of the meta substituents); (rplc derived hydrophobic substituent constant, defined by chen and horv th, and extrapolated to x methanol); (indicator variable for the present for the absence of hydrogen bonding substituents). results: logk' values were determined for the benzenesulfonamides and correlated with chemical descriptors. a highly significant linear relationship between logk' and logp was calculated (eq. ): ( pk, (negative logarithm of the acidic dissociation constant); logp (logarithm of the partition coefficient in i-octanol/water). results: relationships between ki values and the chemical descriptors were investigated for cpz and its listed metabolites. relationship between log(l/ki) and logp was calculated (eq. ) no numerical intercept (c) is given: ( in spite of the complexity of the full mechanism of inhibition involving at least six transition states and five distinct intermediates, a significant linear regression equation was calculated for ki (eq. ): since the crystal structure of the acyl-enzyme complex, the acylation and deacylation rate were available, it was concluded that the inhibition begins with the histidine catalyzed attack of serine , at the benzoxazinone c , while the carbonyl oxygen occupies the oxyanion hole formed by glycine and serine . title: antifolate and antibacterial activities of -substituted authors: harris, n.v.; smith, c.; bowden, k. rhone results: it was shown earlier that binding of diaminoquinazolines to dhfr correlated with the torsional angle of the -amino group of the quinazoline nucleus. it was postulated that the interaction between the adjacent -substituent and the -amino group was very important in determining dhfr binding of the compounds possibly, because of the influence on the hydrogen-bond formed between the -amino group and a residue at the active site. the existence of such interaction in -substituted , -diaminoquinazolines were shown by measuring a , , and ~~ values. the ui and uor electronic parameters correlated well with chemical shifts of the -nh, groups (eq. ) but showed poor correlation for the -nh, group (eq. ), respectively: ( ) the equations suggest that the through-ring resonance interactions between the -substituent and the adjacent -amino group are disrupted by some other effects which might have significance for binding. a) an extensive set of compounds based on the nalidixic acid structure of type i. where r', r , r and r are various substituents; x and x* = c, n (for nalidixic acid: x = c, x = n, r' = et, r' = cooh. r = h, r = me); b) subset of (i) (set a) containing fifty two , -disubstituted -alky l- , -dihydro- -oxoquinoline- -carboxylic acids; compounds: abtstr. - quant. struct.-act. relat. , - ( ) c) subset of (i) (set b) containing one hundred and sixty two xylic acids; d) subset of (i) (set c) containing eighty five , -dihydr - -oxo- , -naphthyridine- -carboxylic acids with substituted azetidinyl, pyrrolidinyl and piperidinyl rings at position , fluorine at position and ethyl, vinyl or -fluoroethyl substituent at position . biological material: ps. aeruginosa v- , e. coli nihj jc- , s. the study showed that the most active compounds have fluorine in position , r can be a wide variety of nitrogen containing substituent and the best predictor for r is its lipophilicity. compounds: phytoalexins: pisatin, , a-dihydroxy- , -(methylenedioxy)pterocarpan, a, la-dehydropisatin, -hydroxy- , -(methylenedioxy)- a, a-dehydropterocarpan, (*)- -hydroxy- -methoxypterocarpan, (+)- -hydroxy- -zmethoxypterocarpan, (-)- -zhydroxy- -methoxypterocarpan, vestitol, sativan, formonenetin, coumestrol, '-o-methylcoumestro , phaseoilin, phaseollinisoflavan, '-methoxyphaseollin-isoflavan, glyceollin, a- a-dehydroglyceollin, tuberosin, a, ladehydrotuberosin. (capacity factor determined rp-hplc). calculated for logp of six reference compounds using their k' values (eq. ): ( ) n = r = . s not given f not given the lipophilicity of the phytoalexins were within the range of log p = . - . . it was found that the antifungal activity of similar compounds positively correlated with antifungal activity but no equation could be calculated for the whole set of compounds. it was suggested, however, that compounds with logp values higher than . were retained in the membranes, therefore phytoalexins with slightly lower lipophilicity, as well as greater fungitoxicity and systemic activity should be searched. certain structural features seemed to correlate with antifungal activity such as the presence of phenolic oh and benzylic hydrogen. it was suggested that the ability of the ortho oh group to form fairly stable intramolecular hydrogen bond may contribute to the greater stability of the shiff base hnctional group and the higher biological activity of the substances (various subsets required different equations). results showed that compounds with increasing lipophilicity and electron donating substituents at the -and -positions have high inhibitory activity. i-[( '-allyl- '-hydroxybenzilidene)amino]- -hydroxyguanidine was found to be the most active compound. the use of parameter focusing of the substituent hydrophobic constant and electronic constants was suggested for the selection of further substituents to design effective compounds. biological material: a) rabbits; b) rats; c) guinea pig. data taken from the literature: analogue results: prp, ecjoh, ecsob, ecsot values were measured and presented for the c,, paf analogue and compared with that of other analogues. c,, paf analogue was less potent than the c or cis paf analogues and equivalent to c,, paf analogue, showing that the activity decreased with lipophilicity. a highly significant parabolic relationship was calculated between log(rps) and cf (eq. ): the maximum activity was calculated cf = . , this corresponds to the cl paf. (energy minimizatipn of the compounds were calculated using the free valence geometry energy minimization method); (molecular shape analysis according to hopfinger was used to quantitatively compare the shape similarity of analogs in their minimum energy conformer states (within kcal/mol of their global minimum energy ( fig. shows the superposition of the reference conformations of the phenylalanine and tryptophane analogues). quant. struct.-act. relat. , - ( ) chemical descriptors: logp (logarithm of the partition coefficient in -octanol/water); (hansch-fujita's substituent constant characterizing hydrophobicity of a substituent on the aromatic ring and the hydrophobicity of the aromatic ring itself, respectively) ; [common overlap steric volumes (a ) between pairs of superimposed molecules in a common low energy conformation]; [dipole moment (debeyes) of the whole molecule and of the aromatic ring, respectively, calculated using the cndoi method] ; quantum chemical indices (partial atomic charges calculated by the cndoi method); - [torsion angles (deg) (fig. ) rotated during the conformational analysis of the compounds]. results: significant parabolic regression equations were calculated for the antigelling activity of the phenylalanine and tryptophan analogues (eq. and eq. , respectively): the different qsar for the phenylalanine and tryptophan analogues indicated that they interact with hemoglobin in different ways or at different sites. for the phenylalanine analogues the hydrophobicity of the side chain, the aromatic dipole moment and the steric overlap volume explained about %, % and % of the variance in antigelling activity, respectively. for the tryptophan analogues the square of the dipole moment or the steric overlap volume explained % or % of the variance in ra, respectively, being the two descriptors highly correlated. the results show that the tryptophan analogs have a relatively tight fit with the receptor site. title: s-aryl (tetramethyl) isothiouronium salts as possible antimicrobial agents, iv. in both eq. and eq. , log(l/c) depended primarily on electronic factors (eu' ) and only secondarily on hydrophobicity (ctobsd). a threshold logp value for the active isothiuronium salts was indicated, as the compounds with logp values between - . and - . were found to be totally inactive with the exception of the nitro-derivatives. title: comparative qsar study of the chitin synthesis inhibitory activity of benzoyl-ureas versus benzoyl-biurets. source: tagungsbericht , no. \ r* ponents explaining . %, . % and . % of the variance. fig. shows the minimum energy conformation of a highly active representative of the urea analogs (dimilin) with . a distance between the and carbon atoms. fig. shows the low energy conformation of the corresponding biuret analog with the two benzene rings in appro:imately the same plane and with the same c -c distance ( . a) allowing to fit a hypothetical benzoylurea phamacophore. the similarity of the regression equations and the modelling study supported the hypothesis that the benzoylbiurets act by the same mechanism as the benzoylureas. biological material: insect species: aedes aegypti, musca domestica, chilo suppressalis, hylemya platura, oncopeltus suppressalis, oncopeltus fasciatus, pieris brassicae, leptinotarsa decemlineata. [concentration of the benzoylurea derivative (various dimensions) required to kill % of insect larvae (a. aegypti, m. domestica, c. suppressalis, h. platura, . suppressalis, . fasciatus, p. brassicae or l. decemlineata]. data determined: lcso [concentration of the biuret analogue (ppm) required to kill % of insect larvae (a. aegypti or m. domestica]; molecular modeling (models of the compounds were built using molidea); conformational analysis (minimum energy conformations of the compounds were calculated using molecular mechanics method). chemical descriptors: the thesis is devoted to the quantitative analysis of the uncoupling activity of substituted phenols using chemical descriptors in order to obtain further information on the mode of action of phenol uncouplers: the study of the partition coefficient of substituted phenols in liposomelwater system [p(l/w)] showed that (i) p(l/w) depended primarily on the logp value; (ii) influence of steric and electronic parameters depended on the type of the lipid involved; qsar analysis of uncoupling phenols in rat-liver mitochondria identified the relevant physicochemical parameters required for phenols being protonophore in inner mitochondrial membrane and quantitatively separated the potency as the protonophore in the inner mitochondrial membrane and the incorporation factor (iogp); protonophoric potency of substituted phenols was linearly related to uncoupling activity when certain critical physicochemical parameters of the experiment were taken into account; linear relationship was calculated between uncoupling activities of substituted phenols and related uncouplers in the mitochondria from the flight muscles of house flies and in spinach chloroplasts; the results indicated a shuttle type mechanism for the uncoupling action of substituted phenols. title: uncoupling properties of a chlorophenol series on acer cell - ( ) compounds: chlorinated phenols substituted with -c , -c , , , -cl, , , -ci, pentachlorophenol, -ci- -me. -c - -me, -c - , -me, -c - , -me, -ci- -ally , -c - -pr- -me, z , , -c , , , , -c , , -cl, , -c , , -ci, , , -c , -cl- -no , , -c - -no,, -ci- , -no,. biological material: acer pseudoplatanus l. cell suspensions. data determined: dso [concentration of the compound (pmolll) required for % uncoupling effect registered by measuring the oxygen consumption rate by polarography]; [minimal concentration of the compound (pnol/l) required for giving a full uncoupling effect]. chemical descriptors: logp (logarithm of the partition coefficient in -octanol/water); mr (molar refractivity); ed (steric parameter representing the perimeter of coplanary molecules projected onto the aromatic plane); a (angular parameter expressing the hindrance in the neighborhood of the hydroxyl group in positions and , respectively) ; ui, (hammett's constants, characterizing the electron-withdrawing power of the para-substituent and the ortho-or -nitro substituents, respectively). results: highly significant linear regression equations were calculated for the uncoupling effects of chlorophenols in acer cell suspensions: the equations for the uncoupling effects in the whole cells and those calculated previously for isolated mitochondria or chloroplasts possess similar structures. . (* . ) a, - . ( ) title: effects of ' substituents on diphenyl ether compounds. results: sar suggested that the space for the n' and nz substituents in the psi binding site is relatively large. the variation of the number of the carbon atoms of r on the photosynthetic inhibitory activity is shown in fig. (hansch-fujita's substituent constant characterizing hydrophobicity); chemical descriptors: . (* . ) ior + . (& . ) hb + . the biological activity of three out of the (dpe- , and ) substituted diphenyl esters were measured and listed. igr values were measured for the three compounds and compared with that of a- and methoprene. it was found that the position of acetamido group in the phenol moiety when it is in the ortho position abtstr. - - ( ) increases the lipophilicity of the compound with a logp value of . . if the same group is in mr para position, the logp values are . and . , respectively and they are comparatively ineffective. when both the ortho positions are substituted with tertiary butyl groups (dpe- ) the logp value is relatively higher ( . ) which increases the lipophilicity of the compound and explains the pronounced idr activity at relatively low concentrations. abstr. results: a highly significant linear regression equation was calculated for the descriptors of r' (r' = i-pro was eliminated as an outlier) (eq. ): the compound with r' = eto, r = me and z = was found to be an effective, broad spectrum insecticide. the replacement of the quaternary carbon with a silicon atom cansimplify the synthesis of test compounds and thus can be advantageously utilized for the preparation of large compound sets for qsar studies. the data suggest that the initial electron loss from the given compounds is the preeminent factor effecting the reaction rate. a single mechanism is suggested over the entire range of reactivities, where a transition state with a considerable positive charge is involved. title: connection models of structure and activity: ii. estimation of electronoacceptor and electronodonor functions of active centers in the molecules of physiologically active materials. research institute of physiology active materials chernogolovka, moskow district, ussr. engl. summary). authors chemical descriptors: logp (logarithm of hydrophobicity). results: calculations for electronoacceptor and electronodonor entharpic and free energy factors on the base of functional groups were made according to the principle of independence of active centers: data determined: linear correlation was found between the calculated and measured characteristics: the accuracy of the fitting was the same as the measurement error of ah,,, and agm.the entropy might be calculated from enthalpy, gibbs energy and temperature: the good linear correlations between the measured and calculated data show that the functional group approaches might be used for these compound types. the substituent effects for the a-acceptorlr-donor substituents (f, c , br, i) were found to be very much larger for the c fsr relative to the nitrobenzenes. these results indicate that the extra electron enters a o*-orbital, which is localized on the c-r atoms. for the structure-solubility relationship of aliphatic alcohols. the study indicated that solubility of aliphatic alcohols depends primarily on molecular connectivity ('x), the number of carbon atoms in the alkyl chain (n'), the number of hydrogens on the a-carbon atom (normal, iso, secondary, ternary) and the degree of branching (vg): ( ) n not given r not given s not given f not given eq. was found to be a highly significant predictor of s (eq. ): the result support kier's, furthermore kier and hall's earlier models on the structural dependence of water solubility of alcohols. -log(s) = 'x + ( )* sg - . title: linear free energy relationships for peroxy radical-phenol reactions. influence of the para-substituent, the orthodi-tert-butyl groups and the peroxy radical. k (reaction rate constant (m -'s -i ) of the reaction between the reaction of cumyl-, -phenylethyl-and t-butyl-peroxy radicals and ortho-para-substituted phenol inhibitors). data taken from the literature: chemical descriptors: u+ r. ui, ur (charton's electronic substituent constant and its decomposition to inductive and resonance components, respectively for the characterization of the para substituent); (indicator variable for the presence or absence of the t-bu groups in , -position of the phenols). results: highly significant linear regression equations were calculated by stepwise regression analysis for logk in spite of the diverse data set originating from different laboratories using different peroxy radicals (eq. , eq. ): quant. struct.-act. relat. , - ( ) ( ) n = r = . s = . f = . i c~" was not selected by stepwise regression indicating that the orthodi-t-bu substitution had no significant effect on the rate of hydrogen abstraction from phenols by the radicals. the form of the equations for different subsets of the phenols and radicals indicated that the reaction mechanism was the same for the different peroxy radicals. title: a fractal study of aliphatic compounds. a quantitative structure-property correlation through topological indices and bulk parameters. the following descriptors are considered as 'bulk parameters': vw (van der waals volume, calculated from the van der waals radii of the atoms); mw (molecular weight); sd (steric density of the functional group). results: highly significant equations are presented for calculating vw, sd and mw r values ranging from . to . , other statistics and the number of investigations are not given. q and values calculated by these equations were introduced to the equation given above and the physicochemical properties were calculated. the observed and calculated iogv,, d and p values are presented and compared for the alkanes, alcohols, acids and nitriles. the observed and calculated physicochemical parameters agreed well. fractal nature of the alkyl chain length was discussed and a relationship was presented between the fractal-dimensioned alkyl chain length and a generalized topological index. title: application of micellar liquid chromatography to modeling of organic compounds by quantitative structure-activity relationships. chemical descriptors: logp (logarithm of the partition coefficient in -octanol/water). results: in a series of experiment with the listed compounds micellar liquid chromatography has been applied to model hydrophobicity of organic compounds in a biological system. the measured logk' values of the substituted benzenes were found to be superior predictors of logp. fig. shows the plot of logp versus logk' of the substituted benzenes. highly significant correlation was calculated for the logk' values of phenols (open squares) (n = , r = . ), for the rest of the compounds (full squares) (n = , r = . ) and for the entire set (n = , r = . ). further experiments using various surfactant types in the mobil phase suggested that logk' values generated on a lamellar phase may be better predictors of hydrophilicity than logp obtained from binary solvent systems. title: isoxazolinyldioxepins. . the partitioning characteristics and the complexing ability of some oxazolinyldioxepin diastereoisomers. authors quant. struct.-act. relat. , - ( ) source: j. chem. soc. perkin trans. i . no. , - compounds: oxazolinyldioxepin derivatives of type i and , where x = h, f, ci, cf or ch . data determined: logk' [logarithm of the capacity factor, measured by reversed-phase liquid chromatography (rplc)]; mep (molecular electrostatic computed by geesner-prettre and pullman's vsspot procedure). chemical descriptor: logp (logarithm of the partition coefficient in -octanollwater). results: the logk' and logp values were measured for the two type of diastereomers and a highly significant linear relationship between logk' and logp was presented (r = . ): the meps of i and 's f-derivatives were determined and presented, "a" for type i, "b" for type ii: the complex forming ability of the diastereoisomers with mono-cations was investigated and explained in terms of the structures and electronic properties of the compounds. results: linear relationships are presented plotting y versus n for the hydrophobic sorbents (fig. ) and the slopes of these straight lines are suggested for experimental determination of . q, values. ~ values determined by the suggested method are listed. while no linear relationships were found between kd and n, y depend linearly on n for the test compounds [alkanols (i). alkane diols ( ) results: three linear models were fittedwith independent variabies of log(p), mr and o x . the best fitting parameters (independent of composition) were obtained from the following models (no statistical characteristics is presented): ( ) ( ) the two types of correlations (with structural and with moving phase parameters) together might be used for the optimization of chromatographic separation of complex mixtures of sulphur-containing substances. (zero order molecular bonding type connectivity index); ig(k) = a + a p' + a logp + a p' logp ig(k) = a + a tg(cm) + a, logp + a tg(cm) logp - - the kd values derived by the suggested method were compared by kd values calculated by martin's rule and a good agreement was found. title: mathematical description of the chromatographic behaviour of isosorbide esters separated by thin layer chromatography. compounds: isosorbide esters: isosorbide (l), - -monoacetate, - -monoacetate, - -mononitrate, - -mononitrate, l-diacetate, - -nitro- -acetate, - -nitro- -acetate, l-dinitrate. rn, r~i [retention factors obtained by thin-layer chromatography in benzene/ethylacetate/isopropanol/ ( : : . ) and in dichloromethane/diisopropylether/isopropanol ( : : : ) eluent systems, respectively]. data determined: chemical descriptors: (information index, based on the distribution of the elements in the topological distance matrix); (the geometrical analogue); (randic connectivity index); (maximum geometric distance in the molecule); compounds: highly diverse chemicals. grouped according to the following properties: contains (ester or amide or anhydride) or (heterocyclic n) or ( bound to c) or (unbranched alkyl group with greater than carbons). data determined: aerud chemical descriptors: (aerobic ultimate degradation in receiving waters). v x x, nci m, (molecular weight). results: the paper has aimed at developing a model for predicting aerud. the data sets were collected from biodegradation experts. the experts estimated the biodegradationtime that might be required for aerud on the time scales of days, weeks, months and longer. (valence second order molecular connectivity index); (fourth order path/cluster connectivity index); (number of covalently bound chlorine atoms); highly diverse chemicals but typical in wastewater treatment systems were examined. zero to six order molecular and cluster connectivity indexes were calculated using computer programs wrinen in for-tran for ibm pc/xt. the best fitted linear regression model is: [first order rate constant: transport or transformation parameter (mol/pa. h)]. results: the qwasi fugacity model describes the fate of a (contaminating) chemical, such as organo-chlorine compounds, pesticides or metals. the lake model consists of water, bottom and suspended sediments, and air. the model includes the following processes: advective flow, volatilization, sediment deposition, resuspension and burial, sediment-water diffusion, wet and dry atmospheric deposition, and degrading reactions. the steady state solution of the model is illustrated by application to pcbs in lake ontario using the equilibrium criterion of fugacity as the variable controlling environmental fate of the chemical. the applications are based upon inaccurate data. use of fugacity is inappropriate for involatile chemicals, such as metals, or ionic species, because fugacities are calculated from a basis of vapor phase concentrations. for these materials the use of the equilibrium concentration activity is more appropriate since activities are calculated from a water phase base. thus, a new equilibrium criterion, termed the "aquivalent" concentration (or equivalent aqueous concentration) is suggested as being preferable. this concentration has the advantage of being applicable in all phases, such as water, air and sediments. the formalism developed in the qwasi approach can also be applied, making possible a ready comparison of the relative rates (and thus, the importance) of diverse environmental fate processes. all these are illustrated by applying the model on a steady state basis to quant. struct.-act. relat. , - ( ) abstr. - the pcb example and to the fate of lead in lake ontario. the estimated and observed concentrations of pcbs and lead in lake ontario agree well: the largest difference in the case of pcbs in rain amounts to a factor of three. in other phases, and especially in the case of lead, the difference is usually less than per cent. although in order to judge the biological effects of a contaminant it is of fundamental importance to know its transport and transformations, and the present model has been proven to useful to describe this; direct biological implications are not deduced at the present stage. the similar slopes of the equations show that these compounds exert their cytotoxicity primarily by alkylation. while the majority of the tested compounds showed no hypoxia-selective cytotoxicity (ratio awa .o), the -n and -no substituted compounds were more toxic to uv cells under hypoxic conditions (ratio = . for the compound with r = -n ), indicating cellular reduction of the nitro-group. the measured hypoxic selectivity of the -no and -n , substituted compounds was a fraction of the calculated ratio (measured fold and calculated fold by eq. between the -n and -nh, substituted compounds). the main reason for the difference between the calculated and measured hypoxic selectivity is suggested to be the low reduction potential of the -n and -no, groups (e = - mv and e = - mv, respectively). title: quantitative structure-activity relationships for the cytotoxici- (hammett's constant, characterizing the electron-withdrawing power of the substituent); (hammett's polar electronic constant characterizing the electron withdrawing power of the substituent for anilines). results: significant linear regression equations were calculated for the halflife (t / ), growth inhibition ( ) and clonogenicity data (ctlo) using hammett constants (eq. , eq. , eq. ): ( ) n = r = . s = . f not given - ( ) type, test animals, the mean level of toxicity and the form of the equation. e.g. analysis of the toxicity of phenols showed a transition between simple dependence from logp to exclusive dependence to reactivity factors indicating two separate classes of phenol toxicity (eq. for mouse i.p. toxicity, and eq. for rat oral toxicity): results: an additivity model, plc = cni a ti -to, where ni is the number of ith substituents in a benzene derivative, at; is the toxicity contribution of the ilh substituent and to is the toxicity of the parent compound (benzene), was used for predicting toxicity of lo similar correlation was found between mutagenicity and u (fig. ) indicating that both biochemical and chemical processes involve a ph dependent nucleophilic ring opening (including the protonation of the aziridin nitrogen as rate controlling step) and in.,uenced by electronic and steric factors (equation not given). resonance effect). (electron density on n in the homo calculated by the mndo). results: highly significant linear relationships between log( ic) and logp, &homo (eq. ); iogp, qhomo (eq. ) are presented indicating that the more hydrophobic and more electron-rich triazines are more active according to the ames test: substructures [a total of fragments were generated from the compounds using the program case (computer-automated structure evaluation) system]. results: a comparative classification of the compounds were performed using case for identifying molecular fragments associated with cancerogenic activity (biophores) as well as deactivating fragments (biophobes). case identified biophores and biophobes from the fragments of the compounds with a less than . % probability of being associated with carcinogenicity as a chance. the sensitivity and specificity of the analysis was unexpectedly high: . and . , respectively. the predictive power of case biological material: chemical descriptors: was tested using the identified biophores and biophobes on a group of chemicals not present in the data base. the ability of case to correctly predict carcinogens and presumed non-carcinogens was found to be very good. it was suggested that non-genotoxic carcinogens may act by a broader mechanism rather than being chemical specific. compounds: compounds of type i where r = h, ch , c,h , czh , c h , czh , c h , c h , c h , sc h ; compounds of t y p i where r = h, c~heoh, sc h h. data determined: p t pi" (a priori probability of appearance of the i-th active compound); (a priori probability of appearance of the i-th nonactive compound). (the first order molecular connectivity index); (the second order molecular connectivity index); (information-theoretic index on graph distances calculated by the wiener index according to gutmann and platt); chemical descriptors: (rank of smell, where the rank is defined to equal with one for the most active compound). results: the authors' previously proposed structure-activity relationship approach was applied for structure-odor relationship. different compounds of groups i and i were examined using the topological indices w, r, i, x as independent variables and v as the dependent variable. the best correlation was obtained between r and v ( fig. i) results: logp and iogp, values were determined for the nitroimidazole derivatives. significant linear equations were calculated, the best one related for logp and logp,r (eq. ): ( ) logp = . logp,i + . n = r = . s not given f not given chemical descriptors: logp descriptors (logarithm of the partition coefficient in i-octanoll water); ( indicator variables taking the value of for the presence of cr/p-hydroxy , a-fluoro, a-methy , afluoro, -hydroxy, i a-fluoro. , -acetonide, -deoxy, -acetate, -propionate, i-butyrate or -isobutyrate, respectively). results: a data set of steroids were compiled after removing those ones containing unique substituents. the set was divided into two categories of approximately equal membership by defining a threshold logp value of . . a descriptor set was created and the non-significant ones were eliminated using the weight-sign change feature selection technique. linear leaning machine was applied to calculate the weight vectors and complete convergence was achieved in the training procedure. the predictive ability of the linear pattern classifier thus obtained was tested using the leave one out procedure. the predictive ability was found to be i . %. the predictive ability of the approach was found to be good and improvement was expected with larger data set. steroids, however, containing new substituents would have to be subjected to a repeated pattern-recognition calculation. lengthlbreadth descriptors ( descriptors)]. results: for modeling the shape of the compounds, simca was used: the approach was to generate disjoint principal models of clustered points in a multidimensional space. the number of clusters for each structure was determined by using hierarchical cluster analysis. fig. shows the orthogonal views of a schematic representation of the sim-ca models for the atom clusters in senecionine: each compound in turn was used as a reference structure. every other structure was superimposed on the reference using the ends of the corresponding binding moment vector plus the ring nitrogen atom. canonical correlation analysis was used for calculating the correlation between the five biological activity data and shape descriptors of structures. the best correlation was observed for jurs' shadow descriptors. the msa and simca descriptors were comparable. the model was able to express both the amount and direction of shape the differences, and also for encoding relevant information for correlation with the biological activity. compounds: n-substituted -methyl- -nitropyrazole- -carboxamides ( ), n-substituted -amino- -methylpyrazole- -carboxamides (iii), n-substituted -methyl- -diazopyrazole- -carboxamides and n-piperidiny -n-( , -dimethyl- -nitrosopyrazol- -yl)-urea (vii) . title: structure-activity correlations for psychotomimetics. . phenylalkylamines: electronic, volume, and hydrophobicity parameters. abtstr. quant. struct.-act. relat. , - ( ) data determined: edso conformational analysis g [dose of the compound (mg/kg) which causes % of the rats which were trained on rngfkg reference compound to respond as they would to the training drug]; (geometries of the compounds were calculated using mmf from starting geometries determined by the program euclid). discriminant analysis resulted in a function containing six variables which misclassified only one compound in the training set. when the data was repeatedly split randomty into a training and a test sb, the misclassification rate was % ( out of classifications). fig. shows the plot of the two canonical varieties from discriminant analysis visualizing the separation of hallucinogenic and nonhallucinogenic derivatives (meaning of symbols are the same as in fig. ). multiple regression analysis (mra) was found to be the most useful for identifying relevant and discarding redundant variables. highly significant parabolic regression equations were calculated for the human activity data (a) (n = , r ranging from . to . , f not given) and for animal data (edso) (n = , r = . and r = . , f not given). eight descriptors were found to be highly significant. among these the importance of directional hydrophobicity and volume effects indicated that steric and hydrophobic interactions participate in the interaction with the receptor.mra indicated a strong interaction between the meta-and para-substituents and the presence of the formation of charge transfer complex by accepting charge. data did not support the hypothesis that the human activity data and animal ( ) data taken from the literature: sweet(n) (sweet taste of the compound, where n represent the number of times a sample has to be diluted to match the taste of % sucrose solution). [class fit distances of a compound to sweet and nonsweet class (dimension not given) calculated by principal component analysis]. chemical descriptors: mr (molar refractivity); bi, l (sterimol steric parameters, characterizing the steric effect of the substituent); r (hansch-fujita's substituent constant characterizing hydrophobicity); urn, up (hammett's constants, characterizing the electron-withdrawing power of the substituent in meta-and para-position, respectively). results: no statistically significant regression equation was obtained by the hansch-fujita approach using the chemical descriptors listed. d', d quant. stact.-act. relat. , - ( ) abstr. - principal component analysis of the data set extracted principal components, explaining % of the variance of the sweet compounds. the sweet compounds clustered in a relatively confined region of the d space whereas the tasteless and bitter compounds were scattered around the sweet compounds. a coomans plot, however., indicated, when plotting d' versus d , that sweet and nonsweet compounds could be well separated along the d' axis ( fig. , title: conformation of cyclopeptides. factor analysis. a convenient tool for simplifying conformational studies of condensed poly-ring systems. prolyl-type cyclopeptides. authors conformations of the six-membered dop-ring family may be reproduced by means of a superposition of the canonical twist (t), boat (b) and chair (c) forms. physically, the coefficients have the meaning of relative contributions (amplitudes) of the t, b and c forms into the total conformation of the ring. here factor analysis (fa) and principal component analysis was used in conformational studies of various x-ray conformers of dop/pyr. a correspondence was found between factors identified and rpt, when the rings are considered separately. this fact allows a physical interpretation of the fa results: two or three puckering variables were found for the dop and pyr rings expressing the absolute amplitudes of the basic pucker modes. subse-quent fa treatment of the condensed system revealed five conformational variables necessary and sufficicnt to describe the twolring puckering completely. each of the basic pucker modes defines a unique pattern of conformational variation of the whole two-ring system. the results demonstrate that fa is a powerful technique in analysing condensed poly-ring systems, not amenable to the rpt treatment. title: preprocessing, variable selection, and classification rules in the application of simca pattern recognition to mass spectral data. authors: dunn m*, w.j.; emery, s.l.; glen, g.w; scott, d.r. college of pharmacy, the university of illinois at chicago south wood, chicago il , usa. source: environ. sci. technol. , ( ) , - . compounds: a diverse set of compounds observed in ambient air classified as ( ) nonhalogenated benzenes; ( ) chlorine containing compounds; ( ) bromo-and bromochloro compounds; ( ) aliphatic hydrocarbons; ( ) miscellaneous oxygen-containing hyhocarbon-like compounds (aliphatic alcohols, aldehydes and ketones). pattern recognition was applied to autocorrelation-transformed mass spectra of the compounds using providing chemical class assignment for an unknown]; m/z;, mlz,, m/zl (first three principal components scores of simca). results: simca pattern recognition method was applied on a training set of toxic compounds targeted for routine monitoring in ambient air. the analysis resulted in very good classification and identification of the compounds ( % and %, respectively). however, the training procedure proved to be inadequate as a number hydrocarbons from field samples (gcims analysis) were incorrectly classified as chlorocarbons. a new approaches for the preprocessing (scaling the ms data by taking the square root of the intensitiesfollowed by autocorrelation transform), variable selection (only the most intense ions in the ms spectrum were taken), and for the classification rules of simca has been introduced to improve results on real data. fig, and as a result of the revised rules the classification performance has been greatly improved for field data ( - %). title: a qsar model for the estimation of carcinogenicity. - ( ) it was suggested that the mechanism of mutagenicity of the cimeb[a]ps measured in the ames test is probably more complex than the simple reactivity of carbocation intermediates. [dipole interaction potential (dimension not given)]; [molecular electrostatic potential (kcallmol) in a plane]; [molecular electrostatic field map (kcall mol), mapping the e(r)j values of a molecule surface in a plane, predicting the directions and energies of the interactions with small polar molecules at distances greater than the van der waals sphere]; (construction of surfaces corresponding to a given value of potential); d mep and mef maps ( d maps weregenerated by superimposing the equipotential curves corresponding to a value of kcallmol in the case of mep. and . kcallmol in the case of mef, computed in several planes perpendicular to the mean plane of the analogues in low energy conformations, stacking over each other in a distance). results: the three vasopressin analogues differ significantly in their biological activities. both mep and mef maps of the of the biologically active (mpa')-avp and (cpp')-avp are similar, but they are different from that of the inactive (ths')-avp. fig. , fig. abtstr. - quant. struct.-act. relat. , - ( ) a new method for calculating the points of the equipotential curves was also presented. crystal structure (crystal coordinates of the molecules were determined by x-ray diffraction methods). data taken from the literature: [electrostatic molecular potential (ev) were calculated using am- type semiempirical mo calculations]; conformational analysis [minimum energy conformations were calculated using x-ray structures as input geometries followed by am -method (fletcher-powell algorithm)]. chemical descriptors: ui, u [rotational angles of the n-ally group (deg)]. results: four similar energy minima were located by am-i calculations for both namh+ and nlph+. the energy minima for the protonated nam + and nlph + were the most populated ones with conformational enantiomers relative to the involved n-allyl-piperidine moiety ( % and %, respectively). it was shown that the isopotential curve localization of emp contour maps were very similar for the corresponding conformations of both nlph * and namh + indicating that both molecules should interact a the same anionic sites of the opioid receptor, ( p morphine receptor). fig. and fig. shows the emp contour maps of namh' and nlph + , respectively, in their preferred conformations: compounds: esfenvalerate (ss and sr isomers) of type i, -phenoxybenzyl -( -ethoxyphenyl)- , , -trifluoropropyl ether (r quant. struct.-act. relat. , - ( ) abstr. isomer) ( ). a-cyano- -phenoxybenzyl -( -chlorophenyl)- -methylpropionate (s isomer) (iii) and deltamethrin (iv). " cn data determined: conformational analysis (minimum energy conformations of the compounds in vacuum were calculated using am molecular orbital method and broyden-fletcher-goldfarb-shanno method integrated into mopac); (root mean square, indicating the goodness of fit between two conformers in d); (logarithm of the partition coefficient in i-octanol/water estimated using clogp program); [heat of formation of the most stable conformer (kcallmol)]. rms logp e chemical descriptors: - results: it was assumed that the d positions of the benzene rings of the pyrethroids are decisive for good insecticidal activity. the lower energy conformers of (i) (ss and sr isomers), ( ) (r isomer), ( ) (s isomer) and deltamethrin (iv) were compared by superimposition. inspite of their opposite configuration, esfenvalerate (i) (ss isomer) and the new type pyrethroid i (r isomer) were reasonably superimposed, indicating that the positions of the benzene rings in space are important and the bonds between them are not directly determinant (fig. ) crystal structure (x-ray crystal coordinates of penicillopepsin was obtained from the protein data bank). data determined: electrostatic potential [electrostatic potential of the protein atoms (kcallmol) is calculated using the partial charges in the amber united atom force field); docking (the dock program was used to find molecules that have a good geometric fit to the receptor). results: a second generation computer-assisted drug design method has been developed utilizing a rapid and automatic algorithm of locating sterically reasonable orientations of small molecules in a receptor site of known d structure. it includes also a scoring scheme ranking the orientations by how well the compounds fit the receptor site. in the first step a large database (cambridge crystallographic database) is searched for small molecules with shapes complementary to the receptor structure. the second step is a docking procedure investigating the electrostatic and hydrogen bonding properties of the receptor displayed by the midas graphics package. the steps of the design procedure is given. the algorithm includes a simple scoring function approximating a soft van der waals potential summing up the interaction between the receptor and ligand atoms. directional hydrogen bonding is localized using electrostatic potential of the receptor at contact points with the substrate. the shape search of (i) was described in detail. a new method has been developed for the construction of a hypothetical active site (hasl), and the estimation of the binding of potential inhibitors to this site. the molecules were quantitatively compared to one another through the use of their hasl representations. after repeated fitting one molecule lattice to another, they were merged to form a composite lattice reflecting spatial and atomic requirements of all the molecules simultaneously. the total pki value of an inhibitor was divided to additive values among its lattice points presumed to account for the binding of every part of the molecule. using an iterative method, a self consistent mathematical model was produced distributing the partial pki values of the training set in a predicting manner in the lattice. the hasl model could be used quantitatively and predictively model enzyme-inhibitor interaction. a lattice resolution of - a was found to be optimal. a learning set of e. coli dhfr inhibitors were chosen to test the predictive power of the hasl model at various resolutions. binding predictions (pki values) were calculated for the entire inhibitor set at each resolution and plotted separately for the learning and test set members at . a resolution (fig. i) : . ala-). data determined: molecular models ( d structure of the molecules have been constructed and displayed using the program geom communicating with cambridge x-ray data bank, brookhaven protein data bank, sandoz x-ray data bank sybyl and disman; quant. struct.-act. relat. , - ( ) abstr. - distance geometry [nuclear overhauser enhancements (noe) and spin-spin coupling constants were measured by d nmr methods, semiempirically calibrated as proton-proton distance (a) and dihedral angle (deg) constrains and used in distance geometry calculations (disman) andlor in restrained molecular dynamics calculations to determine d structure of molecules in solution]; crystal structure (atomic coordinates of the compounds were determined by x-ray crystallography); rms [root mean square deviation (a) of the corresponding atoms of two superimposed molecular structures]. chemical descriptors: results: distance geometry calculations were carried out using geom and disman, to identify all conformations of the compounds in solution which were consistent with experimental data obtained by noe measurements. the application of geom was demonstrated by modelling cycbsporin a with and without a limited set of h-bond constrains and with a full nmr data set. in case of cyclosporin a, randomly generated linear analogues of the cyclic structure were formed from the monomers. geometric cyclization was achieved using disman, resulting in many different but stereochemically correct conformations of cyclosporin a. superposition of the backbones of the best cyclic conformers showed rms deviations between . a and . a. fig. shows the superposition of a disman generated ring conformation (thick line) with its x-ray structure of cyclosporin a (thin line) with h-bond constraints (rms = . a): fig. distance and dihedrl-angle constraints have been extracted from noe and vicinal coupling data and used to generate the conformation and the cyclization conditions of the hexapeptide (fig. ) (position of residual distance violations and their direction is shown by arrows): although the described method is not exhaustive, it explores a much greater variety of initial structures than had been previously possible. title: a new model parameter set for @-lactams. authors: durkin, k.a.; sherrod, m.j.; liotta*, d. department of chemistry, emory university atlanta gl , usa. source: j. org. chem. , ( ) , - . compounds: @lactam antibiotics of diverse structure. data taken from the literature: crystal structures (crystal coordinates of the p-lactams were determined using x-ray diffraction method). results: superposition of the x-ray structures and the calculated geometries of -lactams using the original parameter set in the mm force field in model gave satisfactory rms values. a lack of planarity of the -lactam ring and significant differences in the calculated bond lengths and anglesaround the &lactam nitrogen were detected, however. = , s, so, so,]. in order to improve fit, a new atom type with new parameters has been developed for the p-lactam nitrogen (wild atom type coded with symbol in model). the new parameters were evaluated by comparison of the calculated and x-ray geometries of the -lactams. using the new parameter set, the x-ray data were satisfactorily reproduced except for the sulfone -lactams. it was indicated that the ampac data were not suitable for the sulfones as the hypervalent sulfur compounds are not well described in the am hamiltonian. an additional parameters was, however, derived giving good structural data unrelated to the ampac information. it is not known which the new parameter sets is the best for the sulfone / -lactams. title: a molecular modelling study of the interaction of compounds noradrenalin. biological material: a) cdna of the hamster lung ,-adrenergic receptor and &-adrenergic receptor kinase; b) bacterio-ovine-and bovine-rhodopsin and rhodopsin kinase. protein primary sequence (amino acid sequence of the hamster lung p,-adrenergic receptor has been deduced by cloning the gene and the cdna of the hamster lung &adrenergic receptor);, (the cosmic molecular modeling program was used for modeling a-helices in a hydrophobic environment using p and w torsion angles of - " and - ", respectively, according to blundell et al.); (the two highest lying occupied and the two lowest lying unoccupied orbitals, respectively, calculated using indo molecular orbital calculation); crystal structure (crystal coordinates of noradrenaline has been determined by x-ray diffractometry); conformation analysis (minimum energy conformation of the &-adrenergic receptor model has been calculated using molecular mechanics method). results: strong experimental evidences suggested that rhodopsin and ,-adrenergic receptor had similar secondary structure. thus, it was assumed, that similarly to bacterio-ovine-and bovine-rhodopsins, d -adr ener gic receptor -. . b -adrenergic receptor possesses a structure consisting of seven ahelices traversing the cell membrane. fig. shows the postulated arrangements of the a-helices of rhodopsin and the &-receptor. using the experimental data, a model of the &-adrenergic receptor has been generated for the study of its interaction with noradrenaline. a possible binding site was created. successful docking indicated that homo and lumo orbitals contributed to the binding in a chargetransfer interaction between trp- and noradrenaline. a hydrogen bond was detected between the threonine residue of the model receptor and the noradrenaline side chain hydroxyl explaining why chirality was found to be important for the activity of adrenergic substances. title: three-dimensional steric molecular modeling of the [binding affinity (nm) of the compounds to the -ht receptor]. data determined: molecular modeling ( d molecular models of each compound were made using camseqlm molecular modeling system); [distance (a) from the center of the aromatic ring to the ring-embedded nitrogen, when the nitrogen is placed in the same plane as the aromatic ring]. results: in order to derive rules for the -ht pharmacophore, a molecular graphics-based analysis was made using six core structures. the structures were aligned so as to overlay the aromatic rings and to place the ring embedded nitrogen atom in the same plane as the aromatic ring. nine steric rules were derived from the analysis common to all potent -ht agents. fig. shows the d representation of the six overlaid -ht core structures using camseqim: the -ht inactivity of atropine could be explained because its steric properties differed from those the active ics - only by a single atom and failed to meet two of the nine hypothetical criteria. uv-visible spectra [spectrophotometric studies of mixtures of the dyes and nicotine in % (vlv) aqueous ethanol mixture at "ci. results: cyanine dyes demonstrate a multitude of biological activities which may be due to the interference of the adsorbed dye molecule on active sites of the living cell. it was shown by uv-and visible spectrophotometry that the hydroxy styryl cyanine dyes and nicotine formed : charge-transfer complexes. the absorption band of the complex formed between nicotine and dye was detected at wavelengths longer than those of the individual pure substances having identical concentrations to those in mixture. fig. shows that the two partially positive centres of the dye ( -( -hydroxystyryl)-pyridinium-i-ethyliodide) were located at a similar distance than the two nitrogen atoms of pyridine or pyrrolidinyl moieties of nicotine allowing the suggested : i parallel stucking interaction between the two molecule: molecular modeling ( conformations were calculated using a distance geometry algorithm and energy minimized by a modified mm force field in moledit). results: all conformers within kcallmol of the lowest energy conformer were superposed on the x-ray structure of mk- . crystal structure (crystal coordinates of the proteins were determined using x-ray diffraction method). data taken from the literature: was less than a); (probability that a given tetrapeptide sequence is superimposable on the ribonuclease a structure); [probability that the ith residue (amino acid) will occur in the jth conformational state of the tetrapeptide which is superimposable to ribonuclease a]. results: it was suggested that the five tetrapeptides were essential components of larger peptides and might be responsible for their biological activity (binding to the cd receptor). earlier it was hypothesized that the critical tetrapeptide located in a segment of ribonuclease a, would assume low energy conformations (residues - , a @-bend, having a segment homologous to the sequence of peptide t). low energy conformers of the tetrapeptides could be superimposed to the native structure of segment - of ribonuclease a. fig. shows the superimposition of peptide t (full square): many low energy conformers could be calculated for the tetrapeptides but for the polio sequence. the p, value for most tetrapeptides were - times higher that the value of the less active polio sequence. the results supported the hypothesis that the active peptide t adopts the native ribonuclease @-bend. title: potential cardiotonics. . synthesis, cardiovascular activity, molecule-and crystal structure of -phenyl-and -(pyrid- -yl)- data determined: [dose of the compound (mollkg) required for % increase of the heart beat frequency of guinea pig or dog heart]; [dose of the compound (mollkg) required for % decrease of systolic or diastolic blood pressure of dog]; crystal structure (atomic coordinates of the compounds were determined by x-ray diffraction); molecular modeling (molecule models were built using molpac); mep [molecular electrostatic potential (mep) (dimension not given) was calculated using cndoiz]. results: milrinon and its oxygen containing bioisoster possess highly similar crystal structure and mep isopotential maps ( fig. and fig. ) both compounds show strong positive inotropic and vasodilatoric activity. it was suggested that the negative potential region around the thiocarbonyl group such as the carbonyl group in milrinon imitates the negative potential field around the phosphate group of camp. title: molecular mechanics calculations of cyclosporin a analogues. effect of chirality and degree of substitution on the side chain conformations of ( s, r, r, e)- -hydroxy- -methyl- -(meth~lamino)- octenoic acid and related derivatives. [solution conformation of csa in cdch has been elucidated via molecular dynamics simulation incorporating distance constrains obtained from ir spectroscopy and nuclear overhauser effect (noe) data]; (conformational analysis was performed using the search subroutine within sybyl); energy minimization (low energy conformers were calculated using molecular mechanics withinmacromodel ver. . applying an all-atom version of the amber force field). results a total of conformations of csa have been identified within kcalfmol of the minimum energy conformer. population analysis showed that one conformer dominates in solution. fig. shows the superposition of the peptide backbone of the crystal and solution structures of csa (crystal structure is drawn with thick line and the solution structure with thin line). it was shown that the boltzmann distribution between active and inactive conformers correlated with the order of the immunosuppressive activity. a common bioactive conformer serving as a standard for further design has been proposed for csa and its analogs. abtstr. quant. struct.-act. relat. , - ( ) data determined: molecular modeling (models of (i), ( ) and ( ) were built using sybyl based on x-ray coordinates of the compounds); conformational analysis (minimum energy conformations of the compounds were calculated using the search option of sybyl and mndo method; [interaction energy of the molecules (kcall mol) with a hypothetical receptor probe (negatively charged oxygen atom) calculated by grid]. results: the specific receptor area of the sodium channel was modeled with a negatively charged oxygen probe (carboxyl group), interacting with the positively charged (protonated) ligand. fig. shows areas for energetically favorable interaction (areas i, oh ho c *. . quant. struct.-act. relat. , - ( ) abstr. - biological material: a) aspergillus ochraceus; b) carbopeptidase a. data determined kobr [first order rate coefficient (io- /sec) of the hydrochloric acid hydrolysis of ochratoxin a and b]; x-ray crystallography (coordinates of the crystal structure of ochratoxin a and b was obtained using x-ray diffraction); (models of ochratoxin a and b was built using alchemy); ["c nmr chemical shifts (ppm) of the amide and ester carbonyls of the ochratoxins]. chemical descriptors: pka (negative logarithm of the acidic dissociation constant). results: a reversal of the hydrolysis rate between ochratoxin a and b was observed comparing the hydrolysis rates obtained in vitro (carbopeptidase a) and in vivo (hydrochloric acid). the difference in hydrolysis rates cannot be due to conformation since the two toxins have the same conformation in both crystal and in solution. fig. shows the fit of ochratoxin a and b based on superimposing the phenolic carbon atoms. it is suggested that the relative large steric bulk of the chloro atom hinders the fit between ochratoxin a and the receptor site of carbopeptidase a. thus, probably the slower metabolism is the reason, why ochratoxin a is more toxic than ochratoxin b. title: inhibitors of cholesterol biosynthesis. . trans- -( -pyrrol- charge distribution studies showed that compactin had two distinct regions of relatively large partial charges corresponding to the pyrrol ring and the isobutyric acid side chain. experiments for more closely mimicking the polar regions associated with the high activity of compactin indicated that potency of the new compounds was relatively insensitive to the polarity of the r' group. it was also suggested that an electron deficient pyrrole ring was required for high potency. title: synthesis and biological activity of new hmg-coa reductase inhibitors. . lactones of pyridine-and pyrimidine-substituted . dihydroxy- -heptenoicf-heptanoic) acids. chemical descriptors: results: an attempt was made to correlate electrophysiological activity with the effect of the position of the aryl group on the conformation of the side chain using molecular modeling. the study suggested that the compounds with class activity prefer a gauche (a in fig. ) and compounds in which class i activity prefer trans relationship of the nitrogens (b in fig. ) : the study indicated that the point of attachment of the aryl moiety had an effect on the side chain conformation which appeared to be a controlling factor of the electrophysiological profile of these compounds. title: a molecular mechanics analysis of molecular recognition by cyclodextrin mimics of a-chymotrypsin. authors ( ) quant. struct.-act. relat. . - ( ) biological material: chymotrypsin. data taken from the literature: crystal structure (crystal coordinates of the macrocycles determined using x-ray diffraction analysis). data determined: molecular modeling structure superposition (models of b-cd and in chains by nmethylformamide and n-dimethyl-formamide substituted (capped) b-cd were built using the amber program and the coordinates for building the n-methylformamide substituent were calculated using mndo in the mopac program); (energy minimization of the molecules were calculated in vacuo using molecular mechanics program with the amber force field); (energy minimized structures of b-cd and capped b-cd were separately fit to the xray structure of the b-cd complex); [molecular electrostatic potential (kcallmol) of b-cd and capped b-cd were approximated by the coulombic interaction between a positive point charge and the static charge distribution of the molecule, modeled by the potential derived atomic point charges at the nuclei and visualized as d mep map]. results: b-cd and capped b-cd were analyzed as biomimetic models of the active site of chymotrypsin. capped b-cd was shown to be the more effective biomimetic catalyst. capping also altered certain structural features of molecular recognition. the orientation of the secondary hydroxyls were altereddue to twisting of some of the glucose units. secondary hydroxyl oxygen mimics the ser- of chymotrypsin in initiating the acyl transfer event through nucleophilic attack on the substrate. fig. shows the energy minimized structures of b-cd (a) and capped b-cd (b) (fragment number is given in parenthesis). the mep maps of b-cd and capped b-cd showed that the qualitative features of the electrostatic recognition were practically the same in the two mimics. biologicai material: four monocotyledonous (johnson grass, yellow foxtail, barnyard grass, yellow millet) and four dicotyledonous weed species (velvetleaf, morning glory, prickly sida, sicklepod). data determined: [pre-emergence and postemergence herbicidal activities of the compounds were measured and rated using a scale ranging from (no activity) to (complete kill]; [measure of the compound's ability (dimension not given) to translocate upwards in plants through xylem vessels); [soil sorption coefficient calculated by the formula k, = c,/c,, where c, is the concentration of the compound (pg compoundlg soil) and c. is the concentration of the compound (pg compoundlml) in water solution in equilibrium with the soil]; (models of the compounds were built using maccs and prxbld programs); tscf kd molecular modeling quant. struct.-act. relat. , - ( ) abstr. - conformational analysis (minimum energy conformations of the compounds were calculated using mm molecular mechanics method); (molecules were visualized using program mogli on an evans and sutherland picture system ); [total energies, orbital eigenvalues, atomic charges and dipole moments of simple model analogs of type i were calculated using prddo (partial retention of diatomic overlap) level of approximation]. electronic structure chemical descriptors: logp (logarithm of the partition coefficient in -octanollwater). results: conformational analyses and high level quantum mechanical calculations of the conformational preferences showed that the compounds with r = -c and -ci substituents adopt a coplanar structure stabilized by intramolecular hydrogen bond, whereas the -c analogue does not (fig. ): higher logp values ( . - . logarithmic unit difference), higher kd and tscf values of the -ci and -ci substituted compounds relative to the -ci analog were interpreted as the result of the intramolecular hydrogen bond and were consistent with the observation that the -ci and -ci analogs were active as post-emergence but not pre-emergence herbicides while the -ci derivative was active in both modes. title: application of molecular modeling techniques to pheromones of the marine brown algae cutleria multifida and ectocarpus siliculosus (phaeophyceae). metalloproteins as chemoreceptors? (geometrical models of the compounds were constructed using information from the cambridge structural data base (csd) and calculated using molecular mechanics methods in sybyl); (minimum energy conformations of the compounds were calculated using molecular mechanics method within sybyl). chemical descriptors: kfcq [partition coefficient in fc /water (fc = fluorocarbon results: as both ectocarpene (i) and multifidene ( ) trigger mutual cross reactions between male gametes of ectocarpus siliculosus and cutleria multifida males it was supposed that a common mode of binding should exist for the two structurally different pheromones. the active analogue approach was applied to model the pheromone receptor by superposing the minimum energy conformations of active structural analogues (hi, iv, v, vi) on ectocarpene and multifidene. the common active conformation of (i) and ( ) was extracted by systematic superimposition of the analogues. to explain the function of the double bonds in the pheromones, the presence of a receptor bound metal cation was assumed. simultaneous optimization,of both structures without and with a receptor bound metal cation resulted in virtually the same conformations. fig. shows the mapping of multifidene onto ectocarpene in their biologically relevant conformations. solvent)]. title: critical differences in the binding of aryl phosphate and carbamate inhibitors of acetylcholinesterases. conformational analysis [minimum energy conformations of (asn-ala-asn-pro) was calculated using charmm (chemistry at harvard macromolecular mechanics), amber (assisted model building with energy refinement) and ecepp (empirical conformational energy program for peptides) potential energy functions]; [root mean square deviation (a) of the position of the corresponding atoms of two superimposed molecular sructures], results: low energy conformations of (asn-ala-asn-pro) has been determined using charmm, ecepp and amber in order to determine their final conformations and relative energies. the final conformations were compared calculating the rms values of their c" atoms and matching the parameters of the energy minimized (asn-ala-asn-pro), peptide to that of the ideal helix or coiled coil. the similarity of the final conformations obtained by using any two different potentials starting from the same conformation varied from the satisfactory to highly unacceptable. the extent of difference between any pairs of the final conformations generated by two different potential energy functions were not significantly different. the lowestenergy conformation calculated by each of the energy potentials for any starting conformation was a left handed helix and pair-wise superposition of the c" atoms in the final conformations showed small rms values ( .o - . a) . it was suggested that the native conformation of (asn-ala-asn-pro), in the cs protein may be a left-handed helix, since all three potential energy functions generated such conformation. - ( ) source: proteins proteins , ( ), - , . biological material: crambin. data determined phi-psi probability plot (probabilities of the occurrences of phi-psi dihedral angle pairs for each amino acid were determined and plotted using the data of approximately proteins from the brookhaven protein data bank); (optimization technique for the reproduction the folding process converging to the native minimum energy structure by dynamically sampling many different conformations of the simplified protein backbone). chemical descriptors: phi-psi values [dihedral angles (deg) defined by the bonds on either side of the a-carbon atom of the amino acid residue in a protein]. results: a simplified model has been developed for the representation of protein structures. protein folding was simulated assuming a freely rotating rigid chain where the effect of each side chain approximated by a single atom. phi-psi probabilities were used to determine the potentials representing the attraction or repulsion between the different amino acid residues. many characteristics of native proteins have been successfully reproduced by the model: (i) the optimization was started from protein models with random conformations and led to protein models with secondary structural features (a-helices and strands) similar by nature and site to that of the native protein; (ii) the formation of secondary structure was found to be sequence specific influenced by long-range interactions; (iii) the association of certain pairs of cysteine residues were preferred compared to other cysteine pairs depending on folding; (iv) the empirical potentials obtained from phi-ps probabilities led to the formation of a hydrophobic core of the model peptide. x [dihedral angle (deg) ]. results: four kinds of monte carlo simulations of about , steps of the conformations of crambin were carried out by using the second derivative matrix of energy functions (starting from native and unfolded conformations both in two kinds of systems, in vacuo and in solution). fig. shows the native (a) and theunfolded (b) conformation of crambin. starting from native conformation, the differences between the mean properties of the simulated crambin conformations obtained from in vacuo and solution calculations were not very large. the fluctuations around the mean conformation during simulation were smaller in solution than in vacuo, however. simulation starting from the unfolded conformation resulted in a more intensive fluctuation of the structure in solution than in vacuo indicating the importance of the hydration energy term in the model. the conformations generated in the simulations starting from the native conformation deviate slightly from the xray conformation (rms = . a and i . a for in vacuo and solution simulations, respectively). the results indicate that the simulations of the protein with hydration energyare more realistic that the simulations without hydration energy. fig. results: earlier studies overestimated the catalytic rate decrease of the hypothetical d a point mutant of thrombin ( orders of magnitude decrease calculated instead of the order of magnitude measured). the source of error was due to an overestimation of v and neglecting the effects of the surrounding water molecules and induced dipoles. to compensate for these errors, a scale factor of . was introduced into the calculations. as aresult of the rescaling, one magnitude increase of tat for the d mutant and two magnitudes decrease of k,, of the k mutant of ribonuclease a was predicted. it was shown that the effect of the mutations on the catalytic rate depended almost entirely on steric factors. it was suggested that in mutants of serine proteases where the buried asp is replaced by ala or asp, the kcat value will decrease between - orders of magnitude. title: high-resolution structure of an hiv zinc fingerlike domain via a new nmr-based distance geometry approach. authors: summers*, m.f.; south, t.l.; kim [root mean square deviation (a) of the corresponding atoms of two superimposed molecular sructures]. results: the atomic resolution structure of an hiv zinc fingerlike domain has been generated by a new nmr-based dg method using d noesy backcalculation. the quality of the structures thus obtained were evaluated on the basis of the consistence with the experimental data (comparison of measured and back-calculated nmr spectra) rather than comparing it tostructural informations from other sources (e.g. x-ray data). the method provided a quantitative measure of consistence between experimental and calculated data which allowed for the use of tighter interproton distance constraints. the folding of the c( l)-f( )-n( )-c( )-g(s)-k( ) residues were found to be virtually identical with the folding of the related residues in the x-ray structure of the iron domain of rubredoxin (rms values . and . a). the backbone folding of the peptide was found to be. significantly different from that of the "classical" dna-binding zn-finger. fig. shows the wire frame model of all the back%ne atoms and certain side chain atoms of the peptide (dg struciure) (dashed lines indicate hydrogen atoms): active site of the protease dimer in an extended conformation with extensive van der waals and hydrogen bonding and was more than % excluded from contact with the surrounding water (fig. i , where the inhibitor is shown in thicker lines and the hydrogen bonds in dashed lines): data determined: ago,, ago, [standard free energy (callmol) of transfer of a molecule from an apolar phase to an aqueous phase, observed or calculated by eq. : ago, = c aui ai, where aoi is the atomic solvation parameter of atomic group i, ai is the accessible surface area of atom i, respectively]. results: atomic solvation parameters (asps) characterizing the free energy change per unit area for transfer of a chemical group from the protein interior to aqueous surroundings were determined. ago, and ago, were determined and compared, and a highly significant linear relationship is presented (fig. ) . one letter symbols indicate amino acid side chains: fig . the binding of the inhibitor induced substantial movement in the en-' zyme around the residues to in both subunits at places exceeding i - the structure of glutamine synthetase is discussed. it was established that hydrophobic interactions are important for the intersubunit interactions, and the hydrophobic interactions between the two rings of subunits are stronger than between the subunits within a ring. the cterminal helix contribute strongly to the inter-ring hydrophobic interaction. asps are suggested to estimate the contribution of the hydrophobic energy to protein folding and subunit assembly and the binding of small molecules to proteins. title: determination of the complete three-dimensional structure of the trypsin inhibitor from squash seeds in aqueous solution by nuclear magnetic resonance and a combination of distance geometry and dynamical simulated annealing. authors: holak*, t.a.; gondol, d.; otlewski, j.; wilusz, t. max-planck-hstitut fiir biocbemie d- martinsried bei miinchen, federal republic of germany. title: interpretation of protein folding and binding with atomic crystal structure (atomic coordinates of cmti-i was determined by solvation parameters. x-ray diffraction method). results: in order to obtain information of the d structure of the free cmti-i in solution, a total of inhibitor structures were calculated by a combination of distance-geometry and dynamical simulated annealing methods, resulting in well defined d positions for the backbone and side-chain atoms. fig. shows the superposition of the backbone (n, c", c, ) atoms of the structures best fitted to residues to (binding loop): the average rms difference between the individual structures and the minimized mean stfucture was . (* . ) a for the backbone atoms and . (+ . ) a for all heavy atoms. title: electron transport in sulfate reducing bacteria. molecular modeling and nmr studies of the rubredoxin-tetraheme-cytochrome-c complex. biological material: a) sulfate reducing bacterium (desulfovibrio vulgaris); b) rubredoxin (iron-sulfur protein); c) tetraheme cytochrome c from d. vulgaris; e) flavodoxin. detected in the segments from the residues to and - . fig. shows the best superposition (residues to ) of the nmr and crystal structure of cmti-i indicating the backbone c, c", n, , as well as the disulfide c and s atoms: fig. it was demonstrated that uncertainty in nmr structure determination can be eliminated by including stereospecific assignments and precise distance constraints in the definition of the structure. crystal structure (coordinates of the crystal structure of the compounds were determined by x-ray crystallography). results: the speed of the homolysis of the organometallic bond is " times higher in the apoenzyme bound coenzyme biz than in a homogenous solution. structural changes occurring during the co-c bond homolysis of the coenzyme biz leading from cobalt(ii ) corrin to cobalt(i ) corrin were investigated. fig. shows the superposition of structures of the cobalt corrin part of the biz (dotted line) and of cob(ii)alamin (solid line): biological material: apoenzyme, binding the coenzyme biz and the data determined: . shows that the crystal structure of biz and cob(i )alamin are strikingly similar and offers no explanation for the mechanism of the protein-induced activation of homolysis. it was suggested that the co-c bond may be labiiized by the apoenzyme itself and in addition to a substrate-induced separation of the homolysis fragments (which mights be supported by a strong binding of the separated fragments to the protein). 'h nmr (complete stereospecific assignments were carried out and proton-proton distance constrains were determined by the analyses of dqf-cosy, hohaha and noesy spectra); nh, ah, oh [chemical shifts (ppm) of proton resonances of human eti ; d structure ( d structure of et was calculated using the distance geometry program dadas based upon the noesy proton-proton distance constrains determined by nmr spectroscopy); [root mean square distance (a) between et conformers calculated by distance geometry (dadas)]. results: the solution conformation of et has been determined by the combined use of d 'h nmr spectroscopy and distance geometry calculations. five structures of et have been calculated from different initial conformations. the superposition of the backbone atoms the calculated structures is shown in fig. . the average rms value in the core region for the main-cahin atoms was . a. quant. struct.-act. relat. , - ( ) the lack of specific interactions between the core and tail portions of et and a characteristic helix-like conformation in the region from lys' to cys" was shown. literature data indicated that neither the eti - nor the etl - truncated derivatives of et showed constricting or receptor binding activity suggesting that the et receptor recognizes an active conformation consisting of both the tail and core portion. the present study, however, suggested that the receptor bound conformation of et is probably different from that in solution because the lack of interaction between tail and core. the hydrophobic nature of the tail suggested the importance of a hydrophobic interaction with the receptor. compounds: triphenyl-methyl-phosphit cation (tpmp+). biological material: nicotinic acetylcholine receptor (achr), a prototype of the type i of membrane receptor protein from the electric tissue of torpedo and electrophorus. results: a computer model of the achr ion channel has been proposed. fig. shows the side view of the ion channel model with five pore-forming mz-helices and the channel blocking photoaffinity label (tpmp +) represented by the dotted sphere: fig. it was supported by electronmicroscopy, electrophysiological-and biochemical experiments that the mz-helices were formed by homologous amino acid sequences containing negatively charged amino acid side chains which act as the selectivity filter. the amino acid side chains may undergo conformational changes during the permeation of the cation. the predicted transmembrane folding of four transmembrane a-helices of type i receptors is shown in fig. : fig. energy profile calculations indicate that other transmembrane sequences of the receptor protein besides m may affect the ion channel. source: cabios , ( ), - . results: an interactive computer program tefoojj has been developed for drug design on ibm/pc and compatible computers. the program contains the following modules and performs the following calculations: a) series design for selecting an optimal starting set of compounds using a modified version of austel's method; b) regression analysis calculating the hansch's equation; c) hansch searching method using the equation calculated by the regression analysis routine or the use of an input equation for the identification of the most active compounds; d) geometrical searching methods for finding the optimum substituents in the parameter space using the sphere, ellipse, quadratic or polyhedric cross algorithms with or without directionality factors; e) space contraction for reducing the dimension of the parameter space by eliminating non-significant parameters; f) an example is given for the lead optimization of an aliphatic lead compound correctly predicting the n-pentane to be the optimum substituent. results: a new expert system sparc is being developed at epa and at the university of georgia to develop quantitative structure-activity relationships for broad compound classes: a) classical qsar approaches predict therapeutic response, environmental fate or toxicity from structure/property descriptors quantifying hydrophobicity, topological descriptors, electronic descriptors and steric effects; b) sparc (sparc performs automated reasoning in chemistry), an expert system written in prolog, models chemistry at the level of physical organic chemistry in terms of mechanism of interaction that contribute to the phenomena of interest; c) sparc uses algorithms based on fundamental chemical structure theory to estimate parameters such as acid dissociation constants (pk,s), hydrolysis rate constants, uv, visible and ir absorption spectra, and other properties; d) the information required to predict input data for broad classes of compounds is dispersed throughout the entire ir spectrum and can be extracted using fourier transforms; e ) the accuracy of sparc algorithm was demonstrated on the close match of calculated and experimental pk, values of carboxylic acid derivatives near to the noise level of measurement. abtstr. - quant. struct.-act. relat. , - ( ) results: a new stand-alone molecular simulation program, nmrgraf integrating molecular modeling and nmr techniques has been introduced by biodesign inc. a) nmrgraf is a molecular modeling program utilizing force fields which incorporate empirical properties such as bond lengths and angles, dihedral, inversion and nonbonded interactions, electrostatic charges and van der waals interactions; b) the molecular structural properties are combined with nuclear overhouser effect (noe) and j-coupling nmr data (experimental interproton distance constrains and bond angle data); c) the nmr proton-proton distance data are accurate only at relatively short distances ( to a) which restricts the use of nmr noe approaches only for the analysis of molecules with known x-ray structure; d) the combination of nmr and molecular modeling approaches, however, makes it possible to model virtually any molecule even if its structure does not exist in the databases. title: electronic structure calculations on workstation computers. results: the main features of the program system turbomole for large-scale calculation of scf molecular electronic structure on workstation computers is described: a) the program system allows for scf level treatments of energy, first-and second-order derivatives with respect to nuclear coordinates, and an evaluation of the me? correlation energy approximation; b) the most important modules of turbomole are (i) dscf performing closed and open shell rhf calculations; (ii) egrad used for analytical scf gradient evaluations; (iii) kora calculating direct two-electron integral transformation (iv) force for the computation and processing of integral derivatives and the solution of cphf equations; c) comparison and evaluation of timings of representative applications of turbomole on various workstations showed that apollo ds o.ooo and iris d were the fastest and comparable to the convex c in scalar mode. results: a new algorithm has been developed for the calculating and visualizing space filling models of molecules.the algorithm is about times faster than a conventional one and has an interesting transparency effect when using a stereo viewer. a) the algorithm is briefly described and and the result is visualized on modeling a ribonucleotide unit; b) in the order of increasing atomnumbers, the (x,y) sections of the hemispherical disks of the atoms are projected on the screen with decreasing value of the azimuthal angle (p) of the van der wads radius and as the value of p decreases the projection is increasingly whitened to obtain shading effect on the surfaces of the spheres; c) the transparency of the van der waals' surfaces of atoms of a molecule makes possible to perceive almost the whole space filling structure and not only the surface, hiding the underlying atoms. title: supercomputers and biological sequence comparison algo-authors: core*, n.g. ; edmiston, e.w. ; saltz, j.h. ; smith, r.m. rithms. yale university school of medicine new haven ct - , usa. source: computers biomed. res. , ( ) , - . compounds: dna and protein fragments. chemical descriptors: sequences of monomers. results: a dynamic programming algorithm to determine best matches betweenpairs of sequences or pairs of subsequences has been used on the intel ipsc/l hypercube and ,the connection machine (cm-i). parallel processing of the comparison on cm-i results in run times which are to times as fast as the vax , with this factor increasing as the problem size increases. the cm-i and the intel ipsc hypercube are comparable for smaller sequences, but the cm-i is several times quicker for larger sequences. a fast algorithm by karlin and his coworkers designed to determine all exact repeats greater than a given length among a set of strings has been tried out on the encore multimax/ o. the dynamic programming algorithms are normally used to compare two sequences, but are very expensive for multiple sequences. the karlin algorithm is well suited to comparing multiple sequences. calculating a multiple comparison of dna sequences each - nucleotides long results in a speedup roughly equal to the number of the processors used. source: cabios , ( ), . results: a program has been developed for the prediction and display of the secondary structure of proteins using the primary amino acid sequence as database. a) the program calculates and graphically demonstrates four predictive profiles of the proteins allowing interpretation and comparison with the results of other programs; b) as a demonstration the sliding averages of n sequential amino acids were calculated and plotted for four properties of human interleukin : (i) plot of the probabilities of a-helix, p-structure and p-turns according to chou and fasman; (ii) @-turn index of chou and fasman; (iii) plot of the hydrophobicity index of hopp and woods; (iv) flexibility index of karplus and schulz; c) the regions of primary structure having properties which usually go together agreed reasonably well with each other, i.e. loops and turns with bend probability and hydrophilicity with flexibility. title: dsearch. a system for three-dimensional substructure searching. quant. struct.-act. relat. , - ( ) sci. , ( ) , - . results: the search for threedimensional substructures is becoming widely used in d modeling and for the construction of pharmacophores for a variety of biological activities. a system ( dsearch) for the definition and search of three-dimensional substructures is described: a) representation of atom types consists of five fields (i) element (he-u); (ii) number of non hydrogen neighbors (bonded atoms); (iii) number of 'k electrons; (iv) expected number of attached hydrogens; (v) formal charge; (vi) four type of dummy atoms are also used to define geometric points in space (e.g. centroid of a b) definition of queries (i) definition of spatial relationship between atoms; (ii) matches in atom type (iii) preparation of keys (constituent descriptors); (iv) execution of key search; (v) geometric search including the handling of angleldihedral constrains and takes into account "excluded volume"; c) time tests showed that a search of d structures with to atoms in large databases with more than , entries took only a few minutes ( - s). results: a database containing about , compounds in connection tables and , experimentally determined structures from the cambridge structural database has been transformed into a database of low energy d molecular structures using the program concord. the strategy for building the d database consisted of the following four steps: a) generation of approximate d coordinates from connection tables (hydrogens were omitted); b) assignment of atom types from connection table information characterized by five descriptors: (i) element type (he-u); (ii) number of attached non-hydrogen neighbors ( - ); (iii) number of 'k electrons ( - ); (iv) calculated number of attached hydrogens ( - ); (v) formal charge ( - , , ); c) addition of three types of chemically meaningful dummy atoms for purposes of d substructure searching: (i) centroids of planar and -membered rings; (ii) dummy atoms representing the lone electron pairs; (iii) ring perpendiculars positioned orthogonal to and . a above and below each planar ring; d) efficient storage of the resultant coordinate database indexing the compounds with identification number; e) the database can be used among others to deduce pharmacophores essential for biological activity and to search for compounds containing a given pharmacophore. title source: c&en , ( ) , - . results: a complex carbohydrate structure database (ccsd) has been developed by the complex carbohydrate research center at the university of georgia, having more than structures and related text files, with about more records to be added over the next two years. the following are the most important features of ccsd: a) in ccsd, database records include full primary structures for each complex carbohydrate, citations to papers in which sequences were published, and suplementary information such as spectroscopic analysis, biological activity, information about binding studies, etc; b) structural display format visualize branching, points of attachment between glycosyl residues and substituents, anomeric configuration of glycosyl linkages, absolute configuration of glycosyl residues, ring size, identity of proteins or lipids to which carbohydrates are attached and other data; c) it is planned that ccsd will provide threedimensional coordinates, to visualize and rotate the structures in stereo and study their interaction with proteins or other biopolimers. probing bioactive mechanisms p commercial carbamate insecticides of type ii, where r' = h, s-bu biological material: acetylcholinesterase. data determined: d [distance (a) between the serine oxygen of acetylcholinesterase and the methyl substituents of the carbamate and phosphate inhibitors molecular modeling (models and minimum energy conformations of acetylcholine, aryl carbamate and phosphate ester inhibitors were created using the draw mode of a maccs database and the prxbld modeling program) results: transition state modeling of the reaction of the serine hydroxyl ion of acetylcholinesterase with the methylcarbamoyl and dimethyl phosphoryl derivatives of , -dimethyl-phenol showed that the active site binding for these two classes of acetylcholinesterase inhibitors should be different. the model shows that the distances between the serine oxygen and the ring substituents (meta-and para-me-thy groups) are different in both spacing and direction. fig. shows the transition state models of serine hydroxyl d values for the meta-and para-methyl substituents of n-methylcarbamate and dimethylphosphate were meta = . , para = . and meta = . , para = . a, respectively title: a comparison of the charmm, amber and ecepp potentials for peptides. . conformational predictions for the tandemly repeated peptide (asn-ala-asn-pro) biological material: tandemly repeated peptide (asn-ala-asn-pro) which is a major immunogenic epitope in the circumsporozoite (cs) protein of plasmodium falciparum conformational analysis dream or reality? a authors: nbray-szab *, g.; nagy, j.; bcrces, a. priori predictions for thrombin and ribonuclease mutants molecular modeling (geometric model of subtilisin and trypsin was built using protein data bank coordinates and model of thrombin was built using the theoretical coordinate set of graphic representations of the triad of the tetrahedral intermediate for the enzymes formed on the active side chain and residues (ser- , his- and asp in subtilisin; ser- , his- and asp-i in trypsin and thrombin; his- , lys- and his-i in ribonuclease a) were created using pcgeom electrostatic properties (electrostatic surfaces and fields of the molecules were calculated and displayed using amber and associated programs); [chemical shift (ppm) measured by nmr results: a hypothetical model between rubredoxin and cytochrome c was built as a model for the study of electron transfer between different redox centers, as observed in other systems. fig. i shows the main chains atoms of the proposed complex where the hemes of the cytochromes are shown along with the center of rubredoxin and stabilized by hydrogen bonds and charge-pair interactions (the nonheme iron of the rubredoxin is in close proximity to heme of cytochrome c ): spectroscopy].the model was consistent with the requirements of steric factors, complementary electrostatic interactions and nmr data of the complex. comparison of the new model and the nmr data of the previously proposed flavodoxin-cytochrome c complex showed that both proteins interacted with the same heme-group of cytochrome c . title: nuclear magnetic resonance solution and x-ray structures of squash trypsin inhibitor exhibit the same conformation of the proteinase binding loop.authors: holak*, t.a.; bode, w.; huber, r.; otlewski, j.; wilusz, t. max-planck-institut fiir biochemie d- martinsried bei munchen, federal republic of germany.source: j. mol. biol. , no. , - . biological material: a) trypsin inhibitor from the seeds of the squash cucurbita maxima; b) p-trypsin and trypsin inhibitor complex.title: retention prediction of analytes in reversed-phase high-performance liquid chromatography based on molecular structure. . quant. struct.-act. relat. , - ( ) results: an expert system (cripes) has been developed for the prediction of rp-hplc retention indices from molecular structure by combining a set of rules with retention coefficients stored in a database. the method underlying the system is based on the "alkyl aryl retention index scale" and aims to improve the reproducibility of prediction and compatibility between various instruments and column materials. the vp-expert system shell from microsoft was used for the development. the performance of cripes was demonstrated on several subtypes of substituted benzenes (phenacyl halides, substituted arylamines, arylamides and other types). in general the calculated and measured retention indices agreed well but relatively large deviations were observed between the ie and ic values for phenacyl bromides and chlorides, o-and p-bromo anilines, n-methylbenzamide and n,n-dimethylbenzamide and phthalate esters. the extension of the database with further interaction values was regarded as necessary for a consistently high accuracy at prediction. [out-of-plane bending energy (kcal/mol) given by the formula e, = kd', where d is the distance from the atom to the plane defined by its three attached atoms and k is a force constant]; eb e, [torsional energy (kcal/mol .deg ) associated with four consecutive bonded atoms i,j,k,l given by the formula e, = ki,j.k,l(l fs/lslcos(lslbi,j,k,~)), where b is the torsion angle between atoms i j , k and , s and k are constants]; e, [potential energy (kcallmol) (nonbonded energy term) associated with any pair of atoms which are neither directly bonded to a common atom or belong to substructures more than a specified cutoff distance away given by the formula e, = kij(l. la" - . /a ), where a is the distance between the two atoms divided by the sum of their radii, and k is the geometric mean of the k constants associated with each atom]. results: model geometries produced by the tripos . force field have been assessed by minimizing the crystall structures of three cyclic hexapeptides, crambin and diverse complex organic compounds. comparative force field studies of the tripos . , amber and amberlopls force fields were carried out by energy minimization of three cyclic hexapeptides starting from the crystal structures showed the tripos . force field superior to the others with the exception of amber, ecep and levb force fields as published by other workers. a direct comparison between the performance of tripos . and . amber using isolated crambin showed that the bond and torsion angles of tripos . averaged closer to the crystal structure than the angles calculated by amber (rms = . a, . deg and . deg for bond lengths, angles, and torsions, respectively, and rms = . a for heavy atoms). fig. shows the superimposed structures of crambin before and after energy minimization:fi . tripos . was assessed for general purpose applications by minimizing organic compounds starting from their crystal structures. the test showed that tripos . had a systematic error in overestimating the bond lengths of atoms in small rings. statistical analysis of the results showed that t r i p s . had an acceptable overall performance with both peptides and various organic molecules, however, its performance was not equal to the best specialized force fields. title: new software weds molecular modeling, nmr. author: krieger. j. c&en sixteenth st., n.w., washington dc , usa. source: c&en , ( ) , . key: cord- - dhfptw authors: uribe-sánchez, andrés; savachkin, alex title: predictive and reactive distribution of vaccines and antivirals during cross-regional pandemic outbreaks date: - - journal: influenza res treat doi: . / / sha: doc_id: cord_uid: dhfptw as recently pointed out by the institute of medicine, the existing pandemic mitigation models lack the dynamic decision support capability. we develop a large-scale simulation-driven optimization model for generating dynamic predictive distribution of vaccines and antivirals over a network of regional pandemic outbreaks. the model incorporates measures of morbidity, mortality, and social distancing, translated into the cost of lost productivity and medical expenses. the performance of the strategy is compared to that of the reactive myopic policy, using a sample outbreak in fla, usa, with an affected population of over four millions. the comparison is implemented at different levels of vaccine and antiviral availability and administration capacity. sensitivity analysis is performed to assess the impact of variability of some critical factors on policy performance. the model is intended to support public health policy making for effective distribution of limited mitigation resources. as of july , who has reported confirmed human cases of avian influenza a/(h n ) which resulted in deaths worldwide [ ] . at the same time, the statistics for the h n outbreak has so far included countries with a total reported number of infections and deaths of , and , , respectively [ ] . today, an ominous expectation exists that the next pandemic will be triggered by a highly pathogenic virus, to which there is little or no pre-existing immunity in humans [ ] . the nation's ability to mitigate a pandemic influenza depends on the available emergency response resources and infrastructure, and, at present, challenges abound. predicting the exact virus subtype remains a difficult task, and even when identified, reaching an adequate vaccine supply can currently take up to nine months [ , ] . even if the existing vaccines prove to be potent, their availability will be limited by high production and inventory costs [ , ] and also will be constrained by the supply of antiviral drugs, healthcare providers, hospital beds, medical supplies, and logistics. hence, pandemic mitigation will have to be done amidst limited availability of resources and supporting infrastructure. this challenge has been acknowledged by who [ ] and echoed by the hhs and cdc [ , ] . the existing models on pandemic influenza (pi) containment and mitigation aims to address various complex aspects of the pandemic evolution process including: (i) the mechanism of disease progression, from the initial contact and infection transmission to the asymptomatic phase, manifestation of symptoms, and the final health outcome [ ] [ ] [ ] , (ii) the population dynamics, including individual susceptibility [ , ] and transmissibility [ , [ ] [ ] [ ] , and behavioral factors affecting infection generation and effectiveness of interventions [ ] [ ] [ ] , (iii) the impact of pharmaceutical and nonpharmaceutical measures, including vaccination [ ] [ ] [ ] , antiviral therapy [ ] [ ] [ ] , social distancing [ ] [ ] [ ] [ ] [ ] and travel restrictions, and the use of low-cost measures, such as face masks and hand washing [ , [ ] [ ] [ ] . recently, the modeling efforts have focused on combining pharmaceutical and nonpharmaceutical interventions in search for synergistic strategies, aimed at better resource utilization. most of such approaches attempt implementing a form of social distancing followed by application of pharmaceutical measures. for significant contributions in this area see [ , [ ] [ ] [ ] [ ] [ ] [ ] [ ] . one of the most notable among these efforts is a - initiative by midas [ ] , which cross-examined independent simulation models of pi spread in rural areas of asia [ , ] , usa and uk [ , ] , and the city of chicago [ ] , respectively. midas crossvalidated the models by simulating the city of chicago, with . m inhabitants and implementing a targeted layered containment [ , ] . the research findings of midas and some other groups [ , ] were used in a recent "modeling community containment for pandemic influenza" report by iom, to formulate a set of recommendations for pi mitigation [ ] . these findings were also used in a pandemic preparedness guidance developed by cdc [ ] . at the same time, the iom report [ ] points out several limitations of the midas models, observing that "because of the significant constraints placed on the models . . . the scope of models should be expanded." the iom recommends "to adapt or develop decision-aid models that can . . . provide real-time feedback . . . and include the costs and benefits of intervention strategies." our literature review yields a similar observation that most existing approaches focus on assessment of a priori defined strategies, and virtually none of the models are capable of "learning," that is, adapting to changes in the pandemic progress, or even predicting them, to generate dynamic strategies. such a strategy has the advantage of being developed dynamically, as the pandemic spreads, by selecting a mix of available mitigation options at each decision epoch, based on both the present state of the pandemic and its predicted evolution. in an attempt to address the iom recommendations, we present a simulation optimization model for developing predictive resource distribution over a network of regional outbreaks. the underlying simulation model mimics the disease and population dynamics of each of the affected regions (sections . and . ). as the pandemic spreads from region to region, the optimization model distributes mitigation resources, including stockpiles of vaccines and antiviral and administration capacities (section . ). the model seeks to minimize the impact of ongoing outbreaks and the expected impact of potential outbreaks, using measures of morbidity, mortality, and social distancing, translated into the cost of lost productivity and medical expenses. the methodology is calibrated and implemented on a sample outbreak in fla, usa with over m inhabitants (section ). the strategy is compared to the reactive myopic policy, which allocates resources from one actual outbreak region to the next, each time trying to cover the entire regional population at risk, regardless of the resource availability. the comparison is done at different levels of vaccine and antiviral availability and administration capacity. we also present a sensitivity analysis for assessing the impact of variability of some critical factors, including: (i) antiviral efficacy, (ii) social distancing conformance, and (iii) cdc response delay. the objective of our methodology is to generate a progressive allocation of the total resource availability over a network of regional outbreaks. the methodology incorporates (i) a cross-regional simulation model, (ii) a set of single-region simulation models, and (iii) an embedded optimization model. we consider a network of regions with each of which classified as either unaffected, ongoing outbreak, or contained outbreak ( figure ) . the cross-regional simulation model connects the regions by air and land travel. the single-region simulation models mimic the population and disease dynamics of each ongoing region, impacted by intervention measures. the pandemic can spread from ongoing to unaffected regions by infectious travelers who pass through regional border control. at every new regional outbreak epoch, the optimization model allocates available resources to the new outbreak region (actual distribution) and unaffected regions (virtual distribution). daily statistics is collected for each ongoing region, including the number of infected, deceased, and quarantined cases, for different age groups. as a regional outbreak is contained, its societal and economic costs are calculated. in sections . - . , we present the details of the simulation and optimization models. a testbed illustration and a comparison of our strategy to the myopic policy is given in section . a schematic of the cross-regional simulation model is shown in figure . the model is initialized by creating population entities and mixing groups, for each region. a pandemic is started by an infectious case injected into a randomly chosen region. the details of the resulting regional contact dynamics and infection transmission are given in section . . as the infected cases start seeking medical help, a new regional outbreak is detected. a resource distribution is then determined and returned to the single-region model. the outbreak can begin cross-regional simulation the single-region model subsumes the following components (see figure ): (i) population dynamics (mixing groups and schedules), (ii) contact and infection process, (iii) disease natural history, and (iv) mitigation strategies, including social distancing, vaccination, and antiviral application. the model collects detailed statistics, including number of infected, recovered, deceased, and quarantined cases, for different age groups. for a contained outbreak, its societal and economic costs are calculated. the societal cost includes the cost of lost lifetime productivity of the deceased; the economic cost includes the cost of medical expenses of the recovered and deceased and the cost of lost productivity of the quarantined [ ] . each region is modeled as a set of population centers formed by mixing groups or places where individuals come into contact with each other during the course of their social interaction. examples of mixing groups include households, offices, schools, universities, shopping centers, entertainment centers, and so forth, [ ] . each individual is assigned a set of attributes such as age, gender, parenthood, workplace, infection susceptibility, and probability of travel, among others. each person is also assigned Δt time-discrete (e.g., Δt = hour) weekday and weekend schedules, which depend on: (i) person's age, parenthood, and employment status, (ii) disease status, (iii) travel status, and (iv) person's compliance to social distancing decrees [ ] . as their schedules advance, the individuals circulate throughout the mixing groups and come into contact with each other (see section . . ). it is assumed that at any point of time, an individual belongs to one of the following compartments (see figure ): susceptible, contacted (by an infectious individual), infected (asymptomatic or symptomatic), and recovered/deceased. in what follows, we present the infection transmission and disease natural history model, which delineates the transitions between the above compartments. process. infection transmission occurs during contact events between susceptible and infectious cases, which take place in the mixing groups. at the beginning of every Δt period (e.g., one hour), for each mixing group g, the simulation tracks the total number of infectious cases, n g , present in the group. it is assumed that each infectious case generates r g per Δt unit of time new contacts [ ] , chosen randomly (uniformly) from the pool of susceptibles present in the group. we also assume the following: (i) during Δt period, a susceptible may come into contact with at most one infectious case and (ii) each contact exposure lasts Δt units of time. once a susceptible has started her contact exposure at time t, she will develop infection at time t + Δt with a certain probability that is calculated as shown below. let l i (t) be a nonnegative continuous random variable that represents the duration of contact exposure, starting at time t, required for susceptible i to become infected. we assume that l i (t) is distributed exponentially with mean /λ i (t), where λ i (t) represents the instantaneous force of infection applied to susceptible i at time t [ ] [ ] [ ] . the probability that susceptible i, whose contact exposure has started at time t, will develop infection at time t + Δt is then given as a schematic of the disease natural history is shown in figure . during the incubation phase, the infected case stays asymptomatic. at the end of the latency phase, she enters the infectious phase [ , , ] . she becomes symptomatic at the end of the incubation period. at the end of the infectious phase, she enters the period leading to a health outcome, which culminates in her recovery or death. mortality for influenza-like diseases is a complex process affected by many factors and variables, most of which have limited accurate data support available from past pandemics. furthermore, the time of death can sometimes be weeks following the disease episode (which is often attributable to pneumonia-related complications [ ] ). because of the uncertainty underlying the mortality process, we adopted an age-based form of the mortality probability of infected i, as follows: where μ i is the age-dependent base mortality probability of infected i, ρ i is her status of antiviral therapy ( or ), and τ is the antiviral efficacy measured as the relative decrease in the base probability [ ] . we assume that a recovered case develops full immunity but continues circulating in the region. mitigation is initiated upon detection of a critical number of confirmed infected cases [ ] , which triggers resource distribution and deployment. the model incorporates a certain delay for deploying field responders. pharmaceutical intervention (phi) includes vaccination and antiviral application. vaccination is targeted at individuals at risk [ ] to reduce their infection susceptibility. the vaccine takes a certain period to become effective [ ] . vaccination is constrained by the allocated stockpile and administration capacity, measured in terms of the immunizer-hours. we assume that as some symptomatic cases seek medical help [ , ] , those at risk of them will receive an antiviral. the process is constrained by the allocated stockpile and administration capacity, measured in terms of the number of certified providers. both vaccination and antiviral application are affected by a number of sociobehavioral factors, including conformance of the target population, degree of risk perception, and compliance of healthcare personnel [ ] [ ] [ ] . the conformance level of the population at risk can be affected, among other factors, by the demographics and income level [ ] [ ] [ ] [ ] [ ] as well as by the quality of public information available [ ] . the degree of risk perception can be influenced by the negative experience developed during previous pharmaceutical campaigns [ , ] , as well as by public fear and rumors [ , ] . nonpharmaceutical intervention (npi) includes social distancing and travel restrictions. we adopted a cdc guidance [ ] , which establishes five categories of pandemic severity and recommends quarantine and closure options according to the category. the categories are determined based on the value of the case fatality ratio (cfr), the proportion of fatalities in the total infected population. for cfr values lower than . % (category ), voluntary at-home isolation of infected cases is implemented. for cfr values influenza research and treatment between . % and . % (categories and ), in addition to at-home isolation, the following measures are recommended: (i) voluntary quarantine of household members of infected cases and (ii) child and adult social distancing. for cfr values exceeding . % (categories and ), all the above measures are implemented. as the effectiveness of social distancing is affected by some of the behavioral factors listed above [ ] , we assume a certain social distancing conformance level. travel restrictions considered in the model included regional air and land border control for infected travelers. figure , the optimization model is invoked at the beginning of every nth new regional outbreak epoch (n = , , . . .), starting from the initial outbreak region (n = ). the objective of the model is to allocate some of the available mitigation resources to the new outbreak region (actual distribution) while reserving the rest of the quantities for potential outbreak regions (virtual distribution). by doing so, the model seeks to progressively minimize the impact of ongoing outbreaks and the expected impact of potential outbreaks, spreading from the ongoing locations. mitigation resources can include stockpiles of vaccines and antivirals, administration capacity, hospital beds, medical supplies, and social distancing enforcement resources, among others. the predictive mechanism of the optimization model is based on a set of regression equations obtained using single-region simulation models. in what follows, we present the construction of the optimization model and explain the solution algorithm for the overall simulation-based optimization methodology. we introduce the following general terminology and notation: the optimization criterion (objective function) of the model incorporates measures of expected societal and economic costs of the pandemic: the societal cost includes the cost of lost lifetime productivity of the deceased; the economic cost includes the cost of medical expenses of the recovered and deceased and the cost of lost productivity of the quarantined. to compute these costs, the following impact measures of morbidity, mortality, and quarantine are used, for each region k: to estimate these measures, we use the following regression models obtained using a single-region simulation of each region k: where δ i ·· denotes the regression coefficient associated with resource i and δ im ·· is the regression coefficient for the interaction between resources i and m. similar models are used for y hk , d hk , and v hk . the above relationships between the impact measures and the resource distributions ought to be determined a priori of implementing a cross-regional scenario (see section ). here, we consider each region k as the initial outbreak region. we assume, however, that as the pandemic evolves, the disease infectivity will naturally subside. hence, the regression equations need to be re-estimated at every new outbreak epoch, for each region k ∈ c n , using the singleregion simulation models, where each simulation must be initialized to the current outbreak status in region k in the cross-regional simulation. as an alternative to using a computationally burdensome approach of re-estimating the regression equations, a modeler may choose to use a certain decay factor α n [ ] to adjust the estimates of the regional impact measures at every nth outbreak epoch, in the following way: in addition, we use the following regression model to estimate the probability of pandemic spread from affected region l to unaffected region k, as a function of resources allocated to region l, which, in turn, impact the number of outgoing infectious travelers from the region: where γ i ·· denotes the regression coefficient associated with resource i, γ im ·· is the regression coefficient associated with interaction between resources i and m, and γ ·· represents the intercept. consequently, the total outbreak probability for unaffected region k can be found as p k = l∈b n p lk . as in the case of the impact measures, the estimates of the regional outbreak probabilities need to be progressively re-estimated or adjusted using a scheme similar to ( ), as follows: finally, we calculate the total cost of an outbreak in region k at the nth decision epoch as follows: where m h is total medical cost of an infected case in age group h over his/her disease period, w h is total cost of lost wages of an infected case in age group h over his/her disease period, w h is cost of lost lifetime wages of a deceased case in age group h, and w h is daily cost of lost wages of a non-infected case in age group h who complies with quarantine. the model. the optimization model has the following form. minimize tc n j q j , q j , . . . , q r j + s∈c n tc n s q s , q s , . . . , q rs · p n s subject to the first term of the objective function represents the total cost of the new outbreak j, estimated at the nth outbreak epoch, based on the actual resource distribution {q j , q j , . . . , q r j } (see ( )). the second term represents the total expected cost of outbreaks in currently unaffected regions, based on the virtual distributions {q s , q s , . . . , q rs } ( ) and the regional outbreak probabilities p n s ( ) . the set of constraints assures that for each resource i, the total quantity allocated (current and virtual, both nonnegative) does not exceed the total resource availability at the nth decision epoch. note that both the objective function and the availability constraints are nonlinear in the decision variables. ( ) estimate regression equations for each region using the single-region simulation model. ( ) begin the cross-regional simulation model. ( ) select randomly the initial outbreak region j. set n = . (c) re-estimate regression equations for each region k ∈ b n ∪ c n using the single-region simulations, where each simulation is initialized to the current outbreak status in the region (alternatively, use ( ) and ( )). (d) solve the resource distribution model for region j. (e) update the total resource availabilities. ( ) calculate the total cost for each contained region and update the overall pandemic cost. to illustrate the use of our methodology, we present a sample h n outbreak scenario including four counties in fla, usa: hillsborough, miami dade, duval, and leon, with populations of . , . , . , and . million people, respectively. a basic unit of time for population and disease dynamics models was taken to be Δt = hour. regional simulations were run for a period (up to days) until the daily infection rate approached near zero (see section . ). below, we present the details on selecting model parameter values. most of the testbed data can be found in the supplement [ ] . models. demographic and social dynamics data for each region [ ] were extracted from the u.s. census [ ] and the national household travel survey [ ] . daily (hourly) schedules [ ] were adopted from [ ] . each infected person was assigned a daily travel probability of . % [ ] , of which % was by air and % by land. the probabilities of travel among the four regions were calculated using traffic volume data [ ] [ ] [ ] [ ] , see table . infection detection probabilities for border control for symptomatic cases were assumed to be % and %, for air and land, respectively [ ] . the instantaneous force of infection applied to contact i at time t (( ), [ ] ) was modeled as influenza research and treatment where α i is the age-dependent base instantaneous infection probability of contact i, θ i (t) is her status of vaccination at time t ( or ), and δ is the vaccine efficacy, measured as the reduction in the base instantaneous infection probability (achieved after days [ ] ). the values of age-dependent base instantaneous infection probabilities were adopted from [ ] (see table ). the disease natural history included a latent period of hours ( . days), an incubation period of hours ( . days), an infectiousness period from to hours ( . to . days), and a period leading to health outcome from to hours ( . to days) [ ] . base mortality probabilities (μ i in ( )) were found using the statistics recommended by the working group on pandemic preparedness and influenza response [ ] . this data shows the percentage of mortality for age-based high-risk cases (hrc) ( table , columns - ). mortality probabilities (column ) were estimated under the assumption that highrisk cases are expected to account for % of the total number of fatalities, for each age group [ ] . single-region simulation models were calibrated using two common measures of pandemic severity [ , , ] : the basic reproduction number (r ) and the infection attack rate (iar). r is defined as the average number of secondary infections produced by a typical infected case in a totally susceptible population. iar is defined as the ratio of the total number of infections over the pandemic period to the size of the initial susceptible population. to determine r , all infected cases inside the simulation were classified by generation of infection, as in [ , ] . the value of r was calculated as the average reproduction number of a typical generation in the early stage of the pandemic, with no interventions implemented (the baseline scenario) [ ] . historically, r values for pi ranged between . and . [ , ] . to attain similar values, we calibrated the hourly contact rates of mixing groups [ ] (original rates were adopted from [ ] ). for the four regions, the average baseline value of r was . , which represented a high transmissibility scenario. the values of regional baseline iar averaged . . mitigation resources included stockpiles of vaccines and antiviral and administration capacities (section . ). a -hour delay was assumed for deployment of resources and filed responders [ ] . phi. the vaccination risk group included healthcare providers [ ] , and individuals younger than years (excluding younger than months old) and older than years [ ] . the risk group for antiviral included symptomatic individuals below years and above years [ , ] . the efficacy levels for the vaccine (δ in ( )) and antiviral (τ in ( )) were assumed to be % [ , ] and %, respectively. we did not consider the use of antiviral for a mass prophylactic reduction of infection susceptibility due to the limited antiviral availability [ ] and the risk of emergence of antiviral resistant transmissible virus strains [ ] . we assumed a % target population conformance for both vaccination and antiviral treatment [ ] . the immunity development period for the vaccine was taken as days [ ] . a version of the cdc guidance for quarantine and isolation for category was implemented (section . . , [ ] ). once the reported cfr value had reached . %, the following policy was declared and remained in effect for days [ ] : (i) individuals below a certain age ξ ( years) stayed at home during the entire policy duration, (ii) of the remaining population, a certain proportion φ [ ] stayed at home and was allowed a one-hour leave, every three days, to buy essential supplies, and (iii) the remaining ( − φ) noncompliant proportion followed a regular schedule. all testbed scenarios considered the quarantine conformance level φ equal to % [ ] . an outbreak was considered contained, if the daily infection rate did not exceed five cases, for seven consecutive days. once contained, a region was simulated for an additional days for accurate estimation of the pandemic statistics. a statistical design of experiment [ ] was used to estimate the regression coefficient values of the significant decision factors and their interactions (see section . ; the values of adjusted r ranged from . % to . %). the simulation code was developed using c++. the running time for a cross-regional simulation replicate involving over four million inhabitants was between and minutes (depending on the initial outbreak region, with a total of replicates) on a pentium . ghz with . gb of ram. the performance of the dpo and myopic policies is compared at different levels of resource availability. table summarizes the total vaccine and antiviral requirements for each region, based on the composition of influenza research and treatment average daily cost of lost productivity of a non-infected quarantined case ( - ) $ . theregional risk groups (see section . ). table shows the per capita costs of lost productivity and medical expenses, which were adopted from [ ] and adjusted for inflation for the year of [ ] . comparison of the two strategies is done at the levels of %, %, and % of the total resource requirement shown in table . figures (a) and (b) show the policy comparison in the form of the % confidence intervals (ci) for the average number of infected and deceased, respectively. figure also shows the policy comparison using the % ci for the average total pandemic cost, calculated using the pandemic statistics, and the per capita costs from table . for illustrative purposes, we also show the average number of regional outbreaks, for each policy, at different levels of resource availability, in the testbed scenario involving four regions, with the hillsborough as the initial outbreak region ( table ) . it can be observed that the values of all impact measures exhibit a downward trend, for both dpo and myopic policies, as the total resource availability increases from % to %. an increased total resource availability not only helps alleviating the pandemic impact inside the ongoing regions but also reduces the probability of spread to the unaffected regions. for both policies, as the total resource availability approaches the total resource requirement (starting from approximately %), the impact numbers show a converging behavior, whereby the marginal utility of additional resource availability diminishes. this behavior can be explained by noting that the total resource requirements were determined assuming the worst case scenario when all (four) regions would be affected and ought to provided with enough resources to cover their respective regional populations at risk. it can also be seen that on average, the dpo policy outperforms the myopic approach at all levels, which can attest to a more efficient resource utilization achieved by the dpo policy (see also table ). the difference in the policy performance is particularly noticeable at the lower levels of resource availability, and it gradually diminishes, as the resource availability increases and becomes closer to be sufficient to cover the entire populations at risk in all regions. it can also be noted that the variability in the performance of the dpo strategy is generally smaller than that of the myopic policy. in general, for both strategies, the performance variability decreases with higher availability of resources. in this section, we assess the marginal impact of variability of some of the critical factors. the impact was measured separately by the change in the total pandemic cost and the number of deaths (averaged over multiple replicates), resulting from a unit change in a decision factor value, one factor at a time. factors under consideration included: (i) antiviral efficacy, (ii) social distancing conformance, and (iii) cdc response delay. we have used all four regions, separately, as initial outbreak regions for each type of sensitivity analysis. the results (patterns) were rather similar. due to limited space, we have opted to show the results for only one initial region, chosen arbitrarily, for each of the three types of sensitivity studies. while duval county was selected as the initial outbreak region to show the sensitivity results on antiviral efficacy, hillsborough and miami dade were used as the initial regions to show the results on, respectively, social distancing conformance and cdc response delay. figure depicts the sensitivity of the average total cost and average total deaths to antiviral efficacy values between % and %. as expected, for both policies, the curves for the average number of deaths exhibit a decreasing trend which is almost linear for the values of τ between % and %. as the value of τ approaches %, the curves start exhibit a converging behavior. the curves for the average total pandemic cost exhibit a similar pattern for both policies. it can be noted that the performance of both policies is somewhat identical for low antiviral efficacy (between % and %). however, the performance of the dpo policy improves consistently as τ increases which can be attributed to a more discretionary allocation of the antiviral stockpile by the dpo policy. reduction of the contact intensity through quarantine and social distancing has proven to be one of the most effective containment measures, especially in the early stages of the pandemic [ , , , ] . figure shows the sensitivity of the average total cost and average total deaths to the social distancing conformance ranging between % and %. we observed that for both impact measures, the dpo policy demonstrated a better performance with the difference ranging from $ b to $ b in the total cost and from , to , in the number of fatalities. the biggest difference in performance was achieved at the lower-to-medium levels of conformance (between % and %). as the conformance level approached %, the dominating impact of social distancing masked the effect of better utilization of vaccines and antivirals achieved by the dpo strategy. the cdc response delay corresponds to the interval of time from the moment an outbreak is detected to a complete deployment of mitigation resources. depending on the disease infectivity, cdc response delay may represent one of the most critical factors in the mitigation process. figure shows how the performance of both policies was significantly impacted by this factor. the dpo policy showed a uniformly better performance with the difference ranging between $ b to $ b in the average total cost, and between to , in the average number of mortalities, over the range ( - figure : sensitivity analysis for cdc response delay. as recently pointed by the iom, the existing models for pi mitigation fall short of providing dynamic decision support which would incorporate "the costs and benefits of intervention" [ ] . in this paper, we present a large-scale simulation optimization model which is attempted at filling this gap. the model supports dynamic predictive resource distribution over a network of regions exposed to the pandemic. the model aims to balance both the ongoing and potential outbreak impact, which is measured in terms of morbidity, mortality, and social distancing, translated into the cost of lost productivity and medical expenses. the model was calibrated using historic pandemic data and compared to the myopic strategy, using a sample outbreak in fla, usa, with over million inhabitants. summary of the main results. in the testbed scenario, for both strategies, the marginal utility of additional resource availability was found to be diminishing, as the total resource availability approached the total requirement. in the testbed scenario, the dpo strategy on average outperformed the myopic policy. as opposed to the dpo strategy, the myopic policy is reactive, rather than predictive, as it allocates resources regardless of the remaining availability and the overall cross-regional pandemic status. in contrast, the dpo model distributes resources trying to balance the impact of actual outbreaks and the expected impact of potential outbreaks. it does so by exploiting regionspecific effectiveness of mitigation resources and dynamic reassessment of pandemic spread probabilities, using a set of regression submodels. hence, we believe that in scenarios involving regions with a more heterogeneous demographics, the dpo policy will likely to perform even better and with less variability than the myopic strategy. we also note that the difference in the model performance was particularly noticeable at lower levels of resource availability, which is in accordance with a higher marginal utility of additional availability at that levels. we thus believe that the dpo model can be particularly useful in scenarios with very limited resources. contributions of the paper. the simulation optimization methodology presented in this paper is one of the first attempts to offer dynamic predictive decision support for pandemic mitigation, which incorporates measures of societal and economic costs. our comparison study of the dpo versus myopic cross-regional resource distribution is also novel. additionally, our simulation model represents one of the first of its kind in considering a broader range of social behavioral aspects, including vaccination and antiviral treatment conformance. the simulation features a flexible design which can be particularized to a broader range of phi and npi and even more granular mixing groups. we also developed a decision-aid simulator which is made available to the general public through our web site at http://imse.eng.usf.edu/pandemics.aspx. the tool is intended to assist public health decision makers in implementing what-if analysis for assessment of mitigation options and development of policy guidelines. examples of such guidelines include vaccine and antiviral risk groups, social distancing policies (e.g., thresholds for declaration/lifting and closure options), and travel restrictions. limitations of the model. lack of reliable data prevented us from considering geo-spatial aspects of mixing group formation. we also did not consider the impact of public education and the use of personal protective measures (e.g., face masks) on transmission, again due to a lack of effectiveness data [ ] . we did not study the marginal effectiveness of individual resources due to a considerable uncertainty about the transmissibility of an emerging pandemic virus and efficacy of vaccine and antiviral. for the same reason, the vaccine and antiviral risk groups considered in the testbed can be adjusted, as different prioritization schemes have been suggested. the form of social distancing implemented in the testbed can also be modified as a variety of schemes can be found in the literature, including those based on geographical and social targeting. effectiveness of these approaches is substantially influenced by the compliance factor, for which limited accurate data support exists. it will thus be vital to gather the most detailed data on the epidemiology of a new virus and the population dynamics early in the evolution of a pandemic, and expeditiously analyze the data to adjust the interventions accordingly. cumulative number of confirmed human cases of avian influenza a(h n ) reported to who world health organization spanish flu pandemic influenza and the global vaccine supply vaccine production pandemic (h n ) vaccine deployment pandemic influenza preparedness and response preparing for pandemic influenza department of health & human services towards a quantitative understanding of the within-host dynamics of influenza a infections initial human transmission dynamics of the pandemic (h n ) virus in north america quantifying the routes of transmission for pandemic influenza little evidence for genetic susceptibility to influenza a (h n ) from family clustering data estimating variability in the transmission of severe acute respiratory syndrome to household contacts in hong kong, china the transmissibility and control of pandemic influenza a (h n ) virus detecting human-to-human transmission of avian influenza a (h n ) a bayesian mcmc approach to study transmission of influenza: application to household longitudinal data the role of the airline transportation network in the prediction and predictability of global epidemics coupled contagion dynamics of fear and disease: mathematical and computational explorations invited commentary: challenges of using contact data to understand acute respiratory disease transmission optimal vaccination policies for stochastic epidemics among a population of households optimal vaccination strategies for a community of households repeated influenza vaccination of healthy children and adults:borrow now, pay later? a population-dynamic model for evaluating the potential spread of drug-resistant influenza virus infections during community-based use of antivirals economic analysis of influenza vaccination and antiviral treatment for healthy working adults antiviral resistance and the control of pandemic influenza a small community model for the transmission of infectious diseases: comparison of school closure as an intervention in individual-based models of an influenza pandemic measures against transmission of pandemic h n influenza in japan in : simulation model responding to simulated pandemic simulation suggests that rapid activation of social distancing can arrest epidemic development due to a novel strain of influenza analysis of the effectiveness of interventions used during the a/h n influenza pandemic on estimation of vaccine efficacy using validation samples with selection bias targeted social distancing design for pandemic influenza living with influenza: impacts of government imposed and voluntarily selected interventions containing pandemic influenza with antiviral agents reducing the impact of the next influenza pandemic using household-based public health interventions transmissibility of pandemic influenza finding optimal vaccination strategies for pandemic influenza using genetic algorithms modeling the worldwide spread of pandemic influenza: baseline case and containment interventions using influenza-like illness data to reconstruct an influenza outbreak flute, a publicly available stochastic influenza epidemic simulation model models of infectious disease agent study strategies for containing an emerging influenza pandemic in southeast asia containing pandemic influenza at the source strategies for mitigating an influenza pandemic mitigation strategies for pandemic influenza in the united states modelling disease outbreaks in realistic urban social networks modeling targeted layered containment of an influenza pandemic in the united states report from the models of infectious disease agent study (midas) steering committee modeling community containment for pandemic influenza: a letter report interim pre-pandemic planning guidance: community strategy for pandemic influenza mitigation in the united states the economic impact of pandemic influenza in the united states: priorities for intervention a large scale simulation model for assessment of societal risk and development of dynamic mitigation strategies pandemic influenza: quarantine, isolation and social distancing spatial considerations for the allocation of pre-pandemic influenza vaccination in the united states statistical models and methods for lifetime data mathematical epidemiology of infectious diseases: model building, analysis and interpretation deaths from bacterial pneumonia during - influenza pandemic cdc influenza operational plan who guidelines on the use of vaccine and antivirals during influenza pandemics influenza a(h n ) monovalent vaccine public response to community mitigation measures for pandemic influenza precautionary behavior in response to perceived threat of pandemic influenza the immediate psychological and occupational impact of the sars outbreak in a teaching hospital the psychosocial effects of being quarantined following exposure to sars: a qualitative study of toronto health care workers influenza vaccination of health-care personnel confidence in vaccination: a parent model parental decisionmaking for the varicella vaccine exploring hepatitis b vaccination acceptance among young men who have sex with men: facilitators and barriers hepatitis b vaccine acceptance among adolescents and their parents why do parents decide against immunization? the effect of health beliefs and health professionals reassessment of the association between guillain-barré syndrome and receipt of swine influenza vaccine in - : results of a two-state study psychosocial determinants of immunization behavior in a swine influenza campaign the fear factor doctors swamped by swine flu vaccine fears a simulation based learning automata framework for solving semi-markov decision problems under long run average reward supplemental data and model parameter values for cross-regional simulation-based optimization testbed american community survey national household travel survey (nths) tampa international airport: daily traffic volume data miami international airport: daily traffic volume data miami international airport: daily traffic volume data tallahassee regional airport: daily traffic volume data how thermal-imaging cameras spot flu fevers writing committee of the world health organization (who) antivirals for pandemic influenza: guidance on developing a distribution and dispensing program safety and immunogenicity of an inactivated subvirion influenza a (h n ) vaccine attitudes toward the use of quarantine in a public health emergency in four countries design and analysis of experiments inflation calculator nonpharmaceutical interventions for pandemic influenza, national and community measures the authors would like to acknowledge with thanks the many helpful suggestions made by professor yiliang zhu, department of epidemiology and biostatistics at the university of south florida, tampa, fla, usa. key: cord- -qw tusd authors: krishna, smriti m.; omer, safraz mohamed; li, jiaze; morton, susan k.; jose, roby j.; golledge, jonathan title: development of a two-stage limb ischemia model to better simulate human peripheral artery disease date: - - journal: sci rep doi: . /s - - - sha: doc_id: cord_uid: qw tusd peripheral arterial disease (pad) develops due to the narrowing or blockage of arteries supplying blood to the lower limbs. surgical and endovascular interventions are the main treatments for advanced pad but alternative and adjunctive medical therapies are needed. currently the main preclinical experimental model employed in pad research is based on induction of acute hind limb ischemia (hli) by a -stage procedure. since there are concerns regarding the ability to translate findings from this animal model to patients, we aimed to develop a novel clinically relevant animal model of pad. hli was induced in male apolipoprotein e (apoe(−/−)) deficient mice by a -stage procedure of initial gradual femoral artery occlusion by ameroid constrictors for days and subsequent excision of the femoral artery. this -stage hli model was compared to the classical -stage hli model and sham controls. ischemia severity was assessed using laser doppler perfusion imaging (ldpi). ambulatory ability was assessed using an open field test, a treadmill test and using established scoring scales. molecular markers of angiogenesis and shear stress were assessed within gastrocnemius muscle tissue samples using quantitative polymerase chain reaction. hli was more severe in mice receiving the -stage compared to the -stage ischemia induction procedure as assessed by ldpi (p = . ), and reflected in a higher ischemic score (p = . ) and lower average distance travelled on a treadmill test (p = . ). mice undergoing the -stage hli also had lower expression of angiogenesis markers (vascular endothelial growth factor, p = . ; vascular endothelial growth factor- receptor , p = . ) and shear stress response mechano-transducer transient receptor potential vanilloid (p = . ) within gastrocnemius muscle samples, compared to animals having the -stage hli procedure. mice subjected to the -stage hli receiving an exercise program showed significantly greater improvement in their ambulatory ability on a treadmill test than a sedentary control group. this study describes a novel model of hli which leads to more severe and sustained ischemia than the conventionally used model. exercise therapy, which has established efficacy in pad patients, was also effective in this new model. this new model maybe useful in the evaluation of potential novel pad therapies. major amputation, renal failure and death) and poor long-term durability [ ] [ ] [ ] [ ] . there is great interest in developing novel medical therapies for the leg symptoms of pad. recent efforts have focused on stimulating development of new blood vessels within the leg through angiogenesis or by encouraging the remodelling of existing small vessels into improved collateral channels (arteriogenesis). promising results for novel treatments, such as viral vectors carrying angiogenesis promoting agents and stem cells, in pre-clinical models of pad have not been consistently replicated in large clinical trials - . in patients that have pad, atherosclerosis-associated arterial narrowing develops gradually over many years allowing the legs to adjust to the gradual decrease in blood flow through compensatory mechanisms within the blood vessels and muscle fibres . in contrast, the most commonly used animal model for initial testing of novel therapies for pad is a model of acute blood supply interruption through ligation or excision of the femoral artery (referred to here as the -stage hind limb ischemia (hli) model) , . previous studies report that the ligation and excision of the femoral artery in the -stage model leads to increased fluid shear stress within the limb collateral arteries resulting in altered gene expression patterns through shear stress responsive elements which promote arterio-and angio-genesis [ ] [ ] [ ] . hind limb blood supply in this -stage model therefore usually naturally recovers over a period of approximately weeks . this model does not therefore simulate the clinical presentation of pad. patients typically present with a history of acute exacerbation of chronic symptoms of leg pain on walking and have ongoing ischemic symptoms. the -stage hli model may therefore not be an ideal model to study therapeutic angiogenesis and arteriogenesis , . another approach to inducing hli is the placement of an ameroid constrictor around the femoral artery to induce gradual occlusion , , . previous studies suggest that this approach leads to mild ischemia and that blood flow recovery occurs within - weeks , . furthermore, previous pre-clinical pad research has mainly focused on assessing hind limb blood supply with limited assessment of ambulatory ability , . on the other hand, the assessment of novel treatments in pad patients usually involves measures of walking ability using treadmill or corridor walking tests , . there is therefore a need for an increased focus on functional tests of the limb within clinically-relevant rodent models. we hypothesised that the limb ischemia produced by the current -stage hli model would be more severe and sustained if the model was modified to include an initial more slowly progressive arterial narrowing over days prior to the induction of acute ischemia (i.e. a -stage model). our overall aim was to develop a more clinically relevant rodent model that could incorporates stable on-going limb ischemia in order to test therapeutic interventions. mice. male apolipoprotein e deficient (apoe −/− ) mice (n = , obtained from animal resources centre, western australia) were used for the experiments. mice were housed in a temperature-controlled room ( ± °c) with an automatic : -h light/dark cycle ( : to : hours). mice were singly housed in a clear individually-ventilated, temperature and humidity-controlled cage system (aero ivc green line; tecniplast) with enrichment. all experiments were performed during the light phase ( : - : hours) and mice were fed with standard rodent chow and water ad libitum during the course of these experiments. approval for the animal studies was obtained from the institutional ethics committee (animal ethics committee, james cook university) and experimental work performed in accordance with the institutional and ethical guidelines of james cook university, australia, and conforming to the guide for the care and use of laboratory animals (national institutes of health, usa). hli models. the first phase of the study utilised two hli models: the most commonly used unilateral acute hli model ( -stage hli) , , and the newly developed -stage hli model. male apoe −/− mice aged months were randomly divided into groups as follows: group = -stage hli model (n = ), group = -stage sham (n = ), group = -stage hli model (n = ) and group = -stage sham (n = ). body weight and primary outcome measures were recorded at regular intervals as illustrated in fig. a . all functional assessments were performed in a subset of mice randomly selected from each experimental group (n = - ). the creation of the -stage hli model involved exposure of the left femoral artery through a vertical . - cm skin incision under a stereotactic microscope (leica). the femoral artery and its side branches were then ligated with - silk sutures (ethicon) immediately distal to the inguinal ligament and proximal to the popliteal bifurcation before being excised ( supplementary fig. s a,b) . femoral nerves were carefully preserved. the wound was irrigated with sterile saline and then the overlying skin was closed using - vicryl sutures (ethicon). post-operative pain was reduced using lignocaine (troy laboratories). a similar surgery without ligation or excision of the femoral artery was performed on the -stage sham controls. the -stage hli model was performed using a -stage surgical procedure. the left femoral artery was exposed as described above and custom made miniature ameroid constrictors of . mm internal diameter (research instruments sw) were positioned on the artery. one was placed on the femoral artery immediately distal to the inguinal ligament and one was positioned proximal to the sapheno-popliteal bifurcation ( supplementary fig. s c ). after days, a new incision was made and the femoral artery ligated and excised, as described for the -stage hli model. a similar two-stage surgery was performed without placement of ameroids, nor ligation and excision of the femoral artery for the -stage sham controls. assessment of the effect of exercise training in the -stage hli model. during the second phase of the study, the effect of an exercise program on male apoe −/− mice aged months undergoing the -stage hli model (male apoe −/− mice, n = ) was tested. mice subjected to -stage hli were randomly allocated to an exercise training or control group (n = per group). mice in the exercise group received to m (between - mins of running wheel access) of exercise each day on a running wheel ( station home cage running the movement of animals were monitored and functional scores of the various groups were assessed according to the scoring criteria detailed in the materials and methods section. the -stage hli model showed reduced function throughout the experimental period compared to the -stage hli model. data shown as mean ± sem and analysed by repeated measures -way anova, and p value significant at ≤ . . (e) graph showing modified ischemia scores. the animals were monitored and scored for signs of ischemia according to previously published criteria detailed in the materials and methods. the -stage hli model showed a higher level of ischemia compared to the -stage hli depicted by a lower scoring throughout the study period. data shown as median ±sem and analysed by repeated measures -way anova, and p value significance set at ≤ . . suggested that the hli typically resolved over the course of days, with hind limb blood supply reaching a plateau between and days after surgery . the ldpi measurements were therefore performed at the following time points for -stage hli model: day prior to surgery, immediately after surgery, days , , , and after surgery. for the -stage hli model, the ldpi measurements were performed at the following time points: day prior to the first operation (ameroid placement), immediately after the first operation, days , , and after the first operation, immediately after the second operation (femoral artery excision), and , , and days after the second operation (i.e. , , , and days after the first operation). a schematic illustration of the experimental design is shown in fig. a . body mass was also measured on the same day as the ldpi measurements. in clinical practice pad patients are treated to improve pain free walking capacity and resolve rest pain and tissue loss (critical limb ischemia, cli). in clinical trials, these are usually investigated by walking tests and assessment of pain. similar to clinical trials, in the current study ambulatory ability was assessed with a treadmill test, voluntary physical activity examined through an open field test and foot pain estimated through a mechanical allodynia test. all outcomes were assessed by an assessor blinded to mice group. semi-quantitative assessments of limb function and ischemia were performed at the same time points as the blood flow measurements ( fig. a ; supplementary tables s , s ). limb function was assessed using the clinical use score (tarlov scale) as: = no movement; = barely perceptible movement, no weight bearing; = frequent and vigorous movement, no weight bearing; = supports weight, may take or steps; = walks with only mild deficit; = normal but slow walking and = full and fast walking , . limb ischemia was scored using the ischemia scoring scale as previously reported: = auto-amputation of leg; = leg necrosis; = foot necrosis; = two or more toe discoloration; = one toe discoloration; = two or more nail discolorations; = one nail discoloration and = no necrosis . all scoring was performed by two independent observers and found to be identical. treadmill test. mice were run on a six lane excer / treadmill (columbus instruments) without incline. mice were acclimatised to the treadmill by ambulating on it at m/min for min once daily on three consecutive days prior to any testing. before each treadmill test, mice were fasted for h. the speed of the treadmill was controlled using the software and calibrated using an inbuilt speedometer mounted on the treadmill platform. a treadmill test involved an initial warm up at m/min for min followed by a progressive speed increase from to m/min, accelerated at m/min. following this the treadmill speed remained at m/min for up to a total running time of min. during the test a stimulus grid of hz was kept on until mouse exhaustion as previously reported . exhaustion of the mouse was defined if the mouse returned to the stimulus grid times despite a hz electrical stimulus to encourage walking on the belt. the treadmill software recorded the total distance walked by a mouse until exhaustion. a blinded observer supervised the experiment to assess outcomes. the treadmill belt and lanes were cleaned with water and % alcohol and dried with paper towel after each test to remove any body scent. treadmill testing was carried out before ameroid placement, days after ameroid placement, and and weeks after completion of the -stage hli (fig. a) . for the -stage hli model, treadmill testing was performed before ligation and excision of the femoral artery and and weeks after ischemia induction. voluntary physical activity test. the open field test is a common measure of voluntary physical activity in rodents suggested to be similar to a -min walk test used in humans . to ensure consistency prior to the test, mice were brought to the testing room in their home cages at least hr prior to the start of behavioural testing. the mice were fasted during the acclimatisation period and given free access to water under normal lighting. the open field box was made of opaque plastic ( × × cm), divided into an outer field (periphery) and a central field ( × cm) which was primarily used for analysis. mice were individually placed in the centre of the arena and movements of the mice were recorded using a video camera (logitech) supported with acquisition software (capture star ver. ; cleversys inc) and analysed by the topscan lite software (high throughput version . ; cleversys inc). the test protocol used was identical for each mouse assessed. after each test the open field box was cleaned with water and % alcohol and dried with a paper towel to remove the body scent, which could be a cue to movement of the mice. room lighting, temperature, and noise levels were kept consistent for all tests. the mouse movements were recorded for min, to mimic the short timed nature of the -min walk test. rest time was recorded as motion measure score < . in the software and average speed was calculated only for motion measure score > . . total distance travelled (m), frequency of movement, time spent in the arena (s) and average velocity in the arena (mm/s) were measured. mechanical allodynia test. the paw pressure transducer and the pressure application measurement device (pam; ugo basile) is a non-invasive tool for measuring mechanical allodynia threshold and hypersensitivity in rodents. the pam device allows an accurate measurement of primary mechanical hypersensitivity in rodents , . a gradually increasing squeeze force is applied across the joint at a rate of approximately gms/sec until a behavioural response (paw withdrawal, freezing of whisker movement, wriggling or vocalization) is observed with a cut-off of sec. the peak gram force (gf) applied immediately prior to limb withdrawal was recorded by the base unit, and this value was designated the limb withdrawal threshold (lwt). lwt was measured twice in both the ipsilateral and contralateral limbs by two independent observers. the measurements were averaged and presented as a ratio of operated left limb to the un-operated right limb. blood tests. blood was collected by cardiac puncture at the completion of the experiments. platelet poor plasma was separated as described previously , . the plasma concentrations of interleukin (il)- , interferon (ifn)-γ, monocyte chemoattractant protein- (mcp- ) and tumour necrosis factor (tnf)-α were determined scientific reports | ( ) : | https://doi.org/ . /s - - - www.nature.com/scientificreports www.nature.com/scientificreports/ using a cytometric bead array kit (cba, bd biosciences). the inflammatory markers were assessed in samples (n = /group) selected from each group using a random number generator. briefly, μl of mixed capture beads and μl of serially diluted standard or plasma sample and μl of phycoerythrin (pe) detection reagent, were incubated in the dark for h in sample assay tubes. samples were then washed twice with ml of the wash buffer, resuspended, and acquired on the cyan adp flow cytometer (beckman coulter). results were analysed and quantified by fcap array ™ software (v , bd biosciences). we previously reported this method to have good reproducibility with an inter-assay coefficient of variation of - % (n = - ) . total nitrate was measured in plasma samples by a nitrate/nitrite colorimetric assay kit following the manufacturer's protocol (inter-assay coefficient of variation . %; cayman chemicals) as reported previously . briefly, nitrate was converted to nitrite using nitrate reductase. subsequently, addition of the griess reagents converted nitrite into a deep purple azo compound and the absorbance was measured at nm using an omega plate reader. histological assessments. low capillary density has been reported in the gastrocnemius muscle of pad patients and animal hli models and associated with functional impairment , . hence at the end of experiments gastrocnemius muscle samples were collected from the mice and stored in optimal cutting compound (oct, proscitech) which was progressively frozen in isopentane (sigma) suspended in liquid nitrogen. sections ( µm-thick) were obtained on poly l-lysine coated slides (proscitech) from each sample with muscle fibres oriented in the transverse direction. all histological assessments were performed on sections that were examined in a blinded fashion at x magnification. muscle fibre structure. fixed cryostat sections were stained with hematoxylin and eosin (h&e, proscitech), examined at magnifications of x or x to assess the integrity of the tissues. degenerating muscle fibres were identified in the h&e stained sections by morphological assessment. assessors looked for the presence of mature skeletal muscle fibres (small peripheral nuclei) versus immature skeletal muscle myoblasts (large lobulated central nuclei) . muscle fibre number and size were examined in separate fields in distinct areas in each specimen. muscle fibrosis. the extent of skeletal muscle fibrosis was assessed by staining the cryostat sections ( µm-thick) with picrosirius (proscitech). briefly, tissue sections were stained and examined under x power light microscope and skeletal muscle fibrosis was analysed using the image analysis software (zeiss axio imager z ). quantification of fibrosis was expressed as the percentage of fibrotic tissue present within the section ( mm tissue area) using a previously published protocol . immunohistochemistry and morphometric analysis of capillary and arteriolar density. these were performed as previously reported , . the gastrocnemius muscles from ischaemic and non-ischaemic hind-limbs were collected and embedded in oct compound (proscitech), frozen, and cut into µm-thick sections. the slides were fixed at − °c in % ethanol for hr. slides were washed three times in cold pbs with % horse serum ( min/ wash) and blocked overnight with % horse serum in pbs at °c. immunohistochemistry was performed using primary antibodies against cd ( : dilution; abcam) and smooth muscle α-actin (α-sma, : dilution; abcam). bound primary antibodies were detected by using appropriate secondary antibodies (biotinylated anti-goat igg and biotinylated anti-rat igg, all at : dilutions, vector labs) using avidin-biotin-peroxidase (vector labs) as described previously . pictures from four random areas of each section and three sections per mouse were taken by using a digital camera (nikon eclipse sci epifluorescence microscope, nikon corporation) at × magnification. capillary density were quantified by measuring the percentage of cd and α-sma staining out of the total area as previously described. western blotting. gastrocnemius muscles were mainly harvested at the end of studies (i.e. weeks after full ischemia induction). gastrocnemius muscles were also harvested from a subset of mice (n = ) subjected to the -stage hli (n = /time-point) prior to and days and after ameroid placement. samples were frozen in liquid nitrogen, and stored in oct compound (proscitech) at − °c. tissues were pulverised in ripa buffer ( mm sodium chloride, . % np- or triton x- , . % sodium deoxycholate, . % sodium dodecyl sulfate, mm tris, ph . ) with protease inhibitors (roche diagnostics, australia) and phostop (roche diagnostics, australia) to extract proteins and quantitated using the bradford protein assay kit (biorad, usa). samples ( μg of protein/lane) were loaded onto a % sds-polyacrylamide electrophoresis gel. after electrophoresis ( v, min), the separated proteins were transferred ( ma, min) to a polyvinylidene difluoride membrane (biorad, usa). non-specific sites were blocked with % non-fat dry milk for min, and the blots were then incubated with following antibodies: anti-vascular endothelial growth factor (anti-vegf www.nature.com/scientificreports www.nature.com/scientificreports/ mrna analysis by quantitative real-time pcr. at the end of the study gastrocnemius muscle samples were harvested, placed in rna later (qiagen) and stored at − °c. samples (n = /group) were selected from each group using a random number generator and were processed for gene expression analysis. total rna was isolated using an rneasy mini kit (qiagen) according to manufacturer's instructions and quantified spectrophotometrically using nanodrop . rna samples ( ng) were subjected to quantitative real time pcr (qrt-pcr) analysis of genes of interest using the quantitect sybr green one-step rt-pcr assay (qiagen). qrt-pcr was performed using primers for mouse vegf-r (ppm f), vegf-r (ppm a), trpv (ppm a), klf (ppm b) and gapdh (qt ). the relative expression of these genes were calculated by using the concentration-ct-standard curve method and normalized using the average expression of mouse gapdh for each sample using the rotor-gene q operating software (version . . ) as previously reported , . statistical analyses. all data were tested for normality using the d' agostino-pearson normality test. data with normal distribution were expressed as mean ± standard error of mean (sem) and analysed using parametric tests. non-normally distributed data were expressed as median and interquartile ranges (iqr) and analysed using non-parametric tests. statistical significance was determined using the unpaired student t test for comparison between two groups or analysis of variance followed by student-newman-keuls post-hoc analysis for comparison between multiple groups. comparison of the time course of ldpi indices, clinical scores, open field tests and treadmill exercise tests were done by -way anova for repeated measures, followed by bonferroni post hoc analysis or by linear mixed effect method using r studio software. difference in the clinical ischemia score were determined by fisher's exact test. analyses were performed using prism (graphpad software, san diego, ca) or r software. a p value of ≤ . was considered to be statistically significant. mice undergoing -stage hli had more severe ischemia than those undergoing -stage hli. immediately after femoral artery excision, limb perfusion assessed by ldpi was similarly reduced in both hli models by approximately %. mice subjected to the -stage hli had more rapid recovery of hind limb perfusion than those subjected to the -stage procedure (p = . , fig. b,c, supplementary fig. s ). by days after ischemia induction limb perfusion was similar in mice subjected to the -stage procedure and sham controls (p = . ) but still reduced in mice subjected to the -stage procedure by comparison to sham controls (p < . ). there was no change in overall body mass after ischemia induction (supplementary fig. s ). mice subjected to -stage hli showed more severely impaired hind limb use than those undergoing -stage hli. after ischemia induction, mice subjected to both methods of hli developed limb oedema, paleness of skin and occasional muscle necrosis. mice in all experimental groups exhibited a severe functional deficit after surgery (fig. d) . functional score was significantly worse in mice subjected to the -stage hli than those subjected to the -stage hli (p = . ). both the hli models showed increased ischemia compared to the respective shams. there were no cases of auto-amputation or foot or limb necrosis (supplementary fig. s ). mice subjected to the -stage hli showed a significantly worse ischemic score compared to those subjected to the -stage hli (repeated measures way anova, p = . , fig. e ). mice subjected to -stage hli had reduced treadmill performance. mice subjected to the -stage hli showed no significant difference in total distance travelled during the study period on a treadmill test when compared to shams (repeated measures way anova, p = . ; fig. a ). after the first procedure of the -stage model (ameroid placement) the treadmill ambulatory performance of mice was not significantly affected ( supplementary fig. s a ). in contrast mice subjected to -stage hli had a significant reduction in total distance travelled on the treadmill compared to their sham controls (p = . ) and mice subjected to -stage hli (p = . ; fig. a ). supplementary fig. s b-d) . this reduction in physical activity was maintained after completing the -stage hli by comparison to sham controls and also mice subjected to -stage hli (fig. b) . the reduction in physical activity of mice subjected to -stage hli was reflected in less distance travelled, less total time in motion and lower velocity of the movement compared to sham controls (fig. c-e) . when compared to the -stage hli, the stage-hli model showed a reduction in the total distance travelled in the open field arena (linear mixed effect model, p = . ). mice subjected to hli had enhanced mechanical allodynia. mice subjected to both -stage and -stage hli showed significantly increased sensitivity to pressure compared their respective sham controls and there was no significant difference in pressure sensitivity between the two models (fig. f ). hli induces systemic inflammation. the plasma concentrations of the cytokines assessed were below the detectable ranges in both the sham control groups (table ) . mice subjected to hli, irrespective of model, had plasma cytokine concentration significantly higher than the sham controls although levels were not significantly different between models ( table ). the plasma concentrations of nitric oxide (no) metabolites were higher in mice undergoing -stage hli than the sham control group (p < . ) and mice undergoing -stage hli (p = . , table ). myofibers which are healthy and functionally active are characterised by peripheral nuclei, while myofibrils with central nuclei are immature and do not show optimal contraction . in gastrocnemius muscle samples removed from sham controls, myocytes were angular with peripheral nucleus (fig. a) . at day after ischemia induction, mice subjected to -stage hli showed microscopic changes such as cellular swelling, focal necrosis and interstitial oedema. there were also numerous infiltrating inflammatory cells. gastrocnemius muscle samples from mice subjected to -stage hli showed more homogenous appearance with all myocytes showing peripheral nuclei and limited inflammatory cell infiltration (fig. a) . histological evaluation revealed that tissues of mice subjected to -stage hli had fewer immature myofibers cells than the tissues from -stage hli model. furthermore, muscle samples from mice subjected to -stage had prominent oedema, myofibre separation and multifocal neutrophilic infiltration. neutrophils were observed throughout the tissue sections. necrotic muscle fibres were prominent and formed confluent areas (fig. a) . muscle fibrosis was assessed by picrosirius red staining which suggested that -stage hli led to increased skeletal muscle fibrosis compared to -stage hli (p = . ) or sham controls (p = . ; fig. b ,e). mice subjected to -stage hli had fewer hind limb collaterals. consistent with the reduced perfusion as assessed by ldpi, both arteriogenesis and angiogenesis was inhibited in the ischemic gastrocnemius muscles of the -stage hli model (fig. c,d) . measurement of angiogenesis by cd immunostaining showed that the presence of arterioles was significantly reduced in samples from the -stage hli compared to the -stage hli (p = . ) or sham controls (p = . ; fig. ). the arteriolar density was also significantly less in samples from the -stage hli model compared to the -stage hli (p = . ) or sham control (p = . ; fig. ). protein concentrations of angiogenesis and shear stress response markers were downregulated in the gastrocnemius muscles of mice undergoing -stage hli. the relative total vegf and vegfr- (but not vegfr- , p-enos/enos and hif-α) protein levels in gastrocnemius tissue collected from mice weeks after -stage hli were significantly less than in sham controls and mice undergoing -stage hli (fig. a-d, supplementary fig. s ). analysis of tissues from the -stage hli model prior to ameroid placement and day and day after ameroid placement suggested no significant changes in concentrations of trpv or vegf in response to ameroid constriction ( supplementary fig. s ). at the end of the experiment, protein concentrations of trpv and klf were significantly less in the gastrocnemius muscle samples from mice undergoing -stage hli compared to sham controls and mice undergoing -stage hli (fig. e,f) . qrt-pcr showed that the relative expressions of vegf-r , trpv and klf in gastrocnemius muscle of mice subjected to -stage hli were significantly lower than within the gastrocnemius muscle of mice undergoing -stage hli or in sham controls (fig. g-j) . exercise training improved functional capacity but not limb perfusion in mice subjected to -stage hli. supervised exercise training is an established method to improve functional capacity in pad patients , and previous studies show that mice respond positively to exercise training , . in order to examine whether an established clinically effective therapy was effective in the novel animal model, the effect of exercise training (using a running wheel) on functional capacity of mice subjected to -stage hli was assessed. exercise training was commenced days after ischemia induction. exercise training did not affect limb perfusion as shows the data from the ischemic limbs from the -stage and -stage hli models and the respective sham controls (n = /group, scale bars in all images = µm). (e-g) quantitative bar graphs showing the effect of hli on (e) skeletal muscle fibrosis by picrosirius red staining, (f) angiogenesis by immunohistochemical staining against cd and (g) arteriogenesis by immunohistochemical staining against α-sma. all values are median and interquartile ranges (n = /group) and p value significance set at ≤ . . ( ) : | https://doi.org/ . /s - - - www.nature.com/scientificreports www.nature.com/scientificreports/ www.nature.com/scientificreports www.nature.com/scientificreports/ assessed by ldpi (linear mixed effect test, p = . ; fig. b ,c). mice subjected to exercise training showed a significant increase in average treadmill walking capacity compared to sedentary controls (p = . ; fig. d ). exercise training upregulated gastrocnemius muscle vegf and trpv levels. since improvement in ambulatory performance as a result of exercise training could be due to enhanced angiogenesis, the expression of angiogenesis and shear stress responsive proteins vegf, vegf-r , vegf-r , trpv and klf were assessed in the gastrocnemius muscles (fig. ). vegf (p = . , fig. b ) and trpv (p = . , fig. e ) but not vegf-r , vegf-r and klf , were highly upregulated following exercise training (fig. c,d,f ). this report describes the development of a novel model of hli which results in more severe and prolonged ischemia than the traditional model. mice subjected to the -stage hli had functional and ambulatory impairment and a positive response to exercise training as has been reported for pad patients . previous rodent studies suggest that placing of ameroid constrictors alone without manipulating the femoral artery results in mild ischemia and blood flow recovers within weeks . hence, the novel hli model was based on placement of two ameroid constrictors on the femoral artery to promote gradual occlusion followed by excision of intervening segment after days to induce severe ischemia. a previous report suggests that recovery of hind limb blood flow is reduced in apoe −/− compared to c bl/ j mice due to their limited collateral arteries and hence apoe −/− mice were used in the current study . pad patients are generally older and exhibit metabolic derangements that limit angio-and arterio-genesis and previous studies suggest that apoe −/− mice have delayed skeletal muscle healing , reduced angiogenesis responses and impaired functional recovery after hli further supporting the rationale for choosing this mice species , , . mice subjected to -stage hli had rapid recovery of limb blood flow and reached a perfusion level similar to the sham controls within days as has been reported by other investigators , , . ligation and sudden excision of the femoral artery is believed to generate a pressure gradient between the proximal and distal ends of the occluded vessel, resulting in increased shear stress and a redirection of blood flow towards the collaterals and through numerous branches arising from the internal iliac artery, resulting in rapid improvement in blood flow , . ameroid constrictors have been shown to cause luminal occlusion within days, however, the blood flow slowly increases in the next - weeks , . hence, in the new model, we superimposed a secondary acute event by excising the intervening femoral artery segment along with the ameroid constrictors. in contrast to the -stage hli, mice subjected to -stage hli had on-going limb ischemia and a prolonged functional deficit on both forced and voluntary ambulation tests. these findings support the value of the novel model for the testing of interventions aimed at achieving clinical improvements in pad patients. since angiogenesis is an inflammation-driven process , the concentrations of circulating cytokines were measured. these markers of systemic inflammation increased in response to hli in both models examined. twenty eight days after hli induction, the concentrations of circulating cytokines were similar in the two models studied. gastrocnemius muscle samples obtained from the -stage hli model had marked neutrophilic infiltration. it has been previously reported that inflammatory cells accumulate in hypoxic tissues and promote angiogenesis . it is possible that the systemic concentrations of cytokines were not reflective of the level of inflammation within the hind limb. markers of angiogenesis and arteriogenesis in gastrocnemius tissue, such as cd and α-smc, were found to be less evident in the -stage than the -stage hli model. furthermore, gastrocnemius vegf and vegf-r protein levels and amount of total plasma no metabolites were significantly lower in the -stage than -stage hli model. vegf promotes angiogenesis through binding to vegf-r expressed on endothelial cells . vegf induces the release of no thereby promoting microvascular perfusion and endothelial progenitor cell mobilization [ ] [ ] [ ] . endothelial cell derived microrna, such as mir- , have also been implicated in controlling angiogenesis through inhibiting rho gdp dissociation inhibitor (rhogdi)-α, an important regulator of enos phosphorylation . it appears likely that the low levels of pro-angiogenic markers in the -stage hli model reflect less activation of endothelium-dependent pro-angiogenesis signalling pathways which were stimulated by collateral flow within the -stage model. shear stress promotes arteriogenesis by stimulating remodelling of collaterals , . endothelial cells transduce changes in shear stress into intracellular signals which promote expression of a distinct set of genes which can control the response to ischemia [ ] [ ] [ ] . previous studies suggest that increased shear stress promotes phosphorylation and upregulation of mechano-sensors, such as trpv [ ] [ ] [ ] [ ] . it was postulated that the distinct ways of inducing hli in the two models studied might be reflected in different trpv expression. mice subjected to -stage hli had lower expression of trpv compared to those undergoing -stage hli. furthermore, in the -stage hli model, there was no change in trpv protein levels after ameroid placement, suggesting that ameroid constriction is a gradual process that does not lead to shear stress changes capable of stimulating mechano-sensors like trpv . these findings also suggest that -stage hli results in more limited collateral flow than the -stage approach. the acute reduction in arterial pressure gradient following femoral artery excision in the -stage hli model it thought to be registered by endothelial shear stress response elements resulting in upregulation of angiogenesis and arteriogenesis promoting genes . simulation of trpv , for example by α-phorbol , of vegf-r (g), vegf-r (h), trpv (i) and klf (j) assessed in the gastrocnemius muscles of mice undergoing hli. quantitative real time pcr (qrt pcr) was performed on extracted total mrna using specific primers and normalised to glyceraldehyde phosphate dehydrogenase (gapdh) expression. data analysed by mann-whitney u test (n = samples/group). ( ) : | https://doi.org/ . /s - - - www.nature.com/scientificreports www.nature.com/scientificreports/ -didecanoate, has been reported to promote increase no release and increased hind limb blood flow . trpv deficient rodents have impaired vasodilatation [ ] [ ] [ ] [ ] . this supports the important role of trpv in promoting adaptation to acute limb ischemia. four weeks of exercise training led to an approximate -fold increase in treadmill ambulatory distance in the mice subjected to -stage hli, paralleling findings from pad patients [ ] [ ] [ ] [ ] . these findings suggest the novel -stage hli mouse model simulates the walking impairment experienced by patients. in support of the relevance of this model to patients, we found that exercise therapy improved treadmill performance without improving limb perfusion, a finding similar to that described in pad patients , , . since exercise training increases shear stress in pre-existing collaterals we examined the expression of trpv and angiogenesis markers . compared to sedentary controls, mice undergoing exercise training showed increased total protein levels of vegf and trpv , which was in accordance with previous reports showing enhanced expression of pro-angiogenesis markers after exercise training [ ] [ ] [ ] . overall these findings suggest flow-mediated upregulation of shear stress responsive genes is important in stimulating angiogenesis responses in the new model. this study have several strengths and weaknesses. the -stage hli model which is usually utilised for pad research has many limitations including its disparate pathophysiological mechanisms compared to patient presentations, the temporary nature of the ischemia and its relative responsiveness to a variety of therapies which are not effective in patients . this study suggests the novel -stage model has clear advantages over the -stage model since ischemia is more severe and prolonged and does not naturally recover making it suitable to access the longer term effects of potential therapies. the study also has some limitations. these include the absence of a detailed time course of the genes and proteins assessed and the absence of a definitive test of the role of shear stress response elements in the differences in the models found. the importance of shear stress response elements in the response to hli has however been previously reported in prior studies . it is therefore logical to assume similar mechanisms were involved in the current study given the gene expression differences demonstrated. the study also lacked the use of new high resolution imaging systems, such as fluorescent microsphere angiography. lower-limb arterial disease comparison of global estimates of prevalence and risk factors for peripheral artery disease in and : a systematic review and analysis the global pandemic of peripheral artery disease global and regional burden of death and disability from peripheral artery disease: world regions endovascular versus open revascularization for peripheral arterial disease outcomes of polytetrafluoroethylene-covered stent versus bare-metal stent in the primary treatment of severe iliac artery obstructive lesions trial of a paclitaxel-coated balloon for femoropopliteal artery disease drug-coated balloon versus standard percutaneous transluminal angioplasty for the treatment of superficial femoral and popliteal peripheral artery disease: -month results from the in.pact sfa randomized trial autologous cell therapy for peripheral arterial disease: systematic review and meta-analysis of randomized, nonrandomized, and noncontrolled studies growth factors for angiogenesis in peripheral arterial disease trends in the national outcomes and costs for claudication and limb threatening ischemia: angioplasty vs bypass graft the australian national workplace health project: design and baseline findings muscle fiber characteristics in patients with peripheral arterial disease evaluation of the clinical relevance and limitations of current pre-clinical models of peripheral artery disease the efficacy of extraembryonic stem cells in improving blood flow within animal models of lower limb ischaemia the effect of gradual or acute arterial occlusion on skeletal muscle blood flow, arteriogenesis, and inflammation in rat hindlimb ischemia blood flow and vascular gene expression: fluid shear stress as a modulator of endothelial phenotype biomechanical activation of vascular endothelium as a determinant of its functional phenotype a review of the pathophysiology and potential biomarkers for peripheral artery disease subacute limb ischemia induces skeletal muscle injury in genetically susceptible mice independent of vascular density cellular and molecular mechanism regulating blood flow recovery in acute versus gradual femoral artery occlusion are distinct in the mouse physical activity during daily life and mortality in patients with peripheral arterial disease skeletal muscle perfusion in peripheral arterial disease a novel end point for cardiovascular imaging impaired collateral vessel development associated with reduced expression of vascular endothelial growth factor in apoe −/− mice murine model of hindlimb ischemia delayed arteriogenesis in hypercholesterolemic mice vascular growth factor expression in a rat model of severe limb ischemia mouse model of angiogenesis critical function of bmx/etk in ischemia-mediated arteriogenesis and angiogenesis limb ischemia after iliac ligation in aged mice stimulates angiogenesis without arteriogenesis exercise performance and peripheral vascular insufficiency improve with ampk activation in high-fat diet-fed mice altered behavioural adaptation in mice with neural corticotrophin-releasing factor overexpression amitriptyline reverses hyperalgesia and improves associated mood-like disorders in a model of experimental monoarthritis validation of the digital pressure application measurement (pam) device for detection of primary mechanical hyperalgesia in rat and mouse antigen-induced knee joint arthritis a peptide antagonist of thrombospondin- promotes abdominal aortic aneurysm progression in the angiotensin ii-infused apolipoprotein-e-deficient mouse influence of apolipoprotein e, age and aortic site on calcium phosphate induced abdominal aortic aneurysm in mice impaired acetylcholine-induced endothelium-dependent aortic relaxation by caveolin- in angiotensin ii-infused apolipoprotein-e (apoe −/− ) knockout mice relationship between leg muscle capillary density and peak hyperemic blood flow with endurance capacity in peripheral artery disease chronically ischemic mouse skeletal muscle exhibits myopathy in association with mitochondrial dysfunction and oxidative damage detrimental effect of class-selective histone deacetylase inhibitors during tissue regeneration following hindlimb ischemia chronic sodium nitrite therapy augments ischemia-induced angiogenesis and arteriogenesis sildenafil promotes ischemia-induced angiogenesis through a pkg-dependent pathway fenofibrate increases high-density lipoprotein and sphingosine phosphate concentrations limiting abdominal aortic aneurysm progression in a mouse model osteoprotegerin deficiency limits angiotensin ii-induced aortic dilatation and rupture in the apolipoprotein e-knockout mouse wnt signaling pathway inhibitor sclerostin inhibits angiotensin ii-induced aortic aneurysm and atherosclerosis nuclear positioning in muscle development and disease the effect of exercise on haemodynamics in intermittent claudication: a systematic review of randomized controlled trials exercise for intermittent claudication impaired aerobic capacity in hypercholesterolemic mice: partial reversal by exercise training hypercholesterolemia impairs exercise capacity in mice exercise for intermittent claudication apolipoprotein e −/− mice have delayed skeletal muscle healing after hind limb ischemia-reperfusion peripheral nerve ischemia: apolipoprotein e deficiency results in impaired functional recovery and reduction of associated intraneural angiogenic response differential necrosis despite similar perfusion in mouse strains after ischemia time course of skeletal muscle repair and gene expression following acute hind limb ischemia in mice promoting blood vessel growth in ischemic diseases: challenges in translating preclinical potential into clinical success inflammation and wound healing: the role of the macrophage post-ischaemic neovascularization and inflammation vegf receptor signalling -in control of vascular function glycaemic control improves perfusion recovery and vegfr protein expression in diabetic mice following experimental pad high affinity vegf binding and developmental expression suggest flk- as a major regulator of vasculogenesis and angiogenesis vascular endothelial growth factor/vascular permeability factor augments nitric oxide release from quiescent rabbit and human vascular endothelium recovery of disturbed endothelium-dependent flow in the collateral-perfused rabbit ischemic hindlimb after administration of vascular endothelial growth factor nitric oxide synthase modulates angiogenesis in response to tissue ischemia essential role of endothelial nitric oxide synthase for mobilization of stem and progenitor cells hindlimb ischemia impairs endothelial recovery and increases neointimal proliferation in the carotid artery hemodynamic stresses and structural remodeling of anastomosing arteriolar networks: design principles of collateral arterioles wall remodeling during luminal expansion of mesenteric arterial collaterals in the rat shear-induced force transmission in a multicomponent, multicell model of the endothelium oscillatory shear stress and hydrostatic pressure modulate cell-matrix attachment proteins in cultured endothelial cells effects of biaxial oscillatory shear stress on endothelial cell proliferation and morphology ultrastructure and molecular histology of rabbit hind-limb collateral artery growth (arteriogenesis) trpv induces collateral vessel growth during regeneration of the arterial circulation calcium-dependent signalling is essential during collateral growth in the pig hind limb-ischemia model effects of endogenous nitric oxide and of deta nonoate in arteriogenesis trpv -dependent dilation of peripheral resistance arteries influences arterial pressure transient receptor potential vanilloid type -deficient mice exhibit impaired endothelium-dependent relaxation induced by acetylcholine in vitro and in vivo activation mediates flow-induced nitric oxide production in the rat thick ascending limb trpv -mediated endothelial ca + influx and vasodilation in response to shear stress esc guidelines on the diagnosis and treatment of peripheral arterial diseases, in collaboration with the european society for vascular surgery (esvs): document covering atherosclerotic disease of extracranial carotid and vertebral, mesenteric, renal, upper and lower extremity arteriesendorsed by: the european stroke organization (eso)the task force for the diagnosis and treatment of peripheral arterial diseases of the european society of cardiology (esc) and of the european society for vascular surgery (esvs) cilostazol for intermittent claudication a systematic review of the uptake and adherence rates to supervised exercise programs in patients with intermittent claudication exercise for intermittent claudication a systematic review of treatment of intermittent claudication in the lower extremities cardiovascular effects of exercise: role of endothelial shear stress exercise-induced expression of angiogenesis-related transcription and growth factors in human skeletal muscle the influence of physical training on the angiopoietin and vegf-a systems in human skeletal muscle exercise linked to transient increase in expression and activity of cation channels in newly formed hind-limb collaterals this research was funded by a faculty administered grant and a research infrastructure block grant from james cook university and funding from the queensland government. jg holds a practitioner fellowship from the national health and medical research council, australia ( ) and a senior clinical research fellowship from the queensland government. smo was supported by funding from the graduate research school and college of medicine, james cook university. we would like to acknowledge with thanks the help of prof. zoltan sarnyai (james cook university) who provided access to open field test assessment facility, dr. joseph moxon (james cook university) who assisted with the lme used to analyse data from the exercise study and dr. pacific huynh (james cook university) who provided the mouse images in figs. a and a. s.m.k. lead the design of the research, undertaking of experiments, and interpretation of data and writing of the manuscript. s.m.o., j.l., s.m. and r.j.j. performed parts of the experiments, contributed to interpretation of the data and gave critical comments on the manuscript. j.g. contributed rationales for the studies, led funding applications, co-wrote the manuscript and contributed to project supervision and data interpretation. the authors declare no competing interests. supplementary information is available for this paper at https://doi.org/ . /s - - - .correspondence and requests for materials should be addressed to j.g.reprints and permissions information is available at www.nature.com/reprints.publisher's note springer nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations. license, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the creative commons license, and indicate if changes were made. the images or other third party material in this article are included in the article's creative commons license, unless indicated otherwise in a credit line to the material. if material is not included in the article's creative commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. to view a copy of this license, visit http://creativecommons.org/licenses/by/ . /. key: cord- -d im p e authors: helbing, dirk title: challenges in economics date: - - journal: social self-organization doi: . / - - - - _ sha: doc_id: cord_uid: d im p e in the same way as the hilbert program was a response to the foundational crisis of mathematics [ ], this article tries to formulate a research program for the socio-economic sciences. the aim of this contribution is to stimulate research in order to close serious knowledge gaps in mainstream economics that the recent financial and economic crisis has revealed. by identifying weak points of conventional approaches in economics, we identify the scientific problems which need to be addressed. we expect that solving these questions will bring scientists in a position to give better decision support and policy advice. we also indicate, what kinds of insights can be contributed by scientists from other research fields such as physics, biology, computer and social science. in order to make a quick progress and gain a systemic understanding of the whole interconnected socio-economic-environmental system, using the data, information and computer systems available today and in the near future, we suggest a multi-disciplinary collaboration as most promising research approach. static where the world was dynamic, it assumed competitive markets where few existed, it assumed rationality when we knew full well that economic agents were not rational . . . economics had no way of dealing with changing tastes and technology . . . econometrics was equally plagued with intractable problems: economic observations are never randomly drawn and seldom independent, the number of excluded variables is always unmanageably large, the degrees of freedom unacceptably small, the stability of significance tests seldom unequivocably established, the errors in measurement too large to yield meaningful results . . . " [ ] . in the following, we will try to identify the scientific challenges that must be addressed to come up with better theories in the near future. this comprises practical challenges, i.e. the real-life problems that must be faced (see sect. . ), and fundamental challenges, i.e. the methodological advances that are required to solve these problems (see sect. . ) . after this, we will discuss, which contribution can be made by related scientific disciplines such as econophysics and the social sciences. the intention of this contribution is constructive. it tries to stimulate a fruitful scientific exchange, in order to find the best way out of the crisis. according to our perception, the economic challenges we are currently facing can only be mastered by large-scale, multi-disciplinary efforts and by innovative approaches [ ] . we fully recognize the large variety of non-mainstream approaches that has been developed by "heterodox economists". however, the research traditions in economics seem to be so powerful that these are not paid much attention to. besides, there is no agreement on which of the alternative modeling approaches would be the most promising ones, i.e. the heterogeneity of alternatives is one of the problems, which slows down their success. this situation clearly implies institutional challenges as well, but these go beyond the scope of this contribution and will therefore be addressed in the future. since decades, if not since hundreds of years, the world is facing a number of recurrent socio-economic problems, which are obviously hard to solve. before addressing related fundamental scientific challenges in economics, we will therefore point out practical challenges one needs to pay attention to. this basically requires to classify the multitude of problems into packages of interrelated problems. probably, such classification attempts are subjective to a certain extent. at least, the list presented below differs from the one elaborated by lomborg et al. [ ] , who identified the following top ten problems: air pollution, security/conflict, disease control, education, climate change, hunger/malnutrition, water sanitation, barriers to migration and trade, transnational terrorism and, finally, women and development. the following (non-ranked) list, in contrast, is more focused on socio-economic factors rather than resource and engineering issues, and it is more oriented at the roots of problems rather than their symptoms: . demographic change of the population structure (change of birth rate, migration, integration. . . ) . financial and economic (in)stability (government debts, taxation, and inflation/ deflation; sustainability of social benefit systems; consumption and investment behavior. . . ) . social, economic and political participation and inclusion (of people of different gender, age, health, education, income, religion, culture, language, preferences; reduction of unemployment. . . ) . balance of power in a multi-polar world (between different countries and economic centers; also between individual and collective rights, political and company power; avoidance of monopolies; formation of coalitions; protection of pluralism, individual freedoms, minorities. . . ) . collective social behavior and opinion dynamics (abrupt changes in consumer behavior; social contagion, extremism, hooliganism, changing values; breakdown of cooperation, trust, compliance, solidarity. . . ) . security and peace (organized crime, terrorism, social unrest, independence movements, conflict, war. . . ) . institutional design (intellectual property rights; over-regulation; corruption; balance between global and local, central and decentral control. . . ) . sustainable use of resources and environment (consumption habits, travel behavior, sustainable and efficient use of energy and other resources, participation in recycling efforts, environmental protection. . . ) . information management (cyber risks, misuse of sensitive data, espionage, violation of privacy; data deluge, spam; education and inheritance of culture. . . ) . public health (food safety; spreading of epidemics [flu, sars, h n , hiv], obesity, smoking, or unhealthy diets. . . ) some of these challenges are interdependent. in the following, we will try to identify the fundamental theoretical challenges that need to be addressed in order to understand the above practical problems and to draw conclusions regarding possible solutions. the most difficult part of scientific research is often not to find the right answer. the problem is to ask the right questions. in this context it can be a problem that people are trained to think in certain ways. it is not easy to leave these ways and see the problem from a new angle, thereby revealing a previously unnoticed solution. three factors contribute to this: . we may overlook the relevant facts because we have not learned to see them, i.e. we do not pay attention to them. the issue is known from internalized norms, which prevent people from considering possible alternatives. . we know the stylized facts, but may not have the right tools at hand to interpret them. it is often difficult to make sense of patterns detected in data. turning data into knowledge is quite challenging. . we know the stylized facts and can interpret them, but may not take them seriously enough, as we underestimate their implications. this may result from misjudgements or from herding effects, i.e. from a tendency to follow traditions and majority opinions. in fact, most of the issues discussed below have been pointed out before, but it seems that this did not have an effect on mainstream economics so far or on what decision-makers know about economics. this is probably because mainstream theory has become a norm [ ] , and alternative approaches are sanctioned as norm-deviant behavior [ , ] . as we will try to explain, the following fundamental issues are not just a matter of approximations (which often lead to the right understanding, but wrong numbers). rather they concern fundamental errors in the sense that certain conclusions following from them are seriously misleading. as the recent financial crisis has demonstrated, such errors can be very costly. however, it is not trivial to see what dramatic consequences factors such as dynamics, spatial interactions, randomness, non-linearity, network effects, differentiation and heterogeneity, irreversibility or irrationality can have. despite of criticisms by several nobel prize winners such as reinhard selten ( ), joseph stiglitz and george akerlof ( ) , or daniel kahneman ( ) , the paradigm of the homo economicus, i.e. of the "perfect egoist", is still the dominating approach in economics. it assumes that people would have quasi-infinite memory and processing capacities and would determine the best one among all possible alternative behaviors by strategic thinking (systematic utility optimization), and would implement it into practice without mistakes. the nobel prize winner of , milton friedman, supported the hypothesis of homo economicus by the following argument: "irrational agents will lose money and will be driven out of the market by rational agents" [ ] . more recently, robert e. lucas jr., the nobel prize winner of , used the rationality hypothesis to narrow down the class of empirically relevant equilibria [ ] . the rational agent hypothesis is very charming, as its implications are clear and it is possible to derive beautiful and powerful economic theorems and theories from it. the best way to illustrate homo economicus is maybe a company that is run by using optimization methods from operation research, applying supercomputers. another example are professional chess players, who are trying to anticipate the possible future moves of their opponents. obviously, in both examples, the future course of actions can not be fully predicted, even if there are no random effects and mistakes. it is, therefore, no wonder that people have repeatedly expressed doubts regarding the realism of the rational agent approach [ , ] . bertrand russell, for example, claimed: "most people would rather die than think". while this seems to be a rather extreme opinion, the following scientific arguments must be taken seriously: . human cognitive capacities are bounded [ , ] . already phone calls or conversations can reduce people's attention to events in the environment a lot. also, the abilities to memorize facts and to perform complicated logical analyses are clearly limited. . in case of np-hard optimization problems, even supercomputers are facing limits, i.e. optimization jobs cannot be performed in real-time anymore. therefore, approximations or simplifications such as the application of heuristics may be necessary. in fact, psychologists have identified a number of heuristics, which people use when making decisions [ ] . . people perform strategic thinking mainly in important new situations. in normal, everyday situation, however, they seem to pursue a satisficing rather than optimizing strategy [ ] . meeting a certain aspiration level rather than finding the optimal strategy can save time and energy spent on problem solving. in many situation, people even seem to perform routine choices [ ] , for example, when evading other pedestrians in counterflows. . there is a long list of cognitive biases which question rational behavior [ ] . for example, individuals are favorable of taking small risks (which are preceived as "chances", as the participation in lotteries shows), but they avoid large risks [ ] . furthermore, non-exponential temporal discounting may lead to paradoxical behaviors [ ] and requires one to rethink, how future expectations must be modeled. . most individuals have a tendency towards other-regarding behavior and fairness [ , ] . for example, the dictator game [ ] and other experiments [ ] show that people tend to share, even if there is no reason for this. leaving a tip for the waiter in a restaurant that people visit only once is a typical example (particularly in countries where tipping is not common) [ ] . such behavior has often been interpreted as sign of social norms. while social norms can certainly change the payoff structure, it has been found that the overall payoffs resulting from them do not need to create a user or system optimum [ ] [ ] [ ] . this suggests that behavioral choices may be irrational in the sense of non-optimal. a typical example is the existence of unfavorable norms, which are supported by people although nobody likes them [ ] . . certain optimization problems can have an infinite number of local optima or nash equilibria, which makes it impossible to decide what is the best strategy [ ] . . convergence towards the optimal solution may require such a huge amount of time that the folk theorem becomes useless. this can make it practically impossible to play the best response strategy [ ] . . the optimal strategy may be deterministically chaotic, i.e. sensitive to arbitrarily small details of the initial condition, which makes the dynamic solution unpredictable on the long run ("butterfly effect") [ , ] . this fundamental limit of predictability also implies a limit of control-two circumstances that are even more true for non-deterministic systems with a certain degree of randomness. in conclusion, although the rational agent paradigm (the paradigm of homo economicus) is theoretically powerful and appealing, there are a number of empirical and theoretical facts, which suggest deficiencies. in fact, most methods used in financial trading (such as technical analysis) are not well compatible with the rational agent approach. even if an optimal solution exists, it may be undecidable for practical reasons or for theoretical ones [ , ] . this is also relevant for the following challenges, as boundedly rational agents may react inefficently and with delays, which questions the efficient market hypothesis, the equilibrium paradigm, and other fundamental concepts, calling for the consideration of spatial, network, and time-dependencies, heterogeneity and correlations etc. it will be shown that these points can have dramatic implications regarding the predictions of economic models. the efficient market hypothesis (emh) was first developed by eugene fama [ ] in his ph.d. thesis and rapidly spread among leading economists, who used it as an argument to promote laissez-faire policies. the emh states that current prices reflect all publicly available information and (in its stronger formulation) that prices instantly change to reflect new public information. the idea of self-regulating markets goes back to adam smith [ ] , who believed that "the free market, while appearing chaotic and unrestrained, is actually guided to produce the right amount and variety of goods by a so-called "invisible hand"." furthermore, "by pursuing his own interest, [the individual] frequently promotes that of the society more effectually than when he intends to promote it" [ ] . for this reason, adam smith is often considered to be the father of free market economics. curiously enough, however, he also wrote a book on "the theory of moral sentiments" [ ] . "his goal in writing the work was to explain the source of mankind's ability to form moral judgements, in spite of man's natural inclinations towards self-interest. smith proposes a theory of sympathy, in which the act of observing others makes people aware of themselves and the morality of their own behavior . . . [and] seek the approval of the "impartial spectator" as a result of a natural desire to have outside observers sympathize with them" [ ] . such a reputation-based concept would be considered today as indirect reciprocity [ ] . of course, there are criticisms of the efficient market hypothesis [ ] , and the nobel prize winner of , joseph stiglitz, even believes that "there is not invisible hand" [ ] . the following list gives a number of empirical and theoretical arguments questioning the efficient market hypothesis: . examples of market failures are well-known and can result, for example, in cases of monopolies or oligopolies, if there is not enough liquidity or if information symmetry is not given. . while the concept of the "invisible hand" assumes something like an optimal self-organization [ ] , it is well-known that this requires certain conditions, such as symmetrical interactions. in general, however, self-organization does not necessarily imply system-optimal solutions. stop-and-go traffic [ ] or crowd disasters [ ] are two obvious examples for systems, in which individuals competitively try to reach individually optimal outcomes, but where the optimal solution is dynamically unstable. . the limited processing capacity of boundedly rational individuals implies potential delays in their responses to sensorial inputs, which can cause such instabilities [ ] . for example, a delayed adaptation in production systems may contribute to the occurrence of business cycles [ ] . the same applies to the labor market of specially skilled people, which cannot adjust on short time scales. even without delayed reactions, however, the competitive optimization of individuals can lead to suboptimal individual results, as the "tragedy of the commons" in public goods dilemmas demonstrates [ , ] . . bubbles and crashes, or more generally, extreme events in financial markets should not occur, if the efficient market hypothesis was correct (see next subsection). . collective social behavior such as "herding effects" as well as deviations of human behavior from what is expected from rational agents can lead to such bubbles and crashes [ ] , or can further increase their size through feedback effects [ ] . cyclical feedbacks leading to oscillations are also known from the beer game [ ] or from business cycles [ ] . the efficient market paradigm implies the equilibrium paradigm. this becomes clear, if we split it up into its underlying hypotheses: . the market can be in equilibrium, i.e. there exists an equilibrium. . there is one and only one equilibrium. . the equilibrium is stable, i.e. any deviations from the equilibrium due to "fluctuations" or "perturbations" tend to disappear eventually. . the relaxation to the equilibrium occurs at an infinite rate. note that, in order to act like an "invisible hand", the stable equilibrium (nash equilibrium) furthermore needs to be a system optimum, i.e. to maximize the average utility. this is true for coordination games, when interactions are well-mixed and exploration behavior as well as transaction costs can be neglected [ ] . however, it is not fulfilled by so-called social dilemmas [ ] . let us discuss the evidence for the validity of the above hypotheses one by one: . a market is a system of extremely many dynamically coupled variables. theoretically, it is not obvious that such a system would have a stationary solution. for example, the system could behave periodic, quasi-periodic, chaotic, or turbulent [ - , - , ] . in all these cases, there would be no convergence to a stationary solution. . if a stationary solution exists, it is not clear that there are no further stationary solutions. if many variables are non-linearly coupled, the phenomenon of multistability can easily occur [ ] . that is, the solution to which the system converges may not only depend on the model parameters, but also on the initial condition, history, or perturbation size. such facts are known as path-dependencies or hysteresis effects and are usually visualized by so-called phase diagrams [ ] . . in systems of non-linearly interacting variables, the existence of a stationary solution does not necessarily imply that it is stable, i.e. that the system will converge to this solution. for example, the stationary solution could be a focal point with orbiting solutions (as for the classical lotka-volterra equations [ ] ), or it could be unstable and give rise to a limit cycle [ ] or a chaotic solution [ ] , for example (see also item ). in fact, experimental results suggest that volatility clusters in financial markets may be a result of over-reactions to deviations from the fundamental value [ ] . . an infinite relaxation rate is rather unusual, as most decisions and related implemenations take time [ , ] . the points listed in the beginning of this subsection are also questioned by empirical evidence. in this connection, one may mention the existence of business cycles [ ] or unstable orders and deliveries observed in the experimental beer game [ ] . moreover, bubbles and crashes have been found in financial market games [ ] . today, there seems to be more evidence against than for the equilibrium paradigm. in the past, however, most economists assumed that bubbles and crashes would not exist (and many of them still do). the following quotes are quite typical for this kind of thinking (from [ ] ): in , the federal reserve chairman of the u.s., alan greenspan, stated that the rise in house values was "not enough in our judgment to raise major concerns". in july when asked about the possibility of a housing bubble and the potential for this to lead to a recession in the future, the present u.s. federal reserve chairman ben bernanke (then chairman of the council of economic advisors) said: "it's a pretty unlikely possibility. we've never had a decline in housing prices on a nationwide basis. so, what i think is more likely is that house prices will slow, maybe stabilize, might slow consumption spending a bit. i don't think it's going to drive the economy too far from it's full path though." as late as may bernanke stated that the federal reserve "do not expect significant spillovers from the subprime market to the rest of the economy". according to the classical interpretation, sudden changes in stock prices result from new information, e.g. from innovations ("technological shocks"). the dynamics in such systems has, for example, been described by the method of comparative statics (i.e. a series of snapshots). here, the system is assumed to be in equilibrium in each moment, but the equilibrium changes adiabatically (i.e. without delay), as the system parameters change (e.g. through new facts). such a treatment of system dynamics, however, has certain deficiencies: . the approach cannot explain changes in or of the system, such as phase transitions ("systemic shifts"), when the system is at a critical point ("tipping point"). . it does not allow one to understand innovations and other changes as results of an endogeneous system dynamics. . it cannot describe effects of delays or instabilities, such as overshooting, self-organization, emergence, systemic breakdowns or extreme events (see sect. . . ). . it does not allow one to study effects of different time scales. for example, when there are fast autocatalytic (self-reinfocing) effects and slow inhibitory effects, this may lead to pattern formation phenomena in space and time [ , ] . the formation of settlements, where people agglomerate in space, may serve as an example [ , ] . . it ignores long-term correlations such as memory effects. . it neglects frictional effects, which are often proportional to change ("speed") and occur in most complex systems. without friction, however, it is difficult to understand entropy and other path-dependent effects, in particular irreversibility (i.e. the fact that the system may not be able to get back to the previous state) [ ] . for example, the unemployment rate has the property that it does not go back to the previous level in most countries after a business cycle [ ] . comparative statics is, of course, not the only method used in economics to describe the dynamics of the system under consideration. as in physics and other fields, one may use a linear approximation around a stationary solution to study the response of the system to fluctuations or perturbations [ ] . such a linear stability analysis allows one to study, whether the system will return to the stationary solution (which is the case for a stable [nash] equilibrium) or not (which implies that the system will eventually be driven into a new state or regime). in fact, the great majority of statistical analyses use linear models to fit empirical data (also when they do not involve time-dependencies). it is know, however, that linear models have special features, which are not representative for the rich variety of possible functional dependencies, dynamics, and outcomes. therefore, the neglection of non-linearity has serious consequences: . as it was mentioned before, phenomena like multiple equilibria, chaos or turbulence cannot be understood by linear models. the same is true for selforganization phenomena or emergence. additionally, in non-linearly coupled systems, usually "more is different", i.e. the system may change its behavior fundamentally as it grows beyond a certain size. furthermore, the system is often hard to predict and difficult to control (see sect. . . ). . linear modeling tends to overlook that a strong coupling of variables, which would show a normally distributed behavior in separation, often leads to fat tail distributions (such as "power laws") [ , ] . this implies that extreme events are much more frequent than expected according to a gaussian distribution. for example, when additive noise is replaced by multiplicative noise, a number of surprising phenomena may result, including noise-induced transitions [ ] or directed random walks ("ratchet effects") [ ] . . phenomena such as catastrophes [ ] or phase transition ("system shifts") [ ] cannot be well understood within a linear modeling framework. the same applies to the phenomenon of "self-organized criticality" [ ] (where the system drives itself to a critical state, typically with power-law characteristics) or cascading effects, which can result from network interactions (overcritically challenged network nodes or links) [ , ] . it should be added that the relevance of network effects resulting from the on-going globalization is often underestimated. for example, "the stock market crash of , began with a small drop in prices which triggered an avalanche of sell orders in computerized trading programs, causing a further price decline that triggered more automatic sales." [ ] therefore, while linear models have the advantage of being analytically solvable, they are often unrealistic. studying non-linear behavior, in contrast, often requires numerical computational approaches. it is likely that most of today's unsolved economic puzzles cannot be well understood through linear models, no matter how complicated they may be (in terms of the number of variables and parameters) [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] . the following list mentions some areas, where the importance of non-linear interdependencies is most likely underestimated: • collective opinions, such as trends, fashions, or herding effects. • the success of new (and old) technologies, products, etc. • cultural or opinion shifts, e.g. regarding nuclear power, genetically manipulated food, etc. • the "fitness" or competitiveness of a product, value, quality perceptions, etc. • the respect for copyrights. • social capital (trust, cooperation, compliance, solidarity, . . . ). • booms and recessions, bubbles and crashes. • bank panics. • community, cluster, or group formation. • relationships between different countries, including war (or trade war) and peace. another common simplification in economic modeling is the representative agent approach, which is known in physics as mean field approximation. within this framework, time-dependencies and non-linear dependencies are often considered, but it is assumed that the interaction with other agents (e.g. of one company with all the other companies) can be treated as if this agent would interact with an average agent, the "representative agent". let us illustrate this with the example of the public goods dilemma. here, everyone can decide whether to make an individual contribution to the public good or not. the sum of all contributions is multiplied by a synergy factor, reflecting the benefit of cooperation, and the resulting value is equally shared among all people. the prediction of the representative agent approach is that, due to the selfishness of agents, a "tragedy of the commons" would result [ ] . according to this, everybody should free-ride, i.e. nobody should make a contribution to the public good and nobody would gain anything. however, if everybody would contribute, everybody could multiply his or her contribution by the synergy factor. this example is particularly relevant as society is facing a lot of public goods problems and would not work without cooperation. everything from the creation of public infrastructures (streets, theaters, universities, libraries, schools, the world wide web, wikipedia etc.) over the use of environmental resources (water, forests, air, etc.) or of social benefit systems (such as a public health insurance), maybe even the creation and maintainance of a commonly shared language and culture are public goods problems (although the last examples are often viewed as coordination problems). even the process of creating public goods is a public good [ ] . while it is a well-known problem that people tend to make unfair contributions to public goods or try to get a bigger share of them, individuals cooperate much more than one would expect according to the representative agent approach. if they would not, society could simply not exist. in economics, one tries to solve the problem by introducing taxes (i.e. another incentive structure) or a "shadow of the future" (i.e. a strategic optimization over infinite time horizons in accordance with the rational agent approach) [ , ] . both comes down to changing the payoff structure in a way that transforms the public good problem into another one that does not constitute a social dilemma [ ] . however, there are other solutions of the problem. when the realm of the mean-field approximation underlying the representative agent approach is left and spatial or network interactions or the heterogeneity among agents are considered, a miracle occurs: cooperation can survive or even thrive through correlations and co-evolutionary effects [ ] [ ] [ ] . a similar result is found for the public goods game with costly punishment. here, the representative agent model predicts that individuals avoid to invest into punishment, so that punishment efforts eventually disappear (and, as a consequence, cooperation as well). however, this "second-order free-rider problem" is naturally resolved and cooperation can spread, if one discards the mean-field approximation and considers the fact that interactions take place in space or social networks [ ] . societies can overcome the tragedy of the commons even without transforming the incentive structure through taxes. for example, social norms as well as group dynamical and reputation effects can do so [ ] . the representative agent approach implies just the opposite conclusion and cannot well explain the mechanisms on which society is built. it is worth pointing out that the relevance of public goods dilemmas is probably underestimated in economics. partially related to adam smith's belief in an "invisible hand", one often assumes underlying coordination games and that they would automatically create a harmony between an individually and system optimal state in the course of time [ ] . however, running a stable financial system and economy is most likely a public goods problem. considering unemployment, recessions always go along with a breakdown of solidarity and cooperation. efficient production clearly requires mutual cooperation (as the counter-example of countries with many strikes illustrates). the failure of the interbank market when banks stop lending to each other, is a good example for the breakdown of both, trust and cooperation. we must be aware that there are many other systems that would not work anymore, if people would lose their trust: electronic banking, e-mail and internet use, facebook, ebusiness and egovernance, for example. money itself would not work without trust, as bank panics and hyperinflation scenarios show. similarly, cheating customers by selling low-quality products or selling products at overrated prices, or by manipulating their choices by advertisements rather than informing them objectively and when they want, may create profits on the short run, but it affects the trust of customers (and their willingness to invest). the failure of the immunization campaign during the swine flu pandemics may serve as an example. furthermore, people would probably spend more money, if the products of competing companies were better compatible with each other. therefore, on the long run, more cooperation among companies and with the customers would pay off and create additional value. besides providing a misleading picture of how cooperation comes about, there are a number of other deficiencies of the representative agent approach, which are listed below: . correlations between variables are neglected, which is acceptable only for "well-mixing" systems. according to what is known from critical phenomena in physics, this approximation is valid only, when the interactions take place in high-dimensional spaces or if the system elements are well connected. (however, as the example of the public goods dilemma showed, this case does not necessarily have beneficial consequences. well-mixed interactions could rather cause a breakdown of social or economic institutions, and it is conceivable that this played a role in the recent financial crisis.) . percolation phenomena, describing how far an idea, innovation, technology, or (computer) virus spreads through a social or business network, are not well reproduced, as they depend on details of the network structure, not just on the average node degree [ ] . . the heterogeneity of agents is ignored. for this reason, factors underlying economic exchange, perturbations, or systemic robustness [ ] cannot be well described. moreover, as socio-economic differentiation and specialization imply heterogeneity, they cannot be understood as emergent phenomena within a representative agent approach. finally, it is not possible to grasp innovation without the consideration of variability. in fact, according to evolutionary theory, the innovation rate would be zero, if the variability was zero [ ] . furthermore, in order to explain innovation in modern societies, schumpeter introduced the concept of the "political entrepreneur" [ ] , an extra-ordinarily gifted person capable of creating disruptive change and innovation. such an extraordinary individual can, by definition, not be modeled by a "representative agent". one of the most important drawbacks of the representative agent approach is that it cannot explain the fundamental fact of economic exchange, since it requires one to assume a heterogeneity in resources or production costs, or to consider a variation in the value of goods among individuals. ken arrow, nobel prize winner in , formulated this point as follows [ ] : "one of the things that microeconomics teaches you is that individuals are not alike. there is heterogeneity, and probably the most important heterogeneity here is heterogeneity of expectations. if we didn't have heterogeneity, there would be no trade." we close this section by mentioning that economic approaches, which go beyond the representative agent approach, can be found in refs. [ , ] . another deficiency of economic theory that needs to be mentioned is the lack of a link between micro-and macroeconomics. neoclassical economics implicitly assumes that individuals make their decisions in isolation, using only the information received from static market signals. within this oversimplified framework, macro-aggregates are just projections of some representative agent behavior, instead of the outcome of complex interactions with asymmetric information among a myriad of heterogeneous agents. in principle, it should be understandable how the macroeconomic dynamics results from the microscopic decisions and interactions on the level of producers and consumers [ , ] (as it was possible in the past to derive micro-macro links for other systems with a complex dynamical behavior such as interactive vehicle traffic [ ] ). it should also be comprehensible how the macroscopic level (the aggregate econonomic situation) feeds back on the microscopic level (the behavior of consumers and producers), and to understand the economy as a complex, adaptive, self-organizing system [ , ] . concepts from evolutionary theory [ ] and ecology [ ] appear to be particularly promising [ ] . this, however, requires a recognition of the importance of heterogeneity for the system (see the the previous subsection). the lack of ecological thinking implies not only that the sensitive network interdependencies between the various agents in an economic system (as well as minority solutions) are not properly valued. it also causes deficiencies in the development and implementation of a sustainable economic approach based on recycling and renewable resources. today, forestry science is probably the best developed scientific discipline concerning sustainability concepts [ ] . economic growth to maintain social welfare is a serious misconception. from other scientific disciplines, it is well known that stable pattern formation is also possible for a constant (and potentially sustainable) inflow of energy [ , ] . one of the great achievements of economics is that it has developed a multitude of methods to use scarce resources efficiently. a conventional approach to this is optimization. in principle, there is nothing wrong about this approach. nevertheless, there are a number of problems with the way it is usually applied: . one can only optimize for one goal at a time, while usually, one needs to meet several objectives. this is mostly addressed by weighting the different goals (objectives), by executing a hierarchy of optimization steps (through ranking and prioritization), or by applying a satisficing strategy (requiring a minimum performance for each goal) [ , ] . however, when different optimization goals are in conflict with each other (such as maximizing the throughput and minimizing the queue length in a production system), a sophisticated timedependent strategy may be needed [ ] . high profit? best customer satisfaction? large throughput? competitive advantage? resilience? [ ] in fact, the choice of the optimization function is arbitrary to a certain extent and, therefore, the result of optimization may vary largely. goal selection requires strategic decisions, which may involve normative or moral factors (as in politics). in fact, one can often observe that, in the course of time, different goal functions are chosen. moreover, note that the maximization of certain objectives such as resilience or "fitness" depends not only on factors that are under the control of a company. resilience and "fitness" are functions of the whole system, in particularly, they also depend on the competitors and the strategies chosen by them. . the best solution may be the combination of two bad solutions and may, therefore, be overlooked. in other words, there are "evolutionary dead ends", so that gradual optimization may not work. (this problem can be partially overcome by the application of evolutionary mechanisms [ ] ). . in certain systems (such as many transport, logistic, or production systems), optimization tends to drive the system towards instability, since the point of maximum efficiency is often in the neighborhood or even identical with the point of breakdown of performance. such breakdowns in capacity or performance can result from inefficiencies due to dynamic interaction effects. for example, when traffic flow reaches its maximum capacity, sooner or later it breaks down. as a consequence, the road capacity tends to drop during the time period where it is most urgently needed, namely during the rush hour [ , ] . . optimization often eliminates reduncancies in the system and, thereby, increases the vulnerability to perturbations, i.e. it decreases robustness and resilience. . optimization tends to eliminate heterogeneity in the system [ ] , while heterogeneity frequently supports adaptability and resilience. . optimization is often performed with centralized concepts (e.g. by using supercomputers that process information collected all over the system). such centralized systems are vulnerable to disturbances or failures of the central control unit. they are also sensitive to information overload, wrong selection of control parameters, and delays in adaptive feedback control. in contrast, decentralized control (with a certain degree of autonomy of local control units) may perform better, when the system is complex and composed of many heterogeneous elements, when the optimization problem is np hard, the degree of fluctuations is large, and predictability is restricted to short time periods [ , ] . under such conditions, decentralized control strategies can perform well by adaptation to the actual local conditions, while being robust to perturbations. urban traffic light control is a good example for this [ , ] . . further on, today's concept of quality control appears to be awkward. it leads to a never-ending contest, requiring people and organizations to fulfil permanently increasing standards. this leads to over-emphasizing measured performance criteria, while non-measured success factors are neglected. the engagement into non-rewarded activities is discouraged, and innovation may be suppressed (e.g. when evaluating scientists by means of their h-index, which requires them to focus on a big research field that generates many citations in a short time). while so-called "beauty contests" are considered to produce the best results, they will eventually absorb more and more resources for this contest, while less and less time remains for the work that is actually to be performed, when the contest is won. besides, a large number of competitors have to waste considerable resources for these contests which, of course, have to be paid by someone. in this way, private and public sectors (from physicians over hospitals, administrations, up to schools and universities) are aching under the evaluationrelated administrative load, while little time remains to perform the work that the corresponding experts have been trained for. it seems naïve to believe that this would not waste resources. rather than making use of individual strengths, which are highly heterogeneous, today's way of evaluating performance enforces a large degree of conformity. there are also some problems with parameter fitting, a method based on optimization as well. in this case, the goal function is typically an error function or a likelihood function. not only are calibration methods often "blindly" applied in practice (by people who are not experts in statistics), which can lead to overfitting (the fitting of meaningless "noise"), to the neglection of collinearities (implying largely variable parameter values), or to inaccurate and problematic parameter determinations (when the data set is insufficient in size, for example, when large portfolios are to be optimized [ ] ). as estimates for past data are not necessarily indicative for the future, making predictions with interpolation approaches can be quite problematic (see also sect. . . for the challenge of time dependence). moreover, classical calibration methods do not reveal inappropriate model specifications (e.g. linear ones, when non-linear models would be needed, or unsuitable choices of model variables). finally, they do not identify unknown unknowns (i.e. relevant explanatory variables, which have been overlooked in the modeling process). managing economic systems is a particular challenge, not only for the reasons discussed in the previous section. as large economic systems belong to the class of complex systems, they are hard or even impossible to manage with classical control approaches [ , ] . complex systems are characterized by a large number of system elements (e.g. individuals, companies, countries, . . . ), which have non-linear or network interactions causing mutual dependencies and responses. such systems tend to behave dynamic rather than static and probabilistic rather than deterministic. they usually show a rich, hardly predictable, and sometimes paradoxical system behavior. therefore, they challenge our way of thinking [ ] , and their controllability is often overestimated (which is sometimes paraphrased as "illusion of control") [ , , ] . in particular, causes and effects are typically not proportional to each other, which makes it difficult to predict the impact of a control attempt. a complex system may be unresponsive to a control attempt, or the latter may lead to unexpected, large changes in the system behavior (so-called "phase transitions", "regime shifts", or "catastrophes") [ ] . the unresponsiveness is known as principle of le chatelier or goodhart's law [ ] , according to which a complex system tends to counteract external control attempts. however, regime shifts can occur, when the system gets close to so-called "critical points" (also known as "tipping points"). examples are sudden changes in public opinion (e.g. from pro to anti-war mood, from a smoking tolerance to a public smoking ban, or from buying energy-hungry sport utilities vehicles (suvs) to buying environmentally-friendly cars). particularly in case of network interactions, big changes may have small, no, or unexpected effects. feedback loops, unwanted side effects, and circuli vitiosi are quite typical. delays may cause unstable system behavior (such as bull-whip effects) [ ] , and over-critical perturbations can create cascading failures [ ] . systemic breakdowns (such as large-scale blackouts, bankruptcy cascades, etc.) are often a result of such domino or avalanche effects [ ] , and their probability of occurrence as well as their resulting damage are usually underestimated. further examples are epidemic spreading phenomena or disasters with an impact on the socio-economic system. a more detailed discussion is given in refs. [ , ] . other factors contributing to the difficulty to manage economic systems are the large heterogeneity of system elements and the considerable level of randomness as well as the possibility of a chaotic or turbulent dynamics (see sect. . . ) . furthermore, the agents in economic systems are responsive to information, which can create self-fulfilling or self-destroying prophecy effects. inflation may be viewed as example of such an effect. interestingly, in some cases one even does not know in advance, which of these effects will occur. it is also not obvious that the control mechanisms are well designed from a cybernetic perspective, i.e. that we have sufficient information about the system and suitable control variables to make control feasible. for example, central banks do not have terribly many options to influence the economic system. among them are performing open-market operations (to control money supply), adjustments in fractional-reserve banking (keeping only a limited deposit, while lending a large part of the assets to others), or adaptations in the discount rate (the interest rate charged to banks for borrowing short-term funds directly from a central bank). nevertheless, the central banks are asked to meet multiple goals such as: • to guarantee well-functioning and robust financial markets. • to support economic growth. • to balance between inflation and unemployment. • to keep exchange rates within reasonable limits. furthermore, the one-dimensional variable of "money" is also used to influence individual behavior via taxes (by changing behavioral incentives). it is questionable, whether money can optimally meet all these goals at the same time (see sect. . . ) . we believe that a computer, good food, friendship, social status, love, fairness, and knowledge can only to a certain extent be replaced by and traded against each other. probably for this reason, social exchange comprises more than just material exchange [ ] [ ] [ ] . it is conceivable that financial markets as well are trying to meet too many goals at the same time. this includes: • to match supply and demand. • to discover a fair price. • to raise the foreign direct investment (fdi). • to couple local economies with the international system. • to facilitate large-scale investments. • to boost development. • to share risk. • to support a robust economy, and • to create opportunities (to gamble, to become rich, etc.). therefore, it would be worth stuyding the system from a cybernetic control perspective. maybe, it would work better to separate some of these functions from each other rather than mixing them. another aspect that tends to be overlooked in mainstream economics is the relevance of psychological and social factors such as emotions, creativity, social norms, herding effects, etc. it would probably be wrong to interpret these effects just as a result of perception biases (see sect. . . ) . most likely, these human factors serve certain functions such as supporting the creation of public goods [ ] or collective intelligence [ , ] . as bruno frey has pointed out, economics should be seen from a social science perspective [ ] . in particular, research on happiness has revealed that there are more incentives than just financial ones that motivate people to work hard [ ] . interestingly, there are quite a number of factors which promote volunteering [ ] . it would also be misleading to judge emotions from the perspective of irrational behavior. they are a quite universal and a relatively energy-consuming way of signalling. therefore, they are probably more reliable than non-emotional signals. moreover, they create empathy and, consequently, stimulate mutual support and a readiness for compromises. it is quite likely that this creates a higher degree of cooperativeness in social dilemma situations and, thereby, a higher payoff on average as compared to emotionless decisions, which often have drawbacks later on. finally, there is no good theory that would allow one to assess the relevance of information in economic systems. most economic models do not consider information as an explanatory variable, although information is actually a stronger driving force of urban growth and social dynamics than energy [ ] . while we have an information theory to determine the number of bits required to encode a message, we are lacking a theory, which would allow us to assess what kind of information is relevant or important, or what kind of information will change the social or economic world, or history. this may actually be largely dependent on the perception of pieces of information, and on normative or moral issues filtering or weighting information. moreover, we lack theories describing what will happen in cases of coincidence or contradiction of several pieces of information. when pieces of information interact, this can change their interpretation and, thereby, the decisions and behaviors resulting from them. that is one of the reasons why socio-economic systems are so hard to predict: "unknown unknowns", structural instabilities, and innovations cause emergent results and create a dynamics of surprise [ ] . the problems discussed in the previous two sections pose interesting practical and fundamental challenges for economists, but also other disciplines interested in understanding economic systems. econophysics, for example, pursues a physical approach to economic systems, applying methods from statistical physics [ ] , network theory [ , ] , and the theory of complex systems [ , ] . a contribution of physics appears quite natural, in fact, not only because of its tradition in detecting and modeling regularities in large data sets [ ] . physics also has a lot of experience how to theoretically deal with problems such as time-dependence, fluctuations, friction, entropy, non-linearity, strong interactions, correlations, heterogeneity, and many-particle simulations (which can be easily extended towards multi-agent simulations). in fact, physics has influenced economic modeling already in the past. macroeconomic models, for example, were inspired by thermodynamics. more recent examples of relevant contributions by physicists concern models of self-organizing conventions [ ] , of geographic agglomeration [ ] , of innovation spreading [ ] , or of financial markets [ ] , to mention just a few examples. one can probably say that physicists have been among the pioneers calling for new approaches in economics [ , , [ ] [ ] [ ] [ ] [ ] . a particularly visionary book beside wolfgang weidlich's work was the "introduction to quantitative aspects of social phenomena" by elliott w. montroll and wade w. badger, which addressed by mathematical and empirical analysis subjects as diverse as population dynamics, the arms race, speculation patterns in stock markets, congestion in vehicular traffic as well as the problems of atmospheric pollution, city growth and developing countries already in [ ] . unfortunately, it is impossible in our paper to reflect the numerous contributions of the field of econophysics in any adequate way. the richness of scientific contributions is probably reflected best by the econophysics forum run by yi-cheng zhang [ ] . many econophysics solutions are interesting, but so far they are not broad and mighty enough to replace the rational agent paradigm with its large body of implications and applications. nevertheless, considering the relatively small number of econophysicists, there have been many promising results. the probably largest fraction of publications in econophysics in the last years had a data-driven or computer modeling approach to financial markets [ ] . but econophyics has more to offer than the analysis of financial data (such as fluctuations in stock and foreign currency exchange markets), the creation of interaction models for stock markets, or the development of risk management strategies. other scientists have focused on statistical laws underlying income and wealth distributions, nonlinear market dynamics, macroeconomic production functions and conditions for economic growth or agglomeration, sustainable economic systems, business cycles, microeconomic interaction models, network models, the growth of companies, supply and production systems, logistic and transport networks, or innovation dynamics and diffusion. an overview of subjects is given, for example, by ref. [ ] and the contributions to annual spring workshop of the physics of socio-economic systems division of the dpg [ ] . to the dissatisfaction of many econophysicists, the transfer of knowledge often did not work very well or, if so, has not been well recognized [ ] . besides scepticism on the side of many economists with regard to novel approaches introduced by "outsiders", the limited resonance and level of interdisciplinary exchange in the past was also caused in part by econophysicists. in many cases, questions have been answered, which no economist asked, rather than addressing puzzles economists are interested in. apart from this, the econophysics work was not always presented in a way that linked it to the traditions of economics and pointed out deficiencies of existing models, highlighting the relevance of the new approach well. typical responses are: why has this model been proposed and not another one? why has this simplification been used (e.g. an ising model of interacting spins rather than a rational agent model)? why are existing models not good enough to describe the same facts? what is the relevance of the work compared to previous publications? what practical implications does the finding have? what kind of paradigm shift does the approach imply? can existing models be modified or extended in a way that solves the problem without requiring a paradigm shift? correspondingly, there have been criticisms not only by mainstream economists, but also by colleagues, who are open to new approaches [ ] . therefore, we would like to suggest to study the various economic subjects from the perspective of the above-mentioned fundamental challenges, and to contrast econophysics models with traditional economic models, showing that the latter leave out important features. it is important to demonstrate what properties of economic systems cannot be understood for fundamental reasons within the mainstream framework (i.e. cannot be dealt with by additional terms within the modeling class that is conventionally used). in other words, one needs to show why a paradigm shift is unavoidable, and this requires careful argumentation. we are not claiming that this has not been done in the past, but it certainly takes an additional effort to explain the essence of the econophysics approach in the language of economics, particularly as mainstream economics may not always provide suitable terms and frameworks to do this. this is particularly important, as the number of econophysicists is small compared to the number of economists, i.e. a minority wants to convince an established majority. to be taken seriously, one must also demonstrate a solid knowledge of related previous work of economists, to prevent the stereotypical reaction that the subject of the paper has been studied already long time ago (tacitly implying that it does not require another paper or model to address what has already been looked at before). a reasonable and promising strategy to address the above fundamental and practical challenges is to set up multi-disciplinary collaborations in order to combine the best of all relevant scientific methods and knowledge. it seems plausible that this will generate better models and higher impact than working in separation, and it will stimulate scientific innovation. physicists can contribute with their experience in handling large data sets, in creating and simulating mathematical models, in developing useful approximations, in setting up laboratory experiments and measurement concepts. current research activities in economics do not seem to put enough focus on: • modeling approaches for complex systems [ ] . • computational modeling of what is not analytically tractable anymore, e.g. by agent-based models [ ] [ ] [ ] . • testable predictions and their empirical or experimental validation [ ] . • managing complexity and systems engineering approaches to identify alternative ways of organizing financial markets and economic systems [ , , ] , and • an advance testing of the effectiveness, efficiency, safety, and systemic impact (side effects) of innovations, before they are implemented in economic systems. this is in sharp contrast to mechanical, electrical, nuclear, chemical and medical drug engineering, for example. expanding the scope of economic thinking and paying more attention to these natural, computer and engineering science aspects will certainly help to address the theoretical and practical challenges posed by economic systems. besides physics, we anticipate that also evolutionary biology, ecology, psychology, neuroscience, and artificial intelligence will be able to make significant contributions to the understanding of the roots of economic problems and how to solve them. in conclusion, there are interesting scientific times ahead. it is a good question, whether answering the above list of fundamental challenges will sooner or later solve the practical problems as well. we think, this is a precondition, but it takes more, namely the consideration of social factors. in particular, the following questions need to be answered: . how do costly punishment, antisocial punishment, and discrimination come about? . how can the formation of social norms and conventions, social roles and socialization, conformity and integration be understood? . how do language and culture evolve? . how to comprehend the formation of group identity and group dynamics? what are the laws of coalition formation, crowd behavior, and social movements? . how to understand social networks, social structure, stratification, organizations and institutions? . how do social differentiation, specialization, inequality and segregation come about? . how to model deviance and crime, conflicts, violence, and wars? . how to understand social exchange, trading, and market dynamics? we think that, despite the large amount of research performed on these subjects, they are still not fully understood. the ultimate goal would be to formulate mathematical models, which would allow one to understand these issues as emergent phenomena based on first principles, e.g. as a result of (co-)evolutionary processes. such first principles would be the basic facts of human capabilities and the kinds of interactions resulting from them, namely: . birth, death, and reproduction. . the need of and competition for resources (such as food and water). . the ability to observe their environment (with different senses). . the capability to memorize, learn, and imitate. . empathy and emotions. . signaling and communication abilities. . constructive (e.g. tool-making) and destructive (e.g. fighting) abilities. . mobility and (limited) carrying capacity. . the possibility of social and economic exchange. such features can, in principle, be implemented in agent-based models [ ] [ ] [ ] [ ] [ ] [ ] . computer simulations of many interacting agents would allow one to study the phenomena emerging in the resulting (artificial or) model societies, and to compare them with stylized facts [ , , ] . the main challenge, however, is not to program a seemingly realistic computer game. we are looking for scientific models, i.e. the underlying assumptions need to be validated, and this requires to link computer simulations with empirical and experimental research [ ] , and with massive (but privacy-respecting) mining of social interaction data [ ] . in the ideal case, there would also be an analytical understanding in the end, as it has been recently gained for interactive driver behavior [ ] . as well as for inspiring discussions during a visioneer workshop in zurich from january how to understand human decision-making? how to explain deviations from rational choice theory and the decision-theoretical paradoxes? why are people risk averse? . how does consciousness and self-consciousness come about? how to understand creativity and innovation? . how to explain homophily, i.e. the fact that individuals tend to agglomerate, interact with and imitate similar others? how to explain social influence, collective decision making, opinion dynamics and voting behavior? why do individuals often cooperate in social dilemma situations? how do indirect reciprocity, trust and reputation evolve? references how did economists get it so wrong? the financial crisis and the systemic failure of academic economics (dahlem report address of the governor of the central bank of barbados on the futurict knowledge accelerator: unleashing the power of information for a sustainable future global solutions: costs and benefits economics as religion: from samuelson to chicago and beyond the counter revolution of science the case of flexible exchange rates adaptive behavior and economic theory bounded rationality. the adaptive toolbox the bounds of reason: game theory and the unification of the behavioral sciences on information efficiency and financial stability simon models of man simple heuristics that make us smart judgment under uncertainty: heuristics and biases physics of risk and uncertainty in quantum decision making a theory of fairness, competition, and cooperation foundations of human sociality: economic experiments and ethnographic evidence from fifteen small-scale societies social distance and other-regarding behavior in dictator games measuring social value orientation economics and restaurant gratuities: determining tip rates the cement of society: a study of social order the rewards of punishment. a relational theory of norm enforcement the emergence of homogeneous norms in heterogeneous populations the emperor's dilemma: a computational model of selfenforcing norms non-explanatory equilibria: an extremely simple game with (mostly) unattainable fixed points the myth of the folktheorem deterministic chaos for example, three-body planetary motion has deterministic chaotic solutions, although it is a problem in classical mechanics, where the equations of motion optimize a lagrangian functional on formally undecidable propositions of principia mathematica and related systems on computable numbers, with an application to the entscheidungsproblem the behavior of stock market prices an inquiry into the nature and causes of the wealth of nations the theory of moral sentiments ( ) evolution of indirect reciprocity revisiting market efficiency: the stock market as a complex adaptive system there is no invisible hand. the guardian optimal self-organization traffic and related self-driven many-particle systems from crowd dynamics to crowd safety: a video-based analysis stability and stabilization of time-delay systems network-induced oscillatory behavior in material flow networks and irregular business cycles the tragedy of the commons altruistic punishment in humans switching phenomena in a system with no switches animal spirits: how human psychology drives the economy, and why it matters for global capitalism testing behavioral simulation models by direct experiment a mathematical model for behavioral changes by pair interactions multistability in a dynamic cournot game with three oligopolists evolutionary establishment of moral and double moral standards through spatial interactions abzweigungen einer periodischen lösung von einer stationären lösung eines differentialgleichungssystems dynamic decision behavior and optimal guidance through information services: models and experiments the dynamics of general equilibrium modeling the stylized facts in finance through simple nonlinear adaptive systems predicting moments of crisis in physics and finance" during the workshop "windows to complexity the chemical basis of morphogenesis lectures on nonlinear differential equation-models in biology settlement formation ii: numerical simulation self-organization in space and induced by fluctuations stokes integral of economic growth: calculus and the solow model hysteresis and the european unemployment problem pattern formation and dynamics in nonequilibrium systems critical phenomena in natural sciences: chaos, fractals, selforganization and disorder the science of disasters noise-induced transitions: theory and applications in physics, chemistry, and biology brownian motors: noisy transport far from equilibrium catastrophe theory introduction to phase transitions and critical phenomena managing complexity: an introduction. pages - in systemic risks in society and economics systemic risk in a unifying framework for cascading processes on networks how nature works: the science of self-organized criticality why economists failed to predict the financial crisis physics and social science-the approach of synergetics nonlinear economic dynamics the self-organizing economy self-organization of complex structures: from individual to collective dynamics complex economic dynamics sociodynamics: a systematic approach to mathematical modelling in the social sciences business dynamics: systems thinking and modeling for a complex world growth theory, non-linear dynamics and economic modelling complexity hints for economic policy coping with the complexity of economics handbook of research on complexity the rise and decline of nations: economic growth, stagflation, and social rigidities the evolution of cooperation a general theory of equilibrium selection phase transitions to cooperation in the prisoner's dilemma evolutionary games and spatial chaos social diversity promotes the emergence of cooperation in public goods games the outbreak of cooperation among success-driven individuals under noisy conditions governing the commons. the evolution of institutions for collective action topological traps control flow on real networks: the case of coordination failures modelling supply networks and business cycles as unstable transport phenomena the selforganization of matter and the evolution of biological macromolecules the theory of economic development the changing face of economics. conversations with cutting edge economists beyond the representative agent economics with heterogeneous interacting agents modeling aggregate behavior and fluctuations in economics collection of papers on an analytical theory of traffic flows in the economy as an evolving complex system ii the economy as an evolving complex system iii evolutionary economics ecology for bankers origin of wealth: evolution, complexity, and the radical remaking of economics introduction to forest ecosystem science and management models of biological pattern formation multi-objective management in freight logistics: increasing capacity, service level and safety with optimization algorithms handbook of research on nature-inspired computing for economics and management self-control of traffic lights and vehicle flows in urban road networks biologistics and the struggle for efficiency: concepts and perspectives analytical calculation of critical perturbation amplitudes and critical densities by non-linear stability analysis of a simple traffic flow model complexity cube for the characterization of complex production systems verfahren zur koordination konkurrierender prozesse oder zur steuerung des transports von mobilen einheiten innerhalb eines netzwerkes (method for coordination of concurrent processes for noise sensitivity of portfolio selection under various risk measures the logic of failure: recognizing and avoiding error in complex situations complexity and the enterprise: the illusion of control nurturing breakthroughs: lessons from complexity theory monetary relationships: a view from threadneedle street. (papers in monetary economics, reserve bank of australia, ); for applications of le chatelier's principle to economics see also p. a. samuelson, foundations of economic analysis structures of social life: the four elementary forms of human relations understanding and assessing the motivations of volunteers: a functional approach happiness: a revolution in economics the wisdom of crowds: why the many are smarter than the few and how collective wisdom shapes business swarm intelligence. introduction and applications economics as a science of human behaviour: towards a new social science paradigm growth, innovation, scaling and the pace of life in cities uncertainty and surprise in complex systems scale-free networks: a decade and beyond economic networks: the new challenges hyperselection and innovation described by a stochastic model of technological evolution introduction to econophysics: correlations and complexity in finance minority games: interacting agents in financial markets economics needs a scientific revolution the economy needs agent-based modelling meltdown modelling introduction to quantitative aspects of social phenomena econophysics forum fifteen years of econophysics: worries, hopes and prospects worrying trends in econophysics econophysics and sociophysics: trends and perspectives aims and scope of the physics of socio-economic systems division of the german physical society pluralistic modeling of complex systems handbook of computational economics simulation modeling in organizational and management research developing theory through simulation methods understanding complex social dynamics: a plea for cellular automata based modelling platforms and methods for agent-based modeling from factors to actors: computational sociology and agent-based modeling artificial societies. multiagent systems and the micro-macro link in sociological theory generative social science: studies in agent-based computational modeling the handbook of experimental economics managing complexity: concepts, insights, applications engineering economy statistical physics of social dynamics the future of social experimenting the authors are grateful for partial financial support by the eth competence center "coping with crises in complex socio-economic systems" (ccss) through eth research grant ch - - and by the future and emerging technologies programme fp -cosi-ict of the european commission through the project visioneer (grant no.: ). they would like to thank for feedbacks on the manuscript by kenett dror, tobias preis and gabriele tedeschi key: cord- -fwz chzf authors: myserlis, pavlos; radmanesh, farid; anderson, christopher d. title: translational genomics in neurocritical care: a review date: - - journal: neurotherapeutics doi: . /s - - - sha: doc_id: cord_uid: fwz chzf translational genomics represents a broad field of study that combines genome and transcriptome-wide studies in humans and model systems to refine our understanding of human biology and ultimately identify new ways to treat and prevent disease. the approaches to translational genomics can be broadly grouped into two methodologies, forward and reverse genomic translation. traditional (forward) genomic translation begins with model systems and aims at using unbiased genetic associations in these models to derive insight into biological mechanisms that may also be relevant in human disease. reverse genomic translation begins with observations made through human genomic studies and refines these observations through follow-up studies using model systems. the ultimate goal of these approaches is to clarify intervenable processes as targets for therapeutic development. in this review, we describe some of the approaches being taken to apply translational genomics to the study of diseases commonly encountered in the neurocritical care setting, including hemorrhagic and ischemic stroke, traumatic brain injury, subarachnoid hemorrhage, and status epilepticus, utilizing both forward and reverse genomic translational techniques. further, we highlight approaches in the field that could be applied in neurocritical care to improve our ability to identify new treatment modalities as well as to provide important information to patients about risk and prognosis. electronic supplementary material: the online version of this article ( . /s - - - ) contains supplementary material, which is available to authorized users. translational genomics represents a diverse collection of research approaches that leverage human genomics and model systems to identify new approaches to treat and prevent disease and improve healthcare ( , ) . rooted by the central dogma of dna to rna to protein, genomic research examines the entire genome concurrently, and may include analyses of dna variants in association with traits of interest as well as the impact of genomic variation on gene transcription and translation. genomic research has been enabled by technological advances to accurately and cost-effectively study variation across the genome at scale, as well as computational techniques to store and analyze genomic data quickly and efficiently ( ) . while translational research is often defined in terms of the traditional "bench to bedside" techniques that advance discoveries from model systems through biomarkers and mechanisms ultimately to clinical applications, genomic research offers a strong use-case for an alternative approach. termed "reverse translation," this approach starts with humans as the model system, utilizing genomic associations to derive new information about biological mechanisms that can be in turn studied further in vitro and in animal models for target refinement (fig. ) . both of these approaches possess advantages and drawbacks ( , ) . forward translation depends on the relevance of the model system to human disease, both in terms of the physiologic responses to disease or insult, as well as the approach taken to perturb the system. for instance, the human applicability of genomic studies of the response to traumatic brain injury (tbi) in a mouse model require that the mouse's response to tbi is analogous to a human's, and that the approach taken to pavlos myserlis and farid radmanesh contributed equally to this work. create a tbi in the mouse provokes a similar pattern of injury seen in human tbi ( ) ( ) ( ) ( ) . as such, a great deal of careful work is required to demonstrate the validity of these model systems before the results arising from them can be judged relevant to human disease. the challenges of bridging this divide are illustrated by the universal failure of neuroprotection mechanisms that reached human trials in the last several decades, essentially all of which had promising model system data in preclinical development ( ) ( ) ( ) ( ) ( ) . reverse genomic translation, in contrast, begins with humans ( fig. ). as such, there are few concerns as to the relevance of the system for discovery of biomarkers and mechanisms of disease. however, this approach carries a new series of challenges in study design and data acquisition ( ) . compared to isogenic cell lines or carefully bred animals in a controlled setting, humans are highly variable in both their environmental and genetic exposures. this is advantageous in identifying genetic susceptibility to disease risk and outcomes, but teasing out these small genetic effects from highly variable non-genetic exposures requires both careful computational techniques as well as large sample sizes. furthermore, because genomic data is both identifiable and can potentially lead to discrimination, human genomic studies require complex consent and data management procedures ( ) . in neurocritical care, the relative rarity of many of the diseases we encounter, coupled with the challenges of critical illness and surrogate consent make human genomic studies all the more difficult to execute effectively ( ) ( ) ( ) ( ) . neurointensivists routinely encounter diseases and complications for which there are a dearth of effective treatments, or even foundational knowledge of their underlying pathophysiologic mechanisms ( , ) . in this review, we will highlight some of the approaches being taken to apply translational genomics to the study of diseases commonly found in neurocritical care, utilizing both forward and reverse genomic translational techniques. further, we will highlight some of the best practices in the field that could be applied in neurocritical care to improve our ability to identify new treatment modalities as well as risk and prognosis information to patients and their families. in advance of the human genome project and the hapmap consortium, genetic studies were confined to the study of candidate genes and lower-resolution genome-wide techniques such as categorization of restriction fragment length polymorphisms (rflp), tandem repeats, and microsatellites ( ) . these genomic features enabled early efforts to perform linkage analyses in families with related traits and disorders, as well as selected populations of unrelated individuals. careful work in this arena led to validated discoveries that have survived replication in the common era, such as chromosome in late-onset alzheimer disease (ad), ultimately mapped to the apoe locus that has become a target for a great deal of genetic research in ad, as well as many other diseases including tbi and intracerebral hemorrhage (ich) ( ) ( ) ( ) ( ) . still, much of the pregwas era was characterized by candidate gene studies that suffered from low statistical power and multiple sources of confounding that led to a failure to replicate many reported associations in the gwas era that followed ( , ) . the most substantial source of confounding in candidate gene analyses is population stratification, in which differences in allele frequency due to ancestral imbalance between cases and controls introduces spurious associations (positive or negative) between genotype and trait based solely on these cryptic ancestral imbalances ( , ) . even in studies of apoe in european ancestry populations, uncontrolled variation in the percentages of individuals of northern vs. southern european ancestry between cases and controls can mask true associations between apoe and ich, for instance ( ) . the gwas era, in which variants across the genome could be reliably genotyped and mapped to a common reference template by chromosomal location, ushered in a new system of best practices that could minimize the contribution of many of the sources of confounding in describing associations between genomic variation and traits or diseases. the international hapmap consortium obtained genotypes on individuals across ancestral populations around the globe, creating a resource that described the patterns of allele frequency variation across diverse populations ( ) . with these breakthroughs and a number of landmark evolutions that followed, case/control and population-based gwas have led to the identification of over , associations with human diseases and other traits (https://www.ebi.ac.uk/gwas/). obviously there is an enormous disconnect between the discovery of genetic loci and leveraging of this information for human benefit, which is where the translational genomic work that serves as the topic of the present review becomes relevant ( ) . post-gwas, in addition to functional and translational efforts, the movement has been towards so-called "next-generation sequencing" methodologies consisting of whole exome sequencing (wes) and whole genome sequencing (wgs). using these approaches, each nucleotide in the exome or genome is ascertained with high reliability, permitting the identification of rare and de novo variants that escape detection in traditional gwas ( ) . wes captures within-gene coding variation only, offering detection of variants that may more directly impact protein structure and function than non-coding variation detected by wgs ( ) . because the coding exome is only~ % of the overall genome, it is more cost-effective than wgs, but debate continues as to which is the more appropriate tool for large-scale study of the human genome ( ) . regardless, both wes and wgs remain orders of magnitude more expensive than traditional gwas approaches at this time, and as such well-powered sequencing studies remain unreachable for many diseases in the current pricing models. less common diseases and conditions that one may find in a neurocritical care unit are doubly disadvantaged, as even larger sample sizes are required for sequencing analyses than gwas, due to the need for many observations to identify rare exonic or intronic variants associated with disease ( , ) . as pricing models improve and larger and larger community or hospital-based cohorts receive sequencing through clinical or biobanking efforts, it is hoped that even uncommon conditions such as subarachnoid hemorrhage or status epilepticus will benefit from the insights achievable through sequencing analysis, where case/control and smaller sequencing studies have shown promise ( , ) . obviously genomic research need not be limited solely to human studies. a wealth of information about disease pathogenesis and response to injury can be gleaned from model systems of human conditions using genomic and transcriptomic approaches. because animal models and isogenic tissue cultures are specifically designed to limit genetic differences between individual animals or plated cells, dna-based association tests typically do not offer insight in the same way that they do in humans. as such, many model system studies start with rna, examining how the genome responds to perturbation through the transcriptome. however, there are substantial genomic differences between model systems and humans, as coding sequences are not necessarily conserved, promoter and enhancer control of gene expression can vary, and in the case of immortalized tissue and cell-based assays, the chromosomal architecture itself can be quite different from the organism from which it was derived ( , ) . these differences can be highly relevant when determining whether observed transcriptomic and proteomic results from model systems are likely to be shared in humans. with those caveats, the dynamic nature of the transcriptome in model systems offers opportunities to assess the way in which the genome responds to noxious insults or drug exposures, and in animal models this can even be done across specific organs or tissues of interest ( ) . as one example, traumatic brain injury researchers have obtained insight into both the initial injury cascade as well as brain response to potential injury modulators such as valproate using animal models and transcriptional microarrays, in which rna expression patterns in brain tissue can be rapidly and replicably assessed across the transcriptome ( ) . using more recent technological advancements such as drop-seq, rna expression can be assessed in single cells, as has been done in individual hippocampal neurons in a mouse model of tbi ( ) . at a minimum, these elegant studies can help to identify relevant cell types important in the response to injury, highlighting testable hypotheses that may be important in human conditions, all with access to tissues and control over experimental conditions that would never be possible in human-based research. given that diseases common to the neurocritical care population so rarely afford access to brain tissue for pathologic or genomic analysis antemortem, model system genomic studies offer an important adjunct for translational research. forward genomic translation begins with model systems with the goal of using the measured associations in these models to derive insight into biological mechanisms that may also be relevant in human disease. forward translation requires wellcharacterized models that are often designed to mimic the human exposures of interest as closely as possible. this is often challenging given the natural differences between humans and many of the animals chosen to serve as models. in this section, we will highlight several model systems in current use for translational genomics relevant to neurocritical care, but the field of translational modeling in neurologic disease is suitably large to prevent an exhaustive review herein. malignant cerebral edema is a highly lethal complication of ischemic stroke, with mortality of - % ( ) . currently, hemicraniectomy is the only available option to prevent death and yet it does not address the underlying pathophysiology. hyperosmolar therapy is potentially useful as a bridge to surgery. preclinical data based on a forward translation approach has been useful in highlighting mechanisms underlying postinfarct edema as potential targets for therapeutic manipulation. the sulfonylurea receptor (sur ) is encoded by the abcc gene that is upregulated after cns injury, forming an ion channel in association with transient receptor potential melastatin (trpm ). continuous activation of this complex can lead to cytotoxic edema and neuronal cell death, which has been demonstrated in both animal and human models ( , ) . sur is also found in pancreatic beta cells, constituting the target for the oral hypoglycemic agent, glyburide. studies of rodent and porcine stroke models demonstrated that in the first few hours after an ischemic insult, both sur and trpm are upregulated ( , ) . limited case series of human postmortem specimens also demonstrated upregulation of sur in infarcted tissue ( ) . therefore, intravenous glyburide has been proposed for treatment of malignant cerebral edema. targeting sur in rat models of ischemia have consistently resulted in reduced edema and better outcomes ( ) . in particular, glyburide infusion starting h after complete middle cerebral artery occlusion resulted in decreased swelling by two thirds and % reduction in mortality ( ) . one desirable characteristic of glyburide is that it cannot penetrate intact blood-brain barrier, but that is facilitated following brain injury ( ) . the effect of glyburide for treatment of cerebral edema has also been studied in tbi with promising data obtained from animal studies ( ) . limited randomized trials in human using oral glyburide have shown promising results; however, use of oral formulation and study design limitations prohibit generalizability of results ( , ) . building on this preclinical data, the phase randomized clinical trial (games-rp) showed that the iv preparation of glyburide, glibenclamide, is associated with reduction in edema-related deaths, less midline shift, and reduced rate of nih stroke scale deterioration. however, it did not significantly affect the proportion of patients developing malignant edema ( ) . the phase charm trial, sponsored by biogen, is currently enrolling patients with large hemispheric infarction to determine whether iv glibenclamide improves -day modified rankin scale scores. if this trial proves successful, this vignette will represent a dramatic success story for the forward translation paradigm in genomic research. in the light of recent advances in revascularization therapy, the national institute of neurological disorders and stroke has supported an initiative aiming to develop neuroprotective agents to be used as adjunctive therapy to extend the time window for reperfusion and to improve long-term functional outcome. this stroke preclinical assessment network (span) supports late-stage preclinical studies of putative neuroprotectants to be administered prior to or at the time of reperfusion, with long-term outcomes and comorbidities constituting the endpoint. the goal is to determine if an intervention can improve outcome as compared to reperfusion alone and/or extend the therapeutic window for reperfusion. span directly applies to forward translation efforts in preclinical models of neuroprotection after stroke and is an outstanding opportunity to stimulate research efforts in a field more remembered for its past failures than the promise it holds for the future of therapeutic development in the area. other societies have also begun to endorse more comprehensive modeling approaches in areas with few therapeutic options with the hope of implementing a paradigm shift. for example, the neurocritical care society has initiated "curing coma" campaign with the -to -year mission to improve the understanding of the mechanisms and to ultimately develop preventative and therapeutic measures. traumatic brain injury (tbi) is among the leading causes of disability and death worldwide, particularly in the young. the type of tbi is in part determined by the attributes of mechanical forces, including objects or blasts striking the head, rapid acceleration-deceleration forces, or rotational impacts. following the primary injury, an intricate cascade of neurometabolic and physiological processes initiates that can cause secondary or additional injury ( , ) . intensive care management has improved the prognosis of tbi patients; however, specific targeted treatments informed by pathophysiology could have a tremendous impact on recovery. the period of secondary tissue injury is the window of opportunity when patients would potentially benefit from targeted interventions, given that in tbi, the primary injury cannot be intervened upon by the neurologist or intensivist. the goal of therapy is therefore to reduce secondary damage and enhance neuroplasticity. the utility of animal models of tbi primarily depends on the research question, as each model emulates specific aspects of injury and has selective advantages and disadvantages. these include biomechanics of initial injury, molecular mechanisms of tissue response, and suitability for high-throughput testing of therapeutic agents, to name a few. although phylogenetically higher species are likely more representative models for human tbi, rodent models are more commonly used given the feasibility to generate and measure outcomes, as well as ethical and financial limitations of higher-order models. table summarizes some common and representative tbi models [ table ]. in contrast with the rodent models described in table , other model systems in tbi have been selected specifically to study other aspects of the physiologic response to tbi. for example, a swine model of controlled cortical impact offers the opportunity to readily monitor systemic physiologic parameters such as tissue oxygen and acid-base status while investigating therapeutic interventions, which is argued to provide greater insight into human response to injury ( ) . translation of preclinical studies using these animal tbi models to humans is inherently challenging. differences in brain structure, including geometry, craniospinal angle, gyral complexity, and white-gray matter ratio, particularly in the rodent models, can result in different responses to trauma ( ) . the limitation of extrapolating animal studies to human is also manifested at the genetic level, as differences in gene structure, function, and expression levels may suggest genetic mechanisms that are incompletely correlated with humans. as an example, female sex may be associated with better outcome through the neuroprotective effect of progesterone in animal models, but these observations did not carry over to humans in the protect-iii trial ( , ) . variable outcome measures, including neurobehavioral functional tests, glasgow outcome scale correlates, and high-resolution mri have been used in attempts to correlate animal responses to injury with those of humans. the lack of a large cache of standardized tools further limits comparison or pooling the results of different studies that use variable models of tbi or outcome measurement. transcriptomics, a genomic technique in which global rna expression is quantified through either expression microarrays or rna sequencing, has been employed to characterize specific inflammatory states following tbi. many studies have assessed the transcriptome in the acute post-tbi interval within - days after injury, with some showing upregulation of inflammation and apoptosis genes. gene ontology analysis at months post-tbi have shown similar changes, with upregulation of inflammatory and immune-related genes ( ) . importantly, late downregulation of ion channel expression in the peri-lesional cortex and thalamus suggests that this delayed examination of the transcriptome could be valuable for revealing mechanisms relevant to chronic tbi morbidities, including epileptogenesis and prolonged cognitive impairment ( ) . in addition, tissue-specific analysis of gene diffuse axonal injury reproduces human tbi needs standardization, e.g., location of animal within shock tube and heard immobilization ( ) expression across cell types in brain could provide useful insight into cell-specific pathways. for example, temporal trending of microglial expression profile indicates a biphasic inflammatory pattern that transitions from downregulation of homeostasis genes in the early stages to a mixed proinflammatory and anti-inflammatory states at subacute and chronic phases ( ) . the list of antiepileptic drugs has expanded significantly in the past decade, reflecting substantial investment in the search for new therapeutics with better efficacy and tolerability. however, the list of options with demonstrated efficacy in status epilepticus (se) has remained limited. the utility of benzodiazepines, often deployed in the field as a first-line agent, decreases with increasing duration of se. in addition, - % of patients with se develop refractory se when they fail to respond to first-and second-line therapy, posing a significant management and prognostic challenge ( ) . the development of aeds has relied substantially on preclinical animal models to establish efficacy and safety prior to proceeding to human trials. different epilepsy models exist that are each useful for different aspects of drug development and no model is suitable for all purposes. the majority of animal models induce epilepsy using electroshock or chemical seizure induction. nearly all recent aeds have been discovered by the same conventional models, and the reliance on these common screening models has been implicated as one of the reasons for the low yield of drugs with efficacy in refractory epilepsy ( ) . the pros and cons for each epilepsy model are discussed in detail in several excellent reviews ( , ) . some of the chemicals used include kainic acid, pilocarpine, lithium, organophosphates, and flurothyl ( ) . sustained electrical stimulation to specific sites, including the perforant path, the ventral hippocampus, the anterior piriform cortex can induce se ( ) . the latency, length, and mortality of convulsive se are more variable in chemoconvulsant as compared to electrical models, which are in turn determined by the drug and route of administration, species, sex, age, strain, and genetic background among other factors ( ) . it should also be noted that the presence of behavioral convulsion does not correlate fully with the electrographic data and vice versa. this can have c r i t i c a l i m p l i c a t i o n s w h e n s t u d y i n g d r u g s f o r pharmacoresistant se. therefore, it has been suggested that electroencephalographic quantification be used to measure the severity of se ( ) . furthermore, the genetic background and expressivity of animals can have a significant effect on seizure susceptibility, even between batches of inbred mice ( ) . proteomic and transcriptomic approaches have been utilized for assessment of alterations in expression profile following se, demonstrating that certain subsets of genes are upregulated at each timepoint following the onset of se. specifically, upregulation of genes regulating synaptic physiology and transcription, homeostasis and metabolism and, cell excitability and morphogenesis occur at immediate, early, and delayed timepoints. in addition, related studies have demonstrated changes in expression of micrornas related to epileptogenesis, including mirna- and mi-rna- following se ( , ) . selective rna editing post-transcription is yet another potential source of proteomic diversity in preclinical models of se, and merits further investigation as a modulator of protein levels that may be less closely tethered to gene expression ( ) . aneurysmal subarachnoid hemorrhage (sah) has an earlier age of onset and is associated with higher morbidity compared with other stroke subtypes. the pathophysiology of insult has traditionally been studied under two time-intervals, early brain injury (ebi) and, cerebral vasospasm (cv) and delayed cerebral ischemia (dci). the prime goal of translational research in this arena is to identify the mechanisms and targets related to the risk, severity, evolution and outcome. about % of patients die immediately following sah ( ) . thereafter, early brain injury within the first days, followed by dci are the most feared complications. cv is the phenomenon with strongest association with the development of dci, which - % of patients experience between day to ( ) . the underlying mechanisms leading to cv remain poorly understood and have therefore been a prime focus of preclinical studies. the majority have used rodent models, but primate, swine, and dog models have also been employed ( ) . cerebral aneurysms are difficult to model and hence two common approaches to modeling sah use alternative strategies. the first is direct injection of blood into the subarachnoid space, specifically into either the prechiasmatic cistern or cisterna magna, to generate sah predominantly in the anterior or posterior circulation territories, respectively ( ) . the second model, endovascular suture, passes a suture or filament through the internal carotid artery, creating a hole in one of the major branches resulting in egress of variable amount of blood into the subarachnoid space ( ) . variations in some parameters of the first method, including injected blood volume, csf removal prior to injection to prevent egress of blood into the spinal canal, and replenishing intravascular volume to keep cerebral perfusion pressure constant through maintenance of mean arterial pressure, as well as the rapidity of injection have raised questions about comparability and biofidelity of the results ( ) ( ) ( ) ( ) . the latter model appears to remove some of the mentioned confounding factors, as the hemorrhage occurs at physiologic mean arterial pressure (map) and intracranial pressure (icp), but is limited by variable puncture site and ultimate hemorrhage volume. another potential drawback is the period of ischemia caused by the intraluminal suture, although the occlusion period is typically not judged to be long enough to cause significant ischemia. the missing element in these models is the absence of aneurysm formation and rupture, and consequently the vascular processes intrinsic to the aneurysm itself that influence dci. as such, some studies have used combinations of interventions to generate aneurysms, including induced hypertension via unilateral nephrectomy and administration of angiotension ii or deoxycorticosterone acetate, as well as elastase injection. the downside of these models is that the timing of aneurysm rupture cannot be reliably predicted, which limits close monitoring and physiologic assessments in the early phase following sah, blurring the timing of dci ( ) ( ) ( ) . the immediate hemodynamic changes following the hemorrhage are monitored via a variety of methods. regardless of the method chosen, reports on the direction and range of values of cpp, cbf, and map can be quite variable, both within the same model and between different models. a common technique to measure blood flow is laser doppler flowmetry that provides a continuous measure of cortical perfusion. although it does not measure global cerebral blood flow and has spatial limitation, it appears to be relatively reliable and technically reproducible. other methods of flow measurement include radiolabeling methods and mri with the latter has the advantage of capturing the dynamic nature of the condition, as well as global and region-specific blood flows. as noted, cv and dci are responsible for delayed morbidity and mortality. given that these manifestations typically occur while patients are inpatient for care of their sah, therapeutic interventions are more feasible compared to the hyperacute phase when the processes leading to initial damage may have already occurred. however, monitoring for cv in animal models is not straightforward. one method of identifying cv is measuring the intraluminal diameter of vessels on histological samples. in addition to being an end-measure and therefore precluding measurements at different time points in the same animal, varying degrees of tissue desiccation among samples may yield numbers different from actual in vivo values. digital subtraction angiography and magnetic resonance angiography can provide a real-time evaluation, but the severity of cvand its timing, as well as neuronal cell death varies depending on the model and the affected vessels ( ) . the foundational molecular pathways that orchestrate cv are complex and remain incompletely elucidated. however, translational research using many of the above models has demonstrated that endothelin- , nitric oxide, and an inflammatory cascade ignited by breakdown of blood products play predominant roles. endothelin- is a potent vasoconstrictor produced by infiltrated leukocytes, and based on this notion, clazosentan was developed as an endothelin- receptor antagonist to combat cv. in human trials, clazosentan was found to significantly reduce the incidence of the dci without improving the functional outcome, and this or a related approach could ultimately prove beneficial if off-target drug effects, including pulmonary complications, hypotension, and anemia can be mitigated ( , ) . hemoglobin and its degradation products are also a strong stimulus for cv through direct oxidative stress on arterial smooth muscle, decreased nitric oxide production and, increasing endothelin and free radical production ( ) . this suggests that facilitating clearance of hemoglobin degradation products from the csf may be a potential therapeutic target. modulating the intense inflammatory response is also intuitive and while preclinical results support this notion in general, the evidence has thus far not been judged adequate to justify clinical trials. for example, il- receptor antagonist (il- ra) reduces blood-brain barrier (bbb) breakdown, a biomarker that is itself correlated with the severity of brain injury, and work continues to determine whether this or related pathways mediating bbb permeability might have therapeutic promise ( ) . given these numerous and likely interconnected mechanisms of delayed brain injury, further research is needed to understand their relative applicability to humans, and whether targeting a single pathway or a number of pathways simultaneously is likely to be the most adaptive strategy to reduce cv and dci in humans. the results of genome-wide rna sequencing analysis have supported the primary role of neuroinflammation in the pathogenesis of early brain injury. some studies have specifically found a key role for long non-coding rna (lncrna), a type of rna without protein-coding potential that are particularly abundant in the brain, in modulating the inflammatory behaviors of microglial cells ( ) . high-throughput mass spectrometry has also been utilized in demonstration of differential expression of proteins in the cerebral vessels after sah, as well as for monitoring the effect of experimental therapeutics ( ). we will not cover these proteomic studies in detail here, as they typically fall outside the rubric of what is classically considered "genomics", but their approach, which leverages global protein signatures rather than restricting observations to specific compounds, shares many similarities with genomics. as mentioned above, reverse genomic translation refers to an approach to the study of a disease by starting with humans using either cohort-based or case/control genomic studies. the observations made through the course of these studies then inform on the best approach for target validation and refinement to prioritize candidate mechanisms and related endophenotypes for therapeutic development. it has been shown that candidate compounds with independent confirmation of their therapeutic target via human genomics are more than twice as likely to prove effective in clinical trials ( ) . therefore, the reverse translation approach would seem an adaptive strategy to identify disease-associated mechanisms and therapeutic targets with the best chance of impacting clinical care in the near term. however, the approach to reverse translation requires large sample sizes with well-characterized patient data in order to achieve a statistically confident result. these large sample sizes raise the issue of variability in risk and treatment exposures between participants, which could impact patient outcomes independently of genomic effects and therefore erode power to detect genetic risk. the utility of reverse translation in target refinement and mechanism exploration in model systems can be highlighted using an example from the stroke community. recent gwas and subsequent meta-analyses of ischemic stroke and stroke subtypes in very large case/control datasets have validated the histone deacetylase (hdac ) region in chromosome p . as a major risk locus for stroke due to large artery atheroembolism (laa). this locus was also previously discovered in association with coronary artery disease (cad) ( , ) . based on these findings, azghandi et al. sought to investigate the role of the leading single nucleotide polymorphism (snp) in this genomic region (rs ) in increasing laa stroke risk ( ) . they found that rs , both in heterozygotes as well as in homozygote human carriers, is associated with increased expression of hdac in peripheral blood mononuclear cells on a dose-dependent manner, suggesting that the effect of this locus in stroke risk may be mediated by increased hdac expression. additionally, they demonstrated that hdac deficiency in mice is associated with smaller and less advanced atherosclerotic lesions in the aortic valves, curvature, and branching arteries, suggesting that hdac may increase atherogenesis and therefore represents a novel target for atherosclerosis and laa stroke prevention. notably, recent studies have suggested that both nonspecific (e.g. sodium valproate) as well as specific hdac inhibitors can have a positive impact on both stroke recurrence risk, as well as other phenotypes, including cancer. this highlights the central role that reverse translation can have in therapeutic target investigation and refinement, with potential beneficial off-target properties ( , ) . while acute stroke care is a vital component of neurocritical care at many institutions, reverse genomic translation successes in other relevant traits also merit mention. acute respiratory distress syndrome (ards) is a frequent complication of severe neurologic injury due to sah or neurotrauma. in a recent gwas by bime et al., variation in the selectin p ligand gene (selpg), encoding p-selectin glycoprotein ligand (psgl- ) was found to be associated with increased susceptibility to ards ( ) . the most significant snp in this locus, rs , which results in a missense mutation, has been successfully replicated in independent cohorts. further functional analyses have demonstrated that selpg expression was significantly increased in mice with ventilator (vili)-and lipoprotein (lps)-induced lung injury, and that psgl- inhibition with a neutralizing polyclonal antibody led to an attenuation of inflammatory response and lung injury. in selpg knockout mice, inflammatory response as well as lung injury scores were significantly reduced compared to wild-type mice ( ) . these results highlight the value of reverse genomic translation in first identifying human-relevant genetic risk factors for disease, and using model systems to understand the pathways impacted by their introduction to select rationally-informed modalities for potential treatment. intracranial aneurysms (ia) are commonly encountered in the neurocritical care setting, albeit most commonly after rupture. even so, inroads leading to a better understanding of aneurysm formation may ultimately reveal opportunities for treatments to prevent acute re-rupture or prevent future aneurysm formation after sah. the strongest associations with ia have been reported in the region near cdkn a/cdkn b in p . as well as in a nearby intragenic region known as cdkn bas or anril ( , ). anril is a long noncoding region responsible for the regulation of cdkn a and cdkn b and has also been implicated in the pathogenesis of cad and atherosclerosis, among other traits ( ) . overexpression of anril in mouse models of cad has been associated with negative atherosclerosis outcomes including increased atherosclerosis index, unfavorable lipid profiles, thrombus formation, endothelial cell injury, overexpression of inflammatory factors in vascular endothelial cells, increased apoptosis of endothelial cells, and upregulation of apoptosis-related genes. notably, reduced anril expression has been associated with reduced inflammatory, biochemical and molecular markers of atherosclerosis, indicating a potential target for atherosclerosis and ia prevention ( ) . when utilizing the reverse translation approach in genomic studies, the aforementioned examples highlight two distinct but equally important considerations for a successful implementation of such approaches. the first major consideration is that large populations of well-characterized individuals must be selected to ensure adequate statistical power to detect meaningful associations. thorough and standardized phenotyping of study subjects is one of the main predictors of the success of a gwas ( , ) . careful assignment of cases based on strict phenotypic criteria permits well-executed gwas even in diseases with heterogeneous presentations and multiple pathogenic features, such as multiple sclerosis (ms) and stroke ( ) . in neurocritical care populations where subtle characteristics of disease presentation and intermediate outcomes may represent important phenotypes for genomic investigation, such as sah, these traits should be closely defined and recorded to the greatest degree possible in all participants. this initial step is critically important in the greater scheme of reverse translational genomics, as these associations with subclasses and endophenotypes of disease often provide the biological insights needed to continue translational efforts using model systems tailored to refine observations. the second major consideration is that the execution of genomic studies needs to be comprehensive and thorough so as to permit association testing in a hypothesis-free environment. at the moment, gwas array-based studies seem to remain a favorable option of the genome, considering the lower cost associated with their utilization and proven track record in discovery, but over time, wes and wgs studies will become reachable even on more modest research budgets. for the transcriptome, rna sequencing and rna microarrays both offer unbiased surveys of global transcriptional variation, but because gene expression varies substantially by tissue it is critical that rational choices are made regarding the suitability of specific tissues for specific conditions. in uncommon conditions with necessarily small sample sizes, including neurocritical care-relevant diseases like se, sah, and ards, external validation studies can strengthen associations from an initial small discovery dataset, and in many cases these follow-up studies can make use of freely available resources. for example, in a recent expression-based gwas (egwas), microarray data for ards from the gene expression omnibus (geo) were collected and combined in an effort to identify novel genetic targets ( ) . the study not only validated previously known lung injury-and ardsrelated genes, but also discovered new candidate genes that may prove to be useful in future translational work. identifying loci, variants, expression patterns, and genenetworks with the use of human genomic studies is only the initial step in the reverse translation process. these discoveries must inform and guide the research to further understand and refine the phenotypic effects of these variants in model systems, including some of those described above. there are several techniques with which we can utilize the discoveries made from case/control genomic studies to build or modify model systems. one approach is transgenesis, in which a larger dna sequence including a human gene containing a mutation of interest, called a transgene, is injected into the pronucleus of a mouse fertilized egg. the fertilized egg is then inserted into the oviduct of a pseudopregnant female mouse, which is a female who has been mated with vasectomized male in order to achieve the hormonal profile of a pregnancy state. the offspring produced from this female can create an animal line that contains the human gene and allele of interest ( ) . however, because the transgene is inserted randomly at one or more genetic locations as either one or more copies, the level of expression and regulatory influences of the gene of interest may not initially be well-controlled across animals. as such, there are several intermediate steps that can allow more specific genetic alteration using transgenesis, involving embryonic stem cells (escs). the first step is the introduction of regulatory sequences (such as expression cassettes) into escs. then, by injecting the transgene first into these modified escs, gene expression can be more closely controlled. the escs with the transgene can then be inserted into blastocysts and give rise to new strains, using the same methods previously described ( ) . there are multiple variations on the transgenic approach which are uniquely suited to the model system being employed and can give rise to models that express transgenes in response to a particular stimulus, or in particular tissues of interest. a newer method utilizing programmable endonucleases has allowed researchers to bypass more traditional escbased methods for direct and precise gene editing. endonucleases are enzymes that cause double stranded dna (dsdna) breaks that can further be repaired either with non-homologous end-joining (nhej), an imprecise method for rejoining the dna breaks that involves various enzymes and may result in inactivating mutations, or with homologydirected repair (hdr), in which the dna breaks are repaired based on a co-injected template. four categories of programmable endonucleases have been used for direct and precise gene editing: homing endonucleases (he), zinc-finger nucleases (zfns), transcription activator-like effector nucleases (talens) and the clustered regularly interspaced short palindromic repeats/crispr-associated (crispr/cas ) system. the common characteristic of these enzymes is that they possess sequence-specific nuclease activity, allowing researchers to cleave dsdna at desired, pre-specified sites. the crispr/ cas system has proven to be the most successful so far, in terms of efficiency, cost, and simplicity of use. perhaps the most important advantage of this approach is that programmable endonucleases do not require the use of escs and can directly be inserted into one-or two-cell stage embryos, thus allowing more specific and direct gene-editing in a single step ( ) . drawbacks include enzymatic limitations as to where dna breaks can be reliably introduced, as well as off-target endonuclease activity at other sites across the genome which can disrupt gene activity in unintended ways. work is ongoing to refine these tools, improving the number of sites where gene editing can occur while also improving the specificity of the system ( ). one illustrative example of human genomic studies being used to refine models to understand disease processes is the case of human ich-associated mutations in col a and col a . col a and col a are the most abundant proteins in basement membranes. they form heterotrimers consisting of one col a and two col a peptides and are produced and modified in the endoplasmic reticulum (er). after their production, they are packaged into vesicles in the golgi apparatus and transferred to vascular endothelial basement membranes ( ) ( ) ( ) . the initial identification of mutations in this region in familial forms of cerebral small vessel disease, coupled with the subsequent detection of common col a /col a variants associated with sporadic deep ich led to the development of animal model systems to explore their effects ( ) ( ) ( ) ( ) . through mouse models, representative col a /col a mutations were found to recapitulate human disease phenotypes, with multifocal ich in subcortical regions of the forebrain and the cerebellum, as well as porencephaly, small vessel disease, recurrent intraparenchymal and intraventricular hemorrhages, agerelated ich, and macro-angiopathy ( , ) . using cellular assays and tissue derived from mouse models, mutations in col a /col a have been associated with decreased ratio of intracellular to extracellular col a , retention of abnormal collagen proteins in the er, er stress, and activation of the unfolded protein response ( ) ( ) ( ) , suggesting that the intracellular accumulation and er-stress could be an important molecular mechanism underlying ich related to col a and col a mutations. notably, treatment with the molecular chaperone sodium -phenylbutyrate resulted in decreased intracellular accumulation and significant decrease of ich severity in vivo, which could point the way towards eventual forms of treatment for both familial and sporadic col a and col a -associated ich ( ) . another recent example of model system refinement for neurocritical care-relevant disorders is status epilepticus (se). pyridoxal phosphate binding protein (plpbp) variants have been associated with a rare form of b -dependent epilepsy, which, if left untreated can lead to se. in a recent study, johnstone et al. utilized crispr/cas to create a zebrafish model lacking its encoded protein ( ) . they observed that plpbp-deficient zebrafish experienced significantly increased epileptic activity compared to their wild type counterparts, in terms of physical activity (high-speed movements), biochemistry (c-fos expression) and electrophysiologicallyrecorded neuronal activity. additionally, treatment of plpbp −/ − larvae with plp and pyridoxine led to an increase in their lifespan, and a decrease in their epileptic movements and neuronal activity. lastly, in these plpbp-deficient zebrafish, systemic concentrations of plp and pyridoxine were significantly reduced, as well as concentrations of plp-dependent neurotransmitters. collectively, these results provide insights for biomarker development and preclinical target refinement in b -dependent epilepsy. understanding how novel treatments might impact rare disease presentations could ultimately lead to new insights for common forms of disease as well, just as the discovery of rare pcsk variants in patients with very low cholesterol ultimately led to pcsk -inhibitors to treat more common forms of familial hypercholesterolemia. however, the use of animal models is not always the ideal approach to describing the effects of genetic variation, as the phenotypic alterations may be too subtle to observe or require impractical prolonged observation in late-life animals to ultimately exhibit relevant phenotypes. in these cases, tissuebased systems can provide a useful tool to study these effects. for example, ia formation, as previously described, has been associated with variants in anril. although the direct impact of these variants in human tissue or animal models is difficult to discern, work with mutations of anril in endothelial models have provided valuable insight. specifically, upregulation of anril has been associated with increased expression of inflammatory and oxidative markers in the vascular tissue such as il- , il- , nf-κb, tnf-a, inos, icam- , vcam- , and cox- ( , ) . these observations provide vital information about cellular mechanisms impacted by human disease-associated genetic risk factors without requiring the expense and time investment of creating, validating, and studying animal models. ultimately such models may still be required, but prior knowledge about cellular phenotypes associated with genetic variation may be highly valuable in choosing the right model system and selecting efficient approaches to validate these systems. the aforementioned examples highlight significant contributions of the field of translational genomics in identifying novel therapeutic targets, developing biomarkers of disease severity and elucidating disease-relevant pathophysiology. undoubtedly, these contributions are valuable in application to existing model systems of disease, or through refinement of models informed by the reverse translation process. given that many of our current models have proven to be ineffective in many cases, the reverse translation approach offers a significant advantage in that the translational discoveries arising in established or refined model systems have already been proven to be relevant to human disease. this advantage provides us with reasonable expectation that observed effects in model systems will also remain relevant to human disease, providing a substrate for therapeutic development. certainly, the ultimate goal of translational genomics is to be able to transfer the discoveries found from experimental models into clinically useful information in order to improve human health. this aim, with regard to the translational genomic approach, can be satisfied with two distinct approaches. one is concerned with improving our understanding of the mechanisms of disease, providing novel targets for therapeutic development. the other is concerned with leveraging the conclusions of translational genomics through more direct applications to clinical care. we will discuss these in order. once genomic discovery and translational exploration have confirmed the mechanism and relevance of a particular genomic association, translational genomics offers the opportunity to use these same translational approaches to derive highthroughput assays for screening of compound libraries, which are collections of small molecules useful for early-stage drug-discovery ( ) . the same in vitro assays used to identify cellular phenotypes associated with genetic risk factors can be tested for amelioration or "rescue" of wild-type features after exposure to library compounds. this is particularly advantageous for the reverse genomic translation approach, as these assays are often critical components of the overall discovery cycle, and with optimization to provide ideal readouts, screening can proceed quickly. an example of success here is the identification of molecular chaperones that can ameliorate the unfolded protein response detrimental to cell survival in col a -mediated cerebrovascular disease ( ) . identifying hits in these assays has the potential to accelerate drug-discovery, provided that the mechanism can be targeted by a small molecule and not a designed biologic entity such as a monoclonal antibody. while screening can be performed using novel compound libraries, it can also be accomplished using libraries containing already-approved drugs, providing an innovative way for compound repurposing based on genetic interactions. numerous tools already exist for in silico evaluation of existing compounds based on known mechanisms, so this step can begin even in the genomic discovery phase prior to translational validation ( ) . the second approach in which translational genomics has proven to be of great potential is the rapidly evolving and highly anticipated field of precision medicine. the observations arising from translational genomics, even when not informing us about the specific mechanisms associated with the phenotype in question, may be of predictive value. this finds application in two relevant translational genomics tools: polygenic risk scores (prs) and biomarker development based on rna expression profiles. while common genetic variation can provide valuable information about disease-relevant mechanisms and help refine disease models, they are relatively weak in explaining a significant proportion of the genetic basis of complex polygenic disorders, such as cad, diabetes, stroke, or sah ( ) . by summarizing the impact of many variants of small effect across the genome simultaneously, a polygenic risk score (prs) can be developed which explains far more of the genetic risk of a disease than any common variant can individually (fig. ) . application of these prs in independent clinical populations as a predictive tool represents a novel translational approach. in a recent study examining stroke, a prs combining snps associated with atrial fibrillation (af) was found to be significantly associated with cardioembolic (ce) stroke risk and no other stroke subtypes, paving the way for a potentially useful tool to discriminate ce stroke from other etiologies without reliance on expert adjudication or longitudinal monitoring ( ) . another recent study compiled a prs of cad, demonstrating that individuals in the highest quantiles of the prs exhibited cad risk on par with known mendelian cardiac diseases ( ) . these studies highlight the potential uses of prs as a genetic biomarker of disease, capturing orthogonal risk information compared to clinical risk factors alone. much work is still needed in this arena, ranging from derivation of readily accessible clinical genomic testing, dissemination of prs results in an interpretable format, disclosure of off-target results that may be clinically meaningful in their own right, and, critically, the validation of prs in ancestrally-diverse populations ( ) . despite these challenges, utilization of polygenic risk data to directly inform patient risk independent of our understanding of the underlying mechanisms is an exciting and rapidly evolving use-case for translational genomics. development of biomarkers is another approach in translational genomics that focuses more on predictive utilization than on elucidating mechanisms, and critical care has seen some early potential applications of this approach. in sepsis, where clinical prs percentile across the population distribution. plotting percentiles by disease risk, patients in higher prs percentiles (red dots) are at correspondingly highest risk for the disease outcomes are highly heterogenous, tools that might identify patients who are more likely to respond to certain treatments or identify individuals at highest risk for morbidity and mortality would be highly useful. in a recent study by scicluna et al., the authors categorized sepsis patients based on peripheral bloodderived genome-wide expression profiles and identified four distinct molecular endotypes (mars - ) ( ) . the mars expression profile was the only category that was significantly associated with -day and -year mortality. in addition, combination of the apache iv clinical score with this genetic scoring system resulted in significant improvement in -day mortality risk prediction, compared to apache iv alone. to further aid translation to clinical application, the authors used expression ratios of combinations of genes to stratify patients to the four molecular endotypes. bisphosphoglycerate mutase (bpgm): transporter , atp binding cassette subfamily b member (tap ) ratio reliably stratified patients to mars endotype; while other protein ratios were able to assign individuals to the other three mars categories. using this approach, not only could bpgm and tap transcripts potentially be used to identify patients with increased risk of mortality, but if these categorizations can be demonstrated to be causal, these molecular pathways could also be explored for therapeutic target identification and validation. further work is required to extend these findings across clinical populations, but this approach could ultimately yield new tools for prognostication in sepsis. in ischemic stroke, tissue plasminogen activator (t-pa) response and risk for hemorrhagic transformation (ht) are highly correlated with functional outcomes, and biomarkers to predict each of these would have obvious clinical utility. in a recent study, del rio-espinola et al. found that two genetic variations (rs and rs ) were associated with increased risk for ht and mortality after t-pa administration in stroke patients ( ) . specifically, rs in a m was associated with ht and rs was associated with in-hospital mortality. in a subsequent validation study, researchers created a genetic-clinical regression score that was successfully used to stratify stroke patients treated with t-pa based on risk for ht and parenchymal hemorrhage (ph) ( ) . while in the current clinical landscape the vast majority of patients do not have readily accessible genome-wide genotypes prior to events like acute stroke, increasing uptake of clinical genomics and genomically-enabled electronic health record systems could soon enable real-time risk prediction calculations incorporating both clinical and genetic information, providing more accurate tools for clinicians to incorporate into medical decision-making. a separate set of tools that could potentially become diagnostically useful in the clinical setting is the transcriptomic approaches to identify biomarkers, using array-based screening or rna sequencing. in a recent systematic review, a total of mirnas were reported to be differentially expressed in the blood cells of patients with acute ischemic stroke within h after stroke ( ) . some studies reported the area under the curve (auc) ranging from . to . , indicating a potential for clinical utility as early diagnostic markers when neuroimaging is not immediately available or is limited by feasibility. subsequent studies were able to partially replicate these findings, showing three mirnas (mir- a- p, mir- b- p, and mir- - p) that were upregulated in the acute poststroke period, an effect independent of stroke pathophysiology and infarct volume ( ) . these transcripts were associated with an auc of . in differentiating ischemic stroke and healthy controls, a metric that significantly outperformed computed tomography, as well as previously reported bloodbased biomarkers. in ich, a recent report identified up to and transcripts from whole blood that are differentially expressed between ich and controls and, ich and ischemic stroke, respectively ( ) . when comparing ich and ischemic stroke transcriptomes in the first h, and transcripts were differentially expressed compared to controls, respectively. ich transcriptome was over-represented by t cell receptor genes compared to none for ischemic stroke and underrepresented by non-coding and antisense transcripts. t cell receptor expression successfully differentiated between ich, ischemic stroke, and controls. similarly, rna-seq of whole blood rna successfully differentiated between not only ich, ischemic stroke and controls, but also between different stroke subtypes ( ). the list of genetic mutations that can cause se is extensive with most genes are associated with infantile-onset or childhood epilepsy syndromes. only a minority are seen in adultonset status epilepticus ( ) . the patients in the former group usually have accompanying intellectual disability related to their epilepsy syndromes. however, the evidence supporting a genetic etiology in the latter group may be absent, posing a diagnostic challenge. the available options include gene panel sequencing, whole exome sequencing, or whole genome sequencing. sequencing a pre-selected panel of genes is more common, but with decreasing cost, exome and genome sequencing are being used with increasing frequency. bioinformatic filtering and genotype-phenotype correlation are the main challenges, particularly with the large number of genetic variants identified during whole exome or genome sequencing. the yield of sequencing studies depends on pretest probability that is determined by early age of onset, consanguinity, or affected siblings. as such, to date, only a few genes associated with adult-onset se have been identified, posing a practical limitation that predominantly limits next-generation sequencing to pediatric patients at present ( ) . as clinical tools for determination of putative functional significance and deleteriousness of variants identified through sequencing are refined, it is hoped that sequencing approaches find a home in the armamentarium of the clinicians treating refractory or recurrent se in the neurocritical care unit. translational genomics undoubtedly represents an important component to overall efforts to improve our understanding of the diseases we treat, and in principle should improve our ability to identify therapeutic approaches to improve outcomes and, in some cases, prevent disease altogether. given the inherent complexity and inaccessibility of the human brain and its tissues, combined with the relative infrequency of the conditions we treat at the overall population level, progress has been modest when compared to conditions such as hyperlipidemia ( ) , coronary artery disease ( ) , or atrial fibrillation ( ) . nevertheless, the observation-based, hypothesisfree experimental process inherent to translational genomics lends itself well to conditions such as stroke and tbi in which the search for the "master regulator" that governs response to injury has remained elusive despite carefully designed and executed hypothesis-driven studies. an important component to future translational genomic studies in neurocritical care is the pressing need for collaboration across centers with access to large, well-characterized patient populations. the success of the international stroke genetics consortium, and the track-tbi and center-tbi consortia in amassing large human populations with stroke and traumatic brain injury, respectively, is a proven model to accelerate the human genomic studies that serve as the basis for reverse genomic translational research ( , ( ) ( ) ( ) . similar efforts through the critical care eeg monitoring research consortium and other partners could lead to biorepositories of specific conditions relevant to neurocritical care that could provide sample sizes sufficient to drive unbiased genomic discoveries ( ) . close alliances with model systems researchers are another critical component to accelerating translational genomics in neurocritical care. as characterization of human disease through multimodal and continuous physiological monitoring, electrophysiology, medical imaging, and biomarker sampling continues to evolve, it is imperative that this information is shared and explored with allied model systems researchers to ensure that models are re-evaluated for their correlation with these endophenotypes, and potentially for dedicated exploration of how these human-derived phenotypes inform on the utility of specific model systems to investigate disease. finally, building relationships with biotechnology and pharmaceutical industry partners will be essential to efforts to extend therapeutic targets arising from translational genomic discoveries towards drug development ( ) . while repurposing existing drug compounds for new indications is an important consideration, small molecule and biologic targets are likely to require extensive research and development in the preclinical and clinical space, and industry partners are often optimized for these phases of the therapeutic development process ( , ) . relatedly, development of polygenic risk scores for assessment of risk, prognosis, or treatment response will also require commercial investment and infrastructure, as few academic environments exist that can manage cliacertified genotyping, quality control, and result reporting and interpretation for on-target and clinically relevant secondary results ( , ) . particularly in rarer or particularly challenging disease indications like those commonly encountered in neurocritical care populations, academic-industry partnerships are important to raise awareness of and interest in important clinical indications where investment could yield a large impact on a relatively small population of patients. translational genomics, in which genomic associations with risk, outcome, or treatment response are systematically identified and explored for functional relevance in humans or model systems of disease, is a valuable tool for identification of mechanisms, risk factors, therapeutic targets, and risk estimates in multiple diseases that are highly relevant to clinicians and scientists operating in the neurocritical care space. while there are undoubtedly challenges to studying some of the most complex diseases that affect the most complex organ in the body, translational genomic approaches may be uniquely suited to this task. coordinated investments in the collaborations, consortia, and infrastructures that enable these studies are likely to contribute to the novel treatments and biomarkers that are so sorely needed in the highly morbid and often poorly understood conditions in the patient populations we serve. translational genomics reengineering translational science: the time is right human genome project: twenty-five years of big biology it's time to reverse our thinking: the reverse translation research paradigm defining translational research: implications for training a systematic review of large animal models of combined traumatic brain injury and hemorrhagic shock sex differences in animal models of traumatic brain injury animal models of traumatic brain injury and assessment of injury severity differences in pathological changes between two rat models of severe traumatic brain injury neuroprotection in stroke: the importance of collaboration and reproducibility neuroprotection for ischemic stroke: two decades of success and failure stroke foundation of ontario centre of excellence in stroke r. toward wisdom from failure: lessons from neuroprotective stroke trials and new therapeutic directions very early administration of progesterone for acute traumatic brain injury embracing failure: what the phase iii progesterone studies can teach about tbi clinical trials mitochondria-wide association study of common variants in osteoporosis genetic association studies in cardiovascular diseases: do we have enough power? discovery of rare variants for complex phenotypes stroke genetics: discovery, biology, and clinical applications raredisease genetics in the era of next-generation sequencing: discovery to translation review: animal models of acquired epilepsy: insights into mechanisms of human epileptogenesis pathological mechanisms underlying aneurysmal subarachnoid haemorrhage and vasospasm tools for genomics a brief history of alzheimer's disease gene discovery association of apolipoprotein e with intracerebral hemorrhage risk by race/ethnicity: a meta-analysis apolipoprotein e polymorphism and outcomes from traumatic brain injury: a living systematic review and meta-analysis apoe genotype specific effects on the early neurodegenerative sequelae following chronic repeated mild traumatic brain injury evaluating historical candidate genes for schizophrenia current concepts and clinical applications of stroke genetics unbiased methods for population-based association studies robust inference of population structure for ancestry prediction and correction of stratification in the presence of relatedness variants at apoe influence risk of deep and lobar intracerebral hemorrhage international hapmap c. the international hapmap project advancements in next-generation sequencing opportunities and challenges of whole-genome and -exome sequencing whole-genome sequencing is more powerful than wholeexome sequencing for detecting exome variants statistical power and significance testing in large-scale genetic studies rare coding variants in angptl are associated with familial forms of intracranial aneurysm de novo variants in rhobtb , an atypical rho gtpase gene, cause epileptic encephalopathy multi-omic measurements of heterogeneity in hela cells across laboratories mouse regulatory dna landscapes reveal global principles of cis-regulatory evolution resuscitation with valproic acid alters inflammatory genes in a porcine model of combined traumatic brain injury and hemorrhagic shock single cell molecular alterations reveal target cells and pathways of concussive brain injury malignant' middle cerebral artery territory infarction: clinical course and prognostic signs trpm inhibition promotes angiogenesis after ischemic stroke sur -trpm cation channel expression in human cerebral infarcts malignant infarction of the middle cerebral artery in a porcine model. a pilot study newly expressed sur -regulated nc(ca-atp) channel mediates cerebral edema after ischemic stroke sulfonylurea receptor expression in human cerebral infarcts does inhibiting sur complement rt-pa in cerebral ischemia? glibenclamide is superior to decompressive craniectomy in a rat model of malignant stroke. stroke; a journal of cerebral circulation atp-dependent potassium channel blockade strengthens microglial neuroprotection after hypoxia-ischemia in rats glibenclamide pretreatment protects against chronic memory dysfunction and glial activation in rat cranial blast traumatic brain injury effects of oral glibenclamide on brain contusion volume and functional outcome of patients with moderate and severe traumatic brain injuries: a randomized double-blind placebo-controlled clinical trial effect of iv glyburide on adjudicated edema endpoints in the games-rp trial brain injury: the pathophysiology of the first hours cell death mechanisms and modulation in traumatic brain injury lateral fluid percussion brain injury: a -year review and evaluation temporal and spatial characterization of neuronal injury following lateral fluidpercussion brain injury in the rat experimental traumatic brain injury. experimental & translational stroke medicine animal models of head trauma a new model of diffuse brain injury in rats. part i: pathophysiology and biomechanics responses to cortical injury: i. methodology and local effects of contusions in the rat characterization of a new rat model of penetrating ballistic brain injury blast related neurotrauma: a review of cellular injury animal models of traumatic brain injury moderate controlled cortical contusion in pigs: effects on multiparametric neuromonitoring and clinical relevance progesterone in the treatment of acute traumatic brain injury: a clinical perspective and update the role of progesterone in traumatic brain injury. the journal of head trauma rehabilitation acute response of the hippocampal transcriptome following mild traumatic brain injury after controlled cortical impact in the rat analysis of post-traumatic brain injury gene expression signature reveals tubulins, nfe l , nfkb, cd , and s a as treatment targets time-dependent changes in microglia transcriptional networks following traumatic brain injury. frontiers in cellular neuroscience treatment of refractory status epilepticus with pentobarbital, propofol, or midazolam: a systematic review animal models of seizures and epilepsy: past, present, and future role for the discovery of antiseizure drugs experimental models of status epilepticus and neuronal injury for evaluation of therapeutic interventions loscher w. critical review of current animal models of seizures and epilepsy used in the discovery and development of new antiepileptic drugs animal models of status epilepticus and temporal lobe epilepsy: a narrative review kindling and status epilepticus models of epilepsy: rewiring the brain. progress in neurobiology status epilepticus: behavioral and electroencephalography seizure correlates in kainate experimental models differences in sensitivity to the convulsant pilocarpine in substrains and sublines of c bl/ mice. genes, brain, and behavior identification of micrornas with dysregulated expression in status epilepticus induced epileptogenesis dual and opposing roles of microrna- in epilepsy are mediated through inflammatory and nrsf-dependent gene networks genome-wide analysis of differential rna editing in epilepsy a comparison of pathophysiology in humans and rodent models of subarachnoid hemorrhage cerebral vasospasm after aneurysmal subarachnoid hemorrhage and traumatic brain injury. current treatment options in neurology a novel swine model of subarachnoid hemorrhage-induced cerebral vasospasm experimental subarachnoid hemorrhage: subarachnoid blood volume, mortality rate, neuronal death, cerebral blood flow, and perfusion pressure in three different rat models a murine model of subarachnoid hemorrhage learning deficits after experimental subarachnoid hemorrhage in rats a murine model of subarachnoid hemorrhage-induced cerebral vasospasm a modified double injection model of cisterna magna for the study of delayed cerebral vasospasm following subarachnoid hemorrhage in rats. experimental & translational stroke medicine a new percutaneous model of subarachnoid haemorrhage in rats pharmacological stabilization of intracranial aneurysms in mice: a feasibility study elastase-induced intracranial aneurysms in hypertensive mice a mouse model of intracranial aneurysm: technical considerations clazosentan, an endothelin receptor antagonist, in patients with aneurysmal subarachnoid haemorrhage undergoing surgical clipping: a randomised, doubleblind, placebo-controlled phase trial (conscious- ) randomized trial of clazosentan in patients with aneurysmal subarachnoid hemorrhage undergoing endovascular coiling. stroke; a journal of cerebral circulation cerebral vasospasm following subarachnoid hemorrhage: time for a new world of thought interleukin- receptor antagonist is beneficial after subarachnoid haemorrhage in rat by blocking haem-driven inflammatory pathology. disease models & proteomic expression changes in large cerebral arteries after experimental subarachnoid hemorrhage in rat are regulated by the mek-erk / pathway the support of human genetic evidence for approved drug indications genetic risk factors for ischaemic stroke and its subtypes (the metastroke collaboration): a meta-analysis of genome-wide association studies genome-wide association study identifies a variant in hdac associated with large vessel ischemic stroke deficiency of the stroke relevant hdac gene attenuates atherosclerosis in accord with allele-specific effects at p . sodium valproate, a histone deacetylase inhibitor, is associated with reduced stroke risk after previous ischemic stroke or transient ischemic attack identification of hdac as a viable therapeutic target for the treatment of gastric cancer genome-wide association study in african americans with acute respiratory distress syndrome identifies the selectin p ligand gene as a risk factor genome-wide association study of intracranial aneurysms confirms role of anril and sox in disease risk genome-wide association study of intracranial aneurysm identifies three new risk loci anril: a lncrna at the cdkn a/b locus with roles in cancer and metabolic disease effect of circular anril on the inflammatory response of vascular endothelial cells in a rat model of coronary atherosclerosis. cellular physiology and biochemistry : international journal of experimental cellular physiology, biochemistry, and pharmacology genome-wide association studies recommendations from the international stroke genetics consortium, part : standardized phenotypic data collection genes associated with multiple sclerosis: and counting. expert review of molecular diagnostics identification of new biomarkers for acute respiratory distress syndrome by expression-based genome-wide association study technical approaches for mouse models of human disease generating mouse models for biomedical research: technological advances. disease models & mechanisms potential pitfalls of crispr/cas -mediated genome editing the role of secretory granules in the transport of basement membrane components: radioautographic studies of rat parietal yolk sac employing h-proline as a precursor of type iv collagen monoclonal antibodies against chicken type iv and v collagens: electron microscopic mapping of the epitopes after rotary shadowing basement membrane (type iv) collagen is a heteropolymer col a mutations as a monogenic cause of cerebral small vessel disease: a systematic review common variation in col a /col a is associated with sporadic cerebral small vessel disease review: molecular genetics and pathology of hereditary small vessel diseases of the brain genome-wide association study of cerebral small vessel disease reveals established and novel loci molecular and genetic analyses of collagen type iv mutant mouse models of spontaneous intracerebral hemorrhage identify mechanisms for stroke prevention col a mutations impair col a and col a secretion and cause hemorrhagic stroke abnormal expression of collagen iv in lens activates unfolded protein response resulting in cataract col a mutation causes endoplasmic reticulum stress and genetically modifiable ocular dysgenesis. human molecular genetics plphp deficiency: clinical, genetic, biochemical, and mechanistic insights interfering with long chain noncoding rna anril expression reduces heart failure in rats with diabetes by inhibiting myocardial oxidative stress the interplay of lncrna anril and mir- b on the inflammation-relevant coronary artery disease through mediating nf-kappab signalling pathway compound libraries: recent advances and their applications in drug discovery. current drug discovery technologies multiancestry genome-wide association study of , subjects identifies loci associated with stroke and stroke subtypes atrial fibrillation genetic risk differentiates cardioembolic stroke from other stroke subtypes genome-wide polygenic scores for common diseases identify individuals with risk equivalent to monogenic mutations predicting polygenic risk of psychiatric disorders classification of patients with sepsis according to blood genomic endotype: a prospective cohort study. the lancet respiratory medicine a predictive clinicalgenetic model of tissue plasminogen activator response in acute ischemic stroke validation of a clinical-genetics score to predict hemorrhagic transformations after rtpa journal of stroke and cerebrovascular diseases : the official journal of national stroke association rna-seq identifies circulating mir- a- p, mir- b- p, and mir- - p as potential biomarkers for acute ischemic stroke the intracerebral hemorrhage blood transcriptome in humans differs from the ischemic stroke and vascular risk factor control blood transcriptomes intracerebral hemorrhage and ischemic stroke of different etiologies have distinct alternatively spliced mrna profiles in the blood: a pilot rna-seq study genetic mutations associated with status epilepticus association of genetic variants related to cetp inhibitors and statins with lipoprotein levels and cardiovascular risk pcsk : from basic science discoveries to clinical trials rare truncating variants in the sarcomeric protein titin associate with familial and early-onset atrial fibrillation recommendations from the international stroke genetics consortium, part : biological sample collection and storage investigators t-t. outcome prediction after mild and complicated mild traumatic brain injury: external validation of existing models and identification of new predictors using the track-tbi pilot study the center-tbi core study: the making-of comparison of machine learning models for seizure prediction in hospitalized patients developing interactions with industry in rare diseases: lessons learned and continuing challenges. genetics in medicine : official journal of the american college of medical genetics drug repurposing from the perspective of pharmaceutical companies insights into computational drug repurposing for neurodegenerative disease clinical providers' experiences with returning results from genomic sequencing: an interview study recommendations for reporting of secondary findings in clinical exome and genome sequencing, update (acmg sf v . ): a policy statement of the american college of medical genetics and genomics publisher's note springer nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations key: cord- -rb sr r authors: koutsomitropoulos, dimitrios a.; andriopoulos, andreas d. title: automated mesh indexing of biomedical literature using contextualized word representations date: - - journal: artificial intelligence applications and innovations doi: . / - - - - _ sha: doc_id: cord_uid: rb sr r appropriate indexing of resources is necessary for their efficient search, discovery and utilization. relying solely on manual effort is time-consuming, costly and error prone. on the other hand, the special nature, volume and broadness of biomedical literature pose barriers for automated methods. we argue that current word embedding algorithms can be efficiently used to support the task of biomedical text classification. both deep- and shallow network approaches are implemented and evaluated. large datasets of biomedical citations and full texts are harvested for their metadata and used for training and testing. the ontology representation of medical subject headings provides machine-readable labels and specifies the dimensionality of the problem space. these automated approaches are still far from entirely substituting human experts, yet they can be useful as a mechanism for validation and recommendation. dataset balancing, distributed processing and training parallelization in gpus, all play an important part regarding the effectiveness and performance of proposed methods. digital biomedical assets include a variety of information ranging from medical records to equipment measurements to clinical trials and research outcomes. the digitization and availability of biomedical literature is important at least in two aspects: first, this information is a valuable source for open education resources (oers) that can be used in distance training and e-learning scenarios; second, future research advancements can stem from the careful examination and synthesis of past results. for both these directions to take effect, it is critical to consider automatic classification and indexing as a means to enable efficient knowledge management and discovery for these assets. in addition, the sheer volume of biomedical literature is continuously increasing and puts excessive strain on manual cataloguing processes: for example, the us national library of medicine experiences daily a workload of approximately , articles for processing [ ] . research in the automatic indexing of literature is constantly advancing and various approaches are recently proposed, a fact that indicates this is still an open problem. these approaches include multi-label classification using machine learning techniques, training methods and models from large lexical corpora as well as semantic classification approaches using existing thematic vocabularies. to this end, the medical subject headings (mesh) is the de-facto standard for thematically annotating biomedical resources [ ] . in this paper we propose and evaluate an approach for automatically annotating biomedical articles with mesh terms. while such efforts have been investigated before, in this work we are interested in the performance of current state-of-the-art algorithms based on contextualized word representations or word embeddings. we suggest producing vectorized word and paragraph representations of articles based on context and existing thematic annotations (labels). consequently, we seek to infer the most similar terms stored by the model without the need and overhead of a separate classifier. moreover, we combine these algorithms with structured semantic representations in web ontology language format (owl), such as the implementation of the mesh thesaurus in owl simple knowledge organization systems (skos) [ ] . finally, we investigate the effect and feasibility of employing distributed data manipulation and file system techniques for dataset preprocessing and training. the rest of this paper is organized as follows: in sect. we summarize current word embedding approaches as the main background and identify the problem of automated indexing; in sect. we review relevant literature in the field of biomedical text classification; sect. presents our methodology and approach, by outlining the indexing procedure designed, describing the algorithms used and discussing optimizations regarding dataset balancing, distributed processing and training parallelization. section contains the results of the various experiments and their analysis, while sect. outlines our conclusions and future work. word embedding techniques [ ] convert words into word vectors. the following approaches have emerged in recent years with the performance of text recognition as the primary objective. the start has been made with the word vec algorithm [ ] where unique vector word representations are generated by means of shallow neural networks and the prediction method as well as by the explicit extension doc vec (document to vector) [ ] , where a unique vector representation can also be given for whole texts. next, the global vectors (glove) algorithm [ ] manages to transfer the words into a vector space by making use of the enumeration method. then, the fasttext algorithm [ ] achieves not only the management of a large bulk of data in optimal time but also better word embedding due to the use of syllables. in addition, the elmo algorithm [ ] uses deep neural networks, lstms, and a different vector representation for a word the meaning of which differentiates. lastly, the bert algorithm [ ] also generates different representations of a word according to its meaning, but instead of lstms it uses transformer elements [ ] . assigning a topic to text data is a demanding process. nevertheless, if approached correctly, it ensures easier and more accurate access for the end user. a typical subcase of the topic assignment problem is the attempt to create mesh indexes in repositories with biomedical publications. this particular task, which improves the time and the quality of information retrieved from the repositories, is usually undertaken by field experts. however, this manual approach is a time consuming process (it takes two to three months to incorporate new articles), but also a costly one (the cost for each article is approximately $ ) [ ] . a plethora of research approaches have attempted to tackle with the problem of indexing. to do this, they use word embeddings in combination with classifiers. typical cases are discussed below. mesh now [ ] classifies the candidate terms based on the relevance of the targetarticle and selects the one with the highest ranking, thus achieving a . f-score. to do this, researchers are using k-nn and support vector machine (svm) algorithms. another approach, named deepmesh [ ] , deals with two challenges, that is, it attempts to examine both the frequency characteristics of the mesh tags and the semantics of the references themselves (citations). for the first it proposes a deep semantic representation called d v-tfidf and for the latter a classification framework. a k-nn classifier is used to rate the candidate mesh headings. this system achieves an f-score of . . another study [ ] uses the word vec algorithm on all the abstracts of the pubmed repository, thereby generating a complete dictionary of , , unique words. the use of these vectors as a method to reduce dimensionality is examined by allowing greater scaling in hierarchical text classification algorithms, such as k-nn. by selecting a skip-gram neural network model (fasttext) vectors of size are generated with different windows from to , which the authors call mesh-gram with an f-score of . [ ] . moreover, taking into consideration the assumption that similar documents are classified under similar mesh terms, the cosine similarity metric and a representation of the thesaurus as a graph database, scientists proceed to an implementation with an f-score of . [ ] . a problem-solving approach is considered starting from converting texts into vectors with the use of elastic search and identifying the most similar texts with the help of the cosine similarity metric. then, by deriving the tags from these texts and calculating the frequency of occurrence in conjunction with similarity, an evaluation function is defined which classifies documents. biowordvec [ ] is an open set of biomedical word vectors/embeddings which combines subword information from unlabeled biomedical text with mesh. there are two steps in this method: first, constructing mesh term graph based on its rdf data and sampling the mesh term sequences and, second, employing the fasttext subword embedding model to learn the distributed word embeddings based on text sequences and mesh term sequences. in this way, the value of the f-score metric is improved to . and . for cnn and rnn models respectively. the proposed approach for the indexing of biomedical resources starts with assembling the datasets to be used for training. we then proceed by evaluating and reporting on two prominent embedding algorithms, namely doc vec and elmo. the models constructed with these algorithms, once trained, can be used to suggest thematic classification terms from the mesh vocabulary. in an earlier work we have shown how to glean together resources from various open repositories, including biomedical ones, in a federated manner. user query terms can be reverse-engineered to provide additional mesh recommendations, based on query expansion [ ] . finally, we can combine and assess these semantic recommendations by virtue of a trained embeddings model [ ] . each item's metadata is scraped for the title and abstract of the item (body of text). this body of text is next fed into the model and its vector similarity score is computed against the list of mesh terms available in the vocabulary. training datasets comprise biomedical literature from open access repositories including pubmed [ ], europepmc [ ] and clinicaltrials [ ] along with their handpicked mesh terms. for those terms that may not occur at all within the datasets, we fall back to their scopenote annotations within the mesh ontology. as a result, the model comes up with a set of suggestions together with their similarity score ( fig. ). for the application of the doc vec and elmo methods, a dataset from the pubmed repository with records of biomedical citations and abstracts was used. in december of every year, the core pubmed dataset integrates any updates that have occurred in the field. each day, the national library of medicine produces updated files that include new, revised and deleted citations. about m records, which are collected annually, can be accessed by researchers as of december . another source is europepmc, which is a european-based database that mirrors pubmed abstracts but also provides free access to full-texts and an additional m other relevant resources. each entry in the dataset contains information, such as the title and abstract of the article, and the journal which the article was published in. it also includes a list of subject headings that follow the mesh thesaurus. these headings are selected and inserted after manual reading of the publication by human indexers. indexers typically select - mesh terms to describe every indexed paper. mesh is a specialized solution for achieving a uniform and consistent indexing of biomedical literature. in addition, it has already been implemented in skos [ ] . it is a large and dense thesaurus consisting of , concepts. the clinicaltrials repository also provides medical data, that is, records with scientific research studies in xml. clinicaltrials contains a total of , records and each record is also mesh indexed. we include these records with the ones obtained from pubmed and europepmc for the purposes of variability and dataset diversity. our initial methodology for collecting data for training followed a serial approach [ ] . access to pubmed can be done easily via file transfer protocol (ftp). the baseline folder includes zip files, up to a certain date (december ). each file is managed individually with the information being serially extracted, thus rendering the completion of the entire effort a time costly process. in addition to delays, this problem makes the entire process susceptible to the need for constant internet connection to the repositories and any interruptions that may occur. algorithm : dataset preparation procedure input: xml files from repository output: two csv files step . for each file repository do step . connect to ftp server step . get file to local disk step . store file as line in rdd step . delete file from local disk step . end for step . parse file -useful information is extracted step . convert rdd to dataframe step . write useful information to csv files to solve this particular problem, we investigate the use of a distributed infrastructure at the initial stage of data collection. for this purpose, apache spark , a framework for parallel data management, was used. xml files are now stored as a whole in a dataframe. in detail, all the information in an xml file is read as a line and then converted into resilient distributed dataset (rdd). the useful information is then extracted as in the previous procedure. finally, the rdd is converted into a dataframe, from which the information is easily extracted, for example, into csv files. with this process, although extracting data from repositories is still not avoided, parsing can be now performed on a distributed infrastructure. an essential part of the dataset preparation process is to cover the whole thesaurus as thoroughly as possible. thus, apart from the full coverage of the thesaurus terms, the model must also learn each term with an adequate number of examples (term annotations) and this number should be similar between the terms. otherwise, there will be a bias towards specific terms that happen to have several training samples vs. others which might have only a few. therefore, to achieve a balanced dataset for training, when a term is incorporated into the set, its number of samples is restricted by an upper limit (algorithm ). if there are fewer samples the term is ignored; if there are more, exceeding annotations are cut off. the final dataset, as shown in table , does not fully cover the thesaurus, but the terms it contains are represented by an adequate and uniform number of samples. doc vec model. to create the model, the input is formed from the "one-hot" vectors of the fixed-size body of text, which is equal to the dictionary size. the hidden plane includes nodes with linear triggering functions and with the same size as the resulting vector dimensions. at the output level there will also be a vector equal to the dictionary size, while the activation function will be softmax. the training of the doc vec model, with the help of the gensim library , is performed by using the following parameters: train epochs , size vector , learning parameter . and min count . a variety of tests were performed to estimate these values. tests have shown that when there are only a few samples per term, a larger number of epochs can compensate for the sparsity of training samples. in addition, removing words with less than occurrences also creates better and faster vector representations for thesaurus terms. the created model is stored so that it can be called directly when needed. this model, with the adopted weights of the synapses that have emerged, is in fact nothing more than a dictionary. the content of this dictionary is the set of words used in the training along with their vector representations as well as a vector representation for each complete body of text. initially, all the vectors are extracted through a url connection . then, all the words related to the topic and mentioned in labels, are converted into classes, that is, numeric values, e.g. , , , etc., which in turn become "one-hot" vectors. to create the model, we have used the keras library . the input layer receives one body of text (title and abstract) at a time. the next layer is the lambda, which is supplied by the input layer, and uses the elmo embeddings, with an output count of . to create the elmo embeddings, the vectors derived from the url are used and the full body of text is converted into string format and compressed at the same time with the help of the tensorflow package. selecting the default parameter in this process ensures that the vector representation of the body of text will be the average of the word vectors. the next layer will be the dense one, which in turn is powered by the lambda layer and contains nodes and the relu activation function. finally, we have the output layer, which is also a dense layer with as many nodes as classes, and in this case the activation function will be softmax. in the final stage of the system (compile), the loss parameter is categorical_crossentropy, because of the categorization being attempted. the adam optimization, and the accuracy metric are chosen. the training of the elmo model is done by starting a tensorflow session with parameters: epochs and batch_size . this ensures that training will not take longer than epochs and the data will be transferred in packages of samples. upon completion, the finalized weights are stored, for the purpose of immediately creating a model either for evaluation or any other process required. in the context of the implementation of the above two models, the most significant stage is the execution of the algorithm responsible for their training. this process, depending on the architecture of the model, is particularly demanding on computing resources. most experiments are conducted on average commodity hardware (intel i , . ghz, -cores cpu with gb of ram). in a shallow neural network, such as doc vec, execution can be completed relatively quickly without significant processing power demands. our test configuration appears sufficient for at least epochs to be achieved, so that the model can be trained properly. however, this is not the case for deep neural networks, such as elmo. in these models, the multiple layers with many connections add complexity and increase computational power requirements in order to be adequately trained. therefore, finding a way to parallelize the whole process is considered necessary, not only for the optimization of the training time, but also for the completion of the entire effort. specifically, options associated with increasing the size of the training set in combination with the number of epochs can well lead either to a collapse of the algorithm execution or to prohibitive completion times. as an example, training elmo on a dataset of , samples ( labels with items each) took about min in the above hardware configuration. based on this concern, we have conducted experiments on infrastructures with a large number of graphics processing units (gpus). gpus owe their speed to high bandwidth and generally to hardware which enables them to perform calculations at a much higher rate than conventional cpus. we experimented on high-performance hardware with two intel xeon cpus including -cores, gb of ram and a nvidia v gpu. the v has gb of memory and , cores and supports cuda v. . , an api that allows the parallel use of gpus by machine learning algorithms. in this configuration, training with the , samples dataset took only s for the same model. this faster execution ( x) can help run experiments with larger datasets that would otherwise be prohibitive to achieve. to evaluate the doc vec approach, a total of , samples was used. this test set is balanced with the same procedure as the k dataset we have used previously for training. therefore, each mesh label occurring in the test set appears within the annotations of bibliographic items vs. in the training set. in both sets, a total of , distinct mesh labels, out of the available k, are considered. a body of text (title and abstract) is given as input to the model, which in turn generates a vector. the model then searches through the set of vectors, already incorporated from training, to find those that are close to the generated one. the process is quite difficult due to the plethora of labels in total, but also individually per sample as each one contains a finite non-constant number of labels. the threshold value reported is the similarity score above which suggestions are considered. at most suggestions are produced by the model. the following fig. plots precision (p) and recall (r) for various threshold values. for our purposes, a single information need is defined as a single-term match with an item ( - ) and we report the mean value of these metrics over all these matches, i.e. they are micro-averaged. precision takes into account how many of these - suggestions are correct out of the total suggestions made, while recall considers correct suggestions out of the total number of suggestions contained in the ground truth. other than the standard tradeoff between precision and recall, we first notice that the higher the threshold, the better the quality of results is for p. however, an increase in the threshold causes recall to drop. this makes sense, because there may be considerably fewer than suggestions for higher thresholds, i.e. only few of the terms pass the similarity threshold, thus leaving out some relevant terms. for middle threshold values, more that % of predictions are correct and also cover over % of the ground truth. similarly balanced test sets were also used to evaluate the elmo model, with a varying number of labels and samples per label. the results obtained are shown in fig. . initially, equally comparable or better results to previous related work are observed for precision and recall, when a small number of labels is selected. given that the dataset is balanced, these considerably improve when allowing x more samples for each label, thus allowing for better embeddings to be learned by the model. raising the number of labels to increases the dimensionality of the problem; now, each item must be classified among classes rather than simply . consequently, we notice a decrease in both metrics which, however, is ameliorated with the use of more samples ( , per label), as expected. nonetheless, further increasing the complexity of the dataset, by allowing , labels, causes absolute values to decrease considerably. even with the better performing hardware, any further efforts to improve the results in the case of , labels, notably by increasing the number of samples, do not succeed, as the execution of the algorithm is interrupted each time due to the need for additional memory. since our elmo-based model performs multi-class classification (but not multilabel), micro-averaged precision and recall would be identical . now, because our test set is balanced, both micro-and macro-averaged recall will also be equal (the denominator is the fixed number of samples per label). figure plots macro-averaged precision i.e. averages precision equally for each class/label. this macro-averaged precision will always be higher than the micro averaged one, thus explaining why p appears better than r, but only slightly, because of the absence of imbalance. p and r are virtually the same. the doc vec model is capable of making matching suggestions on its own, by employing similarity scores. a threshold value of . is where precision and recall acquire their maximum values. certainly, fewer than suggestions may pass this limit, but these are more accurate, apparently because of their higher score; they also occur more frequently within the terms suggested by experts in the ground truth: on average, each biomedical item in the test set hardly contains mesh terms, let alone (see table ). this observation serves as validation of the fact that the similarity measure produced by the model is highly relevant and correlated with the quality of suggestions. the elmo model, in turn, has allowed us to construct a multi-class classification pipeline that is built around the elmo embeddings. this manages to achieve very good results when dimensionality is kept low, i.e. a relatively small number of labels or classes is selected. it also seems to outperform the doc vec approach, even though classes are fewer and the recommendation problem is reduced to multi-class classification. however, doc vec surpasses the elmo classification pipeline when there is a need to choose labels from a broader space. one the other hand, the threshold existence favors doc vec, in the sense that fewer ground truth annotations pass its mark and are, therefore, considered when computing retrieval metrics. further attempts for improvement where not possible for elmo, a fact that confirms it is a very computationally expensive module compared to word embedding modules that only perform embedding lookups. contextualized word representations have revolutionized the way traditional nlp used to operate and perform. neural networks and deep learning techniques combined with evolving hardware configurations can offer efficient solutions to text processing tasks that would be otherwise impossible to perform on a large scale. we have shown that word embeddings can be a critical part and deserve careful consideration when approaching the problem of automated text indexing. especially in the biomedical domain, the shifting nature of research trends and the complexity of authoritative controlled vocabularies still pose challenges for fully automated classification of biomedical literature. to this end, we have investigated both deep-and shallow learning approaches and attempted comparison in terms of performance. dimensionality reduction, either implied by setting a threshold or directly posing a hard cap over available classification choices, appears necessary for automated recommendations to be feasible and of any practical value. still, a careful dataset balancing as well as the capability of deep networks to leverage distributed gpu architectures are demonstrated beneficial and should be exercised whenever possible. as a next step, we intend to further evaluate our elmo implementation and design a model that would perform multi-label classification using the elmo embeddings. in addition, we are planning to release a web-based service offering access to the recommendations provided by the two models. this would facilitate interoperability, for example, with learning management systems and other repositories. finally, we see improvements in the way classification suggestions are being offered, especially in view of the density of the thesaurus used: other ontological relations, such as generalization or specialization of concepts can be taken into account in order to discover and prune hierarchy trees appearing in the recommendations list. the mesh-gram neural network model: extending word embedding vectors with mesh concepts for umls semantic similarity and relatedness in the biomedical domain bert: pre-training of deep bidirectional transformers for language understanding metadata of all full-text europe pmc articles bag of tricks for efficient text classification biomedical semantic indexing using dense word vectors subject classification of learning resources using word embeddings and semantic thesauri semantic annotation and harvesting of federated scholarly data using ontologies distributed representations of sentences and documents word embedding for understanding natural language: a survey mesh now: automatic mesh indexing at pubmed scale via learning to rank efficient estimation of word representations in vector space the nlm medical text indexer system for indexing biomedical literature deepmesh: deep semantic representation for improving large-scale mesh indexing glove: global vectors for word representation deep contextualized word representations search and graph database technologies for biomedical semantic indexing: experimental analysis a method to convert thesauri to skos attention is all you need biowordvec, improving biomedical word embeddings with subword information and mesh. sci. data key: cord- - ndb rm authors: iwasa, yoh; sato, kazunori; takeuchi, yasuhiro title: mathematical studies of dynamics and evolution of infectious diseases date: journal: mathematics for life science and medicine doi: . / - - - - _ sha: doc_id: cord_uid: ndb rm nan the practical importance of understanding the dynamics and evolution of infectious diseases is steadily increasing in the contemporary world. one of the most important mortality factors for the human population is malaria. every year, hundreds of millions of people suffer from malaria, and more than a million children die. one of the obstacles of controlling malaria is the emergence of drug-resistant strains. pathogen strains resistant to antibiotics pose an important threat in developing countries. in addition, we observe new infectious diseases, such as hiv, ebora, and sars. the mathematical study of infectious disease dynamics has a long history. the classic work by kermack and mckendrick ( ) established the basis of modeling infectious disease dynamics. the variables indicate the numbers of host individuals in several different states -susceptive, infective and removed. this formalism is the basis of all current modeling of the dynamics and evolution of infectious diseases. since then, the number of theoretical papers on infectious diseases has increased steadily. especially influential was a series of papers by roy anderson and robert may, summarized in their book (anderson and may ) . anderson and may have developed population dynamic models of the host engaged in reproduction and migration. in a sense, they treated epidemic dynamics as a variant of ecological population dynamics of multiple species community. combining the increase of our knowledge of nonlinear dynamical systems (e. g. chaos), anderson and may also demonstrated the usefulness of simple models in understanding the basic principles of the system, and sometimes even in choosing a proper policy of infectious disease control. the dynamical systems for epidemics are characterized by nonlinearity. the systems include many processes at very different scales, from the population on earth to the individual level, and further to the immune system within a patient. hence, mathematical studies of epidemics need to face this dynamical diversity of phenomena. tools of modeling and analysis for situations including time delay and spatial heterogeneity are very important. as a consequence, there is no universal mathematical model that holds for all problems in epidemics. when we are given a set of epidemiological phenomena and questions to answer, we must "construct" mathematical models that can describe the phenomena and answer our questions. this is quite different from studies in "pure" mathematics, in which usually the models are given beforehand. one of the most important questions in mathematical studies of epidemics is the possibility of the eradication of disease. the standard local stability analysis of the endemic equilibrium and disease-free equilibrium is often not enough to answer the question, because it gives us information only on the local behavior, or the solution in the neighborhood of those equilibria. on the other hand, it is known that global stability analysis of the models is often very difficult, and even impossible in general cases, because the dynamics are highly nonlinear. even if the endemic equilibrium were unstable and the disease-free equilibrium were locally stable, the diseases can remain endemic and be sustained forever. sometimes, rather simple models show periodic or chaotic behavior. recently, the concept of "permanence" was introduced in population biology and has been studied extensively. this concept is very important in mathematical epidemiology as well. permanence implies that the disease will be maintained globally, irrespective of the initial composition. even if the endemic equilibrium were unstable, the disease will last forever, possibly with perpetual oscillation or chaotic fluctuation. since the epidemiological data supplied by medical and public health sectors are abundant, epidemiological models are in general much better tested than similar population models in ecology developed for wild animals and plants. the diversity of models is also extensive, including all the different levels of complexity. rather simple and abstract models are suitable to discuss general properties of the system, while more complex and realistic computerbased simulators are adopted for policy decision making incorporating details of the structure closely corresponding to available data. mathematical modeling of infectious diseases is the most advanced subfield of theoretical studies in biology and the life sciences. what is notable in this development is that, even if many computer-based detailed simulators become available, the rigorous mathematical analysis of simple models remains very useful, medically and biologically, in giving a clear understanding of the behavior of the system. recently, the evolutionary change of infectious agents in the host population or within a patient has attracted an increasing attention. mutations during genome replication would create pathogens that may differ slightly from the original types. this gives an opportunity for a novel strain to emerge and spread. as noted before, emergence of resistant strains is a major obstacle of infectious disease control. essentially the same evolutionary process occurs within the body of a single patient. a famous example is hiv, in which viral particles change and diversify their nucleotide sequences after they infect a patient. this supposedly reflects the selection by the immune system of the host working on the virus genome. a similar process of escape is involved in carcinogenesis -a process in which normal stem cells of the host become cancerous. the papers included in this volume are for mathematical studies of models on infectious diseases and cancer. most of them are based on presentations in the first international symposium on dynamical systems theory and its applications to biology and environmental sciences, held in hamamatsu, japan, on - march . this introductory chapter is followed by four papers on infectious disease dynamics, in which the roles of time delay (chaps. and ) and spatial structures (chaps. and ) are explored. then, there are two chapters that discuss competition between strains and evolution occurring in the host population (chap. ) and within a single patient (chap. ). finally, there are papers on models of the immune system and cancer (chaps. and ). below, we briefly summarize the contents of each chapter. in chap. , zhien ma and jianquan li give an introduction to the mathematical modeling of disease dynamics. then, they summarize a project of modeling the spread of sars in china by the authors and their colleagues. in chap. , yasuhiro takeuchi and wanbiao ma introduce mathematical studies of models with time delay. they first review past mathematical studies on this theme during the last few decades, and then introduce their own work on the stability of the equilibrium and the permanence of epidemiological dynamics. in chaps and , wendi wang and shigui ruan discuss the spatial aspect of epidemiology. the spread of a disease in a population previously not infected may appear as "wave of advance". this is often modeled as a reaction diffusion system, or by other models handling spatial aspects of population dynamics. the speed of disease propagation is analogous to the spread of invaders in a novel habitat in spatial ecology (shigesada and kawasaki ) . since microbes have a shorter generation time and huge numbers of individuals, they have much faster evolutionary changes, causing drug resistance and immune escape, among the most common problems in epidemiology. by considering the appearance of novel strains with different properties from those of the resident population of pathogens, and tracing their abundance, we can discuss the evolutionary dynamics of infectious diseases. in chap. , horst thieme summarized the work on the competition between different and competing strains, and the possibility of their coexistence and replacement. an important concept is the "maximal basic replacement ratio". if a host once infected and then recovered from a single strain is perfectly immune to all the other strains (i. e. cross immunity is perfect), then the one with the largest basic replacement ratio will win the competition among the strains. the author explores the extent to which this result can be generalized. he also discusses the coexistence of strains considering the aspect of maternal transmission as well. in chap. , yoh iwasa and his colleagues analyze the result of evolutionary change occurring within the body of a single patient. some of the pathogens, especially rna viruses have high mutation rates, due to an unreliable replication mechanism, and hence show rapid genetic change in a host. the nucleotide sequences just after infection by hiv will be quite different from those hiv occurring after several years. by mutation and natural selection under the control of the immune system, they become diversified and constantly evolve. iwasa and his colleagues derive a result that, without cross-immunity among strains, the pathogenicity of the disease tends to increase by any evolutionary changes. they explore several different forms of cross-immunity for which the result still seems to hold. in chap. , edoardo beretta and his colleagues discuss immune response based on mathematical models including time delay. the immune system has evolved to cope with infectious diseases and cancers. they have properties of immune memory and, once attached and recovered, they will no longer be susceptive to infection by the same strain. to achieve this, the body has a complicated network of diverse immune cells. beretta and his colleagues summarize their study of modeling of an immune system dynamics in which time delay is incorporated. in the last chapter, h.i. freedman studies cancer, which originates from the self-cells of the patient, but which then become hostile by mutations. there is much in common between cancer cells and pathogens originated from outside of the host body. freedman discusses the optimal chemotherapy, considering the cost and benefit of chemotherapy. this collection of papers gives an overview of theoretical studies of infectious disease dynamics and evolution, and hopefully will serve as a source in future studies of different aspects of infectious disease dynamics. here, the key words are time delay, spatial dynamics, and evolution. toward the end of this introductory chapter, we would like to note one limitation -all of the papers in this volume discuss deterministic models, which are accurate when the population size is very large. since the number of microparasites, such as bacteria, or viruses, or cancer cells, is often very large, the neglect of stochasticity due to the finiteness of individuals seems to be acceptable. however, when we consider the speed of the appearance of novel mutants, we do need stochastic models, because mutants always start from a small number. according to studies on the timing of cancer initiation, which starts from rare mutations followed by population growth of cancer cells, the predictions of deterministic models differ by several orders of magnitude from those of stochastic models and direct computer simulations. infectious diseases of humans a contribution to the mathematical theory of epidemics biological invasions: theory and practice key: cord- - l d ew authors: lv, yang; hu, guangyao; wang, chunyang; yuan, wenjie; wei, shanshan; gao, jiaoqi; wang, boyuan; song, fangchao title: actual measurement, hygrothermal response experiment and growth prediction analysis of microbial contamination of central air conditioning system in dalian, china date: - - journal: sci rep doi: . /srep sha: doc_id: cord_uid: l d ew the microbial contamination of central air conditioning system is one of the important factors that affect the indoor air quality. actual measurement and analysis were carried out on microbial contamination in central air conditioning system at a venue in dalian, china. illumina miseq method was used and three fungal samples of two units were analysed by high throughput sequencing. results showed that the predominant fungus in air conditioning unit a and b were candida spp. and cladosporium spp., and two fungus were further used in the hygrothermal response experiment. based on the data of cladosporium in hygrothermal response experiment, this paper used the logistic equation and the gompertz equation to fit the growth predictive model of cladosporium genera in different temperature and relative humidity conditions, and the square root model was fitted based on the two environmental factors. in addition, the models were carried on the analysis to verify the accuracy and feasibility of the established model equation. with the large-scale use of central air conditioning system and the improvement of people's living standard, more and more attention has been paid to the increasingly serious problem of indoor air pollution. studies showed that air handing unit is an important source of microorganisms for indoor biological pollution, and some microorganisms tend to stay in the dust of air conditioning units with the appropriate temperature and humidity environment. the microorganisms grow and then enter the indoor space through the air, resulting in the destruction of indoor air quality [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] . national institute for occupational safety and health (niosh) conducted a study of buildings, and the research results showed that among the sources of indoor air pollution, central air-conditioning system pollution accounted for % . summary of measurement by professor fanger showed that air-conditioning systems accounted for % in the indoor pollution sources . based on the supervision and inspection of air conditioning and ventilation system in the public places of china's provinces, municipalities and autonomous regions by china academy of building research and the chinese center for disease control and prevention, it was found that more than % of the central air conditioning systems could not meet china's national sanitary standard . thus, air conditioning system should eliminate negative impact caused by its own, on this basis, it may relate to positive effect of the ventilation. in recent years, h n , sars and other popular virus spread [ ] [ ] , and some researches showed that the hygienic, reasonable air conditioning systems were important to reduce damage [ ] [ ] . therefore, microbial pollution in central air conditioning system has become a critical topic in the field of indoor air pollution. studies showed that the filter, cooling coil, humidifier, and cooling tower in central air-conditioning system were easy to microbial breeding [ ] [ ] [ ] [ ] [ ] [ ] . in this study, a venue in dalian was selected as the research object. as the working condition of the air conditioning system was down, the environment parameters were measured, and microorganisms existing on the wind pipe, filtering net, surface cooler, and condensate water, on the floor and in the air were collected. besides, according to the tested microbial density and the identified genome sequence of collected microorganisms, the hygrothermal response experiment of dominant fungal was detected, and the fitting analysis was carried out based on the prediction model, followed by a series of statistical analysis. the aim of the present study was to clarify characteristics of the microorganisms in air conditioning systems, and the study would be helpful for policymakers and hvac engineers to develop the appropriate strategies and to conduct the bacteria and fungi characteristic assessments in hvac system. preliminary survey. the object of study is a venue in dalian, which covers a total area of m and building area is m . the aboveground part includes a swimming pool, ball training venues, gymnasium, the lounge room and the clinic. the underground part consists of a table tennis hall, air conditioning equipment rooms and reservoir area. the whole building is centralized-controlled by the central air conditioning room, which includes two air handling units. two measured units were all air system, which only had a coarse efficiency filter, and the unit is also provided with a heater, cooler and fan etc. both units are the primary return air system and the filters are removable types. the running time of the air conditioning system is from may to october, and the daily operation period is : - : . all components are cleaned every two years. when the measurement was carried on, the unit a and b were both cleaned a year ago. both units were closed during the sample collection. measurement method. the actual measurement is divided into two parts: the environment parameter measurement and air unit sampling. first, the temperature, humidity, and co concentration were automatically recorded by the temperature, humidity and co monitor (mch- sd, japan). second, the disinfected planktonic microorganism sampler (hkm series dw- , china) was installed where fresh air and return air mixed. once installing the sampler, we loaded medium in the sampler and set the parameter of air flow in sampler as l loaded medium. after the sample collection, the petri dishes must be sealed for preservation. finally, according to the hygienic specification of central air conditioning ventilation system in public buildings of china , we sampled the dust by using sterile non-woven ( mm* mm) on components of unit a and b, respectively, and each sampling point covered a cm* cm area at the sampling area. the non-woven fabrics were put into sterile water and stirred to make sure that organic substance on the non-woven was fully dissolved in sterile water. then, the samples of sterile water containing organic substances were prepared in times and times diluted concentration, respectively. there are sampling points in unit a and points in b, and two measuring point positions of the units are shown in fig. . the microorganisms collected in the air were directly cultured, and the samples in the dust were times and times diluted and μ l of the sample was inoculated into the two kinds of solid culture media. beef extract peptone medium was used for cultivating bacteria and potato dextrose agar was used for cultivating fungus , . each dilution was done in parallel samples to reduce the error, and final results showed the average value. the blank samples and test samples were set up for each of the cultures. if there is no colony detected on the blank sample, the test results are valid. both field test and laboratory measurements were performed in accordance with the hygienic specification of central air conditioning ventilation system in public buildings requirements . genome sequencing. only a small part of microorganisms are cultivable. therefore, the traditional cultivation method can not harvest all the species in ecological samples . fungal genome sequencing is an emerging method to identify the microbial genome, which could directly indicate related species information from environment samples . fungal amplicon sequencing analysis was used in this study, because the existing research showed that fungal spores have stronger vitality than other microorganisms in the air, and fungi dominated the microorganism in air conditioning systems. therefore, this method was mainly used to identify fungi in this study , - . environment parameters in air handling units. temperature, humidity and co concentration of unit a and b are shown in table . unit a is located in the ground floor (b ), and the unit b is located on the ground floor. compared to the unit b, the humidity of unit a is higher, and the temperature is lower. microbial colony analysis. the distribution density of bacteria and fungi in the unit a is obtained through statistics, as shown in fig. . the concentration of airborne fungus was cfu/m , and the concentration of airborne bacteria was cfu/m . the unit a showed the obvious microbial contamination status, though all components and airborne microorganism meet the hygienic specification of central air conditioning ventilation system in public buildings of china . the microbial distribution in filter net is central < edge < bottom and bacteria accounted for a larger proportion; the microbial distribution in surface cooler is center > against the wall > edge, and fungi accounted for a large. the fungal contamination in the air is more serious than the bacteria. the distribution density of bacteria and fungi in the unit b were obtained through statistics, as shown in fig. . the concentration of airborne fungus was cfu/m , and the concentration of airborne bacteria was cfu/m . parts of the measuring point in the unit b were polluted seriously. the bacterial colonies in the corner and the ground of the surface cooler were beyond the hygienic index (≤ cfu/cm ) in the hygienic specification of central air conditioning ventilation system in public buildings of china regulates . limited by unit placement, there were less measuring points in unit b, and we chose the same measuring points in both units for comparison (centre of surface cooler, surface cooler against wall, corner of surface, and ground of surface cooler). the comparison between unit a and b indicates that the bacterial density in unit a was less than that in the same sampling point in unit b, but the fungal density in unit a was more than that in the same sampling point in unit b. if the cleaning and disinfection is not enough before the air conditioning system running, it may make the fungus to enter the indoor environment, which results in make the pollution of indoor air. compared with cooling coil, the fungus contamination is worse in the floor dust and the air suspension. during the actual measurement, it is found that the unit internal is unprecedentedly narrow and low intensity of illumination in a closed state. according to the description by technicians, it is easy to trample damage to the underground pipes, which leads to the disinfection and cleaning work rarely in the unit. fungal genome sequencing analysis. in this study, we analysed the samples from the sampling points a , b , and b by amplicon sequencing information analysis, respectively named a a, b a, and b a. all collected samples in the air conditioner were transferred to the eppendorf tubes and processed with the extraction step. samples were resusponded in tens buffer with sds and proteinase k as described by vinod . after incubation at °c, phenol/chloroform/isoamyl alcohol was added to remove proteins, and the nucleic acid was precipitated with isopropanol and sodium acetate ( . m). total dna was dissolved in × te after washing with % ethanol. and then the quality and quantity tests were conducted by agarose gel electrophoresis, . for pcr product, the jagged ends of dna fragment would be converted into blunt ends by using t dna polymerase, klenow fragment and t polynucleotide kinase. then add an ' a' base to each ' end to make it easier to add adapters. after all that, fragments too short would be removed by ampure beads. for genomics dna, we use fusion primer with dual index and adapters for pcr, fragments too short would be removed by ampure beads too. in both cases, only the qualified library can be used for sequencing. the quality and quantity of libraries were assessed using the bioanaylzer (agilent technologies) and the steponeplus real-time pcr system (applied biosystems). the raw data generated by miseq and hiseq sequencers was processed to eliminate the adapter pollution and low quality to obtain clean reads. the qualified clean data was used for the further bioinformatics analysis. firstly, paired-end reads with overlap were merged to tags by software flash (v . . ) , and then tags were clustered to otu at % sequence similarity using usearch (v . . ) . secondly, taxonomic ranks were assigned to otu representative sequence using ribosomal database project (rdp) na, e bayesian classifier v. . . at last, alpha diversity, beta diversity and the different species screening were analyzed based on otu and taxonomic ranks by mothur (v . . ) . in order to fully understand the community structure of fungal sample and analyse fungus microbial diversity, while excluding errors that human operation brings, genome sequencing method in fields of molecular biology was employed in this study to obtain micro biological information. illumina company developed miseq method with higher flux and simple operation and lower cost for genome sequencing. besides, the synthesis of side edge sequencing method with higher reliability is more suitable for laboratory community structure. the high-throughput sequencing was found to be useful to characterize compositions and diversities of moulds. the gene sequence of the test samples from genome sequencing was dealed with, such as stitching and matching, and the sample had a total of high quality fungal sequences, with an average length of bp. the optimized sequence and the average length of the sample are shown in table . otu and abundance analysis. stitching and optimising the tags, in order to be the otu (operational taxonomic units) for species classification, gathering in the % similarity, and the statistics of each sample in the abundance of information in each otu were done , [ ] [ ] . rank otu curve is a form of species diversity in the sample, which can explain two aspects of sample diversity, that is, the richness and evenness of species in the sample. the richness of species in the samples represented by the horizontal length of the curve is wide, so that the sample species is more abundant. the uniformity of species in the samples from the curve reflects the longitudinal axis of the shape. that the curve is flat means that the sample has a higher composition of the species evenness. from fig. , the species composition of b a is the most abundant, and the uniformity is the highest. sample diversity analysis of observed species. alpha diversity is the analysis of species diversity in a single sample , including the species observed index, shannon index, etc. the greater the two indices are, the more abundant species is in the sample. the species observed index reflects the richness of the community in the sample, which also refers to the number of species in the community, without taking into account the abundance of each species in the community. shannon index reflects the diversity of the community, and the species richness and evenness of species in the sample community. in the case of the same species richness, the greater the evenness of the species is in the community, the greater the diversity of the community is. observed species exponential dilution curve. random sample in the processed sequence, draw the sequence number for the abscissa and the corresponding species number can be detected as the ordinate, so as to form a curve production dilution curve, shown in fig. (a) . with the increase of the sample quantity, the number of species increase and gradually become stabilized. when the curve reaches the plateau, it can be considered that the depth of the sequencing has been basically covered all the species in the sample. at the same time, the observed species index can reflect the richness of community in the sample, that is, the number of species in the community. it can be seen that the distribution of fungal species richness is b a > b a > a a. shannon exponential dilution curve. shannon index is affected not only by species richness in the sample community, but also by the evenness of species. in the case of the same species richness, the greater the evenness of the species is in the community, the more abundant diversity of the community is. it can be seen in the fig. (b) that the fungal species diversity of the unit b is significantly more complex than the unit a, and the similarity of species diversity in two sampling points of unit b was very high. composition of microbial samples. figure illustrates the species composition proportion of the three sampling points, and the proportion was redrew by removing the strains which were not detected in the sample. the results are shown in table . the species with the largest proportion is the dominant fungi. according to the fungal genome sequencing analysis results, fungal components in different units at the same sampling were different, and that in the same unit at different sampling points were roughly similar. they were caused by the different environmental conditions. on the center of air cooling coil in unit a, candida accounted for %; on the center and against the wall of the air cooling coil in unit b, cladosporium accounted for %, accompanied by alternaria, emericella and other fungus. cladosporium is usually rich in outdoor air, but they will also grow on the indoor surfaces when the humidity is high. existing research shows that the cladosporium spore is an extremely important allergen in the airborne transmission, which could cause asthma attacks or similar respiratory disease in patients with allergic reactions . some species of candida is a kind of conditional pathogenic fungi in the human body. growth prediction analysis of models. traditional microbial detection generally have the characteristics of hysteresis, which cannot play the role of prediction, but the use of mathematical models to predict the growth of microorganisms can timely and effectively predict the growth of microorganisms. therefore, it is very important to study the growth prediction model of the fungi in the air conditioning system. according to environmental conditions mentioned before, we established growth kinetics prediction model of cladosporium spp. to predict the rapid fungal growth in the experimental conditions, which can provide a theoretical basis for air microbial contamination prediction system and help evaluate the health risk inside buildings. the models were fitted by origin software (version ) and matlab r a, and the fitting conditions of logistic model and gompertz model were compared under different temperature and humidity conditions. the fitting effect between these two models and the fitting results of the two models were compared, and the corresponding model parameters were obtained. in addition, the square root model was fitted based on the two environmental factors. experimental study on the hygrothermal response of fungus. laboratory studies have revealed that fungal growth and reproduction are affected by water, temperature, nutrients, micro-elements, ph, light, carbon dioxide, and oxygen tension .the most relevant determinants of fungal proliferation in the building context are water/moisture and temperature, and to a certain extent those factors affect other environmental factors such as substrate ph, osmolarity, nutrient, material properties etc , .in order to lay the foundation for the fitting model, and to study the growth characteristics of fungi in different temperature and relative humidity, we set an experimental study on the hygrothermal response of fungus. from the results of fungal genome sequencing and literature research - , we selected cladosporium spp. and penicillium spp. as the research objects which are both common in air conditioning systems.this paper mainly studied the status of microbial contamination in air handling units so that the air temperature of each part of the air handling unit should be considered. the temperature gradient of °c − °c − °c and relative humidity gradient of %− %− %− % were selected as experimental hygrothermal conditions. the results of hygrothermal experiments are shown in figs , , . it can be known that growth rate of cladosporium spp. is faster than that of penicillium spp., in any experimental conditions, which is the essential characteristics of a strain, is hygrothermal response control method cannot change. these data indicated that low rh environments can reduce or even inhibit fungal growth. this observation agrees with findings by w. tang and pasanen , . growth prediction analysis based on logistic model. logistic model is a typical model of biological ecology, which has been widely used in the field of biological ecology . according to the actual research, the following formula equation ( ) was obtained after the appropriate change of the logistic equation. n was the colony growth diameter, cm; t was the microbial growth culture time, h; a , a , x , p as the model parameters. it can be seen from the table , the fitting curve of logistic model is similar to the experimental results. at °c and °c temperature conditions, the model's fitting effect is excellent, and r is greater than . ; at °c temperature conditions, the model fitting effect is not as good as other temperature conditions. predicting the growth of microorganisms. the pmp program developed by the ministry of agriculture to establish the model of pathogenic bacteria is the basic model for the study of gompertz equation. gompertz model has been widely used in the field of biology. gompertz model expression was as equation ( ): c n was the colony growth diameter, cm; t was the microbial growth culture time, h; a, a, k, x c as the model parameters. it can be seen from the table that the fitting curve of gompertz model is better fitted to the measured parameters. at the same temperature, with the increase of relative humidity, gompertz model fitting effect is better; the model is well fitted at the temperature of °c and the fitting effect is better than °c and °c temperature conditions. the fitting of logistic model to the growth of the fungus is better than that of the gompertz model. the two models are tested by the deviation factor b f and the accuracy factor a f in the mathematical model test. staphylococcus xylosus were studied by mcmeekin . they found that when t min is fixed, for each ϕ , the relationship between growth rate and temperature can be described by using the square root model. the combined effects of these two variables can be expressed by the modified equation ( ): in the formula, u is the growth rate of fungus, cm/h; b is the coefficient; t is the culture temperature, °c; t min is the most important parameter of square root equation, and it refers to the lowest temperature when the growth rate is zero, °c; ϕ is relative humidity of the cultivation, %. by using the logistic primary model, the predictive value of the growth rate of the cladosporium colony growth rate (instantaneous velocity) was obtained, as table shows. through the model fitting, the parameters of the square root model could be obtained, as table shows, and the model fitting of predicting growth of cladosporium was shown as fig. . the model equation of b f value was between . - . , indicating that the model used to predict the range of the experimental environment in cladosporium colony growth condition. at the same time, the a f value of the model was . that is closed to , which shows that the model has high accuracy. table . model fitting and model parameters of double factor square root. this study selected two central air conditioning systems at a venue in dalian as the objects. actual measurement and a series of studies were carried out on microbial pollution characteristic, and the results are shown as below: ( ) the bacterial colony forming units of the two measuring points in unit b were cfu/cm and cfu/cm , respectively, which exceeded the hygienic specification of central air conditioning ventilation system in public buildings of china (≤ cfu/cm ), and the rest of the test points met the relevant standards of china. the distribution of bacteria was more than fungi, and the concentration was higher. with the total characteristics of different distribution density, the area of dust associated microorganisms and the air pollution were more serious. ( ) alternaria spp., candida spp., cercospora spp. and cladosporium spp. existed in both units. the candida spp. accounted for % in unit a, and the cladosporium spp. occupied % in unit b. the composition of fungi in b was more complicated. two dominant fungi are both deleterious to health, so the timely maintenance and cleaning are required. it is suggested that the operating space should be reserved in the air conditioning room, so as to avoid incomplete cleaning and disinfection. ( ) within the experimental temperature and relative humidity, with the increase of relative humidity or temperature, the colony growth of the same strain showed an increasing trend. for the prediction model of the fungus growth, the study found that the overall fitting effect of logistic model is better, and r values were greater than . . logistic model for the cladosporium spp. growth was better than gompertz model. at the same time, considering the influence of temperature and relative humidity, the square root model can well predict the growth of cladosporium spp. it provides a theoretical basis for the growth of fungi in the air conditioning system under the hygrothermal environment conditions. why, when and how do hvac-systems pollute the indoor environment and what to do about it? the european airless project the control technology of microbial contamination in air conditioning system overview of biological pollution prevention methods in air conditioning system. heating ventilating and air conditioning relationships between air conditioning, airborne microorganisms and health. bulletin de l'academie nationale de medecine a review of the biological pollution characteristics, standards and prevention and control technologies of central air-conditioning system risks of unsatisfactory airborne bacteria level in air-conditioned offices of subtropical climates development of a method for bacteria and virus recovery from heating, ventilation, and air conditioning (hvac) filters microbial air quality at szczawnica sanatorium the national institute for occupational safety and health indoor environmental evaluation experience air pollution sources in offices and assembly halls quantified by the olf unit control and improvement of building indoor biological pollution a major outbreak of severe acute respiratory syndrome in hong kong epidemiological investigation of an outbreak of pandemic influenza a (h n ) in a boarding school: serological analysis of cases a study on the effective removal method of microbial contaminants in building according to bioviolence agents performance analysis of a direct expansion air dehumidification system combined with membrane-based total heat recovery hvac systems as emission sources affecting indoor air quality: a critical review testing and analysis for microbes and particles in central air conditioning systems of public buildings. heating ventilating and air conditioning endotoxins and bacteria in the humidifier water of air conditioning systems for office rooms investigation and review of microbial pollution in air conditioning systems of public buildings. heating ventilating and air conditioning study on the microbial secondary pollution and control in air conditioning system research on the dust deposition in ventilation and air conditioning pipe. heating ventilating and air conditioning hygienic specification of central air conditioning ventilation system in public buildings indoor air part : detection and enumeration of moulds-culture-based method macro genomics approach in environmental microbial ecology and genetic search application study of the microbiome in waiting rooms of a japanese hospital fungal colonization of automobile air conditioning systems changes in airborne fungi from the outdoors to indoor air; large hvac systems in nonproblem buildings in two different climates research on the microbial diversity analysis in the sediment and macro genomic library at the south pole, xiamen university isolation of vibrio harveyi bacteriophage with a potential for biocontrol of luminous vibriosis in hatchery environments fast length adjustment of short reads to improve genome assemblies highly accurate otu sequences from microbial amplicon reads ribosomal database project: data and tools for high throughput rrna analysis introducing mothur: open-source, platform-independent, community-supported software for describing and comparing microbial communities the silva ribosomal rna gene database project: improved data processing and web-based tools naive bayesian classifier for rapid assignment of rrna sequences into the new bacterial taxonomy study on the relationship between dominant fungi in the air and the allergic asthma in children field guide for the determination of biological contaminants in environmental samples separate effects of moisture content and water activity on the hyphal extension of penicillium rubens on porous media investigation and review of microbial pollution in air conditioning systems of public buildings. heating ventilating and air conditioning microorganisms and particles in ahu systems: measurement and analysis the indoor fungus cladosporium halotolerans survives humidity dynamics markedly better than aspergillus niger and penicillium rubens despite less growth at lowered steady-state water activity effects of temperature, humidity and air flow on fungal growth rate on loaded ventilation filters fungal growth and survival in building materials under fluctuating moisture and temperature conditions mathematical modeling of growth of salmonella in raw ground beef under isothermal conditions from to °c model for combined effect of temperature and salt concentration/water activity on the growth rate of staphylococcus xylosus the study is supported by the national nature science foundation of china ( ), beijing key lab of heating, gas supply, ventilating and air conditioning engineering (nr k ), the fundamental research funds for the central universities (dut qy ) and the urban and rural housing construction science and technology plan project ( -k - ). key: cord- -vgfv pi authors: hall, graeme m. j.; hollinger, david y. title: simulating new zealand forest dynamics with a generalized temperate forest gap model date: - - journal: ecol appl doi: . / - ( ) [ :snzfdw] . .co; sha: doc_id: cord_uid: vgfv pi a generalized computer model of forest growth and nutrient dynamics (linkages) was adapted for the temperate evergreen forests of new zealand. systematic differences in species characteristics between eastern north american species and their new zealand counterparts prevented the initial version of the model from running acceptably with new zealand species. several equations were identified as responsible, and those modeling available light were extended to give more robust formulations. the resulting model (linknz) was evaluated by comparing site simulations against independent field measurements of stand sequences and across temperature and moisture gradients. it successfully simulated gap dynamics and forest succession for a range of temperate forest ecosystems in new zealand, while retaining its utility for the forests of eastern north america. these simulations provided insight into new zealand conifer–hardwood and beech species forest succession. the adequacy of the ecological processes, such as soil moisture balance, decomposition rates, and nutrient cycling, embodied in a forest simulation model was tested by applying it to new zealand forest ecosystems. this gave support to the model’s underlying hypothesis, derived from linkages, that interactions among demographic, microbial, and geological processes can explain much of the observed variation in ecosystem carbon and nitrogen storage and cycling. the addition of a disturbance option to the model supported the hypothesis that large‐scale disturbance significantly affects new zealand forest dynamics. individual-based forest simulation models predict the dynamics and structure of complex forest ecosystems. worldwide, long-term forest composition and forest species distributions are under pressure from continuing large-scale anthropogenic effects. because forest simulation models are ecosystem based, they can provide both predictions of forest response to these impacts and a consistent synthesis of the ecological processes involved (rastetter ) . such properties make them a vital part of any assessment of ecosystem response to global change (reynolds et al. , shugart and smith ) . the structure of most simulation models of forest succession can be traced back to those developed to reproduce the population dynamics of trees in mixedspecies forests of northeastern north america (botkin et al. , shugart and west ) . this approach tracked the development of each individual plant throughout its life cycle, with forest dynamics simulated by calculating the competitive interrelationships among trees in a restricted area, similar to that resulting manuscript received april ; accepted january ; final version received march . from the gap in a forest canopy formed by the death or removal of a large canopy tree. by simulating a sufficient number of gaps, the dynamics of the forest are reproduced (yamamoto ) . this concept is supported by various plant succession studies, which show that changes in a forest ecosystem may be described by averaging the growth dynamics in gaps of different successional ages (watt , bray , curtis , forman and godron . forest gap simulation models have been developed to predict long-term impacts on forest ecosystems caused by blight, harvest management, past climates, animal browse, pollution, and large-scale disturbance by fire or storm, and to predict transients in species composition and forest structure due to changing climate, (e.g., shugart and west , aber et al. , solomon et al. , pastor and post , bugmann . shugart and smith ( ) compiled a list of such models developed to simulate vegetation dynamics in environments ranging from cool northern hemisphere boreal forest to warm subtropical australian rain forest. almost half of these models are dedicated to north american vegetation ( % of these in eastern forests), with the other models predicting forest composition and dynamics in central and northern europe, australasia, africa, and asia. comparisons have shown that a forest gap model developed for one geographical area is unlikely to contain all of the ecological processes required to successfully simulate forest composition and structure for another area . this is partly because, despite a common lineage, models formulate the basic processes of species establishment, growth, and mortality variously. models also vary by the way in which resource limitations alter growth, by the species' life history attributes employed, and by the depth of physiology incorporated. not all models explicitly maintain a soil moisture balance or a litter decomposition-soil nutrient cycle. in an effort to better determine the potential role of climate change and other exogenous factors on new zealand forest development, we extended an eastern north american simulation model (pastor and post ) for new zealand's forests. in doing so, we evaluated the ecological generality of functions and algorithms developed originally for the eastern north american forests, and gained insight into several longrunning debates about ecological processes in new zealand forests. the nature of new zealand's topography, climate, and forest characteristics indicated which ecosystem processes to model. new zealand's small group of islands (landmass km ) have variations in climate, geology, and soil that offer a wider range of habitats than many much larger landmasses (wardle , molloy ). the three main islands (north, south, and stewart) span km ( Њ) of latitude and have a maritime climate, with sea level temperatures ranging from warm temperate in the north (mean annual temperature (mat) ϳ Њc) to cold temperate in the south (mat ϳ Њc). new zealand is tectonically active, with mountain building continuing at rates of up to mm/yr (whitehouse ) . about % of the land is Ͼ m above sea level, and % is defined as hilly or steep (molloy ) . the varied topography cuts across the prevailing westerly winds, modifying land temperatures and causing strong rainfall gradients. the north island has a central volcanic plateau with an average elevation of m and main axial mountain ranges running northeast to south. rainfall exceeds mm/yr in these areas and can reach mm/yr. there are also sizable areas of lowlands and coastal plains. the south island has a young mountain range running km north to south. the many summits Ͼ m create strongly differentiated climatic gradients. in the eastern lee of the south island range, rainfall drops to a low of mm/yr, whereas on the windward side in westland and fiordland, it may reach mm/yr near the ranges. the forests of new zealand are dominated by longlived evergreen species (wardle , wardle . nothofagus (beech) species characterize many forest types, occurring either in pure associations ( % of the remaining forested area) or in mixed forest ( % of remaining forest area). beech species predominate in mountain regions of both main islands. their wide ecological ranges enable their frequent occurrence in montane zones as well as lowland areas. in the northern warmer areas (north of Њ s), the conifer agathis australis dominates the forests in association with a mixture of hardwoods, podocarp, and beech species. further south, lowland forests are characterized by emergent, long-lived, evergreen podocarp species dacrydium, podocarpus, and prumnopitys. these are associated with a diverse group of broad-leaved hardwoods (beilschmiedia, metrosideros, and weinmannia) , characteristic of the main canopy (ogden et al. ) . the strong rainfall gradients and dry summer climates influence species composition and led us to consider models with a site water balance (e.g., post , botkin and nisbet ) . the young landscapes and generally infertile forest soils in new zealand (molloy , wardle ) similarly required models that relate soil nutrient status to species composition and forest stature (e.g., aber et al. , pastor and post , bonan . we chose the linkages gap model post , post and pastor ) because it included these ecosystem processes without excessive data requirements. it has explicit feedbacks between light, water, and nitrogen availability, and their effects on stand composition and productivity. previous gap models have related soil nutrient status to tree growth by species-specific sigmoid equations that reach maximum values at the highest reported basal area, or biomass, for a given region. the linkages model eliminates these site-specific maxima by explicitly simulating water and nutrient availability and using them to influence tree growth (post and pastor ) . however, the model lacks some recent modifications in allometric relationships, growth equations, and spatially explicit modeling of the light environment (e.g., leemans and prentice , martin , pacala et al. , prentice et al. , bugmann . a complete description of linkages is given in pastor and post ( ) . our model, linknz, is a version of linkages generalized for new zealand forest conditions, with species parameters obtained from an alternative database of new zealand species (g. m. j. hall and d. y. hollinger, unpublished manuscript) . here, we examine how well linknz simulates forest patterns and reproduces forest characteristics in a range of broadleaved hardwood and conifer forest types throughout new zealand. the model presented is intended as a basis for future development. some characteristics of new zealand tree species differ profoundly from those of eastern north american species. therefore, we were careful to extend, rather than modify, the mechanisms of competition or nutrient cycling. this preserved the model's original ability to reproduce dynamics of the forests of eastern north america and allowed a com-generalized forest gap model parison of results with linkages. our intention was to produce a more generally applicable forest gap model. the linkages model post , ) shares a common structure with the jabowa/foret class of stochastic tree population models that predict ecosystem dynamics through interactions between the forest and available resources (botkin et al. , shugart and west ) . the model simulates, on a yearly cycle, the establishment, growth, and mortality of all individual trees in a / ha plot ( . ha), adjusted by the effects of climate, soil properties, and competition. plot size corresponds to the average gap created by a dominant tree in eastern north american forests (shugart and west ) . initial diameter at breast height (dbh) is stochastically set between . cm and . cm. monthly rainfall and temperature variables, together with soil moisture capacity and wilting point, determine available site moisture. as the canopy forms and develops, light availability to each tree changes, affecting growth rates and the establishment of new trees. available soil nitrogen is initialized for the site and then determined annually by external inputs, losses due to leaching, and dynamics associated with processes of immobilization and mineralization during litter decomposition (post and pastor ) . available soil nitrogen then affects tree growth and stand composition, which, in turn, alters litter quantity and quality and modifies decomposition rates (pastor and post ) . a sufficient number of growing seasons, or annual cycles, was set to allow modeled forest biomass at each site to settle into an approximate steady state. the longevity of several widespread, dominant new zealand tree species (g. m. j. hall and d. y. hollinger, unpublished manuscript) led us to run the model for annual cycles, rather than the - considered sufficient for eastern north american species by pastor and post ( ) , or the -yr period used by pacala et al. ( ) . we retained this -yr time frame when simulating stochastic whole-stand disturbances, to allow long-lived individuals in any undisturbed stands to complete at least one life cycle. the stochastic nature of the model requires that more than one plot be simulated to obtain an adequate description of forest composition and structure (yamamoto ) . we generated plot successions on each site to smooth anomalous events. because of its linkages heritage, linknz also tracks details of soil organic matter and nitrogen pools, as well as site water balance. typically, both soil organic matter and soil n accumulate on a plot for several centuries until reaching an approximate steady state. by contrast, the estimates of transpiration depend only on the physical environment. these outputs assist in the understanding of factors regulating vegetation at different sites, but insufficient new zealand data prevented further evaluation. latitude, monthly temperature, monthly rainfall, and growing season data were obtained from new zealand meteorological service climate reports ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) . soil moisture holding capacity and wilting point were set according to broad soil type and soil moisture deficit maps (molloy ) , and knowledge of the seven test sites. the initial organic matter ( . mg/ha) and n levels ( . mg/ha) were left as the model default values. we selected tree species considered fundamental to the structure and functioning of new zealand native forest ecosystems, using a reference file of all woody species found on surveys (hall ) . these species included early successional species, understory species, and many major canopy species occurring throughout new zealand forests. we also included four common and widespread tree fern species, cyathea smithii, c. dealbata, c. medullaris, and dicksonia squarrosa, because of their influence on patterns of succession in new zealand forests (wardle ) . they are treated by the model in the same way as tree species. optimal growth constants were calculated from the equations of botkin et al. ( ) from species maximum dimension and longevity. species used are listed in the appendix. model parameters generated from the species' life history attributes included maximum height, maximum diameter, maximum longevity, limits of annual growing degree-day sums, shade and nutrient tolerances, establishment conditions and rates, and various canopy, foliage, and litter properties. data and methods of obtaining parameters are described in g. m. j. hall and d. y. hollinger (unpublished manuscript) . the linkages model, as presented by pastor and post ( ) , required modifications to its slow-growth, available-light, and decay-rate conditions to reproduce forests characteristic of new zealand sites. many major forest species did not feature in the simulations. for example, at riverhead near auckland, linkages limited species composition to the large conifer agathis australis and two hardwoods, beilschmiedia tawa and b. tarairi, and excluded the common podocarp species dacrydium cupressinum, prumnopitys ferruginea, and p. taxifolia. further south at taupo, linkages predicted that the hardwood b. tawa would dominate and exclude all the common podocarp species. this pattern was repeated at other sites throughout the country, with site occupation being captured by one or two hardwood ecological applications vol. , no. or beech species and other widespread species failing to establish. the linkages model tests the growth of a tree independently of age and employs conditions common to many other models. if the resource-limited diameter increment is less than a fixed minimum of mm/yr (botkin et al. , shugart , botkin and nisbet , or Ͻ % of the maximum increment for its size (solomon ) , the model defines this as a poor growth year for the stem. if a tree grows poorly for two consecutive years, the probability of survival is reduced so that it has a Ͻ % chance of surviving consecutive years of poor growth. this ''poor-growth'' condition is too restrictive for modeling several slow-growing new zealand species. the clearest case is that of the long-lived conifer halocarpus biformis, with a highest recorded growth increment of . mm/yr (wardle ) , preventing linkages from establishing it at all. under link-ages, the widespread, dominant conifers of the podocarpaceae with recorded maximum annual diameter increments Ͻ mm/yr also struggle to survive in slightly less than optimal conditions. slow growth is common in the understory for other new zealand species as well. seedlings of nothofagus species, for example, remain in a quiescent state (Ͻ . m tall), making only limited growth for decades until an opportunity is provided by the death of canopy trees (wardle ) . nothofagus fusca, normally one of the faster growing dominant species (stewart and rose ) , may grow in diameter by only . mm/yr under a dense canopy, with % of the poles in a typical stand passing through this stage (kirkland ) . baxter and norton ( ) show a similar growth release behavior for dacrydium cupressinum and quintinia acutifolia, in which ring widths of young trees are Ͻ . mm/yr under an intact canopy, but increase to widths of - mm/yr after the overstory trees are removed. to minimize alterations to the model, we retained the same slow-growth conditions of linkages when maximum diameter increments exceeded . mm/yr, or when maximum longevity was attained. for maximum increments Ͻ . mm/yr, growth was defined as slow only when resource-limited increments were Ͻ % of the maximum increment. the . mm/yr threshold was chosen by comparison against optimal growth increments recorded for a range of new zealand tree species (g. m. j. hall and d. y. hollinger, unpublished manuscript) . species with a maximum diameter increment exceeding . mm/yr will have, at some point, the same tests for slow growth as in linkages, whereas those that never attain this increment have the mm/yr fixed minimum waived. these conditions allowed the slowest growing species, including several in the podocarpaceae, to establish and move through the slowgrowing phase without excessive mortality. the light passing through a canopy can be modeled, using the beer-lambert law, as where i is the light intensity below the canopy; i o is the light intensity above the canopy; k is an extinction coefficient that is a function of foliage angle distribution, spatial dispersion, and optical properties; and lai is the leaf area index in square meters of foliage per square meter of ground (monsi and saeki ). the linkages model calculates i as a percentage of full sunlight. this value is used to determine whether there is sufficient light for new individuals to become established on the plot, and as a growth multiplier for seedlings of those species that can establish. in link-ages, it is calculated (aber et al. ) as: where f is the foliage mass in grams per plot, and the divisor is a factor that accounts for the size of the plot ( . m ) and converts the mass of the foliage into an effective leaf area index. the per meter conversion factor of . ( / . ) is thus the product of the leaf mass per unit area (slm) and the reciprocal of the extinction coefficient, k. whittaker et al. ( ) used a value of ϳ g/m for northern hardwoods foliage, implying that the implicit value of k from aber et al. ( ) is ϳ . . new zealand forests are predominantly evergreen, with high specific leaf mass, in contrast to the predominantly deciduous eastern north american forests. consequently, linkages underestimates available light in new zealand sites for a given foliage mass, with the result that canopy species prevent any new establishment. we rewrote eq. as to take account of the variation in slm, which was an implied constant in linkages, and then used our measured slm values. within linkages, canopy openings are assumed to increase decay rates because of microclimatic changes. the model relates closed-canopy leaf production, l c , to available water for plant growth, w s (pastor and post ) , as it then compares this year's leaf litter (l a ) with l c to construct a decay multiplier that increases decomposition under canopies with low leaf area (aber et al. ) . for soils of high water-holding capacity, the multiplier ranges from . (l a ϭ l c ) to . (no canopy); for low water-holding capacity, it ranges from . to . (pastor and post ) . in linkages, the l c value for canopy leaf pro-duction is too low for most new zealand forests; l a exceeds l c , causing the model to set the decay multiplier to . and so nullify any gap effect. field examples show that daniel ( ) estimated l c as . mg·ha Ϫ ·yr Ϫ in a new zealand podocarp-broad-leaved forest; benecke and evans ( ) established that l c ϭ . mg·ha Ϫ ·yr Ϫ from a nothofagus truncata forest; and hollinger et al. ( ) obtained l c ϭ . mg·ha Ϫ ·yr Ϫ from a mature n. fusca stand. leaf production may be greater in new zealand forests because of a long growing season, combined with the evergreen habit of most tree species. over a -yr run of the model, simulations at a warm-temperate species-rich forest site at riverhead, auckland showed that the original decay multiplier was nullified in . % of years on soils of low water-holding capacity, and in . % of years on soils of high water-holding capacity. to compensate for differing new zealand forest leaf production values and to retain eq. for american species, a scaling factor was applied to l a . this was obtained by multiplying each plant's litter mass by the average northern hardwood foliage mass of g/m (whittaker et al. ) and dividing by the species' slm. over the simulated period, the model calculated mean values for l a of . mg·ha Ϫ ·yr Ϫ on low-water-holding capacity soils and . mg·ha Ϫ ·yr Ϫ on high-capacity soils. the slm adjustment rescaled these mean values downward by nearly %, making them comparable with the l c of old-growth american forests in eq. , and activating the decay rate multiplier in new zealand forest sites. after this adjustment, the decay-rate multiplier on the simulated riverhead forest plot was negated on just . % of the annual cycles. pollen, charcoal, and fossilized plant fragments point to a long history of change in new zealand indigenous forests. forest disturbances have been due to physical factors associated with steep mountain slopes, and include volcanism, periodic fire, forest dieback, windthrow, drought, flooding, and snow damage (wardle ) . we added a basic mechanism to linknz to simulate disturbance. the type (wind or fire) and the mean disturbance return time can be set in the site parameters. the linknz model will trigger that type of disturbance with a probability annually that is the reciprocal of the disturbance frequency. after windthrow, all trees on the plot are assumed to be dead and all biomass, including belowground root mass, is returned to the site for decomposition-nutrient cycling. after fire, all trees are assumed to be dead and the larger biomass components are returned to the site. the biomass and n in foliage and twigs are presumed to be volatilized and lost from the site. revisions made to the linkages code (pastor and post ) gave a threefold improvement in execution speed. the linknz program, species data, and site data will be made available, subject to a ''fair use'' policy, on the web site ͗http:/www.landcare.cri.nz͘. evaluation of forest ecosystem gap models is not straightforward. to construct definitive tests of simulated forest dynamics, several stands would have to be monitored for long periods while they were returning to old-growth forest. even in well-observed eastern north american forests, a lack of historical data on succession has been acknowledged (pacala et al. ). in addition, rastetter ( ) has noted that each alternative used to evaluate modeled ecosystem response to global change fails to provide a severe and crucial test. partly, this is due to difficulties in locating past or present ecosystem states comparable to those expected under climate change. for instance, how can valid data be obtained for evaluating long-term vegetation responses to increased co , when short-term chamber experiments are still inconclusive? we adopt recent evaluation methods for forest gap models to evaluate the ability of the model under current climates, but acknowledge that theoretical difficulties remain. shugart ( ) and shugart and smith ( ) discuss procedures used for testing the results of gap models. most include assessing the model's abilities to reproduce ''target patterns'' of stand or tree biomass increments, stand structure (basal area, density, stem diameter distributions) or composition (relative basal area, relative density) for stands of known age, successional trends in a chronosequence of stands, and forest response to disturbance as a ''natural'' experiment. bugmann ( ) tested model predictions of species composition, biomass, and distribution on a range of sites in the european alps. we assess linknz similarly by comparing output with general characteristics of forest vegetation at sites throughout new zealand. these sites are located across both temperature and precipitation gradients in habitats varying from diverse-species warm-temperate forests to limited-floristic cool-temperate forests (table ) . pacala et al. ( ) tested their spatially explicit model against data from a short chronosequence (up to yr), and against a long-term succession, by comparing species composition and basal area. we compare our model through time against studies of forest successional development on landslides in southwestern new zealand (mark et al. , stewart ). comparisons of models and methods can give confidence if underlying methods are independent (rastetter ) . we compare simulated long-term dynamics against successional sequences deduced by methods based on empirical data and models unrelated to linknz (e.g., ogden , wardle , burns and smale . bugmann et al. ( ) compare versions of forest gap models showing changes due to additional features. we briefly compare results between our model and a test of an allometry- finally, we present and discuss model results for forests where large-scale disturbance appears to play a role in shaping forest structure. the model reproduces the broad patterns of forest succession and composition at a variety of test sites. at riverhead, auckland, a silty clay loam soil texture (field moisture capacity . cm, wilting point of cm) was chosen to represent moister, valley soil conditions. from the simulations, mean relative stem densities at -yr intervals for the main species (with mean biomass Ͼ . mg·ha Ϫ ·yr Ϫ ) were clustered using a group-average linkage with a gamma similarity coefficient (systat ) . this gave three groups of species at the . similarity level. the early-arriving species (group ) separated at the . similarity level into primarily short-lived, small trees (group a) and longer lived tall trees (group b). initially, modeled primary succession proceeds through the fast-growing group a species, especially leptospermum scoparium and kunzea ericoides, with aristotelia serrata also present for the first yr (fig. ) . included in group a is the slow-growing phyllocladus trichomanoides, which can maintain a longer presence. the relative density of these pioneer species drops rapidly after - yr, returning only when large canopy gaps occur after year . the massive conifer agathis australis attains dominance in Ͼ % of the simulated plots by year and, like the group a pioneers, begins at a high density. over the first ϳ yr, agathis also replaces group b, early-establishing hardwood species including weinmannia silvicola, knightia excelsa, elaeocarpus dentatus, and the beech nothofagus truncata (fig. ) . the model suggests, however, that agathis does not regen-erate well in situ, with its relative density declining steadily for Ͼ yr and its biomass dropping until it disappears after - yr. as agathis declines after - yr, the forest becomes co-dominated by species from groups b and , forming an agathis/ hardwood community with an increasing podocarp component (figs. and a) . the group species, beilschmiedia tawa and b. tarairi, rise in numbers, reaching % of total biomass between years and , and then decline to hold a near-constant % of total biomass after yr, as the longer lived group podocarps emerge. of these, dacrydium cupressinum gains slowly in relative density and biomass after group a species disappear, while prumnopitys ferruginea and p. taxifolia increase rapidly as agathis wanes. in the absence of a large disturbance, such as fire, these podocarp species (with a small hardwood component) are predicted to eventually characterize the forest. our simulations indicate that, at this warm new zealand site, maximum forest biomass is reached between and yr, while agathis dominates. during that period, the model predicts a mean basal area of m / ha (fig. b) , of which agathis contributes m /ha. this compares well with a mean basal area of m / ha for agathis, observed in mature (mean age yr) stands . with a drier soil (moisture capacity . cm, wilting point . cm), depicting ridge conditions at riverhead, agathis is predicted to persist at the expense of the group hardwoods. the simulation (not graphed) produces a . % higher mean agathis biomass, with a lower maximum and a smoother decline. the hardwood species beilschmiedia tawa and b. tarairi, which favor more fertile soils and are less tolerant of water stress than agathis, drop in mean biomass by Ͼ %, from . % to . %. in comparison, the three major podocarp species retain - % of total biomass. the droughttolerant prumnopitys taxifolia prospers at the expense of both d. cupressinum and p. ferruginea. in the cooler climate further to the south, near taupo generalized forest gap model fig. . relative stem densities of key species on modeled plots using climate and soil data from riverhead, auckland. other species of lesser importance on the modeled plots include: elaeocarpus hookerianus in group b, quintinia serrata in group , and podocarpus totara and podocarpus hallii in group . in this and all subsequent simulations, the values shown are the means from simulated / ha plots. minimum and maximum mean percentages of total stems are given for each species. ( fig. a) , modeled primary succession on silty clay loam soil proceeds again through kunzea ericoides with leptospermum scoparium, and aristotelia serrata. weinmannia silvicola is replaced by the cooler climate species w. racemosa, and warmer temperate species such as the hardwood beilschmiedia tarairi and the dominant conifer agathis fail to establish. the common north island hardwood b. tawa retains a small, constant biomass (ϳ %) throughout the simulation period. in the first yr, modeled plots are dominated by weinmannia racemosa, k. ericoides, and elaeocarpus species. these species make up Ͼ % of total biomass at year , reduce to % by year as the podocarps dacrydium cupressinum, prumnopitys ferruginea, and p. taxifolia increase, and become a minor component at % by year . modeled community composition is similar to that at lakeside sites in the taupo area (wardle , clarkson and nicholls ) . the successional patterns resemble a sequence described by wardle ( ) for parts of this central north island volcanic plateau area with a deep tephra soil, characterized by d. cupressinum-dominated mixed-podocarp forest establishing by - yr and developing by - yr into a large, mature podocarp-broadleaved forest. by contrast, simulations carried out using the cooler climate conditions for reefton (typical of the south island west coast of new zealand) suggest that the emergent podocarp dacrydium cupressinum, in association with the common hardwood weinmannia racemosa, will more quickly dominate plots in this area (after the initial establishment of aristotelia serrata, leptospermum scoparium, and kunzea ericoides). the model shows prumnopitys and podocarpus species, followed by nothofagus species, beginning to establish after ϳ yr (fig. b) species. apart from n. fusca (to be discussed), forest composition agrees with descriptions of the area (wardle ). the drop-off in biomass over yr at about year represents mortality of the last of the original cohort of the long-lived, dominant d. cupressinum. the gap model predicts that the eventual ''steady-state'' forest composition of the major species at taupo and reefton may be similar (fig. a, b) , with comparable patterns for w. racemosa and the podocarp species d. cupressinum, p. ferruginea, and p. taxifolia at both sites. slight differences are evident, with reefton predicted to have a larger beech component and b. tawa restricted to taupo, as observed in nature. the model generates a similar forest km south at franz josef (fig. c) , where the mean annual temperature is only slightly lower than in reefton (table ) . early succession at this site is started by leptospermum scoparium and aristotelia serrata, rather than kunzea, as at the more northern sites. a similar amount of nothofagus menziesii is predicted in forest at franz josef, as for reefton (fig. b, c) . however, franz josef is located within the -km stretch of the south island west coast where nothofagus is absent (referred to as the ''beech gap''). excluding beech species from the model at franz josef does not alter the forest dynamics greatly, and correctly predicts a dacrydium cupressinum-dominated podocarp forest (fig. d ). other species include weinmannia racemosa, quintinia acutifolia, pseudowintera colorata, and cyathea smithii. still omitting beech species from the model, we simulated climate conditions several hundred meters upslope of the franz josef meteorological station by reducing mean monthly temperatures by Њc. at this upslope site, a mixed hardwood-podocarp forest is sim-ulated with metrosideros umbellata, weinmannia racemosa, podocarpus hallii, and phyllocladus aspleniifolius var. alpinus (fig. e ). other species present include griselinia littoralis, libocedrus bidwillii, and pseudowintera colorata. this change in species composition with elevation corresponds with that commonly observed along the western slopes of the southern alps in new zealand (e.g., wardle ) . the total biomass in these simulated slope forests is about two-thirds that estimated for the lowland podocarp forests. this reduction is caused partly by a change in species composition from the large, lowland podocarp, dacrydium cupressinum, to the smaller-statured podocarp, p. hallii, with the broad-leaved hardwood w. racemosa, and is exacerbated by the decline in biomass of the dominant m. umbellata. metrosideros initially dominates the forest, reaching ϳ mg/ha after yr, but is gradually replaced by podocarpus hallii after yr. on the drier east side of the southern alps, the model simulates very different forests from those on the wetter west side (fig. f, g) . at the driest site (mean precipitation mm/yr) near twizel, on a sandy soil, the model generates a forest dominated by podocarpus hallii, with a small amount of nothofagus solandri var. cliffortioides, phyllocladus aspleniifolius var. alpinus, and prumnopitys taxifolia. early succession at this site is dominated by leptospermum scoparium and n. solandri var. cliffortioides. the simulated biomass of these plots is ϳ mg/ha. although there is no forest at present around twizel, on adjacent slopes there is abundant charcoal evidence for a p. hallii forest before the arrival of polynesian settlers in new zealand (molloy et al. , wells ). on drier, cooler sites, pollen and charcoal evidence from the foot of the ben ohau range near twizel record a p. alpinus-dominated scrub with a lesser p. hallii component and traces of n. solandri var. cliffortioides (mcglone and moar ). at the higher elevation craigieburn site (fig. g) , on a sandy-loam soil, the model generates a mixed nothofagus solandri var. cliffortioides-n. menziesii forest where there is presently solely n. solandri var. cliffortioides. this modeled forest exhibits interesting dynamic behavior, and will be discussed in more detail (see natural monocultures). simulated forest succession at lake thompson (using climate data from the west arm, manapouri station) allowed comparison with several detailed studies of succession in the area (mark et al. , stewart . qualitatively, much of the early pattern of succession observed by mark et al. ( mark et al. ( , was reproduced by linknz, with aristotelia serrata and tually n. menziesii. other species that occur in forested sites close to lake thompson (stewart ), such as pseudowintera colorata, a valley floor forest co-dominant in the understory, and the conifers podocarpus hallii and prumnopitys ferruginea, are also present in our simulations (fig. a, b) . quantitative predictions were more variable. the linknz model reproduced the dynamics of the short-lived pioneer aristotelia serrata, a co-dominant in the -yr-old slip-face plot (exceeding % relative density), and absent from stands Ն yr old (mark et al. ). the principal seral species, leptospermum scoparium, was prolific in the -yr-old stand ( % relative density), much less abundant in the - yr old stands ( %), and not recorded in the mature forest. our simulation reproduced the initial relative density and persistence of l. scoparium, but predicted only % relative density after yr. predictions for the lightdemanding beech species nothofagus solandri var. cliffortioides at year and year were within % and %, respectively of the field observations. the model properly predicted that the mature-forest dominant n. menziesii would establish at an early stage, but gave it only slight increases in relative density until year , whereas mark et al. ( ) found that it occurred in moderate amounts on both -yr-old and -yr-old slip-face plots. the linknz model predicted that high numbers of podocarpus hallii and prumnopitys ferruginea would occur after year . these affect predicted relative densities of other species. weinmannia racemosa was found by mark et al. ( ) to contribute % relative density at year and % in the mature forest. the linknz model predicted % w. racemosa relative density at year , dropping to % at year as the podocarps became abundant (fig. a) . mark et al. ( ) recorded a relatively high density of the hardwood metrosideros umbellata after yr, whereas the model predicted just % for this species after yr. despite these differences, the simulation of an adjacent mature nothofagus menziesii-weinmannia racemosa-pseudowintera colorata forest is still very acceptable (e.g., stewart , mark et al. ). with reference to species' relative densities, we note that our modeled results are the average of stands; individual stands can follow quite distinct trajectories from the mean. introducing relatively infrequent stochastic disturbance ( / ha total plot blow-down on the average of - yr) results in simulated forests with a greater representation of early successional species and lower biomass than in forests where only individual tree gap replacement dynamics are allowed. this regime has the effect of not only increasing the number of gaps over time, but also dramatically altering the stand structure. over time, a plot generally carries several trees so that, when one dies, any remaining individuals are free to respond, but under the blow-down scenario, all remaining trees on the plot are killed (e.g., fig. a vs. fig. a, and fig. b vs. fig. b ). in these simulations, the long-lived podocarp species decline in absolute as well as relative importance, while other species tend to maintain absolute biomass and increase in relative representation. this is illustrated at the reefton site, which has a history of mass disturbances including two major earthquakes in the area during this century (wardle ) . when we imposed a mean disturbance return time of yr and allowed fallen material to remain on site, predicted total biomass of nothofagus species at reefton increased from . % without disturbance to . %. the beech n. fusca is common in the area (wardle ) , but in simulations without disturbance, it was virtually absent (fig. b) . the introduction of disturbance allowed n. fusca to capture Ͼ % of total biomass, increased n. menziesii from % to % of total biomass, and led to the establishment of a small amount of n. solandri. with this disturbance regime, the earlyestablishing and common hardwood weinmannia racemosa retained a constant presence and more than doubled its share of biomass from . % to . %. these gains came at the expense of the podocarps, with the biomass of prumnopitys species reduced by onehalf and dacrydium cupressinum reduced by one-third to %, compared to a scenario without random disturbance. the lists of species and successional patterns generated by the model for different sites in new zealand agree closely with those observed at the local sites (e.g., wardle , wardle ). at the sites tested, the model does not establish any species where it does not belong, with the exception of the ''beech gap,'' to be discussed. for example, the commonly described pattern of colonization, which has the fast-colonizing kunzea ericoides and leptospermum scoparium acting as ''nurse'' species for agathis australis or podocarps (ogden ) , is reproduced in the riverhead, auckland simulation ( fig. a) , as is the eventual replacement of agathis by hardwoods and podocarps . in the initial - yr, the riverhead simulation also shows the common hardwoods weinmannia silvicola, knightia excelsa, elaeocarpus dentatus, the beech nothofagus truncata, and phyllocladus trichomanoides establishing in numbers, and then gradually being overtopped and replaced by agathis (fig. ) . this successional sequence in these warmer temperate forests is described by ecroyd ( ) and burns and smale ( ) . the model also reproduces observed changes in species composition for different soils. dri-er soil sites favor longer term agathis occupation, whereas moister soils suit hardwood species (ecroyd ) . the biomass and basal area estimates produced by the model are more difficult to evaluate, but agree in general with published estimates. from a harvesting trial in a typical -yr-old agathis stand in the hunua ranges south of auckland, madgwick et al. ( ) estimated the agathis biomass component at ϳ mg/ ha. on agathis-occupied plots, with hunua climate data, the model predicts a mean ϳ mg/ha agathis component at yr. huge biomass is possible in clumped stands of mature agathis forest (wardle ) . hinds and reid ( ) estimate that a representative area of a typical mature agathis forest with merchantable stems/ha could produce Ͼ mg/ha, on average, of commercial timber. madgwick et al. ( ) found that stemwood made up % of total agathis biomass in their study; this factor generates an approximate total biomass estimate of mg/ha. this compares with the model estimate in which total agathis biomass reaches a peak of mg/ha in - yr old stands. near riverhead, ogden ( ) recorded basal areas of agathis of m /ha for a ϳ -yr-old stand and m /ha for a young ϳ -yr-old ''ricker'' stand. in comparison, our model simulations predict lower mean basal areas of m /ha on -yr-old agathis-dominated stands and m /ha at yr (fig. b ). burns and smale ( ) , on an intermediate-stage -yr-old site on the coromandel peninsula, obtained a total basal area of m /ha, of which % was contributed by agathis. our model, with their coromandel site climate data, predicted a similar total stand basal area, but with an % agathis component. for reefton, we estimated undisturbed mature podocarp-beech forest aboveground biomass at ϳ mg/ ha (fig. b) and belowground biomass at ϳ mg/ ha. for periodically disturbed mature podocarp-beech forest, above-and belowground biomass was ϳ and ϳ mg/ha, respectively. beets ( ) recorded aboveground (living) and belowground (excluding logs) biomass values of mg/ha and mg/ha, which included mg/ha of roots, at a mature mixedpodocarp-beech site near reefton. in the craigieburn range, several studies have investigated nothofagus solandri var. cliffortioides aboveground biomass, finding values that range between and mg/ha (benecke and nordmeyer , schoenenberger , harcombe et al. ). our simulated values range from to mg/ha. although the model successfully simulates broad successional patterns within new zealand forests, detailed patterns may not be exactly reproduced, particularly during the early-establishment stages. forest gap models incorporate stochastic elements to mimic many ecosystem processes and produce multiple simulations to obtain average results and calculate confidence limits. uncertainty in model data and possible errors in ecological applications vol. , no. field data also obscure reconstructions of past events (rastetter ) . deviations between model and reality result from a number of sources. these include potential errors in site parameters such as climate or soil water-holding capacity, errors in species parameters, and flaws in our understanding (and in modeling this understanding) of how site and species characteristics interact and affect forest growth. the results from lake thompson are instructive. our climate estimates for lake thompson are taken from a site km distant, in an area of high relief and climatic extremes; our estimate of initial site fertility may also be imprecise. we use the same set of growth parameters for each species throughout new zealand (ignoring ecotypic variation) and some of these parameters are relatively imprecise estimates. a chronic problem with testing gap models in this way is that the species potentially available on a plot exceed those that naturally occur (the model provides for an omnipresent seed source, ignoring seed dispersal mechanisms and differing arrival times). this results in a relatively high percentage of ''other'' species that may exist for only several years before dying off. in addition, during early establishment, the model initiates all individuals as equal-aged saplings with a mean stem dbh of . cm. time required to reach this point is not explicitly accounted for and there can already be considerable age differences between individuals, depending upon their species' growth characteristics. despite these problems, the detailed pattern of succession simulated for landslides near lake thompson is reasonable. the species characteristics given to linknz were not altered nor were the species limited to those present in the lake thompson area. develice ( ) presented a basic foret-type model with allometric parameters set for the five tree species of greatest importance at the lake thompson landslides. even so, his model overestimated the initial density of nothofagus solandri var. cliffortioides and the subsequent density of n. menziesii. in these develice ( ) simulations, n. menziesii accounted for nearly % of total stand density by year , whereas mark et al. ( mark et al. ( , found that its density was generally Ͻ % of the total. the linknz model produced a better approximation to field counts of these nothofagus species, and predicted the increases in pseudowintera colorata over time that were noted by mark et al. ( ) . it did deviate from field observations by initially establishing podocarp species (not included by develice ); although these species did not become a significant part of the site biomass until year (fig. b) , they lowered relative density predictions for the hardwood species. the simulated early and numerous establishment of these bird-dispersed podocarp species lends support to the contention of mark et al. ( ) that the initial floristics model of primary succession (egler ) , in which species arrive simultaneously and successively gain dominance according to their life history attributes, may not fully account for the early dynamics on these slip faces. in summary, although our results and those of develice ( ) deviate in some details from the short-term pattern of succession reported by mark and co-workers, the mature forest simulation of linknz corresponded closely to that described by stewart ( ) and mark et al. ( ) . having established the validity of the model for reproducing general successional patterns across a range of sites in new zealand, we then used the model to provide some insights on several ongoing debates concerning ecological patterns and processes in new zealand. there has been debate concerning the ''lack'' of regeneration in agathis australis forests (for discussion, see ecroyd and ogden ) . our results, based solely on gap-phase replacement dynamics inherent in the model, support the primary succession theory of egler ( ) and the early belief (cockayne ) that agathis is successional to a climax podocarp forest (figs. , a) and that strongly agathis-dominated forests would occur only in the first yr after largescale disturbance. the gap size and frequencies produced by the model are a consequence of the comparative life history attributes set for the trees that occupy the site. thus, long-lived trees will produce gaps only infrequently; in the case of agathis, this frequency is so low that there is sufficient time for the shade-tolerant hardwoods and podocarps to become well established, reducing the likelihood that enough light would penetrate through the understory of a tree-sized gap to permit abundant agathis regeneration. in a study of agathis treefall gaps, ogden et al. ( ) found that agathis established in only a few, enough to maintain a presence but not dominance. they estimated that, owing to its longevity (mean Ͼ yr), agathis could survive on any site up to - yr. the corresponding simulation ( fig. a) , without large-scale disturbance, shows agathis declining from an initial dominance, but remaining a significant component of the forest for nearly yr. many workers have pointed out the importance of larger scale, infrequent disturbance to new zealand forest dynamics (e.g., stewart , ogden ) , and have urged acceptance of ''kinetic'' models (e.g., veblen et al. ) in which stochastic disturbance is accepted as a selective force. ahmed and ogden ( ) inferred from their study of agathis population structure that episodic regeneration occurred at intervals of - yr. our introduction of stochastic disturbance of this frequency into the dynamics of the model (fig. a) have gone, declining to zero by year (fig. a) . in the stochastically disturbed scenario (fig. a) , agathis biomass remains at ϳ % of total biomass over the entire interval between and yr. our results are consistent with the conclusion of ahmed and ogden ( ) : agathis is a successional species that maintains a strong presence in the forests of northland because of repeated disturbance. nothofagus species are completely absent from a -km stretch along the central west coast of the south island, and are also absent from stewart island, km south of the south island. cockayne ( ) suggested that this absence could be the result of insufficient time for beech to have recolonized the area since the end of the last glaciation (ϳ yr bp). our results for sites within the ''beech gap,'' such as franz josef and hokitika (data not shown), lend support to cockayne's hypothesis. when beech is part of the available species pool, it can establish and become a permanent, if lowbiomass, component of the forest. even when beech availability on the site was delayed yr to allow the forest to fully establish first, linknz indicated that a small amount of beech could establish. this suggests that poor modes of dispersal in beech may play a more significant role than any total inability to compete. wardle ( ) also suggests that new zealand beech may compete less effectively with existing vegetation where the rainfall is high, such as on the west coast, than in the drier conditions to the east of the axial ranges. the linknz model supports this contention, because the simulations show that beech requires a longer period of time to become established in the ''beech gap'' sites than on the drier, cooler east side of the south island in the craigieburn range (compare fig. b , c with fig. f, g) . beech establishment in this region may also be influenced by the effects of stand-level disturbance on soil fertility. the model suggests that stochastic disturbances that create gaps larger than normal treefall size promote the establishment of some beech species. wardle ( ) points out that beech can be outcompeted in low-elevation, high-rainfall areas owing to difficulties in finding suitable sites and increasing competition in the high-density, species-rich understories. ogden ( ) describes the role of synchronized cohort mortality in the dynamics of beech and how disturbance promotes nothofagus fusca establishment. when we included beech in the species pool and introduced wind-blow disturbances ( -yr mean return time) to the simulations at franz josef located in the ''beech gap,'' beech biomass increased by a factor of to reach ϳ % of the total; n. menziesii biomass doubled, and the fast-growing n. fusca established and attained Ͼ % of total biomass. by contrast, a disturbance regime that removed organic matter from the site inhibited beech species establishment and produced a more typical low-elevation hardwood-podocarp forest. ogden ( ) notes that n. fusca prefers higher fertility sites and would be expected to have difficulty establishing. at this low-elevation, low-fertility site, the model predicts that other species, such as the common hardwood weinmannia racemosa, would capture the area and lead to the establishment of large, dominant podocarps such as dacrydium cupressinum. our results for the craigieburn forest show that biomass contributions of the beech species nothofagus solandri var. cliffortioides and n. menziesii tend to oscillate out of phase with one another in a damped cycle of ϳ yr (fig. a ). an ecological interpretation of this behavior is that, as the fast-growing, light-demanding, even-aged stands of n. solandri var. cliffortioides thin, they become replaced by relatively evenaged stands of the shade-tolerant n. menziesii. when these n. menziesii stands senesce, the faster growing n. solandri var. cliffortioides begin to recapture the site. because n. menziesii is longer lived and can continue to regenerate under its own canopy, these oscillations lengthen and decrease in amplitude over subsequent generations. this ''counter-cyclical succession'' is a consequence of the life history attributes of the two species most suited to the cool craigieburn climate. both species have wide, overlapping soil fertility, soil moisture, and climatic tolerances (wardle , benecke and allen ; g. m. j. hall and d. y. hollinger, unpublished manuscript) . in new zealand, these species form almost continuous alpine and subalpine forests throughout the axial mountain ranges of both main islands. wardle ( ) found that in mixed stands, dense, smalldiameter (young) nothofagus menziesii usually occur in conjunction with large-diameter (old) n. solandri var. cliffortioides and, conversely, stands with low numbers of large n. menziesii trees often have high densities of young n. solandri var. cliffortioides trees. ogden ( ) further suggested that n. solandri var. cliffortioides will gradually be replaced by n. menziesii unless the stand is severely disturbed. in fact, the craigieburn forests are essentially monocultures of n. solandri var. cliffortioides, with few other canopy species. disturbance events such as wind, earthquake-triggered landslips, and heavy snowfall (wardle ) frequently disrupt these subalpine and alpine forests on the eastern slopes of the south island axial range. age-diameter distributions indicate that stands may be severely damaged by gales at periods of ϳ - yr, and stands suffer minor damage at intervals of - yr (wardle , jane , harcombe et al. ). when we investigated the influence of disturbance on the dynamics and biomass of n. solandri var. cliffortioides-n. menziesii forests by adding stochastic disturbance (blowdown of whole-plot biomass) to the dynamics of the model, the counter-cyclical pattern of succession was removed (fig. b) . furthermore, as the mean interval between disturbances decreased, the relative percentage and absolute amount of n. menziesii in the resulting stands decreased from a % dominance to Ͻ % for mean intervals greater than one event every yr (fig. ) . this model behavior provides support for wardle's ( ) conclusion that stability favors n. menziesii and disturbance favors n. solandri var. cliffortioides, and suggests why n. solandri var. cliffortioides can be so dominant in disturbance-prone subalpine forests. we evaluated the degree to which principles and relationships derived from north american studies could be used to simulate the structure and dynamics of new zealand forest ecosystems. the characteristics of new zealand tree species are significantly different from those of eastern north american trees. new zealand species are generally long-lived evergreens with low n and high specific mass foliage that is retained in the canopy for several years. yet, a model that was based on ecological processes of tree competition and growth, litter decomposition, and n cycling that originated primarily in north america and was designed to simulate the ecology of eastern north american species, only required minor modifications to acceptably simulate general forest patterns and processes across climatic gradients in a range of new zealand forest types. the most significant modifications improved the way in which the predecessor model, linkages, calculated the forest floor light environment and modified the slow growth rate conditions that trigger mortality. overall, our results tested the adequacy of the ecological processes embodied in forest simulators such as linkages. they provided support for the model's underlying hypothesis: interactions among demographic processes determining plant population structure, microbial processes determining n availability, and geological processes determining water availability explain much of the observed variation in ecosystem c and n storage and cycling (pastor and post ) . in this framework, geology and climate act as constraints within which feedbacks between vegetation and light availability, and between vegetation and n availability, operate. by incorporating a simple disturbance regime into the model, we also supported the hypothesis that large-scale disturbance is of importance in shaping the dynamics and current composition of new zealand forests (veblen and stewart , ogden , wardle ). the linknz model is a versatile simulation model of vegetation patterns and processes in new zealand forests. it may find additional practical applications in investigating the impacts of climatic change, forest harvesting practices, forest restoration, or introduced animal impacts on the dynamics of indigenous vegetation. furthermore, the modifications that we have incorporated should also improve the performance of the model in its original domain, the northeastern united states. acknowledgments many colleagues at landcare research and at the department of conservation provided assistance with this project. predicting the effects of different harvesting regimes on productivity and yield in northern hardwoods predicting the effects of rotation length, harvest intensity, and fertilization on fiber yield from northern hardwood forests in new england population dynamics of the emergent conifer agathis australis (d.don) lindl. (kauri) in new zealand. . population structures and tree growth rates in mature stands forest recovery after logging in lowland dense rimu forest, westland, new zealand amount and distribution of dry matter in a mature beech/podocarp community new zealand's mountain forests: their use and abuse. pages - in yang yupo and zhang jiangling, editors. protection and management of mountain forests. iufro project group p . - : ecology of subalpine zones growth and water use in nothofagus truncata (hard beech) in temperate hill country carbon uptake and allocation by nothofagus solandri var. cliffortioides (hook. f.) poole and pinus contorta douglas ex loudon ssp. contorta at montane and subalpine altitudes some ecological consequences of a computer model of forest growth forest response to climatic change: effects of parameter estimation and choice of weather patterns on the reliability of projections gap phase replacement in a maple-basswood forest a simplified forest model to study species composition along climate gradients a comparison of forest gap models structure and behavior changes in structure and composition over fifteen years in a secondary kauri (agathis australis)-tanekaha (phyllocladus trichomanoides) forest stand the distribution of beech (nothofagus) species in the east taupo zone after the bp volcanic eruptions monograph on the new zealand beech forests. part . the ecology of the forests and the taxonomy of the beeches the vegetation of new zealand the vegetation of wisconsin. an ordination of plant communities preliminary account of litter production in a new zealand lowland podocarp-rata-broadleaf forest test of a forest dynamics simulator in new zealand biological flora of new zealand. . agathis australis (d. don) lindl. (araucariaceae). kauri vegetation science concepts. i. initial floristic composition, a factor in old-field vegetation development patches and structural components for a landscape ecology pc-recce vegetation inventory data analysis spatial and temporal patterns in structure, biomass, growth, and mortality in a monospecific nothofagus solandri var. cliffortioides forest in new zealand forest trees and timbers of new zealand carbon dioxide exchange between an undisturbed old-growth temperate forest and the atmosphere wind damage as an ecological process in mountain beech forests of canterbury preliminary notes on seeding and seedlings in red and hard beech forests of north westland and the silvicultural implications description and simulation of tree-layer composition and size distributions in a primaeval picea-pinus forest above-ground biomass, nutrients, and energy content of trees in a second-growth stand of agathis australis forest succession on landslides in the fiord ecological region, southwestern new zealand forest succession on landslides above lake thompson, fiordland exe: a climatically sensitive model to study climate change and co enrichment effects on forests dryland holocene vegetation history distribution of subfossil forest remains, eastern south island soils in the new zealand landscape. mallinson rendel, in association with the new zealand society of soil science Ü ber den lichtfaktor in den pflanzengesellschaften und seine bedeutung fü r die stoffproduktion meteorological observations community matrix model predictions of future forest composition at russell state forest an introduction to plant demography with special reference to new zealand trees forest dynamics and stand-level dieback in new zealand's nothofagus forests ecology of new zealand nothofagus forests population dynamics of the emergent conifer agathis australis (d. don) lindl. (kauri) in new zealand. ii. seedling population sizes and gap-phase regeneration forest models defined by field measurements: i. the design of a northeastern forest simulator calculating thornthwaite's and mather's aet using an approximating function development of a linked forest productivity-soil process model. oak ridge national laboratory ornl/tm- response of northern forests to co -induced climate change linkages: an individualbased forest ecosystem model a simulation model for the transient effects of climate change on forest landscapes validating models of ecosystem response to global change above ground biomass of mountain beech (nothofagus solandri (hook.f) oerst. var. cliffortioides (hook.f.) poole) in different stands near timber a theory of forest dynamics a review of forest patch models and their application to global change research development of an appalachian forest succession model and its application to assessment of the impact of the chestnut blight transient responses of forests to co -induced climatic change: simulation modelling experiments in eastern north america simulating the role of climate change and species immigration on forest succession forest dynamics and disturbance in a beech/hardwood forest the significance of life history strategies in the developmental history of mixed beech (nothofagus) forests systat manual, version . . spss, chicago structure and dynamics of old growth nothofagus forests in the valdivian andes on the conifer regeneration gap in new zealand: the dynamics of libocedrus bidwillii stands on south island the new zealand beeches: ecology, utilisation and management facets of the distribution of forest vegetation in new zealand vegetation of new zealand pattern and process in the plant community ecology of podocarpus hallii in central otago geomorphology of the central southern alps, new zealand: the interaction of plate collision and atmospheric circulation the hubbard brook ecosystem study: forest biomass and production the gap theory in forest dynamics. the botanical magazine in particular, larry burrows, glenn stewart, robert allen, ian payton, and bruce burns assisted with comments about the sites and forest ecosystems. matt mcglone, glenn stewart, and two unnamed referees kindly reviewed the manuscript and suggested improvements. this project was funded by the foundation for research, science and technology under contract number c . a list of the new zealand forest species selected for input to the linknz forest gap model is available in esa's electronic data archive: ecological archives a - . key: cord- -owzhv xy authors: tkacz, magdalena a.; chromiński, kornel title: advantage of using spherical over cartesian coordinates in the chromosome territories d modeling date: - - journal: computational science - iccs doi: . / - - - - _ sha: doc_id: cord_uid: owzhv xy this paper shows results of chromosome territory modeling in two cases: when the implementation of the algorithm was based on cartesian coordinates and when implementation was made with spherical coordinates. in the article, the summary of measurements of computational times of simulation of chromatin decondensation process (which led to constitute the chromosome territory) was presented. initially, when implementation was made with the use of cartesian coordinates, simulation takes a lot of time to create a model (mean . [sec] with the median . [sec]) and additionally requires restarts of the algorithm, also often exceeds acceptable (given a priori) time for the computational experiment. because of that, authors attempted changing the coordinate system to spherical coordinates (in a few previous projects it leads to improving the efficiency of implementation). after changing the way that d point is represented in d space the time required to make a successful model reduced to the mean . [sec] with a median . [s] (alongside with lowering the number of necessary algorithm restarts) which gives a significant difference in the efficiency of model’s creation. therefore we showed, that a more efficient way for implementation was the usage of spherical coordinates. computational power gives very powerful support in the life sciences today. a lot of experiments can be done -they are cheaper to conduct, their parameters can be easily modified. they are also in most cases reproducible and ethical (no wronging living creatures). according [ ] the term modeling is defined as "to design or imitate forms: make a pattern" or "producing a representation or simulation" and model is defined as "a system of postulates, data, and inferences presented as a mathematical description of an entity or state of affairs". in fact, in a case of insilico experiments more precious would be "computational model" and "computational modeling", but in this paper, it will be referred to in a shorter form. sometimes term "modeling" is used in the context of "running computational model" -but in this paper, it will be referred to as "simulation". for many disciplines creating a model is important -it allows to re-scale (extend or reduce) object, slow down or speed up modeled process, examine almost any aspect of object or process (separating parameters or taking into account a given quantity of parameters). with the use of computers, it is also possible to make visualizations and animations. this paper describes some aspects of the modeling process that occurs in all organisms -precisely speaking occurs in almost every living cell. this process also occurs just right now -in my body while i'm writing this text, as well as in your -when you read this. this is the process of transferring genetic material, dna-during cell division. this process is difficult to examine -we can only observe living cells during a relatively short time. another difficulty here is its microscale -to observe it we have to use microscopes. and, besides of a scalewhen we want to examine the interior of the cell -we have to destroy (and kill) it . . . there are attempts to create an "artificial" [ ] (or "synthetic" [ ] ) cell, but this is not an easy task. to face this up, using "divide and conquer" strategy there are attempts to create models of certain cell components and processes. this paper shows some new knowledge that we discover while trying to model chromosome territories (ct's) being a final result of modeling and simulation chromatin decondensation (cd) process and documents some problems (and the way we took to solve them) to make the working model. some time ago we are asked if we can help in the creation of a probabilistic model of ct's (in short -ct's are the distinct d space occupied by each chromosome after cell division, see also sect. . ). we agreed and something that we supposed to be a project for a few months of work, becomes the true mine of many different problems to be solved. the first one we focused on, was the problem of creating appropriate model of chromatin and the model of the chromatin decondensation process (to be able to implement and simulate this process) in a phase just right after cell division. in eukaryotic cells, genetic material is not stored in a well-known form of a helix because dna strand is too long (and too vulnerable to damage). it is stored as a complex of dna strand and proteins -altogether called chromatin which is being rolled-up in a very sophisticated way [ ] . this allows taking much less space and store dna untangled. probably it also helps in preventing random breaks and changes in dna sequences. researches concerning chromatin organization are important because of its influence on gene transcription [ ] . there are levels of chromatin organization (depending on the level of packing) ( [ , ] ). the two extreme levels of packing are condensed and decondensed ones [ ] . the one -somewhere in between extreme ones, that we are interested in, is called euchromatin. this level of organization is often referred to as "beads on a strand" (see fig. ). the level of chromatin condensation depends on different factors. it can be cell state (during cell cycle) but it is also known that it can be controlled by epigenetic modifications [ ] or biological process [ ] . the risk of dna damage [ ] or modification varies depending on the chromatin condensation level. during the cell division, chromatin fibers condense into structures called chromosomes. in the period between two subsequent divisions, called interphase, chromosomes decondense and occupy -d areas within the nucleus. those distinct areas -called "chromosome territories" (ct's) -are regarded as a major feature of nuclear structure ( [ , ] ). chromosome territories can be visualized directly using in-situ hybridization with fluorescently labeled dna probes that paint specifically individual chromosomes ( [ , ] ). researches concerning ct's are: studying the relationship between the internal architecture of the cell nucleus and crucial intranuclear processes such as regulation of gene expression, gene transcription or dna repair ( [ , , ] ). those studies are related to spatial arrangement, dynamics (motion tracking) [ ] , frequency of genomic translocations ( [ ] ) and even global regulation of the genome [ ] . possibility of making experiments in-silicowould speed up and make some of the experiments easier and cheaper. the euchromatin was the starting point to model chromatin structure for us: we decided to model chromatin (and arms of chromosomes) as a sequence of tangent spheres (fig. ) -visually very similar to euchromatin (see fig. ). because euchromatin is observed as "beads on a strand" and beads (sometimes also called "domains") are its basic structural units, we decided to make a single sphere our basic part of the chromatin chain component (and the basic structural units building up cts). this allows also to make our model scalable -by changing the size of the sphere we can easily change the level of chromatin packing. a sphere can be also easily rendered as graphical primitive in most graphical libraries which were very important to guarantee the possibility of further ct's visualization. our modeling process was very closely related to geometrical, visible objects, because it was very important, that the final models could be visualized -to allow visual comparison with real images from confocal microscopy. we also decide to model the decondensation process by adding tangent spheres around existing ones. this effects in gradually expanding volume of the initial strand of spheres. the process continues until the stop condition was met (volume or size of decondensed chromatin). the computational problem was as follows: starting from the initial (condensed) chromatin model (in a form "beads-on-strand"), consisting of a sequence of mutually tangent spheres find coordinates for next n spheres (where n denotes the size -number of beads of chromatin after decondensation). geometrically it is a problem of finding (x, y, z) being the center of a new sphere with the condition of being tangent to the previous one and not in collision in any other (previously generated). our first goal was to make a fully probabilistic model -that means that we do not add additional conditions like the position of centromeres, telomeres, nucleoplasm stickiness and so on (extending model and making it more "real data-driven" are in our current field of interest and research). the modeled process of decondensation can be somehow regarded as a markov process -the subsequent state i + of decondensation strictly depends on the previous one i. the very basic component of our model was a sphere s((x, y, z), r). this notation should be read as a sphere s with a center in the point that has (x, y, z) coordinates and a radius with the length of r, (r ≥ ). the ordered chain of spheres -makes our model of a chromosome, a set of indexed spheres makes a model of ct. the very general algorithm for ct's modeling is presented in algorithm . line and reflect creating initial chromatin strand, line simulation of decondensation. altogether, they led to the generation of the model of the certain ct. the last step of the algorithm (line ) proved to be the most demanding and challenging, which is described in the next section. in the following section, we document the way we take to successfully made the probabilistic model of ct's. at first, we used the cartesian coordinates (cc). first, the algorithm generates coordinates for the sphere that are denoted as the centromere, and next add to it the next ones until it reaches the (given a-priori) length of arms for certain chromosome. having model of the entire chromosome algorithm draw a id of one of the present spheres s i ((x i , y i , z i ), r) (from those composed the chromosome) and then draws "candidate coordinates": x i+ , y i+ and z i+ for the center for the next sphere. the new coordinates are to be from limited range -not too far current sphere's (as they should be tangent). to allow small flexibility, the ε value to the drawn coordinates was introduced. when we had coordinates drawn, the distance dist(s i , s i+ ) was calculated to check whether a new sphere can be added. the distance was computed by calculating ordinary euclidean distance. if dist(s i , s i+ ) was appropriate, the conditions to not collide with existing elements were checked. if all conditions are met -new sphere were added (for details see [ ] ). there were no problems with the generation of the initial chromatin strand as a sequence of spheres (chromosome). the problem emerges when we tried to simulate the decondensation of chromatin: generation of a model takes a lot of time, and we noticed that sometimes simulation was unsuccessful. we discovered (after log analysis) that the algorithm got stuck trying to find coordinates for s i+ . so, we added additional function that triggers restart of algorithm after unsuccessful attempts for placing s i+ sphere (see algorithm lines - ). if s i+ cannot be placed -algorithm starts over and searches possibility to add s i+ , but for another sphere forming chromosome. the pseudocode for this version of the algorithm is shown in algorithm . in the first step it generates the "candidate coordinates" for s i+ center (algorithm lines - ). thanks to ε a possibility that new sphere could be a little too far, or too close the previous s i . the fine-tuning is made by an additional function that checks the distance from the previous sphere (algorithm lines - ). additional code for stuck detection that triggers restarting computations are in (algorithm lines - ) . new sphere x = previous sphere x ± random( , · r + · ε) new sphere y = previous spherey ± random( , · r + · ε) new sphere z = previous spherez ± random( , · r + · ε); check distance from previous sphere this makes the simulation of cd process long and inefficient, and the result was disappointed: the algorithm got stuck relatively often. the measured number of necessary restarts to complete model creation is shown in table . in one model creation, about spheres should be placed as the tangent ones, so it was easy to asses the number of inefficient searches -they are presented in the last column of table . table showed the time needed to generate one ct model. time was measured in seconds, basic statistics were also given, measurements were made on generated models. that was not a satisfactory result. we had to rethink the way we implement the decondensation of chromatin. we decided to try to add -at first sightadditional computations: shifting (change location) of the center of coordinate systems. then we were able to use the notion of the neighborhood with a fixed radius (inspired by a topology) and use spherical coordinates (see fig. ). we were aware of the fact that shifting the coordinate system takes additional time -but the solution with cc works so bad, that we hope that this approach will work a little better. the result of this change beats our expectations -which is described and documented in the next sections. we decided to try spherical coordinates (sc) [ ] instead of cc (for those, who are not familiar with different coordinate systems we recommend to take a look at [ ] , [ ] ). when we wanted to add sphere s i+ to the certain one s i , we first made a shift of the center of the coordinate system in such a way, that the center of coordinate system was situated in the middle of s i sphere (see fig. ). this let us search for the s i+ by drawing two angles and using just one parameter: r. after switching to the sc, we got rid of the problem of looping the simulation during attempts of finding the location for the s i+ . therefore, the function that restarts ct model creation could be removed. we made measurements -time necessary to generate ct models (equivalent to the time of cd simulation) with shifting coordinate system and using spherical coordinates is presented in table . time of creating ct models decreases significantly in comparison to the use of cc. this had a direct and significant impact on the time of the model creation. new sphere x = previous sphere x + · r · cos(ψ) · sin(φ) new sphere y = previous sphere y + · r · sin(ψ) · sin(φ) new sphere z = previous sphere z + · r · cos(φ); is inside nucleus; to follow the rigor for scientific publications (despite very clear difference between times showed in table and table ) we made an analysis, presented in this section. for the purpose of visual comparison of the times of ct model creation we prepared a boxplot (see fig. ) for general view. in fig. the difference, in general, is easy to notice. there is even no single element of the chart (neither whiskers nor dots (outliers)) that overlaps each other. it is easy to notice a huge difference between computing time (and its stability) in both cases. for the record we made statistical test -the result is presented in table . we calculated the value of the t-test, to confirm that the difference in creation times of model (cc and sc) is statistically significant (p-value below . means that the difference is statistically significant). this proves the statistical significance between modeling time in described two methods. based on presented in this paper results we can conclude that when you model in d space, using spherical coordinates may lead to a more efficient implementation of the algorithm, even when you have to shift the center of the coordinate systems. the solution when using euclidean distance in the cartesian coordinate system in implementation was much more time-consuming. what is more important -it often does not finish modeling process in an acceptable time (sometimes we have to break simulation after weeks of computing on a computer with gb ram and i processor), if it finishes at all (do not got stuck). as future work, knowing that using a spherical coordinate system is helpful we want to examine the effectiveness of quaternion-based implementation as a way to represent coordinates in d space. we also want to check in a more detailed way, what has an impact: only changing the center of the coordinate system, only changing the way of point representation -or both. because it is not the first time when we noticed significant change (in plus) after using spherical (or hyperspherical -in more dimensions) coordinates instead of the cartesian ones, we plan (after finishing actual projects with deadlines) design and conduct a separate experiment. we want to investigate in a more methodological and ordered way to answer the question: why spherical coordinates give better results in computational implementations? our case study also shows that it is possible that geometrical and visual thinking while modeling in d space can be helpful. with the "pure algebraic" thinking (based on the calculation on coordinates) finding the idea -to search in the neighborhood, shifting the center of the coordinate system and next using direction (angles) and fixed distance -would be more difficult (if even possible). synthetic cell project the genome-seeing it clearly now maternal prenatal stress and the developmental origins of mental health. the epigenome and developmental origins of health and disease multiple aspects of gene dysregulation in huntington's disease. front. neurol nucleosome positioning in saccharomyces cerevisiae chromatin condensation modulates access and binding of nuclear proteins activation of dna damage response signaling by condensed chromatin nucleoplasmin regulates chromatin condensation during apoptosis chromosome condensation and decondensation during mitosis chromosome territories how do chromosome territory dynamics affect gene redistribution? https://www. mechanobio.info/genome-regulation/how-do-chromosome-territory-dynamicsaffect-gene-redistribution chromosome territory formation attenuates the translocation potential of cells. elife chromosome territories and the global regulation of the genome chromatin spheres and the interchromatin compartment form structurally defined and functionally interacting nuclear networks inheritance of gene density-related higher order chromatin arrangements in normal and tumor cell nuclei chromosome territories, nuclear architecture and gene regulation in mammalian cells nuclear organization of the genome and the potential for gene regulation fluorescence in situ hybridization with human chromosome-specific libraries: detection of trisomy and translocations of chromosome spatial organization of the mouse genome and its role in recurrent chromosomal translocations cell biology -chromosome territories chromosome territories -a functional nuclear landscape chromosome territories: the arrangement of chromosomes in the nucleus chromosome territory modeler and viewer spherical coordinates. from math insight cartesian, polar, cylindrical, and spherical coordinates key: cord- -zhgjmt j authors: tang, min; xie, qi; gimple, ryan c.; zhong, zheng; tam, trevor; tian, jing; kidwell, reilly l.; wu, qiulian; prager, briana c.; qiu, zhixin; yu, aaron; zhu, zhe; mesci, pinar; jing, hui; schimelman, jacob; wang, pengrui; lee, derrick; lorenzini, michael h.; dixit, deobrat; zhao, linjie; bhargava, shruti; miller, tyler e.; wan, xueyi; tang, jing; sun, bingjie; cravatt, benjamin f.; muotri, alysson r.; chen, shaochen; rich, jeremy n. title: three-dimensional bioprinted glioblastoma microenvironments model cellular dependencies and immune interactions date: - - journal: cell res doi: . /s - - - sha: doc_id: cord_uid: zhgjmt j brain tumors are dynamic complex ecosystems with multiple cell types. to model the brain tumor microenvironment in a reproducible and scalable system, we developed a rapid three-dimensional ( d) bioprinting method to construct clinically relevant biomimetic tissue models. in recurrent glioblastoma, macrophages/microglia prominently contribute to the tumor mass. to parse the function of macrophages in d, we compared the growth of glioblastoma stem cells (gscs) alone or with astrocytes and neural precursor cells in a hyaluronic acid-rich hydrogel, with or without macrophage. bioprinted constructs integrating macrophage recapitulate patient-derived transcriptional profiles predictive of patient survival, maintenance of stemness, invasion, and drug resistance. whole-genome crispr screening with bioprinted complex systems identified unique molecular dependencies in gscs, relative to sphere culture. multicellular bioprinted models serve as a scalable and physiologic platform to interrogate drug sensitivity, cellular crosstalk, invasion, context-specific functional dependencies, as well as immunologic interactions in a species-matched neural environment. brain tumors are complex tissues with multicomponent interactions between multiple cell types. precision medicine efforts based solely on genomic alterations and molecular circuitries driving neoplastic cells have translated into relatively limited benefit in clinical practice for brain cancers, including glioblastoma, the most prevalent and lethal primary intrinsic brain tumor. crosstalk between neoplastic cells and the surrounding stroma contributes to tumor initiation, progression, and metastasis. however, most cancer research studies investigate cancer cells in isolation, cultured in non-physiologic adherent conditions containing species-mismatched serum. massive efforts have interrogated functional dependencies of cancer cell lines. [ ] [ ] [ ] [ ] while these studies provide valuable insights into cancer cell dependencies, they lack the capacity to investigate interactions of cancer cells with stromal cells or the microenvironment in an appropriate physiological context. patient-derived xenografts (pdxs) and genetically engineered mouse models are informative and can better recapitulate the genomic and transcriptomic profiles of patient brain tumors than two-dimensional ( d) culture. however, challenges with engraftment, the low throughput nature of animal experiments, and the lack of normal human cellular interactions, limit their broad applications in clinical settings. in tumors with significant immune cell involvement, such as glioblastoma, pdxs are limited as immunocompromised animals prevent investigation of immune cells in cancer biology. methods to construct self-organizing three-dimensional ( d) coculture systems, termed organoids, have been developed to interrogate physiological and pathophysiological processes. , in cancer research, organoid systems serve as models of colorectal cancer, , breast cancer, , hepatocellular and cholangiocarcinomas, pancreatic cancers, and glioblastomas, among others. , in glioblastoma, we first described organoid systems that recapitulate tumor architecture, microenvironmental gradients, and tumor cellular heterogeneity. additional glioblastoma models utilize human-embryonic stem cell (hesc)-derived cerebral organoids to investigate interactions between glioblastoma stem cells (gscs) and normal brain components including infiltration, microenvironmental stimuli, and response to therapies. however, organoid modeling is labor intensive, relatively low throughput, and highly variable in terms of cellular composition and structure due to the process of self-assembly. further development of tissue engineering approaches informs new d culture systems with improved scalability and capacity to tune specific biological parameters, including cellular composition and extracellular matrix stiffness. the development of physiologically relevant brain tumor microenvironments requires careful consideration of the biophysical and biochemical properties of the matrix and cellular composition of specific tumor types, which can be achieved with recent advances in d bioprinting and biomaterials designed specifically for the bioprinting process. [ ] [ ] [ ] [ ] biocompatible scaffolds for tumor microenvironments include the naturally occurring extracellular matrix products chitosan-alginate (ca) and hyaluronic acid (ha)-based hydrogels, , but also synthetic polymers, including poly lactide-co-glycolide (plga), and polyethylene-glycol (peg), or polyacrylamide hydrogels. d printing with biocompatible materials is emerging to advance the fields of regenerative medicine and tissue modeling, with notable relevance and applicability to cancer research. d bioprinting models microenvironmental interactions and drug sensitivities, reciprocal interactions with macrophages, and patient-specific screening tools in microfluidics-based systems. among many d printing technologies, digital light processing (dlp)-based d bioprinting provides superior scalability and printing speed in addition to versatility and reproducibility. several biomimetic tissue models have been developed using this technology, creating tissue-specific architecture and cellular composition that could be used for functional analyses, metastasis studies, and drug screening. , here, we employ a rapid d bioprinting system and photocrosslinkable native ecm derivatives to create a biomimetic d cancer microenvironment for the highly lethal brain tumor, glioblastoma. the model is comprised of patient-derived gscs, macrophages, astrocytes, and neural stem cells (nscs) in a ha-rich hydrogel. one major microenvironmental feature of glioblastoma is the prominent infiltration of tumor masses by macrophage and microglia. in progressive or recurrent glioblastoma, macrophage and microglia account for a substantial fraction of the tumor bulk. using genetic depletion, co-implantation, and pharmacologic depletion, macrophage/microglia have been shown to be functionally important for glioblastoma growth, but each of these approaches may have broader effects beyond direct tumor cellmacrophage interactions. using our rapid d bioprinting platform, we can interrogate functional dependencies and multicellular interactions in a physiologically relevant manner. dlp-based rapid d bioprinting generates glioblastoma tissue models brain tumors are composed of numerous distinct populations of malignant and supporting stromal cells, and these complex cellular interactions are essential for tumor survival, growth, and progression. glioblastomas display high levels of intratumoral heterogeneity, with contributions from astrocytes, neurons, npcs, macrophage/ microglia, and vascular components. to move beyond serum-free sphere culture-based models, we utilized a dlp-based rapid d bioprinting system to generate d tri-culture or tetra-culture glioblastoma tissue models, with a background "normal brain" made up of npcs and astrocytes and a tumor mass generated by gscs, with or without macrophage, using brain-specific extracellular matrix (ecm) materials (fig. a ). leveraging this system with exquisite control of cellular constituents in specific locations, we selected macrophage for additional study, as we hypothesized that dlp-based d bioprinting could enable precise spatial arrangement of cells and matrix, and selection of any cell type. the key components of the bioprinting system were a digital micromirror device (dmd) chip and a motorized stage where prepolymer cellmaterial mixtures were sequentially loaded. the dmd chip with approximately × micromirrors controlled the light projection of the brain-shaped patterns onto the printing materials (fig. b) . the elliptical pattern corresponded to the core region and the coronal slice pattern corresponded to the peripheral region. each pattern was printed with s of light exposure. in the d tri-culture model, a central tumor core composed of gscs was surrounded by a less dense population of astrocytes and npcs. in the d tetra-culture model, we mixed m macrophages with gscs within the central core to mimic the immune cell infiltrated tumor mass (fig. c) . the ecm composition of the glioblastoma microenvironment was modeled with gelatin methacrylate (gelma) and glycidyl methacrylate-ha (gmha) hydrogels. cells were encapsulated into a material mixture of % gelma (at % degree of methacrylation) and . % gmha (at % degree of methacrylation), which generated a hydrogel matrix that resembled glioblastoma tissue (supplementary information, fig. s a , b). gelma has good biocompatibility and serves as a stiffness modulator that provided desirable mechanical properties and little intervention in biochemical cues. ha is the most abundant ecm component in healthy brain tissue and promotes glioblastoma progression, including regulating glioblastoma invasion through the receptor for hyaluronan-mediated motility (rhamm) and cd , as well as other mechanical and topographical cues. we used a physiologically relevant concentration of ha ( . %) determined from clinical analysis of a diverse population of biopsy specimens from patients with different brain tumors. while a range of molecular weight has are present in the brain, low molecular weight ha promotes gsc stemness and resistance. thus, in this study, low molecular weight ha ( kda) was used to synthesize gmha to model the pro-invasive brain tumor microenvironment. the mechanical properties of the model were characterized by the compressive modulus and pore sizes. the stiffness of the acellular hydrogel remained stable over a week of incubation at °c (data not shown). the stiffness of cell-encapsulated tumor core was . ± . kpa, while the less populated peripheral region containing npcs and astrocytes was . ± . kpa. the peripheral region stiffness was designed to match that of healthy brain tissue reported to be~ kpa. glioblastoma displays enhanced migration and proliferation in stiffer materials. the stiffness of the tumor core was modulated with the light exposure time during printing to have higher modulus than the healthy region. the hydrogel had a porosity of % and an average pore size of μm. with these microscale features, small molecules, such as drug molecules, freely diffuse through the matrix. cells closely interacted with other cells and the matrix (fig. d) . at a macro scale, the model had a thickness of mm, and . mm by . mm in width and length, which allowed gradients of oxygen and nutrition diffusion to be formed within the tissue. cells were precisely printed into two prearranged regions to provide more physiologically relevant features: a non-neoplastic peripheral region composed of npcs and astrocytes surrounding a tumor core composed of either gscs alone or gscs with macrophage ( fig. e) . following optimization for cell density (supplementary information, fig. s a, b) , the tumor core in the d tri-culture consisted of . × gscs/ml, while the tetra-culture tumor core contained . × gscs/ml and . × macrophages/ml. d bioprinted models recapitulate glioblastoma transcriptional profiles traditionally grown cell lines have been extensively characterized in glioblastoma, revealing that these conditions fail to replicate article patient tumors in cellular phenotypes (e.g., invasion) or transcriptional profiles. while patient-derived glioblastoma cells grown under serum-free conditions enrich for stem-like tumor cells (gscs) that form spheres and more closely replicate transcriptional profiles and invasive potential than standard culture conditions, we previously demonstrated that spheres display differential transcriptional profiles and cellular dependencies in an rna interference screen compared to in vivo xenografts. based on this background, we interrogated the transcriptional profiles from a large cohort of patient-derived gscs grown in serum-free, sphere cell culture that we recently reported. gscs grown as spheres were transcriptionally distinct from primary glioblastoma surgical resection tissue specimens, when compared through either principal component analysis (pca) or uniform manifold approximation and projection (umap) (fig. a, b) . to determine whether the d bioprinted culture systems more closely resemble primary glioblastoma tumors, we performed global transcriptional profiling through rna extraction followed by next-generation sequencing (rna-seq) on gscs isolated from the bioprinted models and on gscs in sphere culture (fig. c) . upregulation of a fig. d bioprinting enables generation of glioblastoma tri-culture and tetra-culture tissue environment model. a schematic diagram of in vitro d glioblastoma model containing gscs, macrophages, astrocytes, and neural stem cells (nscs). b schematic diagram of digital micromirror device (dmd) chip-based d bioprinting system used to produce the d glioblastoma model. c diagram of tri-culture (left) and tetra-culture (right) model system. d (left) scanning electron microscope (sem) images of acellular glycidyl methylacrylate-hyaluronic acid and gelatin methacrylate extracellular matrix. (center and right) sem images of the cells encapsulated in the extracellular matrix. scale bars, μm (left), μm (center), and μm (right). e brightfield and immunofluorescence images of the tri-culture and tetra-culture d glioblastoma models. gscs are labeled with green fluorescent protein (gfp) while macrophages are labeled with mcherry. nuclei are stained with dapi. scale bars, mm. core set of glioblastoma tissue-specific genes defined a "glioblastoma tissue" gene signature (fig. d) . when compared to gscs grown in sphere culture, the tetra-culture bioprinted model displayed upregulation of the glioblastoma tissue-specific gene set (fig. e) , suggesting that the bioprinted model recapitulates transcriptional states present in patient-derived glioblastoma tissues. gscs in d tetra-culture displayed upregulation of genes specifically expressed in orthotopic intracranial xenografts (fig. f, g) and, to a lesser extent, genes specifically expressed in subcutaneous flank xenografts (supplementary information, fig. s c ) compared to sphere culture. additionally, signatures that distinguish gscs from their differentiated counterparts were upregulated in the tetraculture system compared to sphere culture (fig. h, i) , suggesting that the physiologic tissue environment promotes stem-like transcriptional states. we further interrogated the gene expression profiles that distinguish gscs grown in sphere culture from the d tetraculture bioprinted models (fig. a) . while cells grown in sphere culture displayed enrichment for gene sets involved in ion transport, protein localization, and vesicle membrane function, cells in the tetra-culture d model displayed transcriptional upregulation of cell adhesion, extracellular matrix, cell and structure morphogenesis, angiogenesis, and hypoxia signatures ( fig. s b ). hypoxia response genes, ca , ndrg , angptl , and egln family members, were upregulated in the tetra-culture system, while various ion transporters, including slc a and slc a , were downregulated (fig. d , e). by qpcr, gscs isolated from either d system days after printing displayed elevated levels of the stemness marker olig and decreased levels of the differentiation markers map and tuj compared to their sphere counterparts grown in parallel (fig. f) . additionally, gsc levels of map and tuj were decreased to a greater degree in tetraculture (i.e., with macrophage) compared to tri-culture. we further evaluated the protein expression of stemness, hypoxia, and proliferative markers in the tetra-culture system compared to sphere culture. the hypoxia marker ca was upregulated in the tetra-culture model compared to sphere culture (fig. g) . the heightened hypoxia level more closely resembled pathologic in vivo conditions, in which the tumor core had a higher hypoxia expression compared to the peripheral region of neurons and astrocytes. in the d culture model, cells also showed increased levels of the proliferative marker ki and increased protein expression of the stemness markers olig and sox ( fig. h-j) . macrophages promote hypoxic and invasive signatures in bioprinted models to understand the relative contributions of each cell type incorporated into bioprinted models, we performed rna-seq on gscs derived from tri-cultures and tetra-cultures. given that thp derived macrophages display distinct expression profiles as primary macrophages, we built tetra-cultures containing thp derived macrophage, human induced pluripotent stem cell (hipsc)-derived macrophage generated from an established protocol, and primary human volunteer-derived macrophage. both hipsc-derived macrophage and primary macrophage integrated into the tetra-culture models. umap clustering revealed that the transcriptional outputs of sphere cultured gscs are distinct from that of gscs in bioprinted models (fig. a, b) . concordantly, we detected differentially expressed genes between sphere cultured cells and any of the bioprinted models ( - differentially expressed genes), while there were fewer genes that distinguished the bioprinted models ( - differentially expressed genes) (fig. c) . bioprinted models were characterized by activation of invasion, extracellular matrix, cell surface interaction, and hypoxia signatures, while gscs in sphere culture expressed cell cycle, dna replication, rna processing, and mitochondrial translation signatures (supplementary information, we next investigated differentially expressed pathways between bioprinted models to interrogate the contributions of cellular components. tri-culture-derived gscs upregulated extracellular matrix and biological adhesion pathways compared to gscs in sphere culture (supplementary information, fig. s a -e). addition of macrophage further increased activation of hypoxia and glycolytic metabolism signatures, with enrichment for invasiveness signatures (fig. d-h) . tetra-cultures constructed with hipsc-derived macrophage expressed higher levels of extracellular matrix and wound healing and platelet activation signatures and decreased levels of neuron and glial development and differentiation pathways compared to tetra-cultures containing thp -derived macrophages (supplementary information, fig. s a , b). incorporation of primary human macrophages did not affect levels of ki or sox compared to use of thp -derived cells (supplementary information, fig. s c , d). consistent with our previous findings, use of hipsc-derived macrophages reduced gsc expression of map and tuj differentiation markers and increased expression of ca and ndrg hypoxia markers (supplementary information, fig. s e ). taken together, gscs upregulate extracellular matrix interaction signatures in response to growth in a bioprinted model. the addition of macrophage further accentuates these gene activation signatures and increases activation of hypoxia and pro-invasive transcriptional profiles. d bioprinted tissues model complex cellular interactions and migration interactions between malignant cells and stromal components shape tumor tissue with each cell type impacting the other tissue components. to understand these changes, we investigated how macrophage responded to the d brain tumor microenvironment by isolating thp -derived macrophages from d bioprinted constructs and performing rna-seq (fig. a, b) . for the d printed tissue, macrophage were mixed with gscs at a : ratio to form the tumor core, while the periphery was formed by astrocytes and npcs using the same composition described previously. the transcriptional output of macrophage grown in traditional culture displayed enrichment for prc complex targets, amino acid biosynthesis, protein metabolism signatures and ribosomal pathways, while macrophage exposed to gscs in the fig. d tetra-culture models better recapitulate transcriptional signatures found in glioblastoma tissues than standard sphere culture. a pca of the global transcriptional landscape of glioma stem cells in culture (gscs in culture, n = ) vs primary glioblastoma surgical resection tissues (gbm tissue, n = ) as defined by rna-seq. the top differential genes were used for the analysis. data was derived from mack et al. b umap of the global transcriptional landscape of glioma stem cells in culture (gscs in culture, n = ) vs primary glioblastoma surgical resection tissues (gbm tissue, n = ) as defined by rna-seq. analysis parameters include: sample size of local neighborhood, number of neighbors = ; learning rate = . ; initialization of low dimensional embedding = random; metrics for computation of distance in high dimensional space = manhattan. data was derived from mack et al. c schematic diagram of experimental approach for gsc rna-seq experiments. d volcano plot of transcriptional landscape profiled by rna-seq comparing gscs in sphere culture (n = ) vs glioblastoma primary surgical resection tissues (n = ). the x-axis depicts the log transformed fold change, while the y-axis shows the log transformed p value adjusted for multiple test correction. e gene set enrichment analysis (gsea) of the glioblastoma tissue vs cell culture signature as defined in d when applied to rna-seq data comparing the d tetra-culture system with sphere cell culture. f volcano plot of transcriptional landscape profiled by rna-seq comparing gscs in sphere culture (n = biological samples with technical replicates each) vs matched orthotopic intracranial xenograft specimens (n = biological samples with technical replicates each). the x-axis depicts the log transformed fold change, while the y-axis shows the log transformed p value adjusted for multiple test correction. data was derived from miller et al. g gsea of the glioblastoma tissue vs cell culture signature as defined in f when applied to rna-seq data comparing the d tetraculture system with sphere cell culture. h volcano plot of transcriptional landscape profiled by rna-seq comparing gscs in sphere culture (n = biological samples with technical replicates each) vs differentiated glioma cells (dgcs) in sphere culture (n = biological samples with technical replicates each). the x-axis depicts the log transformed fold change, while the y-axis shows the log transformed p value adjusted for multiple test correction. data was derived from suva et al. i gsea of the glioblastoma tissue vs cell culture signature as defined in h when applied to rna-seq data comparing the d tetra-culture system with sphere cell culture. bioprinted construct showed elevation of pathways involved in leukocyte activation and innate immune response, cytokine signaling and inflammatory responses, and tlr-stimulated signatures ( fig. c ; supplementary information, fig. s a -d). defense response genes, including ch , pla g , and alox , were upregulated in macrophage derived from the tetra-culture system, while genes involved in amino acid restriction, including il , cd , and vldlr, were downregulated (fig. d , e). m macrophage-related markers were upregulated in the d tetracultures, with cd increased by -fold and il- increased by -fold compared to traditional suspension culture, as measured by qpcr. m -related markers, including tnf-α and nos , did not increase, demonstrating that the d printed microenvironment preferentially polarized macrophage towards the m phenotype ( fig. f ). this is consistent with the m polarization of macrophage in glioblastoma tumors. , gene expression signatures defining peripherally-derived tumor-associated macrophage in glioma , were selectively enriched in macrophage derived from tetraculture models compared to those grown in d culture (supplementary information, fig. s ). collectively, macrophage grown in our d bioprinted tetra-culture model expressed gene expression signatures consistent with patient-derived tumorassociated macrophage. we interrogated the functional consequences of the addition of immune components to the d bioprinted model. in four patientderived gscs spanning three major glioblastoma transcriptional subtypes (proneural, classical, and mesenchymal), the addition of thp -derived m macrophage increased gsc invasion into the surrounding brain-like parenchyma ( fig. g-j) . consistent with our gene expression analyses, m macrophage increased the area of invasion by % for cw , % for gsc , % for gsc , and % for gsc . collectively, these results support the tetraculture model as an effective tool to study cancer cell invasion and the mechanisms by which cellular interactions impinge upon these processes. as numerous stromal compartments, including neural progenitor cells, astrocytes, and neurons, [ ] [ ] [ ] interact with glioblastoma cells within patient tumors, we interrogated the effects of the bioprinted model on neuronal and oligodendrocyte differentiation of the non-neoplastic npcs. in d culture, most npcs expressed the proliferative npc marker sox . the high expression and frequency of sox was retained in tri-cultures and tetra-cultures containing macrophage derived from thp cells or primary human macrophage (supplementary information, fig. s a ). in d culture, npcs expressed the neuronal marker tubb , but retained a progenitor-like cellular morphology. in bioprinted models, npcs adopted a neuronal morphology with the appearance of elongated cellular projections (supplementary information, fig. s b ). expression of map was reduced in npcs in bioprinted models compared to d culture (supplementary information, fig. s a ). olig staining revealed oligodendrocytelike cells in tri-cultures (supplementary information, fig. s b ). taken together, npcs partially differentiate in our bioprinted system, but are unlikely to form mature functional neurons or oligodendrocytes. the d bioprinted model serves as a platform for drug response modeling we next investigated the ability of our d bioprinted constructs to model drug responses and the capacity for cellular interactions within the d bioprinted constructs to affect drug sensitivity of gscs. fluorescent dextran molecules ( kda) modeled drug penetration into d bioprinted models. , dextran molecules rapidly entered bioprinted constructs when the hydrogel was soaked in a dextran solution, with rapid increases in average fluorescence intensity measured from the hydrogel. the fluorescence intensity plateaued after min of incubation and displayed a uniform spatial intensity across the hydrogel, demonstrating that drug compounds can effectively permeate the d bioprinted model (fig. a-c) . egfr is commonly amplified, overexpressed, or mutated in glioblastoma, so we evaluated the treatment efficacy of two egfr inhibitors, erlotinib and gefitinib, and the glioblastoma standardof-care alkylating agent temozolomide in our models. d tricultures and tetra-cultures were cultured for days before drug treatment. despite activated egfr in glioblastomas, egfr inhibitors have shown little benefit for glioblastoma patients. gsc in either d model displayed enhanced resistance to egfr inhibitors and temozolomide compared to sphere culture. inclusion of m macrophage further increased resistance of gsc to egfr inhibitors ( glioblastomas are highly lethal cancers for which current therapy is palliative. , therefore, we explored the potential utility of d bioprinted systems to inform drug responses in glioblastoma. overlaying gene expression data from the d tetraculture model with drug sensitivity and gene expression data from the cancer cell line encyclopedia (ccle) and the cancer therapeutic response platform (ctrp) enabled prediction of drug sensitivity and resistance in our d tetra-culture model based on transcriptional signatures (fig. f) . [ ] [ ] [ ] consistent with our studies of erlotinib, gefitinib, and temozolomide, high expression of genes upregulated in gscs in the d tetra-culture model was predicted to be associated with drug resistance for the majority of compounds across all cancer cell lines tested (fig. g) or when restricted to brain cancer cell lines (supplementary information, fig. s a ). drugs predicted to be ineffective included gsk-j (jmjd /kdm b inhibitor), cytarabine (nucleotide antimetabolite), and decitabine (dna methyltransferase inhibitor), while drugs predicted to be effective included abiraterone (cyp a inhibitor), fig. gscs grown in d tetra-culture models upregulate transcriptional signatures of cellular interaction, hypoxia, and cancer stem cells. a volcano plot of transcriptional landscape profiled by rna-seq comparing the cw gsc grown in standard sphere culture vs gscs in the d tetra-culture model. the x-axis depicts the log transformed fold change, while the y-axis shows the log transformed p value adjusted for multiple test correction. n = technical replicates per condition. b pathway gene set enrichment connectivity diagram displaying pathways enriched among gene sets upregulated (red) and downregulated (blue) in gscs in the d tetra-culture system vs standard sphere culture. c normalized single sample gene set enrichment analysis (ssgsea) scores of glioblastoma transcriptional subtypes as previously defined for the cw gsc when grown in in standard sphere culture vs gscs in the d tetra-culture model. bars are centered at the mean value and error bars represent standard deviation. d mrna expression of representative genes in hypoxia response pathways between standard sphere culture vs gscs in the d tetra-culture model as defined by rna-seq. p values were calculated using deseq with a wald test with benjamini and hochberg correction. ****p < e− . bars are centered at the mean value and error bars represent standard deviation. e mrna expression of representative genes in ion transport pathways between standard sphere culture vs gscs in the d tetra-culture model as defined by rna-seq. p values were calculated using deseq with a wald test with benjamini and hochberg correction. ****p < e− . bars are centered at the mean value and error bars represent standard deviation. f mrna expression of stem cell and differentiation markers between standard sphere culture vs gscs in the d tetra-culture model as defined by quantitative pcr (qpcr). three technical replicates were used and ordinary two-way anova with dunnett multiple comparison test was used for statistical analysis, *p < . ; **p < . ; ***p < . . bars indicate mean, with error bars showing standard deviation. g immunofluorescence staining of ca in cells grown in standard sphere culture (top) vs gscs in the d tetra-culture model (bottom). scale bars, μm. h immunofluorescence staining of ki in cells grown in standard sphere culture (top) vs gscs in the d tetra-culture model (bottom). scale bars, μm. i immunofluorescence staining of olig in cells grown in standard sphere culture (top) vs gscs in the d tetra-culture model (bottom). scale bars, μm. j immunofluorescence staining of sox in cells grown in standard sphere culture (top) vs gscs in the d tetra-culture model (bottom). scale bars, μm. fig. addition of macrophages activates extracellular matrix and invasiveness signatures. a umap analysis of rna-seq data from gscs grown in ( ) sphere culture, ( ) tri-culture, ( ) tetra-culture with thp -derived macrophage, and ( ) tetra-culture with hipsc-derived macrophages. b heatmap displaying mrna expression of differentially expressed genes between conditions. c upset plot showing the number of differentially expressed genes between conditions. for conditions containing sphere cultured cells, genes were considered differentially expressed if the log fold change of mrna expression was greater than . (or < − . ) with an adjusted p value of e− . for other conditions, genes were considered differentially expressed if the log fold change of mrna expression was greater than . (or < − . ) with an adjusted p value of e− . d volcano plot of transcriptional landscapes profiled by rna-seq comparing the cw gsc grown in tetraculture containing thp -derived macrophages vs gscs in the tri-culture model. the x-axis depicts the log transformed fold change, while the y-axis shows the log transformed p value adjusted for multiple test correction. n = technical replicates per condition. e pathway gene set enrichment connectivity diagram displaying pathways enriched among gene sets upregulated (red) and downregulated (orange) in gscs in the d tetra-culture system vs tri-culture system. f gsea of the extracellular matrix structural constituent pathway between tetra-culture and tri-culture models. fdr q value = . . g gsea of the anastassiou multicancer invasiveness pathway between tetra-culture and tri-culture models. fdr q value = . . h gene set enrichment analysis (gsea) of the collagen degradation pathway between tetra-culture and tri-culture models. fdr q value = . . vemurafenib and plx- (raf inhibitors), ml (nrf activator), and ifosfamide (akylating agent) (fig. g-j) . the drug sensitivity predictions were similar, but not entirely overlapping, when a glioblastoma orthotopic xenograft expression signature was used (supplementary information, fig. s b ). investigation of the library of integrated network-based cellular signatures (lincs) dataset showed that compounds predicted to recapitulate the d tetra-culture signature included hypoxia inducible factor activators, caspase activators, and hdac inhibitors, while raf inhibitors and immunosuppressive agents may impair expression of this gene signature (supplementary information, fig. s c ). these findings suggest that interactions with the local microenvironment affect gsc sensitivity to therapeutic compounds and that the d bioprinted tissue model can interrogate these context-dependent effects. further, as the tetra-culture model expresses genes associated with poor sensitivity to a variety of therapeutic compounds, this system may be a more realistic model for drug discovery in glioblastoma. to validate these predictions, we treated gscs with three of the predicted compounds, abiraterone, vemurafenib, and ifosfamide in triculture and tetra-culture bioprinted models. when treated at the sphere culture ic value (supplementary information, fig. s d-f ), gscs in tetra-culture displayed enhanced sensitivity to abiraterone and ifosfamide compared to gscs in tri-culture, while sensitivity to vemurafenib was unchanged ( fig. i-k) . this suggests that abiraterone and ifosfamide may be effective in targeting tetra-culture derived gscs. further validating these findings in an in vivo subcutaneous glioblastoma xenograft model, ifosfamide therapy reduced tumor growth compared to vehicle (supplementary information, figs. s a-c). d bioprinted tissues uncover novel context-dependent essential pathways and serve as a platform for crispr screening given widespread therapeutic resistance in glioblastoma, we leveraged the d bioprinted construct as a discovery platform for glioblastoma dependencies. parallel whole-genome crispr-cas loss-of-function screening was performed in gscs in sphere culture as well as in the d tetra-culture system ( fig. a; supplementary information, fig. s ). functional dependencies segregated gscs based on their method of growth ( fig. b; supplementary information, fig. s f ). guide rnas were enriched (indicating that the targeted gene enhances viability when deleted) or depleted (indicating that the targeted gene reduces cell viability when deleted) in each platform (fig. c, d) . genes essential in each context, as well as pan-essential genes common to both platforms, included core pathways involved in translation, ribosome functions, and rna processing, cell cycle regulation, protein localization, and chromosomes and dna repair ( fig. e; supplementary information, fig. s g, h) . gene hits were stratified to identify context-specific dependencies (fig. f) . genes selectively essential in sphere culture were enriched for cell cycle, endoplasmic reticulum, golgi and glycosylation, lipid metabolism, and response to oxygen pathways. gscs grown in the d tetra-culture model were more dependent on transcription factor activity, cell development and differentiation, nf-κb signaling, and immune regulation pathways (fig. g-k) . thus, the d bioprinted model allowed for interrogation of functional dependencies of brain tumor cells in physiological settings and in combination with stromal fractions and revealed a more complex functional dependency network than that observed in sphere culture. to further validate d bioprinted-specific dependencies, we stratified our whole-genome crispr screening results, selecting genes predicted to be essential in d tetra-culture (fig. a, b) . individual gene knockout in luciferase-labeled gscs of pag , znf , atp h, and rnf a with two independent sgrnas reduced gsc viability in both sphere culture and d tetra-culture models (fig. c-m) . additionally, knockout of pag or znf in gscs delayed the onset of neurological signs in orthotopic glioblastoma xenografts compared to gscs treated with a nontargeting sgrna (fig. n-q) . pag and znf are upregulated at the mrna level in glioblastomas compared to normal brain tissue and high expression is associated with poor patient prognosis in primary glioblastomas from the chinese glioma genome atlas (cgga) dataset, highlighting the clinical relevance of these factors in glioblastoma (supplementary information, fig. s a-d) . taken together, this screening approach has identified novel candidates for future investigation and potential therapeutic development. d bioprinted cultures express transcriptional signatures associated with poor glioblastoma patient prognosis to determine the clinical relevance of the d bioprinted construct, we investigated the transcriptional profiles relative to glioblastoma patients. signatures of genes upregulated either in intracranial orthotopic xenografts or in d tetra-culture compared to sphere culture were elevated in glioblastomas compared to low-grade gliomas in the cancer genome atlas (tcga), cgga, and the rembrandt dataset (fig. a-d) . the d tetra-culture gene signature was elevated in recurrent glioblastomas compared to primary tumors (fig. e) and in the mesenchymal subtype compared to classical or proneural glioblastomas (fig. f) . in the tcga and cgga datasets, the orthotopic xenograft signature and the d tetra-culture signature were associated with poor glioblastoma patient prognosis (fig. g-j) . many genes with individual poor prognostic significance were upregulated in the intracranial xenograft signature, including chi l , postn, and ndrg (fig. k) , while dennd a, maob, and igfbp were upregulated in the d bioprinted cultures (fig. l) . genes with poor prognostic significance were enriched among all genes in the d tetra-culture signature, when compared to a background of all genes (fig. m) . thus, d bioprinting enabled investigation of gene pathways associated with more aggressive glioblastomas, suggesting that this model can serve as a more realistic therapeutic discovery platform for the most lethal classes of glioblastoma. fig. macrophages grown in d tetra-culture models upregulate immune activation signatures, increase m polarization, and promote gsc invasion. a schematic diagram of experimental approach for macrophage rna-seq experiments. b volcano plot of transcriptional landscape profiled by rna-seq comparing macrophages grown in standard sphere culture vs macrophages in the d tetra-culture model. the x-axis depicts the log transformed fold change, while the y-axis shows the log transformed p value adjusted for multiple test correction. c pathway gene set enrichment connectivity diagram displaying pathways enriched among gene sets upregulated (red) and downregulated (blue) in macrophages in the d tetra-culture system vs standard sphere culture. d mrna expression of representative genes in defense response and macrophage function pathways between standard sphere culture vs macrophages in the d tetra-culture model as defined by rna-seq. p values were calculated using deseq with a wald test with benjamini and hochberg correction. ****p < e− . bars are centered at the mean value and error bars represent standard deviation. e mrna expression of representative genes in amino acid deprivation pathways between standard sphere culture vs macrophages in the d tetra-culture model as defined by rna-seq. p values were calculated using deseq with a wald test with benjamini and hochberg correction. ****p < e− . bars are centered at the mean value and error bars represent standard deviation. f mrna expression of m and m macrophage polarization markers between standard sphere culture vs macrophages in the d tetra-culture model as defined by qpcr. three technical replicates were used and ordinary two-way anova with dunnett multiple comparison test was used for statistical analysis, ***p < . ; ****p < . . bars indicate mean, with error bars showing standard deviation. g fluorescence imaging of cw gscs (green) and macrophages (red) grown in the d tri-culture model without macrophages (top) vs the d tetra-culture model with macrophages (bottom). scale bars, mm. h fluorescence imaging of gscs (green) and macrophages (red) grown in the d tri-culture model without macrophages (top) vs the d tetra-culture model with macrophages (bottom). scale bars, mm. i fluorescence imaging of gsc gscs (green) and macrophages (red) grown in the d tri-culture model without macrophages (top) vs the d tetra-culture model with macrophages (bottom). scale bars, mm. j fluorescence imaging of gscs (green) and macrophages (red) grown in the d tri-culture model without macrophages (top) vs the d tetra-culture model with macrophages (bottom). scale bars, mm. to improve modeling of a highly lethal brain cancer for which current therapies are limited, we utilized a dlp-based d bioprinting system to model glioblastoma, the most common and highly lethal type of brain tumor. studies have reported using d printing to create coculture models of glioblastoma cells with other stromal cells or fabricate ha-based hydrogel to mimic brain ecm. , , however, most prior models focus on only one aspect of the in vivo situation or used non-human cells, which reduced their capacity to be applied to actual clinical settings. to the best article of our knowledge, this is the first report of a human cell-based d glioblastoma model that recapitulates the complex tumor microenvironment with inclusion of normal brain, immune components, stromal components, and essential mechanical and biochemical cues from the extracellular matrix. the tumor microenvironment provides essential signals to guide tumor growth and survival; however, these cues are inefficiently modeled in standard d culture, even in the absence of serum. hypoxic signaling contributes to glioblastoma aggressiveness by remodeling gsc phenotypes. , our d tetra-culture brain tumor model expressed hypoxia response signatures, allowing for investigation of hypoxic signaling in a physiologic environment, unlike standard cell culture systems. critical growth factor signaling elements are provided from neurons, [ ] [ ] [ ] , npcs, ecm components, , and immune fractions, including macrophages. , the perivascular niche provides a variety of signals including wnts, ephrins, and osteopontins to promote glioblastoma invasion, growth, and maintenance of gscs. future studies will be required to integrate vascular components into the d printed model system to further study these important components of the brain tumor microenvironment. the d tetra-culture tissue environment presented here enables controlled, reproducible, and scalable interrogation of these various cellular interactions that drive brain tumor biology. while microenvironmental components supply critical niche factors to sustain the tumor ecosystem, stromal elements are also actively remodeled by malignant cells. here, we observed the role of immune cells in glioblastoma growth, including changes in gene expression, invasive behaviors, and response to treatments. reciprocally, we also find that the d glioblastoma microenvironment promoted polarization of macrophages towards a protumoral m macrophage phenotype, highlighting this bidirectional crosstalk. the bioprinting approach generates a spatially separated tumor region and surrounding non-neoplastic neural tissue with defined cell density which allows the cells to interact in a more realistic manner, providing a highly reproducible platform for the interrogation of cell-cell interactions with several key advantages. first, this d glioblastoma tissue model allows for investigation of tumor-immune interactions in a fully human species-matched system, which is not possible in xenograft or genetically engineered mouse model. this may facilitate understanding of human-specific immune interactions and advance the field of neuro-oncoimmunology by providing insights into immunotherapy efficacy. second, combining tumoral and non-neoplastic neural components within one model will propel drug discovery efforts by enabling measurements of therapeutic efficacy, toxicities, and therapeutic index. the scalability and reproducibility of this d bioprinted model also allows for more high-throughput compound screening efforts. our findings suggest that the d bioprinted model displays transcriptional signatures closer to patient-derived glioblastoma tissue, and that local stromal interactions present within our model promotes broad therapeutic resistance, enabling compound discovery efforts in a challenging environment. third, the d bioprinted model is amenable to largescale whole-genome crispr-cas -based screening methods to uncover novel functional dependencies in a physiologic setting. this model extends previous approaches by characterizing context-dependent target essentiality in cancer cells and allowing for investigation of multivalent stromal cell dependencies. in conclusion, we report a controlled, reproducible, and scalable d engineered glioblastoma tissue construct that serves as a more physiologically accurate brain tumor model, facilitates interrogation of the multicellular interactions that drive brain tumor biology, and acts as a platform for discovery of novel functional dependencies. gelma and gmha synthesis and characterization gelma and gmha were synthesized using type a, gel strength gelatin from porcine skin (sigma aldrich cat #: g ) and , da hyaluronic acid (lifecore), respectively, as described previously. , briefly, for the gelma synthesis of % degree of methacrylation, % (w/v) gelatin was dissolved in . m : carbonate-bicarbonate buffer solution (ph~ ) at °c. methacrylic anhydride was added dropwise at a volume of . ml/(gram gelatin). the reaction was left to run for h at °c. after synthesis, the solutions were dialyzed, frozen overnight at − °c, and lyophilized. freeze-dried gelma and gmha were stored at − °c and reconstituted immediately before printing to stock solutions of % (w/vol) and % (w/vol), respectively. all materials were sterilized by syringe filters before mixing with cells (millipore). the degree of methacrylation of gelma and gmha were quantified using proton nmr (bruker, mhz). cell culture xenografted tumors were dissociated using a papain dissociation system according to the manufacturer's instructions. gscs were then cultured in neurobasal medium supplemented with % b , fig. d bioprinting enables a drug discovery platform and microenvironmental interactions contribute to drug resistance. a (top) schematic diagram of drug diffusion experiment. (bottom) images of fitc-dextran diffusion through the d hydrogel over a time course. scale bars, mm. b average intensity of fitc-dextran signal through the d tetra-culture model over a time course. three replicates were used. bars indicate mean with error bars showing standard deviation. ordinary one-way anova with tukey correction for multiple comparisons was used for statistical analysis. c spatial intensity of fitc-dextran signal through the d tetra-culture model over a time course. d cell viability of the gsc gsc following treatment with the egfr inhibitors, erlotinib and gefitinib, and the alkylating agent temozolomide (tmz) in standard sphere culture conditions, the d tri-culture model, and the d tetra-culture model. three replicates were used, ordinary two-way anova with dunnett multiple test correction was used for statistical analysis. bars indicate mean, while error bars show standard deviation. **p < . ; ****p < . . e cell viability of the cw gsc following treatment with the egfr inhibitors, erlotinib and gefitinib, and the alkylating agent tmz in standard sphere culture conditions, the d tri-culture model, and the d tetra-culture model. three replicates were used, ordinary two-way anova with dunnett multiple test correction was used for statistical analysis. bars indicate mean, while error bars show standard deviation. **p < . ; ***p < . ; ****p < . . f schematic diagram of process to determine drug sensitivity based on the d tetra-culture gene expression signature from the ccle and ctrp datasets. [ ] [ ] [ ] g therapeutic efficacy prediction of drugs in all cancer cells in the ctrp dataset based on differentially expressed genes between the d tetra-culture model and gscs grown in sphere culture as defined by rna-seq. h correlation of (top) abiraterone and (bottom) gsk-j sensitivities based on the d tetra-culture signature expression across all cancer cell lines in the ccle dataset. compounds are ranked based on the correlation between the tetra-culture gene expression signature and compound area under the curve (auc). i normalized cell viability of gscs in tri-culture and tetra-culture models following treatment with μm of abiraterone. ***p < . . bar shows mean of six technical replicates and error bars indicate standard deviation. unpaired two-tailed t-test was used for statistical analysis. j normalized cell viability of gscs in tri-culture and tetra-culture models following treatment with μm of vemurafenib. ns, p > . . bar shows mean of six technical replicates and error bars indicate standard deviation. unpaired two-tailed t-test was used for statistical analysis. k normalized cell viability of gscs in tri-culture and tetra-culture models following treatment with μm of ifosfamide. ***p < . . bar shows mean of six technical replicates and error bars indicate standard deviation. unpaired two-tailed t-test was used for statistical analysis. % l-glutamine, % sodium pyruvate, % penicillin/streptomycin, ng/ml basic human fibroblast growth factor (bfgf), and ng/ ml human epidermal growth factor (egf) for at least h to recover expression of surface antigens. gsc phenotypes were validated by expression of stem cell markers (sox and olig ) functional assays of self-renewal (serial neurosphere passage), and tumor propagation using in vivo limiting dilution. thp- monocytes were cultured in rpmi (gibco) medium supplemented with % heat-inactivated fetal bovine serum (fbs, invitrogen) and % penicillin/streptomycin. to obtain monocytederived m macrophage, thp- monocytes were first seeded in well plates at a density of × cells/ml ( ml/well). polarization to m macrophage was induced by ( ) incubating cells in ng/ ml phorbol -myristate -acetate (pma, sigma aldrich) for h, ( ) replacing with thp complete medium for h, and then ( ) incubating in ng/ml interleukin (il , peprotech) and ng/ ml interleukin (il , peprotech) for h. hnp neural progenitor cells (neuromics) were cultured on matrigel-coated plates using the complete nbm medium for gscs. human astrocytes (thermofisher) were cultured with astrocyte medium (sciencell) supplemented with % penicillin/streptomycin. d bioprinting process before printing, gscs, hnp s, and astrocytes were digested by accutase (stemcell technology), and macrophages were digested with tryple (thermofisher). for the d tetra-culture samples, the cell suspension solution for the tumor core consisted of . × cells/ml gscs and . × cells/ml macrophages (gscs:m = : ). for the d tri-culture samples, the core cell suspension solution consisted of . × cells/ml gscs only (supplementary information, fig. s a, b) . the cell suspension solution for the peripheral region for both models consisted of × cells/ml hnp s and × cells/ml astrocytes. all cell suspensions were aliquoted into . ml eppendorf tubes and stored on ice before use. the prepolymer solution for bioprinting was prepared with % (w/v) gelma, . % (w/v) gmha, and . % (w/v) lithium phenyl( , , -trimethylbenzoyl) phosphinate (lap) (tokyo chemical industry). prepolymer solution was kept at °c in dark before use. cell suspension was mixed with prepolymer solution at : ratio immediately before printing to maximize viability. the two-step bioprinting process utilized a customized lightbased d printing system. components of the system included a digital micromirror device (dmd) chip (texas instruments), a motion controller (newport), a light source (hamamatsu), a printing stage, and a computer with software to coordinate all the other components. the thickness of the printed samples was precisely controlled by the motion controller and the stage. cellmaterial mixture was loaded onto the printing stage, and the corresponding digital mask was input onto the dmd chip. light was turned on for an optimized amount of exposure time ( s for the core and s for the periphery). the bioprinted d tri-culture/ tetra-culture samples were then rinsed with dpbs and cultured in maintenance medium at °c with % co . maintenance medium was made of % of complete nbm medium, % of thp medium, and % of astrocyte medium. hipsc-derived macrophage generation hipsc-derived macrophage differentiation protocol was adapted from yanagimachi et al. and modified from mesci et al. briefly, ipsc cell lines were generated as previously described, by reprogramming fibroblast from a healthy donor. the ipsc colonies were plated on matrigel-coated (bd biosciences) plates and maintained in mtesr media (stem cell technologies). the protocol of myeloid cell lineage consisted of sequential steps. in the first step, primitive streak cells were induced by bmp addition, which in step , were differentiated into hemangioblast-like hematopoietic precursors (vegf ( ng/ml, peprotech), scf ( ng/ml, gemini) and basic fibroblast growth factor (bfgf), ( ng/ml, life technologies)). then, in the third step, the hematopoietic precursors were pushed towards myeloid differentiation (flt- ligand ( ng/ml, humanzyme), il- ( ng/ml, gemini), scf ( ng/ml, gemini), thrombopoietin, tpo ( ng/ml), m-csf ( ng/ml)) and finally into the monocytic lineage in step [flt -ligand ( ng/ml), m-csf ( ng/ml), gm-csf ( ng/ml)]. cells produced in suspension in step were recovered, sorted by using anti-cd magnetic microbeads (macs, miltenyi) and then integrated into d bioprinted models as described above. isolation and generation of primary human macrophages human blood was obtained from healthy volunteers from the scripps research institute normal blood donor service. mononuclear cells were isolated by gradient centrifugation using lymphoprep (# stemcell), washed with pbs, and treated with red blood cell lysis buffer. cells were plated to adhere monocytes and cultured in % heat inactivated fbs in rpmi with hepes, glutamax, mm sodium pyruvate, and pen/strep with ng/ml m-csf for days as described by ogasawara et al. unpolarized m macrophages were collected and integrated into d bioprinted models as described above. mechanical testing compressive modulus of the d printed constructs was measured with a microsquisher (cellscale). pillars with mm in diameter and mm in height were printed with same conditions used for the tissue models and incubated overnight at °c. both acellular and cell-encapsulated constructs were tested. the microsquisher utilized stainless steel beams and platens to compress the constructs at % displacement of their height. customized matlab scripts were used to calculate the modulus from the force and displacement data collected by microsquisher. sem surface patterns of the materials and cell-material interactions on micron-scale were imaged with a scanning electron microscope (zeiss sigma ). acellular samples were snapfrozen in liquid nitrogen and immediately transferred to the freeze drier to dry overnight. cell-encapsulated samples were dried based on a chemical dehydration protocol. briefly, samples were fixed using . % glutaraldehyde solution for h at room temperature and then overnight at °c. on the next day, the samples were rinsed with dpbs for three times and soaked in % ethanol, % ethanol, and % ethanol subsequently, each for min. then the solution was replaced with % ethanol for min, and the step was repeated two more times. hexamethyldisilazane (hdms) was mixed with % ethanol at : ratio and : ratio. samples were first transferred to hdms: fig. whole-genome crispr-cas screen reveals context-specific functional dependencies. a schematic diagram of whole-genome crispr-cas loss-of-function screening strategy in standard sphere culture conditions and the d tetra-culture model. b pca of functional dependencies defined by whole genome crispr-cas screening as defined in (a). c volcano plot demonstrating genes that enhance (blue) or inhibit (red) cell proliferation in sphere culture when inactivated by a specific sgrna in a whole genome crispr-cas loss-of-function screen. the x-axis displays the z-score and the y-axis displays the p value as calculated by the mageck-vispr algorithm. d volcano plot demonstrating genes that enhance (blue) or inhibit (red) cell proliferation in the d tetra-culture model when inactivated by a specific sgrna in a whole genome crispr-cas loss-of-function screen. the x-axis displays the z-score and the y-axis displays the p value as calculated by the mageck-vispr algorithm. e pathway gene set enrichment connectivity diagram displaying pathways enriched among functional dependency genes common to both sphere culture and d culture in the tetra-culture model. f plot comparing the functional dependency zscores between sphere culture and d culture in the tetra-culture model. g pathway gene set enrichment connectivity diagram displaying pathways enriched among functional dependency genes that are specific to sphere culture, as defined in f. h pathway gene set enrichment connectivity diagram displaying pathways enriched among functional dependency genes that are specific to growth in the d tetra-culture, as defined in f. i volcano plot displaying differential functional dependency scores between sphere culture and the d tetra-culture system as defined by mageck-vispr. j pathway gene set enrichment connectivity diagram displaying pathways enriched among functional dependency genes that are more essential in sphere culture compared to in the d tetra-culture system, as defined in i. k pathway gene set enrichment connectivity diagram displaying pathways enriched among functional dependency genes that are more essential in the d tetraculture system compared to in sphere culture, as defined in i. etoh ( : ) for min, then hdms:etoh ( : ) for min. then the solution was replaced with % hdms for min, and the step was repeated two more times. the samples were left uncovered in chemical hood overnight to dry. the freeze-dried or chemically dried samples were coated with iridium by a sputter coater (emitech) prior to sem imaging. immunofluorescence staining and image acquisition of tumor model d bioprinted samples and sphere cultured cells were fixed with % paraformaldehyde (pfa; wako) for min and min, respectively, at room temperature. all samples were blocked and permeabilized using % (w/v) bovine serum albumin (bsa, gemini bio-products) solution with . % triton x- (promega) for h at room temperature on a shaker. samples were then incubated with the respective primary antibody (listed below) overnight at °c. on the next day, samples were rinsed by dpbs with . % tween (pbst) for three times on the shaker. samples were incubated with fluorophore-conjugated goat antirabbit or goat anti mouse secondary antibodies ( : ; biotium) and hoechst ( : ; life technologies) counterstain in dpbs with % (w/v) bsa for h at room temperature in dark. after incubation, samples were rinsed three times in pbst and stored in dpbs with . % sodium azide (alfa aesar) at °c before imaging. fluorescence images of d samples and their sphere cultured counterparts were taken with a confocal microscope (leica sp ) using consistent settings for each antibody (supplementary information, table s ). fluorescence images of egfp-or mcherry-abeled cells in the d samples were also acquired using the confocal microscope. tile scan merging was completed by the automated program on the leica microscope and the z-stack projection was completed by imagej. quantification of the migration was based on the fluorescence images processed by imagej. rna isolation and rt-pcr egfp-labeled gscs and mcherry-labeled thp s were isolated from d printed tri-culture and tetra-culture samples using flow cytometry (bd facsaria ii). cells isolated from d and sphere cultured cells were treated with trizol reagent (life technologies) before rna extraction. total rna of each sample was extracted using direct-zol rna miniprep kit (zymo) and immediately stored at − °c. to perform rt-pcr, cdna was first obtained by rna reverse transcription using the protoscript® first strand cdna synthesis kit (new england biolabs) with input rna of ng per sample. the primers were purchased from integrated dna technologies. rt-pcr was performed using powerup sybr green master mix (applied biosystems) and detected with quantstudio rt-pcr system. gene expression was determined by the threshold cycle (ct) values normalized against the housekeeping gene (supplementary information, table s ). rna-seq and data analysis rna was purified as described above and subjected to rna-seq. paired-end fastq sequencing reads were trimmed using trim galore version . . (https://www.bioinformatics.babraham.ac.uk/ projects/trim_galore/) using cutadapt version . . transcript quantification was performed using salmon version . . in the quasi-mapping mode from transcripts derived from human gencode release (grch . ). salmon "quant" files were converted using tximport (https://bioconductor.org/packages/ release/bioc/html/tximport.html) and differential expression analysis was performed using deseq in the r programming language. data from gscs and primary glioblastoma surgical resection tissues were derived from mack et al. and were processed using the same analysis pipeline. data from matched gscs grown in serum-free sphere culture and orthotopic intracranial xenografts were derived from miller et al. and were processed using the same analysis pipeline. processed data from matched gscs and differentiated tumor cells were derived from suva et al. and differentially expressed genes were calculated using the limma-voom algorithm in the limma package in the r programming language. pca was performed within the deseq package using the top differentially expressed genes. umap analysis was performed using the umapr package (https://github.com/ropenscilabs/umapr) and uwot (https://cran.r-project.org/web/packages/uwot/index. html). for comparisons of glioblastoma tissue samples with gscs grown in standard sphere culture, analysis parameters include: sample size of local neighborhood, number of neighbors = ; learning rate = . ; initialization of low dimensional embedding = random; metrics for computation of distance in high dimensional space = manhattan. for comparisons of gscs derived from sphere culture or d bioprinted models, analysis parameters include: sample size of local neighborhood, number of neighbors = ; initialization of low dimensional embedding = random; metrics for computation of distance in high dimensional space = cosine. gene set enrichment analysis was performed using the online gsea webportal (http://software.broadinstitute.org/ gsea/msigdb/annotate.jsp) and the gsea desktop application fig. pag and znf are potential therapeutic targets in glioblastoma. a d tetra-culture specific target identification approach. graph showing gene dependency z-score in sphere culture (x-axis) vs tetra-culture (y-axis). red color indicates genes with a sphere culture z-score of > − . and a tetra-culture z-score of < − . . b red genes from (a) ranked based on the dependency significance in tetra-culture models (−log of the p value). c luminescent signal in gscs transfected with a luciferase expression vector (red) or un-transfected cells following treatment with luciferin reagent for min. ***p < . . unpaired, two-tailed t test was used for statistical analysis. d western blot for pag and flag-tagged cas following treatment with two independent sgrnas targeting pag in luciferase-expressing cw cells or a nontargeting control (sgcont). tubulin was used as a loading control. e western blot for znf and flag-tagged cas following treatment with two independent sgrnas targeting znf in luciferase expressing cw cells or a sgcont. tubulin was used as a loading control. f western blot for atp h (atp pd) and flag-tagged cas following treatment with two independent sgrnas targeting atp h in luciferase expressing cw cells or a sgcont. tubulin was used as a loading control. g western blot for rnf a and flag-tagged cas following treatment with two independent sgrnas targeting rnf a in luciferase expressing cw cells or a sgcont. tubulin was used as a loading control. h cell viability of cw luciferase expressing gscs in sphere culture following treatment with two independent sgrnas targeting pag or a sgcont. ****p < . . two-way repeated measures anova with dunnett multiple comparison testing was used for statistical analysis. i cell viability of cw luciferase expressing gscs in sphere culture following treatment with two independent sgrnas targeting znf or a sgcont. ****p < . . two-way repeated measures anova with dunnett multiple comparison testing was used for statistical analysis. j cell viability of cw luciferase-expressing gscs in sphere culture following treatment with two independent sgrnas targeting atp h or a sgcont. ****p < . . two-way repeated measures anova with dunnett multiple comparison testing was used for statistical analysis. k cell viability of cw -luciferase expressing gscs in sphere culture following treatment with two independent sgrnas targeting rnf a or a sgcont. ****p < . . two-way repeated measures anova with dunnett multiple comparison testing was used for statistical analysis. l cell viability of cw luciferase expressing gscs in d tetra-culture models after editing with two independent sgrnas targeting pag , znf , or a non-targeting sgrna after seven days. ****p < . . bars show mean and standard deviation of two biological replicates with technical replicates. ordinary one-way anova with dunnett multiple comparison correction was used for statistical analysis. m cell viability of cw luciferase expressing gscs in d tetra-culture models after editing with two independent sgrnas targeting atp h, rnf a, or a nontargeting sgrna after seven days. *p < . ; **p < . . bars show mean and standard deviation of two biological replicates with technical replicates. ordinary one-way anova with dunnett multiple comparison correction was used for statistical analysis. n western blot for pag and flag-tagged cas following treatment with two independent sgrnas targeting pag in cw gscs or a sgcont. tubulin was used as a loading control. o kaplan-meier plot showing mouse survival following orthotopic implantation of gscs edited with one of two sgrnas targeting pag or a sgcont. sgpag . vs sgcont, p = . . sgpag . vs sgcont = . . log-rank test was used for statistical analysis. p western blot for znf and flag-tagged cas following treatment with two independent sgrnas targeting znf in cw gscs or a sgcont. tubulin was used as a loading control. q kaplan-meier plot showing mouse survival following orthotopic implantation of gscs edited with one of two sgrnas targeting znf or a sgcont. sgznf . vs sgcont, p = . . sgznf . vs sgcont, p > . . log-rank test was used for statistical analysis. (http://software.broadinstitute.org/gsea/downloads.jsp). , pathway enrichment bubble plots were generated using the bader lab enrichment map application and cytoscape (http://www. cytoscape.org). glioblastoma transcriptional subtypes were calculated using a program written by wang et al. and implemented in r. gene signatures were calculated using the single sample gene set enrichment analysis projection (ssgseaprojection) module on genepattern (https://cloud.genepattern.org). crispr editing crispr editing was performed on cw gscs as well as luciferase-labeled cw gscs (cw -luc). for unlabeled cells, sgrnas were cloned into the lenticrisprv plasmid containing a puromycin selection marker (addgene plasmid # ), while luciferase-labeled cells were edited with sgrnas cloned into the lenticrisprv plasmid containing a hygromycin selection marker (addgene plasmid # ). sgrna sequences were chosen from the human crispr knockout pooled library (brunello) (supplementary information, table s ). western blot analysis cells were collected and lysed in ripa buffer ( mm tris-hcl, ph . ; mm nacl; . % np- ; mm naf with protease inhibitors) and incubated on ice for min. lysates were centrifuged at °c for min at , rpm, and supernatant was collected. the pierce bca protein assay kit (thermo scientific) was utilized for determination of protein concentration. equal amounts of protein samples were mixed with sds laemmli loading buffer, boiled for min, and electrophoresed using nupage bis-tris gels, then transferred onto pvdf membranes. tbs-t supplemented with % non-fat dry milk was used for blocking for a period of h followed by blotting with primary antibodies at °c for h (supplementary information, table s ). blots were washed times for min each with tbs-t and then incubated with appropriate secondary antibodies in % non-fat milk in tbs-t for h. for all western immunoblot experiments, blots were imaged using biorad image lab software and subsequently processed using adobe illustrator to create the figures. molecular diffusion assessment d printed hydrogels were printed and incubated in dpbs overnight at °c. fluorescein isothiocyanate (fitc)-dextran with average molecular weight of da was dissolved in dpbs at concentration of µg/ml. dpbs was removed and fitc-dextran solutions were added to the wells with d printed hydrogels. hydrogels were incubated in fitc-dextran solution at °c for , , , , , and min; rinsed three times with dpbs; and then imaged using a fluorescence microscope. fluorescence intensities of the hydrogel were measured by imagej. the average intensities and the spatial intensities at each time point were calculated in excel and plotted using prism. drug response assessment d tri-culture/tetra-culture samples were printed as described above, with regular gscs substituted with luciferase-labeled gscs. d samples and sphere cultured cells plated on matrigel-coated slides were treated with drugs after days in culture. drug effects were evaluated h later for erlotinib and gefitinib. for temozolomide, medium was replaced with fresh medium with temozolomide h after first treatment, and the drug response was evaluated h after second treatment. luciferase readings were obtained using using the promega luciferase assay system (e ) based on the provided protocol and a tecan infinite m plate reader. abiraterone (hy- ), vemurafenib (hy- ), and ifosfamide (hy- ), erlotinib (hy- ), and gefitinib (hy- ) from medchemexpress was used to generate dose response curves in vitro. fig. d bioprinting contributes to upregulation of genes with poor prognostic significance in glioblastoma. a heatmap displaying mrna expression signatures of intracranial xenografts (vs sphere cell culture) and d bioprinted tetra-cultures (vs sphere cell culture) as defined by the tcga glioma hg-u a microarray. various clinical metrics, patient information and information on tumor genetics are also displayed. b mrna expression signature of (left) d bioprinted tetra-cultures (vs sphere cell culture) and (right) intracranial xenografts (vs sphere cell culture) in tcga glioma hg-u a microarray. grade ii (n = ), grade iii (n = ), grade iv (n = ). the box-and-whisker plot indicates the lower quartile, median, and upper quartile. error bars represent the %- % confidence interval. ordinary one-way anova with tukey multiple comparison test was used for statistical analysis, ****p < . . c mrna expression signature of d bioprinted tetra-cultures (vs sphere cell culture) in cgga. grade ii (n = ), grade iii (n = ), grade iv (n = ). the box-and-whisker plot indicates the lower quartile, median, and upper quartile. error bars represent the %- % confidence interval. ordinary one-way anova with tukey multiple comparison test was used for statistical analysis, ****p < . . d mrna expression signature of d bioprinted tetra-cultures (vs sphere cell culture) in the rembrandt glioma dataset. grade ii (n = ), grade iii (n = ), grade iv (n = ). the box-and-whisker plot indicates the lower quartile, median, and upper quartile. error bars represent the %- % confidence interval. ordinary one-way anova with tukey multiple comparison test was used for statistical analysis, ****p < . . e mrna expression signature of d bioprinted tetra-cultures (vs sphere cell culture) in the chinese glioma genome atlas (cgga). data presented is restricted to glioblastomas (grade iv glioma). primary (n = ), recurrent (n = ). the box-and-whisker plot indicates the lower quartile, median, and upper quartile. error bars represent the %- % confidence interval. ordinary one-way anova with tukey multiple comparison test was used for statistical analysis, ****p < . . f mrna expression signature of d bioprinted tetra-cultures (vs sphere cell culture) in the rembrandt glioma dataset. data presented is restricted to glioblastomas (grade iv glioma). proneural (n = ), mesenchymal (n = ), classical iv (n = ). the box-and-whisker plot indicates the lower quartile, median, and upper quartile. error bars represent the %- % confidence interval. ordinary one-way anova with tukey multiple comparison test was used for statistical analysis, ****p < . . g kaplan-meier survival analysis of glioblastoma patients in the tcga dataset based on the mrna expression signature of intracranial xenografts (vs sphere cell culture). patients were grouped into "high" or "low" signature expression groups based on the median signature expression score. low (n = ), high (n = ). log rank analysis was used for statistical analysis, p = . . h kaplan-meier survival analysis of glioblastoma patients in the tcga dataset based on the mrna expression signature of d bioprinted tetra-cultures (vs sphere cell culture). patients were grouped into "high" or "low" signature expression groups based on the median signature expression score. low (n = ), high (n = ). log rank analysis was used for statistical analysis, p = . . i kaplan-meier survival analysis of glioblastoma patients in the cgga dataset based on the mrna expression signature of intracranial xenografts (vs sphere cell culture). patients in the top / of the expression signature score were grouped into the "high" group, while those in the bottom / of the expression signature score were grouped into the "low" group. low (n = ), high (n = ). log rank analysis was used for statistical analysis, p = . . j kaplan-meier survival analysis of glioblastoma patients in the cgga dataset based on the mrna expression signature of d bioprinted tetra-cultures (vs sphere cell culture). patients in the top / of the expression signature score were grouped into the "high" group, while those in the bottom / of the expression signature score were grouped into the "low" group. low (n = ), high (n = ). log rank analysis was used for statistical analysis, p = . . k plot showing genes in the intracranial xenograft signature ranked by (x-axis) the mean survival difference between the "high" expressing group and the "low" expressing group and (y-axis) the statistical significance of the survival difference as calculated by the log-rank test. patients were grouped into "high" or "low" signature expression groups based on the median gene expression. l plot showing genes in the d bioprinted tetra-cultures (vs sphere cell culture) signature ranked by (x-axis) the mean survival difference between the "high" expressing group and the "low" expressing group and (y-axis) the statistical significance of the survival difference as calculated by the log-rank test. patients were grouped into "high" or "low" signature expression groups based on the median gene expression. m the outer pie chart displays the fraction of genes with prognostic significance in the d bioprinted tetra-cultures gene signature as calculated by the log-rank test. patients were grouped into "high" or "low" signature expression groups based on the median gene expression. the inner pie chart displays the number of total prognostically significant genes as a fraction of all genes. the chi-squared test was used for statistical analysis, p < . . sphere culture cell proliferation experiments were conducted by plating cells of interest at a density of cells per well in a -well plate with replicates. cell titer glo (promega) was used to measure cell viability. data is presented as mean ± standard deviation. drug sensitivity prediction therapeutic sensitivity and gene expression data were accessed through the cancer therapeutics response portal (https://portals. broadinstitute.org/ctrp/). [ ] [ ] [ ] gene signature scores were calculated for each cell line in the dataset using the single sample gene set enrichment analysis projection (ssgseaprojection) module on genepattern (https://cloud.genepattern.org). gene signature score was then correlated with area under the curve (auc) values for drug sensitivity for each compound tested. correlation r-value was plotted and statistical analyses were corrected for multiple test correction. crispr screening and data analysis whole-genome crispr-cas loss-of-function screening was performed with the human crispr knockout pooled library (brunello), which was a gift from david root and john doench (addgene # ). the library was used following the instructions on addgene website (https://www.addgene.org/pooled-library/ broadgpp-human-knockout-brunello). briefly, the library was stably transduced into gscs by lentiviral infection with a multiplicity of infection (moi) around . - . , after puromycin selection, cells were propagated in either standard sphere cell culture conditions or in a d tetra-culture system. after days, genomic dna was extracted from gscs and the sequencing library was generated using the protocol on addgene website (https://media.addgene.org/cms/filer_public/ / / f - - a -b c -e a ecf f /broadgpp-sequencing-protocol. pdf). sequencing quality control was performed using fastqc (http://www.bioinformatics.babraham.ac.uk/projects/fastqc) and enrichment and dropout were calculated using the mageck-vispr pipeline , using the mageck-mle pipeline. in vivo tumorigenesis assays intracranial xenografts experiments were generated by implanting , patient-derived gscs (cw ) following treatment with sgrnas targeting pag or znf or a sgcont into the right cerebral cortex of nsg mice (nod.cg-prkdcscid il rgtm wjl/szj, the jackson laboratory, bar harbor, me, usa) at a depth of . mm under a university of california, san diego institutional animal care and use committee (iacuc) approved protocol. all murine experiments were performed under an animal protocol approved by the university of california, san diego iacuc. healthy, wild-type male or female mice of nsg background, - weeks old, were randomly selected and used in this study for intracranial injection. mice had not undergone prior treatment or procedures. mice were maintained in h light/ h dark cycle by animal husbandry staff with no more than mice per cage. experimental animals were housed together. housing conditions and animal status were supervised by a veterinarian. animals were monitored until neurological signs were observed, at which point they were sacrificed. neurological signs or signs of morbidity included hunched posture, gait changes, lethargy and weight loss. survival was plotted using kaplan-meier curves with statistical analysis using a log-rank test. subcutaneous xenografts were established by implanting million luciferase-labeled cw gscs into the right flank of nsg mice and maintained as described above. two weeks after implantation, treatment was initiated with mg/kg of ifosfamide (hy- , medchemexpress) dissolved in % safflower oil (spectrum laboratory products) and % dmso or vehicle alone by μl intraperitoneal injection once per day for days. luminescence signal was assessed at days , , , , and after initiation of treatment using bioluminescence imaging following injection of luciferin reagent intraperitoneally. tumor size was normalized based on the day time point for each mouse individually. statistical analysis statistical analysis parameters are provided in each figure legend. multiple group comparisons were compared by one-way anova with tukey's post-hoc analysis (by graphpad prism). p < . was designated as the threshold value for statistical significance. all data were displayed as mean values with error bars representing standard deviation. all raw sequencing data and selected processed data is available on geo at the accession number gse (https://www.ncbi.nlm.nih.gov/geo/query/acc.cgi? acc=gse ). there are no restrictions on data availability, and all data will be made available upon request directed to the corresponding authors. all biological materials used in this manuscript will be made available upon request to the corresponding authors. distribution of human patient-derived gscs may be distributed following completion of a material transfer agreement (mta) with the appropriate institutions if allowed. all computational algorithms utilized in the manuscript have been referenced in the corresponding figure legend and described in the methods section. additional details can be made available upon request. co-evolution of tumor cells and their microenvironment project drive: a compendium of cancer dependencies and synthetic lethal relationships uncovered by large-scale, deep rnai screening the cancer cell line encyclopedia enables predictive modelling of anticancer drug sensitivity next-generation characterization of the cancer cell line encyclopedia the landscape of cancer cell line metabolism patient-derived xenograft models: an emerging platform for translational cancer research organogenesis in a dish: modeling development and disease using organoid technologies organoids as an in vitro model of human development and disease sequential cancer mutations in cultured human intestinal stem cells modeling colorectal cancer using crispr-cas -mediated engineering of human intestinal organoids a living biobank of breast cancer organoids captures disease heterogeneity brca-deficient mouse mammary tumor organoids to study cancer-drug resistance human primary liver cancer-derived organoid cultures for disease modeling and drug screening ductal pancreatic cancer modeling and drug screening using human pluripotent stem cell-and patient-derived tumor organoids a three-dimensional organoid culture system derived from human glioblastomas recapitulates the hypoxic gradients and cancer stem cell heterogeneity of tumors found in vivo organoids in cancer research organoid cultures derived from patients with advanced prostate cancer modeling patient-derived glioblastoma with cerebral organoids modeling physiological events in d vs. d cell culture the microenvironmental landscape of brain tumors d bioprinting of tissues and organs bioprinting for cancer research d-bioprinted mini-brain: a glioblastoma model to study cellular interactions and therapeutics a bioprinted human-glioblastoma-on-a-chip for the identification of patient-specific responses to chemoradiotherapy chitosan-alginate d scaffolds as a mimic of the glioma tumor microenvironment elucidating the mechanobiology of malignant brain tumors using a brain matrix-mimetic hyaluronic acid hydrogel platform brain-mimetic d culture platforms allow investigation of cooperative effects of extracellular matrix features on therapeutic resistance in glioblastoma engineering tumors with d scaffolds differential response of patient-derived primary glioblastoma cells to environmental stiffness d bioprinting of functional tissue models for personalized drug screening and in vitro disease modeling rapid d bioprinting of decellularized extracellular matrix with regionally varied mechanical properties and biomimetic microarchitecture deterministically patterned biomimetic human ipsc-derived hepatic model via rapid d bioprinting dissecting and rebuilding the glioblastoma microenvironment with engineered materials hyaluronan and hyaluronectin in the extracellular matrix of human brain tumour stroma tumor stem cells derived from glioblastomas cultured in bfgf and egf more closely mirror the phenotype and genotype of primary tumors than do serum-cultured cell lines transcription elongation factors represent in vivo cancer dependencies in glioblastoma chromatin landscapes reveal developmentally encoded transcriptional states that define human glioblastoma modeling neuro-immune interactions during zika virus infection possible involvement of the m anti-inflammatory macrophage phenotype in growth of human gliomas glioma grade is associated with the accumulation and activity of cells bearing m monocyte markers single-cell profiling of human gliomas reveals macrophage ontogeny as a basis for regional differences in macrophage activation in the tumor microenvironment decoupling genetics, lineages, and microenvironment in idh-mutant gliomas by single-cell rna-seq neuronal activity promotes glioma growth through neuroligin- secretion electrical and synaptic integration of glioma into neural circuits targeting neuronal activity-regulated neuroligin- dependency in high-grade glioma tumor vascular permeability, accumulation, and penetration of macromolecular drug carriers cbtrus statistical report: primary brain and other central nervous system tumors diagnosed in the united states in - radiotherapy plus concomitant and adjuvant temozolomide for glioblastoma correlating chemical sensitivity and basal gene expression reveals mechanism of action harnessing connectivity in a large-scale small-molecule sensitivity dataset an interactive resource to identify cancer genetic and lineage dependencies targeted by small molecules lincs data portal . : next generation access point for perturbation-response signatures regulation of glioma cell phenotype in d matrices by hyaluronic acid hypoxic induction of vasorin regulates notch turnover to maintain glioma stem-like cells an id -dependent mechanism for vhl inactivation in cancer glutamatergic synaptic input to glioma cells drives brain tumour progression synaptic proximity enables nmdar signalling to promote brain metastasis neural precursor-derived pleiotrophin mediates subventricular zone invasion by glioma activation of notch signaling by tenascin-c promotes growth of human brain tumor-initiating cells a tension-mediated glycocalyx-integrin feedback loop promotes mesenchymal-like glioblastoma tumour-associated macrophages secrete pleiotrophin to promote ptprz signalling in glioblastoma stem cells for tumour growth macrophage-associated pgk phosphorylation promotes aerobic glycolysis and tumorigenesis a glial signature and wnt signaling regulate glioma-vascular interactions and tumor microenvironment ephrinb drives perivascular invasion and proliferation of glioblastoma stem-like cells osteopontin-cd signaling in the glioma perivascular niche enhances cancer stem cell phenotypes and promotes aggressive tumor growth cancer stem cells: the architects of the tumor ecosystem photocrosslinked hyaluronic acid hydrogels: natural, biodegradable tissue engineering scaffolds precise tuning of facile one-pot gelatin methacryloyl (gelma) synthesis robust and highly-efficient differentiation of functional monocytic cells from human pluripotent stem cells under serum-and feeder cellfree conditions a model for neural development and treatment of rett syndrome using human induced pluripotent stem cells selective blockade of the lyso-ps lipase abhd stimulates immune responses in vivo salmon provides fast and bias-aware quantification of transcript expression gencode reference annotation for the human and mouse genomes differential analyses for rna-seq: transcript-level estimates improve gene-level inferences moderated estimation of fold change and dispersion for rna-seq data with deseq reconstructing and reprogramming the tumor-propagating potential of glioblastoma stem-like cells limma powers differential expression analyses for rnasequencing and microarray studies gene set enrichment analysis: a knowledge-based approach for interpreting genome-wide expression profiles pgc- alpha-responsive genes involved in oxidative phosphorylation are coordinately downregulated in human diabetes enrichment map: a network-based method for gene-set enrichment visualization and interpretation tumor evolution of glioma-intrinsic gene expression subtypes associates with immunological changes in the microenvironment optimized sgrna design to maximize activity and minimize off-target effects of crispr-cas quality control, modeling, and visualization of crispr screens with mageck-vispr mageck enables robust identification of essential genes from genome-scale crispr/cas knockout screens supplementary information accompanies this paper at https://doi.org/ . / s - - - . competing interests: a.r.m. is a co-founder and has equity interest in tismoo, a company dedicated to genetic analysis focusing on therapeutic applications customized for the autism spectrum disorder and other neurological disorders origin genetics. the terms of this arrangement have been reviewed and approved by the university of california, san diego in accordance with its conflict of interest policies. the remaining authors declare no potential conflicts of interest. key: cord- -jms hrmp authors: liu, chunmei; song, yinglei; malmberg, russell l.; cai, liming title: profiling and searching for rna pseudoknot structures in genomes date: journal: transactions on computational systems biology ii doi: . / _ sha: doc_id: cord_uid: jms hrmp we developed a new method that can profile and efficiently search for pseudoknot structures in noncoding rna genes. it profiles interleaving stems in pseudoknot structures with independent covariance model (cm) components. the statistical alignment score for searching is obtained by combining the alignment scores from all cm components. our experiments show that the model can achieve excellent accuracy on both random and biological data. the efficiency achieved by the method makes it possible to search for structures that contain pseudoknot in genomes of a variety of organisms. searching genomes with computational models has become an effective approach for the identification of genes. during recent years, extensive research has been focused on developing computationally efficient and accurate models that can find novel noncoding rnas and reveal their associated biological functions. unlike the messenger rnas that encode the amino acid residues of protein molecules, noncoding rna molecules play direct roles in a variety of biological processes including gene regulation, rna processing, and modification. for example, the human sk rna binds and inhibits the transcription elongation factor p-tefb [ ] [ ] and the rnase p rna processes the ' end of precursor trnas and some rrnas [ ] . noncoding rnas include more than different families [ ] . genome annotation based on models constructed from homologous sequence families could be a reliable and effective approach to enlarging the known families of noncoding rnas. the functions of noncoding rnas are, to a large extent, determined by the secondary structures they fold into. secondary structures are formed by bonded base pairs between nucleotides and may remain unchanged while the nucleotide sequence may have been significantly modified through mutations over the course of evolution. profiling models based solely on sequence content such as hidden markov model (hmm) [ ] may miss structural homologies when directly used to search genomes for noncoding rnas containing complex secondary structures. models that can profile noncoding rnas must include both the content and the structural information from the homologous sequences. the covariance model (cm) developed by eddy and durbin [ ] extends the profiling hmm by allowing the coemission of paired nucleotides on certain states to model base pairs, and introduces bifurcation states to emit parallel stems. the cm is capable of modeling secondary structures comprised of nested and parallel stems. however, pseudoknot structures, where at least two structurally interleaving stems are involved, cannot be directly modeled with the cm and have remained computationally intractable for searching [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] . so far, only a few systems have been developed for profiling and searching for rna pseudoknots. one example is erpin developed by gautheret and lambert [ ] [ ] . erpin searches genomes by sequentially looking for single stem loop motifs contained in the noncoding rna gene, and reports a hit when significant alignment scores are observed for all the motifs at their corresponding locations. since erpin does not allow the presence of gaps when it performs alignments, it is computationally very efficient. however, alignments with no gaps may miss distant homologies and thus result in a lower sensitivity. brown and wilson [ ] proposed a more realistic model comprised of a number of stochastic context free grammar (scfg) [ ] [ ] components to profile pseudoknot structures. in their model, the interleaving stems in a pseudoknot structure are derived from different components; the pseudoknot structure is modeled as the intersection of components. the optimal alignment score of a sequence segment is computed by aligning it to all the components iteratively. the model can be used to search sequences for simple pseudoknot structures efficiently. however, a generic framework for modeling interleaving stems and carrying out the search was not proposed in their work. for pseudoknots with more complex structure, more than two scfg components may be needed and the extension of the iterative alignment algorithm to k components may require k! different alignments in total since all components are treated equally in their model. in this paper, we propose a new method to search for rna pseudoknot structures using a model of multiple cms. unlike the model of brown and wilson, we use independent cm components to profile the interleaving stems in a pseudoknot. based on the model, we have developed a generic framework for modeling interleaving stems of pseudoknot structures; we propose an algorithm that can efficiently assign stems to components such that interleaving stems are profiled in different components. the components with more stems are associated with higher weights in determining the overall conformation of a sequence segment. in order to efficiently perform alignments of the sequence segment to the model, instead of iteratively aligning the sequence segment to the cm components, our searching algorithm aligns it to each component independently following the descending order of component weights. the statistical log-odds scores are computed based on the structural alignment scores of each cm component. stem contention may occur such that two or more base pairs obtained from different components require the participation of the same nucleotide. due to the conformational constraints inherently imposed by the cm components, stem contentions occur infrequently (less than %) and can be effectively resolved based on the conformational constraints from the alignment results on components with higher weight values. the algorithm is able to accomplish the search with a worst case time complexity of o((k − )w l) and a space complexity of o(kw ), where k is the number of cm components in the model, w and l are the size of the searching window and the length of the genome respectively. we used the model to search for a variety of rna pseudoknots inserted in randomly generated sequences. experiments show that the model can achieve excellent sensitivity (se) and specificity (sp) on almost all of them, while using only slightly more computation time than searching for pseudoknot-free rna structures. we then applied the model and the searching algorithm to identify the pseudoknots on the ' untranslated region in several rna genomes from the corona virus family. an exact match between the locations found by our program and the real locations is observed. finally, in order to test the ability of our program to cope with noncoding rna genes with complex pseudoknot structures, we carried out an experiment where the complete dna genomes of two bacteria were searched to find the locations of the tmrna genes. the results show that our program identified the location with a reasonable amount of error (with a right shift of around nucleotide bases) for one bacterial genome and for the other bacteria search was perfect. to the best of our knowledge, this is the first experiment where a whole genome of more than a million nucleotides is searched for a complex structure that contains pseudoknots. to test the performance of the model, we developed a search program in c language and carried out searching experiments on a sun/solaris workstation. the workstation has dual processors and gb main memory. we evaluated the accuracy of the program on both real genomes and randomly generated sequences with a number of rna pseudoknot structures inserted. the rnas we choose to test the model are shown in table . model training and testing are based on the multiple alignments downloaded from the rfam database [ ] . for each rna pseudoknot, we divided the available data into a training set and a testing set, and the parameters used to model it are estimated based on multiple structural alignments among − homologous training sequences with a pairwise identity less than %. the emission probabilities of all nucleotides for a given state in a cm component are estimated by computing their frequencies to appear in the corresponding column in the multiple alignment of training sequences; transition probabilities are computed similarly by considering the rel- table . the performance of the model on different rna pseudoknots inserted into a background (of nucleotides) randomly generated with different c+g concentrations. tn is the total number of pseudoknotted sequence segments inserted; ci is the number of sequence segments correctly identified by the program (with a positional error less than ± bases); nh is the number of sequence segments returned by the program; se and sp are sensitivity and specificity respectively. the thresholds of log-odds score are predetermined using the z-score value of . . ative frequencies for different types of transitions that occur between the corresponding consecutive columns in the alignment. pseudocounts, dependent on the number of training sequences, are included to prevent overfitting of the model to the training data. to measure the sensitivity and specificity of the searching program within a reasonable amount of time, for each selected pseudoknot structure, we selected − sequence segments from the set of testing data and inserted them into each of the randomly generated sequences of nucleotides. in order to test whether the model is sensitive to the base composition of the background sequence, we varied the c+g concentration in the random background. the program computes the log-odds, the logarithmic ratio of the probability of generating sequence segment s by the null (random) model r to that by our model m . it reports a hit when the z-score of s is greater than . . the computation of z-scores requires knowing the mean and standard deviation for the distribution of log-odd scores of random sequence segments; both of them can be determined with methods similar to the ones introduced by klein and eddy [ ] before the search starts. as can be seen in table , the program correctly identifies more than % of inserted sequence segments with excellent specificity in most of the experiments. the only exception is the srprna, where the program misses more than % inserted sequence segments in one of the experiments. the relatively lower sensitivity in that particular experiment can be partly ascribed to the fact that the pseudoknot structure of srprna contains fewer nucleotides; thus its structural and sequence patterns have a larger probability to occur randomly. the running time for srprna, however, is also significantly shorter than that needed by most of other rna pseudoknots due to the smaller size of the model. additionally, while the alpha−rbs pseudoknot has a more complex structure and three cm components are needed to model it, our searching algorithm efficiently identifies more than % of the inserted pseudoknots with high specificities. a higher c+g concentration in the background does not adversely affect the specificity of the model; it is evident from table that the program achieves better overall performance in both sensitivity and specificity in a background of higher c+g concentrations. we therefore conjecture that the specificity of the model is partly determined by the base composition of the genome and is improved if the base composition of the target gene is considerably different from its background. to test the accuracy of the program on real genomes, we performed experiments to search for particular pseudoknot structures in the genomes for a variety of organisms. table shows the genomes on which we have searched with our program and the locations annotated for the corresponding pseudoknot structures. the program successfully identified the exact locations of known 'utr pseudoknot in four genomes from the family of corona virus. this pseudoknot was recently shown to be essential for the replication of the viruses in the family [ ] . in addition, the genomes of the bacteria, haemophilus influenzae and neisseria meningitidis mc , were searched for their tmrna genes. the haemophilus influenzae dna genome contains about . × nucleotides and neisseria meningitidis mc dna genome contains about . × nucleotides. the tmrna functions in the transtranslation process to add a c-terminal peptide tag to the incomplete protein product of table . the results obtained with our searching program on the genomes of a variety of organisms. ga is the accession number of the genome; rl specifies the real location of the pseudoknot structure in the genome; sl is the one returned by the program; rt is the running time needed to perform the searching in hours; gl is the length of the genome in its number of bases. a defective mrna [ ] . the central part of the secondary structure of tmrna molecule consists of four pseudoknot structures. figure shows the pseudoknot structures on the tmrna molecule. in order to search the bacterial dna genomes efficiently, the combined pseudoknots and were used to search the genome first; the program searches for the whole tmrna gene only in the region around the locations where a hit for pk and pk is detected. we cut the genome into segments with shorter lengths (around nucleotide bases for each), and ran the program in parallel on ten of them in two rounds. the result for neisseria meningitidis mc shows that we successfully identified the exact locations of tmrna. however, the locations of tmrna obtained for haemophilus influenzae have a shift of around nucleotides with respect to its real location ( % of the length of the tmrna). this slight error can probably be ascribed to our "hit-andextend" searching strategy to resolve the difficulty arising from the complex structure and the relatively much larger size of tmrna genes; positional errors may occur during different searching stages and accumulate to a significant value. our experiment on the dna genomes also demonstrates that, for each genome, it is very likely there is only one tmrna gene in it, since our program found only one significant hit. to our knowledge, this is the first computational experiment where a whole genome of more than a million nucleotides was successfully searched for a complex structure that contains pseudoknot structures. the covariance model (cm) proposed by eddy and durbin [ ] [ ] can effectively model the base pairs formed between nucleotides in an rna molecule. similarly to the emission probabilities in hmms, the emission probabilities in the cm for both unpaired nucleotides and base pairs are positional dependent. the profiling of a stem hence consists of a chain of consecutive emissions of base pairs. parallel stems on the rna sequence are modeled with bifurcation transitions where a bifurcation state is split into two states. the parallel stems are then generated from the transitions starting with the two states that result respectively. the genome is scanned by a window with an appropriate length. each location of the window is scored by aligning all subsequence segments contained in the window to the model with the cyk algorithm. the maximum log-odds score of them is determined as the log-odds score associated with the location. a hit is reported for a location if the computed log-odds score is higher than a predetermined threshold value. pseudoknot structures are beyond the profiling capability of a single cm due to the inherent context sensitivity of pseudoknots. models for pseudoknot structures require a mechanism for the description of their interleaving stems. previous work by brown and wilson [ ] and cai et al. [ ] has modeled the pseudoknot structures with grammar components that intersect or cooperatively communicate. a similar idea is adopted in this work; a number of independent cm components are combined to resolve the difficulty in profiling that arises from the interleaving stems. interleaving stems are profiled in different cm components and the alignment score of a sequence segment is determined based on a combination of the alignment scores on all components. however, the optimal conformations from the alignments on different components may violate some of the conformational constraints that a single rna sequence must follow. for example, a nucleotide rarely forms two different base pairs simultaneously with other nucleotides in an rna molecule. this type of restriction is not considered by the independent alignments carried out in our model and thus may lead to erroneous searching results if not treated properly. in our model, stem contention may occur. we break the contention by introducing different priorities to components; base pairs determined from components with the highest priority win the contention. we hypothesize that, biochemically, components profiling more stems are likely to play more dominant roles in the formation of the conformation and are hence assigned higher priority weights. in order to profile the interleaving stems in a pseudoknot structure with independent cm components, we need an algorithm that can partition the set of stems on the rna sequence into a number of sets comprised of stems that mutually do not interleave. based on the consensus structure of the rna sequence, an undirected graph g = (v, e) can be constructed where v , the set of vertices in g, consists of all stems on the sequence. two vertices are connected with an edge in g if the corresponding stems are in parallel or nested. the set of vertices v needs to be partitioned into subsets such that the subgraph induced by each subset forms a clique. we use a greedy algorithm to perform the partition. starting with a vertex set s initialized to contain a arbitrarily selected vertex, the algorithm iteratively searches the neighbors of the vertices in s and computes the set of vertices that are connected to all vertices in s. it then randomly selects one vertex v that is not in s from the set and modifies s by assigning v to s. the algorithm outputs s as one of the subsets in the partition when s can not be enlarged and randomly selects an unassigned vertex and repeats the same procedure. it stops when every vertex in g has been included in a subset. although the algorithm does not minimize the number of subsets in the partition, our experiments show that it can efficiently provide optimal partitions of the stems on pseudoknot structures of moderate structural complexity. the cm components in the profiling model are generated and trained based on the partition of the stems. the stems in the same subset are profiled in the same cm component. for each component, the parameters are estimated by considering the consensus structure formed by the stems in the subset only. the optimal alignments of a sequence segment to the cm components are computed with the dynamic programming based cyk algorithm. as we have mentioned before, higher priority weights are assigned to components with more stems profiled. the component with the maximum number of stems thus has the maximum weight and is the dominant component in the model. the algorithm performs alignments in the descending order of component weights. it selects the sequence segment that maximizes the log-odds score from the dominant component. the alignment scores and optimal conformations of this segment on other components are then computed and combined to obtain the overall log-odds score for the segment's position on the genome. more specifically, we assume that the model contains k cm components m , m , ..., m k− in descending order of component weights. the algorithm considers all possible sequence segments s d that are enclosed in the window and uses equation ( ) to determine the sequence segment s to be the candidate for further consideration, where w is the length of the window used in searching, and equation ( ) to compute the overall log-odds score for s. we use sm i to denote the parts of s that are aligned to the stems profiled in cm component m i . basically, log odds(sm i |m i ) accounts for the contributions from the alignment of sm i to m i . the log-odds score of sm i is counted in both m and m i and must be subtracted from the sum. log odds(s|m ) = log odds(s|m ) the conformations corresponding to the optimal alignments of a sequence segment to all cm components are obtained by tracing back the dynamic programming matrices and checking to ensure that no stem contention occurs. since each nucleotide in the sequence is represented with a state in a cm component, the cm inherently imposes constraints on the optimal conformations of sequence segments aligned to it. we hence expect that stem contention occurs with a low frequency. in order to verify this intuition, we tested the model on sequences randomly generated with different base compositions and evaluated the frequencies of stem contentions for pseudoknot structures on which we have performed an accuracy test; the results are shown in figure . the presence of stem contention increases the running time of the algorithm, because the alignment of one of the involved components must be recomputed to resolve the contention. based on the assumption that components with more stems contribute more to the stability of the optimal conformation, we resolve the contention in favor of such components. we perform recomputation on the component with a lower number of stems by incorporating conformational constraints inherited from components with more stems into the alignment algorithm, preventing them from forming the contentious stems. specifically, we assume that stem s j ∈ m i and stem contention occurs between s j and other stems profiled in m i− ; the conformational constraints from the component m i− are in the format of (l , l ) and (r , r ). in other words, to avoid the stem con- tmrna-pk telomerase-vert tombus- -iv alpha-rbs srprna fig. . random sequences were generated at each given base composition and aligned to the corresponding profiling model. the sequences are of about the same length as the length of the pseudoknot structure. the stem contention rates for each pseudoknot structure were measured and plotted. they were the ratio of the number of random sequences in which stem contentions occurred to the number of total random sequences. left: plots of profiling models observed to have a stem contention rate lower than %, right: plots of these with slightly higher stem contention frequencies. the experimental results demonstrate that, in all pseudoknots where we have performed accuracy tests, stem contention occurs with a rate lower than % and is insensitive to the base composition of sequences. tention, the left and right parts of the stem must be the subsequences of indices (l , l ) and (r , r ) respectively. the dynamic programming matrices for s j are limited to the rectangular region that satisfies l ≤ s ≤ l and r ≤ t ≤ r . the stem contention frequency depends on the conformational flexibilities of the components in the covariance model. more flexibilities in conformation may improve the sensitivity of the model but cause higher contention frequency and thus increase the running time for the algorithm. in the worst case, recomputation is needed for all nondominant components in the model and the time complexity of the algorithm becomes o((k − )w l), where k is the number of components in the model, w and l are the window length and the genome length respectively. in this paper, we have introduced a new model that serves as the basis for a generic framework that can efficiently search genomes for the noncoding rnas with pseudoknot structures. within the framework, interleaving stems in pseudoknot structures are modeled with independent cm components and alignment is performed by aligning sequence segments to all components following the descending order of their weight values. stem contention occurs with a low frequency and can be resolved with a dynamic programming based recomputation. the statistical log-odds scores are computed based on the alignment results from all components. our experiments on both random and biological data demonstrate that the searching framework achieves excellent performance in both accuracy and efficiency and can be used to annotate genomes for noncoding rna genes with complex secondary structures in practice. we were able to search a bacterial genome for a complete structure with a pseudoknot in about one week on our sun workstation. it would be desirable to improve our algorithm so that we could search larger genomes and databases. the running time, however, could be significantly shortened if a filter can be designed to preprocess dna genomes and only the parts that pass the filtering process are aligned to the model. alternatively, it may be possible to devise alternative profiling methods to the covariance model that would allow faster searches. dynamic programming algorithms for rna secondary structure prediction with pseudoknots rna pseudoknot modeling using intersections of stochastic context free grammars with applications to database search small subunit ribosomal rna modeling using stochastic context-free grammars stochastic modeling of pseudoknot structures: a grammatical approach biological sequence analysis: probabilistic models of proteins and nucleic acids rna sequence analysis using covariance models ribonuclease p: unity and diversity in a trna processing ribozyme direct rna motif definition and identification from multiple sequence alignments using secondary structure profiles characterization of the rna components of a putative molecular switch in the ' untranslated region of the murine coronavirus genome rfam: an rna family database rsearch: finding homologs of single structured rna sequences hidden markov models in computational biology. applications to protein modeling prediction of rna pseudoknots-comparative study of genetic algorithms rna pseudoknot prediction in energy based models rnamotif, an rna secondary structure definition and search algorithm functional and structural analysis of a pseudoknot upstream of the tag-encoded sequence in e. coli tmrna sk small nuclear rna binds to and inhibits the activity of cdk /cyclin t complexes design, implementation and evaluation of a practical pseudoknot folding algorithm based on thermodynamics the language of rna: a formal grammar that includes pseudoknots a dynamic programming algorithm for rna structure prediction including pseudoknots an iterated loop matching approach to the prediction of rna secondary structures with pseudoknots stochastic context-free grammars for trna modeling an expanding universes of noncoding rnas tree adjoining grammars for rna structure prediction the sk small nuclear rna inhibits the cdk /cyclin t kinase to control transcription key: cord- -od fr l authors: liu, ming; cao, jie; liang, jing; chen, mingjun title: epidemic-logistics network considering time windows and service level date: - - journal: epidemic-logistics modeling: a new perspective on operations research doi: . / - - - - _ sha: doc_id: cord_uid: od fr l in this chapter, we present two optimization models for optimizing the epidemic-logistics network. in the first one, we formulate the problem of emergency materials distribution with time windows to be a multiple traveling salesman problem. knowledge of graph theory is used to transform the mtsp to be a tsp, then such tsp route is analyzed and proved to be the optimal hamilton route theoretically. besides, a new hybrid genetic algorithm is designed for solving the problem. in the second one, we propose an improved location-allocation model with an emphasis on maximizing the emergency service level. we formulate the problem to be a mixed-integer nonlinear programming model and develop an effective algorithm to solve the model. in this chapter, we present two optimization models for optimizing the epidemic-logistics network. in the first one, we formulate the problem of emergency materials distribution with time windows to be a multiple traveling salesman problem. knowledge of graph theory is used to transform the mtsp to be a tsp, then such tsp route is analyzed and proved to be the optimal hamilton route theoretically. besides, a new hybrid genetic algorithm is designed for solving the problem. in the second one, we propose an improved location-allocation model with an emphasis on maximizing the emergency service level. we formulate the problem to be a mixed-integer nonlinear programming model and develop an effective algorithm to solve the model. burg hemorrhagic fevers in angola, sars in china, anthrax mail in usa, ebola in congo,smallpox and so on. bioterrorism threats are realistic and it has a huge influence on social stability, economic development and human health. without question, nowadays the world has become a risk world, filling with all kinds of threaten from both nature and man made. economy would always be the most important factor in normal materials distribution network. however, timeliness is much more important in emergency materials distribution network. to form a timeliness emergency logistics network, a scientific and rational emergency materials distribution system should be constructed to cut down the length of emergency rescue route and reduce economic loss. in s, america had invested lots of money to build and perfect the emergency warning defense system of public health, aiming to defense the potential terrorism attacks of biology, chemistry and radioactivity material. metropolitan medical response system (mmrs) is one of the important parts, which played a crucial role in the " . " event and delivered tons medicine materials to new york in h [ ] . in october , suffered from the bioterrorism attack of anthrax, the federal medicine reserve storage delivered a great deal of medicine materials to local health departments [ ] . khan et al. [ ] considered that the key challenge of anti-bioterrorism is that people don't know when, where and in which way they would suffer bioterrorism attack, and what they can do is just using vaccine, antibiotics and medicine to treat themselves after disaster happened. because of this, venkatesh and memish [ ] mentioned that what a country most needed to do is to check its preparation for bioterrorism attacks, especially the perfect extent of the emergency logistics network which includes the reserve and distribution of emergency rescue materials, and the emergency response ability to bioterrorism attacks. other anti-bioterrorism response relative researches can be found in kaplan et al. [ ] . emergency materials distribution is one of the major activities in anti-bioterrorism response. emergency materials distribution network is driven by the biological virus diffusion network, which has different structures from the general logistics network. quick response to the emergency demand after bioterrorism attack through efficient emergency logistics distribution is vital to the alleviation of disaster impact on the affected areas, which remains challenges in the field of logistics and related study areas [ ] . in the work of cook and stephenson [ ] , importance of logistics management in the transportation of rescue materials was firstly proposed. references ray [ ] and rathi et al. [ ] introduced emergency rescue materials transportation with the aim of minimizing transportation cost under the different constraint conditions. a relaxed vrp problem was formulated as an integer programming model and proved that's a np-hard problem in dror er al. [ ] other scholars have also carried out much research on the emergency materials distribution models such as fiedrich et al. [ ] , ozdamar et al. [ ] and tzeng et al. [ ] . during the actual process of emergency material distribution, the emergency command center(ecc) would always supply the emergency materials demand points(emdp) in groups based on the vehicles they have. besides, each route wouldn't repeat, which made any demand point get the emergency materials as fast as possible. to the best of our knowledge, this is a very common experience in china. under the assumption that any demand point would be satisfied after once replenishment, the question researched would be turn into a multiple traveling salesman problem (mtsp) with an immovable origin. in the work of bektas [ ] , the author gave a detailed literature review on mtsp from both sides of model and algorithm. malik et al. [ ] , carter and ragsdale [ ] illustrate some more results on how to solve the mtsp. to summarize, our model differs from past research in at least three aspects. first, nature disaster such as earthquake, typhoons, flood and so on was always used as the background or numerical simulation in the past research, such kind of disaster can disrupt the traffic and lifeline systems, obstructing the operation of rescue machines, rescue vehicles and ambulances. but situation in anti-bioterrorism system is different, traffic would be normal and the epidemic situation could be controlled with vaccination or contact isolation. second, to the best of our knowledge, this is the first time to combine research on the biological epidemic model and the emergency materials distribution model, and we assume that emergency logistics network is driven by the biological virus diffusion network. therefore, it has a different structure from the general logistics network. third, the new hybrid genetic algorithm we designed and applied in this study is different from all the traditional ways, we improved the two-part chromosome which proposed by carter and ragsdale [ ] , and custom special set order function, crossover function and mutation function, which can find the optimal result effectively. although rule of the virus diffusion isn't the emphasis in our research, it is the necessary part when depicting the emergency demanded. figure . illustrates sir epidemic model with natural birth and death of the population. then we can get the mathematic formulas as follows. where s, i and r, represent number of the susceptible, infective and recovered population, respectively. b and d, stand for the natural birth and death coefficient, α is the death coefficient for disease, β is the proportion coefficient from s to i in unit time, and last, γ is the proportion coefficient from i to r. note that number of the susceptible and the infective persons would be gotten via computer simulation with value of the other parameters preset. then, the emergency materials each point demanded can be also calculated based on the number of sick person. figure . is the roadway network of a city in south china, numbers beside the road are the length of the section (unit: km). point o is the ecc and the other nodes - are the emdps. now, there are some emergency materials arrived at the ecc by air transport and we need to send it to each demand point as fast as possible. we assumed that all the emdps are divided into groups, and any demand point in the network would be satisfied after once replenishment, then the question researched was turn into a mtsp with an immovable origin. however, time windows constraint wasn't considered. in this study, we use the new hybrid ga to solve the mtsp with time windows. using sir epidemic model in sect. . , number of the susceptible and infective people can be forecasted before emergency distribution. then symbol t i is set to show how much time is consumed in demand point i, i = , , . . . , . we assume it has a simple linear functional relationship with number of the infective person as follows. where i i is number of the infective people in demand point i, v vac is the average speed of vaccination. another assumption for this research is that vehicle speed is the same as in any roadway section in the network, which noted as a symbol v . so, question researched in this study is: based on the epidemic model analysis, how can we distribute the emergency materials to the whole emdps with a time windows constraint? how many groups should be divided? and, how can we get the optimization route? with the analysis above, objective function model can be formulated as follows. i / ∈s j∈s where x i j = means that the emergency materials is distributed to point j immediately after point i, otherwise, x i j = . s i j represent the shortest route between point i and j. n is number of the distribution groups. t k is time consumed in group k. t t w is the time windows. equations ( . ) and ( . ) are the grouping constraints, ( . ) and ( . ) insure that each demand point would be supplied once. equation ( . ) assures that there is no sub loop in the optimal route. equation ( . ) is the time windows constraint. and last, eq. ( . ) is the parameter specification. the hybrid genetic algorithm are presented as follows. step : using sir epidemic model in sect. . to forecast number of the susceptible and infective people, and then, confirm the emergency distribution time in each emdp; step : generate the original population based on the code rule; step : using the custom set order function to optimize the original population and make the new population have finer sequence information; step : estimate that whether the results satisfy the constraints ( ) to ( ) in the model, if yes, turn to the next step, if not, delete the chromosome; step : using the fitness function to evaluate fitness value of the new population; step : end one fall and the best one doubled policy are used to copy the population; step : crossover the population using the custom crossover function; step : mutate the population using the custom mutation function; step : repeat the operating procedures ( )- ( ) until the terminal condition is satisfied; step : approximate optimal routes would be found by the new hybrid genetic algorithm and then the best equilibrium solution would be selected by the local search algorithm. in order to evaluate the practical efficiency of the proposed methodology, parameters of the sir epidemic model are given as follows, b = d = − , β = − , α = . , γ = . , and initializing s = , , i = , show the fitness and route length vary with iterate times using the new hybrid ga, respectively. from the figures we can see that each group would be converged effectively, approximate optimal routes would be obtained. comparison of the approximate optimal routes is illustrated in fig. . , and the best equilibrium solution of emergency materials distribution is shown in fig. . . from figs. . and . , though length of the route in group is the shortest one, it isn't the best equilibrium solution. in other words, some demand points can be supplied immediately but others should wait for a long time. this is not the objective we pursue. from fig. . , inside deviation of group is the minimum one, which means route in group is the best equilibrium solution, though it isn't the shortest route. in other words, all the demand points can be supplied in the minimum time difference at widest possibility. another problem worthy to be pointed out is that group is the suboptimal to group , and this can be used as a candidate choice for commander under the emergency environment. in fact, results in the prior section are too idealized, for we just considered emergency materials distribution at the beginning of the virus diffusion (day = ) and we assume that each emdp has the same situation. in fact, it is impossible. each parameter preset would affect the result at last immensely. some of them are discussed as follows. ( ) time consumed with different initial size of s there are emdps in this distribution network, actually, each point has a different number of the susceptible people to others, and we can assume they are distributed from , to , . with the sir epidemic model in sect. . . , different size of the initial susceptible people will bring different size of infective people at last, and then, time consumed in these emdps would be varied. figure . illustrates that time consumed in one emdp with different initial size of s as date increased. there is almost no distinguish among them in the first days (a month), however, distinguish is outstanding in the following days. the larger the initial size of s is, the faster increment speed of the time consumed. in sect. . . , s = , is taken for each emdp and the time consumed almost no more than h, this is a very simple situation, and the optimal route with time windows can be depicted easily. but when initial size of s increased, the problem would become much more trouble for satisfying the time window constraint, and then, we should divided the distribution route in much more groups. ( ) time consumed with different initial size of i as mentioned before, each emdp also has a different number of the infective people to others, and we can assume they are distributed from to . figure . illustrates that time consumed in one emdp with different initial size of i as date increased. it also can get that time consumed in the first days is smoothly, but distinguish is outstanding in the following days. similar as before, the larger the initial size of i is, the faster increment speed of the time consumed. another interesting result is that vary i from to , distinguish of the time consumed in each situation isn't very outstanding, and size of the time consumed is around h. in other words, model in sect. . . is still serviceable and we needn't change the grouping design. ( ) time consumed with different initial size of β β is one of the most important parameters in sir epidemic model, it affects number of the infective people in the population directly, and then, it affects the time consumed in emdp accordingly. vary value of β from − to × − , and we get time consumed with it changed as show in fig. . . still we have conclusion that time consumed in the first days is more or less in different situations, but distinguish is outstanding in the following days. similar as before, the larger the initial size of β is, the faster increment speed of the time consumed. with initial size of β increased, distribution groups should be adjusted for satisfying the time windows. based on the analysis above, we can see that time consumed in the first days always stay in a lower level. it is important information for emergency relief in the anti-bioterrorism system, which means the earlier the emergency materials distributed, the less affect would be brought by parameters varied. this also answers the actual question that why emergency relief activities always get the best effectiveness at the beginning. emergency materials distribution problem with a mtsptw characteristic in the antibioterrorism system is researched in this study, and the best equilibrium solution is obtained by the new hybrid ga. modeling the mtsp using the new two-part chromosome proposed has clear advantages over using either of the existing one chromosome or the two chromosome methods. besides, combined with the sir epidemic model, relationship between the parameters and the result are discussed at last, which makes methods proposed in this study more practical. a problem worthy to be pointed out is that the shortest route between any two emdps in the new hybrid ga is calculated by dijkstra algorithm, so, the optimal result would be gotten even if some sections of the roadway are disrupted, which makes applicability range of the method projected in this study expanded. research on the emergency materials distribution is a very complex work, only some idealized situations are analyzed and discussed in this study, and some other constraints such as loading capacity of the vehicles, death coefficient for disease, distribution mode and so on, which could be directions of further research. emergency logistics network design is extremely important when responding to an unexpected epidemic pandemic. in this study, we propose an improved locationallocation model with an emphasis on maximizing the emergency service level (esl). we formulate the problem to be a mixed-integer nonlinear programming model (minlp) and develop an effective algorithm to solve the model. the numerical test shows that our model can provide tangible recommendations for controlling an unexpected epidemic. over to satisfy the emergency demand of epidemic diffusion, an efficient emergency service network, which considers how to locate the regional distribution center (rdc) and how to allocate all affected areas to these rdcs, should be urgently designed. this problem opens a wide range for applying the or/ms techniques and it has attracted many attentions in recent years. for example, ekici et al. [ ] proposed a hybrid model, which estimated the spread of influenza and integrated it with a location-allocation model for food distribution in georgia. chen et al. [ ] proposed a model which linked the disease progression, the related medical intervention actions and the logistics deployment together to help crisis managers extract crucial insights on emergency logistics management from a strategic standpoint. ren et al. [ ] presented a multi-city resource allocation model to distribute a limited amount of vaccine to minimize the total number of fatalities due to a smallpox outbreak. he and liu [ ] proposed a time-varying forecasting model based on a modified seir model and used a linear programming model to facilitate distribution decision-making for quick responses to public health emergencies. liu and zhang [ ] proposed a time-space network model for studying the dynamic impact of medical resource allocation in controlling the spread of an epidemic. further, they presented a dynamic decision-making framework, which coupled with a forecasting mechanism based on the seir model and a logistics planning system to satisfy the forecasted demand and minimize the total operation costs [ ] . anparasan and lejeune [ ] proposed an integer linear programming model, which determined the number, size, and location of treatment facilities, deployed medical staff, located ambulances to triage points, and organized the transportation of severely ill patients to treatment facilities. büyüktahtakın et al. [ ] proposed a mixed-integer programming (mip) model to determine the optimal amount, timing and location of resources that are allocated for controlling ebola in west-africa. moreover, literature reviews on or/ms contributions to epidemic control were conducted in dasaklis et al. [ ] , rachaniotis et al. [ ] and dasaklis et al. [ ] . in this study, we propose an improved location-allocation model for emergency resources distribution. we define a new concept of emergency service level (esl) and then formulate the problem to be a mixed-integer nonlinear programming (minlp) model. more precisely, our model ( ) identifies the optimal number of rdcs, ( ) determines rdcs' locations, ( ) decides on the relative scale of each rdc, ( ) allocates each affected area to an appropriate rdc, and ( ) obtains esl for the best scenario, as well as other scenarios. ( ) in this study, esl includes two components. esl is constructed to reflect the level of demand satisfaction and esl is proposed for the relative level of emergency operation cost. these two aspects are given by the weight coefficient α and − α respectively. the influence of these two factors on the esl is illustrated in fig. . . figure . a represents that esl increases as the level of demand satisfaction raised. as we can see that it is a piecewise curve. before demand is completely met, it is an s-shape curve from zero to α. after that, it becomes a constant, which means the additional emergency supplies cannot improve the esl. figure . b identifies that esl decreases as the relative total cost increases. when emergency operation cost is minimized, the esl arrives at its best level of − α. similarly, when emergency operation cost is maximized, the esl is zero. our model depicts the problem of location and allocation for emergency logistics network design. the network is a three-echelon supply chain of strategic national stockpile (sns), rdcs, and affected areas. the core problem is to determine the number and locations for the rdcs. in each affected area, there is a point of dispensing (pod). to model the problem, we first present the relative parameters and variables as follows. i : set of snss, i ∈ i . j : set of rdcs, j ∈ j . k : set of affected areas, k ∈ k . α: weight coefficient for the two parts of esl. variables: d i j : distance from sns i to rdc j. for simplify, the euclidean distance is adopted. d jk : distance from rdc j to affected area k. ε jk : binary variable. if rdc j provides emergency supplies to affected area k, ε jk = ; otherwise, ε jk = . z j : binary variable. if rdc j is opened, z j = ; otherwise, z j = . x jk : amount of emergency supplies from rdc j to affected area k. y i j : amount of emergency supplies from sns i to rdc j. (x j , y j ): coordinates of rdc j. according to the above notations, we can define the optimization model as follows. max e sl = e sl + e sl ( . ) herein, e sl is defined as ( . )-( . ). these equations reflect that the less the unsatisfied demand is, the higher e sl is. e sl is defined as follows. first, we formulate the total operation cost as ( . ): ε jk x jk d jk + j j= z j c r dc j ( . ) where c r dc j is the operating cost for rdc j when it is opened. it is decided by the relative size of the rdc, which can be expressed as: second, to non-dimensionalize the cost function f , we calculate the following two extreme values for eq. ( . ) . where var represents all variables and s represents the following constraints. f min and f max are the minimum and maximum values obtained by solving ( . ) without considering the e sl . the definition of e sl means that the lower the total operation cost is, the higher the esl is. the constraints for the optimization model are given as follows: x jk , y i j ∈ z + , ∀i ∈ i, j ∈ j, k ∈ k ( . ) (x j , y j ), ∀ j ∈ j are continuous variables. ( . ) constraint ( . ) indicates that each affected area is serviced by a single rdc. constraint ( . ) specifies that the supplies to each affected area should not be more than its demand. constraint ( . ) is a flow conservation constraint. constraint ( . ) shows that only the opened rdc can provide distribution service. constraint ( . ) the proposed model for emergency services network design is a minlp model as it involves multiplication of two variables (i.e., ε jk x jk ). more importantly, the optimization model includes a continuous facility location-allocation model with unknown number of rdcs. to avoid the complexity of such minlp model, we modify it by adding two auxiliary variables. the detail of the modification was introduced in mccormick [ ] . our solution procedure integrates an enumeration search rule and a genetic algorithm (ga), which are applied iteratively. as ga is a mature algorithm [ ] , details of the ga process are omitted here. we summarize the proposed solution methodology as below. step : data input and parameters setting, which includes i, j, k, α, d k , (x k , y k ), (x i , y i ), c t l , c lt l , and c r dc j and the related parameters for ga. step : initialization. generate the original population according to the constraints. step : evaluation. fitness function is defined as the reciprocal of esl. step : selection. use roulette as the select rule. step : crossover. single-point rule is used. step : mutation. a random mutation is applied. step : if termination condition is met, go to the next step; else, return to step . step : output the results. ( to clarify the effect of the model, we conduct a numerical test. assuming there is an unexpected epidemic outbreak in a × square region with affected areas in it. in the square region, only three snss can provide emergency supplies. because at the early stage of the outbreak, there is a large demand for emergency supplies. the supply capacity of these snss is less than the total demand in affected areas, which are set at , and respectively. the coordinates of the snss and the affected areas are obtained in advance. the upper bound of rdc number is set to be . the cost of operating a rdc is defined as × √ s j . the demand in each affected area is randomly generated. finally, unit transportation cost from sns to rdc is set to be and unit transportation cost from rdc to affected area is . based on the above data setting, we solve our model by using matlab software and obtain the results in fig. . . as it shows in this figure, one can observe that there is a trade-off between the two components of the esl. in our example, we test the parameter α from . to . , which means the demand satisfaction is more and more important in our decision making. the result shows that when α is equal to . , the total esl can arrive at its best value ( . ). beyond which it decreases again. in practice, the decision makers may have different value of α according to the actual needs. correspondingly, it will lead to different esl. our model also determines the optimal number, locations and relative sizes for the rdcs. the test results are shown in table . . for example, rdc deliver emergency supplies to affected areas , and . its relative size is . %, which means emergency supplies transshipped in this rdc occupies the corresponding proportion in total emergency distribution. table . illustrates the proportion of demand satisfaction for each affected area. for an example, demand for emergency supplies in affected area is , and all this area's demand is totally satisfied. however, one can also observe that demands in some areas are partly supplied due to the supply capacity limitation. for example, only . % of the demand in affected area is delivered. ( ) sensitivity analysis ( ) impact of α on the esl to understand the impact of α on the esl, we solve our model with different values of this parameter, meaning that decision makers have different considerations of the two components of the esl. we compare the test result in table . . it can be observed that esl increases along with the emphasis on demand satisfaction. however, the actual proportion of esl is always staying at % of the setting of α. as to the esl , one can note that it increases at first and then decreases as α varied from . to . . ( ) sensitivity analysis on different demand in each affected area we also examined the impact of different demand in each affected area. the test results are shown in fig. . . we change the original demand in each affected area for five scenarios. that means different demand situations when an unexpected infectious epidemic happened. one can easily observe the more the demand is, the lower the optimal esl is. that is because when the demand increases, the supplies of snss remain original, which leads a reduction in esl . when the demand in each affected area changes, esl varies slightly. which shows that the change of the total operation cost for the emergency logistics is not obvious when the scale of disease becomes smaller. in this study, we propose an improved location-allocation model with an emphasis on maximizing the emergency service level (esl). we formulate the problem to be a mixed-integer nonlinear programming model and develop an effective algorithm to solve the model. moreover, we test our model through a case study and sensitivity analysis. the main contribution of this research is the function of esl, which considers demand satisfaction and emergency operation cost simultaneously. our definition of esl is different from the existing literature and has a significant meaning for guiding the actual operations in emergency response. future studies could address the limitations of our work in both the disease forecasting and logistics management. for example, the dynamics of epidemic diffusion could be considered and thus our optimization problem can be extended to a dynamic programming model. american metropolitan medical response system and it's inspire for the foundation of the public healthy system in our country the discussion of orientation of chinese public health in th century public-health preparedness for biological terrorism in the usa bioterrorism-a new challenge for public health analyzing bioterror response logistics: the case of smallpox an emergency logistics distribution approach for quick response to urgent relief demand in disasters lessons in logistics from somalia a multi-period linear programming modal for optimally scheduling the distribution of food-aid in west africa allocating resources to support a multicommodity flow with time windows vehicle routing with split deliveries optimized resource allocation for emergency response after earthquake disasters emergency logistics planning in natural disasters multi-objective optimal planning for designing relief delivery systems the multiple traveling salesman problem: an overview of formulations and solution procedures an approximation algorithm for a symmetric generalized multiple depot, multiple travelling salesman problem a new approach to solving the multiple traveling salesperson problem using genetic algorithms modeling influenza pandemic and planning food distribution modeling the logistics response to a bioterrorist anthrax attack optimal resource allocation response to a smallpox outbreak methodology of emergency medical logistics for public health emergencies a dynamic allocation model for medical resources in the control of influenza diffusion a dynamic logistics model for medical resources allocation in an epidemic control with demand forecast updating resource deployment and donation allocation for epidemic outbreaks a new epidemics-logistics model: insights into controlling the ebola virus disease in west africa emergency supply chain management for controlling a smallpox outbreak: the case for regional mass vaccination controlling infectious disease outbreaks: a deterministic allocation-scheduling model with multiple discrete resources epidemics control and logistics operations: a review computability of global solutions to factorable nonconvex programs: part i-convex underestimating problems distribution network redesign for marketing competitiveness key: cord- -jmjolo p authors: pulliam, juliet r. c.; dushoff, jonathan title: ability to replicate in the cytoplasm predicts zoonotic transmission of livestock viruses date: - - journal: j infect dis doi: . / sha: doc_id: cord_uid: jmjolo p understanding viral factors that promote cross-species transmission is important for evaluating the risk of zoonotic emergence. weconstructed a database of viruses of domestic artiodactyls and examined the correlation between traits linked in the literature to cross-species transmission and the ability of viruses to infect humans. among these traits-genomic material, genome segmentation, and replication without nuclear entry-the last is the strongest predictor of cross-species transmission. this finding highlights nuclear entry as a barrier to transmission and suggests that the ability to complete replication in the cytoplasm may prove to be a useful indicator of the threat of cross-species transmission. previous studies have compared emerging human pathogens to nonemerging human pathogens and looked for characteristics typical of those considered to be emerging [ ] [ ] [ ] . to ask which characteristics predict host jumps requires a different approach. specifically, we must examine the pool of other hosts' pathogens that a target species regularly encounters. from this pool we can compare the characteristics of microbes that are able to infect the target host versus those that manifest no evidence of an ability to infect the target host. molecular characteristics that facilitate cross-species transmission are likely to be substantially different between viruses, bacteria, and protozoa, because of large differences in the pathobiology of these different taxa. here, we focus on cross-species transmission of viral infections and examine the effects of characteristics that are described in the literature as expected to affect the ability of a viral group to infect a novel host species: genome segmentation, genomic material, and site of replication. the ability to rapidly explore genetic state space is expected to increase the probability of a host jump, so we expect that viruses with rna genomes will have a higher probability of jumping than viruses with dna genomes [ , ] and that viruses with segmented genomes will have a higher probability of jumping than viruses with nonsegmented genomes [ ] . complex interactions with a host's cellular machinery, on the other hand, are expected to decrease the probability of a host jump, so we expect that viruses that are able to complete replication in the cytoplasm will have a higher probability of jumping than viruses requiring nuclear entry [ ] . to examine the effects of these characteristics, we should choose a target species that will maximize the chance that viral infection due to cross-species transmission events will have been detected; the obvious choice is humans. likewise, we should minimize differences in exposure of the target host to infectious virions produced by the source hosts. humans have regular contact with all potentially infectious bodily fluids of domestic food animals; we thus ensure that the target species has contact with all viral groups infecting the source hosts by analyzing the pool of viral species known to infect sheep, goats, cattle, and pigs. methods. we constructed a database containing taxonomic and molecular data on known viruses of domestic artiodactyls. to determine which viruses to include in the database, we searched the primary literature for references documenting infection of these species with all recognized species in all viral genera known to infect mammals. for each viral species infecting sheep, goats, pigs, or cattle, we then searched the literature to determine whether human infections have been documented (see table a in appendix a, which appears only in the electronic edition of the journal). viruses dependent on coinfection with other viral species, known to be maintained through continuous transmission within humans (see appendix a), or for which documented instances of artiodactyl infection resulted from human-to-animal transmission or experimental infection were excluded from the database. all literature searches were performed between january and february using web of science. the database contains information on the molecular characteristics hypothesized to influence the potential of a virus to cross host species: site of replication (x sr ; whether replication is completed in the cytoplasm or requires nuclear entry), genomic material (x gm ; rna or dna), and segmentation of the viral genome (x seg ; segmented or nonsegmented). these characteristics are conserved at the family level, and classifications were made on the basis of standard reference books [ , ] . we used a combination of hypothesis testing and modelbased prediction to analyze the database. hypothesis testing allowed us to determine how likely it was that the observed patterns were due to chance, whereas model-based prediction allowed us to determine what trait or set of traits was the best predictor of a livestock virus's ability to infect humans and to estimate the probability that a particular virus species would be able to jump host species, given knowledge of the traits of interest. computer code and data are available at http://lalashan .mcmaster.ca/hostjumps/ or from the authors. to determine the statistical significance of the effect of each trait on zoonotic transmission independent of the other traits of interest, we performed a series of randomization tests. for a particular trait, we held the values of the other traits and the ability of the viral species in the database to infect humans constant and permuted the values of the trait of interest within each subset defined by the other traits (thereby preserving the full cross-correlational structure of the data with regard to the viral traits) , times. the p value was given by the proportion of permutations that allowed the model to predict outcomes as well as or better than the model that was constructed using the observed data, and an ␣ level of % was used to determine the statistical significance of results. we compared model fit by use of a logistic regression model that predicted the ability to infect humans as a function of replication site, genomic material, and segmentation. the logistic model was fit in the r statistics package [ ] , and fits were compared on the basis of likelihood. because the traits examined are conserved at the family level for all species in our database, treating species as independent may bias our results. we therefore repeated our analysis at the genus and family levels. permutations of the data set were constructed by permuting the values of the trait under consideration at the taxonomic level examined and assigning species within a genus (or family) the corresponding trait value after permutation. p values were calculated as in the species-level analysis. to examine the magnitude and relative importance of the effects that the molecular characteristics of interest have on the ability of the viral species in the database to infect humans, we developed a set of logistic regression models. each model included some combination of viral traits as independent variables and the ability to infect humans as the dependent variable. traits not having a statistically significant effect on the ability of livestock viruses to infect humans were still considered for modelbased predictions, because sample sizes were limited and small-but real-effects may not be detected via hypothesis testing. we estimated parameter values for each model in r and compared models using akaike's information criterion adjusted for small sample size (aic c ) [ ] . results. a total of viral species were found to infect the livestock species of interest and meet other criteria for inclusion in the database. of these, species (representing genera in families) fulfilled the criteria for inclusion in the analysis. the effect of site of replication was significant at all taxonomic levels examined (p Ͻ . , p ϭ . , and p ϭ . for the species, genus, and family levels, respectively), with nearly half of the virus species completing replication in the cytoplasm able to infect humans. neither genomic material nor segmentation showed a significant effect on the ability of livestock viruses to infect humans at any taxonomic level. logistic regression model comparisons are summarized in table . the models are given in order as ranked by aic c . figure compares the observed data with the results of the best model. the best model included site of replication as the only variable (odds ratio, . [ % confidence interval, . - . ), and the top models were the that included site of replication. each of these models showed a positive correlation between replication in the cytoplasm and the ability to infect humans, as expected. segmentation appears in models , , and , and all models showed a positive correlation between having a segmented genome and the ability to infect humans. genomic material appears in models , , , and . again, all models showed a correlation in the expected direction. it is interesting to note that both of the viral species that caused major pandemics in humans in the th century (hiv and influenza virus a) require nuclear entry for replication. because influenza virus a infects domestic artiodactyls but was excluded from our database because it is maintained through continuous transmission in humans, we confirmed the robustness of our results to this exclusion; we also confirmed that our findings were robust to the inclusion of viral species for which human infection data were based solely on serology (see table b in appendix b, which appears only in the electronic edition of the journal). discussion. our analyses indicate that viral species infecting domestic artiodactyls are more likely to infect humans if they complete replication in the cytoplasm without nuclear entry. the observed effect of cytoplasmic replication on host-jumping ability is not surprising given the complex molecular pathways regulating nuclear entry. viral species that are unable to complete replication in the cytoplasm require intracellular transport from the site of penetration, targeting of the nucleus through nuclear localization signals, and importation of genetic material, proteins, and/or whole virions through the nuclear pore complex [ ] . the combination of molecular mechanisms governing this chain of events is likely to be highly host specific, because of strong selective pressure against admission of foreign particles into the nucleus. to date, discussion of barriers to viral replica-tion has largely focused on receptors for cellular entry. the concentration on this aspect of the viral life cycle exists for substantive reasons. first, the inability to enter a cell obviously precludes viral replication; second, several well-documented viral host jumps have been shown to occur after point mutations that modify interactions between viral particles and cellular receptors [ ] [ ] [ ] . the effect of nuclear entry seen in our data set emphasizes that cellular entry, while a necessary step, is insufficient for completion of the viral life cycle. the ability to produce genetic diversity is the factor most widely discussed as expected to increase viral host-jumping ability [ , [ ] [ ] [ ] ] . although the observed effects of genomic mate- note. x sr , x gm , and x seg are variables indicating the molecular characteristics of a viral species (see methods). ln(ᐉ ) is the log likelihood of the best-fit parameter combination for a given model. k is the no. of model parameters for a given model. aic c is the value of akaike's information criterion with small sample size correction for each model; thus, ⌬aic c is the difference in aic c value between a given model and the best model (i.e., the model with the lowest aic c value). w i is the akaike weight of the model. ␤ seg , ␤ gm , and ␤ sr are regression coefficients for genome segmentation, genomic material, and site of replication, respectively. ␤ i represents the estimated intercept for the best-fit parameter combination for each model. rial and segmentation were not statistically significant, our data do not necessarily contradict this expectation. the hypothesized effect of segmentation, in particular, may be obscured in our data set by a combination of the small number of viral species with segmented genomes and the absence of segmented dna viruses. on the other hand, the lack of predictive power associated with genomic material and segmentation in our data set may indicate that consideration of these traits alone is insufficient to capture the potential to generate useful genetic diversity. the degree to which the pool of viruses infecting domestic artiodactyls is typical of all potentially zoonotic viral species is uncertain. other pools of viral species should be examined to determine the generality of our results. similarly, further studies should examine whether the observed patterns hold for crossspecies transmission of viruses to other target host species, including wildlife and domestic animals. given the rapid rates at which ecological relationships between species are changing as a result of anthropogenic landscape changes, global warming, and globalization of both human and animal populations, the development of indicators of the risk of cross-species pathogen transmission is an increasingly important goal. as humans, domestic animals, and wildlife are brought into contact with species from which they were formerly isolated, they inevitably encounter the pathogens that these species carry. the finding that the ability to complete replication in the cytoplasm is the best predictor of zoonotic transmission and that nearly half of domestic artiodactyl viruses that are able to complete replication in the cytoplasm can infect humans suggests that cytoplasmic replication will be a useful indicator of the ability of a newly encountered virus species to jump hosts, an essential prerequisite to epidemic or pandemic emergence [ ] . it should be noted, however, that the present analysis focused exclusively on the ability to infect the target host, and the viral traits influencing this step in the emergence process may differ from those that predispose a virus to cause severe disease in a novel host as well as from those that facilitate transmission within a novel host species. diseases of humans and their domestic mammals: pathogen characteristics, host range, and the risk of emergence risk factors for human disease emergence host range and emerging and reemerging infectious diseases evolvability of emerging viruses viral host jumps: moving toward a predictive framework virus taxonomy: eighth report of the international committee on taxonomy of viruses the springer index of viruses r: a language and environment for statistical computing. vienna: r foundation for statistical computing in: model selection and multimodel inference: a practical information-theoretic approach viral entry into the nucleus the natural host range shift and subsequent evolution of canine parvovirus resulted from virus-specific binding to the canine transferrin receptor structure of sars coronavirus spike receptor-binding domain complexed with receptor a single amino acid in the pb gene of influenza a virus is a determinant of host range molecular constraints to interspecies transmission of viral pathogens origins of major human infectious diseases key: cord- -ejk wjr authors: crilly, colin j.; haneuse, sebastien; litt, jonathan s. title: predicting the outcomes of preterm neonates beyond the neonatal intensive care unit: what are we missing? date: - - journal: pediatr res doi: . /s - - - sha: doc_id: cord_uid: ejk wjr abstract: preterm infants are a population at high risk for mortality and adverse health outcomes. with recent improvements in survival to childhood, increasing attention is being paid to risk of long-term morbidity, specifically during childhood and young-adulthood. although numerous tools for predicting the functional outcomes of preterm neonates have been developed in the past three decades, no studies have provided a comprehensive overview of these tools, along with their strengths and weaknesses. the purpose of this article is to provide an in-depth, narrative review of the current risk models available for predicting the functional outcomes of preterm neonates. a total of studies describing separate models were considered. we found that most studies used similar physiologic variables and standard regression techniques to develop models that primarily predict the risk of poor neurodevelopmental outcomes. with a recently expanded knowledge regarding the many factors that affect neurodevelopment and other important outcomes, as well as a better understanding of the limitations of traditional analytic methods, we argue that there is great room for improvement in creating risk prediction tools for preterm neonates. we also consider the ethical implications of utilizing these tools for clinical decision-making. impact: based on a literature review of risk prediction models for preterm neonates predicting functional outcomes, future models should aim for more consistent outcomes definitions, standardized assessment schedules and measurement tools, and consideration of risk beyond physiologic antecedents. our review provides a comprehensive analysis and critique of risk prediction models developed for preterm neonates, specifically predicting functional outcomes instead of mortality, to reveal areas of improvement for future studies aiming to develop risk prediction tools for this population. to our knowledge, this is the first literature review and narrative analysis of risk prediction models for preterm neonates regarding their functional outcomes. preterm infants have long been recognized as a population at high risk for mortality and adverse functional outcomes, including cerebral palsy and intellectual impairment. as mortality rates for preterm neonates decline and more survive to childhood, , attention has increasingly turned towards measuring longer-term morbidities and related functional impairments during childhood and young-adulthood, as well as identifying risk factors related to these complications. , while child-specific characteristics, such as gestational age, birth weight, and sex, are well established as predictors of adverse neurodevelopmental outcomes, - recent work has identified additional factors, including bronchopulmonary dysplasia and family socioeconomic status, that are correlated with relevant outcomes, such as poor neuromotor performance and low intelligence quotient at school age. in clinical settings, the assessment of prognosis can vary widely across neonatologists, making a valid and reliable predictive model for long-term outcomes a highly sought-after clinical tool. moreover, predicting outcomes is vital when making decisions regarding which therapeutic interventions to apply, when providing critical data to parents for informed decision-making, and when matching infants with outpatient services to best meet their needs. in addition, prediction models are useful in evaluating neonatal intensive care unit (nicu) performance and allowing for between-center comparisons with proper adjustment for the severity of cases being treated. numerous prediction tools have been developed to quantify the risk of death for preterm neonates in the nicu setting, including the score for neonatal acute physiology (snap) and the clinical risk index for babies (crib). the national institute of child health and human development (nichd) risk calculator, predicting survival with and without neurosensory impairment, is widely used to counsel families in the setting of threatened delivery at the edges of viability. furthermore, there are numerous other models that use clinical data from the nicu stay to predict risk for poor functional outcomes in infancy and school age. , while several studies have categorized and evaluated the risk prediction models developed and validated in recent decades for mortality, , no studies have compared and contrasted risk prediction models for non-mortality outcomes. recently, linsell et al. published a systematic review of risk factor models for neurodevelopmental outcomes in children born very preterm or very low birth weight (vlbw). however, this review focused primarily on overall trends in model development and validation rather than a detailed consideration of individual models. in this article, we conduct an in-depth, narrative review of the current risk models available for predicting the functional outcomes of preterm neonates, evaluating their relative strengths and weaknesses in variable and outcome selection, and considering how risk model development and validation can be improved in the future. towards this, we first provide an overview of the different risk models developed since . we then frame our review of these models in terms of the outcomes predicted, the range of predictors considered, and the statistical methods used to select the variables included in the final model, as well as to assess the predictive performance of the model. finally, the ethical implications of integrating risk stratification into standard clinical care for preterm neonates are considered. we conducted a manual search for relevant literature via pubmed, entering combinations of key terms synonymous with "prediction tool," "preterm," and "functional outcome" and reading the abstracts of resulting studies (table ). studies with abstracts that appeared related to our review were then read in full to identify prediction models that were eligible for inclusion. reference lists of included studies were also reviewed, as were articles that later cited these original studies. prediction tools were defined as multivariable risk factor analyses (> variables) aiming to predict the probability of developing functional outcomes beyond months corrected age. models that solely investigated associations between individual risk factors and outcomes were excluded, as were models that were not evaluated for predictive ability in terms of either a validation study or an assessment for performance, discrimination, or calibration. tests used to evaluate a model's overall performance were r , adjusted r , and the brier score. the use of a receiver operating characteristic (roc) curve or a c-index evaluated a model's discrimination, and the hosmer-lemeshow test was considered to evaluate a model's calibration. preterm neonates were defined as < weeks of completed gestational age. models with vlbw neonates < g were also included, since in the past birth weight served as a substitute for measuring prematurity when gestational age could not be accurately determined. models were excluded if they used a cohort entirely composed of infants born prior to january ; those born after were likely to have had surfactant therapy available in the event of respiratory distress syndrome, which significantly reduced the morbidity and mortality rates among preterm neonates nationwide. , models were also excluded if they limited their prediction to the outcome of survival, if they incorporated variables measured after initial nicu discharge, or if they included subjects who were not necessarily transferred to a nicu for further care following delivery. finally, we excluded tools that only predicted outcomes to an age of < months corrected age, as well as case reports, narrative reviews, and tools reported in languages other than english. overview of risk prediction models table lists all studies with risk prediction models that meet the inclusion and exclusion criteria. [ ] [ ] [ ] from these, a total of distinct models were reported. from mortality to neurodevelopmental impairment since , several mortality prediction tools have been evaluated in regards to their ability to predict the likelihood of neurodevelopmental impairment (ndi) among neonates surviving to nicu discharge. one such model is the crib, which incorporates six physiologic variables collected within the first h of the preterm infant's life: birth weight, gestational age, presence of congenital malformations, maximum base excess, and minimum and maximum fio requirement. fowlie et al. evaluated how crib models obtained at differing time periods over the first days of life predicted severe disability among a group of infants born > weeks gestational age or vlbw. in another study, fowlie et al. incorporated cranial ultrasound findings on day of life along with crib scores between and h of life into their prediction model. subsequent studies analyzed the crib in its original -h form and, with only one exception, determined that it was not a useful tool for predicting long-term ndi or other morbidities. [ ] [ ] [ ] [ ] a second example is the snap score. snap uses physiologic parameters collected over the first h of life to predict survival to nicu discharge, and was modified to predict ndi at year and - years of age. a subsequent assessment of both the snap and the snap with perinatal extension showed a poor predictive value for morbidity at years of age for children born vlbw and/or with gestational age ≤ weeks. finally, the neonatal therapeutic intervention scoring system, a comprehensive exam-based prediction tool for mortality, was found to have a poor predictive value for adverse outcomes at years of age in children born very preterm or vlbw. shortened forms of the early physiology-based scoring systems were developed and assessed for their ability to predict outcomes in childhood. application of the crib-ii on a small cohort (n = ) of infants born < g predicted significant ndi at years of age. however, a subsequent evaluation in a much larger cohort (n = ) of preterm infants < weeks gestational age concluded that the crib-ii did no better than gestational age or birth weight alone in predicting moderate to severe functional disability at - years of age. studies have supported an association between the snap-ii and snappe-ii scores and neurodevelopmental outcomes and small head circumference at months corrected age. high snap-ii scores were shown to correlate with adverse neurological, cognitive, and behavioral outcomes up to years of age within a large cohort (n = ) of children born very preterm. antenatal risk factors several groups have used data from the nichd's neonatal research network (nrn) to design and test various risk prediction models for extremely low birth weight (elbw) newborns. one of the most widely used risk prediction tools developed from this cohort was by tyson et al., postnatal morbidity a large cohort study (n = ) from schmidt et al. , used data from elbw neonates - g enrolled in the international trial of indomethacin prophylaxis in preterms (tipp). they found that the presence of three morbidities at weeks post-menstrual age -bronchopulmonary dysplasia, serious brain injury, and severe retinopathy of prematurity-had a significant and additive effect on the risk for death or poor neurologic outcome at months corrected age. they developed a model from this relationship that has been corroborated in two studies with smaller samples and by schmidt et al. in a separate, large cohort in which the definition of poor outcome was expanded from solely ndi to "poor general health." , letting the machines decide some innovative work has been recently performed by ambalavanan et al. , in creating several risk prediction models. along with studies developing risk prediction tools with data from the nrn and the tipp to predict the outcomes of death and ndi or solely ndi, the group made the only risk prediction tool for the outcome of rehospitalization, both general and specifically for respiratory complications, using a combination of physiologic and socioeconomic variables incorporated into a decision tree approach. they have also been the only group to create neural network-trained models, using the same small cohort to predict major handicap, low mental development index (mdi), or low psychomotor development index (pdi). the advantage of using neural networks-algorithms that can "learn" mathematical relationships between a series of independent variables and a set of outcomes-is the ability to model complex or nonlinear relationships that can be elucidated by the model without having to consider these relationships a priori (as is typically required when using multiple regression models). despite the use of innovative approaches, however, none of these models differed from other studies in predictive strength or even had high predictive efficacy. limitations of prior approaches the above literature review highlights the substantial interest in developing a clinically useful risk prediction model and the limits of efforts to date. notwithstanding their differing inclusion and exclusion criteria, existing risk prediction models are relatively similar in terms of variables selected, outcomes analyzed, and statistical strategies employed. with few exceptions, the limitations of existing risk prediction models are especially apparent in their reliance on solely biologic variables and traditional analytic methods ill-equipped to handle the statistical complexity necessary for risk modeling. identifying important outcomes. the majority of risk prediction models defined ndi as their primary outcome of interest. making a determination of impairment often relies on standardized measures of cognition in concert with neurosensory deficits. yet, researchers often define ndi in different ways, making betweenstudy comparisons difficult. ndi is a construct relating to global abilities encompassing cognition, language, motor function, and vision and hearing. while the tools used to identify ndi are often also used to make diagnoses of developmental delay, ndi is not a clinical term or diagnosis in and of itself. many of the remaining studies also predicted functional outcomes, such as academic performance, executive function, language ability, and autism spectrum disorder (asd). these outcomes may be more meaningful to parents and providers than ndi. to date, only four studies have considered outcomes unrelated to neurodevelopment, such as impaired pulmonary function, "poor general health," and rehospitalization rates. , , , while the emphasis on ndi is unsurprising given the high-risk population, moderate to severe ndi only affects a minority of the preterm population. , studies have revealed numerous additional adverse outcomes that preterm individuals are more likely to experience compared to their full-term counterparts, such as impaired respiratory, cardiovascular, and metabolic function. [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] neurodevelopment has been linked to chronic health problems in later childhood. limiting risk prediction to moderate to severe ndi therefore ignores other, more common complications that preterm infants are likely to face that have an impact on neurodevelopment. this represents a missed opportunity for researchers to better understand what variables influence the likelihood that these problems occur. the impact of developmental disability on the child and family is completely absent from current risk models. health-related quality of life (hrql), which distinguishes itself as a personal rather than third-party valuation of a patient's physical and emotional well-being, is being increasingly appreciated as an important metric necessary to fully understand the impact of prematurity. in a french national survey, the majority of neonatologists, obstetricians, and pediatric neurologists stated that predicting hrql in the long term for preterm infants would be beneficial for consulting parents about what additional responsibilities they can anticipate in caring for their child. the trajectory of hrql from childhood to young-adulthood appears to improve in both vlbw and extremely low gestational age populations. prediction modeling might aid in determining which factors could positively or negatively impact hrql in this vulnerable population. finally, we must consider the age at which outcomes are being predicted. it is evident that lower gestational age is inversely proportional to rates of ndi and academic achievement in adolescence. , however, the vast majority of risk prediction models assessed outcomes at the age of years or less, with only three studies doing so at years of age or above. although early childhood outcomes may give clues about later development, many problems do not manifest until later in childhood, such as learning disabilities and certain psychiatric disorders. developmental disability severity can fluctuate throughout childhood, with catch-up occurring in early preterm children and worsening delay in some moderate and late preterm children. , although cohorts of preterm infants are not usually followed for more than several years, likely due to lack of resources and expense, recent studies have used data from national registries to link neonatal clinical data to sampled adults, providing evidence of increased rates of adverse neurodevelopmental, behavioral, and educational outcomes among adults born preterm. , opportunities are therefore available to use long-term data to extend risk prediction models beyond the first few years of life. variable selection. most of the risk models reviewed relied primarily on physiologic and clinical measures obtained during the nicu stay. while an emphasis on biologic risk factors is clearly reasonable given the known associations between perinatal morbidities and long-term outcomes, there is strong evidence in the literature suggesting associations between sociodemographic factors like parental race, education, and age, and outcomes such as cognitive impairment, cerebral palsy, and mental health disorders in children born preterm. more specific socioeconomic variables such as lower parental education, maternal income, insurance status, foreign country of birth by a parent, and socioeconomic status as defined by the elly-irving socioeconomic index have been repeatedly correlated with reduced mental development index, psychomotor development index, intelligence quotient, and social competence throughout childhood. , , [ ] [ ] [ ] [ ] [ ] [ ] the geographic area in which preterm neonates are raised could also have a profound influence on their development. neighborhood poverty rate, high school dropout rate, and place of residence (metropolitan vs. non-metropolitan) have all been correlated with academic skills and rate of mental health disorders among low birth weight children. , only of the models reviewed included socioeconomic variables. this may be due, at least in part, to the difficulty in obtaining social, economic, and demographic data; these variables are often not collected upon hospital admission. additionally, socioeconomic information is often poorly, inaccurately, and variably recorded or is largely missing. some risk prediction models collected socioeconomic variables at the follow-up visit when outcomes were assessed. this is an imperfect method given that factors such as household setting and family income may change substantially in the years following nicu discharge and affect children's health. , in some models, socioeconomic variables were not included because they did not significantly improve the model's predictive ability. testing the effects of social factors on infant and child outcomes requires samples that are socially and economically diverse. even large, diverse study populations may become more homogeneous over time, as subjects of lower socioeconomic status and non-white race are more likely to drop out of studies dependent on long-term follow-up. and treating socioeconomic variables as statistically independent factors rather than interrelated might minimize the impact of contextual information on neurodevelopmental outcomes. model development. of the papers included in the review, reported on de novo risk prediction tools. the other studies either evaluated a previous model or adjusted a prior model by changing the times at which data were collected or by adding additional variables. the approach to prediction tool development was almost uniform among the studies, with nine of the models solely using regression techniques to select variables. ambalavanan et al. deviated from this method in three separate studies: two using classification tree analysis, , and one using a four-layer back-propagation neural network. each new model-with the exception of the neural networkbased model by ambalavanan et al. , -depended on an approach in which individual variables were selected and treated as independent of one another as they were analyzed in their ability to predict the outcome of interest. yet, variables may, in fact, not act independently. while parsing the roles of potential interrelationships may be computationally onerous and treating them independently may lead to a more parsimonious model, this may be at the expense of accuracy. alternative computational approaches are needed to account for the differential likelihoods of certain outcomes on the causal pathway from preterm birth to later childhood outcome. nonlinear statistical tools should be further utilized in risk prediction model development to examine the relationships between variables and outcomes of interest. machine learning, for instance, is a method of inputting a group of variables and generating a predictive model without making assumptions of independence between the factors or that specific factors would contribute the most to the model. different forms of machine learning have already been employed in nicu's to extract the most important variables for predicting outcomes such as days to discharge. the non-independence of risk factors is also complicated by the role of time in models of human health and development. the lifecourse framework describes how an accumulation or "chains" of risk experienced over time and at certain critical periods impact later health outcomes. in the context of preterm birth, the risk of being born early is not uniform across populations and dependent on a given set of maternal risks. in turn, the degree of prematurity imparts differential risk for developing complications such as bronchopulmonary dysplasia, necrotizing enterocolitis, or retinopathy of prematurity. these morbidities then, in turn, increase risks for further medical and developmental impairment. these time-varying probabilities can be modeled and incorporated into prediction tools to more accurately capture the longitudinal and varying relationships between exposures and outcomes and improve thereby estimations of risk. [ ] [ ] [ ] a final methodological concern regarding model development is whether and how the competing risk of death is considered when the outcome being predicted is non-terminal. consider, for example, the task of developing a model for the risk of ndi at years of age. how one handles death can have a dramatic effect on the model, especially since mortality is relatively high among preterm infants. moreover, if death is treated simply as a censoring mechanism, as it is often done in time-to-event analyses such as those based on the cox model, then the overall risk of ndi will be artificially reduced; those children who die before being diagnosed with ndi will be viewed as remaining at risk even though they cannot possibly be subsequently diagnosed with ndi. while an alternative to this would be to use a composite outcome of time the first of ndi or death, doing so may result in a model that is unable to predict either event well. instead, one promising avenue is to frame the development of a prediction model for ndi within the semi-competing risks paradigm. , briefly, semicompeting risks refer to settings where one event is a competing risk for the other, but not vice versa. this is distinct from standard competing risks, where each event is competing for the other (e.g., death due to one cause or another). to the best of our knowledge, however, semi-competing risks have not been applied to the study of long-term outcomes among preterm infants. model evaluation. waljee et al. provide a summary of methods for assessing the performance of a predictive model, categorizing them into three types: overall model performance, which focuses on the extent of variation in risk explained by the model; calibration, which assesses differences between observed and predicted event rates; and discrimination, which assesses the ability to distinguish between patients who do and do not experience the outcome of interest. the majority of studies in our review assessed their models with roc curve analysis, a method of assessing discrimination. while widely used, there is some debate with regard to roc-based assessments, specifically in regard to its lack of sensitivity in assessing differences between good predictive models. although several novel performance measures for comparing discrimination among models have been proposed, none have been employed in the context of comparing risk prediction tools for preterm neonates. , few studies employed analyses other than roc. only six in our review assessed overall performance with r or partial r , and five evaluated calibration using the hosmer-lemeshow test. another four studies assessed internal validation with either an internal validation set or bootstrapping techniques. there were nine studies meeting inclusion criteria solely because they had models that were externally validated via other studies. schmidt et al. reported odds ratio associations for their -morbidity model, which are not a reliable method of determining the strength of risk prediction tools. future risk model assessments for preterm neonates should at minimum include an roc curve analysis, although assessments of overall performance and calibration would also be helpful. validation with a different sample from the development set is also advised, ideally with a population outside the original cohort. conclusion risk assessment and outcomes prediction are valuable tools in medical decision-making. fortunately, infants born prematurely enjoy ever-increasing likelihood of survival. research over the past several decades has highlighted the many influences, physiologic and psychosocial, affecting neurodevelopment, hrql, and health services utilization. yet, the wealth of knowledge gained from longitudinal studies of growth and development is not reflected in current risk prediction models. moreover, some of the most wellknown and widely used tools today, such as tyson et al.'s fivefactor model, were developed nearly two decades ago. as advances in neonatal intensive care progressively reduce the risk of certain outcomes, it is clear that these older models require updating if they are to be of continued clinical use. it should be recognized that there are potential ethical ramifications to incorporating more psychosocial factors and outcomes into risk prediction models, such as crossing the line from risk stratification to "profiling" patients and offering different treatment decisions based on race or class. however, physician predictions without the aid of prediction tools are highly inconsistent during counseling at the margins of viability, and further research is needed regarding the level of influence that physicians actually have on caregiver decision-making during counseling, as well as the extent to which risk prediction tools would change their approach to counseling. in addition, despite recent innovation in statistical approaches to risk modeling, such as machine learning, most prediction tools rely on standard regression techniques. insofar that risk prediction models will continue to be developed for preterm neonatal care, making use of the clinical data available in most modern electronic health records and taking into consideration the analytic challenges related to unequal prior probabilities of exposures, non-independence of variables, and semi-competing risk can only strengthen our approach to predicting outcomes. we therefore recommend taking a broader view of risk, incorporating these concepts in creating stronger risk prediction tools that can ultimately serve to benefit the long-term care of preterm neonates. c.j.c. and j.s.l. designed and carried out this literature review. c.j.c., j.s.l., and s.h. worked jointly in the analysis and interpretation of the literature review results, as well as the drafting and revision of this article. all three authors gave final approval of the version to be published. on the influence of abnormal parturition, difficult labours, premature birth, and asphyxia neonatorum, on the mental and physical condition of the child, especially in relation to deformities trends in care practices, morbidity, and mortality of extremely preterm neonates survival of infants born at periviable gestational ages outcomes of preterm infants: morbidity replaces mortality institute of medicine committee on understanding premature birth and assuring healthy outcomes. preterm birth: causes, consequences, and prevention influence of birth weight, sex, and plurality on neonatal loss in united states preterm neonatal morbidity and mortality by gestational age: a contemporary cohort gestational age and birthweight for risk assessment of neurodevelopmental impairment or death in extremely preterm infants neurodevelopmental outcome at years of age of a national cohort of extremely low birth weight infants who were born in - comparing neonatal morbidity and mortality estimates across specialty in periviable counseling prognosis and prognostic research: what, why, and how? neonatal disease severity scoring systems intensive care for extreme prematurity-moving beyond gestational age outcome trajectories in extremely preterm infants prediction of late death or disability at age years using a count of neonatal morbidities in very low birth weight infants prediction of mortality in very premature infants: a systematic review of prediction models risk factor models for neurodevelopmental outcomes in children born very preterm or with very low birth weight: a systematic review of methodology and reporting a primer on predictive models pulmonary surfactant therapy the future of exogenous surfactant therapy nursery neurobiologic risk score and outcome at months evaluation of the ability of neurobiological, neurodevelopmental and socio-economic variables to predict cognitive outcome in premature infants. child care health dev increased survival and deteriorating developmental outcome in to week old gestation infants, - compared with - measurement properties of the clinical risk index for babies-reliabilty, validity beyond the first hours, and responsiveness over days predicting the outcomes of preterm neonates beyond the neonatal intensive predicting outcome in very low birthweight infants using an objective measure of illness severity and cranial ultrasound scanning is the crib score (clinical risk index for babies) a valid tool in predicting neurodevelopmental outcome in extremely low birth weight infants? the crib (clinical risk index for babies) score and neurodevelopmental impairment at one year corrected age in very low birth weight infants can severity-of-illness indices for neonatal intensive care predict outcome at years of age? neurodevelopment of children born very preterm and free of severe disabilities: the nord-pas de calais epipage cohort study chronic physiologic instability is associated with neurodevelopmental morbidity at one and two years in extremely premature infants prediction of neurologic morbidity in extremely low birth weight infants impact of bronchopulmonary dysplasia, brain injury, and severe retinopathy on the outcome of extremely low-birth-weight infants at months: results from the trial of indomethacin prophylaxis in preterms impact at age years of major neonatal morbidities in children born extremely preterm effect of severe neonatal morbidities on long term outcome in extremely low birthweight infants early prediction of poor outcome in extremely low birth weight infants by classification tree analysis consequences and risks of < -g birth weight for neuropsychological skills, achievement, and adaptive functioning clinical data predict neurodevelopmental outcome better than head ultrasound in extremely low birth weight infants infant outcomes after periviable birth; external validation of the neonatal research network estimator with the beam trial clinical risk index for babies score for the prediction of neurodevelopmental outcomes at years of age in infants of very low birthweight nsw and act neonatal intensive care units audit group. can the early condition at admission of a high-risk infant aid in the prediction of mortality and poor neurodevelopmental outcome? a population study in australia autism spectrum disorders in extremely preterm children snap-ii and snappe-ii and the risk of structural and functional brain disorders in extremely low gestational age newborns: the elgan study early postnatal illness severity scores predict neurodevelopmental impairments at years of age in children born extremely preterm high prevalence/low severity language delay in preschool children born very preterm identification of extremely premature infants at high risk of rehospitalization screening for autism spectrum disorders in extremely preterm infants perinatal risk factors for neurocognitive impairments in preschool children born very preterm correlation between initial neonatal and early childhood outcomes following preterm birth bronchopulmonary dysplasia and perinatal characteristics predict -year respiratory outcomes in newborns born at extremely low gestational age: a prospective cohort study the international neonatal network. the crib (clinical risk index for babies) score: a tool for assessing initial neonatal risk and comparing performance of neonatal intensive care units score for neonatal acute physiology: a physiologic severity index for neonatal intensive care neonatal therapeutic intervention scoring system: a therapy-based severity-of-illness index prediction of death for extremely premature infants in a population-based cohort parental perspectives regarding outcomes of very preterm infants: toward a balanced approach risk of developmental delay increases exponentially as gestational age of preterm infants decreases: a cohort study at age years preterm birth-associated neurodevelopmental impairment estimates at regional and global levels for late respiratory outcomes after preterm birth respiratory health in pre-school and school age children following extremely preterm birth preterm delivery and asthma: a systematic review and metaanalysis preterm birth, infant weight gain, and childhood asthma risk: a meta-analysis of , european children preterm birth: risk factor for early-onset chronic diseases preterm heart in adult life: cardiovascular magnetic resonance reveals distinct differences in left ventricular mass, geometry, and function right ventricular systolic dysfunction in young adults born preterm elevated blood pressure in preterm-born offspring associates with a distinct antiangiogenic state and microvascular abnormalities in adult life preterm birth and the metabolic syndrome in adult life: a systematic review and meta-analysis prevalence of diabetes and obesity in association with prematurity and growth restriction prematurity: an overview and public health implications measurement of quality of life of survivors of neonatal intensive care: critique and implications quality of life assessment in preterm children: physicians' knowledge, attitude, belief, practice -a kabp study health-related quality of life and emotional and behavioral difficulties after extreme preterm birth: developmental trajectories prognostic factors for poor cognitive development in children born very preterm or with very low birth weight: a systematic review prognostic factors for cerebral palsy and motor impairment in children born very preterm or very low birthweight: a systematic review evidence for catchup in cognition and receptive vocabulary among adolescents born very preterm the economic burden of prematurity in canada changing definitions of long-term followup: should "long term" be even longer? functional outcomes of very premature infants into adulthood social competence of preschool children born very preterm prediction of cognitive abilities at the age of years using developmental follow-up assessments at the age of and years in very preterm children predicting the outcomes of preterm neonates beyond the neonatal intensive perinatal risk factors of adverse outcome in very preterm children: a role of initial treatment of respiratory insufficiency? the relationship between behavior ratings and concurrent and subsequent mental and motor performance in toddlers born at extremely low birth weight prognostic factors for behavioral problems and psychiatric disorders in children born very preterm or very low birth weight: a systematic review neurodevelopmental outcomes of extremely low birth weight infants < weeks' gestation between neighborhood influences on the academic achievement of extremely low birth weight children mental health outcomes in us children and adolescents born prematurely or with low birthweight measurement of socioeconomic status in health disparities research family income trajectory during childhood is associated with adiposity in adolescence: a latent class growth analysis family income trajectory during childhood is associated with adolescent cigarette smoking and alcohol use machine learning in medicine: a primer for physicians predicting discharge dates from the nicu using progress note data a life course approach to chronic diseases epidemiology nd edn. a life course approach to adult health scientists rise up against statistical significance the asa's statement on p-values: context, process, and purpose time for clinicians to embrace their inner bayesian? reanalysis of results of a clinical trial of extracorporeal membrane oxygenation semi-competing risks data analysis: accounting for death as a competing risk when the outcome of interest is nonterminal beyond composite endpoints analysis: semicompeting risks as an underutilized framework for cancer research use and misuse of the receiver operating characteristic curve in risk prediction assessing the performance of prediction models: a framework for traditional and novel measures novel metrics for evaluating improvement in discrimination: net reclassification and integrated discrimination improvement for normal variables and nested models multivariable prognostic models: issues in developing models, evaluating assumptions and adequacy, and measuring and reducing errors limitations of the odds ratio in gauging the performance of a diagnostic, prognostic, or screening marker just health: on the conditions for acceptable and unacceptable priority settings with respect to patients' socioeconomic status auc: . sensitivity: . % specificity: . % competing interests: the authors declare no competing interests.publisher's note springer nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations. key: cord- -nl gsr c authors: tan, chunyang; yang, kaijia; dai, xinyu; huang, shujian; chen, jiajun title: msge: a multi-step gated model for knowledge graph completion date: - - journal: advances in knowledge discovery and data mining doi: . / - - - - _ sha: doc_id: cord_uid: nl gsr c knowledge graph embedding models aim to represent entities and relations in continuous low-dimensional vector space, benefiting many research areas such as knowledge graph completion and web searching. however, previous works do not consider controlling information flow, which makes them hard to obtain useful latent information and limits model performance. specifically, as human beings, predictions are usually made in multiple steps with every step filtering out irrelevant information and targeting at helpful information. in this paper, we first integrate iterative mechanism into knowledge graph embedding and propose a multi-step gated model which utilizes relations as queries to extract useful information from coarse to fine in multiple steps. first gate mechanism is adopted to control information flow by the interaction between entity and relation with multiple steps. then we repeat the gate cell for several times to refine the information incrementally. our model achieves state-of-the-art performance on most benchmark datasets compared to strong baselines. further analyses demonstrate the effectiveness of our model and its scalability on large knowledge graphs. large-scale knowledge graphs(kgs), such as freebase [ ] , yago [ ] and dbpedia [ ] , have attracted extensive interests with progress in artificial intelligence. real-world facts are stored in kgs with the form of (subject entity, relation, object entity), denoted as (s, r, o), benefiting many applications and research areas such as question answering and semantic searching. meanwhile, kgs are still far from complete with missing a lot of valid triplets. as a consequence, many researches have been devoted to knowledge graph completion task which aims to predict missing links in knowledge graphs. knowledge graph embedding models try to represent entities and relations in low-dimensional continuous vector space. benefiting from these embedding models, we can do complicated computations on kg facts and better tackle the kg completion task. translation distance based models [ ] [ ] [ ] [ ] [ ] regard predicting a relation between two entities as a translation from subject entity to tail entity with the relation as a media. while plenty of bilinear models [ ] [ ] [ ] [ ] [ ] propose different energy functions representing the score of its validity rather than measure the distance between entities. apart from these shallow models, recently, deeper models [ , ] are proposed to extract information at deep level. though effective, these models do not consider: . controlling information flow specifically, which means keeping relevant information and filtering out useless ones, as a result restricting the performance of models. . the multi-step reasoning nature of a prediction process. an entity in a knowledge graph contains rich latent information in its representation. as illustrated in fig. , the entity michael jordon has much latent information embedded in the knowledge graph and will be learned into the representation implicitly. however, when given a relation, not all latent semantics are helpful for the prediction of object entity. intuitively, it is more reasonable to design a module that can capture useful latent information and filter out useless ones. at the meantime, for a complex graph, an entity may contain much latent information entailed in an entity, one-step predicting is not enough for complicated predictions, while almost all previous models ignore this nature. multi-step architecture [ , ] allows the model to refine the information from coarse to fine in multiple steps and has been proved to benefit a lot for the feature extraction procedure. in this paper, we propose a multi-step gated embedding (msge) model for link prediction in kgs. during every step, gate mechanism is applied several times, which is used to decide what features are retained and what are excluded at the dimension level, corresponding to the multi-step reasoning procedure. for partial dataset, gate cells are repeated for several times iteratively for more finegrained information. all parameters are shared among the repeating cells, which allows our model to target the right features in multi-steps with high parameter efficiency. we do link prediction experiments on public available benchmark datasets and achieve better performance compared to strong baselines on most datasets. we further analyse the influence of gate mechanism and the length of steps to demonstrate our motivation. link prediction in knowledge graphs aims to predict correct object entities given a pair of subject entity and relation. in a knowledge graph, there are a huge amount of entities and relations, which inspires previous work to transform the prediction task as a scoring and ranking task. given a known pair of subject entity and relation (s, r), a model needs to design a scoring function for a triple (s, r, o), where o belongs to all entities in a knowledge graph. then model ranks all these triples in order to find the position of the valid one. the goal of a model is to rank all valid triples before the false ones. knowledge graph embedding models aim to represent entities and relations in knowledge graphs with low-dimensional vectors (e s , e r , e t ). transe [ ] is a typical distance-based model with constraint formula e s + e r − e t ≈ . many other models extend transe by projecting subject and object entities into relationspecific vector space, such as transh [ ] , transr [ ] and transd [ ] . toruse [ ] and rotate [ ] are also extensions of distance-based models. instead of measuring distance among entities, bilinear models such as rescal [ ] , distmult [ ] and complex [ ] are proposed with multiplication operations to score a triplet. tensor decomposition methods such as simple [ ] , cp-n [ ] and tucker [ ] can also be seen as bilinear models with extra constraints. apart from above shallow models, several deeper non-linear models have been proposed to further capture more underlying features. for example, (r-gcns) [ ] applies a specific convolution operator to model locality information in accordance to the topology of knowledge graphs. conve [ ] first applies -d convolution into knowledge graph embedding and achieves competitive performance. the main idea of our model is to control information flow in a multi-step way. to our best knowledge, the most related work to ours is transat [ ] which also mentioned the two-step reasoning nature of link prediction. however, in transat, the first step is categorizing entities with kmeans and then it adopts a distance-based scoring function to measure the validity. this architecture is not an end-to-end structure which is not flexible. besides, error propagation will happen due to the usage of kmeans algorithm. we denote a knowledge graph as g = {(s, r, o)} ⊆ e × r × e , where e and r are the sets of entities, relations respectively. the number of entities in g is n e , the number of relations in g is n r and we allocate the same dimension d to entities and relations for simplicity. e ∈ r ne * d is the embedding matrix for the schematic diagram of our model with length of step . es and er represent embedding of subject entity and relation respectively. e i r means the query relation are fed into the i-th step to refine information.ẽs is the final output information, then matrix multiplication is operated betweenẽs and embedding matrix of entities e. at last, logistic sigmoid function is applied to restrict the final score between and . entities and r ∈ r nr * d is the embedding matrix for relations. e s , e r and e o are used to represent the embedding of subject entity, relation and subject entity respectively. besides, we denote a gate cell in our model as c. in order to obtain useful information, we need a specific module to extract needed information from subject entity with respect to the given relation, which can be regarded as a control of information flow guided by the relation. to model this process, we introduce gate mechanism, which is widely used in data mining and natural language processing models to guide the transmission of information, e.g. long short-term memory (lstm) [ ] and gated recurrent unit (gru) [ ] . here we adopt gating mechanism at dimension level to control information entailed in the embedding. to make the entity interact with relation specifically, we rewrite the gate cell in multi-steps with two gates as below: two gates z and r are called update gate and reset gate respectively for controlling the information flow. reset gate is designed for generating a new e s or new information in another saying as follows: update gate aims to decide how much the generated information are kept according to formula ( ):ẽ hardmard product is performed to control the information at a dimension level. the values of these two gates are generated by the interaction between subject entity and relation. σ-logistic sigmoid function is performed to project results between and . here means totally excluded while means totally kept, which is the core module to control the flow of information. we denote the gate cell as c. besides, to verify the effectiveness of gate mechanism, we also list the formula of a cell that exclude gates as below for ablation study: with the gate cell containing several gating operations, the overall architecture in one gate cell is indeed a multi-step information controlling way. in fact, a single gate cell can generate useful information since the two gating operations already hold great power for information controlling. however, for a complex dataset, more fine and precise features are needed for prediction. the iterative multi-step architecture allows the model to refine the representations incrementally. during each step, a query is fed into the model to interact with given features from previous step to obtain relevant information for next step. as illustrated in fig. , to generate the sequence as the input for multi-step training, we first feed relation embedding into a fully connected layer: we reshape the output as a sequence [e r , e r , ..., e k r ] = reshape(e r ) which are named query relations. this projection aims to obtain query relations of different latent aspects such that we can utilize them to extract diverse information across multiple steps. information of diversity can increase the robustness of a model, which further benefits the performance. query relations are fed sequentially into the gate cell to interact with subject entity and generate information from coarse to fine. parameters are shared across all steps so multi-step training are performed in an iterative way indeed. our score function for a given triple can be summarized as: where c k means repeating gate cell for k steps and during each step only the corresponding e i r is fed to interact with output information from last step. see fig. for better understanding. after we extract the final information, it is interacted with object entity with a dot product operation to produce final score. in previous rnn-like models, a cell is repeated several times to produce information of an input sequence, where the repeating times are decided by the length of the input sequence. differently, we have two inputs e s and e r with totally different properties, which are embeddings of subject entity and relation respectively, which should not be seen as a sequence as usual. as a result, a gate cell is used for capturing interactive information among entities and relations iteratively in our model, rather than extracting information of just one input sequence. see fig. for differences more clearly. training. at last, matrix multiplication is applied between the final output information and embedding matrix e, which can be called -n scoring [ ] to score all triples in one time for efficiency and better performance. we also add reciprocal triple for every instance in the dataset which means for a given (s, r, t), we add a reverse triple (t, r − , s) as the previous work. we use binary crossentropy loss as our loss function: we add batch normalization to regularise our model and dropout is also used after layers. for optimization, we use adam for a stable and fast training process. embedding matrices are initialized with xavier normalization. label smoothing [ ] is also used to lessen overfitting. in this section we first introduce the benchmark datasets used in this paper, then we report the empirical results to demonstrate the effectiveness of our model. analyses and ablation study are further reported to strengthen our motivation. language system) are biomedical concepts such as disease and antibiotic. • kinship [ ] contains kinship relationships among members of the alyawarra tribe from central australia. the details of these datasets are reported in table . the evaluation metric we use in our paper includes mean reciprocal rank(mrr) and hit@k. mrr represents the reciprocal rank of the right triple, the higher the better of the model. hit@k reflects the proportion of gold triples ranked in the top k. here we select k among { , , }, consistent with previous work. when hit@k is higher, the model can be considered as better. all results are reported with 'filter' setting which removes all gold triples that have existed in train, valid and test data during ranking. we report the test results according to the best performance of mrr on validation data as the same with previous works. table . link prediction results on umls and kinship. for different datasets, the best setting of the number of iterations varies a lot. for fb k and umls the number at provides the best performance, however for other datasets, iterative mechanism is helpful for boosting the performance. the best number of iterations is set to for wn , for wn rr, for fb k- and for kinship. we do link prediction task on benchmark datasets, comparing with several classical baselines such as transe [ ] , distmult [ ] and some sota strong baselines such as conve [ ] , rotate [ ] and tucker [ ] . for smaller datasets umls and kinship, we also compare with some non-embedding methods such as ntp [ ] and neurallp [ ] which learn logic rules for predicting, as well as minerva [ ] which utilizes reinforcement learning for reasoning over paths in knowledge graphs. the results are reported in table and table . overall, from the results we can conclude that our model achieves comparable or better performance than sota models on datasets. even with datasets without inverse relations such as wn rr, fb k- which are more difficult datasets, our model can still achieve comparable performance. to study the effectiveness of the iterative multi-step architecture, we list the performance of different number of steps on fb k- in table . the model settings are all exactly the same except for length of steps. from the results on fb k- we can conclude that the multi-step mechanism indeed boosts the performance for a complex knowledge graph like fb k- , which verify our motivation that refining information for several steps can obtain more helpful information for some complex datasets. we report the convergence process of tucker and msge on fb k- dataset and wn rr dataset in fig. . we re-run tucker with exactly the same settings in table , we report the parameter counts of conve, tucker and our model for comparison. our model can achieve better performance on most datasets with much less parameters, which means our model can be more easily migrated to large knowledge graphs. as for tucker, which is the current sota method, the parameter count is mainly due to the core interaction tensor w , whose size is d e * d r * d e . as the grow of embedding dimension, this core tensor will lead to a large increasing on parameter size. however, note that our model is an iterative architecture therefore only a very few parameters are needed apart from the embedding, the complexity is o(n e d + n r d). for evaluating time efficiency, we re-run tucker and our model on telsa k c. tucker needs s/ s to run an epoch on fb k- /wn rr respectively, msge needs s/ s respectively, which demonstrate the time efficiency due to few operations in our model. to further demonstrate our motivation that gate mechanism and multi-step reasoning are beneficial for extracting information. we do ablation study with the following settings: • no gate: remove the gates in our model to verify the necessity of controlling information flow. • concat: concatenate information extracted in every step together and feed them into a fully connected layer to obtain another kind of final information, which is used to verify that more useful information are produced by the procedure of multi-step. • replicate: replicate the relation to gain k same query relations for training. this is to prove that extracting diverse information from multi-view query relations is more helpful than using the same relation for k times. the experiment results are reported in table . all results demonstrate our motivation that controlling information flow in a multi-step way is beneficial for link prediction task in knowledge graphs. especially a gated cell is of much benefit for information extraction. in this paper, we propose a multi-step gated model msge for link prediction task in knowledge graph completion. we utilize gate mechanism to control information flow generated by the interaction between subject entity and relation. then we repeat gated module to refine information from coarse to fine. it has been proved from the empirical results that utilizing gated module for multiple steps is beneficial for extracting more useful information, which can further boost the performance on link prediction. we also do analysis from different views to demonstrate this conclusion. note that, all information contained in embeddings are learned across the training procedure implicitly. in future work, we would like to aggregate more information for entities to enhance feature extraction, for example, from the neighbor nodes and relations. freebase: a collaboratively created graph database for structuring human knowledge yago : a knowledge base from multilingual wikipedias dbpedia: a nucleus for a web of open data translating embeddings for modeling multi-relational data knowledge graph embedding by translating on hyperplanes learning entity and relation embeddings for knowledge graph completion knowledge graph embedding on a lie group rotate: knowledge graph embedding by relational rotation in complex space a three-way model for collective learning on multi-relational data embedding entities and relations for learning and inference in knowledge bases complex embeddings for simple link prediction simple embedding for link prediction in knowledge graphs tensor factorization for knowledge graph completion convolutional d knowledge graph embeddings modeling relational data with graph convolutional networks reasonet: learning to stop reading in machine comprehension gated-attention readers for text comprehension knowledge graph embedding via dynamic mapping matrix canonicaltensor decomposition for knowledge base completion translating embeddings for knowledge graph completion with relation attention mechanism long short-term memory learning phrase representations using rnn encoder-decoder for statistical machine translation rethinking the inception architecture for computer vision observed versus latent features for knowledge base and text inference statistical predicate invention end-to-end differentiable proving differentiable learning of logical rules for knowledge base reasoning go for a walk and arrive at the answer: reasoning over paths in knowledge bases using reinforcement learning key: cord- - pm rpzj authors: parnell, gregory s.; smith, christopher m.; moxley, frederick i. title: intelligent adversary risk analysis: a bioterrorism risk management model date: - - journal: risk anal doi: . /j. - . . .x sha: doc_id: cord_uid: pm rpzj the tragic events of / and the concerns about the potential for a terrorist or hostile state attack with weapons of mass destruction have led to an increased emphasis on risk analysis for homeland security. uncertain hazards (natural and engineering) have been successfully analyzed using probabilistic risk analysis (pra). unlike uncertain hazards, terrorists and hostile states are intelligent adversaries who can observe our vulnerabilities and dynamically adapt their plans and actions to achieve their objectives. this article compares uncertain hazard risk analysis with intelligent adversary risk analysis, describes the intelligent adversary risk analysis challenges, and presents a probabilistic defender–attacker–defender model to evaluate the baseline risk and the potential risk reduction provided by defender investments. the model includes defender decisions prior to an attack; attacker decisions during the attack; defender actions after an attack; and the uncertainties of attack implementation, detection, and consequences. the risk management model is demonstrated with an illustrative bioterrorism problem with notional data. toward risk-based regulations, specifically using pra to analyze and demonstrate lower cost regulations without compromising safety. ( , ) research in the nuclear industry has also supported advances in human reliability analysis, external events analysis, and common cause failure analysis. ( − ) more recently, leaders of public and private organizations have requested risk analyses for problems that involve the threats posed by intelligent adversaries. for example, in , the president directed the department of homeland security (dhs) to assess the risk of bioterrorism. ( ) homeland security presidential directive (hspd- ): biodefense for the st century, states that " [b] iological weapons in the possession of hostile states or terrorists pose unique and grave threats to the safety and security of the united states and our allies" and charged the dhs with issuing biennial assessments of biological threats to "guide prioritization of our on-going investments in biodefense-related research, development, planning, and preparedness." a subsequent homeland security presidential directive (hspd- ): medical countermeasures against weapons of mass destruction directed an integrated risk assessment of all chemical, biological, radiological, and nuclear (cbrn) threats. ( ) the critical risk analysis question addressed in this article is: are the standard pra techniques for uncertain hazards adequate and appropriate for intelligent adversaries? as concluded by the nrc ( ) study on bioterrorism risk analysis, we believe that new techniques are required to provide credible insights for intelligent adversary risk analysis. we will show that treating adversary decisions as uncertain hazards is inappropriate because it can provide a different risk ranking and may underestimate the risk. in the rest of this section, we describe the difference between natural hazards and intelligent adversaries and demonstrate, with a simple example, that standard pra applied to attacker's intent may underestimate the risk of an intelligent adversary attack. in the second section, we describe a canonical model for resource allocation decision making for an intelligent adversary problem using an illustrative bioterrorism example with notional data. in the third section, we describe the illustrative analysis results obtained from the model and discuss the insights they provide for risk assessment, risk communication, and risk management. in the fourth section, we describe the benefits and limitations of the model. finally, we discuss future work and our conclusions. we believe that risk analysis of uncertain hazards is fundamentally different than risk analysis of intelligent adversaries. ( , ) some of the key differences are summarized in table i . ( ) a key difference is historical data. for many uncertain events, both natural and engineered, we have not only historical data of extreme failures or crises, but many times we can replicate events in a laboratory environment for further study (engineered systems) or analyze using complex simulations. intelligent adversary attacks have a long historical background, but the aims, events, and effects we have recorded may not prove a valid estimate of future threat because of changes in adversary intent and capability. both uncertain hazard risks of occurrence and geographical risk can be narrowed down and identified concretely. intelligent adversary targets vary by the goals of the adversary and can be vastly dissimilar between adversary attacks. information sharing between the two events differs dramatically. after hurricanes or earthquakes, engineers typically review the incident, publish results, and improve their simulations. sometimes after intelligent adversary attacks, or near misses, the situation and conduct of the attack may involve critical state vulnerabilities and protected intelligence means. in these cases, intelligence agencies may be reluctant to share complete information even with other government agencies. the ability to influence the event is also different. though we can prepare, we typically have no way of influencing the natural event to occur or not occur. on the other hand, governments may be able to affect the impact of terrorism attacks by a variety of offensive, defensive, and recovery measures. in addition, adversary attacks can take on so many forms that one cannot realistically defend/respond/recover/etc. against all types of attacks. although there have been efforts to use event tree technologies in intelligent adversary risk analysis (e.g., btra), many believe that this approach is not credible. ( ) the threat from intelligent adversaries comes from a combination of both intent and capability. we believe that pra still has an important role in intelligent adversary risk analysis for assessment of the capabilities of adversaries, the vulnerabilities of potential targets, and potential consequences of attacks. however, intent is not a some cities may be considered riskier than others (e.g., new york city, washington), but terrorists may attack anywhere, any time. information sharing: asymmetry of information: new scientific knowledge on natural hazards can be shared with all the stakeholders. governments sometimes keep secret new information on terrorism for national security reasons. natural event: intelligent adversary events: to date, no one can influence the occurrence of an extreme natural event (e.g., an earthquake). governments may be able to influence terrorism (e.g., foreign policy; international cooperation; national and homeland security measures). government and insureds can invest in well-known mitigation measures. attack methodologies and weapon types are numerous. local agencies have limited resources to protect potentially numerous targets. federal agencies may be in a better position to develop better offensive, defensive and response strategies. modified from kunreuther. ( , ) - . ( ) factor in natural hazard risk analysis. in intelligent adversary risk analysis, we must consider the intent of the adversary. the adversary will make future decisions based on our preparations, its objectives, and information about its ability to achieve its objectives that is dynamically revealed in a scenario. bier et al. provides an example of addressing an adversary using a defender-attacker game theoretic model. ( ) nrc provides three examples of intelligent adversary models. ( ) we believe it will be more useful to assess an attacker's objectives (although this is not a trivial task) than assigning probabilities to their decisions prior to the dynamic revelation of scenario information. we believe that modeling adversary objectives will provide greater insight into the possible actions of opponents rather than exhaustively enumerating probabilities on all the possible actions they could take. furthermore, we believe the probabilities of adversary decisions (intent) should be an output of, not an input to, risk analysis models. ( ) this is a principal part of game theory as shown in aghassi et al. and jain et al. ( , ) to make our argument and our proposed alternative more explicit, we use a bioterrorism illustrative example. in response to the hspd, in october , the dhs released a report called the bioterrorism risk assessment (btra). ( ) the risk assessment model contained a -step event tree ( steps with consequences) that could lead to the deliberate exposure of civilian populations for each of the most dangerous pathogens that the center for disease control tracks (emergency.cdc.gov/bioterrorism) plus one engineered pathogen. the model was extremely detailed and contained a number of separate models that fed into the main btra model. the btra resulted in a normalized risk for each of the pathogens, and rank-ordered the pathogens from most risky to least risky. the national research council (nrc) conducted a review of the btra model and provided specific recommendations for improvement to the model. ( ) in our example, we will use four of the recommendations: model the decisions of intelligent adversaries, include risk management, simplify the model by not assigning probabilities to the branches of uncertain events, and do not normalize the risk. the intelligent adversary technique we developed builds on the deterministic defenderattacker-defender model and is solved using decision trees. ( ) because the model has been simplified to reflect the available data, the model can be developed in a commercial off-the-shelf (cots) software package, such as the one we used for modeling, dpl (www.syncopation.org). other decision analysis software may work as well. ( ) event trees have been useful for modeling uncertain hazards. ( ) however, there is a key difference in the modeling of intelligent adversary decisions that event trees do not capture. as norman c. rasmussen, the director of the reactor safety study that validated pra for use in nuclear reactor safety, states in a later article, while the basic assumption of randomness holds true for nuclear safety, it is not valid for human action. ( ) the attacker makes decisions to achieve his or her objectives. the defender makes resource allocation decisions before and after an attack to try to mitigate vulnerabilities and consequences of the attacker's actions. this dynamic sequence of decisions made by first the defender, then an attacker, then again by the defender should not be modeled solely by assessing probabilities of the attacker's decisions. for example, when the attacker looks at the defender's preparations for their possible bioterror attack, it will not assign probabilities to its decisions; it chooses the agent and the target based on perceived ability to acquire the agent and successfully attack the target that will give it the effects it desires to achieve its objectives. ( ) representing an attacker decision as a probability may underestimate the risk. consider the simple bioterrorism event tree given in fig. with notional data. using an event tree, for each agent (a and b) there is a probability that an adversary will attack, a probability of attack success, and an expected consequence for each outcome (at the terminal node of the tree). the probability of success a useful reference for decision analysis software is located on the orms website (http://www.lionhrtpub.com/orms/surveys/das/ das.html). involves many factors, including the probability of obtaining the agent and the probability of detection during attack preparations and execution. the consequences depend on many factors, including agent propagation, agent lethality, time to detection, and risk mitigation; in this example, the consequences range from or no consequences to , the maximum consequences (on a normalized scale of consequences). calculating expected values in fig. , we would assess expected consequences of . we would be primarily concerned about agent b because it contributes % of the expected consequences ( × . = for b out of the total of ). looking at extreme events, we would note that the worst-case consequence of has a probability of . . however, adversaries do not assign probabilities to their decisions; they make decisions to achieve their objectives, which may be to maximize the consequences they can inflict. ( ) if we use a decision tree as in fig. , we replace the initial probability node with a decision node because this is an adversary decision. we find that the intelligent adversary would select agent a, and the expected consequences are , which is a different result than with the event tree. again, if we look at the extreme events, the worstcase event ( consequences) probabilities are . for the agent a decision and . for the agent b decision. the expected consequences are greater and the primary agent of concern is now a. in this simple example, the event tree approach underestimates the expected risk and provides a different risk ranking. furthermore, the event tree example underestimates the risk of the extreme events. however, while illustrating important differences, this simple decision tree model does not sufficiently model the fundamental structure of intelligent adversary risk. this model has a large number of applications for homeland security. for example, it would be easy to see the use of this canonical model applied to a dirty bomb example laid out in rosoff and von winterfeldt ( ) or any other intelligent adversary scenario. in this article, we show a use of a bioterrorism application. we believe that the canonical risk management model must have at least six components: the initial actions of the defender to acquire defensive capabilities, the attacker's uncertain acquisition of the implements of attack (e.g., agents a, b, and c), the attacker's target selection and method of attack(s) given implement of attack acquisition, the defender's risk mitigation actions given attack detection, the uncertain consequences, and the cost of the defender actions. from this model, one could also conduct baseline risk analysis by looking at the status quo. in general, the defender decisions can provide offensive, defensive, or information capabilities. we are not considering offensive decisions such as preemption before an attack; instead, we are considering decisions that will increase our defensive capability (e.g., buy vaccine reserves) ( ) or provide earlier or more complete information for warning of an attack (add a bio watch city). ( ) in our defenderattacker-defender decision analysis model, we have the two defender decisions (buy vaccine, add a bio watch city), the agent acquisition for the attacker is uncertain, the agent selection and target of attack is another decision, the consequences (fatalities and economic) are uncertain, the defender decision after attack to mitigate the maximum possible casualties, and the costs of defender decisions are known. the defender risk is defined as the probability of adverse consequences and is modeled using a multiobjective additive model similar to multiobjective value models. ( ) we have assumed that the defender minimizes the risk and the attacker maximizes the risk. we implemented this model as a decision tree (fig. ) and an influence diagram (fig. ) using dpl. the mathematical formulation of our model and the notional data are provided in the appendix the illustrative decision tree model (figs. and ) begins with decisions that the defender (united states) makes to deter the adversary by reducing the vulnerabilities or be better prepared to mitigate a bioterrorism attack of agents a, b, or c. we modeled and named the agents to represent notional bioterror agents using the cdc's agent categories in table ii . for example, agent a represents a notional agent from category a. table iii provides a current listing of the agents by category. there are many decisions that we could model; however, for our simple illustrative example, we chose to model notional decisions about the bio watch program for agents a and b and the bioshield vaccine reserve for agent a. bio watch is a program that installs and monitors a series of passive sensors within a major metropolitan city. ( ) the bioshield program is a plan to purchase and store vaccines for some of the more dangerous pathogens. ( ) the defender first decides whether or not to add another city to the bio watch program. if that city is attacked, this decision could affect the warning time, which influences the response and, ultimately, the potential consequences of an attack. of course, the bio watch system does not detect every agent, so we modeled agent c to be the most effective agent that the bio watch system does not sense and provide additional warning. adding a city will also incur a cost in dollars for the united states. the second notional defender decision is the amount of vaccine to store for agent a. agent a is the notional agent that we have modeled with the largest probability of acquisition and potential consequences. the defender can store a percentage of what experts think we would need in a largescale biological agent attack. the more vaccine the united states stores, the fewer consequences we will have if the adversaries use agent a and we have sufficient warning and capability to deploy the vaccine reserve. however, as we store more vaccine, the costs for purchasing and storage increase. for ( ) category definition a the u.s. public health system and primary healthcare providers must be prepared to address various biological agents, including pathogens that are rarely seen in the united states. high-priority agents include organisms that pose a risk to national security because they: can be easily disseminated or transmitted from person to person; result in high mortality rates and have the potential for major public health impact; might cause public panic and social disruption; and require special action for public health preparedness. b second highest priority agents include those that: are moderately easy to disseminate; result in moderate morbidity rates and low mortality rates; and require specific enhancements of cdc's diagnostic capacity and enhanced disease surveillance. c third highest priority agents include emerging pathogens that could be engineered for mass dissemination in the future because of: availability; ease of production and dissemination; and potential for high morbidity and mortality rates and major health impact. simplicity's sake, each of the defender decisions cost the same amount; therefore, at the first budget level of us$ million, the defender can only choose to one decision. after the defender has made its investment decisions, which we assume are known to the attacker, the attacker makes two decisions: the type of agent and the target. we will assume that the attacker has already made the decision to attack the united states with a bioterror agent. in our model, there are three agents it can choose, although this can be increased to represent the other pathogens listed in table iii . as stated earlier, if we only looked at the attacker decision, agent a would appear to be the best choice. agents b and c are the next two most attractive agents to the attacker, respectively. again, agents a and b can be detected by bio watch whereas agent c cannot. the attacker has some probability of acquiring each agent. if the agent is not acquired, the attacker cannot attack with that agent. in addition, each agent has a lethality associated with it, represented by the agent casualty factor. finally, each agent has a different probability of being detected over time. generally, the longer it takes for the agent to be detected, the more consequences the united states will suffer. the adversary also decides what size of population to target. generally, the larger the population targeted, the more potential consequences could result. the attacker's decisions affect the maximum possible casualties from the scenario. however, regardless of the attacker's decisions, there is some probability of actually attaining a low, medium, or high percentage of the maximum possible casualties. this sets the stage for the next decision by the defender. after receiving warning of an attack, the defender decides whether or not to deploy the agent a vaccine reserve. this decision depends upon how much of the vaccine reserve the united states chose to store, whether the attacker used agent a, and the potential effectiveness of the vaccine given timely attack warning. in addition, there is a cost associated with deploying the vaccine reserve. again, for simplicity's sake, the cost for every defender decision is the same, thus forcing the defender to only choose the best option(s) for each successive us$ million increase in budget up to the maximum of us$ million. in our model (fig. ) , we have two types of consequences: casualties and economic impact. given the defender-attacker-defender decisions, the potential casualties and the economic impact are assessed. casualties are based on the agent, the population attacked, the maximum potential casualties, the warning time given to the defender, and the effectiveness of vaccine for agent a (if the agent a is the agent and the vaccine is used). economic effects are modeled using a linear model with a fixed economic cost that does not depend on the number of casualties and a variable cost of the number of casualties multiplied by the cost per casualty. of course, the defender would like potential consequences (risk) given an attack to be low, whereas the attacker would like the potential consequences (risk) to be high. our economic consequences model was derived using a constant and upper bound from wulf et al. ( ) the constant cost we used is $ billion, and from the upper bound, also given in wulf et al., we derived the cost per casualty. ( ) we believe this fixed cost is reasonable because when looking at the example of the anthrax letters of , experts estimate that although there were only infected and five killed, there was a us$ billion cost to the united states. ( ) in this tragic example, there was an extremely high economic impact even when the casualties were low. each u.s. defender decision incurs a budget cost. the amount of money available to homeland security programs is limited by a budget determined by the president and congress. the model will track the costs incurred and only allows spending within the budget (see the appendix). we considered notional budget levels of us$ million, us$ million, us$ million, and us$ million. normally, a decision tree is solved by maximizing or minimizing the expected attribute at the terminal branches of the tree. in our model however, the defender risk depends on the casualty and economic consequences given an attack. we use multiple objective decision analysis with an additive value (risk) model to assign risk to the defender consequences. the defender is minimizing risk and the attacker is maximizing risk. we assign a value of . to no consequences and a value of . to the worst-case consequences in our model. we model each consequence with a linear risk function and a weight (see the appendix). the risk functions measure returns to scale on the consequences. of course, additional consequences could be included and different shaped risk curves could be used. some of the key assumptions in our model are listed in table iv (the details are in the appendix) along with some possible alternative assumptions. given different assumptions, the model may produce different results. we model the uncertainty of the attacker's capability to acquire an agent with a probability distribution and we vary detection time by agent. clearly, other indications and warnings exist to detect possible attacks. these programs could be included in the model. we model three defender decisions: add a bio watch city for agents a and b, increase vaccine reserve for agent a, and deploy agent a. we assume limited decision options (i.e., % storage of vaccine a, % storage, % storage), but other decisions could be modeled (e.g., other levels of storage, storing vaccines for other agents, etc). we used one casualty model for all agents. other casualty and economic models could be used. finally, our model makes some assumptions about objectives. in the first of these we assume that the risks important to the defender are the number of casualties and the economic impact, but additional measures could be used. second, we assume defenders and attackers have a diametrically opposed view of all of the objectives. clearly, we could model additional objectives. in addition, we made some budget assumptions, which could be improved or modified. we assumed a fixed budget, but this budget could be modeled with more detailed cost models (e.g., instead of a set amount to add a city to the bio watch program, the budget could reflect different amounts depending upon the city and the robustness of the sensors installed). finally, our model results in a risk of a terrorist attack; the same methodology for a defender-attacker-defender decision tree can be used to determine a utility score instead of a risk; an example of this is in keeney. ( ) one thing to consider when altering or adding to the assumptions is the number of strategies the model evaluates. currently, the canonical model has different strategies to evaluate (table v) . with more complexity, the number of strategies that would need to be evaluated could grow significantly. largescale decision trees can be solved with monte carlo simulation. after modeling the canonical problem, we obtained several insights. first, we found that in our model economic impact and casualties are highly correlated. higher casualties result in higher economic impact. other consequences, for example, psychological consequences, could also be correlated with casualties. second, a bioterror attack could have a large economic impact (and psychological impact), even if casualties are low. the major risk analysis results are shown in fig. . risk shifting occurs in our decision analysis model. in the baseline (with no defender spending), agent a is the most effective agent for the attacker to select and, therefore, the agent against which the defender can protect if the budget is increased. as we improve our defense against agent a, at some point the attacker will choose to attack using agent b. the high-risk agent will have shifted from agents a to b. as the budget level continues to increase, the defender adds a city to the bio watch program and the attackers choose to attack with agent c, which bio watch cannot detect. we use notional data in our model, but if more realistic data were used, the defender could determine the cost/benefit ratios of additional risk reduction decisions. this decision model uses cots software to quantitatively evaluate the potential risk reductions associated with different options and make cost-benefit decisions. fig. provides a useful summary of the expected risk. however, it is also important to look at the complementary cumulative distribution (fig. ) to better understand the likelihood of extreme outcomes. for example, the figure shows that spending us$ or us$ million gives the defender a % chance of zero risk, whereas spending us$ or us$ million gives the defender an almost % chance of having zero risk. the best risk management result would be that option deterministically or stochastically dominates (sd) option , option sd option , and option sd option . the first observation we note from fig. is that options , , and stochasically dominate because option has a higher probability for each risk outcome. a second observation is that while option sd option , option does not sd option because option has a larger probability of yielding a risk level of . . along the x-axis, one can see the expected risk (er) of each alternative. this expected risk corresponds to the expected value of risk from the budget versus risk rainbow diagram in fig. . this example illustrates a possibly important relationship necessary for understanding and communicating how the budget might affect the defender's risk and choice of options. risk managers can run a value of control or value of correlation diagram to see which nodes most directly affect the outcomes and which are correlated (fig. ) . because we only have two uncertainty nodes in our canonical model, the results are not surprising. the graphs show that the ability to acquire the agent is positively correlated with the defender risk. as the probability of acquiring the agent increases, so does defender risk. in addition, the value of control shows the amount of risk that could be reduced given perfect control over each probabilistic node, and that it is clear that acquiring the agent would be the most important variable for risk managers to control. admittedly, this is a basic example, but with a more complex model, analysts could determine which nodes are positively or negatively correlated with risk and which uncertainties are most important. using cots software also allows us to easily perform sensitivity analysis on key model assumptions. from the value of correlation and control above, the probability of acquiring the agent was highly and positively correlated with defender risk and had the greatest potential for reducing defender risk. we can generate sensitivity analysis such as rainbow diagrams. the rainbow diagram (fig. ) shows the decision changes as our assumption about the probability of acquiring agent a increases. the different shaded regions represent different decisions, for both the attacker and the defender. this rainbow diagram was produced using a budget level of us$ million, so in the original model, the defender would choose not to add a city to bio watch, store % of vaccine for agent a, but not choose to deploy it because the attacker chose to use agent b. if the probability of acquiring agent a was low enough (in section a from fig. ) , we see that the attacker would choose to use agent c because we have spent our money on adding another city to bio watch, which is the only thing that affects both agents a and b, but not agent c. as the probability of acquiring agent a increases, both the attacker's and the defender's optimal strategies change. our risk management decision depends on the probability that the adversary acquires agent a. risk analysis of intelligent adversaries is fundamentally different than risk analysis of uncertain hazards. as we demonstrated in section . , assigning probabilities to the decisions of intelligent adversaries can underestimate the potential risk. decision tree models of intelligent adversaries can provide insights into the risk posed by intelligent adversaries. the defender-attacker-defender decision analysis model presented in this article provides four important benefits. first, it provides a risk assessment (the baseline or status quo) based on defender and attacker objectives and probabilistic assessment of threat capabilities, vulnerabilities, and consequences. second, it provides information for risk-informed decision making about potential risk management options. third, using cots software, we can provide a variety of very useful sensitivity analysis. fourth, although the model would be developed by a team, the risk analysis can be conducted by one risk analyst with an understanding of decision trees and optimization and training on the features of the cots software. the application of risk assessment and risk management techniques should be driven by the goals of the analysis. in natural hazard risk analysis, there is value in performing risk assessment without risk management. some useful examples are "unfinished business," a report from the epa and the u.k. national risk register. ( , ) in intelligent adversary risk analysis, the defender-attacker-defender decision analysis can provide essential information for risk management decision making. in our example, risk management techniques are important, and this type of model provides insights about resource allocation decisions to reduce or shift risk. in addition, with budget set to us$ , the model can be used to assess the baseline risk. as the budget increases, the model clearly shows the best risk management decisions and the associated risk reduction. this model enables the use of cots risk analysis software. in addition, the use of cots software enables the use of standard sensitivity analysis tools to provide insights into areas in which the assumptions are critical or where the model should be improved or expanded. currently, many event tree models including the dhs btra event tree require extensive contractor support to run, compile, and analyze. ( ) although one would still need a multidisciplinary team to create the model, once completed the defenderattacker-defender decision analysis model is usable by a single risk analyst who can provide near realtime analysis results to stakeholders and decisionmakers as long as the risk analyst understands the risk management options, decision trees, optimization, and has training in the cots tool. the technique we advocate in this article has limitations. some of the limitations of this model are the same as those of event trees. there are limitations on the number of agents used in the models. we easily modeled bioagents with reasonable run times, but more agents could be modeled. in addition, there are challenges in assessing the probabilities of uncertain events, for example, the probability that the attacker acquires agent a. next, there is a limitation in the modeling of the multiple consequences. another limitation may be that to get more realistic results, we may have to develop "response surface" models of more complex consequence models. these limitations are shared by event trees and decision trees. however, decision trees also have some limitations that are not shared by event trees. first, only a limited number of risk management decisions can realistically be modeled. therefore, care must be used to choose the most appropriate set of potential decisions. ( , ) in addition, there may be an upper bound on the number of decisions or events that can be modeled in cots software. it is important to note that it may be difficult to determine an objective function for the attacker. as mentioned before, there is a tradeoff in replacing the probabilities assigned to what an attacker might do (event tree approach) with attacker objectives (decision tree approach). we believe it is easier to make informed assessments about the objectives of adversaries than to assess the probabilities of their future actions. however, we need more research on assessing the robustness of risk management decisions to assumptions about adversary objectives. finally, successful model operation and interpretation requires trained analysts who understand decision analysis and defenderattacker-defender optimization. this article has demonstrated the feasibility of modeling intelligent adversary risk using defenderattacker-defender decision analysis. table iv and section . identified several alternative modeling assumptions that could be considered. we can modify and expand our assumptions to increase the complexity and fidelity of the model. the next step is to use the model with the best data available on the agents of concern and a proposed set of potential risk management options. use of our defender-attacker-defender model does not require a major intelligent adversary research program; it requires only the willingness to change. ( ) much of the data used for event tree models can be used in the decision analysis model. assessing probabilities of attacker decisions will not increase our security but defender-attacker-defender decision analysis models can provide a sound assessment of risk and the essential information our nation needs to make risk-informed decisions. g.s.p. is grateful for the many helpful discussions on intelligent adversary risk analysis with his colleagues on the nrc committee and the defender-attacker-defender research of jerry brown and his colleagues at the naval postgraduate school. the authors are grateful for the dpl modeling advice provided by chris dalton of syncopation. the authors thank roger burk at the united states military academy for his useful reviews and suggestions. finally, the authors thank the area editor and reviewers for very detailed comments and suggestions that have helped us improve our article. this model is a multiobjective decision analysis/game theory model that allows for risk management at the u.s. governmental level in terms of budgeting and certain bioterror risk mitigation decisions. the values for probabilities as well as factors are notional and could easily be changed based on more accurate data. it uses the starting u.s. (defender) decisions of adding a city to the bio watch program (or not) and the percent of storing an agent in the nation's vaccine reserve program to set the conditions for an attacker decision. the attacker can choose which agent to use as well as what size of population to target. there is some unpredictability in the ability to acquire the agent as well as the effects of the agent given the defender and attacker decisions. finally, the defender gets to choose whether to deploy the vaccine reserve to mitigate casualties. the model tracks the cost for each u.s. decision and evaluates them over a specified budget. the decisions cannot violate the budget without incurring a dire penalty. the objectives that the model tracks are u.s. casualties and impact to the u.s. economy. they are joined together using a value function with weights for each objective. we outline our model using a method suggested by brown and rosenthal. ( ) probabilistic risk assessment: reliability engineering, design, and analysis risk analysis in engineering and economics risk modeling, assessment, and management reactor safety study: assessment of accident risk in u.s. commercial nuclear plants nuclear regulatory commission (usnrc) fault tree handbook understanding risk management: a review of the literature and industry practice a survey of risk assessment methods from the nuclear, chemical, and aerospace industries for applicability to the privatized vitrification of hanford tank wastes procedural and submittal guidance for the individual plant examination of external events (ipeee) for severe accident vulnerabilities procedure for analysis of common-cause failures in probabilistic safety analysis a technique for human error analysis (atheana) homeland security presidential directive homeland security presidential directive [hspd- ]: medical countermeasures against weapons of mass destruction guiding resource allocations based on terrorism risk modeling values for anti-terrorism analysis department of homeland security's bioterrorism risk assessment: a call for change, committee on methodological improvements to the department of homeland security's biological agent risk analysis insurability of (mega)-terrorism risk: challenges and perspectives. report prepared for the oecd task force on terrorism insurance, organization for economic cooperation and development integrating risk management with homeland security and antiterrorism resource allocation decision-making. chapter in kamien d (ed). the mcgraw-hill handbook of homeland security fort detrick, md: biological threat characterization center of the national biodefense analysis and countermeasures center choosing what to protect: strategic defensive allocation against an unknown attacker robust game theory robust solutions in stackelberg games: addressing boundedly rational human preference models. association for the advancement of artificial intelligence workshop: - syncopation software. available at probabilistic modeling of terrorist threats: a systems analysis approach to setting priorities among countermeasures probabilistic risk assessment: its possible use in safeguards problems. presented at the institute for nuclear materials management meeting nature plays with dice-terrorists do not: allocating resources to counter strategic versus probabilistic risks a risk and economic analysis of dirty bomb attacks on the ports of los angeles and long beach project bioshield: protecting americans from terrorism the bio watch program: detection of bioterrorism strategic decision making: multiobjective decision analysis with spreadsheets bioterrorist agents/diseases definitions by category center for disease control (cdc) emerging and re-emerging infectious diseases strategic alternative responses to risks of terrorism world at risk: report of the commission on the prevention of wmd proliferation and terrorism unfinished business: a comparative assessment of environmental problems optimization tradecraft: hard-won insights from real-world decision support. interfaces key: cord- -vlv x l authors: ul-haq, zaheer; saeed, maria; halim, sobia ahsan; khan, waqasuddin title: d structure prediction of human β -adrenergic receptor via threading-based homology modeling for implications in structure-based drug designing date: - - journal: plos one doi: . /journal.pone. sha: doc_id: cord_uid: vlv x l dilated cardiomyopathy is a disease of left ventricular dysfunction accompanied by impairment of the β( )-adrenergic receptor (β( )-ar) signal cascade. the disturbed β( )-ar function may be based on an elevated sympathetic tone observed in patients with heart failure. prolonged adrenergic stimulation may induce metabolic and electrophysiological disturbances in the myocardium, resulting in tachyarrhythmia that leads to the development of heart failure in human and sudden death. hence, β( )-ar is considered as a promising drug target but attempts to develop effective and specific drug against this tempting pharmaceutical target is slowed down due to the lack of d structure of homo sapiens β( )-ar (hsβadr ). this study encompasses elucidation of d structural and physicochemical properties of hsβadr via threading-based homology modeling. furthermore, the docking performance of several docking programs including surflex-dock, fred, and gold were validated by re-docking and cross-docking experiments. gold and surflex-dock performed best in re-docking and cross docking experiments, respectively. consequently, surflex-dock was used to predict the binding modes of four hsβadr agonists. this study provides clear understanding of hsβadr structure and its binding mechanism, thus help in providing the remedial solutions of cardiovascular, effective treatment of asthma and other diseases caused by malfunctioning of the target protein. g-protein coupled receptor (gpcr) superfamily constitutes the largest family of receptors in cell responsible for mediating the effects of over % of drugs in the market now-a-days [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] . gpcrs are involved in the transmission of a variety of signals to the interior of the cell and can be activated by a diverse range of small molecules including nucleotides, amino acids, peptides, proteins and odorants. activation of gpcrs results in a conformational change followed by a signal cascade that passes information to the inside of the cell by interacting with a protein known as heterotrimeric g-proteins. there are three main classes of gpcrs (a, b and c) depending on their sequence similarity to rhodopsin (rho) (class a). class a gpcrs is the largest group and encompasses a wide range of receptors including receptors for odorants, adenosine, β-adrenergic and rhodopsin [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] . the β-adrenergic receptors (β-ars) are g s protein-coupled receptors that play important roles in cardiovascular function and disease, through serving as receptors for the neurohormones: norepinephrine and epinephrine. norepinephrine released from cardiac sympathetic nerves activates myocyte β -ars, which activates adenylyl cyclase via stimulatory g-protein (g s ). the rise in the intracellular [camp] level causes the phosphorylation of several intracellular proteins by means of camp-dependent protein kinase a. such type of activated β -ar results in an increased cardiac inotropy, lusitropy, and chronotropy and the secretion of rennin, all of which contribute to regulate the cardiac functions and blood pressure [ ] [ ] . β -ar predominates in the heart, representing about % of the myocardial β-ars; thus, they tend to be viewed as the most significant β-ars with respect to the cardiovascular system. β -and β -ars in kidneys stimulate the release of renin, thereby playing a role in the activation of renin-angiotensin-aldosterone system [ ] [ ] . the role of β-ars in cardiovascular function and disease is also highlighted by the significant roles of drugs whose actions are based on binding to the β-ars blockers (β-blockers). βblockers represent first line therapy for the management of chronic heart failure, hypertension, acute and post myocardial infarction patients, chronic stable angina, and unstable angina [ ] . they are also commonly used to control the symptoms of atrial fibrillation and other arrhythmias [ ] . there are no cardiovascular drugs that have a wider range of indications than βblockers, making them a critical drug class for the management of cardiovascular disease. the availability of uses for β-blockers also suggests that the activation of β-ars, or the sympathetic nervous system (sns), plays an essential function in most cardiovascular diseases. the fact that β -ar selective antagonists are equivalent to non-selective blockers in essentially all situations provides additional evidence that β -ars are the more important β-receptors with respect to cardiovascular disease. the development of a large number of rational inhibitors that have the ability to modulate the activity of such receptors has been a major goal for the pharmaceutical industries to improve the clinical treatment of various disease including hypertension, heart failure and asthma [ ] . however, finding specific drug against a particular β-ars drug target is a slow and laborious process. furthermore, the lack of d structure of hsadrb is an obstacle in the identification of specific drug like molecules. on the other hand, the development of computational approaches for drug designing can be effectively carried out with low cost [ ] [ ] . the use of computational techniques in drug discovery and development process is rapidly gaining popularity, implementation and appreciation. there will be an intensifying effort to apply computational power to combine biological and chemical space in order to rationalize the drug discovery, designing and optimization phenomena. today, computer aided drug design (cadd) is based on the knowledge of structure, either of the receptor, or that of the ligand. the former is described as structure-based while the later as ligand-based drug designing. because it is difficult and time-consuming to obtain experimental structures from methods such as x-ray crystallography and protein nmr for every protein of interest, homology modeling is a widely used in silico technique providing the useful structural models for generating hypotheses about a protein's function and directing further experimental work [ ] . the main objective of this study is to employ "in silico" homology modeling technique to construct the d structure of hsadrb that will be used to identify and characterize new inhibitors for hsadrb by structure-based computational approaches. this model serves as a starting point to gain knowledge of protein-ligand interactions and the structural requirements of active site of protein. computational studies were performed on intel xeon quad core ( . ghz processor) server installed with linux os (opensuse version . ). multiple sequence alignment was carried out by clustalw of the closest homologue identified by ncbi p-blast to find out the identity, similarity and gap region between the target and template [ ] . homology modeling was accomplished by orchestrar [ ] implemented in biopolymer module of sybyl . [ ] . an online server, i-tasser [ ] , was used for modeling a region absent in template structure. the finally selected model of hsβadr was minimized by amber (version . ) [ ]. stereochemical properties of modeled protein structure were validated by procheck [ ] , veri-fy d [ ] and errat [ ] . molecular docking experiments were conducted by surflex-dock implemented in sybyl (version . ) [ ] , fred (version . . ) [ ] and gold (version . ) [ ] . ucsf chimera [ ] [ ] and moe [ ] were used for visualization purpose. the flowchart of work plan is illustrated in (fig ) . search the closest homologue. top-ranked template sequences as determined by blast were subjected for multiple sequence alignment on the basis of optimized e-value of the specified target sequence (table ) . however, meleagris gallopavo β -ar (mgβadr , pdb id: y ) retrieved as the closest homologue, was manually edited for optimal alignment along with the target sequence. best alignment was selected based on alignment score and the reciprocal position of the conserved amino acid residues across the members of class a gpcr superfamily. the confined ballesteros and weinstein numbering scheme [ ] was used to identify the transmembrane (tm) segments relative to the conserved position of amino acids in tm helices assigned as locant. shareing the common features in all class a gpcr superfamily. this is followed by the representation of amino acids tm helix numbers. the immediately preceding and following the . residue are numbered . and . , respectively. orchestrar is specifically designed for homology or comparative protein modeling that identifies structurally conserved regions (scrs), models loops using model-based and ab-initio methods, models side chains, and combine them all to prepare a final model. initially, a homology model was generated by orchestrar that lacks a region of amino acid residues ( - ) of the cytoplasmic loop of tm located within the target sequence but absent in the template structure. this region was modeled by i-tasser, an integrated platform for automated protein structure and function whose prediction is based on sequence-to-structureto-function paradigm as per multiple threading alignments by lomets [ ] . the model generated by i-tasser was named as sub-model . five sub-models were evaluated by replica-exchange monte carlo simulations with low free-energy states, spatial restrains and alignments tm regions [ ] to identify the best structural alignment almost closed to the structural analogs on the basis of structural similarity. any further steric clashes were removed to refine the coordinates, and the final results of all sub-models were based on sequence-structurefunction paradigm obtained from the consensus of structural similarity and confidence score (c-score) of i-tasser server. c-score value is the quality for the predicted sub-model on the basis of threading method. stereochemical properties of each sub-model were evaluated and the best selected sub-model was incorporated to the homology model of hsβadr , generated previously by orchestrar and after insertion of the model the finalized modelled is subjected for optimization. homology model of hsβadr generated by orchestrar was minimized by sybyl using conjugate gradient and steepest descent method with , iterations each. the selected submodel generated by i-tasser was also individually minimized to , cycles by amber , followed by the insertion of sub-model into the homology model of hsβadr by chain joining option in sybyl. the finally generated model is minimized further to , cycles using ff sb force field by amber . selection of complexes for re-docking and cross-docking validation. to identify a suitable docking program for the docking of hsβadr agonists, re-docking and cross-docking experiments were performed by surflex-dock, fred, and gold. six βadr -ligand complexes, three βadr -ligand complexes and two rhodopsin-ligand complexes were retrieved from pdb. the details of the protein-ligand complexes used in this study are summarized in table and s fig. selection of complexes was based on following criteria: availability of the protein-ligand complexes, the crystallographic resolution of protein-ligand complexes should be Ǻ, the binding interaction of the protein-ligand complexes should be known. cross-docking experiments conducted in using multiple docking methods with their scoring function are utilized in this study mentioned in s , s and s tables. the details of docking methodology are discussed in supporting informations. the re-docking results were analyzed to check the ability of docking programs to correctly identify the bound conformation of co-crystallized ligand in the top-ranked solution. rmsds were calculated between the corresponding co-crystallized ligand against its predicted docked pose. cross-docking experiments were conducted to identify which docking program exactly identified its cognate ligand among the diverse set of ligands within the top-ranked solution. for cross-docking, complexes were extracted from pdb in which eight proteins are homodimers (chain a and chain b) while the rest of three are monomers (chain a). for those proteins that are present as homodimers, ligands were docked into both chains. overall, complexes were evaluated for cross-docking. the results were quantified as best ( - position), moderate ( - position) and worst when the cognate ligand ranks position lowers than within its cognate protein, respectively. blast results and multiple sequence alignments blast predicted mgβadr (pdb id: y ) [ ] as the best match for hsβadr with % identity and % positivity (with an e-value of × - ). r r, kj , p g and rh have % while r s and sn have % query coverage, more sequence coverage than observed for y ( %). since r r, kj and rs are available as apo form with fewer scores, identity and positive values, these structures were not used in this study. similarly, the complexes p g, rh and sn were not used for the modeling of hsβadr structure due to their lower scores. hence, y is used according to the blast results but to establish more confidence on the top-ranked search, we opted for two sorts of multiple sequence alignments: raw multiple sequence alignment and manually-edited multiple sequence alignment. for raw alignment, the ten top-ranked templates sequences ( y , vt , r r, kj , r s, sn , gbr, p g, rh and pds) were aligned against the target sequence illustrated in s fig and s table. for manually-edited alignment, both the target and template ( y ) sequences were truncated. the first residues from n-terminus and residues ( - ) from c-terminus were omitted from the target sequence due to the absence of corresponding homologous sequence in the template and has no important residue which is necessary to be in helical segments. the template sequence has amino acid residues whereas the structure of - residues has been resolved (total residues as some residues are missing). the first residues ( - ) from n-terminus and residues ( - ) from c-terminus were omitted. finally, residues of target sequence was aligned with ten top-ranked blast search, y ( residues), vt , r r, kj , r s, sn , gbr, p g, rh and pds the average alignment score for manually edited multiple sequence alignment is better ( . ) than the score obtained by raw multiple sequence alignment ( . ). overall, there are instances where alignments are improved, alignments are improved when the target sequence is aligned with the rest of the sequences and times the alignments have better quality when the template sequence is aligned with the remaining sequences. the generalized ballesteros and weinstein numbering scheme is beneficial for the understanding, recognition and structural alignments of gpcrs family [ ] . the ballesteros and weinstein numbering is illustrated in (fig ) and the conserved amino acid residues of class a gpcrs is tabulated in table . ballesteros and weinstein numbering is useful for the understanding of integrating methods for the construction of d models and computational probing of structure-function relations in gpcrs. these criteria pertain to the selection of correct inputs for the a ignment programs and to structural considerations applicable to checking and refining the sequence alignments generated by alignment programs. this selection criterion depends on the information that is determined by the extent of homology among the compared sequences. alignment of sequences with intermediate homologies (i.e., - %) can identify continuous patterns of conservation distributed over the entire sequence. such patterns provide structural inferences based on conservation. the hsβadr model is selected after structural comparison, superimposition and procheck results ( fig a) . orchestrar generated homology model using template y was incomplete since the structure of residues - was missing. orchestrar fills the gap but not more than - residues long, therefore, an ab-initio based threading method is used to predict the structure of missing region (s fig). subsequently, five sub-models were generated. each sub-model is further analyzed by ramachandran plot (table ). among them, sub-model is selected on the basis of highest c-score (- . ) and stereochemical properties. the c-score value being lower than - . likely indicates a lack of an appropriate template within the i-tasser library. the selected sub-model was subsequently inserted into the homology model of hsβadr by sybyl. the c-terminus val and the n-terminus lys of the homology model is connected with the n-terminus val and the c-terminus thr of submodel , respectively ( fig b) . according to the ramachandran plot,~ %, . % and . % residues are located within the favorable, allowed and the generously allowed regions, respectively while only one residue (ile ) is found to be in the disallowed region. the visual inspection revealed that ile is far away from the active site region and do not lie within Å of active site. additionally, stereochemical properties of the model were validated by verify d web server. verify d evaluated the local environment and inter-residue contacts in the model. ideally, the d- d profile for each of the amino acids should be in range of - . . values less than zero are considered as inaccurate for the homology model. the verify d plot of hsβadr model shows that the average score of all amino acid residues is . which is relatively closed to . . moreover, errat, a protein structure verification web server was used to verify the model on the basis of model building and refinement, and is extremely useful in making decisions about reliability of the homology model. errat results showed that the overall quality factor for the hsβadr model is . %., suggesting that the generated model is robust and can be use for virtual screening purpose in future. the d model of hsβadr revealed an excellent agreement with the experimentally determined d structure of mgβadr . (fig ) shows the superimposed view of hsβadr model and mgβadr structure. the calculated polypeptide backbone (cα, these pdbs have comparable sequence similarities, identities and source as that of the template but some conformational changes has been observed for helix [ ] . however, we found no observable conformational changes especially for those amino acid residues that are involved in molecular interactions with high-affinity antagonists i , p and cau located within h and cl- . finally, the hsβadr model is subjected to the sequence manipulation suite ident and sim [ ] to observe the similarity and identity of the model with respect to the template structure. the results are better but afterwards much improved after manual editing of the target and template sequences, similarity and identity ratios are increased from % to . % and % to . %, respectively. the overall topology and secondary structural elements particular for the class a gpcr family remain quite conserved in the model of hsβadr , that is, an extracellular n-terminus domain, seven -tm domains linked by three intracellular cytoplasmic loops (cl- , cl- and cl- ), three extracellular loops (el- , el- and el- ), and a cytoplasmic c-terminus domain. the nterminus domain comprises of nine amino acids residues ( - ) that are located outside the membrane. the tm- -tm- helices spans from - , - , - , - , - , - and - amino acid residues, respectively, while the c-terminus domain comprises of amino acid residues ranging from to at the inner face of membrane. the cytoplasmic loops, (cl- , cl- and cl- ) comprise of residues - , - and - , respectively. the cytoplasmic loops cl- and cl- are believed to be important in the binding, selectivity or specificity and activation of g-proteins [ ] . the extracellular loops, (el- , el- and el- ) comprising - , - and - residues, respectively. two conserved disulfide bridges which are important for cell surface expression, ligand binding, receptor activation and maintenance of the secondary structure are located in el- and el- regions at positions cys -cys and cys -cys , respectively (table ). conserved motifs of hsβadr homology model dry motif also known as ionic lock switch [ ] is observed at position asp ( . ), arg ( . ) and tyr ( . ) in helix of hsβadr model. the conserved asp in dry motif at the cytoplasmic end of helix believes to regulate the transition state of active state, while the adjacent arg is crucial for g-protein activation [ ] . another conserved penta-peptide npxxy motif known as tyrosine toggle switch (where x usually represents a hydrophobic residue and n is rarely exchanged against d) located at the c terminus of tm- which contributes to gpcr internalization and signal transduction. several site-directed mutagenesis studies revealed the importance of this motif in signaling [ ] . the npxxy motif is present at position arg ( . ), pro ( . ), ileu ( . ), ileu ( . ) and tyr ( . ) in the [ ] . these domains help anchor tm to the cytoskeleton and hold together signaling complexes. pdz domain have many functions, from regulating the trafficking and targeting of proteins to assembling signaling complexes, and networks designed for efficient and specific signal transduction [ ] . the amphipathic amino acid residues present in helix are conserved among all human gpcrs (residues - ), located between the tm bound with helix . the palmitoylation occurs at n-terminus while the biosynthesis of receptor and the proper regulation of surface expression occur at c-terminus of hsβadr . the side chain of two crucial residues of helix , asp ( . ) and arg ( . ) , are projected within the hydrophilic interface involved in stimulatory g-protein (g s ) activation while the residue phe ( . ) and phe ( . ) are buried in the hydrophobic core of the helix [ ] . salt bridges play important roles in protein structure and function. disruption and the introduction of a salt bridge reduce and increase the stability of the protein, respectively [ ] . in membrane proteins, one expects salt bridges to be particularly important because of a smaller dehydration penalty (loss of favorable contacts with water) on salt bridge formation [ ] . charged groups become largely dehydrated when inserted into membranes, and therefore, experience a smaller change in hydration between non-salt-bridging and salt-bridging states. there should also be a smaller effect because of solvent screening, strengthening salt-bridge interactions [ ] . multiple salt bridges are observed in the homology model of hsβadr ; asp :arg , asp :arg , asp :lys , glu :arg , glu :lys and glu : arg . in addition, salt bridges can also serve as key interactions in much the same way as disulfide bonds (s fig) . re-docking analysis. the success of docking is usually scrutinized by its accurate pose prediction ability [ ] [ ] , hence prior to the docking of βadr agonists into the homology model of hsβadr , the reliability of three docking programs including surflex-dock, fred, and gold was assessed. the re-docking results were quantified on the basis of rmsd between the top-ranked docked conformation and the co-crystallized (termed as ˈreferenceˈ) ligand and visual analysis. the prediction is termed as good when rmsd > or < Å and the docked pose is superimposed on the ligand's co-crystallized position, fair when rmsd > and < Å and the docked pose is in active site but not superimposed on its native conformation, and poor or inaccurate when rmsd > Å and the docked pose is inverted or far away from the active site. the re-docking results showed that gold outperformed surflex-dock and fred (fig ) . among the complexes used, gold, surflex-dock, and fred generated %, %, and % good solutions in the top-ranked position, respectively. surflex-dock and fred identified % and % fair poses, respectively in the top-ranked docked poses. while both the programs generated % inaccurate solutions in the top-ranked docked pose. the results are summarized in (table ) . furthermore, docking methods utilized in cross-docking is illustrated in (table ) , was conducted to find out which program utilized in correctly ranks ligands into their cognate binding site. the prediction was quantified on the basis of ligand's ranking (s , s and s ) tables. the cross-docking results indicates that surflex-dock is superior with % best results in ranking the ligand in top - position in their cognate receptors. gold and fred are returned with % and % best results, respectively (fig ) . the position and the interaction of each ligand within the cognate receptor are visually analyzed. the results showed that the conformation of each ligand generated by surflex-dock is much better than the docked conformations generated by gold and fred. most of the interactions generated by surflex-dock are similar to the interactions present in the x-ray conformation. hence, surflex-dock was found to be more appropriate for the docking of gpcr's ligands and it is further used in this study to explore the binding mode of hsβadr agonists into the active site of hsβadr model. table and table . binding mode of y , whj, fw, h the docked pose of y reveals that multiple hydrogen bonding interactions are formed between the surrounding amino acid residues that stabilize y in the catechol binding pocket. the −oh group at the phenol moiety is involved in hydrogen bonding with the γ carboxylate side chain of asp ( . Å). the substituted −oh group at meta and para positions of ring b shows hydrogen bonding interactions with the side chains γ −oh of thr ( . Å) and ser ( . Å), respectively. furthermore, the side chain phenyl ring of phe and the carboxylate of asp provide cation-π stacking interactions to the phenolic moiety of y that further helps to stabilize the orientation of agonist. (fig a) displays the binding mode of compound y . the binding mode of whj demonstrates that the amino group of whj mediates hydrogen bond with the side chain carboxylate of asp at a distance of . Ǻ. similarly, thr γ -oh group probes hydrogen bonding interactions with multiple ligand atoms including n atom and o atom at a distance of . Ǻ and . Ǻ, respectively. the same thr is also involved in forming hydrogen bond at a distance of . Ǻ, the most significant hydrogen bonding interaction for whj. phe forms cation-π interaction with one of the fused aromatic ring of whj. the binding orientation of compound whj is shown in (fig b) . the binding mode of fw shows that the para −oh moiety of fw establishes hydrogen bonding interaction with the side chain carboxylate of ser at a distance of . Ǻ. additionally; ser forms bi-dentate hydrogen bonding with the para and meta −oh groups at distances of . Ǻ and . Ǻ, respectively. the main chain carbonyl moiety of phe mediate hydrogen bond with the amino group of fw ( . Ǻ). the docked binding mode of compound fw is depicted in (fig c) . as revealed in (fig d) , the -oh of h shows similar interactions as observed for compound fw. the para substituted −oh group of h mediates bi-dentate hydrogen bonds with the side chain −oh groups of ser and ser at distances of . Ǻ and . Å, respectively. furthermore, asp mediates bi-dentate interaction with the linear chain amino and −oh groups of h at a distance of . Ǻ and . Ǻ, respectively. the binding mode analysis of agonists y , fw, h displays that the ser plays crucial role in stabilizing the agonists within the catechol binding pocket of hsβadr homology model. the docking results reveals that ser and phe are crucial residues in ligand binding by providing h-bonding, and π-π interactions, respectively, thus helps in the activation of hsβadr . we intend to incorporate molecular dynamic simulation studies in order to investigate the dynamic behavior of protein-inhibitor complex formation in the near future; and the role of most important residues will be determined. the study will be helpful to pursue structure based drug design of hsβadr blockers. human βadr is found to be involved in several cardiovascular diseases. the lack of crystal structure of hsβadr provoked us to apply in silico techniques to initiate the drug discovery process for hsβadr . hence, to understand the characteristics structural features of hsβadr and to execute the structure-based drug design strategy for hsβadr , threading-based homology modeling of mammalian origin were applied in this study. the model possesses acceptable structural profiles. furthermore, the binding modes of four hsβadr agonist were determined via molecular docking simulation. h- , h- , and el- regions were found to be important in ligand binding. several residues including trp , asp , val , asp , phe , thr , ser , and ser are involved in direct interactions with the ligand. among all, ser , and phe provides h-bonding, π-π interactions, respectively, hence found to be crucial residue in ligand binding and for the activation of hsβadr . we are also investigating the dynamic behaviour of the apo and ligand bound forms of hsβadr that will be published in future. note: the coordinate file of hsadrb is submitted to the publicly accessible protein model database (pmdb) [ ] ; www.caspur.it/pmdb). the pmdb id of hsadrb is pm respectively. table and (fig ) ). (tif) s table. alignment scores (a) alignment scores obtained from the alignment scores raw multiple sequence alignment (b) alignment scores obtained from the manually edited multiple sequence alignment (c) alignment scores obtained from the raw target and template pair wise sequence alignment. (doc) s table. cross-docking results of surflex-dock analyzed the basis of ranking of the cognate ligand in their respective receptor. criteria for ranking: - position is best (green cell), - is moderate (blue cell) and > is inaccurate (red cell). (doc) s table. cross-docking results of fred analyzed on the basis of ranking of the cognate ligand in their respective receptor. criteria for ranking: - position is best, - is moderate and > is inaccurate. (doc) s table. cross-docking results of gold analyzed on the basis of ranking of the cognate ligand in their respective receptor. (doc) location and nature of the residues important for ligand recognition in g-protein coupled receptors identification of g protein-coupled receptor genes from the human genome sequence the structure and function of g-protein-coupled receptors the year in g protein-coupled receptor research g protein-coupled receptors: the inside story an overview on gpcrs and drug discovery: structure-based drug design and structural biology on gpcrs i want a new drug: g-protein-coupled receptors in drug development the impact of gpcr structures on pharmacology and structure-based drug design role of β adrenergic receptor polymorphisms in heart failure: systematic review and meta-analysis functional responses of human beta adrenoceptors with defined haplotypes for the common r>g and s>g polymorphisms medical therapy can improve the biological properties of the chronically failing heart. a new era in the treatment of heart failure pharmacogenetics of the human beta-adrenergic receptors predicting molecular interactions in silico: i. a guide to pharmacophore identification and its applications to drug design virtual screening for sars-cov protease based on kz pharmacophore points structure-based d-qsar models and dynamics analysis of novel n-benzyl pyridinone as p α map kinase inhibitors for anticytokine activity clustalw and clus-talx version . comparison of composer and orchestrar sybyl molecular modeling software version . , tripos associates i-tasser: a unified platform for automated protein structure and function prediction procheck: a program to check the stereochemical quality of protein structures verify d: assessment of protein models with three-dimensional profiles verification of protein structures: patterns of nonbonded atomic interactions surflex: fully automatic flexible molecular docking using a molecular similarity-based search engine fred pose prediction and virtual screening accuracy gold version . . cambridge crytallographic data center chimera: an extensible molecular modeling application constructed using standard components molecular operating environment (moe), . chemical computing group inc. sherbooke st. west, suite # uniprot website basic local alignment search tool integrated methods for the construction of three-dimensional models and computational probing of structure-function relations in g protein-coupled receptors i-tasser: fully automated protein structure prediction in casp tm-align: a protein structure alignment algorithm based on the tm-score the structural basis for agonist and partial agonist action on a beta -adrenergic receptor two distinct conformations of helix observed in antagonist-bound structures of a β -adrenergic receptor the sequence manipulation suite: javascript programs for analyzing and formatting protein and dna sequences structure of a beta -adrenergic g-protein-coupled receptor the significance of g protein-coupled receptor crystallography for drug discovery the effect of mutations in the dry motif on the constitutive activity and structural instability of the histamine h receptor mutation of tyrosine in the conserved npxxy sequence leads to constitutive phosphorylation and internalization, but not signaling, of the human b bradykinin receptor hallucinogen actions on -ht receptors reveal distinct mechanisms of activation and signaling by g protein-coupled receptors homology modeling of g-protein-coupled receptors and implications in drug design pdz domains and their binding partners: structure, specificity, and modification evolutionary expansion and specialization of the pdz domains characterization of the residues in helix of the human β -adrenergic receptor that are involved in coupling the receptor to g proteins esbri: a web server for evaluating salt bridges in proteins salt bridges: geometrically specific, designable interactions. proteins structure-based design, synthesis and biological evaluation of β-glucuronidase inhibitors identification of novel interleukin- inhibitors through computational approaches the pmdb protein model database structure of bovine rhodopsin in a trigonal crystal form advances in determination of a highresolution three-dimensional structure of rhodopsin, a model of g-protein-coupled receptors (gpcrs) high resolution crystal structure of an engineered human β -adrenergic g protein coupled receptor a specific cholesterol binding site is established by the . Å structure of the human β -adrenergic receptor conserved binding mode of human β adrenergic receptor inverse agonists and antagonist revealed by x-ray crystallography we are thankful to prof. bernd m. rode (university of innsbruck) for the computational software support during this research work. financial support required to conduct this scientific work from higher education commission (hec), is highly acknowledged. supporting information s fig. d representation of the bound ligands of gpcrs complexes utilized in this study (see also table ). key: cord- -s v bmd authors: subramanian, vigneshwar; kattan, michael w. title: editorial: why is modeling covid- so difficult? date: - - journal: chest doi: . /j.chest. . . sha: doc_id: cord_uid: s v bmd nan as the covid- pandemic continues with no clear end in sight, clinicians, policymakers, and the public alike are searching for answers about where we're headed. several independent models of disease spread and mortality have been published, and they often came to different conclusions. modeling covid- has proven to be a complex and difficult endeavor. it is useful for this discussion to differentiate between epidemiologic and individual prediction models. the former category predicts outcomes for a population, such as a state or country, whereas the latter category predicts outcomes for individual patients. each type of model has a set of constraints involved in its design and use. the cdc publishes an ensemble forecast of national deaths based on nineteen independently developed models, all of which fall into the epidemiologic category . predicting mortality in a population over time is a very complicated task and requires a number of assumptions to be made. of note, the biological properties of this virus are imperfectly understood given how novel it is. for instance, the duration and degree of immunity conferred by exposure will determine the likelihood and intensity of recurrent outbreaks. presumably to mitigate this issue, the models in the cdc ensemble make projections over a four-week period. mortality is fundamentally a function of the fatality rate and spread of disease, each of which is driven by a multitude of factors. fatality rates depend in part on population demographics and availability of testing, and currently available estimates vary widely across countries . disease spread depends heavily on the prevalence of covid- , which is not precisely known, and on policy interventions such as social distancing, which are a moving target and not intrinsically measurable. policy changes rapidly, and as memorial day crowds show, the degree of compliance with restrictions is also unknown. emerging studies using antibody tests suggest that prevalence is likely highly underestimated , but these serologic tests are themselves imperfect, and validated performances can vary broadly depending on the specific assay that is used . to get around this, models either use proxies or make assumptions about the degree and effectiveness of social distancing. for example, the university of texas model uses phone geolocation data as a proxy for social distancing and assumes the intervention remains constant across the forecasted time period . conversely, the columbia model considers three scenarios, all of which assume a flat % reduction in contact rates for each week of stay-at-home orders, with either a one-time % increase in contact rates, a weekly % increase in contact rates, or no change in contact rates after stay-at-home orders are lifted . these design choices significantly alter predicted outcomes, and it is impossible to know which assumptions will most closely match reality. assumptions may also change over time as information emerges and their performance is reassessed; for example, the columbia model updated contact tracing assumptions to the current parameters to model loosening social distancing restrictions as states reopen . individual predictions models, which make diagnostic or prognostic predictions about a disease based on a patient's unique set of characteristics, have their own set of challenges. the general workflow involved in developing such a model is as follows: first, the outcome of interest is defined; second, relevant predictors or risk factors are identified; third, the effects of each predictor variable are estimated, for example in a regression analysis; and finally, the model is validated . validation is comprised of discrimination and calibration, which are the ability to separate individuals who experience the outcome from those who do not and how well the predicted probabilities match reality, respectively. each of these steps presents a challenge in the context of any emerging disease, such as covid- . first, many clinically useful outcomes of interest, such as progression to the icu, need for a ventilator, or death, are time-to-event, and therefore we may need to consider censoring. in addition, relevant risk factors are still emerging, and many are difficult to measure, such as travel history and symptom duration (symptom onset is often uncertain). we can identify candidate predictors by generalizing from models of other respiratory viruses, but these need to be validated, and their relevance may vary depending on the population that is studied. once an outcome is defined and predictors are selected, we need a large data set of patients that includes all the desired features, so that we can estimate regression coefficients or train a machine learning algorithm. thanks to electronic medical records, an enormous amount of information will eventually be available, but these data needs to be collected and cleaned for analysis. dimensionality of the data also becomes an issue: as more predictors are considered, a greater number of observations is required to assess significance (a common rule of thumb being ten observations per predictor). validation of the model is also critical. we can use the training data set itself for internal validation using methods such as bootstrapping and ten-fold crossvalidation. ideally, we would also externally validate models on an independent cohort, but this takes time and is not always done. when considering the value of covid- models, there are several questions we must consider. the strength of epidemiologic models is that they allow us to flexibly consider a range of possible scenarios and examine how outcomes change as our assumptions shift. no single model represents truth, but the combined spectrum of projections can guide us. similarly, individual models can help clinicians design a management plan that has the best predicted outcome for each unique patient. a model cannot and does not have to be perfect, but it should be a useful approximation. the pandemic raises an interesting question: is it more ethical to use a model that is not validated, or to press forward with no model? as a corollary, if a clinical model has been extensively internally validated, but not externally validated, should it be used in clinical practice? models can be helpful decision-making aids during our uncertain future but we must keep in mind the assumptions that were built in and the quality of the underlying data. covid- forecasts johns hopkins coronavirus resource center covid- antibody seroprevalence in santa clara county, california. medrxiv eua authorized serology test performance projections for first-wave covid- deaths across the us using social-distancing measures derived from mobile phones. medrxiv projection of covid- cases and deaths in the us as individual states re-open. medrxiv personalized and precision medicine informatics: a workflow-based view. health informatics key: cord- - vjg e authors: awan, ammar ahmad; jain, arpan; anthony, quentin; subramoni, hari; panda, dhabaleswar k. title: hypar-flow: exploiting mpi and keras for scalable hybrid-parallel dnn training with tensorflow date: - - journal: high performance computing doi: . / - - - - _ sha: doc_id: cord_uid: vjg e to reduce the training time of large-scale deep neural networks (dnns), deep learning (dl) scientists have started to explore parallelization strategies like data-parallelism, model-parallelism, and hybrid-parallelism. while data-parallelism has been extensively studied and developed, several problems exist in realizing model-parallelism and hybrid-parallelism efficiently. four major problems we focus on are: ) defining a notion of a distributed model across processes, ) implementing forward/back-propagation across process boundaries that requires explicit communication, ) obtaining parallel speedup on an inherently sequential task, and ) achieving scalability without losing out on a model’s accuracy. to address these problems, we create hypar-flow—a model-size and model-type agnostic, scalable, practical, and user-transparent system for hybrid-parallel training by exploiting mpi, keras, and tensorflow. hypar-flow provides a single api that can be used to perform data, model, and hybrid parallel training of any keras model at scale. we create an internal distributed representation of the user-provided keras model, utilize tf’s eager execution features for distributed forward/back-propagation across processes, exploit pipelining to improve performance and leverage efficient mpi primitives for scalable communication. between model partitions, we use send and recv to exchange layer-data/partial-errors while allreduce is used to accumulate/average gradients across model replicas. beyond the design and implementation of hypar-flow, we also provide comprehensive correctness and performance results on three state-of-the-art hpc systems including tacc frontera (# on top .org). for resnet- , an ultra-deep model, hypar-flow provides: ) up to . [formula: see text] speedup over horovod-based data-parallel training, ) [formula: see text] speedup over single-node on stampede nodes, and ) [formula: see text] speedup over single-node on frontera nodes. recent advances in machine/deep learning (ml/dl) have triggered key success stories in many application domains like computer vision, speech comprehension and recognition, and natural language processing. large-scale deep neural networks (dnns) are at the core of these state-of-the-art ai technologies and have been the primary drivers of this success. however, training dnns is a compute-intensive task that can take weeks or months to achieve state-of-theart prediction capabilities (accuracy). these requirements have led researchers to resort to a simple but powerful approach called data-parallelism to achieve shorter training times. various research studies [ , ] have addressed performance improvements for data-parallel training. as a result, production-grade ml/dl software like tensorflow and pytorch also provide robust support for data-parallelism. while data-parallel training offers good performance for models that can completely reside in the memory of a cpu/gpu, it can not be used for models larger than the memory available. larger and deeper models are being built to increase the accuracy of models even further [ , ] . figure highlights how memory consumption due to larger images and dnn depth limits the compute platforms that can be used for training; e.g. resnet- k [ ] with the smallest possible batch-size of one (a single × image) needs . gb memory and thus cannot be trained on a gb pascal gpu. similarly, resnet- k on image size × needs gb of memory, which makes it out-of-core for most platforms except cpu systems that have gb memory. these out-of-core models have triggered the need for model/hybrid parallelism. however, realizing model-parallelism-splitting the model (dnn) into multiple partitions-is non-trivial and requires the knowledge of best practices in ml/dl as well as expertise in high performance computing (hpc). we note that model-parallelism and layer-parallelism can be considered equivalent terms when the smallest partition of a model is a layer [ , ] . little exists in the literature about model-parallelism for state-of-the-art dnns like resnet(s) on hpc systems. combining data and model parallelism, also called hybridparallelism has received even less attention. realizing model-parallelism and hybrid-parallelism efficiently is challenging because of four major problems: ) defining a distributed model is necessary but difficult because it requires knowledge of the model as well as of the underlying communication library and the distributed hardware, ) implementing distributed forward/back-propagation is needed because partitions of the model now reside in different memory spaces and will need explicit communication, ) obtaining parallel speedup on an inherently sequential task; forward pass followed by a backward pass, and ) achieving scalability without losing out on a model's accuracy. proposed approach: to address these four problems, we propose hypar-flow: a scalable, practical, and user-transparent system for hybrid-parallel training on hpc systems. we offer a simple interface that does not require any modeldefinition changes and/or manual partitioning of the model. users provide four inputs: ) a model defined using the keras api, ) number of model partitions, ) number of model replicas, and ) strategy (data, model, or hybrid). unlike existing systems, we design and implement all the cumbersome tasks like splitting the model into partitions, replicating it across processes, pipelining over batch partitions, and realizing communication inside hypar-flow. this enables the users to focus on the science of the model instead of system-level problems like the creation of model partitions and replicas, placement of partitions and replicas on cores and nodes, and performing communication between them. hypar-flow's simplicity from a user's standpoint and its complexity (hidden from the user) from our implementation's standpoint is shown in fig. . from a research and novelty standpoint, our proposed solution is both model-size as well as model-type agnostic. it is also different compared to all existing systems because we focus on high-level and abstract apis like keras that are used in practice instead of low-level tensors and matrices, which would be challenging to use for defining state-of-the-art models with hundreds of layers. hypar-flow's solution to communication is also novel because it is the first system to exploit standard message passing interface (mpi) primitives for inter-partition and inter-replica communication instead of reinventing single-use libraries. to the best of our knowledge, there are very few studies that focus on hybridparallel training of large dnns; especially using tensorflow and keras in a user-transparent manner for hpc environments where mpi is a dominant programming model. we make the following key contributions in this paper: [ ] is a language for distributed dl with an emphasis on tensors distributed across a processor mesh. mtf only works with the older tf apis (sessions, graphs, etc.). furthermore, the level at which mtf distributes work is much lower compared to hypar-flow, i.e., tensors vs. layers. users of mtf need to re-write their entire model to be compatible with mtf apis. unlike mtf, hypar-flow works on the existing models without requiring any code/model changes. we summarize these related studies on data, model, and hybrid-parallelism and their associated features in table . out-of-core methods like [ , ] take a different approach to deal with large models, which is not directly comparable to model/hybrid-parallelism. several data-parallelism only studies have been published that offer speedup over sequential training [ , , , , ] . however, all of these are only limited to models that can fit in the main memory of the gpu/cpu. we provide the necessary background in this section. training itself is an iterative process and each iteration happens in two broad phases: ) forward pass over all the layers and ) back-propagation of loss (or error) in the reverse order. the end goal of dnn training is to obtain a model that has good prediction capabilities (accuracy). to reach the desired/target accuracy in the fastest possible time, the training process itself needs to be efficient. in this context, the total training time is a product of two metrics: ) the number of epochs required to reach the target accuracy and ) the time required for one epoch of training. in data-parallel training, the complete dnn is replicated across all processes, but the training dataset is partitioned across the processes. since the model replicas on each of the processes train on different partitions of data, the weights (or parameters) learned are different on each process and thus need to be synchronized among replicas. in most cases, this is done by averaging the gradients from all processes. this synchronization is performed by using a collective communication primitive like allreduce or by using parameter servers. the synchronization of weights is done at the end of every batch. this is referred to as synchronous parallel in this paper. model and hybrid-parallelism: data-parallelism works for models that can fit completely inside the memory of a single gpu/cpu. but as model sizes have grown, model designers have pursued aggressive strategies to make them fit inside a gpu's memory, which is a precious resource even on the latest volta gpu ( gb). this problem is less pronounced for cpu-based training as the amount of cpu memory is significantly higher ( gb) on the latest generation cpus. nevertheless, some models can not be trained without splitting the model into partitions; hence, model-parallelism is a necessity, which also allows the designers to come up with new models without being restricted to any memory limits. the entire model is partitioned and each process is responsible only for part (e.g. a layer or some layers) of the dnn. model-parallelism can be combined with data-parallelism as well, which we refer to as hybrid-parallelism. we expand on four problems discussed earlier in sect. and elaborate specific challenges that need to be addressed for designing a scalable and usertransparent system like hypar-flow. to develop a practical system like hypar-flow, it is essential that we thoroughly investigate apis and features of dl frameworks. in this context, the design analysis of execution models like eager execution vs. graph (or lazy) execution is fundamental. similarly, analysis of model definition apis like tensorflow estimators compared to keras is needed because these will influence the design choices for developing systems like hypar-flow. furthermore, the granularity of interfaces needs to be explored. for instance, using tensors to define a model is very complex compared to using a high-level model api like keras and onnx that follow the layer abstraction. finally, we need to investigate the performance behavior of these interfaces and frameworks. specific to hypar-flow, the main requirement from an api's perspective is to investigate a mechanism that allows us to perform user-transparent model partitioning. unlike other apis, keras seems to provide us this capability via the tf.keras.model interface. data-parallelism is easy to implement as no modification is required to the forward pass or the back-propagation of loss (error) in the backward pass. however, for model-parallelism, we need to investigate methods and framework-specific functionalities that enable us to implement the forward and backward pass in a distributed fashion. to realize these, explicit communication is needed between model partitions. for hybrid-parallelism, even deeper investigation is required because communication between model replicas and model partitions needs to be well-coordinated and possibly overlapped. in essence, we need to design a distributed system, which embeds communication primitives like send, recv, and allreduce for exchanging partial error terms, gradients, and/or activations during the forward and backward passes. an additional challenge is to deal with newer dnns like resnet(s) [ ] as they have evolved from a linear representation to a more complex graph with several types of skip connections (shortcuts) like identity connections, convolution connections, etc. for skip connections, maintaining dependencies for layers as well as for model-partitions is also required to ensure deadlock-free communication across processes. even though model-parallelism and hybrid-parallelism look very promising, it is unclear if they can offer performance comparable to data-parallelism. to achieve performance, we need to investigate if applying widely-used and important hpc techniques like ) efficient placement of processes on cpu cores, ) pipelining via batch splitting, and ) overlap of computation and communication can be exploited for improving performance of model-parallel and hybrid-parallel training. naive model-parallelism will certainly suffer from under-utilization of resources due to stalls caused by the sequential nature of computation in the forward and backward passes. we propose hypar-flow as an abstraction between the high-level ml/dl frameworks like tensorflow and low-level communication runtimes like mpi as shown in fig. (a) . the hypar-flow middleware is directly usable by ml/dl applications and no changes are needed to the code or the dl framework. the four major internal components of hypar-flow, shown in fig. (b) , are ) model generator, ) trainer, ) communication engine (ce), and ) load balancer. the subsections that follow provide details of design schemes and strategies for hypar-flow and challenges (c -c ) addressed by each scheme. the model generator component is responsible for creating an internal representation of a dnn (e.g. a keras model) suitable for distributed training (fig. ) . in the standard single-process (sequential) case, all trainable variables (or weights) of a model exist in the address space of a single process, so calling tape.gradients() on a tf.gradienttape object to get gradients will suffice. however, this is not possible for model-parallel training as trainable variables (weights) are distributed among model-partitions. to deal with this, we first create a local model object on all processes using the tf.keras.model api. next, we identify the layers in the model object that are local to the process. finally, we create dependency lists that allow us to maintain layer and rank dependencies for each of the local model's layers. these three components define our internal distributed representation of the model. this information is vital for realizing distributed backpropagation (discussed next) as well as for other hypar-flow components like the trainer and the communication engine. having a distributed model representation is crucial. however, it is only the first step. the biggest challenge for hypar-flow and its likes are: "how to train a model that is distributed across process boundaries?". we deal with this challenge inside the trainer component. first, we analyze how training is performed on a standard (non-distributed) keras to realize distributed back-propagation, we need ) partial derivative (d ) of loss l with respect to the weight w , and ) partial derivative (d ) of loss l with respect to the weight w . the challenge for multi-process case is that the term called "partial error" shown in eqs. and can only be calculated on partition- (fig. ) as y only exists there. to calculate d , partition- needs this "partial error" term in addition to d . because we rely on accessing gradients using the dl framework's implementation, this scenario poses a fundamental problem. tensorflow, the candidate framework for this work, does not provide a way to calculate gradients that are not part of a layer. to implement this functionality, we introduce the notion of grad layer in hypar-flow, which acts as a pseudo-layer inserted before the actual layer on each model-partition. we note that tensorflow's gradienttape cannot be directly used for this case. grad layers ensure that we can call tape.gradients() on this grad layer to calculate the partial errors during back-propagation. specifically, a grad layer is required for each recv operation so that partial error can be calculated for each preceding partition's input. a call to tape.gradients() will return a list that contains gradients as well as partial errors. the list is then used to update the model by calling optimizer.apply gradients(). we note that there is no need to implement distributed back-propagation for the data-parallel case as each model-replica is independently performing the forward and backward pass. the gradients are only synchronized (averaged) at the end of the backward pass (back-propagation) using allreduce to update the model weights in a single step. in sects. . and . , we discussed how the distributed model definition is generated and how back-propagation can be implemented for a model that is distributed across processes. however, trainer and model generator only provide an infrastructure for distributed training. the actual communication of various types of data is realized in hypar-flow's communication engine (ce). the ce is a light-weight abstraction for internal usage and it provides four simple apis: ) send, ) recv, ) broadcast and ) allreduce. for pure data-parallelism, we only need to use allreduce. however, for model-parallelism, we also need to use point-to-point communication between model-partitions. in the forward pass, the send/recv combination is used to propagate partial predictions from each partition to the next partition starting at layer . on the other hand, send/recv is used to backpropagate the loss and partial-errors from one partition to the other starting at layer n. finally, for hybrid-parallelism, we need to introduce allreduce to accumulate (average) the gradients across model replicas. we note that this is different from the usage of allreduce in pure data-parallelism because in this case, the model itself is distributed across different partitions so allreduce cannot be called directly on all processes. one option is to perform another p p communication between model replicas for gradient exchange. the other option is to exploit the concept of mpi communicators. we choose the latter one because of its simplicity as well as the fact the mpi vendors have spent considerable efforts to optimize the allreduce collective for a long time. to realize this, we consider the same model-partition for all model-replicas to form the allreduce communicator. because we only need to accumulate the gradients local to a partition across all replicas, allreduce called on this communicator will suffice. please refer back to fig. (sect. ) for a graphical illustration of this scheme. the basic ce design described above works but does not offer good performance. to push the envelope of performance further, we investigate two hpc optimizations: ) we explore if the overlap of computation and communication can be exploited for all three parallelization strategies and ) we investigate if pipelining can help overcome some of the limitations that arise due to the sequential nature of the forward/backward passes. finally, we also handle some advanced cases for models with non-consecutive layer connections (e.g. resnet(s)), which can lead to deadlocks. to achieve near-linear speedups for data-parallelism, the overlap of computation (forward/ backward) and communication (allreduce) has proven to be an excellent choice. horovod, a popular data-parallelism middleware, provides this support so we simply use it inside hypar-flow for pure data-parallelism. however, for hybridparallelism, we design a different scheme. we create one mpi communicator per model partition whereas the size of each communicator will be equal to the number of model-replicas. this design allows us to overlap the allreduce operation with the computation of other partitions on the same node. an example scenario clarifies this further: if we split the model across partitions, then we will use allreduce operations (one for each model-partition) to get optimal performance. this design allows us to overlap the allreduce operation with the computation of other partitions on the same node. because dnn training is inherently sequential, i.e., the computation of each layer is dependent on the completion of the previous layer. this is true for the forward pass, as well as for the backward pass. to overcome this performance limitation, we exploit a standard technique called pipelining. the observation is that dnn training is done on batches (or mini-batches) of data. this offers an opportunity for pipelining as a training step on samples within the batch is parallelizable. theoretically, the number of pipeline stages can be varied from all the way to batch size. this requires tuning or a heuristic and will vary according to the model and the underlying system. based on hundreds of experiments we performed for hypar-flow, we derive a simple heuristic: use the largest possible number for pipeline stages and decrease it by a factor of two. in most cases, we observed that num pipeline stages = batch size provides the best performance. figure shows a non-consecutive model with skip connections that requires communication ) between adjacent model-partitions for boundary layers and ) non-adjacent model-partitions for the skip connections. to handle communication dependencies among layers for each model-partition, we create two lists: ) forward list and ) backward list. each list is a list of lists to store dependencies between layers as shown in fig. . "f" corresponds to the index of the layer to which the current layer is sending its data and "b" corresponds to the index of the layer from which the current layer is receiving data. an arbitrary sequence of sending and receiving messages may lead to a deadlock. for instance, if partition- sends the partial predictions to partition- when partition- is waiting for predictions from partition- , a deadlock will occur as partition- is itself blocked (waiting for results from partition- ). to deal with this, we sort the message sequence according to the ranks so that the partition sends the first message to the partition which has the next layer. the models we used did not show any major load imbalance but we plan to design this component in the future to address emerging models from other application areas that require load balancing capabilities from hypar-flow. we have used three hpc systems to evaluate the performance and test the correctness of hypar-flow: ) frontera at texas advanced computing center (tacc), ) stampede (skylake partition) at tacc, and ) epyc: a local system with dual-socket amd epyc -core processors. inter-connect: frontera nodes are connected using mellanox infiniband hdr- hcas whereas stampede nodes are connected using intel omni-path hfis. [ ] . the design schemes proposed for hypar-flow are architecture-agnostic and can work on cpus and/or gpus. however, in this paper, we focus only on designs and scale-up/scale-out performance of manycore cpu clusters. we plan to perform in-depth gpu-based hypar-flow studies in the future. we now present correctness related experiments followed by a comprehensive performance evaluation section. because we propose and design hypar-flow as a new system, it is important to provide confidence to the users that hypar-flow not only offers excellent performance but also trains the model correctly. to this end, we present the correctness results based on two types of accuracy-related metrics: ) train accuracy (train acc)-percentage of correct predictions for the training data during the training process and ) test accuracy (test acc)-percentage of correct predictions for the testing data on the trained model. both metrics are covered for small scale training using vgg- on the cifar- dataset. we train vgg- for epochs using model-partitions on two stampede nodes with a batch size of and pipeline stages as shown in fig. (a) . next, we show test accuracy for resnet- -v in fig. (b) and resnet- -v in fig. (c) . the learning rate (lr) schedule was used from keras applications [ ] for both resnet(s) and was kept similar for sequential as well as parallel training variants. training for resnet- and resnet- was performed for and epochs, respectively. the following variants have been compared: discussion: clearly, model-parallel training with hypar-flow is meeting the accuracy of the sequential model for and epochs of training for resnet- and resnet- , respectively. we note that training is a stochastic process and there are variations in earlier epochs whether we use the sequential version or the model-parallel version. however, the significance is of the end result, which in this case peaks at . % for all the configurations presented. we ran multiple training jobs to ensure that the trends presented are reproducible. we use the term "process" to refer to a single mpi process in this section. the actual mapping of the process to the compute units (or cores) varies according to the parallelization strategy being used. images/second (or img/sec) is the metric we are using for performance evaluation of different types of training experiments. number of images processed by the dnn during training is affected by the depth (number of layers) of the model, batch size (bs), image size (w × h), and number of processes. higher img/sec indicates better performance. some important terms are clarified further: -horovod (dp): dnn training using horovod directly (data-parallel). we train various models on a single stampede node-dual-socket xeon skylake with cores and threads (hyper-threading enabled). the default version of tensorflow relies on underlying math libraries like openblas and intel mkl. on intel systems, we tried the intel-optimized version of tensorflow, but it failed with different errors such as "function not implemented" etc. for the amd system, we used the openblas available on the system. both of these platforms offer very slow sequential training. we present single-node results for vgg- , resnet- -v , and resnet- -v . vgg- has layers so it can be split in to as many as partitions. we try all possible cases and observe the best performance for num partitions = . as shown in fig. (a) , we see that hf (mp) offers better performance for small batch sizes and hf/horovod (dp) offers better performance for large batch sizes. hf (mp) offers better performance compared to sequential ( . × better at bs ) as well as to data-parallel training ( . × better at bs ) for vgg- on stampede . resnet- -v has layers so we were able to exploit up to modelpartitions within the node as shown in fig. (b) . we observe the following: ) hf (mp) is up to . × better than sequential at bs = , ) hf (mp) is up to . × better than horovod (dp) and hf (dp) at bs = , and ) hf (mp) is % slower than hf (dp) at bs = . the results highlight that modelparallelism is better at smaller batch sizes and data-parallelism are better only when large batch-size is used. figure (a) shows that hf (mp) can offer up to . × better performance than sequential training for resnet- -v on epyc ( cores). epyc offered better scalability with increasing batch sizes compared to stampede nodes ( fig. (b) vs. (a)) the performance gains suggest that hf (mp) can better utilize all cores on eypc compared to sequential training. to push the envelope of model depth and stress the proposed hypar-flow system, we also perform experiments for resnet- -v , which has , layers and approximately million parameters. figure (b) shows the performance for resnet- -v . it is interesting to note that dataparallel training performs poorly for this model. this is because the number of parameters increases the synchronization overhead for hf (dp) and horovod (dp) significantly. hence, even for large batch sizes, the computation is not enough to amortize the communication overhead. thus, hf (mp) offers much better performance compared to sequential ( . × better at bs = ) as well as to data-parallel training ( . × better at bs = ). two-node results for model parallelism are presented using vgg- and resnet- -v . figure (a) shows the performance trends for vgg- training across two nodes. as mentioned earlier, we are only able to achieve good performance with model-parallelism for up to model-partitions for the layers of vgg- . we also perform experiments for model-partitions but observe performance degradation. this is expected because of the lesser computation per partition and greater communication overhead in this scenario. we scale resnet- -v on two nodes using model-partitions in the model-parallelism-only configuration on stampede . the result is presented in fig. (b) . we observe that model-parallel hf (mp) training provides . × speedup (at bs = ) over hf (dp) and horovod (dp). on the other hand, a data-parallel-only configuration is not able to achieve good performance for resnet- due to significant communication (allreduce) overhead during gradient aggregation. emerging models like amoebanet [ ] are different compared to vgg and resnet(s). in order to show the benefit of hypar-flow as a generic system for various types of models, we show the performance of training a , -layer amoebanet variant in fig. . we provide results for four different conditions: ) sequential training using keras and tensorflow on one node, ) hf (mp) with partitions on one node, ) hf (mp) with partitions on two nodes, and ) hf (hp), where hp denotes hybrid parallelism on two nodes. as shown in fig. , we observe that hybrid parallelism offers the best possible performance using the same set of nodes. the most comprehensive coverage of hypar-flow's flexibility, performance, and scalability are presented in fig. (a). the figure shows performance for various combinations of hybrid-parallel training of resnet- -v on stampede nodes. the figure has three dimensions: ) the number of nodes on the x-axis, ) performance (img/sec) on y-axis, and ) batch size using the diameter of the circles. the key takeaway is that hybrid-parallelism offers the user to make trade-offs between high-throughput (img/sec) and batch size. from an accuracy (convergence) standpoint, the goal is to keep the batch-size small so model updates are more frequent. however, larger batch-size delays synchronization and thus provides higher throughput (img/sec). hypar-flow offers the flexibility to control these two goals via different configurations. for instance, the large blue circle with diagonal lines shows results for nodes using modelreplicas where the model is split into partitions on the single -core node. this leads to a batch-size of just , , which is × smaller than the expected , if pure data-parallelism is used. it is worth noting that the performance of pure data-parallelism even with × larger batch-size will still be lesser than the hybrid-parallel case, i.e., img/sec (= . × -considering ideal scaling for data-parallel case presented earlier in fig. (b) ) vs. img/sec (observed value- fig. (a) ). this is a significant benefit of hybrid-parallel training, which is impossible with pure model and/or data parallelism. in addition to this, we also present the largest scale we know of for any model/hybrid-parallel study on the latest frontera system. figure (b)) shows near-ideal scaling on frontera nodes. effectively, every single core out of the , cores on these nodes is being utilized by hypar-flow. the resnet- model is split into partitions as frontera nodes have a dual-socket cascade-lake xeon processor for a total of cores/node. we run one model-replica per node with a batch size of . to get the best performance, pipeline stages were tuned and the best number was found to be . today, designers develop models accounting for the restriction of memory consumption. however, with hypar-flow, this restriction no longer exists, and designers can come up with models with as many layers as needed to achieve the desired accuracy. to illustrate this, we present resnet- , an experimental model with layers. resnet- is massive and requires a lot of memory so we were able to train it with a batch-size of only. beyond that, it is not trainable on any existing system. we stress-test hypar-flow to scale the training of resnet- to two nodes and were able to train for bigger batch sizes. we note that training resnet- and investigation of its accuracy and finding the right set of hyper-parameters is beyond the scope of this paper. the objective is to showcase hypar-flow's ability to deal with models that do not exist today. model and data-parallelism can be combined in a myriad of ways to realize hybrid-parallel training. e.g. model-parallelism on a single node with multiple cores with data-parallelism across nodes. there are non-trivial and modeldependent trade-offs involved when designing hybrid schemes. model-parallelism and data-parallelism have different use cases; model-parallelism is beneficial when we have a large model, or we want to keep a small effective batch size for training. on the other hand, data-parallelism gives a near-linear scale-out on multiple nodes but it also increases batch size. in our experiments, we observe that single-node model-parallelism is better than single-node data-parallelism. theoretically, the number of model-partitions can not be larger than the number of layers in the model; we can not have more than partitions for resnet- . in practice, however, we observe that one layer per model-partition will not be used because it suffers from performance degradation. to conclude, hypar-flow's flexible hybrid-parallelism offers the best of both worlds; we can benefit from both model and data parallelism for the same model. we summarize the key observations below: -models like resnet- offer better performance for model-parallelism on smaller batch sizes (< ). -newer and very-deep models like resnet- benefit from model-parallelism for any batch size ( fig. (b) ). -hypar-flow's model-parallel training provides up to . × speedup over sequential training and . × speedup over data-parallel training ( fig. (a) ). -hypar-flow's hybrid-parallel training offers flexible configurations and provides excellent performance for resnet- ; × speedup over single-node training on stampede (xeon skylake) nodes ( fig. (a) ). -hypar-flow's hybrid-parallel training is highly scalable; we scale resnet- to frontera nodes ( , cores) as shown in fig. (b). deep learning workloads are going through a rapid change as newer models and larger, more diverse datasets are being developed. this has led to an explosion of software frameworks like tensorflow and approaches like data and modelparallelism to deal with ever-increasing workloads. in this paper, we explored a new approach to train state-of-the-art dnns and presented hypar-flow: a unified framework that enables user-transparent and parallel training of ten-sorflow models using multiple parallelization strategies. hypar-flow does not enforce any specific paradigm. it allows the programmers to experiment with different parallelization strategies without requiring any changes to the model definition and without the need for any system-specific parallel training code. instead, hypar-flow trainer and communication engine take care of assigning the partitions to different processes and performing inter-partition and interreplica communication efficiently. for resnet- training using hypar-flow, we were able to achieve excellent speedups: up to . × over data-parallel training, up to × over single-node training on stampede nodes, and up to × over single-node on frontera nodes. we also tested the ability of hypar-flow to train very large experimental models like resnet- , which consists of , layers. we believe that this study paves new ways to design models. we plan to publicly release the hypar-flow system so that the community can use it to develop and train next-generation models on large-scale hpc systems. extremely large minibatch sgd: training resnet- on imagenet in minutes oc-dnn: exploiting advanced unified memory capabilities in cuda and volta gpus for out-of-core dnn training s-caffe: co-designing mpi runtimes and caffe for scalable deep learning on modern gpu clusters legion: expressing locality and independence with logical regions demystifying parallel and distributed deep learning: an in-depth concurrency analysis improving strong-scaling of cnn training by exploiting finer-grained parallelism integrated model, batch, and domain parallelism in training neural networks accurate, large minibatch sgd: training imagenet in hour pipedream: fast and efficient pipeline parallel dnn training identity mappings in deep residual networks gpipe: efficient training of giant neural networks using pipeline parallelism beyond data and model parallelism for deep neural networks one weird trick for parallelizing convolutional neural networks imagenet classification with deep convolutional neural networks dragon: breaking gpu memory capacity limits with direct nvm access imagenet/resnet- training in seconds regularized evolution for image classifier architecture search mesh-tensorflow: deep learning for supercomputers optimizing network performance for distributed dnn training on gpu clusters: imagenet/alexnet training in . minutes key: cord- -bqlf fe authors: rydell-törmänen, kristina; johnson, jill r. title: the applicability of mouse models to the study of human disease date: - - journal: mouse cell culture doi: . / - - - - _ sha: doc_id: cord_uid: bqlf fe the laboratory mouse mus musculus has long been used as a model organism to test hypotheses and treatments related to understanding the mechanisms of disease in humans; however, for these experiments to be relevant, it is important to know the complex ways in which mice are similar to humans and, crucially, the ways in which they differ. in this chapter, an in-depth analysis of these similarities and differences is provided to allow researchers to use mouse models of human disease and primary cells derived from these animal models under the most appropriate and meaningful conditions. although there are considerable differences between mice and humans, particularly regarding genetics, physiology, and immunology, a more thorough understanding of these differences and their effects on the function of the whole organism will provide deeper insights into relevant disease mechanisms and potential drug targets for further clinical investigation. using specific examples of mouse models of human lung disease, i.e., asthma, chronic obstructive pulmonary disease, and pulmonary fibrosis, this chapter explores the most salient features of mouse models of human disease and provides a full assessment of the advantages and limitations of these models, focusing on the relevance of disease induction and their ability to replicate critical features of human disease pathophysiology and response to treatment. the chapter concludes with a discussion on the future of using mice in medical research with regard to ethical and technological considerations. although the genetic lineages of mice and humans diverged around million years ago, these two species have evolved to live together, particularly since the development of agriculture. for millennia, mice (mus musculus) were considered to be pests due to their propensity to ravenously consume stored foodstuff (mush in ancient sanskrit means "to steal" [ ] ) and their ability to adapt to a wide range of environmental conditions. since the s, domesticated mice have been bred and kept as companion animals, and in victorian england, "fancy" mice were prized for their variations in coat color and comportment; these mouse strains were the forerunners to the strains used in the laboratory today. robert hooke performed the first recorded inquiry-driven experiments on mice in , when he investigated the effects of changes in air pressure on respiratory function [ ] . more recently, with data from the human genome project and sequencing of the mus musculus genome showing remarkable genetic homology between these species, as well as the advent of biotechnology and the development of myriad knockout and transgenic mouse strains, it is clear why the mouse has become the most ubiquitous model organism used to study human disease. in addition, their small size, rapid breeding, and ease of handling are all important advantages to scientists for practical and financial reasons. however, keeping in mind that mice are fellow vertebrates and mammals, there are ethical issues inherent to using these animals in medical research. this chapter will provide an overview of the important similarities and differences between mus musculus and homo sapiens and their relevance to the use of the mouse as a model organism and provide specific examples of the quality of mouse models used to investigate the mechanisms, pathology, and treatment of human lung diseases. we will then conclude with an assessment of the future of mice in medical research considering ethical and technological advances. as a model organism used to test hypotheses and treatments related to human disease, it is important to understand the complex ways in which mice are similar to humans, and crucially, the ways in which they differ. a clear understanding of these aspects will allow researchers to use mouse models of human disease and primary cells derived from mice under the most appropriate and meaningful conditions. in , the encyclopedia of dna elements (encode) program published a comparative analysis of the genomes of homo sapiens and mus musculus [ ] , as well as an in-depth analysis of the differences in the regulatory landscape of the genomes of these species [ ] . encode, a follow-up to the human genome project, was implemented by the national human genome research institute (nhgri) at the national institutes of health in order to develop a comprehensive catalog of protein-encoding and nonproteincoding genes and the regulatory elements that control gene expression in a number of species. this was achieved using a number of genomic approaches (e.g., rna-seq, dnase-seq, and chip-seq) to assess gene expression in over mouse cell types and tissues; the data were then compared with the human genome. overall, these studies showed that although gene expression is fairly similar between mice and humans, considerable differences were observed in the regulatory networks controlling the activity of the immune system, metabolic functions, and responses to stress, all of which have important implications when using mice to model human disease. in essence, mice and humans demonstrate genetic similarity with regulatory divergence. specifically, there is a high degree of similarity in transcription factor networks but a great deal of divergence in the cis-regulatory elements that control gene transcription in the mouse and human genomes. moreover, the chromatin landscape in cell types of similar lineages in mouse and human is both developmentally stable and evolutionarily conserved [ ] . of particular relevance regarding modeling human diseases involving the immune system, in its assessment of transcription factor networks, the mouse encode consortium revealed potentially important differences in the activity of ets in the mouse and human genome. although conserved between the two species, divergence in ets regulation may be responsible for discrepancies in the function of the immune system in mouse and human [ ] . certainly, the biological consequences of these differences in gene expression and regulation between human and mouse invite further investigation. the anatomical and physiological differences between model organisms and humans can have profound impacts on interpreting experimental results. virtually every biological process under investigation in experimental studies involves at least one anatomical structure. to aid in interpretation, many anatomy compendia have been developed for model organisms; the most useful organize anatomical entities into hierarchies representing the structure of the human body, e.g., the foundational model of anatomy developed by the structural informatics group at the university of washington [ ] . although an analysis of the myriad differences between mouse and human anatomy is beyond the scope of this chapter, a few of the most critical issues that have an impact on the interpretation of data from mouse experiments should be mentioned. the most obvious difference between mice and humans is size; the human body is about times larger than that of the mouse. size influences many aspects of biology, particularly the metabolic rate, which is correlated to body size in placental mammals through the relationship bmr ¼  mass ( . ), where bmr is the basal metabolic rate (in kcal/day). thus, the mouse bmr is roughly seven times faster than that of an average-sized human [ ] . this higher bmr has effects on thermoregulation, nutrient demand, and nutrient supply. as such, mice have greater amounts of metabolically active tissues (e.g., liver and kidney) and more extensive deposits of brown fat [ ] . furthermore, mice more readily produce reactive oxygen species than do humans, which is an important consideration when modeling human diseases involving the induction of oxidative stress (i.e., aging, inflammation, and neurodegeneration) [ ] . the lung provides an excellent example of the similarities and differences between human and mouse anatomy. similar to the human organ, the mouse lung is subdivided into lobes of lung parenchyma containing a branching bronchial tree and is vascularized by the pulmonary circulation originating from the right ventricle. there are a number of subtle variations in this general structure between species, i.e., the number of lobes on the right and left, the branching pattern, and the distribution of cartilage rings around the large airways, but the most important differences between the mouse and human lung are related to the organism's size (airway diameter and alveolar size are naturally much smaller in the mouse) and respiratory rate. moreover, there are important differences in the blood supply of the large airways in humans versus mice [ ] . specifically, the bronchial circulation (a branch of the high-pressure systemic circulation that arises from the aorta and intercostal arteries) supplies a miniscule proportion of the pulmonary tissue in mice (the trachea and bronchi) compared to humans; the majority of the lung parenchyma is supplied by the low-pressure, high-flow pulmonary circulation. in the mouse, these systemic blood vessels do not penetrate into the intraparenchymal airways, as they do in larger species [ ] . this difference, although subtle, has important ramifications regarding the vascular supply of lung tumors which, in humans, is primarily derived from the systemic circulation [ ] . these differences may also have profound consequences when modeling human diseases involving the lung vasculature. the adaptive immune system evolved in jawed fish about million years ago, well before the evolution of mammals and the divergence of mouse and human ancestral species [ ] . many features of the adaptive immune system, including antigen recognition, clonal selection, antibody production, and immunological tolerance, have been maintained since they first arose in early vertebrates. however, the finer details of the mouse and human immune systems differ considerably, which is not surprising since these species diverged million years ago [ ] . while some have claimed that these differences mean that research into immunological phenomena in mice is not transferable to humans, as long as these differences are understood and acknowledged, the study of mouse immune responses can continue to be relevant. research on mice has been vital to the discovery of key features of both innate and adaptive immune responses; for example, the first descriptions of the major histocompatibility complex, the t cell receptor, and antibody synthesis were derived from experiments performed on mice [ ] . the general structure of the immune system is similar in mice and humans, with similar mediators and cell types involved in rapid, innate immune responses (complement, macrophages, neutrophils, and natural killer cells) as well as adaptive immune responses informed by antigen-presenting dendritic cells and executed by b and t cells. however, due to the anatomical and physiological differences between these species as described above, divergence in key features of the immune system, such as the maintenance of memory t cells (related to the life span of the organism) and the commensal microbiota (related to the lifestyle of the organism), has arisen [ ] . similar to what has been discovered regarding the genetics of mice and humans, i.e., broad similarities in structure but considerable differences in regulation, there are a number of known discrepancies in the regulation of innate and adaptive immunity in mouse models of human disease mice versus humans, including the balance of leukocyte subsets, t cell activation and costimulation, antibody subtypes and cellular responses to antibody, th /th differentiation, and responses to pathogens (described in detail in table ). in addition to these differences in immune cell functions, the expression of specific genes involved in immune responses also differs, particularly those for toll-like receptors, defensins, nk inhibitory receptors, thy- , and many components of chemokine and cytokine signaling; additionally, differences between mouse strains are known to exist for many of these mediators [ ] . another important consideration when using mice to perform immunological research (with a view to translating these findings to human medicine) is the availability of hundreds of strains of genetically modified mice that have enabled exquisitely detailed studies on immune cell function, regulation, and trafficking. many of these strains involve the expression of inducible cre or cas that allow for targeted knockdown or overexpression of key immune function-related genes in specific cell types at specific moments in time. however, it is important to note that drift between mouse colonies has long been known to occur. in fact, a recent report described the fortuitous discovery of a point mutation in the natural cytotoxicity receptor (ncr ) gene in the c /bl cd . mouse strain, resulting in absent ncr expression. this mutation was found to have profound effects on the response of mice to viral infection, i.e., the mice were resistant to cytomegalovirus infection but more susceptible to influenza virus [ ] . this cautionary tale highlights the importance of understanding the genetic evolution of laboratory strains of mice, the effect of these genetic and immunological changes on mouse biology, and the impact on the translation of these results to human medicine. in addition to the differences between mouse and human genetics, physiology, and immunology highlighted above, several factors must also be taken into account when performing in vitro assays using isolated mouse cells and applying these findings to our understanding of human disease. particularly with regard to stem cell research, it should be noted that the telomeres of mouse cells are five-to tenfold longer than human telomeres, resulting in greater replicative capacity [ ] . there are also important differences in the regulation of pluripotency and stem cell differentiation pathways in humans and mice [ ] . moreover, there are considerable species differences in the longevity of cultured cells; for example, mouse fibroblasts are capable of spontaneous immortalization in vitro, whereas human fibroblasts become senescent and ultimately fail to thrive in culture [ ] . in summary, although there are considerable differences between mice and humans, constant improvement in the analytical techniques used to delineate these differences and their effects on whole organism and cell function have provided vital information and contributed to our understanding of both murine and human biology. experimentation employing mouse models of human disease will continue to provide key insights into relevant disease mechanisms and potential drug targets for further clinical investigation. however, several important considerations must be taken into account when selecting a mouse model of human disease, as described in the following section, using mouse models of human lung disease to illustrate this point. the two most salient features of a mouse model of human disease are the accuracy of its etiology (it employs a physiologically relevant method of disease induction) and its presentation (its ability to recapitulate the features of human disease). the relevance of any given mouse model can be judged on the basis of these two criteria, and there is considerable variation within mouse models of human disease in this regard. as a full assessment of the advantages and limitations of all currently available mouse models of human disease would be prohibitively long and complex, here we have elected to assess the accuracy of currently available models of human lung diseases, i.e., asthma, chronic obstructive pulmonary disease, and pulmonary fibrosis, focusing on the relevance of disease induction in these models and their ability to replicate critical features of human disease pathophysiology and response to treatment. the first and foremost notion when modeling human disease in mice is to acknowledge the species differences, which are significant [ ] . as described above, genetics, anatomy, physiology, and immunology differ between mice and humans, but despite these differences, mouse models of human disease are useful and necessary, as long as data interpretation is performed appropriately. an elegant example of differences between mice and humans that must be considered when designing a mouse model of human inflammatory lung disease is the key effector cell type in human asthma, i.e., mast cells. these leukocytes differ in granule composition as well as localization in the mouse and human airways [ ] . mice mostly lack mast cells in the peripheral lung [ ] , whereas humans have numerous mast cells of multiple subpopulations in the alveolar parenchyma [ ] . another example is anatomy: in contrast to humans, mice lack an extensive pulmonary circulation, which may have significant effects on leukocyte adhesion and migration, and subsequently inflammation [ ] . still, as long as these differences are taken into consideration, mouse models can be powerful tools in the discovery and exploitation of new targets for the treatment of human disease. the world health organization (who) defines asthma as a chronic disease characterized by recurrent attacks of breathlessness and wheezing, which may vary in severity and frequency from person to person. the disease is characterized by airway hyperresponsiveness, airway smooth muscle thickening, increased mucus secretion and collagen deposition, as well as prominent inflammation affecting both large and small airways [ ] . nowadays, it is recognized that asthma is not a single homogenous disease but rather several different phenotypes united by similar clinical symptoms [ , ] . only a few animal species develop asthma naturally, including cats and horses [ , ] , whereas mice do not [ ] . however, mice can be manipulated to develop a type of allergic airway inflammation, which is similar in many ways to the human disease, in response to different aeroallergens [ ] . importantly, these models are capable of recapitulating only the allergic type of human asthma and have less relevance for other types of asthma (i.e., endotypes induced by medication, obesity, and air pollution). as with many human diseases, asthma has a complex and multifaceted etiology, where environmental factors, genetic susceptibility, and microbial colonization all contribute; thus, it is important to take strain differences into consideration. generations of inbreeding have created mouse strains that differ not only in coat color and disposition but also from a physiological, immunological, and genetic perspective. different strains may be more susceptible to allergic airway inflammation or pulmonary fibrosis, whereas others are more or less resistant. choosing the right strain to model a specific disease or pathologic event is thus essential. the most widely used strains for models of allergic airway inflammation are balb/c and c bl/ . these strains differ regarding the type of immune response mounted to an inhaled allergen: c bl/ is generally considered a t h -skewed strain, whereas balb/c is regarded as a t h -skewed strain [ ] . due to their strong t h response, and subsequent development of robust asthmatic responses, balb/c has been commonly used to model asthma [ ] . however, most humans do not express such a strongly t h -skewed immune system, suggesting this strain may not be the best model of human disease; instead, c bl/ may be more suitable as immune responses in this strain are more similar to those of atopic human subjects [ ] . furthermore, as c bl/ is the most commonly used strain for the development of genetically manipulated mice, using these mice allows for very specific investigations into disease pathology; thus, this strain is increasingly used in models of human lung disease. besides the genetic differences in the mouse strains used in these models, the etiology (the method of disease induction) of commonly used models of asthma is highly variable. in humans with allergic asthma, environmental allergen exposure occurs at the airway mucosa; the immune response is coordinated in the bronchopulmonary lymph nodes, and the t cells, macrophages, and eosinophils recruited as part of this response travel to the lung where they mediate the cardinal features of asthma: airway inflammation, structural remodeling of the airway wall, and airway hyperreactivity [ ] . ideally, these features should be found in a physiologically relevant mouse model of asthma. however, for the sake of cost and convenience, early mouse models of asthma used the surrogate protein ovalbumin (ova) [ ] rather than an environmental allergen to induce an immune response, which also requires the use of a powerful t h -polarizing adjuvant such as alum delivered via the intraperitoneal route, followed by ova nebulization-a clear divergence from the etiology of human asthma [ ] . in terms of disease presentation, mice develop some hallmarks of asthma, including airway eosinophilic inflammation, goblet cell metaplasia, and increased airway smooth muscle density [ ] . after the cessation of ova exposure, most of the remodeling resolves, although some structural alterations remain up to month after the last challenge [ ] . based on these attributes, the ova model is primarily a model to investigate the initiation of inflammation, rather than the chronic progression and maintenance of inflammation [ ] . a clear advantage with the ova model is the number of studies where it is used; both the pros and cons are familiar. it is easy to find a suitable protocol, and the model is readily accessible and flexible regarding the number of sensitizations and allergen doses. the model is relatively easy to reproduce, as ova and different adjuvants are easily obtained. however, the resolution of remodeling following the cessation of allergen provocations is a disadvantage, as is the practical problem with the nebulization of an allergen-it ends up in the mouse's coat and is ingested during grooming, potentially resulting in systemic exposure (this is particularly relevant in models employing systemic, intraperitoneal sensitization). in addition, concerns have been raised against the use of adjuvants to induce the immunological response, as well as the clinical relevance of ova as an allergen, which have driven the development of more clinically relevant allergens and models [ ] . the common environmental aeroallergen house dust mite (hdm) extract is increasingly used to initiate disease in mouse models of allergic airway inflammation, as it is a common human allergen (around % of asthmatics are sensitized to hdm [ ] ) that evokes asthma attacks and other allergic responses in susceptible individuals. in addition, hdm has inherent allergenic properties, likely due to components with protease activity [ ] , so there is no need to use an adjuvant, thus improving the etiological similarity of these models with the clinical situation [ ] . in contrast to ova, prolonged exposure of hdm (up to weeks) induces asthma-like severe airway inflammation with prominent eosinophilia, severe hyperreactivity to methacholine, and robust remodeling of the airway wall [ ] , i.e., the presentation of chronic respiratory hdm exposure in mice effectively recapitulates the key features of human allergic asthma. importantly, the airway structural changes induced by chronic hdm exposure, such as increased collagen deposition, airway smooth muscle thickening, and microvascular alterations, persist for at least weeks after the cessation of hdm exposure [ ] , another commonality with human asthma in which airway remodeling is currently considered to be irreversible. thus, the advantages of using hdm as the allergen in mouse models of asthma are the clinical relevance of the allergen [ ] and the route of delivery via the respiratory tract. moreover, studies have shown that the type of inflammation and characteristics of tissue remodeling are relatively similar to those seen in human asthmatics [ , , ] . one disadvantage is the complexity of hdm extract; as a consequence of this complexity, variations exist in some components between batches, particularly regarding the content of lipopolysaccharide, so reproducibility in these studies may be problematic. with similarity to hdm, these models were developed to be as clinically relevant as possible, as many patients suffer from allergy toward cockroach allergen, molds, and other environmental irritants. a common feature of these allergens is their complex nature, as they commonly consist of a mix of different allergic epitopes and fragments. this complexity is most likely why the immunological reaction in mice is relatively similar to that seen in asthmatics [ ] . cockroach allergen (cra) is a common allergen, known to induce asthma in susceptible individuals; thus, it shares with hdm the advantage of being highly clinically relevant [ ] . cra induces peribronchial inflammation with significant eosinophilic inflammation and transient airway hyperresponsiveness, both of which can be increased by repeated administrations of the allergen [ ] . colonization of the airways with aspergillus fumigatus is the cause of allergic bronchopulmonary aspergillosis (abpa), a disease where the lungs are colonized by the fungus, but allergens from aspergillus fumigatus can also induce asthma similar to other allergens [ ] . the reaction to aspergillus allergens is robust, and often no adjuvants are needed to elicit inflammation [ ] . in addition to aspergillus, other fungi such as penicillium and alternaria can also induce asthma in humans and have been used to model disease in mice [ ] . a common difficulty with these allergens is the method of administration, as the physiological route is believed to be the inhalation of dry allergens; mimicking this route with a nebulizer introduces the risk of the animals ingesting the allergen and thus causing systemic responses [ ] . exacerbations of asthma are defined as the worsening of symptoms, prompting an adjustment in treatment, and are believed to be associated with increased inflammation in the distal airways. clinically, exacerbations are believed to be induced by infections (most common), allergen exposure, or pollutants, which can be modeled in different ways [ , ] : . infections with viruses and bacteria or exposure to proteins/ dna/rna derived from these microbes. . administration of a high dose of allergen in a previously sensitized animal. . exposure to environmental pollutants, such as diesel exhaust or ozone. modeling exacerbations adds a layer of complexity, as robust ongoing allergic airway inflammation needs to be established first, before challenge with the exacerbating agent. both the ova and hdm models are used in this respect, and in both cases chronic protocols extending for several weeks before triggering an exacerbation have been used [ ] . chronic obstructive pulmonary disease (copd) is characterized by chronic airway obstruction, in contrast to asthma where the obstruction is reversible (particularly in response to bronchodilator treatment). clinically, in copd, chronic bronchitis and emphysema can occur either separately or in combination. copd is almost always associated with either first-or secondhand tobacco smoking or in rare cases with a deficiency in the production of α antitrypsin (a serpin that prevents elastin breakdown as a result of neutrophil degranulation) [ ] . the etiology of copd is highly complex and is believed to develop after many years of smoking in combination with other known factors such as genetic susceptibility or environmental factors [ ] . in similarity to asthma, inflammation is a major component in copd, but the leukocyte profile is very different: the most prominent players in copd-related inflammation are neutrophils and, to some degree, macrophages [ ] . due to the complex etiology of copd, it is difficult to recapitulate all aspects of this disease in a single model, so in most cases, the aim is to induce copd-like lesions by exposing mice to tissue-damaging substances (usually cigarette smoke) or to mimic emphysema by the administration of tissue-degrading enzymes [ , ] . clearly, mice do not smoke cigarettes on their own, so to model copd by cigarette smoke (cs) inhalation, the mice need to be exposed to unfiltered cs in an induction chamber; moreover, in an attempt to better model the chronic aspects of copd, this needs to be performed for a prolonged period of time. mice are very tolerant to cs, but eventually (over a period of several weeks), cs induces pulmonary neutrophilic inflammation that is associated with some degree of tissue degradation and destruction [ ] . an important advantage of this model is the fact that cs is the actual irritant responsible for disease in humans, and mice develop several features similar to the clinical disease, making this model highly clinically relevant [ ] . a significant drawback is the self-limitation of the model-the pathological changes do not progress after the cessation of cs exposure [ ] . furthermore, the exposure time needed for mice to develop copd-like pathology is extensive, i.e., studies have shown that an exposure protocol of days per week for a minimum of months is needed to generate robust structural changes to the lung [ ]. the pathological image in copd is complex and varies greatly between patients, commonly encompassing chronic bronchitis and bronchiolitis, emphysema, fibrosis, and airway obstruction. although mice develop some of these symptoms when exposed to cs, they do not develop all the symptoms of human disease; thus, cs has advantages as a model but fails to mimic the complexity of the clinical situation and disease presentation [ ] . other models of copd rely on the administration of proteases (protein-degrading enzymes) that are believed to be involved in the pathology of this disease in a subset of patients, such as elastindegrading elastase. this approach mimics the emphysematous changes seen in copd, but the pathological process underlying tissue destruction is likely very different compared to the clinical situation [ ] , as very few patients show evidence of elastase dysregulation [ ] . however, if the aim of the study is to investigate the general effect of protease-induced tissue destruction and regeneration, then this is a highly relevant method [ ] . some studies on copd have also used genetically modified animals, such as mice overexpressing collagenase, which results in tissue destruction without inflammation or fibrosis with an end result fairly similar to the type of emphysema observed in copd [ ] . pulmonary fibrosis, the accumulation of fibrotic tissue within the alveolar parenchyma, is merely a symptom of disease, and the etiology of this pathology in humans varies greatly [ ] . the most enigmatic class is perhaps the idiopathic interstitial pneumonias, especially idiopathic pulmonary fibrosis (ipf). ipf is a debilitating and progressive disease with a grave prognosis, characterized by progressive fibrosis believed to reflect aberrant tissue regeneration [ ] . as the reason behind this defective repair is unknown, although a combination of immunological, genetic, and environmental factors are suspected, it is very difficult to model disease in a clinically relevant fashion [ ] . the most common method used to model pulmonary fibrosis in mice is administration of the chemotherapeutic agent bleomycin; this agent is known to cause pulmonary fibrosis in humans as well, but this may not accurately reflect the true etiology of most cases of human disease. the strain of choice is c bl/ , as it is prone to developing pulmonary fibrosis, whereas balb/c is relatively resistant, a feature believed to reflect the cytokine response following cellular stress and damage [ ] . bleomycin administration can be performed locally or systemically, producing very different results. the most common model of pulmonary fibrosis is a single intranasal or intratracheal administration of bleomycin, with analysis to weeks later. during this time, the drug causes acute tissue damage in a restricted area of the lung (where the solution ends up during administration), followed by intense inflammation in this area and subsequent fibrosis, which gradually resolves within weeks. however, if older mice are used, the fibrosis will persist longer than in younger mice, which is in accordance with clinical ipf, where the majority of the patients are years of age or older [ , ] . a great advantage of this model is how well-characterized it is. in addition, local administration is labor-effective, as only one administration is required and the result is highly reproducible. the fibrosis is robust, only affects the lungs, and the accumulation of extracellular matrix can be easily measured using standard techniques [ ] . furthermore, as it is used throughout the world, studies performed in different labs and by different groups can be compared relatively easily. unfortunately, the intense pulmonary inflammation may be lethal, and fatalities are to be expected with this model [ ] , representing an important ethical limitation. furthermore, fibrosis is heterogeneous-it develops where the bleomycin solution is deposited. the solution usually deposits within the central lung, a localization that is not in agreement with the clinical situation where fibrosis is located in the more distal regions of the lung parenchyma. in addition, the fibrosis that develops as a result of severe tissue damage is self-limiting and reversible, unlike what is observed clinically [ ] . the severe degree of tissue damage induced by bleomycin may in fact be more relevant for modeling acute lung injury (ali) or acute respiratory distress syndrome (ards). bleomycin can also be administered systemically, through intravenous or subcutaneous injection. in contrast to local administration, this route requires multiple administrations and is thus more laborintensive [ ] . some studies have described the usage of osmotic mini-pumps, where bleomycin is slowly administered over a short period of time, and then fibrosis continues to develop over subsequent weeks [ ] . irrespective of the route of delivery, systemic administration results in more homogenous fibrosis, affecting the entire lung through the pulmonary endothelium and persisting much longer than following local administration [ ] . the main advantages of systemic administration are that inflammation is limited, while the fibrosis is more apparent and displays a more distal pattern, all of which mimics the clinical situation relatively well. the multiple administrations allow for lower doses with each injection; this is less stressful to the animals and results in little to no mortality [ ] and is thus more ethically acceptable. a major disadvantage with this model is that it takes time for fibrosis to develop [ ] , which may be the reason it is used relatively scarcely, and thus the pathological development is less well-understood. in addition, as ipf is a local disease, local administration of the etiologic agent may better mimic the clinical reality [ ] . the administration of fluorescein isothiocyanate (fitc) induces focal inflammation, primarily involving mononuclear cells and neutrophils, and localizes in areas where the fitc solution is deposited [ ] . antibodies against fitc can be detected after week, and the fibrosis persists for up to months after instillation [ ] . the benefits of this model are mainly related to the persistent fibrosis that does not appear to be self-limiting, thus reflecting the clinical situation, and it is also very easy to determine which part of the lung has been exposed to fitc, as the molecule is fluorescent [ ] . it is also an advantage that both c bl/ and balb/c mice are susceptible and develop fibrosis following fitc administration [ ] . the disadvantages of this model include profound variability due to differences between batches of fitc, as well as in the method used to prepare the solution before instillation. importantly, given the characteristics of the etiologic agent used to induce this model of ipf, this model is considered a very artificial system with limited clinical relevance [ ] . adenovirus vectors have been used to overexpress the pro-fibrotic cytokine transforming growth factor (tgf)-β, which results in pulmonary fibrosis. as tgf-β overexpression in the lungs is known to be crucial in the development of fibrosis in humans [ ] , this model mimics an important feature of disease etiology. however, the delivery system has some drawbacks, as the virus itself initiates an immune response. moreover, adenoviruses display significant tropism for epithelial cells and rarely infect other cell types such as fibroblasts [ ] , which are the cells meant to be targeted in this model. as tgf-β has major effects on fibroblast biology, the main feature of this model is the effect of epitheliumderived tgf-β on fibroblasts and myofibroblasts, resulting in the deposition of ecm proteins and areas of dense fibrosis [ ] . an advantage of this model is the relatively low degree of inflammation, as well as what appears to be a direct effect on fibroblasts/ myofibroblasts [ ] , which is in accordance with the clinical situation (as we understand it today). silica administration induces a similar pathology in mouse lungs as in humans exposed to silica, and as is also observed in human silicainduced fibrosis, structural remodeling persists when administration is halted [ ] . following the administration of silica particles, fibrotic nodules develop in mouse lungs, with considerable resemblance to the human lesions that develop after exposure to mineral fibers [ ] . the fibrotic response is accompanied by a limited inflammatory response, and different pro-fibrotic cytokines such as tgf-β, platelet-derived growth factor, and il- are involved in disease development, which is in accordance with the clinical situation [ ] . another advantage is that nodules develop around silica fibers, and these fibers are easy to identify by light microscopy. the response in this model is strain-dependent, with c bl/ mice being the most susceptible. the main drawbacks are the time required to establish disease, i.e., - days, and the need for special equipment to aerosolize the silica particles. however, since the route of administration, the driving etiologic agent, and the resulting pathobiology are all similar to the characteristics of this subtype of pulmonary fibrosis [ , ] , the silica exposure model can be considered to have very good clinical relevance. what does the future hold for mouse models of human disease? medical research using experimental animals (not only mice but other animals including rats, guinea pigs, zebrafish, and fruit flies) has greatly contributed to many important scientific and medical advances in the past century and will continue to do so into the near future. these advances have contributed to the development of new medicines and treatments for human disease and have therefore played a vital role in increasing the human life span and improving quality of life. despite the acknowledged benefits of performing research using experimental animals, a number of considerations must be made before embarking on this type of research. of course, the financial aspects of conducting this type of work are an important limitation, as the costs of purchasing and housing mice can be prohibitive, especially when genetically modified mice and colony maintenance are required for the study. the practicalities of working with animals such as mice may also be an issue, as this type of work requires specialized facilities, equipment, and staff to ensure studies are carried out in a manner that is safe for both the researchers and the animals. moreover, as discussed in detail in this chapter, the relevance of the selected animal model to human disease must be carefully evaluated to ensure that these experiments provide robust results that are translatable to human health and disease. another important and demanding aspect of biomedical research using animals is the ethics of imposing pain and suffering on live animals. although there has been a considerable reduction in the numbers of animals used in research in the last years, animal research remains a vital part of biomedical research. however, no responsible scientist wants to cause unnecessary suffering in experimental animals if it can be avoided, so scientists have accepted controls on the use of animals for medical research. in the uk, this ethical framework has been enshrined in law, i.e., the animals (scientific procedures) act . this legislation requires that applications for a project license to perform research involving the use of "protected" animals (including all vertebrates and cephalopods) must be fully assessed with regard to any harm imposed on the animals. this involves a detailed examination of the proposed procedures and experiments, and the numbers and types of animal used, with robust statistical calculations to support these numbers. the planned studies are then considered in light of the potential benefits of the project. both within and outside the uk, approval for a study involving protected animals also requires an internal ethical review process, usually conducted by the research institution where the work is taking place, with the aim of promoting animal welfare by ensuring the work will be carried out in an ethical manner and that the use of animals is justified. additionally, the uk has a national animal use reduction strategy supported by the national centre for the replacement, refinement and reduction of animals in research (nc rs; london, uk). this consortium was established in to promote and develop high-quality research that takes the principles of replacement, refinement, and reduction (the rs) into account. replacement strategies often involve the use of alternative, non-protected species (e.g., zebrafish, fruit flies, flatworms) and in vitro correlates (two-dimensional cell culture or threedimensional organoids containing multiple cell types) to test hypotheses and assess the effects of therapeutic interventions. the main obstacle with studies on non-protected animals is the difficulty of accurately mimicking the complex physiological systems involved in human health and disease, as described in detail above. for example, the fruit fly drosophila melanogaster is an excellent model organism for studies on genetic diseases, aging, and pathogen-borne illnesses but may be less relevant for studies on complex lung diseases. importantly, model organisms such as fruit flies, zebrafish, and flatworms do not possess lungs, which somewhat limits the translatability of research on these animals in the field of respiratory disease. as such, it is likely that rodents will remain the model organism of choice for studies into lung disease for some time to come. there has been considerable progress recently in imitating single organs such as the liver, lung, and brain in vitro using multiple cell types and a physical scaffold. as an important advantage, these in vitro tests have replaced a large number of rodents in initial drug discovery experiments, while also speeding up the process [ ] . these studies still require further refinement and validation to establish them as suitable models for an entire organ; importantly, these in vitro organoids cannot take into account interactions between organ systems in complex, multisystem diseases such as copd. refinement involves selecting the most clinically relevant model for the disease available, informed by the discussion above on closely recapitulating the etiologic agent and disease pathobiology associated with clinical cases. another important factor is refining the management of pain. an assessment of the procedures used and the effects of the substance on the animal, as well as the degree of handling, restraint, and analgesia, are other important aspects of refinement. this standard of animal care is achieved through strict regulations and controls on how personnel are trained to carry out experiments on live animals. adequate training is an important aspect of refinement and should be reviewed and improved on an ongoing basis. moreover, refinement can be achieved by improving animal housing by environmental enrichment, e.g., providing a place for mice to hide in the cage and housing social animals such as mice in appropriate-sized groups. these simple changes can improve the physiological and behavioral status of research animals; this not only increases animal well-being but also contributes to the quality of the experimental results by reducing stress levels. the rs aspect of reduction focuses on the statistical power of experiments and by following the animal research: reporting of in vivo experiments (arrive) guidelines, originally published in plos biology in . these guidelines provide a framework to improve the reporting of research performed on live animals by maximizing the quality of the scientific data and by minimizing unnecessary studies. the arrive guidelines provide a checklist of aspects that must be considered in good quality research using live animals. the guidelines are most appropriate for comparative studies involving two or more groups of experimental animals with at least one control group, but they also apply to studies involving drug dosing in which a single animal is used as its own control (within-subject experiments). the guidelines provide recommendations on what should be considered when preparing to report on the results of experiments involving live animals, i.e., by providing a concise but thorough background on the scientific theory and why and how animals were used to test a hypothesis, a statement on ethical approvals and study design including power and sample size calculations, a clear description of the methods used to ensure repeatability, objective measurements of outcomes and adverse effects, and interpretation of the results in light of the available literature and the limitations of the study. in addition to the positive impact of the arrive guidelines on reducing the number of animals used in experiments, this checklist provides an easy-tofollow roadmap on what is required for good quality reporting of experimental results. in conclusion, the use of animals in research will continue to be an important aspect of medical research, and these procedures can be ethically justified provided the proper controls are in place. the benefits of animal research have been vital to the progress of medical science; abandoning these studies would have severe negative consequences on human health. by considering aspects such as the rs and the arrive guidelines in planning experiments involving live animals, the number of animals used and suffering of these animals for the benefit of human health can be minimized. this requires a strong regulatory framework such as that found in the uk and many other countries, as well an ongoing public debate on the advantages and limitations of animal experimentation. use of house mice in biomedical research the laboratory mouse a comparative encyclopedia of dna elements in the mouse genome principles of regulatory information conservation between mouse and human of mice and men: aligning mouse and human anatomies mouse models of human disease: an evolutionary perspective structure and composition of pulmonary arteries, capillaries, and veins angiogenesis in the mouse lung lung cancer perfusion: can we measure pulmonary and bronchial circulation simultaneously? origin and evolution of the adaptive immune system: genetic events and selective pressures the evolutionary basis for differences between the immune systems of man, mouse, pig and ruminants of mice and not men: differences between mouse and human immunology gender dimorphism in differential peripheral blood leukocyte counts in mice using cardiac, tail, foot, and saphenous vein puncture methods species differences in the expression of major histocompatibility complex class ii antigens on coronary artery endothelium: implications for cellmediated xenoreactivity icos is critical for cd -mediated antibody class switching homozygous loss of icos is associated with adultonset common variable immunodeficiency b -h : a costimulatory molecule for t cell activation and ifn-gamma production global gene regulation during activation of immunoglobulin class switching in human b cells selective loss of type i interferon-induced stat activation caused by a minisatellite insertion in mouse stat regulation of human helper t cell subset differentiation by cytokines functional dichotomy in the cd + t cell response to schistosoma mansoni what animal models teach humans about tuberculosis cutting edge: check your mice-a point mutation in the ncr locus identified in cd . congenic mice with consequences in mouse susceptibility to infection comparative computational analysis of pluripotency in human and mouse stem cells the transition from primary culture to spontaneous immortalization in mouse fibroblast populations studying human respiratory disease in animals -role of induced and naturally occurring models role of mast cells in allergic and non-allergic immune responses: comparison of human and murine data allergic airway inflammation induces migration of mast cell populations into the mouse airway novel site-specific mast cell subpopulations in the human lung animal models of asthma: reprise or reboot? emerging molecular phenotypes of asthma allergens in veterinary medicine immunologic responses following respiratory sensitization to house dust mite allergens in mice understanding asthma using animal models use of the cockroach antigen model of acute asthma to determine the immunomodulatory role of early exposure to gastrointestinal infection t cell homing to epithelial barriers in allergic disease allergic airway inflammation initiates long-term vascular remodeling of the pulmonary circulation respiratory allergy caused by house dust mites: what do we really know? continuous exposure to house dust mite elicits chronic airway inflammation and structural remodeling induction of vascular remodeling in the lung by chronic house dust mite exposure lung responses in murine models of experimental asthma: value of house dust mite over ovalbumin sensitization a novel mouse model of experimental asthma temporal role of chemokines in a murine model of cockroach allergen-induced airway hyperreactivity and eosinophilia animal models of allergic bronchopulmonary aspergillosis murine models of airway fungal exposure and allergic sensitization mouse models of acute exacerbations of allergic asthma mouse models of severe asthma: understanding the mechanisms of steroid resistance, tissue remodelling and disease exacerbation novel variants of ser-pin a gene: interplay between alpha -antitrypsin deficiency and chronic obstructive pulmonary disease different lung responses to cigarette smoke in two strains of mice sensitive to oxidants collagenase expression in the lungs of transgenic mice causes pulmonary emphysema tissue remodelling in pulmonary fibrosis idiopathic pulmonary fibrosis exploring animal models that resemble idiopathic pulmonary fibrosis the role of mouse strain differences in the susceptibility to fibrosis: a systematic review murine models of pulmonary fibrosis highly selective endothelin- receptor a inhibition prevents bleomycin-induced pulmonary inflammation and fibrosis in mice (r)-resolvin d ameliorates bleomycin-induced pulmonary fibrosis in mice extracellular matrix alterations and acute inflammation; developing in parallel during early induction of pulmonary fibrosis smad signaling involved in pulmonary fibrosis and emphysema adenovector-mediated gene transfer of active transforming growth factor-beta induces prolonged severe fibrosis in rat lung the ethics of animal research. talking point on the use of animals in scientific research key: cord- -zazr uj authors: taif, khasrouf; ugail, hassan; mehmood, irfan title: cast shadow generation using generative adversarial networks date: - - journal: computational science - iccs doi: . / - - - - _ sha: doc_id: cord_uid: zazr uj we propose a computer graphics pipeline for d rendered cast shadow generation using generative adversarial networks (gans). this work is inspired by the existing regression models as well as other convolutional neural networks such as the u-net architectures which can be geared to produce believable global illumination effects. here, we use a semi-supervised gans model comprising of a patchgan and a conditional gan which is then complemented by a u-net structure. we have adopted this structure because of its training ability and the quality of the results that come forth. unlike other forms of gans, the chosen implementation utilises colour labels to generate believable visual coherence. we carried forth a series of experiments, through laboratory generated image sets, to explore the extent at which colour can create the correct shadows for a variety of d shadowed and un-shadowed images. once an optimised model is achieved, we then apply high resolution image mappings to enhance the quality of the final render. as a result, we have established that the chosen gans model can produce believable outputs with the correct cast shadows with plausible scores on psnr and ssim similarity index metrices. shadow generation is a popular computer graphics topic. depending on the level of realism required, the algorithms can be real-time such as shadow mapping [ ] and shadow projection [ ] , or precomputed such as ray marching techniques, which are an expensive way to generate realistic shadows. thus, pre-computing often become a compelling route to take, such as [ ] , [ ] and [ ] . generative adversarial networks have been implemented widely to perform graphical tasks, as it requires minimum to no human interaction, which gives gans a great advantage over conventional deep learning methods, such as image-to-image translation with single d, g semi-supervised model [ ] or unsupervised dual learning [ ] . we apply image-to-image translation to our own image set to generate correct cast shadows for d rendered images in a semi-supervised manner using colour labels. we then augment a high-resolution image to enhance the overall quality. this approach can be useful in real-time scenarios, such as in games and augmented reality applications, since recalling a pre-trained model is less costly in time and quality compared to d real-time rendering, which often sacrifices realism to enhance performance. our approach eliminates the need to constantly render shadows within a d environment and only recalls a trained model at the image plane using colour maps, which could easily be generated in any d software. the model we use utilises a combination of patchgan and conditional gan, because of their ability to compensate for missing training data, and tailor the output image to the desired task. there are many benefits of applying gans to perform computer graphic tasks. gans can interpret predictions from missing training data, which means a smaller training data set compared to classical deep learning models. gans can operate in multi modal scenarios, and a single input can generalise to multiple correct answers that are acceptable. also, the output images are sharper due to how gans learn the cost function, which is based on real-fake basis rather than traditional deep learning models as they minimise the euclidean distance by averaging all plausible outputs, which usually produces blurry results. finally, gan models do not require label annotations nor classifications. the rest of this paper is structured as follows. section reviews related work in terms of traditional shadow algorithms, machine learning and gans, then in sect. we explain the construction of the generative model we used. in sect. we present our experiments from general gan to cgan and dcgan and ending with pix pix. in sect. we discuss our results. then, in sects. and we present the conclusion and future work, respectively. recent advancements in machine learning has benefited computer graphics applications immensely. in terms of d representation, modelling chairs of different styles from large public domain d cad models was proposed by [ ] . [ ] applied deep belief network to create representation of volumetric shapes. similarly, supervised learning can be used to generate chairs, tables, and cars using upconvolutional networks [ ] , [ ] . furthermore, the application of cnn extended into rendering techniques, such as enabling global illumination fast rendering in a single scene [ ] , and image based relighting from small number of images [ ] . another application filters monte carlo noise using a non-linear regression model [ ] . deep convolutional inverse graphics network (dc-ign) enabled producing variations of lighting and pose of the same object from a single image [ ] . such algorithms can provide full deep shading, by training end to end to produce dense per-pixel output [ ] . one of the recent methods applies multiple algorithms to achieve real-time outputs, which is based on a recommendation system that learns the user's preference with the help of a cnn, and then allow the user to make adjustments using a latent space variant system [ ] . gans have been implemented widely to perform graphical tasks. in cgans, for instance, feeding a condition into both d, g networks is essential to control output images [ ] . training the domain-discriminator to maintain relevancy between input image and generated image, to transfer input domain into a target domain in semantic level and generate target images at pixel level [ ] . semi-supervised and unsupervised models has been approached in various ways such as trading mutual information between observed and predicted categori-cal class information [ ] . this can be done by enabling image translators to be trained from two unlabelled images from two domains [ ] , or by translating both the image and its corresponding attributes while maintaining the permutation invariance property of the instance [ ] , or even by training a generator along with a mask generator that models the missing data distribution [ ] . [ ] proposed an image-to-image translation framework based on coupled gans that learns a joint distribution of images in different domains by using images from the marginal distributions in individual domains. also, [ ] uses an unsupervised model that performs image-to-image translation from few images. [ ] made use of other factors such as structure and style. another method eliminated the need for image pairs, by training the distribution of g : x → y until g(x) is indistinguishable from the y distribution which is demonstrated with [ ] . another example is by discovering cross-domain relations [ ] . another way is forcing the discriminator to produce class labels by predicting which of n + class the input belongs to during training [ ] . another model can generate d objects from probabilistic space with volumetric cnn and gans [ ] . the model we use in this paper is a tensorflow port of the pytorch image-to-image translation presented by isola et al. [ ] . this approach can be generalised to any semi-supervised model. however, this model serves us better for its two network types; patchgan, which allows better learning for interpreting missing data and partial data generation. and conditional gan, which allows semisupervised learning to facilitate control over desired output images by using colour labels. it also replaces the traditional discriminator with a u-net structure with a step, this serves two purposes. first, it solves the known drop model issue in traditional gan structure. second, it helps transfer more features across the bottle neck which reduces blurriness and outputs larger and higher quality images. the objective of gan is to train two networks (see fig. ) to learn the correct mapping function through gradient descent to produce outputs believable to the human eye y. the conditional gan here learns from observed image x and the random noise z, such that, where x is a random noise vector, z is an observed image, y is the output image. the generator g, and a discriminator d operate on "real" or "fake" basis. this is achieved by training both networks simultaneously with different objectives, g is trained to produce as realistic images as possible, while d is trained to distinguish which are fake, thus conditional gan (cgan ) can be expressed as, the aim here for g to minimise the objective against the discriminator d which aims to maximise it such that, by comparing it to an unconditional variant where the discriminator does not observe x it becomes here the distance of l l is used instead of l l to reduce blurring such that, the final objective becomes, both networks follow the convolution-batchnorm-relu structure. however, the generator differs by following the general u-net structure, and the discriminator is based on markov random fields. the application of a u-net model allows better information flow across the network than the encoder-decoder model by adding skip connections over bottle necks between layer i and layer n − i, where n is the total number of layers, by concatenating all channels at layer i with the ones in n − i, thus, producing sharper images. for discriminator d, a patchgan of n × n is applied to minimize blurry results by treating the image in small patches that are classified as real-fake across the image, then averaged to produce the accumulative results of d, such that the image is modelled as a markov random field, isola et al. [ ] refers to this patchgan as a form of texture/style loss. the optimisation process alternates descent steps between d and g, by training the model to maximize log d(x, g(x, z)) and dividing the objective by to slow the learning rate of d, minibatch stochastic gradient descend with adam solver is applied at the rate of : for learning, and its momentum parameters are set to β = . , β = : . this allows the discriminator to compare minibatch samples of both generated and real samples. the g network runs at the same setting as the training phase at inference time. dropout and batch normalization to the test batch is applied at test time with batch size of . finally, random jitter is applied by extending the × input image size to × and then crop back to its original size of × . further processing using photoshop, is applied manually to enhance the quality of output image, by mapping a higher resolution render of the model over the output model image, thus delivering a more realistic final image. here we report our initial experiments for shadow generation as well as minor shading functions to support it. our approach is data driven; it focuses on adjusting the image set in every iteration to achieve the correct output. for that we manually created the conditions we intended to test. also, our image set is created using maya with arnold renderer. all of our experiments are conducted on an hp pavilion laptop with . ghz intel core i processor, ghz of ram and nvidia geforce gtx m graphics card. we start with the assumption that gans can generate both soft and hard shadows on demand, using colour labels and given a relatively small training image set. our evaluation is based on both real-fake basis as well as similarity index matrices. real-fake implies that the images can be clearly evaluated visually, for the network itself does not allow poor quality images by design. the similarity index matrices applied here are proposed by [ ] , namely, psnr which measure the peak signal to noise ratio, which is scored between and , and ssim which computes the ratio between the strength of the maximum achievable power of the reconstructed signal and the strength of the corrupted noisy signal, which is scored between and . for our image-to-image translation approach, we started with a small image set of images: for training, for testing and validation of random cube renders, with different lighting intensities and views. the results showed correct translations as shown in fig. (b) top section. however, the output colours of the background sometimes differed from the target images. also, some of the environments contained patches and tiling artefacts as shown in fig. bottom section. which is understandable given the small number of training images. next, we trained the model with stonehenge images. it is also lit with a single light minus the variable intensity. the camera rotates degrees around the y axis. the total image number is images, for training and for testing and validation. we started with the question; can we generate shadows for non-shadowed images that are not seen during training? we worked around it by designing the colour label to our specific need. in the validation step we fed images with no shadows as in fig. (c) , paired with colour labels that contains the correct shadows (a). as the results show in fig. (b) , the model translated the correct shadow with the appropriate amount of drop off to the input image. next we explore generating only accurate shadows fig. (b) for nonshadowed images (c), which is accomplished by constructing the colour map to only contain shadows (a), while training the network with full shadowed images as previously, paired with shadowed labels. the results show accurate translation of shadow direction and intensity (see fig. for the third and final set of experiments, we used two render setups of the stanford dragon, one for training, the second for testing and validation. the camera setup that rotate degrees around the dragon's y axis with different step angle from training and testing. also, the camera elevates degrees across the x axis to show a more complex and elevated view of the dragon model rather than an orthographic one. the image set is composed of training images, testing images and for validation. the training set (shown in table ) is broken into multiple categories, with each one represented within images with overlapping features sometimes. these images help understand how input image/label affects the behaviour of the output images. for example, if we fed a standard colour map to a coloured image, will it be able to translate the colours across or will the output be of standard colour, this will better inform the training process for future applications. all networks are trained from scratch using our image sets, and the weights are initialized from a gaussian distribution with mean and standard deviation . . the test image set consisted of images that are not seen in the training phase. during this, some of the features learnt from the training set are tested and the weights are adjusted accordingly. for validation, a set of images that are not seen in the training set were used, they are derived from the testing image set, but have been modified heavily. the objective here is to test the ability to generate soft and hard shadows from unseen label images, as well as colour shadows, partial shadows, shadow generation for images with no shadows. also, in some cases we have taken the image set to extremes in order to generate shadows for images that has different orientation and colour maps than their original labels. from here, the experiments progressed in three phases. first, we trained the model with the focus on standard labels to produce soft and hard shadows, using an image set of images, of them are dedicated for standard colours and shadows. the remaining images are an arbitrary collection of cases mentioned in the training section (table ) , with similar arbitrary testing and validation image sets of images each. in this experiment, we had the standard colours and shadows provide comprehensive -degree view of the d model. while purposefully choosing arbitrary coloured samples of - images, we created colour variations for both images and their respective colour labels. our first objective was to observe whether the model can generalise colour information into texture, meaning to fill the details of the image from what has been learnt from the view, and overlaying colour information on top of it. even though the general details can be seen on the models, there were heavy artefacts such as blurring and tiling in some of the coloured images with fewer training images. with that knowledge in mind, a second experiment was carried out. we adjusted the image set to reduce the tiling effect, which is mainly due to the lack of sufficient training images for specific cases. hence, the number of training images was increased to per case, to increase the training set to images. in the validation image set, we pushed the code to extreme cases, such as pairing images of different colour maps and different directions, as well as partial and non-shadowed images. thus, accumulating the validation set to images, assuming beforehand that we will get the same tiling effect from the previous experiment in cases where we have different angles or different colours. the significant training time is one of the main challenges that we face, as the training time for our set of experiments ranged between days to one week using fig. . the initial phase of the third set, which shows direct translation tasks. our focus here is mainly the ability to generate believable soft and hard shadows, as well as inpainting for missing patches. our laptop, which is considered long for training images. this why our image set was limited to × pixels. for this work, we overcome the issue with augmenting the output image with a high-resolution render. once trained, however, with some optimisation the model should be capable of real-time execution, but this issue has not been tested by us. the two biggest limitations for this method are, it is a still semi-supervised model that needs a pair of colour maps and a target image. the second limitation is that the colour maps and image pairs are manually created, and the process is labour intensive. these issues should be considered for future work. our method performed well in almost all cases with minimal to no errors and sharp image reproduction, especially when faced with direct translation tasks, such as fig. , and colour variations fig. . even with partial and non-shadowed images, the colours remained consistent and translated correctly across most outputs. this is promising, given a relatively small training set (approximately images per case) we have used. by examining fig. , we notice that the model generalised correctly in most cases even though the colour maps are non-synchronised. this means our method has the breadth to interpret correctly when training set fails. however, it tends to take a more liberal approach in translating colour difference between the label and target with bias towards the colour map. this was also visible in partial images, non-shadowed images, as well as soft and hard shadows. the network struggled mostly when more than two parameters are changed, for example, a partial image and non-shadowed model will translate well. however, partial image, non-synchronised shadow and position will start to show tiling in the output image. the model seems to struggle the most with position switching than any other change, especially when paired with non-synchronised colour map as well. this is usually manifested in the form of noise, blurring and tiling (see fig. ), while the colours remain consistent and true to training images, and the shadows are correct in shape and intensity but are produced with noise and tiling artefacts. we conducted our quantitative assessment by applying the similarity matrices psnr and ssim [ ] and we can confirm the previous observations. when looking at the table , the lowest score were in the categories with nonsynchronized image pairs such as categories and , while the image pairs that were approximately present in both training and testing performed the highest, which are categories and , with overall performance leaning towards a higher scores spectrum. table . this table shows how each category performed in both psnr and ssim similarity indices between output images and corresponding ground truth images. this paper explored a framework based on conditional gans using a pix pix tensorflow port to perform computer graphic functions, by instructing the network to successfully generate shadows for d rendered images given training images paired with conditional colour labels. to achieve this, a variety of image sets were created using an off-the-shelf d program. the first set targeted soft and hard shadows under standard conditions and coloured labels and backgrounds, using colour variations in the training set to test different variations, such as partial and non-shadowed images. the image set consisted of training images and images for testing and validation. the results were plausible in most cases but showed clear blurring and tiling with coloured samples that did not have enough training images paired with it. next, we updated the image set to images, with training images, providing an equal number of training images for each of the cases. we used images for testing, and for validation, which included more variations such as partial and non-shadowed images. in the validation set, the images included extreme cases such as non-sync pairing of position and colour. the results were believable in all cases, except the extreme cases, which resulted in tiling and blurring. the results are promising for shadow generation especially when challenged to produce accurate partial shadows from training image set. the model is reliable to interpret successful output for images not seen during the training phase, except when paired with different colours viewpoints. however, there are still challenges to resolve. for example, the model requires a relatively long time to be trained and the output images still suffer from minor blurriness. this is only a proof of concept. the next logical step is to optimise the process by training the model to create highly detailed renders from lower poly-count models. this also can be tested with video-based models such as hdgan. it is expected to output flickering results due to its learning nature and current state of the art. another direction of interest may be to automate the generation of colour maps from video or live feed such as the work in [ ] . the main challenge, however, is the computation complexity, especially for higher resolution training. seeing d chairs: exemplar part-based d- d alignment using a large dataset of cad models a hierarchical volumetric shadow algorithm for single scattering me and my (fake) shadow learning to generate chairs, tables and cars with convolutional networks learning to generate chairs with convolutional neural networks proceedings of the acm siggraph symposium on interactive d graphics and games image-to-image translation with conditional adversarial networks a machine learning approach for filtering monte carlo noise learning to discover cross-domain relations with generative adversarial networks deep convolutional inverse graphics network misgan: learning from incomplete data with generative adversarial networks unsupervised image-to-image translation networks few-shot unsupervised image-to-image translation conditional generative adversarial nets instagan: instance-aware image-to-image translation deep shading: convolutional neural networks for screen-space shading conditional image synthesis with auxiliary classifier gans image based relighting using neural networks global illumination with radiance regression functions unsupervised and semi-supervised learning with categorical generative adversarial networks generative image modeling using style and structure adversarial networks image quality assessment: from error visibility to structural similarity casting curved shadows on curved surfaces learning a probabilistic latent space of object shapes via d generative-adversarial modeling d shapenets: a deep representation for volumetric shapes dualgan: unsupervised dual learning for image-to-image translation pixel-level domain transfer precomputed shadow fields for dynamic scenes unpaired image-to-image translation using cycle-consistent adversarial networks gaussian material synthesis key: cord- -hgowgq authors: zhang, ruixi; zen, remmy; xing, jifang; arsa, dewa made sri; saha, abhishek; bressan, stéphane title: hydrological process surrogate modelling and simulation with neural networks date: - - journal: advances in knowledge discovery and data mining doi: . / - - - - _ sha: doc_id: cord_uid: hgowgq environmental sustainability is a major concern for urban and rural development. actors and stakeholders need economic, effective and efficient simulations in order to predict and evaluate the impact of development on the environment and the constraints that the environment imposes on development. numerical simulation models are usually computation expensive and require expert knowledge. we consider the problem of hydrological modelling and simulation. with a training set consisting of pairs of inputs and outputs from an off-the-shelves simulator, we show that a neural network can learn a surrogate model effectively and efficiently and thus can be used as a surrogate simulation model. moreover, we argue that the neural network model, although trained on some example terrains, is generally capable of simulating terrains of different sizes and spatial characteristics. an article in the nikkei asian review dated september warns that both the cities of jakarta and bangkok are sinking fast. these iconic examples are far from being the only human developments under threat. the united nation office for disaster risk reduction reports that the lives of millions were affected by the devastating floods in south asia and that around , people died in the bangladesh, india and nepal [ ] . climate change, increasing population density, weak infrastructure and poor urban planning are the factors that increase the risk of floods and aggravate consequences in those areas. under such scenarios, urban and rural development stakeholders are increasingly concerned with the interactions between the environment and urban and rural development. in order to study such complex interactions, stakeholders need effective and efficient simulation tools. a flood occurs with a significant temporary increase in discharge of a body of water. in the variety of factors leading to floods, heavy rain is one of the prevalent [ ] . when heavy rain falls, water overflows from river channels and spills onto the adjacent floodplains [ ] . the hydrological process from rainfall to flood is complex [ ] . it involves nonlinear, time-varying interactions between rain, topography, soil types and other components associated with the physical process. several physics-based hydrological numerical simulation models, such as hec-ras [ ] , lisflood [ ] , lisflood-fp [ ] , are commonly used to simulate floods. however, such models are usually computation expensive and expert knowledge is required for both design and for accurate parameter tuning. we consider the problem of hydrological modelling and simulation. neural network models are known for their flexibility, efficient computation and capacity to deal with nonlinear correlation inside data. we propose to learn a flood surrogate model by training a neural network with pairs of inputs and outputs from the numerical model. we empirically demonstrate that the neural network can be used as a surrogate model to effectively and efficiently simulate the flood. the neural network model that we train learns a general model. with the trained model from a given data set, the neural network is capable of simulating directly spatially different terrains. moreover, while a neural network is generally constrained to a fixed size of its input, the model that we propose is able to simulate terrains of different sizes and spatial characteristics. this paper is structured as follows. section summarises the main related works regarding physics-based hydrological and flood models as well as statistical machine learning models for flood simulation and prediction. section presents our methodology. section presents the data set, parameters setting and evaluation metrics. section describes and evaluates the performance of the proposed models. section presents the overall conclusions and outlines future directions for this work. current flood models simulate the fluid movement by solving equations derived from physical laws with many hydrological process assumptions. these models can be classified into one-dimensional ( d), two-dimensional ( d) and threedimensional ( d) models depending on the spatial representation of the flow. the d models treat the flow as one-dimension along the river and solve d saint-venant equations, such as hec-ras [ ] and swmm [ ] . the d models receive the most attention and are perhaps the most widely used models for flood [ ] . these models solve different approximations of d saint-venant equations. two-dimensional models such as hec-ras d [ ] is implemented for simulating the flood in assiut plateau in southwestern egypt [ ] and bolivian amazonia [ ] . another d flow models called lisflood-fp solve dynamic wave model by neglecting the advection term and reduce the computation complexity [ ] . the d models are more complex and mostly unnecessary as d models are adequate [ ] . therefore, we focus our work on d flow models. instead of a conceptual physics-based model, several statistical machine learning based models have been utilised [ , ] . one state-of-the-art machine learning model is the neural network model [ ] . tompson [ ] uses a combination of the neural network models to accelerate the simulation of the fluid flow. bar-sinai [ ] uses neural network models to study the numerical partial differential equations of fluid flow in two dimensions. raissi [ ] developed the physics informed neural networks for solving the general partial differential equation and tested on the scenario of incompressible fluid movement. dwivedi [ ] proposes a distributed version of physics informed neural networks and studies the case on navier-stokes equation for fluid movement. besides the idea of accelerating the computation of partial differential equation, some neural networks have been developed in an entirely data-driven manner. ghalkhani [ ] develops a neural network for flood forecasting and warning system in madarsoo river basin at iran. khac-tien [ ] combines the neural network with a fuzzy inference system for daily water levels forecasting. other authors [ , ] apply the neural network model to predict flood with collected gauge measurements. those models, implementing neural network models for one dimension, did not take into account the spatial correlations. authors of [ , ] use the combinations of convolution and recurrent neural networks as a surrogate model of navier-stokes equations based fluid models with a higher dimension. the recent work [ ] develops a convolutional neural network model to predict flood in two dimensions by taking the spatial correlations into account. the authors focus on one specific region in the colorado river. it uses a convolutional neural network and a conditional generative adversarial network to predict water level at the next time step. the authors conclude neural networks can achieve high approximation accuracy with a few orders of magnitude faster speed. instead of focusing on one specific region and learning a model specific to the corresponding terrain, our work focuses on learning a general surrogate model applicable to terrains of different sizes and spatial characteristics with a datadriven machine learning approach. we propose to train a neural network with pairs of inputs and outputs from an existing flood simulator. the output provides the necessary supervision. we choose the open-source python library landlab, which is lisflood-fp based. we first define our problem in subsect. . . then, we introduce the general ideas of the numerical flood simulation model and landlab in subsect. . . finally, we present our solution in subsect. . . we first introduce the representation of three hydrological parameters that we use in the two-dimensional flood model. a digital elevation model (dem) d is a w × l matrix representing the elevation of a terrain surface. a water level h is a w × l matrix representing the water elevation of the corresponding dem. a rainfall intensity i generally varies spatially and should be a matrix representing the rainfall intensity. however, the current simulator assumes that the rainfall does not vary spatially. in our case, i is a constant scalar. our work intends to find a model that can represent the flood process. the flood happens because the rain drives the water level to change on the terrain region. the model receives three inputs: a dem d, the water level h t and the rainfall intensity i t at the current time step t. the model outputs the water level h t+ as the result of the rainfall i t on dem d. the learning process can be formulated as learning the function l: physics-driven hydrology models for the flood in two dimensions are usually based on the two-dimensional shallow water equation, which is a simplified version of navier-stokes equations with averaged depth direction [ ] . by ignoring the diffusion of momentum due to viscosity, turbulence, wind effects and coriolis terms [ ] , the two-dimensional shallow water equations include two parts: conservation of mass and conservation of momentum shown in eqs. and , where h is the water depth, g is the gravity acceleration, (u, v) are the velocity at x, y direction, z(x, y) is the topography elevation function and s fx , s fy are the friction slopes [ ] which are estimated with friction coefficient η as for the two-dimensional shallow water equations, there are no analytical solutions. therefore, many numerical approximations are used. lisflood-fp is a simplified approximation of the shallow water equations, which reduces the computational cost by ignoring the convective acceleration term (the second and third terms of two equations in eq. ) and utilising an explicit finite difference numerical scheme. the lisflood-fp firstly calculate the flow between pixels with mass [ ] . for simplification, we use the d version of the equations in x-direction shown in eq. , the result of d can be directly transferable to d due to the uncoupled nature of those equations [ ] . then, for each pixel, its water level h is updated as eq. , to sum up, for each pixel at location i, j, the solution derived from lisflood-fp can be written in a format shown in eq. , where h t i,j is the water level at location i, j of time step t, or in general as h t+ = Θ (d, h t , i t ) . however, the numerical solution as Θ is computationally expensive including assumptions for the hydrology process in flood. there is an enormous demand for parameter tuning of the numerical solution Θ once with high-resolution two-dimensional water level measurements mentioned in [ ] . therefore, we use such numerical model to generate pairs of inputs and outputs for the surrogate model. we choose the lisflood-fp based opensource python library, landlab [ ] since it is a popular simulator in regional two-dimensional flood studies. landlab includes tools and process components that can be used to create hydrological models over a range of temporal and spatial scales. in landlab, the rainfall and friction coefficients are considered to be spatially constant and evaporation and infiltration are both temporally and spatially constant. the inputs of the landlab is a dem and a time series of rainfall intensity. the output is a times series of water level. we propose here that a neural network model can provide an alternative solution for such a complex hydrology dynamic process. neural networks are well known as a collection of nonlinear connected units, which is flexible enough to model the complex nonlinear mechanism behind [ ] . moreover, a neural network can be easily implemented on general purpose graphics processing units (gpus) to boost its speed. in the numerical solution of the shallow water equation shown in subsect. . , the two-dimensional spatial correlation is important to predict the water level in flood. therefore, inspired by the capacity to extract spatial correlation features of the neural network, we intend to investigate if a neural network model can learn the flood model l effectively and efficiently. we propose a small and flexible neural network architecture. in the numerical solution eq. , the water level for each pixel of the next time step is only correlated with surrounding pixels. therefore, we use, as input, a × sliding window on the dem with the corresponding water levels and rain at each time step t. the output is the corresponding × water level at the next time step t + . the pixels at the boundary have different hydrological dynamic processes. therefore, we pad both the water level and dem with zero values. we expect that the neural network model learns the different hydrological dynamic processes at boundaries. one advantage of our proposed architecture is that the neural network is not restricted by the input size of the terrain for both training and testing. therefore, it is a general model that can be used in any terrain size. figure illustrates the proposed architecture on a region with size × . in this section, we empirically evaluate the performance of the proposed model. in subsect. . , we describe how to generate synthetic dems. subsect. . presents the experimental setup to test our method on synthetic dems as a micro-evaluation. subsect. . presents the experimental setup on the case in onkaparinga catchment. subsect. . presents details of our proposed neural network. subsect. . shows the evaluation metrics of our proposed model. in order to generate synthetic dems, we modify alexandre delahaye's work . we arbitrarily set the size of the dems to × and its resolution to metres. we generate three types of dems in our data set that resembles real world terrains surface as shown in fig. a , namely, a river in a plain, a river with a mountain on one side and a plain on the other and a river in a valley with mountains on both sides. we evaluate the performance in two cases. in case , the network is trained and tested with one dem. this dem has a river in the valley with mountains on both sides, as shown in fig. a right. in case , the network is trained and tested with different synthetic dems. the data set is generated with landlab. for all the flood simulations in landlab, the boundary condition is set to be closed on four sides. this means that rainfall is the only source of water in the whole region. the roughness coefficient is set to be . . we control the initial process, rainfall intensity and duration time for each sample. the different initial process is to ensure different initial water level in the whole region. after the initial process, the system run for h with no rain for stabilisation. we run the simulation for h and record the water levels every min. therefore, for one sample, we record a total of time steps of water levels. table summarises the parameters for generating samples in both case and case . the onkaparinga catchment, located at lower onkaparinga river, south of adelaide, south australia, has experienced many notable floods, especially in and . many research and reports have been done in this region [ ] . we get two dem data with size × and × from the australia intergovernmental committee on surveying and mapping's elevation information system . figure b shows the dem of lower onkaparinga river. we implement the neural network model under three cases. in case , we train and test on × onkaparinga river dem. in case , we test × onkaparinga river dem directly with case trained model. in case , we test × onkaparinga river dem directly with case trained model. we generate the data set for both × and × dem from landlab. the initial process, rainfall intensity and rain duration time of both dem are controlled the same as in case . the architecture of the neural network model is visualized as in fig. . it firstly upsamples the rain input into × and concatenates it with × water level input. then, it is followed by several batch normalisation and convolutional layers. the activation functions are relu and all convolutional layers have the same size padding. the total parameters for the neural network are . the model is trained by adam with the learning rate as − . the batch size for training is . the data set has been split with ratio : : for training, validation and testing. the training epoch is for case and case and for case . we train the neural network model on a machine with a ghz amd ryzen tm - -core processor. it has a gb ddr memory and an nvidia gtx ti gpu card with cuda cores and gb memory. the operating system is ubuntu . os. in order to evaluate the performance of our neural network model, we use global measurements metrics for the overall flood in the whole region. these metrics are global mean squared error: case is to test the scalability of our model for the different size dem. in table b , for global performance, the mape of case is around % less than both case and case , and for local performance, the mape of case is . %. similarly, without retraining the existed model, the trained neural network from case can be applied directly on dem with different size with a good global performance. we present the time needed for the flood simulation of one sample in landlab and in our neural network model (without the training time) in table . the average time of the neural network model for a × dem is around . s, while it takes s in landlab. furthermore, for a × dem, landlab takes more time than the neural network model. though the training of the neural network model is time consuming, it can be reused without further training or tuning terrains of different sizes and spatial characteristics. it remains effective and efficient (fig. ). we propose a neural network model, which is trained with pairs of inputs and outputs of an off-the-shelf numerical flood simulator, as an efficient and effective general surrogate model to the simulator. the trained network yields a mean absolute percentage error of around %. however, the trained network is at least times faster than the numerical simulator that is used to train it. moreover, it is able to simulate floods on terrains of different sizes and spatial characteristics not directly represented in the training. we are currently extending our work to take into account other meaningful environmental elements such as the land coverage, geology and weather. hec-ras river analysis system, user's manual, version the landlab v . overlandflow component: a python tool for computing shallow-water flow across watersheds improving the stability of a simple formulation of the shallow water equations for -d flood modeling a review of surrogate models and their application to groundwater modeling learning data-driven discretizations for partial differential equations a simple raster-based model for flood inundation simulation a simple inertial formulation of the shallow water equations for efficient two-dimensional flood inundation modelling rainfall-runoff modelling: the primer hec-ras river analysis system hydraulic userś manual numerical solution of the two-dimensional shallow water equations by the application of relaxation methods distributed physics informed neural network for data-efficient solution to partial differential equations integrating gis and hec-ras to model assiut plateau runoff flood hydrology processes and their variabilities application of surrogate artificial intelligent models for real-time flood routing extreme flood estimation-guesses at big floods? water down under : surface hydrology and water resources papers the data-driven approach as an operational real-time flood forecasting model analysis of flood causes and associated socio-economic damages in the hindukush region deep fluids: a generative network for parameterized fluid simulations fully convolutional networks for semantic segmentation optimisation of the twodimensional hydraulic model lisfood-fp for cpu architecture neural network modeling of hydrological systems: a review of implementation techniques physics informed data driven model for flood prediction: application of deep learning in prediction of urban flood development application of d numerical simulation for the analysis of the february bolivian amazonia flood: application of the new hec-ras version physics-informed neural networks: a deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations storm water management model-user's manual v. . . us environmental protection agency hydrologic engineering center hydrologic modeling system, hec-hms: interior flood modeling decentralized flood forecasting using deep neural networks flood inundation modelling: a review of methods, recent advances and uncertainty analysis accelerating eulerian fluid simulation with convolutional networks comparison of the arma, arima, and the autoregressive artificial neural network models in forecasting the monthly inflow of dez dam reservoir lisflood: a gis-based distributed model for river basin scale water balance and flood simulation real-time waterlevel forecasting using dilated causal convolutional neural networks latent space physics: towards learning the temporal evolution of fluid flow in-situ water level measurement using nirimaging video camera acknowledgment. this work is supported by the national university of singapore institute for data science project watcha: water challenges analytics. abhishek saha is supported by national research foundation grant number nrf vsg-at dcm - . key: cord- -xdymj dk authors: ranjan, rajesh title: temporal dynamics of covid- outbreak and future projections: a data-driven approach date: - - journal: trans indian natl doi: . /s - - -y sha: doc_id: cord_uid: xdymj dk long-term predictions for an ongoing epidemic are typically performed using epidemiological models that predict the timing of the peak in infections followed by its decay using non-linear fits from the available data. the curves predicted by these methods typically follow a gaussian distribution with a decay rate of infections similar to the climbing rate before the peak. however, as seen from the recent covid- data from the us and european countries, the decay in the number of infections is much slower than their increase before the peak. therefore, the estimates of the final epidemic size from these models are often underpredicted. in this work, we propose two data-driven models to improve the forecasts of the epidemic during its decay. these two models use gaussian and piecewise-linear fits of the infection rate respectively during the deceleration phase, if available, to project the future course of the pandemic. for countries, which are not yet in the decline phase, these models use the peak predicted by epidemiological models but correct the infection rate to incorporate a realistic slow decline based on the trends from the recent data. finally, a comparative study of predictions using both epidemiological and data-driven models is presented for a few most affected countries. in recent days, coronavirus disease (covid- ) has emerged as an unprecedented challenge before the world. this disease is caused by a novel coronavirus sars-cov- , for which there is no specific medication or vaccine approved by medical authorities. this disease is transmitted by inhalation or contact with infected droplets or fomites, and the incubation period may range from to days (wu and mcgoogan ) . this disease can be fatal to the elderly patients (about % for + age groups), and those with underlying co-morbid conditions . as of may , , there have been about . million confirmed cases of covid- and about , reported deaths globally. a realistic estimate of intensity and temporal distribution of this epidemic can be beneficial to design key strategies to regulate the quarantine as well as to prepare for social and economic consequences due to lockdown. however, as seen from the recent literature (roda et al. ) , the predictions by epidemiological models for an ongoing spread are often unreliable as they do not accurately capture the dynamics of covid- in the absence of established parameters. in this work, we propose data-driven models for covid- decay purely based on characteristics of covid- spread, and thus include the effects of lockdown and other key factors. for subsequent discussions, the default year is and all the statistics are based on data till may , , unless otherwise specified. first, we examine the dynamics of covid- , before and after the lockdown as shown in fig. a . the abscissa indicates the days shifted by the date when the lockdown was imposed or other intervention measures were taken (see list in ranjan b) . thus, the four phases indicate: ( ) early slow epidemic growth ( t < t − ), ( ) initial exponential growth ( t − < t < t ) typical of an epidemic, ( ) continuing exponential growth during lockdown based on the incubation period of sars-cov- ( t < t < t ≡ t + ) and ( ) expected deceleration phase ( t > t ). in fig. a , both china and south korea (sk) show a very rapid arrest of the covid- growth post interventions ( t > t ), while other countries display just a slowdown evident by the change in slope. further, the growth curves for india and russia are much more rapid in this phase compared to those for other countries . the differences in covid- spread among geographical regions after the lockdown can be better visualized on a linear scale as shown in fig. a . most of the countries considered in the figure took social distancing measures before the end of march, so it is expected that the effects of interventions should become visible latest by mid-april. both the us and the uk exhibit linear growth in this period, while other european countries show initial linear growth followed by a slow flattening (ansumali and prakash ) . the curves for india and russia are closer to exponential very different trends of the curve after the lockdown indicate a disparity in compliance levels of social distancing measures. for example, in the us, each state follows its norm of intervention, and social distancing measures are imposed on different dates. an implication of this is that when the initially most impacted states like new york and new jersey started showing signs of flattening in late april, other states like illinois, massachusetts and california displayed surges in the number of cases, thereby keeping the overall growth in the us on a linear course. the case of india is also compelling as it first displayed a strong impact of lockdown (close to % compliance, as suggested in ranjan a) despite a few local outbreaks. this led to a linear growth for some time, but an escalation in early may cases put india on a near exponential course. to further examine this, we plot covid- distribution in key affected states in india in fig. b . the time-series data is divided into four periods, with the first three being before, during and after lockdown similar to that in fig. a and the last one from may (green shade) when a surge in the number of cases in many states put india onto the exponential course. fig. b , we note a varying distribution of covid- among indian states much like in the us, just four states -maharashtra, gujarat, tamilnadu, and new delhi contribute to about % of the total cases. among these, the most affected states, maharashtra and gujarat are on the course of exponential growth. delhi, tamilnadu and west bengal show an initial arrest of the growth (blue shade), followed by later and more recent local outbreaks as marked by a discontinuity in the slope (see green shade in inset). several other states, including uttar pradesh, kerala and karnataka display good control over the epidemic, while other states are in the linear regime throughout from t (beginning of the blue shade). since the predictive models depend significantly on data, an important aspect to consider is that the number of reported infections does not truly reflect the actual outbreak of covid- . the data on infection rate are often limited by the countries' testing capability, which in turn is related to the availability of testing kits, size of healthcare professionals per population, and infrastructure. further, the asymptomatic population is often excluded in testing strategies adopted by most countries. to elucidate this, we show the daily number of tests for key countries in fig. b . we generally note that the increase in the number of tests with time is very closely related to the infections shown in fig. a , as expected. a small number of reported cases for india in march could be due to inadequate testing at that time. therefore the predictions by models using data from that period had considerable uncertainty (ranjan b; singh and adhikari ). we briefly discuss the implication of rigorous testing in covid- control by carefully examining the south korean data, shown in the inset in fig. b . unlike most countries with the number of tests increasing slowly during the initial phase of the outbreak, the reverse is seen for south korea. the response of sk to the outbreak was quick and they ran the most comprehensive and well-organized testing program in the world from february (fig. b ) when the outbreak was still not severe. this, combined with large-scale efforts to isolate infected people and trace and quarantine their contacts, lead to successful control of the outbreak. for comparison, sk, the us, and india respectively have , and tests per million inhabitants on may . a final but most important factor affecting the outbreak and the predictions is the epidemiology of covid- in different geographical regions. the values of epidemiological parameters such as transmission rate, recovery rate, and basic reproduction number ) depend on many social and environmental factors and are dissimilar in different regions. for an ongoing outbreak, the epidemiology is not fully established, but available data can provide meaningful insights. we report three characteristic ratios: positivity ratio (pr), case recovery ratio (crr) and case fatality ratio (cfr) to roughly correlate with the epidemiological parameters: the rates of infection , recovery , and mortality respectively (hethcote ) . pr is the total number of infections for a given number of tests. crr and cfr are respectively, the number of recovered and deceased cases as a fraction of total infections. these values are reported in percentages in table , a large proportion of the young in the total population, and possible immunity due to bcg vaccinations (curtis et al. ) and malarial infections (goswami et al. ) . crrs for germany ( ≃ % ) and spain ( ≃ % ) are highest, but it is expected that the value of crr in countries currently in the acceleration phase will improve with time. the case fatality ratio is very high for france(≃ % ), uk(≃ % ) and italy(≃ % ) compared to the world average of - %. a high ratio may be due to a higher percentage of the elderly population in these countries. it is clear from the above discussion that the epidemiologies of covid- , as well as the impact of social distancing for different countries, are very dissimilar. further, there is an inhomogenous covid- spread within a country as seen for the us and india. all of these factors make the modeling of this epidemic during its progress very challenging. typically, epidemiological models such as a logistic or a compartmental model are preferred for modeling the later stages. however, these models are highly dependent on initial conditions and underlying unknown epidemiological parameters, incorrect estimation of which can give completely different results. a further concern with these models is the prediction of the decay rate of infections, which is generally high compared to the recent trends (ranjan a) . therefore, in this work, we propose two data-driven models for the predictions that incorporate the slow decay of the epidemic post-lockdown and provide more realistic estimates. projections for key affected countries are presented using these data-driven as well as epidemiological models. covid- data used in this study are taken from various sources. modeling is based on the time-series data from johns hopkins university coronavirus data stream, which combines world health organization (who) and centers for disease control and prevention (cdc) case data. data on tests are taken from 'our world in data' source that compiles data from the european centre for disease prevention and control (ecdc). time-series data for indian states are taken from github.com/covid india. typically for an ongoing epidemic, epidemiological models estimate the underlying parameters based on fit from available data and then use simple ordinary differential equations to predict the day of the peak and the decay rate. to illustrate the limitations of these models, we show predictions for italy using a logistic model (ma ) , as well as two compartmental epidemiological models-sir (hethcote ) and generalized seir (seiqrdp) (peng et al. ) in fig. a . open-source matlab codes developed by batista ( ) and cheynet ( ) are used for sir and seoqrdp models respectively. as the epidemic has passed its peak in italy, a key parameter for estimation is the decay rate. a close examination of the daily cases in fig. a shows that all the three models predict a faster decay rate as the curve is nearly symmetric around the peak. this distribution leads to an under-prediction of size as well as the duration of the epidemic. although not shown for every geographical region considered in this paper, this is true for most of the predictions. we shall describe the first data-driven model to improve the predictions. as shown in fig. a , the infection rate (daily cases) predicted by epidemiological models follow a nearly normal distribution with where represents the day with the peak number of cases, is spread around this day from the beginning of the epidemic to the end, and a is a constant that determines the number of cases. because of the nearly symmetric distribution of the curve, the decline rate is typically predicted as the negative of the climb rate. to make the predictions closer to actual values in the deceleration phase, we introduce a new parameter , which changes the variance of this distribution after the peak to make the decline rate more realistic. hence, the new distribution in this modified gaussian decay model (mgdm) is where the pre-multiplication factor ensures that the number of infections on the peak day remains unaltered. the dash-dot magenta curve in fig. a shows the distribution of mgdm. the infection rate, in this case, is closer to the actual values during decay and generally provides the upper limit of estimated total cases, as seen from the difference in the cumulative cases. parameters for fitted normal distribution with r = . , rmse= . and % confidence bounds are: a = ( , ), = − . (− . , − . ), = . ( , . ) . the modifications due to data gives = . , = . . the final epidemic size by this model as well as sir and seiqrdp models are given in table . note that, a gaussian fit of the infection rate can be directly used in mgdm for regions with sufficient data in the deceleration phase, and a prediction from epidemiological model is not necessary. we shall now discuss the second data-driven model. as discussed earlier, recent trends indicate that the lockdown arrests the initial exponential growth but a linear regime persists after that, and then a prolonged decay follows. we propose that this decay can be modeled better with several linear segments than an exponential or a gaussian curve that gives a fast decline with a relatively small tail. piecewise-linear decay model (pldm) incorporates these dynamics. the cumulative data in the deceleration phase is collected and then divided into equal segments. an optimal piecewise linear fit in a least-squares sense is then obtained. the slopes of these linear fits are m i , where i = , ..., n , n being the number of segments. the ratios of slopes are then computed, , with i < , and a slope factor ̄ is calculated by taking an average of the last three ratios. this factor is then used to predict the slope of the next future segment such as m n+ =̄m n and so on. figure a includes the predictions from this model for italy. data during the decay phase, between mar and may , have been divided into five equal segments of eleven days. the modeling gives ̄= . , which is used to predict the slopes of future linear segments of the same sizes (see bottom panel in the figure). as evident from fig. a , while the prediction of the final estimate size for both mgdm and pldm is similar, pldm predicts the cumulative curve more closely and has a more gradual decay. table shows the projections for key countries using both the epidemiological (sir and seiqrdp) as well as datadriven models until the middle of august. though not shown for individual cases, it is ensured in every case that the logistic fits are statistically significant with r > . and p-value < . . parameters of data-driven models ( , in mgdm, and ̄, m n in pldm) are directly obtained from the data in the deceleration phase when available. for countries, where the infection rate is still growing, predictions from the seiqrdp model are used as a baseline, and parameters of data-driven models are taken from the fit used in italy. as expected, there is higher uncertainty for these countries. all european countries in table except uk, where the outbreak is already in the decline phase, show a good convergence of epidemic sizes i.e., predictions from the epidemiological models are not very different as shown in the case of italy (fig. a) . likewise, estimates from both the data-driven models are very close and are higher than those from epidemiological models as expected. both mgdm and pldm suggest the equilibrium to be expected towards the end of july. predictions for these regions using the datadriven models are fairly reliable provided there is no new outbreak. for the us, there is an uncertainty due to fluctuations in the recent data, which in turn is due to different epidemiology and a differential impact of stay-at-home order among different states (ranjan a) . for the uk, the epidemic is in the linear growth stage with mild signs of decay recently. therefore, there is high unreliability in the prediction of the peak. nevertheless, forecasts by different models in these cases can provide an estimate of the expected range. for india and russia, the growths are still close to exponential, and therefore there is a significant disparity in predictions by different epidemiological models. figure b illustrates the uncertainty in such cases by showing the projections for india by different models. while the logistic and sir models predict the peak close to each other, seiqrdp model shows continuing growth till the middle of june before the decline begins. this difference leads to a significantly higher epidemic size with seiqrdp ( . million) than those with logistic ( . million) and sir ( . millions) models. a critical difference between the sir and seiqrdp models implemented here is that in sir, the population n considered is just the number of susceptible persons before the outbreak, while the entire population of the region is taken as the population size in seiqrdp. as results from the seiqrdp model are used as a baseline for data-driven models, estimates from the latter are also in the higher range. these models are then used for statewise projections in india. table gives the lower and upper range value of the estimated epidemic size as calculated by all the models. as expected, the highest contribution comes from four key states: maharashtra, gujarat, delhi and tamilnadu, who are on the exponential growth (fig. b) . also, the projections have the highest uncertainty in these regions among the states listed in the table. if these states can control the epidemic and new outbreaks do not appear in other states, it is expected that the optimistic scenario for india shown by the sir model in fig. b can be realized. the final epidemic size of the entire world is difficult to estimate without getting individual estimates of all the countries. this is because the global trend of total infections is still on an accelerating stage with new countries (brazil, peru, canada) reporting surge in the number of cases. epidemiological models such as logistic and compartmental models are generally used to predict the total size and duration of covid- . however, these models generally do not account for the precise change in dynamics due to different interventions, or a new outbreak, and therefore estimate unrealistic epidemic size. we show that the covid- curves for different countries after the lockdown are very dissimilar with four primary distributions: linear, exponential, and slow and fast flattening. further, within a country, the characteristics of spread among states may be different. therefore, to account for differences in dynamics, a locally data-driven approach for modeling may be more suitable. two data-driven models for the decay of covid- based on recent trends-one based on skewed gaussian distribution and the other by using a piecewise linear fit-are proposed. these models generally provide a more realistic estimate of the epidemic size than epidemiological models for regions in the deceleration phase, with the piecewise linear model predicting a more gradual decay. for countries (like india and russia) still in the growth stage, these data-driven models use predictions from epidemiological models as a baseline and impose corrections using parameters obtained from an available data with a realistic decline rate. the uncertainty in predictions for such cases is higher. the paper also highlights that the reported data on infections is not an accurate representation of actual outbreak, and is limited by the testing capacity. therefore, estimations given by these models could still be optimistic and should be used with caution. a periodic evaluation of characteristics of covid- spread, and thus a revision of projections is necessary. a very flat peak: exponential growth phase of covid- is mostly followed by a prolonged linear growth phase, not an immediate saturation ) fitviruscovid , matlab central file exchange generalized seir epidemic model (fitting and computation) considering bcg vaccination to reduce the impact of covid- interaction between malarial transmission and bcg vaccination with covid- incidence in the world map: a changing landscape human immune system? medrxiv the mathematics of infectious diseases the reproductive number of covid- is higher compared to sars coronavirus estimating epidemic exponential growth rate and basic reproduction number effective transmission across the globe: the role of climate in covid- mitigation strategies epidemic analysis of covid- in china by dynamical modeling estimating the final epidemic size forcovid- outbreak using improved epidemiological models predictions for covid- outbreak in india using epidemiologicalmodels.medrxiv why is it difficult to accurately predict the covid- epidemic? age-structured impact of social distancing on the covid- epidemic in india characteristics of and important lessons from the coronavirus disease (covid- ) outbreak in china: summary of a report of cases from the chinese center for disease control and prevention prevalence of comorbidities in the novel wuhan coronavirus (covid- ) infection: a systematic review and metaanalysis the author would like to thank prof. datta gaitonde for his support and encouragement during these unprecedented times. the critical inputs from prof. roddam narasimha in improving the manuscript are gratefully acknowledged. contributions from dr. sudheendra n r rao (scientific advisor, organization for rare key: cord- -kf lxp authors: nayyeri, mojtaba; vahdati, sahar; zhou, xiaotian; shariat yazdi, hamed; lehmann, jens title: embedding-based recommendations on scholarly knowledge graphs date: - - journal: the semantic web doi: . / - - - - _ sha: doc_id: cord_uid: kf lxp the increasing availability of scholarly metadata in the form of knowledge graphs (kg) offers opportunities for studying the structure of scholarly communication and evolution of science. such kgs build the foundation for knowledge-driven tasks e.g., link discovery, prediction and entity classification which allow to provide recommendation services. knowledge graph embedding (kge) models have been investigated for such knowledge-driven tasks in different application domains. one of the applications of kge models is to provide link predictions, which can also be viewed as a foundation for recommendation service, e.g. high confidence “co-author” links in a scholarly knowledge graph can be seen as suggested collaborations. in this paper, kges are reconciled with a specific loss function (soft margin) and examined with respect to their performance for co-authorship link prediction task on scholarly kgs. the results show a significant improvement in the accuracy of the experimented kge models on the considered scholarly kgs using this specific loss. transe with soft margin (transe-sm) obtains a score of . % hits@ for co-authorship link prediction task while the original transe obtains . %, on the same task. in terms of accuracy and hits@ , transe-sm also outperforms other state-of-the-art embedding models such as complex, conve and rotate in this setting. the predicted co-authorship links have been validated by evaluating profile of scholars. with the rapid growth of digital publishing, researchers are increasingly exposed to an incredible amount of scholarly artifacts and their metadata. the complexity of science in its nature is reflected in such heterogeneously interconnected information. knowledge graphs (kgs), viewed as a form of information representation in a semantic graph, have proven to be extremely useful in modeling and representing such complex domains [ ] . kg technologies provide the backbone for many ai-driven applications which are employed in a number of use cases, e.g. in the scholarly communication domain. therefore, to facilitate acquisition, integration and utilization of such metadata, scholarly knowledge graphs (skgs) have gained attention [ , ] in recent years. formally, a skg is a collection of scholarly facts represented in triples including entities and a relation between them, e.g. (albert einstein, co-author, boris podolsky). such representation of data has influenced the quality of services which have already been provided across disciplines such as google scholar , semantic scholar [ ] , openaire [ ] , aminer [ ] , researchgate [ ] . the ultimate objective of such attempts ranges from service development to measuring research impact and accelerating science. recommendation services, e.g. finding potential collaboration partners, relevant venues, relevant papers to read or cite are among the most desirable services in research of research enquiries [ , ] . so far, most of the approaches addressing such services for scholarly domains use semantic similarity and graph clustering techniques [ , , ] . the heterogeneous nature of such metadata and variety of sources plugging metadata to scholarly kgs [ , , ] keeps complex metaresearch enquiries (research of research) challenging to analyse. this influences the quality of the services relying only on the explicitly represented information. link prediction in kgs, i.e. the task of finding (not explicitly represented) connections between entities, draws on the detection of existing patterns in the kg. a wide range of methods has been introduced for link prediction [ ] . the most recent successful methods try to capture the semantic and structural properties of a kg by encoding information as multi-dimensional vectors (embeddings). such methods are known as knowledge graph embedding (kge) models in the literature [ ] . however, despite the importance of link prediction for the scholarly domains, it has rarely been studied with kges [ , ] for the scholarly domain. in a preliminary version of this work [ ] , we tested a set of embedding models (in their original version) on top of a skg in order to analyse suitability of kges for the use case of scholarly domain. the primary insights derived from results have proved the effectiveness of applying kge models on scholarly knowledge graphs. however, further exploration of the results proved that the many-to-many characteristic of the focused relation, co-authorship, causes restrictions in negative sampling which is a mandatory step in the learning process of kge models. negative sampling is used to balance discrimination from the positive samples in kgs. a negative sample is generated by a replacement of either subject or object with a random entity in the kg e.g., (albert einstein, co-author, trump) is a negative sample for (albert einstein, co-author, boris podolsky). to illustrate the negative sampling problem, consider the following case: assuming that n = is the number of all authors in a skg, the probability of generating false negatives for an author with true or sensible but unknown collaborations becomes = %. this problem is particularly relevant when the in/out-degree of entities in a kg is very high. this is not limited to, but particularly relevant, in scholarly kgs with its network of authors, venues and papers. to tackle this problem, we propose a modified version of the margin ranking loss (mrl) to train the kge models such as transe and rotate. the model is dubbed sm (soft margins), which considers margins as soft boundaries in its optimization. soft margin loss allows false negative samples to move slightly inside the margin, mitigating the adverse effects of false negative samples. our main contributions are: -proposing a novel loss function explicitly designed for kgs with many-tomany relations (present in co-authorship relation of scholarly kgs), -showcasing the effect of the proposed loss function for kge models, -providing co-authorship recommendations on scholarly kgs, -evaluating the effectiveness of the approach and the recommended links on scholarly kgs with favorable results, -validating the predicted co-authorship links by a profile check of scholars. the remaining part of this paper proceeds as follows. section represents details of the scholarly knowledge graph that is created for the purpose of applying link discovery tools. section provides a summary of preliminaries required about the embedding models and presents some of the focused embedding models of this paper, transe and rotate. moreover, other related works in the domain of knowledge graph embeddings are reviewed in sect. . . section contains the given approach and description of the changes to the mrl. an evaluation of the proposed model on the represented scholarly knowledge graph is shown in sect. . in sect. , we lay out the insights and provide a conjunction of this research work. a specific scholarly knowledge graphs has been constructed in order to provide effective recommendations for the selected use case (co-authorship). this knowledge graph is created after a systematic analysis of the scholarly metadata resources on the web (mostly rdf data). the list of resources includes dblp , springer nature scigraph explorer , semantic scholar and the global research identifier database (grid) with metadata about institutes. a preliminary version of this kg has been used for experiments of the previous work [ ] where suitability of embedding models have been tested of such use cases. through this research work we will point to this kg as skgold. towards this objective, a domain conceptualization has been done to define the classes and relations of focus. figure shows the ontology that is used for the creation of these knowledge graphs. in order to define the terms, the openresearch [ ] ontology is reused. each instance in the scholarly knowledge graph is equipped with a unique id to enable the identification and association of the kg elements. the knowledge graphs consist of the following core entities of papers, events, authors, and departments. in the creation of the our kg which will be denoted as skgnew a set of conference series have been selected (namely iswc, eswc, aaai, neurips, cikm, aci, kcap and hcai have been considered in the initial step of retrieving raw metadata from the source). in addition, the metadata flitted for the temporal interval of - . the second version of the same kg has been generated directly from semantic scholar. the datasets, used for model training, which in total comprise , triples where , triples are coming from the skgold and , triples are generated in skgnew. in each set of experiments, both datasets are split into triples of training/validation/test sets. table includes the detailed statistics about the datasets only considering three relationships between entities namely hasauthor (paper -author), hascoauthor (author -author), hasvenue (author/papervenue). due to the low volume of data, isaffiliated (author -organization) relationship is eliminated due in skgnew. in this section we focus on providing required preliminaries for this work as well as the related work. the definitions required to understand our approach are: -knowledge graph. let e, r be the sets of entities and relations respectively. a kg is roughly represented as a set the proposed loss is trained on a classical translation-based embedding models named transe and a model for complex space as rotate. therefore, we mainly provide a description of transe and rotate and further focus on other state-ofthe-art models. transe. it is reported that transe [ ] , as one of the simplest translation based models, outperformed more complicated kges in [ ] . the initial idea of transe model is to enforce embedding of entities and relation in a positive triple (h, r, t) to satisfy the following equality: where h, r and t are embedding vectors of head, relation and tail respectively. transe model defines the following scoring function: rotate. here, we address rotate [ ] which is a model designed to rotate the head to the tail entity by using relation. this model embeds entities and relations in complex space. by inclusion of constraints on the norm of entity vectors, the model would be degenerated to transe. the scoring function of rotate is loss function. margin ranking loss (mrl) is one of the most used loss functions which optimizes the embedding vectors of entities and relations. mrl computes embedding of entities and relations in a way that a positive triple gets lower score value than its corresponding negative triple. the least difference value between the score of positive and negative samples is margin (γ). the mrl is defined as follows: where [x] + = max( , x) and s + and s − are respectively the set of positive and negative samples. mrl has two disadvantages: ) the margin can slide, ) embeddings are adversely affected by false negative samples. more precisely, the issue of margin sliding is described with an example. assume that f r (h , t ) = and f r (h , t ) = γ, or f r (h , t ) = γ and f r (h , t ) = γ are two possible scores for a triple and its negative sample. both of these scores get minimum value for the optimization causing the model to become vulnerable to a undesirable solution. to tackle this problem, limited-based score [ ] revises the mrl by adding a term to limit maximum value of positive score: it shows l rs significantly improves the performance of transe. authors in [ ] denote transe which is trained by l rs as transe-rs. regarding the second disadvantage, mrl enforces a hard margin in the side of negative samples. however, using relations with many-to-many characteristic (e.g., co-author), the rate of false negative samples is high. therefore, using a hard boundary for discrimination adversely affects the performance of a kge model. with a systematic evaluation (performance under reasonable set up) of suitable embedding models to be considered in our evaluations, we have selected two other models that are described here. complex. one of the embedding models focusing on semantic matching model is complex [ ] . in semantic matching models, the plausibility of facts are measured by matching the similarity of their latent representation, in other words it is assumed that similar entities have common characteristics i.e. are connected through similar relationships [ , ] . in complex the entities are embedded in the complex space. the score function of complex is given as follows: in whicht is the conjugate of the vector t. here we present a multi-layer convolutional network model for link prediction named as conve. the score function of the conve is defined as below: in which g denotes a non-linear function,h andr are d reshape of head and relation vectors respectively, ω is a filter and w is a linear transformation matrix. the core idea behind the conve model is to use d convolutions over embeddings to predict links. conve consists of a single convolution layer, a projection layer to the embedding dimension as well as an inner product layer. this section proposes a new model independent optimization framework for training kge models. the framework fixes the second problem of mrl and its extension mentioned in the previous section. the optimization utilizes slack variables to mitigate the negative effect of the generated false negative samples. in contrast to margin ranking loss, our optimization uses soft margin. therefore, uncertain negative samples are allowed to slide inside of margin. figure visualizes the separation of positive and negative samples using margin ranking loss and our optimization problem. it shows that the proposed optimization problem allows false negative samples to slide inside the margin by using slack variables (ξ). in contrast, margin ranking loss doesn't allow false negative samples to slide inside of the margin. therefore, embedding vectors of entities and relations are adversely affected by false negative samples. the mathematical formulation of our optimization problem is as follows: where f r (h, t) is the score function of a kge model (e.g., transe or rotate), s + , s − are positive and negative samples sets. γ ≥ is the upper bound of score of positive samples and γ is the lower bound of negative samples. γ − γ is margin (γ ≥ γ ). ξ r h,t is slack variable for a negative sample that allows it to slide in the margin. ξ r h,t helps the optimization to better handle uncertainty resulted from negative sampling. the term ( ξ r h,t ) represented in the problem is quadratic. therefore, it is convex which results in a unique and optimal solution. moreover, all three constraints can be represented as convex sets. the constrained optimization problem ( ) is convex. as a conclusion, it has a unique optimal solution. the optimal solution can be obtained by using different standard methods e.g. penalty method [ ] . the goal of the problem ( ) is to adjust embedding vectors of entities and relations. a lot of variables participate in optimization. in this condition, using batch learning with stochastic gradient descent (sgd) is preferred. in order to use sgd, constrained optimization problem ( ) should be converted to unconstrained optimization problem. the following unconstrained optimization problem is proposed instead of ( ). the problem ( ) and ( ) may not have the same solution. however, we experimentally see that if λ and λ are properly selected, the results would be improved comparing to margin ranking loss. this section presents the evaluations of transe-sm and rotate-sm (transe and rotate trained by sm loss), over a scholarly knowledge graph. the evaluations are motivated for a link prediction task in the domain of scholarly communication in order to explore the ability of embedding models in support of metaresearch enquiries. in addition, we provide a comparison of our model with other state-ofthe-art embedding models (selected by performance under a reasonable set up) on two standard benchmarks (freebase and wordnet). four different evaluation methods have been performed in order to approve: ) better performance and effect of the proposed loss, ) quality and soundness of the results, ) validity of the discovered co-authorship links and ) sensitivity of the proposed model to the selected hyperparameters. more details about each of these analyses are discussed in the remaining part of this section. the proposed loss is model independent, however, we prove its functionality and effectiveness by applying it on different embedding models. in the first evaluation method, we run experiments and assess performance of transe-sm model as well as rotate-sm in comparison to the other models and the original loss functions. in order to discuss this evaluation further, let (h, r, t) be a triple fact with an assumption that either head or tail entity is missing (e.g., (?, r, t) or (h, r, ?) ). the task is to aim at completing either of these triples (h, r, ?) or (?, r, t) by predicting head (h) or tail (t) entity. mean rank (mr), mean reciprocal rank (mrr) [ ] and hits@ have been extensively used as standard metrics for evaluation of kge models on link prediction. in computation of mean rank, a set of pre-processing steps have been done such as: -head and tail of each test triple are replaced by all entities in the dataset, -scores of the generated triples are computed and sorted, -the average rank of correct test triples is reported as mr. let rank i refers to the rank of the i−th triple in the test set obtained by a kge model. the mrr is obtained as follows: the computation of hits@ is obtained by replacing all entities in the dataset in terms of head and tail of each test triples. the result is a sorted list of triples based on their scores. the average number of triples that are ranked at most is reported as hits@ as represented in table . the results mentioned in the table validate that transe-sm and rotate-sm significantly outperformed other embedding models in all metrics. in addition, evaluation of the state-of-the-art models have been performed over the two benchmark datasets namely fb k and wn . while our focus has been resolving problem of kges in presence of many-to-many relationships, the evaluations of the proposed loss function (sm) on other datasets show the effectiveness of sm in addressing other types of relationships. table shows the results of experiments for transe, complex, conve, rotate, transe-rs, transe-sm and rotate-sm. the proposed model significantly outperforms the other models with an accuracy of . % on fb k. the evaluations on wn shows that rotate-sm outperforms other evaluated models. the optimal settings for our proposed model corresponding to this part of the evaluation are λ = , γ = . , γ = . , α = , d = for fb k and λ = , γ = . , γ = . , α = , d = for wn . with the second evaluation method, we aim at approving quality and soundness of the results. in order to do so, we additionally investigate the quality of the recommendation of our model. a sample set of researchers associated with the linked data and information retrieval communities [ ] are selected as the foundation for the experiments of the predicted recommendations. table shows the number of recommendations and their ranks among the top predictions for all of the selected researchers. these top predictions are filtered for a closer look. the results are validated by checking the research profile of the recommended researchers and the track history of co-authorship. in the profile check, we only kept the triples which are indicating: . close match in research domain interests of scholars by checking profiles, . none-existing scholarly relation (e.g., supervisor, student), . none-existing affiliation in the same organization, . none-existing co-authorship. for example, out of all the recommendations that our approach has provided for researcher with id a , of them have been identified sound and new collaboration target. the rank of each recommended connection is shown in the third column. furthermore, the discovered links for co-authorship recommendations have been examined with a closer look to the online scientific profile of two top machine learning researchers, yoshua bengio , a and yann lecun , a . the recommended triples have been created in two patterns of (a , r, ?) and (?, r, a ) and deduplicated for the same answer. the triples are ranked based on scores obtained from transe-sm and rotate-sm. for evaluations, a list of top recommendations has been selected per considered researcher, bengio and lecun. in order to validate the profile similarity in research and approval of not existing earlier co-authorship, we analyzed the profile of each recommended author to "yoshua bengio" and "yann lecun" as well as their own profiles. we analyzed the scientific profiles of the selected researchers provided by the most used scholarly search engine, google citation . due to author nameambiguity problem, this validation task required human involvement. first, the research areas indicated in the profiles of researchers have been validated to be similar by finding matches. in the next step, some of the highlighted publications with high citations and their recency have been controlled to make sure that the profiles of the selected researchers match in the machine learning community close to the interest of "yoshua bengio" -to make sure the researchers can be considered in the same community. as mentioned before, the knowledge graphs that are used for evaluations consist of metadata from till . in checking the suggested recommendations, a co-authorship relation which has happened before or after this temporal interval is considered valid for the recommendation. therefore, the other highly ranked links with none-existed co-authorship are counted as valid recommendations for collaboration. figure b shows a visualization of such links found by analyzing top recommendations to and from "yoshua bengio" and fig. a shows the same for "yann lecun". out of the discovered triples for "yoshua bengio" being head, of them have been approved to be a valid recommendation (relevant but never happened before) and triples have been showing an already existing co-authorship. profiles of other researchers have not been made available by google citation. among the triples with "yoshua bengio" considered in the tail, of triples have been already discovered by the previous pattern. profile of researchers were not available and researchers have been in contact and co-authorship with "yoshua bengio". finally, new profiles have been added as recommendations. out of triples (y annlecun, r, ?), recommendations have been discovered as new collaboration cases for "yann lecun". in analyzing the triples with a pattern of the fixed tail (?, r, y annlecun), there have been cases either without profiles on google citations or have had an already existing co-authorship. by excluding these examples as well as the already discovered ones from the other triple pattern, new researchers have remained as valid recommendations. in this part we investigate the sensitivity of our model to the hyperparameters (γ , γ , λ ). to analyze sensitivity of the model to the parameters γ , we fix γ to . , and . moreover, λ is also fixed to one. then different values for γ are tested and visualized. regarding the red dotted line in fig. a , the parameter γ is set to . and λ = . it is shown that by changing γ from . to , the performance increases to reach the peak and then decreases by around %. therefore, the model is sensitive to γ . the significant waving of results can be seen when γ = , as well (see fig. a ). therefore, proper selection of γ , γ is important in our model. we also analyze the sensitivity of the performance of our model on the parameter λ . to do so, we take the optimal configuration of our model corresponding to the fixed γ , γ . then the performance of our model is investigated in different setting where the λ ∈ { . , . , , , , }. according to fig. b , the model is less sensitive to the parameter λ . therefore, to obtain hyper parameters of the model, it is recommended that first (γ , γ ) are adjusted by validation when λ is fixed to a value (e.g., ). then the parameter λ is adjusted while (γ , γ ) are fixed. the aim of the present research was to develop a novel loss function for embedding models used on kgs with a lot of many-to-many relationships. our use case is scholarly knowledge graphs with the objective of providing predicted links as recommendations. we train the proposed loss on embedding model and examine it for graph completion of a real-world knowledge graph in the example of scholarly domain. this study has identified a successful application of a model free loss function namely sm. the results show the robustness of our model using sm loss function to deal with uncertainty in negative samples. this reduces the negative effects of false negative samples on the computation of embeddings. we could show that the performance of the embedding model on the knowledge graph completion task for scholarly domain could be significantly improved when applied on a scholarly knowledge graph. the focus has been to discover (possible but never happened) co-author links between researchers indicating a potential for close scientific collaboration. the identified links have been proposed as collaboration recommendations and validated by looking into the profile of a list of selected researchers from the semantic web and machine learning communities. as future work, we plan to apply the model on a broader scholarly knowledge graph and consider other different types of links for recommendations e.g, recommend events for researchers, recommend publications to be read or cited. openaire lod services: scholarly communication data as linked data construction of the literature graph in semantic scholar towards a knowledge graph for science translating embeddings for modeling multi-relational data convex optimization a three-layered mutually reinforced model for personalized citation recommendation convolutional d knowledge graph embeddings a comparative survey of dbpedia, freebase, opencyc, wikidata, and yago science of science semantic scholar metaresearch recommendations using knowledge graph embeddings combining text embedding and knowledge graph embedding techniques for academic search engines a review of relational machine learning for knowledge graphs data curation in the openaire scholarly communication infrastructure factorizing yago: scalable machine learning for linked data rotate: knowledge graph embedding by relational rotation in complex space arnetminer: extraction and mining of academic social networks linked data in libraries: a case study of harvesting and sharing bibliographic metadata with bibframe complex embeddings for simple link prediction openresearch: collaborative management of scholarly communication metadata unveiling scholarly communities over knowledge graphs aminer: search and mining of academic social networks knowledge graph embedding: a survey of approaches and applications acekg: a large-scale knowledge graph for academic data mining big scholarly data: a survey researchgate: an effective altmetric indicator for active researchers? pave: personalized academic venue recommendation exploiting copublication networks learning knowledge embeddings by combining limit-based scoring loss acknowledgement. this work is supported by the epsrc grant ep/m / , the wwtf grant vrg - , the ec horizon grant lambda (ga no. ), the cleopatra project (ga no. ), and the german national funded bmbf project mlwin. key: cord- -ovhzult authors: veen, lourens e.; hoekstra, alfons g. title: easing multiscale model design and coupling with muscle date: - - journal: computational science - iccs doi: . / - - - - _ sha: doc_id: cord_uid: ovhzult multiscale modelling and simulation typically entails coupling multiple simulation codes into a single program. doing this in an ad-hoc fashion tends to result in a tightly coupled, difficult-to-change computer program. this makes it difficult to experiment with different submodels, or to implement advanced techniques such as surrogate modelling. furthermore, building the coupling itself is time-consuming. the multiscale coupling library and environment version (muscle ) aims to alleviate these problems. it allows the coupling to be specified in a simple configuration file, which specifies the components of the simulation and how they should be connected together. at runtime a simulation manager takes care of coordination of submodels, while data is exchanged over the network in a peer-to-peer fashion via the muscle library. submodels need to be linked to this library, but this is minimally invasive and restructuring simulation codes is usually not needed. once operational, the model may be rewired or augmented by changing the configuration, without further changes to the submodels. muscle is developed openly on github, and is available as open source software under the apache . license. natural systems consist of many interacting processes, each taking place at different scales in time and space. such multiscale systems are studied for instance in materials science, astrophysics, biomedicine, and nuclear physics [ , , , ] . multiscale systems may extend across different kinds of physics, and beyond into social systems. for example, electricity production and distribution covers processes at time scales ranging from less than a second to several decades, covering physical properties of the infrastructure, weather, and economic aspects [ ] . the behaviour of such systems, especially where emergent phenomena are present, this work was supported by the netherlands escience center and nwo under the e-musc project. may be understood better through simulation. simulation models of multiscale systems (multiscale models for short), are typically coupled simulations: they consist of several submodels between which information is exchanged. constructing multiscale models is a non-trivial task. in addition to the challenge of constructing and verifying a sufficiently accurate model of each of the individual processes in the system, scale bridging techniques must be used to preserve key invariants while exchanging information between different spatiotemporal scales. if submodels that use different domain representations need to communicate, then conversion methods are required to bridge these gaps as well. multiscale models that exhibit temporal scale separation may require irregular communication patterns, and spatial scale separation results in running multiple instances, possibly varying their number during simulation. once verified, the model must be validated and its uncertainty quantified (uq) [ ] . this entails uncertainty propagation (forward uq) and/or statistical inference of missing parameter values and their uncertainty (inverse uq). sensitivity analysis (sa) may also be employed to study the importance of individual model inputs for obtaining a realistic result. such analysis is often done using ensembles, which is computationally expensive especially if the model on its own already requires significant resources. recently, semi-intrusive methods have been proposed to improve the efficiency of uq of multiscale models [ ] . these methods leave individual submodels unchanged, but require replacing some of them or augmenting the model with additional components, thus changing the connections between the submodels. when creating a multiscale model, time and development effort can often be saved by reusing existing submodel implementations. the coupling between the models however is specific to the multiscale model as a whole, and needs to be developed from scratch. doing this in an ad-hoc fashion tends to result in a tightly coupled, difficult-to-change computer program. experimenting with different model formulations or performing efficient validation and uncertainty quantification then requires changing the submodel implementations, which in turn makes it difficult to ensure continued interoperability between model components. as a result, significant amounts of time are spent solving technical problems rather than investigating the properties of the system under study. these issues can be alleviated through the use of a model coupling framework, a software framework which takes care of some of the aspects of coupling submodels together into a coupled simulation. many coupling frameworks exist, originating from a diversity of fields [ , , ] . most of these focus on tightlycoupled scale-overlapping multiphysics simulations, often in a particular domain, and emphasise efficient execution on high-performance computers. the muscle framework has taken a somewhat different approach, focusing on scale-separated coupled simulation. these types of coupled simulations have specific communication patterns which occupy a space in between tightlycoupled, high communication intensity multiphysics simulations, and pleasingly parallel computations in which there is no communication between components at all. the aforementioned methods for semi-intrusive uq entail a similar com-munication style, but require the ability to handle ensembles of (parts of) the coupled simulation. in this paper, we introduce version of the multiscale coupling library and environment (muscle [ ] ), and explain how it helps multiscale model developers in connecting (existing) submodels together, exchanging information between them, and changing the structure of the multiscale model as required for e.g. uncertainty quantification. we compare and contrast muscle to two representative examples: precice [ ] , an overlapping-scale multiphysics framework, and amuse [ ] , another multiscale-oriented coupling framework. muscle is based on the theory of the multiscale modeling and simulation framework (mmsf, [ , ] ). the mmsf provides a systematic method for deriving the required message exchange pattern from the relative scales of the modelled processes. as an example, we show this process for a d simulation of in-stent restenosis (isr d, [ , , , ] ). this model models stent deployment in a coronary artery, followed by a healing process involving (slow) cell growth and (fast) blood flow through the artery. the biophysical aspects of the model have been described extensively in the literature; here we will focus on the model architecture and communication pattern. note that we have slightly simplified both the model (ignoring data conversion) and the method (unifying state and boundary condition updates) for convenience. simple coupled simulations consist of two or more sequential model runs, where the output of one model is used as an input of the next model. this suffices if one real-world process takes place before the next, or if there is otherwise effectively a one-way information flow between the modeled processes. the pattern of data exchange in such a model may be described as a directed acyclic graph (dag)-based workflow. a more complex case is that of cyclic models, in which two or more submodels influence each other's behaviour as the simulation progresses. using a dag to describe such a simulation is possible, but requires modelling each executing submodel as a long sequence of state update steps, making the dag unwieldy and difficult to analyse. moreover, the number of steps may not be known in advance if a submodel has a variable step size or runs until it detects convergence. a more compact but still modular representation is obtained by considering the coupled simulation to be a collection of simultaneously executing programs (components) which exchange information during execution by sending and receiving messages. designing a coupled simulation then becomes a matter of deciding which component should send which information to which other component at which time. designing this pattern of information exchange between the components is non-trivial. each submodel must receive the information it needs to perform its next computation as soon as possible and in the correct form. moreover, in order to avoid deadlock, message sending and receiving should match up exactly between the submodels. figure depicts the derivation of the communication pattern of isr d according to the mmsf. figure a) shows the spatial and temporal domains in which the three processes comprising the model take place. temporally, the model can be divided into a deployment phase followed by a healing phase. spatially, deployment and cell growth act on the arterial wall, while blood flow acts on the lumen (the open space inside the artery). figure b) shows a scale separation map [ ] for the healing phase of the model. on the temporal axis, it shows that blood flow occurs on a scale of milliseconds to a second, while cell growth is a process of hours to weeks. thus, the temporal scales are separated [ ] . spatially, the scales overlap, with the smallest agents in the cell growth model as well as the blood flow model's grid spacing on the order of µm, while the domains are both on the order of millimeters. according to the mmsf, the required communication pattern for the coupled simulation can be derived from the above information. the mmsf assumes that each submodel executes a submodel execution loop (sel). the sel starts with an initialisation step (f init ), then proceeds to repeatedly observe the state (o i ) and then update the state (s). after a number of repetitions of these two steps, iteration stops and the final state is observed (o f ). during observation steps, (some of the) state of the model may be sent to another simulation component, while during initialisation and state update steps messages may be received. for the isr d model, causality dictates that the deployment phase is simulated before the healing phase, and therefore that the final state of the deployment (o f ) is fed into the initial conditions (f init ) of the healing simulation. in the mmsf, this is known as a dispatch coupling template. within the healing phase, there are two submodels which are timescale separated. this calls for the use of the call (o i to f init ) and release (o f to s) coupling templates. figure c) shows the resulting connections between the submodels using the multiscale modeling language [ ] . in this diagram, the submodels are drawn as boxes, with lines indicating conduits between them through which messages may be transmitted. decorations at the end of the lines indicate the sel steps (or operators) between which the messages are sent. note that conduits are unidirectional. figure d) shows the corresponding timeline of execution. first, deployment is simulated, then the cell growth and blood flow models start. at every timestep of the cell growth submodel (slow dynamics), part of its state is observed (o i ) and used to initialise (f init ) the blood flow submodel (fast dynamics). the blood flow model repeatedly updates its state until it converges, then sends (part of) its final state (at its o f ) back to the cell growth model's next state update (s). while the above demonstrates how to design a multiscale model from individual submodels, it does not explain how to implement one. in this section, we introduce muscle and ymmsl, and show how they ease building a complex coupled simulation. muscle is the third incarnation of the multiscale coupling library and environment, and is thus the successor of muscle [ ] and muscle [ ] . muscle consists of two main components: libmuscle and the muscle manager. figure shows how libmuscle and the muscle manager work together with each other and with the submodels to enact the simulation. at start-up, the muscle manager reads in a description of the model and then waits for the submodels to register. the submodels are linked with libmuscle, which offers an api through which they can interact with the outside world using ports, which are gateways through which messages may be sent and received. to start the simulation, the manager is started first, passing the configuration, and then the submodels are all started and passed the location of the manager. at submodel start-up, libmuscle connects to the muscle manager via tcp, describes how it can be contacted by other components, and then receives a description for each of its ports of which other component it should communicate with and where it may be found. the muscle manager derives this information from the model topology description and from the registration information sent by the other submodels. the submodels then set up direct peer-to-peer network connections to exchange messages. these connections currently use tcp, but a negotiation mechanism allows for future addition of faster transports without changes to user code. submodels may use mpi for internal communication independently of their use of muscle for external communication. in this case, muscle uses a spinloop-free receive barrier to allow resource sharing between submodels that do not run concurrently. in muscle , each non-java model instance is accompanied by a java minder process which handles communication, an additional complexity that has been removed in muscle in favour of a native libmuscle implementation with language bindings. the manager is in charge of setting up the connections between the submodels. the model is described to the manager using ymmsl, a yaml-based serialisation of the multiscale modelling and simulation language (mmsl). mmsl is a somewhat simplified successor to the mml [ ] , still based on the same concepts from the mmsf. listing shows an example ymmsl file for isr d. the model is described with its name, the compute elements making up the model, and the conduits between them. the name of each compute element is given, as well as a second identifier which identifies the implementation to use to instantiate this compute element. conduits are listed in the form component .port : component .port , which means that any messages sent by component on its port are to be sent to component on its port . the components referred to in the conduits section must be listed in the compute elements section. muscle reads this file directly, unlike muscle which was configured using a ruby script that could be derived from the mml xml file. the ymmsl file also contains settings for the simulation. these can be global settings, like length of the simulated artery section, or addressed to a specific submodel, e.g. bf.velocity. submodel-specific settings override global settings if both are given. settings may be of types float, integer, boolean, string, and d or d array of float. the submodels need to coordinate with the manager, and communicate with each other. they do this using libmuscle, which is a library currently available in python (via pip), c++ and fortran. unlike in muscle , whose java api differed significantly from the native one, the same features are available in all supported languages. listing shows an example in python. first, an instance object is created and given a description of the ports that this submodel will use. at this point, libmuscle will connect to the manager to register itself, using an instance name and contact information passed on the command line. next, the reuse loop is entered. if a submodel is used as a micromodel, then it will need to run many times over the course of the simulation. the required number of runs equals the macromodel's number of timesteps, which the micromodel should not have any knowledge of if modularity is to be preserved. a shared setting could solve that, but will not work if the macromodel has varying timesteps or runs until it detects convergence. determining whether to do another run is therefore taken care of by muscle , and the submodel simply calls its reuse instance() function to determine if another run is needed. in most cases, muscle relied on a global end time to shut down the simulation, which is less flexible and potentially error-prone. within the reuse loop is the implementation of the submodel execution loop. first, the model is initialised (lines - ). settings are requested from libmuscle, passing an (optional) type description so that libmuscle can generate an appropriate error message if the submodel is configured incorrectly. note that re recovered time is specified without the prefix; libmuscle will automatically resolve the setting name to either a submodel-specific or a global setting. a message containing the initial state is received on the relevant port (line ), and the submodel's simulation time is initialised using the corresponding timestamp. the obtained data is then used to initialise the simulation state in a model-specific way, as represented here by an abstract init model() function (line ). next is the iteration part of the sel, in which the state is repeatedly observed and updated (lines [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] . in addition to the simulation time corresponding to the current state, the timestamp for the next state is calculated here (line ). this is unused here, but is required in case two submodels with overlapping timescales are to be coupled [ ] and so improves reusability of the model. in isr d's o i operator, the current geometry of the artery is calculated and sent on the geom out port (lines [ ] [ ] . next, the wall shear stress is received and used in the model's state update, after which the simulation time is incremented and the next observation may occur (lines - ). once the final state is reached, it is sent on the corresponding port (line ). in this example, this port is not connected, which causes muscle to simply ignore the send operation. in practice, a component would be attached which saves this final state to disk, or postprocesses it in some way, possibly via an in-situ/in-transit analysis framework. message data may consist of floating point numbers, integers, booleans, strings, raw byte arrays, or lists or dictionaries containing these, as well as grids of floating point or integer numbers or booleans, where muscle only supported d arrays of numbers. internally, muscle uses messagepack for encoding the data before it is sent. uncertainty quantification of simulation models is an important part of their evaluation. intrusive methods provide an efficient solution in some cases, but uq is most often done using monte carlo (mc) ensembles. an important innovation in muscle compared to muscle is its flexible support for monte carlobased algorithms. this takes the form of two orthogonal features: instance sets and settings injection. the simulation has been augmented with a sampler and a load balancer, and there are now multiple instances of each of the three submodels. the sampler samples the uncertain parameters from their respective distributions, and generates a settings object for each ensemble member. these objects are sent to the load balancer, which distributes them evenly among the available model instances. the settings are then sent into a special port on the submodel instances named muscle settings in, from where the receiving libmuscle automatically overlays them on top of the centrally provided settings. the settings are then transparently passed on to the corresponding other submodel instances. final results are passed back via the load balancer to the sampler, which can then compute the required statistics. to enable communication with sets of instances, muscle offers vector ports, recognisable by the square brackets in the name. a vector port allows sending or receiving on any of a number of slots, which correspond to instances if the port is connected to a set of them. vector ports may also be connected to each other, in which case each sending slot corresponds to a receiving slot. in the example, the sampler component resizes its parameters[] port to the number of samples it intends to generate, then generates the settings objects and sends one on each slot. the load balancer receives each object on the corresponding slot of its front in port, and passes it on to a slot on its back out port. it is then received by the corresponding ensemble member, which runs and produces a result for the load balancer to receive on back in. the load balancer then reverses the earlier slot mapping, and passes the result back to the sampler on the same slot on its front out that the sampler sent the corresponding settings object on. with the exception of the mapping inside the load balancer, all addressing in this use case is done transparently by muscle , and components are not aware of the rest of the simulation. in particular, the submodels are not aware of the fact that they are part of an ensemble, and can be completely unmodified. an important advantage of the use of a coupling framework is the increase in modularity of the model. in muscle , submodels do not know of each other's existence, instead communicating through abstract ports. this gives a large amount of flexibility in how many submodels and submodel instances there are and how they are connected, as demonstrated by the uq example. modularity can be further improved by inserting helper components into the simulation. for instance, the full isr d model has two mappers, components which convert from the agent-based representation of the cell model to the lattice-based representation of the blood flow model and back. these are implemented in the same way as submodels, but being simple functions only implement the f init and o f parts of the sel. the use of mappers allows submodels to interact with the outside world on their own terms from a semantic perspective as well as with respect to connectivity. separate scale bridging components may be used in the same way, except converting between scales rather than between domain representations. other coupling libraries and frameworks exist. while a full review is beyond the scope of this paper (see e.g. [ ] ), we provide a brief comparison here with two other such frameworks in order to show how muscle relates to other solutions. precice is a framework for coupled multiphysics simulations [ ] . it comes with adapters for a variety of cfd and finite element solvers, as well as scale bridging algorithms and coupling schemes. data is exchanged between submodels in the form of variables defined on meshes, which can be written to by one component and read by another. connections are described in an xml-based configuration file. like muscle , precice links submodels to the framework by adding calls to a library to them. for more generic packages, a more extensive adapter is created to enable more configurability. submodels are started separately, and discover each other via files in a known directory on a shared file system, after which peer-to-peer connections are set up. precice differs from muscle in that it is intended primarily for scaleoverlapping, tightly-coupled physics simulations. muscle can do this as well, but is mainly designed for loosely-coupled multiscale models of any kind. for instance, it is not clear how an agent-based cell simulation as used in isr d would fit in the precice data model. muscle 's central management of model settings and its support for sets of instances allows it to run ensembles, thus providing support for uncertainty quantification. precice does not seem to have any features in this direction. the astrophysical multipurpose software environment (amuse) is a framework for coupled multiscale astrophysics simulations [ , ] . it comprises a library of well-known astrophysics models wrapped in python modules, facilities for unit handling and data conversion, and infrastructure for spawning these models and communicating with them at runtime. data exchange between amuse submodels is in the form of either grids or particle collections, both of which store objects with arbitrary attributes. with respect to linking submodels, amuse takes the opposite approach to muscle and precice. instead of linking the model code to a library, the model code is made into a library, and where muscle and precice have a built-in configurable coupling paradigm, in amuse coupling is done by an arbitrary user-written python script which calls the model code. this script also starts the submodels, and performs communication by reading and writing to variables in the models. linking a submodel to amuse is more complex than doing this in muscle , because an api needs to be implemented that can access many parts of the model. this api enables access to the model's parameters as well as to its state. amuse comes with many existing astrophysics codes however, which will likely suffice for most users. coupling via a python script gives the user more flexibility, but also places the responsibility for implementing the coupling completely on the user. uncertainty quantification could be implemented, although scalability to large ensembles may be affected by the lack of peer-to-peer communication. muscle , as the latest version of muscle, builds on almost fourteen years of work on the multiscale modelling and simulation framework and the mus-cle paradigm. it is mainly designed for building loosely coupled multiscale simulations, rather than scale-overlapping multi-physics simulations. models are described by a ymmsl configuration file, which can be quickly modified to change the model structure. linking existing codes to the framework can be done quickly and easily due to its library-based design. other frameworks have more existing integrations however. which framework is best will thus depend on which kind of problem the user is trying to solve. muscle is open source software available under the apache . license, and it is being developed openly on github [ ] . compared to muscle , the code base is entirely new and while enough functionality exists for it to be useful, more work remains to be done. we are currently working on getting the first models ported to muscle , and we plan to further extend support for uncertainty quantification, implementing model components to support the recently-proposed semi-intrusive uq algorithms [ ] . we will also finish implementing semi-intrusive benchmarking of models, which will enable performance measurement and support performance improvements as well as enabling future static scheduling of complex simulations. other future features could include dynamic instantiation and more efficient load balancing of submodels in order to support the heterogeneous multiscale computing paradigm [ ] . patterns for high performance multiscale computing multiphysics and multiscale software frameworks: an annotated bibliography distributed multiscale computing with muscle , the multiscale coupling library and environment foundations of distributed multiscale computing: formalization, specification, and analysis precice a fully parallel library for multi-physics surface coupling a complex automata approach for in-stent restenosis: twodimensional multiscale modelling and simulations towards a complex automata multiscale model of in-stent restenosis a framework for multi-scale modelling principles of multiscale modeling mml: towards a multiscale modeling language physics-based multiscale coupling for full core nuclear reactor simulation mastering the scales: a survey on the benefits of multiscale computing software survey of multiscale and multiphysics applications and communities an agent-based coupling platform for complex automata towards a complex automata framework for multi-scale modeling: formalism and the scale separation map multiscale modeling in nanomaterials science semi-intrusive multiscale metamodelling uncertainty quantification with application to a model of in-stent restenosis semi-intrusive uncertainty propagation for multiscale models uncertainty quantification of a multiscale model for in-stent restenosis the astrophysical multipurpose software environment a review of modelling tools for energy and electricity systems with large shares of variable renewables a comprehensive framework for verification, validation, and uncertainty quantification in scientific computing multiscale computational models of complex biological systems multi-physics simulations using a hierarchical interchangeable software interface key: cord- -y vsc r authors: peiffer, robert l.; armstrong, joseph r.; johnson, philip t. title: animals in ophthalmic research: concepts and methodologies date: - - journal: methods of animal experimentation doi: . /b - - - - . - sha: doc_id: cord_uid: y vsc r nan the component tissues of the eye are both individually and collectively unique and fascinating to both clinician and researcher. we find in a single organ a variety of complex tissues each with its own particular vascular relationships, from the avascular cornea and lens to the extensively vascularized uvea. charac teristic biochemical features of the tissues exist as well, including the endothelial pump of the cornea, the anaerobic glycolytic pathway of the lens, and the poorly defined transport systems of aqueous humor production and outflow. besides the dynamic aqueous, we find the most substantial tissue of the eye, the vitreous humor, to be apparently a relatively stagnant blob of hydrated connective tissue whose physiology is still largely a mystery. the neurologic control mechanisms of such diverse processes as pupillary size, accommodation, and probably aque ous humor inflow and outflow add another dimension of potential inquiry. the eye is devoid of lymphatics, and the nature of its tissues further contributes to its immunologie uniqueness. the retina incorporates all the complexities of neuroperception and transmission. nowhere else in the body are such diverse structure and function so intimately and intricately related. an appreciation of these perspectives, the importance of vision as a cognitive sense, and consideration of the plethora of infectious, inflammatory, degenera tive, traumatic, toxic, nutritional, and neoplastic diseases that can affect any one or all of the component tissues of the eye, will elucidate the challenge the authors face in attempting to review pertinent conceptual and logistical approaches to eye research in the animal laboratory. while the majority of investigations have had as their objective ultimate correlation with normal and abnormal function and structure of the human eye, laboratory studies have provided an abundance of comparative information that emphasizes that while there are numerous and amazing similarities in the peripheral visual system among the vertebrate (and even the invertebrate) animals, significant differences exist that are important to both researcher and clinician in selection of a research model and in extrapolation of data obtained from one species to another, and even among different species subdivisions. details of comparative anatomy and physiology have been reviewed by several authors (prince, (prince, , a polyak, ; duke-elder, ; prince et al., ; walls, ) . these are classic works and invaluable references. one is reminded that "research" means "to look for again," and while some concepts presented in these earlier texts are indeed outdated, and some facts since disproven, phylogenetic perspectives acquired from familiarity with this literature arms the scientist with valuable information. throughout this chapter, we will dwell on species anatomic and physiologic differences when they have been specifically defined or where they are of value in regard to specific laboratory methodologies. while the subhuman primate is frequently described as the ideal laboratory model for eye research, many of these animals are threatened or endangered, and logistical aspects of procurement, maintenance, and restraint should en courage investigators to explore and define the validity of nonprimate animals. when deemed essential, primates should be utilized with maximum thrift and humane care in mind. the eye is a sensitive organ, and in all species adequate anesthesia for man ipulative procedures as well as postoperative analgesia should be considered in in vivo studies; the rabbit, for instance, is extremely sensitive to barbiturate anesthesia and prolonged procedures that require intraocular manipulations pose a challenging problem in terms of subject mortality. the in vitro culture of ocular tissue cell lines offers increasing potential in research as an alternative to animal models. the lens, retinal pigment epithelium, and corneal epithelium and fibrocytes may be readily maintained and manipu lated. current work with the corneal and trabecular endothelium is perhaps the most exciting advance in ophthalmic research in recent years. in vitro culture techniques offer a number of logistical advantages over animal models. a knowledge of spontaneous ocular diseases is essential to avoid interpreting pathology as experimental that may well be unrelated to the investigation. thorough preoperative ophthalmic examination is a prerequisite of any study, and the laboratory scientist will find it to his benefit to be associated with an interested human or veterinary clinical ophthalmologist to assist in defining and managing spontaneous ocular disease. for instance, electrophysiologic studies of the canine retina will be invalid if the subject has chorioretinal lesions of canine distemper, a common entity in laboratory dogs. examination beyond the target organ is also recommended to detect abnormalities that may directly or indirectly influence results. frequently encountered spontaneous infectious diseases are reviewed in a subsequent section. we have limited our discussions primarily to research methodologies involv ing the globe itself; a paucity of information is available regarding the adnexa, orbit, and extraocular muscles, and described techniques are refinements of general approaches rather than those specifically developed for ocular research. in the same perspective, topics of pure physiology or biochemistry are limited, and we did not embark into the unfamiliar and complex area of central visual mechanisms. our approach, we hope, if not all inclusive, has been comprehen sive, with adequate references to specific detail and additional information for the interested reader. spontaneous infectious diseases that involve the eyes of laboratory animals are significantly prevalent to warrent brief discussion. it is important to consider these entities, especially before using animals derived from poorly defined sources. propensity of a species toward the development of spontaneous infecti-ous diseases that might not only affect general health but also induce ocular pathology in both experimental and control subjects may affect selection of a model of procurement and isolation procedures utilized. preinvestigation ocular examinations are necessary to detect preexisting pathology. of the three large species used in ophthalmic research (cat, dog, and primate) , the primate appears to have a minimal predisposition for spontaneous infectious ocular disease. this may be a result, however, of a lack of ophthalmic examina tion in disease situations. it is also important to note that the hamster and guinea pig rarely develop eye infections. table i summarizes those infectious diseases that more commonly involve the eyes of laboratory species. the use of laboratory animals in the investigation of infectious ocular disease has included rats, hamsters, guinea pigs, rabbits, cats, dogs, and subhuman primates. the disease agents studied have ranged from viruses to protozoa, with the rabbit serving as the most popular model. these studies have been of value in defining pathogenesis and the effects of therapeutic agents. those disease agents known to cause conjunctivitis in man do not necessarily produce a similar disease in laboratory species. a particular species may require immunosuppression to allow the induction of a specific infection (payne et al., ; forster and rebell, ) . in other instances, a model may be specifically susceptible to an organism; for example, the chlamydial inclusion body con junctivitis agent in guinea pigs. an additional and somewhat unusual use of the guinea pig involves the sereny test (formal et al., ) . because their conjunctivae are so susceptible to shigella organisms, this animal, through the use of a drip apparatus, has been used to identify the pathogenic strains (mackel et al., ) . listeria monocytogenes produces a severe keratoconjunctivitis in the guinea pig, but in the rabbit and monkey the disease is much milder and more difficult to reproduce (morris and julianelle, ) . this concept has lent itself to studies concerning epithelial cell phagocytosis in the guinea pig (zimianski et al., ) . organism installation techniques involve either direct swabbing or syringe inoculation into the conjunctival sac. the owl monkey (actus trivirgatus) has proved to be the model of choice for the study of chlamydia trachomatis conjunctivitis; other primate species includ ing orangutans have proved to be only mildly susceptible (fraser, ) . a number of models are available for investigators interested in studying keratitis. the rabbit has proved to be the most feasible subject, especially in herpes simplex studies (kaufman and maloney, ) . many bacterial and fun gal diseases afflicting man have also been examined in this model, including the effects of antibiotics, steroids, interferon, vitamins, and other agents nesburn and ziniti, ; smolin et al., ; pollikoff et al., ) . herpesvirus studies can be significantly influenced by the particular strain of virus utilized. for example, some are more effective in producing stromal keratidities as compared to the epithelial condition (metcalf et al., ) . con comitant in importance may be the route of inoculation. in some studies viral suspensions are placed into the conjunctival sac (nesburn and ziniti, ) , while in others they are dropped onto the cornea . corneal preparation may be important; for example, some studies have employed chalazion currette (stern and stock, ) , while others have employed spatula scrap ing or circular trephine defects (kaufman and maloney, ) . deep inoculations are required to produce a stromal keratitis (fig. ) , while the epithelial disease may be induced through shallow scratches (polli koff et al., ) . the rabbit appears to be the species of choice for bacterial keratitis, although davis et al. ( ) feel that the guinea pig is of equal value. in one study an inbred animal was used (strain ) to reduce the inherent variability seen in some of these experiments (davis and chandler, ) . whether or not the rabbit model has any specific biologic advantage, the guinea pig would appear to be more feasible because of its smaller size, ease of handling, ease to anesthetize, cost of purchase and maintenance, and, perhaps most significantly, resistance to spontaneous pasteur ella conjunctivitis, which frequently occurs in laboratory rabbits. methods of bacterial inoculation usually involve an intracorneal injec tion using a microsyringe and small volumes of quantitated microorganisms. fungal keratidities have been studied in rats, rabbits, and owl monkeys. rats were initially popular (burda and fisher, ) , but rabbits and monkeys have become the select species. fusarium solarti keratitis was first studied in the subhuman primate, but, due to expense, forster and rebell ( ) utilized the pigmented rabbit. this model, having received germinating conidia (as opposed to spores) interlamellarly, developed a sustained progressive infection, but only after pretreatment with subconjunctival steroids. candida albicans keratitis has also been induced in the corticosteroid pretreated rabbit . inoculation techniques for these organisms are identical to those described for bacterial and viral keratidities. ishibashi ( ) suggests that backward flow from the inoculation wound and volume accuracy can be controlled by using a microsyringe and a -gauge flat-faced needle. recently, human corneal transplants have been incriminated in the transmis sion of rabies to recipients. a model has been developed in the guinea pig utilizing the slow virus agent of creutzfeldt-jacob disease (manuelidis et al., ) . using a method developed by greene ( ) , this study involved the heterologous transplantation of infected guinea pig cornea sections into the an terior chamber of six recipients with resultant development of spongiform encephalopathy. studies designed to produce anterior uveitis usually employ inoculation into the anterior chamber. this procedure is best performed with a -to -gauge needle directed obliquely through the limbus into the chamber (smith and singer, ) . ocular histoplasmosis was effectively studied in the rabbit using this method, the result being a well-defined granulomatous uveitis. the same proce dure may be applied to produce bacterial or viral anterior uveitis (kaufman et al., ) . it is advisable, however, to withdraw initially an equivalent amount of aqueous fluid to avoid intraocular pressure elevation (zaitseva et al., ) . aguirre et al. ( ) administered canine adenovirus type i intravenously to young beagles to produce the anterior uveitis characteristic of infectious canine hepatitis. carmichael et al. ( ) demonstrated that this was a complementmediated antigen-antibody complex disease. the induction of endophthalmitis may utilize direct transcorneal chamber in oculation as described above; this method was used in rabbits to produce a bacterial infection in order to assess diagnostic anterior chamber paracentesis technique (tucker and forster, ) . the same procedure was also used in the rat to create a fulminating fusarium solarti endophthalmitis (o'day et al., ) . the intravitreal route of inoculation is utilized to induce posterior segment infection. this technique requires toothed forcep stabilization of the eye and needle introduction through the pars plana (may et al., ) . indirect ophthalmoscopy may be employed to ensure accurate deposition. typically a tuberculin syringe with a /s-inch, -gauge needle is used with the subject under general anesthesia or deep tranquilization. the needle is inserted at o'clock and directed toward the posterior pole. small volumes may be injected without signif icant alteration in the intraocular pressure. avallone et al. ( ) used a similar technique in rabbits to demonstrate the effectiveness of the limulus lysate test for rapid detection of e. coli endophthalmitis. after producing a staphylococcol endophthalmitis by this method, michelson and nozik ( ) tested the perfor mance of a subcutaneously implanted minipump for antibiotic administration in the rabbit. ocular toxocariasis, a human disease caused by larval migration of toxocara canis, was produced in mice by olson et al. ( ) . freshly incubated ova were administered per os by stomach tube. anterior chamber hemorrhage and actual sightings of larval migration were reported. hobbs et al. ( ) found ocular mycobacterium leprae and m. lepraemurium in the nine-banded armadillo and mouse, respectively, following intravenous, intraperitoneal, or foot-pad inoculation of the organisms; the uveal tract was primarily involved, although lesions were observed in other ocular tissues. lesions were more dramatic in mice immunologically depressed by thymectomy and total body irradiation. dur to its susceptibility to toxoplasma gondii, the rabbit has been developed as a model for ocular toxoplasmosis. nozik and o'connor ( ) used the california pigmented rabbit and a variation of a method described by vogel ( ) to study the associated retinochoroiditis. this technique consists of prop osing the eye of an anesthetized rabbit and passing the needle retrobulbarly through the sclera to the suprachoroidal space (fig. ) . the authors point out that while the lesions are less active and heal spontaneously in weeks, features are of comparative value. for detailed information concerning other uses of this model the reader is referred to tabbara's organism recoverability experiments (tabbara et al., ) . histoplasmosis has been studied in mice, rats, rabbits, guinea pigs, dogs, pigeons, chickens, and primates. smith et al. ( ) point out that because it is essentially a macular disease in man, the ideal model should have similar anatomy. basically two orders of animals qualify: the avian species due to their dual fovea and the primates. the above authors used the stumptailed macaque, m. arctoides, to produce a focal choroiditis by inoculating yeast-phase histoplasma capsulatum into the internal carotid artery. they emphasized the desirabil ity to create a model that, in addition to focal choroiditis, also exhibits minimal involvement of the retina, no anterior segment disease, and a late macular lesion. experimental ocular cryptococcosis has been induced in rabbits, primates and cats, the latter species appearing to be the ideal choice because of its susceptibil ity and related high incidence of chorioretinal lesions. blouin and cello ( ) used an intracarotid injection technique which reproduced lesions identical to those seen in the naturally occurring disease. a septic choroiditis of bacterial origin was produced by meyers et al. ( ) by intracarotid injection of staphlococcal and streptococcal organisms. this in turn produced a serous retinal detachment, which was the primary objective of the study. animal studies have provided an abundance of information dealing with the pharmacodynamics of the eye and drugs used to treat ocular disease. because of the accessibility of the ocular structures, there are a number of routes of drug administration which are not available in treating other organ systems. on the other hand, there are barriers to drug penetration in the eye which must be considered, including a hydrophobic corneal epithelial layer over the external ocular structures and an internal blood-ocular barrier similar to the blood-brain barrier. the endothelial cells of the iris and retinal vessels, the nonpigmented epithelium of the ciliary body and the retinal pigment epithelium all contribute to this blood-ocular barrier (cunha-vaz, ) . various agents can compromise these barriers to drug penetration. inflamma tion can also affect any of the ocular barriers. in some experimental designs it may be desirable to create inflammation and/or minimize the barriers. tech niques include mechanical removal of the corneal epithelium with a scalpel blade, inducing infectious keratitis or treating the external eye with enzymes (barza et al., ) . a more generalized inflammation can be created by sen sitizing the animal to a specific antigen and then challenging the eye with that antigen (levine and aronson, ) . certain pharmacologie agents affect spe cific barriers; anticholinesterase inhibitors, for instance, can break down the blood aqueous barrier (von sallmann and dillon, ) , while agents with epithelial toxicity, such as benzalkonium chloride, are added to many topical preparations to increase the permeability of the corneal and conjunctival epithelium. all these factors regarding method of administration and barrier permeability must be considered in an experimental design. once again the rabbit has been most frequently chosen for pharmacologie experiments on the eye. although at first glance the anatomy of the rabbit eye would seem close,enough to the human to make this a valid experimental model, at least for drug penetration studies, significant differences do exist, as will be pointed out. drugs can usually be applied topically to the eye in concentrations much higher than could be tolerated systemically. topical application has many limita tions, however. if a topical drug is to be effective intraocularly it must traverse the hydrophobic epithelial layer and the hydrophilic corneal stroma; ideally, then, it should exist in an equilibrium between ionized and unionized forms with biphasic solubility characteristics. thus, both chemical configuration and ph of the vehicle are important. other important limitations are the small tear volume and rapid clearance of drugs from the tear film (janes and stiles, ) . rabbits tend to blink infrequently; consequently corneal contact time is prolonged com pared to humans. most nonprimate mammals possess a nictitating membrane which may affect tear dynamics. drugs may be applied with an ointment vehicle to prolong contact time, but bioavailability may be affected. finally, the position of the animal may be important. a tenfold difference in aqueous drug levels in rabbits has been reported depending on whether the animal was upright or re cumbent during the experiment (sieg and robinson, ) . injection of a drug under the conjunctiva or tenon's capsule theoretically allows delivery of greater amounts of drug, prolonged contact time, and bypass of the external epithelial barrier. subconjunctival injection may be given in two ways; by ( ) injecting directly through the bulbar conjunctiva or ( ) by inserting the needle through the skin of the lids, leaving the conjunctiva intact. the method of injection may be very important; one study showed that in rabbits leakage of hydrocortisone back through the injection site into the tear film ac counted for most of the intraocular absorption (wine et al., ) . although numerous articles have been written on the kinetics of drug absorption after subconjunctival injection (baum, ) , the subject is still controversial, and recent evidence presented by maurice and ota ( ) indicates that absorption into the anterior chamber by this route may be much lower in rabbits than in man. if so, the rabbit may be an inappropriate model for this type of research. the potential for serious complications makes intraocular injection of drugs a 'last resort" option in clinical practice. in many cases the maximum amount of drug tolerated within the eye is small (leopold, ) . good results have been reported, however, in the treatment of experimental bacterial endophthalmitis in rabbits with intravitreal injection of antibiotics (von sallmann et al., ; peyman, ) . usually . - . ml of antibiotic solution is injected into the nucleus of the vitreous through the pars plana. intramuscular and intravenous are the preferred routes for systemic administra tion of drugs in most laboratory animals. drugs may be given orally in food, water, or by stomach tube, or more reliably and precisely with pills or capsules in the case of cats and dogs. whatever the systemic route, intraocular levels are limited by the body's tolerance for the drug and by the blood-ocular barriers. one method of obtaining high ocular levels is arterial infusion, usually of the ipsilateral carotid artery. the highest levels have been obtained experimentally by retrograde perfusion of the intraorbial artery in dogs; details of the technique are discussed by o'rourke et al. ( ) . in general the assessment of penetration into the various intraocular compart ments is carried out using radio-labeled preparations of the drugs under study. the labeled drug is administered, and after a predetermined length of time samples of aqueous are removed by paracentesis or the whole eye is removed and standard size tissue samples taken for scintillation counting. o' brien and edelhouser ( ) utilized a direct chamber perfusion technique to study penetra tion of radio-labeled antiviral drugs through the excised cornea. in the case of antimicrobials, drug levels in ocular tissues may also be determined using a microbioassay such as the agar diffusion technique described by simon and yin ( ) . other methods have been used less commonly. autoradiography has been used to assess intraocular penetration (mccartney et al., ) , and distribution in the orbital compartments has been studied using diatrizoate, a radio-opaque contrast media (levine and aronson, ) . techniques used to assess drug efficacy depend primarily upon the type of drug under study. most drugs used in the treatment of glaucoma, for instance, work through complicated autonomie mechanisms which alter the inflow and outflow of aqueous humor. tonography, tonometry, and fluorophotometry are used to study these effects and are discussed elsewhere. methods used to study antibiotic efficacy, on the other hand, are usually less objective. results are often based on semiquantitative or subjective impressions of clinical response. actual bacterial counts on samples of tissue following treatment of experimental infec tions are feasible, but the organism and size of original inoculum must be strictly defined (kupferman and leibowitz, ) . steroid efficacy is even more difficult to assess objectively in experimental models. leibowitz et al. ( ) have reported a method of quantifying steroid response based on the number of polymorphonuclear leukocytes labeled with tritiated thymidine remaining in the tissue following treatment of inflammatory keratitis. successful organ and cell culture has been reported with a variety of ocular tissues, including whole eyes. the culture of lens, cornea, and corneal endothelium have become routine techniques in many ophthalmic laboratories. paul ( ) provides a good manual of basic tissue culture techniques. the early work in culture of ocular tissue is the subject of an excellent review by lucas ( ) . a number of significant advances have been made since that time, especially in the areas of retinal pigment epithelial and cornea culture. suspended between the aqueous and vitreous humors with no direct vascular supply, the lens is maintained in a type of natural organ culture. it has always been felt, therefore, that the lens should be ideal for in vitro culture, and lens culture experiments have been utilized extensively to study the metabolism of this tissue. a great deal of research into ideal media and culture conditions has been necessary, however. rabbit, mouse, and rat lenses have been most widely used, but culture of bovine lenses is also feasible despite their large size (owens and duncan, ) . the lens may be cultured in an open (continuous perfusion) or closed system. the advantages and disadvantages of each type of system are discussed by schwartz ( a) . closed systems patterned after that described by merriam and kinsey ( ) are by far the most popular, but for certain types of experiments more elaborate perfusion systems are necessary (schwartz, b; sippel, ) . the composition of the media for lens culture has been found to be very important. kinsey et al. ( ) have provided systematic studies on the optimum concentration of various culture media constituents. presently most lens culture is being done in a modified tci media which has been used by kinoshita and others with good success (von sallmann and grimes, ) . thoft and kinoshita ( ) have found that a calcium concentration somewhat higher than that in the aqueous is desirable in the culture media. more recently, chylack and kinoshita ( ) have shown that removal of the lens with part of the vitreous still attached to the posterior surface improves certain parameters of lens function in vitro. presumably, leaving vitreous attached to the lens prevents damage to the pos terior capsule. in addition to culture of whole lenses, cell cultures of pure lens epithelium have been important in the study of lens metabolism. most of the early work was done with chick epithelium, but more recently cultures of mouse lens epithelium have been shown to retain greater cellular differentation in culture (mann, ; russell et al., ) . three different cell types make up the layers of the cornea: epithelial cells, fibrocytic stromal cells, and a single layer of endothelial (mesothelial) cells. all three corneal cell types, as well as whole corneas, can be maintained in culture. the culture of whole corneas has been investigated as a means of corneal preser vation prior to transplantation. most of these studies have used human rather than animal tissues. endothelial cell culture has become an important technique in the study of this very metabolic ly active cell layer. stocker et al. ( ) described a technique for separating a corneal button by carefully peeling off descemet's membrane with endothelium from the posterior surface and peeling the epithelium from the opposite side to yield three relatively pure cell types which can be explanted onto separate cultures (fig. ). colosi and yanoff ( ) have used trypsin to obtain endothelial cells or epithelial cells from whole corneas. perlman and baum ( ) have used stacker's method to isolate endothelial cells from rabbits and have had good success in maintaining large endothelial cell cultures for several months using a modified eagle's mem supplemented with calf serum, bicarbo nate, glutamine, and kanamycin sulfate. most exciting are the reports of using tissue cultured endothelial cells, seeded on to donor corneas with endothelium removed, for transplantation in rabbits (jumblatt et al., ) . although the corneal stromal cells are the most thoroughly studied ocular fibrocytic cells, other fibrocytes have been successfully cultured, including goniocytes from the anterior chamber angle (francois, ) and hyalocytes from the vitreous (françois et al., ) . tissue culture of immature neural retina has been a popular subject. early experiments are summarized by lucas ( ) . recently, differentation of em bryonic neural retinal cells from the chick has been described both in cell aggre gate cultures (sheffield and moscona, ) and monolayer cultures (combes et al., ) . organ culture of mature neural retina has proved more difficult. mature retina tends to deteriorate rapidly in a convential warburg apparatus (lucas, ) . the organ culture technique of trowell ( ) has been used to maintain mature and nearly mature retina for several days (lucas, ) . this technique requires mechanical support of the retina on a metal grid with receptor cells uppermost, exposed to the gas phase; in this case air as % oxygen was found to be toxic. ames and hastings ( ) described a technique for rapid removal of the rabbit retina, together with a stump of optic nerve, for use in short-term culture experi ments including in vitro studies of retinal response to light (ames and gurian, ) . until recently only a few studies were available on the behavior of retinal pigment epithelium (rpe) cells in culture, but recent work on the multipotential nature of the rpe cell has aroused new interest in this area. as with neural retina, chick embryos have been most often used for rpe cultures. hayashi et al. ( ) used edta and trypsin to isolate chick rpe cells for culture on eagles mem supplemented with % fetal calf serum. a similar technique has been used to culture rpe cells from syrian hamsters (albert et al., a) . mandel-corn et al. ( ) have reported an ingenious experiment in which rpe cells from one eye of an owl monkey were transplanted into the vitreous of the fellow eye through the pars plana where the proliferation and metaplasia of those cells could be studied. the experiment described above is not the first time that investigators have attempted to take advantage of the eye's transparent media and large avascular spaces to create a type of in vivo tissue culture. many successes have been reported in transplanting homologous and autologous tissues into the anterior chamber. markee ( ) transplanted endometrium into the anterior chamber of guinea pigs, rabbits, and monkeys where it continued to undergo cyclic changes. goodman ( ) also reported ovulatory cycles of ovaries transplanted into the anterior chamber of rats. woodruff and woodruff ( ) successfully trans planted homologous thyroid tissue into the anterior chamber of guinea pigs. eifrig and prendergast ( ) found that autologous transplants of lymph node tissue in rabbits were well tolerated. a variety of embryonic tissues, with the exception of liver, have been successfully transplanted into the anterior chamber of rabbits (greene, ) . heterologous transplants of tissue from one species to another have more often than not been unsuccessful. greene ( ) reported good results in transplanting malignant tumors into the anterior chamber of different species. he observed that transplants from human to guinea pig and rabbit to mouse worked best. greene's technique of transplantation involves introducing a small piece of donor tissue in a small canula fitted with a stylet through a limbal incision and firmly implanting the tissue into the anterior chamber angle opposite the incision. morris et al. ( ) reported disappointing results using this technique for heterologous tumor transplants in guinea pigs, but obtained better results by transplanting the tissue into the lens beneath the lens capsule. the reader is referred to woodruff ( ) for further discussion of intraocular transplantation experiments. the eye is unique in that it is devoid of lymphatics and thus has no directly associated lymph nodes. classic animal studies have been utilized to define ocu lar mechanisms of immune response and have demonstrated that the uveal and limbal tissues play a role similar to that of secondary lymphoid tissue elsewhere in the body. bursuk ( ) injected typhoid bacilli and staphylococci into the cornea of rabbits and found that agglutinins and opsonins appeared in the corneal tissues earlier and in greater concentration than in the serum, suggesting local respon siveness. thompson and co-workers ( ) obtained identical results using crys talline egg albumin as the antigen, and further showed that when only one cornea of each animal was injected precipitating antibody could not be detected in the contralateral uninjected cornea. rabbit experiments in which ovalbumin was injected into the right cornea and human serum albumin was injected into the left cornea demonstrated that the respective cornea contained antibodies only to the antigen they had received and not to the antigen present in the contralateral cornea; specific antigen was detectable within days (thompson and olsen, ) . using a'fluorescent antibody technique, witmer ( ) demonstrated immunoglobulin cells in the uvea of the rabbit, and wolkowicz et al. ( ) showed that excised, sensitized uveal tissue in short-term culture would actively produce specific antibody. lymphoid cells migrate to the eye in the presence of antigenic stimulus. thus, x irradiation of the eyes of a rabbit does not impair the ability of either the limbal tissues or the uvea to form antibody, whereas irradiation of the peripheral lym phoid system does (thompson and harrison, ; silverstein, ) . when the primary ocular response to an antigenic challenge has subsided, sensitized 'memory" lymphocytes may persist in the uvea or limbal tissues, since a later exposure to the same antigen introduced at a distant site or injected directly into the circulation results in renewed antibody production within the eye (pribnow and hall, ; silverstein, ) . the dog, cat and monkey have all been used as animal models in cornea research. the rabbit, however, has been exploited to such a degree for this purpose that a few comments on the rabbit cornea are appropriate. the rabbit has a large cornea approximately the same diameter as the human cornea. it is significantly thinner, however, with an average central thickness of about . mm. the thin cornea and shallow anterior chamber have contributed to technical difficulties in performing intraocular lens implantation and penetrating keratoplasty in rabbits (mueller, ) . the corneal epithelium is also thin in rabbits, and the existence of bowman's layer in the rabbit has been disputed (prince, b) . one of the most important qualities of the rabbit cornea is the capacity of the endothelium for regeneration. following injury, rabbit endothelial cells are able to divide and cover the defect with minimal effect on the endothelial cell density. the dog is similar in this respect (befanis and peiffer, ) . cats and monkeys, on the other hand, have been shown to have very little potential for regeneration of endothelium. in these animals, as in man, remaining endothelial cells grow and spread to cover the defect (van horn, ) . thus cats and monkeys are probably better models in studies where response to endothelial injury is impor tant. the recent surge in interest in the nature of the corneal endothelium has been brought about largely by the demonstration that the endothelium is the site of the "fluid pump" transport mechanism essential for maintaining corneal dehydra tion and thus transparency (maurice, ) . localization of the fluid pump and many subsequent studies have been made possible by the endothelial specular microscope developed for viewing the endothelium of enucleated eyes and modi fied for use with excised corneas in the form of the perfusion specular micro scope (maurice, ) (fig. ) . later the endothelial microscope was adopted for in vivo examination of humans and animals, including cats, rabbits, and monkeys (laing et al., ) . there is nothing new about viewing the specular reflection from the corneal endothelium, a long established slit lamp technique. the specular microscope, however, is an instrument designed to optimize this type of illumination to produce a high magnification ( - x) image of the endothelium suitable for photomicrography, accurate endothelial cell counts, and detailed observation of individual cells (fig. ) . a discussion of the optical principles of specular mi croscopy is provided by laing et al. ( ) . commercially available endothelial microscopes are designed primarily for use with humans, but can be easily adapted to any laboratory animal with a sufficiently large cornea (which excludes rats and mice). the smallest eye movements blur the image of the endothelium, and, therefore, general anesthesia is usually required when working with ani mals. technique and observations in dogs (stapleton and peiffer, ) and cats (peiffer et al., a) have been described. in ussing and zehran described an apparatus consisting of two small lucite chambers separated by a piece of frog skin which could be used to measure transport of materials potential differences across the frog skin. this same technique was modified for study of the cornea, and dual chamber perfu sion experiments form the basis for much of what we know about corneal physiology (donn et al., ; mishima and kudo, ) . the perfusion specu lar microscope has added an extra dimension to this type of in vitro experimenta tion by allowing direct observation of endothelial cell shape and integrity during perfusion. atraumatic removal of the cornea avoiding contact with the en dothelium or excessive bending of the cornea is essential to the success of this type of experiment. the technique of dikstein and maurice ( ) for excision of the cornea with a scierai rim has been used very successfully. for longer term in vitro experiments the cornea can be cultured by convential closed organ culture techniques as described elsewhere in this chapter. the impetus for the intensive investigation of optimum conditions for corneal storage has resulted from the feasibility of corneal transplantation in humans. most of the research has been done using rabbits, however, and is especially applicable, therefore, to use and storage of animal corneas for whatever purpose. the method of corneal storage in most eye banks today is moist-chamber storage, storing the intact globe at °c in a glass bottle with saline-soaked cotton in the bottom. the stagnant fluid quickly becomes loaded with waste prod ucts, making this method unsuitable for long-term storage. a number of alterna tive techniques are used in laboratory storage including storage of excised cor neas at °c, organ culture at °c, and cryopreservation for long-term storage. several different media for storage of excised corneas have been evaluated. the most popular at present is mccary-kaufman (m-k) media, modified tissue culture media containing tc- , % dextran , and a mixture of streptomycin and penicillin (mccarey and kaufman, ) . storage in m-k media probably preserves corneal viability longer than moist chamber storage, endothelial dam age occurring at days in moist chamber stored rabbit corneas compared to - days for m-k stored corneas (geeraets et al., ) . stocker et al. ( ) reported good success with excised corneas stored in autologous serum at °c. subsequent studies, however, show conflicting results as to the advantage of serum storage over moist chamber storage (geeraets et al., ; van horn and schultz, ) . corneal organ culture represents an alternative method of prolonging corneal viability. it is similar to convential storage of excised corneas except for tempera ture, which is maintained at °c and may allow for preservation of endothelial viability for up to weeks (doughman et al., ) . the culture media most often used is eagles mem supplemented with fetal calf serum. cryopreservation is the only method which permits indefinite storage of whole viable corneas. the tissue is passed through a graded series of solutions contain ing dimethyl sulfoxide and frozen at a controlled rate over liquid nitrogen. studies indicate that some endothelial damage may occur during the freezing process but destruction is incomplete (basta et al., ) . in fact, survival of donor endothelial cells following transplantation of cryopreserved corneas has been demonstrated using sex chromatin markers in monkeys (bourne, ) . techniques of cryopreservation have been reviewed and evaluated by ashwood-smith ( ) . methods of estimating corneal viability after various storage regimens are all based on evaluating endothelial integrity. histochemical stains have been used extensively (stocker et al., ) . dye exclusion stains, such as . % trypan blue applied to the endothelium for seconds and then rinsed off with saline, can be used without apparent toxicity to the endothelium. the pattern obtained with trypan blue staining is more difficult to see in rabbits than other species. a modification of this technique using a combination of trypan blue and alizarin red on rabbit corneas has recently been reported (spence and peyman, ) (fig. ). electron microscopy, both scanning and transmission, are also used to assess the condition of the endothelium (fig. ) . the most innocuous method of estimating corneal viability is simply mea surement of corneal thickness at body temperature which is an indirect indication of the condition of the endothelium, healthy endothelium being able to deturgese the cornea to near-normal thickness. the measurement of corneal thickness, or pachometry, has traditionally been done using an optical technique developed by von bahr ( ) . the specular microscope, however, provides for a more direct method of measuring corneal thickness by registering the distance from the point of applanation at the corneal surface to the image of the endothelium. viraj and bacterial infections of the cornea have been studied mostly in rab bits, especially in the case of herpes simplex keratitis where attempts to study this disease in other animals have been unsuccessful. techniques for initiating these infections are discussed in section iii. noninfectious forms of keratitis may result from tear deficiencies, vitamin a deficiency, allergic phenomenon, and chemical or mechanical damage. keratoconjunctivitis sicca (kcs) resulting from tear deficiency has been studied in the dog, the disorder being surgically induced by removal of the lacrimai gland and nictitating membrane (helper et al., a; gelatt et al., ) or chemi cally induced with phenazopyridine (pyridium), - mg/day orally for - weeks (slatter, ) . françois et al. ( ) have used rabbits to study kcs. they found it necessary to remove all the accessory lacrimai tissue, including harder's glands and the nictitating membrane as well as the lacrimai gland proper to induce the disease in rabbits. interstitial stromal keratitis, thought to be an immune-mediated phenomenon, has been demonstrated in mice after intravenous and intracutaneous injections of bovine γ-globulin or bovine serum albumin (kopeloff, ) . a simple method of inducing an inflammatory stromal keratitis in rabbits is injection of . ml of clove oil into the corneal stroma (leibowitz et al., ) . phlyctenular keratitis, another allergic form of keratitis, has been reported in animal models, but thygeson et al. ( ) have failed to produce true phlyctenular disease in rabbits and questions whether the previously reported models are valid. most of the clinically significant lesions, hereditary or acquired, affecting the corneal endothelium result in visually disabling corneal edema. spontaneous dystrophies leading to corneal edema are described in the veterinary literature and are reviewed by dice ( ) . corneal edema can be induced by in vivo freezing with a metal probe to destroy endothelial cells. the edema may be temporary or permanent depending on the area, temperature, and duration of cold exposure (chi and kelman, ) . we have found that two successive freezes with a contoured -mm brass probe, cooled to - °c with liquid c , consis tently produces temporary corneal edema in dogs (befanis and peiffer, ) . a slightly smaller probe is used for rabbits. the probe is applied two times for - seconds with time for complete thawing allowed in between. corneal edema and scarring are often irreversible. corneal transplantation is one method of restoring a transparent visual axis, but a number of prosthetic devices, usually fashioned out of methyl methacrylate, have been used experi mentally over the last years and are now being used clinically on a limited basis. stone and herbert ( ) reported a two-stage procedure for implanting a plastic window in rabbit corneas. dohlman and refojo ( ) have reviewed the previous years experience with plastic corneal implants. many of these procedures require the use of a tissue adhesive to hold the plastic prosthesis in place. cyanoacrylate adhesives are used. methyl- cyanoacrylate creates the strongest bond, but is too toxic for use on the cornea. butyl- -cyanoacrylate can be used, and the longer chain substituted forms are even less irritating but do not form as strong a bond (havener, ) . gasset and kaufman ( ) have used octyl- -cyanoacrylate for gluing contact lenses to rabbit and monkey corneas with the epithelium removed. glued-on contact lenses have also been used for treatment of experimental alkali burns in rabbits (zauberman and refojo, ; kenyon et al., ) . conventional contact lenses without adhesive can be used in animal experi ments. most investigators have found that both hard lenses and silicone lenses are well-tolerated in the rabbit (thoft and friend, ) . hard lenses should be specifically fitted to the rabbits, however, and a partial tarsorraphy may be used if necessary to prevent rabbits from expelling lenses (enrich, ) . systemic connective tissue disorders may be associated with local ocular signs of diffuse, nodular, or necrotizing scleritis that can involve the cornea as well. hembry et al. ( ) sensitized rabbits by intradermal ovalbumin plus freund 's adjuvant followed by injection of ovalbumin at the limbus to produce a necrotiz ing corneoscleritis. animal studies of the composition of the aqueous humor, its function, and the processes controlling the dynamic state of its constituents and volume have two main objectives: ( ) appreciation of the physiology, biochemistry, and hence metabolism of the tissues of the anterior segment and ( ) definition of the mechanisms that control rate of aqueous humor production and outflow and hence intraocular pressure. definitive information is available for only a limited number of mammalian species, and these data suggest that species variation in aqueous humor composition and dynamics exists. in addition, anatomical charac teristics of the outflow pathways differ between species. excellent reviews by cole ( ) and tripathi ( ) detail these differences, which emphasize that making inferences from research results using nonprimate mammals to man requires appreciation of these species differences. the maintenance of normal intraocular pressure (iop) is dependent upon a critial balance of dynamic equilibrium between the processes of aqueous humor production and drainage. concepts of aqueous humor formation are based largely upon laboratory work with the rabbit. aqueous humor is produced by the ciliary processes as a result of active transport and passive ultrafiltration processes; approximately one-half of the aqueous is produced by active secretion across the two-layered epithelium. the exact mechanisms of transport have not been de-fined and may demonstrate species differences; active transport of sodium in the presence of a sodium-and potassium-activated adenosine triphosphatase located in the cell membrane of the nonpigmented epithelium is one of the main primary events in the formation of the aqueous fluid (cole, ) . the remaining % of aqueous humor is formed by passive processes, including diffusion, dialysis, and ultrafiltration. the composition of aqueous humor is essentially that of a plasma ultrafiltrate but varies between species. gaasterland et al. ( ) have studied the composition of rhesus monkey aqueous. differences between aqueous and plasma levels of potassium, magnesium, chloride, and bicarbonate have been demonstrated and suggest species differences in transport mechanisms and/or anterior segment metabolism. a saturable active transport system for ascorbate has been documented in the rabbit (barany and langham, ) . transport mechanisms for amino acids demonstrated in the rabbit may not be present in rat, cat, monkey, and dog (reddy, ) . total volume of the aqueous humor will vary among species with the size of the globe and relative proportions of the anterior segment (cole, ) . the rate of aqueous humor formation in the species studied, with the exception of the uniquely high-valued cat, is approximately . - . μ,Ι/minute (cole, ) and is dependent upon ciliary artery blood pressure, the pressure in the ciliary body stroma (essentially equal to iop), and the facility of flow through the ciliary capillary and capillary wall. because of these pressure gradients, passive aqueous humor production is decreased as iop increases; this pressure-dependent component or ' 'pseudofacuity ' ' has been demonstrated to account for up to % of total facility in monkeys (bill and barany, ; brubaker and kupfer, ) . the aqueous humor enters the posterior chamber, flows through the pupil into the anterior chamber due to thermal currents, and exits the globe by passing through the flow holes in the trabecular meshwork and reentering the peripheral venous circulation via thin-walled vascular channels. methodology utilized to quantitate aqueous humor production and flow rates involve the measurement of the turnover of a substance within the aqueous introduced by active and/or passive processes from the peripheral vasculature or intraocular injection. the ideal technique does not involve introduction of a needle into the globe, as this undoubtedly disrupts normal homeostatic mechanisms. man, guinea pig, rabbit, and cat have been studied utilizing fluorescein turnover with slit lamp fluorometry, a variety of isotopes, and other substances. reasonable correlation among investigators utilizing different tech niques within species has been observed (cole, ) . the comparative anatomy of the outflow pathways has been reviewed by calkins ( ) and tripathi ( ) . in man and the primates a well-defined trabecular meshwork spans the scierai sulcus from the termination of descemet's membrane to the base of the iris to the scierai spur. deep to the trabecular meshwork within the sulcus a single large channel, schlemm's canal, drains the aqueous into the episcleral veins (fig. ) . in nonprimate mammals, the scierai spur and sulcus are absent; the trabecular fibers span the ciliary cleft, a division of the ciliary body into inner and outer leaves, deep to the fibers of the pectinate ligament. these fibers consist of uveal tissue and extend from the termination of descemet's membrane to the iris base. aqueous drains into small saculated vessels, the aqueous or trabecular veins, which communicate with an extensive scierai venous plexus (figs. and ). van buskirk ( ) utilized plastic lumenal castings and scanning electron mi croscopy to study the canine vessels of aqueous drainage and to demonstrate that in this species the exiting aqueous mixes with uveal venous blood. the mechanisms of aqueous humor outflow through the trabecular meshwork are not completely understood. passive pressure mechanisms certainly play an important role; increases in episcleral venous pressure result in decreased outflow and increased iop. active transport mechanisms for both organic and inorganic anions have been demonstrated. transmission electron microscopic studies of the endothelium of the scierai venous plexus have revealed giant cytoplasmic vac uoles that are suggestive of transcellular transport mechanisms and indicate that this is the main site of resistance to outflow (tripathi, ) . some aqueous humor may exit through the ciliary body, entering the suprachoroidal space and into the choroidal circulation and sclera. this pressureindependent uveoscleral outflow accounts for up to - % of bulk flow in subhuman primates and has been demonstrated in cats, rabbits, and man to be quantitatively less than in two species of monkeys (bill, (bill, , (bill, , bill and phillips, ; fowlks and havener, ) . qualitative uveoscleral routes have been suggested in the dog utilizing dextran-fluorescein studies (gelati et al., b) . because of the radius of curvature of the cornea and the presence of an overlying scierai ledge, light rays from the base of the iris, the angle recess, and the trabecular meshwork undergo total internal reflection, preventing direct visualization of the outflow structures without the use of a contact lens to elimi nate the corneal curve. two general types of gonioscopic contact lenses are available; indirect lenses, which contain mirrors and allow examination of the angle by reflected light, and direct lenses, through which the angle is observed directly. a magnifying illuminated viewing system, ideally the slit-lamp biomicroscope, is essential for critical evaluation. gonioscopic examination of the canine iridocomeal angle was first reported by troncoso and castroviejo ( ) , although nicholas had previously depicted the canine angle by drawings in . troncoso in compared the gross and gonioscopic appearance of the angles of the dog, cat, pig, rabbit, rhesus monkey, and man. the clinical application of gonioscopy in comparative ophthalmology is rela tively recent. lescure and amalric ( ) described its use in the dog, and subsequently numerous investigators have stressed the value of the technique in the diagnosis of glaucoma in the dog (vainsi, ; gelatt and ladds, ; bedford, bedford, , . martin ( ) has correlated the microscopic structure and gonioscopic appearance of the normal and abnormal canine iridocorneal angle. the technique is straightforward and involves topical anesthesia and minimal physical restraint in dog, cat, and rabbit and ketamine sedation in primates (fig. ). the concave surface of the lens is filled with artificial tears and placed on the corneal surface; the franklin goniolens with a circumferential flange is re tained by the eyelids and enables the examiner to have both hands free. the troncoso or koeppe lens may also be utilized; vacuum lenses and the swan lens are smaller and thus adaptable to younger and smaller animals (fig. ) . the structures observed during gonioscopic examination of the nonprimate mammal, using the dog as an example, include, from posterior to anterior, the following: . the anterior surface and base of the iris . the pectinate ligament . deep to the pectinate ligament, the ciliary cleft, and the trabecular meshwork . the deep or inner pigmented zone representing the anterior extension of the outer leaf of the ciliary body . the outer or superficial pigmented zone which is variable in presence and density and represents melanocytes in the limbus . the corneal dome (fig. ) species variations in appearance do exist. the sub-human primates present a gonioscopic appearance of the iridocorneal angle identical to man. in the cat, the pectinate fibers are thin and nonbranching, while in the dog they tend to be deeply pigmented, stout and arbiform. in the rabbit the pectinate fibers are short and broad (fig. a-c) . these tests have been used in man to detect suspicious or borderline glaucoma patients as well as to investigate the heredity of open angle glaucoma by stressing the homeostatic mechanisms of the globe. provocation results in an increase in iop that can be characterized by extent and duration. the water drinking and corticosteroid provocative tests have been most useful in open angle glaucoma, whereas the mydriatic and dark room tests have been valuable in narrow angle glaucoma (kolker and hetherington, ) . the water provocative test in man, rabbits, and subhuman primates (macaca irus) has been studied with schiotz and applanation tonometry and in combina tion with tonography and constant pressure perfusion (swanlijung and biodi, ; galin et al., galin et al., , mcdonald et al., ; galin, ; thorpe and kolker, ; casey, ) . the procedure has been used in rabbits to test the effects of different drugs on pressure-regulating mechanisms (mcdonald et al., ) . the test in dogs has defined normal values and demonstrated significant dif ferences in american cocker spaniels and beagles with spontaneous glaucoma (lovekin, ; gelatt et al., a) (fig. ). it is generally accepted that the increase in intraocular pressure after the administration of a substantial volume of water is primarily related to a sudden decrease in the osmolarity of the blood; the related influx of water into the eye is presumed to increase intraocular pressure proportional to the volume of water administered (galin et al., galin, ) . hemodilution may not be solely responsible; in man the increase in iop in % of the patients occurs before the fall in serum osmolarity (spaeth, ) . the subsequent rate of decay of iop assesses the ability of the outflow system to cope with the increase inflow. clinical measurement of the facility of outflow is accomplished by raising iop by placing a tonometer on the eye and determining the subsequent rate of volume loss and pressure decrease; as resistance to outflow of aqueous humor increases, the pressure changes will decrease. the principle of the test may be traced to the massage effect, whereby pressure on the eye leads to a softening of the globe due to increased outflow. schiotz indentation tonography employs placement of an electronically recording schiotz tonometer on the eye; the weight of the tonome ter will indent the cornea, reducing ocular volume and increasing iop. the instrument is left on the cornea for a -minute period. tables are utilized to derive the rate of aqueous humor outflow based upon the observed changes in iop over the -minute period as recorded by the tonograph (grant, ; kronfeld kronfeld , garner, ; drews, ; podos and becker, ) . in , grant showed that the coefficient of outflow facility (c) is related to the change in ocular volume (Δν) occurring over the time interval (t), as a result of the difference between the average pressure during tonography (p t av) a n d the iop prior to placement of the tonometer on the globe (p ): c = av/t(p tav -p ). the coefficient is expressed in microliters of aqeous humor outflow per minute per millimeters of mercury pressure (grant, ) . tonography has certain limitations. the accuracy is dependent upon the accu racy of the tonometer calibration, since Δν, ptav» **nd pq are derived from this data. in addition, six physiologic assumptions are made upon which accurate quantitative results are dependent (grant, ) . . that there is a constant continuous flow of aqueous humor . that the process of tonography does not alter this flow . that the process does not alter outflow facility . that the process does not change the uveal vascular volume . that tonography does in fact measure the outflow of aqueous humor through the trabecular meshwork . that the eye exists in a steady state in regards to aqueous humor dynamics during the -minute period. previous discussions of pseudofacility-the decrease in aqueous humor pro duction that occurs with increased iop-and uveoscleral outflow demonstrate that assumptions and are not valid. evidence exists that outflow facility is depen dent upon iop. in addition, uveal blood volume is probably influenced by the placement of the tonometer on the cornea (podos and becker, ) . however, tonographic c values determined for the human eye correlated well with values obtained by constant pressure perfusion, aqueous humor fluorescein disappear ance time, and aqueous humor turnover of certain substances (becker and costant, ; françois et al., ; bill and barany, ) . additional factors warrant consideration when evaluating tonographic proce dures in species other than man; animal globes differ significantly from the human eye in terms of ocular volume, radius of corneal curvature, vascular dynamics, and tissue characteristics upon which the accuracy of schiotz identation tonometry or tonography is dependent; schiotz tonometric conversion tables are inaccurate if applied to the canine eye (peiffer et al., ) . the technique cannot be utilized in animals without pharmacologie restraint, and the effect of the drugs utilized on the steady state of iop must be taken into account. helper and his associates ( b) used xylazine and ketamine sedation and found c values from . to . in normal dogs, . to . in basset hounds, and . to . in a small number of glaucomatous patients. gelatt and his associates ( a) observed a combination of acetylpromazine and ketamine to have mini mal effect on the steady-state iop of the normal canine eye, and demonstrated impairment of outflow in beagles with inherited glaucoma. peiffer et al. ( ) reported mean tonographic values of . using this anesthetic combination in normal mongrel dogs (fig. ) . applantation tonography may prove more accurate and versatile than schiotz indentation tonography in the animal laboratory. a -minute tonography period is adequate, and the effect of anatomic variables is minimized ). invasive laboratory techniques that have been described to measure facility of outflow involve direct measurement of intraocular pressure via a cannula inserted into the globe and one of three techniques: ( ) injection of a known volume of fluid and observing the rapid increase and subsequent slow decrease in iop (pressure decay curves), ( ) perfusion of the globe with fluid at a constant rate and observing the related pressure changes (constant rate perfusion), and ( ) perfusion at a variable rate necessary to maintain a given iop (constant pressure perfusion) (barany and scotchbrook, ; grant and trotter, ; macri, ; melton and hayes, ; armaly, ; melton and deville, ; grant, ; langham and eisenlohr, ; barany, ; brubaker, ; wickham etal., ) (fig. ) . these invasive techniques eliminate the variables of scierai creep and changes in uveal blood volume and species differences in corneal anatomy. all require methods in which intraocular pressure, fluid volume in the eye, and flow reach steady state during the measurement. cannulation of the anterior chamber will induce qualitative and quantitative aqueous humor changes. despite mathemati cal equivalence, the methods are quite different in practice because they differ greatly in the time it takes for the eye to go from one steady state to another. the time required to reach steady state during constant pressure perfusion is less than minutes compared to much longer times for the other techniques. all are based on the assumptions of pressure-independent facility and secretion rate, which have been criticized (langham, ) . facility of outflow can be determined in constant pressure perfusion by measuring average flow from an external reservoir into the eye; time period is a compromise between the desire to achieve accuracy by using a long period and to achieve high temporal resolution by using a short one. a -minute averaging period is utilized arbitrarily to facilitate comparison to tonographic values. facil ity may be calculated at any given pressure utilizing the formula c = fip, where f is the rate of perfusion flow in microliters of mercury per minute, p is the intraocular pressure in millimeters of mercury, and c is facility expressed in microliters of fluid per millimeter of mercury per minute. in an in vivo system, however, this equation does not consider aqueous humor production and thus provides values lower than actual facility. barany showed that facility can be calculated at two levels of pressure, Ρ λ and p , utilizing the formula c = (f -f\)k? ~ p\)· this formula necessitates assumption of a constant episcleral venous pressure and aqueous humor production. if one estimates a value of mm hg for the former and ul/minute for the latter, similar results between the two equations are obtainable by dividing rate of secretion by the episcleral venous pressure and adding the result ( . ul/minute/mmhg) to the values obtained by the equation c = f ip . character of the perfusate can influence facility; . % unbuffered saline causes a decline in resistance on prolonged infusion of the anterior chamber. barany ( ) utilized phosphate-buffered saline with calcium and glucose added, and brubaker and kupfer ( ) perfused the heparinized mammalian tissue culture medium to minimize this "washout" factor. gaasterland et al. ( ) used pooled heterologous aqueous humor to perfuse rhesus monkey eyes in vivo. melton and deville ( ) studied enucleated canine eyes using constant pressure perfusion and found an average c value of . at mm hg iop. peiffer and his associates ( ) found the facility of outflow in normal dogs anesthetized with sodium pentobarbital to have a mean value of . ± . sd which increased as iop increased. perfusion values for outflow were less than those determined tonographically ( . ± . sd with acetylpromazineketamine hydrocholoride sedation, . ± . sd with pentobarbital anes thesia) or in the enucleated globe. van buskirk and brett ( a,b) perfused enucleated canine eyes and found a pressure-dependent facility of outflow of . ± . sd at mm hg iop which increased to . at mm hg iop. outflow increased significantly during perfusion with hyaluronidase-containing solution and with time in the non-hyaluronidase-perfused eyes. peiffer et al. ( a) perfused beagles with inherited glaucoma in vivo and found impairment of outflow facility compared to normal dogs (fig. ). the limitations of tonographic and perfusion techniques to quantitate aqueous humor outflow in animal species must be appreciated; neither is ideal. the refinement and development of more accurate methodology remains a challenge to the investigator. either technique is certain to provide more relevant informa tion if performed in vivo rather than on enucleated globes. the iop is a differential fluid pressure that measures the vector sum of the forces generated by the intraocular fluids acting at the interface between the fluid and the fibrous coats of the globe. the accurate determination of the iop is difficult because all the techniques utilized to measure it in some way necessitate altering the parmeter from its original value. in the laboratory, the anterior chamber can be cannulated and iop determined directly by the fluid level of an open-air manometer; this situation, of course, is not applicable to clinical situa tions where a noninvasive technique is required. any noninvasive technique, however, must ultimately be compared to simultaneous readings from the cannu lated globe. quantitative determination of iop is achieved by one of two types of tonometry. clinical estimations depend on subjecting the cornea to a force that either indents (impresses) or flattens (applanates) it. tonometers that indent the cornea are referred to as indentation tonometers, and those that flatten it are referred to as applanation tonometers. the cornea is utilized because other areas, such as the anterior sclera, have a nonuniform thickness and the variability of additional tissues including conjunctiva bulbar fascia and the underlying anterior uvea. normal iop in many of the animal species, notably the nonprimate mammals, appears to be higher and more variable than that observed in persons (bryan, ; heywood, ) . a number of physiologic variables may affect the iop. these include the nature of the subject, the time of day, and the position of the subject. intraocular pressure is related to blood pressure, and it has been demon strated that animals that are excited will have higher iop. accurate values are obtained in animals that have been handled and previously subjected to the technique. in persons, a diurnal variation of iop has been observed with the lowest iop occurring early in the morning. this variation is probably related to changes in endogenous corticosteriod levels. diurnal variation has been demon strated in the new zealand white rabbit (katz et al., ) and in beagles with inherited glaucoma, but not in normal dogs (gelatt et al., a) . in persons significant differences in iop are observed with the patient in a sitting position as compared to the prone position. in addition, the variable of technique may contribute to the wide range of normal intraocular pressures observed in animals. sedation and anesthesia, because of associated cardiovascular effects, are likely to affect iop. this is especially true of the barbiturates. bito and co-workers ( ) found that ketamine hydrochloride had minimal effect on iop of rhesus monkeys (mucaca mulatta) and reported a mean iop of . ± . sd with higher values and a greater diurnal variation in young animals. schiotz indentation tonometry estimates iop by applying a carefully stan dardized instrument on the cornea and measuring the depth of indentation of the cornea by a weighter plunger. the schiotz tonometer has the advantages of simple construction, reasonable cost, portability, and relative simplicity of tech nique. in schiotz tonometry, a force from a small curved solid surface with a spheri cal curvature which indents the cornea will be balanced by a fluid pressure separated from the solid surface by a thin flexible membrane, the cornea in this case, when the applied force or weight of the tonometer equals the resultant force from the fluid pressure measured in a direction parallel to the direction of the applied force times the area of distortion to the membrane. the surface tension of the tear film exerts a small force parallel to the corneal surface. the instrument consists of a footplate that approximates the radius of curvature of the human cornea, a plunger, a holding bracket, a recording scale, and . , . , . and occasionally . gm weights (fig. ) . the cornea is in dented with a relatively frictionless weighted plunger; the amount of plunger protruding from the plate depends upon the amount of indentation of the cornea. the tonometer scale is adjusted so that . mm of the plunger equals one scale unit. calibration tables are used to derive the actual iop from the observed tonometer reading. most accurate estimations of iop are obtained with the lighter weights within the middle scale ranges. the technique for schiotz tonometry is relatively simple (bryan, ; vainisi, ). the cornea is anesthetized with a drop of topical anesthesia. while allowing a few seconds for the anesthesia to take effect, check the recording arm on the tonometer by placing it on the solid convex metal surface provided. the tonometer should read , indicating that no indentation of the plunger is occur ring and that the plunger surface is flush with the footplate. with a bit of practice, schiotz tonometry can be performed without assistance in the dog and cat. rabbit tonometry is facilitated with an assistant holding the animal in lateral recumbancy. the animal should be relaxed, and care should be taken not to compress the jugular veins. the first, second, and/or third fingers of the left hand or the thumb may be used to simultaneously retract the lower eyelid of the right or left eyes, respectively. the tonometer is grasped between the thumb and first finger of the right hand and the tonometer scale rotated such that it is easily observed by the examiner. the fourth finger of the right hand is rested on the frontal bone to provide stability and retract the upper eyelid. care must be taken in retracting the lids so that pressure is applied only to the bony orbital rim and not the globe itself. excessive eyelid retraction may also create abnormal forces on the globe and should be avoided. the footplate is placed on the cornea as central as possible and gentle pressure applied until the holding bracket glides freely around the footplate shaft (fig. ) . the scale reading is noted, and the tonometer is removed. the procedure is repeated two more times; each scale readings should be within one full unit to another. if the readings are to the low end of the scale, additional weights may be applied to the tonometer and the procedure repeated to obtain midscale readings. one potential source of error is the position of the tonometer in relationship to the perpendicular; deviation from the perpendicular will result in an overestimation of iop directly proportional to the degree of deviation. the process of tonometry will induce an increase in aqueous humor outflow and each repeated measure ment may be slightly lower than the previous estimate of iop. following use, the tonometer should be disassembled and cleaned carefully with a pipe cleaner. free movement of the plunger within the casing is essential for proper function, as is smooth working of the lever system and the recording arm. the process of placing the tonometer on the eye will increase iop between to mm hg; the schiotz tonometric conversion tables enable the clinician to correlate the tonometer reading with the iop prior to the placement of the fig. . use of the schiotz tonometer in a dog. the instrument is applied in a perpendicular fashion to the central cornea. its use is limited to larger primates, dogs, cats, and rabbits, and calibration tables must be devised for each species for maximally accurate determination of intraocular pressure. instrument on the cornea. because of species differences in corneal curvature, ocular rigidity, and tissue characteristics, the use of a human conversion table will result in inaccurate estimation of iop. conversion tables have been de veloped for rabbit (best et al., ) and canine (peiffer et al., ) globes. limitations of the schiotz indentation technique depend upon the species, the clinician, and the instrument. use in animals is limited to those species with relatively large corneas and animals that can be adequately restrained and posi tioned. it is most useful in larger primates, dog, cat, and rabbit. ocular rigidity, or the ability of the cornea and sclera to stretch, will vary with age and from species to species and animal to animal (best et al., ; peiffer et al., b) . ocular rigidity also increases as the tonometer is placed closer to the limbus. increases in ocular rigidity provide schiotz recordings that are higher than actual iop; with increased ocular rigidity there is less indentation by the schiotz tonometer, creating a false impression of increased iop. applanation tonometry is based upon the principles of fluid pressures; pressure equals force divided by the area. a force from a plane solid surface applied to a fluid contained by a thin membranous surface will be balanced when the area of contact times the pressure of the fluid equals the force applied by the plane solid surface. this is known as the imbert-fick law. simply stated, in applanation tonometry one may either measure the force necessary to flatten a constant area of the corneal surface or measure the area of cornea flattened by the constant applied force. electronic applanation tonometers, notably the mackay-marg, are the most versatile and accurate in a wide variety of species. probe tips are smaller than indentation tonometers, a minimum of intraocular fluid is displaced, iop is not significantly increased by the procedure, and the technique is independent of ocular rigidity and corneal curvature. it is applicable to smaller laboratory species, and its use has been reported in the chinchilla (peiffer and johnson, experimental animal models of glaucoma have been developed to study the effects of elevated iop on other ocular tissues, to determine the efficacy of medical and/or surgical treatment in reducing iop, and to define mechanisms of the glaucomatous process itself. spontaneous animal models of glaucoma have been utilized for the above purposes in addition to investigation to determine their similarities to the disease in man. historically, experimental glaucoma has been produced in rabbits by the injec tion of % kaolin into the anterior chamber to obstruct the outflow channels, with iop reaching - mm hg within days (voronina, ) . skotnicki ( ) enclosed rabbit globes with cotton threads to induce glaucoma with resultant optic disc cupping. flocks and his associates ( ) utilized a similar technique and rubber bands; while initial increases in iop reached - mm hg, pres sures dropped to - mm hg within hours. one-third of the eyes developed panophthalmitis and only % showed cupping of the optic disc. kupfer ( ) threaded polyethylene tubing into rabbit iridocorneal angles; iop increase was observed within hours and remained elevated for months. loss of retinal ganglion cells and cupping of the optic disc occurred. samis ( ) and kazdan ( ) produced glaucoma in rabbits by the intraocular injection of methylcellulose. huggert ( ) blocked the outflow of aqueous humor in rabbits using three different techniques but failed to cause significant increases in iop. de carvalho ( ) injected cotton fragments into rabbit anterior chambers, which elevated iop to - mm hg; in those animals in which the glaucoma persisted longer than days, retinal and optic disc pathology was observed. injection of talcum powder or dental cement (kalvin et al., b) produced elevated iop in monkeys. the injection of the enzyme α-chymotrypsin into the globe will produce a variable increase in iop that may or may not be prolonged; the enzyme particles dissolve the zonules of the lens, the fragments of which collect in and obstruct the trabecular mesh work. the technique has been utilized in the rhesus and owl monkey to study optic nerve and ocular vascular changes (kalvin et al., a; zimmerman et al., ; lambert et al., ; lesseil and kuwabara, ) . the enzyme may have direct toxic effects on the retina which must be considered in studying the morphologic and functional effects of elevated iop on this tissue. cyclocryotherapy will cause an acute elevation of iop in rhesus monkeys (minckler and tso, ) , and the technique has been utilized to study axoplasmic transport of the axons of the ganglion cells and optic nerve (minckler et al., ) ; the same parameters have been studied by controlled elevation of iop by cannulation and perfusion (minckler et al., ) . the use of repeated circumferential argon laser photocoagulation of the iridocorneal angle as described by gaasterland and kupfer ( ) results in a predictable sustained elevation of iop, marked reduction of outflow facility, and progressive cupping of the optic nerve head. the technique has the advantages of being noninvasive and associated with minimal intraocular inflammation and unrelated pathology. in chickens raised under continuous light exposure from the day of birth onward, iop rises related to increased resistance to aqueous humor outflow that is detectable as early as weeks of age (lauber et al., ) . morphologic studies of the trabecular meshwork of affected animals reveals an increase of intercellular collagen and elastic trabecular tissue with resultant densification of the meshwork. there was an absence of endothelial vacuoles, pores, and microchannels observed by transmission electron microscopy in normal birds (tripathi and tripathi, a,b; rohen, ) . while sporadic spontaneous cases of animal glaucoma have been described in a variety of species, only two models are reliably producable by controlled breedings, inherited glaucoma in the rabbit and beagle. these two models will be discussed briefly emphasizing methodologies utilized to define the disease pro cesses. buphthalmia (hydrophthalmus, congenital infantile glaucoma) in rabbits is due to an autosomal recessive gene with incomplete prentrance (hanna et al., ) . histologie abnormalities of the eye are observed at birth; elevated iop and buphthalmus may be observed as early as to weeks of age in some animals with progressive clouding and flattening of the cornea; ectasia of the globe, particularly in the sclero-cornea region; deepening of the anterior chamber with detachment and fragmentation of the iris membrane; partial atrophy of the ciliary body; and glaucomatous excavation of the optic disk (hanna et al., ) . the primary defect responsible for the development of glaucoma probably in volves impairment of facility of outflow. it has been postulated that the glaucoma may be part of a primary systemic disorder (hanna et al., ) . the gross ocular enlargement which characterizes buphthalmia is accompanied by fibrosis of the filtering angle (mcmaster and macri, ) , a decrease in facility of aqueous humor outflow (mcmaster, ; kolker et al., ) , and a hyposecretion of aqueous humor (smith, ; auricchio and wistrand, ; mcmaster and macri, ; greaves and perkins, ) . anatomic studies have em phasized changes at the angle and have postulated that they constitute at least one site of obstruction to aqueous humor outflow. hanna et al. ( ) demonstrate an absence of the space of fontana, the iris pillars, and either total absence or a rudimentary development of the trabecular canals and intrascleral channels. mcmaster and macri ( ) observed that the obstruction to outflow lies be tween the trabeculum and the episcleral veins. the angle according to hanna et al. ( ) is open in the adult buphthalmia but appears closed in the newborn. the combination of hyposecretion and reduced outflow explains why buphthalmic rabbits may have a normal iop. rabbits that are genetically buphthalmic but phenotypically normal appear to have an iop approximately mm hg higher than normal, and rabbits with clinical signs of buphthalmia may have iop as high as mm hg greater. the iop tends to increase a few weeks prior to observable distention of the globe. elevated iop will subsequently return to normal levels. the cause and effect relationship of the striking inverse correlation of aqueous humor ascorbate concentration with the severity of the buphthalmia is not clear; a marked drop in ascorbate levels is present in early preclinical stages of buphthalmia (lam et al., ; fox et al., ; lee et al., ) . the actual sequence of events, involving alterations in iop, outflow facility, ascorbate concentration of the aqueous humor, and clinical signs are not clearly defined. the rate of progression has been shown to vary with the genetic background. sheppard et al. ( ) reported that the corneal endothelial cells in a flat preparation from a buphthalmic rabbit were enlarged and of variable size. he postulated that the cells expand to cover the increased corneal area. van horn et al. ( ) utilized scanning electron microscopy to confirm this report, but also indicated that there is a loss of endothelial cells in the disease as well. gelati ( ) published a report on a familial glaucoma in the beagle; pathologic increases in iop occurred from months to years of age, and affected dogs demonstrated open iridocorneal angles upon gonioscopy. additional observations were summarized in subsequent papers (gelatt et al., b (gelatt et al., , c . controlled breedings suggested an autosomal recessive mode of inheritance. the glaucomatous process was divided into early ( to months of age), moderate ( to months of age), and advanced ( months of age and greater) and was evaluated clinically by tonometry, gonioscopy, and anterior and posterior segment examination. in early glaucoma, the iridocorneal angle was open and without anomalies, and iop was elevated. with moderate glaucoma, variable optic disc atrophy, elongation of the ciliary processes, and focal disin sertion of the zonules from the lens were seen in addition to elevated iop and open iridocorneal angles. advanced glaucoma was characterized by increased iop, narrow to closed iridocorneal angles, lens dislocation, optic atrophy, and progression to phthisis bulbus. scanning electron microscopy was performed in a small number of dogs and correlated with the gonioscopy observations. affected dogs responded positively to water provocation (gelatt et al., a) and dem onstrated decreased facility of aqueous humor outflow at all stages of the glaucomatous process when compared to normal dogs, both tonographically (gelatt et al., c) and by constant pressure perfusion (peiffer et al., b) . responses to topical autonomie agents whitley et al., ) and carbonic anhydrase inhibitors (gelatt et al., ) have been studied, and histochemical studies of adrenergic and cholinergic receptor sites have been performed (gwin et al., a) . peiffer and gelatt ( ) described gross and light microscopic observations of the iridocorneal angle; data sup ported gonioscopic observations that the disease appeared to be an open angle glaucoma, with secondary pathology of the angle structures noted. inflammation of the uveal tract is a common and challenging enigmatic clini cal entity that encompasses the variables of inciting stimulus, host response, and associated alteration of ocular structure. the infections uveidites are discussed elsewhere; this section will review noninfectious uveal inflammation (primarily immune-mediated in nature) in animal models, which have proved useful in defining the etiopathogenesis of disease processes; enhancing of our understand ing of the immune response in general and specifically in regards to the eye, a rather unique organ, immunologically speaking; and investigating pharmaco logie mediation of the disease processes. limitations of these models should be noted: inflammation of the uveal tract in the animal model, regardless of species, tends to be an acute, self-limited dis ease. in addition to antigens derived from ocular tissue, complete freund 's adjuvant must be given to induce experimental allergic uveitis (eau). models of chronic uveitis have been particularly difficult to develop and require repeated immunizations with the inciting antigen. even in such models, the inflammation is usually restricted to the anterior segment, and there are minimal retrograde changes compared to chronic human uveitis. several studies have shown that the eye is not an immunological privileged site. allogeneic tissue implanted into the anterior chamber triggers an im munologie reaction. franklin and pendergrast ( ) observed that allogeneic thyroidal implants were rejected by the rabbit eye in a histologie manner and chronologic sequence similar to that for implants to other parts of the body. kaplan and his associates ( ; kaplan and streilein, ) have reported that although the rejection of allogeneic implants placed in the anterior chamber of inbred rats is delayed, the immunologie recog nition via the afferent limb of the immune response is intact. since ocular implant vascularization is coincident with that of implants in other areas of the body, immunologie suppression is probably responsible for the delay in the transplanta tion rejection observed (raju and grogan, ) . in studies of anterior chamber immunization, kaplan and streilein ( ) have demonstrated that the allogeneic antigens present in the anterior chamber are processed immunologically by the spleen via a vascular route, since afferent lymphatic channels do not drain the anterior chamber. they suggest that immune deviation occurs as a result of splenic suppressor factors which delay anterior chamber graft rejection. work by vessela and his co-workers ( ) in the rat has supported this theory. their model suggests that antigen processing and recognition occur, but that effector response is delayed. a possible delay mechanism could be low dose antigen sensitization with tolerance because of the lack of lymphatics, but a more likely explanation would be the presence of suppressor factors including nonspecific serologie blocking factors, specific antigen-antibody blocking complexes, sup pressor macrophages, or suppressor t cells. the majority of experimental animal investigations have been performed using models of sympathetic ophthalmitis or lens-induced uveitis. the role of immunologie mechanisms in the sympathetic ophthalmitis models has been fairly well characterized. the vast majority of this literature does not distinguish whether the immunologie response is primary or whether it is merely an epiphenomenon resulting from an alteration in the ocular antigens produced by another, nonimmunologic, insult. the physical location and biochemical characterization of antigens responsible for the induction of sympathetic ophthalmitis models of eau have been studied extensively. earlier studies dealt with the identification of uveal antigens; bilat eral uveitis can be produced by homologous immunization (aronson, ) . the inflammation tends to be nongranulomatous; however, occasional reports (col lins, (col lins, , describe lesions histopathologically similar to the granulomatous process seen in human sympathetic ophthalmitis. vannas et al. ( ) pro duced experimental uveitis in rabbits by enucleating one eye and implanting it in the peritoneal cavity. aronson and co-workers ( a,b,c) demonstrated that uveal preparations from albino guinea pigs are antigenic, suggesting that melanin is not a vital antigenic component of the disease process. a number of workers have demonstrated that antigens from the rod outer segments and retinal pigment epithelium can induce experimental allergic uveitis in primates, guinea pigs, and rabbits (wong et al., ; meyers and pettit, ; hempel et al., ) . while previous workers had demonstrated that crude extracts of uveal tissue can also produce eau, faure et al. ( ) have suggested that the activity observed with these uveal preparations was probably due to contamination by retinal antigens, and most investigators accept their hypothesis that retinal antigens are more important than uveal antigens in eau. wacker ( ) and wacker and his associates ( ) have demonstrated that antigens present in the photoreceptor and retinal pigment epithelial layers can elicit the development of chorioretinitis and have partially characterized the responsible antigens. there is a soluble antigen (s antigen) located throughout the photoreceptor layer. this s antigen is most active in the production of eau; animals with the disease often develop delayed hypersensitivity responses to it. the s antigen appears biochemically similar to a protein subunit of retinolbinding lipoglycoprotein present throughout the photoreceptor layer. s antigen is apparently tissue-specific; crude bovine s antigen was relatively ineffective in the induction of guinea pig eau. a more purified preparation, however, resulted in a histopathologic lesion, with cellular infiltration in the anterior uvea and destruction of the photoreceptor layer. thus a purified extract of zenogeneic antigenic material had characteristics similar to those of the tissue-specific allogeneic extract. the s antigen is probably not rhodopsin, since a number of physical characteristics, including solubility, molecular weight, amino acid se quence, specific location, and the absorption spectrum, are different. the authors also identified a particulate antigen (p antigen), located in the rod outer seg ments, that does not elicit a delayed hypersensitivity response and has a lower eau induction rate. it does elicit the development of antibodies, even in the absence of disease. in most models of experimental uveitis, cell-mediated, rather than humoral, immune responses are most important in the pathophysiology of these diseases. delayed hypersensitivity reactions to inciting antigens have correlated with the onset and course of disease in eau; passive transfer experiments using lymphoid cells have been conducted; cellular mediators (lymphokines) have produced var iations of eau; and lymphoid cell depletion experiments have been performed. friedlander and his associates ( ) defined a predominantly eosinphilic re sponse to delayed hypersensitivity in the guinea pig. meyers and pettit ( ) demonstrated in the guinea pig both cutaneous de layed hypersensitivity and macrophage migration inhibition reactions toward rod outer segments and retinal pigment epithelial antigens in eau. in their study, cellular immunity appeared to correlate with clinical disease; specific antibody to the inciting antigen was absent in a number of animals that developed eau, and there was no correlation with humoral immunity. in a similar study, wacker and lipton ( ) demonstrated that delayed hypersensitivity toward the inciting antigen is usually present when eau is induced; in a number of animals who developed uveitis, no antibody response was detected. experimental uveitis can be passively tranferred to normal animals with lym phoid cells but not with serum from animals with eau (arnonson and mcmaster, ). chandler and his co-workers ( ) demonstrated that an ocular inflammation resembling uveitis can be induced with lymphokines produced by sensitized lymphocytes, and other investigators have demonstrated that experi mental uveitis can be abrogated using anti-lymphocyte serum (bürde et ai, ) . in a mouse viral model the induction and maintenance of lymphocytic choriomeningitis (lcm) virus-induced uveitis is dependent on the thymusderived lymphocytes . immune spleen cells adaptively trans ferred to immunosuppressed animals did not receive immune lymphoid cells, no uveitis was produced; if the immune spleen cells were treated with anti-theta serum to eliminate the t cells, the uveitis was also aborted. while cell-mediated immunologie alteration appears to be responsible for the majority of experimental allergic uveitis models, in experimental lens-induced granulatomatous uveitis (elgu) humoral immunity is important. needling of the lens will provoke a uveitis in rabbits provided they also receive an intramus cular injection of freund's adjuvant (müller, ) . homologous lens antigens injected into the eye have minimal effect in the absence of adjuvant (müller, ; goodner, ) , and even animals previously sensitized by repeated injec tions of lens material into the footpad fail to mount more than a limited uveitis when challenged with an intravitreal injection of either homologous or heterologous lens antigens (selzer et al., ) . findings such as these prompted the search for a naturally occurring adjuvant which might account for the spon taneous disease; halbert et al. ( a,b) were able to show that streptococcal extracts are able to potentiate lens reactions, although to a lesser degree com pared to freund's adjuvant, while burky ( ) had already demonstrated a similar effect in rabbits with staphylococcal extracts. clinical lgu, however, is not associated with bacterial infection. as lens-induced uveitis is essentially a feature of old and cataractous lenses containing a high proportion of insoluble material, behrens and manski ( a) focused attention on the possible adjuvant effect of albuminoid lens protein. they found that a single injection of albuminoid into the vitreous of inbred rats produced after about days a uveitis characterized by an initial neutrophil and macrophage response, whereas a similar injection of crystallins was essentially without effect; this suggested that albuminoid may have an effect similar to freund's adjuvant. other experiments in rats previously sensitized with whole lens preparations showed that intraocular challenge with albuminoid evokes a cellular response consisting initially of neutrophils followed by macrophages whereas crystallins, particularly a-crystallin, give rise to round cell exudation (behrens and manski, b) . these findings suggest that the soluble antigens are responsible for humoral antibody formation and that insoluble albuminoid accounts for cell-mediated responses. müller ( ) has drawn attention to the enhancing effect of previous sensitization on the uveal response to lens proteins, having found that injection of homologous lens tissue in the presence of freund's adjuvant some time before needling the lens gives rise to a uveitis of marked proportions. marak and his associates ( ) have demonstrated that elgu can be transferred with serum. immunofluorescent studies of traumatized lenses of sensitized animals have demonstrated the presence of igg and complement, components of an immune complex-mediated reaction. depletion of the third component of complement using cobra venom factor or anti-leukocyte serum to inactivate polymorphonuclear lymphocytes (pmns) decreased the incidence of elgu. while it is most likely that elgu is an immune complex disease, immune complexes in the serum or aqueous of experimental animals have not been demonstrated. carmichael et al. ( ) reproduced the uveitis that accompanied infectious canine hepatitis adenovirus infection or vaccination demonstrating an immune complex disease resulting form soluble virus-antibody complexes and the asso ciated pmn cell response. there is significant variance in the incidence of experimental allergic uveitis produced in different species of animals. genetic influences can markedly alter immune responses; in congenic animals which differ only in the region of the chromosome containing immune response (ir) genes, marked differences in the incidence of autoimmune disease, malignancy, and infections have been noted, and it is possible that genetically determined differences in immunologie reactiv ity may be important in the development of uveitis in both humans and animals. there is a paucity of data using experimental animal models to determine the importance of immune response genes in the development of uveitis. in guinea pigs, the incidence of uveitis differs markedly between different strains. while it is relatively easy to induce uveitis in the hartley or nih strains, it is slightly more difficult to induce it in strain and almost impossible to induce it in strain (mcmaster et al., ) . mice that have well-characterized histocompatibility and immune response gene systems are a potential model to further study this phenomenon. (silverstein, ; hall and o'connor, ) . is there a specific immunologie reaction by these cells toward uveal antigens that is important in the production of human endogenous uveitis, or is the reactivity observed merely as an epiphenomenon? an equally feasible mechanism for the prolongation of an ocular inflammatory response could be the structural alteration induced by a nonspecific immunologie reaction. in the rabbit the development of systemic immune complex disease, with nonocular antigens, results in a change in ocular vascular permeability (gamble et al., howes and mckay, ) . it is conceivable that, once an animal receives this type of insult to its vascular system, the structural alteration of the ocular tissue is such that a chronic uveitis either develops or continues despite the lack of specific immune reactivity toward ocular antigens. the structure, composition, and physiology of the lens has been studied in a wide variety of animal species. investigators in specific areas of lens research, however, have tended to concentrate on one type of animal model. thus, much of what is known about the embryology of the lens is derived from research with chick embryos. likewise, the characterization of lens proteins is based largely on research with bovine lenses, and the physiology and metabolism of the lens has been studied mainly in rabbits and rats. some caution must be exercised, there fore, in extrapolating data from one animal model to another since significant differences undoubtedly exist. the chick lens, for instance, has a very different protein composition than that of the mammalian lens (rabaey, ) . there are a number of important differences between the lens of all the com monly studied laboratory animals and the lens of humans and other higher pri mates as have been pointed out by van heyningen ( ) . the rat, rabbit, and bovine lens all demonstrate minimal growth during adulthood, whereas the human lens continues to grow throughout life. also, the human lens tends to maintain its water content at about %, while lower animals have a decrease in water content during old age. the composition of the primate lens is different from all others in that it contains a group of tryptophan derivatives, hydroxykynurinines, which are highly absorbant in the ultraviolet range, and, fi nally, only the higher primates appear to have the ability to change the shape of their lens in accommodation. in addition to the higher vertebrates, the amphibians also deserve mention for their contribution to lens research. newts and salamanders have long been recog nized for their ability to regenerate a lens from the pigmented epithelium of the iris (reyer, ) . toads, on the other hand, demonstrate a similar ability to regenerate a lens from cells derived from the cornea. some of the techniques involved in this type of study are discussed by stone ( ) . more recently eguchi and okada ( ) have caused renewed interest in the relationship be tween pigmented epithelium and lens by demonstrating lenslike structures in cultures of retinal pigment epithelium from chick embryos. the dry weight of the lens is comprised almost entirely of proteins, the trans parency of which is the sine qua non of lens function. understandably, then, a large part of lens research has been aimed at analyzing these proteins. the techniques used include the complete armamentarium of sophisticated tools available to the protein chemist. as previously mentioned, the bovine lens, because of its large size and availability, has been most extensively used, al though the rabbit has also been studied in some depth. mammalian lenses consist of three major classes of soluble proteins, known as α-, ß, and γ-crystallins, and an insoluble fraction, predominantly albuminoid, most likely derived from a-crystallin. the separation of lens proteins can be achieved in a variety of ways depending on the objectives of the experiment. most recent studies, however, begin by separating the lens crystallins into major classes by gel filtration chromatography. specifically, sephadex g , ultragel ac a , and biogel p all yield four major protein peaks corresponding to a-, ßn-, ßh~y and γ-crystallin (bloemendal, ) . sephacryl s- has also been used to separate the / -crystallins into four subclasses (mostafapour and reddy, ) . further separation into subclasses is often done utilizing polycrylamide electrophoresis, and still more refined separations can be achieved using twodimensional electrophoresis and immunochemical techniques. an important concept of lens composition which has been elicited by the use of immunochemical identification of proteins is that of organ specificity, as opposed to the more usual rule of species specificity. immunoelectrophoresis can be used to demonstrate cross-reactions between lens proteins of species which are widely separated on the phylogenetic scale. in general, antibodies to the lens of one species will show extensive cross-reactivity to lens proteins of species lower on the evolutionary scale, indicating that these antigens have been carried over as evolution has progressed. this technique has been useful in defining some basic evolutionary relationships. immunoelectrophoresis and other modern methods for analysis of lens proteins are discussed by kuck ( ) . more recently a technique for radioimmunoassay of a-and γ-crystillin which is sensitive to very minute amounts of protein has been described (russell et al., ) . systems for in vitro study of the intact lens may be of the open type, where the lens is continually perfused with fresh media, or the closed system in which there is no exchange or only intermittent exchange of culture media. very elaborate open systems have been designed for accurately measuring such functions as metabolite production and oxygen consumption by the lens (schwartz, a ,b, sippel, . although the open systems theoretically approximate the physiologic state more closely, closed systems similar to the one described by merriam and kinsey ( ) continue to be much more widely used because of simplicity and the potential for maintaining several lenses simultaneously. since a long series of articles has been published by kinsey et al. describing a variety of in vitro studies primarily using closed culture systems. these articles may be traced by referring to a recent addition to the series (kinsey and hightower, ) . if measuring exchange of materials between lens and culture media is not an important part of the experimental design, a compromise between open and closed system has been described by bito and harding ( ) which involves culturing lenses in dialysis bags with intermittent changes of the outer media. there is an overwhelming amount of literature describing agents capable of inducing experimental cataracts in animals and the subsequent chemical and metabolic effects of those agents. an attempt will be made here to review some of the most important and interesting methods of producing experimental cataracts. for a more complete list of experimental cataractogenic agents the reader is referred to reviews by van heyningen ( ) and kuck ( kuck ( , . rabbits and rat have been the most popular animals for this type of research. it is interesting that despite the vast amount of information accumulated on the mechanisms of cataractogenesis, the etiopathogenesis of senile cataracts so com mon in man remains undefined. a variety of physical insults ranging from the relatively gross technique of sticking a needle through the lens capsule to bombarding the lens with a neutron beam will induce cataracts in animals. by far the most widely studied types of cataract under this category is that produced by ionizing radiation. radiation cataracts have been studied in many different laboratory animals, most of which make satisfactory models with the exception of the bird which seems to be resistant to radiation cataract (pirie, ) . the lenses of younger animals are generally more susceptible to radiation, and the periphery of the lens which contains the actively dividing cells is much more sensitive than the nucleus (pirie and flanders, ) (fig. ). x rays, y rays, and neutrons are all potent cataractogens. beta radiation in sufficient doses can also be used to cause cataracts (von sallmann et al., ) . much of the early work on ocular effects of ionizing radiation is reviewed by duke-elder ( ) . because of the recent boom in microwave technology, cataracts caused by this type of radiant energy have received increased attention. the cataractogenic potential for low doses of microwave radiation is probably minimal, however. hirsch et al. ( ) and appleton et al. ( ) found that lethal doses of microwaves were often required to produce cataract in rabbits. ultraviolet radiation, although not ordinarily thought of as cataractogenic, can cause cataracts experimentally in animals when a photosensitizing agent is ad ministered simultaneously (see toxic cataracts), which illustrates the important concept that two or more cataractogenic factors may act synergistically to pro duce lens opacities (hockwin and koch, ) . diets consisting of more than % galactose will consistently produce cataracts in young, rats. this type of experimental cataract has obvious clinical relevance to galactosemia in humans and the secondary cataract which often develops in this disease. the work of kinoshita et al. ( ) and others in explaining the mechanism of this cataract have demonstrated that this model is also applicable to the study of diabetic cataracts. effects of cataractogenic sugars (galactose, xylose, and glucose) have also been studied in vitro using rabbit lenses (chylack and kinoshita, ) . diabetic cataracts may also be studied more directly by surgical or chemical induction of diabetes in laboratory animals. patterson ( ) has used alloxan, mg/kg in a single intravenous dose, to induce diabetes in rats but other tech niques have also been effective including % pancreatectomy and intravenous dehydroascorbate (patterson, ) , and intravenous streptozotocin in some spe cies (white and Çinottim, ) . a large number of drugs and toxins have been demonstrated to cause cataracts in both man and animals. toxic cataracts are reviewed in depth by kuck ( ) . many of these agents have been of particular interest because of their therapeutic use in humans. metabolic inhibitors, not surprisingly, frequently cause cataracts. one such inhibitor, iodoacetate, has been shown to produce cataracts in rabbit lenses in vivo, in vitro, and, interestingly, by direct injection into the lens (zeller and shoch, ) . dinitrophenol, a metabolic inhibitor used for weight loss in humans until a high incidence of cataracts was recognized, will also induce cataracts in the chicken (buschke, ) . sodium cyanate, used experimentally in sickle cell anemia, has recently been shown to cause cataracts in the beagle (kern et al., ) . cytotoxin agents are another group of drugs associated with cataracts. busulfan and triethylene melamine cataracts have been studied in the rat as has dibromomannitol grimes, , ) . recently bleomycin has been added to the list of cytotoxins with cataractogenic potential, causing cataracts in to -day-old rats injected intraperitoneally with a dosage of - gm/logm (weill et al., ) . drugs which cause cataracts through more subtle effects on lens metabolism include corticosteroids, anticholinesterose inhibitors, and a number of drugs which interfere with lipid metabolism. steroid cataracts, although relatively common in humans, have been difficult to reproduce in animals, but tarkkanen et al. ( ) and wood et al. ( ) have reported success in rabbits using subconjunctival and topical steroid preparations. anticholinesterase cataracts have also proved challenging to reproduce in animals, with the monkey being the only animal model available at present (kaufman et al., ) . triparanol, which blocks cholesterol synthesis, and chloroquine and chlorphentermine, which also affect lipid metabolism, have all been shown to produce cataracts. a reversible cataract is seen in rats receiving a diet containing . % triparanol; the cataract could not be reproduced in rabbits (harris and gruber, ). an un usual anterior polar cataract is produced with chloraquine and chlorphentermine (drenckhahn, ) . a group of drugs with the potential to produce cataracts by photosensitizing the lens to ultraviolet light has generated considerable attention. chlorpromazine and methozyproralen have been used in rats to study this effect (howard et al., ; jose and yielding, ) . of the nontherapeutic agents known to cause cataract, naphthalene has been studied most thoroughly, and lens opacities have been induced in the rabbit, rat, and dog. the mechanism of this type of cataract has been studied extensively by van heyningen and pirie ( ) . naphthalene can be used to create congential cataracts in rabbits (kuck, ) . toxic congential cataracts have also been seen in mice secondary to corticosteroids (ragoyski and trzcinska-dabrowska, ) . congenital cataracts can be due to infections as well as toxic and metabolic insults in vitro. although no animal model of rubella cataract is available for study, three models of viral-induced cataracts in animals have been reported. the enders strain of mumps virus, injected into the chick blastoderm prior to dif ferentiation of the lens pläcode, has been reported to cause cataracts in the survivors (robertson et al., ) . hanna et al. ( ) showed that subviral particles of st. louis encephalitis virus injected intracerebrally into -day-old rats would cause cataracts in survivors. finally, a poorly categorized infectious agent believed to be a type of slow virus, the suckling mouse cataract agent (smca), produces cataracts when injected intracerebrally in mice less than -hours old (olmsted et al., ) . smc a has also been studied in vitro by infecting cultured rabbit lenses (fabiyi et al., ) . cataracts due to amino acid deficiencies can cause cataracts in rats (hall et al., ) . the most reproducable form of this type of cataract is that due to tryptophan deficiency which can be demonstrated in guinea pigs by feeding a diet containing less than . % tryptophan (von sallmann et al., ) . despite the skyrocketing proliferation of intraocular lenses for use in humans following cataract extraction, very little animal experimentation has been done in *ff:##p s *" this area. animal selection should consider size and shape of the anterior chamber, pupil, and lens as well as species-related reactivity to surgery and lens implantation. in general, nonprimate mammals respond to intraocular manipula tions with increased exudation and inflammation compared to primates and hu mans. sampson ( ) implanted ridley-type lenses in nine dogs undergoing cataract extraction. schillinger et al. ( ) reported disappointing results with crude lenses implanted in rabbit and dog eyes. eifrig and doughman ( ) used modern iris fixation lenses in rabbits (fig. ) . rabbits were prone to develop corneal edema after lens extraction with or without lens implantation, but edema was prolonged in eyes receiving lenses. cats have been used for implantation of lenses with slightly less technical difficulty than rabbits, but their vertical slit pupil is not ideal. because of their large anterior chamber, cats are not suitable for implantation of standard-sized anterior chamber lenses. degeneration of the retina has been a leading cause of visual impairment and partial or complete blindness in both man and animals. many causal factors have been defined and may be categorized as either genetic, chemical, developmen tal, or environmental. the latter category has received the majority of experi mental attention, being primarily concerned with the induction of retrograde changes by chemicals, radiant energy, physical trauma, nutritional deficiency, and viral infections. the multiplicity of genetic models will be dealt with briefly because rather than specific techniques they are entities within themselves. the wag/rij rat is described by lai ( ) as the "retinitis pigmentosa model," featuring a slowly progressive photoreceptor degeneration thought to be an autosomal dominant trait. at present the genetic model for human retinal dystrophy is the rca (tan-hooded) or 'dystrophie rat" (herron, ) . in this animal rod photorecep tor degeneration occurs because the retinal pigment epithelial (rpe) cells do not phagocytize shed rod outer segment photopigment discs. the (rd) mutant mouse has proved to be helpful to evaluate the effects of photoreceptor degeneration on the bipolar cell terminal synapse (blanks et al., ) . feline central retinal degeneration involving a diffuse atrophy of retinal cones has been described and is a potential model for human macular degeneration (bellhorn and fischer, ) (fig. ) ; the genetics have not been defined (fischer, ) . a tool for studying retinal degeneration in mice and rats involves the use of chimera pro- inherited canine retinal degenerative disease collectively described as "pro gressive retinal atrophy" has been recognized in a number of breeds (magnusson, ; hodgmen et al, , parry, barnett, a) (fig. ) (table ii) . genetic studies have revealed the majority to be inherited as a simple autosomal recessive trait, with three genotypes in the population: the homozygous normal, the heterozygous normal carrier, and the homozygous affected dog. clinical, electroretinographic, and transmission electron microscopic studies have been utilized to define distinctly different conditions in various breeds including rod-cone dystrophy in poodles (aguirre and rubin, ) and central progressive retinal atrophy in other species (parry, ; barnett, b; aguirre and laties, ). evans rats and produced a retinal degeneration that involved both photoreceptor cells and rpe. fluorescein angiography and electron microscopy were utilized to assess altered retinal vascularity. these findings were likened to retinitis pigmentosa in man. to evaluate retinal changes caused by phenylalinine, the drug was adminis- tered subcutaneously to newborn rats for week, producing profound damage to the bipolar and ganglion cell layers. the presence of the immature blood-brain barrier was suggested to be significant in the development of the lesions (colmant, ) . hamsters were used by another investigator to produce retinal pigmentary degeneration with yv-methyl-jz-nitrosourea, this model being suggested for use in screening the retinotoxicity of carcinogenic drugs. it should be pointed out, however, that the fundus was difficult to visualize, and the results were derived from histopathology (herrold, ) . male dutch-belted rabbits were given different levels of oxalate compounds subcutaneously to study the "flecked retina" condition seen in oxalate retinopathy. this model is also pro claimed valuable for assessing b vitamin deficiencies, ethylene glycol poisoning, and methoxyfluorane anesthesia toxicity (caine et al., ) . four-to month-old pigmented rabbits were used by brown in to study the effect of lead on the rpe. the electroretinogram, electrooculogram, and flat mount and histopathological preparations were utilized. the erg and eog recordings were normal, suggesting no major disturbance of the rpe. one of the smaller new world primate species, the squirrel monkey, was used by mascuilli et al. ( ) to evaluate the effects of various elemental iron salts on the retina. ocular siderosis with rpe and photoreceptor cell damage was evident as early as hours after intravitreal administration. the erg was used to study these effects. synthetic antimalarials have long been recognized as producing ocular com plications in humans, especially macular pigmentary disturbances. berson ( ) administered chloroquine to cats to assess acute damage to the retina, utilizing intravitreal and posterior ciliary artery injections (figure ). hey wood ( ) mentions the affinity of chloroquine and related drugs for melanin pig ment. miyata et al. ( ) used the long evans pigmented rat to evaluate the effects of intramuscular fenthion on the retina. erg as well as histopathologic changes were extensive months after administration. a brief review of retinotoxic drug effects illustrated in color is available in the "atlas of veterinary ophthalmoscopy" (rubin, ) . included are funduscopic representations of ethambutol effects in the dog, naphthalene in the rabbit, chloroquine and related drugs in rabbits and cats, and a selection of other re tinotoxic agents. in the study of induced retinal degeneration, especially these conditions with rpe involvement, pigmented animals are more desirable than albinos. the sub- human primate is the only species with a macula comparable to the human, although birds may possess dual foveae. some mammals have an area centralis, a zone of high cone concentration that, while similar, is neither functionally nor anatomically identical to the macula. species differences in relative numbers of rod and cone photoreceptors varies with the animals' ecological niche, with nocturnal creatures exhibiting a preponderance of rods and diurnal animals more likely to possess a cone dominant retina. with these exceptions (not including variations in retinal vascular patterns, discussed elsewhere), the structural and functional relationships of the neurosensory retina are remarkably similar thorughout the vertebrates. a multitude of variations producing this type of degeneration have been stud ied including fluorescent, incandescent, ultraviolet, infrared, and colored light excesses with continuous or interrupted exposure patterns, high and low level intensities, and other variations. additional factors such as age, sex, body tem perature, species, and nutrition have also been simultaneously evaluated. lai et al. ( ) used fisher albino rats exposed to a high-intensity fluorescent light to produce a severe peripheral retinal degeneration. a continuous highintensity fluorescent light exposure regimen for albino and hooded rats produced a more severe photoreceptor cell degeneration in the former (reuter and hobbelen, ). low-intensity continuous colored lights disclosed that white and blue bulbs produced the most damage in adult albino rats . it is apparent that continuous light sources are the most damaging to the albino animal, the higher the intensity the more severe the degeneration. newborn albino species, however, will show initial outer segment growth before damage eventually occurs (kuwabara and funahashi, ) . in , berso evaluated the -striped ground squirrel, which has an all-cone retina, under high intensity illumination. essentially times the amount of light required to produce cone degeneration in the albino rat produced "retinitis punctata albescens" in the ground squirrel. zigman et al. ( ) a more specific study was performed by tso et al. ( ) in rhesus monkeys to assess the effects of a retinally focused, low-intensity light beam under normal and hypothermie conditions. although no difference was detected due to body temperature, a progressive degeneration that involved the rpe and photoreceptor cells was identified in both study groups. the recent development of laser application in ocular therapy prompted study in laboratory species. gibbons et al. ( ) used rhesus monkeys to study visible and non visible lasers at high-and low-power levels and found that a -hour exposure of the latter was equivalent in rpe damage to hours of the former. in another study tso and fine ( ) exposed the foveolas of rhesus monkeys to an argon laser for - minutes and found cystoid separations of the rpe and bruchs membrane at - years post-insult. good anesthesia and subject stabiliza tion are obvious requirements for these procedures. the advent of microwave radiation for everyday use has initiated considerable evaluation of its safety for the human eye. retinal damage in the form of synaptic neuron degeneration was produced in rabbits by exposure to mhz of microwave radiation (paulsson et al., ) . ultrastructural changes were dis covered that had not previously been seen, and these resulted from lower levels of energy than anticipated. solar retinitis has long been a problem in man but recently has increased due to greater numbers of unprotected people being exposed to snow-reflected sunlight. ham et al. ( ) produced photochemical (short wavelength) and thermal (long wavelength) lesions in rhesus monkeys by exposure to a w xenon lamp, stimulating in the former case the solar spectrum at sea level. this model demon strated the short wavelength causal nature of sunlight overexposure, resulting in solar retinitis and eclipse blindness. higher energy radiation effects have been evaluated by other means including high-speed particle acceleration. proton irradiation of discrete areas of the retina in owl monkeys produced full thickness retinal damage (gragoudas et al., ) . rhesus monkey retinal irradiation by highly accelerated ( mev) oxygen nuclei produced extensive retinal vascular damage followed by degrees of degen eration and necrosis of the neural layers (bonney et al., ) . traumatic optic nerve injuries in humans are not so unusual as diseases that attack and destroy the optic nerve. in either case, however, retinal degeneration frequently follows. ascending and descending optic atrophy was produced in squirrel monkeys by anderson ( ) to study the degenerative process. total optic nerve neurectomy by razor blade knife via lateral orbitotomy was per formed to assess the descending condition, while xenon arc photocoagulation was performed to create the ascending condition. electron microscopy demon strated degenerative changes effecting the nerve fiber and ganglion cell layers. both methods produced morphologically similar retrograde processes; photocoagulation, however, provided evidence of ascending atrophy within weeks. commotio retinae or traumatic retinopathy was initially induced in rabbits by berlin ( ) who used an elastic stick to produce retinopathy, sipperley et al. ( ) employed a bb gun technique that delivered a standard sized metallic bb which struck the cornea between the limbus and pupillary axis to produce the desired contrecoup injury of the posterior pole. the segments of the outer nuclear photoreceptor layer were specifically effected, the authors using fluorescein an-giographic, histopathologic, and electron microscopic results to demonstrate the disease. retinal degenerative conditions caused by nutritionally related problems have been studied in rats, cats, and monkeys. these studies have mostly been con cerned with vitamin deficiencies, and for economic reasons have lent themselves well to the rat as a model. the subhuman primate may be geneologically ideal for certain investigations, but it is equally as impractical for fiscal reasons. the cat has been used because of a specific susceptibility to a deficiency of the amino acid taurine (fig. ) . in hayes et al. used young and adult cats to assess the effects of taruine deficiency by feeding a diet of casein, which contains very little cystine, the major precursor for taurine. in months a photoreceptor cell degeneration was produced, the initial signs including a hyperreflective granular zone in the area centralis. the erg recordings indicated a photoreceptor degenerative process, while morphologically the outer nuclear and plexiform layers were destroyed. recently, wen et al. ( ) discovered that taurine deficiency in cats produces in addition to the photoreceptor cell a tapetal cell degeneration, thus suggesting that this amino acid also plays a role in maintaining structural integrity in this tissue. these investigators employed the use of the visual evoked potential (vep) technique to arrive at their conclusions. due to its known unusually high concentration in the retina, numerous species, such as chickens, rabbits, rats, and frogs, have been used to define the biologic function of taurine. pourcho ( ) utilized intravenous and intravitreal radiolabeled s-taurine. the cat, mouse, rabbit, and monkey exhibited higher concentrations of taurine in müller cells, whereas, by comparison, the chick and frog showed very little. vitamin a deficiency has been studied in various species, the rat having been utilized most extensively because it is susceptible to night blindness as well as being less expensive and readily available. carter-dawson et al. ( ) used the offspring of vitamin a-deficient pregnant female rats to accelerate the desired condition of tissue depletion. under low levels of cyclic illumination ( . - ft-c) both rhodopsin and opsin levels decreased, the latter requiring a longer period of time. in addition it was determined that rod cells degenerated before the cone cells. in another experiment, autoradiographic techniques were used to assess the influence of vitamin a deficiency on the removal of rod outer segments in weanling albino rats (herron and riegel, ) . the process was distinctly retarded by the absence of the vitamin. in a combination study, robison et al. ( ) evaluated the effects of relative deficiencies of vitamin a and e in rats. autofluorescent and histochemical techniques revealed that the vitamin e-free diets produced significant increases in fluorescence and lipofucsin granule stain ing regardless of the vitamin a levels. in conclusion vitamin e was thought to prevent light damage by scavenging oxygen radicals and thereby providing pro tection via lipid peroxidation. two different species of subhuman primates (ce bus and m acaca) were used by hayes ( ) to compare (separately) vitamin a and e deficiencies. after years on a vitamin e-deficient diet a macular degeneration developed, charac terized by a focal massive disruption of cone outer segments. vitamin a defi ciency, on the other hand, produced xerophthalmia, keratomalacia, and impaired vision in six ce bus monkeys, the latter due to cone degeneration in the macula and midperipheral retina. information describing these techniques is available from a previous publication (ausman and hayes, ) . the ocular effects of two slow virus diseases have recently been studied: scrapie agent in hamsters (buyukmihci et al., ) and borna disease in rabbits (krey et al., ) . the scrapie agent when inoculated intracerebrally in young hamsters produced a diffuse thinning of the retina, the photoreceptor cell layer being most severely affected (fig. ) . volvement in each foci. viral antigen was detected in these lesions as well as in nonaffected areas, thus suggesting an immune component involvement. this condition was likened to ocular pathology seen in subacute sclerosing panencephalitis in humans. friedman et al. ( ) injected newborn albino rats with simian virus (sv ) and demonstrated viral antigen by immunofluorescense at hours. adult rats injected at birth developed retinal neovascularization, folds, and gliosis, features similar to retinal dysplasia. in comparison, an extensive retinopathy was produced by monjan et al. ( ) by infecting newborn rats with lymphocytic choriomenigitis virus. the outer nuclear layers of the retina were initially effected, followed by the inner nuclear and ganglion cell layers and finally total destruction. due to a modest inflammatory infiltrate, this condition was also suspected to be immunopathologic in nature. the critical assessment of retinal degeneration has been done through the implementation of five basic techniques: ( ) electroretinography (erg), ( ) electrooculography (eog), ( ) visual evoked response (ver), ( ) autoradiography, and ( ) light and electron microscopy. the concept of electroretinogram (erg) was first demonstrated by holmgren in using the enucleated eye of a frog. this technique is in principle applica ble to both animal and man and provides a detailed assessment of the rod-cone (visual cell) nature of the retina as well as its functional status. it is based upon the summation of retinal electrical potential changes which occur in response to light stimulation and are measured via corneal or skin surface electrodes (fig. ) . detailed analyses of the erg can be obtained from brown ( ) as well as information concerning the use of microelectrodes. additional information describing the local electroretinogram (lerg), a more precise but invasive technique, is available from rodieck ( ) . the organization of the vertebrate retina and the origin of the erg potential change has been elucidated by the use of intracellular readings. the mudpuppy, necturia maculosa, has been a favorite subject because of its large retinal cells. excellent reviews of comparative retinal organization and function are provided by dowling ( ) and witkovsky ( ) . the electrooculogram (eog) is a clinically applied test of retinal function that was first measured in the eye of a tench by du bois-reymond in . this test assesses the standing or corneofundal potential which exists between the anterior and posterior poles of the eye as a subject is taken from a dark to a light adapted state. for detailed information concerning this technique refer to krogh ( ) (rabbit) or arden and ikeda ( ) (rat) . a newer technique, the visual evoked potential (vep), has been adapted for use msec in animals by wen et al. ( ) . this enables the clinical assessment of visual function in the area centralis as compared to the peripheral retina by measuring potential changes in the visual cortex in response to focal or diffuse retinal stimulation. the techniques described for the cat are also applicable to other species. autoradiography has been available and successfully used in ophthalmologic research for several years (cowan et al. y ) . tritiated amino acids are readily available in various concentrations and forms and may be injected into the vitreous or other ocular tissues where they may become imcorporated into cellu lar protein metabolism pathways. distribution may be studied by light or electron microscopy. ogden ( ) utilized the concept of axoplasmic transport together with autoradiography to trace the course of peripheral retinal nerve fibers to the optic disc. others have utilized this technique to study outer segment photopigment metabolism. rods have been shown to regenerate photopigment discs con tinuously with the rpe participating in the process by phagocytizing the shed products. (young, ; norton et al., ; mandelcorn et al., ) . cones also have mechanisms for photopigment renewal. specific information concern ing autoradiographic techniques may be obtained from cowan et al. ( ) , rogers ( ) , and kopriwa and leblond ( ) . light and electron microscopy have expanded the morphologic study of the retina beyond the limitations of light microscopy. in laties and liebman discovered by chance that the chlortriazinyl dye, procion yellow, was a selective stain for the outer segments of retinal cones in the mud puppy and frog. further investigation by laties et al. ( ) demonstrated the value of this dye in distinguishing between the outer segments of rods and cones and its use in monitoring outer segment renewal. these investigators studied the dye in rat, dog, rabbit, and monkey eyes as well and discovered that the rod basal saccules could not be visualized in the rat and dog with the same certainty as the gekko (nocturnal lizard) and mud puppy. specimen preparation involves the injection of aqueous procion yellow ( . to %) into the vitreous using a /e-inch, -gauge needle. although originally used to described a syndrome consisting of cns and systemic anomalies in the human infant, retinal dysplasia (rd) now applies specifically to any abnormal differentiation of the retina after the formation of the anläge. the histopathologic features normally constituting rd include: ( ) rosette formation, ( ) retinal folds, and ( ) gliosis (lahav and albert, ) . table iii lists those agents that have either been associated with or used in the induction of rd. the administration of blue tongue virus vaccine to fetal lambs appears to be the one technique that offers all of the features described for rd. this technique involves the use of a live, attenuated virus vaccine, and it is effective only during the first half of gestation (silverstein, et al., ) . the rd inducing effects of x-ray exposure on the retina of the primate fetus have been described (rugh, ) . the erg was used in this experiment as well as histopathology to demonstrate the lesions. shimada et al. ( ) and used suckling rats to evaluate the effects of antitumor and antiviral drugs, respectively, the results in both being the induction of rosette formation. in silverstein used the fetal lamb and intrauterine trauma to demonstrate the relationship of the rpe in the organizational histogenesis of the retina (fig. ). experimental retinal detachment was first described by chodin in . the rabbit, dog, and subhuman primate have since become species of choice; the pigmented rabbit model probably the most practical, it being less expensive, more genetically uniform, and easier to manipulate. a spontaneous inherited detachment has been reported in the collie dog, ac companied by choroidal hypoplasia and posterior staphylomas (freeman etal., ) . experimental methods have utilized specifically designed techniques to induce detachment including: ( ) traction detachment by perforating injury (cleary and ryan, a,b) , ( ) intravitreal hyperosmotic injection (marmor, (norton et al., ) , ( ) detachment by experimental circulatory embolization (algvere, ) , and ( ) detachment by blunt needle rotation and suction without the use of hyaluronidase (johnson and foulds, ) . traction detachments are dependent on vitreal hemorrhage and scarring affects (cleary and ryan, a,b), the rabbit and rhesus monkey being successfully utilized in these studies. vitreal changes are significant in most of these techniques, producing detachments of varying dimensions and duration (marmor, ) . embolization of the retinal and choroidal circulations with resultant retinal ischemia produced a long-lasting retinal detachment (algvere, ) ; the technique used plastic (polystyrene) beads injected into the central retinal artery and the supratemporal vortex veins of owl monkeys (aotus trivirgatus). another approach which produced septic choroiditis and multifocal serous retinal detachments in dogs involved the intracarotid injection of pathogenic bacteria (meyers et al., ) . experimental studies of the optic nerve have explored two aspects of this tissue; the pathogenesis of the degenerative changes resulting from increased iop, and the development of an experimental model for allergic optic nearitis, similar to that seen in multiple sclerosis. the former topic is discussed in the section on glaucoma. in regards to the latter, several workers have observed optic neuritis as well as retinal vasculitis and uvertis in guinea pigs, rabbits, and monkeys affected with experimental allergic encephalomyelitis. (raine et al. t ; von sallmann et al., ; bullington and waksman, ) . rao et al ( ) induced papilledema and demyelination of the optic nerve in strain guinea pigs by sensitization with isogenic spinal cord emulsion in complete freunds adjuvant. unlike many other areas of ocular research where evidence obtained from animal models has been accepted with allowances for differences between the human and animal eye, the striking contrasts in ocular blood supply from one species to the next are impossible to ignore. as a result a variety of animal models have been studied in an attempt to sort out those aspects of the vascular anatomy and physiology which are highly variable from those which seem to be generally similar for a number of different species. the rabbit is sometimes used in this line of research, but more often it is supplanted by the monkey, cat, pig, or rat, all of which have an ocular blood supply which is more analogous to that of man. there are some general trends in the comparative anatomy of laboratory ani mals with respect to the ocular circulation. all of the commonly studied animals have a dual circulation to the eye consisting of ( ) the uveal and ( ) the retinal blood vessels. in higher primates, including man, both the uveal and retinal vessels are branches of the ophthalmic artery, itself a branch of the internal carotid; the external carotid system contributes very little to the ocular blood supply. in lower animals, on the other hand, the external carotid system, by way of the external ophthalmic artery, supplies the major portion of blood to the eye. many animals, including the dog, cat, and rat (but not the rabbit), have a strong anastomosing ramus between the external ophthalmic artery and the circle of willis, and since a relatively large part of their cerebral blood supply comes from the vertebral arteries, the common carotid artery may often be totally occluded with no apparent ill effects to the eye (jewell, ) . the retinal circulation in primates is supplied by the central retinal artery which branches off from the ophthalmic artery to enter the optic nerve close to the globe (fig. ) . in lower animals the retinal circulation is more often derived from a network of anastomosing branches of the short ciliary arteries usually referred to as the circle of haller-zinn. in fact, of the nonprimate mammals, the rat is the only animal with a central retinal artery homologous to that of the primates. the presence of a central retinal artery in dogs and cats is disputed (françois and neetens, ) , but even if present it is not the main source of blood to the retina (fig. ) . the extent and configuration of the retinal vessels is even more variable than their source. in most laboratory animals, including primates, dogs, cats, rats and mice, the retina is more or less completely vascularized, or holangiotic. rabbits, however, have only two wing-shaped horizontal extensions of vessels from the optic nerve which are accompanied by myelinated nerve fibers (the medullary rays), and do not actually extend into the retina at all. horses have only a few small retinal vessels scattered around the optic disc (paurangiotic), and birds have a retina which is completely devoid from blood vessels (anangiotic). in summary then the primates have an ocular blood supply which is almost identical to that of man. the rat should also be considered as an animal model in studies of the retinal circulation. the ocular blood supply in cats and dogs is somewhat less analogous to that of the human. the rabbit has some very atypical features, especially of the retinal circulation, which should be considered very carefully before including it in experiments on the ocular circulation. one feature of the orbital vascular anatomy in rabbits which has not been mentioned is a peculiar anastomosis between the two internal ophthalmic arteries which has been implicated in the consensual irritative response in rabbits following an injury to one eye (forster et al., ) . because of the complexity and small size of the intraocular blood vessels, dissection, even under the microscope, is usually inadequate to study the anatom ical relationships of these vessels. corrosion casting techniques, involving infu-sion of the eye with neoprene (prince, l' b) or plastic resin followed by digestion of the tissues, usually with concentrated sodium hydroxide, have been used for study of the uveal circulation and less often the retinal circulation. in our laboratory corrosion casting with a special low-viscosity plastic has been used with excellent results to study both the anterior and posterior uveal as well as the retinal circulation in cats and monkeys (risco and noapanitaya, ) . castings are studied with the scanning electron microscope (fig. ) . the retina, because it is thin and relatively transparent, is amenable to simpler studies of vascular morphology. the classic text on this subject is by michaelson ( ) , who based his observations primarily on flat mounts of retina perfused with india ink. other dyes have been utilized in a similar fashion (prince, b) . another popular technique introduced bykuwabara and cogan ( ) involves digestion of the retina in % trypsin at °c for - hours until the tissue shows signs of disintegration followed by careful cleaning away of the partially digested tissue from the vascular tree and then staining with periodic acid-schiff (pas). other investigators believe that trypsin digestion may destroy some of the more delicate or immature vessels and advocate pas staining alone to delineate retinal vessels in immature retina (engerman, a,b; henkind, et al., ) . so far we have mentioned only in vitro studies of the ocular circulation. in vivo studies can provide information about the dynamics of blood flow as well as morphology. the most widely used in vivo technique is fluorescein angiography. fluorescein dye is excited by blue light with a wavelength of mm, emitting a yellow-green light with a wavelength of mm. the blood vessels of the retina and iris are relatively impermeable to fluorescein because of their endothelial tight junctions. thus fluorescein angiography has been useful not only in the study of normal anatomy and time sequence of blood flow in these vessels as demonstrated by the excellent studies in the monkey by hayreh ( ; hayreh and baines, a,b) but also as a test of integrity of the blood-ocular barrier in certain pathological states, especially neovascular lesions. the diffusion of fluorescein across the blood ocular barrier can be assessed quantitatively by fluorophotometric techniques (cunha-vaz, ) . there is evidence that, in mammals at least, the blood ocular barriers are similar for different species (rodriguez-peralta, ) . the basic equipment required is a fundus camera capable of rapid sequence flash photographs and appropriate filters. satisfactory fluorescein angiograms can be obtained in most laboratory animals, even mice and rats, although general anesthesia is usually required. angiograms of dogs, cats, and other carnivora are technically more difficult because the reflection from the tapetum interferes with the observation of retinal blood flow. in these animals color fluorescein angiog raphy may give better results. many of the technical aspects of fluorescein angiography in animals are discussed by bellhorn ( ) . anterior segment angiography of the iris vessels has the advantage of requiring less expensive equipment; a mm camera with a macro lens and rapid recycling flash are essential (rumelt, ) . the rabbit, the most popular experimental animal for anterior segment angiograms, has an anterior segment circulation similar to humans, with the exception of the absence of a contribution from the anterior ciliary arteries to the anterior uveal blood supply, a situation possibly unique among mammals (ruskell, ) . the technique and interpretation of anterior segment angiography is discussed in detail by kottow ( ) . choroidal angiograms are also possible but present a technical problem in that in pigmented animals the retinal pigment epithelium (rpe) absorbs light strongly in both the excitation and emission wavelengths for fluorescein. the rpe is transparent to wavelengths in the infared range and the use of cyanine dyes for infared absorption and emission angiograms of the choroid have been described (hochheimer, ) as well as a technique for simultaneous angiography of the retinal and choroidal circulations using both fluorescein and indocyanine green (flower and hochheimer, ) . tsai and smith ( ) have used fluorescein for choridal angiograms by injecting dye directly into a vortex vein. angiograms of the ciliary circulation in rabbits have been obtained using conventional radio-opaque media and dental x-ray film (rothman et al., ) . the measurement of absolute and relative blood flow, velocity of blood flow, and oxygen content in various ocular tissues has been key to the understanding of nervous control of ocular physiology as well as the ocular effects of many drugs. unfortunately, these measurements have been fraught with technical difficulties. many of the techniques utilized have been indirect methods requiring compli cated mathematical analysis often based on tenuous assumptions. as a result large discrepancies have resulted from the use of different techniques or even the same technique in different hands. to complicate matters further different inves tigators have expressed results in different terms which are not easily interconver tible. the results of many of these studies have been reviewed and tabulated for comparison by henkind et al. ( ) . measurements of total or localized ocular blood flow have been made based on the washout of various gases including nitrous oxide, kr and xe (o'rourke, ) . a heated thermocouple has been used to measure relative blood flow (armaly and araki, ) , and bill ( a,b) measured choroidal flow directly in rabbits by cannulating a vortex vein and in cats by cannulation of the anterior ciliary vein. the most recent technique is the use of radio-labeled microspheres (o'day et al., ) . the labeled microspheres are injected into the arterial system, and shortly thereafter the animal is sacrificed and samples of ocular tissues are taken for quantitation of radioactivity. the microspheres are presumed to embolize in the small vessels of the various ocular tissues in amounts propor tional to the blood flow in that tissue. photographic analysis following injection of fluorescein and other dyes has been used to determine retinal oxygen saturation (hickam et al., ) , mean circulation time (hickam and frayser, ) , and choroidal blood flow (trokel, ; ernest and goldstick, ) . oxygen saturation has also been measured directly in monkeys using a microelectrode inserted through the pars plana (flower, ) . velocity of blood flow has been estimated using doppler tech niques (riva et al., ; stern, ) and thermistors (takats and leister, ) . while absolute measurements may not agree from one study to another, relative measurements may still be valid. most studies confirm that the choroid receives about % of the total ocular blood flow while the anterior uvea receives from to % and the retina about % of total ocular flow. one of the most important problems facing clinical human ophthalmologists is the control and treatment of vasoproliferative diseases in the eye. these neovascular proliferations may be confined to the retina or grow into the over lying vitreous and eventually result in retinal detachment or vitreous hemorrhage. they may also involve the iris (rubeosis iridis) and lead to neovascular glaucoma. diabetes, retrolental fibroplasia, and retinal vein occlusion are all commonly associated with ocular neovascularization. all of the above conditions have in common some degree of vasoobliteration followed by a period of retinal ischemia and subsequent vasoproliferative response. the most widely accepted theory is that growth of new vessels is stimulated by elaboration of a vasoproliferative factor from ischemie retina, a diffusable substance similar to the embryologie inducing agents discovered by spemann ( ) . this concept was supported by the isolation of a diffusable factor from certain tumors which stimulated neovascularity (folkman et al., ) . ryu and albert ( ) demonstrated a variable nonspecific neovascular response to viable or nonviable melanoma or retinoblastoma cells in a rabbit corneal model. the response was negligible in immune-deficient animals. more recently experiments by federman et al. ( ) have indicated that implanta tion of ischemie ocular tissues into the rabbit cornea can stimulate a non inflammatory neovascular response distant to the site of implantation. in diabetes capillary dropout in the retina is a well-documented phenomenon and is probably the precursor of proliferati ve retinopathy. reports of eye changes, especially capillary changes, in animals with spontaneous or induced diabetes are common (table iv) , but many of these reports represent sporadic findings. engerman ( a,b) has reviewed the subject and feels that the re tinopathy associated with alloxan-induced diabetes in dogs is the best animal model based on morphology and reproducibility of the lesions. alloxan diabetes produces a similar retinopathy in monkeys but microaneurysms require about years to develop (r. l. engerman, unpublished communication, ) . the retinopathic changes seen in these animals consists of microaneurysms and other structural capillary changes. we are not aware of any animal model of proliferative diabetic retinopathy with extraretinal neovascularization. patz and maumenee ( ) ; patz et al. ( ) sibay and hausier ( ) gepts and toussaint ( ) bloodworm ( ) engerman et al. ( ) ; engerman and bloodworth ( retrolental fibroplasia (rlf) is a result of oxygen toxicity to the immature retina. thus premature babies treated with oxygen are most susceptible, and the most immature part of the retina, usually the temporal periphery, is the most common area of involvement. the disease can be induced in newborn puppies, kittens, rats, and mice by exposure to high oxygen concentrations because these animals have an immature incompletely vascularized retina at birth. the kitten has been the most popular animal model for rlf, but the changes seen are comparable only to the early stages of rlf in humans with severe extraretinal disease complicated by retinal detachment being a rare finding in any of the animal models. there is some evidence that there are subtle differences in the pathophysiology of human rlf and that seen in animal models (ashton, ) . isolated retinal branch vein occlusion produced with the argon laser has been shown to cause a neovascular response in monkeys (hamilton et al., ) . once again, however, the new vessels are intraretinal and do not grow into the vitreous. the injection of inflammatory or toxic substances, such as blood (yamashita and cibis, ) , ammonium chloride (sanders and peyman, ), or immunogenemic substances including insulin (shabo and maxwell, ) into the vitreous may result in intravitreal vessels and fibrous membranes. these eyes show similarities to the late stages of diabetes or rlf in humans, but the pathophysiology is most likely different. the difficulty in producing intravitreal neovascular growth in animals may be due in part to an inhibitory effect of vitreous on proliferating blood vessels (brem et al., ; felton et al., ) . rabbit v- carcinoma implanted into rabbit vitreous results in a neovascular response only after the tumor has grown into contact with the retina (finkelstein et al., ) . the same tumor implanted into the corneal stroma stimulates a neovascular ingrowth toward the tumor (brem et al., ) , and, as already mentioned, federman et al. ( ) have reported that corneal implantation of ischemie retina stimulates a similar response. rubeosis, or neovascularization of the iris, may accompany diseases which cause retinal neovascularization and often produces glaucoma which is difficult to treat. there is apparently no reproducible way to induce true rubeosis in animals at present. attempts to produce the condition experimentally have been reviewed by gartner and henkind ( ) . occlusive vascular disease of the eye includes not only the relatively spectacu lar central retinal artery and vein occlusions but also a number of disorders with more subtle or insidious signs and symptoms. ischemie optic neuropathy, geo-graphic choroiditis, some forms of persistent uveitis (knox, ) , and possibly acute posterior multifocal placoid pigment epitheliopathy (ueno et al., ) are all examples of ocular disorders which can result from vasoocclusive events. experimental occlusion of the larger ocular vessels can be accomplished by occlusion of the vessel with a ligature or clamp. this includes the long and short posterior ciliary arteries (hayreh, ; hayreh and baines, a,b; anderson and davis, ) and the central retinal artery and vein (hayreh et al., (hayreh et al., , . lesseil and miller ( ) reported effects on the optic nerve and retina in surviving monkeys following complete circulatory arrest for to minutes. experimental occlusion of the smaller vessels has been produced by embolization of the arterial system with plastic beads or various other particulate matter. most of the early embolization experiments were limited by the fact that the particulate material was injected into the carotid artery, thus indiscriminately embolizing small vessels throughout the eye (reinecke et al., ; hollenhorst et al., ; ashton and henkind, ) . more recent reports describe tech niques for selective embolization of the choroidal and retinal circulations based on injection site (kloti, ) or temporary occlusion of the retinal circulation during injection (stern and ernest, ) . the most selective occlusion of retinal vessels has been accomplished with photocoagulation using the argon laser (hamilton et al., ) or xenon photocoagulator (ernest and archer, ) . the concepts of using the globe as an in vivo tissue culture medium has been discussed and exemplified in a previous section; the technique is applicable to a variety of autologous, homologous, and heterologous tumors of a variety of cell line origins. specifics are more within the realm of the oncologist rather than the ophthalmologist and as such will not be dwelt on. several experimental tumor models have been well defined and have been utilized to eludiate spontaneous tumorogenic processes and to study the biological behavior of the induced tumors. intraocular injection of viruses will induce tumors in a number of laboratory animals. injection of human adenovirus type , a dna virus, into newborn rat vitreous produced retinoblastoma-like tumors in of animals between and days postinjection . albert and his associates ( ) had earlier described similar observations in hamsters injected subcutaneously with tissue-cultured ocular cells lines exposed to the virus. muaki and kobayashi ( ) produced extraocular orbital tumors in of newborn hamsters to days following intravitreal injection of the same virus; the tumors resembled a neuroepithelioma, suggesting that they were derived from neurogenic primordia in the retrobulbar space. further investigations demon strated the retinoblastoma in mice (muaki et al., ) . jc polyoma virus was injected into the eyes of newborn hamsters by ohashi and his associates ( ); % of the animals developed retinoblastoma-or ependymoblastoma-like intraocular tumors vi-\?> months postinjection. eleven percent developed extraocular tumors, including schwannomas, fibrosarcomas, lacrimai gland carcinomas, and ependymal tumors. feline leukemia virus, an oncornavirus, was injected systemically and intraocularly into fetal and newborn kittens; a tumor of apparent retinal origin developed in of subjects (albert et al., a) . the same virus injected into the anterior chamber of cats results in the development of iris and ciliary body melanomas or, less commonly, fibrosarcomas, in a high percentage of injected animals (shadduck et al., ; albert et al., b) . in taylor and his associates ( ) utilized intravenous ra to induce ciliary body melanomas in dogs; the tumors appeared to originate from the pigmented epithelium. albert et al. ( ) transplanted human choriodal melanoma into the vitreous of the "nude" mouse, a homozygous mutant (nu/nu) with a defect in cellular immunity. serial passage transplantation was possible. the fact that the animal is immunodeficient limits the value of this model to study biologic behavior, but this in vivo system provides abundant tissue for morphologic, biochemical, immunologie, and therapeutic studies. ophthalmic res. , . baum the eye manual of tonography system of ophthalmology. the eye in evolution the biology of the laboratory rabbit the eye tonography and the glaucomas immunopathology of uveitis ocular pharmacology proc. am. coll /« "physiology of the human eye and visual system cataract and abnormalities of the lens becker-schaffer's diagnosis and therapy of the glaucomas anterior segment fluorescein angiography drugs and ocular tissues cells and tissues in culture retinal circulation in man and animals nuclear ophthalmology cell and tissue culture the vertebrate visual system comparative anatomy of the eye anatomy and histology of the eye and orbit in domestic animals textbook of veterinary internal medicine techniques in autoradiography atlas of veterinary ophthalmoscopy the rabbit in eye research immunopathology of uveitis the eye a treatise on gonioscopy the eye the vertebrate eye and its adaptive radiation transplantation of tissues and organs proc. am. coll key: cord- - gpasqtr authors: wild, karoline; breitenbücher, uwe; képes, kálmán; leymann, frank; weder, benjamin title: decentralized cross-organizational application deployment automation: an approach for generating deployment choreographies based on declarative deployment models date: - - journal: advanced information systems engineering doi: . / - - - - _ sha: doc_id: cord_uid: gpasqtr various technologies have been developed to automate the deployment of applications. although most of them are not limited to a specific infrastructure and able to manage multi-cloud applications, they all require a central orchestrator that processes the deployment model and executes all necessary tasks to deploy and orchestrate the application components on the respective infrastructure. however, there are applications in which several organizations, such as different departments or even different companies, participate. due to security concerns, organizations typically do not expose their internal apis to the outside or leave control over application deployments to others. as a result, centralized deployment technologies are not suitable to deploy cross-organizational applications. in this paper, we present a concept for the decentralized cross-organizational application deployment automation. we introduce a global declarative deployment model that describes a composite cross-organizational application, which is split to local parts for each participant. based on the split declarative deployment models, workflows are generated which form the deployment choreography and coordinate the local deployment and cross-organizational data exchange. to validate the practical feasibility, we prototypical implemented a standard-based end-to-end toolchain for the proposed method using tosca and bpel. in recent years various technologies for the automated deployment, configuration, and management of complex applications have been developed. these deployment automation technologies include technologies such as chef, terraform, or ansible to name some of the most popular [ ] . additionally, standards such as the topology and orchestration specification for cloud applications (tosca) [ ] have been developed to ensure portability and interoperability between different environments, e.g., different cloud providers or hypervisors. these deployment automation technologies and standards support a declarative deployment modeling approach [ ] . the deployment is described as declarative deployment model that specifies the desired state of the application by its components and their relations. based on this structural description a respective deployment engine derives the necessary actions to be performed for the deployment. although most of these technologies and standards are not limited to a specific infrastructure and able to manage multi-cloud applications, they all use a central orchestrator for the deployment execution. this central orchestrator processes the declarative deployment model and either forwards the required actions in order to deploy and orchestrate the components to agents, e.g., in the case of chef to the chef clients running on the managed nodes, or executes them directly, e.g., via ssh on a virtual machine (vm), as done by terraform [ ] . however, today's applications often involve multiple participants, which can be different departments in a company or even different companies. especially in industry . the collaboration in the value chain network is of great importance, e.g., for remote maintenance or supply chain support [ ] . all these applications have one thing in common: they are cross-organizational applications that composite distributed components, whereby different participants are responsible for different parts of the application. the deployment and management of such applications cannot be automated by common multi-cloud deployment automation technologies [ ] , since their central orchestrators require access to the internal infrastructure apis of the different participants, e.g., the openstack api of the private cloud, or their credentials, e.g., to login to aws. there are several reasons for the involved participants to disclose where and how exactly the application components are hosted internally: new security issues and potential attacks arose, legal and compliance rules must be followed, and the participant wants to keep the control over the deployment process [ ] . this means that common centralized application deployment automation technologies are not suitable to meet the requirements of new emerging application scenarios that increasingly rely on cross-organizational collaborations. in this paper, we address the following research question: "how can the deployment of composite applications be executed across organizational boundaries involving multiple participants that do not open their infrastructure apis to the outside in a fully automated decentralized manner?" we present a concept for the decentralized cross-organizational application deployment automation that (i) is capable of globally coordinating the entire composite application deployment in a decentralized way while (ii) enabling the involved participants to control their individual parts locally. therefore, we introduce a global multi-participant deployment model describing the composite crossorganizational application, which is split into local parts for each participant. based on the local deployment models a deployment choreography is generated, which is executed in a decentralized manner. based on the tosca and bpel [ ] standards the existing opentosca ecosystem [ ] is extended for the proposed method and validated prototypically. for application deployment automation two general approaches can be distinguished: declarative and imperative deployment modeling approaches [ ] . for our decentralized cross-organizational application deployment automation concept both approaches are combined. most of the deployment automation technologies use deployment models that can be processed by the respective deployment engine. deployment models that specify the actions and their order to be executed, e.g., as it is done by workflows, are called imperative deployment models, deployment models that specify the desired state of an application are called declarative deployment models [ ] . we explain the declarative deployment models in a technology-independent way based on the essential deployment meta model (edmm) that has been derived from investigated deployment technologies in previous work [ ] . the meta model for declarative deployment models presented in sect. is based on the edmm and is the basis for the declarative part of the presented concept. in edmm an application is defined by its components and their relations. for the semantic of these components and relations reusable component and relation types are specified. for example, it can be defined that a web application shall be hosted on an application server and shall be connected to a queue to publish data that are processed by other components. for specifying the configuration of the components properties are defined, e.g., to provide the credentials for the public cloud or to set the name of the database. for instantiating, managing, and terminating components and relations executable artifacts such as shell scripts or services are encapsulated as operations that can be executed to reach the desired state defined by the deployment model. the execution order of the operations is derived from the deployment model by the respective deployment engine [ ] . in contrast, imperative deployment models explicitly specify the actions and their order to be executed to instantiate and manage an application [ ] . actions can be, e.g., to login to a public cloud or to install the war of a web application on an application server. especially for complex applications or custom management behavior imperative deployment models are required, since even if declarative models are intuitive and easy to understand, they do not enable to customize the deployment and management. imperative deployment technologies are, e.g., bpmn tosca [ ] , and general-purpose technologies such as bpel, bpmn [ ], or scripting languages. in general, declarative deployment models are more intuitive but the execution is less customizable, while imperative deployment models are more complex to define but enable full control of the deployment steps. therefore, there are hybrid approaches for using declarative models that are transformed into imperative models to get use of the benefits of both approaches [ ] . in this paper, we follow this hybrid approach by transforming declarative models to imperative choreography models. this means, the user only has to specify the declarative model and, thus, we explain the declarative modeling approach in sect. using a motivating scenario. first, in the next section the meta model for declarative deployment models is introduced. our approach presented in sect. is based on declarative deployment models that are transformed into imperative choreographies. based on edmm and inspired by the declarative application management modeling and notation (dmmn) [ ] , the gentl meta model [ ] , and tosca, a definition of declarative deployment models d ∈ d is introduced: model) . a declarative deployment model d ∈ d is a directed, weighted, and possibly disconnected graph and describes the structure of an application with the required deployment operations: the elements of the tuple d are defined as follows: declarative deployment model specifying all details of the desired application. the notation is based on vino tosca with components as nodes, relations as edges, and the types in brackets [ ] . in addition, sample operations are shown as dots. following the design cycle by wieringa [ ] , we first examined the current situation in various research projects with industrial partners, namely in the projects ic f , sepia.pro , and smartorchestra . with regard to horizontal integration through the value chain network in the context of industry . , we focused on the requirements and challenges of collaboration between different companies [ ] . based on our previous research focus, the deployment and management of applications, the following research problems have emerged: (a) how can the deployment of composite applications across organizational boundaries be automated in a decentralized manner? (b) what is the minimal set of data to be shared between the involved participants to enable the automated decentralized deployment? in fig. queue and database, respectively. in addition, three operations are exemplary shown: a connectsto to establish a connection to the queue, a connectsto to connect to the database, and an install operation to install the jar artifact on the order vm. the other properties and operations are abstracted. assuming that a single organization is responsible for deploying the entire application and has full control over the openstacks and aws, the common deployment automation technologies examined by wurster et al. [ ] fit perfectly. however, in the depicted scenario two participants, p and p , who may be different departments or companies, intend to realize a cross-organizational application so that common deployment automation technologies are no longer applicable. while all participants must agree on the application-specific components, the underlying infrastructure is the responsibility of each participant. for security reasons, participants typically do not provide access to internal apis, share the credentials for aws, or leave the control over deployment to others. to address the research problems, we propose a decentralized concept to enable the cross-organizational application deployment automation ensuring that (i) only as little data as necessary is exchanged between participants and (ii) each participant controls only his or her local deployment while the overall deployment is coordinated. the proposed solution is described in detail in the following section and in sect. the implementation and validation is presented. the motivating scenario in fig. serves as use case for the validation. for the decentralized cross-organizational application deployment automation with multiple participants, it has to be considered that (i) the participants want to exchange as little data as necessary and (ii) each participant controls only his or her local deployment while the global coordination of the deployment of the entire application is ensured. taking these requirements into account, we have developed the deployment concept depicted in fig. . in the first step, the application-specific components are modeled representing the use case to be realized. they typically include the business components such as the order app, storage components such as the database component, and communication components such as the order queue in fig. . in the second step, the global multi-participant deployment model (gdm) is generated, a declarative deployment model containing all publicly visible information that is shared between the participants. this publicly visible information contains also data that must be provided by the respective infrastructure. for example, to execute the operation to establish a connection between order processor and database in fig. , the ip of the database vm is required as input. subgraphs, so called local parts of the gdm, are then assigned to participants responsible for the deployment of the respective components. the gdm is then processed by each participant. first, in step three, for each application-specific component a hosting environment is selected and the adapted model stored as local multi-participant deployment model (ldm). in the motivating scenario in fig. participant p selected aws for the order queue and the openstack for the order app. however, this individual placement decision is not shared. for the deployment execution we use an hybrid approach: based on the ldm a local deployment workflow model is generated in step four that orchestrates the local deployment and cross-organizational information exchange activities. all local workflows form implicitly the deployment choreography which enables the global coordination of the deployment across organizational boundaries. each step is described in detail in the following. in the initial step, the application-specific components representing the use case to be realized have to be modeled. they typically include business components, storage components, and communication components. in the motivating scenario in fig. the set of application-specific components contains the order app, the order queue, the order processor, and the database. in addition, the lifecycle operations, e.g., to install, start, stop, or terminate the components and relations, have to be defined for each of these components and their relations, since all input parameters of these operations must be provided as globally visible information in the gdm. application-specific components are defined as follows: c s ⊆ c d in d, where all r s = (c s , c t ) ∈ r d with {c s , c t } ∈ c s are of type d (r s ) = connectst o and for each c i ∈ c s : cap(type d (c i )) = ∅, to ensure that the application-specific components can be deployed across organizational boundaries, the gdm is generated in the second step which contains the minimal set of necessary information that have to be globally visible, i.e., that have to be shared. thus, the gdm is defined as follows: the elements of the tuple g are defined as follows: -d ∈ d: declarative deployment model that is annotated with participants. -p g ⊆ ℘(Σ + ) × ℘(Σ + ): set of participants with p i = (id , endpoint) ∈ p , whereby Σ + is the set of characters in the ascii table. -participant g : the mapping assigning a component c i ∈ c d to a participant p i ∈ p g participant g : c d → p g . the example in fig. depicts a simplified gdm. the application-specific components, depicted in dark gray, specify requirements, e.g., the order queue requires a message queue middleware. these requirements have to be satisfied by the respective hosting environment. furthermore, for these components as well as their connectsto-relations operations with input parameters are defined. to establish a connection to the order queue the url and q-name of the queue are required. either the target application-specific component provides respective matching properties such as the q-name property exposed by the order queue component or the environment has to provide it such as the input parameter url. for this, in this step placeholder host components are generated that contain all capabilities and properties that have to be exposed by the hosting environment. -for each r j ∈ r d : π (r j ) = c j with type d (r j ) = connectsto and for each operation op r ∈ operations s (r j ) all data elements v r ∈ π (op r ) \ properties s (c j ) are added to properties d (c h ). in the example in fig. the host order queue component provides the capability messagequeue and exposes the property url, which is required as input parameter for the connectsto operations. before the deployment model is processed by each participant, subgraphs of the gdm are assigned to the participants. this subgraph is called local part and indicates who is responsible for this part of the application. this is done by annotating the gdm as shown in fig. on the right. since participants typically do not want to share detailed information about their hosting environment, the gdm is given to each participant for further processing. each participant p i has to select a hosting environment for all c s ∈ c s with participant g (c s ) = p i . in fig. fig. is valid because the property url is covered and the sqs exposes the required capability messagequeue. the substitution is automated by our prototype described in sect. components and matching to existing infrastructure and middleware several approaches exist [ , , ] . soldani et al. [ ] introduced the toscamart method to reuse deployment models to derive models for new applications, hirmer et al. [ ] introduced a component wise completion, and we presented in previous work [ ] how to redistribute a deployment model to different cloud offerings. these approaches use a requirement-capability matching mechanism to select appropriate components. we extended this mechanism to match the properties as well. the resulting local multi-participant deployment model (ldm) is a partially substituted gdm with detailed middleware and infrastructure components for the application-specific components managed by the respective participant. up to this point we follow a purely declarative deployment modeling approach. the core step of our approach is the generation of local deployment workflow models that form the deployment choreography. they are derived from the ldms by each participant and (i) orchestrate all local deployment activities and (ii) coordinate the entire deployment and data exchange to establish crossparticipant relations. while centralized deployment workflows can already be generated [ ] , the global coordination and data exchange are not covered yet. cross-participant relations are of type connectsto and between components managed by different participants. to establish cross-participant relations, the participants have to exchange the input parameters for the respective connectsto-operations. in the example in fig. the relation con establishes a connection from the order processor managed by p to the order queue managed by p . the connectsto-operation requires the url and the q-name as input. both parameters have to be provided by p . since this information is first available during deployment time, this data exchange has to be managed during deployment: for each cross-participant relation a sending and receiving activity is required to exchange the information after the target component is deployed and before the connection is established. in addition, the deployment of the entire application must be ensured. independent which participant initiates the deployment, all other participants have to deploy their parts as well. this is covered by three cases that have to be distinguished for the local deployment workflow generation as conceptually shown in fig. . in the upper part abstracted ldms and in the lower part generated activities from the different participants perspectives are depicted. on the left (a) activities from a crossparticipant relation target perspective, in the middle (b) from a cross-participant relation source perspective, and on the right (c) activities generated to ensure the initiation of the entire deployment are depicted. first, a definition of local deployment workflow models based on the production process definition [ , ] is provided: for each participant p i ∈ p a local deployment workflow model w i based on the ldm is defined as: the elements of the tuple w i are defined as follows: set of control connectors between activities, whereby each e y = (a s , a t ) ∈ e wi represents that a s has to be finished before a t can start. set of data elements, whereby Σ + is the set of characters in the ascii table and v y = (datatype, value) ∈ v wi . -i wi : the mapping assigns to each activity a y ∈ a wi its input parameters and it is called the input container i wi : a wi → ℘(v wi ). -o wi : the mapping assigns to each activity a y ∈ a wi its output parameters and it is called the output container o wi : a wi → ℘(v wi ). -type wi : the mapping assigns each a y ∈ a wi to an activity type, type wi : based on this definition, local deployment workflow models can be generated based on specific rules. in fig. the resulting activities are depicted: (a) for each component c t ∈ c d that is target of a cross-participant relation r c = (c s , c t ) with participant g (c t ) = p i and participant g (c s ) = p j , an activity a t ∈ a wi : type wi (a t ) = invoke is added that invokes the start operation of c t . after a component is started, a connection to it can be established [ ] . thus, a c : type wi (a c ) = send is added to w i that contains all input parameters of the connectsto-operation of r c provided by p i in o wi (a c ). (b) for the component c s ∈ c d , the source of the cross-participant relation r c , an activity a c : type wj (a c ) = receive is add to w j of p j . with the control connector e(a init , a c ) added to w j it is ensured that the activity is activated after the initiate activity of p j . after the input values are received and the start operation of c s is successfully executed, the actual connectstooperation can be executed. (c) each workflow w i starts with the initiate activity a init ∈ a wi : type wi (a init ) = receive. to ensure that after a init is called the entire application deployment is initiated, a notification is sent to all other participants. for each p j ∈ p \ {p i } an activity a n : type wi (a n ) = send with a control connector e(a init , a n ) is added to w i . since each participant notifies all others, for n participants, each participant has to discard n- messages. since the payloads are at most a set of key-value pairs this is not critical. each participant generates a local deployment workflow model, which together implicitly form the deployment choreography. as correlation identifier the gdm id and application instance id are sufficient. while the gdm id is known in advance, the application instance id is generated by the initiating participant. the approach enables a decentralized deployment while each participant controls only his or her deployment and shares only necessary information. to demonstrate the practical feasibility of the approach we extended the tosca-based open-source end-to-end toolchain opentosca [ ] . it consists of a modeling tool winery, a deployment engine opentosca container, and a self-service portal. in tosca, deployment models are modeled as topology templates, the components as node, and the relations as relationship templates with their types. the types define properties, operations, capabilities, and requirements. plans are the imperative part of tosca, for which standard workflow languages such as bpmn or bpel can be used. all tosca elements and executables, implementing operations and components, are packaged as cloud service archive (csar). in fig. the system architecture for two participants is depicted. winery is extended by the placeholder generation and the placeholder substitution. either p or p models the application-specific components and generates the gdm using the placeholder generation that generates node types with the respective properties and capabilities. the resulting gdm is then packaged with the csar im-/exporter and sent to each participant. the substitution mapping detects the local part of managed by the respective participant in the gdm and selects topology templates from the repository to substitute the placeholder host components. the substituted topology template is then uploaded to the opentosca container. the plan builder generates a deployment plan based on the declarative model. we use bpel for the implementation. either p or p can then initiate the deployment. the plan runtime instantiates the plan and invokes the operations. the actual operation, e.g., to create a vm, is executed by the operation runtime. the communication between the opentosca containers is managed by the management bus. the management bus is the participant's endpoint in our setup. however, also arbitrary messaging middleware or any other endpoint that can process the messages can be used. we used the deployment model presented in fig. with two and three participants for the validation. in contrast to general workflow approaches [ , ] , we do not have to deal with splitting workflows according to the participants, since we can completely rely on the declarative deployment model and only implicitly generates a choreography. however, a prerequisite is that each participant only uses the predefined interfaces so that the choreography can be executed. at present, we also limit ourselves to the deployment aspect and do not consider the subsequent management. while management functionalities such as scaling are often covered by the cloud providers themselves, other functionalities such as testing, backups, or updates are not offered. management increases the complexity of automation, especially when local management affects components managed by other participants. we currently only support tosca as a modeling language and opentosca as a deployment engine. so far, we lack the flexibility to support technologies like kubernetes, terraform, or chef, which are often already in use in practice. however, this is part of the planned future work. the research in the field of multi-cloud, federated cloud, and inter-cloud [ , ] focuses on providing unified access to different cloud providers, making placement decisions, migration, and management. all these approaches consider multiple cloud providers satisfying the requirements of a single user. the cloud forms differ in whether the user is aware of using several clouds or not. however, the collaboration between different users each using and controlling his or her environment, whether it is a private, public, or multi-cloud, is not considered, but this is highly important, especially in cross-company scenarios which arose with new emerging use cases in the fourth industrial revolution. arcangeli et al. [ ] examined the characteristics of deployment technologies for distributed applications and also considered the deployment control, whether it is centralized or decentralized. however, also the decentralized approaches with a peer-to-peer approach does not consider the sovereignty of the involved peers and the communication restrictions. in previous work [ ] , we introduced an approach to enable the deployment of parts of an application in environments that restrict incoming communication. however, the control is still held by a central orchestrator. kopp and breitenbücher [ ] motivated that choreographies are essential for distributed deployments. approaches for modeling choreographies, e.g., with bpel [ ] or to split orchestration workflows into multiple workflows [ , ] have been published. however, most of the deployment technologies are based on a declarative deployment models [ ] , since defining the individual tasks to be performed in the correct order to reach a desired state are error-prone. thus, instead of focusing on workflow choreographies we implicitly generated a choreography based on declarative deployment models. breitenbücher et al. [ ] demonstrated how to derive workflows from declarative deployment models. however, their approach only enables to generate orchestration workflows which cannot be used for decentralized cross-organizational deployments. herry et al. [ ] introduced a planning based approach to generate a choreography. however, they especially focus on generating an overall choreography that can be executed by several agents. for us the choreography is only an implicit artifact, since we mainly focus on enabling the cross-organizational deployment by minimizing the globally visible information and obtaining the sovereignty of the participants. in this paper, we presented an approach for the decentralized deployment automation of cross-organizational applications involving multiple participants. a cross-organizational deployment without a central trusted third-party is enabled based on a declarative deployment modeling approach. the approach facilitates that (i) each participant controls the local deployment, while the global deployment is coordinated and (ii) only the minimal set of information is shared. a declarative global multi-participant deployment model that contains all globally visible information is generated and split to local deployment models that are processed by each participant. each participant adapts the local model with internal information and generates an imperative deployment workflow. these workflows form the deployment choreography that coordinates the entire application deployment. we implemented the concept by extending the opentosca ecosystem using tosca and bpel. in future work the data exchange will be optimized since each participant sends notification messages to all other participant and thus for n participants n- messages have to be discarded. we further plan not only to enable multi-participant deployments but also multi-technology deployments by enabling to orchestrate multiple deployment technologies. a gentl approach for cloud application topologies automatic deployment of distributed software systems: definitions and state of the art eine musterbasierte methode zur automatisierung des anwendungsmanagements. dissertation vino tosca: a visual notation for application topologies based on tosca combining declarative and imperative cloud application provisioning based on tosca the opentosca ecosystem -concepts & tools collaborative networks as a core enabler of industry . bpel chor: extending bpel for modeling choreographies declarative vs. imperative: two modeling patterns for the automated deployment of applications inter-cloud architectures and application brokering: taxonomy and survey choreographing configuration changes automatic topology completion of tosca-based cloud applications deployment of distributed applications across public and private networks supporting business process fragmentation while maintaining operational semantics: a bpel perspective e role-based decomposition of business processes using bpel bpmn tosca: a domainspecific language to model management plans for composite applications choreographies are key for distributed cloud application provisioning production workflow: concepts and techniques oasis: web services business process execution language version . ( ) . oasis: tosca simple profile in yaml version . ( ) . omg: bpmn version . . object management group (omg multi-cloud: expectations and current approaches topology splitting and matching for multi-cloud deployments toscamart: a method for adapting and reusing cloud applications a taxonomy and survey of cloud resource orchestration techniques design science methodology for information systems and software engineering the essential deployment metamodel: a systematic review of deployment automation technologies acknowledgments. this work is partially funded by the bmwi project ic f ( ma g), the dfg project distopt ( ), and the dfg's excellence initiative project simtech (exc - ). key: cord- -hsw dk d authors: thys, séverine title: contesting the (super)natural origins of ebola in macenta, guinea: biomedical and popular approaches date: - - journal: framing animals as epidemic villains doi: . / - - - - _ sha: doc_id: cord_uid: hsw dk d in december , a two-year-old child died from viral haemorrhagic fever in méliandou village in the south-east of guinea, and constituted the likely index case of a major epidemic. when the virus was formally identified as ebola, epidemiologists started to investigate the chains of transmission, while local people were trying to make sense out of these deaths. the epidemic control measures taken by national and international health agencies were soon faced by strong reluctance and a sometimes aggressive attitude of the affected communities. based on ethnographic work in macenta (forest region) in the autumn of for the global outbreak and alert response network (goarn) of the world health organization, this chapter shows that while epidemiologists involved in the outbreak response attributed the first ebola deaths in the forest region to the transmission of a virus from an unknown animal reservoir, local citizens believed these deaths were caused by the breach of a taboo. epidemiological and popular explanations, mainly evolving in parallel, but sometimes overlapping, were driven by different explanatory models: a biomedical model embodying nature in the guise of an animal disease reservoir, which in turn poses as threat to humanity, and a traditional-religious model wherein nature and culture are not dichotomized. the chapter will argue that epidemic responses must be flexible and need to systematically document popular discourse(s), rumours, codes, practices, knowledge and opinions related to the outbreak event. this precious information must be used not only to shape and adapt control interventions and health promotion messages, but also to trace the complex biosocial dynamics of such zoonotic disease beyond the usual narrow focus on wild animals as the sources of infection. at the end of december with the death of a two-year-old child in the village of méliandou in guéckédou prefecture, four days after the onset of symptoms (fever, black stools and vomiting). this patient would be considered from now on as the 'case zero', the index case stemming the severe ebola virus disease (evd) epidemic of west africa from apparently a single zoonotic transmission event. but then, with the idea of the spillover taking central stage the question arises: which animal species, the mythic 'animal zero', came to bear the burden of epidemic blame this time? while this retrospective epidemiological study was perceived as essential for limiting high-risk exposures and for quickly implementing the most appropriate control interventions, these investigations (biomedical experts deployed from the rich north) were tempted to mimic and fulfil the 'outbreak narrative' imposed by the global health governance. in this endeavour, rather than discovering the epidemiological origin, what becomes crucial is to quickly identify the carriers-'these vehicles necessary to drive forward the plot', which often function as the outbreak narrative's scapegoats. historically always located at the boundary of the human social body, the ideal candidate to carry this role in the evd epidemic of - was once again the wild and villainous non-human animal. because the pathways for emergence are in any way 'natural' or 'sylvatic', according to the dominant western biomedical model, the inclusion of wildlife in the epidemiology and the evolution of emerging infectious diseases is justified, yet its role is often misrepresented. although the probability of a humans contracting the disease from an infected animal still remains very low, certain cultural practices sometimes linked with poverty, especially 'bushmeat' hunting, continue to be seen as the main source of transgression of species boundaries. in the african context, research into emerging infections from animal sources implicates nonhuman primate 'bushmeat' hunting as the primary catalyst of new diseases. since the virus of ebola was identified for the first time in zaïre in and qualified as the first 'emerging' virus according to the new world clinic called 'global health', the link between animal and human health appears based on an 'us vs. them'. after the formal confirmation of the aetiological agent in march , the epidemic quickly took on an unprecedented scale and severity in several respects. it was declared by the who as an 'extraordinary event' because of its duration, the number of people infected, and its geographical extent which made it the largest ebola epidemic recorded in history until then. to these quantifiable impact measures were added sociological, ecological, political and economic phenomena that are much more complex to decrypt. these have had a profound impact on society, well beyond the remote rural environment that was typically affected by preceding epidemics. by threatening major urban areas, these 'geographies of blame' or 'hotspots' (usually at the margin of modern civilisation and configuring specific areas of the world or the environment into the breeding grounds of viral ontogenesis) have been mapped by 'virus-hunters' to update 'predictions about where in africa wild animals may harbour the virus and where the transmission of the virus from these animals to humans is possible'. in addition to this epidemic's extraordinary character, by spreading beyond the capacities of humanitarian aid, this new biomedically unsolved complexity conferred upon it a status of 'exceptionality' also by 'proclaiming the danger of putting the past in (geographical) proximity with the present'. this status had the effect, among others, of the most intense involvement, perhaps more visibly than before, of different disciplines, from human and animal health to the social sciences, in the international response. anthropology's response in particular was 'one of the most rapid and expansive anthropological interventions to a global health emergency in the discipline's history'. yet it is very critical that the collective social science experiences acquired during this west african ebola epidemic remained engaged to addressing future outbreaks and beyond. they translated and shared anthropological knowledge between scholars by including translation for public health specialists, transmitting that knowledge to junior scientists, and engaging in ongoing work to develop relevant methodology and theory. among the three west african countries most affected by the epidemic, guinea-conakry has been more marked by this dual 'exceptionality', that is to say, both epidemiological and social. beside the exceptionalism described by the senegalese anthropologist faye on the strong and sometimes violent demonstrations of popular reticence with regard to the activities of the 'riposte', guinea was also marked by a higher case fatality rate, as shown in the who report of march . globally raised up to more than % (while knowing that the number of cases and deaths was probably underreported), this case fatality rate confirmed the seriousness of the disease in a guinean context where the ebola virus had never hit before. neither the medical community, nor the population, nor the authorities had so far experienced it. despite all the measures implemented, to the question, why did we observe a higher case fatality rate in guinea compared to that of other countries, a multitude of factors can be advanced. the latter deserve to be the subject of multidimensional analyses, especially as this global lethality has manifested itself differently according to the geographical region of the country. the highest fatality rate was observed in forest guinea ( . %, / ), the region of origin of the index case and main epicentre of the epidemic. was this due to exclusively biomedical factors, such as a lower level of immunity among the guinean population? or was it because of late care that would have given patients less chance of surviving and fighting the virus? but then, why did people infected with the virus later arrive at ebola treatment centres (etc) in guinea? was it due to a poorer and more limited health system and frailer medical and health infrastructure than liberia and sierra leone at the time of the epidemic? or was it due to less effective coordination work by international and national teams in responding to the epidemic? or simply because in guinea the local communities were much more reluctant and intentionally opposed to the deployment of humanitarian and health assistance? although sharing broadly similar cultural worlds, what can therefore explain this notable difference of social resistance between the affected countries? combined with a divergent political practice and lived experiences of the state, especially between sierra leone and guinea, the working hypothesis drawn from my ethnographic observations in macenta and related literature review is that part of the continuing episodes of hostility and social resistance manifested by guinean communities regarding the adoption of the proposed control measures against the scourge of ebola has its origins in the divergence between explanatory systems of the disease; on the one hand, biomedical explanatory systems, and, on the other hand, popular explanatory systems. in march , when ebola hemorrhagic fever was formally identified a few months after the first death, epidemiologists and local populations each actively began to trace and understand this first human-to-human transmission chain of the disease, as well as its triggering event. evolving most often in parallel, and overlapping at times, these epidemiological and popular investigations generally refer to different explanatory models, some more biomedical ('natural') and others more mysticoreligious ('supernatural'). the purpose of this chapter is to trace and reflect on the interpretations of the origin and transmission of the ebola disease, as perceived and explained by the population, and to contrast them with the explanatory model of epidemiologists. in order to interrupt the two routes of evd transmission, namely from animal reservoirs to humans and between human infection, humanitarian responses followed the following public health logic: 'bushmeat' hunting, butchering and consumption should be banned and the ill should be isolated within etcs and burials should be made safe. yet, the interventions related to this reasoning had unattended consequences and, together with the ebola disease itself, they 'disrupted several intersecting but precarious social accommodations that had hitherto enabled radically different and massively unequal worlds to coexist'. carriers, in the case of human-to-human transmission, are generally perceived as the ones promulgating the epidemics and are marked with transgressive attributes intrinsic to their 'contagiousness' (e.g. wanton or deviant sexuality for the hiv epidemic, uncleanliness for the cholera epidemic, immigration for typhoid). however, in zoonosis-related diagnostic discourses, pathogens have the potential to reverse relations between humans and animals in such a way that the carrier becomes the victim. located at the 'interface' between humans, animals and the (natural) environmentalready proved to be a virtual place where deadly pandemic risks lie waiting for humanity-'forest people' from guinea were rendered both carriers of the disease and victims of the villainous role of nonhuman animals. the response to the fear of pandemics has been made unmistakable: we have to shield off humanity from nature. this mindset strongly adheres to the prevailing 'culture-nature divide' which is also depicted through zoonotic cycles diagrams further operating both as pilots of human mastery over human-animal relations and as crucial sites of unsettlement for the latter. wild animals became public enemy number one, together with those who were supposedly facilitating the transgression of the boundaries between the cultural and natural world with (or because of) their culturally 'primitive' or 'underdeveloped' practices. by framing 'bushmeat' hunting, as well as local burials, as the main persisting cultural practices among the 'forest people' to explain (or to justify) the maintenance of the evd transmission during the west african epidemic, the notion of culture that fuelled sensational news coverage has strongly stigmatised this 'patient zero' community both globally and within guinea, and has been employed to obscure the actual, political, economic and political-economic drivers of infectious disease patterns. appointed by my former institute, the institute of tropical medicine of antwerp, belgium (itm), to the who, i was sent to guinea-conakry from the end of october to the end of november for a four-week mission by the global outbreak alert and response network (goarn). since august , the country had been in the largest and longest phase of the epidemic, the second recrudescence which would also be the most intense one up until january . i first spent a week in conakry to follow the implementation of a social mobilisation project (project of monitoring committees at the level of each commune in the urban area). then, following an evaluation of the situation qualified as catastrophic by the national coordinator of the who, it was in macenta, forest guinea, where i was deployed. macenta, located east of guéckédou, was the prefecture considered to be the epicentre of this new outbreak of ebola and where transmission was the most intense. this district would remain one of guinea's most affected regions. by october , macenta, where catastrophic scenarios seemed possible, had already a cumulative number of almost cases since the beginning of the epidemic. the epidemiological situation was out of control, with a lack of material, human and financial resources. on arrival, there was still only one transit center (cdt). a new etc was being finalised by msf belgium. its management would be taken over a few weeks later by the french red cross. due to the long rainy season, the road used for bringing confirmed cases from macenta to the guéckédou treatment centre was in a deplorable state, slowing down the start of treatment and increasing the risk of transmission during transportation. it is as a medical anthropologist that i have been involved in guinea's national coordination platform for the fight against ebola and this within the commission of 'social mobilization and communities engagement', also named locally the 'communication' unit, in order to document, better understand and help to address the reluctance manifested by the local community. without going into the debate about the instrumentalisation of anthropologists as simple 'cultural mediators' at the service of humanitarians, i will simply recall here the specific objectives assigned for the mission. they consisted, on the one hand, in an analysis of rumours and crisis situations in order to propose responsive actions and, on the other hand, in adapting the responses and protocols of the various national and international institutions to local conditions, giving priority to comprehensive and participatory approaches. by integrating the 'communication' unit, i tried to support and animate the meticulous and sensitive work of a whole team working to rebuild trust with communities and to 'open' villages reluctant to receive care interventions. under the authority of unicef guinea, this communication team also hosted many local associations previously working for the prevention of infectious diseases, such as hiv/aids, in the region. the latter had already been mobilised to serve as a relay and to mitigate the unpredictable consequences of the epidemic not foreseen by the riposte, such as, among other things, sensitisation and reception of healed people and orphans of ebola, food distribution, and support for people and villages stigmatised by the disease for whom access to the market-purchase and sale of products-was forbidden. religious representatives of protestant and muslim communities also voluntarily joined this platform to learn and then preach preventive behaviour, to comfort the population, as well as to deconstruct and addressed rumours. their main message was to convince the public that ebola did indeed exist and 'was a real disease'. subsequently, the communication unit was finally able to associate the prefectural direction of traditional medicine of macenta counting traditional healers and distributed in the subprefectures of macenta. the main objective of this new activity was to engage all traditional healers in the fight against evd by raising the awareness of their patients and their entourage thanks to their high level of credibility in their respective communities. they also undertook to refer their patients directly to the tc if they came to present even one of the symptoms of evd (fever, diarrhoea [with blood], vomiting [with blood], loss of appetite). a 'health promotion' team managed and financed by msf belgium also acted on the ground. each morning, the different commissions and stakeholders of the riposte present in macenta were meeting at the prefectural health directorate (dps) to discuss and coordinate their activities in the field. alongside a guinean sociologist, consultant for the who and the assistant coordinator of the mission philafricaine, i was quickly immersed in the realities of the field and in the local strategies elaborated with respect of traditional hierarchies, despite the emergencies. their goal was to restore dialogue with the various village representatives who, since the officialisation of the epidemic, had decided to resist ebola interventions. this was, for instance, the case of the village of dandano, where deaths had risen to ; a village whose access was authorised the day after my arrival in macenta. although tragic, this coincidence made me earn some legitimacy from the other national and international 'fighters'. it is in this intense and difficult context that the ethnographic observations and their preliminary analysis, presented in this chapter, were collected. the methods employed are based on participant observation, including many informal discussions during meetings with villagers (representatives of youth/notables/sages/women), with religious representatives (protestant pastors, and imams), with drivers and partners of the coordination community (e.g. doctors without borders, guinean red cross, unicef among others). some formal interviews were also conducted with key informants such as healed individuals (ebola survivors), traditional healers, pastoralists and local actors in the fight. biomedical scientific literature and reports on epidemiological data, as well as observational notes, photographs and audio recordings collected in the field, allowed me to trace the interpretations of the origin and transmission of ebola in a dual perspective: that of epidemiologists, on the one hand, and that of the population on the other. it is through the concept of explanatory models or 'cultural models of the disease' developed by arthur kleinman that i attempted to interpret the observations ( fig. . ) . this is a conceptual framework that has already been used by barry and bonnie hewlett, alain epelboin and pierre formenty in their respective interventions during the previous outbreak of ebola haemorrhagic fever in the congo in . to be able to adapt the response and interrupt transmission, it is essential to know and understand how the population perceives the introduction of a disease, especially when it is such a deadly one. explanatory or cultural models refer to the explanations of an individual or a culture and to predictions about a particular disease. these are social and cultural systems that construct the clinical reality of the disease. culture is not the only factor that shapes their forms: political, economic, social, historical and environmental factors also play an important role in disease knowledge construction. in kleinmann's model, care systems are composed of three sectors (popular, professional, and traditional) that overlap. in each healthcare system, the disease is perceived, named and interpreted, and a specific type of care is applied. the sick subject encounters different discourses about the illness as she or he moves from one sector to another. kleinman defines the existence, in each sector, of explanatory models of the disease for the sick individual, for his/her family and for the practitioner, whether professional or not. in general, only one part of an explanatory model is conscious, the other is not. although the explanatory models seek from the health district mbomo in congo in , they identified five different cultural models including a sorcery model (sorcerer sending spiritual objects into victims), a religious sect (la rose croix, a christian sect devoted to study of mystical aspects of life), an illness model (fever, vomiting, diarrhoea with blood), an epidemic model (illness that comes rapidly with the air/wind and effects many people) and a biomedical model (ebola haemorrhagic fever). interestingly, none of the integrated non-biomedical models identified a specific non-human animal as potential source and/or carrier of evd or hunting and butchering as specific health risk activities for such illness. this further supports the epistemic dissonance observed during many epidemics (including the west african evd epidemic in this case), between the public health framing of wild meat as hazardous and the practical and social significance of the activities that occasion contact with that hazard. in the case of evd, it is the biomedical cultural model that prevails among western health workers. when the alert was launched by the local health authorities on march , two and a half months after the beginning of the disease of the index case, it was virologic investigations that were conducted at first, following the many deaths that occurred during this socalled silent phase. when the zaïre ebolavirus was identified as the causative agent, retrospective epidemiological investigations of the cases took place, which are crucial during the outbreak of an infectious disease responsible for such high mortality rate. the first chains of transmission of evd are presented in the below graph adapted from baize et al. ( ) (fig. . ). these investigations are mainly based on the identification of patients and the analysis of hospital documents and reports (results of blood tests carried out in the laboratory), as well as on testimonies and interviews with the affected families, the inhabitants of the villages where the cases occurred, suspected patients and their contacts, funeral participants, public health authorities and hospital staff members. virologic analyses suggest a single introduction of the virus into the human population. but the exact origin of the infection of this two-year-old child has not yet been definitively identified, even though the role of bats as natural hosts of the ebola virus, including this time also the insectivorous species, remains one of the most probable scientific hypotheses. up to now, the precise nature of the initial zoonotic event in guinea remains undetermined and the natural reservoir of the ebola virus more generally is not yet certain, beside for three species of fruit bat and other insectivorous african bat species known to carry ebola antibodies and rna. therefore, this informational gap was from the start filled with assumptions during the west african outbreak. among these assumptions, the elusive link between bats, wild animals and humans triggered high concerns over handling, butchering and consuming wild animals, commonly referred to as 'bushmeat'. consequently, these concerns were integrated into public health messages on disease prevention and were translated into a 'bushmeat ban' by governments across the region and enforced during the entire outbreak. this raises the question of the value of focusing on zoonotic transmission, in particular by fruit bats and non-human primates, which was quickly (s ) child, years-old onset dec. dotted arrows: the epidemiological links have not been well established deemed to be of minimal risk, when the biggest threat of infection was from other humans. furthermore, it raises the question of whether there is evidence to indicate and confirm that 'bushmeat'-related information included in public health campaigns in the region actually reduced ebola transmission. first, hunting and consuming 'bushmeat' for food have long been a part of human history occurring worldwide, serving as an important source of protein, and household income, especially where the ability to raise domestic animals is limited. the term itself encompasses an extensive list of taxa that are harvested in the wild (ranging from cane rats to elephants and including duiker, squirrels, porcupine, monkeys, non-human primates, bats and hogs) for food, medicine, trophies and other traditional, cultural uses. yet, designating the consumption of wild animal meat through the use of the term 'bushmeat' for west africans instead of 'game', as is the case for europeans and americans, by the media, scientific literature and public health campaigns that prohibit this practice, participates in 'semiotics of denigration' and has the effect of perpetuating 'exotic' and 'primitive' stereotypes of africa. although involuntary, the immediate and visceral effect produced in western minds by the thought of someone eating a chimpanzee, a dog or a bat, for instance, creates a feeling of disgust which downgrades this person, his/her needs and his/her claims on us. this issue has led to calls to replace the term with 'wild meat' or 'meat from wild animals'. secondly, while the term 'bushmeat' typically refers to the practice in the forests of africa, the trade of 'bushmeat', which has expanded over the past two decades, is considered as an example of an anthropogenic factor that provides opportunities for the transmission of diseases from wildlife to humans. the unsolved reconciliation between present policies and practices and the different values at stake (ecological, nutritional, economic and intrinsic values of wildlife hunted for food) in the actual 'bushmeat crisis', have accentuated the national and global conservation, development and health (infectious disease transmission related) concerns over hunting, eating and trading wild meat. thirdly, because of the many competing interests and realities involved, the proscription of hunting and consuming certain species of wild animals-in particular fruit bats and nonhuman primates during the west africa ebolavirus outbreak-has resulted in several unintended consequences, has incurred great cost and has had only a limited effect. in addition to being vague, inconsistent with scientific research and targeted to the wrong audience, messaging that unilaterally stressed the health risk posed by wild meat and fomite consumption contradicted the experiences of target publics, who consume wild meat without incident. consequently, in addition to having a negative impact on the livelihoods of people living at the frontlines of animal contact, the ban ran the risk of eroding public confidence in the response efforts and fuelling rumours as to the cause of evd (e.g. that the government was attempting to weaken villages in areas supporting the opposition party, as wild meat is considered an important source of physical 'strength' and energy). by focusing exclusively on the risk of spillover, we are distorting and concealing aspects of the dynamics at play. what if species boundaries are not perceived in the same way by everyone? what if the transgression of this 'invisible enemy' is spotted at a different intersection, beyond the nature/society binary? the first chains of human-to-human transmission led to the conclusion that the main vector of contamination was a health professional (s ) who spread the ebola virus in macenta, nzérékoré and kissidougou in february . the fifteenth patient, a doctor (s ), would have also contaminated his relatives in the same areas. the aetiological agent of this deadly disease (the ebola virus for some, the transgression of a taboo for others) remained hidden until then and finally became apparent because of clusters of cases in the hospitals of guéckédou and macenta. indeed, even though the high risk of exposures was elucidated, the problem remained hidden for a number of months, mainly because no doctor or health official had previously witnessed a case of ebola and because its clinical presentation was similar to many other endemic diseases experienced in guinea, such as cholera, which affects the region regularly. but these signals could also have been blurred by another narrative of the causative agent of these same symptoms. this is very similar to what genese marie sodikoff has identified during the recent bubonic plague epidemic in madagascar, when scientists elicited survivors' memories of dead rats in the vicinity to reconstruct the transmission chain. not only were these clues imperceptible to most, but residents had also constructed an alternative outbreak narrative based on different evidence. indeed, the mystico-religious beliefs deeply rooted in this region, even within the medical profession, have offered a different interpretation of causality according to a cultural model other than the biomedical model used by epidemiologists. following james fairhead, it is important to note that 'cultural' model does not tend here to slip into more totalising ideas of 'culture', such as a model being a 'kissi culture' (see below) nor its strict symmetrical opposite (e.g. a model of the 'humanitarian culture' or of a 'western culture'). origin and transmission chain according to an 'animist' model at the beginning of the epidemic, for some, the first deaths in forest guinea were due to the transmission of the filovirus through contact with animals' and/or patients' body fluids; while for others, these deaths originated from the transgression of a taboo related to the touch of a fetish belonging to a sick person, a member of a secret society belonging to one of the ethnic groups of the region. as a result, susceptibility to ebola was initially perceived to be restricted to this particular ethnic group, labelling ebola as an 'ethnic disease'. i decided to name this explanatory model of evd in forest guinea, the 'animist' model, not to further racialise this epidemic, but because it refers to the genies and fetishes that constitute principal aspects of the ancient religions of west africa and also because it describes a belief in a dual existence for all things-a physical, visible body and a psychic, invisible soul. according to a young pastor from macenta who i interviewed, and as confirmed by several other sources of key informants, the population of macenta initially attributed the origin of the disease (in this region at least) to a curse that was only affecting the kissi ethnic group because the first deaths solely affected people belonging to this ethnic group. here is what was stated: … on arrival with all the rumours we heard in conakry, i really did not believe in the beginning that it [the ebola virus disease] must be true because i thought it was an issue of the kissi (…) because it had started in macenta with the kissi, the first deaths were almost only kissi. so we thought it was something related to it … and so we, as toma, it was not going to touch us, it is like that at the beginning we perceived things (…) not something genetic, we thought about the fetishism and idolatry activities that people exercised and that can influence them in one way or another … the first rumour that was there, in macenta, the first death was the doctor who was dead in front of everyone's views. people said they have an idol called 'doma' and so when a person dies of that according to the tradition and according to what is done. and those who are on the thing [those who belong to the secret society of 'doma'] have no right to touch, to manipulate the corpse, or to see it otherwise they may die (…) and that, it existed before. it is a kind of secret society, so they have told us that it can certainly be that, that it is why they [the kissi] are just dying successively. according to these discourses, a health worker from guéckédou hospital (s ), who had gone to seek treatment at his friend's house at macenta hospital (s ), belonged, like him, to a secret initiation society called 'doma' which is also the name of a very powerful fetish; so powerful that it can cause a very fast death for its owner if it has been touched by someone else belonging to the same secret society. when the guéckédou health worker's body was moved, the doctor from macenta would have touched this fetish, idol, sacred object, often hidden in the owner's boubou (traditional clothing). by touching the sacred, the fetish got upset causing the brutal death of the director of macenta's hospital very soon after this event. at that point, in order to repair this transgression and calm the anger of the fetish, six more deaths must succeed each other to reach the symbolic number of seven. if the number of sudden and rapid deaths reaches eight, it means that the fetish is very powerful, and, as a result, seven additional deaths must occur to reach deaths to restore harmony and repair sacrilege. if we reach deaths, we must go to deaths before the disturbed order is restored and moreover that the stain is 'washed', and so on. since the first deaths of this second chain were indeed members of this kissi ethnic group (fig. . ), the 'animist' explanatory model of the disease was quite consistent with people's observations and gained legitimacy among the population at the expense of the biomedical discourse of the existence of evd. as the susceptibility of dying from ebola was initially and predominantly perceived as restricted to this particular ethnic group, no preventive measures were adopted by the non-kissi population of the region. among the kissi, the consequent epistemic dissonance between the public health logic and the transgression to be restored led between june and july twenty-six kissi-speaking villages in guéckedou prefecture to isolate themselves from ebola response, cutting bridges and felling trees to prevent vehicle access, and stoning intruding vehicles. because it is a disease of the social-of those who look after and visit others, and of those who attend funerals-there are of course many reasons why the ebola phenomenon was likely to be associated with sorcery. it is also not a coincidence that the triggering event, the transgression, in this explanatory model was attributed to medical doctors. as elite africans generally educated in european ways and relatively wealthy, this social group displays many characteristics of sorcerers (they lead a secluded life, do not share their gains, exchange abrupt greetings, eat large quantities of meat and eat alone). moreover, the intense preoccupation throughout this region with 'hidden evil in the world around you that finds dramatic expression in the clandestine activities of witches and the conspiracies of enemies' is exacerbated by tiny pathogens remaining largely invisible to our routine social practices, hence attracting suspicions of sorcery (fig. . ) . following the investigation of this 'animist' model in relation to the strong community resistance manifested in forest guinea, i interviewed a member of the riposte communication unit originating from macenta about the dandano case : yes, there is the specificity of dandano. (…) [in] dandano there was a great witch doctor who had gone to greet his counterpart witch doctor where there were a lot of cases. and that is where he got infected. he returned to dandano. three days later he developed the disease and died. afterwards, as he is a great, recognised witch doctor, people said to themselves, because he died, it was not ebola that killed him but his fetish that is taking revenge on him because it is a betrayal to leave one's domain to greet one's friend. maybe he went to spy on his friend and his friend hit him … well, there have been many versions. (…) among the old people who knew the drug he had, euh… his fetish, the grigri that he had, and that if it was his grigri who killed him, it means that all those who saw him, who saw his body, must also suffer. (…) [we could] see his dead body because he was not protected, because we had to wash him and there were medicines that had to be poured to annihilate his fetishes' power before burying him. so there must have been deaths, hence it was already premeditated. then there were deaths, as it was said, and they were successive deaths. that means there were deaths, two days, three days, so people put more anathema on what happened. and that is how dandano lived things. so there were deaths, we said it is the fetish that woke up because dandano is known as a village of powerful fetishes, that is known. (…) even all the sensitisation we do, we never stop in dandano on a manager, a notable, otherwise they can do something to you … so it is well recognised (…) dandano, is not where you have to go joking. (…) at the end, with a lot of deaths, a lot of funerals, they saw that no, it is not that [the fetish anger] anymore, and with the information here and there, it is ebola. and it is like that with all the negotiations (…). notably, these explanatory models are distinct from general beliefs about diseases and care techniques in the region. we cannot argue then that 'biomedicine' and 'kissi culture' are somehow distinct and opposed. chain of transmission according to the 'animist' cultural model. s and s are the two suspect cases as presented in the 'biomedical' chain of transmission (see fig. . ); the grey blocks are the kissi people of the 'animist' transmission chain these beliefs belong to the ideology of different sectors of the care system and exist independently of the illness of a subject. explanatory models are collected in response to a particular episode of illness in a given subject in a given sector and can evolve over time, depending on how the experience, knowledge and risk exposure of the concerned individual develop. this is precisely what has been reported to us and what has been observed in forest guinea. as the number of deceased progressed, and according to the religious and/or ethnic affiliation of the deceased, a new explanatory model was put in place as stated in this conversation: yes, at first it was said, when i was in conakry, since our country is predominantly muslim, it was said that it is a matter for christians since muslims do not eat apes. muslims do not eat the bat. it's only the foresters who eat that. and that's why this disease hits only the kissi and toma who are from the forest. so it's a kaf disease. -kaf? (séverine thys) -from unbelievers, pagans who do not know god. we call kaf, all those who do not believe in the god of muslims. this last extract particularly highlights the fact that these explanatory models are not fixed in time and space and are not impervious to each other either. indeed, the first health messages communicated to the population and built on the biomedical model were intensely focused on the need to avoid the consumption of 'bushmeat', especially wild animals identified as potential primary sources of contamination, namely monkeys and bats. the content of these messages gave birth to another popular model, in which the food taboos or eating habits observed by members affiliated to a certain religion allowed them to explain why this disease was affecting certain groups and not others. this quote also perfectly illustrates how popular discourses have integrated medical interpretations or public health messages. in the study conducted by bonwitt et al. about the local impact of the wild meat ban during the outbreak, all participants, irrespective of age or gender, were aware of wild mammals acting as a source of transmission for ebola. yet a confusion remained about which species in particular could transmit the ebola virus, which may be due to the content of public health messages that were inconsistent as regards the species shown to be potentially hazardous. messages are being absorbed, but in such chaos and fear, people process information according to their own worldview, according to the sources available to them, and following their personal experiences and instincts. furthermore, the criminalisation of wild meat consumption, which fuelled fears and rumours within communities, did entrench distrust towards outbreak responders and also exacerbated pre-existing tensions within villages, ethnicities and religions. following the kissi, it seemed that it was the muslim community that was hit by sudden and numerous deaths. to cope with this new upheaval, this new incomprehension, the operated explanatory model of these deaths' origin was, consequently, first that of a 'maraboutage': it started like that until a certain moment. and then it turned upside down. there have always been upheavals. it turned upside down, and instead of being weighed at a certain moment on the toma and the kissi, it was rather on the manyas, who are entirely, %, % even muslims. and so people started saying 'ha! that only attacks muslims, why not christians?'. so there has always been upheaval in all the procedures of this disease evolution. as noted by hewlett et al., 'patients, physicists, caregivers and local people in different parts of the world have cultural patterns for different diseases. providing care and appropriate treatment for a particular disease is often based on negotiation between these different models'. to be able to negotiate, it is necessary that each one, doctor and patient, partakes in the knowledge of the explanatory model of the other. while most health professionals rarely assume that people have and construct their own interpretation of the causal chain, my ethnographic observations presented in this chapter demonstrate that the a priori on which all interventions of sensitisation are based is not only incorrect, but also a source of blockages for the adoption of prescribed behaviours. this is because, to return to hewlett et al., 'people do not just follow the continuous thread of learning; they also develop an ability to articulate adherence to prescribed behaviours with the refusal of others, to cooperate at certain times and to show reluctance to others, inviting the analysis to move towards a sociology of compromise'. through the example of funerals, wilkinson and leach have also cast light on the presumption that the knowledge needed to stop the epidemic is held by public health experts and scientists, and not by local people. this very often leads to the development of protocols and procedures that completely negate the contribution of communities. this asymmetrical reflection between caregivers and care receivers, the structural violence that has cultivated inequalities in this region, the heterogeneity of experiences seen by the populations as fundamental contradictions between words and facts, the confidence and trust crisis since the 'demystification' programme initiated during sékou touré's time, and the traumas inflicted by a transgression of usages in the name of urgency and the exceptional nature of the ebola epidemic, are all realities that have fueled community reluctance and resistance. the late involvement of traditional healers, primarily consulted by guineans when experiencing illness, in the activities of the response in macenta, is another example of this asymmetry, which too often omits to acknowledge and relate to these other categories that support the social fabric, even if since alma ata in these stakeholders should no longer be on the margins of the health system. although the concept of explanatory models is not sufficient to explain all the failures of response in the context of guinea, or the bordering regions with sierra leone and liberia, nevertheless it allows to move past linear technical discussions of 'weak health systems' as the main reason for the scale of the disaster. the use of this conceptual framework for understanding popular interpretations of the origin of the disease and its transmission reveals the complex, historically rooted and multidimensional picture of the ebola crisis. several authors agree that, 'in any case, it is not a question of archaic beliefs or outlier depictions, but good answers -which can be called rational in this context -to a vital emergency situation, interpreted in the light of past and present experiences'. a better knowledge and comparison of these discourses and different cultural models of the disease, sometimes incorporated, sometimes hermetic, could nevertheless contribute considerably to the success of the fight against the epidemic, especially when it concerns the improvement of knowledge of the chains of disease transmission, the identification and understanding of the behaviours of local populations, and of the sources of denials and rumours. explanatory models proposed by the biomedical sciences are very often in competition and in contradiction with diagnoses made by traditional healers and especially with rumours involving divine punishments, breaches of prohibitions, the misdeeds of wizards or genies, or virologic warfare. if this 'animist' model is not identified nor recognised as making sense for others at the key moment, there will also be no negotiation and no understanding of the distances and proximities existing between the thought systems present in the concerned ecosystems. an anthropological approach remains essential to adapting this response to local realities. epelboin further argues that 'local models of causation regarding misfortune, often the most predominant, involve not only the virulence of the virus and human behaviour, but the evil actions of human and non-human individuals. the virologic model is then only one explanatory model among others, leaving the field open to all social, economic and political uses of misfortune'. following the re-emergence of this infectious disease of zoonotic origin in a whole new social ecosystem, a cross-sectoral research agenda, the so-called one health integrated approach, has finally emerged in the field of viral haemorrhagic fevers, also enabling the role of anthropology to be expanded to times of epidemic outbreak. until then, anthropologists were mandated to contribute to the adaptation and improvement of immediate public health interventions in relation to human-to-human transmission. yet, the growing interest of anthropologists in the interaction between humans and non-humans has made it possible to extend their research topic to the complex dynamics of the primary and secondary transmission of the virus. in addition, this anthropological interest has provided a new cross-cultural perspective on the movement of pathogens and has therefore improved knowledge about the mechanisms of emergence, propagation and amplification of a disease located at the interface between humans and wildlife. such was the role of almudena marí saéz and colleagues who, in a multidisciplinary team, conducted an ethnographic study in the village of the ebola epidemic's origin, the index case village, to better understand local social hunting practices and the relationships between bats and humans. however, the realm of the human-animal-disease interaction has been limited to 'natural versus cultural' domains and frequently conceived as a biological phenomenon in one health studies instead of a biocultural one integrating the social and cultural dimensions generated by human-animal relations. incorporating anthropology into one health approaches should provide a more nuanced and expanded account of the fluidity of bodies, categories and boundaries as drawn up by existing ethnographies on cattle in east and southern africa for example. epelboin et al. have stressed that, 'the anthropological approach in previous epidemics has confirmed that the urgency and severity of an epidemic must not prevent people from listening to them and thinking throughout the epidemic of taking into account indigenous codes, customs, knowledge, skills and beliefs'. by taking seriously the possibility that affected people in the places where we do research or implement control measures might not see things in the same way, we have to be willing to have our categories (such as culture/nature, human/animal, mind/body, male/female, caregivers/care receivers) unsettled, and to grapple with the practical implications of this for engagement in field sites, for knowledge-sharing and for the design of interventions, in the hope that such improvements might contribute to a future prevention of ebola and to public health policies more suitable to respond to people's basic needs. it also allows the affected people themselves to have a say in the matter. as philippe descola and other anthropologists have argued, on the basis of a comparative analysis of a wide range of ethnographic work across the continents, native classificatory systems usually offer a continuum, rather than sharp divisions, among humans and other animal species. indeed, human dispositions and behaviours are attributed not only to animals but also to spirits, monsters and artefacts, contrasting to modern western models, which generally see the categories of human and non-human as clearly defined and mutually exclusive. the ability to sense and avoid harmful environmental conditions is necessary for the survival of all living organisms and, as paul slovic has argued, 'humans have an additional capability that allows them to alter their environment as well as respond to it'. as regards the emerging violence in conservation as either against nature (e.g. culling bats) or in defence of it (e.g. rearranging landscapes within an inclusive 'one health' approach), james fairhead proposes that such violence is increasingly between 'the included' and 'rogues' in ways that transcend the nature/society binary. while the 'white', and african elites were seen by the affected population as 'antisocial' intruders or rogues, suspected of sorcery and using ebola as a tool for political manipulation, those involved in the struggle to address the ebola epidemic were not fighting just against the virus but also against the natural world that harboured it: the rogues which included villainous bats but moreover habitat destroyers, namely hunters, bushmeat traders and deforesters. these were the humans casted as the ones invading the habitat of the virus. since evd will be constantly reconceptualised, and because of new scientific discoveries (e.g. on natural reservoir, or vaccine development), control interventions must listen to and take into account popular perceptions as well as the socio-cultural and political context and their respective evolution. rumours must be identified and managed on a case-by-case basis without global generalisation that could reinforce misinterpretations on the assumption that ignorance alone generates these rumours, con-flicts, lack of trust and resistance. moreover, zoonotic epidemic fighters should follow macgregor's and waldman' recommendations by starting to think differently with and about animals and about species boundaries in order to generate novel ways of addressing zoonotic diseases, allowing for closer integration with people's own cultural norms and understandings of human-animal dynamics. and medicine ( ) ebola virus disease in guinea-update (situation as of aspects épidémiologiques de la maladie à virus ebola en guinée (décembre -avril emergence of zaire ebola virus disease in guinea investigating the zoonotic origin of the west african ebola epidemic zoonosis: prospects and challenges for in this outbreak story, a disease emerges in a remote location and spreads across a world highly connected by globalisation and air travel to threaten 'us all'-read the globally powerful north: see a entre science et fiction' contagious: cultures, carriers, and the outbreak narrative: wald priscilla the global focus on wildlife as a major contributor to emerging pathogens and infectious diseases in humans and domestic animals is due to reports which are not based on field, experimental or dedicated research but rather on surveys of literature and research regarding human immunodeficiency virus (hiv) and aids, severe acute respiratory syndrome (sars) and highly pathogenic avian influenza (hpai), all of which have an indirect wildlife link: r. kock, 'drivers of disease emergence and spread: is wildlife to blame on how and why 'bushmeat' hunting leads to the emergence of novel zoonotic pathogens see bushmeat hunting, deforestation, and prediction of zoonoses emergence uncovering zoonoses awareness in an emerging disease "hotspot"'. social science attempt of the zoonotic niche of evd, see contagious: cultures, carriers, and the outbreak narrative the term 'exceptionality' is borrowed from s. l. faye, 'l' "exceptionnalité" d'ebola et les "réticences" populaires en guinée-conakry. réflexions à partir d'une approche d'anthropologie symétrique to-plague-and-beyond-how-can-anthropologistsbest-engage-past-experience-to-prepare-for-new-epidemics. for the policy relevance of anthropological expertise and a (self-)critical reflection on ebola and on anthropological (and more broadly social scientific) engagements with humanitarian response, see a. menzel and a. schroven the term 'riposte' is the french name used to designate the official national mobilisation settled to respond to the evd crisis, structured into two poles, an inter-ministerial committee and a national coordination committee grouping together the international actors and the national non-governmental organisations; see m. fribault heterogeneities in the case fatality ratio in the west african ebola outbreak challenges in controlling the ebola outbreak in two prefectures in guinea: why did communities continue to resist? comparison of social resistance to ebola response in sierra leone and guinea suggests explanations lie in political configurations not culture understanding social resistance to the ebola response in the forest region of the republic of guinea: an anthropological perspective contagious: cultures, carriers, and the outbreak narrative zoonosis: prospects and challenges for medical anthropology the good, the bad and the ugly: framing debates on nature in a one health community understanding social resistance to the ebola response in the forest region of the republic of guinea: an anthropological perspective sustainability and contemporary man-nature divide: aspects of conflict and alienation on the visual ethnographic examination of the ebola zoonotic cycle transformed into tools of public health communication by the us cdc during the outbreak of medical anthropology and ebola in congo unintended consequences of the "bushmeat ban emergence of zaire ebola virus disease in guinea maladie à virus ebola: une zoonose orpheline?'. bulletin de l'académie vétérinaire de france inclusivity and the rogue bats and the war against "the invisible enemy about the natural reservoir for ebola virus see the evolution of ebola virus: insights from the - epidemic' a review of the role of food and the food system in the transmission and spread of ebolavirus mammalian biogeography and the ebola virus in africa for information on the 'bushmeat ban', see bonwitt et al., 'unintended consequences of the "bushmeat ban ebola virus disease epidemic emergence of zaire ebola virus disease in guinea'; world health organization, 'one year into the ebola epidemic: a deadly, tenacious and unforgiving virus caring for critically ill patients with ebola virus disease ebola-myths, realities, and structural violence'; and olival and hayman the threat to primates and other mammals from the bushmeat trade in africa, and how this threat could be diminished origins of major human infectious diseases'; centers for disease control and prevention take-a-semiotician-or-what-we-talk-aboutwhen-we-talk-about-bush-meat-by-adia-benton/. the kellogg institute on the feeling of disgust as a sentiment with powerful political valences, see also j. livingston, 'disgust, bodily aesthetics and the ethic of being human in botswana the anatomy of disgust world organisation for animal health the bushmeat trade: increased opportunities for transmission of zoonotic disease bushmeat crisis' is caused by the dual threats of wildlife extinctions and declining food and livelihood security of some of the poorest people on earth and whether the hunting of bushmeat is primarily an issue of biodiversity conservation or human livelihood, or both, varies according to perspective, place and over time; see unintended consequences of the "bushmeat ban impact of the ebola virus disease outbreak on market chains and trade of agricultural products in west africa'. food and agriculture organization of the united nations sending the right message: wild game and the west africa ebola outbreak bushmeat ban" in west africa during the - ebola virus disease epidemic'; p. richards, ebola: how a people's science helped end an epidemic les errances de la communication sur la maladie à virus ebola zoonotic semiotics: plague narratives and vanishing signs in madagascar' understanding social resistance to the ebola response in the forest region of the republic of guinea: an anthropological perspective one year on: why ebola is not yet over in guinea encyclopedia of medical anthropology : health and illness in the world's cultures extracts of the individual interview conducted with the pastor on the cultural and political role of initiation societies in the forest region and the related experiences of local citizens in relation to both the manding (often islamic) world to the north, and to the 'white' (often christian) colonial and neo-colonial order, see fairhead purity and danger, an analysis of concepts of pollution and taboo communication with rebellious communities during an outbreak of ebola virus disease in guinea: an anthropological approach understanding social resistance to the ebola response in the forest region of the republic of guinea: an anthropological perspective memories of the slave trade: ritual and historical imagination in sierra leone lifeworlds: essays in existential anthropology for more information about dandano village 'surrendering their sick and dead after being battered by the virus', see a. nossiter extracts of the individual interview conducted with a voluntary of the communication unit of macenta new therapeutic landscapes in africa: parental categories and practices in seeking infant health in republic of guinea extracts of the individual interview conducted with the pastor for similar narrative about muslim communities and food taboos regarding bats, see f. batty, 'reinventing "others" in a time of ebola unintended consequences of the "bushmeat ban extracts of the individual interview conducted with the pastor medical anthropology and ebola in congo: cultural models and humanistic care ebola en guinée: violences historiques et régimes de doute briefing: ebola-myths, realities, and structural violence l' "exceptionnalité" d'ebola et les "réticences" populaires en guinée-conakry ebola en guinée: violences historiques et régimes de doute'; wilkinson and leach traiter les corps comme des fagots' production sociale de l'indifférence en contexte ebola (guinée approche anthropologique de l'épidémie de fhv ebola en guinée conakry zoonosis: prospects and challenges for medical anthropology extending the "social": anthropological contributions to the study of viral haemorrhagic fevers investigating the zoonotic origin of the west african ebola epidemic views from many worlds: unsettling categories in interdisciplinary research on endemic zoonotic diseases animal spirits and mimetic affinities: the semiotics of intimacy in african human/animal identities annexe . contribution de l'anthropologie médicale à la lutte contre les épidémies de fièvres hémorragiques à virus ebola et marburg'. in world health organisation, Épidémies de fièvres hémorragiques à virus ebola et marburg: préparation, alerte, lutte et évaluation the morning after: anthropology and the ebola hangover beyond nature and culture biosecurity and the topologies of infected life: from borderlines to borderlands nature and society: anthropological perspectives by perception of risk inclusivity and the rogue bats and the war against "the invisible enemy views from many worlds acknowledgements i would like to thank tenin traoré, a guinean sociologist and consultant to who, and joseph kovoïgui, assistant coordinator of the philafrican mission and then consultant to who, for their commitment and engagement in the fight against ebola, their generosity, their knowledge, their experience and our fruitful collaboration in many respects. i would also like to thank the coordination team and the dps (prefectural health direction) of macenta for their welcome and sincere attention; goarn/who, antwerp institute of tropical medicine, and in particular prof. marleen boelaert for emotional, financial and logistical support; dr. alain epelboin for field preparation and numerous sharing with the francophone anthropological platform; and christos lynteris for his invitation to connect and exchange with the anglophone 'anthro-zoonoses' network and contribute to this timely collection. key: cord- -sw pdnh authors: aksyonov, konstantin; aksyonova, olga; antonova, anna; aksyonova, elena; ziomkovskaya, polina title: development of cloud-based microservices to decision support system date: - - journal: open source systems doi: . / - - - - _ sha: doc_id: cord_uid: sw pdnh intelligent systems of simulation become a key stage of the scheduling of companies and industries work. most of the existing decision support systems are desktop software. today there is a need to use durability, flexibility, availability and crossplatforming information technologies. the paper proposes the idea of working cloud based decision support system bpsim.web and this one consists of some set of services and tools. the model of the multiagent resources conversion process is considered. the process of the simulation model developing via bpsim.web is described. an example of the real process model is given. creation of simulation systems (sim) [ ] is one of the promising directions for the development of decision-making systems for business processes (bp), supply chains and logistics [ , ] , technological processes (for example metallurgy [ ] ). currently, the presence in sim of communities of interaction agents that are identified with decision makers (dm) is significant [ ] [ ] [ ] [ ] . currently, commercial simulation solutions based on the market (such as anylogic, aris, g ) are desktop applications. aris system allows you to create html-pages with the results of experiments and upload them to the internet. anylogic system is able to compile java applets with developed models and place them on the network. to start working with the model, it is necessary to fully download it to the user's electronic device, playing the simulation experiment of the model applet takes place on the user's device and requires significant computing resources. the analysis showed that the greatest functionality of sim of business processes is provided by anylogic and bpsim products. in the direction of service-oriented architecture, only g is developing. thus, the urgent task is to choose a dynamic model of a business process and build on its basis a web-service of simulation. comparative analysis of sim is presented in table . ) accounting for various types of resources [ , ] ; ) accounting for the status of operations and resources at specific times; ) accounting for the conflicts on common resources and means [ , ] ; ) modeling of discrete processes; ) accounting for complex resources (resource instances with properties, in the terminology of queuing systems -application (transaction)); ) application of a situational approach (the presence of a language for describing situations (a language for representing knowledge) and mechanisms for diagnosing situations and finding solutions (a logical inference mechanism according to the terminology of expert systems); ) implementation of intelligent agents (dm models); ) description of hierarchical processes. consider the following approaches and models of multi-agent systems and bp: ) model of a multi-agent process of resource conversion; ) sie-model a.u. filippovich; ) models of active and passive converters (apc) b.i. klebanov, i.m. moskalev. the dynamic model of multi-agent resource conversion processes (marcp) [ , ] is designed to model organizational and technical, bp and support of management decisions. the marcp model was developed on the basis of the following mathematical schemes: petri nets, queuing systems and system dynamics models. the key concept of the marcp model is a resource converter having the following structure: input, start, conversion, control, and output. "start-up" determines the moment the converter is started on the basis of: the state of the conversion process, input and output resources, control commands, means. at the time of launch, the conversion execution time is determined based on the parameters of the control command and available resource limitations. the marcp model has a hierarchical structure. agents manage the objects of the process based on the content of the knowledge base (kb). integrated situational, simulation, expert model a.u. filippovich (sie-model) is presented in [ ] . due to the fact that this model is focused on the problematic area of prepress processes (printing), some of its fragments will be described in terms of the marcp model. sie-model is presented in the form of several different levels, corresponding to the imitation, expert and situational presentation of information [ ] . the first level of the model is intended to describe the structure of the system. for this, a block is associated with each object (subject). blocks are interconnected by interaction channels. each block processes a transaction for a certain time and delays it for a time determined by the intensity. we formulate the following conclusions: . the sie-model can serve as the basis for creating a multi-agent bp model. . the sie-model has the following advantages: apparatus/ mechanism for diagnosing situations; a combination of simulation, expert and situational approaches. . the sie model does not satisfy the requirements of the bp model: the presence of a dm (agent) model; problem orientation to business processes. in the work of i.m. moskalev, b.i. klebanov [ ] [ ] [ ] presents a mathematical model of resource conversion process, the specificity of which is the allocation of active and passive converters (apc). in this model, the vertices of graph x are formed by passive transducers, active transformers, stock of instruments and resource storages, and many arcs are represented by resource and information flows, flows to funds. the model of active and passive converters is focused on solving production-planning problems and based on scheduling theory. this model has not worked the possibility of implementing intelligent agents (models of dm) with a production knowledge base, as well as the implementation of a language for describing mechanisms and situations for diagnosing situations and finding solutions. the results of the analysis of the considered approaches and models of dynamic modeling of situations are given in table . as follows from the table, all the requirements of the multi-agent business process model are met by the marcp model. as the theoretical basis of the implemented method you can use the sie-model, the advantage of which is the study of integration issues of simulation, expert and situational modeling. to implement the simulation modeling service, it was decided to use the marcp concept. the simulation modeling service is based on asp.net core technology in the c# programming language. asp.net core is a cross-platform, high-performance, open-source environment for building modern, cloud-based, internet-connected applications. asp.net core provides the following benefits: • a single solution for creating a web user interface and web api. • integration of modern client platforms and development workflows. • cloud-based configuration system based on the environment. • built-in dependency injection. • simplified, high-performance, modular http request pipeline. • ability to host in iis, nginx, apache, docker or in your own process. • parallel version control of the application focused on .net core. • tools that simplify the process of modern web development. • ability to build and run on windows, macos and linux. • open source and community oriented. asp.net core comes fully in nuget packages. using nuget packages allows you to optimize applications to include only the necessary dependencies. asp.net core .x applications targeting .net core require only one nuget package. due to the small size of the application's contact area, benefits such as higher security, minimal maintenance and improved performance are available. figure shows the architecture of the simulation service. the service manages entities: models, simulation processes, and execution reports. it receives commands from the integration interface (api) and, depending on the command, receives or stores data in the database, performs internal transformations and calculations, and starts modeling. the simulation modeling service has an interaction and integration interface in the form of cross-platform http webapi. the http protocol allows you to not only provide web pages. the simulation service describes a rest-style interface (representation-alstatetransfer). rest is an architectural style of software that defines the interaction of components of a distributed application on a network or the integration of multiple applications. one of the requirements of rest is a unified programming interface for all entities with which the web service works. the crud (createreadupdatedelete) interface of operations is described using the http request verb (get, post, put, etc.), the names of entities and, if necessary, the identifiers of these entities. consider some possible queries on the models that the simulation service handles: here, all commands to the service have a single prefix "model". if there is an appeal to many models at once -to get all models, add a new one to many models -only the prefix is used. if you need an action associated with a specific model -get, change, delete -you must continue the request url with the model identifier. an http verb is used to determine the action to be performed on the request object. in the rest style, a special approach is laid down for requests that work with some service entities at once. this is how the imitation task is created: the url describes the relationship of one particular model with its tasks. a service should take one from the model domain with a specific identifier and refer to its many tasks for simulation. the http verb indicates that a new task should be added to this set. the receipt of all tasks for simulating a model is described in a similar way: • get model/{id}/task; however, work with many tasks for imitation can be carried out not in the context of any particular model, but immediately with the whole set. the following methods are used for this: • get task -get all simulation tasks; • get task/{id} -get information on a simulation task; • delete task/{id} -stop simulation. for each task, many reports are generated. the task report is strongly related to the essence of the task itself. therefore, deleting a task deletes all reports for this task. the reports do not have unique identifiers and can be received by the whole set. mongodb was chosen as the data storage system of the sim service. mongodb -dbms that uses a document-oriented data model. this allows mon-godb to carry out crud operations very quickly, to be more prepared for changes in stored data and to be more understandable for the user. it is an open-source project that is written in c ++. all libraries and drivers for programming languages and platforms are also available in the public domain. the storage method in mongodb is similar to json (javascriptobjectnotation), although formally json is not used. mongodb uses a format called bson (binaryj-son) for storage. the bson format has a certain structure and stores some additional information about a specific key and value pair (for example, data type and hash). therefore, usually an object in bson takes up a bit more space than json. however, bson allows you to work with data faster: faster search and processing. in mongodb, each model is stored as a document. this document stores the entire structure of the model. figure shows the simulation model saved by the service in mongodb. all models are saved in the models collection. the structure of the model object itself, in addition to a unique identifier, consists of the following fields: • name -model name. • resources -an object in the form of a key-value, where key is the name of the resource, and value is the default value of the resource. • orders -the key-value object, where the key is the name of the request, and the value is an array of the names of the fields of the request. • nodes -a key-value object, where key is the name of the node, and value is a complex object that describes either the agent or the operation. the "agent" node has an array of global rules (globalrules) and a lot of knowledge (knowledges). an array of rules stores the same rule entities as operations. knowledge is a complex key-value object, where key is the name of knowledge, and value is an array of rules that describe this knowledge. as mentioned earlier, simulation tasks also end up in the database. they are saved to the tasks collection. the task consists of the following fields. • modelid -identifier of the model with which the task is associated. • modelrevision -version of the model by which the task was created. if the model has changed between the creation of the task and its immediate imitation, unforeseen incidents may occur. • state -task state. • subjectarea -the subject area with which the simulation will occur. • configuration -the configuration with which the simulation will be launched. it stores information about when to stop the simulation, as well as some other information. • reports -an array of reports that were created for this task. the task may be in one of the states. this is determined by the life cycle of the task. the states are as follows: • -open -the task is created and ready to simulate; • -scheduled -the task has already been planned, but has not yet been simulated (the engine is being prepared, objects are being created, etc.); • -pending -the task is in simulation; • -failed -the task failed; • -successful -task completed successfully; • -aborted -the task was interrupted; • -deleted -the task was deleted before it was taken into imitation. the task created as a result of the http request does not begin to be simulated immediately. it is saved to the database and enters the task queue. service within a given period of time will check the queue for tasks. if there are no tasks for imitation, the cycle will be repeated after some time. if there is a task in the queue, imitation begins on it. the simulator can simulate several tasks in parallel. the degree of parallelism is not strictly defined and can be configured by the service settings. if the simulator does not have free resources to simulate, he does not look at the queue until they appear. as soon as some task completes the simulation, the simulator looks through the queue for new tasks. when the task falls into the simulation, it saves to the database with the status (open). as soon as the simulator takes the task from the queue, it is set to status (scheduled). this is necessary so that a parallel process or another instance of the service does not begin to simulate the same task. when the simulator finishes preparations, he will begin the simulation, setting the task status (pending). upon successful completion of the task, it will receive the status (successful), and if an error occurs - (failed). the task can be canceled by a special team. in this case, she will have the status (aborted) or (delete). figure shows the pattern of product movement between the hot rolling mill lpc- and the cold rolling mill lpc- . the task is to recreate the production process of sheet steel coils and conduct a series of studies in which it is necessary to evaluate a set of key parameters within three -h working days. the parameters are as follows: . the minimum number of slabs in the warehouse of cast slabs at the beginning of the simulation, providing a continuous supply of slabs every three minutes. . the current, minimum, maximum and average number of objects in each warehouse of the system during the simulation time. . the load of all units in percent during the simulation time and the current load. to create such a simulation model, you need to use the post/api/v /model service interface and put a json object describing model in the http request body. figure shows a small part of this object. you can see the first nodes of the model on it: two operations "slab store" and "batch generator" and the beginning of the description of the "in bake" agent. nine experiments were conducted with the model. in each experiment, the minimum value of the slabs in the warehouse was changed. table presents a partial result of the experiments, including the minimum value of slabs in the warehouse, the final output of the model and the waiting time-the total idle time of the furnaces. according to the results of the experiments, it was decided that the th experiment was the best. starting from it, a continuous flow of slabs in workshops was obtained with the minimum quantity in the warehouse, the best effect was obtained. the data obtained in the course of this work made it possible to analyze the current development of business process simulation systems (such as anylogic, aris, bpsim, g ) and highlight the requirements for a new system oriented to work on the internet. a comparative analysis of the existing dynamic bp models was carried out and the model of the multi-agent resource conversion process was taken as a basis. a prototype webservice for bp simulation bpsim.web was developed. the web service has been tested in solving the problem of analyzing the processes of two workshops. cloud technology in simulation studies, gpss cloud project analysis of position optimization method applicability in supply chain management problem theoretical and technological foundations of complex objects proactive monitoring management and control on design of domain-specific query language for the metallurgical industry the architecture of the multi-agent resource conversion processes analysis of the electric arc furnace workshop logistic processes using multiagent simulation agents and data mining interaction intelligent agent: theory and practice reengineering the corporation: a manifesto for business revolutions project scheduling with multiple modes: a comparison of exact algorithms simulation tool based on a memetic algorithm to solve a real instance of a dynamic tsp a model of co-evolution in multi-agent system integration of situational, simulation and expert modeling systems. publisher "ooo elix+ system for analysis and optimization of resource conversion processes the principles of multi-agent models of development based on the needs of the agents bases of imitation model of artificial society construction accounting of the agents' needs recursion the reported study was funded by rfbr according to the research project № - - . key: cord- -nu ok w authors: varshney, deeksha; ekbal, asif; nagaraja, ganesh prasad; tiwari, mrigank; gopinath, abhijith athreya mysore; bhattacharyya, pushpak title: natural language generation using transformer network in an open-domain setting date: - - journal: natural language processing and information systems doi: . / - - - - _ sha: doc_id: cord_uid: nu ok w prior works on dialog generation focus on task-oriented setting and utilize multi-turn conversational utterance-response pairs. however, natural language generation (nlg) in the open-domain environment is more challenging. the conversations in an open-domain chit-chat model are mostly single-turn in nature. current methods used for modeling single-turn conversations often fail to generate contextually relevant responses for a large dataset. in our work, we develop a transformer-based method for natural language generation (nlg) in an open-domain setting. experiments on the utterance-response pairs show improvement over the baselines, both in terms of quantitative measures like bleu and rouge and human evaluation metrics like fluency and adequacy. conversational systems are some of the most important advancements in the area of artificial intelligence (ai). in conversational ai, dialogue systems can be either an open-domain chit-chat model or a task-specific goal-oriented model. task-specific systems focus on particular tasks such as flight or hotel booking, providing technical support to users, and answering non-creative queries. these systems try to generate a response by maximizing an expected reward. in contrast, an open-domain dialog system operates in a non-goal driven casual environment and responds to the all kinds of questions. the realization of rewards is not straightforward in these cases, as there are many factors to model in. aspects such as understanding the dialog context, acknowledging user's personal preferences, and other external factors such as time, weather, and current events need consideration at each dialog step. in recent times, there has been a trend towards building end-to-end dialog systems such as chat-bots which can easily mimic human conversations. [ , , ] developed systems using deep neural networks by training them on a large amount of multi-turn conversational data. virtual assistants in open-domain settings usually utilize single-turn conversations for training the models. chitchat bots in such situations can help humans to interact with machines using natural language, thereby allowing humans to express their emotional states. in dialogue systems, generating relevant, diverse, and coherent responses is essential for robustness and practical usages. generative models tend to generate shorter, inappropriate responses to some questions. the responses range from invalid sentences to generic ones like "i don't know". the reasons for these issues include inefficiency of models in capturing long-range dependencies, generation of a large number of out-of-vocabulary (oov) words, and limitations of the maximum likelihood objective functions for training these models. transformer models have become an essential part of most of the state-of-the-art architectures in several natural language processing (nlp) applications. results show that these models capture long-range dependencies efficiently, replacing gated recurrent neural network models in many situations. in this paper, we propose an efficient end-to-end architecture based on the transformer network for natural language generation (nlg) in an open-domain dialogue system. the proposed model can maximize contextual relevancy and diversity in generated responses. our research reported here contributes in three ways: (i) we build an efficient end-to-end neural architecture for a chit-chat dialogue system, capable of generating contextually consistent and diverse responses; (ii) we create a singleturn conversational dataset with chit-chat type conversations on several topics between a human and a virtual assistant; and (iii) empirical analysis shows that our proposed model can improve the generation process when trained with enough data in comparison to the traditional methods like retrieval-based and neural translation-based. conversational artificial intelligence (ai) is currently one of the most challenging problems of artificial intelligence. developing dialog systems that can interact with humans logically and can engage them in having long-term conversations has captured the attention of many ai researchers. in general, dialog systems are mainly of two types -task-oriented dialog systems and open-domain dialog systems. task-oriented dialog systems converse with the users to complete a specific task such as assisting customers to book a ticket or online shopping. on the other hand, an open-domain dialog system can help users to share information, ask questions, and develop social etiquette's through a series of conversations. early works in this area were typically rule-based or learning-based methods [ , , , ] . rule-based methods often require human experts to form rules for training the system, whereas learning-based methods learn from a specific algorithm, which makes it less flexible to adapt to the other domains. data from various social media platforms like twitter, reddit, and other community question-answering (cqa) platforms have provided us with a large number of human-to-human conversations. data-driven approaches developed by [ , ] can be used to handle such problems. retrieval based methods [ ] generate a suitable response from a predefined set of candidate responses by ranking them in the order of similarity (e.g., by matching the number of common words) against the input sentence. the selection of a random response from a set of predefined responses makes them static and repetitive. [ ] builds a system based on phrase-based statistical machine translation to exploit single turn conversations. [ ] presented a deep learning-based method for retrieval-based systems. a brief review of these methods is presented by [ ] . lately, generation based models have become quite popular. [ , , , ] presented several generative models based on neural network for building efficient conversational dialog systems. moreover, several other techniques, for instance generative adversarial network (gan) [ , ] and conditional variational autoencoder (cvae) [ , , , , , ] are also implemented for dialog generation. conversations generated from retrieval-based methods are highly fluent, grammatically correct, and are of good quality as compared to dialogues generated from the generative methods. their high-quality performance is subjected to the availability of an extensive repository of human-human interactions. however, responses generated by neural generative models are random in nature but often lack grammatical correctness. techniques that can combine the power of both retrieval-based methods and generative methods can be adapted in such situations. on the whole hybrid methods [ , , , ] first find some relevant responses using retrieval techniques and then leverages them to generate contextually relevant responses in the next stage. in this paper, we propose a novel method for building an efficient virtual assistant using single-turn open-domain conversational data. we use a self-attention based transformer model, instead of rnn based models to get the representation of our input sequences. we observe that our method can generate more diverse and relevant responses. our goal is to generate contextually relevant responses for single-turn conversations. given an input sequence of utterance u = u , u , ..., u n composed of n words we try to generate a target response y = y , y , ..., y m . we use pre-trained glove [ ] embeddings to initialize the word vectors. glove utilizes two main methods from literature to build its vectors: global matrix factorization and local context window methods. the glove model is trained on the non-zero entries of a global word to word co-occurrence matrix, which computes how frequently two words can occur together in a given corpus. the embeddings used in our model are trained on common crawl dataset with b tokens and . m vocab. we use -dimensional sized vectors. we formulate our task of response generation as a machine translation problem. we define two baseline models based on deep learning techniques to conduct our experiments. first, we build a neural sequence to sequence model [ ] based on bi-directional long short term memory (bi-lstm) [ ] cells. the second model utilizes the attention mechanism [ ] to align input and output sequences. we train these models using the glove word embeddings as input features. to build our first baseline, we use a neural encoder-decoder [ ] model. the encoder, which contains rnn cells, converts the input sequence into a context vector. the context vector is an abstract representation of the entire input sequence. the context vector forms the input for a second rnn based decoder, which learns to output the target sequence one word at a time. our second baseline uses an attention layer [ ] between the encoder and decoder, which helps in deciding which words to focus on the input sequence in order to predict the next word correctly. the third model, which is our proposed method, is based on the transformer network architecture [ ] . we use glove word embeddings as input features for our proposed model. we develop the transformer encoder as described in [ ] to obtain the representation of the input sequence and the transformer decoder to generate the target response. figure shows the proposed architecture. the input to the transformer encoder is both the embedding, e, of the current word, e(u n ), as well as positional encoding pe(n) of the nth word: there are a total of n x identical layers in a transformer encoder. each layer contains two sub-layers -a multi-head attention layer and a position-wise feedforward layer. we encode the input utterances and target responses of our dataset using multi-head self-attention. the second layer performs linear transformation over the outputs from the first sub-layer. a residual connection is applied to each of the two sub-layers, followed by layer normalization. the following equations represent the layers: where m is the hidden state returned by the first layer of multi-head attention and f is the representation of the input utterance obtained after the first feed forward layer. the above steps are repeated for the remaining layers: where n = , ..., n x . we use c to denote the final representation of the input utterance obtained at n x -th layer: similarly, for decoding the responses, we use the transformer decoder. there are n y identical layers in the decoder as well. the encoder and decoder layers are quite similar to each other except that now the decoder layer has two multihead attention layers to perform self-attention and encoder-decoder attention, respectively. r y = [y , ..., y m ] y m = e(y m ) + p e(m) ( ) to make prediction of the next word, we use softmax to obtain the words probabilities decoded by the decoder. in this section, we present the details of the datasets used in our experiments, along with a detailed overview of the experimental settings. our dataset comprises of single-turn conversations from ten different domains -data about user, competitors, emotion, emergency, greetings, about bixby, entertainment, sensitive, device, and event. professional annotators with a linguistics background and relevant expertise created this dataset. the total dataset comprises of , utterance and response pairs with an average of . and . words for utterance and response, respectively. we first split the data into a train and test set in a : ratio. we then use % of the training data for preparing the validation set. the dataset details are given in table . some examples from the dataset are shown in table . we use two different types of models for our experiments -recurrent and transformer-based sequence-to-sequence generative models. all data loading, model implementations, and evaluation were done using the opennmt [ ] as the code framework. we train a seq seq model where the encoder and decoder are parameterized as lstms [ ] . we also experiment with the seq seq model with an attention mechanism [ ] between the decoder and the encoder outputs. the encoder and decoder lstms have layers with -dimensional hidden states with a dropout rate of . . the layers of both encoder and decoder are set to with -dimensional hidden states with a dropout of . . there are multihead attention heads and nodes in the feed-forward hidden layers. the dimension of word embedding is empirically set to . we use adam [ ] for optimization. when decoding the responses, the beam size is set to . automatic evaluation: we use the standard metrics like bleu [ ] , rouge [ ] and perplexity for the automatic evaluation of our models. perplexity is reported on the generated responses from the validation set. lower perplexity indicates better performance of the models. bleu and rouge measure the ngram overlap between a generated response and a gold response. higher bleu and rouge scores indicate better performance. to qualitatively evaluate our models, we perform human evaluation on the generated responses. we sample random responses from our test set for the human evaluation. given an input utterance, target response, and predicted response triplet, two experts with post-graduate exposure were asked to evaluate the predicted responses based on the given two criteria: . fluency: the predicted response is fluent in terms of the grammar. . adequacy: the predicted response is contextually relevant to the given utterance. we measure fluency and adequacy on a - scale with ' ' indicating an incomplete or incorrect response, ' ' indicating acceptable responses and ' ' indicating a perfect response. to measure the inter-annotator agreement, we compute the fleiss kappa [ ] score. we obtained a kappa score of . for fluency and a score of . for adequacy denoting "good agreement. in this section we report the results for all our experiments. the first two experiments (seq seq & seq seq attn) are conducted with our baseline models. our third experiment (c.f fig. ) is carried out on our proposed model using word embeddings as the input sequences. table and table show the automatic and manual evaluation results for both the baseline and the proposed model. our proposed model has lower perplexity and higher bleu and rouge scores than the baselines. the improvement in each model is statistically significant compared to the other models . for all the evaluation metrics, seq seq attn has the highest score among the baselines, and our model outperforms those scores by a decent margin. for adequacy, we find that our seq seq model achieves the highest score of . among the baseline models. our proposed model outperforms the baselines with a score of . . for fluency, we observe that the responses generated by all the models are quite fluent in general. to observe our results in more details, we perform an error analysis on the predicted response. in table as seen in the example, the predicted response would not be the best fit reply to the utterance "you are online" as the response falls out of context for the given utterance. in this paper, we propose an effective model for response generation using singleturn conversations. firstly, we created a large single-turn conversational dataset, and then built a transformer-based framework to model the short-turn conversations effectively. empirical evaluation, in terms of both automatic and humanbased metrics, shows encouraging performance. in qualitative and quantitative analyses of the generated responses, we observed the predicted responses to be highly relevant in terms of context, but also observed some in-corrections as discussed in our results and analysis section. overall we observed that our proposed model attains improved performance when compared with the baseline results. in the future, apart from improving the architectural designs and training methodologies, we look forward to evaluating our models on a much larger dataset of single-turn conversation. neural machine translation by jointly learning to align and translate deep retrieval-based dialogue systems: a short review variational autoregressive decoder for neural response generation measuring nominal scale agreement among many raters long short-term memory an information retrieval approach to short text conversation generating informative responses with controlled sentence function adam: a method for stochastic optimization opennmt: open-source toolkit for neural machine translation adversarial learning for neural dialogue generation rouge: a package for automatic evaluation of summaries njfun-a reinforcement learning spoken dialogue system reinforcement learning of questionanswering dialogue policies for virtual museum guides bleu: a method for automatic evaluation of machine translation glove: global vectors for word representation data-driven response generation in social media a survey of statistical user simulation techniques for reinforcement-learning of dialogue management strategies a hierarchical latent variable encoder-decoder model for generating dialogues neural responding machine for short-text conversation improving variational encoder-decoders in dialogue generation an ensemble of retrieval-based and generation-based human-computer conversation systems a neural network approach to context-sensitive generation of conversational responses sequence to sequence learning with neural networks attention is all you need a neural conversational model the generalization of student's' problem when several different population variances are involved retrieve and refine: improved sequence generation models for dialogue partially observable markov decision processes for spoken dialog systems diversity-promoting gan: a cross-entropy based generative adversarial network for diversified text generation learning to respond with deep neural networks for retrieval-based human-computer conversation system a hybrid retrieval-generation neural conversation model unsupervised discrete sentence representation learning for interpretable neural dialog generation learning discourse-level diversity for neural dialog models using conditional variational autoencoders the design and implementation of xiaoice, an empathetic social chatbot acknowledgement. the research reported in this paper is an outcome of the project "dynamic natural language response to task-oriented user utterances", supported by samsung research india, bangalore. key: cord- -echxuw authors: modarresi, kourosh title: detecting the most insightful parts of documents using a regularized attention-based model date: - - journal: computational science - iccs doi: . / - - - - _ sha: doc_id: cord_uid: echxuw every individual text or document is generated for specific purpose(s). sometime, the text is deployed to convey a specific message about an event or a product. other occasions, it may be communicating a scientific breakthrough, development or new model and so on. given any specific objective, the creators and the users of documents may like to know which part(s) of the documents are more influential in conveying their specific messages or achieving their objectives. understanding which parts of a document has more impact on the viewer’s perception would allow the content creators to design more effective content. detecting the more impactful parts of a content would help content users, such as advertisers, to concentrate their efforts more on those parts of the content and thus to avoid spending resources on the rest of the document. this work uses a regularized attention-based method to detect the most influential part(s) of any given document or text. the model uses an encoder-decoder architecture based on attention-based decoder with regularization applied to the corresponding weights. the main purpose of nlp (natural language processing) and nlu (natural language understanding) is to understand the language. more specifically, they are focused on not just to see the context of text but also to see how human uses the language in daily life. thus, among other ways of utilizing this, we could provide an optimal online experience addressing needs of users' digital experience. language processing and understanding is much more complex than many other applications in machine learning such as image classification as nlp and nlu involve deeper context analysis than other machine learning applications. this paper is written as a short paper and focuses on explaining only the parts that are contribution of this paper to the state-of-the art. thus, this paper does not describe the state-of-the-art works in details and uses those works [ , , , , , , , , , ] to build its model as a modification and extension of the state of the art. therefore, a comprehensive set of reference works have been added for anyone interested in learning more details of the previous state of the art research [ , , , , , , , - , - , , , , , ] . deep learning has become a main model in natural language processing applications [ , , , , , , , , , - , , , ] . among deep learning models, often rnn-based models like lstm and gru have been deployed for text analysis [ , , , , , - , , , , ] . though, modified version of rnn like lstm and gru have been improvement over rnn (recurrent neural networks) in dealing with vanishing gradients and long-term memory loss, still they suffer from many deficiencies. as a specific example, a rnn-based encoder-decoder architecture uses the encoded vector (feature vector), computed at the end of encoder, as the input to the decoder and uses this vector as a compressed representation of all the data and information from encoder (input). this ignores the possibility of looking at all previous sequences of the encoder and thus suffers from information bottleneck leading to low precision, especially for texts of medium or long sequences. to address this problem, global attention-based model [ , ] where each of the encoder sequence uses all of the encoder sequences. figure shows an attention-based model. where i = :n is the encoder sequences and, t = :m represents the decoder sequences. each of the encoder states looks into the data from all the encoder sequences with specific attention measured by the weights. each weight, w ti , indicates the attention decoder network t pays for the encoder network i. these weights are dependent on the previous decoder and output states and present encoder state as shown in fig. . given the complexity of these dependencies, a neural network model is used to compute these weights. two layers ( ) of fully connected layers and relu activation function is used. where h is the state of the encoder networks, s t− is the previous state of the decoder and v t− is the previous decoder output. also, w t is the weights of the encoder state t. since w t are the output from softmax function, then, this section overviews of the contribution of this paper and explains the extension made over the state-of-the-art model. a major point of attention for many texts related analysis is to determine which part(s) of the input text has had more impact in determining the output. he length of input text could be very long combining of potentially hundreds and thousands of words or sequences, i.e., n could be very large number. thus, there are many weights (w ti ) in determining any part of output v t , and also since many of these weights are correlated, it's difficult to determine the significance of any input sequence in computing any output sequence v t . to make these dependencies clearer and to recognize the most significant input sequences for any output sequence, we apply a zero-norm penalty to make the corresponding weight vector to become a sparse vector. to achieve the desired sparsity, zero-norm (l ) is applied to make any corresponding w t vector very sparse as the penalty leads to minimization of the number of non-zero entries in w t . the process is implemented by imposing the constraint of, since l is computationally intractable, we could use surrogate norms such as l norm or euclidean norm, l . to impose sparsity, the l norm, lasso [ , , , , ] is used in this work, or, as the penalty function to enforce sparsity on the weight vectors. this penalty, β w t , is the first extension to the attention model [ , ] . here, β is the regularization parameter which is set as a hyperparameter where its value is set before learning. higher constraint leads to higher sparsity with higher added regularization biased error and lower values of the regularization parameter leads to lower sparsity and lesser regularization bias. the main goal of this work is to find out which parts of encoder sequences are most critical in determining and computing any output. the output could be a word, a sentence or any other subsequence. the goal is critical especially in application such as machine translation, image captioning, sentiment analysis, topic modeling and predictive modeling such as time series analysis and prediction. to add another layer of regularization, this work imposes embedding error penalty to the objective function (usually, cross entropy). this added penalty also helps to address the "coverage problem" (the phenomenon of often observed dropping or frequently repeating words --or any other subsequence --by the network). the embedding regularization is, α embedding error ( ) input to any model has to be a number and hence the raw input of words or text sequence needs to be transformed to continuous numbers. this is done by using one-hot encoding of the words and then using embedding as shown in fig. . whereu i is the raw input text, is the one-hot encoding representation of the raw input and u i is the embedding of the i-th input or sequence. also, α is the regularization parameter. the idea of embedding is based on that embedding should preserve word similarities, i.e., the words that are synonyms before embedding, should remain synonyms after embedding. using this concept of embedding, the scaled embedding error is, or, after scaling the embedding error, which could be re-written, using a regularization parameter (α), as, where l is the measure or metric of similarity of words representations. here, for all similarity measures, both euclidean norm and cosine similarity (dissimilarity) have been used. in this work, the embedding error using the euclidean norm is used, alternatively, we could include the embedding error of the output sequence in eq. ( ) . when the input sequence (or the dictionary) is too long, to prevent high computational complexity of computing similarity of each specific word with all other words, we choose a random (uniform) sample of the input sequences to compute the embedding error. the regularization parameter, α, is computed using cross validation [ ] [ ] [ ] [ ] [ ] [ ] . alternatively, adaptive regularization parameters [ , ] could be used. this model was applied on wikipedia datasets for english-german translation (one-way translation) with sentences. the idea was to determine which specific input word (in english) is the most important one for the corresponding german translation. the results were often an almost diagonal weight matrix, with few non-zero off diagonal entries, indicating the significance of the corresponding word(s) in the original language (english). since the model is an unsupervised approach, it's hard to evaluate its performance without using domain knowledge. the next step in this work would be to develop a unified and interpretable metric for automatic testing and evaluation of the model without using any domain knowledge and also to apply the model to other applications such as sentiment analysis. inverse problems: principles and applications in geophysics, technology, and medicine polosukhin: attention is all you need domain adaptation via pseudo in-domain data selection multiple object recognition with visual attention neural machine translation by jointly learning to align and translate the dropout learning algorithm deep learning an unsupervised feature learning nips workshop nesta, a fast and accurate first-order method for sparse recovery learning long-term de-pendencies with gradient descent is difficult a neural probabilistic language model theano: a cpu and gpu math expression compiler audio chord recognition with recurrent neural networks a singular value thresholding algorithm for matrix completion exact matrix completion via convex optimization compressive sampling long short-term memory-networks for machine reading learning phrase representations using rnn encoder-decoder for statistical machine translation framewise phoneme classification with bidirectional lstm and other neural network architectures generating sequences with recurrent neural networks the elements of statistical learning; data mining, inference and prediction handwritten digit recognition via deformable prototypes gene shaving' as a method for identifying distinct sets of genes with similar expression patterns matrix completion via iterative soft-thresholded svd package 'impute'. cran multilingual distributed representations without word alignment advances in natural language processing long short-term memory gradient flow in recurrent nets: the difficulty of learning long-term dependencies regularization for applied inverse and ill-posed problems compositional attention networks for machine reasoning two case studies in the application of principal component principal component analysis rotation of principal components: choice of normalization constraints a modified principal component technique based on the lasso recurrent continuous translation models statistical machine translation structured attention networks statistical phrase-based translation learning phrase representations using rnn encoder-decoder for statistical machine translation conditional random fields: probabilistic models for segmenting and labeling sequence data neural networks: tricks of the trade a structured self-attentive sentence embedding effective approaches to attention-based neural machine translation learning to recognize features of valid textual entailments natural logic for textual inference encyclopedia of language & linguistics introduction to information retrieval the stanford corenlp natural language processing toolkit. computer science, acl computational linguistics and deep learning differentiating language usage through topic models effective approaches to attention based neural machine translation application of dnn for modern data with two examples: recommender systems & user recognition. deep learning summit standardization of featureless variables for machine learning models using natural language processing generalized variable conversion using k-means clustering and web scraping an efficient deep learning model for recommender systems effectiveness of representation learning for the analysis of human behavior an evaluation metric for content providing models, recommendation systems, and online campaigns combined loss function for deep convolutional neural networks a randomized algorithm for the selection of regularization parameter. inverse problem symposium a local regularization method using multiple regularization levels a decomposable attention model on the difficulty of training recurrent neural networks. in: icml on the difficulty of training recurrent neural networks how to construct deep recurrent neural networks fast curvature matrix-vector products for second-order gradient descent bidirectional recurrent neural networks continuous space translation models for phrase-based statistical machine translation continuous space language models for statistical machine translation sequence to sequence learning with neural networks google's neural machine translation system: bridging the gap between human and machine translation adadelta: an adaptive learning rate method key: cord- -iioqkydg authors: zhong, qi; zhang, leo yu; zhang, jun; gao, longxiang; xiang, yong title: protecting ip of deep neural networks with watermarking: a new label helps date: - - journal: advances in knowledge discovery and data mining doi: . / - - - - _ sha: doc_id: cord_uid: iioqkydg deep neural network (dnn) models have shown great success in almost every artificial area. it is a non-trivial task to build a good dnn model. nowadays, various mlaas providers have launched their cloud services, which trains dnn models for users. once they are released, driven by potential monetary profit, the models may be duplicated, resold, or redistributed by adversaries, including greedy service providers themselves. to mitigate this threat, in this paper, we propose an innovative framework to protect the intellectual property of deep learning models, that is, watermarking the model by adding a new label to crafted key samples during training. the intuition comes from the fact that, compared with existing dnn watermarking methods, adding a new label will not twist the original decision boundary but can help the model learn the features of key samples better. we implement a prototype of our framework and evaluate the performance under three different benchmark datasets, and investigate the relationship between model accuracy, perturbation strength, and key samples’ length. extensive experimental results show that, compared with the existing schemes, the proposed method performs better under small perturbation strength or short key samples’ length in terms of classification accuracy and ownership verification efficiency. as deep learning models are more widely deployed and become more valuable, many companies, such as google, microsoft, bigml, and amazon, have launched cloud services to help users train models from user-supplied data sets. although appealing simplicity, this process poses essential security and legal issues. the customer can be concerned that the provider who trains the model for him might resell the model to other parties. say, for example, an inside attacker can replicate the model with little cost and build a similar pay-per-query api service with a lower charge. once that happens, the market share of the model holder may decrease. in another scenario, a service provider may be concerned that customers who purchase a deep learning network model may distribute or even sell the model to other parties with a lower fee by violating the terms of the license agreement. undoubtedly, these can threaten the provider's business. as a result, endowing the capability of tracing illegal deep neural network redistribution is imperative to secure a deep learning market and provides fundamental incentives to the innovation and creative endeavours of deep learning. in the traditional literature, watermarking [ ] is mainly used for copyright protection [ , ] of multimedia data. applying the idea of watermarking to protect the intellectual property (ip) of deep neural network (dnn) models is first introduced by uchida et al. [ ] in . after that, researchers have proposed several dnn watermarking schemes, which can be mainly categorized into two types according to their watermark extraction/verification method: white-box and black-box watermarking. the works in [ ] and [ ] are the typical examples of white-box watermarking, which are built on the assumption that the internal details of the suspicious model are known to the model owner and the entire watermark needs to be extracted. the authorship verification is done by comparing the bit error between the extracted watermark and the embedded one. however, their range of application has been restricted by the inherent constraint, i.e., the internal details is known to the owner, and recent works are more focused on the black-box setting. the black-box setting only assumes access to the remote dnn api but not its internal details. the frameworks of white-box and black-box dnn watermarking schemes are the same, i.e., they both consist of a watermark embedding stage and an extraction/verification stage. typical examples of black-box watermarking are the works in [ , ] , where the authors utilized the back-door property of neural network models [ , ] to embed ownership information when building the model. more specifically, in these works, the watermark embedding is achieved by training with, besides normal samples, some extra crafted samples, or the so-called trigger set (both are referred to as key samples in this work). in the verification stage, the watermarked model will return the predefined labels upon receiving the key samples (compared to the watermark-free model who returns random labels) while performing as normal on non-key samples. according to the key samples they used, these methods can be further categorized into two main classes as follows. the first category is to use crafted key samples, that is, key samples are obtained by superimposing perturbation to training samples. taking image classification as an example, one can embed a readable logo or noise pattern into the normal training images. then these key images are assigned to a specific wrong label [ ] . in merrer et al.'s work [ ] , some normal images that close to the decision frontier are modified imperceptibly, and part of them are assigned to wrong labels, while others inherit their original correct ones. different from [ , ] , the authors in [ ] employed an autoencoder to embed an exclusive logo into ordinary samples and get the key samples. the second category is to use clean key samples. for instance, in [ ] , one kind of key images are chosen from unrelated classes and marked to a specific wrong label. in [ ] , the key samples are sampled from the ordinary images, which can be correctly recognized by the watermarked model but misclassified by the corresponding watermark-free model. another typical example is the work proposed by adi et al. in [ ] , in which they chose some abstract images that are uncorrelated to each other to serve as key samples, and these abstract images are randomly labeled (so the probability that this random label equals the output label of an watermark-free model is low). the underlying rationale is, once again, that only the protected model can correctly recognize the key samples with overwhelming probability since they contribute to the training process. to summarize and to the best of our knowledge, all the existing blackbox dnn watermarking schemes are back-door based, and they are key sample dependent since assigning key samples with wrong labels will inevitably, more or less, twist the original decision boundary. from this sense, the functionality (i.e., classification accuracy) and robustness of the watermarked model are directly related to the characteristics of the used key samples. say for example, if crafted key samples are used for watermarking a dnn model, and a fixed perturbation is superimposed to certain key sample and this very crafted key sample is far away from the original classification frontier (of the watermark-free dnn model), then the decision boundary will be twisted heavily (e.g., become a fractal-like structure) to meet the accuracy criteria, while the robustness or the generality will decrease correspondingly. our key observation to mitigate this problem is simple but effective: adding a new label's to the key samples will minimize, if not eliminate, the effect of boundary twisting. the rationale lies in the fact that, instead of treating key samples are drawn from the marginal distribution of the sample space, we consider the superimposed perturbation to the samples or unrelated natural samples as a new feature that dominates the classification of a new class. theoretically, after adding a new label, the boundary will not be twisted, and all the merits of the corresponding watermark-free model will be preserved. from another point of view, the required number of key samples for watermark embedding, ownership verification, and the false-positive rate will be minimized when compared with boundary-twisted kind dnn watermarking schemes [ ] . in a nutshell, we regard the contributions of this work are as follows: -we propose a novel black-box dnn watermarking framework that has high fidelity, high watermark detection rate, and zero false-positive rate, and robust to pruning attack and fine-tuning attack. -we evaluate the proposed framework on three benchmark datasets, i.e., mnist, cifar and cifar , to quantify the relationship among classification accuracy, perturbation strength, and length of the key samples used during training. the rest of this paper is structured as follows. in sect. , we briefly introduce some background knowledge of deep neural networks, watermarking and dnn watermarking. section presents the formal problem formulation and algorithmic details of the proposed dnn watermarking approach. the experimental results and analyses are presented in sect. , and some further security considerations are discussed in sect. . we make a conclusion in sect. . conceptually, the basic premise of a dnn model is to find a function f : x → y that can predict an output value y ∈ y upon receiving a specific input data x ∈ x. a dnn model generally consists of three parts: an input layer, one or more hidden layers, and an output layer. each layer has several nodes that are customarily called neurons that connect to the next layer. generally speaking, the more hidden layers, the better the performance of the model. however, it is not an easy task to train and learn a good model f that predicts well on unseen samples. typically, the training requires a vast number of , while the labeling requires expert knowledge in most applications. with the data available, the real training, which involves minimizing a loss function l that is dependent on millions of parameters in the case of dnn, also relies on powerful computing resources. this observation motivates us to design mechanisms to protect the intellectual property of dnn. digital multimedia watermarking, which makes use of the redundancy among the data of multimedia to hide information, is a long-studied research area. one popular application of it is to provide ownership verification of digital content, including audio, video, images, etc. the ownership verification process can be achieved in two different ways depending on the embedding methods: ) extracting data from a suspicious copy and comparing the similarity between the extracted data and the embedded watermarks; ) confirming the existence of an ownership-related watermark does exist in a suspicious copy. typically, the verification is executed by a trusted third party, for example, a judge. for dnn watermarking, the watermark extraction/verification process can be executed in either a white-box or black-box way. the white-box setting assumes that the verifier has full access to all of the parameters of the suspicious model, which is similar to the first kind of digital watermarking verification. while in the black-box setting, it assumes that the verifier can only access the api of the remote suspicious model, i.e., sending queries through the api of the suspicious model who will output a class tag. most recent dnn watermarking schemes focused on the black-box verification as it is more practical than a white-box one. this work also lies in the domain of the black-box setting. similar to the current literature and for easy presentation, we only focus on image classification dnn model ip protection. without loss of generality, we only consider the first category of black-box dnn watermarking, i.e., crafting image samples by superimposing perturbation to them. but it is noteworthy to mention that the proposed model can also be applied for classification models of other data formats, and it is also compatible with the second category of dnn watermarking. there is no essential distinction between these two kinds of key samples in terms of classification tasks since both of them can be viewed as the perturbed version of the original images. we consider the scenario in which three parties are involved: a service provider, who helps the customer to train a watermarked model f w ; a customer alice, who is the model owner that provides the training data; and an adversary bob, who is the business competitor of alice that has obtained a copy of alice's model f w . after receiving the model of alice, bob has the incentive to modify the model from f w slightly to get f w , say for example, by model compression, to avoid ip tracing under the condition that the model accuracy does not decrease. we study the problem of how to prove the model f w from bob is an illegal copy of alice's model f w via black-box accessing f w . the overall workflow of the service is depicted in fig. . ideally, a good watermarked dnn model needs to have the following desirable properties: -fidelity: the classification accuracy of the watermarked model f w for normal test data should be close to that of the original model f; -effectiveness and efficiency: the false positive rate for key samples should be minimized, and a reliable ownership verification result needs to be obtained with few queries to the remote dnn api; -robustness: the watermarked model can resist several known attacks, for example, pruning attack and fine-tuning attack. from a high-level point of view, a dnn watermarking scheme Π consists of three algorithms: ksgen, tremb, and ver. ksgen takes as input a subset of the original dataset d and a secret s, and outputs a key sample dataset. tremb takes as input the original dataset d and the result from ksgen, and outputs a watermarked model f w . and ver takes as input a suspicious copy f w and the result from ksgen, and conclude whether f w is pirate or not. the dnn watermarking scheme Π is superior (to the literature works) if it achieves better trade-off among the above mentioned three properties. before diving into the details of the method, we present a motive example first. for illustration, we extract the output layer to form a toy network (the left part of fig. (a) ). then we add a new label to the extracted network to observe the boundary twist of the expanded network (the right part of fig. (a) ). as is clear from fig. (a) , the change caused by adding a new label is quite small. we run more experiments on this toy network and the expanded network and depict the results in fig. (b) for clear comparison. for ease of presentation and without loss of generality, assume the original goal is to predict (Δ − ) classes by training a model f from the dataset d = after adding a new label, we alternatively train a model f w from d and some crafted samples (by running ksgen) to predict Δ different classes. with these notations, the details of the three algorithms ksgen, tremb, and ver are given as follows. key samples generation ksgen: for a given subset of d, say d , the algorithm crafts samples by calculating where is the perturbation pattern determined by the secret s, and α = |d | |d| and β, the perturbation strength, are system parameters that will be studied later in sect. . assigning all the crafted samples to the Δ-th label, ksgen outputs the key sample dataset k = {k (i) , Δ} |d | i= . dnn training and watermark embedding tremb: with the datasets available, the service provider trains a dnn model f w . different from the watermark-free model f that classifies (Δ − ) classes, the watermarked model f w learns from the crafted dataset k to classify one more class, i.e., the class Δ. aligning with the literature works [ , , , , ] , we also employ the softmax function for the output layer. ver: upon detecting a suspicious dnn service f w of bob, alice will ask the judge to verify whether a secret watermark can be identified from bob's model. the judge will choose a subset of d, say it is d , and produce k = ksgen(s, d ) and send query image k ∈ k to f w to check the output label is Δ or not. it is easy to understand that, after adding a new label, a watermarkfree model cannot output a nonexistent class label Δ, that is, the probability it holds no matter x ∈ d or x ∈ k, which implies zero false-positive rate and it is desirable as discussed in sect. . . correlating with more properties from sect. . , fidelity essentially requires where θ is the number of appearance of the label that is not Δ. then the mean value of q is determined by which is bounded by the reciprocal of the accuracy on key samples. for example, if p = . , we have e[q] = , which is small enough for verification purpose. in this section, we evaluate the performance of our proposed dnn watermarking method using three datasets: mnist, cifar and cifar . the back-door based dnn watermarking scheme proposed by zhang et al. [ ] serves as the main test-bed to evaluate our proposal. we train and test all models using the tensorflow package on a machine equipped with xtesla v gpus. to eliminate the impact of randomness, every experiments are executed by times, and the mean is calculated. datasets: three different benchmark datasets are used for the evaluation of our proposal, which are mnist, cifar , and cifar , respectively. according to our definition of key samples, they can be viewed as the modified version of the ordinary samples, and the differences lie in the location and strength of the perturbation. in [ ] , the authors validated that the key samples generated by adding noise to normal images are the best choice in terms of different assessment metrics. for this reason, and also to facilitate experiments and comparisons, we use gaussian noise mode, which can be easily obtained from a secret random number generator under s. in [ ] , the key samples are labeled as one of the existing classes, say, for example, class "airplane". so the key samples should be generated from normal samples that do not belong to the class "airplane". it is worth mentioning that the aim of this work is not to achieve superior classification accuracy, but to compare the performances between watermarked networks trained with key samples that predefined with a new label or not. these dnns are relatively shallow but have a fast training speed, which meets our requirements. we using the normal dataset, without key samples to train the watermark-free models f, and the their accuracy for the three benchmark datasets are %, %, and %, respectively. fidelity: the main purpose of fidelity is to test whether the classification accuracy of the watermarked model f w , when testing on non-key images, is deteriorated or not after embedding. to assess this property, we test the classification accuracy of the watermarked model f w on original test dataset (the original functionality of f) and newly generated key sample dataset (the judge will need to use it at the ver stage). in addition, we, by comparing with the work in [ ] , experimentally investigate the relationship among performance, the ratio of the perturbed samples for training α, and the perturbation strength β, as shown in fig. . from the dotted line in fig. , it is easy to come to the conclusion that both of the proposed method and zhang et al.'s method achieve high classification accuracy on normal samples. in fact, they are similar to the ground truth of the original watermark-free model f. the goal of effectiveness is to measure the credibility of watermark existence provided by the output results of the verification process, while efficiency is to test how many queries are needed to get a credible watermark existence result under the pay-per-query api service. obviously, the fewer queries the better, as it can not only save time & money for verification, but also prevent arousing bob's suspicion. from fig. (a)-(c) , we can see that the model accuracy of both methods is increasing with the perturbation strength of key samples. as shown in fig. (e) , when perturbation strength β = . , our method achieves the testing accuracy higher than % with only . % of key samples for training. for comparison, in zhang et al.'s method, to get the same accuracy, more than . % of key samples are needed. to conclude, our method performs better under small α or β for all datasets. once again, we regard this improvement is due to adding a new label. when α and β are small, number of crafted key samples is small and they are very similar to normal samples. under this circumstance, if the key samples are predefined to wrong classes, the learned weights that contribute to the outputs of key samples cannot change too much due to the fidelity constraint. conversely, if a new label is added, the weights associated with this exact new class can be modified without breaking the fidelity constraint. for efficiency, as discussed in sect. . , in our method, only queries are needed on average to determine the existence of a watermark in a suspicious dnn model with p = . , which is just the case for most choices of α and β as shown in fig. . for zhang et al.'s approach, since it is not false-positive free, so query a watermark-free model with key samples may still trigger the predefined label (of key samples) as the output of the api. to mitigate this bias, a larger number of queries should be used and ver should be re-defined as where τ is a pre-defined threshold and θ is the number of appearance of the label that is not equal to the predefined label (of key samples). then the accuracy, after submitting the whole set k to the api as a batch, of ver is for example, with p = . , acc = % and τ = . , |k | = queries should be used for ver. clearly, it is not as efficient as the proposed method. the goal of robustness is to check if the proposed model can resist to attacks, and following the literature, we mainly consider pruning attack (or compression attack in the literature) and fine-tuning attack here. as discussed in sect. . , the adversary has incentive to modify the model to prevent ownership verification. obviously, the adversary does not want to affect the model's classification accuracy with such modification. and pruning and fine-tuning attacks exactly fit this requirement. by saying robust in the scenario of ip protection of dnn using watermarking, essentially, we expect the classification accuracy of key samples is insensitive after such attacks. in the experiments, we test the classification accuracy of the watermarked model for ordinary samples and key samples, separately, under different pruning rates, and the results are shown in fig. . it can be observed from this figure that the model accuracy for classifying newly generated key samples under both this proposal and zhang et al.'s design does not decrease too much with the increasing of pruning rate. but in general, our method performs slightly better than the one in [ ] , especially the pruning rate is relatively high. it is reported from [ ] that deep neural networks, besides their incremental learning capability, are also prone to forget previous tasks. the fine-tuning is a useful method that utilises the catastrophy forgetting of dnns to retrain a watermarked model to invalidate the ownership verification. to measure the testing accuracy of clean samples and key samples of our method under finetuning, we employ the same experimental settings as used in [ ] . the results of the fine-tuning attack are tabulated in table . for fair comparison, the parameters used in the three datasets are: (α = . , β = . × − ) for mnist, and (α = . , β = . ) for cifar and cifar . under these settings, both our method and the one in [ ] can achieve the ground-truth accuracy on each dataset, as shown from the values in parenthesis of table . from this table, it is easy to see that after fine-tuning, both our method and the method in [ ] still preserve good classification accuracy on normal samples. this is due the generalization property of dnn and it is well accepted in the machine learning field. for the classification of key samples after fine-tuning, we expect accuracy loss. for sure the generalization property still holds in this case, but the watermarked label is learnt from insufficient data and weak features. it is observed from this table, for the mnist dataset, the accuracy of both methods is still as high as the ground truth. it may be due to the reason that the mnist dataset is relatively simple, so the weak features (from the key samples) are learnt perfectly during the training process, and the generalization property dominates classification accuracy. for the other two datasets, the acccracy decreases as expected. to conclude, although our method cannot fully prevent the fine-tuning attack, compared with the literature work [ ] , it mitigate the attack to large extent. apart from the pruning attack and fine-tuning attack we mentioned above, recently, several new attacks [ , , ] are proposed against black-box dnn watermarking techniques. we discuss the most related type of attacks in brief in this section. this attack considers the scenario that, given a query, bob first judges whether or not the query issued by someone works as a key sample for verification. in this way, the verification ver will be invalidated by rejecting service [ ] . in [ ] , the authors adopted an autoencoder to serve as the key sample detector. as discussed in sect. . , our method works with a smaller number of training key samples and weaker perturbation strength, which makes the detection harder. in this paper, we proposed a novel black-box dnn watermarking method: assigning a new label to key samples to minimize the distortion of the decision boundary. compared with the existing dnn framework, it achieves zero false-positive rates and performs better when the number of training key samples are small and the perturbation is weak. for security, it is validated that the new proposal is more robust than existing schemes, and we leave the investigation of its resistance to query rejection attack for further study. turning your weakness into a strength: watermarking deep neural networks by backdooring an overview of digital video watermarking deepmarks: a digital fingerprinting framework for deep neural networks leveraging unlabeled data for watermark removal of deep neural networks deepsigns: an end-to-end watermarking framework for ownership protection of deep neural networks badnets: identifying vulnerabilities in the machine learning model supply chain have you stolen my model? evasion attacks against deep neural network watermarking techniques measuring catastrophic forgetting in neural networks. in: thirty-second aaai conference on artificial intelligence adversarial frontier stitching for remote neural network watermarking how to prove your model belongs to you: a blind-watermark based framework to protect intellectual property of dnn secure and robust digital image watermarking scheme using logistic and rsa encryption robust watermarking of neural network with exponential weighting embedding watermarks into deep neural networks protecting intellectual property of deep neural networks with watermarking you can access but you cannot leak: defending against illegal content redistribution in encrypted cloud media center acknowledgements. this work was supported in part by the australian research council under grant lp , in part by the national natural science foundation of china under grant , and in prat by the nvidia corporation. key: cord- -yk cs up authors: paciorek, mateusz; bogacz, agata; turek, wojciech title: scalable signal-based simulation of autonomous beings in complex environments date: - - journal: computational science - iccs doi: . / - - - - _ sha: doc_id: cord_uid: yk cs up simulation of groups of autonomous beings poses a great computational challenge in terms of required time and resources. the need to simulate large environments, numerous populations of beings, and to increase the detail of models causes the need for parallelization of computations. the signal-based simulation algorithm, presented in our previous research, prove the possibility of linear scalability of such computations up to thousands of computing cores. in this paper further extensions of the signal-based models are investigated and new method for defining complex environments is presented. it allows efficient and scalable simulation of structures which cannot be defined using two dimensions, like multi-story buildings, anthills or bee hives. the solution is applied for defining a building evacuation model, which is validated using empirical data from a real-life evacuation drill. the research on modeling and simulation of groups of autonomous beings, like crowds of pedestrians, cars in traffic or swarms of bees, has a great need for efficient simulation methods. the need for simulating numerous groups of beings in large environment, the growing complexity of models and the desire to collect the results fast increase the complexity of the computational task. therefore, new, scalable simulation methods are constantly pursued by the researchers. in our previous research [ ] , a novel method for parallelization of autonomous beings simulation has been presented. it is based on the concept of information propagation, which replaces the need for searching the required data by individual beings. the signal is modeled in a d grid of cells, it is directional and autonomously spreads to adjacent cells in the grid. this simple concept removes the need for analyzing remote locations during model update, and therefore, allows splitting the simulated environment between many computing processes, which communicate with fixed number of neighbors once in a single time step, which is the crucial requirement of super-scalable algorithm defined in [ ] . the basic version of the method allows representing two-dimensional environments as grids of cells, which are split into equal fragments and updated in parallel. after each time step, the processes updating neighbor fragments exchange information about signal and beings located in common borders. the method requires defining dedicated simulation models, however, in return it can provide linear scalability of simulation. linear scalability has been achieved in several tests performed on hpc hardware, which involved up to computing cores. linearly scalable method for autonomous beings simulation encouraged further research on signal-based models and their validation. in [ ] three simple models have been presented in order to demonstrate the capabilities of the method. in [ ] a model of pedestrian behaviour based on proxemics rules have been developed and tested. further work on signal-based models for pedestrians led to the problem considered in this paper: the need for modeling complex, multi-level buildings, which cannot be represented in two dimensions. similar problem can be encountered in modeling of other species habitations, like anthills, bee hives or mole tunnel systems. this class of modeling problems can be addressed by using three-dimensional models, however, efficiency of such approach would be doubtful. only a small part of the d space is accessible and significant so the d model would introduce a huge waste in memory and computing resources. in this paper we propose the extension of the signal propagation method addressing the requirements of environments that cannot be represented by d grid. the extension introduces an abstraction over the relation of adjacency, which enables flexibility in defining the shape of the environment while preserving values of distance and direction. this concept is further explored by implementing a evacuation scenario of a multi-story building as an example of such an environment. finally, we compare the metrics collected during the simulation with the available real-life data to validate the resulting model. the problem of autonomous beings simulation and computer-aided analysis of phenomena taking place in large groups of such beings has been studied for many decades. for example, first significant result of traffic simulation using computers can be found in the fifties of xx century [ ] . since then countless reports on methods for modeling and simulation of different types of beings and different phenomena have been published. specific problems, like urban traffic or pedestrian dynamics, attracted so much attention, that several different classifications of models can be found in the literature [ , , ] . the taxonomy is based on the considered level of detail [ ] can be found in several different problems, where macro-, mezo-and mico-scale models are being distinguished. in recent years the vast majority of research focus on micro-scale models, which distinguish individual entities and allow differences in their characteristics and behavior. one of the basic decisions which has to be made while defining the model is the method of representing the workspace of the beings. the environment model can be discrete or continuous, it can represent or dimensions. the decision to discretize the workspace significantly simplifies the algorithms for model execution, which allows simulating larger groups in bigger environments. in many cases a discrete workspace model is sufficient to represent desired features of the beings and reproduce phenomena observed in real systems, like in the wellrecognized nagel-schreckenberg freeway traffic model [ ] . many other researchers use inspirations from cellular automata in simulation of different types of beings (swarms of bees [ ] or groups of pedestrians [ ] ) because of simplicity, elegance and sufficient expressiveness. the common challenge identified in the vast majority of the publications in the area is the performance of the simulations. the need for increasing the complexity of models, simulating larger populations of beings and getting results faster is visible in almost all of the considered approaches. in many cases the performance issues prevent further development, and therefore a lot of effort is being put into creating parallel versions of simulation algorithms and generalpurpose simulation frameworks. scalable communication mechanisms have been added to repast [ ] , dedicated frameworks are being built (like pandora [ ] ). the results show that by defining models dedicated for parallel execution, scalability can be achieved, like in [ ] , where almost linear scaling is demonstrated up to processing units with the flame on hpc framework. efficient parallelization of this type of computational task is non-trivial. the algorithm executed in each time step of the simulation operates on a single data structure, which represents the environment state. parallel access to the data structure requires complex synchronization protocols, which imply significant and non-scalable overhead. therefore, in our solution presented in [ ] , we focused on removing the need for accessing the remote parts of the data structure. the modeling method assumes that the information is explicitly pushed to computing units that might need it for updating their state. implemented xinuk simulation platform proves the scalability of the approach, offering linear scalability up to cores of a supercomputer. in this paper we present important extensions to the modeling methods supported by the platform, which allow representing complex environments, not possible to model with a d grid of cells. the scalable simulation algorithm, implemented by the xinuk framework, is capable of simulating any d environment while providing a flexible approach to the interpretation of the grid. cells might represent a terrain with qualities appropriate for the simulation, actors exclusively occupying a specific place in the grid or a group of actors amassed in one location. each cell is adjacent to its eight neighbor cells and is able to interact with any of them (e.g. move its contents or a part thereof), if the logic defined by the simulation designer allows such an action. however, the simulations that cannot be represented in simple d environment are difficult, if not impossible, to properly model using the framework. while some d layouts can be mapped to d grid utilizing simplifications or other compromises (e.g. modelling upward/downward slopes or stairs as special cells that modify movement speed or behavior of the agent in such a cell), most terrain configurations would greatly benefit from-or even require-more general solution. one example of such configuration, which will be the main focus of this work, is a multi-story building with staircases located in the center. each floor can be represented as an independent part of the d grid, but the connections between the floors would have to overlap with them. the standard, d moore neighborhood is not sufficient to correctly represent aforementioned environments. one solution would be to generalize the method to represent d grid of cells, with each cell having neighbors. however, in the considered class of problems, this approach would result in significant waste of memory and computational resources, as only few of the cells would be important for the simulation results. from this problem stems the idea of the abstraction of the cell neighborhood. the proposed version of the modeling method introduces a neighborhood mapping for each direction. each cell can connect to: top, top-right, right, bottomright, bottom, bottom-left, left and top-left. given a grid of dimensions h × w , this mapping can be declared as in eq. : where x × y is a set of all possible coordinates in initial grid, d is a set of mentioned directions and n is a function mapping coordinates and direction to another set of coordinates or none, representing the absence of the neighbor in given direction. likewise, the signal has been updated to be stored as a similar map containing signal strength in given direction. as a result, the signal propagation algorithm required reformulation to make use of the new representation. firstly, the idea of the function of adjacent direction ad of a direction was necessary, which is presented in eq. : with use of this function, and assuming the function s that returns the current signal in the cell at given coordinates in given direction (eq. ), the new signal propagation function sp can be described as in eq. : where sp f ∈ [ , ] is a global suppression factor of the signal. it is worth noting that, should the need arise, this mechanism can be extended to any collection of directions, as long as the signal propagation function is updated to properly represent distribution in all directions. introduction of the new neighbor resolution allows seamless adaptation of the previous approach: by default, all the neighbors of the cell are the adjacent cells. the neighbor mapping function for such a case is defined as in eq. , with an exception of the grid borders, where neighbors are nonexistent (none, as in ( )): additionally, in the step of the grid creation any neighbor relation can be replaced to represent non-grid connection between cells. while the concept of remote connections appears trivial, it is critical to consider the possibility that in the process of the grid division neighbors are distributed to the separate computational nodes. such a situation is certain to occur on the line of the division. in our previous approach, we applied buffer zones mechanism as a solution. the new concept required this part of the framework to be redesigned, to allow more flexible cell transfer. as a result, new type of cells was introduced: remote cells. each represents a cell that is not present in the part of grid processed in this worker and contains information regarding: -the identifier of the worker responsible for processing the part of grid containing target cell, -the coordinates of target cell, -the contents of the cell awaiting the transfer to the target cell. following each simulation step, contents of all remote cells are sent to their respective workers, using the logic previously associated with the buffer zones. the modification of the framework did not introduce any alterations in the overall complexity of the simulation process. communication between processes was not altered and utilizes the same methods as the previous synchronization of the buffer zones. creation of the grid containing large number of non-standard neighborhood relations does introduce additional cell contents that need to be transmitted to the target worker, however it is the minimal volume of data required for the simulation to be processed. summarizing, as a result of all the mentioned modifications, the simulation algorithm acquired the ability to model environments that are not possible to be represented on the d grid. buffer zones mechanism has been abstracted to allow more flexible communication without assuming that the communication can only occur at the grid part borders. the scalability of the framework has been preserved, since the amount of the data sent between workers remains unchanged for the same simulation model represented in the previous and the proposed approach. it is possible to define more complex communication schemes, however, the number of communication targets remains fixed and relatively low for each worker. signal-based methods can be used to simulate evacuations of people from buildings. in such a signal-based model, it is enough to place a signal sources in exits, so beings will move accordingly to egress routes created by signal emitted by the sources, leaving the building. a negative signal can be used to make beings stay away from potential threats, for instance fire or smoke. in the presented model, the repelling signal was used for representing the reluctance for creating large crowds when alternative routes are possible. in xinuk framework, there are two basic cell types: -obstacle -a cell that does not let signal through, -emptycell -an empty cell traversable by a signal and accessible by any being. in the proposed model, obstacle type of cells was used to create walls. in addition, the following cells were added to create the evacuation model: -personcell -representing a person that was to evacuate, -exitcell -representing an exit from a building, -teleportationcell -a remote cell, described in the previous section, that was moving a being to a destination cell, -evacuationdirectioncell -source of a static signal. a being of personcell type was able to move to any accessible adjacent cell, that is an emptycell, an exitcell or a teleportationcell. movements were possible in directions if there were no walls or other people in the neighborhood, as shown in fig. . in the created model, two types of signals were used: -static signal field -a snapshot of a signal propagated in a particular number of initial iterations, where cells of evacuationdirectioncell type were the signal sources. propagated signal created egress routes that were static during the simulation, -dynamic signal field -signal emitted by moving personcell beings. the static signal field can be compared to a floor field described in [ ] or potential field [ ] . two different models of evacuating people behaviors were implemented and tested: -moving variant -always move if a movement is possible. -standing variant -move only when the best movement is possible. in the moving variant, a person's destination was calculated as follows: . signals in directions of neighbor cells were calculated based on dynamic and static signals by summing both signals, . calculated signals were sorted in a descending order, . from sorted destinations, a first destination was chosen that was currently available (the cell was not occupied by any other person). in the standing variant, a person did not move if the best direction was not available, preventing unnatural movements to directions further away from targeted exit. thus the rd step from the moving variant algorithm was changed as follows: . from sorted destinations, a first destination was chosen. if the destination was not available, the being would not move. in a high congestion of beings trying to get to a particular location, conflicts are highly possible. one solution to this problem is to let two beings enter one cell, creating a crowd. a simpler solution, implemented in our model, is to check if a destination cell chosen by a being is empty both in current and in the next iteration. this way, all conflicts will be avoided. each floor of a building was mapped onto a d grid reflecting its shape and dimensions. to simulate a multi-story buildings using d grids, a special teleportationcell was created. a being that entered the teleportationcell at the end of a staircase on a floor n, was moved to a cell corresponding to the beginning of a staircase at floor n − . to validate the evacuation model, a simulation of a real-life evacuation described in [ ] was implemented. the evacuation drill took place in a -story tower connected to a -story low-rise structure. the highest floor of the tower was ignored in the research, thus in the implementation there is no xii floor in the tower, as shown on the building plan (fig. ) . the highest floor of a low-rise structure was ignored as well and the data focused on the tower, which is a reason why the low-rise structure is shown as one floor on the building plan. each floor was connected to two staircases, each , m wide, allowing only a single line of pedestrians to be created. the staircase length was not mentioned in the paper. evacuation started earlier on floors: v, vi and vii. after min, there was a general alarm on the remaining floors. pre-evacuation time curve was approximately linear (fig. in [ ] ). evacuation rate was over one person per second for the first % of people in the building. afterwards, the rate was slightly smaller because of discontinuation of use of one of exits (the reason not explained in the article). results from the drill can be seen in fig. . based on the above data, an implementation of evacuation model described in the previous section was created in xinuk platform. crucial parameters were set as follows: -gridsize = ; size of a grid. the whole grid was a square of × cells, each cell representing a square with side length of , m. -iterationsnumber = ; number of iterations completed in a single simulation. iteration was corresponding to s. -evacuationdirectioninitialsignal = ; signal intensity that was emitted by evacuationdirectioncell. -personinitialsignal = - . ; signal intensity that was emitted by person-cell. in contrast to evacuationdirectioninitialsignal, the value was negative so people were repelling each other slightly. in the simulation, iteration corresponds to s of an evacuation. number of people on each floor was set accordingly to table from the source paper. the number of people placed on low-rise structure's floor was equal to the number of all people in low-rise structure stated in the source paper. the implemented simulation had three phases: -phase - st to th iteration -creating static signal field by placing signal sources in the cells corresponding to the exits and corridors' ends, -phase - th to th iteration -evacuation after the initial alarm, evacuating people from v, vi and vii floors, -phase - th to th iteration -evacuation after the general alarm, evacuating people from the whole building. to achieve linear pre-evacuation times, all of the people were not placed on a grid in the first iteration of the nd and rd phase of the simulation, but they were placed on a grid linearly -one person per iteration on each floor (relying on the fig. in [ ] ). validation of the model was based on visual analysis of simulation results and on the evacuation curve metric which is a dependency between the number of people evacuated from the building and time. empirical data is shown in fig. . during the observations, two anomalies were visible: . problem : crowds of people next to the upper and lower entrance of staircases on each floor were consistently different-lower entrance seemed to generate larger crowds-and an evacuation using upper stairs tended to take longer (fig. ), . problem : people in corridors that could not move forward but could go back, were moving back and forth waiting to go further or were choosing to go back to staircase entrances. problem was a result of a sequential order of updating cells. a person that was trying to enter an upper staircase was above the people in the staircase, thus the person was updated before the people in the corridor. this way, the person's movement was favored and congestion was visible in the corridors and there were no crowds close to the upper entrances to the upper staircases. similarly, trying to enter a lower staircase, the movement of people already in the staircase was favored. a person trying to enter the staircase was updated after people in the staircase managed to occupy all empty spaces, preventing the person from entering the staircase. figure shows both of the situations. a simple solution to this problem was to update locations of people in a random order. figure shows the results of such an approach. crowds are distributed equally close to both of staircase entrances and evacuation in both staircases progresses similarly. problem was a result of the decision making algorithm. according to the algorithm, people would make any movement if it was possible, even a bad one, rather than stay in place. a solution was to use the standing variant: choose only the most attractive destination or not to move at all. the visual analysis of the final model behavior is satisfactory. the formation of crowds and selection of the exits resembled the expected phenomena. the people were eager to follow the shortest path towards the exits, while avoiding excessive crowding which might lead to trampling. the four combinations of the model update algorithm, moving/standing variant and sequential/random variant, were executed times. we used the favoring movements of people that are upper in a grid in sequential way of updating cells. on the left -a situation when person a is trying to enter a staircase that is below. on the right -a situation when person b is trying to enter a staircase that is above prometheus supercomputer (part of pl-grid infrastructure ), which allowed completion of all computation within several minutes. the resulting chart (fig. ) shows that this particular metrics is not influenced significantly by the selected variant. an average rate of people leaving the building is a little over person per second, which matches the data on . after an evacuation of % of people, the rate did not change, as in the simulation people continued using two exits. on a source chart (fig. ) people have already reached exits in first seconds while on the resulting chart (fig. ) the first person left the building after th second of evacuation. according to [ ] , people on the floors other than v, vi and vii did not evacuate before the general alarm. thus, it is not clear why the chart shown in fig. suggests that some people have left the building in such a short time overcoming at least floor long stairs. the results of the experiments are not perfectly matching the empirical data from [ ] . nonetheless, taking into consideration the ambiguity, contradictions or lack of some details of the described evacuation, we conclude that the resulting model yielded realistic and consistent outcomes. in the context of validation of the presented method as a building block of similar environments, the results are satisfactory. in this paper we proposed an extension to the signal propagation modeling method and the xinuk framework, addressing the limitations of the existing approach in the context of complex d environments. we introduced the alternative to the standard, grid-based neighborhood as a means to generalizing the idea of buffer zones present in the framework. the new mechanism greatly increased the flexibility of the method, while avoiding the massive growth of complexity that would result from switching to full d grid environment. at the same time, the scalability of the original solution remained unhindered, as the alteration did not increase the amount of data exchanged between computational nodes in the synchronization step. the evacuation scenario used as a demonstration of the new capabilities of the method provided results confirming that such a scenario can be accurately represented in the resulting framework. as a followup of this work, we intend to further explore the possibilities arising from the flexible neighborhood declaration, especially the environments that would benefit from additional directions, e.g. layers of an anthill utilizing the vertical directions. research dedicated to further reduction in the communication might yield interesting results as well, e.g. investigating a trade-off between accuracy and performance of the system while performing the synchronization after several simulation steps. beehave: a systems model of honeybee colony dynamics and foraging to explore multifactorial causes of colony failure high-performance computing framework with desynchronized information propagation for large-scale simulations exploitation of high performance computing in the flame agent-based simulation framework repast hpc: a platform for large-scale agent-based modeling super-scalable algorithms for computing on , processors simulation of freeway traffic on a general-purpose discrete variable computer pre-evacuation data collected from a mid-rise evacuation exercise a cellular automaton model for freeway traffic hpc large-scale pedestrian simulation based on proxemics rules verification and validation of simulation models cellular automaton model for evacuation process with obstacles genealogy of traffic flow models towards realistic and effective agent-based models of crowd dynamics scalable agent-based modelling with cloud hpc resources for social simulations crowd analysis: a survey acknowledgments. the research presented in this paper was supported by the polish ministry of science and higher education funds assigned to agh university of science and technology. the authors acknowledge using the pl-grid infrastructure. key: cord- -tbfg vmd authors: brauer, fred; castillo-chavez, carlos title: epidemic models date: - - journal: mathematical models in population biology and epidemiology doi: . / - - - - _ sha: doc_id: cord_uid: tbfg vmd communicable diseases such as measles, influenza, and tuberculosis are a fact of life. we will be concerned with both epidemics, which are sudden outbreaks of a disease, and endemic situations, in which a disease is always present. the aids epidemic, the recent sars epidemic, recurring influenza pandemics, and outbursts of diseases such as the ebola virus are events of concern and interest to many people. the prevalence and effects of many diseases in less-developed countries are probably not as well known but may be of even more importance. every year millions, of people die of measles, respiratory infections, diarrhea, and other diseases that are easily treated and not considered dangerous in the western world. diseases such as malaria, typhus, cholera, schistosomiasis, and sleeping sickness are endemic in many parts of the world. the effects of high disease mortality on mean life span and of disease debilitation and mortality on the economy in afflicted countries are considerable. (german measles), and chicken pox, confer immunity against reinfection, while diseases transmitted by bacteria , such as tuberculosis, meningitis, and gonorrhea, confer no immunity against reinfection. other human diseases, such as malaria, are transmitted not directly from human to human but by vectors, agents (usually insects) that are infected by humans and who then transmit the disease to humans. there are also diseases such as west nile virus, that are transmitted back and forth between animals and vectors. heterosexual transmission of hiv/aids is also a vector process in which transmission goes back and forth between males and females. we will focus on the transmission dynamics of an infection from individual to individual in a population, but many of the same ideas arise in transmission of a genetic characteristic, such as gender, race, genetic diseases, a cultural "characteristic," such as language or religion, an addictive activity, such as drug use, and the gain or loss of information communicated through gossip, rumors, and so on. similarly, many of the ideas arise also with different characterizations of what is meant by an individual, including the types of cells in the study of disease dynamics of the immune system. in the study of chagas disease, a "house" (infested houses may correspond to "infected" individuals) may be chosen as an epidemiological unit; in tuberculosis, a household or community or a group of strongly linked individuals ("cluster") may be the chosen unit. an epidemic, which acts on a short temporal scale, may be described as a sudden outbreak of a disease that infects a substantial portion of the population in a region before it disappears. epidemics usually leave many members untouched. often these attacks recur with intervals of several years between outbreaks, possibly diminishing in severity as populations develop some immunity. this is an important aspect of the connection between epidemics and disease evolution. the historian w.h. mcneill argues, especially in his book plagues and peoples ( ) , that the spread of communicable diseases frequently has been an important influence in history. for example, there was a sharp population increase throughout the world in the eighteenth century; the population of china increased from million in to million in , and the population of europe increased from million in to million in . there were many factors involved in this increase, including changes in marriage age and technological improvements leading to increased food supplies, but these factors are not sufficient to explain the increase. demographic studies indicate that a satisfactory explanation requires recognition of a decrease in the mortality caused by periodic epidemic infections. this decrease came about partly through improvements in medicine, but a more important influence was probably the fact that more people developed immunities against infection as increased travel intensified the circulation and cocirculation of diseases. there are many biblical references to diseases as historical influences. the book of exodus describes the plagues that were brought down upon egypt in the time of moses. another example is the decision of sennacherib, the king of assyria, to abandon his attempt to capture jerusalem about bc because of the illness of his soldiers (isaiah , - ), and there are several other biblical descriptions of epidemic outbreaks. the fall of empires has been attributed directly or indirectly to epidemic diseases. in the second century ad, the so-called antonine plagues (possibly measles nomic hardships leading to disintegration of the empire because of disorganization, which facilitated invasions of barbarians. the han empire in china collapsed in the third century ad after a very similar sequence of events. the defeat of a population of millions of aztecs by cortez and his followers can be explained, in part, by a smallpox epidemic that devastated the aztecs but had almost no effect on the invading spaniards, thanks to their built-in immunities. the aztecs were not only weakened by disease but also confounded by what they interpreted as a divine force favoring the invaders. smallpox then spread southward to the incas in peru and was an important factor in the success of pizarro's invasion a few years later. smallpox was followed by other diseases such as measles and diphtheria imported from europe to north america. in some regions, the indigenous populations were reduced to one-tenth of their previous levels by these diseases: between and the indian population of mexico was reduced from million to million. the black death (probably bubonic plague) spread from asia throughout europe in several waves during the fourteenth century, beginning in , and is estimated to have caused the death of as much as one-third of the population of europe between and . the disease recurred regularly in various parts of europe for more than years, notably as the great plague of london of - . it then gradually withdrew from europe. since the plague struck some regions harshly while avoiding others, it had a profound effect on political and economic developments in medieval times. in the last bubonic plague epidemic in france ( - ), half the population of marseilles, percent of the population in nearby toulon, per cent of the population of arles, and percent of the population of aix and avignon died, but the epidemic did not spread beyond provence. expansions and interpretations of these historical remarks may be found in mcneill ( ) , which was our primary source on the history of the spread and effects of diseases. the above examples depict the sudden dramatic impact that diseases have had on the demography of human populations via disease-induced mortality. in considering the combined role of diseases, war, and natural disasters on mortality rates, one may conclude that historically humans who are more likely to survive and reproduce have either a good immune system, a propensity to avoid war and disasters, or, nowadays, excellent medical care and/or health insurance. descriptions of epidemics in ancient and medieval times frequently used the term "plague" because of a general belief that epidemics represented divine retribution for sinful living. more recently, some have described aids as punishment for sinful activities. such views have often hampered or delayed attempts to control this modern epidemic. there are many questions of interest to public health physicians confronted with a possible epidemic. for example, how severe will an epidemic be? this question may be interpreted in a variety of ways. for example, how many individuals will be affected and require treatment? what is the maximum number of people needing and) invaded the roman empire, causing drastic population reductions and eco-care at any particular time? how long will the epidemic last? how much good would quarantine of victims do in reducing the severity of the epidemic? scientific experiments usually are designed to obtain information and to test hypotheses. experiments in epidemiology with controls are often difficult or impossible to design, and even if it is possible to arrange an experiment, there are serious ethical questions involved in withholding treatment from a control group. sometimes data may be obtained after the fact from reports of epidemics or of endemic disease levels, but the data may be incomplete or inaccurate. in addition, data may contain enough irregularities to raise serious questions of interpretation, such as whether there is evidence of chaotic behavior [ellner, gallant, and theiler ( ) ]. hence, parameter estimation and model fitting are very difficult. these issues raise the question whether mathematical modeling in epidemiology is of value. mathematical modeling in epidemiology provides understanding of the underlying mechanisms that influence the spread of disease, and in the process, it suggests control strategies. in fact, models often identify behaviors that are unclear in experimental data-often because data are nonreproducible and the number of data points is limited and subject to errors in measurement. for example, one of the fundamental results in mathematical epidemiology is that most mathematical epidemic models, including those that include a high degree of heterogeneity, usually exhibit "threshold" behavior, which in epidemiological terms can be stated as follows: if the average number of secondary infections caused by an average infective is less than one, a disease will die out, while if it exceeds one there will be an epidemic. this broad principle, consistent with observations and quantified via epidemiological models, has been used routinely to estimate the effectiveness of vaccination policies and the likelihood that a disease may be eliminated or eradicated. hence, even if it is not possible to verify hypotheses accurately, agreement with hypotheses of a qualitative nature is often valuable. expressions for the basic reproductive number for hiv in various populations is being used to test the possible effectiveness of vaccines that may provide temporary protection by reducing either hiv-infectiousness or susceptibility to hiv. models are used to estimate how widespread a vaccination plan must be to prevent or reduce the spread of hiv. in the mathematical modeling of disease transmission, as in most other areas of mathematical modeling, there is always a trade-off between simple models, which omit most details and are designed only to highlight general qualitative behavior, and detailed models, usually designed for specific situations including short-term quantitative predictions. detailed models are generally difficult or impossible to solve analytically and hence their usefulness for theoretical purposes is limited, although their strategic value may be high. for public health professionals, who are faced with the need to make recommendations for strategies to deal with a specific situation, simple models are inadequate and numerical simulation of detailed models is necessary. in this chapter, we concentrate on simple models in order to establish broad principles. furthermore, these simple models have additional value since they are the building blocks of models that include detailed structure. a specific goal is to compare the dynamics of simple and slightly more detailed models primarily to see whether slightly different assumptions can lead to significant differences in qualitative behavior. many of the early developments in the mathematical modeling of communicable diseases are due to public health physicians. the first known result in mathematical epidemiology is a defense of the practice of inoculation against smallpox in by over three generations) who had been trained as a physician. the first contributions to modern mathematical epidemiology are due to p.d. en'ko between and [dietz ( ) ], and the foundations of the entire approach to epidemiology based on compartmental models were laid by public health physicians such as sir r.a. ross, w.h. hamer, a.g. mckendrick, and w.o. kermack between and , along with important contributions from a statistical perspective by j. brownlee. a particularly instructive example is the work of ross on malaria. dr. ross was awarded the second nobel prize in medicine for his demonstration of the dynamics of the transmission of malaria between mosquitoes and humans. although his work received immediate acceptance in the medical community, his conclusion that malaria could be controlled by controlling mosquitoes was dismissed on the grounds that it would be impossible to rid a region of mosquitoes completely and that in any case, mosquitoes would soon reinvade the region. after ross formulated a mathematical model that predicted that malaria outbreaks could be avoided if the mosquito population could be reduced below a critical threshold level, field trials supported his conclusions and led to sometimes brilliant successes in malaria control. unfortunately, the garki project provides a dramatic counterexample. this project worked to eradicate malaria from a region temporarily. however, people who have recovered from an attack of malaria have a temporary immunity against reinfection. thus elimination of malaria from a region leaves the inhabitants of this region without immunity when the campaign ends, and the result can be a serious outbreak of malaria. we formulate our descriptions as compartmental models, with the population under study being divided into compartments and with assumptions about the nature and time rate of transfer from one compartment to another. diseases that confer immunity have a different compartmental structure from diseases without immunity and from diseases transmitted by vectors. the rates of transfer between compartments are expressed mathematically as derivatives with respect to time of the sizes of the compartments, and as a result our models are formulated initially as differential equations. models in which the rates of transfer depend on the sizes of compartments over the past as well as at the instant of transfer lead to more general types of functional equations, such as differential-difference equations and integral equations. in this chapter we describe models for epidemics, acting on a sufficiently rapid time scale that demographic effects, such as births, natural deaths, immigration into and emigration out of a population may be ignored. in the next chapter we will describe models in which demographic effects are included. daniel bernouilli, a member of a famous family of mathematicians (eight spread throughout history, epidemics have had major effects on the course of events. for example, the black death, now identified as probably having been the bubonic plague which had actually invaded europe as early as the sixth century, spread from asia throughout europe in several waves during the fourteenth century, beginning in , and is estimated to have caused the death of as much as one third of the population of europe between and . the disease recurred regularly in various parts of europe for more than years, notably as the great plague of london of - . it then gradually withdrew from europe. more than % of the population of london died in the great plague ( - ). it appeared quite suddenly, grew in intensity, and then disappeared, leaving part of the population untouched. one of the early triumphs of mathematical epidemiology was the formulation of a simple model by kermack and mckendrick ( ) whose predictions are very similar to this behavior, observed in countless epidemics. the kermack-mckendrick model is a compartmental model based on relatively simple assumptions on the rates of flow between different classes of members of the population. in order to model such an epidemic we divide the population being studied into three classes labeled s, i, and r. we let s(t) denote the number of individuals who are susceptible to the disease, that is, who are not (yet) infected at time t. i(t) denotes the number of infected individuals, assumed infectious and able to spread the disease by contact with susceptibles. r(t) denotes the number of individuals who have been infected and then removed from the possibility of being infected again or of spreading infection. removal is carried out through isolation from the rest of the population, through immunization against infection, through recovery from the disease with full immunity against reinfection, or through death caused by the disease. these characterizations of removed members are different from an epidemiological perspective but are often equivalent from a modeling point of view that takes into account only the state of an individual with respect to the disease. we will use the terminology sir to describe a disease that confers immunity against reinfection, to indicate that the passage of individuals is from the susceptible class s to the infective class i to the removed class r. epidemics are usually diseases of this type. we would use the terminology sis to describe a disease with no immunity against re-infection, to indicate that the passage of individuals is from the susceptible class to the infective class and then back to the susceptible class. usually, diseases caused by a virus are of sir type, while diseases caused by bacteria are of sis type. in addition to the basic distinction between diseases for which recovery confers immunity against reinfection and diseases for which recovered members are susceptible to reinfection, and the intermediate possibility of temporary immunity signified by a model of sirs type, more complicated compartmental structure is possible. for example, there are seir and seis models, with an exposed period between being infected and becoming infective. when there are only a few infected members, the start of a disease outbreak depends on random contacts between small numbers of individuals. in the next section we will use this to describe an approach to the study of the beginning of a disease outbreak by means of branching processes, but we begin with a description of deterministic compartmental models. the independent variable in our compartmental models is the time t, and the rates of transfer between compartments are expressed mathematically as derivatives with respect to time of the sizes of the compartments, and as a result our models are formulated initially as differential equations. we are assuming that the epidemic process is deterministic, that is, that the behavior of a population is determined completely by its history and by the rules that describe the model. in formulating models in terms of the derivatives of the sizes of each compartment we are also assuming that the number of members in a compartment is a differentiable function of time. this assumption is plausible once a disease outbreak has become established but is not valid at the beginning of a disease outbreak when there are only a few infectives. in the next section we will describe a different approach for the initial stage of a disease outbreak. the basic compartmental models to describe the transmission of communicable diseases are contained in a sequence of three papers by w.o. kermack and a.g. mckendrick in mckendrick in , mckendrick in , and . the first of these papers described epidemic models. what is often called the kermack-mckendrick epidemic model is actually a special case of the general model introduced in this paper. the general model included dependence on age of infection, that is, the time since becoming infected, and can be used to provide a unified approach to compartmental epidemic models. the special case of the model proposed by kermack and mckendrick in , which is the starting point for our study of epidemic models, is a flow chart is shown in figure . . it is based on the following assumptions: (i) an average member of the population makes contact sufficient to transmit infection with β n others per unit time, where n represents total population size (mass action incidence). (ii) infectives leave the infective class at rate αi per unit time. (iii) there is no entry into or departure from the population, except possibly through death from the disease. (iv) there are no disease deaths, and the total population size is a constant n. according to (i), since the probability that a random contact by an infective is with a susceptible, who can then transmit infection, is s/n, the number of new infections in unit time per infective is (β n)(s/n), giving a rate of new infections (β n)(s/n)i = β si. alternatively, we may argue that for a contact by a susceptible the probability that this contact is with an infective is i/n and thus the rate of new infections per susceptible is (β n)(i/n), giving a rate of new infections (β n)(i/n)s = β si. note that both approaches give the same rate of new infections; in models with more complicated compartmental structure one may be more appropriate than the other. we need not give an algebraic expression for n, since it cancels out of the final model, but we should note that for an sir disease model, n = s + i + r. later, we will allow the possibility that some infectives recover while others die of the disease. the hypothesis (iii) really says that the time scale of the disease is much faster than the time scale of births and deaths, so that demographic effects on the population may be ignored. an alternative view is that we are interested only in studying the dynamics of a single epidemic outbreak. the assumption (ii) requires a fuller mathematical explanation, since the assumption of a recovery rate proportional to the number of infectives has no clear epidemiological meaning. we consider the "cohort" of members who were all infected at one time and let u(s) denote the number of these who are still infective s time units after having been infected. if a fraction α of these leave the infective class in unit time, then u = −αu , and the solution of this elementary differential equation is thus, the fraction of infectives remaining infective s time units after having become infective is e −αs , so that the length of the infective period is distributed exponentially with mean ∞ e −αs ds = /α, and this is what (ii) really assumes. if we assume, instead of (ii), that the fraction of infectives remaining infective a time τ after having become infective is p(τ), the second equation of ( . ) would be replaced by the integral equation where i (t) represents the members of the population who were infective at time t = and are still infective at time t. the assumptions of a rate of contacts proportional to population size n with constant of proportionality β and of an exponentially distributed recovery rate are unrealistically simple. more general models can be constructed and analyzed, but our goal here is to show what may be deduced from extremely simple models. it will turn out that that many more realistic models exhibit very similar qualitative behaviors. in our model r is determined once s and i are known, and we can drop the r equation from our model, leaving the system of two equations together with initial conditions we think of introducing a small number of infectives into a population of susceptibles and ask whether there will be an epidemic. we remark that the model makes sense only so long as s(t) and i(t) remain nonnegative. thus if either s(t) or i(t) reaches zero, we consider the system to have terminated. we observe that s < for all t and i > if and only if s > α/β . thus i increases so long as s > α/β , but since s decreases for all t, i ultimately decreases and approaches zero. if s < α/β , i decreases to zero (no epidemic), while if s > α/β , i first increases to a maximum attained when s = α/β and then decreases to zero (epidemic). the quantity β s /α is a threshold quantity, called the basic reproduction number [heesterbeek ( ) ] and denoted by r , which determines whether there is an epidemict. if r < , the infection dies out, while if r > , there is an epidemic. the definition of the basic reproduction number r is that it is the number of secondary infections caused by a single infective introduced into a wholly susceptible population of size n ≈ s over the course of the infection of this single infective. in this situation, an infective makes β n contacts in unit time, all of which are with susceptibles and thus produce new infections, and the mean infective period is /α; thus the basic reproduction number is actually β n/α rather than β s /α. another way to view this apparent discrepancy is to consider two ways in which an epidemic may begin. one way is an epidemic started by a member of the population being studied, for example by returning from travel with an infection acquired away from home. in this case we would have i > , s + i = n. a second way is for an epidemic to be started by a visitor from outside the population. in this case, we would have s = n. since ( . ) is a two-dimensional autonomous system of differential equations, the natural approach would be to find equilibria and linearize about each equilibrium to determine its stability. however, since every point with i = is an equilibrium, the system ( . ) has a line of equilibria, and this approach is not applicable (the linearization matrix at each equilibrium has a zero eigenvalue). fortunately, there is an alternative approach that enables us to analyze the system ( . ). the sum of the two equations of ( . ) is thus s + i is a nonnegative smooth decreasing function and therefore tends to a limit as t → ∞. also, it is not difficult to prove that the derivative of a nonnegative smooth decreasing function must tend to zero, and this shows that integration of the sum of the two equations of ( . ) from to ∞ gives division of the first equation of ( . ) by s and integration from to ∞ gives ) is called the final size relation. it gives a relationship between the basic reproduction number and the size of the epidemic. note that the final size of the epidemic, the number of members of the population who are infected over the course of the epidemic, is n − s ∞ . this is often described in terms of the attack rate [technically, the attack rate should be called an attack ratio, since it is dimensionless and is not a rate.] the final size relation ( . ) can be generalized to epidemic models with more complicated compartmental structure than the simple sir model ( . ), including models with exposed periods, treatment models, and models including quarantine of suspected individuals and isolation of diagnosed infectives. the original kermack-mckendrick model ( ) included dependence on the time since becoming infected (age of infection), and this includes such models. integration of the first equation of ( . ) from to t gives and this leads to the form this implicit relation between s and i describes the orbits of solutions of ( . ) in the (s, i) plane. in addition, since the right side of ( . ) is finite, the left side is also finite, and this shows that s ∞ > . the final size relation ( . ) is valid for a large variety of epidemic models, as we shall see in later sections. it is not difficult to prove that there is a unique solution of the final size relation ( . ). to see this, we define the function then, as shown in figure . , is monotone decreasing from a positive value at x = + to a negative value at x = n. thus there is a unique zero s ∞ of g(x) with s ∞ < n. is monotone decreasing from a positive value at x = + to a minimum at x = n/r and then increases to a negative value at x = n . thus there is a unique zero s ∞ of g(x) with in fact, it is generally difficult to estimate the contact rate β , which depends on the particular disease being studied but may also depend on social and behavioral factors. the quantities s and s ∞ may be estimated by serological studies (measurements of immune responses in blood samples) before and after an epidemic, and from these data the basic reproduction number r may be estimated using ( . ). this estimate, however, is a retrospective one, which can be derived only after the epidemic has run its course. the maximum number of infectives at any time is the number of infectives when the derivative of i is zero, that is, when s = α/β . this maximum is given by since detailed records were preserved and the community was persuaded to quarantine itself to try to prevent the spread of disease to other communities, the disease in eyam has been used as a case study for modeling [raggett ( ) ]. detailed examination of the data indicates that there were actually two outbreaks, of which the first was relatively mild. thus we shall try to fit the model ( . ) over the period from mid-may to mid-october , measuring time in months with an initial population of infectives and susceptibles, and a final population of . raggett ( ) gives values of susceptibles and infectives in eyam on various dates, beginning with s( ) = , i( ) = , shown in table . . the final size relation with s = , i = , s ∞ = gives β /α = . × − , α/β = . the infective period was days, or . month, so that α = . . then β = . . the relation ( . ) gives an estimate of . for the maximum number of infectives. we use the values obtained here for the parameters β and τ in the model ( . ) for simulations of both the phase plane, here the (s, i)-plane, and for graphs of s and i as functions of t (figures . , . , . ). figure . plots these data points together with the phase portrait given in figure . for the model ( . ). the actual data for the eyam epidemic are remarkably close to the predictions of this very simple model. however, the model is really too good to be true. our model assumes that infection is transmitted directly between people. while this is possible, bubonic plague is transmitted mainly by rat fleas. when an infected rat is bitten by a flea, the flea becomes extremely hungry and bites the host rat repeatedly, spreading the infection in the rat. when the host rat dies, its fleas move on to other rats, spreading the disease further. as the number of available rats decreases, the fleas move to human hosts, and this is how plague starts in a human population (although the second phase of the epidemic may have been the pneumonic form of bubonic plague, which can be spread from person to person). one of the main reasons for the spread of plague from asia into europe was the passage of many trading ships; in medieval times ships were invariably infested with rats. an accurate model of plague transmission would have to include flea and rat populations, as well as movement in space. such a model would be extremely complicated, and its predictions might well not be any closer to observations than our simple unrealistic model. very recent study of the data from eyam suggests that the rat population may not have been large enough to support the epidemic and human to human transmission was also a factor. raggett ( ) also used a stochastic model to fit the data, but the fit was rather poorer than the fit for the simple deterministic model( . ). in the village of eyam the rector persuaded the entire community to quarantine itself to prevent the spread of disease to other communities. one effect of this policy was to increase the infection rate in the village by keeping fleas, rats, and people in close contact with one another, and the mortality rate from bubonic plague was much higher in eyam than in london. further, the quarantine could do nothing to prevent the travel of rats and thus did little to prevent the spread of disease to other communities. one message this suggests to mathematical modelers is that control strategies based on false models may be harmful, and it is essential to distinguish between assumptions that simplify but do not alter the predicted effects substantially, and wrong assumptions that make an important difference. in order to prevent the occurrence of an epidemic if infectives are introduced into a population, it is necessary to reduce the basic reproductive number r below one. this may sometimes be achieved by immunization, which has the effect of transferring members of the population from the susceptible class to the removed class and thus of reducing s( ). if a fraction p of the population is successfully immunized, the effect is to replace s( ) by s( )( − p), and thus to reduce the basic reproductive number to β s( ) a large basic reproductive number means that the fraction that must be immunized to prevent the spread of infection is large. this relation is connected to the idea of herd immunity, which we shall introduce in the next chapter. initially, the number of infectives grows exponentially because the equation for i may be approximated by i = (β n − α)i and the initial growth rate is this initial growth rate r may be determined experimentally when an epidemic begins. then since n and α may be measured, β may be calculated as however, because of incomplete data and underreporting of cases, this estimate may not be very accurate. this inaccuracy is even more pronounced for an outbreak of a previously unknown disease, where early cases are likely to be misdiagnosed. because of the final size relation, estimation of β or r is an important question that has been studied by a variety of approaches. there are serious shortcomings in the simple kermack-mckendrick model as a description of the beginning of a disease outbreak, and a very different kind of model is required. exercises . the same survey of yale students described in example reported that . percent were susceptible to influenza at the beginning of the year and . percent were susceptible at the end of the year. estimate the basic reproductive number β /α and decide whether there was an epidemic. . what fraction of yale students in exercise would have had to be immunized to prevent an epidemic? . what was the maximum number of yale students in exercises and suffering from influenza at any time? . an influenza epidemic was reported at an english boarding school in that spread to of the students. estimate the basic reproductive number β /α. . what fraction of the boarding school students in exercise would have had to be immunized to prevent an epidemic? . what was the maximum number of boarding school students in exercises and suffering from influenza at any time? . a disease is introduced by two visitors into a town with inhabitants. an average infective is in contact with . inhabitants per day. the average duration of the infective period is days, and recovered infectives are immune against reinfection. how many inhabitants would have to be immunized to avoid an epidemic? . consider a disease with β = / , /α = days in a population of members. suppose the disease conferred immunity on recovered infectives. how many members would have to be immunized to avoid an epidemic? . a disease begins to spread in a population of . the infective period has an average duration of days and the average infective is in contact with . persons per day. what is the basic reproductive number? to what level must the average rate of contact be reduced so that the disease will die out? . european fox rabies is estimated to have a transmission coefficient β of km years/fox and an average infective period of days. there is a critical carrying capacity k c measured in foxes per km , such that in regions with fox density less than k c , rabies tends to die out, while in regions with fox density greater than k c , rabies tends to persist. estimate k c . [remark: it has been suggested in great britain that hunting to reduce the density of foxes below the critical carrying capacity would be a way to control the spread of rabies.] . a large english estate has a population of foxes with a density of . foxes/km . a large fox hunt is planned to reduce the fox population enough to prevent an outbreak of rabies. assuming that the contact number β /α is km /fox, find what fraction of the fox population must be caught. . following a complaint from the spca, organizers decide to replace the fox hunt of exercise by a mass inoculation of foxes for rabies. what fraction of the fox population must be inoculated to prevent a rabies outbreak? . what actually occurs on the estate of these exercises is that percent of the foxes are killed and percent are inoculated. is there danger of a rabies outbreak. . here is another approach to the analysis of the sir model ( . ). (i) divide the two equations of the model to give (ii) integrate to find the orbits in the (s, i)-plane, (iv) show that no orbit reaches the i-axis and deduce that s ∞ = lim t→∞ s(t) > , which implies that part of the population escapes infection. the kermack-mckendrick compartmental epidemic model assumes that the sizes of the compartments are large enough that the mixing of members is homogeneous, or at least that there is homogeneous mixing in each subgroup if the population is stratified by activity levels. however, at the beginning of a disease outbreak, there is a very small number of infective individuals, and the transmission of infection is a stochastic event depending on the pattern of contacts between members of the population; a description should take this pattern into account. our approach will be to give a stochastic-branching process description of the beginning of a disease outbreak to be applied as long as the number of infectives remains small, distinguishing a (minor) disease outbreak confined to this stage from a (major) epidemic, which occurs if the number of infectives begins to grow at an exponential rate. once an epidemic has started, we may switch to a deterministic compartmental model, arguing that in a major epidemic, contacts would tend to be more homogeneously distributed. implicitly, we are thinking of an infinite population, and by a major epidemic we mean a situation in which a nonzero fraction of the population is infected, and by a minor outbreak we mean a situation in which the infected population may grow but remains a negligible fraction of the population. there is an important difference between the behavior of branching process models and the behavior of models of kermack-mckendrick type, namely, as we shall see in this section that for a stochastic disease outbreak model if r < , the probability that the infection will die out is , but if r > , there is a positive probability that the infection will increase initially but will produce only a minor outbreak and will die out before triggering a major epidemic. we describe the network of contacts between individuals by a graph with members of the population represented by vertices and with contacts between individuals represented by edges. the study of graphs originated with the abstract theory of erdős and rényi of the s and s [erdős and rényi ( rényi ( , rényi ( , ]. it has become important in many areas of application, including social contacts and computer networks, as well as the spread of communicable diseases. we will think of networks as bidirectional, with disease transmission possible in either direction along an edge. an edge is a contact between vertices that can transmit infection. the number of edges of a graph at a vertex is called the degree of the vertex. the degree distribution of a graph is {p k }, where p k is the fraction of vertices having degree k. the degree distribution is fundamental in the description of the spread of disease. we think of a small number of infectives in a population of susceptibles large enough that in the initial stage, we may neglect the decrease in the size of the susceptible population. our development begins along the lines of that of [diekmann and heesterbeek ( ) ] and then develops along the lines of [callaway, newman, strogatz, and watts ( ), newman ( ), newman, strogatz, and watts ( )]. we assume that the infectives make contacts independently of one another and let p k denote the probability that the number of contacts by a randomly chosen individual is exactly k, with ∑ ∞ k= p k = . in other words, {p k } is the degree distribution of the vertices of the graph corresponding to the population network. for the moment, we assume that every contact leads to an infection, but we will relax this assumption later. it is convenient to define the probability generating function since ∑ ∞ k= p k = , this power series converges for ≤ z ≤ , and may be differentiated term by term. thus it is easy to verify that the generating function has the properties the mean degree, which we denote by k or z , is more generally, we define the moments when a disease is introduced into a network, we think of it as starting at a vertex (patient zero) that transmits infection to every individual to whom this individual is connected, that is, along every edge of the graph from the vertex corresponding to this individual. we may think of this individual as being inside the population, as when a member of a population returns from travel after being infected, or as being outside the population, as when someone visits a population and brings an infection. for transmission of disease after this initial contact we need to use the excess degree of a vertex. if we follow an edge to a vertex, the excess degree of this vertex is one less than the degree. we use the excess degree because infection cannot be transmitted back along the edge whence it came. the probability of reaching a vertex of degree k, or excess degree (k − ), by following a random edge is proportional to k, and thus the probability that a vertex at the end of a random edge has excess degree (k − ) is a constant multiple of kp k with the constant chosen to make the sum over k of the probabilities equal to . then the probability that a vertex has excess degree (k − ) is this leads to a generating function g (z) for the excess degree, and the mean excess degree, which we denote by k e , is we let r = g ( ), the mean excess degree. this is the mean number of secondary cases infected by patient zero and is the basic reproduction number as usually defined; the threshold for an epidemic is determined by r . the quantity k e = g ( ) is sometimes written in the form our next goal is to calculate the probability that the infection will die out and will not develop into a major epidemic, proceeding in two steps. first we find the probability that a secondary infected vertex (a vertex that has been infected by another vertex in the population) will not spark a major epidemic. suppose that the secondary infected vertex has excess degree j. we let z n denote the probability that this infection dies out within the next n generations. for the infection to die out in n generations, each of the j secondary infections coming from the initial secondary infected vertex must die out in (n − ) generations. the probability of this is z n− for each secondary infection, and the probability that all secondary infections will die out in (n − ) generations is z j n− . now z n is the sum over j of these probabilities, weighted by the probability q j of j secondary infections. thus since g (z) is an increasing function, the sequence z n is an increasing sequence and has a limit z ∞ , which is the probability that this infection will die out eventually. then z ∞ is the limit as n → ∞ of the solution of the difference equation thus z ∞ must be an equilibrium of this difference equation, that is, a solution of z = g (z). let w be the smallest positive solution of z = g (z). then, because it follows by induction that from this we deduce that z ∞ = w. the equation g (z) = z has a root z = , since g ( ) = . because the function g (z) − z has a positive second derivative, its derivative g (z) − is increasing and can have at most one zero. this implies that the equation g (z) = z has at most two and the equation g (z) = z has only one root, namely z = . on the other hand, if r > , the function g (z) − z is positive for z = and negative near z = since it is zero at z = , and its derivative is positive for z < and z near . thus in this case the equation g (z) = z has a second root z ∞ < . this root z ∞ is the probability that an infection transmitted along one of the edges at the initial secondary vertex will die out, and this probability is independent of the excess degree of the initial secondary vertex. it is also the probability that an infection originating outside the population, such as an infection brought from outside into the population under study, will die out. next, we calculate the probability that an infection originating at a primary infected vertex, such as an infection introduced by a visitor from outside the population under study, will die out. the probability that the disease outbreak will die out eventually is the sum over k of the probabilities that the initial infection in a vertex of degree k will die out, weighted by the degree distribution {p k } of the original infection, and this is to summarize this analysis, we see that if r < , the probability that the infection will die out is . on the other hand, and there is a probability − g (z ∞ ) > that the infection will persist, and will lead to an epidemic. however, there is a positive probability g (z ∞ ) that the infection will increase initially but will produce only a minor outbreak and will die out before triggering a major epidemic. this distinction between a minor outbreak and a major epidemic, and the result that if r > there may be only a minor outbreak and not a major epidemic, are aspects of stochastic models not reflected in deterministic models. if contacts between members of the population are random, corresponding to the assumption of mass action in the transmission of disease, then the probabilities p k are given by the poisson distribution the commonly observed situation that most infectives do not pass on infection but there are a few "superspreading events" [riley et al. ( ) ] corresponds to a probability distribution that is quite different from a poisson distribution, and could give a quite different probability that an epidemic will occur. for example, if r = . , the assumption of a poisson distribution gives z ∞ = . and g (z ∞ ) = . , so that the probability of an epidemic is . . the assumption that nine out of ten infectives do not transmit infection while the tenth transmits infections gives from which we see that the probability of an epidemic is . . another example, possibly more realistic, is to assume that a fraction ( − p) of the population follows a poisson distribution with constant r, while the remaining fraction p consists of superspreaders each of whom makes l contacts. this would give a generating function for example, if r = . , l = , p = . , numerical simulation gives so that the probability of an epidemic is . . these examples demonstrate that the probability of a major epidemic depends strongly on the nature of the contact network. simulations suggest that for a given value of the basic reproduction number, the poisson distribution is the one with the maximum probability of a major epidemic. it has been observed that in many situations there is a small number of long-range connections in the graph, allowing rapid spread of infection. there is a high degree of clustering (some vertices with many edges), and there are short path lengths. such a situation may arise if a disease is spread to a distant location by an air traveler. this type of network is called a small-world network. long range connections in a network can increase the likelihood of an epidemic dramatically. these examples indicate that the probability of an epidemic depends strongly on the contact network at the beginning of a disease outbreak. we will not explore network models further here, but we point out that this is an actively developing field of science. some basic references are [newman ( ( , strogatz ( ]. contacts do not necessarily transmit infection. for each contact between individuals of whom one has been infected and the other is susceptible, there is a probability that infection will actually be transmitted. this probability depends on such factors as the closeness of the contact, the infectivity of the member who has been infected, and the susceptibility of the susceptible member. we assume that there is a mean probability t , called the transmissibility, of transmission of infection. the transmissibility depends on the rate of contacts, the probability that a contact will transmit infection, the duration time of the infection, and the susceptibility. the development in section . assumed that all contacts transmit infection, that is, that t = . in this section, we will continue to assume that there is a network describing the contacts between members of the population whose degree distribution is given by the generating function g (z), but we will assume in addition that there is a mean transmissibility t . when disease begins in a network, it spreads to some of the vertices of the network. edges that are infected during a disease outbreak are called occupied, and the size of the disease outbreak is the cluster of vertices connected to the initial vertex by a continuous chain of occupied edges. the probability that exactly m infections are transmitted by an infective vertex of degree k is we define Γ (z, t )to be the generating function for the distribution of the number of occupied edges attached to a randomly chosen vertex, which is the same as the distribution of the infections transmitted by a randomly chosen individual for any (fixed) transmissibility t . then in this calculation we have used the binomial theorem to see that note that for secondary infections we need the generating function Γ (z, t ) for the distribution of occupied edges leaving a vertex reached by following a randomly chosen edge. this is obtained from the excess degree distribution in the same way, the basic reproduction number is now the calculation of the probability that the infection will die out and will not develop into a major epidemic follows the same lines as the argument for t = . the result is that if r = t g ( ) < , the probability that the infection will die out is . and a probability −Γ (z ∞ (t ), t ) > that the infection will persist, and will lead to an epidemic. however, there is a positive probability Γ (z ∞ (t ), t ) that the infection will increase initially but will produce only a minor outbreak and will die out before triggering a major epidemic. another interpretation of the basic reproduction number is that there is a critical transmissibility t c defined by in other words, the critical transmissibility is the transmissibility that makes the basic reproduction number equal to . if the mean transmissibility can be decreased below the critical transmissibility, then an epidemic can be prevented. the measures used to try to control an epidemic may include contact interventions, that is, measures affecting the network such as avoidance of public gatherings and rearrangement of the patterns of interaction between caregivers and patients in a hospital, and transmission interventions such as careful hand washing or face masks to decrease the probability that a contact will lead to disease transmission. in each exercise, assume that the transmissibility is . . show that it is not possible for a major epidemic to develop unless at least one member of the contact network has degree at least . . what is the probability of a major epidemic if every member of the contact network has degree . estimate (numerically) the probability of a major epidemic if c = . . . show that the probability generating function for an exponential distribution, given by for what values of α is it possible to normalize this (i.e., choose c to make ∑ p k = ? compartmental models for epidemics are not suitable for describing the beginning of a disease outbreak because they assume that all members of a population are equally likely to make contact with a very small number of infectives. thus, as we have seen in the preceding section, stochastic branching process models are better descriptions of the beginning of an epidemic. they allow the possibility that even if a disease outbreak has a reproduction number greater than , it may be only a minor outbreak and may not develop into a major epidemic. one possible approach to a more realistic description of an epidemic would be to use a branching process model initially and then make a transition to a compartmental model when the epidemic has become established and there are enough infectives that mass action mixing in the population is a reasonable approximation. another approach would be to continue to use a network model throughout the course of the epidemic. in this section we shall indicate how a compartmental approach and a network approach are related. we assume that there is a known static configuration model (cm) network in which the probability that a node u has degree k u is p(k u )). we let g (z) denote the probability generating function of the degree distribution, the per-edge from an infected node is assumed to be β , and it is assumed that infected nodes recover at a rate α. we use an edge-based compartmental model because the probability that a random neighbor is infected is not necessarily the same as the probability that a random individual is infected. we let s(t) denote the fraction of nodes that are susceptible at time t, i(t) the fraction of nodes that are infective at time t, and r(t) the fraction of nodes that are recovered at time t. it is easy to write an equation for r , the rate at which infectives recover. if we know s(t), we can find i(t), because a decrease in s gives a corresponding increase in i. since we need only find the probability that a randomly selected node is susceptible. we assume that the hazard of infection for a susceptible node u is proportional to the degree k u of the node. each contact is represented by an edge of the network joining u to a neighboring node. we let ϕ i denote the probability that this neighbor is infective. then the per-edge hazard of infection is assuming that edges are independent, u's hazard of infection at time t is consider a randomly selected node u and let θ (t) be the probability that a random neighbor has not transmitted infection to u. then the probability that u is susceptible is θ k u . averaging over all nodes, we see that the probability that a random node u is susceptible is ( . ) we break θ into three parts, with ϕ s the probability that a random neighbor v of u is susceptible, ϕ i the probability that a random neighbor v of u is infective but has not transmitted infection to u, and ϕ r the probability that a random neighbor v has recovered without transmitting infection to u. then the probability that v has transmitted infection to u is − θ . since infected neighbors recover at rate α, the flux from ϕ i to ϕ r is αϕ i . thus it is easy to see from this that r = αi. ( . ) since edges from infected neighbors transmit infection at rate β , the flux from to obtain ϕ i we need the flux into and out of the ϕ i compartment. the incoming flux from ϕ s results from infection of the neighbor. the outgoing flux to ϕ r corresponds to recovery of the neighbor without having transmitted infection, and the outgoing flux to ( −θ ) corresponds to transmission without recovery. the total outgoing flux is (α + β )ϕ i . to determine the flux from ϕ s to ϕ i , we need the rate at which a neighbor changes from susceptible to infective. consider a random neighbor v of u; the probability that v has degree k iskp(k)/ k . since there are (k − ) neighbors of v that could have infected v, the probability that v is susceptible is θ k− . averaging over all k, we see that the probability that a random neighbor v of u is susceptible is to calculate ϕ r , we note that the flux from ϕ i to ϕ r and the flux from ϕ i to ( −θ ) are proportional with proportionality constant α/β . since both ϕ r and ( − θ ) start at zero, now, using ( . ), ( . ), ( . ), and we now have a dynamic model consisting of equations ( . ), ( . ), ( . ), and s + i + r = . we wish to show a relationship between this set of equations and the simple kermack-mckendrick compartmental model ( . ). in order to accomplish this, we need only show under what conditions we would have s = −β si. differentiating ( . ) and using ( . ), we obtain consider a large population with n members, each making c ≤ n − contacts, so that we now let c → ∞ (which implies n → ∞) in such a way that we will now show that ϕ i θ ≈ , and this will yield the desired approximation the probability that an edge to a randomly chosen node has not transmitted infection is θ (assuming that the given target node cannot transmit infection), and the probability that in addition it is connected to an infected node is ϕ i . becausê β = βc is constant and therefore bounded as c grows, only a fraction no greater than a constant multiple of i/c of edges to the target node may have transmitted infection from a node that is still infected. for large values of c, ϕ i is approximately i. similarly, θ is approximately as c → ∞. thus ϕ i /θ ≈ i as c → ∞. this gives the desired approximate equation for s. the result remains valid if all degrees are close to the average degree as the average degree grows. the edge-based compartmental modeling approach that we have used can be generalized in several ways. for example, heterogeneity of mixing can be included. in general, one would expect that early infections would be in individuals having more contacts, and thus that an epidemic would develop more rapidly than a mass action compartmental model would predict. when contact duration is significant, as would be the case in sexually transmitted diseases, an individual with a contact would play no further role in disease transmission until a new contact is made, and this can be incorporated in a network model. the network approach to disease modeling is a rapidly developing field of study, and there will undoubtedly be fundamental developments in our understanding of the modeling of disease transmission. in the remainder of this chapter, we assume that we are in an epidemic situation following a disease outbreak that has been modeled initially by a branching process. thus we return to the study of compartmental models. we have established that the simple kermack-mckendrick epidemic model ( . ) has the following basic properties: . there is a basic reproduction number r such that if r < , the disease dies out while if r > , there is an epidemic. . the number of infectives always approaches zero and the number of susceptibles always approaches a positive limit as t → ∞. . there is a relationship between the reproduction number and the final size of the epidemic, which is an equality if there are no disease deaths. in fact, these properties hold for epidemic models with more complicated compartmental structure. we will describe some common epidemic models as examples. in many infectious diseases there is an exposed period after the transmission of infection from susceptibles to potentially infective members but before these potential infectives develop symptoms and can transmit infection. to incorporate an exposed period with mean exposed period /κ, we add an exposed class e and use compartments s, e, i, r and total population size n = s + e + i + r to give a generalization of the epidemic model ( . ) ( . ) a flow chart is shown in figure . . the analysis of this model is the same as the analysis of ( . ), but with i replaced by e + i. that is, instead of using the number of infectives as one of the variables, we use the total number of infected members, whether or not they are capable of transmitting infection. in some diseases there is some infectivity during the exposed period. this may be modeled by assuming infectivity reduced by a factor ε during the exposed period. a calculation of the rate of new infections per susceptible leads to a model we take initial conditions for this model, integration of the sum of the equations of ( . ) from to ∞ gives integration of the third equation of ( . ) gives and division of the first equation of ( . ) by s followed by integration from to ∞ gives in this final size relation there is an initial term β i /α, caused by the assumption that there are individuals infected originally who are beyond the exposed stage in which they would have had some infectivity. in order to obtain a final size relation without such an initial term it is necessary to assume i( ) = , that initial infectives are in the first stage in which they can transmit infection. if i( ) = , the final size relation has the form ( . ). one form of treatment that is possible for some diseases is vaccination to protect against infection before the beginning of an epidemic. for example, this approach is commonly used for protection against annual influenza outbreaks. a simple way to model this would be to reduce the total population size by the fraction of the population protected against infection. in reality, such inoculations are only partly effective, decreasing the rate of infection and also decreasing infectivity if a vaccinated person does become infected. this may be modeled by dividing the population into two groups with different model parameters, which would require some assumptions about the mixing between the two groups. this is not difficult, but we will not explore this direction here. if there is a treatment for infection once a person has been infected, this may be modeled by supposing that a fraction γ per unit time of infectives is selected for treatment, and that treatment reduces infectivity by a fraction δ . suppose that the rate of removal from the treated class is η. this leads to the sitr model, where t is the treatment class, given by a flow chart is shown in figure . . it is not difficult to prove, much as was done for the model ( . ), that in order to calculate the basic reproduction number, we may argue that an infective in a totally susceptible population causes β n new infections in unit time, and the mean time spent in the infective compartment is /(α + γ). in addition, a fraction γ/(α + γ) of infectives are treated. while in the treatment stage the number of new infections caused in unit time is δ β n, and the mean time in the treatment class is /η. thus r is it is also possible to establish the final size relation ( . ) by means very similar to those used for the simple model ( . ). we integrate the first equation of ( . ) to obtain integration of the third equation of ( . ) gives integration of the sum of the first two equations of ( . ) gives combination of these three equations and ( . ) gives ( . ). in some diseases, such as influenza, at the end of a stage individuals may proceed to one of two stages. there is a latent period after which a fraction p of latent individuals l proceeds to an infective stage i, while the remaining fraction ( − p) proceeds to an asymptomatic stage a, with infectivity reduced by a factor δ and a different period /η. a flow chart is shown in figure . . the model ( . ) is an example of a differential infectivity model. in such models, also used in the study of hiv/aids [hyman, li and stanley ( )], individuals enter a specific group when they become infected and stay in that group over the course of the infection. different groups may have different parameter values. for example, for influenza infective and asymptomatic members may have different infectivities and different periods of stay in the respective stages. . exposed members may be infective with infectivity reduced by a factor ε e , ≤ ε e < . . exposed members who are not isolated become infective at rate κ e . . we introduce a class q of quarantined members and a class j of isolated (hospitalized) members, and exposed members are quarantined at a proportional rate γ q in unit time (in practice, a quarantine will also be applied to many susceptibles, but we ignore this in the model). quarantine is not perfect, but it reduces the contact rate by a factor ε q . the effect of this assumption is that some susceptibles make fewer contacts than the model assumes. . infectives are diagnosed at a proportional rate γ j per unit time and isolated. isolation is imperfect, and there may be transmission of disease by isolated members, with an infectivity factor of ε j . . quarantined members are monitored, and when they develop symptoms at rate κ q they are isolated immediately. . infectives leave the infective class at rate α i and isolated members leave the isolated class at rate α j . these assumptions lead to the seqijr model [gumel et al. ( )]: the model before control measures are begun is the special case of ( . ). it is the same as ( . ). a flow chart is shown in figure . . we define the control reproduction number r c to be the number of secondary infections caused by a single infective in a population consisting essentially only of susceptibles with the control measures in place. it is analogous to the basic reproduction number, but instead of describing the very beginning of the disease outbreak it describes the beginning of the recognition of the epidemic. the basic reproduction number is the value of the control reproduction number with we have already calculated r for ( . ), and we may calculate r c in the same way but using the full model with quarantined and isolated classes. we obtain each term of r c has an epidemiological interpretation. the mean duration in e is /d with contact rate ε e β , giving a contribution to r c of ε e β n/d . a fraction κ e /d goes from e to i, with contact rate β and mean duration /d , giving a contribution of β nκ e /d d . a fraction γ q /d goes from e to q, with contact rate ε e ε q β and mean duration /κ q , giving a contribution of ε e ε q β nγ q /d κ q . a fraction κ e γ j /d d goes from e to i to j, with a contact rate of ε j β and a mean duration of /α j , giving a contribution of ε j β nκ e γ j /α j d d . finally, a fraction γ q /d goes from e to q to j with a contact rate of ε j β and a mean duration of /α j , giving a contribution of ε j β nγ q /d α j . the sum of these individual contributions gives r c . in the model ( . ) the parameters γ q and γ j are control parameters, which may be chosen in the attempt to manage the epidemic. the parameters ε q and ε j depend on the strictness of the quarantine and isolation processes and are thus also control measures in a sense. the other parameters of the model are specific to the disease being studied. while they are not variable, their measurements are subject to experimental error. the linearization of ( . ) at the disease-free equilibrium (n, , , , ) has matrix ⎡ the corresponding characteristic equation is a fourth-degree polynomial equation whose leading coefficient is and whose constant term is a positive constant multiple of − r c , thus positive if r c < and negative if r c > . if r c > , there is a positive eigenvalue, corresponding to an initial exponential growth rate of solutions of ( . ). if r c < , it is possible to show that all eigenvalues of the coefficient matrix have negative real part, and thus solutions of ( . ) die out exponentially [van den driessche and watmough ( )]. in order to show that analogues of the relation ( . ) and s ∞ > derived for the model ( . ) are valid for the management model ( . ), we begin by integrating the equations for s + e, q, i, j, of ( . ) with respect to t from t = to t = ∞, using the initial conditions we continue by integrating the equation for s, and then an argument similar to the one used for ( . ) but technically more complicated may be used to show that s ∞ > for the treatment model ( . ) and also to establish the final size relation thus the asymptotic behavior of the management model ( . ) is the same as that of the simpler model ( . ). in the various compartmental models that we have studied, there are significant common features. this suggests that compartmental models can be put into a more general framework. in fact, this general framework is the age of infection epidemic model originally introduced by kermack and mckendrick in . however, we will not explore this generalization here. these models represent an sir epidemic model and an seir epidemic model respectively with a mean infective period of days and a mean exposed period of days. do numerical simulations to decide whether the exposed period affects the behavior of the model noticeably. use the parameter values if you feel really ambitious, formulate and analyze an seir model with infectivity in the exposed period and treatment. . consider an sir model in which a fraction θ of infectives is isolated in a perfectly quarantined class q with standard incidence (meaning that individuals make a contacts in unit time of which a fraction i/(n − q) are infective), given by the system . isolation/quarantine is a complicated process because we don't live in a perfect world. in hospitals, patients may inadvertently or deliberately break from isolation and in the process have casual contacts with others including medical personnel and visitors. taking this into account, we are led to the model (i) determine all the parameters in the system and define each parameter. (ii) show that the population is constant. (iii) find all equilibria. (iv) find the reproductive number r . (v) describe the asymptotic behavior of the model, including its dependence on the basic reproduction number. . formulate a model analogous to ( . ) for which treatment is not started immediately, but begins at time τ > . can you say anything about the dependence of the reproduction number on τ? in the simple model ( . ) studied in section . we have assumed that the infective period is exponentially distributed. now let us consider an sir epidemic model in a population of constant size n with mass action incidence in which p(τ) is the fraction of individuals who are still infective a time τ after having become infected. the model is ( . ) here, i (t) is the number of individuals who were infective initially at t = who are still infective at time t. then because if all initial infectives were newly infected we would have equality in this relation, and if some initial infectives had been infected before the starting time t = , they would recover earlier. we assume that p(τ) is a nonnegative, nonincreasing function with p( ) = . we assume also that the mean infective period ∞ p(τ)dτ is finite. since a single infective causes β n new infections in unit time and ∞ p(τ)dτ is the mean infective period, it is easy to calculate since s is a nonnegative decreasing function, it follows as for ( . ) that s(t) deceases to a limit s ∞ as t → ∞, but we must proceed differently to show that i(t) → . this will follow if we can prove that t i(s) ds is bounded as t → ∞. we have since ∞ p(τ)dτ is assumed to be finite, it follows that t i(s) ds is bounded, and thence that i(t) → . now integration of the first equation in ( . ) from to ∞ gives and this shows that s ∞ > . if all initially infected individuals are newly infected, so that i (t) = (n −s )p(t), integration of the second equation of ( . ) gives and this is the final size relation, identical to ( . ). if there are individuals who were infected before time t = , a positive term the general epidemic model described by kermack and mckendrick ( ) included a dependence of infectivity on the time since becoming infected (age of infection). we let s(t) denote the number of susceptibles at time t and let ϕ(t) be the total infectivity at time t, defined as the sum of products of the number of infected members with each infection age and the mean infectivity for that infection age. we assume that on average, members of the population make a constant number a of contacts in unit time. we let b(τ) be the fraction of infected members remaining infected at infection age τ and let π(τ) with ≤ π(τ) ≤ be the mean infectivity at infection age τ. then we let the mean infectivity of members of the population with infection age τ. we assume that there are no disease deaths, so that the total population size is a constant n. the basic reproduction number is integration with respect to t from to ∞ gives here, ϕ (t) is the total infectivity of the initial infectives when they reach age of infection t. if all initial infectives have infection age zero at t = , then ϕ then ( . ) takes the form and this is the general final size relation. if there are initial infectives with infection age greater than zero, let u(τ) be the fraction of these individuals with infection age τ, ∞ u(τ)dτ = . at time t these individuals have infection age t + τ and mean infectivity a(t + τ). thus thus, the initial term satisfies the final size relation is sometimes presented in the form example . the seir model ( . ) can be viewed as an age of infection model with ϕ = εe + i. to use the age of infection interpretation, we need to determine the kernel a(τ) in order to calculate its integral. we let u(τ) be the fraction of infected members with infection age τ who are not yet infective and v(τ) the fraction of infected members who are infective. then the rate at which members become infective at infection age τ is κu(τ), and we have the solution of this system is and it is easy to calculate this gives the same value for r as was calculated directly. the age of infection model also includes the possibility of disease stages with distributions that are not exponential [feng ( ), feng, xu, and zhao ( )]. example . consider an seir model in which the exposed stage has an exponential distribution but the infective stage has a period distribution given by a function p, ( . ) with initial conditions if we define u(τ), v(τ) as in example , we again obtain u(τ) = e −κτ , and v satisfies for period distributions that are not exponential, it is possible to calculate ∞ a(τ)dτ without having to calculate the function a(τ) explicitly. example . consider an seir model in which the exposed period has a distribution given by a function q and the infective period has a distribution given by a function p. then in order to obtain an equation for i, we differentiate the equation for e, obtaining thus the input to i at time t is and the first term in this expression may be written as i (t), and the second term may be simplified, using interchange of the order of integration in the iterated integral, to yield we obtain then the model is which is in age of infection form with ϕ = i, and we have an explicit expression for a(τ). . interpret the models ( . ), ( . ), and ( . ) introduced earlier as age of infection models and use this interpretation to calculate their reproduction numbers. . calculate the basic reproduction number for the model ( . ) but with infectivity in the exposed class having a reduction factor ε. the assumption in the model ( . ) of a rate of contacts per infective that is proportional to population size n, called mass action incidence or bilinear incidence, was used in all the early epidemic models. however, it is quite unrealistic, except possibly in the early stages of an epidemic in a population of moderate size. it is more realistic to assume a contact rate that is a nonincreasing function of total population size. for example, a situation in which the number of contacts per infective in unit time is constant, called standard incidence, is a more accurate description for sexu-ally transmitted diseases. if there are no disease deaths, so that the total population size remains constant, such a distinction is unnecessary. we generalize the model ( . ) by dropping assumption (iv) and replacing assumption (i) by the assumption that an average member of the population makes c(n) contacts in unit time with c (n) ≥ [castillo-chavez, cooke, huang, and levin ( a), dietz ( )], and we define it is reasonable to assume β (n) ≤ to express the idea of saturation in the number of contacts. then mass action incidence because the total population size is now present in the model, we must include an equation for total population size in the model. this forces us to make a distinction between members of the population who die of the disease and members of the population who recover with immunity against reinfection. we assume that a fraction f of the αi members leaving the infective class at time t recover and the remaining fraction ( − f ) die of disease. we use s, i, and n as variables, with n = s + i + r. we now obtain a three-dimensional model ( . ) since n is now a decreasing function, we define n( ) = n = s + i . we also have the equation r = − f αi, but we need not include it in the model, since r is determined when s, i, and n are known. we should note that if f = , the total population size remains equal to the constant n, and the model ( . ) reduces to the simpler model ( . ) with β replaced by the constant β (n ). we wish to show that the model ( . ) has the same qualitative behavior as the model ( . ), namely that there is a basic reproduction number that distinguishes between disappearance of the disease and an epidemic outbreak, and that some members of the population are left untouched when the epidemic passes. these two properties are the central features of all epidemic models. for the model ( . ) the basic reproduction number is given by because a single infective introduced into a wholly susceptible population makes c(n ) = n β (n ) contacts in unit time, all of which are with susceptibles and thus produce new infections, and the mean infective period is /α. we assume that β ( ) is finite, thus ruling out standard incidence (standard incidence does not appear to be realistic if the total population n approaches zero, and it would be more natural to assume that c(n) grows linearly with n for small n). if we let t → ∞ in the sum of the first two equations of ( . ), we obtain the first equation of ( . ) may be written as − s (t) s(t) = β (n(t))i(t). since we now obtain a final size inequality if the disease death rate is small, the final size inequality is an approximate equality. it is not difficult to show that n(t) ≥ f n , and then a similar calculation using the inequality β (n) ≤ β ( f n ) < ∞ shows that from which we may deduce that s ∞ > . . for the model ( . ) show that the final total population size is given by to cope with annual seasonal influenza epidemics there is a program of vaccination before the "flu" season begins. each year, a vaccine is produced aimed at protecting against the three influenza strains considered most dangerous for the coming season. we formulate a model to add vaccination to the simple sir model ( . ) under the assumption that vaccination reduces susceptibility (the probability of infection if a contact with an infected member of the population is made). we consider a population of total size n and assume that a fraction γ of this population is vaccinated prior to a disease outbreak. thus we have a subpopulation of size n u = ( − γ)n of unvaccinated members and a subpopulation of size n v = γn of vaccinated members. we assume that vaccinated members have susceptibility to infection reduced by a factor σ , ≤ σ ≤ , with σ = describing a perfectly effective vaccine and σ = describing a vaccine that has no effect. we assume also that vaccinated individuals who are infected have infectivity reduced by a factor δ and may also have a recovery rate α v that is different from the recovery rate of infected unvaccinated individuals α u . we let s u , s v , i u , i v denote the number of unvaccinated susceptibles, the number of vaccinated susceptibles, the number of unvaccinated infectives, and the number of vaccianted infectives respectively. the resulting model is the initial conditions prescribe s u ( ), s v ( ), i u ( ), i v ( ), with since the infection now is beginning in a population that is not fully susceptible, we speak of the control reproduction number r c rather than the basic reproduction number. however, as we will soon see, calculation of the control reproduction number will require a more general definition and a considerable amount of technical computation. the computation method is applicable to both basic and control reproduction numbers. we will use the term reproduction number to denote either a basic reproduction number or a control reproduction number. we are able to obtain final size relations without knowledge of the reproduction number, but these final size relations do contain information about the reproduction number, and more. since s u and s v are decreasing nonnegative functions they have limits s u (∞) and s v (∞) respectively as t → ∞. the sum of the equations for s u and i u in ( . ) is from which we conclude, just as in the analysis of ( . ), that i u (t) → as t → ∞, and that α similarly, using the sum of the equations for s v and i v , we see that integration of the equation for s u in ( . ) and use of (??) gives a similar calculation using the equation for s v gives this pair of equations ( . ), ( . ) are the final size relations. they make it possible to calculate s u (∞), s v (∞) if the parameters of the model are known. it is convenient to define the matrix then the final size relations ( . ), ( . ) may be written the matrix k is closely related to the reproduction number. in the next section we describe a general method for calculating reproduction numbers that will involve this matrix. . suppose we want to model the spread of influenza in a city using an sliar model (susceptible-latent-infectious-asympotmatic-recovered, respectively). then our system of equation would be where β is the transmission coefficient, δ is the reduced transmissibility factor from asymptomatic contacts, κ is the rate of disease progression from the latent class, p is the proportion of individuals that are clinically diagnosed, η is the recovery rate from the asymptomatic class, γ is the recovery rate from the infectious (clinically diagnosed) class, and n is the total population size. (i) add a vaccination class to the model. assume that the vaccine imparts partial protection until it becomes fully effective. is the population of the new system constant? are there any endemic equilibria? (ii) vary the vaccination rate from . to . and determine how the number of infected individuals changes compared with the model without vaccination. does vaccination prevent the outbreak? up to this point, we have calculated reproduction numbers by following the secondary cases caused by a single infective introduced into a population. however, if there are subpopulations with different susceptibilities to infection, as in the vaccination model introduced in section . , it is necessary to follow the secondary infections in the subpopulations separately, and this approach will not yield the re-production number. it is necessary to give a more general approach to the meaning of the reproduction number, and this is done through the next generation matrix [diekmann and heesterbeek ( ) , diekmann, heesterbeek, and metz ( ), van den driessche and watmough ( )]. the underlying idea is that we must calculate the matrix whose (i, j) entry is the number of secondary infections caused in compartment i by an infected individual in compartment j. the procedure applies both to epidemic models, as studied in this chapter, and to models with demographics for endemic diseases, to be studied in the next chapter. in a compartmental disease transmission model we sort individuals into compartments based on a single, discrete state variable. a compartment is called a disease compartment if the individuals therein are infected. note that this use of the term disease is broader than the clinical definition and includes stages of infection such as exposed stages in which infected individuals are not necessarily infective. suppose there are n disease compartments and m nondisease compartments, and let x ∈ r n and y ∈ r m be the subpopulations in each of these compartments. further, we denote by f i the rate at which secondary infections increase the i − th disease compartment and by v i the rate at which disease progression, death, and recovery decrease the i − th compartment. the compartmental model can then be written in the form note that the decomposition of the dynamics into f and v and the designation of compartments as infected or uninfected may not be unique; different decompositions correspond to different epidemiological interpretations of the model. the definitions of f and v used here differ slightly from those in [van den driessche and watmough ( )]. the derivation of the basic reproduction number is based on the linearization of the ode model about a disease-free equilibrium. for an epidemic model with a line of equilibria, it is customary to use the equilibrium with all members of the population susceptible. we assume: • f i ( , y) = and v i ( , y) = for all y ≥ and i = ,..., n. • the disease-free system y = g( , y) has a unique equilibrium that is asymptotically stable, that is, all solutions with initial conditions of the form ( , y) approach a point ( , y o ) as t → ∞. we refer to this point as the disease-free equilibrium. the first assumption says that all new infections are secondary infections arising from infected hosts; there is no immigration of individuals into the disease compartments. it ensures that the disease-free set, which consists of all points of the form ( , y), is invariant. that is, any solution with no infected individuals at some point in time will be free of infection for all time. the second assumption ensures that the disease-free equilibrium is also an equilibrium of the full system. the uniqueness of the disease-free equilibrium in the second assumption is required for models with demographics, to be studied in the next chapter. although it is not satisfied in epidemic models, the specification of a specific disease-free equilibrium with all memebers of the population susceptible is sufficient to validate the results. next, we assume: • f i (x, y) ≥ for all nonnegative x and y and i = ,..., n. y) ≥ for all nonnegative x and y. the reasons for these assumptions are that the function f represents new infections and cannot be negative, each component v i represents a net outflow from compartment i and must be negative (inflow only) whenever the compartment is empty, and the sum ∑ n i= v i (x, y) represents the total outflow from all infected compartments. terms in the model leading to increases in ∑ n i= x i are assumed to represent secondary infections and therefore belong in f . suppose that a single infected person is introduced into a population originally free of disease. the initial ability of the disease to spread through the population is determined by an examination of the linearization of ( . ) about the disease-free equilibrium ( , y ). it is easy to see that the assumption for every pair (i, j). this implies that the linearized equations for the disease compartments x are decoupled from the remaining equations and can be written as where f and v are the n × n matrices with entries because of the assumption that the disease-free system y = g( , y) has a unique asymptotically stable equilibrium, the linear stability of the system ( . ) is completely determined by the linear stability of the matrix (f −v ) in ( . ). the number of secondary infections produced by a single infected individual can be expressed as the product of the expected duration of the infectious period and the rate at which secondary infections occur. for the general model with n disease compartments, these are computed for each compartment for a hypothetical index case. the expected time the index case spends in each compartment is given by is the solution of ( . ) with f = (no secondary infections) and nonnegative initial conditions x representing an infected index case: in effect, this solution shows the path of the index case through the disease compartments from the initial exposure through to death or recovery with the i − th component of ϕ(t, x ) interpreted as the probability that the index case (introduced at time t = ) is in disease state i at time t. the solution of ( . ) is φ (t, x ) = e −v t x , where the exponential of a matrix is defined by the taylor series this series converges for all t (see, for example, [hirsch and smale ( ) ]. thus ∞ ϕ(t, x ) dt = v − x , and the (i, j) entry of the matrix v − can be interpreted as the expected time an individual initially introduced into disease compartment j spends in disease compartment i. the (i, j) entry of the matrix f is the rate at which secondary infections are produced in compartment i by an index case in compartment j. hence, the expected number of secondary infections produced by the index case is given by following diekmann and heesterbeek ( ), the matrix k = fv − is referred to as the next generation matrix for the system at the disease-free equilibrium. the (i, j) entry of k is the expected number of secondary infections in compartment i produced by individuals initially in compartment j, assuming, of course, that the environment experienced by the individual remains homogeneous for the duration of its infection. shortly, we will describe some results from matrix theory that imply that the matrix k l = fv − , called the next generation matrix with small domain, is nonnegative and therefore has a nonnegative eigenvalue, r = ρ(fv − ), such that there are no other eigenvalues of k with modulus greater than r and there is a nonnegative eigenvector ω associated with r [berman and plemmons ( ), theorem . . ]. this eigenvector is in a sense the distribution of infected individuals that produces the greatest number r of secondary infections per generation. thus, r and the associated eigenvector ω suitably define a "typical" infective, and the basic reproduction number can be rigorously defined as the spectral radius of the matrix k l . the spectral radius of a matrix k l , denoted by ρ(k l ), is the maximum of the moduli of the eigenvalues of k l . if k l is irreducible, then r is a simple eigenvalue of k l and is strictly larger in modulus than all other eigenvalues of k l . however, if k l is reducible, which is often the case for diseases with multiple strains, then k l may have several positive real eigenvectors corresponding to reproduction numbers for each competing strain of the disease. we have interpreted the reproduction number for a disease as the number of secondary infections produced by an infected individual in a population of susceptible individuals. if the reproduction number r = ρ(fv − ) is consistent with the differential equation model, then it should follow that the disease-free equilibrium is asymptotically stable if r < and unstable if r > . this is shown through a sequence of lemmas. the spectral bound (or abscissa) of a matrix a is the maximum real part of all eigenvalues of a. if each entry of a matrix t is nonnegative, we write t ≥ and refer to t as a nonnegative matrix. a matrix of the form a = si − b, with b ≥ , is said to have the z sign pattern. these are matrices whose off-diagonal entries are negative or zero. if in addition, s ≥ ρ(b), then a is called an m-matrix. note that in this section, i denotes an identity matrix, not a population of infectious individuals. the following lemma is a standard result from [berman and plemmons ( )]. lemma . . if a has the z sign pattern, then a − ≥ if and only if a is a nonsingular m-matrix. the assumptions we have made imply that each entry of f is nonnegative and that the off-diagonal entries of v are negative or zero. thus v has the z sign pattern. also, the column sums of v are positive or zero, which, together with the z sign pattern, implies that v is a (possibly singular) m-matrix [berman and plemmons ( ), condition m of theorem . . ]. in what follows, it is assumed that v is nonsingular. in this case, v − ≥ , by lemma . . hence, k l = fv − is also nonnegative. theorem . . consider the disease transmission model given by ( . ). the diseasefree equilibrium of ( . ) is locally asymptotically stable if r < , but unstable if r > . proof. let f and v be as defined as above, and let j and j be the matrices of partial derivatives of g with respect to x and y evaluated at the disease-free equilibrium. the jacobian matrix for the linearization of the system about the disease-free equilibrium has the block structure the disease-free equilibrium is locally asymptotically stable if the eigenvalues of the jacobian matrix all have negative real parts. since the eigenvalues of j are those of (f −v ) and j , and the latter all have negative real parts by assumption, the diseasefree equilibrium is locally asymptotically stable if all eigenvalues of (f − v ) have negative real parts. by the assumptions on f and v , f is nonnegative and v is a nonsingular m-matrix. hence, by lemma all eigenvalues of (f −v ) have negative real parts if and only if ρ(fv − ) < . it follows that the disease-free equilibrium is locally asymptotically stable if r = ρ(fv − ) < . instability for r > can be established by a continuity argument. if r ≤ , then for any ε > , (( + ε)i − fv − ) is a nonsingular m-matrix and by lemma . , (( + ε)i − fv − ) − ≥ . by lemma . , all eigenvalues of (( + ε)v − f) have positive real parts. since ε > is arbitrary, and eigenvalues are continuous functions of the entries of the matrix, it follows that all eigenvalues of (v − f) have nonnegative real parts. to reverse the argument, suppose all the eigenvalues of (v − f) have nonnegative real parts. for any positive ε, (v + εi − f) is a nonsingular m-matrix, and by lemma . , ρ(f(v + εi) − ) < . again, since ε > is arbitrary, it follows that ρ(fv − ) ≤ . thus, (f −v ) has at least one eigenvalue with positive real part if and only if ρ(fv − ) > , and the disease-free equilibrium is unstable whenever r > . these results validate the extension of the definition of the reproduction number to more general situations. in the vaccination model ( . ) of the previous section we calculated a pair of final size relations that contained the elements of a matrix k. this matrix is precisely the next generation matrix with large domain k l = fv − that has been introduced in this section. example . consider the seir model with infectivity in the exposed stage, ( . ) here the disease states are e and i, then we may calculate since fv − has rank , it has only one nonzero eigenvalue, and since the trace of the matrix is equal to the sum of the eigenvalues, it is easy to see that the element in the first row and first column of fv − . if all new infections are in a single compartment, as is the case here, the basic reproduction number is the trace of the matrix fv − . in general, it is possible to reduce the size of the next generation matrix to the number of states at infection [diekmann and heesterbeek ( )]. the states at infection are those disease states in which there can be new infections. suppose that there are n disease states and k states at infection with k < n. then we may define an auxiliary n×k matrix e in which each column corresponds to a state at infection and has in the corresponding row and elsewhere. then the next generation matrix is it is easy to show, using the fact that ee t k l = k l , that the n × n matrix k l and the k × k matrix k have the same nonzero eigenvalues and therefore the same spectral radius. construction of the next generation matrix that has lower dimension than the next generation matrix with large domain may simplify the calculation of the basic reproduction number. in example above, the only disease state at infection is e, the matrix a is , and the next generation matrix k is the × matrix example . consider the vaccination model ( . ). the disease states are i u and i v . then it is easy to see that the next generation matrix with large domain is the matrix k calculated in section . . since each disease state is a disease state at infection, the next generation matrix is k, the same as the next generation matrix with large domain. as in example , the determinant of k is zero and k has rank . thus the control reproduction number is the trace of k, example . (a multi-strain model of gonorrhea) the following example comes from lima and torres ( ). the system of equations is where s is the susceptible class, i is the class infected with strain , and i is the class of individuals infected with a mutated strain. the "birth" rate of the population is ρ, μ is the natural mortality rate, c is the probability of successful contact, λ i is the of strain i, γ i is the recovery rate of strain i, and p is the proportion of the original infected population that become infected by the mutated strain. the disease-free equilibrium for this model is [s = n, i = , i = ]. next we will reorder our variables: di dt di dt t and note that we need only the infected classes to calculate r . then the new infection terms are cλ s(t)i (t) in the di dt equation and in the di dt equation, pγ i (t) enter the i class, but only after they have been infected with strain . then and v = (μ + γ )i (t) (μ + γ )i (t) − pγ i (t) . since we have only two infected classes, n = , and our jacobian matrices are then we can calculate the inverse of v, to calculate the spectral radius of fv − we find the eigenvalues of the matrix: we often call r = cλ μ+γ the reproductive number for strain and r = cλ μ+γ the reproductive number for strain . then the basic reproductive number is where s i refers to the susceptible class of species i, and i i refers to the infected class of species i for i = m, b, h, mosquitoes, birds, and humans, respectively. then p i is the mosquito biting preference for species i, μ i is the natural mortality rate of species i, b is the number of bites per mosquito per unit time, θ is the human recovery rate, β m is the transmission probability from mosquito to host per bite, β b is the transmission probability from birds to mosquito, and β h is the transmission probability from humans to mosquito. from which we calculate the eigenvalues and determine that the spectral radius is we have described the next generation matrix method for continuous models. there is an analogous theory for discrete systems, described in [allen and van den driessche ( )]. there are some situations in which r < in which it is possible to show that the asymptotic stability of the disease -free equilibrium is global, that is, all solutions approach the disease -free equilibrium, not only those with initial values sufficiently close to this equilibrium. we will say that a vector is nonnegative if each of its components is nonnegative, and that a matrix is nonnegative if each of its entries is non -negative. we rewrite the system ( . ) as ( . ) y j = g j (x, y) , j = ,..., m. if r < , we have shown that the disease -free equilibrium is asymptotically stable, and that −a = −(f −v ) is a non -singular m -matrix. theorem . (castillo-chavez, feng, and huang ( )). if −a is a nonsingular m-matrix andf ≥ , if the assumptions on the model ( . ) made earlier in this section are satisfied, and if r < , then the disease-free equilibrium of ( . ) is globally asymptotically stable. proof. the variation of constants formula for the first equation of ( . ) gives there are examples to show that the disease-free equilibrium may not be globally asymptotically stable if the conditionf ≥ is not satisfied. exercises . use the next generation approach to calculate the basic reproduction number for the model ( . ) but with infectivity in the exposed class having a reduction factor ε. . formulate an seitr model and calculate its reproduction number. . for each of the examples in this section determine whether the disease-free equilibrium is globally asymptotically stable when r < . a fundamental assumption in the model ( . ) is homogeneous mixing, that all individuals are equivalent in contacts. a more realistic approach would include separation of the population into subgroups with differences in behavior. for example, in many childhood diseases the contacts that transmit infection depend on the ages of the individuals, and a model should include a description of the rate of contact between individuals of different ages. other heterogeneities that may be important include activity levels of different groups and spatial distribution of populations. network models may be formulated to include heterogeneity of mixing, or more complicated compartmental models can be developed. an important question that should be kept in mind in the formulation of epidemic models is the extent to which the fundamental properties of the simple model ( . ) carry over to more elaborate models. an epidemic model for a disease in which recovery from infection brings only temporary immunity cannot be described by the models of this chapter because of the flow of new susceptibles into the population. this effectively includes demographics in the model, and such models will be described in the next chapter. many of the important underlying ideas of mathematical epidemiology arose in the study of malaria begun by sir r.a. ross ( ). malaria is one example of a disease with vector transmission, the infection being transmitted back and forth between vectors (mosquitoes) and hosts (humans). other vector diseases include west nile virus and hiv with heterosexual transmission. vector transmitted diseases require models that include both vectors and hosts. an actual epidemic differs considerably from the idealized models ( . ) and ( . ). some notable differences are these: . when it is realized that an epidemic has begun, individuals are likely to modify their behavior by avoiding crowds to reduce their contacts and by being more careful about hygiene to reduce the risk that a contact will produce infection. . if a vaccine is available for the disease that has broken out, public health measures will include vaccination of part of the population. various vaccination strategies are possible, including vaccination of health care workers and other first line responders to the epidemic, vaccination of members of the population who have been in contact with diagnosed infectives, and vaccination of members of the population who live in close proximity to diagnosed infectives. . isolation may be imperfect; in-hospital transmission of infection was a major problem in the sars epidemic. . in the sars epidemic of - , in-hospital transmission of disease from patients to health care workers or visitors because of imperfect isolation accounted for many of the cases. this points to an essential heterogeneity in disease transmission that must be included whenever there is any risk of such transmission. the discrete analogue of the continuous-time epidemic model ( . ) is s j+ = s j g j , where s j and i j denote the numbers of susceptible and infective individuals at time j, respectively, g j is the probability that a susceptible individual at time j will remain susceptible to time j + , and σ = e −α is the probability that an infected individual at time j will remain infected to time j + . assume that the initial conditions are s( ) = s > , i( ) = i > , and s + i = n. exercise . consider the system ( . ). (a) show that the sequence {s j + i j } has a limit s ∞ + i ∞ = lim j→∞ (s j + i j ). with r = β −σ . next, consider the case that there are k infected stages and there is treatment in some stages, with treatment rates that can be different in different stages. assume that selection of members for treatment occurs only at the beginning of a stage. let i (i) j and t (i) j denote the numbers of infected and treated individuals respectively in stage i (i = , ,..., k) at time j. let σ i i denote the probability that an infected individual in the i (i) stage continues on to the next stage, either treated or untreated, and let σ t i denote the probability that an individual in the t (i) stage continues on to the next treated stage. in addition, of the members leaving an infected stage i (i) , a fraction p i enters treatment in t (i+ ) , while the remaining fraction q i continues to i (i+ ) . let m i denote the fraction of infected members who go through the stage i (i) , and n i the fraction of infected members who go through the stage t (i) . then, ..., m k = q q · · ·q k , n = p , n = p + q p , ..., n k = p + q p + . . . + q q · · ·q k− p k . the discrete system with treatment is s j+ = s j g j , [i = ,..., k, j ≥ ], with where ε i is the relative infectivity of untreated individuals at stage i and δ i is the relative infectivity of treated individuals at stage i. consider the initial conditions exercise . consider the system ( . ). show that hint. equation ( . ) can be proved by showing the following equalities first: . project: fitting data for an influenza model consider an sir model ( . ) with basic reproduction number . . ), and ( . ) use the next generation approach to calculate their reproduction numbers describe the qualitative changes in (s, i, r) as a function of time for different values of β and α with β ∈ { discuss the result of part (a) in terms of the basic reproductive number (what is β /γ?). use a specific disease such as influenza to provide simple interpretations for the different time courses of the disease for the different choices of β and γ , , . }, and for each value of r , choose the best pair of values (β , α) that fits the slope before the first peak in the data found in the table for reported h n influenza cases in méxico below reluctant" means the class of mbts that come into the door as new hires without a disposition to learn new stuff from ℜ , discuss what would be the impact of changing parameters q, γ, and δ what are your conclusions from this model? cases assume that n(t) = r(t) + p(t) + m(t) +u(t) + i(t) and that the total number of mbts is constant, that is, n(t) = k μ for all t, where k is a constant. the model iswhere q, β , δ , μ, γ, and α are constants and ≤ q ≤ . . interpret the parameters. key: cord- -sh kehrh authors: jurj, sorin liviu; opritoiu, flavius; vladutiu, mircea title: deep learning-based computer vision application with multiple built-in data science-oriented capabilities date: - - journal: proceedings of the st eann (engineering applications of neural networks) conference doi: . / - - - - _ sha: doc_id: cord_uid: sh kehrh this paper presents a data science-oriented application for image classification tasks that is able to automatically: a) gather images needed for training deep learning (dl) models with a built-in search engine crawler; b) remove duplicate images; c) sort images using built-in pre-trained dl models or user’s own trained dl model; d) apply data augmentation; e) train a dl classification model; f) evaluate the performance of a dl model and system by using an accuracy calculator as well as the accuracy per consumption (apc), accuracy per energy cost (apec), time to closest apc (ttcapc) and time to closest apec (ttcapec) metrics calculators. experimental results show that the proposed computer vision application has several unique features and advantages, proving to be efficient regarding execution time and much easier to use when compared to similar applications. data is at the core of every dl application. because the machine learning lifecycle consists of four stages such as data management, model learning, model verification and model deployment [ ] , in order to collect, analyze, interpret and make use of this data, e.g. training accurate models for real-life scenarios, in recent years, new specializations were introduced in universities around the world such as machine learning and data science, to name only a few. additionally, also new career positions were created recently such as machine learning engineer and data scientist, being some of the top paid positions in the industry [ ] . regarding computer vision applications for image classification tasks, a major bottleneck before training the necessary dl models is considered to be the data collection which consists mainly of data acquisition, data labeling and improvement of the existing data in order to train very accurate dl models [ ] . another bottleneck is that, because the amount of data needed to train a dl model is usually required to be very large in size and because most of this important data is not released to the general public but is instead proprietary, the need of an original dataset for a particular dl project can be very crucial. in general, data can be acquired either by a) buying it from marketplaces or companies such as quandl [ ] and ursa [ ] ; b) searching it for free on platforms like kaggle [ ] ; c) crawling it from internet resources with the help of search engine crawlers [ ] ; d) paying to a  workforce on amazon mechanical turk [ ] like the creators of the imagenet dataset did to have all of their images labeled [ ] ; e) creating it manually for free (e.g. when the user takes all the photos and labels them himself), which can be impossible most of the time because of a low-budget, a low-quality camera or time constraints. the importance of image deduplication can be seen in the fields of computer vision and dl where a high number of duplicates can create biases in the evaluation of a dl model, such as in the case of cifar- and cifar- datasets [ ] . it is recommended that before training a dl classification model, one should always check and make sure that there are no duplicate images found in the dataset. finding duplicate images manually can be very hard for a human user and a time-consuming process, this being the reason why a software solution to execute such a task is crucial. some of the drawbacks of existent solutions are that they usually require the user to buy the image deduplication software or pay monthly for a cloud solution, they are big in size or are hard to install and use. despite all of these options, especially in the case of scraping the images from the internet, once stored they can still be unorganized or of a lower quality than expected, with images needed to be sorted out each in their respective class folder in order for the user (e.g. data scientist) to be able later to analyze and use this data for training a performant dl model. this kind of sorting task can take a tremendous amount of time even for a team, from several days or weeks to even months [ ] . another difficult task is that once the data is cleaned, organized and ready to be trained from scratch or using transfer learning, because of the variety of dl architectures, each with different sizes and training time needed until reaching convergence [ ] , it can be very difficult to know from the beginning which dl architecture fits the best a given dataset and will, at the end of the training, result in a dl model that has high accuracy. because energy consumption in dl started to become a very debated aspect in recent months, especially regarding climate change [ ] [ ] [ ] [ ] [ ] , the necessity of evaluating the performance of dl models also by their energy consumption and cost is very crucial. considering these aspects, our work introduces a dl-based computer vision application that has multiple unique built-in data science-oriented capabilities which give the user the ability to train a dl image classification model without any programming skills. it also automatically searches for images on the internet, sort these images each in their individual class folder and is able to remove duplicate images as well as to apply data augmentation in a very intuitive and user-friendly way. additionally, it gives the user an option to evaluate the performance of a dl model and hardware platform not only by considering its accuracy but also its power consumption and cost by using the environmentally-friendly metrics apc, apec, ttcapc and ttcapec [ ] . the paper is organized as follows. in sect. we present the related work. section describes the proposed dl-based computer vision application. section presents the experimental setup and results. finally, sect. concludes this paper. considering the advancements of dl in recent years, there is a growing interest in computer vision applications in the literature, such as regarding the automatic sorting of images, as shown by the authors in [ ] . the authors propose a solution called imagex for sorting large amounts of unorganized images found in one or multiple folders with the help of a dynamic image graph and which successfully groups together these images based on their visual similarity. they also created many similar applications, e.g. imagesorter [ ] , which besides sorting images based on their color similarity, is also able to search, download and sort images from the internet with a built-in google image search option. a drawback of their applications is that the user is able to only visualize similar images, without also having these images automatically cleaned and sorted in their respective class folder with high accuracy. also, the authors in [ ] created an application called sharkzor that combines user interaction with dl in order to sort large amounts of images that are similar. by comparison, regarding sorting, their solutions only sort images by grouping them based on how similar they are to each other after a human interacted and sorted these images initially, whereas our application sorts them automatically by using in-built pre-trained dl models or gives the user an option to use his own trained dl models. an on-device option that uses dl capabilities and helps users find similar photos (e.g. finding photos that contain certain objects such as flowers, trees, food, to name only a few) is presented also by apple in their newest version of photos app [ ] . regarding the detection of duplicate images, this technique has practical applications in many domains such as social media analysis, web-scale retrieval as well as digital image forensics [ , ] , with several works in the literature applying it for the detection of copyright infringements [ ] and fraud detection [ ] . recently, a python package that makes use of hashing algorithms and convolution neural networks (cnns) that finds exact or near-duplicates in an image collection called image deduplicator (imagededup) was released in [ ] . in our computer vision application, we make use of this package in order to offer a user the option to remove duplicate images from the images dataset (e.g. right before training a dl model). when training dl models from scratch or by using transfer learning, usually frameworks such as tensorflow and pytorch are used [ ] , either locally (e.g. on a personal laptop or desktop pc that contains a powerful gpu) or in cloud services such as cloud automl [ , ] , amazon aws [ ] or microsoft azure [ ] , with the work in [ ] even assessing the feasibility and usefulness of automated dl in medical imaging classification, where physicians with no programming experience can still complete such tasks successfully. the problem when training locally is that the user still has to research on his own which size the images should have for a given dl architecture, which dl architecture to choose for his dataset and if it is necessary to apply fine-tuning and image augmentation. regarding using the cloud services for training a dl model, even though these may solve most of the problems mentioned above, they still have some drawbacks such as that they can be affected by latency, can be difficult to manage (not user-friendly) and most importantly, they can be very expensive when training for several hours (e.g. cloud automl from google costs around $ per hour when used for computer vision tasks [ ] ). similar work to ours is presented by the authors in [ ] where they created the image atm (automated tagging machine) tool that automatizes the pipeline of training an image classification model (preprocessing, training with model tweaking, evaluation, and deployment). regarding preprocessing, the image atm tool just resizes the images to fit the model input shape. for training, it uses transfer learning with pre-trained cnns from keras by firstly training the last dense layer followed by the whole network. for evaluation, it calculates the confusion matrix and other metrics. a few disadvantages of image atm: the tool is aimed at people with programming knowledge (developers) and is focused mainly on the training function. also, in order to use the image atm tool, the user must take the work of preparing the data in a specific folder structure, e.g. the user must create a .yml file with some of the parameters desired, path to images and destination path. the user must also create a . json file containing the classification of each image. some advantages of the image atm are that the tool offers the possibility for cloud training, has access to more models (although all are trained with the same dataset) and that the evaluation errors can be visualized. when compared to image atm, our computer vision application has several advantages such as that it is accessible to more kinds of people and offers more functionalities such as image web scraping and sorting, deduplication, calculators for accuracy as well as for the apc, apec, ttcapc and ttcapec metrics, all in a user-friendly graphical user interface (gui). the proposed dl-based computer vision application is summarized in fig. and is built using the python programming language. it is composed of the most common features needed in the computer vision field and facilitate them in the form of a gui, without requiring the user to have any knowledge about coding or dl in order to be able to fully use it. regarding the system, the compilation dependencies and installation requirements of the proposed application are python , windows (or later version) or linux (ubuntu or later version). regarding the python libraries, we use pyqt for creating the gui, hdf for loading dl model files, tensorflow for training and inference, opencv for image processing, numpy for data processing, shutil for copying images in the system, tqdm for showing the terminal progress bar, imagededup [ ] for deduplication of images, icrawler for crawling the images and fman build system (fbs) for creating installers. there are certain conventions that are common in all the features of the proposed application: . model files: these are .h files that contain the architecture of a keras model and the weights of its parameters. these are used to load (and save) a previously trained model in order to be able to use it. . model class files: these are extensionless files that contain the labels of each of the classes of a dl model. it contains n lines, where n is the number of classes in the model, and the line i contains the label corresponding to the ith element of the output of the dl model. . preprocessing function: in this convention, a preprocessing function is a function that takes as input the path to an image and a shape, loads the image from the input path, converts the image to an array and fits it to the input of the model. images folders structures: we use two different folders structures: unclassified structures and classified structures. the unclassified images folders structure is the simplest one, consisting of just one folder containing images, presumably to be classified or deduplicated. the classified images folders structure consists of a folder which in turn contains subfolders. each subfolder represents a class of images, is named the same as the label for that class, and contains images tagged or classified belonging to that class. following, we will present all the built-in features: automatic web crawler assisted by inference classification, images deduplication, images sorter assisted by inference classification, dl model trainer with data augmentation capabilities, accuracy calculator as well as the apc and apec [ ] calculators. the purpose of this feature is to collect images related to a keyword (representing a class) from the web and by using a classification algorithm, to make sure that the images are indeed belonging to this class. during the inference process needed for cleaning the images, a preprocessing is happening in the background, which, depending on the pretrained or custom dl model that is chosen, will resize the images, making them have the correct input shape (e.g.   for mnist and   for imagenet) for the dl model. a summarized view of the implemented image crawler feature can be seen in fig. and is composed of the following elements: 'model' -a combo box containing all the existent pretrained in-built dl models such as "mnist" or "resnet " as well as the 'custom' option which gives the user the possibility to load his own previously trained dl model; confidence slider ('confidence required') -a slider to select the minimum accuracy value to be used when classifying the images and which ranges from to ; image class selector ('select a class of images') -a combo box containing the labels of all the classes from the pretrained built-in selected dl model (e.g. classes for when the "mnist" model is selected and classes when the "resnet " model is selected). additionally, the box contains an autocomplete search function as well; images amount ('max amount to get') -a slider to select the number of images that should be crawled from the internet and which ranges from to and 'destination folder' -a browser to select the path for the final location of the obtained images. the options under 'custom model configuration' only apply when the dl model selected is "custom" and is not built-in in the proposed computer vision application, e.g. when it was trained by the user itself. these options are: 'model file' -a browser to select the .h file the user wishes to use for inference and model classes -a browser to select the extensionless file containing the name of each output class on which the selected dl model (.h file) was trained on. finally, this feature's gui interface has a button ('add images!') that begins the web crawling process. with the help of this feature, images are automatically crawled by the crawler and downloaded to a temporal folder location. after that, each image is classified with the selected dl model, and if the classification coincides with the selected class and the confidence is higher than the selected threshold, the image is moved to the 'destination folder', where each image will be saved in its own class folder. this feature automatizes the population of image classification datasets by providing a reliable way of confirming that the downloaded images are clean and correctly organized. the purpose of this feature is to remove duplicate images found in a certain folder. for this, we incorporated the imagededup techniques found in [ ] . a summarized view of the implemented images deduplication feature can be seen in fig. and is composed of the following elements: 'images folder' -a browser to select the location of the folder containing the images that need to be analyzed for duplicate images; 'destination folder' -a browser to select the location of the folder where the deduplicated images will be stored; 'duplicates folder' -a browser to select the location of the folder where the found duplicate images will be stored. each duplicate image found will be stored in a subfolder. regarding advanced settings, it is composed of: hashing method selector ('select a hashing method') -a combo box containing hashing methods that can be used for deduplication (perceptual hashing (default), difference hashing, wavelet hashing, and average hashing) as well as a 'max distance threshold' -the maximum distance by which two images will be considered to be the same (default value is ). finally, this interface has a button ('deduplicate!') that begins the deduplication process according to the selected parameters. following, we will shortly describe the types of hashes we are using in the images deduplication feature: a) average hash: the average hash algorithm first converts the input image to grayscale and then scales it down. in our case, as we want to generate a -bit hash, the image is scaled down. next, the average of all gray values of the image is calculated and then the pixels are examined one by one from left to right. if the gray value is larger than the average, a value is added to the hash, otherwise a value; b) difference hash: similar to the average hash algorithm, the difference hash algorithm initially generates a grayscale image from the input image. here, from each row, the pixels are examined serially from left to right and compared to their neighbor to the right, resulting in a hash; c) perceptual hash: after gray scaling, it applies the discrete cosine transform to rows and as well as to columns. next, we calculate the median of the gray values in this image and generate, analogous to the median hash algorithm, a hash value from the image; d) wavelet hash: analogous to the average hash algorithm, the wavelet hash algorithm also generates a gray value image. next, a twodimensional wavelet transform is applied to the image. in our case, we use the default wavelet function called the haar wavelet. next, each pixel is compared to the median and the hash is calculated. regarding this deduplication feature, first, the hasher generates hashes for each of the images found in the images folder. with these hashes, the distances between hashes (images) are then calculated and if they are lower than the maximum distance threshold (e.g. ), then they are considered duplicates. secondly, for each group of duplicates, the first image is selected as "original" and a folder is created in the duplicates folder with the name of the "original" folder. then all duplicates of this image are stored on that folder. this feature successfully integrates the image deduplication technique [ ] and provides a simple and quick way to utilize it. this feature helps a user to sort an unsorted array of images by making use of dl models. a summarized view of the implemented images sorter feature assisted by inference classification can be seen in fig. and is composed of elements similar to the ones presented earlier for the image crawler feature, but in this case with the function of selecting the path to the folders from which and where images should be sorted. in the destination folder, a new folder is created for each possible class, with the name extracted from the extensionless file that contains all the names of the classes, plus a folder named 'undetermined'. then, each image from the 'images folder' is automatically preprocessed, feed as input to the selected dl model and saved in the corresponding class folder. the highest value from the output determines the predicted class of the image: if this value is less than the minimum 'confidence required', value, then the image will be copied and placed in the 'undetermined' folder, otherwise, the image will be copied to the folder corresponding to the class of the highest value from the output. we took the decision of copying the files instead of moving them, for data security and backup reasons. this feature heavily reduces the amount of time required to sort through an unclassified dataset of images by not only doing it automatically but also removing the need to set up coding environments or even write a single line of code. this feature gives the user a simple gui to select different parameters in order to train and save a dl image classifier model. a summarized view of the implemented dl model trainer feature assisted by inference classification can be seen in fig. and is composed of the following elements: 'model'as described earlier for the image crawler feature; 'sorted images folder' -a browser to select the folder that contains the classified folder structure with the images to be trained on; 'number of training batches' -an integer input, to specify the number of batches to train and 'size of batches'an integer input, to specify the number of images per batch. regarding the custom options, they are the same as mentioned earlier regarding the image crawler feature. next, this interface has a button ('train model') that, when clicked on, prompts a new window for the user to be able to visualize in a very user-friendly way all the image transformations that can be applied to the training dataset in a random way during training. more exactly, as can be seen in fig. , the user can input the following parameters for data augmentation: horizontal flip -if checked the augmentation will randomly flip or not images horizontally; vertical flip -if checked the augmentation will randomly flip or not images horizontally; max width shift -slider (%), maximum percentage (value between and ) of the image width that it can be shifted left or , the maximum amount of degrees (value between and ) that an image might be rotated and max shear shift -slider (%), maximum shear value (value between and ) for image shearing. the data augmentation feature allows the user to visualize the maximum possible changes that can be made to an image in real-time, without the need of guessing the right parameters. following, a training generator is defined with the selected parameters; the generator randomly takes images from the folder structure and fills batches of the selected size, for the number of batches that are selected. these batches are yielded as they are being generated. regarding the training, first, the selected dl model is loaded, its output layer is removed, the previous layers are frozen and a new output layer with the size of the number of classes in the folder structure is added. the model is then compiled with the adam optimizer and the categorical cross-entropy as the loss function. finally, the generator is fed to the model to be fitted. once the training is done, the total training time is shown to the user and a model file (.h ) is created on a prompted input location. this feature achieves the possibility of training a custom dl model on custom classes just by separating images in different folders. there is no knowledge needed about dl and this feature can later also be easily used by the image sorting feature described earlier in order to sort future new unsorted images. this section of the application gui gives a user the option to compute the accuracy of a dl model on the given dataset in the classified images folder structure. a summarized view of the implemented accuracy calculator feature can be seen in fig. and is composed of the following elements: 'model' -as described earlier for the image crawler feature; 'test images folder' -a browser to select the folder that contains the classified folder structure to measure the accuracy of a dl classification model; 'size of batches'an integer input, to specify the number of images per batch. the custom options are the same as mentioned earlier regarding the image crawler feature. finally, this interface has a button ('calculate accuracy') that starts the accuracy evaluation process. after loading the dl model and the list of classes, it searches for the classes as subfolders names in the classified images folder structure. then, for each class (or subfolder) it creates batches of the selected batch size, feeds them to the dl model and counts the number of accurate results as well as the number of images. with these results, it calculates the total accuracy of the dl model and shows it to the user directly in the application gui. this feature provides a simple and intuitive gui to measure the accuracy of any dl image classification model. this gui feature makes use of our apc metric [ ] and which is a function that takes into account not only the accuracy of a system (acc) but also the energy consumption of the system (c). the apc metric can be seen in eq. ( ) below: where c stands from energy consumption of the system and it's measured in watt/hour (wh) and acc stands for accuracy; a is the parameter for the wc a function, the default value is . ; b is a parameter (ranges from to infinity) that controls the influence of the consumption in the final result: higher values will lower more heavily the value of the metric regarding the consumption. the default value is . the application gui gives a user the option to define the values for a and b as well as to specify and calculate the accuracy and energy consumption of a dl model using the above apc metric equation. a summarized view of the implemented apc calculator feature can be seen in fig. and is composed of the following elements: 'model test accuracy (%)' -this widget gives a user the option to input the accuracy or use the previously described accuracy calculator feature to measure the accuracy of a dl model and 'energy consumption (wh)' -float input to specify the power consumption of a user's dl model. regarding the advanced options, it has: alpha (a) -float input to specify the desired value of a (default . ) and beta ðb) -float input to specify the desired value of b (default ). for simplicity, a table is shown with the following columns: accuracy, energy consumption, alpha, beta, and apc. whenever a value is changed, the table is automatically updated as well. finally, the application gui has a button ('calculate apc') to begin the calculation of the apc metric. the function itself is a numpy implementation of our previously defined apc metric [ ] seen in eq. ( ) and takes as input parameters the values defined in the application gui. the implemented feature brings this new apc metric to any user by allowing them to easily calculate the accuracy per consumption and know the performance of their dl model with regards to not only the accuracy but also to the impact it has on the environment (higher energy consumption = higher negative impact on nature). however, the drawback of the current version of this apc calculator feature in the proposed application gui is that the user has to measure the energy consumption of the system manually. we plan to implement automatic readings of the power consumption in future updates (e.g. by using the standard performance evaluation corporation (spec) ptdaemon tool [ , ] , which is also planned to be used for power measurements by the mlperf benchmark in their upcoming mid- update). this metric is a function that takes into account not only the accuracy of a system (acc) but also the energy cost of the system (c). the apec metric can be seen in eq. ( ) below: where c stands for the energy cost of the system and it's measured in eur cents per inference and acc stands for accuracy. a is the parameter for the wc a function, the default value is . ; b is a parameter (ranges from to infinity) that controls the influence of the cost in the final result: higher values will lower more heavily the value of the metric regarding the cost. the default value is . the apec feature is presented in fig. and lets a user define the values for a and b, specify or calculate the accuracy of a dl model, specify the energy consumption and the cost of wh of the dl as well as calculate the apec using the formula seen earlier in ( ) . the apec feature of the proposed computer vision application is composed of the following elements: 'model test accuracy (%)'works similar to the apc widget described earlier; 'energy consumption (wh)' -works also similar to the apc widget described earlier and watt-hour cost -float input to specify the cost in eur cents of a wh. regarding the advanced options, we have: alpha (a) -float input to specify the desired value of a(default . ) and beta bfloat input to specify the desired value of b(default ). a similar table like the one for apc calculator is shown also here, with the following columns: accuracy, energy cost, alpha, beta, and apec. whenever a value is changed, the table is automatically updated here as well. finally, the application gui has a button ('calculate apec') to begin the calculation of the apec metric. the function itself is an implementation on numpy of our previously defined apec metric [ ] seen in eq. ( ) and takes as input parameters the values defined in the application gui. the implemented feature brings this new apec metric to any user by allowing them to easily calculate the accuracy per energy cost and evaluate the performance of their dl model with regards to the impact it has on the environment (higher energy consumption = higher cost = negative impact on nature). however, the drawback of the current version of this apec calculator feature is that the user has to measure the energy consumption of the system and calculate its wh cost manually. the objective of the ttapc metric [ ] is to combine training time and the apc inference metric in an intuitive way. the ttcapc feature is presented in fig. and is composed of the following elements: 'model test accuracy (%)' and 'energy consumption (wh)', both working similar to the apec widget described earlier; 'accuracy delta' -float input to specify the granularity of the accuracy axis; 'energy delta'float to specify the granularity of the energy axis. regarding the advanced options, they are the same as the ones presented earlier regarding the apec feature. a similar table like the one for apec calculator is shown also here, with the following columns: accuracy, energy consumption, alpha, beta, accuracy delta, energy delta, rounded accuracy, rounded energy, training time and closest apc. whenever a value is changed, the table is automatically updated here as well. finally, the application gui has a button ('calculate ttcapc') to begin the calculation of the ttcapc metric. the objective of the ttcapec metric [ ] is to combine training time and the apec inference metric. the ttcapec feature is presented in fig. and is composed of the same elements like the ttcapc feature presented earlier and one additional element called 'energy cost (eur cents per wh)' which is similar to the one presented earlier regarding the apec metric calculator and where the user can specify the cost in eur cents of a wh. a similar table like the one for ttcapc calculator is shown also here, with the following columns: accuracy, energy cost, alpha, beta, accuracy delta, energy delta, rounded accuracy, rounded energy, training time and closest apec. finally, the application gui has a button ('calculate ttcapec') to begin the calculation of the ttcapec metric. following, we will show the experimental results regarding all the implemented features in comparison with existing alternatives found in the literature and industry. we run our experiments on a desktop pc with the following configuration: on the hardware side we use an intel(r) core(tm) i - x cpu @ . ghz, core(s), logical processor(s) with gb ram and an nvidia gtx ti as the gpu; on the software side we use microsoft windows pro as the operating system with cuda . , cudnn . . and tensorflow . . using the keras . . framework. as can be seen in table , our proposed image crawler feature outperforms existent solutions and improves upon them. even though the crawling took the same amount of time, this is not the case regarding the cleaning part, where, because this feature is not available in any of the existent solutions, this needed to be done manually and took s for a folder containing images as compared to only s for our proposed solution which executed the task automatically. a comparison between "dirty" images and clean images can be seen in fig. where, for simplicity, we searched for pictures of "cucumber", which is one class from the total of classes found in the imagenet dataset [ ] . it can be easily observed how the existent solutions provide images that don't represent an actual cucumber, but products (e.g. shampoos) that are made of it. after automatically cleaning these images with a confidence rate of % with the proposed feature, only clean images remained in the folder. for the experiments seen in table , we tested the speed time of the proposed built-in image deduplication feature that uses the imagededup python package [ ] . we run these experiments on finding only exact duplicates on the same number of images with a maximum distance threshold of for all four hashing methods. as can be seen, the average speed is about s for finding duplicates in a folder containing . images, with difference hashing being the fastest hashing method from all four. for our experiments regarding the sorting of images with the proposed images sorter feature, we used both the mnist as well as the imagenet pre-trained models with a confidence rate of % and presented the results in table . regarding mnist experiments, we converted the mnist dataset consisting of . images of  pixels to png format by using the script in [ ] and mixed all these images in a folder. after that, we run our image sorter feature on them and succeeded to have only . % of undetermined images, with a total speed time of around min. regarding imagenet, we used the imagenet large scale visual recognition challenge (ilsvrc ) dataset containing . images belonging to classes with a confidence rate of %. here we successfully sorted regarding the custom model, we used one of our previously trained dl models (resnet- ) that can classify animal classes [ ] on a number of . images of  ratio pixels ( images for each of the animal classes) with a confidence rate of %. here we succeeded to have . % undetermined images, with a total speed time of almost min. the percentage of the undetermined images for all cases can be improved by modifying the confidence rate, but it is out of this paper's scope to experiment with different confidence values. the time that a dl prediction task takes depends on a few variables, mainly the processing power of the machine used to run the model, the framework used to call the inference of the model and the model itself. since processing power keeps changing and varies greatly over different machines, and all the frameworks are optimized complexity wise and keep evolving, we find that among these three the most important to measure is, therefore, the model itself used in the prediction. models vary greatly in their architecture, but all dl models can be mostly decomposed as a series of floating points operations (flops). because, generally, more flops equal more processing needed and therefore more time spent in the whole operation, we measured the time complexity of the built-in imagenet and mnist models in flops and presented the results in table . for the experiments regarding the dl model training feature, because we want to evaluate the application on a real-world problem, we will attempt to show that this feature could be very useful for doctors or medical professionals in the aid of detecting diseases from imaging data (e.g. respiratory diseases detection with x-ray images). in order to prove this, we will attempt to automatically sort between the images of sick patients versus healthy patients regarding, firstly, pneumonia [ ], and secondly, covid- [ ] , all within our application and doing it only with the training feature that the application provides. for this, first, in order to classify between x-ray images of patients with pneumonia versus x-ray images of healthy patients, we made use of transfer learning and trained a 'resnet ' architecture for around h without data augmentation on pneumonia [ ] dataset containing . train images by selecting as the value for the number of training batches and as the value for the size of batches (amount of images per batch) and achieved . % train accuracy after epochs. secondly, in order to classify between x-ray images of patients with covid- versus x-ray images of negative patients, we again made use of transfer learning and trained a 'resnet ' for the experiments regarding the accuracy calculator feature, we used the two custom dl models trained earlier to classify x-ray images of patients with pneumonia versus x-ray images of healthy patients and between x-ray images of patients with covid- versus x-ray images of negative patients, with as the size of batches ( images per batch). the evaluation took in both cases around s with a test accuracy of . % regarding the pneumonia model on test images and % regarding the covid- model on test images, proving that the proposed computer vision application can easily be used by any medical personal with very basic computer knowledge in order to train and test a dl classification model for medical work purposes. regarding the experiments with the proposed apc [ ] calculator feature, we presented the simulated results for different model test accuracy (%) and energy consumption (wh) values in table . we run all the experiments with . as the alpha value and with . as the beta value. it is important to mention that our recommendation for a correct comparison between two dl models, is that it is always necessary that they are both tested with the same alpha and beta values. as can be seen in table where we experimented with random energy consumption and test accuracy values, our apc calculator feature is evaluating the performance of a dl model by considering not only the accuracy but also the power consumption. therefore, dl models that consume around wh (e.g. when running inference on a laptop) instead of wh (e.g. when running inference on a low-cost embedded platform such as the nvidia jetson tx ) [ ] , are penalized more severely by the apc metric. regarding the experiments with the proposed apec [ ] calculator feature, we presented the simulated results for different model test accuracy (%) and energy cost in table . we run all the experiments with . as the alpha value and with . as the beta value. for simplicity, regarding electricity costs, we took germany as an example. according to "strom report" (based on eurostat data) [ ] , german retail consumers paid . euro cents for a wh of electricity in . we used this value to calculate the cost of energy by plugging it in the equation presented in ( )", where "c" in this case stands for the energy cost. as can be seen, the apec metric favors lower power consumption and cost, favoring the use of green energy (free and clean energy). regarding the experiments with the proposed ttcapc [ ] calculator feature, we simulated a custom dl model on two platforms and presented the results in table . as can be seen, even though the accuracy and training time is the same for both platforms, the ttcapc feature favors the platform which has less power consumption. regarding the experiments with the proposed ttcapec [ ] calculator feature, we simulated with the same dl model values used also in the experiments regarding the ttcapc calculator earlier and presented the results in table . as can be also seen in this case, the ttcapec feature favors the lower power consumption of a system because it results in a lower cost. additionally and more importantly, it favors dl-based systems that are powered by green energy, because they have electricity costs and no negative impact on our environment. in this paper, we present a computer vision application that succeeds in bringing common dl features needed by a user (e.g. data scientist) when performing image classification related tasks into one easy to use and user-friendly gui. from automatically gathering images and classifying them each in their respective class folder in a matter of minutes, to removing duplicates, sorting images, training and evaluating a dl model in a matter of minutes, all these features are integrated in a sensible and intuitive manner that requires no knowledge of programming and dl. experimental results show that the proposed application has many unique advantages and also outperforms similar existent solutions. additionally, this is the first computer vision application that incorporates the apc, apec, ttcapc and ttcapec metrics [ ] , which can be easily used to calculate and evaluate the performance of dl models and systems based not only on their accuracy but also on their energy consumption and cost, encouraging new generations of researchers to make use only of green energy when powering their dl-based systems [ ] . assuring the machine learning lifecycle: desiderata, methods, and challenges impact of artificial intelligence on businesses: from research, innovation, market deployment to future shifts in business models a survey on data collection for machine learning: a big data -ai integration perspective imagenet: a large-scale hierarchical image database do we train on test data? purging cifar of near-duplicates snapshot serengeti, high-frequency annotated camera trap images of mammalian species in an african savanna deep double descent: where bigger models and more data hurt energy and policy considerations for deep learning in nlp efficient implementation of a self-sufficient solar-powered real-time deep learning-based system environmentally-friendly metrics for evaluating the performance of deep learning models and systems tackling climate change with machine learning dynamic construction and manipulation of hierarchical quartic image graphs sharkzor: interactive deep learning for image triage, sort, and summary apple photos benchmarking unsupervised near-duplicate image detection recent advance in content-based image retrieval: a literature survey effective and efficient global context verification for image copy detection image forensics: detecting duplication of scientific images with manipulation-invariant image similarity performance analysis of deep learning libraries: tensorflow and pytorch automl: a survey of the state-of-the-art automated deep learning design for medical image classification by healthcare professionals with no coding experience: a feasibility study measuring and benchmarking power consumption and energy efficiency standard performance evaluation corporation (spec) power mnist converted to png format real-time identification of animals found in domestic areas of europe covid- image data collection key: cord- -mckqp v authors: ksieniewicz, paweł; goścień, róża; klinkowski, mirosław; walkowiak, krzysztof title: pattern recognition model to aid the optimization of dynamic spectrally-spatially flexible optical networks date: - - journal: computational science - iccs doi: . / - - - - _ sha: doc_id: cord_uid: mckqp v the following paper considers pattern recognition-aided optimization of complex and relevant problem related to optical networks. for that problem, we propose a four-step dedicated optimization approach that makes use, among others, of a regression method. the main focus of that study is put on the construction of efficient regression model and its application for the initial optimization problem. we therefore perform extensive experiments using realistic network assumptions and then draw conclusions regarding efficient approach configuration. according to the results, the approach performs best using multi-layer perceptron regressor, whose prediction ability was the highest among all tested methods. according to cisco forecasts, the global consumer traffic in the internet will grow on average with annual compound growth rate (cagr) of % in years - [ ] . the increase in the network traffic is a result of two main trends. firstly, the number of devices connected to the internet is growing due to the increasing popularity of new services including internet of things (iot ). the second important trend influencing the traffic in the internet is popularity of bandwidth demanding services such as video streaming (e.g., netflix ) and cloud computing. the internet consists of many single networks connected together, however, the backbone connecting these various networks are optical networks based on fiber connections. currently, the most popular technology in optical networks is wdm (wavelength division multiplexing), which is expected to be not efficient enough to support increasing traffic in the nearest future. in last few years, a new concept for optical networks has been deployed, i.e., architecture of elastic optical networks (eons). however, in the perspective on the next decade some new approaches must be developed to overcome the predicted "capacity crunch" of the internet. one of the most promising proposals is spectrally-spatially flexible optical network (ss-fon) that combines space division multiplexing (sdm) technology [ ] , enabling parallel transmission of co-propagating spatial modes in suitably designed optical fibers such as multi-core fibers (mcfs) [ ] , with flexible-grid eons [ ] that enable better utilization of the optical spectrum and distanceadaptive transmissions [ ] . in mcf-based ss-fons, a challenging issue is the inter-core crosstalk (xt) effect that impairs the quality of transmission (qot ) of optical signals and has a negative impact on overall network performance. in more detail, mcfs are susceptible to signal degradation as a result of the xt that happens between adjacent cores whenever optical signals are transmitted in an overlapping spectrum segment. addressing the xt constraints significantly complicates the optimization of ss-fons [ ] . besides numerous advantages, new network technologies bring also challenging optimization problems, which require efficient solution methods. since the technologies and related problems are new, there are no benchmark solution methods to be directly applied and hence many studies propose some dedicated optimization approaches. however, due to the problems high complexity, their performance still needs a lot of effort to be put [ , ] . we therefore observe a trend to use artificial intelligence techniques (with the high emphasis on pattern recognition tools) in the field of optimization of communication networks. according to the literature surveys in this field [ , , , ] , the researchers mostly focus on discrete labelled supervised and unsupervised learning problems, such as traffic classification. regression methods, which are in the scope of that paper, are mostly applied for traffic prediction and estimation of quality of transmission (qot ) parameters such as delay or bit error rate. this paper extends our study initiated in [ ] . we make use of pattern recognition models to aid optimization of dynamic mcf-based ss-fons in order to improve performance of the network in terms of minimizing bandwidth blocking probability (bbp), or in other words to maximize the amount of traffic that can be allocated in the network. in particular, an important topic in the considered optimization problem is selection of a modulation format (mf) for a particular demand, due to the fact that each mf provides a different tradeoff between required spectrum width and transmission distance. to solve that problem, we define applicable distances for each mf (i.e., minimum and maximum length of a routing path that is supported by each mf). to find values of these distances, which provide best allocation results, we construct a regression model and then combine it with monte carlo search. it is worth noting that this work does not address dynamic problems in the context of changing the concept over time, as is often the case with processing large sets, and assumes static distribution of the concept [ ] . the main novelty and contribution of the following work is an in-depth analysis of the basic regression methods stabilized by the structure of the estimator ensemble [ ] and assessment of their usefulness in the task of predicting the objective function for optimization purposes. in one of the previous works [ ] , we confirmed the effectiveness of this type of solution using a regression algorithm of the nearest weighted neighbors, focusing, however, much more on the network aspect of the problem being analyzed. in the present work, the main emphasis is on the construction of the prediction model. its main purpose is: -a proposal to interpret the optimization problem in the context of pattern recognition tasks. the rest of the paper is organized as follows. in sect. , we introduce studied network optimization problem. in sect. , we discuss out optimization approach for that problem. next, in sect. we evaluate efficiency of the proposed approach. eventually, sect. concludes the work. the optimization problem is known in the literature as dynamic routing, space and spectrum allocation (rssa) in ss-fons [ ] . we are given with an ss-fon topology realized using mcfs. the topology consists of nodes and physical link. each physical link comprises of a number of spatial cores. the spectrum width available on each core is divided into arrow and same-sized segments called slices. the network is in its operational state -we observe it in a particular time perspective given by a number of iterations. in each iteration (i.e., a time point), a set of demands arrives. each demand is given by a source node, destination node, duration (measured in the number of iterations) and bitrate (in gbps). to realize a demand, it is required to assign it with a light-path and reserve its resources for the time of the demand duration. when a demand expires, its resources are released. a light-path consists of a routing path (a set of links connecting demand source and destination nodes) and a channel (a set of adjacent slices selected on one core) allocated on the path links. the channel width (number of slices) required for a particular demand on a particular routing path depends on the demand bitrate, path length (in kilometres) and selected modulation format. each incoming demand has to be realized unless there is not enough free resources when it arrives. in such a case, a demand is rejected. please note that the selected light-paths in i -th iteration affect network state and allocation possibilities in the next iterations. the objective function is defined here as bandwidth blocking probability (bbp) calculated as a summed bitrate of all rejected demands divided by the summed bitrate of all offered demands. since we aim to support as much traffic as it is possible, the objective criterion should be minimized [ , ] . the light-paths' allocation process has to satisfy three basic rssa constraints. first, each channel has to consists of adjacent slices. second, the same channel (i.e., the same slices and the same core) has to be allocated on each link included in a light-path. third, in each time point each slice on a particular physical link and a particular core can be used by at most one demand [ ] . there are four modulation formats available for transmissions- -qam, -qam, qpsk and bpsk. each format is described by its spectral efficiency, which determines number of slices required to realize a particular bitrate using that modulation. however, each modulation format is also characterized by the maximum transmission distance (mtd) which provides acceptable value of optical signal to noise ratio (osnr) at the receiver side. more spectrally-efficient formats consume less spectrum, however, at the cost of shorter mtds. moreover, more spectrally-efficient formats are also vulnerable to xt effects which can additionally degrade qot and lead to demands' rejection [ , ] . therefore, the selection of the modulation format for each demand is a compromise between spectrum efficiency and qot. to answer that problem, we use the procedure introduced in [ ] to select a modulation format for a particular demand and routing path [ ] . let m = , , , denote modulation formats ordered in increasing mtds (and in decreasing spectral efficiency at the same time). it means that m = denotes -qam and m = denotes bpsk. let mt d = [mtd , mtd , mtd , mtd ] be a vector of mtds for modulations -qam, -qam, qpsk, bpsk respectively. moreover, let at d = [atd , atd , atd , atd ] (where atd i <= mtd i , i = , , , ) be the vector of applicable transmission distances. for a particular demand and a routing path we select most spectrally-efficient modulation format i for which atd i is grater of equal to the selected path length and the xt effect is on an acceptable level. for each candidate modulation format, we asses the xt level based on the adjacent resources' (i.e., slices and cores) availability using procedure proposed in [ ] . it is important to note that we do not indicate atd (for bpsk) since we assume that this modulation is able to support transmission on all candidate routing paths regardless of their length. please also note that when xt level is too high for all modulation formats, the demand is rejected regardless of the light-paths' availability. in sect. we have studied rssa problem and emphasised the importance of efficient modulation selection task. for that task we have proposed solution method whose efficiency strongly depends on the applied atd vector. therefore, we aim to find atd * vector that provides best results. the vector elements have to be positive and have upper bounds given by vector mtd. moreover, the following condition have to be satisfied: atd i < atd i+ , i = , . since solving rssa instances is a time consuming process, it is impossible to evaluate all possible atd vectors in a reasonable time. we therefore make use of regression methods and propose a scheme to find atd * depicted in fig. . a representative set of different atd vectors is generated. then, for each of them we simulate allocation of demands in ss-fon (i.e., we solve dynamic rssa). for the purpose of demands allocation (i.e., selection of light-paths), we use a dedicated algorithm proposed in [ ] . for each considered atd vector we save obtained bbp. based on that data, we construct a regression model, which predicts bbp based on an atd vector. having that model, we use monte carlo method to find atd * vector, which is recommended for further experiments. to solve an rssa instance for a particular atd vector, we use heuristic algorithm proposed in [ ] . we work under the assumption that there are candidate routing paths for each traffic demand (generated using dijkstra algorithm). since the paths are generated in advance and their lengths are known, we can use an atd vector and preselect for these paths modulation formats based on the procedure discussed in sect. . therefore, rssa is reduced to the selection of one of the candidate routing paths and a communication channel with respect to the resource availability and assessed xt levels. from the perspective of pattern recognition methods, the abstraction of the problem is not the key element of processing. the main focus here is the representation available to construct a proper decision model. for the purposes of considerations, we assume that both input parameters and the objective function take only quantitative and not qualitative values, so we may use probabilistic pattern recognition models to process them. if we interpret the optimization task as searching for the extreme function of many input parameters, each simulation performed for their combination may also be described as a label for the training set of supervised learning model. in this case, the set of parameters considered in a single simulation becomes a vector of object features (x n ), and the value of the objective function acquired around it may be interpreted as a continuous object label (y n ). repeated simulation for randomly generated parameters allows to generate a data set (x) supplemented with a label vector (y). a supervised machine learning algorithm can therefore gain, based on such a set, a generalization abilities that allows for precise estimation of the simulation result based on its earlier runs on the random input values. a typical pattern recognition experiment is based on the appropriate division of the dataset into training and testing sets, in a way that guarantees their separability (most often using cross-validation), avoiding the problem of data peeking and a sufficient number of repetitions of the validation process to allow proper statistical testing of mutual model dependencies hypotheses. for the needs of the proposal contained in this paper, the usual -fold cross validation was adopted, which calculates the value of the r metric for each loop of the experiment. having constructed regression model, we are able to predict bbp value for a sample atd vector. please note that the time required for a single prediction is significantly shorter that the time required to simulate a dynamic rssa. the last step of our optimization procedure is to find atd * -vector providing lowest estimated bbp values. to this end, we use monte carlo method with a number of guesses provided by the user. the rssa problem was solved for two network topologies-dt ( nodes, links) and euro ( nodes, links). they model deutsche telecom (german national network) and european network, respectively. each network physical link comprised of cores wherein each of the cores offers frequency slices of . ghz width. we use the same network physical assumptions and xt levels and assessments as in [ ] . traffic demands have randomly generated end nodes and birates uniformly distributed between gbps and tbps, with granularity of gbps. their arrival follow poisson process with an average arrival rate λ demands per time unit. the demand duration is generated according to a negative exponential distribution with an average of /μ. the traffic load offered is λ/μ normalized traffic units (ntus). for each testing scenario, we simulate arrival of demands. four modulations are available ( -qam, -qam, qpsk, bpsk) wherein we use the same modulation parameters as in [ ] . for each topology we have generated different datasets, each consists of samples of atd vector and corresponding bbp. the datasets differ with the xt coefficient (μ = · − indicated as "xt ", μ = · − indicated as "xt ", for more details we refer to [ ] ) and network links scaling factor (the multiplier used to scale lengths of links in order to evaluate if different lengths of routing paths influence performance of the proposed approach). for dt we use following scaling factors: . , . , . , . . . , . . for euro the values are as follows: . , . , . , . , . , . , . , . , . . we indicate them as "sx.xxx " where x.xxx refers to the scaling factor value. using these datasets we can evaluate whether xt coefficient (i.e., level of the vulnerability to xt effects) and/or average link length influence optimization approach performance. the experimental environment for the construction of predictive models, including the implementation of the proposed processing method, was implemented in python, following the guidelines of the state-of-art programming interface of the scikit-learn library [ ] . statistical dependency assessment metrics for paired tests were calculated according to the wilcoxon test, according to the implementation contained in scipy module. each of the individual experiments was evaluated by r score -a typical quality assessment metric for regression problems. the full source code, supplemented with employed datasets is publicly available in a git repository . five simple recognition models were selected as the base experimental estimators: knr-k-nearest neighbors regressor with five neighbors, leaf size of and euclidean metric approximated by minkowski distance, -dknr-knr regressor weighted by distance from closest patterns, mlp-a multilayer perceptron with one hidden layer of one hundred neurons, with the relu activation function and adam optimizer, dtr-cart tree with mse split criterion, lin-linear regression algorithm. in this section we evaluate performance of the proposed optimization approach. to this end, we conduct three experiments. experiment focuses on the number of patterns required to construct a reliable prediction model. experiment assesses the statistical dependence of built models. eventually, experiment verifies efficiency of the proposed approach as a function of number of guesses in the monte carlo search. the first experiment carried out as part of the approach evaluation is designed to verify how many patterns -and thus how many repetitions of simulations -must be passed to individual regression algorithms to allow the construction of a reliable prediction model. the tests were carried out on all five considered regressors in two stages. first, the range from to patterns was analyzed, and in the second, from to patterns per processing. it is important to note that due to the chosen approach to cross-validation, in each case the model is built on % of available objects. the analysis was carried out independently on all available data sets, and due to the non-deterministic nature of sampling of available patterns, its results were additionally stabilized by repeating a choice of the objects subset five times. in order to allow proper observations, the results were averaged for both topologies. plots for the range from to patterns were additionally supplemented by marking ranges of standard deviation of r metric acquired within the topology and presented in the range from the . value. the results achieved for averaging individual topologies are presented in figs. and . for dt topology, mlp and dtr algorithms are competitively the best models, both in terms of the dynamics of the relationship between the number of patterns and the overall regression quality. the linear regression clearly stands out from the rate. a clear observation is also the saturation of the models, understood by approaching the maximum predictive ability, as soon as around patterns in the data set. the best algorithms already achieve quality within . , and with patterns they stabilize around . . the relationship between each of the recognition algorithms and the number of patterns takes the form of a logarithmic curve in which, after fast initial growth, each subsequent object gives less and less potential for improving the quality of prediction. this suggests that it is not necessary to carry out further simulations to extend the training set, because it will not significantly affect the predictive quality of the developed model. very similar observations may be made for euro topology, however, noting that it seems to be a simpler problem, allowing faster achievement of the maximum model predictive capacity. it is also worth noting here the fact that the standard deviation of results obtained by mlp is smaller, which may be equated with the potentially greater stability of the model achieved by such a solution. the second experiment extends the research contained in experiment by assessing the statistical dependence of models built on a full datasets consisting of a thousand samples for each case. the results achieved are summarized in tables a and b. as may be seen, for the dt topology, the lin algorithm clearly deviates negatively from the other methods, in absolutely every case being a worse solution than any of the others, which leads to the conclusion that we should completely reject it from considering as a base for a stable recognition model. algorithms based on neighborhood (knr and dknr) are in the middle of the rate, in most cases statistically giving way to mlp and dtr, which would also suggest departing from them in the construction of the final model. the statistically best solutions, almost equally, in this case are mlp and dtr. for euro topology, the results are similar when it comes to lin, knr and dknr approaches. a significant difference, however, may be seen for the achievements of dtr, which in one case turns out to be the worst in the rate, and in many is significantly worse than mlp. these observations suggest that in the final model for the purposes of optimization lean towards the application of neural networks. what is important, the highest quality prediction does not exactly mean the best optimization. it is one of the very important factors, but not the only one. it is also necessary to be aware of the shape of the decision function. for this purpose, the research was supplemented with visualizations contained in fig. . algorithms based on neighborhood (knn, dknn) and decision trees (dtr) are characterized by a discrete decision boundary, which in the case of visualization resembles a picture with a low level of quantization. in the case of an ensemble model, stabilized by cross-validation, actions are taken to reduce this property in order to develop as continuous a border as possible. as may be seen in the illustrations, compensation occurs, although in the case of knn and dknn leads to some disturbances in the decision boundary (interpreted as thresholding the predicted label value), and for the dtr case, despite the general correctness of the performed decisions, it generates image artifacts. such a model may still retain high predictive ability, but it has too much tendency to overfit and leads to insufficient continuity of the optimized function to perform effective optimization. clear decision boundaries are implemented by both the lin and mlp approaches. however, it is necessary to reject lin from processing due to the linear nature of the prediction, which (i ) in each optimization will lead to the selection of the extreme value of the analyzed range and (ii ) is not compatible with the distribution of the explained variable and must have the largest error in each of the optimas. summing up the observations of experiments and , the mlp algorithm was chosen as the base model for the optimization task. it is characterized by (i ) statistically best predictive ability among the methods analyzed and (ii ) the clearest decision function from the perspective of the optimization task. the last experiment focuses on the finding of best atd vector based on the constructed regression model. to this end, we use monte carlo method with different number of guesses. tables and present the obtained results as a function of number of guesses, which changes from up to . the results quality increases with the number of guesses up to some threshold value. then, the results do not change at all or change only a little bit. according to the presented values, monte carlo method applied with guesses provides satisfactory results. we therefore recommend that value for further experiments. the following work has considered the topic of employing pattern recognition methods to support ss-fon optimization process. for a wide pool of generated cases, analyzing two real network topologies, the effectiveness of solutions implemented by five different, typical regression methods was analyzed, starting from logistic regression and ending with neural networks. conducted experimental analysis shows, with high probability obtained by conducting proper statistical validation, that mlp is characterized by the greatest potential in this type of solutions. even with a relatively small pool of input simulations, constructing a data set for learning purpouses, interpretable in both the space of optimization and machine learning problems, simple networks of this type achieve both high quality prediction measured by the r metric, and continuous decision space creating the potential for conducting optimization. basing the model on the stabilization realized by using ensemble of estimators additionally allows to reduce the influence of noise on optimization, whichin a state-of-art optimization methods -could show a tendency to select invalid optimas, burdened by the nondeterministic character of the simulator. further research, developing ideas presented in this article, will focus on the generalization of the presented model for a wider pool of network optimization problems. high-capacity transmission over multi-core fibers a comprehensive survey on machine learning for networking: evolution, applications and research opportunities visual networking index: forecast and trends elastic optical networking: a new dawn for the optical layer on the efficient dynamic routing in spectrally-spatially flexible optical networks on the complexity of rssa of any cast demands in spectrally-spatially flexible optical networks machine learning assisted optimization of dynamic crosstalk-aware spectrallyspatially flexible optical networks survey of resource allocation schemes and algorithms in spectrally-spatially flexible optical networking data stream classification using active learned neural networks artificial intelligence (ai) methods in optical networks: a comprehensive survey an overview on application of machine learning techniques in optical networks scikit-learn: machine learning in python machine learning for network automation: overview, architecture, and applications survey and evaluation of space division multiplexing: from technologies to optical networks modeling and optimization of cloud-ready and content-oriented networks. ssdc classifier selection for highly imbalanced data streams with minority driven ensemble key: cord- -ydnxotsq authors: chen, jiarui; cheong, hong-hin; siu, shirley weng in title: bestox: a convolutional neural network regression model based on binary-encoded smiles for acute oral toxicity prediction of chemical compounds date: - - journal: algorithms for computational biology doi: . / - - - - _ sha: doc_id: cord_uid: ydnxotsq compound toxicity prediction is a very challenging and critical task in the drug discovery and design field. traditionally, cell or animal-based experiments are required to confirm the acute oral toxicity of chemical compounds. however, these methods are often restricted by availability of experimental facilities, long experimentation time, and high cost. in this paper, we propose a novel convolutional neural network regression model, named bestox, to predict the acute oral toxicity ([formula: see text]) of chemical compounds. this model learns the compositional and chemical properties of compounds from their two-dimensional binary matrices. each matrix encodes the occurrences of certain atom types, number of bonded hydrogens, atom charge, valence, ring, degree, aromaticity, chirality, and hybridization along the smiles string of a given compound. in a benchmark experiment using a dataset of observations (train/test / ), bestox achieved a squared correlation coefficient ([formula: see text]) of . , root-mean-squared error (rmse) of . , and mean absolute error (mae) of . . despite of the use of a shallow model architecture and simple molecular descriptors, our method performs comparably against two recently published models. measuring the chemical and physiological properties of chemical compounds are fundamental tasks in biomedical research and drug discovery [ ] . the basic idea of modern drug design is to search chemical compounds with desired affinity, potency, and efficacy against the biological target that is relevant to the disease of interest. however, not only that there are tens of thousands known chemical compounds existed in nature, but many more artificial chemical compounds are being produced each year [ ] . thus, the modern drug discovery pipeline is focused on narrowing down the scope of the chemical space where good drug candidates are [ , ] . potential lead compounds will be subjected to further experimental validation on their pharmacodynamics and pharmacokinetic (pd/pk) properties [ , ] ; the latter includes absorption, distribution, metabolism, excretion, and toxicity (adme/t) measurements. traditionally, chemists and biologists conduct cell-based or animal-based experiments to measure the pd/pk properties of these compounds and their actual biological effects in vivo. however, these experiments are not only high cost in terms of both time and money, the experiments that involve animal testings are increasingly subjected to concerns from ethical perspectives [ ] . among all measured properties, toxicity of a compound is the most important one which must be confirmed before approval of the compound for medication purposes [ ] . there are different ways to classify the toxicity of a compound. for example, based on systemic toxic effects, the common toxicity types include acute toxicity, sub-chronic toxicity, chronic toxicity, carcinogenicity developmental toxicity and genetic toxicity [ ] . on the other hand, based on the toxicity effects area, toxicity can also be classified as hepatotoxicity, ototoxicity, ocular toxicity, etc. [ ] . therefore, there is a great demand for accurate, low-cost and time-saving toxicity prediction methods for different toxicity categories. toxicity of a chemical compound is associated with its chemical structure [ ] . a good example is the chiral compounds. this kind of compounds and their isomers have highly similar structures but only slight differences in molecular geometry. their differences cause them to possess different biological properties. for example, the drug dopa is a compound for treating the parkinson disease. the d-isomer form of this compound has severe toxicity whereas the l-isomer form does not [ ] . therefore, only its levorotatory form can be used for medical treatments. this property-structure relationship is often described as quantitative structure-activity relationship (qsar) and have been widely used in the prediction of different properties of compounds [ , ] . based on the same idea, toxicities of a compound, being one of the most concerned properties, can be predicted via computational means as a way to select more promising candidates before undertaking further biological experiments. the simplified molecular input line entry system, also called smiles [ , ] , is a linear representation of a chemical compound. it is a short ascii string describing the composition, connectivity, and charges of atoms in a compound. an example is shown in fig. . the compound is called morphine; it is originated from the opiate family and is found to exist naturally in many plants and animals. morphine has been widely used as a medication to relief acute and chronic pain of patients. nowadays, compounds are usually converted into their smiles strings for the purpose of easy storage into databases or for other computational processing such as machine learning. common molecular toolkits such as rdkit [ ] and openbabel [ ] can convert a smiles string to its d and d structures, and vice versa. in recent years, machine learning has become the mainstream technique in natural language processing (nlp). among all machine learning applications for nlp, text classification is the most widely studied. based on the input text sentences, a machine learning-based nlp model analyzes the organization of words and the types of words in order to categorize the given text. two pioneering nlp methods are textcnn [ ] and convnets [ ] . the former method introduced a pretrained embedding layer to encode words of input sentences into fixed-size feature vectors with padding. then, feature vectors of all words were combined to form a sentence matrix that was fed into a standard convolutional neural network (cnn) model. this work was considered a breakthrough at that time and accumulated over citations since (as per google scholar). another spotlight paper in nlp for text classification is convnets [ ] . instead of analyzing words in a sentence, this model exploited simple one-hot encoding method at the character level for unique characters in sentence analysis. the success of these methods in nlp shed lights to other applications that have only texts as raw data. compound toxicity prediction can be considered as a classification problem too. recently, hirohara et al. [ ] proposed a new cnn model for toxicity classification based on character-level encoding. in this work, each smiles character is encoded into a -dimensional feature vector. the cnn model based on this simple encoding method achieved an area-under-curve (auc) value of . for classification of endpoints using the tox dataset [ ] . the best auc score in tox challenge is . which is achieved by deeptox [ ] . despite of its higher accuracy, the deeptox model is extremely complex. it requires heavy feature engineering from a large pool of static and dynamic features derived from the compounds or indirectly via external tools. the classification model is ensemble-based combining deep neural network (dnn) with multiple layers of hidden nodes ranging from to nodes. the train dataset for this highly complex model was comprised of over , observations and superior predictive performance was demonstrated. besides classification, toxicity prediction can be seen as a regression problem when the compound toxicity level is of concern. like other qsar problems, toxicity regression is a highly challenging task due to limited data availability and noisiness of the data. with limited data, the use of simpler model architecture is preferred to avoid the model being badly overfitted. in this work, we have focused on the regression of acute oral toxicity of chemical compounds. two recent works [ , ] were found to solve this problem where the maximally achievable r is only . [ ] . in this study, we developed a regression model for acute oral toxicity prediction. the prediction task is to estimate the median lethal dose, ld , of the compound; this is the dose required to kill half the members of the tested population. a small ld value indicates high toxicity level whereas a large ld value indicates low toxicity level of the compound. based on the ld value, compounds can be categorized into four levels as defined by the united states environmental protection agency (epa) (see table ). category iii slightly toxic and slightly irritating < ld ≤ category iv practically non-toxic and not an irritant < ld the rat acute oral toxicity dataset used in this study was kindly provided by the author of toptox [ ] . this dataset was also used in the recent study of computational toxicity prediction by karim et al. [ ] . for ld prediction task, the dataset contains samples; out of which samples are for training and samples are for testing. the original train/test split was deliberately made to maintain similar distribution of the train and test datasets to facilitate learning and model validation. it is noteworthy that as the actual ld values were in a wide range (train set: . mg/kg to . mg/kg, test set: . mg/kg to . mg/kg), the ld values were first transformed to mol/kg format, and then scaled logarithmically to −log (ld ). finally, the processed experimental values range from . to . in the train set and . to . in the test set. as a smiles string is not an understandable input format for general machine learning methods, it needs to be converted or encoded into a series of numerical values. ideally, these values should capture the characteristics of the compound and correlates to the interested observables. the most popular way to encode a smile is to use molecular fingerprints such as molecular access system (maccs) and extended connectivity fingerprint (ecfp). however, fingerprint algorithms generate high dimensional and sparse matrices which make learning difficult. here, in order to solve the regression task for oral toxicity prediction. inspired by the work of hirohara et al. [ ] , we proposed the modified binary encoding method for smiles, named bes for short. in bes, each character is encoded by a binary vector of bits. among them bits are for encoding the smiles alphabets and symbols by the one-hot encoding approach; bits are for encoding various atomic properties including number of bonded hydrogens, formal charge, valence, ring atom, degree, aromaticity, chirality, and hybridization. the feature types and corresponding size of the feature is listed in table . as the maximum length of smiles strings in our dataset is , the size of the feature matrix for one smiles string was defined to be × . for a smiles string that is shorter than in length, zero padding was applied. figure illustrates how bes works. our prediction model is a conventional cnn model with convolutional layers to extract features, pooling layers to reduce dimensionality of the feature matrix and to prevent overfitting, and a multi-layer neural network to correlate features to ld values. to decide the model architecture and to tune hyperparameters of the model, a grid search method was employed. table shows the hyperparameters and their ranges of values within which the model was optimized. in each grid search process, the model training was run for epochs and the mean-squared error (mse) loss of the model in -fold cross validation was used as a criteria for model selection. the optimal parameters are also presented in table . the final production model was trained using the optimal parameters and the entire train dataset. the maximum training epoch was ; early stop method was used to prevent the problem of overfitting. the architecture of our optimized cnn model is presented in fig. . the model contains two convolutional layers (conv) with and filters respectively. after each convolutional layer is an average pooling layer and a batch normalization layer (bn). then, a max pooling layer is used before the learned features fed into the fully connected layers (fc). four fcs containing , , , and hidden nodes were found to be the optimal combination for toxicity prediction and the relu function is used to generate the prediction output. all implementations were done using python . . with the following libraries: anaconda . . , rdkit v . . . , pytorch . . and cuda . . we used gettotalnumhs, getformalcharge, getchiraltag, gettotaldegree, isinring, getisaromatic, gettotalvalence and gethybridization functions from rdkit to calculate atom properties. our model was trained and tested in a workstation equipped with two nvidia tesla p gpus. training of the final production model was performed using the optimal parameters obtained from the result of our extensive grid search. figure shows the evolution of mse over the number of training cycles. the training stopped at the -th epoch with mse of . . table shows the performances of our model table . in the train and test sets. the training performance is excellent which gives r of . as all the data was used to construct the model. for the test set, the model predicts with r of . , rmse of . , and mae of . . figure shows the scatterplot of bestox prediction on the test data. we can see that prediction is better for compounds with lower toxicity (lower −log (ld )) and worse for those with higher toxicity. this may be due to fewer data available in the train set for higher toxicity compounds. thus, we also tested our model on samples with target values less than . in the test set ( samples out of total samples, the sample coverage is more than %). in this case, the performance of our model is improved: rmse is decreased from . to . and mae is reduced from . to . . table . performance comparison of our model to two existing acute oral toxicity prediction methods: toptox [ ] and dt+snn [ ] . performance data of these methods were obtained from the original literature. table presents the comparative performance of bestox to two existing acute oral toxicity prediction models, the st-dnn model from toptox and the dt+snn model from karim et al. [ ] . results show that our model is slightly better than st-dnn with respect to r and mae. the best performed model is dt+snn which has a correlation of . ; but rmse and mae were not provided in the original study. the closeness of the performance metrics of bestox to two existing models suggest that our model performs on par with them. nevertheless, it should be mentioned that while our model has employed simple features and relatively simple model architecture, st-dnn and dt+snn relied on highly engineered input features and complex ensemble-based model architectures. for st-dnn [ ] , they combined element specific topological descriptors (estd) and auxiliary descriptors as candidates to generate the feature vectors for prediction (our model uses only features). in addition, their model included ensemble of two different types of classifiers, namely, deep neural network (dnn) and gradient boosted decision tree (gbdt). combining predictions from several classifiers is an easy way to improve prediction accuracy, however, the complexity introduced into the model makes the already "black box model" more difficult to understand. for the recent dt+snn model [ ] , they used decision trees (dt) to select different descriptors generated from the padel tools [ ] . although their shallow neural network (snn) architecture required short model training time, more time was spent on feature generation and selection. different combination of features were used depending on the tasks to be predicted, which had high computational cost. here, bestox has achieved results comparable to these more complex models with simple binary features and model architecture, showing the power of our method. in this paper, we present our new method bestox for acute oral toxicity prediction. inspired by nlp techniques for text classification, we have designed a simple character-level encoding method for smiles called the binary-encoded smiles (bes). we have developed a shallow cnn to learn the bes matrices to predict the ld values of compounds. we trained our model on the rat acute oral toxicity data, tested and compared to two other existing models. despite the simplicity of our method, bestox has achieved a good performance with r of . , comparable to the single-task model proposed by toptox [ ] but slightly inferior to the hybrid decision tree and shallow neural network model by karim et al. [ ] . future improvement of bestox will be focused on extending the scope of datasets. as shown in the work of wu et al. [ ] , multitask learning can improve performance of prediction models due to availability of more data on different toxicity effects. the idea of multitask technique is to train a model with multiple training sets; each set corresponds to one toxicity prediction task. feeding the learners with different toxicity data helps them to learn common latent features of molecules offered by different datasets. recent efforts to elucidate the scientific validity of animalbased drug tests by the pharmaceutical industry, pro-testing lobby groups, and animal welfare organisations screening: methods for experimentation in industry, drug discovery, and genetics convolutional neural network based on smiles representation of compounds for detecting chemical motif a review on machine learning methods for in silico toxicity prediction efficient toxicity prediction via simple features using shallow neural networks and decision trees convolutional neural networks for sentence classification virtual screening for bioactive molecules rdkit: open-source cheminformatics exploration of the chemical space and its three historical regimes deeptox: toxicity prediction using deep learning virtual screening strategies in drug discovery chiral drugs: an overview open babel: an open chemical toolbox integrating virtual screening in lead discovery new promising approaches to treatment of chemotherapy-induced toxicities in silico toxicology: computational methods for the prediction of chemical toxicity understanding the basics of qsar for applications in pharmaceutical sciences and risk assessment improving the human hazard characterization of chemicals: a tox update dose finding in drug development smiles, a chemical language and information system. . introduction to methodology and encoding rules smiles. . algorithm for generation of unique smiles notation encyclopedia of toxicology quantitative toxicity prediction using topology based multitask deep neural networks machine learning based toxicity prediction: from chemical structural description to transcriptome analysis padel-descriptor: an open source software to calculate molecular descriptors and fingerprints character-level convolutional networks for text classification acknowledgments. this work was supported by university of macau (grant no. myrg - -fst). key: cord- -d rz dc authors: webb, b.; eswar, n.; fan, h.; khuri, n.; pieper, u.; dong, g.q.; sali, a. title: comparative modeling of drug target proteins date: - - journal: reference module in chemistry, molecular sciences and chemical engineering doi: . /b - - - - . - sha: doc_id: cord_uid: d rz dc in this perspective, we begin by describing the comparative protein structure modeling technique and the accuracy of the corresponding models. we then discuss the significant role that comparative prediction plays in drug discovery. we focus on virtual ligand screening against comparative models and illustrate the state-of-the-art by a number of specific examples. structure-based or rational drug discovery has already resulted in a number of drugs on the market and many more in the development pipeline. [ ] [ ] [ ] [ ] structure-based methods are now routinely used in almost all stages of drug development, from target identification to lead optimization. [ ] [ ] [ ] [ ] central to all structure-based discovery approaches is the knowledge of the threedimensional ( d) structure of the target protein or complex because the structure and dynamics of the target determine which ligands it binds. the d structures of the target proteins are best determined by experimental methods that yield solutions at atomic resolution, such as x-ray crystallography and nuclear magnetic resonance (nmr) spectroscopy. while developments in the techniques of experimental structure determination have enhanced the applicability, accuracy, and speed of these structural studies, , structural characterization of sequences remains an expensive and time-consuming task. the publicly available protein data bank (pdb) currently contains $ structures and grows at a rate of approximately % every years. on the other hand, the various genome-sequencing projects have resulted in over million sequences, including the complete genetic blueprints of humans and hundreds of other organisms. , this achievement has resulted in a vast collection of sequence information about possible target proteins with little or no structural information. current statistics show that the structures available in the pdb account for less than % of the sequences in the uniprot database. moreover, the rate of growth of the sequence information is more than twice that of the structures, and is expected to accelerate even more with the advent of readily available next-generation sequencing technologies. due to this wide sequence-structure gap, reliance on experimentally determined structures limits the number of proteins that can be targeted by structure-based drug discovery. fortunately, domains in protein sequences are gradually evolving entities that can be clustered into a relatively small number of families with similar sequences and structures. , for instance, - % of the sequences in the uniprot database have been grouped into fewer than domain families. , similarly, all the structures in the pdb have been classified into about distinct folds. , computational protein structure prediction methods, such as threading and comparative protein structure modeling, , strive to bridge the sequence-structure gap by utilizing these evolutionary relationships. the speed, low cost, and relative accuracy of these computational methods have led to the use of predicted d structures in the drug discovery process. , the other class of prediction methods, de novo or ab initio methods, attempts to predict the structure from sequence alone, without reliance on evolutionary relationships. however, despite progress in these methods, [ ] [ ] [ ] especially for small proteins with fewer than amino acid residues, comparative modeling remains the most reliable method of predicting the d structure of a protein, with an accuracy that can be comparable to a low-resolution, experimentally determined structure. the basis of comparative modeling the primary requirement for reliable comparative modeling is a detectable similarity between the sequence of interest (target sequence) and a known structure (template). as early as , chothia and lesk showed that there is a strong correlation between sequence and structural similarities. this correlation provides the basis of comparative modeling, allows a coarse assessment of model errors, and also highlights one of its major challenges: modeling the structural differences between the template and target structures (figure ). comparative modeling stands to benefit greatly from the structural genomics initiative. structural genomics aims to achieve significant structural coverage of the sequence space with an efficient combination of experimental and prediction methods. this goal is pursued by careful selection of target proteins for structure determination by x-ray crystallography and nmr spectroscopy, such that most other sequences are within 'modeling distance' (e.g., > % sequence identity) of a known structure. , , , the expectation is that the determination of these structures combined with comparative modeling will yield useful structural information for the largest possible fraction of sequences in the shortest possible timeframe. the impact of structural genomics is illustrated by comparative modeling based on the structures determined by the new york structural genomics research consortium. for each new structure without a close homolog in the pdb, on average, protein sequences without any prior structural characterization could be modeled at least at the level of the fold. thus, the structures of most proteins will eventually be predicted by computation, not determined by experiment. in this review, we begin by describing the various steps involved in comparative modeling. next, we emphasize two aspects of model refinement, loop modeling and side-chain modeling, due to their relevance in ligand docking and rational drug discovery. we then discuss the errors in comparative models. finally, we describe the role of comparative modeling in drug discovery, focusing on ligand docking against comparative models. we compare successes of docking against models and x-ray structures, and illustrate the computational docking against models with a number of examples. we conclude with a summary of topics that will impact on the future utility of comparative modeling in drug discovery, including an automation and integration of resources required for comparative modeling and ligand docking. comparative modeling consists of four main steps (figure (a)): ( ) fold assignment that identifies similarity between the target sequence of interest and at least one known protein structure (the template); ( ) alignment of the target sequence and the template(s); ( ) building a model based on the alignment with the chosen template(s); and ( ) predicting model errors. although fold assignment and sequence-structure alignment are logically two distinct steps in the process of comparative modeling, in practice almost all fold assignment methods also provide sequence-structure alignments. in the past, fold assignment methods were optimized for better sensitivity in detecting remotely related homologs, often at the cost of alignment accuracy. however, recent methods simultaneously optimize both the sensitivity and alignment accuracy. therefore, in the following discussion, we will treat fold assignment and sequence-structure alignment as a single protocol, explaining the differences as needed. as mentioned earlier, the primary requirement for comparative modeling is the identification of one or more known template structures with detectable similarity to the target sequence. the identification of suitable templates is achieved by scanning structure databases, such as pdb, scop, dali, and cath, with the target sequence as the query. the detected similarity is usually quantified in terms of sequence identity or statistical measures, such as e-value or z-score, depending on the method used. sequence-structure relationships are coarsely classified into three different regimes in the sequence similarity spectrum: ( ) the easily detected relationships characterized by > % sequence identity; ( ) the 'twilight zone,' corresponding to relationships figure average model accuracy as a function of sequence identity. as the sequence identity between the target sequence and the template structure decreases, the average structural similarity between the template and the target also decreases (dashed line, triangles). structural overlap is defined as the fraction of equivalent c a atoms. for the comparison of the model with the actual structure (filled circles), two c a atoms were considered equivalent if they belonged to the same residue and were within . Å of each other after least-squares superposition. for comparisons between the template structure and the actual target structure (triangles), two c a atoms were considered equivalent if they were within . Å of each other after alignment and rigid-body superposition. the difference between the model and the actual target structure is a combination of the target-template differences (green area) and the alignment errors (red area). the figure was constructed by calculating comparative models based on a single template of varying similarity to the targets. all targets had known (experimentally determined) structures. with statistically significant sequence similarity in the - % range; and ( ) the 'midnight zone,' corresponding to statistically insignificant sequence similarity. for closely related protein sequences with identities higher than - %, the alignments produced by all methods are almost always largely correct. the quickest way to search for suitable templates in this regime is to use simple pairwise sequence alignment methods such as ssearch, blast, and fasta. brenner et al. showed that these methods detect only $ % of the homologous pairs at less than % sequence identity, while they identify more than % of the relationships when sequence identity is between % and %. another benchmark, based on reference structural alignments with - % sequence identity, indicated that blast is able to correctly align only % of the residue positions. the sensitivity of the search and accuracy of the alignment become progressively difficult as the relationships move into the twilight zone. , a significant improvement in this area was the introduction of profile methods by gribskov and co-workers. the profile of a sequence is derived from a multiple sequence alignment and specifies residue-type occurrences for each alignment position. the information in a multiple sequence alignment is most often encoded as either a position-specific scoring matrix (pssm) , , or as a hidden markov model (hmm). , to identify suitable templates for comparative modeling, the profile of the target sequence is used to search against a database of template sequences. the profile-sequence methods are more sensitive in detecting related structures in the twilight zone than the pairwise sequence-based methods; they detect approximately twice the number of homologs under % sequence identity. , , the resulting profile-sequence alignments correctly align approximately - % of residues in the - % sequence identity range; , this number is almost twice as large as that of the pairwise sequence methods. frequently used programs for profile-sequence alignment are psi-blast, sam, hmmer, hhsearch, hhblits, and build_profile. as a natural extension, the profile-sequence alignment methods have led to profile-profile alignment methods that search for suitable template structures by scanning the profile of the target sequence against a database of template profiles, as opposed to a database of template sequences. these methods have proven to include the most sensitive and accurate fold assignment and figure comparative protein structure modeling. (a) a flowchart illustrating the steps in the construction of a comparative model. (b) description of comparative modeling by extraction of spatial restraints as implemented in modeller. by default, spatial restraints in modeller include: ( ) homology-derived restraints from the aligned template structures; ( ) statistical restraints derived from all known protein structures; and ( ) stereochemical restraints from the charmm- molecular mechanics force field. these restraints are combined into an objective function that is then optimized to calculate the final d model of the target sequence. alignment protocols to date. , [ ] [ ] [ ] profile-profile methods detect $ % more relationships at the superfamily level and improve the alignment accuracy by - % compared to profile-sequence methods. , there are a number of variants of profile-profile alignment methods that differ in the scoring functions they use. , , [ ] [ ] [ ] [ ] [ ] [ ] [ ] however, several analyses have shown that the overall performances of these methods are comparable. , [ ] [ ] [ ] some of the programs that can be used to detect suitable templates are ffas, sp , salign, hhblits, hhsearch, and ppscan. sequence-structure threading methods as the sequence identity drops below the threshold of the twilight zone, there is usually insufficient signal in the sequences or their profiles for the sequence-based methods discussed above to detect true relationships. sequence-structure threading methods are most useful in this regime as they can sometimes recognize common folds, even in the absence of any statistically significant sequence similarity. these methods achieve higher sensitivity by using structural information derived from the templates. the accuracy of a sequence-structure match is assessed by the score of a corresponding coarse model and not by sequence similarity, as in sequence comparison methods. the scoring scheme used to evaluate the accuracy is either based on residue substitution tables dependent on structural features such as solvent exposure, secondary structure type, and hydrogen bonding properties, , - or on statistical potentials for residue interactions implied by the alignment. [ ] [ ] [ ] [ ] [ ] the use of structural data does not have to be restricted to the structure side of the aligned sequence-structure pair. for example, sam-t makes use of the predicted local structure for the target sequence to enhance homolog detection and alignment accuracy. commonly used threading programs are genthreader, , d-pssm, fugue, sp , sam-t multitrack hmm, , , and muster. iterative sequence-structure alignment yet another strategy is to optimize the alignment by iterating over the process of calculating alignments, building models, and evaluating models. such a protocol can sample alignments that are not statistically significant and identify the alignment that yields the best model. although this procedure can be time-consuming, it can significantly improve the accuracy of the resulting comparative models in difficult cases. regardless of the method used, searching in the twilight and midnight zones of the sequence-structure relationship often results in false negatives, false positives, or alignments that contain an increasingly large number of gaps and alignment errors. improving the performance and accuracy of methods in this regime remains one of the main tasks of comparative modeling today. it is imperative to calculate an accurate alignment between the target-template pair. although some progress has been made recently, comparative modeling can rarely recover from an alignment error. after a list of all related protein structures and their alignments with the target sequence have been obtained, template structures are prioritized depending on the purpose of the comparative model. template structures may be chosen purely based on the targettemplate sequence identity or a combination of several other criteria, such as experimental accuracy of the structures (resolution of x-ray structures, number of restraints per residue for nmr structures), conservation of active-site residues, holo-structures that have bound ligands of interest, and prior biological information that pertains to the solvent, ph, and quaternary contacts. it is not necessary to select only one template. in fact, the use of several templates approximately equidistant from the target sequence generally increases the model accuracy. , model building once an initial target-template alignment is built, a variety of methods can be used to construct a d model for the target protein. , , [ ] [ ] [ ] [ ] the original and still widely used method is modeling by rigid-body assembly. , , this method constructs the model from a few core regions, and from loops and side chains that are obtained by dissecting related structures. commonly used programs that implement this method are composer, - d-jigsaw, rosettacm, and swiss-model. another family of methods, modeling by segment matching, relies on the approximate positions of conserved atoms from the templates to calculate the coordinates of other atoms. [ ] [ ] [ ] [ ] [ ] an instance of this approach is implemented in segmod. the third group of methods, modeling by satisfaction of spatial restraints, uses either distance geometry or optimization techniques to satisfy spatial restraints obtained from the alignment of the target sequences with the template structures. , [ ] [ ] [ ] [ ] specifically, modeller, , , our own program for comparative modeling, belongs to this group of methods. modeller implements comparative protein structure modeling by the satisfaction of spatial restraints that include: ( ) homologyderived restraints on the distances and dihedral angles in the target sequence, extracted from its alignment with the template structures; ( ) stereochemical restraints such as bond length and bond angle preferences, obtained from the charmm- molecular mechanics force field; ( ) statistical preferences for dihedral angles and nonbonded interatomic distances, obtained from a representative set of known protein structures; and ( ) optional manually curated restraints, such as those from nmr spectroscopy, rules of secondary structure packing, cross-linking experiments, fluorescence spectroscopy, image reconstruction from electron microscopy, site-directed mutagenesis, and intuition ( figure (b) ). the spatial restraints, expressed as probability density functions, are combined into an objective function that is optimized by a combination of conjugate gradients and molecular dynamics with simulated annealing. this model-building procedure is similar to structure determination by nmr spectroscopy. accuracies of the various model-building methods are relatively similar when used optimally. , other factors, such as template selection and alignment accuracy, usually have a larger impact on the model accuracy, especially for models based on less than % sequence identity to the templates. however, it is important that a modeling method allows a degree of flexibility and automation to obtain better models more easily and rapidly. for example, a method should allow for an easy recalculation of a model when a change is made in the alignment; it should be straightforward to calculate models based on several templates; and the method should provide tools for incorporation of prior knowledge about the target (e.g., cross-linking restraints and predicted secondary structure). protein sequences evolve through a series of amino acid residue substitutions, insertions, and deletions. while substitutions can occur throughout the length of the sequence, insertions and deletions mostly occur on the surface of proteins in segments that connect regular secondary structure segments (i.e., loops). while the template structures are helpful in the modeling of the aligned target backbone segments, they are generally less valuable for the modeling of side chains and irrelevant for the modeling of insertions such as loops. the loops and side chains of comparative models are especially important for ligand docking; thus, we discuss them in the following two sections. loop modeling is an especially important aspect of comparative modeling in the range from % to % sequence identity. in this range of overall similarity, loops among the homologs vary while the core regions are still relatively conserved and aligned accurately. loops often play an important role in defining the functional specificity of a given protein, forming the active and binding sites. loop modeling can be seen as a mini protein folding problem because the correct conformation of a given segment of a polypeptide chain has to be calculated mainly from the sequence of the segment itself. however, loops are generally too short to provide sufficient information about their local fold. even identical decapeptides in different proteins do not always have the same conformation. , some additional restraints are provided by the core anchor regions that span the loop and by the structure of the rest of the protein that cradles the loop. although many loop-modeling methods have been described, it is still challenging to correctly and confidently model loops longer than approximately - residues. , , two classes of methods there are two main classes of loop-modeling methods: ( ) database search approaches that scan a database of all known protein structures to find segments fitting the anchor core regions , ; and ( ) conformational search approaches that rely on optimizing a scoring function. [ ] [ ] [ ] there are also methods that combine these two approaches. , database-based loop modeling the database search approach to loop modeling is accurate and efficient when a database of specific loops is created to address the modeling of the same class of loops, such as b-hairpins, or loops on a specific fold, such as the hypervariable regions in the immunoglobulin fold. , there are attempts to classify loop conformations into more general categories, thus extending the applicability of the database search approach. [ ] [ ] [ ] however, the database methods are limited because the number of possible conformations increases exponentially with the length of a loop, and until the late s only loops up to residues long could be modeled using the database of known protein structures. , however, the growth of the pdb in recent years has largely eliminated this problem. optimization-based methods there are many optimization-based methods, exploiting different protein representations, objective functions, and optimization or enumeration algorithms. the search algorithms include the minimum perturbation method, dihedral angle search through a rotamer library, , molecular dynamics simulations, , genetic algorithms, monte carlo and simulated annealing, [ ] [ ] [ ] multiple-copy simultaneous search, self-consistent field optimization, robotics-inspired kinematic closure and enumeration based on graph theory. the accuracy of loop predictions can be further improved by clustering the sampled loop conformations and partially accounting for the entropic contribution to the free energy. another way of improving the accuracy of loop predictions is to consider the solvent effects. improvements in implicit solvation models, such as the generalized born solvation model, motivated their use in loop modeling. the solvent contribution to the free energy can be added to the scoring function for optimization, or it can be used to rank the sampled loop conformations after they are generated with a scoring function that does not include the solvent terms. , [ ] [ ] [ ] side-chain modeling two simplifications are frequently applied in the modeling of side-chain conformations. first, amino acid residue replacements often leave the backbone structure almost unchanged, allowing us to fix the backbone during the search for the best side-chain conformations. second, most side chains in high-resolution crystallographic structures can be represented by a limited number of conformers that comply with stereochemical and energetic constraints. this observation motivated ponder and richards to develop the first library of side-chain rotamers for the types of residues with dihedral angle degrees of freedom in their side chains, based on high-resolution protein structures determined by x-ray crystallography. subsequently, a number of additional libraries have been derived. [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] rotamers rotamers on a fixed backbone are often used when all the side chains need to be modeled on a given backbone. this approach reduces the combinatorial explosion associated with a full conformational search of all the side chains, and is applied by some comparative modeling and protein design approaches. however, $ % of the side chains cannot be represented well by these libraries. in addition, it has been shown that the accuracy of side-chain modeling on a fixed backbone decreases rapidly when the backbone errors are larger than . Å . earlier methods for side-chain modeling often put less emphasis on the energy or scoring function. the function was usually greatly simplified, and consisted of the empirical rotamer preferences and simple repulsion terms for nonbonded contacts. nevertheless, these approaches have been justified by their performance. for example, a method based on a rotamer library compared favorably with that based on a molecular mechanics force field, and new methods continue to be based on the rotamer library approach. [ ] [ ] [ ] the various optimization approaches include a monte carlo simulation, simulated annealing, a combination of monte carlo and simulated annealing, the dead-end elimination theorem, , genetic algorithms, neural network with simulated annealing, mean field optimization, and combinatorial searches. , , several studies focused on the testing of more sophisticated potential functions for conformational search , and development of new scoring functions for side-chain modeling, reporting higher accuracy than earlier studies. the major sources of error in comparative modeling are discussed in the relevant sections above. the following is a summary of these errors, dividing them into five categories (figure ) . this error is a potential problem when distantly related proteins are used as templates (i.e., less than % sequence identity). distinguishing between a model based on an incorrect template and a model based on an incorrect alignment with a correct template is difficult. in both cases, the evaluation methods (below) will predict an unreliable model. the conservation of the key functional or structural residues in the target sequence increases the confidence in a given fold assignment. the single source of errors with the largest impact on comparative modeling is misalignments, especially when the target-template sequence identity decreases below %. alignment errors can be minimized in two ways. using the profile-based methods discussed above usually results in more accurate alignments than those from pairwise sequence alignment methods. another way of improving the alignment is to modify those regions in the alignment that correspond to predicted errors in the model. segments of the target sequence that have no equivalent region in the template structure (i.e., insertions or loops) are one of the most difficult regions to model. again, when the target and template are distantly related, errors in the alignment can lead to incorrect positions of the insertions. using alignment methods that incorporate structural information can often correct such errors. once a reliable alignment is obtained, various modeling protocols can predict the loop conformation, for insertions of fewer than - residues. , , , distortions and shifts in correctly aligned regions as a consequence of sequence divergence, the main-chain conformation changes, even if the overall fold remains the same. therefore, it is possible that in some correctly aligned segments of a model, the template is locally different (< Å ) from the target, resulting in errors in that region. the structural differences are sometimes not due to differences in sequence, but are a consequence of artifacts in structure determination or structure determination in different environments (e.g., packing of subunits in a crystal). the simultaneous use of several templates can minimize this kind of error. , figure typical errors in comparative modeling. shown are the typical sources of errors encountered in comparative models. two of the major sources of errors in comparative modeling are due to incorrect templates or incorrect alignments with the correct templates. the modeling procedure can rarely recover from such errors. the next significant source of errors arises from regions in the target with no corresponding region in the template, i.e., insertions or loops. other sources of errors, which occur even with an accurate alignment, are due to rigid-body shifts, distortions in the backbone, and errors in the packing of side chains. as the sequences diverge, the packing of the atoms in the protein core changes. sometimes even the conformation of identical side chains is not conserved -a pitfall for many comparative modeling methods. side-chain errors are critical if they occur in regions that are involved in protein function, such as active sites and ligand-binding sites. the accuracy of the predicted model determines the information that can be extracted from it. thus, estimating the accuracy of a model in the absence of the known structure is essential for interpreting it. as discussed earlier, a model calculated using a template structure that shares more than % sequence identity is indicative of an overall accurate structure. however, when the sequence identity is lower, the first aspect of model evaluation is to confirm whether or not a correct template was used for modeling. it is often the case, when operating in this regime, that the fold assignment step produces only false positives. a further complication is that at such low similarities the alignment generally contains many errors, making it difficult to distinguish between an incorrect template on one hand and an incorrect alignment with a correct template on the other hand. there are several methods that use d profiles and statistical potentials, , , which assess the compatibility between the sequence and modeled structure by evaluating the environment of each residue in a model with respect to the expected environment, as found in native high-resolution experimental structures. these methods can be used to assess whether or not the correct template was used for the modeling. they include verify d, and tsvmod. even when the model is based on alignments that have > % sequence identity, other factors, including the environment, can strongly influence the accuracy of a model. for instance, some calcium-binding proteins undergo large conformational changes when bound to calcium. if a calcium-free template is used to model the calcium-bound state of the target, it is likely that the model will be incorrect, irrespective of the target-template similarity or accuracy of the template structure. the model should also be subjected to evaluations of self-consistency to ensure that it satisfies the restraints used to calculate it. additionally, the stereochemistry of the model (e.g., bond lengths, bond angles, backbone torsion angles, and nonbonded contacts) may be evaluated using programs such as procheck and whatcheck. although errors in stereochemistry are rare and less informative than errors detected by statistical potentials, a cluster of stereochemical errors may indicate that there are larger errors (e.g., alignment errors) in that region. when multiple models are calculated for the target based on a single template or when multiple loops are built for a single or multiple models, it is practical to select a subset of models or loops that are judged to be most suitable for subsequent docking calculations. if some known ligands or other information for the desired model is available, model selection should be guided by this known information. if this extra information is not available, model selection should aim to select the most accurate model. while models or loops can be selected by the energy function used for guiding the building of comparative models or the sampling of loop configurations, using a separate statistical potential for selecting the most accurate models or loops is often more successful. , , , it is crucial for method developers and users alike to assess the accuracy of their methods. an attempt to address this problem has been made by the critical assessment of techniques for proteins structure prediction (casp) and in the past by the critical assessment of fully automated structure prediction (cafasp) experiments, which is now integrated into casp. however, casp assesses methods only over a limited number of target protein sequences, and is conducted only every years. , to overcome this limitation, the new cameo web server continuously evaluates the accuracy and reliability of a number of comparative protein structure prediction servers, in a fully automated manner. every week, cameo provides each tested server with the prerelease sequences of structures that are to be shortly released by the pdb. each server then has days to build and return a d model of these sequences. when pdb releases the structures, cameo compares the models against the experimentally determined structures, and presents the results on its web site. this enables developers, non-expert users, and reviewers to determine the performance of the tested prediction servers. cameo is similar in concept to two prior such continuous testing servers, livebench and eva. , there is a wide range of applications of protein structure models ( figure ) . , [ ] [ ] [ ] [ ] [ ] [ ] [ ] for example, high-and medium-accuracy comparative models are frequently helpful in refining functional predictions that have been based on a sequence match alone because ligand binding is more directly determined by the structure of the binding site than by its sequence. it is often possible to predict correctly features of the target protein that do not occur in the template structure. , for example, the size of a ligand may be predicted from the volume of the binding site cleft and the location of a binding site for a charged ligand can be predicted from a cluster of charged residues on the protein. fortunately, errors in the functionally important regions in comparative models are many times relatively low because the functional regions, such as active sites, tend to be more conserved in evolution than the rest of the fold. even low-accuracy comparative models may be useful, for example, for assigning the fold of a protein. fold figure accuracy and applications of protein structure models. shown are the different ranges of applicability of comparative protein structure modeling, threading, and de novo structure prediction, their corresponding accuracies, and their sample applications. assignment can be very helpful in drug discovery, because it can shortcut the search for leads by pointing to compounds that have been previously developed for other members of the same family. , the remainder of this review focuses on the use of comparative models for ligand docking. [ ] [ ] [ ] comparative protein structure modeling extends the applicability of virtual screening beyond the atomic structures determined by x-ray crystallography or nmr spectroscopy. in fact, comparative models have been used in virtual screening to detect novel ligands for many protein targets, including the g-protein coupled receptors (gpcr), , [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] protein kinases, [ ] [ ] [ ] [ ] nuclear hormone receptors, and many different enzymes. [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] nevertheless, the relative utility of comparative models versus experimentally determined structures has only been relatively sparsely assessed. , , , [ ] [ ] [ ] the utility of comparative models for molecular docking screens in ligand discovery has been documented with the aid of protein targets selected from the 'directory of useful decoys' (dud). for each target sequence, templates for comparative modeling were obtained from the pdb, including at least one holo (ligand bound) and one apo (ligand free) template structure for each of the eight % sequence identity ranges from % to %. in total, models were generated based on templates for the test proteins using modeller v . dud ligands and decoys ( molecules) were screened against the holo x-ray structure, the apo x-ray structure, and the comparative models of each target using dock . . . the accuracy of virtual screening was evaluated by the overall ligand enrichment that was calculated by integrating the area under the enrichment plot (logauc). a key result was that, for % and % of the targets, at least one comparative model yielded ligand enrichment better or comparable to that of the corresponding holo and apo x-ray structure. this result indicates that comparative models can be useful docking targets when multiple templates are available. however, it was not possible to predict which model, out of all those used, led to the highest enrichment. therefore, a 'consensus' enrichment score was computed by ranking each library compound by its best docking score against all comparative models and/or templates. for % and % of the targets, the consensus enrichment for multiple models was better or comparable to that of the holo and apo x-ray structures, respectively, suggesting that multiple comparative models can be useful for virtual screening. despite problems with comparative modeling and ligand docking, comparative models have been successfully used in practice in conjunction with virtual screening to identify novel inhibitors. we briefly review a few of these success stories to highlight the potential of the combined comparative modeling and ligand-docking approach to drug discovery. comparative models have been employed to aid rational drug design against parasites for more than years. , , , as early as , ring et al. used comparative models for computational docking studies that identified low micromolar nonpeptidic inhibitors of proteases in malarial and schistosome parasite lifecycles. li et al. subsequently used similar methods to develop nanomolar inhibitors of falcipain that are active against chloroquine-resistant strains of malaria. in a study by selzer et al. comparative models were used to predict new nonpeptide inhibitors of cathepsin l-like cysteine proteases in leishmania major. sixty-nine compounds were selected by dock . as strong binders to a comparative model of protein cpb, and of these, had experimental ic values below mmol l À . finally, in a study by que et al., comparative models were used to rationalize ligand-binding affinities of cysteine proteases in entamoeba histolytica. specifically, this work provided an explanation for why proteases acp and acp had substrate specificity similar to that of cathepsin b, although their overall structure is more similar to that of cathepsin d. enyedy et al. discovered new inhibitors of matriptase by docking against its comparative model. the comparative model employed thrombin as the template, sharing only % sequence identity with the target sequence. moreover, some residues in the binding site are significantly different; a trio of charged asp residues in matriptase correspond to tyr and trp residues in thrombin. thrombin was chosen as the template, in part because it prefers substrates with positively charged residues at the p position, as does matriptase. the national cancer institute database was used for virtual screening that targeted the s site with the dock program. the best-scoring compounds were manually inspected to identify positively charged ligands (the s site is negatively charged), and compounds were experimentally screened for inhibition, identifying the inhibitors. one of them, hexamidine, was used as a lead to identify additional compounds selective for matriptase relative to thrombin. the wang group has also used similar methods to discover seven new, low-micromolar inhibitors of bcl- , using a comparative model based on the nmr solution structure of bcl-x l . schapira et al. discovered a novel inhibitor of a retinoic acid receptor by virtual screening using a comparative model. in this case, the target (rar-a) and template (rar-g) are very closely related; only three residues in the binding site are not conserved. the icm program was used for virtual screening of ligands from the available chemicals directory (acd). the high-scoring compounds identified in the first round were subsequently docked into a full atom representation of the receptor with flexible side chains to obtain a final set of good-scoring hits. these compounds were then manually inspected to choose the final for testing. two novel agonists were identified, with -nanomolar activity. zuccotto et al. identified novel inhibitors of dihydrofolate reductase (dhfr) in trypanosoma cruzi (the parasite that causes chagas disease) by docking into a comparative model based on $ % sequence identity to dhfr in l. major, a related parasite. the virtual screening procedure used dock for rigid docking of over selected compounds from the cambridge structural database (csd). visual inspection of the top hits was used to select compounds for experimental testing. this work identified several novel scaffolds with micromolar ic values. the authors report attempting to use virtual screening results to identify compounds with greater affinity for t. cruzi dhfr than human dhfr, but it is not clear how successful they were. following the outbreak of the severe acute respiratory syndrome (sars) in , anand et al. used the experimentally determined structures of the main protease from human coronavirus (m pro ) and an inhibitor complex of porcine coronavirus (transmissible gastroenteritis virus, tgev) m pro to calculate a comparative model of the sars coronavirus m pro . this model then provided a basis for the design of anti-sars drugs. in particular, a comparison of the active site residues in these and other related structures suggested that the ag inhibitor of the human rhinovirus type c protease is a good starting point for design of anticoronaviral drugs. comparative models of protein kinases combined with virtual screening have also been intensely used for drug discovery. , , [ ] [ ] [ ] the > kinases in the human genome, the relatively small number of experimental structures available, and the high level of conservation around the important adenosine triphosphate-binding site make comparative modeling an attractive approach toward structure-based drug discovery. g protein-coupled receptors are another interesting class of proteins that in principle allow drug discovery through comparative modeling. , [ ] [ ] [ ] [ ] approximately % of current drug targets belong to this class of proteins. however, these proteins have been extremely difficult to crystallize and most comparative modeling has been based on the atomic resolution structure of the bovine rhodopsin. despite this limitation, a rather extensive test of docking methods with rhodopsin-based comparative models shows encouraging results. the applicability of structure-based modeling and virtual screening has recently been expanded to membrane proteins that transport solutes, such as ions, metabolites, peptides, and drugs. in humans, these transporters contribute to the absorption, distribution, metabolism, and excretion of drugs, and often, mediate drug-drug interactions. additionally, several transporters can be targeted directly by small molecules. for instance, methylphenidate (ritalin) inhibiting the norepinephrine transporter (net) and, consequently, inhibiting the reuptake of norepinephrine, is used in the treatment of attention-deficit hyperactivity disorder (adhd). schlessinger et al. predicted putative ligands of human net by docking drugs from the kyoto encyclopedia of genes (kegg drug) into a comparative model based on $ % sequence identity to leucine transporter (leut) from aquifex aeolicus. of these predicted ligands, ten were validated by cis-inhibition experiments; five of them were chemically novel. close examination of the predicted primary binding site helped rationalize interactions of net with its primary substrate, norepineprhine, as well as positive and negative pharmacological effects of other net ligands. subsequently, schlessinger et al. modeled two different conformations of the human gaba transporter (gat- ), using the leut structures in occluded-outward-facing and outward-facing conformations. enrichment calculations were used to assess the quality of the models in molecular dynamics simulations and side-chain refinements. the key residue, glu , interacting with the substrate was identified during the refinement of the models and validated by site-directed mutagenesis. docking against two conformations of the transporter enriches for different physicochemical properties of ligands. for example, top-scoring ligands found by docking against the outward-facing model were bulkier and more hydrophobic than those predicted using the occluded-outward-facing model. among twelve ligands validated in cis-inhibition assays, six were chemically novel (e.g., homotaurine). based on the validation experiments, gat- is likely to be a high-selectivity/low-affinity transporter. following these two studies, a combination of comparative modeling, ligand docking, and experimental validation was used to rationalize toxicity of an anti-cancer agent, acivicin. the toxic sideeffects are thought to be facilitated by the active transport of acivicin through the blood-brain-barrier (bbb) via the large-neutral amino acid transporter (lat- ). in addition, four small-molecule ligands of lat- were identified by docking against a comparative model based on two templates, the structures of the outward-occluded arginine-bound arginine/agmatine transporter adic from e. coli and the inward-apo conformation of the amino acid, polyamine, and organo-cation transporter apct from methanococcus jannaschii. two of the four hits, acivicin and fenclonine, were confirmed as substrates by a trans-stimulation assay. these studies clearly illustrate the applicability of combined comparative modeling and virtual screening to ligand discovery for transporters. although reports of successful virtual screening against comparative models are encouraging, such efforts are not yet a routine part of rational drug design. even the successful efforts appear to rely strongly on visual inspection of the docking results. much work remains to be done to improve the accuracy, efficiency, and robustness of docking against comparative models. despite assessments of relative successes of docking against comparative models and native x-ray structures, , relatively little has been done to compare the accuracy achievable by different approaches to comparative modeling and to identify the specific structural reasons why comparative models generally produce less accurate virtual screening results than the holo structures. among the many issues that deserve consideration are the following: • the inclusion of cofactors and bound water molecules in protein receptors is often critical for success of virtual screening; however, cofactors are not routinely included in comparative models • most docking programs currently retain the protein receptor in a rigid conformation. while this approach is appropriate for 'lock-and-key' binding modes, it does not work when the ligand induces conformational changes in the receptor upon binding. a flexible receptor approach is necessary to address such induced-fit cases , • the accuracy of comparative models is frequently judged by the c a root mean square error or other similar measures of backbone accuracy. for virtual screening, however, the precise positioning of side chains in the binding site is likely to be critical; measures of accuracy for binding sites are needed to help evaluate the suitability of comparative modeling algorithms for constructing models for docking • knowledge of known inhibitors, either for the target protein or the template, should help to evaluate and improve virtual screening against comparative models. for example, comparative models constructed from holo template structures implicitly preserve some information about the ligand-bound receptor conformation • improvement in the accuracy of models produced by comparative modeling will require methods that finely sample protein conformational space using a free energy or scoring function that has sufficient accuracy to distinguish the native structure from the nonnative conformations. despite many years of development of molecular simulation methods, attempts to refine models that are already relatively close to the native structure have met with relatively little success. this failure is likely to be due in part to inaccuracies in the scoring functions used in the simulations, particularly in the treatment of electrostatics and solvation effects. a combination of physics-based energy function with the statistical information extracted from known protein structures may provide a route to the development of improved scoring functions • improvements in sampling strategies are also likely to be necessary, for both comparative modeling and flexible docking given the increasing number of target sequences for which no experimentally determined structures are available, drug discovery stands to gain immensely from comparative modeling and other in silico methods. despite unsolved problems in virtually every step of comparative modeling and ligand docking, it is highly desirable to automate the whole process, starting with the target sequence and ending with a ranked list of its putative ligands. automation encourages development of better methods, improves their testing, allows application on a large scale, and makes the technology more accessible to both experts and non-specialists alike. through large-scale application, new questions, such as those about ligand-binding specificity, can in principle be addressed. enabling a wider community to use the methods provides useful feedback and resources toward the development of the next generation of methods. there are a number of servers for automated comparative modeling (table ) . however, in spite of automation, the process of calculating a model for a given sequence, refining its structure, as well as visualizing and analyzing its family members in the sequence and structure spaces can involve the use of scripts, local programs, and servers scattered across the internet and not necessarily interconnected. in addition, manual intervention is generally still needed to maximize the accuracy of the models in the difficult cases. the main repository for precomputed comparative models, the protein model portal, , , begins to address these deficiencies by serving models from several modeling groups, including the swiss-model and modbase databases. it provides access to web-based comparative modeling tools, cross-links to other sequence and structure databases, and annotations of sequences and their models. a number of databases containing comparative models and web servers for computing comparative models are publicly available. the protein model portal (pmp) , , centralizes access to these models created by different methodologies. the pmp is being developed as a module of the protein structure initiative knowledgebase (psi kb) and functions as a meta server for comparative models from external databases, including swiss-model and modbase, additionally to being a repository for comparative models that are derived from structures determined by the psi centers. it provides quality estimations of the deposited models, access to web-based comparative modeling tools, cross-links to other sequence and structure databases, annotations of sequences and their models, and detailed tutorials on comparative modeling and the use of their tools. the pmp currently contains . million comparative models for . million uniprot sequences (august ). a schematic of our own attempt at integrating several useful tools for comparative modeling is shown in figure . , modbase is a database that currently contains $ million predicted models for domains in approximately . million unique sequences from uniprot, ensembl, genbank, and private sequence datasets. the models were calculated using modpipe , and modeller. the web interface to the database allows flexible querying for fold assignments, sequence-structure alignments, models, and model assessments. an integrated sequence-structure viewer, chimera, allows inspection and analysis of the query results. models can also be calculated using modweb, , a web interface to modpipe, and stored in modbase, which also makes them accessible through the pmp. other resources associated with modbase include a comprehensive database of multiple protein structure alignments (dbali), structurally defined ligand-binding sites, structurally defined binary domain interfaces (pibase), , predictions of ligand-binding sites, interactions between yeast proteins, and functional consequences of human nssnps (ls-snp). , , a number of associated web services handle modeling of loops in protein structures (modloop), , evaluation of models (modeval), fitting of models against small angle x-ray scattering (saxs) profiles (foxs), - modeling of ligand-induced protein dynamics such as allostery (allosmod), , prediction of the ensemble of conformations that best fit a given saxs profile (allosmod-foxs), prediction of cryptic binding sites, scoring protein-ligand complexes based on a , compared to protein structure prediction, the attempts at automation and integration of resources in the field of docking for virtual screening are still in their nascent stages. one of the successful efforts in this direction is zinc, , a publicly available database of commercially available drug-like compounds, developed in the laboratory of brian shoichet. zinc contains more than million 'ready-to-dock' compounds organized in several subsets and allows the user to query the compounds by molecular properties and constitution. the shoichet group also provides a dockblaster service that enables end-users to dock the zinc compounds against their target structures using dock. , in the future, we will no doubt see efforts to improve the accuracy of comparative modeling and ligand docking. but perhaps as importantly, the two techniques will be integrated into a single protocol for more accurate and automated docking of ligands against sequences without known structures. as a result, the number and variety of applications of both comparative modeling and ligand docking will continue to increase. cameo http://cameo d.org/ casp http://predictioncenter.llnl.gov figure an integrated set of resources for comparative modeling. various databases and programs required for comparative modeling and docking are usually scattered over the internet, and require manual intervention or a good deal of expertise to be useful. automation and integration of these resources are efficient ways to put these resources in the hands of experts and non-specialists alike. we have outlined a comprehensive interconnected set of resources for comparative modeling and hope to integrate it with a similar effort in the area of ligand docking made by the shoichet group. , drug disc comparative protein structure modeling modeller comparative protein structure modeling and its applications to drug discovery comparative protein structure modeling this article is partially based on papers by jacobson and sali, fiser and sali, and madhusudhan et al. we also acknowledge the funds from sandler family supporting foundation, nih r gm , p gm , p a , and u gm , as well as sun, ibm, and intel for hardware gifts. key: cord- -ywz w po authors: maus, carsten title: component-based modelling of rna structure folding date: journal: computational methods in systems biology doi: . / - - - - _ sha: doc_id: cord_uid: ywz w po rna structure is fundamentally important for many biological processes. in the past decades, diverse structure prediction algorithms and tools were developed but due to missing descriptions in clearly defined modelling formalisms it’s difficult or even impossible to integrate them into larger system models. we present an rna secondary structure folding model described in ml-devs, a variant of the devs formalism, which enables the hierarchical combination with other model components like rna binding proteins. an example of transcriptional attenuation will be given where model components of rna polymerase, the folding rna molecule, and the translating ribosome play together in a composed dynamic model. single stranded ribonucleic acids (rna) are able to fold into complex threedimensional structures like polypeptide chains of proteins do. the structure of rna molecules is fundamentally important for their function, e.g. the well studied structures of trna and the different rrna variants. but also other transcripts of the dna, i.e. mostly mrnas, perform structure formation which has been shown to be essential for many regulatory processes like transcription termination and translation initiation [ , , ] . the shape of a folded rna molecule can also define binding domains for proteins or small target molecules which can be found for example within riboswitches [ ] . the enormous relevance for many biological key processes led to raised research efforts in identifying various rna structures over the past decades. unfortunately the experimental structure identification with nmr and x-ray techniques is difficult, expensive, and highly time-consuming. therefore, many in silico methods for rna structure prediction were developed which cover different requirements. diverse comparative methods exist using alignments of similar rna sequences to predict structures [ , ] , but also many single sequence prediction algorithms work very well. some of them predict the most stable rna structure in a thermodynamical equilibrium, e.g. [ , , ] , whereas some other simulate the kinetic folding pathway over time [ , , , ] . the latter is also in the focus of the presented modelling approach here. results of rna structure predictions as well as kinetic folding simulations have reached a high level of accuracy and thus in silico folding became a widely used and well established technique in the rna community. however, none of the existing tools and programs provides a flexible integration into larger system models which is also due to the fact that they are written in proprietary formalisms and do not distinguish between model description and simulation engine. to illuminate the importance of the folding processes and the possibility to integrate them into larger models, lets take a look at a concrete example of gene regulation. the tryptophan (trp) operon within bacterial genomes represents one of the best understood cases of gene regulation and has been subject to various modelling approaches [ , ] . tryptophan is an amino acid, a building block for the production of proteins. the trp operon includes five consecutive genes, coding for five proteins. the joint action of the proteins permits the synthesis of trp through a cascade of enzymatic reactions. this ability is vital since the bacterium may be unable to feed on trp from the environment. as long as trp is obtained from the surrounding medium, its costly synthesis is impaired by a threefold control mechanism: repression of transcription initiation, transcriptional attenuation, and inactivation of the cascade of enzymatic reactions actually producing trp. each of these are triggered by trp availability. transcriptional attenuation follows if transcription starts although trp was available. this has a small but non-negligible chance. as soon as rna polymerase (rnap) has transcribed the operon's leader region into an mrna molecule, a ribosome can access this. the ribosome starts translating the mrna content into a growing sequence of amino acids. the speed of the ribosome depends on trp availability. the ribosome advances quickly as long as trp is abundant, which prevents rnap from proceeding into the operon's coding region. the attenuation is caused by the formation of a certain constellation of rna hairpin loops in presence of a trp molecule at a distinct segment of the mrna molecule (figure ). attenuation depends on the synchronised advance of both rnap and ribosome, and their relative positioning with respect to mrna. in [ ] a model of the tryptophan operon was developed, in which repressors, the operon region, and the mrna were modelled individually. however, the latter in not much detail. only the repression of transcription initiation was included in the model. consequently, the simulation result showed stochastic bursts in the trp level, caused by the repressor falling off, which started the transcription. integrating a transcriptional attenuation model would have prevented this non-realistic increase in trp concentration, and mimicked the threefold regulation of the trp more realistically. however, the question is how to model the attenuation. as this regulating process depends largely on structure formation, modelling of rna folding would be a big step in the right direction for reflecting attenuation dynamics. additionally, modelling interactions between mrna and rnap as well as mrna and the ribosome are needed because both influence the kinetic folding process and rna termination structures break up gene transcription. the focus of this paper is the rna folding process, but at the end we will also give a detailed outlook how the composed model of tryptophan attenuation looks like and how the individual model components act together. the reason for rna folding is the molecules' general tendency to reach the most favourable thermodynamical state. complementary bases of rna nucleotides can form base pairs by building hydrogen bonds similar to dna double helices. adenine (a) and uracil (u) are complementary bases as well as cytosine (c) and guanine (g). in addition, the wobble base pair g-u is also frequently found in rna structure folding. each additional hydrogen bond of base pairs affords a small energy contribution to the overall thermodynamic stability, but there is another chemical interaction which is even more important for the rna structure than just the number and type of base pairs. it's called base stacking and describes the interactions between the aromatic rings of adjacent bases by van-der-waals bonds. base pair stacking is the cause why an uninterrupted long helix is thermodynamically more favourable than a structure of multiple single base pairs or short helices interrupted by loop regions, even if the number and type of base pairs are equal. since the s, significant progress has been done on identifying thermodynamic parameters of different base pair neighbourhoods and structural elements like hairpin loops, e.g. [ , ] . this was a precondition to develop rna structure prediction algorithms based on energy minimisation, i.e. finding the thermodynamical most stable structure. rna structures are hierarchically organised (see figure ). the most simple hierarchy level is the primary structure which is nothing else than the linear sequence of nucleotides. two nucleotides are linked over the ' and ' carbon atoms of their ribose sugar parts resulting in a definite strand direction. the secondary structure consists of helices formed by base pairs and intersecting loop regions. such structural elements are formed rapidly within the first milliseconds of the folding process [ ] . interacting secondary structure elements finally build the overall three-dimensional shape of rna molecules. although they are formed by simple base pairs like secondary structures, helices inside loop regions are often seen as tertiary structures. such pseudoknots and higher order tertiary interactions are, due to their complexity and analog to many other rna structure prediction methods, not covered by our model. however, it should not retain unstated here that there are some existing tools which can predict pseudoknots quite well, e.g. [ ] . primary structure secondary structure tertiary structure as already mentioned, typically kinetic rna folding simulations, as e.g. [ , , ] , are aimed at efficiently and accurately simulating the molecules structure formation in isolation rather than supporting a reuse of rna folding models and a hierarchical construction of models. for approaching such model composition, we use the modelling formalism ml-devs [ ] , a variant of the devs formalism [ ] . as devs does, it supports a modular-hierarchical modelling and allows to define composition hierarchies. ml-devs extends devs by supporting variable structures, dynamic ports, and multi-level modelling. the latter is based on two ideas. the first is to equip the coupled model with a state and a behaviour of its own, such that macro behaviour the macro level does not appear as a separate unit (an executive) of the coupled model. please recall that in traditional devs coupled models do not have an own state nor a behaviour. secondly, we have to explicitly define how the macro level affects the micro level and vice versa. both tasks are closely interrelated. we assume that models are still triggered by the flow of time and the arrival of events. obviously, one means to propagate information from macro to micro level is to exchange events between models. however, this burdens typically modelling and simulation unnecessarily, e.g. in case the dynamics of a micro model has to take the global state into consideration. therefore, we adopt the idea of value couplings. information at macro level is mapped to specific port names at micro level. each micro model may access information about macro variables by defining input ports with corresponding names. thus, downward causation (from macro to micro) is supported. in the opposite direction, the macro level needs access to crucial information at the micro level. for this purpose, we equip micro models with the ability to change their ports and to thereby signalise crucial state changes to the outside world. upward causation is supported, as the macro model has an overview of the number of micro models being in a particular state and to take this into account when updating the state at macro level. therefore, a form of invariant is defined whose violation initiates a transition at macro level. in the downward direction, the macro level can directly activate its components by sending them events -thereby, it becomes possible to synchronously let several micro models interact which is of particular interest when modelling chemical reactions. these multi-level extensions facilitate modelling, figure depicts the basic idea, see also [ ] . the central unit in composed models using rna structure information is an rna folding model. therefore, we first developed a model component which describes the folding kinetics of single stranded rna molecules. it consists of a coupled ml-devs model representing the whole rna molecule and several atomic models. each nucleotide (nt) of the rna strand is represented by an instance of the atomic model nucleotide which is either of the type a, c, g, or u meaning its base. they are connected via ports in the same order as the input sequence (primary structure) and have knowledge about their direct neighbours. for example, the nt at sequence position is connected with the nt number on its ' side and on the ' location it is connected with the nt at position (see figure ). state variables hold rudimentary information about the neighbours, to be exact their base type and current binding partners. "binding partner" means a secondary structure defining base pair and the term is used only in this context here and does not mean the primary backbone connections. if a partner of a nucleotide changes, an output message will be generated and the receiving (neighboured) nucleotides will update their state variables. holding information about other atomic model states is normally not the case in devs models as they are typically seen as black boxes. however, here it is quite useful because of some dependencies concerning base pair stability. base pairs are modelled by wide range connections of nucleotides via additional interfaces. whereas the rna backbone bonds of adjacent nucleotides are fixed after model initialisation, the connections between base pairing nucleotides are dynamically added and removed during simulation (figure ). therefore, two different major states (phases) of nucleotides exist: they can be either unpaired or paired. as already stated in section . , base pair stability depends on the involved bases and their neighbourhood, especially stacking energies of adjacent base pairs provide crucial contributions for structure stabilisation. in our kinetic folding model, base pair stability is reflected by binding duration, i.e. the time advance function of the paired phase. thus, pairing time depends on thermodynamic parameters for nucleic acids base stacking which were taken from [ ] and are also be used by mfold version . [ ] . this thermodynamic data set not only provides the free energy (gibbs energy) for a given temperature of • c, but also the enthalpy change Δh of various stacking situations. the enthalpy together with the free energy and the absolute temperature allows us to calculate the entropy change Δs which allows us further to calculate the activation energy Δe a for base pair dissociation at any temperature t between and • c: Δe a is directly used as one parameter for base pair opening time, i.e. the duration of a paired phase is directly dependent on the activation energy for base pair disruption. to allow rna structures to escape from local energy minima and refold to more stable structures, the base pair dissociation time will be randomised, which leads to very short bonding times in some cases although the activation energy needed for opening the base pair is quite large. for base pair closing an arbitrary short random time is assigned with the unpaired phase of nucleotides assuming that rna base pair formation is a very fast and random process. after unpaired time has expired, the nucleotide model tries to build a base pair with another nucleotide randomly chosen from within a set of possible pairing partner positions. this set is determined by sterically available rna strand regions and thus an abstraction of spatial constraints. for example, a hairpin loop smaller than nucleotides is sterically impossible, but also many other nucleotide positions with larger distance can be excluded for secondary structure folding (see figure ). an unpaired nucleotide is not able to choose another nt for base pairing by its own. it has no global information about the rna shape which is strongly needed here. therefore, an implicit request by adding a new input port will be made to the coupled model that holds such macro knowledge and can therefore choose a valid position with which the requesting nt will try to pair next. for choosing this position at macro level two different model variants exist. the first and more simple one picks a position totally random from within the set of possible partners, whereas the second variant takes the entropy change into account when a pairing partner will be selected. the latter method prefers helix elongation in contrast to introducing new interior loops by single base pair formations and small loops will be more favourable than large ones [ ] . correct rna folding with the first method is dependent on the base pair stabilities and the random folding nuclei which are the first appearing base pairs of helical regions. this last point is less important for rna folding with model variant because the chosen binding partners are more deterministic due to loop entropy consideration. a comparison of both approaches with respect to simulation results is given in section . once a nucleotide received an input message by the macro model containing the number of a potential pairing partner, it tries to form a base pair with this nt by adding a coupling and sending a request. for a successful pairing, the partners must be of complementary base type and they must be able to pair in principle, e.g. bases can be modified so that they can not bind to others. figure illustrates the whole state flow for base pair formation and disruption of the nucleotide model component. the role of the macro model and its interactions with the micro level (nucleotides) shows the schematic organisation of the whole rna folding model in figure . already mentioned in the previous section, high level information about the whole rna molecule is needed to take sterical restrictions for base pairing into account. therefore, the coupled model holds the overall rna secondary structure which will be updated every time the state of a nucleotide changes from unpaired to paired and vice versa. this will be triggered by nucleotide port adding and removal recognised by the macro level. the same functionality is used to signalise the macro level the wish to try pairing with another nucleotide. the macro model detects a port adding, calculates the sterically possible partner set, chooses a position from within the set, and after all sends this position to the just now added nucleotide input port ( figure , nt ) . the coupled macro model is further responsible for sequence initialisation on the micro level by adding and connecting nucleotide models, i.e. it generates the primary rna structure. another task of the macro model is to observe the current folding structure for special structural patterns. this could be for example a specific binding domain for a protein or small ligand. also transcription termination or pausing structures can be of interest for observation. if observed structures are present during folding simulation, the macro level model can signalise this information and thus trigger dynamics to other components by adding new ports to itself (representing docking sites) or send messages over existing ports. a composed example model which uses this capability can be found in section . for evaluating the model's validity we simulated the folding of different rna molecules with known structure and analysed the results. three different types of experiments were done: native structure-correlates the majority of formed structures with the native structure after sufficient long simulation time? structure distribution-is the equilibrium ratio between minimum free energy (mfe) and suboptimal structures as expected? structure refolding-are molecules able to refold from suboptimal to more stable structural conformations? unfortunately, only few time-resolved folding pathways are experimentally derived and most of them treat pseudoknots [ ] and higher order tertiary structure elements [ , , ] which can not be handled by our folding model and are therefore out of the question for a simulation study. hence, some comparisons with other in silico tools were also made, although we know that one has to be careful with comparing different models for validating a model as it is often unclear how valid the other models are. because the folding model is highly stochastic, every simulation experiment was executed multiple times. typically replications were made. structural analysis of the cis-acting replication element from hepatitis c virus revealed a stem hairpin loop conformation where the helix is interrupted by an internal or bulge loop region [ ] . figure shows simulation results of its structure formation. the three-dimensional base pair lifetime plots indicate correct folding of both helical regions and only few misfolded base pairs. only small differences can be seen between simulations with the two base pair formation variants described in section . . without taking entropy into account for pairing, a bit more noise of misfolded base pairs can be observed which is not surprising due to the absolutely random partner choice. another well known rna structure is the cloverleaf secondary structure of trnas [ , ] consisting of four helical stems: the amino acid arm, d arm, anticodon arm, and the t arm. some base modifications and unusual nucleotides exist in trna which stabilise its structure formation, e.g. dihydrouridine and pseudouridine. such special conditions are not considered by our folding model as well as tertiary interactions leading to the final l-shaped form. however, folding simulations result in significant cloverleaf secondary structure formation (figure ). although there is much misfolded noise, the four distinct helix peaks are the most stable structural elements introduced during simulation, especially the amino acid arm. no fundamental difference can be seen between base pair formation model and . a third native structure validation experiment treats the corona virus s m motif which is a relatively long hairpin structure with some intersecting internal and bulge loops [ , ] . simulation of the sars virus s m rna folding indicates only for one of the native helix regions a conspicuous occurrence ( figure ). the other helices closer to the hairpin loop show no significant stabilisation. competing misfolded structural elements can be observed equally frequent or even more often. base pair formation model provides a slightly better result than the first one, but it is unsatisfying too. a reason for the result can be the multiple internal and bulge loops, which destabilise the stem and thus allow locally more stable structure elements to form. in [ ] a quantitative analysis of different rna secondary structures by comparative imino proton nmr spectroscopy is described. the results indicate that a small -nt rna has two equally stable structures in thermodynamic equilibrium, one with short helices and the other with a single hairpin. folding simulations of the same rna strand show an equal ratio of the probed structures as well ( figure ) . however, both are representing just % of all present structures which was not detected in the nmr experiments. many base pairs were introduced during simulation which are competing with base pairs of the two stable structures and thus reduce their appearance. this can be easily seen in the d matrix of figure where some additional peaks show high misfolded base pair lifetimes. simulating the rna folding with kinfold [ ] results in a five times higher amount of the -helix conformation than the single hairpin, but their total sum is about % of all molecules and thus less misfolded structures can be observed. real-time nmr spectroscopy was used by wenter et al. to determine refolding rates of a short -nt rna molecule [ ] . the formation of its most stable helix was temporarily inhibited by a modified guanosine at position . after photolytic removal of this modification a structure refolding was observed. to map such forced structure formation to relatively unstable folds at the beginning of an experiment, most rna folding tools have the capability to initialise simulations with specified structures. we used, much closer to the original wetlab experiment, a different strategy, i.e. at first g was not capable to bind any other base. the time course after removing this prohibition during simulation is shown in figure . wenter et al. detected structure refolding by measuring the imino proton signal intensity of u and u , which show high signal intensity if they are paired with other bases. accordingly we observed the state of both uracils over simulation time as well. after removal of g unpaired locking, a logarithmic decrease of structures with paired u and uniform increase of folds fig. . refolding of an deliberately misfolded small rna molecule [ ] . wetlab measurements are drawn by a x. simulation parameters: time seconds, temperature . k, base pair formation model , replications. simulations with kinfold were made with the same temperature and replication number but over a time period of seconds. with paired u can be observed reaching a complete shift of the conformational equilibrium after seconds. a very similar refolding progression was experimentally measured (single spots in figure ), but with a strong deviating time scale of factor . this could be a remaining model parameter inaccuracy or due to special experimental conditions which are not covered by our model, e.g. unusual salt concentrations. however, our model allows a quite realistic refolding from a suboptimal to a more stable rna structure. identical in silico experiments with kinfold [ ] by contrast, do not show any significant refolding (figure , nonchanging curves). the same counts for seqfold [ ] . with both approaches the energy barrier seems to be too high to escape from the misfolded suboptimal structure. local optima can more easily be overcome in comparison to other traditional pure macro methods (see figure ). we assume that even "stable" base pairs might be subject to changes, and let the nucleotides "searching" for a stable structure at micro level. this proved beneficial and emphasised the role of the micro level. however, the simulation revealed the importance of macro constraints for the folding process, and the implications of a lack of those. macro constraints that have been considered are for example the relative positioning of the nucleotides, particularly within spatial structures like hairpin or internal loops. the interplay between macro and micro level allowed us to reproduce many of the expected structure elements, e.g. figures and , although macro constraints have been significantly relaxed. these simplifications lead to "wrongly" formed structures and maybe could have been prevented by integrating terminal base stacking for pairing stability as well as less abstract base pair closing rules as macro constraints. a comparison of the two implemented base pair formation methods indicate only few differences. without taking entropy into account the noise of unstable single base pairs and short helices increases, but not dramatically. the same stable structures are formed based on both rules. having a working folding model we are now able to combine it with other model components that are influenced by or are influencing the rna structure formation and come back to the motivation, the attenuation of tryptophan synthesis. at least two further models are needed to reflect transcription attenuation: the rna polymerase and the ribosome (figure ). rna molecules are products of the transcription process which is the fundamental step in gene expression. once the rna polymerase enzyme complex (rnap) has successfully bound to dna (transcription initiation), it transcribes the template sequence into an rna strand by sequentially adding nucleotides to the ' end and thus elongates the molecule. to reflect this synthesising process, in the rna model, new nucleotide models and their backbone connections are added dynamically during simulation. this process is triggered by the rnap model component which interacts with rna. this dynamic rna elongation allows the simulation of sequential folding, where early synthesised parts of the rna molecule can already fold whereas other parts still have to be added. please note that this is not a unique feature of the model presented here, as kinetic folding tools typically realise sequential folding by just adding a new nt after a certain time delay. however, a component design allows to combine the rnap with further models (e.g. the dna template), or to model it in more detail (e.g. diverse rnap subunits), and to exchange model components on demand. the pattern observation function of the rna folding model, which is realised at macro level, allows us to look for an intrinsic transcription termination structure [ ] during simulation. if such structure is formed, the folding model removes its elongation input port meaning the release of the rna from the polymerase enzyme. at this time point the elongation stops, but structure folding and interactions with other components proceed. the ribosome enzyme complex translates rna sequences into protein determining amino acid sequences (peptides). translation starts at the ribosome binding site of mrna molecules which is reflected by a pair of input and output ports of the rna models. the translation begins after connecting it with a ribosome model. the current ribosome position with respect to the rna sequence is known by the rna model. a triplet of three rna bases (codon) encodes for one amino acid. the ribosome requests for the next codon ' of its current rna location when peptide chain elongation has proceeded. this is the case when the correct amino acid of the last codon entered the enzyme. the speed of the translation process depends strongly on the availability of needed amino acids. if an amino acid type is not sufficiently available, the ribosome stalls at the corresponding codon and thus pauses translation. a ribosome is quite big and thus - nucleotides are covered by its shape [ ] . therefore, a region upstream and downstream of the ribosome location is not able to form base pairs. as the rna model knows the ribosome location, this is handled by the rna macro level model which sends corresponding events to its nucleotide micro model components. the same counts for the helicase activity of the ribosome [ ] . for sequence translation, the macro level model will disrupt a base paired structure element when it is reached by the enzyme. whether those additional models are realised as atomic models, or coupled models depends on the objective of the simulation study. referring to the operon model presented in [ ] , the rnap, the mrna, and the ribosome would replace the simplistic mrna model, to integrate the attenuation process into the model. we presented a component-based model of rna folding processes. unlike traditional approaches which focus on the results of the folding process, e.g. stable structures in thermodynamical equilibrium, our objective has been different. the idea was to develop an approach that allows to integrate the folding processes into larger models and to take the dynamics into account, that has shown to be crucial in many regulation processes. therefore, the formalism ml-devs was used. at macro level, certain constraints referring to space and individual locations were introduced, whereas at micro level, the nucleotides were responsible for a successful base pairing and for the stability of the structure. a model component for the nucleotides and one model component for the entire rna molecule have been defined. the simulation results have been compared to wetlab experiments. therefore, the model components can be parametrised for different rna sequences (base types) as well as environmental conditions (e.g temperature). the evaluation revealed an overall acceptable performance, and in addition, insights into the role of micro level dynamics and macro level constraints. the integration of the rna folding model into a model of transcription attenuation has been sketched. next steps will be to realise this integration and to execute simulation experiments to analyse the impact of this more detailed regulation model on the synthesis of tryptophan. translation initiation and the fate of bacterial mrnas the mechanism of intrinsic transcription termination transcription attenuation: once viewed as a novel regulatory strategy genetic control by a metabolite binding mrna multiple structural alignment and clustering of rna sequences secondary structure prediction for aligned rna sequences mfold web server for nucleic acid folding and hybridization prediction fast folding and comparison of rna secondary structures a dynamic programming algorithm for rna structure prediction including pseudoknots prediction and statistics of pseudoknots in rna structures using exactly clustered stochastic simulations description of rna folding by "simulated annealing rna folding at elementary step resolution beyond energy minimization: approaches to the kinetic folding of rna dynamic regulation of the tryptophan operon: a modeling study and comparison with experimental data a variable structure model -the tryptophan operon rna hairpin loop stability depends on closing base pair coaxial stacking of helixes enhances binding of oligoribonucleotides and improves predictions of rna folding rapid compaction during rna folding combining micro and macro-modeling in devs for computational biology theory of modeling and simulation one modelling formalism & simulator is not enough! a perspective for computational biology based on james ii single-molecule rna folding rna folding at millisecond intervals by synchrotron hydroxyl radical footprinting a single-molecule study of rna catalysis and folding a cis-acting replication element in the sequence encoding the ns b rna-dependent rna polymerase is required for hepatitis c virus rna replication transfer rna: molecular structure, sequence, and properties the crystal structure of trna a common rna motif in the ' end of the genomes of astroviruses, avian infectious bronchitis virus and an equine rhinovirus the structure of a rigorously conserved rna element within the sars virus genome bistable secondary structures of small rnas and their structural probing by comparative imino proton nmr spectroscopy kinetics of photoinduced rna refolding by real-time nmr spectroscopy mrna helicase activity of the ribosome acknowledgments. many thanks to adelinde m. uhrmacher for her helpful comments and advice on this work. i will also thank roland ewald and jan himmelspach for their instructions for using james ii. the research has been funded by the german research foundation (dfg). key: cord- - i ye authors: bielecki, andrzej; gierdziewicz, maciej title: simulation of neurotransmitter flow in three dimensional model of presynaptic bouton date: - - journal: computational science - iccs doi: . / - - - - _ sha: doc_id: cord_uid: i ye in this paper a geometrical model for simulation of the nerve impulses inside the presynaptic bouton is designed. the neurotransmitter flow is described by using partial differential equation with nonlinear term. the bouton is modeled as a distorted geosphere and the mitochondrion inside it as a highly modified cuboid. the quality of the mesh elements is examined. the changes of the amount of neurotransmitter during exocytosis are simulated. the activity of neurons [ ] , including those taking place in the presynaptic bouton of the neuron, may be described theoretically with differential equations [ , , ] or, more specifically, with partial differential equations (pde) [ , , ] . numerical methods have been commonly used to solve pde [ , , , ] , which may be a proof that scientists are constantly interested in this phenomenon and, on the other hand, that results are variable and depend strongly on the biological assumptions and on the way of modeling. to obtain the solution, the appropriate design of the mesh of the studied object (in this paper: of a presynaptic bouton of the neuron) is necessary, which may be a demanding task [ , ] . some experiments in the cited papers have already been performed in order to answer the question: how the neurotransmitter (nt) mediates the process of conducting nerve impulses, which is connected with the distribution of nt in the synapse and with its exocytosis. this paper is intended to expand that knowledge by using more complicated mathematical model in three dimensions together with a geometric model which is more complex than it was in some previous works [ , ] . there are also examples of the description of a realistic geometric model [ ] , but without performing simulations. however, the nt flow in the realistically shaped model of the presynaptic bouton affected by the presence of a mitochondrion inside it has not been studied yet. therefore, the objective of this paper was to examine the process of nt flow in a realistic model of the presynaptic bouton with a mitochondrion partly occupying its volume. in this section the model based on partial differential equations is presented briefly. such approach allows us to study the dynamics of the transport both in time and in spatial aspect. the model is nonlinear, diffusive-like -see the paper [ ] for details. there are a few assumptions about the model. the bouton location is a bounded domain Ω ⊂ r . the boundary (the membrane) of the bouton is denoted as ∂Ω. the total amount of neurotransmitter in the bouton is increasing when new vesicles are supplied inside the bouton and decreasing when their contents are released through the bouton membrane. the proper subset of the bouton, Ω , is the domain of vesicles supply. the release site, Γ d , is a proper subset of the membrane. the function f : Ω → r models the synthesis of the vesicles that contain neurotransmitter. the value of this function is f (x) = β > on Ω and f (x) = on Ω \ Ω . the neurotransmitter flow in the presynaptic bouton was modeled with the following partial differential equation: where (x, t) is the density of neurotransmitter at the point x and time t and a is the diffusion coefficient. the last term contains the function f (x), presented before, and the threshold nt density, above which synthesis dose not take place, denoted by¯ . for any x ∈ r the "x + " symbol means max( , x). the boundary where ∂/∂ν is a directional derivative in the outer normal direction ν. the function η(t) depends on the time t and takes the value of for t ∈ [t n , t n + τ ] (with the action potential arriving at t n and with the release time τ ), and η(t) = otherwise [ , ] . multiplying ( ) by a smooth function v : Ω → r, and integrating by parts gives (see [ ] for details): provided that is sufficiently smooth. piecewise linear c tetrahedral finite elements are used in the numerical scheme. the unknown function is approximated by h , computed with the formula in which the basis functions of the finite element space are denoted by v i . using let a(t) = a + η(t)a . let us also introduce the nonlinear operator b : r k → r k in the following way the eq. ( ) may be rewritten as the time step is defined as t k = kΔt where Δt is at least several times smaller than τ . if we assume that k h is approximated by h (t k ) we can use the crank-nicolson approximative scheme valid for k = , , . . .. in the finite element basis {v k } k k= , is approximated by h . the scheme must be solved for k h . the problem of the nonlinearity in b is dealt with by the iterative scheme: in each step of the iteration the linear system ( ) is solved. the iterative procedure is stopped for the value m for which the correction of the solution due to one step of the scheme is sufficiently small. note that the non-linear term is calculated at the previous iteration (m) and the linear terms at iterations (m + ). such scheme reduces to solving the linear system in each step of the iteration. it is an alternative to the newton method, and it has the advantage of not needing to cope with the non-differentiable nonlinearity introduced with the last term (taking the positive part) of the equation. from physiological point of view, the time step for the eq. ( ) should be less than the time scale of the modeled phenomenon. we do not have the proof of the scheme stability but the numerical experiments show that the iterations of the fixed point scheme with respect to index m converge, and the limit is the solution of the crank-nicolson scheme. this scheme is known to be unconditionally stable. still, to avoid numerical artifacts, the time-step needs to be sufficiently small. our numerical experiments show that no numerical artifacts are seen in the obtained solution for Δt = . µs and for Δt = . µs. for the numerical simulations, the bouton is modeled as a bounded polyhedral domain Ω ⊂ r . the boundary of the bouton, ∂Ω, is represented by a set of flat polygons. the proper polyhedral subset of the bouton, Ω , is the domain of vesicles supply. the release site Γ d is modeled as a finite sum of flat polygons. various geometrical models of the presynaptic bouton have been used by the authors to simulate the process of conducting nerve impulses so far. the natural reference point for assessing their quality is to compare them to the images of real boutons which have been thoroughly examined, for example, several years ago [ , ] . one of the models was based of the skeleton made up of two concentric spheres, the outer one having the diameter of - µm and the inner one free from synaptic vesicles, and also with numerous release sites [ ] . the other structure consisting of two concentric globes and a single release site was filled with a tetrahedral mesh and it has been used to examine the distribution of neurotransmitter inside the bouton [ ] . a very similar structure has been utilized to find the connection between the number and location of synthesis domains and the amount of neurotransmitter in different locations in the bouton [ ] . another, more complicated, surface model intended for the discrete simulation of vesicle movement consisted of a realistically shaped bouton and a mitochondrion partly occupying its space [ ] . the amplitudes of evoked excitatory junctional potentials and the estimated number of the synaptic vesicles released during exocytosis were examined in [ ] with the use of the model with empty central region. the study discussed, among others, the way the synaptic depression was reflected in the results of the simulations. the model made up from two concentric spheres filled with a tetrahedral mesh was used to demonstrate synaptic depression [ ] . the volume of the model was about , µm and the surface about , µm . the diffusion coefficient was µm /s, synthesis rate and the exocytosis rate around µm /s. the results confirmed that during . s of hz stimulation, with the assumed values of parameters, the amount of neurotransmitter in the presynaptic bouton decreased, though only by the fraction of about %. the next studies with the globe model [ ] were performed to examine the influence of the location and number of synthesis zones on the amount of neurotransmitter in particular locations of the bouton. the simulation parameters were similar as in [ ] but the number of synthesis zones was or (however, with the same total volume) and the location was also variable. the chief conclusion was that the closer the synthesis (or supply) domain to the release site, the faster the bouton becomes depleted of neurotransmitter. the geometric aspects of the models of the bouton used so far, mentioned in the previous section, were extremely simplified. in this paper, the simulations have been conducted on the basis of the realistic model of the bouton in which the mitochondrion is modeled as well. the structure used to model the presynaptic bouton in this paper was based on the real shape of such a bouton and it was composed of geometric shapes modified so that it resembled the original. therefore, the outer of the bouton was modeled as a moderately modified sphere, whereas a part of the mitochondrion inside it was a highly distorted cuboid, as it has been presented in [ ] . the parameters of the surface mesh are described therein in detail. it should be stressed that the mesh described in this paper was designed to shorten the computing time of simulations of neurotransmitter flow, and therefore contains, among others, relatively large elements. the number of tetrahedra in the mesh was less then and the number of the surface triangles (faces) was less than × . the parameters of the input surface mesh were as follows. the total surface of the bouton was s ≈ . µm , and the area of the active zone (of the release site) was s az ≈ . µm . the surface mesh is presented in fig. . the tetrahedral mesh was generated with the tetgen program [ , ] . the result is depicted in fig. . the input parameters for tetgen were chosen in order to minimize the number of mesh elements. as a result of this, the bouton three-dimensional mesh contained tetrahedra with the total volume v ≈ . µm . the volume of the part of the bouton located in the direct neighborhood of the mitochondrion (where the endings of the microtubules are sometimes found), assumed as the theoretical nt supply zone ("synthesis domain") was v s ≈ . µm . the quality of the mesh was assessed by computing several measures for each tetrahedra and by taking, for each measure, its maximal value i.e. the value for the worst mesh element. the values are collected in table . the relatively high values of the mesh quality parameters may suggest a cautious approach to the analysis of the results. however, further stages of the experiment revealed that the mesh proved sufficient for performing the planned simulation. the initial density of the neurotransmitter was calculated by using the for- where a = [vesicles/µm ] is the theoretical maximal value of the function, b = . [ /µm ], and r[µm] is the distance from the "center" of the bouton i.e. from the center of the minimal cuboid containing the model. the simulation time was . s and during that time there were impulses. the results are depicted in the following figures: fig. and fig. , and in table . the program was designed primarily for validating the accuracy of the calculations; therefore the algorithm was implemented as a single threaded code in python, with numeric and graphic modules included, and the calculations were relatively slow. the approximate speed was - iterations per hour. therefore one simulation run lasted about - days. the total amount of neurotransmitter in the presynaptic bouton calculated during simulation was almost constant. the relative decrease of its value did not exceed %. those values refer to the situation when the neuron is not very intensively stimulated, and synaptic depression is not likely to occur. from the analysis of fig. it may be concluded that the activity of the release site is very moderate. however, closer examination of fig. reveals that the spatial distribution of neurotransmitter does change during stimulation. in the region directly adjacent to the release site the amount of neurotransmitter drops a little, and also in its vicinity a slight decrease may be noticed, which is visible in the last picture in fig. . the amount of neurotransmitter increased between stimuli, though at a very low speed, which may be attributed to the low value of exocytosis rate. the changes in these parameters can result in neurotransmitter amount sinking rapidly during exocytosis, thus leading to synaptic depression. to verify the reliability of the results, two control simulations were run. the first one, with two times larger mesh, confirmed that the relative differences in the results did not exceed . %. for the second one, with two times larger mesh (as before) and with halved time step, the results were almost the same; the relative difference between two control runs did not exceed − . the process described throughout this paper is similar to which has been found before [ ] where, in fig. , one can notice the almost constant number of released vesicles in every time step at the end of the simulation. taking into account the fact that the authors of the cited paper assumed that the total number of synaptic vesicles in the bouton was more than , and that the number of synaptic vesicles released in each time step was about , we may conclude that if with m ≈ vesicles in the bouton (the value assumed in this paper) the amount of neurotransmitter released during one time step is approximately . , our results are similar to those found in literature; the proportions of the released amount to the total vesicle pool are of the same order of magnitude. in another study [ ] the equilibrium state has been achieved, though at a lower level, indicating synaptic depression. the model of realistically shaped presynaptic bouton with a mitochondrion partly blocking its volume, presented in this paper, proved its ability to be used in simulation of synaptic depression. it should be stressed that the values of the parameters chosen for the initial tests of the proposed structure refer to the situation of a very regular activity, not threatened by depression. however, in silico tests may reveal the changes in distribution of neurotransmitter in a presynaptic bouton during a very frequent stimulation, thus allowing us to study depression in detail, the more so because the results of the experiments found in literature, whether or not a strong synaptic depression was detected, confirm that our simulation results reflect the real processes taking place in the presynaptic bouton of a stimulated neuron. simulation and parameter estimation of dynamics of synaptic depression temperature dependence of vesicular dynamics at excitatory synapses of rat hippocampus model of neurotransmitter fast transport in axon terminal of presynaptic neuron dynamical properties of the reaction-diffusion type model of fast synaptic transport three-dimensional model of signal processing in the presynaptic bouton of the neuron three-dimensional simulation of synaptic depression in axon terminal of stimulated neuron construction of a d geometric model of a presynaptic bouton for use in modeling of neurotransmitter flow a study on efficiency of d partial differential diffusive model of presynaptic processes numerical simulation for a neurotransmitter transport model in the axon terminal of a presynaptic neuron compartment model of neuropeptide synaptic transport with impulse control a model of intracellular transport of particles in an axon mixed-element octree: a meshing technique toward fast and real-time simulations in biomedical applications synaptic bouton properties are tuned to best fit the prevailing firing pattern tetgen: a quality tetrahedral mesh generator and d delaunay triangulator, version . user manual. wias -weierstrass institute for applied analysis and stochastics (wias tetgen, a delaunay-based quality tetrahedral mesh generator new trends in neurocybernetics stoichiometric biology of the synapse. dissertation composition of isolated synaptic boutons reveals the amounts of vesicle trafficking proteins high-fidelity geometric modeling for biomedical applications new software developments for quality mesh generation and optimization from biomedical imaging data key: cord- -hinho mh authors: zak, matthew; krzyżak, adam title: classification of lung diseases using deep learning models date: - - journal: computational science - iccs doi: . / - - - - _ sha: doc_id: cord_uid: hinho mh in this paper we address the problem of medical data scarcity by considering the task of detection of pulmonary diseases from chest x-ray images using small volume datasets with less than thousand samples. we implemented three deep convolutional neural networks (vgg , resnet- , and inceptionv ) pre-trained on the imagenet dataset and assesed them in lung disease classification tasks using transfer learning approach. we created a pipeline that segmented chest x-ray (cxr) images prior to classifying them and we compared the performance of our framework with the existing ones. we demonstrated that pre-trained models and simple classifiers such as shallow neural networks can compete with the complex systems. we also validated our framework on the publicly available shenzhen and montgomery lung datasets and compared its performance to the currently available solutions. our method was able to reach the same level of accuracy as the best performing models trained on the montgomery dataset however, the advantage of our approach is in smaller number of trainable parameters. furthermore, our inceptionv based model almost tied with the best performing solution on the shenzhen dataset despite being computationally less expensive. the availability of computationally powerful machines allowed emerging methods like pixel/voxel-based machine learning (pml) breakthroughs in medical image analysis/processing. instead of calculating features from segmented regions, this technique uses voxel/pixel values in input images directly. therefore, neither segmentation nor feature extraction is required. the performance supported by the natural sciences and engineering research council of canada. part of this research was carried out by the second author during his visit of the westpomeranian university of technology while on sabbatical leave from concordia university. of pml's can possibly exceed that of common classifiers [ ] as this method is able to avoid errors caused by inaccurate segmentation and feature extraction. the most popular powerful approaches include convolutional neural networks (including shift-invariant neural networks). they resulted in false positive (fp) rates reduction in computer-aided design framework (cad) for detection of masses and microcalcifications [ ] in mammography and in lung nodule detection in chest x-ray cxr images [ ] , neural filters and massive-training artificial neural networks including massive-training artificial neural networks (mtanns) including a mixture of expert mtanns, laplacian eigenfunction lap-mtann and massive-training support vector regression (mtsvr) for classification, object detection and image enhancement in malignant lung modules detection in ct, fp reduction in cad for polyp detection in ct colonography, bone separation from soft tissue in cxr and enhancement of lung nodules in ct [ ] . chest x-ray is one of the most frequently used diagnostic modality in detecting different lung diseases such as pneumonia or tuberculosis. roughly million of adults require hospitalization because of pneumonia, and about , dies from this disease annually in the us only. examination of lung nodules in cxr can lead to missing of diseases like lung cancer. however, not all of them are visible in retrospect. studies show that - % of lung cancer cases were missed due to occlusions (at least partial) by ribs or clavicle. to address this problem researchers examined dual-energy imaging, a technique which can produce images of two tissues, namely "soft-tissue" image and "bone" image. this technique has many drawbacks, but undoubtedly one of the most important ones is the exposure to radiation. the mtanns models have been developed to address this problem and serve as a technique for ribs/soft-tissue separation. the idea behind training of those algorithms is to provide them with bone and soft-tissue images obtained from a dual-energy radiography system. the mtann was trained using cxrs as input and corresponding boneless images. the ribs contrast is visibly suppressed in the resulting image, maintaining the soft tissue areas such as lung vessels. recent developments in deep neural networks [ ] lead to major improvements in medical imaging. the efficiency of dimensionality reduction algorithms like lung segmentation was demonstrated in the chest x-ray image analysis. recently researchers aimed at improving tuberculosis detection on relatively small data sets of less than images per class by incorporating deep learning segmentation and classification methods from [ ] . we will further explore these techniques in this paper. in this paper we combine two relatively small datasets containing less than images per class for classification (pneumonia and tuberculosis detection) and segmentation purposes. we selected examples per "disease" class ( images with tuberculosis and images with pneumonia) and of healthy patients yielding the set of samples from different patients. sample images from both datasets are shown in fig. . the shenzhen hospital dataset (sh) [ , ] containing cxr images was created by the people's hospital in shenzhen, china. it includes both abnormal (containing traces of tuberculosis) and standard cxr images. unfortunately, the dataset is not well-balanced in terms of absence or presence of disease, gender, or age. we extracted only samples of healthy patients ( from both datasets) and of those labeled with traces of tuberculosis. selecting information about one class from different resources ensures that the model is not contaminated by the features resulting from the method of taking images, e.g., the lens. pneumonia is an inflammatory condition of the lung affecting the little air sacs known as alveoli. standard symptoms comprise of a blend of a dry hacking cough, inconvenience breathing, chest agony, and fever. the labeled optical tomography and chest x-ray images for classification dataset [ ] includes selected images of pneumonia patients from the medical center in guangzhou. it consists of data with two classes -normal and those containing marks of pneumonia. all data come from the patient's routine clinical care. the volume of the complete dataset includes thousands of validated optical coherence tomography (oct) and x-ray images yet for our analysis we wanted to keep the dataset tiny and evenly distributed thus only images were selected (other images come from the tuberculosis dataset) from the resources labeled as healthy and as pneumonia -both chosen randomly. external segmentation of left and right lung images (exclusion of redundant information: bones, internal organs, etc.) was proven to be effective in boosting prediction accuracy. to extract lungs information and exclude outside regions, we used the manually prepared masks included in the extension of the sh dataset, namely, the segmented sh dataset, see fig. . due to nonidentical borders and lung shapes, the segmentation data has high variability although its distribution is quite similar to the regular one when compared to image area distribution. model-based methods greatly improve their predictions when the number of training samples grows. when a limited amount of data is available, some transformations have to be applied to the existing dataset to synthetically augment the training set. researchers in [ ] employed three techniques to augment the training dataset. the first approach was to randomly crop of a × pixel fixed-size window from a × pixel image. the second technique was flipping the image horizontally, which allowed capturing information about reflection invariance. finally, the third method added randomly generated lighting to capture color and illumination variation. transfer learning is a very popular approach in computer vision related tasks using deep neural networks when data resources are scarce. therefore, to launch a new task, we incorporate the pre-trained models skilled in solving similar problems. this method is crucial in medical image processing due to the shortage of real samples. in deep neural networks, feature extraction is carried out but passing raw data through models specialized in other tasks. here, we can refer to deep learning models such as resnet, where the last layer information serves as input to a new classifier. transfer learning in deep learning problems can be performed using a common approach called pre-trained models approach. reuse model states that pre-trained model can produce a starting point for another model used in a different task. this involves incorporation of the whole model or its parts. the adopted model may or may not need to be refined on the input-output data for the new task. the third option considers selecting one of the available models. it is very common that research institutions publish their algorithms trained on challenging datasets which may fully or partially cover the problem stated by a new task. imagenet [ ] is a project that helps computer vision researches in classification and detection tasks by providing them with a large image dataset. this database contains roughly million different images from over thousand classes. imagenet also provides bounding boxes with annotations for over million images, which are used in object localization problems. in this work, we will experiment with three deep models (vgg , resnet- , and inceptionv ) pre-trained on the imagenet dataset. the following deep nets have been considered: vgg , resnet- and incep-tionv . the vgg convolutional network is a model with layers trained on fixed size images. the input is processed through a set of convolution layers which use small-size kernels with a receptive field × . this is the smallest size allowing us to capture the notion of up, down, right, left, and center. the architecture also incorporates × kernels which may be interpreted as linear input transformation (followed by nonlinearity). the stride of convolutions (number of pixels that are shifted in every convolution -step size) is fixed and set to pixel; therefore the spatial resolution remains the same after processing an input through a layer, e.g., the padding is fixed to for × kernels. spatial downsizing is performed by five consecutive pooling (max-pooling) layers, which are followed by some convolution layers. however, not all of them are followed by max-pooling. the max-pooling operation is carried over a fixed × pixel window, with a stride of pixels. this cascade of convolutional layers ends with three fully-connected (fc) layers where the first two consist of nodes each and the third one of as it performs the -way classification using softmax. all hidden layers have the same non-linearity relu (rectified linear unit) [ ] . the resnet convolutional neural network is a -layer deep model trained on more than a million fixed-size images from the imagenet dataset. the network classifies an input image into one of object classes like car, airplane, horse or mouse. the network has learned a large amount of features thanks to training images diversity and achieved . % top- error rate on the imagenet dataset. the resnet- convolutional neural network consists of stages, each having convolutions and identity blocks. every convolution block consists of convolutional layers. resnet- is related to resnet- , however, the idea behind its sibling model remains the same. the only difference is in residual blocks; unlike those in resnet- resnet- replaces every two layers in a residual block with a three-layer bottleneck block and × convolutions, which reduce and eventually restore the channel depth. this allows reducing a computational load when a × convolution is calculated. the model input is first processed through a layer with filters each × and stride and downsized by a max-pooling operation, which is carried over a fixed × pixel window, with a stride of pixels. the second stage consists of three identical blocks, each containing a double convolution with × pixels filters and a skip connection block. the third pile of convolutions starts with a dotted line (image not included) as there is a change in the dimensionality of an input. this effect is achieved through the change of stride in the first convolution bloc from to pixels. the fourth and fifth groups of convolutions and skip connections follow the pattern presented in the third stage of input processing, yet they change the number of filters (kernels) to and , respectively. this model has over million parameters. the researchers from google introduced the first inception (inceptionv ) neural network in during the imagenet competition. the model consisted of blocs called "inception cell" that was able to conduct convolutions using different scale filters and afterward aggregate the results as one. thanks to × convolution which reduces the input channel depth the model saves computations. using a set of × , × , and finally, × size of filters, an inception unit cell learns extracting features of different scale from the input image. although inception cells use max-pooling operator, the dimension of a processed data is preserved due to "same" padding, and so the output is properly concatenated. a follow-up paper [ ] was released not long after introducing a more efficient inceptionv solution to the first version of the inception cell. large filters sized × , and × are useful in extensive spatial features extraction, yet their disadvantage lies in the number of parameters and therefore computational disproportion. the inceptionv model contains over million parameters. the architecture can be divided into modules. the first processing block consists of inception modules. then, information is passed through the effective grid size reduction and processed through four consecutive inception cells with asymmetric convolutions. moving forward, information flows to the × pixels convolution layer connected to an auxiliary classifier and another effective grid size-reduction block. finally, data progresses through a series of two blocs with wider filter banks and consequently gets to a fully-connected layer ended with a softmax classifier. visualization of the network architecture can be found in fig. . many vision-related tasks, especially those from the field of medical image processing expect to have a class assigned to every pixel, i.e., every pixel is associated with a corresponding class. to conduct this process, we propose so-called u-net neural network architecture described in [ ] and in sect. . . this model works well with very few training image examples yielding precise segmentation. the motivation behind this network is to utilize progressive layers instead of a building system, where upsampling layers are utilized instead of pooling operators, consequently increasing the output resolution. high-resolution features are combined with the upsampled output to perform localization. the deconvolution layers consist of a high number of kernels, which better propagate information and result in outputs with higher resolution. owing to the described procedures, the deconvolution path is approximately symmetric to the contracting one and so the architecture resembles the u shape. there are no fully connected layers, therefore, making it possible to conduct the seamless segmentation of relatively large images extrapolating the missing context by mirroring the processed input. the network showed in fig. consists of an expansive path (right) and a contracting one(left). the first part (contracting) resembles a typical convolutional neural network; the repeated × convolutions followed by a non-linearity (here relu), and × poling with stride . each downsampling operation doubles the number of resulting feature maps. all expansive path operations are made of upsampling of the feature channels followed by a × deconvolution (or "up-convolution") which reduces the number of feature maps twice. the result is then concatenated with the corresponding feature layer from the contracting path and convolved with × kernels, and each passed through a relu. the final layers apply a × convolution to map each feature vector to the desired class. following the approaches presented in the literature we wanted to use deep convolutional neural networks to segment lungs [ ] before processing it through the classification models mentioned in sect. . . researchers in [ ] indicate that u-net architecture and its modifications outperform the majority of cnn-based models and achieve excellent results by easily capturing spacial information about the lungs. thus, we propose a pipeline that consists of two stages: first segmentation and then classification. the phase of extracting valuable information (lungs) is conducted with a model presented in sect. . our algorithms trained for epochs on an extension of the sh dataset. the input to our u-shaped deep neural network is a regular chest x-ray image, whereas the output is a manually prepared binary mask of lung shape, matching the input. the code for the transfer-learning models is publicly available through a python api, keras. our algorithms were trained on servers equipped with gpu provided by helios calcul québec, which consists of fifteen computing nodes each having eight nvidia k gpus and additionally six computing nodes with eight nvidia k boards each. every k board includes two gpu's and so the total of gpu's in the cluster. as mentioned before, our model was trained for epochs using a dataset partitioned into %, %, and % bins, for training, validation and test parts, respectively using the models introduced in sect. . using the batch size of samples, augmentation techniques briefed in sect. . , adam optimizer and categorical cross-entropy as a loss function for pixel-wise binary classification. the training results are shown in fig. . as we can easily notice, the validation error is slowly falling throughout the whole training, whereas there is no major change after the th epoch. the final error on the validation set is right below . and slightly above . on the test set. our algorithm learns shape-related features typical for lungs and can generalize well further over unseen data. figure shows the results of our u-net trained models. it is clear that the network was able to learn chest shape features and exclude regions containing internal organs such as heart. these promising results allowed us to process the whole dataset presented earlier and continue our analysis on the newly processed images. we propose a two-stage pipeline for classifying lung pathologies into pneumonia and tuberculosis consisting of two stages: first for chest x-ray image segmentation and second for lung disease classification. the first stage (segmentation) is trained during experiments described in the previous section. the second stage utilizes deep models described in sect. . , whereas we investigate potential improvements in performance depending on the type of model used. our classification models were trained using the same setup as described in sect. . . here, we conduct our experiments using the data described in sect. . . the difference is in prior segmentation, which extracts valuable information for the task, namely lungs. figure shows the training samples; the left and right panels correspond to input and output, respectively. we tried all models with three deep net classifiers (vgg , resnet- , incep-tionv ) in the task of classification of lung images into two classes: pneumonia and tuberculosis. we observed that inceptionv based model performed best and thus due to lack of space we display only its performance results. the confusion matrix in fig. (a) shows that the new model improved the number of true positives (tp) in all classes in comparison with the vgg and resnet- based models. image fig. (b) shows that the auc score for healthy, tuberculosis and pneumonia cases were %, %, and %, respectively. after comparing the results obtained by models without transfer learning we observe that transfer-learning models perform well in lung diseases classification using segmented images tasks even when the data resources are scarce. in this section, we compare the performance of our models to the results achieved in the literature over different datasets (fig. ). the algorithm that scored the best in the majority results was inceptionv trained on the segmented images. what is more, it produced very high scores for the "disease" classes showing that a random instance containing marks of tuberculosis or pneumonia has over % probability to be classified to the correct class. although the scores of the healthy class are worse than the diseased ones, its real cost is indeed lower as it is always worse to classify a sick patient as healthy. the inceptionv based model scored best, reaching better accuracy than vgg algorithms by over %. although the interpretability of our methods is not guaranteed, we can clearly state that using transfer-learning based algorithms on small datasets allows achieving competitive classification scores on the unseen data. furthermore, we compared the class activation maps shown in fig. in order to investigate the reasoning behind decision making. the remaining features, here lungs, force the network to explore it and thus make decisions based on observed changes. that behavior was expected and additionally improved the interpretability of our models as the marked regions might bring attention of the doctor in case of sick patients. in this section, we compare performance of our models with the results in the literature using over different datasets. in order to do so we trained our algorithms on the shenzhen and montgomery datasets [ ] ten times, generated the results for all the models and averaged their scores: accuracy, precision, sensitivity, specificity, f score and auc. table presents comparison of different deep learning models trained on the shenzhen dataset [ ] . although our approach does not guarantee the best performance, it is always close to the highest even though it is typically less complex. researchers in [ ] used various pre-trained models in the pulmonary disease detection task, and the ensemble of them yields the highest accuracy and sensitivity. to compare, our inceptionv -based model achieves accuracy smaller by only one percent and has identical auc, which means that our method gives an equal probability of assigning a positive case of tuberculosis to its corresponding class over a negative sample. although we could not outperform the best methods, our framework is less complicated. furthermore, in table we compared the performance of our framework trained on the montgomery dataset [ ] to the literature. our inceptionv -based model tied with [ ] in terms of accuracy, yet showed higher value of auc. resnet- and vgg based models performed worse, however not by much as they reached accuracies of % and % respectively, which is roughly and % less than the highest score achieved. table . comparison of different deep learning based solutions trained on the shenzhen dataset [ ] . although our result is not the best, it performs better than any single model (excluding ensemble). horizontal line means that corresponding results were not provided in literature. accuracy precision sensitivity specificity f score auc [ ] . our average performance is almost identical to [ ] . we created lung diseases classification pipeline based on transfer learning that was applied to small datasets of lung images. we evaluated its performance in classification of non-segmented and segmented chest x-ray images. in our best performing framework we used u-net segmentation network and inceptionv deep model classifier. our frameworks were compared with the existing models. we demonstrated that models pre-trained by transfer learning approach and simple classifiers such as shallow neural networks can successfully compete with the complex systems. tb detection in chest radiograph using deep learning architecture lung segmentation in chest radiographs using anatomical atlases with nonrigid registration imagenet: a largescale hierarchical image database deep learning with lung segmentation and bone shadow exclusion techniques for chest x-ray analysis of lung cancer abnormality detection and localization in chest x-rays using deep convolutional neural networks. arxiv two public chest x-ray datasets for computer-aided screening of pulmonary diseases automatic tuberculosis screening using chest radiographs nuclei segmentation in histopathological images using two-stage learning large dataset of labeled optical coherence tomography (oct) and chest x-ray images imagenet classification with deep convolutional neural networks computer-aided detection of peripheral lung cancers missed at ct: roc analyses without and with localization a multiple circular path convolution neural network system for detection of mammographic masses artificial convolution neural network for medical image pattern recognition efficient deep network architectures for fast chest x-ray tuberculosis screening and visualization a novel approach for tuberculosis screening based on deep convolutional neural networks. in: medical imaging : computer-aided diagnosis pixel-based machine learning in medical imaging rethinking the inception architecture for computer vision u-net: convolutional networks for biomedical image segmentation key: cord- - og pivv authors: lepenioti, katerina; pertselakis, minas; bousdekis, alexandros; louca, andreas; lampathaki, fenareti; apostolou, dimitris; mentzas, gregoris; anastasiou, stathis title: machine learning for predictive and prescriptive analytics of operational data in smart manufacturing date: - - journal: advanced information systems engineering workshops doi: . / - - - - _ sha: doc_id: cord_uid: og pivv perceiving information and extracting insights from data is one of the major challenges in smart manufacturing. real-time data analytics face several challenges in real-life scenarios, while there is a huge treasure of legacy, enterprise and operational data remaining untouched. the current paper exploits the recent advancements of (deep) machine learning for performing predictive and prescriptive analytics on the basis of enterprise and operational data aiming at supporting the operator on the shopfloor. to do this, it implements algorithms, such as recurrent neural networks for predictive analytics, and multi-objective reinforcement learning for prescriptive analytics. the proposed approach is demonstrated in a predictive maintenance scenario in steel industry. perceiving information and extracting business insights and knowledge from data is one of the major challenges in smart manufacturing [ ] . in this sense, advanced data analytics is a crucial enabler of industry . [ ] . more specifically, among the major challenges for smart manufacturing are: (deep) machine learning, prescriptive analytics in industrial plants, and analytics-based decision support in manufacturing operations [ ] . the wide adoption of iot devices, sensors and actuators in manufacturing environments has fostered an increasing research interest on real-time data analytics. however, these approaches face several challenges in real-life scenarios: (i) they require a large amount of sensor data that already have experienced events (e.g. failures of -ideally-all possible causes); (ii) they require an enormous computational capacity that cannot be supported by existing computational infrastructure of factories; (iii) in most cases, the sensor data involve only a few components of a production line, or a small number of parameters related to each component (e.g. temperature, pressure, vibration), making impossible to capture the whole picture of the factory shop floor and the possible correlations among all the machines; (iv) the cold-start problem is rarely investigated. on the other hand, there is a huge treasure of legacy, enterprise and operational systems data remaining untouched. manufacturers are sitting on a goldmine of unexplored historical, legacy and operational data from their manufacturing execution systems (mes), enterprise resource planning systems (erp), etc. and they cannot afford to miss out on its unexplored potential. however, only - % of the value from such available data-at-rest is currently accrued [ ] . legacy data contain information regarding the whole factory cycle and store events from all machines, whether they have sensors installed or not (e.g. products per day, interruption times of production line, maintenance logs, causalities, etc.) [ ] . therefore, legacy data analytics have the credentials to move beyond kpis calculations of business reports (e.g. oee, uptime, etc.), towards providing an all-around view of manufacturing operations on the shopfloor in a proactive manner. in this direction, the recent advancements of machine learning can have a substantial contribution in performing predictive and prescriptive analytics on the basis of enterprise and operational data aiming at supporting the operator on the shopfloor and at extracting meaningful insights. combining predictive and prescriptive analytics is essential for smarter decisions in manufacturing [ ] . in addition mobile computing (with the use of mobile devices, such as smartphones and tablets) can significantly enable timely, comfortable, non-intrusive and reliable interaction with the operator on the shopfloor [ ] , e.g. for generating alerts, guiding their work, etc. through dedicated mobile apps. the current paper proposes an approach for predictive and prescriptive analytics on the basis of enterprise and operational data for smart manufacturing. to do this, it develops algorithms based on recurrent neural networks (rnn) for predictive analytics, and multi-objective reinforcement learning (morl) for prescriptive analytics. the rest of the paper is organized as follows: sect. presents the background, the challenges and prominent methods for predictive and prescriptive analytics of enterprise and operational data for smart manufacturing. section describes the proposed approach, while sect. shows a walkthrough scenario of the proposed approach in the steel industry. section presents the experimental results, while sect. concludes the paper and outlines the plans for future research. background. intelligent and automated data analysis which aims to discover useful insights from data has become a best practice for modern factories. it is supported today by many software tools and data warehouses, and it is known by the name "descriptive analytics". a step further, however, is to use the same data to feed models that can make predictions with similar or better accuracy than a human expert. in the framework of smart manufacturing, prognostics related to machines' health status is a critical research domain that often leverages machine learning methods and data mining tools. in most of the cases, this is related to the analysis of streaming sensor data mainly for health monitoring [ ] [ ] [ ] , but also for failure prediction [ ] [ ] [ ] as part of a predictive maintenance strategy. however, in all of these approaches, the prediction is produced only minutes or even seconds before the actual failure, which, is not often a realistic and practical solution for a real industrial case. the factory managers need to have this information hours or days before the event, so that there is enough time for them to act proactively and prevent it. one way to achieve this is to perform data mining on maintenance and operational data that capture the daily life-cycle of the shop floor in order to make more high-level predictions [ ] [ ] [ ] . existing challenges. the most notable challenges related to predictive analytics for smart manufacturing include: (a) predictions always involve a degree of uncertainty, especially when the data available are not sufficient quantity-wise or quality-wise; (b) inconsistent, incomplete or missing data with low dimensionality often result into overfitting or underfitting that can lead to misleading conclusions; (c) properly preparing and manipulating the data in order to conclude to the most appropriate set of features to be used as input to the model is the most time-consuming, yet critical to the accuracy of the algorithms, activity; (d) lack of a common "language" between data scientists and domain experts hinders the extraction of appropriate hypothesis from the beginning and the correct interpretation and explainability of results. novel methods. time series forecasting involves prediction models that analyze time series data and usually infer future data trends. a time series is a sequence of data points indexed in time order. unlike regression predictive modeling, time series also adds the complexity of a sequence dependence among the input variables. recurrent neural networks (rnn) are considered to be powerful neural networks designed to handle sequence dependence. long short-term memory network (lstm) is a type of rnn that is typically used in deep learning for its ability to learn long-term dependencies and handle multiple input and output variables. background. prescriptive analytics aims at answering the questions "what should i do?" and "why should i do it?". it is able to bring business value through adaptive, time-dependent and optimal decisions on the basis of predictions about future events [ ] . during the last years, there is an increasing interest on prescriptive analytics for smart manufacturing [ ] , and is considered to be the next evolutionary step towards increasing data analytics maturity for optimized decision making, ahead of time. existing challenges. the most important challenges of prescriptive analytics include [ , , ] : (i) addressing the uncertainty introduced by the predictions, the incomplete and noisy data and the subjectivity in human judgement; (ii) combining the "learned knowledge" of machine learning and data mining methods with the "engineered knowledge" elicited from domain experts; (iii) developing generic prescriptive analytics methods and algorithms utilizing artificial intelligence and machine learning instead of problem-specific optimization models; (iv) incorporating adaptation mechanisms capable of processing data and human feedback to continuously improve decision making process over time and to generate non-intrusive prescriptions; (v) recommending optimal plans out of a list of alternative (sets of) actions. novel methods. reinforcement learning (rl) is considered to be a third machine learning paradigm, alongside supervised learning and unsupervised learning [ ] . rl shows an increasing trend in research literature as a tool for optimal policies in manufacturing problems (e.g. [ , ] ). in rl, the problem is represented by an environment consisting of states and actions and learning agents with a defined goal state. the agents aim to reach the goal state while maximizing the rewards by selecting actions and moving to different states. in interactive rl, there is the additional capability of incorporating evaluative feedback by a human observer so that the rl agent learns from both human feedback and environmental reward [ ] . another extension is multi-objective rl (morl), which is a sequential decision making problem with multiple objectives. morl requires a learning agent to obtain action policies that can optimize multiple objectives at the same time [ ] . the proposed approach consists of a predictive analytics component (sect. . ) and a prescriptive analytics component (sect. . ) that process enterprise and operational data from manufacturing legacy systems, as depicted in fig. . the communication is conducted through an event broker for the event predictions and the actions prescriptions, while other parameters (i.e. objective values and alternative actions) become available through restful apis. the results are communicated to business users and shopfloor operators through intuitive interfaces addressed to both computers and mobile devices. the proposed predictive analytics approach aims to: (i) exploit hidden correlations inside the data that derive from the day-to-day shop floor operations, (ii) create and adjust a predictive model able to identify future machinery failures, and (iii) make estimations regarding the timing of the failure, i.e. when a failure of the machinery may occur, given the history of operations on the factory. this type of data usually contains daily characteristics that derive from the production line operations and are typically collected as part of a world-wide best practice for monitoring, evaluation and improvement of the effectiveness of the production process. the basic measurement of this process is an industry standard known as overall equipment effectiveness (oee) and is computed as: oee(%) = availability(%)  performance(%)  quality (%). availability is the ratio of actual operational time versus the planned operational time, performance is the ratio of actual throughput of products versus the maximum potential throughput, and the quality is the ratio of the not-rejected items produced vs the total production. the oee factor can be computed for the whole production line as an indication of the factory's effectiveness or per machine or a group of machines. the proposed methodology takes advantage of these commonly extracted indicators and processes them in two steps: in predictive model building (learning) and predictive model deployment. predictive model building. the predictive analytics model incorporates lstm and exploits its unique ability to "remember" a sequence of patterns and its relative insensitivity to possible time gaps in the time series. as in most neural network algorithms, lstm networks are able to seamlessly model non-linear problems with multiple input variables through the iterative training of their parameters (weights). since the predictive analytics model deals with time-series, the lstm model is trained using supervised learning on a set of training sequences assigned to a known output value. therefore, an analyst feeds the model with a set of daily features for a given machine (e.g. the factors that produce the oee) and use as outcome the number of days until the next failure. this number is known since historical data can hold this information. nevertheless, when the model is finally built and put in operation, it will use new input data and will have to estimate the new outcome. predictive model deployment. when the lstm model is fed with new data it can produce an estimation of when the next failure will occur (i.e. number of days or hours) and what is the expected interruption duration in the following days. although this estimation may not be % accurate, it could help factory managers to program maintenance actions proactively in a flexible and dynamic manner, compared to an often rigid and outdated schedule that is currently the common practice. this estimation feeds into prescriptive analytics aiming at automating the whole decision-making process and provide optimal plans. the proposed prescriptive analytics approach is able to: (i) recommend (prescribe) both perfect and imperfect actions (e.g. maintenance actions with various degrees of restoration); (ii) model the decision making process under uncertainty instead of the physical manufacturing process, thus making it applicable to various industries and production processes; and, (iii) incorporate the preference of the domain expert into the decision making process (e.g. according to their skills, experience, etc.), in the form of feedback over the generated prescriptions. to do these, it incorporates multi-objective reinforcement learning (morl). unlike most of the multi-objective optimization approaches which result in the pareto front set of optimal solutions [ ] , the proposed approach provides a single optimal solution (prescription), thus generating more concrete insights to the user. the proposed prescriptive analytics algorithm consists of three steps: prescriptive model building, prescriptive model solving, and prescriptive model adapting, which are described in detail below. prescriptive model building. the prescriptive analytics model representing the decision making process is defined by a tuple s; a; t; r ð Þ ; where s is the state space, a is the action space, t is the transition function t : s  a  s ! r and r is the vector reward function r : s  a  s ! r n where the n-dimensions are associated with the objectives to be optimized o n . the proposed prescriptive analytics model has a single starting state s n , from which the agent starts the episode, and a state s b that the agent tries to avoid. each episode of the training process of the rl agent will end, when the agent returns to the normal state s n or when it reaches s b . figure depicts an example including alternative (perfect and/or imperfect maintenance) actions (or sets of actions) s a i ; i ¼ ; ; , each one of which is assigned to a reward vector. the prescriptive analytics model is built dynamically. in this sense, the latest updates on the number of the action states s a i and the estimations of the objectives' values for each state s k are retrieved through apis from the predictive analytics. each action may be implemented either before the breakdown (in order to eliminate or mitigate its impact) or after the breakdown (if this occurs before the implementation of mitigating actions). after the implementation of each action, the equipment returns to its normal state s n . solid lines represent the transitions a i that have non-zero reward with respect to the optimization objectives and move the agent from one state to another. prescriptive model deployment. on the basis of event triggers for predicted abnormal situations (e.g. about the time of the next breakdown) received through a message broker, the model moves from the normal state s n to the dangerous state s d . for each objective, the reward functions are defined according to whether the objective is to be maximized or minimized. on this basis, the optimal policy p o i s; a ð Þ; for each objective o i is calculated with the use of the actor-critic algorithm, which is a policy gradient algorithm aiming at searching directly in (some subset of) the policy space starting with a mapping from a finite-dimensional (parameter) space to the space of policies [ ] . assuming independent objectives, the multi-objective optimal policy is derived from: p opt s; a ð Þ ¼ q i i p o i s; a ð Þ. the time constraints of the optimal policy (prescription) are defined by the prediction event trigger. the prescription is exposed to the operator on the shop-floor (e.g. through a mobile device) providing them the capability to accept or reject it. if accepted, the prescribed action is added to the actions plan. prescriptive model adaptation. the prescriptive analytics model is able to adapt according to feedback by the expert over the generated prescriptions. this approach learns from the operator whether the prescribed actions converge with their experience or skills and incorporates their preference to the prescriptive analytics model. in this way, it provides non-disruptive decision augmentation and thus, achieves an optimized human-machine interaction, while, at the same time, optimizing manufacturing kpis. to do this, it implements the policy shaping algorithm [ ] , a bayesian approach that attempts to maximize the information gained from human feedback by utilizing it as direct labels on the policy. for each prescription, optional human feedback is received as a signal of approval or rejection, numerically mapped to the reward signals and interpreted into a step function. the feedback is converted into a policy p feedback s; a ð Þ, the distribution of which relies on the consistency, expressing the user's knowledge regarding the optimality of the actions, and the likelihood of receiving feedback. assuming that the feedback policy is independent from the optimal multi-objective policy, the synthetic optimal policy for the optimization objectives and the human feedback is calculated as: p opt s; a ð Þ ¼ p opt s; a ð ÞÃp feedback s; a ð Þ. the case examined is the cold rolling production line of m. j. maillis s.a. cold rolling is a process of reduction of the cross-sectional area through the deformation caused by a pair of rotating in opposite directions metal rolls in order to produce rolling products with the closest possible thickness tolerances and an excellent surface finish. in the milling station, there is one pair of back up rolls and one pair of work rolls. the deformation takes place through force of the rolls supported by adjustable strip tension in both coilers and de-coilers. over the life of the roll some wear will occur due to normal processing, and some wear will occur due to extraneous conditions. during replacement, the rolls are removed for grinding, during which some roll diameter is lost, and then are stored in the warehouse for future use. after several regrinding, the diameter of the roll becomes so small that is no longer operational. the lstm model of predictive analytics was created using the keras library with tensorflow as backend and the morl using brown-umbc reinforcement learning and planning (burlap) library, while the event communication between them is performed with a kafka broker. in the m. j. maillis s.a case, the system predicts the time of the next breakdown and the rul of the available rolls. for the latter, the operator can select one of the repaired rollers, having been subject to grinding, or a new one. therefore, the alternative actions are created dynamically according to the available repaired rollers existing in the warehouse. each one has a different rul, according to its previous operation, and a different cost (retrieved from enterprise systems) due to its depreciation. each roller has an id and is assigned to its characteristics/objectives of morl (i.e. cost to be minimized and rul to be maximized) in order to facilitate its traceability. the available rolls along with the aforementioned objectives values are retrieved on the basis of a predicted breakdown event trigger. the alternative actions for the current scenario along with their costs and ruls are shown in table . the action "replace with new roller" represents a perfect maintenance action, while the rest ones represent imperfect maintenance actions. figure depicts an example of the process in which the prescription "replace with repaired roller id " is generated on the basis of a breakdown prediction and previously received feedback and instantly communicated to the operators through a dedicated mobile app. the operators are also expected to provide feedback so that their knowledge and preferences are incorporated in the system and the models are adapted accordingly. the second analysis aimed to predict the expected interruption duration for the following day ('which is the expected interruption duration for the following day?'). the input features used in this lstm model were: availability, performance, minutes of breakdown, real gross production, number of breakdowns, and month (date). again, several lstm parameters and layers were tested and the final model resulted to be a sequential model with a first lstm layer of neurons and an activation function 'relu', a second layer of neurons with a 'relu' activation function, a dropout layer of . rate, and finally a dense layer. the model was trained using data from and ; using a batch size of , epochs, a timestep of and an rmsprop optimizer. predictions were performed in data and results are depicted in fig. . the blue line represents the actual value whereas the orange line represents the predicted value. the overall rmse is . , meaning that there is an average of . min uncertainty in each prediction. for this experiment, the actor-critic algorithm, which calculates the associated optimal policy sequentially within episodes, consists of a boltzmann actor and a td-lambda critic with learning rate = . , lambda = . and gamma = . . the generated policies are then integrated into a single policy taking into account the consistency (c = . ) and likelihood (l = . ) values. table presents five "snapshots" of positive and negative feedback along with the resulting shaped prescriptions and their respective policies. each "snapshot" is compared to the previous one. in this paper, we proposed an approach for predictive and prescriptive analytics aiming at exploiting the huge treasure of legacy enterprise and operational data and to overcome some challenges of real-time data analytics. the potential of the proposed approach is high, especially in traditional industries that have not benefit from the advancements of industry . and that have just started investigating the potential of data analytics and machine learning for the optimization of their production processes. the traditional manufacturing sectors (e.g. textile, furniture, packaging, steel processing) have usually older factories with limited capacity on investing in modern production technologies. since the neural networks are inherently adaptive, the proposed approach could be applied to similar production lines (e.g. at a newly established factory of the same type) overcoming the cold-start problem, due to which other techniques usually fail. it also exploits both the "voice of data" and the "voice of experts". regarding future work, we plan to evaluate our proposed approach in additional use cases, with different requirements, as well as to investigate approaches and algorithms for fusion of the outcomes derived from real-time data analytics and operational data analytics that represent different levels of information. building an industry . analytics platform predictive, prescriptive and detective analytics for smart manufacturing in the information age big data challenges in smart manufacturing: a discussion paper for bdva and effra research & innovation roadmap alignment bdva the age of analytics: competing in a data-driven world predictive maintenance in a digital factory shopfloor: data mining on historical and operational data coming from manufacturers' information systems the internet of things for smart manufacturing: a review recent advances and trends in predictive manufacturing systems in big data environment a full history proportional hazards model for preventive maintenance scheduling a neural network application for reliability modelling and conditionbased predictive maintenance data mining in manufacturing: a review based on the kind of knowledge data mining in manufacturing: a review a practical approach to combine data mining and prognostics for improved predictive maintenance application of data mining in a maintenance system for failure prediction analyzing maintenance data using data mining methods machine learning for predictive maintenance: a multiple classifier approach prescriptive analytics smart manufacturing with prescriptive analytics prescriptive analytics: literature review and research challenges reinforcement learning: an introduction model-free adaptive optimal control of episodic fixedhorizon manufacturing processes using reinforcement learning a reinforcement learning framework for optimal operation and maintenance of power grids human-centered reinforcement learning: a survey multiobjective reinforcement learning: a comprehensive overview many-objective stochastic path finding using reinforcement learning policy shaping: integrating human feedback with reinforcement learning acknowledgments. this work is funded by the european commission project h uptime "unified predictive maintenance system" ( ). key: cord- -yomnqr authors: basile, davide; ter beek, maurice h.; legay, axel title: strategy synthesis for autonomous driving in a moving block railway system with uppaal stratego date: - - journal: formal techniques for distributed objects, components, and systems doi: . / - - - - _ sha: doc_id: cord_uid: yomnqr moving block railway systems are the next generation signalling systems currently under development as part of the shift rail european initiative, including autonomous driving technologies. in this paper, we model a suitable abstraction of a moving block signalling system with autonomous driving as a stochastic priced timed game. we then synthesise safe and optimal driving strategies for the model by applying advanced techniques that combine statistical model checking with reinforcement learning as provided by uppaal stratego. hence, we show the applicability of uppaal stratego in this concrete case study. next generation railway systems are based on distributed inter-organisational entities, such as on-board train computers and wayside radio-block centres and satellites, which have to interact to accomplish their tasks. a longstanding effort in the railway domain concerns the use of formal methods and tools for the analysis of railway (signalling) systems in light of the sector's stringent safety requirements [ , , , , [ ] [ ] [ ] [ ] [ ] , ] . due to their distributed and interorganisational nature, their formal verification is still an open challenge. whilst model-checking and theorem-proving techniques are predominant, to the best of our knowledge, applications of controller synthesis techniques are largely lacking. we describe a formal modelling and analysis experience with uppaal stratego of a moving block railway signalling system. this work was conducted in the context of several projects concerned with the use of formal methods and tools for the development of railway systems based on moving block signalling systems, in which train movement is no longer authorised based on sections of the railway track between fixed points, but computed in real time as safe zones around the trains. most notably, the h shift rail projects astrail: satellite-based signalling and automation systems on railways along with formal method and moving block validation (http://www.astrail. eu) and securail: formal methods and csirt for the railway sector (http://www. securail.eu). the european shift rail initiative (http://shift rail. org) is a joint undertaking of the european commission and the main railway stakeholders to move the european railway industry forward by increasing its competitiveness. this concerns in particular the transition to next generation signalling systems, including satellite-based train positioning, moving-block distancing, and automatic driving. with a budget of nearly billion euro, it is unique in its kind. previously, in [ , ] , we introduced a concrete case study of a satellite-based moving block railway signalling system, which was developed in collaboration with industrial partners of the astrail project and which was modelled and analysed with simulink and uppaal smc (statistical model checker). while those models offered the possibility to fine tune communication parameters that are fundamental for the reliability of their operational behaviour, they did not account for the synthesis of autonomous driving strategies. building on such efforts, in this paper we present a formal model of a satellitebased moving block railway signalling system, which accounts for autonomous driving and which is modelled in uppaal stratego as a stochastic priced timed game. the autonomous driving module is not modelled manually, but it is synthesised automatically as a strategy, after which both standard and statistical model checking are applied under the resulting (safe) strategy. the starting point for deriving the strategy is a safety requirement that the model must respect. we moreover consider reliability aspects, and the autonomous driving strategy also provides guarantees for the minimal expected arrival time. the model and experiments are available at https://github.com/davidebasile/forte . at last year's forte, parametric statistical model checking was applied to unmanned aerial vehicles (uav) [ ] . the model was formalised as a parametric markov chain with the goal of reducing the probability of failure while varying parameters such as precision of the position. the uav follows a predefined flight plan, whereas we aim at automatically synthesising a strategy to safely drive the train. it would be interesting to investigate the possibility of synthesising flight plans under safety constraints. a decade ago at forte' , one of the first applications of statistical model checking (using the bip toolset) to an industrial case study was presented, namely the heterogeneous communication system for cabin communication in civil airplanes [ ] . the goal was to study the accuracy of clock synchronisation between different devices running in parallel on a distributed application, i.e. a time bound within which communication must occur. an implementation of this case study in uppaal smc would allow a comparison of the results. statistical model checking has also been used to verify the reliability of railway interlocking systems [ ] and uppaal has been used to verify railway timetables [ ] . uppaal stratego has been applied to a few other case studies belonging to the transport domain, such as traffic light controllers [ ] , cruise control [ ] , and railway scheduling [ ] . we conjecture that the uppaal stratego model in [ ] could be paired with our model to study railway scheduling for autonomous trains, with the goal of synthesising improved strategies for both the scheduler and the autonomous driver. finally, there have been several recent attempts at modelling and analysing ertms level signalling systems (in particular hybrid level systems with virtual fixed blocks) with promela/spin, mcrl , alloy/electrum, iuml, sysml, prob, event-b, and real-time maude [ , , , , , , , ] . none of these concern quantitative modelling and analysis, typically lacking uncertainty, which is fundamental for demonstrating the reliability of the operational behaviour of next generation satellite-based ertms level moving block railway signalling system models. one of the earliest quantitative evaluations of moving block railway signalling systems can be found in [ ] , based on gsm-r communications. structure of the paper. after some background on uppaal stratego in sect. , we describe the setting of the case study from the railway domain in sect. . in sect. , we present the formal model, followed by an extensive description of the conducted analyses in sect. . finally, we discuss our experience with uppaal stratego and provide some ideas for future work in sect. . in this section, we provide some background of the tools and their input models used in this paper, providing pointers to the literature for more details. uppaal stratego [ ] is the latest tool of the uppaal [ ] suite. it integrates formalisms and algorithms coming from the less recent uppaal tiga [ ] (synthesis for timed games), uppaal smc [ ] (statistical model checking), and the synthesis of near optimal schedulers proposed in [ ] . uppaal tiga [ , ] implements an efficient on-the-fly algorithm for the synthesis of strategies extended to deal with models of timed games. these are automata modelling a game between a player (the controller) and an opponent (the environment). transitions are partitioned into controllable and uncontrollable ones. the controller plays the controllable transitions, while the opponent plays the uncontrollable ones. the controller is only allowed to deactivate controllable transitions. the goal is to synthesise a strategy for the controller such that, no matter the actions of the opponent, a particular property is satisfied. generally, uncontrollable transitions are used to model events such as delays in communication or other inputs from the environment. on the converse, controllable transitions characterise the logic of the controller, generally related to actuators. the strategy synthesis algorithm uses a suitable abstraction of the real-time part of the model, through zones that are constraints over the realtime clocks. strategy synthesis allows an algorithmic construction of a controller which is guaranteed to ensure that the resulting system satisfies the desired correctness properties, i.e. reachability and safety. uppaal smc is a statistical model checker based on models of stochastic timed automata. these are automata enhanced with real-time modelling through clock variables. moreover, their stochastic extension replaces non-determinism with probabilistic choices and time delays with probability distributions (uniform for bounded time and exponential for unbounded time). these automata may communicate via (broadcast) channels and shared variables. statistical model checking (smc) [ , ] is based on running a sufficient number of (probabilistic) simulations of a system model to obtain statistical evidence (with a predefined level of statistical confidence) of the quantitative properties to be checked. smc offers advantages over exhaustive (probabilistic) model checking. most importantly, smc scales better since there is no need to generate and possibly explore the full state space of the model under scrutiny, thus avoiding the combinatorial state-space explosion problem typical of model checking, and the required simulations can be easily distributed and run in parallel. this comes at a price: contrary to (probabilistic) model checking, exact results (with % confidence) are out of the question. the method proposed in [ ] extends the strategy synthesis of [ ] to find near-optimal solutions for stochastic priced timed games, which are basically stochastic timed automata enhanced with controllable and uncontrollable transitions, similarly to timed games. in short, the method starts from the most permissive strategy guaranteeing the time bounds, computed with the algorithms in [ ] . this strategy is then converted into a stochastic one by substituting non-determinism with uniform distributions. finally, reinforcement learning is applied iteratively to learn from sampled runs the effect of control choices, to find the near-optimal strategy. uppaal stratego uses stochastic priced timed games as formalism whilst integrating (real-time) model checking, statistical model checking, strategy synthesis, and optimisation. it thus becomes possible to perform model checking and optimisation under strategies, which are first-class objects in the tool. internally, abstractions that allow to pass from stochastic priced timed games to timed games similar to those in [ ] are used to integrate the various algorithms. the european railway traffic management system (ertms) is a set of international standards for the interoperability, performance, reliability, and safety of modern european rail transport [ ] . it relies on the european train control system (etcs), an automatic train protection system that continuously supervises the train, ensuring to not exceed the safety speed and distance. the current standards distinguish four levels ( - ) of operation of etcs signalling systems, depending largely on the role of trackside equipment and on the way information is transmitted to and from trains. the ertms/etcs signalling systems currently deployed on railways throughout europe concern at most level . level signalling systems are based on fixed blocks starting and ending at signals. the block sizes are determined based on parameters like the speed limit, the train's speed and braking characteristics, drivers' sighting and reaction times, etc. but the faster trains are allowed to run, the longer the braking distance and the longer the blocks need to be, thus decreasing the line's capacity. this is because the railway sector's stringent safety requirements impose the length of fixed blocks to be based on the worst-case braking distance, regardless of the actual speed of the train. for exact train position detection and train integrity supervision, level signalling systems make use of trackside equipment (such as track circuits or axle counters). however, communication of the movement authority (ma), i.e. the permission to move to a specific location with supervision of speed, as well as speed information and route data to and from the train is achieved by continuous data transmission via gsm-r or gprs with a wayside radio block centre. moreover, an onboard unit continuously monitors the transferred data and the train's maximum permissible speed by determining its position in between the eurobalises (transponders on the rails of a railway) used as reference points via sensors (axle transducers, accelerometer and radar). the next generation level signalling systems currently under investigation and development, no longer rely on trackside equipment for train position detection and train integrity supervision. instead, an onboard odometry system is responsible for monitoring the train's position and autonomously computing its current speed. the onboard unit frequently sends the train's position to a radio block centre which, in turn, sends each train a ma, computed by exploiting its knowledge of the position of the rear end of the train ahead. for this to work, the precise absolute location, speed, and direction of each train needs to be known, which are to be determined by a combination of sensors: active and passive markers along the track, and trainborne speedometers. the resulting moving block signalling systems allow trains in succession to close up, since a safe zone around the moving trains can be computed, thus considerably reducing headways between trains, in principle to the braking distance. this allows for more trains to run on existing railway tracks, in response to the ever-increasing need to boost the volume of passenger and freight rail transport and the cost and impracticability of constructing new tracks. furthermore, the removal of trackside equipment results in lower capital and maintenance costs [ ] . one of the current challenges in the railway sector is to make moving block signalling systems as effective and precise as possible, including satellite-based positioning systems and leveraging on an integrated solution for signal outages (think, e.g., of the absence of positioning in tunnels) and the problem of multipaths [ ] . however, due to its robust safety requirements the railway sector is notoriously cautious about adopting technological innovations. thus, while gnss-based positioning systems are in use for some time now in the avionics and automotive sectors, current train signalling systems are still based on fixed blocks. however, experiments are being conducted and case studies are being validated in order to move to level signalling systems [ , , , , , , , , , ] . the components of the moving block railway signalling case study considered in this paper are depicted in fig. . the train carries the location unit and onboard unit components, while the radio block centre is a wayside component. the location unit receives the train's location from gnss satellites, sends this location (and the train's integrity) to the onboard unit, which, in turn, sends the location to the radio block centre. upon receiving a train's location, the radio block centre sends a ma to the onboard unit (together with speed restrictions and route configurations), indicating the space the train can safely travel based on the safety distance with preceding trains. the radio block centre computes such ma by communicating with neighbouring radio block centres and exploiting its knowledge of the positions of switches and other trains (head and tail position) by communicating with a route management system. we abstract from the latter and from communication among neighbouring radio block centres: we consider one train to communicate with one radio block centre, based on a seamless handover when the train moves from one radio block centre supervision area to an adjacent one, as regulated by its functional interface specification [ ] . in this section, we describe the formal model of the case study introduced before. it consists of a number of sptgs, which are basically timed automata with prices (a cost function) and stochasticity, composed as a synchronous product. we briefly describe the model's components, followed by details of the onboard unit. delays in the communications are exponentially distributed with rate : to account for possible delays. this is a common way of modelling communication delays. moreover, all transitions are uncontrollable, except for the controllable actions of the driver in the train_ato_t component, which are used to synthesise the safe and optimal strategy. component obu_main_generatelocationrequest_t initiates system interactions by generating a request for a new location to send to the location unit. the location unit component lu_main_t receives a new position request from the onboard unit, replying with the current train location (computed via gnss). the main component obu_main_sendlocationtorbc_t of the onboard unit performs a variety of operations. it receives the position from the location unit, sends the received position to the radio block centre, and eventually implements a safety mechanism present in the original system specification. in particular, at each instant of time, it checks that the train's position does not exceed the ma received from the radio block centre; if it does, it enters a failure state. once received, the radio block centre repeatedly sends a ma message until the corresponding acknowledgement from the onboard unit is received. also obu_main_receivema_t models the logic of the onboard unit. it receives a ma from the radio block centre, and sends back a corresponding acknowledgement. finally, the train_ato_t component was defined to synthesise a strategy for moving the train in a safe and optimal way. in particular, the position of the train (variable loc) is stated in a unidimensional space and identified by one coordinate (representing the position along its route), and the train is allowed at each cycle to either move one unit or stay idle. to allow state-space reduction, the value of loc represents a range of the space in which the train is located, rather than a specific point in space. next, we describe this component, depicted in fig. , in detail. the initial state of train_ato_t is the nominal state i_go, drawn with two circles. two failure states (failwhilego and failwhilereadloc) are reached in case the ma is exceeded in obu_main_sendlocationtorbc_t. the initial state has an invariant to guarantee that the train has not passed its destination. note that invariants can be constraints on clocks or variables. this is done by checking that the location of the train, which is encoded by the integer variable loc, is less than or equal to the integer constant arrive, which is an input parameter of the model to perform experiments. from the initial state it is possible to transit to state readlocwhilerun, upon a location request coming from lu_main_t, and coming back from readlocwhilerun to i_go by replying to such a request. variable x is a buffer for value-reading messages. to reduce the model's state space, the value transmitted by train_ato is the remaining headway, i.e. the difference between the ma and the location. indeed, such value has a fixed range if compared to the location (under the assumption that the arrival point is greater than the initial headway value, otherwise the train will never exceed its ma before arriving to the destination). in turn, obu_main_sendlocationtorbc_t checks if such transmitted value (headway) is negative for triggering a failure, since in that case the train has exceeded its ma. from both states i_go and readlocwhilerun, an inner loop is used to receive the new ma (movaut) from rbc_main_t. the movaut should be relative to the current location loc of the train, i.e. movaut = loc + fixed number of meters that the train is allowed to travel. however, to reduce the state space, such a message simply resets the headway variable to its initial value, which is an integer constant called ma. thus, movaut is not stored in the state space because its value can be retrieved as loc+ma. the constant ma is another input parameter of the model. the reason such a loop is also present in state readlocwhilerun is that otherwise the ma message would be lost in this state, and similarly for the urgent state (marked with the symbol u, described below). we now discuss the two controllable transitions in the model. the first is used to move a train. in uppaal stratego a controller cannot delay its actions (whereas the environment can), hence the movement of the train is split into an uncontrollable transition followed by a controllable one, with an intermediate urgent state. an intermediate urgent state is such that a transition must be taken without letting time pass. this is a workaround to force the controller to perform an action at that instant of time. from the initial state i_go, an uncontrollable transition targeting the urgent state is used to check that the conditions for moving the train are met. in particular, if the headway is nonnegative and the train has not arrived, the transition for moving the train is enabled. additionally, a test c> on the clock c is used to forbid zeno behaviour. indeed the clock c is reset to zero after the train has moved, hence time is forced to pass before the next movement. such a condition cannot be stated directly on the controllable transition, otherwise a time-lock (i.e. time is not allowed to pass) would be reached in case the condition is not met. the controllable transition (drawn as a solid arc) from the urgent state can either set the integer speed to or to , allowing the train to proceed to the next interval of space or to remain in the previous interval, respectively. recall that loc is not a coordinate but rather an abstraction of a portion of space. the controllable transition also updates the headway. to reduce the state space, the only negative value allowed for the variable headway is − . finally, the second controllable transition is used to reach a sink state done. to further reduce the state space, the train is not allowed to move once loc has reached value arrive. a hybrid clock timer is used as a stop-watch to measure the time it takes for the train to arrive in state done. to this aim, the invariant timer'== in state done sets the derivative of clock timer to zero. a hybrid clock can be abstracted away during the synthesis of a safe strategy. in this section, we report on analysis of the formal model. the main objective is to synthesise a safe strategy such that the train does not exceed the ma. additionally, the train should be as fast as possible, within the limits imposed by the safety requirements. to this aim, an optimal and safe strategy is synthesised. the experiments were carried out on a machine with a processor intel(r) core(tm) i - y cpu at . ghz, mhz, cores, and logical processors with gb of ram, running bit windows . the development version of uppaal stratego (academic version . . - -beta ) was used. indeed, when developing the model and its analysis, minor issues were encountered (more later). this version of uppaal stratego contains some patches resulting from a series of interactions between the first author and the developers team at aalborg university. the set-up of the parameters of the statistical model checker was chosen to provide a good confidence in the results and is as follows: probabilistic deviation set to δ = . , probability of false negatives and false positives set to α = . and β = . , respectively, and the probability uncertainty set to = . . as anticipated in sect. , we focussed on one radio block centre, one onboard unit, and one location unit, i.e. one train communicating with one radio block centre. finally, we set ma = and arrive = . to begin with, we want to check if the hazard of exceeding the ma is possible at all in our model. if such a hazard were never possible, the safe strategy would simply allow all behaviour. to analyse this, we perform standard model checking: this formula checks if for every possible path, the state maexceededfailure of the component obu_main_sendlocationtorbc is never visited. indeed, this particular state is reached exactly when the hazard occurs, i.e. the ma is exceeded, thus triggering a failure. after . s, using , kb of memory, uppaal stratego reports that this formula is not satisfied, thus such hazard is possible without a proper strategy to drive the train. we would like to check the likelihood to reach this failure, given this specific set-up of parameters. first, the average maximal time in which the train reaches its destination is computed. this is important to fine tune the time bound for further simulations. to do so, we use the statistical model checker to evaluate the following formula: this formula computes the average maximal value of the train_ato.timer stopwatch, i.e. the arrival time. it is computed based on simulations with an experimental time bound of s. the computed value is . ± . , and its probability distribution is depicted in fig. . by analysing the probability distribution it is possible to notice that the average value is lower if faults are ignored. indeed, in case of faults the value of timer is equal to the end of the simulation (i.e. s). hence, the time bound for the following simulations is set to s, thus considering also worst cases of arrival time. we now compute the likelihood of the model to reach a failure, using smc to measure the probability of reaching the failure state with the following formula: uppaal stratego executes simulations and the probability is within the range [ . , . ], with confidence . . the probability confidence interval plot for this experiment is depicted in fig. . we conclude that, for this set-up of parameters, there is a relatively high probability for this hazard to occur. this is as expected, due to the absence of a strategy for driving the train and the non-deterministic choice of whether or not to move the train. after these standard and statistical model-checking experiments, we exploit the synthesis capabilities of uppaal stratego to automatically fix the specification to adhere to safety constraints. indeed, no manual intervention to fix the model is needed: it suffices to compute a driving strategy and compose it with the model. recall that the only controllable transitions in the model are those for deciding whether or not to move the train (i.e. related to acceleration/deceleration, accordingly). this in turn depends on the stochastic delays in communication. the strategy prunes controllable transitions such that those previously reachable configurations leading to the failure state are no longer figure shows the trajectory of variable train_ato.loc for simulations, computed in . s, using , kb. we see that in all trajectories the train never stops before reaching its destination, i.e. no failure occurs. however, in some simulations the train is relatively slower, when compared to other simulations. uppaal stratego also allows to model check the synthesised strategies. we ran a full state-space exploration by means of standard model checking to formally verify that after composing the model with the safe strategy the hazard of exceeding the ma is mitigated. this is checked through the following formula: this formula checks that in the model composed with the safe strategy, the 'bad' state is never reached. after . s and using , kb of memory, uppaal stratego reports that the formula is satisfied, thus confirming that we automatically synthesised a strategy for mitigating the hazard. however, even if not showed in fig. , there exist trajectories in the composition where the train never reaches its destination. this can be formally proven with a full state-space exploration of the strategy by standard model checking of the following formula: this formula checks that in all paths eventually state train_ato.done is reached (i.e. the train reached its destination). after . s, using , kb of memory, uppaal stratego reports that the formula does not hold. indeed, as expected, the strategy does not guarantee that such a state is always reached, but it only guarantees to avoid state obu_main_sendlocationtorbc.maexceededfailure. for example, there exists also a safe strategy that allows the train to remain in its starting position. to evaluate the probability to reach state train_ato.done under the safe strategy, we ran the statistical model checker to evaluate the following formula: fig. . we conclude that the likelihood for the train to not reach its destination within time units under the safe strategy is low, and it is mainly due to the possibility of large delays in communications. these delays are indeed the only source of stochastic behaviour in the model. we now show how uppaal stratego can account for dependability parameters other than safety. in particular, reliability of the system can be related to the capacity of the train to reach its destination quickly. we optimise the safe strategy to minimise the arrival time, thus increasing its reliability whilst satisfying safety. this can be done with the following query, computed in . s using , kb of memory: as expected, the optimised safe strategy has improved the arrival time of the safe strategy. the probability distribution of query φ is depicted in fig. . sensitive analysis of maximal headway. up to this point, we evaluated the moving block railway signalling system under analysis with a specific parameter set-up. in this set-up, each time the train receives a fresh ma, its headway is reset to (i.e. ma = ). thus, this is the maximal possible headway. the parameters of the model can be tuned in such a way that the analysed properties are within a desired range of values. in particular, we hypothesise that reducing the maximal headway (i.e. ma) results in a deterioration in performance of the optimal strategy and in an increment of the probability of reaching a failure without strategy. indeed, with a tight headway, the train is forced to move slowly in order not to exceed its ma. in the remainder of this section, we experimentally verify our hypothesis. table reports the evaluation of properties φ -φ in three different experiments, with values for ma taken from the set { , , }, reporting also the computation times and, where appropriate, the number of runs. by reducing the maximal headway (i.e. ma = ), we notice an overall deterioration of the average maximal arrival time (cf. properties φ , φ , and φ ). moreover, without strategy the probability of failure is higher when compared to ma = (cf. property φ ). these results confirm our hypothesis and further corroborate the reliability of our model. as a final experiment, we enlarged the maximal headway (i.e. ma = ) to evaluate the improvement in performance in case of a larger headway. we recall that a large headway is not desirable, since it would result in a lower capacity of the railway network. in this experiment, the values of φ and φ are similar to the case of ma = . however, by observing the values of φ , φ , and φ , we note that there is only a slight improvement in arrival time, even if we have doubled the maximal headway. this experiment confirms our intuition that an excessive increment of the maximal headway does not lead to a better performance. this is because the train cannot go faster than its optimal speed. on the converse, an excessive enlargement of the headway results in a deterioration of the overall track capacity. hence, ma = is a satisfactory set-up for the maximal headway. we have modelled and analysed an autonomous driving problem for a moving block railway signalling system. communication between the train and the radio block centre are modelled such that the train is allowed to proceed only within the limits imposed by the radio block centre via the ma, which is based on the position of the train and updated continuously. the goal is to synthesise a strategy for the train to arrive to its destination as quickly as possible without exceeding its limits. we modelled the problem as a stochastic priced timed game. the controller is in charge of moving the train, playing against uncontrollable stochastic delays in communication. we used uppaal stratego to compute a strategy to enforce safety in the model. the safe strategy was statistically model checked to evaluate the mean arrival time of the train. this quantity was optimised, and the optimised strategy was compared to the safe one. we observed an improvement in the mean arrival time, whilst retaining safety. as far as we know, this is the first application of synthesis techniques to autonomous driving for next generation railway signalling systems. this was our first experience with strategy synthesis and optimisation of a case study from the railway domain and also with uppaal stratego. since this is a very recent tool there has not been much experimentation, in particular not outside of the groups involved in its development. the tool is still undergoing testing, and new versions and patches are released frequently. in fact, while developing the model we ran into corner cases that needed interactions with the developers team at aalborg university. those interactions led to the release of new versions, with patches fixing the issues discovered through our model. we did have experience in modelling and analysing railway case studies with uppaal smc [ , , ] . the original model developed in [ ] and statistically model checked had to be simplified considerably (cf. sect. ) to undergo strategy synthesis and verification. indeed, while uppaal smc scales to large systems by applying simulations rather than full state-space exploration, uppaal stratego requires full state-space exploration of the timed game for strategy synthesis. for example, using the set-up discussed in sect. with ma = , if we double the constant arrive (i.e. instead of ) then during the strategy synthesis the tool terminates with an error message due to memory exhaustion. an interesting future line of research would be to adapt the statistical synthesis techniques described in [ , ] to learn safety objectives, thus avoiding the full state-space exploration (as currently performed in uppaal stratego) while guaranteeing the scalability of smc. this would enable the modelling of more complex ertms case studies. also, further experiments, with different set-ups of the parameters and more trains and radio block centres need to be performed, to investigate the limits of the approach described in this paper in terms of optimisation. finally, we intend to discuss with our railway project partners the impact of the techniques discussed in this paper. a survey of statistical model checking modelling the hybrid ertms/etcs level case study in spin coordinated intelligent traffic lights using uppaal stratego parametric statistical model checking of uav flight plan modelling and analysing ertms hybrid level with the mcrl toolset statistical model checking of a moving block railway signalling scenario with uppaal smc on the industrial uptake of formal methods in the railway domain modelling and analysing ertms l moving block railway signalling with simulink and uppaal smc statistical abstraction and model-checking of large heterogeneous systems adopting formal methods in an industrial setting: the railways case formal methods for transport systems quantitative evaluation of systems (qest) uppaal-tiga: time for playing games! verification of the european rail traffic management system in real-time maude performability evaluation of the ertms/etcs -level partial order reduction for reachability games formal methods applied to industrial complex systems -implementation of the b method abz verification of interlocking systems using statistical model checking efficient on-the-fly algorithms for the analysis of timed games validating the hybrid ertms/etcs level concept with electrum uppaal smc tutorial on time with minimal expected cost! uppaal stratego diagram-led formal modelling using iuml-b for hybrid ertms level ertms/etcs rams requirements specification -chap. -ram, twenty-five years of formal methods and railways: what next? formal methods and safety certification: challenges in the railways domain some trends in formal methods applications to railway signaling model-based development and formal methods in the railway industry comparing formal tools for system design: a judgment study ertms level : the game-changer validation and real-life demonstration of etcs hybrid level principles using a formal b model formal verification of railway timetables -using the uppaal model checker teaching stratego to play ball: optimal synthesis for continuous space mdps dependability checking with stocharts: is train radio reliable enough for trains? safe and time-optimal control for railway games safe and optimal adaptive cruise control statistical model checking: an overview an event-b model of the hybrid ertms/etcs level standard ten diverse formal models for a cbtc automatic train supervision system towards formal methods diversity in railways: an experience report with seven frameworks modeling railway control systems in promela recent progress in application of gnss and advanced communications for railway signaling unisig: fis for the rbc/rbc handover acknowledgements. funding by miur prin ftxr s project it matters (methods and tools for trustworthy smart systems) and h project securail (formal methods and csirt for the railway sector). the securail project received funding from the shift rail joint undertaking under eu's h research and innovation programme under grant agreement .we thank the uppaal developers team, in particular danny poulsen, marius mikucionis, and peter jensen, for their assistance with uppaal stratego. key: cord- -rk fuovf authors: venero, sheila katherine; schmerl, bradley; montecchi, leonardo; dos reis, julio cesar; rubira, cecília mary fischer title: automated planning for supporting knowledge-intensive processes date: - - journal: enterprise, business-process and information systems modeling doi: . / - - - - _ sha: doc_id: cord_uid: rk fuovf knowledge-intensive processes (kips) are processes characterized by high levels of unpredictability and dynamism. their process structure may not be known before their execution. one way to cope with this uncertainty is to defer decisions regarding the process structure until run time. in this paper, we consider the definition of the process structure as a planning problem. our approach uses automated planning techniques to generate plans that define process models according to the current context. the generated plan model relies on a metamodel called metakip that represents the basic elements of kips. our solution explores markov decision processes (mdp) to generate plan models. this technique allows uncertainty representation by defining state transition probabilities, which gives us more flexibility than traditional approaches. we construct an mdp model and solve it with the help of the prism model-checker. the solution is evaluated by means of a proof of concept in the medical domain which reveals the feasibility of our approach. in the last decades, the business process management (bpm) community has established approaches and tools to design, enact, control, and analyze business processes. most process management systems follow predefined process models that capture different ways to coordinate their tasks to achieve their business goals. however, not all types of processes can be predefined at design timesome of them can only be specified at run time because of their high degree of uncertainty [ ] . this is the case with knowledge-intensive processes (kips). kips are business processes with critical decision-making tasks that involve domain-specific knowledge, information, and data [ ] . kips can be found in domains like healthcare, emergency management, project coordination, and case management, among others. kip structure depends on the current situation and new emergent events that are unpredictable and vary in every process instance [ ] . thus, a kip's structure is defined step by step as the process executes, by a series of decisions made by process participants considering the current specific situations and contexts [ ] . in this sense, it is not possible to entirely define beforehand which activities will execute or their ordering and, indeed, it is necessary to refine them as soon as new information becomes available or whenever new goals are set. these kinds of processes heavily rely on highly qualified and trained professionals called knowledge workers. knowledge workers use their own experience and expertise to make complex decisions to model the process and achieve business goals [ ] . despite their expertise, it is often the case that knowledge workers become overwhelmed with the number of cases, the differences between cases, rapidly changing contexts, and the need to integrate new information. they therefore require computer-aided support to help them manage these difficult and error-prone tasks. in this paper, we explore how to provide this support by considering the process modeling problem as an automated planning problem. automated planning, a branch of artificial intelligence, investigates how to search through a space of possible actions and environment conditions to produce a sequence of actions to achieve some goal over time [ ] . our work investigates an automated way to generate process models for kips by mapping an artifact-centric case model into a planning model at run time. to encode the planning domain and planning problem, we use a case model defined according to the metakip metamodel [ ] that encloses data and process logic into domain artifacts. it defines datadriven activities in the form of tactic templates. each tactic aims to achieve a goal and the planning model is derived from it. in our approach, we use markov decision processes (mdp) because they allow us to model dynamic systems under uncertainty [ ] , although our definition of the planning problem model enables using different planning algorithms and techniques. mdp finds optimal solutions to sequential and stochastic decision problems. as the system model evolves probabilistically, an action is taken based on the observed condition or state and a reward or cost is gained [ , ] . thus, an mdp model allows us to identify decision alternatives for structuring kips at run time. we use prism [ ] , a probabilistic model checker, to implement the solution for the mdp model. we present a proof of concept by applying our method in a medical treatment scenario, which is a typical example of a non-deterministic process. medical treatments can be seen as sequential decisions in an uncertain environment. medical decisions not only depend on the current state of the patient, but they are affected by the evolution of the states as well. the evolution of the patient state is unpredictable, since it depends on factors such as preexisting patient illnesses or patient-specific characteristics of the diseases. in addition, medical treatment decisions involve complex trade-offs between the risks and benefits of various treatment options. we show that it is possible to generate different optimal treatment plans according to the current patient state and a target goal state, assuming that we have enough data to accurately estimate the transition probabilities to the next patient state. the resulting process models could help knowledge workers to make complex decisions and structure execution paths at run time with more probability of success and optimizing constraints, such as cost and time. the remainder of this paper is organized as follows: sect. presents a motivating medical scenario. section introduces the theoretical and methodological background. section describes the proposed method to encode a case model as a planning model. section reports on the application of the methodology in a scenario. section discusses the obtained findings and related work. finally, sect. wraps up the paper with the concluding remarks. this section presents a motivating medical case scenario. suppose we have the following medical scenario in the oncology department stored in the electronic medical record (emr). in order to receive the second cycle of r-ice, it is necessary to stabilize mary's health status as soon as possible. thus, at this time the goal is to decrease her body temperature to . • c ≤ t emp ≤ . • c and reduce the level of nausea to zero ln = . for that, physicians need to choose from vast treatment strategies to decide which procedures are the best for mary, in her specific current context. assume that we have statistical data about two possible tactics for achieving the desired goal: fever (fvr) and nausea (nausea) management, shown in table adapted from [ ] . each of these tactics can be fulfilled through multiple activities that have different interactions and constraints with each other, as well as to the specifics of the patient being treated. for example, (a) treating nausea with a particular drug may affect the fever, (b) administration of the drug may depend on the drugs that the patient is taking, (c) drug effectiveness may depend on the patient history with the drug, or (d) giving the drug may depend on whether the drug has already been administered and how much time has elapsed since the last dose. these issues make manual combination of even this simple case challenging, and it becomes much harder for more complex treatments and patient histories. support is therefore needed that can take into account patient data, constraints, dependencies, and patient/doctor preferences to help advise the doctor on viable and effective courses of treatment. this section presents the underlying concepts in our proposal. section . provides an overview of the metakip metamodel; sect. . introduces basic concepts of automated planning; sect. . explains markov decision process (mdp). section . describes the prism tool and language. our previous work proposed an artifact-centric metamodel [ ] for the definition of kips, aiming to support knowledge workers during the decision-making process. the metamodel supports data-centric process management, which is based on the availability and values of data rather than completion of activities. in data-centric processes, data values drive decisions and decisions dynamically drive the course of the process [ ] . the metamodel is divided into four major packages: case, control-flow, knowledge, and decision, in such a way that there is an explicit integration of the data, domain, and organizational knowledge, rules, goals, and activities. the case package defines the base structure of the metamodel, a case. a case model definition represents an integrated view of the context and environment data of a case, following the artifact-centric paradigm. this package is composed of a set of interconnected artifacts representing the logical structure of the business process. an artifact is a data object composed of a set of items, attributes, and data values, defined at run time. the knowledge package captures explicit organizational knowledge, which is encoded through tactic templates, goals, and metrics that are directly influenced by business rules. tactics templates represent best practices and guidelines. usually, they have semi-structured sequences of activities or unstructured loose alternative activities pursuing a goal. the control-flow package defines the behavior of a case. it is composed of a set of data-driven activities to handle different cases. activity definitions are made in a declarative way and have pre-and post-conditions. the metamodel refines the granularity of an activity that could be a step or a task. a task is logically divided into steps, which allows better management of data entry on the artifacts. step definitions are associated with a single attribute of an artifact, a resource, and a role type at most. this definition gives us a tight integration between data, steps and resources. these packages are used to model alternative plans to answer emergent circumstances, reflecting environmental changes or unexpected outcomes during the execution of a kip. the decision package represents the structure of a collaborative decision-making process performed by knowledge workers. we proposed a representation of how decisions can be made by using the principles of strategic management, such as, looking towards goals and objectives and embracing uncertainty by formulating strategies for the future and correct them if necessary. the strategic plan is structured at run time by goals, objectives, metrics and tactic templates. planning is the explicit and rational deliberation of actions to be performed to achieve a goal [ ] . the process of deliberation consists of choosing and organizing actions considering their expected outcomes in the best possible way. usually, planning is required when an activity involves new or less familiar situations, complex tasks and objectives, or when the adaptation of actions is constrained by critical factors such as high risk. automated planning studies the deliberation process computationally [ ] . a conceptual model for planning can be represented by a state-transition system, which formally is a -tuple Σ = (s, a, e, γ), where s = {s , s , ....} is a finite or recursively enumerable set of states; a = {a , a , ...} is a finite or recursively enumerable set of actions; e = {e , e , ...} is a finite or recursively enumerable set of events; and γ : s × a × e → s is a state-transition function. actions are transitions controlled by a plan executor. events are unforeseen transitions that correspond to the internal dynamics of the system and cannot be controlled by the plan executor. both events and actions contribute to the evolution of the system. given a state transition system Σ, the purpose of planning is to deliberate which actions to apply into which states to achieve some goal from a given state. a plan is a structure that gives the appropriate actions. a markov decision process (mdp) is a discrete-time stochastic control process. it is a popular framework designed to make decisions under uncertainty, dealing with nondeterminism, probabilities, partial observability, and extended goals [ ] . in mdps, an agent chooses action a based on observing state s and receives a reward r for that action [ ] . the state evolves probabilistically based on the current state and the action taken by the agent. figure (a) presents a decision network [ ] , used to represent a mdp. the state transition function t (s |s, a) represents the probability of transitioning from state s to s after executing action a. the reward function r(s, a) represents the expected reward received when executing action a from state s. we assume that the reward function is a deterministic function of s and a. an mdp treats planning as an optimization problem in which an agent needs to plan a sequence of actions that maximizes the chances of reaching the goal. action outcomes are modeled with a probability distribution function. goals are represented as utility functions that can express preferences on the entire execution path of a plan, rather than just desired final states. for example, finding the optimal choice of treatment optimizing the life expectancy of the patient or optimizing cost and resources. prism [ ] is a probabilistic model checker that allows the modeling and analysis of systems that exhibit probabilistic behavior. the prism tool provides support for modeling and construction of many types of probabilistic models: discrete-time markov chains (dtmcs), continuous-time markov chains (ctmcs), markov decision processes (mdps), and probabilistic timed automata (ptas). the tool supports statistical model checking, confidence-level approximation, and acceptance sampling with its discrete-event simulator. for nondeterministic models it can generate an optimal adversary/strategy to reach a certain state. models are described using the prism language, a simple, state-based language based on the reactive modules formalism [ ] . figure (b) presents an example of the syntax of a prism module and rewards. the fundamental components of the prism language are modules. a module has two parts: variables and commands. variables describe the possible states that the module can be in at a given time. commands describe the behavior of a module, how the state changes over time. a command comprises a guard and one or more updates. the guard is a predicate over all the variables in the model. each update describes a transition that the module can take if the guard is true. a transition is specified by giving the new values of the variables in the module. each update has a probability which will be assigned to the corresponding transition. commands can be labeled with actions. these actions are used for synchronization between modules. cost and rewards are expressed as real values associated with certain states or transitions of the model. in our approach, plans are fragments of process models that are frequently created and modified during process execution. plans may change as new information arrives and/or when a new goal is set. we advocate the creation of a planner to structure process models at run time based on a knowledge base. the planner synthesizes plans on-the-fly according to ongoing circumstances. the generated plans should be revised and re-planned as soon as new information becomes available. thereby, it involves both computer agents and knowledge workers in a constant interleaving of planning, execution (configuration and enactment), plan supervision, plan revision, and re-planning. an interactive software tool might assist human experts during planning. this tool should allow defining planning goals and verifying emerging events, states, availability of activities and resources, as well as preferences. the run-time generation of planning models according to a specific situation in a case instance requires the definition of the planning domain and then the planning problem itself. definition . let the case model be represented according to the metakip metamodel. the planning domain is derived from the case model that can be described using a state-transition system defined as a -tuple Σ = (s, a, e, γ, c) such as that: s is the set of possible case states. a is the set of actions that are represented by activities inside tactics that an actor may perform. e is the set of events in the context or in the environment. γ : s × a × e → s , is the state-transition function, so the system evolves according to the actions and events that it receives. c : s ×a → [ , ∞) is the cost function that may represent monetary cost, time, risk or something that can be minimized or maximized. the state of a case is the set of values (available data) of the attributes contained in artifacts of the context and the environment. however, since the number of attributes of the artifacts is very large, it is necessary to limit the number of attributes to only the most relevant ones, which determines the current state of the case at a given time t. actions in the metakip metamodel are represented by the activities within a tactic. tactics represent best practices and guidelines used by the knowledge workers to make decisions. in metakip, they serve as tactic templates to be instantiated to deal with some situations during the execution of a case instance. tactics are composed of a finite set of activities pursuing a goal. a tactic can be structured or unstructured. a tactic is a -tuple t = (g, p c, m, a), where: g is a set of variables representing the pursuing goal state, p c is a finite set of preconditions representing a state required for applying the tactic, m is a set of metrics to track and assess the pursuing goal state, and a is a finite set of activities. in metakip, an activity could be a single step or a set of steps (called a task). an activity has some preconditions and post-conditions (effects). we map activities into executable actions. an executable action is an activity in which their effects can modify the values of the attributes inside business artifacts. these effects can be deterministic or non-deterministic. ef ∈eff p ef (i) = . c is the number which represents the cost (monetary, time, etc.) of performing a. as the state-transition function γ is too large to be explicitly specified, it is necessary to represent it in a generative way. for that, we use the planning operators from which it is possible to compute γ. thus, γ can be specified through a set of planning operators o. a planning operator is instantiated by an action. at this point, we are able to define the planning problem to generate a plan as a process model. definition . the planning problem for generating a process model at a given time t is defined as a triple p = (os t , gs t , ro t ), where: os t is the observable situation of a case state at time t. gs t is the goal state at time t, a set of attributes with expected output values. ro t represents a subset of the o that represents only available and relevant actions for a specific situation during the execution of a case instance at a given time t. where the state of c is s t and the set issues in the situation of c is i t . is an attribute with an expected output value, v i belongs to an artifact of c. these attributes are selected by the knowledge workers. some metrics required to asses some goals inside tactics can be added to the goal. gs t represents the expected reality of c. gs t serves as an input for searching an execution path for a specific situation. different goal states can be defined over time. let p = (os t , gs t , ro t ) be the planning problem. a plan π is a solution for p . the state produced by applying π to a state os t in the order given is the state gs t . a plan is any sequence of actions π = (a , ..., a k ), where k ≥ . the plan π represents the process model. our problem definition enables the use of different planning algorithms and the application of automatic planning tools to generate alternatives plans. as we are interested in kips, which are highly unpredictable processes, we use markov decision processes for formulating the model for the planner. mdps allows us to represent uncertainty with a probability distribution. mdp makes sequential decision making and reasons about the future sequence of actions and obstructions, which provides us with high levels of flexibility in the process models. in the following, we show how to derive an mdp model expressed in the prism language from a metakip model automatically. algorithm shows the procedure to automatically generate the mdp model for the prism tool, where the input parameters are: os t , gs t , set of domain t actics, t is the given time, p p minimum percentage of preconditions satisfaction, and p g minimum percentage of goal satisfaction, both p p and p g are according to the rules of the domain. as described in sect. . , a module is composed of variables and commands. variables of the module are the set of attributes from the case artifacts that belong to os t ∪ gs t . commands are represented for the relevant planning operators ro t . the name of the command is the identifier of the action, the guards are the preconditions p c and the effects eff are the updates with associated probabilities. rewards are represented by the cost of actions c and are outside of the module of prism. add necessary metrics to evaluate createp rismmodel(v, c, r) for finding the set of relevant planning operators ro t , first, we select tactics whose preconditions must be satisfied by the current situation os t and whose goal is related to the target state gs t . this can be done by calculating the percentages of both the satisfied preconditions and achievable goals. if these percentages are within an acceptable range according to the rules of the domain, the tactics are selected. second, this first set of tactics is shown to the knowledge workers who select the most relevant tactics. the set of the selected relevant tactics is denoted as rt . from this set of tactics, we verify which activities inside the tactics are available at time t. thus, the set of available actions at time t is denoted by a t = a , a , . . . , a n . finally, the relevant planning operators, ro t , are created by means of a t . to generate plans in prism, it is necessary to define a property file that contains properties that define goals as utility functions. prism evaluates properties over an mdp model and generates all possible resolutions of non-determinism in the model, state graphs, and gives us the optimal state graph. the state graph describes a series of possible states that can occur while choosing actions aiming to achieve a goal state. it maximizes the probability to reach the goal state taking into consideration rewards computed, that is maximizing or minimizing rewards and costs. in our context, a property represents the goal state gs t to be achieved while trying to optimize some criteria. then, prism calculates how desirable an executing path is according to one criterion. thus, plans can be customized according to knowledge workers' preferences (costs and rewards). to generate a plan, we need to evaluate a property. the generated plan is a state graph that represents a process model to be executed at time t. the generated process model shows case states as nodes and states transitions as arcs labeled with actions which outcomes follow probability distribution function. according to this state graph, the knowledge worker could choose which action to execute in a particular state. this helps knowledge workers to make decisions during kips execution. this section formulates a patient-specific mdp model in prism for the medical scenario presented in sect. . in the area of health care, medical decisions can be modeled with markov decisions processes (mdp) [ , ] . although mdp is more suitable for certain types of problems involving complex decisions, such as liver transplants, hiv, diabetes, and others, almost every medical decision can be modeled as an mdp [ ] . we generate the prism model by defining the observable situation os t , goal state gs t , and the set of relevant planning operators ro t . taking in consideration the medical scenario, the observable situation is os = {t emp = • , ln = } and the goal state is gs = { • c ≤ t emp ≤ . • c, ln = } where: temp is the temperature of the patient and ln is the level of nausea, both attributes of the health status artifact. we assume that the set of relevant tactics rt according to the current health status of the patient are fever and nausea management, presented in sect. . table shows the specification of one activity of each tactic, showing their preconditions, effects with their probability, time, and cost of execution. we modeled the activity effects with probabilities related to the probability of the patient to respond to the treatment. for example, the possible effects of applying the activity administer oral antipyretic medication are: (e ) the patient successfully responds to treatment, occurring with a probability . ; (e ) % of the time the patient partially responds to treatment where their temperature decreases by . • or more fails to reach the goal level; and (e ) the patient does not respond at all to treatment or gets worse (occurring with a probability of . ). the other activities are similarly modeled according to the response of the patient. assuming that all activities from both tactics are available, the set of executable actions is a t = {a , a , a , a , a , b , b , b }. then, it is possible to model the set of relevant planning operators ro t . having os t , gs t and ro t , it is possible to generate the mdp model in the language prism. once we created the mdp model, the following utility functions were evaluated: minimize time and cost while reaching the target state. the optimal plan to achieve the goal state gs t while minimizing the cost shows that reachability is eight iterations. the resulting model has states, transitions, and choices. the time for the model construction was . s. figure presents only a fragment of the model generated, highlighting the most probable path from the initial state to the goal state. the first suggested action is b (labeled arc) with possible outcome states with their probabilities. if the most probable next state is achieved, the next action to perform is a which has a probability of . to reach the goal state. knowledge workers can use this generated plan to decide which is the next activity they should perform in a particular state. to make the plan readable to knowledge workers, they could be presented with only the most probable path, and this could be updated according to the state actually reached after activity execution. further studies are necessary to help guiding knowledge workers in interpreting and following the model. in the last decades, there has been a growing interest in highly dynamic process management, with different types of approaches that deal with the variability, flexibility, and customization of processes at design time and at run time. most approaches start from the premise that there is a process model to which different changes have to be made, such as adding or deleting fragments according to a domain model or to generate an alternative sequence of activities due to some customization option. a few approaches use automated planning for synthesizing execution plans. laurent et al. [ ] explored a declarative modeling language called alloy to create the planning model and generate the plans. this approach seems to be very promising for activity-centric processes, but not effective enough for data-centric processes, as data is not well-enough treated to be the driver of the process as required in kips. smartpm [ ] investigated the problem of coordinating heterogeneous components inside cyber-physical systems. they used a pddl (planning domain definition language) planner that evaluates the physical reality and the expected reality, and synthesize a recovery process. similarly, marrella and lespérance proposed an approach [ ] to dynamically generate process templates from a representation of the contextual domain described in pddl, an initial state, and a goal condition. however, for the generation of the process templates, it is assumed that tasks are black boxes with just deterministic effects. on the other hand, henneberger et al. [ ] explored an ontology for generating process models. the generated process models are action state graphs (asg). although this work uses a very interesting semantic approach, they did not consider important aspects such as resources and cost for the planning model. there has been an increasing interest in introducing cognitive techniques for supporting the business process cycle. ferreira et. al. [ ] proposed a new life cycle for workflow management based on continuous learning and planning. it uses a planner to generate a process model as a sequence of actions that comply with activity rules and achieve the intended goal. hull and nezhad [ ] proposed a new cycle plan-act-learn for cognitively-enabled processes that can be carried out by humans and machines, where plans and decisions define actions, and it is possible to learn from it. recently, marrella [ ] showed how automatic planning techniques can improve different research challenges in the bpm area. this approach explored a set of steps for encoding a concrete problem as a pddl planning problem with deterministics effects. in this paper we introduced the notion of the state of a case regarding datavalues in the artifacts of a case instance. from this state, we can plan different trajectories towards a goal state using automated planning techniques. our solution generates action plans considering the non-deterministic effects of the actions, new emerging goals and information, which provides high levels of flexibility and adaptation. as we describe a generic planning model, it is possible to use different planning algorithms or combine other planning models, such as the classical planning model or the hierarchical task network (htn), according to the structuring level of the processes at different moments. thereby, we could apply this methodology to other types of processes, from well-structured processes to loosely or unstructured processes. our approach relies on mdp, which requires defining transition probabilities, which in some situations can be very difficult and expensive to get. nowadays a huge amount of data is produced by many sensors, machines, software systems, etc, which might facilitate the acquisition of data to estimate these transition probabilities. in the medical domain, the increasing use of electronic medical record systems shall provide the medical data from thousands of patients, which can be exploited to derive these probabilities. a limitation in mdps refers to the size of the problem because the size of the state-space explodes, and it becomes more difficult to solve. in this context, several techniques for finding approximate solutions to mdps can be applied in addition to taking advantage of the rapid increase of processing power in the last years. flexible processes could be easily designed if we replan after an activity execution. in fact, our approach suggests a system that has a constant interleaving of planning, execution, and monitoring. in this way, it will help knowledge workers during the decision-making process. process modeling is usually conducted by process designers in a manual way. they define the activities to be executed to accomplish business goals. this task is very difficult and prone to human errors. in some cases (e.g., for kips), it is impossible due to uncertainty, context-dependency, and specificity. in this paper, we devised an approach to continually generate run-time process models for a case instance using an artifact-centric case model, data-driven activities, and automatic planning techniques, even for such loosely-structured processes as kips. our approach defined how to synthesize a planning model from an artifactoriented case model defined according to the metakip metamodel. the formulation of the planning domain and the planning problem rely on the current state of a case instance, context and environment, target goals, and tactic templates from which we can represent actions, states, and goals. as our focus is kips management, we chose to use the mdp framework that allows representing uncertainty, which is one of kips essential characteristics. to automatically generate the action plan, we used the tool prism, which solves the mdp model and provides optimal solutions. future work involve devising a user-friendly software application for knowledge workers to interact with the planner and improve the presentation of plans in such a way that it is more understandable to them. our goal is to develop a planner which combines different types of planning algorithms to satisfy different requirements in business processes, especially regarding the structuring level. this planner will be incorporated into a fully infrastructure for managing knowledge-intensive processes that will be based on the dw-saarch reference architecture [ ] . reactive modules nursing interventions classification (nic)-e-book thinking for a living. how to get better performance and results knowledge-intensive processes: characteristics, requirements and analysis of contemporary approaches mdps in medicine: opportunities and challenges an integrated life cycle for workflow management based on learning and planning automated planning: theory and practice semantic-based planning of process models rethinking bpm in a cognitive world: transforming how we learn and perform business processes decision making under uncertainty: theory and application prism . : verification of probabilistic real-time systems planning for declarative processes towards is supported coordination in emergent business processes automated planning for business process management a planning approach to the automated synthesis of template-based process models smartpm: an adaptive process management system through situation calculus, indigolog, and classical planning a markov decision process model to guide treatment of abdominal aortic aneurysms enabling flexibility in process-aware information systems: challenges, methods technologies dw-saaarch: a reference architecture for dynamic self-adaptation in workflows towards a metamodel for supporting decisions in knowledge-intensive processes key: cord- - l mua authors: menotti-raymond, marilyn; o’brien, stephen j. title: the domestic cat, felis catus, as a model of hereditary and infectious disease date: journal: sourcebook of models for biomedical research doi: . / - - - - _ sha: doc_id: cord_uid: l mua the domestic cat, currently the most frequent of companion animals, has enjoyed a medical surveillance, as a nonprimate species, second only to the dog. with over hereditary disease pathologies reported in the cat, the clinical and physiological study of these feline hereditary diseases provides a strong comparative medicine opportunity for prevention, diagnostics, and treatment studies in a laboratory setting. causal mutations have been characterized in felid genes, with the largest representation from lysosomal storage enzyme disorders. corrective therapeutic strategies for several disorders have been proposed and examined in the cat, including enzyme replacement, heterologous bone marrow transplantation, and substrate reduction therapy. genomics tools developed in the cat, including the recent completion of the -fold whole genome sequence of the cat and genome browser, radiation hybrid map of integrated coding and microsatellite loci, a -cm genetic linkage map, arrayed bac libraries, and flow sorted chromosomes, are providing resources that are being utilized in mapping and characterization of genes of interest. a recent report of the mapping and characterization of a novel causative gene for feline spinal muscular atrophy marked the first identification of a disease gene purely from positional reasoning. with the development of genomic resources in the cat and the application of complementary comparative tools developed in other species, the domestic cat is emerging as a promising resource of phenotypically defined genetic variation of biomedical significance. additionally, the cat has provided several useful models for infectious disease. these include feline leukemia and feline sarcoma virus, feline coronavirus, and type c retroviruses that interact with cellular oncogenes to induce leukemia, lymphoma, and sarcoma. mankind has held a centuries-long fascination with the cat. the earliest arch eological records that have been linked to the domestication of felis catus date to approximately years ago from cyprus, with recent molecular genetic analyses in our laboratory suggesting a middle eastern origin for domestication (c. driscoll et al., unpublished observations) . currently the most numerous of companion animals, numbering close to million in households across the united states (http://www.appma.org/ press_industrytrends.asp), the cat enjoys a medical surveillance second only to the dog and humankind. in this chapter we review the promise of the cat as an important model for the advancement of human hereditary and infectious disease and the genomic tools that have been developed for the identification, and characterization of genes of interest. for many years we have sought to characterize genetic organization in the domestic cat and to develop genomic resources that establish f. catus as a useful animal model for human hereditary disease analogues, neoplasia, genetic factors associated with host response to infectious disease, and mammalian genome evolution. , to identify genes associated with inherited pathologies that mirror inherited human conditions and interesting pheno-types in the domestic cat, we have produced genetic maps of sufficient density to allow linkage or association-based mapping exercises. [ ] [ ] [ ] [ ] [ ] [ ] [ ] [ ] the first genetic map of the cat, a physical map generated from a somatic-cell hybrid panel, demonstrated the cat's high level of conserved synteny with the human genome, which offered much promise for the future application of comparative genomic inference in felid mapping and association exercises. several radiation hybrid (rh) and genetic linkage (gl) maps have since been published. [ ] [ ] [ ] [ ] [ ] [ ] , , although previous versions of the cat gene map, based on somatic cell hybrid and zoo fish analysis, , revealed considerable conservation of synteny with the human genome, these maps provided no knowledge of gene order or intrachromosomal genome rearrangement between the two species, information that is critical to applying comparative map inference to gene dis covery in gene-poor model systems. radiation hybrid (rh) mapping has emerged as a powerful tool for constructing moderate-to high-density gene maps in vertebrates by obviating the need to identify interspecific polymorphisms critical for the generation of genetic linkage maps. the most recent rh map of the cat includes markers: coding loci, selected markers derived from the cat x whole genome sequence targeted at breakpoints in conserved synteny between human and cat, and short tandem repeat (str) loci. the strategy used in developing the current rh map was to target gaps in the feline-human comparative map, and to provide more definition in breakpoints in regions of conserved synteny between cat and human. the markers cover the length of the feline autosomes and the x chromosome at an average spacing of one marker every . mb (megabase), with fairly uniform marker density. an enhanced comparative map demonstrates that the current map provides % and % comparative coverage of the human and canine genomes, respectively. ninety-six percent of the cat markers have identifi able orthologues in the canine and human genome sequences, providing a rich comparative tool, which is critical in linkage mapping exercises for the identification of genes controlling feline phenotypes. figure - presents a graphic display of each cat chromosome and blocks of conserved syntenic order with the human and canine genomes. one hundred and fifty-two cat-human and cat-dog homologous synteny blocks were identified. alignment of cat, dog, and human chromosomes demonstrated different patterns of chromosomal rearrangement with a marked increase in interchromosomal rearrangements relative to human in the canid lineage ( % of all rearrangements), as opposed to the more frequent intrachromosomal rearrangements in the felid lineage ( % of all rearrangements) since divergence from a common carnivore ancestor my ago. with an average spacing of marker every . mb in the feline euchromatic sequence, the map provided a solid framework for the chromosomal assignment of feline contigs and scaffolds during assembly of the cat genome assembly, and served as a comparative tool to aid in the identification of genes controlling feline phenotypes. as a complement to the rh map of the cat, a third generation linkage map of strs is currently nearing completion. the map has been generated in a large multigeneration domestic cat pedigree (n = informative meioses). previous first-and second-generation linkage maps of the cat were generated in a multigeneration interspecies pedigree generated between the domestic cat and the asian leopard cat, prionailurus bengalensis, to facilitate the mapping and integration of type i (coding) and type ii (polymorphic str) loci. the current map, which spans all autosomes with single linkage groups, has twice the str density of previous maps, providing a -cm resolution. there is also greatly expanded coverage of the x chromosome, with some str loci. marker order between the current generation rh and gl maps is highly concordant. approximately % of the strs are mapped in the most current rh map of the cat, which provides reference and integration with type i loci. whereas the third-generation linkage map is composed entirely of str loci, the sequence homology of extended genomic regions adjacent to the str loci in the cat x whole genome sequence, to the dog's homologous region, has enabled us to obtain identifiable orthologues in the canine and human genome sequences for over % of the strs. thus, practically every str acts as a "virtual" type locus, with both comparative anchoring and linkage map utility. combined with the cat rh map, these genomic tools provide us with the comparative reference to other mammalian genomes critical for linkage and association mapping. the domestic cat is one of mammalian species endorsed by the national human genome research institute (nhgri) human genome annotation committee for a "light" -fold whole genome sequence, largely to capture the pattern of genome variation and divergence that characterizes the mammalian radiations (http:// www.hgsc.bcm.tmc.edu/projects/bovine/, http://www.broad.mit. edu/mammals/). although light genome coverage provides limited sequence representation, (∼ %), one of the rationales for these light genome sequences included "enhancing opportunities for research on species providing human medical models." the -fold assembly of the domestic cat genome has recently been completed for a female abyssinian cat, "cinnamon," and a x whole genome sequencing effort is planned in the near future. a total of , , reads were assembled to , contigs, covering . gb with an n (i.e., half of the sequenced base pairs are in contigs