key: cord-0545161-otwiadg3 authors: Skwark, Marcin J.; Carranza, Nicol'as L'opez; Pierrot, Thomas; Phillips, Joe; Said, Slim; Laterre, Alexandre; Kerkeni, Amine; cSahin, Uugur; Beguir, Karim title: Designing a Prospective COVID-19 Therapeutic with Reinforcement Learning date: 2020-12-03 journal: nan DOI: nan sha: d94ece79f0f598b6b8aa9e77bb2ffa0558313d49 doc_id: 545161 cord_uid: otwiadg3 The SARS-CoV-2 pandemic has created a global race for a cure. One approach focuses on designing a novel variant of the human angiotensin-converting enzyme 2 (ACE2) that binds more tightly to the SARS-CoV-2 spike protein and diverts it from human cells. Here we formulate a novel protein design framework as a reinforcement learning problem. We generate new designs efficiently through the combination of a fast, biologically-grounded reward function and sequential action-space formulation. The use of Policy Gradients reduces the compute budget needed to reach consistent, high-quality designs by at least an order of magnitude compared to standard methods. Complexes designed by this method have been validated by molecular dynamics simulations, confirming their increased stability. This suggests that combining leading protein design methods with modern deep reinforcement learning is a viable path for discovering a Covid-19 cure and may accelerate design of peptide-based therapeutics for other diseases. The ongoing worldwide Covid-19 pandemic is compelling researchers to explore multiple routes to develop a cure. While the creation of small molecule drugs remains a promising path [24, 31, 26] , another possibility is the design of a novel peptide that would bind to and neutralize the SARS-CoV-2 spike protein [13, 9] . To replicate, the virus needs to penetrate the cell by hijacking its active transport processes. It does so through a specific set of interactions with the cellular receptor. For SARS-CoV and SARS-CoV-2, it is the spike protein that mediates the cell infection through the binding of its receptor-binding domain (RBD) to the angiotensin converting enzyme 2 (ACE2) on the surface of the cell [28] . SARS-COV-2 only recently crossed the species barrier into humans and developed an affinity towards ACE2, meaning its binding is likely to be suboptimal. A new binder could be engineered to neutralize the SARS-CoV-2 virus by inhibiting its attachment process to the native ACE2. To do so, it would need to be more specific for SARS-COV-2, i.e. have a higher binding affinity than ACE2, while simultaneously remaining sufficiently similar to the native variant in order to be safe for humans. The resulting binder would keep SARS-COV-2 outside the human cells, making it more visible for recognition by the immune system. The interface between ACE2 and the SARS-COV-2 spike protein consists of 34 amino acids, with 20 residues possible at each position. Designing the interface of a novel variant of ACE2 therefore results in a search space over amino acids in the order of 20 34 , or equivalently, 10 45 . To explore this research direction, we have developed a novel protocol, which given a protein complex structure, optimizes the binding interface. This is achieved through the combination of computational structural biology and deep reinforcement learning (RL). The protocol maintains the stability of the interacting partners, and optimizes binding while minimizing the number of redesigned positions, leading to designs that are more stable, and due to minimal, unobtrusive quality of the design, less likely to either misfold, or elicit an immune reaction in vivo. While it is not realistic to directly estimate the binding affinity between interacting partners, one can relatively well approximate the free energy of the system. We leverage the industry standard methodology implemented in the ROSETTA suite [2, 6] to evaluate quality of binding through a combination of standard terms present in the ROSETTA energy function. To the best of our knowledge, the only work applying Reinforcement Learning to protein design can be found in [3] . The authors propose an RL formulation, but postulate that the evaluation cost of the scoring function hinders its applicability. Consequently, to improve the sample efficiency, they developed a model-based approach. The scoring function is approximated using an ensemble of models of various nature and capacities. This reduces the estimation errors, that could have been exploited by the agent. We differ from the resulting method (DynaPPO) not only in terms of the problem formulation -more expressive action space and initial state space, which increase sample efficiency -but also in terms of the scoring function, which balances accuracy and speed, obviating the need for approximation with ensemble models. We formulate the design of high affinity binder proteins as a Markov Decision Process (MDP) (S, A, T , R, γ) where S is the state space, A the action space, T : S × A → S the transition function, R the reward signal and γ a discount factor. One episode corresponds to the design of a protein. An environment state represents a protein, i.e. a sequence of amino acids s = s 1 , s 2 , . . . , s n , where s i refers to the residue at position i and n is the length of the complete sequence. Each amino acid is chosen among a set of m = 20 possible proteinogenic types. A possible formulation (e.g. [3] ), is to initialize each episode with a blank sequence and then generate each amino acid progressively. In this case, the agent chooses n times between m actions during each episode. At the end of an episode, the final protein obtained is scored. The resulting reward signal is sparse, i.e non-zero only at the end of an episode. The reward is normalized to the interval R ∈ (−1.0, 1.0], with ACE2 set as a baseline at R ACE2 = 0. In our work, we refer to this formulation as the generative approach. In this study, we propose another formulation: the iterative approach. In this setting, the agent begins each episode with an initial sequence. The agent then mutates the sequence by selecting positions to change, as well as the new residues to change them to, in such a way that there is always a valid sequence to be scored. As a result, the mutation process can be stopped at any moment, potentially by the agent itself (see figure 1 ). The choice of initial sequence can be arbitrary, though we choose either the native human ACE2 protein sequence, or randomly perturbed versions thereof. Prior initialization conditioned on ACE2 enables the agent to prioritize exploration around the subset of states that represent close mutations of the native ACE2 protein sequence, which is the only protein known to bind to SARS-COV-2 with certainty. Contrary to [3] , our target function is not learned, but computed directly from simulation. Proteinprotein interactions are driven by a large set of intertwined interactions, which are practically impossible to estimate from protein sequence space alone. Our scoring function therefore estimates the change of Gibbs free energy resulting from the action chosen by the agent. This is computed through the combination of knowledge-based statistical potential [2] , and prior knowledge of experimentally observed protein complex conformations derived from Protein Data Bank [5] , as well as binding target variability [10] . The score reflects the binding-oriented problem formulation, which expressly does not capture the entire complexity of the physiology of virus adhesion to the cell. Reinforcement Learning Algorithms For both formulations of the protein design problem, we make use of policy gradient methods [23] to directly optimize a stochastic policy parametrized by vector θ, π θ : S → D A where D A is the space of distributions over the action space A. The iterative formulation defined in section 3, also allows the problem to be solved using sequential action spaces, i.e. the agent must select not one action, but a sequence of p actions. At each step, the agent must first choose a position in the sequence, then select an amino acid to place, given the chosen position, and finally, it must decide whether to take the action to stop the design. When the agent chooses to stop, the episode terminates and the resulting sequence is scored. At each time step, the three actions are selected autoregressively, each one depending on the choices of the previous ones. The factorization of the action space helps to reduce the number of parameters of the policy network. It also improves decision interpretability and eases the learning by splitting one potentially extremely large distribution into several smaller conditional distributions. Empirically, it displays improved exploration capacities as well. We implement the policy gradient method using Proximal Policy Optimization (PPO) [20] , and extend PPO to the sequential setting for the sequential action space configurations. We call the resulting algorithm Sequential PPO (SPPO). We show that SPPO learns in a few time-steps to start from random perturbations of the ACE2 sequence and make only a few changes in its amino acids to obtain good scores. Neural network architecture Attention-based model architectures, such as Transformers [25] have recently achieved state-of-the-art results in NLP across various tasks [8, 30, 7] . In our work, we leverage the ability of transformer models to process sequential information and extract long-range dependencies. The transformer backbone is followed by the appropriate heads, e.g. Policy and Value heads for (S)PPO, as shown in figure 3 . A complete description of the network architecture alongside its hyper-parameters and the input specifications can be found in section A.1 of the supplementary materials. On both formulations of the protein design problem, the RL agents exceed the performance of a set of benchmark methods based on simulated annealing, Monte-Carlo sampling, and random sampling. These results are shown in Figure 2 and in Appendix A.4. For the iterative formulation of the problem, we note that the process of mutating sequences initialized from native ACE2 introduces a source of stability. This formulation reduces the variance in the mean episodic reward by two orders of magnitude, compared to the generative formulation of the problem. While both PPO and SPPO achieve comparable scores, SPPO yields candidate designs above the threshold of native ACE2 in a much more sample efficient manner. The factorization of the action space into sequential conditional policies enables more informative gradient updates during training, compared to those received in the joint action space configuration implemented by PPO. Indeed, autoregressive policies can offer more consistent exploration in RL tasks, especially when there are sparse rewards [15] . This results in a 5-10x reduction in the number of steps required to surpass the native ACE2 score. . Native ACE2 is benchmarked with a score of RACE2 = 0. In all cases, RL agents learn to design superior binders than the native ACE2 and outperform baseline agents based on sampling. We compared the best-scoring designs of our method to designs previously reported in [13, 9] . We applied the industry-standard ROSETTA remodel [12] method to the same reference structures, and optimize the same positions as in this work. To ensure a conservative comparison with our results, we devoted ten times more CPU hours to the Rosetta designs than recommended by [11] , starting the designs from multiple, alternative conformations, including those derived from experimental structures, those used internally in our solution, and their variants obtained by conformational perturbation and minimization. While this protocol diverges from state-of-art ROSETTA approaches, it allows us to estimate the computational effort needed to tackle such a problem with no external expert input. Therefore we claim that, given the typical computation time recommended for such tasks, reinforcement learning delivers better results with greater diversity than traditional methods (see Table 3 ). Knowledge-based assessment For each design candidate (either ours, published, or generated by ROSETTA), we construct a pool of structural models. Each model is subject to the ROSETTA InterfaceAnalyser protocol [22] . This shows that our designs both bind ACE2 (see annex B.1), and retain sufficient similarity to human ACE2 receptor, to evade recognition by immune system, and to make them plausible as drop-in therapeutics. In terms of the Gibbs free energy (δG) of binding, our designs, proposed in an entirely automated manner, prove to be substantially more competent binders than human ACE2. We note, that it is our method that proposed the best performing design in terms of δG in the entire benchmark. The design proposed in [13] outperforms us in terms of EvoEF2 and RW+ metrics, however these are the metrics authors have been explicitly optimizing, while we have not been optimizing for binding explicitly. The analysis above relies on a knowledge-based potential and, as such, involves theoretically optimized structures, which can lead to overfitting. To assess the binding performance of newly designed proteins, we run a series of unrestrained, molecular dynamics simulations using the industry standard, GPU-aided GROMACS 2018.4 [1, 17, 19] software package. In a 50ns simulation (on average over 5650 core-hours per simulation) the native human ACE2 does not bind particularly strongly to SARS-COV-2 RBD. After 15ns of simulation time, partners drift apart, to fully disconnect by 20 ns (see Figure 5) . We see evidence of complex reformation by the end of the trajectory, as partners come back together. Our designs do not exhibit such a behavior and are more stable, as measured by the trajectory RMSD to the bound conformation. Even if started from suboptimally bound pose, partners with newly designed interfaces quickly form an transitory complex and remain stable throughout the simulation. The analyses of molecular dynamics trajectories indicate that our designed proteins bind SARS-COV-2 RBD in a stronger, more affine manner. They result in much more stable complexes as well. This achieves the goal of tight binding for coronavirus neutralization. In this paper we introduce a novel framework for protein design by combining reinforcement learning with computational structural biology and apply it to develop a candidate cure for Covid-19. First, we build a fast reward scoring function that leverages insights from co-crystal structures of the SARS-COV-2 spike protein complexed with the human ACE2. Second, we experiment with several RL algorithms and demonstrate convergence. In particular, we design a sequential version of the Proximal Policy Optimization algorithm for the protein design task, which we name Sequential PPO (SPPO). SPPO displays improved sample efficiency compared to standard PPO and quickly obtains promising ACE2-based protein binders. Third, we subject our best designs to a range of evaluations and report favourable in silico knowledge-based metrics, including for metrics not used in our scoring protocol. Our designs are competitive with those published by well-established groups in computational structural biology. Finally, full scale molecular dynamics simulations confirm that our candidate therapeutics are more stable than the native human ACE2 and bind tightly to the SARS-COV-2 RBD, suggesting our candidate cures are likely to perform their intended task in vivo. We give more details about the algorithms we use in our methods, about the baselines we compare to and about our experimental details in the following sections 2 . The input of the neural network is a n × f sequence, consisting of f = 10 features for each amino acid 3 . Those features form the amino acid vocabulary and correspond to the Atchley factors [4] , a small set of highly interpretable numeric patterns of amino acid variability. The sequence containing n amino acids is embedded as a sequence of n corresponding vectors in R d , where d = 512. Each vector is the sum of a learned token embedding and a sinusoidal positional embedding as in the original Transformer architecture [25] . The embedded sequence is then processed by a GTrXL model [18] , comprising 5 blocks, each composed of a self-attention operation followed by a position-wise multi-layer network. Each block preserves the embedding dimension of each sequence element. The associated hyper-parameters can be found in table 1 and as well as those defining the appropriate policy and value heads for PPO and SPPO. A schema is shown in figure 3 with the policy and value heads for PPO. Figure 3 : Protein sequence design schema of the reinforcement learning agent and environment. A key aspect of this work is that the reward function can directly output scores from ROSETTA simulations, relatively quickly: in less than 1 minute and without the need of a proxy. Each design is evaluated in a series of biologically plausible scenarios, which reflect the current state of knowledge of SARS-COV-2 RBD-ACE2 structures [27] , [16] , [29] , [21] and prior knowledge on the biochemically relevant interactions between partners. This comprises effective use of diverse structural information, leveraging experimental information on evolutionary constraints on the diversity of the partners, as well as the data on the genetic diversity of the population. The function is designed to be progressively more accurate and has a built-in early stop criteria. Therefore, designs that are highly unlikely to result in high scores, are evaluated in a faster and more cursory manner, while the bulk of computational resources are devoted to fine-grained evaluation of prospectively successful designs. The reward function is fully deterministic and provided the same score given the same input. This permitted for in-memory reward caching. For the memory cache, we use Redis, an in-memory key-value database. Depending on the RL algorithm, this allowed speedups of up to 3 times. We ran our simulations in machines ranging from 96 cores to 224 cores. In these conditions, an average distributed reward calculation would take around 1 minute for our ROSETTA configuration. If the score was previously calculated the cache would provide the reward in a few milliseconds. gen -0.020 -0.12 --- Table 2 : Summary statistics for RL algorithms compared to benchmark methods. Rewards displayed for R ∈ (−1.0, 1.0]. Native ACE2 benchmark score is 0: R ACE2 = 0. All supplied RL figures are averaged over 3 random seeds. Simulated annealing performed with temperature 1% and 10%, best results shown here (1%). We note in table 2, the superior performance of the RL methods over the benchmark methods, as measured by R max and R avg . The extension of PPO-generative to SPPO-iterative reduces the number of steps required to generate candidate protein sequences that outperform the native ACE2 at binding to SAR-CoV-2. Each of the tables represents the same designs, in the same order, scored according to diverse criteria, scored per each assessment target, and in terms of a mean of all the columns. Relative binding affinity: δG per 100 Å 2 Method Measures of binding energy: δG for the best design (Best) and mean of top 20% designs in terms of our selection protocol (Good) indicate high affinity. Energy per residue [22] is an estimate of the overall quality of the model. The evoEF2 metric is an independent estimate of binding energy [14] . With respect to every metric (except for energy per residue, which can be alleviated by further minimization), our designs outperform the other methods, despite not optimizing any of them explicitly. Native structure starts from an experimentally determined, bound position, promptly disassociates from the protein and intermittently explores binding. In the bottom row, we demonstrate unrestrained binding of the designed protein, which starts nearly unbound, that is is translated away from the complex and rotated by a random small angle, then mimimized as a soluble monomer and translated back to contact with the receptor. The new design starts forming biochemical contacts, and finally forms a stable complex with the spike protein. Note that the simulation of the native complex cannot be expected to be typical and is provided here for illustrative purpose only. Figure 6 : Interactions between chains in the native structure (left figures) and one of the good designs (right figures). Heat maps depict contact maps of all interacting residues, protein structures visualise the 3D interactions. Blue interactions are favorable and have negative energy, red -unfavorable. Not only does the new design bind tighter, with more residues involved in binding, but also it has fewer detrimental interactions. For each of the methods, we have selected one design, assessed as the most suitable one by our methodology (that is having the best binding and stability). On each of these structures, we have performed a single run of molecular dynamics simulations. These simulations have been performed using GROMACS 2018.4 [1, 17, 19] , with an atomistic AMBER ff03 as a force field. Each of the low-energy poses of a new design based on native ACE2spike protein co-crystal structure has been fully solvated (with a three point water model -SPC/E) in a cubic box with 20Åmargin. The choice of simulation parameters has not been benchmarked, but each of them is a rational choice, according to the literature. We do not expect that the results of simulation will be radically different, were we to make an other choice. Each of the solvated systems has been minimized for the 1 ns (50,000 steps of 2 fs each), equilibrated with respect to temperature and pressure for 1 ns each (with a modified Berendsen thermostat and Parinello-Raman pressure coupling, respectively). Finally we simulated the system unrestrained for 200 ns, in line with the simulations previously reported in literature [9] . The simulations reveal (see Figure 7 , that each of the trajectories is largely stable, with one of [13] starting in a potentially frustrated state, but promptly settling into an energetically favorable pose. However, we have noticed that the pose based on the native state starts with a large amount of potential energy, which is approximately 40% into the simulation converted into kinetic energy. This corresponds to the complex disassociating. The final 10% of simulation, where both potential and kinetic energy are low, illustrates the complex forming anew. From molecular dynamics simulations, we can conclude, that each of the designs, be it proposed by our method, or by others, is biophysically feasible and can result in a high-affinity binding. Figure 7 : Molecular dynamics. The plots illustrate the development of the energy of a simulated system. As each of the systems comprises different number of atoms, with differently shaped periodic box, we have offset each trajectory by its median value, thus 0 corresponds to the median value of the observed energy. Large fluctuations represent a frustrated, dynamic system, low -a system at equilibrium. We note, that -with exception of the native structure -every design appears to settle in a comfortable energy well. Gromacs: High performance molecular simulations through multi-level parallelism from laptops to supercomputers The rosetta all-atom energy function for macromolecular modeling and design Model-based reinforcement learning for biological sequence design Solving the protein sequence metric problem The protein data bank Pyrosetta: a script-based interface for implementing molecular modeling algorithms using rosetta Universal transformers BERT: pre-training of deep bidirectional transformers for language understanding Computational design of ace2-based peptide inhibitors of sars-cov-2 Virus variation resource-improved response to emergent viral outbreaks Protein origami: Function follows form Rosettaremodel: a generalized framework for flexible backbone protein design Computational design of peptides to block binding of the sars-cov-2 spike protein to human ace2. bioRxiv Evoef2: Accurate and fast energy function for computational protein design Autoregressive Policies for Continuous Control Deep Reinforcement Learning Structure of the sars-cov-2 spike receptor-binding domain bound to the ace2 receptor Tackling exascale software challenges in molecular dynamics simulations with gromacs Seb Noury, et al. Stabilizing transformers for reinforcement learning Gromacs 4.5: a high-throughput and highly parallel open source molecular simulation toolkit Proximal policy optimization algorithms Structural basis of receptor recognition by sars-cov-2 A comparison of successful and failed protein interface designs highlights the challenges of designing buried hydrogen bonds Policy gradient methods for reinforcement learning with function approximation Rapid identification of potential inhibitors of sars-cov-2 main protease by deep docking of 1.3 billion compounds Attention is all you need Fast identification of possible drug treatment of coronavirus disease-19 (covid-19) through computational drug repurposing study Structural and functional basis of sars-cov-2 entry by using human ace2 Cryo-em structure of the 2019-ncov spike in the prefusion conformation Structural basis for the recognition of sars-cov-2 by full-length human ace2 Xlnet: Generalized autoregressive pretraining for language understanding Network-based drug repurposing for novel coronavirus 2019-ncov/sars-cov-2 MethodAbsolute binding affinity: evoEF2Method