key: cord-337897-hkvll3xh authors: Yang, Zheng Rong title: Peptide Bioinformatics- Peptide Classification Using Peptide Machines date: 2009 journal: Artificial Neural Networks DOI: 10.1007/978-1-60327-101-1_9 sha: doc_id: 337897 cord_uid: hkvll3xh Peptides scanned from whole protein sequences are the core information for many peptide bioinformatics research subjects, such as functional site prediction, protein structure identification, and protein function recognition. In these applications, we normally need to assign a peptide to one of the given categories using a computer model. They are therefore referred to as peptide classification applications. Among various machine learning approaches, including neural networks, peptide machines have demonstrated excellent performance compared with various conventional machine learning approaches in many applications. This chapter discusses the basic concepts of peptide classification, commonly used feature extraction methods, three peptide machines, and some important issues in peptide classification. Proteins are the major components for various cell activities, including gene transcription, cell proliferation, and cell differentiation. To implement these activities, proteins function only if they interact with each other. Enzyme catalytic activity, acetylation, methylation, glycosylation, and phosphorylation are typical protein functions through binding. Studying how to recognize functional sites is then a fundamental topic in bioinformatics research. As proteins function only when they are bound together, the binding sites (functional sites) and their surrounding residues in substrates are the basic components for functional recognition. Studying consecutive residues around a functional site within a short region of a protein sequence for functional recognition is then a task of peptide classification that aims to find a proper model that maps these residues to functional status. A peptide classification model can be built using a properly selected learning algorithm with similar principles to those used in pattern classification systems. The earlier work was to investigate a set of experimentally determined (synthesized) functional peptides to find some conserved amino acids, referred In protease cleavage site prediction, we commonly use peptides with a fixed length. In some cases, we may deal with peptides with variable lengths. For instance, we may have peptides whose lengths vary from one residue to a couple of hundreds residues long when we try to identify disorder regions (segments) in proteins [ 5 ] . Figure 9 .2 shows two cases, where the shaded regions indicate some disorder segments. The curves represent the probability of disorder for each residue. The dashed lines show 50% of the probability and indicate the boundary between order and disorder. It can be seen that the disorder segments have variable lengths. The importance of accurately identifying disorder segments is related to accurate protein structure analysis. If a sequence contains any disorder segments, X-ray crystallization may fail to give a structure for the sequence. It is then critically important to remove such disordered segments in a sequence before the sequence is submitted for X-ray crystallization experiments. The study of protein secondary structures also falls into the same category as disorder segment identification, where the length of peptides that are constructs for secondary structures varies [ 6 ] . Because the basic components in peptides are amino acids, which are nonnumerical, we need a proper feature extraction method to convert peptides to numerical vectors. The second section of this chapter discusses some commonly used feature extraction methods. The third section introduces three peptide machines for peptide classification. The fourth section discusses some practical issues for peptide classification. The final section concludes peptide classification and gives some future research directions. Currently, three types of feature extraction methods are commonly used for converting amino acids to numerical vectors: orthogonal vector, frequency estimation, and bio-basis function methods. The orthogonal vector method encodes each amino acid using a 20-bit long binary vector with 1 bit assigned a unity and the rest zeros [ 8 ] . Denoted by s a peptide, a numerical vector generated using the orthogonal vector method is x Î b 20×| s | , where b = {0, 1} and | s | is the length of s . The introduction of this method greatly eased the difficulty of peptide classification in the early days of applying various machine learning algorithms like neural networks to various peptide classification tasks. However, the method significantly expands the input variables, as each amino acid is represented by 20 inputs [ 9 , 10 ] . For an application with peptides ten residues long, 200 input variables are needed. Figure 9 .3 shows such an application using the orthogonal vector method to convert amino acids to numerical inputs. It can be seen that data significance, expressed as a ratio of the number of data points over the number of model parameters, can be very low. Meanwhile, the method may not be able to properly code biological content in peptides. The distance (dissimilarity) between any two binary vectors encoded from two different amino acids is always a constant (it is 2 if using the Hamming distance or the square root of 2 if using the Euclidean distance), while the similarity (mutation or substitution probability) between any two amino acids varies [ 11 -13 ] . The other limitation with this method is that it is unable to handle peptides of variable lengths. For instance, it is hard to imagine that this method could be used to implement a model for predicting disorder segments in proteins. An example of using the orthogonal vector for converting a peptide to numerical vector for using feedforward neural networks, where each residue is expanded to 20 inputs. Adapted from [ 14 ] The frequency estimation is another commonly used method in peptide classification for feature extraction. When using single amino acid frequency values as features, we have the conversion as S ‫ۋ‬ x Î ᑬ 20 , meaning that a converted vector always has 20 dimensions. However, it has been widely accepted that neighboring residues interact. In this case, di-peptide frequency values as features may be used, leading to the conversion as S ‫ۋ‬ x Î ᑬ 420 . If tri-peptide frequency values are used, we then have the conversion as S ‫ۋ‬ x Î ᑬ 8420 . This method has been used in dealing with peptides with variable lengths for peptide classification, for instance, secondary structure prediction [ 15 ] and disorder segment prediction [ 5 ] . One very important issue of this method is the difficulty of handling the large dimensionality. For any application, dealing with a data set with 8,420 features is no easy task. The third method, called the bio-basis function , was developed in 2001 [ 9 , 10 ] . The introduction of the bio-basis function was based on the understanding that nongapped pairwise homology alignment scores using a mutation matrix are able to quantify peptide similarity statistically . Figure 9 .4 shows two contours where one functional peptide (SKNYPIVQ) and one nonfunctional peptide (SDGNGMNA) are selected as indicator peptides. We use the term indicator peptides , because the use of some indicator peptides transform a training peptide space to an indicator or feature space for modeling. We then calculate the nongapped pairwise homology alignment scores using the Dayhoff mutation matrix [ 12 ] to calculate the similarities among all the functional/positive and these two indicator peptides as well as the similarities among all the nonfunctional/negative peptides and these two indicator peptides. It can be seen that positive peptides show larger similarities with the positive indicator peptide (SKNYPIVQ) from the left panel, while negative peptides show larger similarities with the negative indicator peptide (SDGNGMNA) from the right panel. We denote by s a query peptide and by r i the i th indicator peptide and each has d residue, the bio-basis function is defined as K(s, r ) = exp (s, r ) (r , r ) Note that s j and r ij are the j th residues in the query and indicator peptides, respectively. The value of M ( s j , r ij ) can be found from a mutation matrix. Figure 9 .5 shows the Dayhoff mutation matrix [ 12 ] . From Eq. 1 , we can see that K ( s , The equality occurs if and only if s = r i . The basic principle of the bio-basis function is normalizing nongapped pairwise homology alignment scores. As shown in Fig. 9 .6 , a query peptide ( IPRS ) is aligned with two indicator peptides ( KPRT and YKAE ) to produce two nongapped pairwise homology alignment scores a ( å 1 = 24 + 56 + 56 + 36 = 172) and b ( å 2 = 28 + 28 + 24 + 32 =112 ) respectively. Because a > b , it is believed that the query peptide is more likely to have the same functional status (functional or nonfunctional) as the first indicator peptide. After the conversion using the bio-basis function, each peptide s is represented by a numerical vector x Î ᑬ m or a point in an m -dimensional feature space, where m is the number of indicator peptides. Note that m = 2 in Fig. 9 .6 . The bio-basis function method has been successfully applied to various peptide classification tasks, for instance, the prediction of trypsin cleavage sites [ 9 ] , the prediction of HIV cleavage sites [ 10 ] , the prediction of hepatitis C virus protease cleavage sites [ 16 ] , the prediction of the disorder segments in proteins [ 7 , 17 ] , the prediction of protein phosphorylation sites [ 18 , 19 ] , the prediction of the O-linkage sites in glycoproteins [ 20 ] , the prediction of signal peptides [ 21 ] , the prediction of factor Xa protease cleavage sites [ 22 ] , the analysis of mutation patterns of HIV-1 Fig. 9 . 5 Dayhoff matrix. Each entry is a mutation probability from one amino acid (in rows) to another amino acid (in columns). Reproduced from [ 14 ] with permission drug resistance [ 23 ] , the prediction of caspase cleavage sites [ 24 ] , the prediction of SARS-CoV protease cleavage sites, [ 25 ] and T-cell epitope prediction [ 26 ] . As one class of machine learning approaches, various vector machines are playing important roles in machine learning research and applications. Three vector machines-the support vector machine [ 27 , 28 ] , relevance vector machine [ 29 ] , and orthogonal vector machine [ 30 ] -have already drawn the attention for peptide bioinformatics. Each studies pattern classification with a specific focus. The support vector machine (SVM) searches for data points located on or near the boundaries for maximizing the margin between two classes of data points. These found data points are referred to as the support vectors . The relevance vector machine (RVM), on the other hand, searches for data points as the representative or prototypic data points referred to as the relevance vectors . The orthogonal vector machine (OVM) searches for the orthogonal bases on which the data points become mutually independent from each other. The found data points are referred to as the orthogonal vectors . All these machines need numerical inputs. We then see how to embed the bio-basis function into these vector machines to derive peptide machines. The support peptide machine (SPM) aims to find a mapping function between an input peptide s and the class membership (functional status). The model is defined as The bio-basis function. As IPRS is more similar to KPRT than YKAE , its similarity with KPRT is larger that that with YKAE , see the right figure. Reproduced from [ 14 ] with permission where w is the parameter vector, y = f ( s , w ) the mapping function, and y the output corresponding to the desired class membership t Î {-1, 1}. Note that -1 and 1 represent nonfunctional and functional status, respectively. With other classification algorithms like neural networks, the distance (error) between y and t is minimized to optimize w . This can lead to a biased hyperplane for discrimination. In Fig. 9 .7 , there are two classes of peptides, A and B . The four open circles of class A and four filled circles of the class B are distributed in balance. With this set of peptides, the true hyperplane separating two classes of peptides, represented as circles, can be found as in Fig. 9 .7(a) . Suppose a shaded large circle belonging to class B as a noise peptide is included, as seen in Fig. 9 .7(b) ; the hyperplane (the broken thick line) is biased because the error (distance) between the nine circles and the hyperplane has to be minimized. Suppose a shaded circle belonging to class A as a noise peptide is included as seen in Fig. 9 .7(c) ; the hyperplane (the broken thick line) also is biased. With theses biased hyperplanes, the novel peptides denoted by the triangles could be misclassified. In searching for the best hyperplane, the SPM finds the set of peptides that are the most difficult training data points to classify. These peptides are referred to as support peptides . In constructing a SPM classifier, the support peptides are closest to the hyperplane within the slat formed by two boundaries ( Fig. 9 .7d ) and are located on the boundaries of the margin between two classes of peptides. The advantage of using an SPM is that the hyperplane is searched through maximizing this margin. Because of this, the SPM classifier is robust; hence, it has better generalization performance than neural networks. In Fig. 9 .7(d) , two open circles on the upper boundary and two filled circles on the lower boundary are selected as support peptides. The use of these four circles can form the boundaries of the maximum margin between two classes of peptides. The trained SPM classifier is a linear Fig. 9 .7 a Hyperplane formed using conventional classification algorithms for peptides with a balanced distribution. b and c Hyperplanes formed using conventional classification algorithms for peptides without balanced distribution. d Hyperplane formed using SPM for peptides without balanced distribution. The open circles represent class A , the filled circles class B , and the shaded circle class A or B . The thick lines represent the correct hyperplane for discrimination and the broken thick lines the biased hyperplanes. The thin lines represent the margin boundaries. The γ indicates the distance between the hyperplane and the boundary formed by support peptides. The margin is 2γ. Reproduced from [ 14 ] with permission combination of the similarity between an input peptide and the found support peptides. The similarity between an input peptide and the support peptides is quantified by the bio-basis function. The decision is made using the following equation, y = sign {Σ w i t i K ( s , r i )}, where t i is the class label of the i th support peptide and w i the positive parameter as a weight for the i th support peptide. The weights are determined by a SPM algorithm [ 31 ] . The relevance peptide machine (RPM) was proposed based on automatic relevance determination (ARD) theory [ 32 ] . In SPM, the found support peptides are normally located near the boundary for discrimination between two classes of peptides. However, the relevance peptides found by RPM are prototypic peptides. This is a unique property of RPM, as the prototypic peptides found by a learning process are useful for exploring the biological inside. Suppose the model output is defined as where T is a similarity vector and w = ( w 1 , w 2 , . . ., w n ) T . The model likelihood function is defined as follows: Using ARD prior to computation can prevent overfitting the coefficients by assuming each weight follows a Gaussian where α = (α 1 , α 2 , . . . , α n ) T . The Bayesian learning method gives the following posterior for the coefficients: where the parameters (covariance matrix Σ , the mean vector u , and the hyperparameters for weights S ), which can be approximated: and The marginal likelihood can be obtained through integrating out the coefficients: In learning, α can be estimated in an iterative way: A formal learning process of the RPM is then conducted. The following condition is checked during learning: where θ is a threshold. If the condition is satisfied, the corresponding expansion coefficient is zeroed. The learning continues until either the change of the expansion coefficients is small or the maximum learning cycle is approached [ 33 ] . Figure 9 .8 shows how the RPM selects prototypic peptides as the relevance peptides, compared with the SPM, which selects boundary peptides as the support peptides. First, we define a linear model using the bio-basis function as y = Kw (16) where K is defined in Eq. 10 . The orthogonal least square algorithm is used as a forward selection procedure by the orthogonal peptide machine (OPM). At each step, the incremental information content of a system is maximized. We can rewrite the design matrix K as a collection of column vectors ( k 1 , k 2 , . . . , k m ), where k i represents a vector of the similarities between all the training peptides and the i th The OPM involves the transformation of the indicator peptides ( r i ) to the orthogonal peptides ( o i ) to reduce possible information redundancy. The feature matrix K is decomposed to an orthogonal matrix and an upper triangular matrix as follows: where the triangular matrix U has ones on the diagonal line: The orthogonal matrix satisfies where H is diagonal with the elements h ii as The space spanned by the set of orthogonal peptides is the same space spanned by the set of indicator peptides, and Eq. 16 can be rewritten as We can define an error model as follows: Suppose e ~ N (0, 1), the pseudoinverse method can be used to estimate a as follows: Because H is diagonal, The element in a is then The quantities a and w satisfy the triangular system The Gram-Schmidt or the modified Gram-Schmidt methods are commonly used to find the orthogonal peptides then to estimate w [ 30 , 34 ] . HIV/AIDS is one of the most lethal and transmittable diseases, with a high mortality rate worldwide. The most effective prevention of HIV infection is to use a vaccine to block virus infection [ 34 ] . However, HIV vaccine is difficult to develop because of the expense and complexity in advancing new candidate vaccines. Although some efficient models and integrated HIV vaccine research enterprises have been proposed [ 36 ] , there is little hope that an HIV vaccine would be developed before 2009 [ 37 ] . HIV is a type of retrovirus. It can enter uninfected cells to replicate itself. Inhibiting viral maturation and viral reverse transcription are then the major methods so far for treating HIV-infected patients. Two groups of inhibitors have since been developed. One group aims to stop protease cleavage activities and is referred to as the protease inhibitor . The other aims to stop reverse transcriptase cleavage activities and is referred to as the reverse transcriptase inhibitor . However, HIV often develops resistance to the drugs applied. Drug-resistant mutants of HIV-1 protease limit the long-term effectiveness of current antiviral therapy [ 38 ] . The emergence of drug resistance remains one of the most critical issues in treating HIV-1-infected patients [ 39 ] . The genetic reason for drug resistance is the high mutation rate along with a very high replication rate of the virus. These two factors work together, leading to the evolution of drug-resistant variants and consequently resulting in therapy failure. The resistance to HIV-1 protease inhibitors can be analyzed at a molecular level with a genotypic or a phenotypic method [ 39 , 40 ] . The genotypic method is used to scan a viral genome for some mutation patterns for potential resistance. As an alternative, the viral activity can be measured in cell culture assays. Yet another alternative, phenotypic testing is based on clinic observations, ithat is, directly measuring viral replication in the presence of increasing drug concentration. Genotypic assays have been widely used as tools for determining HIV-1 drug resistance and for guiding treatment. The use of such tools is based on the principle that the complex impact of amino acid substitutions in HIV reverse transcriptase or protease on the phenotypic susceptibility or clinical response to the 18 available antiretroviral agents is observable [ 41 ] . Genotypic assays are used to analyze mutations associated drug resistance or reduced drug susceptibility. However, this method is problematic, because various mutations and mutational patterns may lead to drug resistance [ 39 ] , and therefore the method occasionally fails to predict the effects of multiple mutations [ 42 ] . In addition, genotypic assays provide only indirect evidence of drug resistance [ 43 ] . Although HIV-1 genotyping is widely accepted for monitoring antiretroviral therapy, how to interpret the mutation pattern associated with drug resistance to make accurate predictions of susceptibility to each antiretroviral drug is still challenging [ 44 ] . Phenotypic assays directly measure drug resistance [ 43 ] , where drug resistance can be experimentally evaluated by measuring the ratio of free drug bound to HIV-1 protease molecules. However, this procedure is generally expensive and time consuming [ 39 , 42 ] . To deal with the difficulties encountered in genotypic assay analysis and phenotypic evaluation, a combination of inhibitor flexible docking and molecular dynamics simulations was used to calculate the protein-inhibitor binding energy. From this, an inhibitory constant is calculated for prediction [ 39 ] . Later, some statistical models were established for predicting drug resistance. For instance, a statistical model was proposed to analyze the viral and immunologic dynamics of HIV infection, taking into account drug-resistant mutants and therapy outcomes [ 45 ] . The factors causing the increase in CD4 cell count and the decrease in viral load from baseline after six months of HAART (highly active antiretroviral therapy) were analyzed. Note that CD4 stands for cluster of differentiation 4, which is a molecule expressed on the surface of T helper cells. In addition to the baseline viral load and CD4 cell count, which are known to affect response to therapy, the baseline CD8 cell count and resistance characteristics of detectable strains are shown to improve prediction accuracy. Note that CD8 stands for cluster of differentiation 8, which is a membrane glycoprotein found primarily on the surface of cytotoxic T cells. A logistic regression model was built for the prediction of the odds of achieving virology suppression after 24 weeks of HAART in 656 antiretroviral-naive patients starting HAART according to their week 4 viral load [ 46 ] . A regression model was built on a set of 650 matched genotype-phenotype pairs for the prediction of phenotypic drug resistance from genotypes [ 47 ] . As it was observed that the range of resistance factors varies considerably among drugs, two simple scoring functions were derived from different sets of predicted phenotypes. The scoring functions were then used for discrimination analysis [ 48 ] . In addition, machine learning algorithms were used. Decision trees as a method for mimicking human intelligence were implemented to predict drug resistance [ 43 ] . A fuzzy prediction model was built based on a clinical data set of 231 patients failing highly active antiretroviral therapy (HAART) and starting salvage therapy with baseline resistance genotyping and virological outcomes after three and six months [ 49 ] . In the model, a set of rules predicting genotypic resistance was initially derived from an expert and implemented using a fuzzy logic approach. The model was trained using the virological outcomes data set, that is, Stanford's HIV drug-resistance database (Stanford HIVdb). Expert systems were also used for this purpose [ 41 ] . Neural networks as a type of powerful nonlinear modelers were utilized to predict HIV drug resistance for two protease inhibitors, indinavir and saquinavir [ 50 ] . Other than using sequence information for prediction, a structure-based approach was proposed to predict drug resistance [ 42 ] . Models of WT complexes were first produced from crystal structures. Mutant complexes were then built by amino acid substitutions in the WT complexes with subsequent energy minimization of the ligand (a small molecule binding to a protein or receptor) and PR (protease receptor) binding site residues. A computational model was then built based on the calculated energies. The data set studied was obtained from an online relational database, the HIV RT and protease database ( http://hivdb.stanford.edu ) [ 51 ] . The susceptibility of the data to five protease inhibitors were determined, saquinavir (SQV), indinavir (IDV), ritonavir (RTV), nelfinavir (NFV) and amprenavir (APV). We obtained 255 genotype-phenotype pairs for each of the protease inhibitor, except for APV, for which we obtained 112 pairs. Phenotypic resistance testing measures in-vitro viral replication of the wild type and the viral sample in increasing drug concentrations [ 52 ] . The resistance factor, defined as the ratio of 50% inhibitory concentration (IC 50 ) of the respective viral sample to IC 50 of the nonresistant reference strain, reports the level of resistance as the fold change in susceptibility to the drug as compared to a fully susceptible reference strain. While genotypic resistance testing is done by scanning the viral genome for resistance-associated mutations. The major and minor mutations observed in a resistant HIV protease can be seen in Figure 1 in [ 53 ] . HIV protease is an aspartyl protease with 99 residues. In this study, we considered major and minor mutation sites, obtaining a window size of 20 residues for each drug. We divided the sample set into two classes by attaching to each peptide a label 1 or -1 for resistant (functional) and susceptible (nonfunctional), respectively. This division depended on whether the resistance factor of a sample exceeded the cutoff value or not. Based on previously published datasets and studies, we assumed the cutoff value for the resistant factor of each sample with respect to each protease inhibitor to be 3.5. We used Matthews correlation coefficient [ 54 ] The larger the Matthews correlation coefficient, the better the model fits the data. If the value of the Matthews correlation coefficient is 1, it represents a complete correlation. If the value is 0, the prediction is completely random. If the value is negative, the prediction is on the opposite side of the target. The evaluation is carried out based on the simulation of fivefold cross-validation for each drug and each algorithm. Figure 9 .9 shows the comparison between vector machines and peptide machines. It can be seen that the peptide machines outperformed the vector machines. We also constructed neural networks models for this task, their performance is far worse than vector machines (data are not shown). In building a computer model for a real application, there are two important issues for model validation in addition to model parameter estimation. The first is how to evaluate a model. The second is how to select a model based on the evaluation. There are always some hyperparameters for determination when using neural networks or machine learning algorithms. For instance, the number of hidden neurons in feedforward neural networks is such a hyperparameter, and the determination of the optimal number of hidden neurons is not an easy task. Using the training data for the evaluation will not deliver meaningful result, as an optimization process can easily overfit the data, leading to poor generation capability. The generalization capability should measure how well a built model works with novel data. Based on this requirement, we then need to consider how to use the available data for proper model validation. There are three commonly used methods for this purpose: cross-validation, resampling, and jackknife. All these methods use the same principle, that the validation data must not involve any process of model parameter estimation. This means that the available data must be divided into two parts. One is for model parameter estimation, which is commonly referred to as training . The other is for model evaluation (validation) and selection. The difference between these three methods is the strategy used for the division of a given data set. With the resampling method, we normally randomly sample a certain percentage of data for training and the rest for validation. Such a process is commonly repeated many times. Suppose there are n data points and we repeat the sampling process for m times. There will be m validation models, each of which has different training and validation data with some possible overlap. Note that all these validation models use the same hyperparameters; for instance, they all use h hidden neurons. The parameters of the i th validation model are estimated using the i th training data set with k i < n data points. The i th validation model is validated on the i th validation data set with n -k i data points. Because we use different training data sets each time, it is expected that the parameters of the i th validation model will be different from those of the j th validation model. The validation performance of the i th validation model is certainly different from that of the j th validation model as well. We denote by ω i the evaluated performance for the i th validation model. The evaluation statistic of the model with designed hyperparameters can follow To determine the proper values for the hyperparameters so as to select a proper model, we can vary the values assigned to the hyperparameters. If we have g hyperparameters for selection, the selection will be taken in a g -dimensional space, where each grid is a combination of hyperparameters. Suppose we need to determine only the number of hidden neurons, we then have a series of m and s 2 . The best model can be selected through It should be noted that, for the resampling method, some data points may be used for multiple times in training or validation. In cross-validation, we normally randomly divide a data set into m folds. Each fold contains distinctive data points. If we denote by W i as the set of data points in i th fold, we have W i Ç W j = φ, meaning that two sets have no elements in common. Every time, we select one fold as the validation set and the remaining m -1 folds are used as the training set for model parameters estimate. Such a process is repeated for m times, until each fold has been used for validation once. This means that there are m validation models. Again, all these validation models use the same hyperparameters. The i th validation model is trained using the folds except for the i th fold and validated on the i th fold. The parameters of the i th validation model also is different from those of the j th validation model, and the validation performance of different validation models will vary. Note that each data point ise validated only once. We denote by When data size is not too large, one commonly prefers using the jackknife (often called leave-one-out cross-validation ) method. In using the jackknife method, we normally pick up one data point for validation and use the remaining data points for training. Such a process is repeated for n times until each data point has been exhausted for validation. This means that there are n validation models for n data points. Obviously, all these validation models use the same hyperparameters. The i th validation model is trained using all the data points except for the i th data point and validated on the i th data point. Equation 31 can be used for evaluation. The best model can be selected through the use of Eq. 30 . Model validation can help us evaluate models and select models with a proper setting of hyperparameters. However, such evaluation values cannot be regarded as the final model evaluation, which can be delivered to users. The evaluated performance on validation data may not be able to reflect the true performance, because the validation data has been used to tune hyperparameters. As a consequence, the evaluated performance on validation data commonly overestimates the true model accuracy. Based on this understanding, a proper peptide classification methodology must contain a blind test stage. This means that, after model evaluation and model selection, we may need to test the "best" model using another data set, which has never been used for estimating model parameters and tuning model hyperparameters or selecting models [ 7 ] . A very important issue must be discussed here, especially for peptide classification. Many peptide classification tasks deal with functional site predictions. Within each protein sequence, there might be a few functional sites, such as protease cleavage sites, protein interaction sites, or protein posttranslational modification sites. Based on the substrate size, we can extract a peptide around one functional site. Such a peptide is commonly regarded as a functional peptide. To conduct proper peptide classification, we have to have a set of nonfunctional peptides. Each nonfunctional peptide has no functional site at the desired residue(s). We normally use a sliding window with a fixed length to scan a protein sequence from the N-terminal to the C-terminal, one residue by one residue to generate no-functional peptides. A commonly used method is to combine both functional and nonfunctional peptides to produce a data set. Suppose we use the cross-validation method, the data set is then randomly divided into m folds for cross-validation. One then builds m validation models for model evaluation and model selection. Now a question arises. When we use such kind of models for testing on novel whole protein sequences, the performance is commonly not as expected. This means that the model is some where overfitted. If model parameter estimation follows a proper procedure, the most probable problem is the data used for training and validation. We mentioned previously that we must maintain the independence of a validation data set from a training data set. When we check the method as described, we can clearly see two interesting facts. First, a training peptide and a validation peptide may come from the same protein. If one protein has conserved amino acids, the validation peptides generated this way may not be independent of the training peptides. Second, a training peptide and a validation peptide may have many identical residues if they are extracted from neighboring sliding widows. Based on this analysis, we proposed to use protein-oriented validation [ 24 ] . Suppose we have n proteins, we may divide these n proteins into m folds. We then generate validation peptides using one fold and generate training peptides from the remaining folds. The validation models are constructed using the training peptides scanned from the sequences of the training proteins and verified on the validation peptides scanned from the sequences of the validation proteins. We always have an important issue when using machine learning approaches for peptide classification, That is, if the model is correct forever. The answer is no, as a peptide data set collected at a certain time can be far less than complete. Based on this understanding, the models built using an incomplete set of peptides may not be able to generalize well forever. When new experimentally determined peptides have been collected, the exiting models must be corrected so that they can conduct classification tasks well. In this section, we investigate an interesting peptide classification topic, where we can use a model built on published peptides for peptide classification, and we can use a model built on both published peptides and newly submitted sequences in NCBI (www.ncbi.nih.gov) for peptide classification. We found there is a significant difference between the two. The case studied in this session is about hepatitis C virus (HCV), which is a member of the flaviviridae family [ 55 ] and is the major agent of parenterally transmitted non-A/non-B hepatitis [ 56 , 57 ] . The nomenclature of Schechter and Berger [ 1 ] is applied to designate the cleavage sites on the peptides, P 6 -P 5 -P 4 -P 3 -P 2 -P 1 -P 1 ' -P 2 ' -P 3 ' -P 4 ' , the asymmetric scissile bond being between P 1 and P 1 ' . Two resources are available for modeling. First, research articles have published experimentally determined peptides, cleaved and noncleaved. Second, some databank like NCBI has collected many new submissions. Each submission contains a whole protein sequence with a number of experimentally determined cleavage sites. In terms of this, there are two strategies for modeling. If we believe that the published peptides are able to represent all the protease cleavage specificity, we can select indicator peptides from these published peptides for modeling. We refer to a model constructed this way as a Type-I model . In fact, viruses mutate very rapidly; hence, the published peptides may not be able to represent the cleavage specificity in the recent submissions to NCBI. We can then select more indicator peptides, which show more viral mutation information from the new sequences downloaded from NCBI to build a more informative model referred to as a Type-II model. From published papers (data are not shown), we collected 215 experimentally determined peptides. They are referred to as published peptides in this study. Among them, 168 are cleaved, while 47 are noncleaved. Twenty-five whole protein sequences were downloaded from NCBI. They are aaa65789, aaa72945, aaa79971, aab27127, baa01583, baa02756, baa03581, baa03905, baa08372, baa09073, baa09075, baa88057, bab08107, caa03854, caa43793, cab46677, gnwvtc, jc5620, np_043570, np_671491, p26663, p26664, p27958, pc2219, and s40770. Within these 25 whole protein peptides (data are not shown) are 123 cleavage sites. The cleavage sites with notes of POTENTIAL, BY SIMILARITY , OR PROBABLE were removed. First, we built Type-I models using the leave-one-out method. Each model is built on 214 published peptides and tested on 1 remaining published peptide. The models are then evaluated in terms of the mean accuracy, which is calculated on 215 leaveone-out testes. The best model is selected and tested on 25 new sequences downloaded from NCBI, which are regarded as the independent testing data. The simulation shows that the noncleaved, cleaved, and total accuracies are 77%, 77%, and 77%, respectively. Shown in Fig. 9 .10 are the prediction results on five whole protein sequences, where the horizontal axis indicates the residues in whole protein sequences and the vertical axis the probability of positive (cleavage). If the probability at a residue is larger than 0.5, the corresponding residue is regarded as the cleavage site P 1 . The simulation shows that there were too many false positives. The average of false positive fraction was 27%, or 722 misclassified noncleavage sites. Second, we built Type-II models. Each model is built and validated using all the published peptides (215) plus the peptides scanned from 24 newly downloaded sequences. In this run, the protein-oriented leave-one-out method is used. With this method, 1 of 24 new sequences is removed for validation on a model built using 215 published peptides and the peptides scanned from the remaining 23 new sequences. This is repeated 24 times and the performance is estimated on the predictions on 24 new sequences. The best model is selected and tested on the peptides on the remaining new sequence. These peptides are regarded as the independent testing peptides. The noncleaved, cleaved, and total accuracies are 99%, 83%, and 99%, respectively. It can be seen that the prediction accuracy has been greatly improved compared with the Type-I models. Shown in Fig. 9 .11 is a comparison between the Type-I and the Type-II models. It shows that the Type-II models greatly outperformed the Type-I models in terms of the performance for noncleaved peptides. More important, the standard deviation in the Type-II models is much smaller, demonstrating high robustness. Shown in Fig. 9 .12 are the prediction results on five protein sequences using the Type-II models, where the horizontal axis indicates the residues in whole protein TPf Total Type-I Type-II Fig. 9 .11 A comparison between the Type-I and Type-II models. It can be seen that the Type-II models performed much better than the Type-I models in terms of the increase of specificity (truenegative fraction). Therefore, the false-positive fraction (false alarm) has been significantly reduced. TNf and TPf stand for true-negative fraction (specificity) and true-positive fraction (sensitivity). Reproduced from [ 58 ] with permission Fig. 9.12 The probabilities of cleavage sites among the residues for five whole protein sequences for the Type-II models. The horizontal axes indicate the residues in whole sequences and the vertical axes the probabilities of positives (cleavage sites). The numbers above the percentages are the NCBI accession numbers and the percentages are the false positive fractions and the true positive fractions. The simulation shows that the Type-II models demonstrated very small falsepositive fractions compared with the Type-I models. It can be seen that the probability of positives are very clean . Reproduced from [ 58 ] with permission sequences and the vertical axis the probability of positive (cleavage). The simulation shows fewer false positives than the Type-I models. The reason is that many of the 25 newly downloaded sequences were published after the reports with the published peptides. The published peptides may not contain complete information for all these 25 new sequences. This chapter introduced the state-of-the-art machines for peptide classification. Through the discussion, we can see the difference between the peptide machines and other machine learning approaches. Each peptide machine combines the feature extraction process and the model construction process to improve the efficiency. Because peptide machines use the novel bio-basis function, they can have biologically well-coded inputs for building models. This is the reason that the peptide machines outperformed the other machine learning approaches. The chapter also discussed some issues related with peptide classification, such as model evaluation, protein-oriented validation, and model lifetime. Each of these issues is important in terms of building correct, accurate, and robust peptide classification models. For instance, the blind test can be easily missed by some new bioinformatics researchers. Protein-oriented validation has not yet been paid enough attention in bioinformatics. There is also little discussion on model lifetime. Nevertheless, these important issues have been evidenced in this chapter. We should note that there is no a priori knowledge about which peptide machine should be used. Research into the link between these three peptide machines may provide some important insights for building better machines for peptide classification. The core component of peptide machines is a mutation matrix. Our earlier work shows that the model performance using various mutation matrices varies [ 10 ] . It is then interesting to see how we can devise a proper learning method to optimize the mutation probabilities between amino acids during model construction. On the active site of proteases, 3. Mapping the active site of papain; specific peptide inhibitors of papain Sequence and structure based prediction of eukaryotic protein phosphorylation sites The carbohydrates of glycoproteins Machine learning algorithms for protein functional site recognition Intrinsic protein disorder in complete genomes Matching protein beta-sheet partners by feedforward and recurrent neural networks RONN: use of the bio-basis function neural network technique for the detection of natively disordered regions in proteins Predicting the secondary structure of globular proteins using neural network models Characterising proteolytic cleavage site activity using bio-basis function neural networks A novel neural network method in mining molecular sequence data Basic local alignment search tool A model of evolutionary change in proteins. matrices for detecting distant relationships A structural basis for sequence comparisons-an evaluation of scoring methodologies Application of support vector machines to biology Linear optimization of predictors for secondary structure: application to transbilayer segments of membrane proteins Reduced bio-basis function neural networks for protease cleavage site prediction Predict disordered proteins using bio-basis function neural networks Reduced bio basis function neural network for identification of protein phosphorylation sites: comparison with pattern recognition algorithms Predicting the phosphorylation sites using hidden Markov models and machine learning methods Bio-basis function neural networks for the prediction of the Olinkage sites in glyco-proteins Predict signal peptides using bio-basis function neural networks A bio-basis function neural network for protein peptide cleavage activity characterisation Bio-kernel self-organizing map for HIV drug resistance classification Prediction of caspase cleavage sites using Bayesian bio-basis function neural networks Mining SARS-coV protease cleavage data using decision trees, a novel method for decisive template searching Predict T-cell epitopes using bio-support vector machines Vapnik V (1995) The nature of statistical learning theory The kernel trick for distances Sparse Bayesian learning and the relevance vector machine Orthogonal least squares learning algorithm for radial basis function networks Bio-support vector machines for computational proteomics A practical Bayesian framework for backpropagation networks Orthogonal kernel machine in prediction of functional sites in proteins Relevance peptide machine for HIV-1 drug resistance prediction How antibodies block HIV infection: paths to an AIDS vaccine The need for a global HIV vaccine enterprise HIV vaccine still out of our grasp Molecular mechanics analysis of drug-resistant mutants of HIV protease Prediction of HIV-1 protease inhibitor resistance using a protein-inhibitor flexible docking approach Correlation between rules-based interpretation and virtual phenotype interpretation of HIV-1 genotypes for predicting drug resistance in HIV-infected individuals Variable prediction of antiretroviral treatment outcome by different systems for interpreting genotypic human immunodeficiency virus type 1 drug resistance Structure-based phenotyping predicts HIV-1 protease inhibitor resistance Diversity and complexity of HIV-1 drug resistance: a bioinformatics approach to predicting phenotype from genotype Comparison of nine resistance interpretation systems for HIV-1 genotyping The role of resistance characteristics of viral strains in the prediction of the response to antiretroviral therapy in HIV infection Use of viral load measured after 4 weeks of highly active antiretroviral therapy to predict virologic putcome at 24 weeks for HIV-1-positive individuals Geno2pheno: estimating phenotypic drug resistance from HIV-1 genotypes Characterizing the relationship between HIV-1 genotype and phenotype: prediction-based classification Construction, training and clinical validation of an interpretation system for genotypic HIV-1 drug resistance based on fuzzy rules revised by virological outcomes Predicting HIV drug resistance with neural networks Human immunodeficiency virus reverse transcriptase and protease sequence database Rapid, phenotypic HIV-1 drug sensitivity assay for protease and reverse transcriptase inhibitors Analysis of the protease sequences of HIV-1 infected individuals after indinavir monotherapy Comparison of the predicted and observed secondary structure of T4 phage lysozyme Classification and nomenclature of virus: fifth report of the International Committee on Taxonomy of Viruses Isolation of a cDNA clone derived from a blood-borne non-A non-B viral hepatitis genome An assay for circulating antibodies to a major etiologic virus of human non-A non-B hepatitis Predicting hepatitis C virus protease cleavage sites using generalised linear indicator regression models