Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Accurate prediction of functional effects for variants by combining gradient tree boosting with optimal neighborhood properties

Abstract

Single amino acid variations (SAVs) potentially alter biological functions, including causing diseases or natural differences between individuals. Identifying the relationship between a SAV and certain disease provides the starting point for understanding the underlying mechanisms of specific associations, and can help further prevention and diagnosis of inherited disease.We propose PredSAV, a computational method that can effectively predict how likely SAVs are to be associated with disease by incorporating gradient tree boosting (GTB) algorithm and optimally selected neighborhood features. A two-step feature selection approach is used to explore the most relevant and informative neighborhood properties that contribute to the prediction of disease association of SAVs across a wide range of sequence and structural features, especially some novel structural neighborhood features. In cross-validation experiments on the benchmark dataset, PredSAV achieves promising performances with an AUC score of 0.908 and a specificity of 0.838, which are significantly better than that of the other existing methods. Furthermore, we validate the capability of our proposed method by an independent test and gain a competitive advantage as a result. PredSAV, which combines gradient tree boosting with optimally selected neighborhood features, can return reliable predictions in distinguishing between disease-associated and neutral variants. Compared with existing methods, PredSAV shows improved specificity as well as increased overall performance.

Introduction

Single amino acid variants (SAVs) are single-base changes that result in amino acid changes of the encoded protein [1]. With the rapid development of sequencing and genomic analysis technologies, substantial SAVs between individuals have been uncovered. The 1000 Genomes project [2] and recent sequencing of whole human genomes [36] have provided a large number of single-nucleotide polymorphisms (SNPs), insertions, deletions and structural variants in humans. Among these variations, SAVs are recognized as the most common type in the human genome [7, 8], and some are often closely related to particular diseases [911]. According to the previous studies, SAVs may be responsible for the initiation or progression of cancer through aberrant proteins [12]. And the amino acid change can affect, for example, protein stability, interactions and enzyme activity, thereby leading to disease. Therefore, the identification of whether a SAV is neutral or disease-associated is playing an increasingly important role in understanding the underlying mechanisms of specific SAV-disease associations and developing treatment strategies for diseases.

However, experimentally determining the SAV-disease relationship of such a large number of variants is time-consuming and costly. Accurate computational approaches are vital for analysis the relationship between SAV and disease. Current prediction methods typically employ machine learning algorithms [1316] such as neural networks [17], random forests(RF) [18] and support vector machines (SVMs) [19], and a large variety of properties, including amino acid sequence features [20], position-specific scoring matrices, residue-contact network features and 3-D structure information. This includes methods such as SIFT [21, 22], SNAP [23], Polyphen2 [24], FunSAV [25] and SusPect [26]. SIFT uses sequence homology to predict phenotypic effect based on the assumption that amino acid variants in the evolutionarily conserved regions are more likely to have functional effects [21, 22]. SNAP [23] combines multiple sequence analysis methods with neural networks to predict the functional effects of variants. Polyphen2 [24] predicts the functional impact of a variant by a Naive Bayes classifier trained using sequence, phylogenetic and structural information. FunSAV utilizes a two-stage random forest with a large number of sequence and structural properties to discriminate the SAV-disease links [25]. Yates et al. combine sequence and structural features to build an SVM classifier named SusPect to predict disease-SAV associations [26].

In this work, we develop a novel approach, termed as PredSAV, to predict the phenotypic effects of SAVs by using the Friedman’s gradient tree boosting [27, 28] algorithm. PredSAV combines both sequence neighborhood features and structural neighborhood features describing not only the properties of the target residue but also the target residue’s neighborhood environment. PredSAV uses a efficient two-step feature selection method to eliminate uninformative properties, which in turn improves the performance and helps to build faster and more cost-effective models. Extensive comparisons of PredSAV with other existing tools on the benchmark dataset and another independent dataset show that PredSAV significantly outperforms the existing state-of-the-art methods, and illustrate the effectiveness and advantage of the proposed approach. The framework of PredSAV is shown in Fig 1.

thumbnail
Fig 1. The framework of PredSAV.

(A) Feature representation. A total of 1521 sequence, Euclidean and Voronoi neighborhood features are initially generated. (B)Two-step feature selection. Stability selection is used as the first step. We select the top 152 features with score larger than 0.2. The second step is performed using a wrapper-based feature selection. Features are evaluated by 5-fold cross-validation with the GTB algorithm. (C) Prediction model. Gradient boosted trees are finally built for prediction.

https://doi.org/10.1371/journal.pone.0179314.g001

Materials and methods

Datasets

The dataset, taken from Wang and Song [25], are mainly gathered from the UniProt [29] human sequence variations and the Ensemble human variation database [30]. Disease-associated variants are obtained from the UniProt human sequence variations and non-Mendelian disease variants without any OMIM entry [31] are removed. Neutral variants are extracted from the Ensemble human variation database. All the SAVs emanate from the HapMap project [32]. We remove the redundant homology sequences at the sequence similarity of 40% by using CD-HIT [33]. Then a BLAST search [34] is used to map the remaining sequences to PDB structures [35]. Structures with resolutions lower than 2.5Å are removed. Ambiguity or invalid SAV-disease associations and neutral variants are deleted. Obsoleted PDB structures, such as 1VJJ and 2HR0, are removed. Eventually, the dataset consists of 670 proteins containing 1006 disease-associated and 963 neutral variants. A total of 816 disease-associated and 776 neutral variants are randomly selected as the benchmark dataset and the rest are used as the independent dataset including of 190 disease-associated and 187 neutral variants.

Performance evaluation

We evaluate the performance of the proposed method using 5-fold cross-validation and several widely used measures. These measures include sensitivity (SEN/Recall), specificity (SPE), precision (PRE), F1-score (F1), accuracy (ACC), the Matthew’s correlation coefficient (MCC) and the area under the ROC curve (AUC). (1) (2) (3) (4) (5) (6)

Features extraction

In our experiment, a wide variety of sequence and structure features are generated for predicting the phenotypic effects of SAVs. Several novel structural features, including residue-contact network features, solvent exposure features and structural neighborhood features, are calculated. The details of these features are listed as follows.

Sequence features.

A large number of sequence features are calculated: 1) position-specific scoring matrices (PSSMs) [34]; 2) predicted solvent accessibility using the SSpro and SSpro8 programs [36]; 3) predicted native disorder by DISOPRED [37]; 4) the dScore that represents the difference between the PSIC [38] scores for the wild type amino acid residue and mutant amino acid residue, calculated by PolyPhen2 [24]; 5) predicted disorder in proteins by DisEMBL [39]; 6) the local structural entropy of a particular residue is computed by LSE [40]; 7) the eight physicochemical properties for each amino acid are obtained from the AAindex database [41]; 8)BLOSUM62 [42] was used to count the relative frequencies of amino acid and their substitution probabilities; 9) solvent accessible surface area, secondary structure and local backbone angles generated by SPIDER2 [43]; 10) predicted the relative solvent accessibility of protein residues by the ACCpro and ACCpro20 from the SCRATCH package [36]; 11)evolutionary conservation scores calculated based on PSSM [34] and Jensen-Shannon divergence [44, 45].

Structure features.

Structural features, including secondary structure, four-body statistical pseudo-potential, solvent accessibility and exposure features, are calculated as candidate features for SAV phenotype prediction. We used DSSP [46] to calculate the secondary structure features, including hydrogen bonds, solvent-accessible surface area, Cα atom coordinates and backbone torsion angles. The four-body statistical pseudo-potential is based on the Delaunay tessellation of proteins [47]. Delaunay tessellation is a effective way to define the structural neighbors of a target protein. The potential is defined as follows: (7) where i, j, k, and l are the residue identities of the four amino acids in a Delaunay tessellation of the target protein. Each residue is represented by a central point among the atoms in the residue. is the observed frequency of the residue composition (ijkl) in a tetrahedron of type α over a set of protein structures. is the expected random frequency.

Energy scores including side-chain energy score, residue energy, conservation, interface propensity, combined1 score, combined2 score and relative solvent accessibility are calculated by using ENDES [48]. Two combined energy scores are also used. The combined1 score is a combination of residue energy, conservation and interface propensity scores. The combined2 score is an optimized combination weights of the three features to get the best prediction of single residue.

Solvent-accessible related features have been shown to be very useful in identifying SAV-disease association [4951]. We use NACCESS [52] and NetSurfP [53] to calculate solvent accessibility for the protein structures, respectively. The NACCESS program is used to calculate the absolute and relative solvent accessibilities of all atoms. For NetSurfP, the absolute and relative surface accessibility, Z-fit score and secondary structure are computed based on the homology proteins obtained from PSI-BLAST search.

Solvent exposure features, include the coordination number (CN), number of Cβ atoms in the upper Half-Sphere (HSEBU), number of Cβ atoms in the lower Half-Sphere(HSEBD) and residue depth (RD), are calculated by HSEpred [54] and the hsexpo program [55]. The hsexpo program uses protein structure information, while the HSEpred uses sequence information to predict these features.

Residue-contact network features.

Residue-residue contact networks have bee proved to be very beneficial for analyzing and predicting SAV-disease associations [56]. If the distance between the centers of two residues in a structure are within 6.5Å, an edge exists between the two residues in the network. We use NAPS [57] to compute the residue-residue contact network properties, which describe the local environment of the target variant in the network, including betweenness, closeness, coreness, degree, clustering coefficient, eigenvector centrality, eccentricity and average nearest neighbor degree.

Structural neighborhood features (SNF).

Conventional features usually describe only the properties of the current residue itself, cannot represent the real environment well, and thus are insufficient to predict functional effects of SAVs with high precision. Here, we calculate two types of structural neighborhood features (SNF) based on Euclidean distance and Voronoi diagram [5860], respectively. Surrounding residues located within a sphere of the radius of 5Å are defined as the Euclidean neighborhood of the central amino acid. The Euclidean distance is computed between any heavy atoms of the surrounding residues and that of the central amino acid. The score of a specific feature i for the central residue r regarding the neighbor n is defined as follows: (8) where dr,n is the minimum Euclidean distance between residue r and residue n. The Euclidean neighborhood feature of the central residue r is defined as: (9) where m is the total number of Euclidean neighbors.

Voronoi neighborhood features are calculated based on Voronoi diagram/Delaunay triangulation. For a 3-D protein structure, individual atoms are devided into Voronoi polyhedra by Voronoi tessellation partition. In the Voronoi diagram (Delaunay triangulation), a pair of residues are defined as Voronoi neighbors if there exists at least one common Voronoi facet between heavy atoms of each residue. The Qhull package [61] is used to calculate Voronoi/Delaunay polyhedra. For the target residue r and its Voronoi neighbors n {n = 1, …, m}, the Voronoi neighborhood property of the feature i is defined as: (10) where Pi(n) is the score of the residue feature i for neighbor n.

Feature encoding with neighborhood properties.

For each sample, a combination of 1,287 (117*11) sequence neighborhood features, 117 Euclidean neighborhood features and 117 Voronoi neighborhood features are calculated. The sequence neighborhood features is generated by applying a sliding window of size 11 to incorporate the evolutionary information from upstream and downstream neighbors in the protein sequence.

Feature selection

The feature selection method improves the performance by removing some redundant features in high-dimensional data [6265]. In this study, we propose a two-step feature selection approach to select the most important features for predicting the phenotypic effects of SAVs. First, we assess the feature elements using the stability selection [66] calculated by the RandomizedLasso package in the scikit-learn [67]. The idea of stability selection is that a feature selection algorithm is employed on subsample datasets and subsample features. The selection results are merged after repeating a certain number of times. Stronger features have higher scores (close to 1), while weaker features have scores close to 0. The score represents the importance of an individual feature for correctly predicting an SAV-disease association. Here, we select the top 152 features with the score larger than 0.2.

The second step is performed using a wrapper-based feature selection method. The features are evaluated by 5-fold cross-validation with the GTB (gradient tree boosting) algorithm, and correlation features are added by sequential forward selection (SFS). In the SFS scheme, features are sequentially added to a null feature set till an optimal feature subset is obtained. Each added feature is the one whose add maximizes the performance of the classifier. This stepwise feature selection process continues until the AUC score no longer increased. As a result, a set of 44 optimal features are selected as the final optimal feature set.

Gradient tree boosting algorithm

The Gradient Tree Boosting (GTB) [27, 28] is an effective machine learning algorithm that can be utilized for both classification and regression problems. In this study, GTB is implemented under the PredSAV framework as shown in Fig 1 and the prediction of the phenotypic effect of single amino acid variants could be considered as a binary classification problem. For a large number of given input feature vectors χi (χi = {x1, x2, …, xn}, i = 1, 2, …, N) with labels yi (yiϵ{−1, +1}, i = 1, 2, …, N, where “-1” represents neutral variant and “+1” denotes disease-associated variant), the purpose of the GTB algorithm is to build an effective classifier to predict whether a variant is disease-associated or neutral. The GTB algorithm is shown in Algorithm 1.

Algorithm 1 Gradient Tree Boosting Algorithm

Input:

 Data set: D = {(χ1, y1), (χ2, y2), …, (χN, yN)}, χiϵχ, χR, yiϵ{−1, +1}; loss function: L(y, Θ(χ)); iterations = M;

Output:

1: Initialize

2: for m = 1 to M do

3: Compute the negative gradient as the working response

4: The input χi is adapted to the classification model ri by Logistic function and get the estimate αm of βh(χ; α)

5: Get the estimate βm by minimizing L(yi, Θm−1(χi) + βh(χi; αm))

6: Update Θm(χ) = Θm−1(χ) + βmh(χ; αm)

7: end for

8: return

In the algorithm, the variable iterations = M should be initialized. The logistic function is used as the loss function, which is defined as: (11) where y is a real class label of variants and Θ(χ) is a decision function. The decision function is initialized by the following equation. (12) where N is the number of SAVs in the benchmark dataset. Then, GTB constructs m different classification trees h(χ, α1), h(χ, α2), …, h(χ, αm) from a number of benchmark datasets. The addictive function Θm(x) can be defined as: (13)

Above, the βm and αm are a weight and vector of parameters for the m-th classification tree h(χ, αm), respectively. In order to minimize the loss function L(y, Θm(χ)), the weight of βm and the parameter of αm need to be iterated from m = 1 to m = M. In the third step, the negative gradient ri as the working response by the following formula: (14) Then, the weight of βm and the parameter of αm for the mth iteration can be defined as: (15)

However, we do not directly calculate the above equation. In the fourth step, the input χi is adapted to the classification model ri by Logistic function and get the estimate αm of βh(χi; α). Therefore, we can obtain (16)

In the fifth step, the estimate parameter of βm is obtained by minimizing the log loss function L(y, Θ(χ)). (17)

Then, in the sixth step, a new addictive function Θm(χ) is updated by in the eq (8). Finally, we obtain a classification function ΘM(χ) and a useful GTB model as follows: (18)

We use a grid search strategy to select the optimal parameters of GTB with 5-fold cross-validation on the benchmark dataset. The optimized number of trees of the GTB is 2000. And the selected depth of the trees is 3. The rest use the default parameters.

The source code and data are available at http://www.leideng.org/PredSAV/.

Results and discussion

Benefits of the two-step feature selection

The selection of informative attributes is critically important for building effective and accurate classification models. In total 1521 sequence, Euclidean and Voronoi neighborhood features are initially generated. We apply a two-step feature selection method, consisting of stability selection and sequential forward selection. Stability selection is used as the first attribute selection step for two reasons. First, stability selection can address the difficult variable selection problem with markedly improved error control and structure estimation, especially for high-dimensional problems. Second, stability selection depends little on the chosen initial regularisation and can reduce the risk of overfitting [66]. To assess the utility of the stability selection method, we evaluate the performance by incorporating the GTB classifier with selected attributes that correspond to different cutoffs of stability selection scores. As shown in Table 1, when the number of selected features decreases from 1521 to 152 (the cutoff increases from 0 to 0.2), the highest accuracy of 81.3% is yielded. The other measurements (SEN, SEP, PRE, MCC and AUC) are observed as 0.810, 0.817, 0.823, 0.628 and 0.896, respectively. We select the top 152 features (stability score > 0.2) as the input of the next sequential forward selection step. A set of 44 optimal features is finally selected with the highest AUC score of 0.908. The results of selected features show ∼2% and ∼5% increase in AUC and MCC over the initial features, respectively.

thumbnail
Table 1. Performance of selected attributes with the two-step feature selection method.

The first column lists different cutoffs of stability selection scores.

https://doi.org/10.1371/journal.pone.0179314.t001

We compare the proposed two-step feature selection method with three widely used feature selection methods: random forests(RF), maximum Relevance Minimum Redundancy(mRMR) [68] and Recursive Feature Elimination(RFE) [69]. The experiment is based on the benchmark dataset with 5-fold cross-validation. Fig 2 shows the ROC curves of the four feature selection methods. The results are shown in Fig 2. Our two-step feature selection approach obtains the best performance. The results indicate that our two-step feature selection algorithm, which is a composite approach combining the merits of both stability selection and sequential forward selection, can substantially boost the prediction performance with less computational expense and lower risk of overfitting.

thumbnail
Fig 2. ROC curves of our two-step algorithm and other three existing feature selection methods.

https://doi.org/10.1371/journal.pone.0179314.g002

Feature importance

The feature importance of these features are calculated by using the gradient tree boosting method. The relative importance and rankings of the optimal features are shown in Fig 3 and Table 2. Among them, the feature with the highest score (>100) is the dScore feature calculated by PolyPhen2. The dScore represents the difference between the PSIC scores of the wild type amino acid residue and mutant amino acid residue. Another important feature that have not been found useful in previous studies is structural neighborhood features (Euclidean and Voronoi). We find that structural neighborhood features achieve the second highest score compared with other features, suggesting that structural neighborhood features are critical in distinguishing disease-associated SAVs from neutral SAVs. Solvent accessibilities are also found to be useful in SAVs phenotype prediction. Many solvent accessibility related features, including the solvent accessibility feature calculated by NACCESS, NetSurfP, DSSP, solvent exposure features, ACC20 and the SSpro score, are important and contributive. The results suggest that sequence and structural neighborhood features complement each other quite well and thus collectively make a contribution to the performance enhancement.

thumbnail
Table 2. Rankings of feature importance for the optimal selected features.

SN, EN and VN represent sequence neighborhood, Euclidean neighborhood and Voronoi neighborhood, respectively. The numbers in the brackets denote the positions in the sliding window for sequence neighborhood features.

https://doi.org/10.1371/journal.pone.0179314.t002

thumbnail
Fig 3. The relative importance and ranking of the optimal feature group, as evaluated by the gradient tree boosting.

The bar represents the importance score of the corresponding feature group.

https://doi.org/10.1371/journal.pone.0179314.g003

Gradient tree boosting improves predictions

PredSAV uses Gradient Tree Boosting (GBT) to build the final model with the 44 optimal features. We compare GBT with Support Vector Machine (SVM) and Random Forests (RF), which are well known to perform fairly well on a variety of tasks. Fig 4 shows the AUC scores of GTB and other machine learning methods on the final optimal feature set. GTB, SVM and RF achieve AUC values of 0.908, 0.894 and 0.890, respectively. Comparing with the other methods, the GTB model can improve the prediction preformance. Note that the GBT algorithm is implemented with scikit-learn [67] in this study.

thumbnail
Fig 4. Comparison of the AUC value of the the three methods using 5-fold cross-validation on the benchmark dataset.

https://doi.org/10.1371/journal.pone.0179314.g004

PredSAV outperforms other predictors

To evaluate the performance of the proposed PredSAV, six existing SAVs prediction methods, including FunSAV [25], Polyphen2 [24], SusPect [26], SIFT [21], SNAP [23] and nsSNPAnalyzer [70], are evaluated on the benchmark dataset.

Table 3 and Fig 5 show the detailed results of comparing our method with the existing methods. Overall, our approach shows dominant advantage over the existing methods in six metrics: ACC, SPE, PRE, F1, MCC and AUC. When comparing the AUC score with that of the existing classifiers, FunSAV (0.814), Polyphen2 (0.813), SusPect (0.800), SIFT (0.760) and SNAP (0.706), our PredSAV classifier (0.908) shows greater improvement by 9%, 9%, 11%, 14% and 20%, respectively. For the remaining measurements ACC, SPE, PRE, F1 and MCC, we can observe similar increases. Especially, the specificity of PredSAV is significant higher than other methods (increased by 10%), which suggests that it has better performance detecting true negatives and may help for reducing experiment cost. Only in SEN, PredSAV is lower than PolyPhen2 and SNAP (0.855 and 0.866 for SNAP and PolyPhen2, respectively). We can observe that PredSAV gains a balanced sensitivity and specificity (0.814 and 0.838, respectively), suggesting that PredSAV has better balance of prediction accuracy between disease-associated and neutral SAVs.

thumbnail
Table 3. Prediction performance of PredSAV classifiers in comparison with six other prediction tools on the benchmark dataset.

https://doi.org/10.1371/journal.pone.0179314.t003

thumbnail
Fig 5. The ROC curves of seven classifiers on the benchmark dataset.

https://doi.org/10.1371/journal.pone.0179314.g005

Performance confirmed for independent test

We also validate the performance of PredSAV on the independent test dataset to avoid over-optimistic performance estimates. Results of the independent test are presented in Table 4, which indicate marked improvements for all the performance measures except SEN comparing PredSAV with the exiting methods. Fig 6 shows the ROC curves. The ROC curves indicate the trade-off between the amounts of true positives (TP) and false positives (FP) generated by the classifiers. We observe that PredSAV produces higher true positive rates of prediction across most of the false positive rates. Overall, these observations suggest that the performance of our PredSAV approach is superior to that of the state-of-the-art approaches.

thumbnail
Table 4. Prediction performance of PredSAV classifiers in comparison with six other prediction tools on the independent test dataset.

https://doi.org/10.1371/journal.pone.0179314.t004

thumbnail
Fig 6. The ROC curves of seven classifiers on the independent test dataset.

https://doi.org/10.1371/journal.pone.0179314.g006

Case study

To further illustrate the effectiveness of PredSAV, we present examples by comparing predictions for variants that are difficult to classify with commonly applied methods. The enzyme phenylalanine hydroxylase (PAH, PDB ID: 1J8U, chain A) [71, 72] is responsible for the conversion of phenylalanine to another amino acid, tyrosine. PAH works with a molecule called tetrahydrobiopterin (BH4) to carry out this chemical reaction. The majority of mutations in PAH result in deficient enzyme activity and cause hyperphenylalaninemia. Some cause phenylketonuria (PKU), others cause non-PKU hyperphenylalaninemia, while still others are silent polymorphisms. As shown in Fig 7A, three PKU-associated SAVs, Q160P (dbSNP:rs199475601), V177L (dbSNP:rs199475602) and V388L (dbSNP:rs62516101), are colored in red. This example illustrates how PredSAV combines gradient tree boosting with optimal neighborhood features to provide better predictions. PredSAV (TP = 3) correctly identified all the three disease-associated variants, compared to Suspect (TP = 0), PolyPhen (TP = 0), SNAP (TP = 2), FunSAV (TP = 0), SIFT (TP = 1) and nsSNP (TP = 0).

thumbnail
Fig 7. Prediction examples of the functional effects of SAVs in two proteins by PredSAV and other methods.

Red color denotes disease-associated variants while blue color represents neutral variants. (A) and (B) represent proteins PAH (PDB ID: 1J8U, chain A) and LSS (PDB ID: 1W6K, chain A), respectively. 3-D structures are rendered using PyMol [75].

https://doi.org/10.1371/journal.pone.0179314.g007

Another example is lanosterol synthase (LSS, PDB ID: 1W6K, chain A) [73, 74], which catalyzes the cyclization of (S)-2,3 oxidosqualene to lanosterol, a reaction that forms the sterol nucleus. Through the production of lanosterol may regulate lens protein aggregation and increase transparency. The variants R614W (dbSNP:rs35785446) and P688L (dbSNP:rs17293705) in LSS are neutral substitutions. From Fig 7B, we can see that PredSAV can predict the neutral SAVs successfully, while other existing methods result in almost completely wrong results (except Suspect in P688L). This suggests that PredSAV has the highest specificity, which is desirable for many biological applications since it allows researchers to identify a short list of SAVs for targeted phenotype studies.

Conclusion

In this study, we present a novel approach named PredSAV for producing reliable predictions in distinguishing between effect and neutral variants. To be able to do this, we first extract a very large collection of imformative and complementary features, including sequence, structure, network and neighborhood features that describe the local environments proximal to the centered variant and neighboring residues. A two-step feature selection approach, which combines stability selection and sequential forward selection, is utilized to select an optimal subset of features within a reasonable computational cost, and thus improves the prediction performance and reduces the risk of overfitting. Importantly, the use of gradient tree boosting algorithm further attains higher levels of prediction accuracy. We evaluate the PredSAV method with both cross-validation and independent test, and the results indicate that the proposed PredSAV is able to identify disease-associated SAVs with higher overall performance, especially in terms of specificity, when compared with other existing approaches.

A limitation of PredSAV is that it requires the 3D protein structure, which may limit its broader application. However, with the increasing solved protein structures, protein homology modeling projects [76] and predicted 3D structures [77], it is expected that PredSAV can be used as a powerful tool to prioritize the disease-associated variants and help towards the phenotypic effect annotation of these targets.

As for future work, we will explore more efficient features to further improve the performance and learn from other methods [7882] to provide a a web-server for the method proposed in this paper.

Supporting information

S1 File. The disease-associated and neutral variant data used in these experiments.

https://doi.org/10.1371/journal.pone.0179314.s001

(ZIP)

Author Contributions

  1. Conceptualization: YP LD.
  2. Data curation: YP.
  3. Formal analysis: YP DL.
  4. Funding acquisition: LD.
  5. Investigation: LD.
  6. Methodology: YP DL LD.
  7. Project administration: LD.
  8. Resources: LD YP.
  9. Software: YP DL.
  10. Supervision: LD.
  11. Validation: YP DL LD.
  12. Writing – original draft: YP LD.
  13. Writing – review & editing: YP DL LD.

References

  1. 1. Yip YL, Scheib H, Diemand AV, Gattiker A, Famiglietti LM, Gasteiger E, et al. The Swiss-Prot variant page and the ModSNP database: A resource for sequence and structure information on human protein variants. Human mutation. 2004;23(5):464–470. pmid:15108278
  2. 2. Consortium GP, et al. A map of human genome variation from population-scale sequencing. Nature. 2010;467(7319):1061–1073. pmid:20981092
  3. 3. Cline MS, Karchin R. Using bioinformatics to predict the functional impact of SNVs. Bioinformatics. 2011;27(4):441–448. pmid:21159622
  4. 4. Schuster SC, Miller W, Ratan A, Tomsho LP, Giardine B, Kasson LR, et al. Complete Khoisan and Bantu genomes from southern Africa. Nature. 2010;463(7283):943–947. pmid:20164927
  5. 5. Ahn SM, Kim TH, Lee S, Kim D, Ghang H, Kim DS, et al. The first Korean genome sequence and analysis: full genome sequencing for a socio-ethnic group. Genome research. 2009;19(9):1622–1629. pmid:19470904
  6. 6. Wang J, Wang W, Li R, Li Y, Tian G, Goodman L, et al. The diploid genome sequence of an Asian individual. Nature. 2008;456(7218):60–65. pmid:18987735
  7. 7. Jiao YS, Du PF. Predicting Golgi-resident protein types using pseudo amino acid compositions: Approaches with positional specific physicochemical properties. Journal of theoretical biology. 2016;391:35–42. pmid:26702543
  8. 8. Du P, Wang L. Predicting human protein subcellular locations by the ensemble of multiple predictors via protein-protein interaction network with edge clustering coefficients. PloS one. 2014;9(1):e86879. pmid:24466278
  9. 9. Li P, Guo M, Wang C, Liu X, Zou Q. An overview of SNP interactions in genome-wide association studies. Briefings in Functional Genomics. 2014; p. elu036. pmid:25241224
  10. 10. Zou Q, Wan S, Ju Y, Tang J, Zeng X. Pretata: predicting TATA binding proteins with novel features and dimensionality reduction strategy. BMC Systems Biology. 2016;10(4):401. pmid:28155714
  11. 11. Du P, Tian Y, Yan Y. Subcellular localization prediction for human internal and organelle membrane proteins with projected gene ontology scores. Journal of theoretical biology. 2012;313:61–67. pmid:22960368
  12. 12. Bignell GR, Greenman CD, Davies H, Butler AP, Edkins S, Andrews JM, et al. Signatures of mutation and selection in the cancer genome. Nature. 2010;463(7283):893–898. pmid:20164919
  13. 13. Yuan LF, Ding C, Guo SH, Ding H, Chen W, Lin H. Prediction of the types of ion channel-targeted conotoxins based on radial basis function network. Toxicology in Vitro. 2013;27(2):852–856. pmid:23280100
  14. 14. Yang H, Tang H, Chen XX, Zhang CJ, Zhu PP, Ding H, et al. Identification of Secretory Proteins in Mycobacterium tuberculosis Using Pseudo Amino Acid Composition. BioMed Research International. 2016;2016. pmid:27597968
  15. 15. Chen XX, Tang H, Li WC, Wu H, Chen W, Ding H, et al. Identification of Bacterial Cell Wall Lyases via Pseudo Amino Acid Composition. BioMed Research International. 2016;2016. pmid:27437396
  16. 16. Zhao YW, Lai HY, Tang H, Chen W, Lin H. Prediction of phosphothreonine sites in human proteins by fusing different features. Scientific reports. 2016;6. pmid:27698459
  17. 17. Specht DF. Probabilistic neural networks. Neural networks. 1990;3(1):109–118.
  18. 18. Breiman L. Random forests. Machine learning. 2001;45(1):5–32.
  19. 19. Cai Yd, Lin SL. Support vector machines for predicting rRNA-, RNA-, and DNA-binding proteins from amino acid sequence. Biochimica et Biophysica Acta (BBA)-Proteins and Proteomics. 2003;1648(1):127–133. pmid:12758155
  20. 20. Wei L, Xing P, Su R, Shi G, Ma ZS, Zou Q. CPPred-RF: a sequence-based predictor for identifying cell-penetrating peptides and their uptake efficiency. Journal of Proteome Research. 2017;. pmid:28436664
  21. 21. Ng PC, Henikoff S. SIFT: Predicting amino acid changes that affect protein function. Nucleic acids research. 2003;31(13):3812–3814. pmid:12824425
  22. 22. Sim NL, Kumar P, Hu J, Henikoff S, Schneider G, Ng PC. SIFT web server: predicting effects of amino acid substitutions on proteins. Nucleic acids research. 2012;40(W1):W452–W457. pmid:22689647
  23. 23. Bromberg Y, Yachdav G, Rost B. SNAP predicts effect of mutations on protein function. Bioinformatics. 2008;24(20):2397–2398. pmid:18757876
  24. 24. Adzhubei I, Jordan DM, Sunyaev SR. Predicting functional effect of human missense mutations using PolyPhen-2. Current protocols in human genetics. 2013; p. 7–20. pmid:23315928
  25. 25. Wang M, Zhao XM, Takemoto K, Xu H, Li Y, Akutsu T, et al. FunSAV: predicting the functional effect of single amino acid variants using a two-stage random forest model. PloS one. 2012;7(8):e43847. pmid:22937107
  26. 26. Yates CM, Filippis I, Kelley LA, Sternberg MJ. SuSPect: enhanced prediction of single amino acid variant (SAV) phenotype using network features. Journal of molecular biology. 2014;426(14):2692–2701. pmid:24810707
  27. 27. Friedman JH. Greedy function approximation: a gradient boosting machine. Annals of statistics. 2001; p. 1189–1232.
  28. 28. Friedman JH. Stochastic gradient boosting. Computational Statistics & Data Analysis. 2002;38(4):367–378.
  29. 29. Bairoch A, Apweiler R, Wu CH, Barker WC, Boeckmann B, Ferro S, et al. The universal protein resource (UniProt). Nucleic acids research. 2005;33(suppl 1):D154–D159. pmid:15608167
  30. 30. Flicek P, Amode MR, Barrell D, Beal K, Brent S, Carvalho-Silva D, et al. Ensembl 2012. Nucleic acids research. 2011; p. gkr991.
  31. 31. Hamosh A, Scott AF, Amberger JS, Bocchini CA, McKusick VA. Online Mendelian Inheritance in Man (OMIM), a knowledgebase of human genes and genetic disorders. Nucleic acids research. 2005;33(suppl 1):D514–D517. pmid:15608251
  32. 32. Frazer KA, Ballinger DG, Cox DR, Hinds DA, Stuve LL, Gibbs RA, et al. A second generation human haplotype map of over 3.1 million SNPs. Nature. 2007;449(7164):851–861. pmid:17943122
  33. 33. Li W, Godzik A. Cd-hit: a fast program for clustering and comparing large sets of protein or nucleotide sequences. Bioinformatics. 2006;22(13):1658–1659. pmid:16731699
  34. 34. Altschul SF, Madden TL, Schäffer AA, Zhang J, Zhang Z, Miller W, et al. Gapped BLAST and PSI-BLAST: a new generation of protein database search programs. Nucleic acids research. 1997;25(17):3389–3402. pmid:9254694
  35. 35. Berman HM, Westbrook J, Feng Z, Gilliland G, Bhat TN, Weissig H, et al. The protein data bank. Nucleic acids research. 2000;28(1):235–242. pmid:10592235
  36. 36. Cheng J, Randall AZ, Sweredoski MJ, Baldi P. SCRATCH: a protein structure and structural feature prediction server. Nucleic acids research. 2005;33(suppl 2):W72–W76. pmid:15980571
  37. 37. Jones DT, Cozzetto D. DISOPRED3: precise disordered region predictions with annotated protein-binding activity. Bioinformatics. 2015;31(6):857–863. pmid:25391399
  38. 38. Sunyaev SR, Eisenhaber F, Rodchenkov IV, Eisenhaber B, Tumanyan VG, Kuznetsov EN. PSIC: profile extraction from sequence alignments with position-specific counts of independent observations. Protein engineering. 1999;12(5):387–394. pmid:10360979
  39. 39. Linding R, Jensen LJ, Diella F, Bork P, Gibson TJ, Russell RB. Protein disorder prediction: implications for structural proteomics. Structure. 2003;11(11):1453–1459. pmid:14604535
  40. 40. Chan CH, Liang HK, Hsiao NW, Ko MT, Lyu PC, Hwang JK. Relationship between local structural entropy and protein thermostabilty. Proteins: Structure, Function, and Bioinformatics. 2004;57(4):684–691. pmid:15532068
  41. 41. Kawashima S, Kanehisa M. AAindex: amino acid index database. Nucleic acids research. 2000;28(1):374–374. pmid:10592278
  42. 42. Henikoff S, Henikoff JG. Amino acid substitution matrices from protein blocks. Proceedings of the National Academy of Sciences. 1992;89(22):10915–10919. pmid:1438297
  43. 43. Heffernan R, Paliwal K, Lyons J, Dehzangi A, Sharma A, Wang J, et al. Improving prediction of secondary structure, local backbone angles, and solvent accessible surface area of proteins by iterative deep learning. Scientific reports. 2015;5. pmid:26098304
  44. 44. Capra JA, Singh M. Predicting functionally important residues from sequence conservation. Bioinformatics. 2007;23(15):1875–1882. pmid:17519246
  45. 45. Miller MP, Kumar S. Understanding human disease mutations through the use of interspecific genetic variation. Human molecular genetics. 2001;10(21):2319–2328. pmid:11689479
  46. 46. Kabsch W, Sander C. Dictionary of protein secondary structure: pattern recognition of hydrogen-bonded and geometrical features. Biopolymers. 1983;22(12):2577–2637. pmid:6667333
  47. 47. Liang S, Grishin NV. Effective scoring function for protein sequence design. Proteins: Structure, Function, and Bioinformatics. 2004;54(2):271–281. pmid:14696189
  48. 48. Liang S, Meroueh SO, Wang G, Qiu C, Zhou Y. Consensus scoring for enriching near-native structures from protein–protein docking decoys. Proteins: Structure, Function, and Bioinformatics. 2009;75(2):397–403. pmid:18831053
  49. 49. Dobson RJ, Munroe PB, Caulfield MJ, Saqi MA. Predicting deleterious nsSNPs: an analysis of sequence and structural attributes. BMC bioinformatics. 2006;7(1):1. pmid:16630345
  50. 50. David A, Sternberg MJ. The contribution of missense mutations in core and rim residues of protein–protein interfaces to human disease. Journal of molecular biology. 2015;427(17):2886–2898. pmid:26173036
  51. 51. Saunders CT, Baker D. Evaluation of structural and evolutionary contributions to deleterious mutation prediction. Journal of molecular biology. 2002;322(4):891–901. pmid:12270722
  52. 52. Hubbard SJ, Thornton JM. Naccess. Computer Program, Department of Biochemistry and Molecular Biology, University College London. 1993;2(1).
  53. 53. Petersen B, Petersen TN, Andersen P, Nielsen M, Lundegaard C. A generic method for assignment of reliability scores applied to solvent accessibility predictions. BMC structural biology. 2009;9(1):1.
  54. 54. Song J, Tan H, Takemoto K, Akutsu T. HSEpred: predict half-sphere exposure from protein sequences. Bioinformatics. 2008;24(13):1489–1497. pmid:18467349
  55. 55. Hamelryck T. An amino acid has two sides: a new 2D measure provides a different view of solvent exposure. Proteins: Structure, Function, and Bioinformatics. 2005;59(1):38–48. pmid:15688434
  56. 56. Li Y, Wen Z, Xiao J, Yin H, Yu L, Yang L, et al. Predicting disease-associated substitution of a single amino acid by analyzing residue interactions. BMC bioinformatics. 2011;12(1):1. pmid:21223604
  57. 57. Chakrabarty B, Parekh N. NAPS: Network Analysis of Protein Structures. Nucleic acids research. 2016; p. gkw383. pmid:27151201
  58. 58. Deng L, Guan J, Dong Q, Zhou S. Prediction of protein-protein interaction sites using an ensemble method. BMC bioinformatics. 2009;10(1):1. pmid:20015386
  59. 59. Deng L, Guan J, Wei X, Yi Y, Zhang QC, Zhou S. Boosting prediction performance of protein–protein interaction hot spots by using structural neighborhood properties. Journal of Computational Biology. 2013;20(11):878–891. pmid:24134392
  60. 60. Chao F, Liu D, Rui H, Chen Z, Lei D. PredRSA: a gradient boosted regression trees approach for predicting protein solvent accessibility. Bmc Bioinformatics. 2016;17 Suppl 1(S1):85–95. pmid:26818760
  61. 61. Barber CB, Dobkin DP, Huhdanpaa H. The quickhull algorithm for convex hulls. ACM Transactions on Mathematical Software (TOMS). 1996;22(4):469–483.
  62. 62. Zou Q, Zeng J, Cao L, Ji R. A novel features ranking metric with application to scalable visual and bioinformatics data classification. Neurocomputing. 2016;173:346–354.
  63. 63. Lin H, Chen W. Prediction of thermophilic proteins using feature selection technique. Journal of microbiological methods. 2011;84(1):67–70. pmid:21044646
  64. 64. Wei L, Xing P, Shi G, Ji ZL, Zou Q. Fast prediction of protein methylation sites using a sequence-based feature selection technique. IEEE/ACM Transactions on Computational Biology and Bioinformatics. 2017;. pmid:28222000
  65. 65. Ding H, Feng PM, Chen W, Lin H. Identification of bacteriophage virion proteins by the ANOVA feature selection and analysis. Molecular BioSystems. 2014;10(8):2229–2235. pmid:24931825
  66. 66. Meinshausen N, Bühlmann P. Stability selection. Journal of the Royal Statistical Society: Series B (Statistical Methodology). 2010;72(4):417–473.
  67. 67. Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, et al. Scikit-learn: Machine learning in Python. The Journal of Machine Learning Research. 2011;12:2825–2830.
  68. 68. Peng H, Long F, Ding C. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Transactions on pattern analysis and machine intelligence. 2005;27(8):1226–1238. pmid:16119262
  69. 69. Guyon I, Weston J, Barnhill S, Vapnik V. Gene selection for cancer classification using support vector machines. Machine learning. 2002;46(1–3):389–422.
  70. 70. Bao L, Zhou M, Cui Y. nsSNPAnalyzer: identifying disease-associated nonsynonymous single nucleotide polymorphisms. Nucleic acids research. 2005;33(suppl 2):W480–W482. pmid:15980516
  71. 71. Flydal MI, Martinez A. Phenylalanine hydroxylase: function, structure, and regulation. IUBMB life. 2013;65(4):341–349. pmid:23457044
  72. 72. Andersen OA, Flatmark T, Hough E. High resolution crystal structures of the catalytic domain of human phenylalanine hydroxylase in its catalytically active Fe (II) form and binary complex with tetrahydrobiopterin. Journal of molecular biology. 2001;314(2):279–291. pmid:11718561
  73. 73. Baker CH, Matsuda SP, Liu DR, Corey E. Molecular-cloning of the human gene encoding lanosterol synthase from a liver cDNA library. Biochemical and biophysical research communications. 1995;213(1):154–160. pmid:7639730
  74. 74. Thoma R, Schulz-Gasch T, D’arcy B, Benz J, Aebi J, Dehmlow H, et al. Insight into steroid scaffold formation from the structure of human oxidosqualene cyclase. Nature. 2004;432(7013):118–122. pmid:15525992
  75. 75. DeLano WL. The PyMOL molecular graphics system. 2002;.
  76. 76. Schwede T, Kopp J, Guex N, Peitsch MC. SWISS-MODEL: an automated protein homology-modeling server. Nucleic acids research. 2003;31(13):3381–3385. pmid:12824332
  77. 77. Hardin C, Pogorelov TV, Luthey-Schulten Z. Ab initio protein structure prediction. Current opinion in structural biology. 2002;12(2):176–181. pmid:11959494
  78. 78. Zhang CJ, Tang H, Li WC, Lin H, Chen W, Chou KC. iOri-Human: identify human origin of replication by incorporating dinucleotide physicochemical properties into pseudo nucleotide composition. Oncotarget. 2016;7(43):69783–69793. pmid:27626500
  79. 79. Lin H, Liang ZY, Tang H, Chen W. Identifying sigma70 promoters with novel pseudo nucleotide composition. IEEE/ACM transactions on computational biology and bioinformatics. 2017;. pmid:28186907
  80. 80. Liang ZY, Lai HY, Yang H, Zhang CJ, Yang H, Wei HH, et al. Pro54DB: a database for experimentally verified sigma-54 promoters. Bioinformatics. 2016; p. btw630.
  81. 81. Deng L, Zhang QC, Chen Z, Meng Y, Guan J, Zhou S. PredHS: a web server for predicting protein-protein interaction hot spots by using structural neighborhood properties. Nucleic Acids Research. 2014;42(Web Server issue):290–5. pmid:24852252
  82. 82. Garzón JI, Deng L, Murray D, Shapira S, Petrey D, Honig B. A computational interactome and functional annotation for the human proteome. Elife. 2016;5:e18715. pmid:27770567