Skip to main content

Validation of protein models by a neural network approach

Abstract

Background

The development and improvement of reliable computational methods designed to evaluate the quality of protein models is relevant in the context of protein structure refinement, which has been recently identified as one of the bottlenecks limiting the quality and usefulness of protein structure prediction.

Results

In this contribution, we present a computational method (Artificial Intelligence Decoys Evaluator: AIDE) which is able to consistently discriminate between correct and incorrect protein models. In particular, the method is based on neural networks that use as input 15 structural parameters, which include energy, solvent accessible surface, hydrophobic contacts and secondary structure content. The results obtained with AIDE on a set of decoy structures were evaluated using statistical indicators such as Pearson correlation coefficients, Z nat , fraction enrichment, as well as ROC plots. It turned out that AIDE performances are comparable and often complementary to available state-of-the-art learning-based methods.

Conclusion

In light of the results obtained with AIDE, as well as its comparison with available learning-based methods, it can be concluded that AIDE can be successfully used to evaluate the quality of protein structures. The use of AIDE in combination with other evaluation tools is expected to further enhance protein refinement efforts.

Background

The very large and continuously increasing amount of data obtained by genome sequencing makes the development of reliable computational methods capable to infer protein structures from sequences a crucial step for functional annotation of proteins. In fact, functional annotation is often strictly dependent on the availability of structural data, which in turn are still difficult to obtain experimentally. As a consequence, efforts and progresses in high throughput X-ray and NMR methods need to be accompanied by computational techniques suitable for three-dimensional structure predictions, such as homology modeling, fold recognition or ab-initio methods [1–7], which are intrinsically characterized by different levels of accuracy.

In parallel to the development and improvement of prediction methods, reliable and accurate evaluation tools are necessary to check the quality of computational protein models [8, 9]. Moreover, in the context of protein structure refinement, which has been recently identified as one of the bottlenecks limiting the quality and usefulness of protein structure prediction [1], it has been noted that improvements in the selection of the most native-like model from an ensemble of closely related alternative conformations can be crucial. The increasing importance of the field of quality assessment methods is demonstrated by the introduction of a dedicated section in the latest CASP edition (CASP7) [10].

To evaluate protein structures, several different scoring functions have been developed, which can be classified into different categories depending on the principles and on the structural features considered in the evaluation. Physical scoring (energy) functions aim to describe the physics of the interaction between atoms in a protein and are generally parameterized on molecular systems smaller than proteins [11]. Knowledge-based scoring functions are designed by evaluating the differences between some selected features of a random protein model and the characteristics of a real protein structure [12–16].

Learning-based functions can be developed by training algorithms to discriminate between correct and incorrect models [17]. Independently by the category, scoring functions are generally tested by examining their capability to detect the native structure among a set of decoys [18], which can be generated in several different ways [19–21].

It is important to note that the performance of learning-based functions are generally strongly dependent on the specific aim for which they were developed, and consequently on the training set used. As an example, ProQ, a neural network based method developed to predict the quality of protein models [17], was specifically designed to discriminate between correct and wrong models, i.e. to recognize folds that are not compatible with a protein sequence. In fact, ProQ was recently combined successfully with the Pcons [22] fold recognition predictor and ranked as one the best methods in a recent survey of quality assessment methods [10]. Other reliable and extensively used computational methods used to validate the quality of protein structures are PROSA [23], ERRAT [24], Verify 3D [25, 26], PROCHECK [27], what-if [28], PROVE [29] and victor/FRST [15].

In the present contribution, we present a computational method (A rtificial I ntelligence D ecoys E valuator: AIDE) that is able to reliably and consistently discriminate between correct and incorrect protein models. In particular, the quality of the protein structure is evaluated with neural networks using as input 15 structural parameters, which include solvent accessible surface, hydrophobic contacts and secondary structure content. In the first section of the paper, the neural network structure and the training procedure are presented and discussed. In the second section, the performance of the neural network is evaluated, compared to available methods, and critically discussed.

Results and Discussion

The evaluation of the quality of protein structures is generally carried out calculating a score which is a function of a set of parameter values computed for the protein model under study. In our computational procedure, the description of the relation between the parameters space and the scoring values is obtained using neural networks, because of their ability to describe complex non-linear relationships among data.

Selection of protein parameters related to structure quality

Among the possible parameters that can be computed for a protein structure, we have selected some properties that are expected to be related to structure quality: solvent accessible surface of hydrophobic and hydrophilic residues, secondary structure content, the fraction of secondary structure content of the model fitting with that predicted by PSIPRED [30], number of hydrophobic contacts, and selected PROCHECK parameters [27] (see Methods for details). It should be noted that other possibly relevant parameters, such as the number of hydrogen bonds, have not been used due to intrinsic difficulties in the normalisation of their values.

Selection of the parameters used to evaluate structure similarity

A key issue for evaluating the quality of a predicted protein structure is the measure of its "distance" relative to the "real" structure, experimentally obtained by X-ray diffraction or NMR. Since AIDE has been developed to evaluate protein models that are often characterized by the correct fold but may differ for local details, the backbone root mean square deviation (RMSD) of the protein model relative to the X-ray structure can be considered a suitable measure of structure similarity [31]. In fact, it is well known that the proper evaluation of the quality of protein structures can be a non-trivial task, often depending on the methods used to generate protein models. Therefore, several other measures of protein structure similarity have been formulated, the most commonly used being: GDT-TS [32], LG-score [32], TM-score [33] and MaxSub [34], which have also been adopted in the present work.

Selection and optimization of the neural networks

A preliminary evaluation of the relative importance of each parameter in the description of structure quality was obtained using a linear models built with the M5-prime attribute selection algorithm [35], as implemented in Weka 3.4.2 [36]. A different linear model was computed for each accuracy measure. Analysis of the linear models revealed that the secondary structure content and the solvent accessible surface have the highest importance in all models. Moreover, results show that it is not possible to exclude any parameter since non negligible weights are associated to all selected parameters, when the accuracy measures chosen are considered as a whole (Additional File 1).

The neural networks forming the core of AIDE are four layers feed forward neural networks with fifteen neurons (corresponding to the selected parameters) in the input layer, two hidden layers formed by two neurons each, and one neuron in the output layer. A linear activation function was chosen for all neurons. Indeed, different combinations of hidden layers (one or two) and different numbers of hidden neurons per layer (from two to ten nodes per layer) were tested. In addition, we tested also different activation functions of the neurons (sigmoid, log-sigmoid and linear functions). It turned out that, among the different combinations, the neural network featuring two hidden layers formed by two neurons gave the best results. In fact, an increase in the number of neurons led to poorer performances, probably due to the increased difficulties in the optimization procedure arising from the augmented network complexity. To carry out the optimization of neural networks, we have implemented the attractive-repulsive particle swarm optimization algorithm (AR-PSO) [37], as explained in Methods. Training of the neural networks using more conventional approaches (Gradient descent, Levenberg-Marquardt), led to slightly lower performances (Additional file 2). This may be due to the greater exploration ability that characterize the PSO methods.

AIDE was trained and tested on datasets of all-atoms protein decoys for which the three-dimensional structures are available. Since it is known that methods used for building decoys may introduce some systematic bias, it is important to benchmark a scoring function on different decoy sets in order to assess its generality. The overall dataset used in the present study is composed by an ensemble of widely used all-atom datasets containing models of different proteins (4state-reduced, fisa, fisa-casp3, rosetta all-atoms, CASP5, CASP7, Livebench2, lmds, and hg-structal [19–21, 38–40]), plus a molecular dynamics set that was generated in our laboratory from X-ray structures (see Methods).

After computation of the structural parameters to be inserted in the neural networks, the overall dataset was subdivided into a training and a test set, which were composed by 13693 and 49126 structures, respectively. The training-set includes only the proteins belonging to the LiveBench2 and CASP7 decoy sets (13693 model structures built on 96 different proteins). The test-set includes the lmds, CASP5, hg_structal, MD, Rosetta and 4state-reduced subsets (49126 models build on 97 proteins). The LiveBench2 and CASP7 decoy sets were chosen as training sets because they contain models build with different methods and of different protein size, ranging from 20 to 500 residues. No protein contained in the training set is present also in the test set.

Then, a population of 50 neural networks was trained starting from different initializations of the structural parameters. The network featuring the best performance (the highest correlation coefficient on the training set) was selected as the working network in AIDE.

A different neural network was trained for each measure of structure similarity chosen to evaluate proteins quality (RMSD, TM-score, GDT-TS, LG-score and MaxSub). Therefore, five different versions of AIDE were obtained from the training procedure, referred to in the following as AIDE RMSD, AIDE TM-score, AIDE GDT-TS, AIDE LG-score and AIDE MaxSub.

Assessment of AIDE performance

The performances of the different version of AIDE have been compared to results obtained from widely used methods developed to evaluate protein models quality.

The performances of the different methods were evaluated using a test-set which includes lmds, CASP5, hg_structal, MD, Rosetta and 4state-reduced subsets. The LiveBench2 and CASP7 sets were already used for training AIDE and therefore were not used in the comparative evaluation.

The Pearson correlation coefficient, Z nat and fraction enrichment (F.E.), which give indications about a method ability to assign good scores to good models, have been computed and results are collected in Tables 1, 2, 3.

Table 1 Pearson correlation coefficients. For each dataset belonging to the test-set the Pearson correlation coefficient between the predicted and the computed values is reported. The performance of AIDE is compared to that of ProQ and Victor/FRST validation softwares.
Table 2 10%-fraction enrichment. The 10%-fraction enrichment is shown for each dataset belonging to the test-set. The performance of AIDE is compared to that of ProQ and Victor/FRST validation softwares.
Table 3 Z nat . Comparison of Z nat values obtained using AIDE and other protein structure validation softwares. ProQ values have been obtained from Ref. 17.

Analysis of Pearson correlation coefficients (Table 1) shows that, according to this statistical indicator, the different AIDE versions behave quite similarly. Most importantly, average AIDE performances are similar or slightly better than those obtained by two state-of-the-art methods such as ProQ [17] and Victor [15]. It is also noteworthy that the performance of AIDE changes significantly moving through the different subsets forming the test-set. In particular, very high correlation coefficients are obtained with the MD and hg_structal datasets (correlation coefficient in the range 0.61–0.89 and 0.48–0.73, respectively), whereas low values of Pearson coefficients are associated to the CASP5 dataset (0.15–0.38). Relatively different values of Pearson correlation coefficients are obtained also with ProQ and Victor. In particular, and differently from AIDE, low correlation coefficients are obtained by ProQ for the Rosetta subset, and by Victor for the fisa subset (Table 1). The factors responsible for such non-homogeneous performances of the methods, when applied to different datasets, could not be unrevealed and might require further dissection of the test-set. In light of these results and observations it can be concluded that, even if the overall performances of AIDE, ProQ and Victor are similar, these methods can behave very differently on protein models obtained using different approaches, suggesting that the combined use of AIDE, ProQ and Victor could be useful to properly evaluate the quality of a protein structure.

Analysis of F. E. values (Table 2) shows again quite similar overall performances of AIDE, ProQ and Victor. However, the average F. E. values obtained using ProQ are consistently higher (by 5–10%) relative to the corresponding values obtained with Victor and AIDE. A more detailed analysis of F. E. values obtained from the different subsets composing the test set highlights some interesting trends. F. E. values obtained from the lmds and fisa subsets are consistently lower than the average. Moreover, AIDE and ProQ versions trained using different parameters to evaluate structure similarity can give quite different results. The latter observation is particularly evident for the lmds subset. It is also interesting to note that the best performances on the different subsets forming the test set are often obtained by different methods. As an example, the best F. E. values for the fisa subset are obtained using AIDE, whereas the best values for the hg_structal subset are obtained with ProQ, further suggesting that the combined use of the different methods can be a good strategy to obtain a more confident evaluation of the quality of a protein structure. Z nat allows to evaluate how (and if) the different methods distinguish the native (X-ray) structure from the ensemble of its models (Table 3). In this case it was possible to extend the comparison to other methods widely used to evaluate protein structures quality (Errat, Prosa II and Verify 3D). Only the lmds and 4state_reduced subsets have been used in this comparison because these are the only datasets in common among all the compared methods for which data are available. Analysis of Z nat values reveals that ProQ and Victor have better performances in this statistical test, whereas AIDE results are generally comparable to those obtained with Errat, Prosa II and Verify 3D. Notably, very low Z nat scores are obtained using AIDE RMDS and AIDE LG-score on the 4state_reduced subset.

It should be noted that Z nat and F.E. do not give information about the ability of a method to assign low scores to bad models, i. e. these statistical indicators do not allow to check if a method is confusing different classes. To explore this issue we have qualitatively compared AIDE and ProQ performances, superposing the ROC plots (see Methods) computed on the test-set for each different performance function (Figure 1). According to this analysis, ProQ MaxSub exhibits the greatest overall accuracy, whereas AIDE GDT-TS has the lowest accuracy.

Figure 1
figure 1

ROC plots. Comparison of AIDE and ProQ as obtained using a receiver operating characteristic (ROC) plot computed on the complete test-set (lmds, CASP5, hg_structal, MD, Rosetta and 4state-reduced datasets). Each line represents the ROC curve obtained with a specific AIDE or ProQ version. The ProQ MaxSub and the ProQ LG-score are plotted as solid:red and dashed:green line, respectively. The different AIDE versions are plotted as follow: GDT_TS dashed:blue; LG-score dotted:purple; MaxSub dashed-dotted:cyan; RMSD dashed-dotted:yellow; TM-score dashed:black.

Considering the different AIDE versions, a clear distinction can be observed when comparing the overall accuracy of AIDE RMSD and AIDE MaxSub relative to AIDE LGscore, AIDE GDT-TS and AIDE TMscore (Figure 1). Notably, a similar difference was not evident when considering the correlation coefficients or the fraction enrichment test. It is also important to note that AIDE LGscore behaves very similarly to ProQ LGscore until about 60% of sensitivity, whereas at higher sensitivity levels AIDE outperforms ProQ LGscore. These observations further corroborate the hypothesis that the combined use of ProQ and AIDE should give improved results in the evaluation of the quality of three-dimensional protein models.

The web interface of AIDE

The availability of five different AIDE versions gives a nice picture of the overall performance of the method. However, the overloading of output information can become a drawback for the user interested only in the most relevant results. In fact, the analysis of AIDE performance has shown that the five different versions of AIDE are generally characterised by similar behaviour (see Table 1, 2, 3). To better evaluate the degree of correlation among different AIDE versions we have carried out a principal component analysis on the Pearson correlation matrix of the descriptors chosen to evaluate models quality. This analysis reveals a strong correlation between TM-score, GDT-TS and MaxSub. The different clustering of TM-score, GDT-TS and MaxSub relative to RMSD and LG-score is mainly due to the inverse relationship between the two families(Additional files 3 and 4). Therefore, two (GDT-TS and the MaxSub) of these highly correlated parameters have been excluded from the output of the AIDE program available on the Internet [41]. Moreover, to help the user in the evaluation of AIDE results, we have defined a threshold for each predicted parameter, in order to discriminate between incorrect and correct models. In particular, correct models should have TM-score ≥ 0.31, RMSD ≤ 4.96 Å and LG-score ≤ 0.35. These thresholds were chosen using a dataset of manually assessed models composed by some CASP5 targets belonging to the new fold and fold recognition categories. According to the visual evaluation of Aloy and coworkers [42], the models were divided into three class: class 2 ("excellent") when the overall fold is correct, class 1 ("good") when the model is considered partway to the correct fold, and class 0 for all the other models. For each model, the TM-score, LG-score and RMSD were computed (Additional files 5, 6, 7) and the average value for the models belonging to the "excellent" class was used as threshold. To further evaluate the classification ability using the chosen thresholds, the sensitivity and the specificity based on the ROC plots were also computed (Additional file 8).

Conclusion

In this paper we have presented AIDE, a neural network system which is able to evaluate the quality of protein structures obtained by prediction methods.

AIDE differs from other evaluation methods mainly for : i) a different choice of the parameters used to describe the protein structure, ii) a different choice of the parameters related to structure quality, iii) a novel strategy used to optimize the neural networks. AIDE overall performances are comparable to recently published state of the art methods, such as ProQ [17] and Victor [15]. However, detailed comparative analysis of results obtained using AIDE, ProQ and Victor reveals that the three methods have different and often complementary ability to properly assess the quality of protein structures. This observation suggests that the combined use of AIDE, ProQ and Victor could increase the reliability in the evaluation of protein structures quality. AIDE is presently available on the Internet [41].

Methods

Protein datasets

The 4state-reduced set is an all-atom version of the models generated by Park & Levitt [19] using a four-state off-lattice model.

The fisa and fisa-casp3 sets contain decoys of four small alpha-helix proteins. In these sets main chains were generated using a procedure of fragment insertion based on simulated annealing: native-like structures were assembled from a combination of fragments of known unrelated protein structures characterized by similar local sequences, using Bayesian scoring functions [20]. The side chains of fisa and fisa-casp3 were modeled with the software package SCWRL [43].

The hg-structal is a set of hemoglobin models generated by homology modelling.

The lmds subset [19, 44] was produced by Keasar and Levitt by geometry optimizations carried out using a complex potential that contains a pairwise component, as well as cooperative hydrogen bonds terms. The Rosetta all-atom decoys were generate with the ROSETTA method developed by David Baker [21]. The molecular dynamics set of decoys was generated by molecular dynamics (MD) simulations carried out in vacuum with the software GROMACS 3.2 [45, 46]. Each protein structure was submitted to 100 ps of simulation using the OPLS force field [47]. MD simulations were performed in the NVT ensemble at 600 K, using an external bath with a coupling constant of 0.1 ps [48]. The LINCS algorithm [49] was adopted to constrain bond lengths of heavy atoms, allowing us to use a 2 fs time step. Van der Waals and Coulomb interactions were truncated at 8 Ã…, while long-range electrostatics interactions were evaluated using the particle mesh ewald summation scheme [50]. The Van der Waals radii were increased to 4 Ã… for all atoms, in order to speed up the unfolding process [51].

Snapshots from the trajectory have been extracted every 0.4 ps, collecting 250 misfolded structures for each protein, with a backbone RMSD (root mean square deviation between the initial structure and each snapshot) ranging from 0 to about 10 Ã….

In addition to these decoy datasets, the CASP5 [38], CASP7 [39] and LiveBench2 [40] sets were also included.

The complete dataset contains 62819 protein models build on 193 proteins.

Training-set and test-set

The dataset was splitted into two disjoint sets : a training-set and a test-set. The training-set includes only proteins belonging to the LiveBench2 and CASP7 decoys sets (13693 model structures built on 96 different proteins). The test-set includes the lmds, CASP5, hg_structal, MD, Rosetta and 4state-reduced datasets (49126 models build on 97 proteins).

Parameters-Descriptors used in the neural network

The relative solvent accessible surface (rSAS) was computed as follow:

rSAS hydrophobic = SAS hydrophobic SAS total
rSAS hydrophobic = SAS hydrophobic SAS total MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGacaGaaiaabeqaaeqabiWaaaGcbaGaemOCaiNaem4uamLaemyqaeKaem4uam1aaSbaaSqaaiabdIgaOjabdMha5jabdsgaKjabdkhaYjabd+gaVjabdchaWjabdIgaOjabdMgaPjabdYgaSjabdMgaPjabdogaJbqabaGccqGH9aqpjuaGdaWcaaqaaiabdofatjabdgeabjabdofatnaaBaaabaGaemiAaGMaemyEaKNaemizaqMaemOCaiNaem4Ba8MaemiCaaNaemiAaGMaemyAaKMaemiBaWMaemyAaKMaem4yamgabeaaaeaacqWGtbWucqWGbbqqcqWGtbWucaWLa8+aaSbaaeaacqWG0baDcqWGVbWBcqWG0baDcqWGHbqycqWGSbaBaeqaaaaaaaa@60A4@

where the residues A, L, V, I, P, F, M, W were considered as hydrophobic and the SAS total is the total solvent accessible surface computed using NACCESS [52].

The secondary structure was evaluated with the DSSP program [53], in which the typical 8-state DSSP definition was simplified according to the following rules : H and G to helix, E and B to strand and all other states considered as coil, in agreement with PSIPRED definition [30].

The fraction of secondary structure (SS) is defined as:

S S = n s s N ∗ 100 , MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGacaGaaiaabeqaaeqabiWaaaGcbaGaem4uamLaem4uamLaeyypa0tcfa4aaSaaaeaacqWGUbGBdaWgaaqaaiabdohaZjabdohaZbqabaaabaGaemOta4eaaOGaey4fIOIaeGymaeJaeGimaaJaeGimaaJaeiilaWcaaa@3A51@

where n ss is the number of residues located in well-defined secondary structure elements, and N is the number of protein residues.

The secondary structure for each decoy was also compared with the corresponding secondary structure predicted by PSIPRED. Accordingly, the relative consensus secondary structure (SSc) was defined as the ratio:

S S c = n c N ∗ 100 , MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGacaGaaiaabeqaaeqabiWaaaGcbaGaem4uamLaem4uamLaem4yamMaeyypa0tcfa4aaSaaaeaacqWGUbGBdaWgaaqaaiabdogaJbqabaaabaGaemOta4eaaOGaey4fIOIaeGymaeJaeGimaaJaeGimaaJaeiilaWcaaa@3A11@

where n c is the number of residues located in corresponding secondary structure elements according to DSSP definition and PSIPRED secondary structure prediction.

Generally, native structures are characterized by hydrophobic residues clustered in buried regions. Therefore, the number of contacts between hydrophobic residues was chosen as a possible relevant parameter to discriminate among correct and incorrect models. According to our definition, a contact is present if the distance between two residues is greater than 2.5 Ã… and lower or equal to 5 Ã… [54]. Given n hydrophobic residues, the number of hydrophobic contacts (Q) is normalized relative to the number of all possible contacts:

r Q = 2 Q n ( n − 1 ) ! , MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGacaGaaiaabeqaaeqabiWaaaGcbaGaemOCaiNaemyuaeLaeyypa0tcfa4aaSaaaeaacqaIYaGmcqWGrbquaeaacqWGUbGBcqGGOaakcqWGUbGBcqGHsislcqaIXaqmcqGGPaqkcqGGHaqiaaGccqGGSaalaaa@3A87@

Moreover, to keep into account the stereochemical quality of the model, some PROCHECK parameters were considered (Table 4).

Table 4 PROCHECK parameters. PROCHECK parameters used in AIDE. The G-factor, which is a log-odds score based on the observed distributions of stereochemical parameters, provides a measure of how "normal", or alternatively how "unusual", a given stereochemical property is.

Model accuracy measures

Quality of protein models was evaluated by means of five different descriptors, using the crystal structure as reference: RMSD on the backbone atoms, TM-score [33], GDT-TS [32], LG-score [32] and MaxSub [34]. RMSD was computed on the backbone atoms after superposing the model structure on the crystal structure, using the program CE [55].

TM-score was developed to evaluate the topology similarity of two protein structures [33]. TM-score values fall into the interval [0, 1]. Scores equal or below 0.17 indicate that the prediction has a reliability compared to a random selection from the PDB library.

GDT-TS gives an estimation of the largest number of residues that can be found in which all distances between the model and the reference structure are shorter than the cutoff D. The number of residues is measured as a percentage of the length of the target structure. The values of GDT-TS fall into the interval [0–1], with a GDT-TS of 1 corresponding to perfect superposition.

The LG-score represents the significance (P-value) of a score (S str [56]) associated to the best subpart of a structural alignment between the model and the correct structure. The value is measured by using a structural P-value ranging from 0 to 1, with a value of 0 corresponding to optimal superposition.

MaxSub is calculated from the largest number of residues that superimpose well over the reference structure, and produces a normalized score that ranges between 0 and 1. A MaxSub value of 1 is associated to perfect superposition.

Neural network

Four layers feed forward neural networks were used, with fifteen neurons in the input layer, two neurons in two hidden layers and one neuron in the output layer. A linear activation function was chosen for all neurons.

For each accuracy measure chosen to evaluate proteins quality (RMSD, TM-score, GDT-TS, LG-score and MaxSub) a different neural network was trained.

The inverse of the Pearson correlation coefficient (CC) between the true and the predicted data was used as performance function.

{ C C } − 1 = { ( t − μ t ) T ( y − μ y ) ( M − 1 ) σ t σ y } − 1 MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGacaGaaiaabeqaaeqabiWaaaGcbaGaei4EaSNaem4qamKaem4qamKaeiyFa03aaWbaaSqabeaacqGHsislcqaIXaqmaaGccqGH9aqpdaGadaqcfayaamaalaaabaGaeiikaGccbmGae8hDaqNaeyOeI0ccciGae4hVd02aaSbaaeaacqWG0baDaeqaaiabcMcaPmaaCaaabeqaaiabdsfaubaacqGGOaakcqWF5bqEcqGHsislcqGF8oqBdaWgaaqaaiabdMha5bqabaGaeiykaKcabaGaeiikaGIaemyta0KaeyOeI0IaeGymaeJaeiykaKIae43Wdm3aaSbaaeaacqWG0baDaeqaaiab+n8aZnaaBaaabaGaemyEaKhabeaaaaaakiaawUhacaGL9baadaahaaWcbeqaaiabgkHiTiabigdaXaaaaaa@54B1@

where t is the vector of predicted values for each decoy, y is the vector of true values, μ t , σ t , μ y , σ y are the averages and the standard deviations of predicted and true values, respectively, and M is the number of decoys.

Optimization of neural networks was carried out using the attractive-repulsive particle swarm optimization algorithm (AR-PSO) [37], which is a modification of the original PSO method [57, 58]. PSO is a stochastic population-based optimization approach which explores the hyper-dimensional parameters space of a population of candidate solutions named particles. Particles fly over the solution space looking for the global optimum. Each particle retains an individual memory of the best position visited and a global memory of the best position visited by all the particles.

A particle calculates its next position combining information from its last movement, the individual memory, the global memory and a random component.

The PSO updating rule is described as follow:

{ v i , t + 1 = μ v i , t + c 1 ( w i , t b e s t − w i , t ) + c 2 ( w t g l o b a l − w i , t ) w i , t + 1 = w i , t + v i , t + 1 MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGacaGaaiaabeqaaeqabiWaaaGcbaWaaiqaaeaafaqaaeGabaaabaacbmGae8NDay3aaSbaaSqaaiabdMgaPjabcYcaSiabdsha0jabgUcaRiabigdaXaqabaGccqGH9aqpiiGacqGF8oqBcqWF2bGDdaWgaaWcbaGaemyAaKMaeiilaWIaemiDaqhabeaakiabgUcaRiabdogaJnaaBaaaleaacqaIXaqmaeqaaOGaeiikaGIae83DaC3aa0baaSqaaiabdMgaPjabcYcaSiabdsha0bqaaiabdkgaIjabdwgaLjabdohaZjabdsha0baakiabgkHiTiab=Dha3naaBaaaleaacqWGPbqAcqGGSaalcqWG0baDaeqaaOGaeiykaKIaey4kaSIaem4yam2aaSbaaSqaaiabikdaYaqabaGccqGGOaakcqWF3bWDdaqhaaWcbaGaemiDaqhabaGaem4zaCMaemiBaWMaem4Ba8MaemOyaiMaemyyaeMaemiBaWgaaOGaeyOeI0Iae83DaC3aaSbaaSqaaiabdMgaPjabcYcaSiabdsha0bqabaGccqGGPaqkaeaacqWF3bWDdaWgaaWcbaGaemyAaKMaeiilaWIaemiDaqNaey4kaSIaeGymaedabeaakiabg2da9iab=Dha3naaBaaaleaacqWGPbqAcqGGSaalcqWG0baDaeqaaOGaey4kaSIae8NDay3aaSbaaSqaaiabdMgaPjabcYcaSiabdsha0jabgUcaRiabigdaXaqabaaaaaGccaGL7baaaaa@7EB7@

in which wi, t+1represents the position vector of the particle i at time t (i.e. the neural network weights), w i , t b e s t MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGacaGaaiaabeqaaeqabiWaaaGcbaacbmGae83DaC3aa0baaSqaaiabdMgaPjabcYcaSiabdsha0bqaaiabdkgaIjabdwgaLjabdohaZjabdsha0baaaaa@36AB@ is the best position identified by the particle i so far (i.e. the neural network weights associated with the best performance value) and w t g l o b a l MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGacaGaaiaabeqaaeqabiWaaaGcbaacbmGae83DaC3aa0baaSqaaiabdsha0bqaaiabdEgaNjabdYgaSjabd+gaVjabdkgaIjabdggaHjabdYgaSbaaaaa@3708@ is the best position identified among all the particles. The vector v represents the particles velocity, which is computed as the difference between two positions and assuming unitary time.

The term ( w i , t b e s t MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGacaGaaiaabeqaaeqabiWaaaGcbaacbmGae83DaC3aa0baaSqaaiabdMgaPjabcYcaSiabdsha0bqaaiabdkgaIjabdwgaLjabdohaZjabdsha0baaaaa@36AB@ - wi, t) represents the individual memory component and ( w t g l o b a l MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGacaGaaiaabeqaaeqabiWaaaGcbaacbmGae83DaC3aa0baaSqaaiabdsha0bqaaiabdEgaNjabdYgaSjabd+gaVjabdkgaIjabdggaHjabdYgaSbaaaaa@3708@ - wi, t) is the global one. These two terms are rescaled by the random coefficients c1 and c2, respectively. The μ coefficient is used to rescale the velocity.

Starting particle positions and velocities were initialized at random. To reduce the problem of premature convergence to relative minima, the Attractive-Repulsive modification has been introduced [37]. This modification defines a measure of global diversity (D) among the particles as:

D = 1 S ∑ i = 1 S ∑ j = 1 N ( w i j − w ¯ j ) 2 MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGacaGaaiaabeqaaeqabiWaaaGcbaGaemiraqKaeyypa0tcfa4aaSaaaeaacqaIXaqmaeaacqWGtbWuaaGcdaaeWbqaamaakaaabaWaaabCaeaacqGGOaakcqWG3bWDdaWgaaWcbaGaemyAaKMaemOAaOgabeaakiabgkHiTiqbdEha3zaaraWaaSbaaSqaaiabdQgaQbqabaGccqGGPaqkdaahaaWcbeqaaiabikdaYaaaaeaacqWGQbGAcqGH9aqpcqaIXaqmaeaacqWGobGta0GaeyyeIuoaaSqabaaabaGaemyAaKMaeyypa0JaeGymaedabaGaem4uamfaniabggHiLdaaaa@49C5@

where S is the number of particles in the swarm, N is space dimension (the number of networks weights) and w ¯ j MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGacaGaaiaabeqaaeqabiWaaaGcbaGafm4DaCNbaebadaWgaaWcbaGaemOAaOgabeaaaaa@2F39@ is the average of the parameter j among the particles.

If D falls below a minimal threshold (t min ) the update rule is inverted as follow

{ v i , t + 1 = μ v i , t + ( − 1 ) c 1 ( w i , t b e s t − w i , t ) + ( − 1 ) c 2 ( w t g l o b a l − w i , t ) w i , t + 1 = w i , t + v i , t + 1 MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGacaGaaiaabeqaaeqabiWaaaGcbaWaaiqaaeaafaqaaeGabaaabaacbmGae8NDay3aaSbaaSqaaiabdMgaPjabcYcaSiabdsha0jabgUcaRiabigdaXaqabaGccqGH9aqpiiGacqGF8oqBcqWF2bGDdaWgaaWcbaGaemyAaKMaeiilaWIaemiDaqhabeaakiabgUcaRiabcIcaOiabgkHiTiabigdaXiabcMcaPiabdogaJnaaBaaaleaacqaIXaqmaeqaaOGaeiikaGIae83DaC3aa0baaSqaaiabdMgaPjabcYcaSiabdsha0bqaaiabdkgaIjabdwgaLjabdohaZjabdsha0baakiabgkHiTiab=Dha3naaBaaaleaacqWGPbqAcqGGSaalcqWG0baDaeqaaOGaeiykaKIaey4kaSIaeiikaGIaeyOeI0IaeGymaeJaeiykaKIaem4yam2aaSbaaSqaaiabikdaYaqabaGccqGGOaakcqWF3bWDdaqhaaWcbaGaemiDaqhabaGaem4zaCMaemiBaWMaem4Ba8MaemOyaiMaemyyaeMaemiBaWgaaOGaeyOeI0Iae83DaC3aaSbaaSqaaiabdMgaPjabcYcaSiabdsha0bqabaGccqGGPaqkaeaacqWF3bWDdaWgaaWcbaGaemyAaKMaeiilaWIaemiDaqNaey4kaSIaeGymaedabeaakiabg2da9iab=Dha3naaBaaaleaacqWGPbqAcqGGSaalcqWG0baDaeqaaOGaey4kaSIae8NDay3aaSbaaSqaaiabdMgaPjabcYcaSiabdsha0jabgUcaRiabigdaXaqabaaaaaGccaGL7baaaaa@85D5@

causing the particles to spread in the phase space. If D reaches a maximal threshold (t max ) the update rule is restored as in the standard PSO method. We choose t min = 0.1 and t max = 5.0.

The parameters c 1, c 2 and μ were set as in the original PSO method as c 1 = c 2 ∈ [0.0, 2.0] and μ = 0.7298. The maximum number of iterations was set to 10000. A population size of 5 particles was chosen. It should be noted that standard training algorithms such as gradient descent back-propagation, Levenberg-Marquardt and BFGS, led to poorer results when compared to the particle swarm optimization (data not shown).

Statistical analysis

The following statistical parameters were used: Pearson correlation coefficient, already described in the neural network section, fraction enrichment (F.E.) and Z_nat.

Fraction enrichment (F.E.) is defined as the fraction of the top 10% conformations featuring best structural resemblance to the native structure among the top 10% best scoring conformations.

Z nat is the Z-score of the X-ray structure compared to the ensemble of decoys structures. It is computed using the following equation:

Z n a t = s c o r e n a t i v e − μ d e c o y s σ d e c o y s MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGacaGaaiaabeqaaeqabiWaaaGcbaGaemOwaO1aaSbaaSqaaiabd6gaUjabdggaHjabdsha0bqabaGccqGH9aqpjuaGdaWcaaqaaiabdohaZjabdogaJjabd+gaVjabdkhaYjabdwgaLnaaBaaabaGaemOBa4MaemyyaeMaemiDaqNaemyAaKMaemODayNaemyzaugabeaacqGHsisliiGacqWF8oqBdaWgaaqaaiabdsgaKjabdwgaLjabdogaJjabd+gaVjabdMha5jabdohaZbqabaaabaGae83Wdm3aaSbaaeaacqWGKbazcqWGLbqzcqWGJbWycqWGVbWBcqWG5bqEcqWGZbWCaeqaaaaaaaa@57D5@

Higher Z nat values correspond to higher capacity to discriminate between the native structure and the corresponding decoys.

The Receiver Operating Characteristic (ROC) graph is a plot of all sensitivity/specificity pairs resulting from continuously varying the decision threshold over the range of results observed. The sensitivity or true positive fraction is reported on the y-axis, while the x-axis represents the 1-specificity or true negative fraction. A test with perfect discrimination (no overlap between the two distribution of results) has a plot curve that passes through the upper left corner, where both specificity and sensitivity are 1.00. The ipotetical plot of a test with no discrimination between the two groups is a 45° line going from the lower left to the upper right corner.

Qualitatively, the closer the plot is to the upper left corner, the higher the overall accuracy of the test.

Availability and requirements

Project name: AIDE

Project home page: http://linux.btbs.unimib.it/cgi-bin/aide.cgi

Operating system(s): Linux, Unix

Programming language: Fortran77/90, Perl

Licence: GNU GPLv3

Any restrictions to use by non-academics: No restrictions

References

  1. Tress M, Ezkurdia I, Grana O, Lopez G, A V: Assessment of predictions submitted for the CASP6 comparative modeling category. Proteins: Structure, Function, and Bioinformatics 2005, 61(Suppl 7):27–45.

    Article  CAS  Google Scholar 

  2. Bradley P, Malmstrom L, Qian B, Schonbrun J, Chivian D, Kim D, Meiler J, Misura K, D B: Free modeling with Rosetta in CASP6. Proteins: Structure, Function, and Bioinformatics 2005, 61(Suppl 7):128–134.

    Article  CAS  Google Scholar 

  3. Soonming J, Eunae K, Seokmin S, P Y: Ab inition folding of helix bundle proteins using molecular dynamics simulations. JACS 2003, 125: 14841–14846.

    Article  Google Scholar 

  4. Andrzej Kolinacuteski JMB: Generalized protein structure prediction based on combination of fold-recognition with de novo folding and evaluation of models. Proteins: Structure, Function, and Bioinformatics 2005, 61(Suppl 7):84–90.

    Google Scholar 

  5. Moult J, Fidelis K, Rost B, Hubbard T, Tramontano A: Critical assessment of methods of protein structure prediction (CASP)-round 6. Proteins: Structure, Function, and Bioinformatics 2005, 61(Suppl 7):3–7.

    Article  CAS  Google Scholar 

  6. Xu J, Yu L, Li M: Consensus fold recognition by predicted model quality. APBC 2005, 73–83.

    Google Scholar 

  7. Xu J: Fold Recognition by Predicted Alignment Accuracy. IEEE/ACM Trans Comput Biology Bioinform 2005, 2(2):157–165.

    Article  CAS  Google Scholar 

  8. Moult J: A decade of CASP: progress, bottlenecks and prognosis in protein structure prediction. Curr Opin Struct Biol 2005, 15: 285–289.

    Article  CAS  PubMed  Google Scholar 

  9. Kryshtafovych A, Venclovas C, Fidelis K, Moult J: Progress Over the First Decade of CASP Experiments. Proteins: Structure, Function, and Bioinformatics 2005, 61(Suppl 7):225–267.

    Article  CAS  Google Scholar 

  10. Tramontano A: An account of the Seventh Meeting of the Worldwide Critical Assessment of Techniques for Protein Structure Prediction. FEBS Journal 2007, 274(7):1651–1654.

    Article  CAS  PubMed  Google Scholar 

  11. Lazaridis T, Karplus M: Effective energy functions for protein structure prediction. Curr Opin Struct Biol 2000, 10: 139–145.

    Article  CAS  PubMed  Google Scholar 

  12. Sippl M: Recognition of errors in three-dimensional structures of proteins. Proteins 1993, 17: 355–362.

    Article  CAS  PubMed  Google Scholar 

  13. Sippl M: Knowledge based potential for proteins. Curr Opin Struct Biol 1995, 5: 229–235.

    Article  CAS  PubMed  Google Scholar 

  14. Melo F, Feytmans : Novel knowledge-based mean force potential at atomic level. J Mol Biol 1997, 267: 207–222.

    Article  CAS  PubMed  Google Scholar 

  15. Tosatto S: The Victor/FRST Function for Model Quality Estimation. Journal of Computational Biology 2005, 12: 1316–1327.

    Article  CAS  PubMed  Google Scholar 

  16. Melo F, Sanchez R, Sali A: Statistical potentials for fold assessment. Protein Science 2002, 11: 430–448.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  17. Wallner B, Elofsson A: Can correct protein models be identified? Protein Science 2003, 12: 1073–1086.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  18. Samudrala R, Levitt M: Decoys R Us: A database of incorrect conformations to improve protein structure prediction. Protein Science 2000, 9: 1399–1401.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  19. Park B, Levitt M: Energy functions that discriminate X-ray and near native folds from well-constructed decoys. J Mol Biol 1996, 258: 367–392.

    Article  CAS  PubMed  Google Scholar 

  20. Simons K, Kooperberg C, Huang E, Baker D: Assembly of protein tertiary structures from fragments with similar local sequences using simulated annealing and bayesian scoring functions. J Mol Biol 1997, 268: 209–225.

    Article  CAS  PubMed  Google Scholar 

  21. Simons K, Bonneau R, Ruczinski I, Baker D: Ab initio protein structure prediction of CASP III targets using ROSETTA Proteins. Proteins 1999, (Suppl 3):171–176.

    Google Scholar 

  22. Lundstrom J, Rychlewski L, Bujnicki J, Elofsson A: Pcons: a neural-network-based consensus predictor that improves fold recognition. Protein Sci 2001, 10: 2354–2362.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  23. Sippl MJ: Recognition of errors in three-dimensional structures of proteins. Proteins 1993, 17: 355–362.

    Article  CAS  PubMed  Google Scholar 

  24. Colovos C, Yeates TO: Verification of protein structures: patterns of nonbonded atomic interactions. Protein Sci 1993, 2: 1511–1519.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  25. Bowie JU, Luthy R, Eisenberg D: A method to identify protein sequences that fold into a known three-dimensional structure. Science 1991, 253: 164–170.

    Article  CAS  PubMed  Google Scholar 

  26. Luthy R, Bowie JU, Eisenberg D: Assessment of protein models with three-dimensional profiles. Nature 1992, 356: 83–85.

    Article  CAS  PubMed  Google Scholar 

  27. Laskowski RA, MacArthur MW, Moss DS, Thornton JM: PROCHECK: a program to check the stereochemical quality of protein structures. J Appl Cryst 1993, 26: 283–291.

    Article  CAS  Google Scholar 

  28. Vriend G: WHAT IF: a molecular modeling and drug design program. J Mol Graph 1990, 8: 52–56.

    Article  CAS  PubMed  Google Scholar 

  29. Pontius J, Richelle J, Wodak SJ: Deviations from standard atomic volumes as a quality measure for protein crystal structures. J Mol Biol 1996, 264: 121–136.

    Article  CAS  PubMed  Google Scholar 

  30. Jones DT: Protein secondary structure prediction based on position-specific scoring matrices. J Mol Biol 1999, 292: 195–202.

    Article  CAS  PubMed  Google Scholar 

  31. Eramian D, yi Shen M, Devos D, Melo F, Sali A, Marti-Renom MA: A composite score for predicting errors in protein structure models. Protein Sci 2006, 15(7):1653–1666.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  32. Cristobal S, Zemla A, Fischer D, Rychlewski L, Elofsson A: A study of quality measures for protein threading models. BMC Bioinformatics 2001, 2: 5.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  33. Zhang Y, Skolnick J: Scoring function for automated assessment of protein structure template quality. Proteins 2004, 57(4):702–710.

    Article  CAS  PubMed  Google Scholar 

  34. Siew N, Elofsson A, Rychlewski L, Fischer D: MaxSub: an automated measure for the assessment of protein structure prediction quality. Bioinformatics 2000, 16(9):776–785.

    Article  CAS  PubMed  Google Scholar 

  35. Induction of model trees for predicting continuous classes 1997.

  36. S G: WEKA: The Waikato Environment for Knowledge Analysis. University of Waikato, Hamilton, New Zealand: University of Waikato; 1995.

    Google Scholar 

  37. Riget J, Vesterstrom S: A Diversity-Guided Particle Swarm Optimizer – the ARPSO. 2002.

    Google Scholar 

  38. CASP5[http://predictioncenter.genomecenter.ucdavis.edu/casp5/]

  39. CASP7[http://www.predictioncenter.org/casp7/Casp7.html]

  40. Bujnicki JM, Elofsson A, Fischer D, Rychlewski L: LiveBench-2: large-scale automated evaluation of protein structure prediction servers. Proteins 2001, (Suppl 5):184–191.

    Google Scholar 

  41. AIDE : Artificial Intelligence Decoys Evaluator[http://linux.btbs.unimib.it/cgi-bin/aide.cgi]

  42. Aloy P, Stark A, Hadley C, Russell R: Prediction wihout templates: new fold, secondary structure, and contacts in CASP5. Proteins 2003, 53(Suppl 6):436–456.

    Article  CAS  PubMed  Google Scholar 

  43. Bower M, Cohen F, Dunbrack R: Prediction of protein side-chain rotamer from a backbone dependent rotamer library: a new homology modelling tool. J Mol Biol 1997, 267: 1268–1282.

    Article  CAS  PubMed  Google Scholar 

  44. Fain B, Xia Y, Levitt M: Design of an optimal Chebyshev-expanded discrimination function for globular proteins. Protein Sci 2002, 11: 2010–2021.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  45. Lindahl E, Hess B, van der Spoel D: GROMACS 3.0: A package for molecular simulation and trajectory analysis. J Mol Biol 2001, 7: 306–317.

    CAS  Google Scholar 

  46. Berendsen H, van der Spoel D, van Drunen R: GROMACS: A message passing parallel molecular dynamics implementation. Comp Phys Comm 1995, 91: 43–56.

    Article  CAS  Google Scholar 

  47. Jorgensen W, Tirado-Rives J: The OPLS potential functions for proteins. energy minimizations for crystals of cyclic peptides and crambin. J Am Chem Soc 1988, 110: 1657–1666.

    Article  CAS  Google Scholar 

  48. Berendsen H, Postma J, Dinola A, JR H: MD with coupling to an external bath. J Phys Chem 1984, 81: 3684–3690.

    Article  CAS  Google Scholar 

  49. Hess B, Bekker H, Berendsen H, JGEM F: LINCS: A linear constraint solver for molecular simulations. J Comp Chem 1997, 18: 1463–1472.

    Article  CAS  Google Scholar 

  50. Essman U, Perela L, Berkowitz M, Darden T, Lee H, Pederson L: A smooth particle mesh Ewald method. J Chem Phys 1995, 103: 8577–8592.

    Article  Google Scholar 

  51. Lazaridis T, Karplus M: Discrimination of the native from misfolded protein models with an energy function including implicit solvation. J Mol Biol 1999, 288: 477–487.

    Article  CAS  PubMed  Google Scholar 

  52. Hubbard SJ, Thornton JM: NACCESS Computer Program. Department of Biochemistry and Molecular Biology, University College London; 1993.

    Google Scholar 

  53. Kabsch W, Sander C: Dictionary of Protein Secondary-Structure: Pattern Recognition of Hydrogen-Bonded and Geometrical Features. Biopolymers 1983, 22: 2577–2637.

    Article  CAS  PubMed  Google Scholar 

  54. Salerno WJ, Seaver SM, Armstrong BR, Radhakrishnan I: MONSTER: inferring non-covalent interactions in macromolecular structures from atomic coordinate data. Nucleic Acids Res 2004, 32: 566–568.

    Article  Google Scholar 

  55. Shindyalov IN, Bourne PE: Protein structure alignment by incremental combinatorial extension (CE) of the optimal path. Protein Eng 1998, 11: 739–747.

    Article  CAS  PubMed  Google Scholar 

  56. Levitt M, Gerstein M: A unified statistical framework for sequence comparison and structure comparison. Proc Natl Acad Sci USA 1998, 95: 5913–5920.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  57. Kennedy J, Eberhart RC: Particle swarm optimization. In Proc IEEE Int'l Conf on Neural Networks, IV, 1942–1948. Piscataway, NJ; 1995:1942–94.

    Chapter  Google Scholar 

  58. Eberhart RC, Kennedy J: A new optimizer using particle swarm theory. In Proceedings of the Sixth International Symposium on Micromachine and Human Science. Nagoya, Japan; 1995:39–43.

    Chapter  Google Scholar 

Download references

Acknowledgements

The authors are grateful to CINECA (Project 562)-2007 for the use of computational facilities

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Luca De Gioia.

Additional information

Authors' contributions

MP carried out the computational work. MP, PE and DGL carried out the data analysis. All authors contributed to the design of the study. All authors read and revised the manuscript.

Electronic supplementary material

12859_2007_2051_MOESM1_ESM.pdf

Additional file 1: Linear models. Linear models obtained with linear regression and M5-prime attribute selection algorithm. (PDF 23 KB)

12859_2007_2051_MOESM2_ESM.pdf

Additional file 2: Comparison of neural network performances. Comparison of neural network performances using different training algorithms. (PDF 15 KB)

12859_2007_2051_MOESM3_ESM.PDF

Additional file 3: Accuracy parameters loading plot. Loading plot of the accuracy parameters correlation matrix obtained by principal component analysis. Given the accuracy parameters matrix, where each row represents a different model and the columns are the descriptors used as quality measure (GDT_TS, LG-score, MaxSub, RMSD and TM-score), the correlation matrix (see additional file 5) was computed and analyzed by principal component analysis. Only the first two principal components are plotted. (PDF 9 KB)

12859_2007_2051_MOESM4_ESM.pdf

Additional file 4: Accuracy parameters correlation matrix. Pearson correlation matrix of predicted accuracy parameters for the test-set. (PDF 10 KB)

12859_2007_2051_MOESM5_ESM.png

Additional file 5: Manual assessment of CASP5 models vs TM-score. Manual assessment of different models of 13 targets of CASP5 belonging to the category of "new fold" and "fold recognition". Each model has been classified into one of the following three classes : "excellent", "good" and "bad", and showed in the figure as blue, green and red circles, respectively [42]. Each target is represented into a different panel, where the horizontal axes indicates the model number and the vertical axes is the TM-score. (PNG 54 KB)

12859_2007_2051_MOESM6_ESM.png

Additional file 6: Manual assessment of CASP5 models vs LG-score. Manual assessment of different models of 13 targets of CASP5 belonging to the category of "new fold" and "fold recognition". Each model has been classified into three classes : "excellent", "good" and "bad", and showed in the figure as blue, green and red circles, respectively [42]. Each target is represented into a different panel where the horizontal axes indicates the model number and the vertical axes is the LG-score. (PNG 236 KB)

12859_2007_2051_MOESM7_ESM.png

Additional file 7: Manual assessment of CASP5 models vs RMSD. Manual assessment of different models of 13 target of CASP5 belonging to the category of "new fold" and "fold recognition". Each model has been classified into three classes : "excellent", "good" and "bad" showed in the figure as blue, green and red circles, respectively [42]. Each target is represented into a different panel where the horizontal axes indicates the model number and the vertical axes is the RMSD. (PNG 234 KB)

12859_2007_2051_MOESM8_ESM.PNG

Additional file 8: ROC curves derived accuracy. Sensitivity and specificity of AIDE TM-score, AIDE RMSD and AIDE LG-score, as obtained from the ROC curves at the chosen threshold. (PNG 220 KB)

Authors’ original submitted files for images

Rights and permissions

Open Access This article is published under license to BioMed Central Ltd. This is an Open Access article is distributed under the terms of the Creative Commons Attribution License ( https://creativecommons.org/licenses/by/2.0 ), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and permissions

About this article

Cite this article

Mereghetti, P., Ganadu, M.L., Papaleo, E. et al. Validation of protein models by a neural network approach. BMC Bioinformatics 9, 66 (2008). https://doi.org/10.1186/1471-2105-9-66

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/1471-2105-9-66

Keywords