Skip to main content

Protein structural class prediction based on an improved statistical strategy

Abstract

Background

A protein structural class (PSC) belongs to the most basic but important classification in protein structures. The prediction technique of protein structural class has been developing for decades. Two popular indices are the amino-acid-frequency (AAF) based, and amino-acid-arrangement (AAA) with long-term correlation (LTC) – based indices. They were proposed in many works. Both indices have its pros and cons. For example, the AAF index focuses on a statistical analysis, while the AAA-LTC emphasizes the long-term, biological significance. Unfortunately, the datasets used in previous work were not very reliable for a small number of sequences with a high-sequence similarity.

Results

By modifying a statistical strategy, we proposed a new index method that combines probability and information theory together with a long-term correlation. We also proposed a numerically and biologically reliable dataset included more than 5700 sequences with a low sequence similarity. The results showed that the proposed approach has its high accuracy. Comparing with amino acid composition (AAC) index using a distance method, the accuracy of our approach has a 16–20% improvement for re-substitution test and about 6–11% improvement for cross-validation test. The values were about 23% and 15% for the component coupled method (CCM).

Conclusion

A new index method, combining probability and information theory together with a long-term correlation was proposed in this paper. The statistical method was improved significantly based on our new index. The cross validation test was conducted, and the result show the proposed method has a great improvement.

Background

Protein function is strongly related to its structure. Analysis of protein functions becomes a fundamental research domain to comprehend its structures. Nowadays, with the increased number of parsed structure entries in bioinformatics databases, it is important to do classification of protein structures in bioinformatics research. Scientists had developed various methodologies for the classification of protein structures. For example, based on the structure types and the arrangements of secondary structural elements, Levitt and Chothia [1] proposed a method to recognize ten protein classes, four principal and six small classes of a protein structure. Biological scientists common focus on the first four principal classes. They are all-α, all-β, α/β, and α+β classes, respectively. Therefore, the prediction of the four principal protein structural classes is the foundation in the field of protein analysis. In the fundamental study, many indices and methods were proposed to predict protein structural class [27]. The commonly-used indices and their corresponding methods are described briefly in the following.

Nishkawa [8] et al. found that protein structural classes are related to their amino acid compositions (AAC). Based on this hypnosis, Chou [9, 10] proposed standard vectors from amino acid composition in proteins. The statistics-based indices are 20-dimensional vectors, through which each variant corresponds to one amino acid occurrence frequency in protein sequence. Although these indices can be considered the eigenvector of a sequence, the information is insufficient enough to reflect the correlation among residues. Another weakness is that the statistics indices can not reflect the biological significance commendably. Accordingly, several methods were proposed such as the distance-based algorithm [11, 12], component-coupled-based algorithm [1315], support vector machine (SVM) based algorithm [16] and others [17, 18].

Alternatively, people can introduce protein-structural-class prediction index, which is based on the residues' arrangement and correlation in analysis of proteins. Such index method that uses various physiochemical properties has been experimented and adopted in the prediction. For example, Bu et al. [19] found that the auto-correlation function (ACF) of average non-bonded energy can represent the protein structural class with a better accuracy of prediction. Although a long-term correlation between different residues was considered, it did not include the statistical characteristics of sequences.

In this paper, a new index method is proposed. The method is based on the information and probability theories. In this method, a residue occurrence frequency is used instead of physiochemical indices for long-term correlation calculation. The statistical strategy of residue occurrence frequency is changed from a single sequence to a whole-training dataset. The results showed that the accuracy is significantly improved.

Methods

Suppose the whole dataset S contains N sequences, and this dataset can be divided into m (in this paper, we set m = 4; without losing generality) subsets S ξ (ξ = 1, 2,......, m), thus,

S = S1 S2 ...... S m (m = 4) (1)

The number of sequences in each subset is given by n ξ ; thus the total number of sequence, N = ξ = 1 m n ξ MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemOta4Kaeyypa0ZaaabmaeaacqWGUbGBdaWgaaWcbaGaeqOVdGhabeaaaeaacqaH+oaEcqGH9aqpcqaIXaqmaeaacqWGTbqBa0GaeyyeIuoaaaa@3867@

Chou et al. [9] proposed an index based on the amino acid composition (AAC) frequency in a sequence (Equation 2–4), i.e,

X k ξ = [ x k , 1 ξ , x k , 2 ξ ...... x k , 20 ξ ] T ( k = 1 , 2 , ... , n ξ ; ξ = 1 , 2 , ... , m ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGceaqabeaacqWGybawdaqhaaWcbaGaem4AaSgabaGaeqOVdGhaaOGaeyypa0ZaamWaaeaacqWG4baEdaqhaaWcbaGaem4AaSMaeiilaWIaeGymaedabaGaeqOVdGhaaOGaeiilaWIaemiEaG3aa0baaSqaaiabdUgaRjabcYcaSiabikdaYaqaaiabe67a4baakiabc6caUiabc6caUiabc6caUiabc6caUiabc6caUiabc6caUiabdIha4naaDaaaleaacqWGRbWAcqGGSaalcqaIYaGmcqaIWaamaeaacqaH+oaEaaaakiaawUfacaGLDbaadaahaaWcbeqaaiabdsfaubaaaOqaaiabcIcaOiabdUgaRjabg2da9iabigdaXiabcYcaSiabikdaYiabcYcaSiabc6caUiabc6caUiabc6caUiabcYcaSiabd6gaUnaaBaaaleaacqaH+oaEaeqaaOGaei4oaSJaeqOVdGNaeyypa0JaeGymaeJaeiilaWIaeGOmaiJaeiilaWIaeiOla4IaeiOla4IaeiOla4IaeiilaWIaemyBa0MaeiykaKcaaaa@6B27@
(2)

Where x k , 1 ξ , x k , 2 ξ ...... x k , 20 ξ MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemiEaG3aa0baaSqaaiabdUgaRjabcYcaSiabigdaXaqaaiabe67a4baakiabcYcaSiabdIha4naaDaaaleaacqWGRbWAcqGGSaalcqaIYaGmaeaacqaH+oaEaaGccqGGUaGlcqGGUaGlcqGGUaGlcqGGUaGlcqGGUaGlcqGGUaGlcqWG4baEdaqhaaWcbaGaem4AaSMaeiilaWIaeGOmaiJaeGimaadabaGaeqOVdGhaaaaa@46DB@ are the normalized occurrence frequencies of 20 residues for the kth protein X k ξ MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemiwaG1aa0baaSqaaiabdUgaRbqaaiabe67a4baaaaa@305D@ in the subset S ξ , and T stands for the transpose symbol.

The average occurrence frequencies or the so-called standard vector for subset S ξ is represented by

X ¯ ξ = [ x ¯ 1 ξ , x ¯ 2 ξ , ...... , x ¯ 20 ξ ] T , ( ξ = 1 , 2 , ... , m ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWaa0aaaeaacqWGybawaaWaaWbaaSqabeaacqaH+oaEaaGccqGH9aqpdaWadaqaamaanaaabaGaemiEaGhaamaaDaaaleaacqaIXaqmaeaacqaH+oaEaaGccqGGSaaldaqdaaqaaiabdIha4baadaqhaaWcbaGaeGOmaidabaGaeqOVdGhaaOGaeiilaWIaeiOla4IaeiOla4IaeiOla4IaeiOla4IaeiOla4IaeiOla4IaeiilaWYaa0aaaeaacqWG4baEaaWaa0baaSqaaiabikdaYiabicdaWaqaaiabe67a4baaaOGaay5waiaaw2faamaaCaaaleqabaGaemivaqfaaOGaeiilaWYaaeWaaeaacqaH+oaEcqGH9aqpcqaIXaqmcqGGSaalcqaIYaGmcqGGSaalcqGGUaGlcqGGUaGlcqGGUaGlcqGGSaalcqWGTbqBaiaawIcacaGLPaaaaaa@57D0@
(3)

where

x ¯ i ξ = 1 n ξ k = 1 n ξ x k , i ξ , ( i = 1 , 2 , ... , 20 ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWaa0aaaeaacqWG4baEaaWaa0baaSqaaiabdMgaPbqaaiabe67a4baakiabg2da9KqbaoaalaaabaGaeGymaedabaGaemOBa42aaSbaaeaacqaH+oaEaeqaaaaakmaaqadabaGaemiEaG3aa0baaSqaaiabdUgaRjabcYcaSiabdMgaPbqaaiabe67a4baaaeaacqWGRbWAcqGH9aqpcqaIXaqmaeaacqWGUbGBdaWgaaadbaGaeqOVdGhabeaaa0GaeyyeIuoakiabcYcaSmaabmaabaGaemyAaKMaeyypa0JaeGymaeJaeiilaWIaeGOmaiJaeiilaWIaeiOla4IaeiOla4IaeiOla4IaeiilaWIaeGOmaiJaeGimaadacaGLOaGaayzkaaaaaa@546F@
(4)

Since Chou's great contribution, many methods that are based on residue composition were proposed. The n-order component coupled method was one of them. When n = 0, this algorithm degenerated to the amino acid composition (AAC) method. In the case when n = 1, the corresponding indices can be expressed in terms of a 20 × 20 conditional probability matrix [20]. And if n > 1, the n-order component coupled components can be expressed in terms of a multi-dimensional conditional probability matrix. In those residue-composition-based methods, the size of statistics samples must be largely enough. However, the present statistical approach requires to calculate the probability of 20 amino acids or the conditional probability for one sequence. In this way, the conditional probabilities, especially the high-order coupled components, can not be calculated accurately since the length of each protein sequence is not long enough. For any n = 0 coupled component, the influence of amino acid that nearby was not considered. With the increase of n, the long-term interaction between the residues at different positions in a same sequence can be reflected; which it is of computational complexity.

In order to solve these problems mentioned above, a new method with an innovative index is proposed in this paper, which can be summarized as follows:

First, a new statistical approach was proposed. The amino acids' component frequency of each entire class (expressed in Equation 5, rather than Equation 4) is calculated, instead of the occurrence frequencies of different residues for a certain protein in each class.

x ¯ i ξ = k = 1 n ξ x k , i ξ × l k ξ k = 1 n ξ l k ξ , ( i = 1 , 2 , ...... , 20 ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWaa0aaaeaacqWG4baEaaWaa0baaSqaaiabdMgaPbqaaiabe67a4baakiabg2da9KqbaoaalaaabaWaaabmaeaacqWG4baEdaqhaaqaaiabdUgaRjabcYcaSiabdMgaPbqaaiabe67a4baacqGHxdaTcqWGSbaBdaqhaaqaaiabdUgaRbqaaiabe67a4baaaeaacqWGRbWAcqGH9aqpcqaIXaqmaeaacqWGUbGBdaWgaaqaaiabe67a4bqabaaacqGHris5aaqaamaaqadabaGaemiBaW2aa0baaeaacqWGRbWAaeaacqaH+oaEaaaabaGaem4AaSMaeyypa0JaeGymaedabaGaemOBa42aaSbaaeaacqaH+oaEaeqaaaGaeyyeIuoaaaGccqGGSaaldaqadaqaaiabdMgaPjabg2da9iabigdaXiabcYcaSiabikdaYiabcYcaSiabc6caUiabc6caUiabc6caUiabc6caUiabc6caUiabc6caUiabcYcaSiabikdaYiabicdaWaGaayjkaiaawMcaaaaa@669D@
(5)

where l k ξ MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemiBaW2aa0baaSqaaiabdUgaRbqaaiabe67a4baaaaa@3085@ is the size of kth sequence length for the subset S ξ , and the other parameters remain the same definitions as in Equation 4.

Secondly, we develop a method to improve the component coupled algorithm. Conditional probabilities of different amino acids that have different correlation lengths can be calculated. To simplify the calculation procedure, only a 2-dimensional (20 × 20) matrix is introduced to express any possible distances between residues. The conditional probability can be expressed as P d (a i /a j ), where the subscript d is the distance between the residue a i and a j , that is, d = i-j. For each d, one has

j = 1 20 i = 1 20 P d ( a i / a j ) = 1 , ( d = 0 , 1 , 2 , ...... ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqbaeqabeGaaaqaamaaqadabaWaaabmaeaacqWGqbaudaWgaaWcbaGaemizaqgabeaakiabcIcaOiabdggaHnaaBaaaleaacqWGPbqAaeqaaOGaei4la8Iaemyyae2aaSbaaSqaaiabdQgaQbqabaGccqGGPaqkaSqaaiabdMgaPjabg2da9iabigdaXaqaaiabikdaYiabicdaWaqdcqGHris5aaWcbaGaemOAaOMaeyypa0JaeGymaedabaGaeGOmaiJaeGimaadaniabggHiLdGccqGH9aqpcqaIXaqmcqGGSaalaeaacqGGOaakcqWGKbazcqGH9aqpcqaIWaamcqGGSaalcqaIXaqmcqGGSaalcqaIYaGmcqGGSaalcqGGUaGlcqGGUaGlcqGGUaGlcqGGUaGlcqGGUaGlcqGGUaGlcqGGPaqkaaaaaa@574B@
(6)

According to the theory of the probability multiplication:

P d (a i /a j ) = P d (a i , a j )/P d (a j ) (7)

In Equation 7, P d (a i , a j ) and P d (a j ) can be easily computed from protein sequences, and the conditional probability P d (a i /a j ) can also be calculated.

For the case that d + j exceeds the length of the sequence, the cyclic boundary condition can be used. The residue at which its position is equal to the remainder of d + j and the length of sequence can be considered.

The third step is to determine the indexation of the conditional probability matrix for prediction. The information content of conditional probability is used as the quantification index. For each residue (a j ) in an undetermined sequence, the index of the d-interval can be calculated as:

I d (a j ) = -log P d (a i /a j ), (j = 1, 2,......l) (8)

In this natural logarithm expression, l is the length of sequence k. For all the residues in the sequence k, the total information content can be obtained by

I d = I d (a1) + I d (a2) + ...... + I d (a l ) (9)

To consider multi-residue effects on some amino acids, the information contents with different distances can be accumulated to form the whole information contents, I w , i.e.,

I w = I a + Ia+1+ ... + I b (a, b = 0, 1,...l, ba) (10)

From Equation 8, we can find that the larger the conditional probability is, the smaller the information content is. Hence, the prediction result with minimum total information content should be considered in a predicted class in our method.

I d = min ( I d 1 , I d 2 , I d 3 , I d 4 ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemysaK0aaSbaaSqaaiabdsgaKbqabaGccqGH9aqpcyGGTbqBcqGGPbqAcqGGUbGBcqGGOaakcqWGjbqsdaqhaaWcbaGaemizaqgabaGaeGymaedaaOGaeiilaWIaemysaK0aa0baaSqaaiabdsgaKbqaaiabikdaYaaakiabcYcaSiabdMeajnaaDaaaleaacqWGKbazaeaacqaIZaWmaaGccqGGSaalcqWGjbqsdaqhaaWcbaGaemizaqgabaGaeGinaqdaaOGaeiykaKcaaa@4697@
(11)
P D ( ξ ) = { α ( I d = I d 1 ) β ( I d = I d 2 ) α + β ( I d = I d 3 ) α / β ( I d = I d 4 ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemiuaaLaemiraqKaeiikaGIaeqOVdGNaeiykaKIaeyypa0Zaaiqaaqaabeqaaiabeg7aHjabcIcaOiabdMeajnaaBaaaleaacqWGKbazaeqaaOGaeyypa0JaemysaK0aa0baaSqaaiabdsgaKbqaaiabigdaXaaakiabcMcaPaqaaiabek7aIjabcIcaOiabdMeajnaaBaaaleaacqWGKbazaeqaaOGaeyypa0JaemysaK0aa0baaSqaaiabdsgaKbqaaiabikdaYaaakiabcMcaPaqaaiabeg7aHjabgUcaRiabek7aIjabcIcaOiabdMeajnaaBaaaleaacqWGKbazaeqaaOGaeyypa0JaemysaK0aa0baaSqaaiabdsgaKbqaaiabiodaZaaakiabcMcaPaqaaiabeg7aHjabc+caViabek7aIjabcIcaOiabdMeajnaaBaaaleaacqWGKbazaeqaaOGaeyypa0JaemysaK0aa0baaSqaaiabdsgaKbqaaiabisda0aaakiabcMcaPaaacaGL7baaaaa@6343@
(12)

where PD is the predicted result.

Dataset and results

Dataset

In order to comprehensively perform our statistical studies, the latest version (version 1.71 updated on 24 January 2007) of the database SCOP [21] was used. Four classes' sequences – including 1267 in α class, 1424 in β class, 1682 in α/β class and 1551 in α+β class – with the similarity less than 30% were selected (the reason why using this dataset will be explained in discussion part in detail). After removing the uncertainty sequences that contain the letter x in sequence, the total numbers are 1250, 1375, 1565 and 1524, respectively (see additional files 1 and 2). According to the cross-validation principle, a whole sequence was divided into two subsets, randomly. The training and prediction sets were non-homologous and we selected a number that is large enough for training and test (about 20 times more than the size of dataset used in [9]).

Results

To test the feasibility, verification, and applicability of our index and method, the cross-validation [22] method was used in our study. The total sequences including 4 classes were randomly divided into 2 datasets, i.e., the training and the prediction datasets. The training dataset contains 2856 sequences, and the prediction dataset contains 2858 sequences.

Two traditional indices, AAC and ACF mentioned above, were used to compare with the results from our method. Three methods, mainly, the Hamming distance algorithm (DH), the Euclidean distance algorithm (DE) and the component coupled algorithm (CC), were used to assess the indices.

For the AAC index, the results of DH, DE and CC method were shown in Table 1 and 2.

Table 1 Training dataset using AAC index
Table 2 Prediction dataset using AAC index

For the auto-correlation based method, we found that our method with hydrophobicity based indices has a higher accuracy value than the one with other physiochemical properties. We used the Kyte and Doolittle [23] hydrophobicity values respectively, and the number of the auto-correlation function length is listed in Table 3 and 4.

Table 3 Training dataset using Kyte and Doolittle ACF index
Table 4 Prediction dataset using Kyte and Doolittle ACF index

In our experiments, different numbers of long-term correlations were tested, and the distance (d) between 2 and 4 shows to have a better result of accuracy. The results for training dataset and prediction dataset were shown in Table 5.

Table 5 Training and prediction dataset using our index*

The comparison of training and prediction results calculated by three different indices was illustratively presented in Figure 1 and 2. We found that our index has the best accuracy in protein structural class prediction. With the same index, the method DE always obtained better accuracy than the method DH.

Figure 1
figure 1

Accuracy of 3 indices for the training dataset

Figure 2
figure 2

Accuracy of 3 indices for the prediction dataset

Discussion

We will discuss the dataset, since it is the most important part in evaluating different indices and methods. People usually use the frequently-used dataset which includes several hundred sequences [10]. It is not relatively reliable enough, relevant to a given dataset scale. Another critical issue is the high sequence similarity. Let's take the 277 dataset [10] as an example. The 277 contains 277 protein domains extracted from the SCOP database.

The remarkable pair-wise similarity can be found in each class after multiple sequence alignment is conducted. For instance, in an alpha class, we found that there are several groups of identical sequences; the biggest one contains about 20 sequences (see additional files 1 and 2). After we conducted pair-wise alignment among these 20 sequences, we found that the sequence similarity was over 85%; indicating that these sequences are very identical to each other. The finding happens when we used other 3 classes. Such a high sequence similarity existed in the both training and test datasets; certainly violating the principle of cross validation. Therefore, we suspended such dataset for a reliable result.

In order to clearly emphasize the importance of selected dataset, we compared the three above methods from two different datasets. The amino acid composition index was used in this comparison study. The re-substitution and cross validation tests were designed and implemented for feature evaluations.

For the dataset including 138 sequences [10], the accuracy for re-substitution test and jack-knife test are shown in Table 6 and 7, and plotted in Figure 3, respectively.

Table 6 The 138 dataset with re-substitution test1
Table 7 The 138 dataset with jack-knife test.
Figure 3
figure 3

Accuracy of amino acid composition index using the 138 dataset. DH, DE and CC mean the Hamming distance method, the Euclidean distance method and the component coupled method respectively. RS means the re-substitution text, and CV corresponds to the cross validation text.

Our dataset is summarized in Table 8 and 9 with a total number of 5714 sequences, 2856 for training dataset and 2856 for testing dataset (see Figure 4).

Table 8 The 2856 dataset with re-substitution test
Table 9 the 2858 dataset with cross validation test*
Figure 4
figure 4

The accuracy of amino acid composition index using the 5714 dataset. DH, DE and CC mean the Hamming distance method, the Euclidean distance method and the component coupled method, respectively. RS means the resubstitution text, and CV corresponds to the cross validation set.

From Table 6 and 7, one can find that the prediction accuracy is very high for all three methods. This is because that the 138 dataset, just like the 277 dataset, is homologous, which means some sequences are almost the same. We can also find an interesting phenomenon that the accuracy of DH and DE are relatively higher in a cross validation test than that is in re-substitution test. It is mainly because these methods are insensitive to dataset, which means that there is a good extrapolating property in these algorithms. Comparing with CC and SVM, the total accuracy of our method is much better. However, like many advanced methods, the accuracies of re-substitution and cross validation tests are significantly different.

Traditional methods are usually based on simple criterions, while new-developed algorithms have more complicated rules. More prior probability information made current methods more accurate. However, this information must strongly rely on dataset. Fortunately, with an increased number of parsed-sequences, scientists can solve this problem commendably.

Generally speaking, using three above methods, the accuracy of dataset 5714 is much lower than one of the dataset 138. The 138 dataset is unreliable due to its high sequence similarity. However, in cross-validation test, the accuracy of DH and DE in 5714 dataset is much higher than that in 138 dataset. This illuminates that with an increase of dataset scale, one can improve the extrapolation of algorithms remarkablely.

From Table 1, 2, 3 and 4, we found that the accuracy is obviously decreased, compared with the result mentioned before. This is mainly because that the dataset we used are now larger and much different from the one used before. Therefore, the traditional methods had to be improved with an increase of sample size.

Table 1, 2, 3 and 4 also tell us that the difference of accuracy between the training and the prediction datasets is quite small. Therefore, the generalization of these methods is pretty good. It is because there are very few restriction conditions and technical manipulations in traditional methods that avoid a fluctuation between the training and test results by some techniques.

Using our method, the accuracy is between 6% and 16% higher than in the traditional methods. This is because long-term concepts are introduced and the conditional probability is used instead of physiochemical indices; thus to avoid the errors influenced by other parameters. In our test, distance (d) value is between 2 and 4, the accuracy is high. This phenomenon is a good accordance with the frequency characteristics of proteins. As we all know, most alpha helices are 3.6 residues per cycle, which means that a hydrogen bond bridges current residue and the residue 3 or 4 positions behind. Most beta strands have 2 residues per strand cycle, which reflects a strong interaction between two residues in a 2-position interval.

The advantage of our method can be concluded into three aspects:

  • In our method, the long-term correlation factor is considered without any other physiochemical parameters.

  • The accuracy is significantly improved for about 6–16% comparing with two traditional indices.

  • The merits in both two traditional methods are inherited. That is, the residue composition frequency and the amino acid arrangement.

However, there still exit some problems, which motivate our future study.

  • In our method, we must calculate the correlation between d residues. For the situation that the residue position is near the end of a sequence, the residue d sites behind may exceed the length of the sequence. In such case, the boundary process is crucial to the final result. For convenience, the cyclic boundary condition is used hereby. However, such approach is not biologically significant, and it is not quite reliable. To solve this problem, we are planning to test different types of extended boundary conditions.

  • The presented method only calculate the correlation between certain residue and the residue d positions behind. This is a "one-side" statistical work, and the information can not be extracted enough. The calculation of the correlation between the target residue and the residues different sites before and after is necessary to solve the problem.

Conclusion

In this paper, a new method by new indices is proposed. A reliable dataset with large number of entries and low sequence similarity is used to train and test our algorithm. The result showed that our method has a higher accuracy than the ones in traditional methods. The application of conditional probability and information content shows that the protein structural prediction can be largely improved by combining the information theory with the probability theory.

References

  1. Levitt M, Chothia C: Structural patterns in globular proteins. Nature. 1976, 261: 552-557. 10.1038/261552a0.

    Article  CAS  PubMed  Google Scholar 

  2. Shen HB, Yang J, Liu XJ, Chou KC: Using supervised fuzzy clustering to predict protein structural classes. Biochem Biophys Res Commun. 2005, 334: 577-581. 10.1016/j.bbrc.2005.06.128.

    Article  CAS  PubMed  Google Scholar 

  3. Chou KC, Cai YD: Predicting protein structural class byfunctional domain composition. Biochem Biophys Res Commun. 2004, 321 (4): 1007-1009. 10.1016/j.bbrc.2004.07.059.

    Article  CAS  PubMed  Google Scholar 

  4. Feng KY, Cai YD, Chou KC: Boosting classifier for predicting protein domain structural class. Biochem Biophys Res Commun. 2005, 334 (1): 213-217. 10.1016/j.bbrc.2005.06.075.

    Article  CAS  PubMed  Google Scholar 

  5. Zhou GP: An intriguing controversy over protein structural class prediction. J Protein Chem. 1998, 17 (8): 729-738. 10.1023/A:1020713915365.

    Article  CAS  PubMed  Google Scholar 

  6. Chou KC: Progress in protein structural class prediction and its impact to bioinformatics and proteomics. Curr Protein Pept Sci. 2005, 6 (5): 423-436. 10.2174/138920305774329368.

    Article  CAS  PubMed  Google Scholar 

  7. Cai Y, Zhou G: Prediction of protein structural classes by neural network. Biochimie. 2000, 82 (8): 783-785. 10.1016/S0300-9084(00)01161-5.

    Article  CAS  PubMed  Google Scholar 

  8. Nishkawa K, Ooi T: Correlation of the amino acid composition of a protein to its structural and biological characters. J Biochem. 1982, 91: 1821-1824.

    Google Scholar 

  9. Chou KC, Zhang CT: Prediction of protein structural classes. Crit Rev Biochem Mol Biol. 1995, 30: 275-349. 10.3109/10409239509083488.

    Article  CAS  PubMed  Google Scholar 

  10. Chou KC, Maggiora GM: Domain structural class prediction. Protein Eng. 1998, 11: 523-538. 10.1093/protein/11.7.523.

    Article  CAS  PubMed  Google Scholar 

  11. Mardia KV, Kent JT, Bibby JM: Multivariate Analysis. 1979, Academic Press, London

    Google Scholar 

  12. Nakashima H, Nishikawa K, Ooi T: The folding type of a protein is relevant to the amino acid composition. J Biochem. 1986, 99: 153-162.

    CAS  PubMed  Google Scholar 

  13. Chou KC: A novel approach to predicting protein structural classes in a (20-1)-D amino acid composition space. Proteins. 1995, 21: 319-344. 10.1002/prot.340210406.

    Article  CAS  PubMed  Google Scholar 

  14. Wang ZX, Yuan Z: How good is prediction of protein structural class by the component – coupled method. Proteins. 2000, 38: 165-175. 10.1002/(SICI)1097-0134(20000201)38:2<165::AID-PROT5>3.0.CO;2-V.

    Article  CAS  PubMed  Google Scholar 

  15. Zhou GP, Assa-Munt N: Some insights into protein structural class prediction. Proteins. 2001, 44 (1): 57-59. 10.1002/prot.1071.

    Article  CAS  PubMed  Google Scholar 

  16. Cai YD, Liu XJ, Xu XB: Support vector machines for predicting protein structural class. BMC Bioinformatics. 2001, 2: 3-10.1186/1471-2105-2-3.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  17. Luo RY, Feng ZP, Liu JK: Prediction of protein structural class by amino acid and ploypeptide composition. Eur J Biochem. 2002, 269: 4219-4225. 10.1046/j.1432-1033.2002.03115.x.

    Article  CAS  PubMed  Google Scholar 

  18. Du QS, Jiang ZQ, He WZ, Li DP, Chou KC: Amino acid principal component analysis (AAPCA) and its application in protein structural class prediction. J Biomol Struct Dyn. 2006, 23: 635-640.

    Article  CAS  PubMed  Google Scholar 

  19. Bu WS, Feng ZP, Zhang ZD, Zhang CT: Prediction of protein (domain) structural classes based on amino-acid index. Eur J Biochem. 1999, 266: 1043-1049. 10.1046/j.1432-1327.1999.00947.x.

    Article  CAS  PubMed  Google Scholar 

  20. Liu WM, Chou KC: Prediction of protein secondary structure content. Protein Eng. 1999, 12: 1041-1050. 10.1093/protein/12.12.1041.

    Article  CAS  PubMed  Google Scholar 

  21. Murzin AG, Brenner SE, Hubbard T, Chothia C: SCOP: a structural classification of protein database for the investigation of sequence and structures. J Mol Biol. 1995, 247: 536-540. 10.1006/jmbi.1995.0159.

    CAS  PubMed  Google Scholar 

  22. Chou KC: Prediction of protein structural classes and subcellular locations. Curr Protein Pept Sci. 2000, 1: 171-208. 10.2174/1389203003381379.

    Article  CAS  PubMed  Google Scholar 

  23. Kyte J, Doolittle RF: A simple method for displaying the hydropathic character of a protein. J Mol Biol. 1982, 157: 105-132. 10.1016/0022-2836(82)90515-0.

    Article  CAS  PubMed  Google Scholar 

Download references

Acknowledgements

This article has been published as part of BMC Bioinformatics Volume 9 Supplement 6, 2008: Symposium of Computations in Bioinformatics and Bioscience (SCBB07). The full contents of the supplement are available online at http://www.biomedcentral.com/1471-2105/9?issue=S6.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Hang Chen.

Additional information

Competing interests

The authors declare that they have no competing interests.

Electronic supplementary material

12859_2008_2616_MOESM1_ESM.xls

Additional file 1: Datasets. The 5172 dataset including 1250 alpha class sequences (alphatotal sheet), 1375 beta class sequences (betatotal sheet), 1565 alpha/beta class sequences (aabtotal sheet) and 1524 alpha+beta class sequences (apbtotal sheet). The training dataset including 625 alpha class sequences (alphatrain sheet), 687 beta class sequences (betatrain sheet), 782 alpha/beta class sequences (aabtrain sheet) and 762 alpha+beta class sequences (apbtrain sheet). The training dataset including 625 alpha class sequences (alphapre sheet), 688 beta class sequences (betapre sheet), 783 alpha/beta class sequences (aabpre sheet) and 762 alpha+beta class sequences (apbpre sheet). (XLS 491 KB)

12859_2008_2616_MOESM2_ESM.txt

Additional file 2: Sequences and sequence alignment result. Traditional datasets were always included the sequences with high sequence similarity. Take the 277 dataset as an example, the biggest identical group for alpha class was shown below: This group contains 20 sequences as in file 'additional file 2.txt'. Then the pair-wise sequence alignment was performed using the program FASTA, version 3.3, the result was also represented in file 'additional file 2.txt'. From this file, we can find a extremely high sequence similarity among these sequences. This situation made the dataset unreliable. (TXT 23 KB)

Rights and permissions

Open Access This article is published under license to BioMed Central Ltd. This is an Open Access article is distributed under the terms of the Creative Commons Attribution License ( https://creativecommons.org/licenses/by/2.0 ), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and permissions

About this article

Cite this article

Gu, F., Chen, H. & Ni, J. Protein structural class prediction based on an improved statistical strategy. BMC Bioinformatics 9 (Suppl 6), S5 (2008). https://doi.org/10.1186/1471-2105-9-S6-S5

Download citation

  • Published:

  • DOI: https://doi.org/10.1186/1471-2105-9-S6-S5

Keywords