Abstract
Protein secondary structure prediction is one of the most important and challenging problems in bioinformatics. Machine learning techniques have been applied to solve the problem and have gained substantial success in this research area. However there is still room for improvement toward the theoretical limit. In this paper, we present a novel method for protein secondary structure prediction based on a data partition and semi-random subspace method (PSRSM). Data partitioning is an important strategy for our method. First, the protein training dataset was partitioned into several subsets based on the length of the protein sequence. Then we trained base classifiers on the subspace data generated by the semi-random subspace method, and combined base classifiers by majority vote rule into ensemble classifiers on each subset. Multiple classifiers were trained on different subsets. These different classifiers were used to predict the secondary structures of different proteins according to the protein sequence length. Experiments are performed on 25PDB, CB513, CASP10, CASP11, CASP12, and T100 datasets, and the good performance of 86.38%, 84.53%, 85.51%, 85.89%, 85.55%, and 85.09% is achieved respectively. Experimental results showed that our method outperforms other state-of-the-art methods.
Similar content being viewed by others
Introduction
Proteins play a key role in almost all biological processes; they are the basis of life. For example, they take part in maintaining the structural integrity of the cell, transport and storage of small molecules, catalysis, regulation, signaling, and the immune system. There are 20 different amino acids that form proteins in nature1. The amino acids of a protein are connected in sequence with the carboxyl group of one amino acid forming a peptide bond with the amino group of the next amino acid. Protein structure is essential for the understanding of protein function. In order to recognize the protein functions of proteins at a molecular level, it is sometimes necessary to determine their 3D structure. Accurately and reliably predicting structures from protein sequences is one of the most challenging tasks in computational biology2. Protein secondary structure prediction provides a significant first step toward tertiary structure prediction, as well as offering information about protein activity, relationships, and functions.
Protein secondary structure refers to the local conformation proteins’ polypeptide backbone. There are two regular secondary structure states, α-helix (H) and β-strand (E), and one irregular secondary structure type, the coil region (C). Sander developed a secondary structure assignment method Dictionary of Secondary Structure of Proteins (DSSP)3, which automatically assigns secondary structure into eight states (H, E, B, T, S, L, G, and I) according to hydrogen-bonding patterns. These eight states are often further simplified into three states of helix, sheet and coil. The most widely used convention is that helix is designated as G, H and I; sheet as B and E; and all other states are designated as a coils. Most commonly, the secondary structure prediction problem is formulated as follows: given a protein sequence with amino acids, predict whether each amino acid is in the α-helix (H), β-strand (E), or coil region (C). Protein secondary structure prediction is usually evaluated by Q3 accuracy, which measures the percentage of residues for three-state secondary structures to determine whether they have been predicted correctly.
Protein secondary structure prediction began in 1951 when Pauling and Corey predicted helical and sheet conformations for protein polypeptide backbones, even before the first protein structure was determined2. Many statistical approaches and machine learning approaches have been developed to predict secondary structure. One of the first approaches for predicting protein secondary structure, uses a combination of statistical and heuristic rules4,5. The GOR6 method formalizes the secondary structure prediction problem within an information-theoretic framework. Position specific scoring matrix (PSSM)7 based on PSIBLAST8 reflects evolutionary information and has made the most significant improvements in protein secondary structure prediction. Many machine learning methods have been developed to predict protein secondary structure, and exhibit good performance by exploiting evolutionary information, as well as statistic information about amino acid subsequences9. For example, many neural network (NN)10,11,12,13,14 methods, hidden Markov model (HMM)15,16,17, support vector machines (SVM)18,19,20,21, and K-nearest neighbors22 have had substantial success, and Q3 accuracy has reached to 80%. The prediction accuracy has been continuously improved over the years, especially by using hybrid or ensemble methods and incorporating evolutionary information in the form of profiles extracted from alignments of multiple homologous sequences23. Recently, several papers used deep learning networks24,25,26,27,28 to predict protein secondary structure and obtained good success. The highest Q3 accuracy without relying on structure templates is now at 82–84%3. DeepCNF27 is a deep learning extension of conditional neural fields (CNF), which integrates conditional random fields and shallow neural networks. The overall performance of DeepCNF is significantly better than other state-of-the-art methods, breaking the long-lasting ~80% accuracy. Recently SPIDER3 improved the prediction of protein secondary structure by capturing non-local interactions using long short-term memory bidirectional recurrent neural networks29. In the paper30, a new deep inception-inside-inception network, called MUFOLD-SS, was proposed for protein secondary structure prediction. SPIDER3 and MUFOLD-SS achieved better performance, compared to DeepCNF.
In this paper, we presented a data partition and semi-random subspace method (PSRSM) for protein secondary structure prediction. The first step was partitioning the protein training dataset into several subsets based on the lengths of proteins sequences. The second step was generating subspaces by the semi-random subspaces method, training base classifiers on the subspaces, and then combining them by majority vote rule on each subset. Fig. 1 demonstrates our PSRSM experimental framework.
A key step of our method was to partition the training dataset into several subsets according to the length of the protein. The length of a protein sequence is the number of amino acids (AAs) in a protein sequence. Then we trained base classifiers in parallel on subspace data generated by using semi-random subspace method and combined them on each subset. In the conventional random subspace method, the low-dimensional subspaces are generated by random sampling of the original high-dimensional spaces. In order to get good performance of the ensemble, in this paper, we proposed a semi-random subspace method for protein secondary structure prediction. This method ensured that the base classifiers were as accurate and diverse as possible. We used support vector machines (SVMs) as the base classifier. Support vector machines are a popular machine learning method for classification, regression, and other learning tasks. Compared to other machine learning methods, SVM has the advantages of high performance, absence of local minima, and ability to deal with multidimensional datasets, in which with complex relationships exist among data elements. Support vector machines (SVMs) have had substantial success in protein secondary structure prediction.
Experimental results show that the overall performance of PSRSM was better than the current state-of-the-art methods.
Results
Datasets
We used six publicly available datasets (1) ASTRAL31, (2) CullPDB32, (3) CASP1033, (4) CASP1134, (5) CASP1235, (6) CB51336, and (7) 25PDB37 (8) a dataset T100 developed in-house. ASTRAL, ASTRAL + CullPDB, and T100 datasets are available from supplement files.
In this research, we combined the ASTRAL dataset and CullPDB dataset to be our training dataset, i.e., the ASTRAL + CullPDB dataset. The CullPDB dataset was selected based on the percentage identity cutoff of 25%, the resolution cutoff of 3 angstroms, and the R-factor cutoff of 0.25. There are 12,288 proteins in the CullPDB dataset. ASTRAL dataset had 6,892 proteins, with less than 25% sequence identity. Our training dataset ASTRAL + CullPDB had 15,696 proteins; we removed all duplicated proteins.
Publicly available datasets CASP10, CASP11, CASP12, CB513, and 25PDB were used to evaluate our method and compared using SPINE-X38, JPRED39, PSIPRED40 and DeepCNF. 99 proteins of the CASP10 dataset, 81 proteins of the CASP11 dataset, and 19 proteins of the CASP12 dataset were selected according to the availability of crystal structure. The CB513 dataset has 513 protein sequences. Any two proteins of CB513 share less than 25% sequence identity with each other. The 25PDB dataset was selected with low sequence similarity of no more than 25%, and has 1673 proteins, consisting of 443 all-α, 443 all-β, 346 α/β and 441 α + β. Note that the number of proteins in these datasets may be different from those reported in other published papers because we only used the available online (http://www.rcsb.org/) or with the PSSM program.
In addition, we randomly downloaded 100 new proteins (T100) released after 1 January 2018 from http://www.rcsb.org/. The dataset (T100) contains 100 proteins with sequence lengths ranging from 18 to 1460. We used T100 to test PSRSM and deepCNF using our online servers and their online server RaptorX-Property which was ranked first in secondary structure prediction.
Because T100 dataset is released after 2018, there is no duplicated proteins with our training dataset. All our training datasets were collected before February 2017.
Performance measures
Several different measures can be used to measure the secondary structure prediction accuracy, the most common being Q3. The Q3 accuracy is defined as the percentage of residues for which the predicted secondary structures are correct, Q3 is calculated as follows:
where, NH, NE, and NC, are the number of correctly predicted secondary structures: helix, strand and coil, respectively. N is the total number of residues (amino acids).
We calculate the average accuracy of the whole test dataset and use average Q3 to evaluate the performance of our model on the test dataset, the average Q3 is defined as
Where n is the number of protein sequences that has the valid predicted results in the test dataset, Xi denotes a protein sequence, and Q3(Xi) is the Q3 accuracy of Xi.
Performance
We used Q3 accuracy to compare our PSRSM method with other state-of-the-art methods, SPINE-X, PSIPRED, JPRED, and DeepCNF, on four publicly available datasets (CASP10, CASP11, CASP12, and CB513). Table 1 shows the Q3 accuracy of PSRSM and the other state-of-the-art methods on the four datasets. The experimental results show that PSRSM is significantly outperforming SPINE-X, PSIPRED, and JPRED. Moreover, PSRSM had 1–3% higher Q3 accuracy than DeepCNF. We also tested our method on 25PDB dataset with 1674 proteins, and Q3 accuracy is 86.38%.
In addition, we compared our proposed method to DeepCNF using our online servers (http://210.44.144.20:82/protein_PSRSM/default.aspx) and their online server RaptorX-Property (http://raptorx.uchicago.edu/StructurePropertyPred/predict/) on T100 dataset. Table 2 lists the Q3 accuracy of PSRSM and DeepCNF for each protein. The average Q3 accuracy of PSRSM was higher 2.5% than that of DeepCNF. In addition, we analyzed Q3 accuracy of predicted secondary structures in internal regions and at boundaries2. Here, we defined a helical/sheet residue as internal if its two nearest neighboring residues were also helical/sheet residues; we defined it as a boundary if one or both of the nearest neighbors had a different secondary structural assignment. The overall Q3 accuracies of PSRSM and DeepCNF, respectively, were 89.89% and 85.68% in internal regions, and 75.33% and 73.30% at boundaries. We also compared our method with other state-of-the-art methods (SPIDER3, MUFOLD, PSIPRED and JPRED) using their online server on T100 dataset in Table 3. The newly updated MUFOLD and SPIDER3 obtained 89.28% and 88.25% in internal regions, and 74.65% and 70.72% at boundaries. We can see that PSRSM was superior to current state-of-the-art methods not only in internal regions, but also at boundaries.
Discussion
Reason for partitioning training datasets according to protein length rather than randomly
Our training data was the ASTRAL+CullPDB dataset, which had 15,696 proteins, and 3,863,231 amino acids (AAs). Since training support vector machines on such a large dataset is a very slow process, the first step of our method was partitioning the training data into several different subsets and training SVMs in parallel. If we partitioned the training data randomly, it would just reduce the computation time, but not increase the prediction accuracy41. The length of a protein sequence is the number of amino acids in a protein sequence. Protein length is an important feature of a protein because it influences protein structure. For example, the short sequence ‘VVDALVR’ formed ‘EEEEEE’ in six proteins: 1by5_A, 1qfg_A, 1qff_A, 1fcp_A, 1fi1_A, and 2fcp_A. Their lengths are 714, 725, 725, 705, 707, and 723 respectively. Meanwhile ‘VVDALVR’ formed ‘HHHHHH’ in one protein (3vtz_A), and its length was 269. This data can be downloaded at prodata.swmed.edu/chseq.42. Identical amino acid sequence has different types of secondary structures in proteins of different lengths; this is because protein length can affect both local and long-range interactions of the protein. Based on the above considerations, we partitioned training datasets according to protein length to cluster proteins in the training data.
In order to validate the effectiveness of our data partitioning strategy, we conducted another experiment. We randomly generated a subset of the ASTRAL+CullPDB dataset randomly instead of according to protein length, and similarly trained SVM base classifiers on the subset. Then we combined them into an ensemble (Classifier_C). We compared Classifier_C with our PSRSM1, and Table 4 shows that the performance of PSRSM1 is quite similar to that of Classifier_C on CB513 dataset, but significantly better on subset with protein length L ∈ [1, 100]. The main difference between the two classifiers was the training set. All training proteins of PSRSM1 were short proteins, they had similar protein lengths, and all lengths belonged to interval [1,100]; conversely, the lengths of Classfier_C training data were randomly distributed.
Table 5 shows the performance of T100 dataset with different lengths based on 6 PSRSMs. 6 protein subsets with different lengths achieved the best performance 79.84%, 84.58%, 87.59%, 87.51%, 83.24%, and 83.93% respectively using their corresponding PSRSM.
Training time analysis
Another advantage of our method is that the training time was short. Because our training data ASTRAL + CullPDB is a large dataset, it was very slow to train the SVM classifier. We failed to train the SVM classifier on ASTRAL + CullPDB using our server.
The computational complexity to train an SVM43 is
Where NS is the number of support vectors, Nf is the feature dimension, and N is the size of the training set.
After data partitioning and sampling, the number of support vectors NS, feature dimension Nf, and the size of the training set N are much smaller. Furthermore, since we trained our base classifiers in parallel, the running time was reduced.
Table 6 shows the training time on each subset of the ASTRAL + CullPDB. D1, D2, …, D5 and D6 were subsets of ASTRAL + CullPDB (Table 7). We failed to train the SVM classifier on the ASTRAL + CullPDB using our server. After data partitioning but before sampling we completed training of SVM classifiers on each subset; more time was required because D3 had more amino acids than other subsets. When we used PSRSM, the feature dimension was decreased, and the training time was reduced.
Conclusion and Future Work
In this paper we proposed a novel method, PSRSM, to predict protein secondary structure. The first step of our method was partitioning of the training set into several subsets based on protein length. In the second step, we generated k ensemble classifiers using the semi-random subspace method. If given a new query protein sequence, our method would select one, and only one, ensemble classifier from k ensemble classifiers according to length to predict the protein secondary structure. Experimental results showed that the overall performance of PSRSM was better than that of other current state-of-the-art methods. In particular, our method PSRSM is superior to other methods not only in internal regions, but also at boundaries.
Methods
Partitioning the training data
We partitioned the training data into k different subsets according to the protein sequence length. Let X denote a protein sequence, and L denote the length of X. We set k−1 partition points of interval \((0,\,\infty )\). Let \({{\rm{r}}}_{0}=0,\,{{\rm{r}}}_{{\rm{k}}}=\infty \), and r1, …, r2 and rk−1 denote partition points that satisfy \({{\rm{r}}}_{0} < {{\rm{r}}}_{1} < \cdots < {{\rm{r}}}_{{\rm{k}}-1} < {{\rm{r}}}_{{\rm{k}}}\). These partition points partition interval (0, ∞) into k intervals without intersection. Let \(R=\{(0,{r}_{1}),({r}_{1},{r}_{2}),\,\cdots ,\,({r}_{k-1},\infty )\}\).
Let D denote the training data ASTRAL + CullPDB. Subsets D1, D2, …, Dk−1 and Dk are defined as follows:
and, D1, D2, …, Dk−1 and Dk satisfy \({\cup }_{{\rm{i}}={\rm{1}}}^{{\rm{k}}}{{\rm{D}}}_{{\rm{i}}}={\rm{D}}\), and \({{\rm{D}}}_{{\rm{i}}}\cap {{\rm{D}}}_{{\rm{j}}}=\rlap{/}{0},\,{\rm{i}}\ne {\rm{j}}{\rm{.}}\)
In our experiment, we set k = 6 and
\(R=\{(0,100],(100,200],(200,300],(300,400],\,(400,500],(\mathrm{500},\infty )\}\)
Table 7 shows the number of proteins and amino acids in \({\{{D}_{i}\}}_{i=1}^{6}\).
Training classifiers
We generated t random subspaces of r-dimension, and trained t SVM base classifiers on each subset D i t feature subsets are used to train t base classifiers, and each subset had r features sampled from the 260-dimensional dataset.
Therefore we got k × t SVM base classifiers on k subsets, we denote these classifiers as a k × t matrix, where k is the number of subsets of the training data.
where Cij is the SVM base classifier trained on the jth subspace data of subset Di.
We combined classifiers \({\{{C}_{ij}\}}_{j=1}^{t}\) into a final ensemble classifier by majority vote rule, and thus got k ensemble classifiers as the final decision on each subset. They are denoted as below.
Here ‘Voting’ means combining classifiers by majority vote rule, PSRSM i represents the final ensemble classifier on subset Di, and,
In this study The parameters t is set to 12 base classifiers, and the dimension of subspaces r is 160 in our experiment.
The publicly available LIBSVM44 software was used to train SVM classifiers. There are several kernel functions, commonly used in SVM: “liner”, “polynomial”, and “radial basis”. In this paper, we used the radial basis function (RBF) as kernel, the form is \({\rm{k}}({\rm{x}},{{\rm{x}}}_{{\rm{i}}})=\exp (-{\rm{\gamma }}{\Vert {\rm{x}}-{{\rm{x}}}_{{\rm{i}}}\Vert }^{2})\), where γ is a parameter. C is another parameter for SVM training; it is the regularization factor that controls the balance between low error and large divided margin. Parameters C and γ were decided using the grid search method. The optimal values of the two parameters are 0.9956 and 0.065, respectively.
Prediction
Given a new query protein sequence X, and protein sequence length L, our method selected one and only one ensemble classifier from k ensemble classifiers ({PSRSM1, PSRSM2, …, PSRSMk}) according to the length L to predict the protein secondary structure of X. Let \(\tilde{{\rm{Y}}}\) denote the prediction output by PSRSM. Then
where, PSRSM i is defined as (7).
For example, if a new query protein sequence X is a short protein and \({\rm{L}}\in (0,{{\rm{r}}}_{1}]\), then the corresponding PSRSM1 trained on the short protein subset is used to predict its secondary structure. In general, if \({\rm{L}}\in ({{\rm{r}}}_{{\rm{i}}-1},{{\rm{r}}}_{{\rm{i}}}]\), the ith classifier PSRSMi will be selected from k ensemble classifiers to predict the protein secondary structure of X.
Semi-Random Subspace Method (SRSM)
The random subspace method (RSM) is an ensemble construction technique. It was proposed by Ho in 199845. RSM randomly samples a set of low-dimensionality subspaces from the whole original high-dimensional features space, then constructs a classifier on each smaller subspace and finally applies a combination rule for the final decision.
We proposed a semi-random subspace method for protein secondary structure prediction. In our research, each protein sequence was represented by a 260 × L matrix. The ith column vector represents features of the ith amino acid residue. We generated t feature subsets to train t base classifiers. Each subset had r features sampled from the 260-dimensional dataset.
Because the original PSSM of the associated residue is an important feature for the base classifier, those 20 dimensions in a central location of 260-dimensional data are fixed for each sampling.
Let S represent the 260-dimensional features vector, and \({\rm{S}}=({v}_{1},{v}_{2},\cdots ,{v}_{260})\). We generated t subspaces (\({\{{{\rm{S}}}_{{\rm{i}}}\}}_{{\rm{i}}}^{{\rm{t}}}\)) from S.S i represents a feature subset sampled from S, and \({S}_{i}=({x}_{i1},\,{x}_{i2},\cdots ,\,{x}_{id},{v}_{121},{v}_{122},\cdots ,{v}_{140},{y}_{i1},{y}_{i2},\cdots ,{y}_{id})\), where (v121, v122, …, v140) are fixed in each S i ;(xi1, xi2, …, x id ) and (yi1, yi2, …, y id ) are sampled from (v1, v2, …, v120) and (v141, v122, …, v260), respectively; here, r and d satisfy \(r=2\times d+20\). Let L i = (xi1, xi2, …, x id ), S0 = (v121, x122, …, v140), and R i = (yi1, yi2, …, y id ), then \({S}_{i}={{\rm{L}}}_{{\rm{i}}}\cup {{\rm{S}}}_{0}\,\cup {{\rm{R}}}_{{\rm{i}}}\).
Additionally, high diversity of base classifiers can make an ensemble with more accurate decisions; this is because different base classifiers make different errors on different patterns. The diversity of base classifiers is negatively correlated with the similarity of the training data. \(|{{\rm{S}}}_{{\rm{i}}}\cup {{\rm{S}}}_{{\rm{j}}}|\) reflects the similarity of Si and Sj to a certain extent. When \(|{{\rm{S}}}_{{\rm{i}}}\cap {{\rm{S}}}_{{\rm{j}}}|\) is smaller, the similarity between Si and Sj is smaller, and the diversity of base classifiers is higher. Let oj be the number of occurrences of vj in \({\{{S}_{i}\}}_{i=1}^{t}\). In our research, it can be proved that, when \({{\rm{o}}}_{{\rm{i}}}=\frac{{td}}{120}\,{\rm{for}}\,{\rm{i}}=1,2,\ldots ,120\,{\rm{and}}\,{\rm{i}}=141,142,\ldots ,260\), the sum \({\sum }_{{\rm{i}}=1}^{{\rm{t}}}{\sum }_{{\rm{j}}=1}^{{\rm{t}}}|{{\rm{S}}}_{{\rm{i}}}\cap {{\rm{S}}}_{{\rm{j}}}|\) becomes the minimum. Therefore our method was to generate t feature subsets randomly, and adjust elements of each subspace to make oi = \(\frac{{\rm{td}}}{120}\) for i = 1, 2, …, 120 and i = 141, 142, …, 260.
The steps of the proposed semi-random subspace method are as follows.
-
1.
Generating semi- random subspaces
-
(1)
Let \({\rm{L}}=({v}_{1},{v}_{2},\cdots ,{v}_{120})\), \({{\rm{S}}}_{0}=({v}_{121},{v}_{122},\cdots ,{v}_{140})\), and \({\rm{R}}=({v}_{141},{v}_{142},\cdots ,{v}_{260})\) and generate d-dimensional random subspaces \({\{{{\rm{L}}}_{{\rm{i}}}\}}_{{\rm{i}}=1}^{{\rm{t}}}\) from L, \({\{{{\rm{R}}}_{{\rm{i}}}\}}_{{\rm{i}}=1}^{{\rm{t}}}\) from R, respectively.
-
(2)
Calculate \({\{{{\rm{o}}}_{{\rm{j}}}\}}_{{\rm{j}}=1}^{120}\), where oj is the number of occurrences of vj in \({\{{{\rm{L}}}_{{\rm{i}}}\}}_{{\rm{i}}}^{{\rm{t}}}\). Let \({\rm{mino}}=\,{\rm{\min }}\,{\{{{\rm{o}}}_{{\rm{j}}}\}}_{{\rm{j}}=1}^{120}\) and \({\rm{maxo}}=\,{\rm{\max }}\,{\{{{\rm{o}}}_{{\rm{j}}}\}}_{{\rm{j}}=1}^{120}\).
-
(3)
Let \({\rm{idmin}}=\{{\rm{j}}|{{\rm{o}}}_{{\rm{j}}}={\rm{mino}}\wedge {\rm{j}}=1,2,\cdots ,120\}\) and
\({\rm{idmax}}=\{{\rm{j}}|{{\rm{o}}}_{{\rm{j}}}={\rm{maxo}}\wedge {\rm{j}}=1,2,\cdots ,120\}\). Then generate a ternary ordered pairs set \({\rm{P}}=\{({\rm{i}},{\rm{j}},{\rm{k}})|{{\rm{v}}}_{{\rm{j}}}\notin {{\rm{L}}}_{{\rm{i}}}\wedge {{\rm{v}}}_{{\rm{k}}}\in {{\rm{L}}}_{{\rm{i}}}\wedge {\rm{j}}\in {\rm{idmin}}\wedge {\rm{k}}\in {\rm{idmax}}\}\).
-
(4)
Randomly select a ternary ordered pair (i, j, k) from P, insert feature vj to Li, and delete vk from Li. Then return to step (1) until \({{\rm{o}}}_{1}={{\rm{o}}}_{2}=\cdots ={{\rm{o}}}_{120}\).
-
(5)
Repeat (2), (3) and (4) on \({\{{{\rm{R}}}_{{\rm{i}}}\}}_{{\rm{i}}=1}^{{\rm{t}}}\) and R.
-
(6)
\({{\rm{S}}}_{{\rm{i}}}={{\rm{L}}}_{{\rm{i}}}\cup {{\rm{S}}}_{0}\cup {{\rm{R}}}_{i},i=1,2,\cdots ,{\rm{t}}\).
-
2.
Construct t classifier \({\{{{\rm{C}}}_{{\rm{i}}}\}}_{{\rm{i}}}^{{\rm{t}}}\) from the corresponding t random subspaces.
-
3.
Combine classifiers \({\{{{\rm{C}}}_{{\rm{i}}}\}}_{{\rm{i}}}^{{\rm{t}}}\) by majority vote rule.
There are two parameters to be determined for the semi-random subspace method, i.e., the number of subspaces t, and dimension of subspaces r.
Since D1 was smaller than other subsets, the training time on D1 was shorter than on other subsets. Therefore we conducted a series of experiments on D1 to determine t and r. We fixed t = 12, because it requires t*d to be divided by 120, it is easy to set d or r. Experimental results on the CB513 dataset showed that with increasing r the Q3 accuracy increased, but when r > 160, the Q3 accuracy increased slowly(Fig. 2) and the training time must be much longer. So we determine r = 160 as the dimension of subspaces in our experiment.
Input features
The PSSM of a protein sequence represents homolog information affiliated with its aligned sequences. We used the PSI-BLAST program to generate the PSSM data. PSI-BLAST used BLOSUM62 evolutionary matrix to search a reduced version of the NCBI’s non-redundant (NR) database filtered at 90% sequence similarity, in order to find the variability of the residue within a multiple sequence alignment. PSI-BLAST parameters was set with threshold h = 0.001 and j = 3 iterations. The resulting PSSMs were a 20 × L matrix, where L is the protein length and 20 is the number of amino acid types.
A sliding window of consecutive amino acids was used to obtain residue sequence information and predict the secondary structure of the central residue. Each residue was encoded by a vector of dimension 20 × w, where w is the sliding window size and is an odd number. The window was shifted from residue to residue through the protein chain. In this paper, the sliding window length w was set to 13. To use the first and last six amino acids, we inserted six zeros before and behind each protein sequence. Therefore each protein sequence was represented by a 260 × L matrix, and the ith column vector represented the protein features associated with the ith residue.
Secondary structure assignment was done with the DSSP. DSSP program defines eight states for secondary structure (H, E, B, T, S, L, G, and I) that are reduced to three states (H, E, and C) by different predictive methods. We used the following reductions: H, G and I to helix (H); E and B to beta strands (E); all the rest to coil (C).
Availability
References
Alberts B. et al. Molecular biology of the cell, 5th ed. New York: Garland Science (2008).
Yang, Y. et al. Sixty-five years of the long march in protein secondary structure prediction: the final stretch? Briefings in Bioinformatics (2016).
Kabsch, W. & Sander, C. Dictionary of protein secondary structure: pattern recognition of hydrogen-bonded and geometrical features. Biopolymers 22, 2577–2637 (1983).
Fasman, G. D. & Chou, P. Y. Prediction of protein conformation: consequences and aspirations. Biochemistry 13, 222–245 (1974).
Chou, P. Y. & Fasman, G. D. Conformational parameters for amino acids in helical, beta-sheet, and random coil regions calculated from proteins. Biochemistry 13, 211–222 (1974).
Garnier, J., Gibrat, J. F. & Robson, B. GOR method for predicting protein secondary structure from amino acid sequence. Methods in Enzymology 266, 540–553 (1996).
Jones, D. T. Protein secondary structure prediction based on position-specific scoring matrices. J. Mol. Biol. 292, 195–202 (1999).
Altschul, S. F. et al. Gapped BLAST and PSI-BLAST: a new generation of protein database search programs. Nucleic Acids Res. 25, 3389–3402 (1997).
Yoo, P. D., Zhou, B. B. & Zomaya, A. Y. Machine learning techniques for protein secondary structure prediction: an overview and evaluation. Current Bioinformatics 3, 74–86 (2008).
Holley, L. H. & Karplus, M. Protein secondary structure prediction with a neural network. Proc. Natl. Acad. Sci. USA 86, 152–156 (1989).
Qian, N. & Sejnowski, T. J. Predicting the secondary structure of globular proteins using neural network models. J. Mol. Biol. 202, 865–884 (1988).
Kneller, D., Cohen, F. & Langridge, R. Improvements in protein secondary structure prediction by an enhanced neural network. J. Mol. Biol. 214, 171–182 (1990).
Malekpour, S. A., Naghizadeh, S., Pezeshk, H., Sadeghi, M. & Eslahchi, C. Protein secondary structure prediction using three neural networks and a segmental semi markov model. Mathematical Biosciences 217, 145–150 (2009).
Wu, Q., Sui, H., Yang, B. & Qian, W. Improving protein secondary structure prediction using a multi-modal bp method. Computers in Biology & Medicine 41, 946–959 (2011).
Asai, K., Hayamizu, S. & Handa, K. Prediction of protein secondary structure by the hidden markov model. Computer Applications in the Biosciences Cabios 9, 141–146 (1993).
Won, K. J. et al. An evolutionary method for learning HMM structure: prediction of protein secondary structure. Bmc Bioinformatics 8, 1–13 (2007).
Aydin, Z., Altunbasak, Y. & Borodovsky, M. Protein secondary structure prediction for a single-sequence using hidden semi-Markov models. BMC Bioinformatics 7, 178 (2006).
Kim, H. & Park, H. Protein secondary structure prediction based on an improved support vector machines approach. Protein Eng. 16, 553–560 (2003).
Ward, J. J., McGuffin, L. J., Buxton, B. F. & Jones, D. T. Secondary structure prediction with support vector machines. Bioinformatics 19, 1650–1655 (2003).
Guo, J., Chen, H., Sun, Z. & Lin, Y. A novel method for protein secondary structure prediction using dual - layer SVM and profiles. Proteins: Struct. Funct. Bioinform. 54, 738–743 (2004).
Hua, S. & Sun, Z. A novel method of protein secondary structure prediction with high segment overlap measure: support vector machine approach. J. Mol. Biol. 308, 397–407 (2001).
Tan, Y. T. & Rosdi, B. A. Fpga-based hardware accelerator for the prediction of protein secondary class via fuzzy k-nearest neighbors with lempel–ziv complexity based distance measure. Neurocomputing 148, 409–419 (2015).
Bouziane, H., Messabih, B. & Chouarfia, A. Profiles and majority voting-based ensemble method for protein secondary structure prediction. Evolutionary Bioinformatics 7, 171–188 (2011).
Zhou, J. & Troyanskaya, O. D. Supervised and Convolutional Generative Stochastic Network for Protein Secondary Structure Prediction. Proceedings of the 31th International Conference on Machine Learning, ICML 2014, Beijing, China, 21-26 June 2014. JMLR Proceedings 32, 745–753 (2014).
Spencer, M., Eickholt, J. & Cheng, J. A Deep Learning Network Approach to ab initio Protein Secondary Structure Prediction. IEEE/ACM Trans. Comput. Biol. Bioinform. 12, 103–112 (2015).
Lee, H., Grosse, R., Ranganath, R. & Ng, A. Y. Convolutional deep belief networks for scalable unsupervised learning of hierarchical representations. Proceedings of the 26th Annual International Conference on Machine Learning, ICML 2009, Montreal, Quebec, Canada, June 14–18 (2009).
Wang, S. et al. Protein secondary structure prediction using deep convolutional neural fields. Scientific Reports, https://doi.org/10.1038/srep18962 (2016).
Wang, S., Li, W., Liu, S. & Xu, J. Raptorx-property: a web server for protein structure property prediction. Nucleic Acids Research 44, W430–W435, https://doi.org/10.1093/nar/gkw306 (2016).
Fang, C., Shang, Y. & Xu, D. MUFOLD-SS:New deep inception-inside-inception networks for protein secondary structure prediction. Proteins 86, 592–598 (2018).
Heffernan, R., Yang, Y., Paliwal, K. & Zhou, Y. Capturing Non-Local Interactions by Long Short Term Memory Bidirectional Recurrent Neural Networks for Improving Prediction of Protein Secondary Structure, Backbone Angles, Contact Numbers, and Solvent Accessibility. Bioinformatics 33, 2842–2849 (2017).
Fox, N. K. SCOPe: Structural classification of proteins-extended, integrating SCOP and ASTRAL data and classification of new structures. Nucleic Acids Research 42, 304–309 (2014).
Wang, G. & R. D. Jr. PISCES: recent improvements to a PDB sequence culling server. Nucleic Acids Research, 33(Web Server issue), W94–W98 (2005).
Moult, J., Fidelis, K., Kryshtafovych, A. & Tramontano, A. Critical assessment of methods of protein structure prediction (CASP)- round X. Proteins: Structure, Function, and Bioinformatics 79, 1–5 (2012).
Moult, J., Fidelis, K., Kryshtafovych, A. & Tramontano, A. Critical assessment of methods of protein structure prediction (CASP)- round XI. Proteins: Structure, Function, and Bioinformatics 82, 1–6 (2014).
Moult, J., Fidelis, K., Kryshtafovych, A., Schwede, T. & Tramontano, A. Critical assessment of methods of protein structure prediction (CASP)- progress and new directions in Round XII. Proteins: Structure, Function, and Bioinformatics 84(S1), 4–14 (2016).
Cuff, J. A. & Barton, G. J. Evaluation and improvement of multiple sequence methods for protein secondary structure prediction. Proteins: Structure, Function, and Bioinformatics 34, 508–519 (1999).
Kedarisetti, K. D., Kurgan, L. & Dick, S. Classifier ensembles for protein structural class prediction with varying homology. Biochem. Biophys. Res. Commu. 348, 981–988 (2006).
Faraggi, E., Zhang, T., Yang, Y., Kurgan, L. & Zhou, Y. SPINE X: improving protein secondary structure prediction by multistep learning coupled with prediction of solvent accessible surface area and backbone torsion angles. J. Comp. Chem. 33, 259–267 (2012).
Drozdetskiy, A., Cole, C., Procter, J. & Barton, G. J. JPred4: a protein secondary structure prediction server. Nucleic Acids Res. gkv332 (2015).
McGuffin, L. J., Bryson, K. & Jones, D. T. The PSIPRED protein structure prediction server. Bioinformatics 16, 404–405 (2000).
Meyer, O., Bischl, B., & Weihs, C. Support Vector Machines on Large Data Sets: Simple Parallel Approaches. Data Analysis, Machine Learning and Knowledge Discovery. Springer International Publishing. 87–95 (2014).
Li, W., Kinch, L. N., Karplus, P. A. & Grishin, N. V. Chseq: a database of chameleon sequences. Protein Science 24, 1075–1086 (2015).
Vapnik, V. N., Statistical learning theory. Encyclopedia of the Sciences of Learning (2008).
Chang, C. & Lin, C. LIBSVM: A library for support vector machines. ACM. 2, 1–27 (2011).
Ho, T. K. The Random Subspace Method for Constructing Decision Forests. IEEE Transactions on Pattern Analysis & Machine Intelligence 20, 832–844 (1998).
Acknowledgements
The work is supported by the National Natural Science Foundation of China (61375013), Shandong Provincial Natural Science Foundation, China (ZR2013FM020).
Author information
Authors and Affiliations
Contributions
Y.M. designed, implemented the algorithm, and wrote the manuscript. Y.L. supervised the whole experiments, collected all the experimental data, and performed part of experiments. J.C. performed part of experiments, and designed our web site. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Competing Interests
The authors declare no competing interests.
Additional information
Publisher's note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Ma, Y., Liu, Y. & Cheng, J. Protein Secondary Structure Prediction Based on Data Partition and Semi-Random Subspace Method. Sci Rep 8, 9856 (2018). https://doi.org/10.1038/s41598-018-28084-8
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-018-28084-8
This article is cited by
-
Plasma Sterilization for Bacterial Inactivation: Studies on Probable Mechanisms and Biochemical Actions
Plasma Chemistry and Plasma Processing (2024)
-
HIV-1 Vif protein sequence variations in South African people living with HIV and their influence on Vif-APOBEC3G interaction
European Journal of Clinical Microbiology & Infectious Diseases (2024)
-
GBDT_KgluSite: An improved computational prediction model for lysine glutarylation sites based on feature fusion and GBDT classifier
BMC Genomics (2023)
-
Structural mechanism of BRD4-NUT and p300 bipartite interaction in propagating aberrant gene transcription in chromatin in NUT carcinoma
Nature Communications (2023)
-
In Silico Analysis: Genome-Wide Identification, Characterization and Evolutionary Adaptations of Bone Morphogenetic Protein (BMP) Gene Family in Homo sapiens
Molecular Biotechnology (2023)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.