Introduction

Proteins play important roles in biological activities, and their functional significance is determined by their three-dimensional structure. However, it is difficult and expensive to experimentally determine protein tertiary structures. Moreover, with the rapid large-scale sequencing technologies, a gap between the huge number of protein sequences and a small number of known structures is being enlarged. Predicting protein three-dimensional structures is an alternative way to narrow the gap. It has been a grand challenge to make an accurate prediction without any structural information in computational biophysics for decades1,2, as there are mainly two difficulties in the prediction: (1) efficient sampling methods to search an astronomically larger conformation space3, and (2) accurate free energy determination to find the most stable shape1. Although it is extremely hard to search the large space of possible structures for the one with the lowest energy, inferred structural constraints, such as contact/distance between pairwise residues, have advanced protein structure prediction and decreased the deviation between the predicted and the authentic structures4,5,6. The backbone torsion angles, as an important structural constraint, also play a critical role in protein structure prediction (e.g., sampling the space of the torsion angles (\(\phi\), \(\psi\)) to investigate protein folding7) and refinement8, and they are also commonly used as constraints in many computational methods, e.g., CNS9, CYANA10, and AMBER11 to determine protein structures. Accurately predicting the torsion angles can considerably advance the tertiary structure prediction by accelerating efficient sampling of the large conformational space for the low-energy structures.

Owing to the larger protein databases and the development of computing resources, as well as advances in machine learning methods and deep neural networks, the accuracy of protein backbone torsion angle prediction has been improved increasingly. Typically, machine learning-based methods including neural networks12,13,14, support vector machines (SVM)13,14,15, and hidden Markov models16,17 predict discrete states of \(\phi\)/\(\psi\) angle values. Recently, computational advances have been developed to predict real values of the torsion angles. The DESTRUCT method uses position-specific scoring matrix (PSSM) to build iterative neural network models for the first time predicting the real value of the angle \(\psi\) although the correlation coefficient between the predicted and the real value is less than 0.518. The Real-SPINE model was designed based on integrated neural networks to improve the correlation coefficient for the angle \(\psi\) to more than 0.612. Based on a composite machine-learning algorithm, Wu et al. developed the ANGLOR13 to separately predict \(\phi\) using the feed-forward neural network and \(\psi\) using the SVM.

Considerable progress has recently been made by leveraging computational advances, especially deep learning (DL), in both protein secondary and tertiary structure prediction. For example, DL-based approaches have been convincingly demonstrated to predict structural constraints that successfully guide protein folding5,6. The SPIDER2 method was developed to predict the torsion angles using iterative neural networks19, while the SPIDER320 takes advantage of the long short-term memory bidirectional recurrent neural networks (LSTM-BRNN21) to remove the effect of the sliding window that was used in the SPIDER219. The DeepRIN22 was designed based on the architectures of the Inception23 and the ResNet24 networks. Gao et al. developed a deep neural network-based model to predict discrete angles within \(5^{\circ }\) bins25. As a hybrid method, the RaptorX-Angle combines K-means clustering and deep learning techniques to predict real-valued angles26, as claimed it takes advantage of both discrete and continuous representation of the torsion angles. The SPOT-1D27 is a hybrid model that employs an ensemble of LSTM-BRNN and ResNet. The OPUS-TASS was developed based on the network architecture of the modified transformer and CNN modules, and it is trained using the additional feature8. The same network in the OPUS-TASS was trained for six different tasks including secondary structure, backbone torsion angles (TA), discrete descriptors of local backbone structure, solvent accessible surface area, and side-chain dihedral angles.

Advances in predicting protein torsion angles have also benefited increasingly from machine learning-based methods. As a growing focus on DL-based methods, accurate prediction of the torsion angle is not only dependent on the architecture of DL but also on information (features) extracted from protein sequences. In most cases, the performances of a DL-based method are highly determined by the information. Generally, classical features, such as PSSM28, physicochemical properties (PP)29, and amino acid (AA), have been widely and successfully used to predict protein secondary/tertiary structure, but the classical features are far from satisfactory for existing DL-based models as a large number of parameters needs to be learned for better predictions. Moreover, the deeper the DL model is, the more difficult we optimize its parameters. To meet these requirements, we firstly present four evolutionary signatures as novel features, including the relative entropy (RE), the degree of conservation (DC), the position-specific substitution probabilities (PSSP), and the Ramachandran basin potential (RBP) statistically derived from protein sequences by removing redundant or unnecessary signals, and we also develop an evolutionary signatures-driven deep neural network (termed ESIDEN) that adopts straightforward architecture of recurrent neural networks to improve the prediction accuracy of the torsion angles.

Methods

In this section, we describe the benchmark datasets and the novel features used in our study, and the proposed method is presented in detail.

Datasets and input features

Two benchmark datasets are used for both training and test in our study, and they include a culled dataset (D2020, Table S1, Fig. S1) from the PISCES30 and the SPOT-1D dataset27,31 (Table S3). The D2020 dataset was culled from the PISCES server30 with less than 25% identity and less than 1.6Å resolution (R-factor is 0.25) (released in December 2020), which contains 8669 protein chains. We filter out the protein whose sequence length is more than 500 and finally obtain 7443 proteins (Table S1), and these proteins are randomly classified into three groups by percentage 0.8:0.1:0.1, that is, 5995 proteins in the training dataset, 744 proteins in the validation dataset, and the rest 744 proteins for the test dataset. For each protein chain, its torsion angles (\(\phi\) and \(\psi\)) are extracted by using the stride32 from its structure file that is downloaded from the Protein Data Bank (PDB)33. The angles in the training dataset are used to train the proposed network, while the validation and test datasets are used to evaluate and measure the performance of the built model. The D2020 is just used to evaluate the significance of each feature in the presented study. We utilize SPOT-1D dataset27,31 (Table S3) as a benchmark dataset to compare the ESIDEN to other best-so-far methods. Briefly, the SPOT-1D dataset (Table S3) contains 12,450 protein chains culled from the PISCES server30 with a high resolution of less than 2.5 Å, R-factor less than 1, and the cutoff of the sequence identity is set to 25%. By removing proteins of more than 700 amino acids, there are 10,029, 983, and 1213 protein chains in training sets, validation, and test (TEST2016) datasets, respectively. An additional test dataset (TEST2018) that includes 250 protein chains is also used for fair comparison among different methods, and each protein is of the resolution less than \(2.5\AA ~\) and R-factor no more than 0.25. To further evaluate different methods, we collected 59 proteins (Table S5) from the template-free modeling (TFM) targets in the Critical Assessment of protein Structure Prediction (CASP), and any of them has sequence identity less than 25% with SPOT-1D training set. There are twenty-seven, eleven, thirteen, and eight proteins from the CASP11, CASP12, CASP13, and CASP14, respectively. We also collected recently released proteins between March 2021 and June 2021 from CAEMO34. Any protein with sequence length \(>500\) was removed, and we also removed proteins with \(>25\%\) sequence identity computed by using EMBOSS tool needle35 against SPOT-1D training set. Finally, the CAMEO dataset consists of 109 proteins.

For all datasets, the features of each protein are extracted from its sequence and corresponding multiple sequence alignment (MSA). The basic features include 20 types of amino acids (AA), seven physicochemical properties (PP)29, including steric parameter, polarizability, normalized van der Waals volume, hydrophobicity, isoelectric point, helix probability, and sheet probability, and PSSM28 that is commonly used for protein secondary structure prediction28, residue contact prediction31 and torsion angles prediction13,19,36. The PSSM of each protein is derived from its MSA by searching its sequence against NCBI non-redundant dataset using PSI-BLAST37 with default parameters (e-value 0.001 and 3 iterations). In the PSSM, each amino acid has a vector that composes of 20-dimension scores, accordingly, the dimension of each PSSM is \(L\times 20\), where L is the number of amino acids of a given sequence.

The four novel features proposed in our study include the degree of conservation (DC), the relative entropy (RE), the position-specific substitution probabilities (PSSP), and the Ramachandran basin potential (RBP). To obtain the three features (DC, RE, and PSSP), we firstly prepare an MSA for each protein by searching its sequence (query) against the Uniclust30 database (as of 2/2020)38 by HHblits39, and the MSA is trimmed using Leri sequence_trim tool40. Relying on the filtered MSA, we compute the DC, RE, and PSSP of each sequence as input features to train the ESIDEN. Second, the RBP is derived based on the query sequence from the potential of torsion angles (\(\phi\) and \(\psi\)) using Leri40. The details of the four new features are presented in the following paragraphs.

The relative entropy (RE) is to measure how a probability distribution of an amino acid in the MSA is different from that of another amino acid. The RE of an amino acid at the ith position is defined as follows,

$$\begin{aligned} RE_i^a=f_i^a\ln \left( \frac{f_i^a}{p^a}\right) +(1-f_i^a) \ln \left( \frac{1-f_i^a}{1-p^a}\right) , \end{aligned}$$
(1)

where \(f_i^a\) is a probability of an amino acid a at the ith position in the MSA, and \(p^a\) is the background probability of the amino acid a. The dimension of each RE is \(L\times 20\).

The degree of conservation (DC) is derived from the same MSA as that of the RE. The DC of a given amino acid a at the ith position in the MSA is defined as

$$\begin{aligned} D_i=\sum \limits _{a=1}^{20}RE_i^a, \end{aligned}$$
(2)

\(D_i\) of the amino acid a is to measure how much conservation at the ith position in the MSA, and, generally, it provides rich evolutionary information (e.g., conservation) of structured regions (e.g., \(\alpha\)-helix and \(\beta\)-strand)41. The dimension of the DC is \(L\times 1\).

Figure 1
figure 1

Schematics of prediction system and the ESIDEN network for protein torsion angles. Starting from the primary sequence, the system extracts and processes the features (\(L\times 232\)) as inputs to feed the ESIDEN network. The ESIDEN mainly composes of LSTM (LSTM-1 is in parallel with LSTM-2 and they are concatenated to another LSTM-2) and FC modules.

The position-specific substitution probabilities (PSSP) is derived from protein sequence profiles using the evolutionary statistical energy (a Markov Random field or a Potts model in statistical physics42,43) that is defined as follows,

$$\begin{aligned} E(\tau )=\sum \limits _{i<j}e_{ij}(\tau _i, \tau _j) + \sum \limits _{i}h_{i}(\tau _i), \end{aligned}$$
(3)

where \(h_i\) and \(e_{ij}\) are site-specific bias terms of a single amino acid and coupling terms between pairwise amino acids, respectively. Without considering inter-dependencies between pairwise amino acids, the site-specific amino acid constraints \(h_i\) are used to construct the PSSP. In our study, we use the same MSA as that of RE and DC to optimize the Markov Random field by Leri40 and obtain the PSSP from the optimized site-specific bias \(h_i\) [Eq. (3)]. Without the gaps, the dimension of the PSSP of each protein is \(L\times 20\).

The Ramachandran basin potential (RBP) Neighboring amino acids have been shown to exert a strong influence on protein structure44,45. In the present study, we report for the first time that a statistical potential derived from the Ramachandran basins is used to predict the torsion angles. Briefly, the potential of torsion angles is computed from proteins with 25% sequence identity, and the potential is divided into \(72\times 72\) bins (\(5^{\circ }\times 5^{\circ }\) of each bin). To generate the RBP of a given sequence, the probability distribution of each amino acid is computed by: (1) taking advantage of two close neighbors of a central residue, that is, the left and right neighbors of the triple residues, and (2) taking predicted secondary structure of the three residues into account, e.g., the Q3 (helix/sheet/coil classes, H/E/C) or Q8 (3-turn helix/4-turn helix/5-turn helix/hydrogen-bonded turn/extended strand in parallel and/or anti-parallel \(\beta\)-sheet conformation/residue in isolated \(\beta\)-bridge/bend/coil, G/H/I/T/E/B/S/C). In the present study, we treat the secondary structure of each residue as A (not predicted), that is, its secondary structure can be any type of Q3. In the present study, as illustrated in Fig. 1, we collected all the data in the CATH40 for generating the statistical potential of the torsion angles (\(\phi\) and \(\psi\)). Each pair of the angles were extracted from the PDB files in the CATH40 database by considering the left and right neighbors of each residue, and the probabilities were computed by their frequency and mapped them to the Ramachandran map to generate the potential, e.g., to compute the probability of residue Asparagine (N) in a sequence ARNCFGD, we extracted its torsion angles by considering the residues Arginine (R) and Cysteine (C) from the structure and counted its frequency to contribute to a statistical potential.

The RBP is employed as a new feature to predict protein torsion angles. As shown in Fig. 1, each amino acid has its own Ramanchnadran basin that is derived from the statistical potential of the torsion angles (\(\phi\) and \(\psi\)). It is generated by the Leri software40, and we leverage the nonlinear dimensionality reduction technique, the t-SNE algorithm46, to reduce the original dimension \(72\times 72\) to \(72\times 2\) for efficient computation. In practice, the reduced RBP is flattened to a one-dimensional vector \(1\times 144\), and the dimension of the RBP feature is \(L\times 144\) of a protein chain.

The proposed system

The prediction system (Fig. 1) is to estimate the torsion angles (\(\phi\) and \(\psi\)) from protein primary sequence, and it mainly consists of two parts: (1) features extraction and processing from the primary sequence; and (2) predicting the torsion angle using the proposed method (ESIDEN). In addition to the basic features, the four novel features extracted from protein sequences are also leveraged in our network for highly accurate torsion angle prediction. There are seven features, including AA, PP, PSSM, RE, DC, PSSP, and RBP, that are derived from the primary sequence without any structural information (Fig. 1). The classic features (AA, PP, and PSSM) are generated and processed into a matrix data of \(L\times 47\) (as discussed above), and four novel evolutionary signatures, the RE, DC, and PSSP are computed from the same MSA of each protein and processed using in-house scripts. In the present study, both the new and the classic features are normalized into the range \([-1,1]\) to enlarge differences and highlight important components of each feature. Accordingly, we achieve a feature matrix of \(L\times 232\) for a protein chain of L, and these features are fed to the proposed network (ESIDEN) to optimize its learnable parameters.

The ESIDEN is developed based on the LSTM and fully connected (FC) modules. As illustrated in Fig. 1, we design the ESIDEN using the LSTM-1 (Eq. S1) and LSTM-2 (Eq. S2) in parallel, and they are concatenated to another LSTM-2 architecture, which can effectively understand the combination of different features of each protein sequence. As a basic LSTM network, each hidden layer cell in the LSTM-1 has an input that depends on the cell at the previous state, the LSTM-2 adopts an architecture of BiLSTM that is composed of the forward and backward LSTMs, which can capture previous and future context information. The FC layer plays an important role in effectively learning the nonlinear combination of components extracted from the input features. In the FC module, there are two different FC layers: the FC-1 and the FC-2 layers. As illustrated in Fig. 1, the FC-1 layer consists of 256 nodes with ReLU activation operator, that is, \(f(x)=x\) if \(x\ge 0\), otherwise \(f(x)=0\), and \(80\%\) dropout are used in the FC-1 layer when the ESIDEN is trained. While the other FC layer (FC-2) has four outputs with the same Sigmoid activation function \(\left( \frac{1}{1+e^{-x}}\right)\), which converts the four outputs of the ESIDEN into real values. Accordingly, the detailed architecture of the ESIDEN is designed based on the above operators, and there are a small number of learnable parameters, about 6.6M in total. Benefiting from those advantages, the ESIDEN developed into a simple and efficient neural network model for predicting the torsion angles.

Outputs

Both the torsion angles (\(\phi\) and \(\psi\)) are formed by continuously connecting four atoms located in the backbone of the protein, that is, \(\phi\) results from atoms \(C_{i-1}-N_{i}-C_{\alpha i}-C_{i}\) while \(\psi\) is computed from \(N_{i}-C_{\alpha i}-C_{i}-N_{i+1}\) (Fig. 1). They are all located in the range \([-180^{\circ }, 180^{\circ }]\), and the torsion angles \(\phi\) at the N-terminal and \(\psi\) at the C-terminal are fixed. In our study, the proposed ESIDEN method can simultaneously predict the torsion angles \(\phi\) and \(\psi\) using four outputs (\(\sin {(\phi )}\), \(\cos {(\phi )}\), \(\sin {(\psi )}\), and \(\cos {(\psi )}\)). To remove the effect of angle’s periodicity, we use the sine and cosine values of each torsion angles as targets instead of directly predicting \(\phi\) and \(\psi\), accordingly, the predicted values of an angle (P) is defined as follows,

$$\begin{aligned} P = \arctan \left( {\frac{\sin {\alpha }}{\cos {\alpha }}}\right) \end{aligned}$$
(4)

where \(\alpha\) is a representation of either the angle \(\phi\) or \(\psi\).

Performance evaluation

We evaluate the accuracy of the predicted torsion angles by the mean absolute error (MAE), which is to measure the average absolute difference between predicted angles (P) and experimental values (E) over all residues in a protein chain. To reduce the periodicity of an angle and the artificial effect, we take the minimum value between \(\left| P_{ij} - E_{ij}\right|\) and \(360^{\circ } - \left| P_{ij} - E_{ij}\right|\), i.e.

$$\begin{aligned} MAE = \frac{1}{\sum _{i=1}^{N}L_i}\sum _{i=1}^{N}\sum _{j=1}^{L_i}\min \left( 360^{\circ } - \left| P_{ij} - E_{ij}\right| , \left| P_{ij} - E_{ij}\right| \right) \end{aligned}$$
(5)

where N is the number of protein chains, \(L_i\) is the total number of residues in the ith protein chain. \(P_{ij}\) and \(E_{ij}\) are the values of predicted and experimental angles of the jth residue in the ith protein chain, respectively.

Results

The developed ESIDEN network is implemented in PyTorch v1.7.047, and it is trained on high-performance computational clusters using one NVIDIA GTX2080Ti Graphics Processing Unit (GPU). During the training, we use the Adam optimization algorithm48 with a learning rate of 0.001 to optimize the parameters of the ESIDEN network, and the mean square error (MSE) between the predicted and experimental values is defined as a loss function, which used to update the weights and biases of the network. The batch size used in the present study is set to 32, and the maximum number of iterations is 5000.

To demonstrate the performance of the developed ESIDEN, firstly we conduct experiments on our independent dataset (D2020). On the D2020 dataset (Table S1), we analyze how much each feature can contribute to improve the prediction accuracy of the ESIDEN and evaluate the importance of each feature. Further, we demonstrate that the four novel features (RE, DC, PSSP, and RBP) can improve the accuracy in predicting the torsion angles when they are combined with the basic features, and the performance of the ESIDEN is accessed by the combinations of different features on the same dataset. As further validation, we implement a fair comparison between the ESIDEN (trained by SPOT-1D training set, Table S3) and the best-so-far state-of-the-arts (Spider320, RaptorX-Angle26, SPOT-1D27, and OPUS-TASS8). Apart from TEST2016 and TEST2018 sets, we validate the ESIDEN network with the same parameters to predict the torsion angles of the fifty-nine CASP TFM targets. On the TFM targets, we compare the predicted torsion angles of the ESIDEN to those of the three best-so-far methods (Spider3, RaptorX-Angle, and SPOT-1D), which are implemented locally using their standalone packages with default configurations based on the same computing resources. The torsion angles estimated by the ESIDEN are leveraged to demonstrate its ability in predicting the tertiary structures of four representative TFM targets of the fifty-nine proteins in the CASP dataset (Table S5), and we also compare the predicted structures to those that are predicted by the other methods under the same folding configuration. Finally, we apply the proposed ESIDEN on the CAMEO dataset and compare the predictions of the torsion angles to those of the other three methods. In all the comparisons, the developed model was built on the SPOT-1D training dataset (before June 2015) for the torsion angles’ predictions of the targets in the TEST2016, TEST2018, the CASPs, and CAMEO targets.

Independent features

In this section, we compare basic features (AA, PP, and PSSM) and novel features (DC, RE, PSSP, and RBP) on the ESIDEN (Figure 1), and we also evaluate the performances measured by the MAE [Eq. (5)] between the predicted and experimental torsion angles on the validation and test dataset of the D2020 (Table S1). As illustrated in Table 1, two of the basic features (AA and PP) achieve comparable MAE of \(\phi\) and \(\psi\) using the developed model, while another basic feature PSSM is better than the two basic features. Compared to the basic features, the MAE \(\phi\) and \(\psi\) of the ESIDEN with the DC is slightly higher than the two basic features (AA and PP), as the DC loses much more information than the two features. Notably, the MAE of the model with either the RE, RBP, or PSSP is lower than those of both the features AA and PP, especially, the features PSSP and RBP achieve comparable MAE to the PSSM, as although the PSSP contains a little more noises than the PSSM, the feature PSSP is not heavily dependent on the MSAs if they have similar diversity of sequences (Table S6). Due to the lack of distinguishable characteristics of amino acids, the RBP is slightly worse than that of PSSM and PSSP. Similarly, the predicted accuracy based on the RE is better than that of the DC as DC loses more information than the RE. They all don’t preserve much discernible information about different residues as the PSSM and PSSP, and it could be the reason why PSSM and PSSP provide better predictions by comparing to that of the RE and DC on both torsion angles. Although the RE loses information, the precision of our method based on the RE is still better than that of the two basic features (AA, PP). Accordingly, we find that a single feature is not sufficient to enhance the performance of the proposed method. We further conduct analyses on different combinations of features and measure the prediction accuracy of the ESIDEN on the combined features.

Table 1 The MAE of each single feature in predicting the torsion angles using ESIDEN on the D2020 dataset.

Combined features

To address how joint contribution the novel features make to the prediction, we produce different combinations of all the features, including the basic and new ones (Table 2), to validate the performance of the ESIDEN in predicting the torsion angles (\(\phi\) and \(\psi\)).

Table 2 The MAE of combined features in predicting the torsion angles using ESIDEN on the D2020 dataset.

As shown in Table 2, we compare the MAE [Eq. (5)] performance of different feature combinations on our validation and test set by using the ESIDEN. Firstly, the combination of all the four novel features (RE + DC + PSSP + RBP) outperforms the basic features (PSSM + PP + AA) with regard to the MAE of both angles. In particular, the MAE of the angle \(\psi\) predicted by the novel features is much better than that of the basic features (PSSM, PP, and AA), that is, the MAE (\(\psi\)) is reduced by about 4 for both the validation and test set. Combined with the basic features, every single new feature can still improve the prediction accuracy on the torsion angles. For example, the RE slightly improves the performance in predicting the angles \(\phi\) and \(\psi\) when compared to that of the combined basic features. The DC outperforms both the basic features and their combination with the RE, and it would be a result of the DC preserves conservation information from protein evolution with much fewer noises than the RE. The combined features (Basic + PSSP) distinctly improve the predicted angle \(\phi\), and it’s worth noting that the combined feature (Basic + RBP) makes a significant contribution to increasing the prediction accuracy of the angle \(\psi\). The RE and DC combined with the basic features (Basic + DC + RE) are a little better than that of each in fusion with the basic features. Similar performances are achieved by the ESIDEN based on the PSSP and RBP combined with the basic features, interestingly, the RBP in the fusion of the basic features plays a significant role in improving the prediction accuracy of the angle \(\psi\). Notably, the fusion of all the basic and the new features remarkably decreases the MAEs from 18.00 to 15.72 and 25.87 to 19.77 for the angles \(\phi\) and \(\psi\), respectively, by comparing to those of the basic features (Fig. 2). These promising results demonstrate that the four novel features can significantly improve the prediction accuracy of the torsion angles, and they can make the prediction much better when combined with the basic features. Therefore, we use the combination of basic and novel features as the input features to build a small but efficient network model in the present study.

Figure 2
figure 2

Comparison of prediction performance using different combinations of features on the D2020 dataset. The MAE as measurements between the predicted and experimental values of (a) the torsion angle \(\phi\) and (b) the torsion angle \(\psi\), respectively.

Comparisons to other the-state-of-the-arts

In this section, we compare the MAE [Eq. (5)] performance of the proposed ESIDEN to those of other best-so-far methods (Spider320, RaptorX-Angle26, SPOT-1D27 and OPUS-TASS8) on the widely used datasets, TEST2016 and TEST2018 (Table S3).

Table 3 Performance of different methods on the TEST2016 and TEST2018.

As shown in Table 3, on the TEST2016 dataset, the Spider3 and the RaptorX-Angle achieve comparable MAE of the angle \(\psi\), but the Spider3 slightly outperforms the RaptorX-Angle in terms of \(\phi\) MAE. The hybrid model, OPUS-TASS, performs better measured by the MAE of \(\phi\) than the SPOT-1D, although the MAE of the predicted \(\psi\) by the SPOT-1D is a little better than that of the OPUS-TASS. Our method, the ESIDEN, achieves the best MAE accuracy on the TEST2016 dataset, as shown, it obtains better MAE of the angle \(\phi\) than that of the OPUS-TASS on the TEST2016 dataset, and the MAE of the angle \(\psi\) inferred by the ESIDEN is much better than all the other four compared methods by reducing the MAE by more than 5 degrees. On the TEST2018 dataset, the RaptorX-Angle has underperformed the other compared methods with regard to the MAE of both \(\phi\) and \(\psi\). The SPOT-1D and OPUS-TASS obtain similar MAEs of the angles \(\phi\) and \(\psi\), and both of them are better than that of the Spider3. On the same dataset, the ESIDEN achieves a better MAE of \(\phi\) than those of all the other four compared methods. It’s worth noting that the MAE of the angle \(\psi\) is much decreased by the ESIDEN compared to those of the rest methods. The results demonstrate that the ESIDEN outperforms among the compared methods, especially in terms of the \(\psi\) MAE, and it significantly improves the prediction accuracy of the angle \(\psi\), decreasing the MAE by 3.8 degrees over the OPUS-TASS (the second best) on both of the benchmark datasets. Benefiting from evolutionary signatures, the delicately designed architecture of the ESIDEN accounts for the distinguishing outperformance on accurately predicting the torsion angles.

Validations on the CASPs and CAMEO

The CASP datasets are widely used to evaluate the performance of different predictors on structural informatics. Here, we collect 59 TFM targets from the recent CASPs (CASP11, CASP12, CASP13, and CASP14, Table S5), and the proposed ESIDEN is validated by comparison to the best-so-far methods (Spider320, RaptorX-Angle26 and SPOT-1D27) in predicting the torsion angles. As shown in Table 4, on the four CASP datasets, the MAE performance of ESIDEN is slightly better than those of all the compared methods, while the decrement of its MAE of \(\psi\) is more than 2 degrees, except 1.7 degrees on the CASP13, by comparing to those of the others. On the CASP11, CASP12, CASP13, and CASP14 datasets, the performance of our method consistently outperforms those of all the other three methods, especially, the \(\psi\) MAE of the ESIDEN is distinguishably better than those of the other methods. The average torsion angle prediction errors (measured by the MAE) of the ESIDEN are demonstrated on the fifty-nine TFM targets across eight types of secondary structures (Q8) (Fig. 3). There are no secondary structures (5-turn helix I, and bend S) in the CASP datasets, and the result shows that the prediction errors of both \(\phi\) and \(\psi\) for H (helix) are the lowest by comparing other Q8 types, yet the secondary structures B/T/C are higher, as variational coiled coils and single \(\beta\) strands are not stable as the \(\alpha\)-helix. On the CAMEO dataset, the MAE of the torsion angle \(\phi\) predicted by the ESIDNE is comparable to that of SPOT-1D and lower than those of Spider3 and RaptorX-Angle, while the MAE of the angle \(\psi\) of ESIDEN outperforms all the other compared methods (Table S4).

Table 4 Comparison among different methods on the TFM targets of the CASP11, CASP12, CASP13, CASP14, and CAMEO.
Figure 3
figure 3

The MAE of the torsion angles (\(\phi\) and \(\psi\)) on the 8-class secondary structures for the TFM targets in the CASPs and CAMEO.

Figure 4
figure 4

The predicted structures of four representative TFM targets using the torsion angles estimated by the ESIDEN. The cartoon structures were made by PyMOL49.

We model four representative TFM targets of the 59 TFM targets in the CASPs using the predicted torsion angles (Figs. S2S5), including T0968s2-D1, T0986s1-D1, T0957s1-D1, and T0969-D1 (Table S5), and the inferred angles are used as a constraint to launch an improved folding module based on the method in the study4. In the presented study, we applied a coarse-grained molecular dynamics simulation to sample the space of torsional angles (\(\phi\) and \(\psi\)). As an improved module, it is embedded as a module of Leri. In the module, we employed and optimized the energy functions in the study7 to rank the predicted structures. The atoms are moved by Newton’s laws of motion on the space of the torsion angles that is derived from the Ramachandran potential using the NDRD TCB library51, and the moving is smooth to sample the space for possible conformations of the query sequence. In a round of the folding simulation, given a query sequence, we first generate a Ramachandran map for each residue and reduce the searching space by the predicted angles from the proposed method. The module moves on the Ramachandran map of each residue and generated a possible conformation, and the conformation is evaluated by the improved energy function using the metropolis criterion for best-so-far structure. When the round is stopped, the top 20% of candidates with the lowest energy are chosen to generate new structural constraints, including torsion angles, residue-distances, and hydrogen-bond constraints. The information is recycled as enhanced constraints for the next round simulation.

Ten folding simulations are launched for each target. For the targets T0986s1-D1, T0968s2-D1, and T0957s1-D1, we implement ten folding simulations with 100,000 iterations and obtain 200 structures with lower energy from each trajectory. As the number of residues of the target T0969-D1 is large, we conduct 200,000 iterations and obtain 200 structures with lower energy from each trajectory. The largest cluster is obtained over the two thousand structures of each target, and the centroid structure in the cluster is compared to its native structure, as well as the best-so-far structure (Fig. 4). The root-mean-square deviation (RMSD) and TM-score of each target are computed by the TMscore software50, and they are presented in Table S7. Although the coiled structures of either the best-so-far or the centroid structures are not always in agreement with those in the native structures, the secondary structures (\(\alpha\)-helix and \(\beta\)-strand) in both the best-so-far and centroid structures share the same ones as those in the native structure of each target. Accordingly, the predicted torsion angles can accelerate protein folding and structure prediction, and they can also improve structural precision.

Discussion

The classical features extracted from a protein sequence are powerful input data for the ML-based methods, which predict protein structural properties, such as residue contacts, residue distances, backbone torsion angles, solvent accessible surface area, protein-protein interaction, and protein function in computational biology. They have several advantages for characterizing amino acid types and sequential order of amino acids that can be able to specify structural properties. Therefore, accurate prediction of structural properties can provide valuable information to infer protein tertiary structure and function. Recently, incorporation of pre-estimated structural features, e.g., secondary structure and solvent accessibility, into the ML-based methods has been a practical way to improve the predictions for either protein structure or function. Nevertheless, it still remains challenging to further improve the prediction accuracy of these key structural properties, such as the torsion angles, merely based on the classical features. Extracting efficient features from the large-scale sequence profiles increasingly advances the capabilities of computational methods, especially DL-based algorithms, to draw out knowledge and address important biological questions, e.g., the quantitative relationship between protein structure and its function.

In the present study, we propose an evolutionary signatures-driven deep neural network-based system (ESIDEN) and four novel features computed from evolutionary signatures and the Ramachandran basins for predicting protein torsion angles. The ESIDEN adopts a straightforward architecture of a small number of learnable parameters, and yet it achieves high accuracy in the torsion angle prediction in comparison to other state-of-the-art approaches. Furthermore, we also developed four novel features that are derived from protein evolutionary signatures to improve the performance of ESIDEN. Generally, it is more difficult to predict the angle \(\psi\) than \(\phi\), as the diversity of \(\psi\) results from widely different secondary structures for variational amino acids in proteins. In contrast to existing methods that leverage the classical features by combining other features to substantially improve performance, the ESIDEN is built upon the different features, especially, the ESIDEN with only the RBP can still achieve similar performance by comparison to that of PSSM. The results demonstrate that the newly developed features distinguishably improve the accuracy in predicting the angle \(\psi\). Moreover, the recurrent architecture in the ESIDEN, can capture sequential motifs hidden in the amino acid residues and their neighbors. On the TEST2016, TEST2018, and the CASP datasets, we demonstrate the ESIDEN achieves higher precision of predictions on the torsion angles by comparison to the best-so-far methods. The accurately predicting the torsion angles is the result of the efficient architecture of the ESIDEN and the new features that contribute to the classical ones.

Limitations of the model include biases that arise from features filtered by the dimension-reduction methods on the Ramachandran basins, which may result in computationally intractable reduction if improper methods are used and evolutionarily younger families of limited diversity that result in many noises to the feature PSSP. Although incorporating evolutionary signatures into the ESIDEN results in a practical improvement over other methods, challenges remain in the precise interpretation of the model.

The success of the ESIDEN is based on deep learning at recapitulating large-scale data from protein sequence profiles. For example, the Ramachandran basins could robustify other deep learning-based methods for many applications in predicting protein contacts/distances, secondary/tertiary structure, and designing proteins. We anticipate that the ESIDEN and the new features developed in the present study can be utilized by other deep-learning-based methods applications ranging from drug discovery to protein design. The consistency of our estimated torsion angles with the authentic angles highlights how the inclusion of evolutionary signatures will facilitate more accurate inferences and aid the prediction/determination of the tertiary structures of protein sequences.