# Derivative-free neural network for optimizing the scoring functions associated with dynamic programming of pairwise-profile alignment

## Abstract

### Background

A profile-comparison method with position-specific scoring matrix (PSSM) is among the most accurate alignment methods. Currently, cosine similarity and correlation coefficients are used as scoring functions of dynamic programming to calculate similarity between PSSMs. However, it is unclear whether these functions are optimal for profile alignment methods. By definition, these functions cannot capture nonlinear relationships between profiles. Therefore, we attempted to discover a novel scoring function, which was more suitable for the profile-comparison method than existing functions, using neural networks.

### Results

Although neural networks required derivative-of-cost functions, the problem being addressed in this study lacked them. Therefore, we implemented a novel derivative-free neural network by combining a conventional neural network with an evolutionary strategy optimization method used as a solver. Using this novel neural network system, we optimized the scoring function to align remote sequence pairs. Our results showed that the pairwise-profile aligner using the novel scoring function significantly improved both alignment sensitivity and precision relative to aligners using existing functions.

### Conclusions

We developed and implemented a novel derivative-free neural network and aligner (Nepal) for optimizing sequence alignments. Nepal improved alignment quality by adapting to remote sequence alignments and increasing the expressiveness of similarity scores. Additionally, this novel scoring function can be realized using a simple matrix operation and easily incorporated into other aligners. Moreover our scoring function could potentially improve the performance of homology detection and/or multiple-sequence alignment of remote homologous sequences. The goal of the study was to provide a novel scoring function for profile alignment method and develop a novel learning system capable of addressing derivative-free problems. Our system is capable of optimizing the performance of other sophisticated methods and solving problems without derivative-of-cost functions, which do not always exist in practical problems. Our results demonstrated the usefulness of this optimization method for derivative-free problems.

## Keywords

Dynamic programming Profile alignment Neural network Evolutionary strategy Derivative-free optimization## Abbreviations

- CMA-ES
covariance matrix adaptation evolution strategy

- DP
dynamic programming

- PSSM
position-specific scoring matrix

- PSSV
position-specific scoring vector

## Background

The profile-comparison alignment method with a position-specific scoring matrix (PSSM) [1] is a highly accurate alignment method. The PSSM is a two dimensional vector (matrix) that stores sequence lengths, with each element in the vector consisting of a 20-dimensional numerical vector where each value represents the likelihood of the existence of each amino acid at a site in a biological sequence. Here, we designed the vector inside a PSSM as a position-specific scoring vector (PSSV). In profile alignment, cosine similarity or the correlation coefficient between two PSSVs is generally computed to measure similarity or dissimilarity between the two sites in the sequences of interest using dynamic programming (DP) [2, 3]. Profile alignment methods using these functions have long been used successfully [4], and the performance of profile alignment has improved in recent decades. As examples, HHalign improved alignment quality using profiles constructed with a hidden Markov model, which provided more information than a PSSM [5], MUSTER incorporated protein-structure information into a profile [3], and MRFalign utilized Markov random fields to improve alignment quality [6]. However, although various methods have been devised from different perspectives, studies to develop the scoring function for PSSV comparison using sophisticated technologies are lacking. Moreover, there remains room for improvement in the performance of sequence alignment, especially for remote sequence alignment [7, 8, 9]; therefore, it is important to continue developing aligners from various perspectives. Although cosine similarity or a correlation coefficient is normally used for comparison of PSSVs, in principle, they are unable to capture nonlinear relationships between vectors. However, the similarity between two amino acid positions is not always explained by linear relationship, which is merely one of a particular case of a nonlinear relationships. Because scoring functions are directly related to the quality of biological-sequence alignment, development of a novel function capable of capturing nonlinear relationships reflecting similarity between two sites in sequences is required.

The expression of nonlinear functions can be realized by neural networks. A neural network is a computing system that mimics biological nervous systems. Theoretically, if a proper activation function is set on middle layer(s) of a network, it can approximate any function including nonlinear functions [10]. Neural networks have attracted interest from various areas of research, including bioinformatics, due to recent advances in computational technologies and the explosive increase in available biological data. In recent years, these algorithms have been vigorously applied for bioinformatics purposes, including several studies associated with application of deep neural network models to predict protein–protein interactions [11, 12], protein structure [13, 14], and various other biological conditions, such as residue-contact maps, backbone angles, and solvent accessibility [15, 16]. These neural networks used backpropagation as a solver, which requires a derivative-of-cost function to search for optimal parameters [17]. However, few studies have implemented derivative-free neural networks.

Since neural networks are capable of implementing nonlinear functions, they are suitable for developing novel scoring functions for PSSV comparison. Therefore, in this study we used a neural network to optimize a nonlinear scoring function associated with PSSV comparison by combining two PSSVs as an input vector. Since we lacked a target vector normally required to implement supervised learning, we calculated the entire DP table for the input sequences, and the difference between the resultant alignment and the correct alignment was used to calculate cost of learning. Due to the nature of the problem, we could not use the backpropagation method as a solver for optimal weight and bias searches, because we lacked the derivative-of-cost function normally required. These issues are common when applying such methods to real-world problems. It is impossible to calculate a derivative for problems where the output vectors are not directly used for computation of cost function such as cross entropy or square error [18]. In this study, the outputs of a neural network were similarity score between two PSSVs and not directly used for computation of the cost function but indirectly used for computation of dynamic programming. The possibility of computing neural network inferences without derivatives would be useful for solving such problems.

Here, we used a covariance matrix adaptation-evolution strategy (CMA-ES) [19] as a solver for the neural network to implement a derivative-free neural network system. CMA-ES is an adaptive-optimization method that modifies the basic evolutionary strategy [20]. As advantages, it requires a smaller number of hyperparameters than other evolutionary strategy methods [19], and when the dimensionality of an objective function is large, it offers higher computation speeds relative to other derivative-free optimization methods, such as the Nelder–Mead method, which requires computation times proportional to the dimensionality of the objective function [21]. In this study, we implemented a derivative-free neural network system using CMA-ES and produced a high-performance scoring function for remote-sequence alignment. Our goal was to develop a novel scoring function for profile alignment method and provide a novel derivative-free learning method useful for optimizing derivative-free problems.

## Methods

### Dataset

We downloaded the non-redundant subset of SCOP40 (release 1.75) [22], in which sequence identity between any sequence pair is < 40%, from ASTRAL [23]. We selected the remote-sequence subset, because we wanted to improve remote-sequence alignment quality, which is generally a difficult problem for sequence aligners. SCOP is a protein-domain database where sequences are classified in a hierarchical manner by *class*, *fold*, *superfamily*, and *family*. To guarantee independence between a learning and test dataset, all notations of *superfamily* in the dataset were sorted in alphabetical order, and all *superfamilies*, the ordered numbers of which were multiples of three, were classified into a learning dataset, whereas the others were classified into a test dataset. This procedure is often used in existing studies for protein sequence analysis [8, 9], in order to cope with a problem of overfitting. We obtained 3726 and 6843 sequences in the learning and test datasets, respectively. We then randomly extracted a maximum of 10 pairs of sequences from each *superfamily* to negate a bias induced by different volumes of each *superfamily* and used these sequence pairs for subsequence construction of a PSSM. We confirmed that sequences in each pair were from the same *family* in order to obtain decent reference alignments. We ultimately obtained 1721 and 3195 sequence pairs in the learning and test datasets, respectively. These datasets are provided at https://github.com/yamada-kd/nepal.

### Construction of profiles and reference alignments

We constructed PSSMs for all sequences in the learning and test datasets using DELTA-BLAST version 2.2.30+ with the Conserved Domain Database for DELTA-BLAST version 3.12 [24]. Reference alignments were constructed through structural alignment of protein steric structures, which corresponded to sequences of interest using TM-align [25]. All structure data were also downloaded from ASTRAL [23].

### Learning network

**x**_{ a }(PSSV A) and

**W**_{1a},

**x**_{ b }(PSSV B) and

**W**_{1b,}and 1 (bias) and

**b**_{1}in the neural network were calculated. Here,

**x**_{ a }and

**x**_{ b }were 20-element vector calculated from a DELTA-BLAST search, where each element of the vector represented the likelihood of existence of each amino acid, and

**W**_{1a},

**W**_{1b,}1, and

**b**_{1}were weight and bias parameters of the neural network. The resultant vector was transformed by an activating function,

*φ*(

*u*). The rectified linear unit [26] was utilized as the activation function:

The summation of the dot products between the transformed vector, *φ*(* u*) and

**w**_{2}, and 1 and

*b*

_{2}was calculated, where

*was a vector representing the middle layer, and*

**u**

**w**_{2}, 1, and

*b*

_{2}were parameters of the neural network. The resultant value was used as the similarity score for the two sites. Namely, the forward calculation was computed by the equation:

*y*, a scalar value, is the similarity score.

The complete DP table was calculated using the similarity score, and a final pairwise alignment was produced. The pairwise alignment and its corresponding reference alignment were compared to each other, and an alignment sensitivity score was calculated. Subtraction of the alignment-sensitivity score from 1 was used as the cost for searching the optimal weight using the neural network with CMA-ES.

**W**_{1a}and

**W**_{1b}equal to each other (shared weight) in order to apply the same value to the network outputs, even though the input order of the two PSSVs was opposite one another:

The number of units of the middle layer was set to 144. To compute backward calculations for the network, we used CMA-ES. As hyperparameters for CMA-ES, we set σ, λ, and μ to 0.032, 70, and 35, respectively. Here, σ is almost equivalent to the step size (learning rate) of the normal gradient-descent method, and λ and μ indicate the number of descendant and survival individuals in the evolutionary process, respectively. We input training datasets into the learning system in a batch manner. The maximum number of epochs was set to a relatively small number (150) to accommodate our computational environment. During learning, the performance of the scoring function was evaluated on the validation dataset starting from the 50th epoch to the final epoch in five steps, and a scoring function that maximized the validation score was selected as the final product of the learning process. The initial weight and bias were derived from parameters that mimicked the correlation coefficient. To generate the initial weight, we randomly generated 200,000 PSSV pairs and learned them using multilayer perceptron with hyperparameters (the dimensions of the weight and activating function) identical to those already described. In addition to the parameters, we simultaneously optimized the open- and extension-gap penalties, the initial values of which were set to − 1.5 and − 0.1, respectively. The source code for our learning method is provided at https://github.com/yamada-kd/nepal.

### Alignment algorithm

In this study, we implemented the semi-global alignment method (global alignment with free-end-gaps) [27, 28].

### Metrics of alignment quality

Alignment quality was evaluated using alignment sensitivity and precision [9]. The alignment sensitivity was calculated by dividing the number of correctly aligned sites by the number of non-gapped sites in a reference alignment. By contrast, alignment precision was calculated by dividing the number of correctly aligned sites by the number of non-gapped sites in a test alignment.

### Calculation of residue interior propensity

The relative accessible surface area (rASA) for residues of all proteins in the learning and test datasets was calculated by areaimol in the CCP4 package version 6.5.0 [29]. The residues associated with rASA < 0.25 were counted as interior residues, and the other residues were counted as surface residues based on methods used previously [30]. We divided the ratio of the interior residues by the background probability associated with these residues to calculate the residue interior propensity, which represented the likelihood of a residue existing inside a protein. A propensity > 1 signified that the probability of the residue being inside the protein was higher than expected.

### Statistical analysis

Statistical tests, including Wilcoxon signed-rank test with Bonferroni correction and Spearman’s rank correlation, were computed using the functions pairwise.wilcox.test() and cor.test() from R version 2.15.3 (https://cran.r-project.org/), respectively.

## Results and discussion

### Gap optimization of existing functions

Gap optimization of the existing scoring function

Open | Extension | Sensitivity | Precision | |
---|---|---|---|---|

Cosine | − 1.0 | − 0.1 | 0.6837 | 0.6550 |

CC | − 1.5 | − 0.1 | 0.6882 | 0.6613 |

### Optimization of the scoring function and gap penalties

We then optimized the scoring function on the neural network with CMA-ES. During learning, we randomly divided the learning dataset into two subsets (training and validation datasets) and observed training and validation curves to confirm overfitting did not occur. The learning and validation dataset included 1536 and 160 pairwise PSSM sets and the corresponding reference alignments as targets, respectively. Because calculation of learning using our parameter settings requires > 100,000 × DP (the size of the training dataset × λ) per epoch, the consumption of computer resources was large, and calculation time was long, even when 24 threads were used with the C++ program. Therefore, we set the maximum limit for epoch to a relatively small number (150). To maximize the learning within the finite learning time, we monitored the performance of intermediate scoring functions on the validation dataset every fifth epoch. According to the validation scores, we ultimately selected a scoring function derived from the 145th epoch, which maximized the validation score, as the final product of learning. In addition to the scoring function, open- and extension-gap penalties are also vital parameters for DP, which outputs optimal alignments against four parameters, including the pairwise sequences, a scoring function, and open- and extension-gap penalties. We optimized the gap penalties along with other parameters, and simultaneously optimized gap penalties using a scoring function to obtain final weight and bias matrices representing the substance of a novel scoring function and optimal gap-penalty combinations, respectively. Our results allowed realization of an optimal combination of open- and extension-gap penalties for the final weight and bias matrices (approximately − 1.7 and − 0.2, respectively).

We implemented a pairwise-profile aligner with the weight and bias matrices as a novel scoring function and named it Neural network Enhanced Profile Alignment Library (Nepal). Nepal accepts pairwise sequences and their corresponding PSSM as an input and outputs a pairwise alignment for the input sequences. The scoring function is performed by a neural network, and the similarity score, *y*, between two PSSVs (**x**_{ a } and **x**_{ b }) is calculated using Eq. 2, with three weight (**W**_{1a}, **W**_{1b}, and **w**_{ 2 }) and two bias (**b**_{1} and *b*_{2}) matrices the final products of learning. Our aligner and scoring function (weight and bias matrices) can be downloaded from https://github.com/yamada-kd/nepal.

### Benchmarking of Nepal and other aligners using an existing function on the test dataset

We then conducted a benchmark test of Nepal and other aligners using an existing function on the test dataset. In addition to profile-comparison methods, we examined the performance of sequence-comparison aligners with different substitution matrices, such as BLOSUM62 [31] and MIQS [32], as references. We used − 10 and − 2 as open- and extension-gap penalties, respectively, based on a previous study [32]. When calculating alignment quality, the test dataset was further categorized into remote and medium subsets depending on the pairwise sequence identity of the reference alignments. The remote and medium subsets included sequence pairs where each sequence identity was not < 0 and < 20% and not < 20 and < 40%, respectively. Generally, a pairwise alignment between sequences of lower identity under the twilight zone is a more difficult problem [7].

Comparison of Nepal with other alignment methods

Remote [0,20) (1405 files) | Medium [20,40) (1790 files) | All [0,40) (3195 files) | |
---|---|---|---|

Sensitivity | |||

Nepal | 0.5317 | 0.8343 | 0.7012 |

Cosine | 0.5045** | 0.8246** | 0.6838** |

CC | 0.5135** | 0.8269** | 0.6891** |

MIQS | 0.2775** | 0.7316** | 0.5319** |

BL62 | 0.2333** | 0.6955** | 0.4923** |

Precision | |||

Nepal | 0.5031 | 0.8102 | 0.6751 |

Cosine | 0.4753** | 0.7999** | 0.6571** |

CC | 0.4858** | 0.8032** | 0.6636** |

MIQS | 0.2654** | 0.7134** | 0.5164** |

BL62 | 0.2317** | 0.6902** | 0.4885** |

### Importance of attributes according to the connection-weight method

Based on these results, the connection weights of hydrophobic residues, such as Leu, Ile, and Val, were of higher value. These residues are located mostly inside the hydrophobic cores of proteins. Additionally, as shown in Fig. 2b, other residues, which often buried within proteins, such as Ala, Cys, and Tyr, were also of higher importance. By contrast, residues often located on the protein surface, such as Asp, Pro, Lys, and Asn, were of lower importance. The Spearman’s rank correlation coefficient between the connection weight and interior propensity was ~ 0.6 (P < 0.05), meaning that the importance of attributes was related to the propensity of residues to be located on the interior of the protein. While residues located at the protein surface are subject to higher mutation rates, buried residues are less susceptible to mutation [35], because protein structure can be disrupted by mutation of residues buried in the core of the protein, which could potentially result in collapse of the hydrophobic core [36]. The scoring function presented in this study was optimized for the alignment of remote homologous sequences. According to a previous study based on substitution matrices [37], residue hydrophobicity was the dominant property of remote sequence substitution rather than simple mutability. This fact partially explains why residues occupying interior locations are considered more meaningful for remote sequence alignment. Because our scoring function was optimized for remote sequence alignment, it considered these amino acids as important attributes. This characteristic of the scoring function represents a superior attribute of our method relative to existing methods.

*) and B (*

**a***), resulting in 0.488207 and 0.387911 when calculated using the correlation coefficient and Nepal methods, respectively (Fig. 3, middle panel). The scores calculated using the correlation coefficient did not change when the 1st and 18th sites or the 4th and 19th sites were swapped. These results could be inappropriate, because the converted PSSV obtained after swapping was not identical to the original, which could represent a potential drawback of using unweighted linear functions, such as cosine similarity and correlation coefficient. By contrast, the Nepal scores changed after swapping and varied along with changes in the PSSV. This expressiveness represents a merit of nonlinear functions. There were ~ 290,000 overlaps following the calculation of similarity scores to six decimal places against 1 million randomly generated PSSVs using the correlation coefficient method, whereas there were ~ 180,000 overlaps when Nepal was used. These overlaps would negatively affect DP computation, because higher overlap scores would cause difficulties in determining the correct path, especially during the computation of a maximum of three values derived from different sides of DP cell. Our results showed that the use of different weights by the connection-weight method and based on amino acid variety is one reason why the Nepal scoring method improved alignment quality as compared with the existing scoring functions.*

**b**## Conclusions

In this study, we optimized a scoring function for pairwise-profile alignment using a machine-learning method mimicking a nonlinear function. Our method enabled computational optimization, regardless of whether given problem involved a derivative-of-cost function, given that this scenario is not always present in real-world problems. In this study, we developed a novel derivative-free neural network with CMA-ES and successfully applied this learning system to optimize a scoring function for pairwise-profile alignment. Nepal significantly improved the alignment quality of profile alignments, especially for alignments based on remote relationships, as compared with existing scoring functions. Moreover, Nepal improved alignment quality based on the adaptation to remote sequence alignment and the increasing expressiveness of the similarity score. This method alone is not practical as a standalone pairwise-profile aligner; however, because the novel scoring function involves a simple matrix operation using parameters provided on the website, the performance of distant homology detection or multiple-sequence-alignment methods for remote homologous sequences might be further improved by incorporation of our scoring function. Finally, the goal of the study was not only to provide an alternative alignment method but also to provide a novel learning system capable of addressing derivative-free problems. Our system will be useful for optimizing the scoring functions of other sophisticated methods such as similarity search, multiple-sequence alignment and etc.

## Notes

### Authors’ contributions

The author did everything. The author read and approved the final manuscript.

### Acknowledgements

We are grateful to Dr. Kentaro Tomii, Dr. Satoshi Omori, and Mr. Tsukasa Nakamura for constructive discussions. Computations were partially performed on the NIG supercomputer at the ROIS National Institute of Genetics.

### Competing interests

The author declares that there is no competing interests.

### Availability of data and materials

The source code of the derivative-free neural network, Nepal and the learned parameters are available at GitHub (https://github.com/yamada-kd/nepal).

### Ethics approval and consent to participate

Not applicable.

### Funding

This work was supported in part by the Top Global University Project from the Ministry of Education, Culture, Sports, Science, and Technology of Japan (MEXT) and Platform Project for Supporting in Drug Discovery and Life Science Research (Basis for Supporting Innovative Drug Discovery and Life Science Research (BINDS)) from AMED under Grant Number 17am0101067.

### Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

## References

- 1.Altschul SF, Madden TL, Schaffer AA, Zhang J, Zhang Z, Miller W, Lipman DJ. Gapped BLAST and PSI-BLAST: a new generation of protein database search programs. Nucleic Acids Res. 1997;25(17):3389–402.CrossRefPubMedPubMedCentralGoogle Scholar
- 2.Tomii K, Akiyama Y. FORTE: a profile-profile comparison tool for protein fold recognition. Bioinformatics. 2004;20(4):594–5.CrossRefPubMedGoogle Scholar
- 3.Wu S, Zhang Y. MUSTER: improving protein sequence profile-profile alignments by using multiple sources of structure information. Proteins. 2008;72(2):547–56.CrossRefPubMedPubMedCentralGoogle Scholar
- 4.Tomii K, Hirokawa T, Motono C. Protein structure prediction using a variety of profile libraries and 3D verification. Proteins. 2005;61(Suppl 7):114–21.CrossRefPubMedGoogle Scholar
- 5.Söding J. Protein homology detection by HMM–HMM comparison. Bioinformatics. 2005;21(7):951–60.CrossRefPubMedGoogle Scholar
- 6.Ma J, Wang S, Wang Z, Xu J. MRFalign: protein homology detection through alignment of Markov random fields. PLoS Comput Biol. 2014;10(3):e1003500.CrossRefPubMedPubMedCentralGoogle Scholar
- 7.Rost B. Twilight zone of protein sequence alignments. Protein Eng. 1999;12(2):85–94.CrossRefPubMedGoogle Scholar
- 8.Angermuller C, Biegert A, Soding J. Discriminative modelling of context-specific amino acid substitution probabilities. Bioinformatics. 2012;28(24):3240–7.CrossRefPubMedGoogle Scholar
- 9.Biegert A, Soding J. Sequence context-specific profiles for homology searching. Proc Natl Acad Sci USA. 2009;106(10):3770–5.CrossRefPubMedPubMedCentralGoogle Scholar
- 10.Cybenko G. Approximations by superpositions of sigmoidal functions. Math Control Signal. 1989;2(4):303–14.CrossRefGoogle Scholar
- 11.Sun T, Zhou B, Lai L, Pei J. Sequence-based prediction of protein protein interaction using a deep-learning algorithm. BMC Bioinform. 2017;18(1):277.CrossRefGoogle Scholar
- 12.Du X, Sun S, Hu C, Yao Y, Yan Y, Zhang Y. DeepPPI: boosting prediction of protein–protein interactions with deep neural networks. J Chem Inf Model. 2017;57(6):1499–510.CrossRefPubMedGoogle Scholar
- 13.Wang S, Peng J, Ma J, Xu J. Protein secondary structure prediction using deep convolutional neural fields. Sci Rep. 2016;6:18962.CrossRefPubMedPubMedCentralGoogle Scholar
- 14.Spencer M, Eickholt J, Cheng J. A deep learning network approach to ab initio protein secondary structure prediction. IEEE/ACM Trans Comput Biol Bioinform. 2015;12(1):103–12.CrossRefPubMedGoogle Scholar
- 15.Di Lena P, Nagata K, Baldi P. Deep architectures for protein contact map prediction. Bioinformatics. 2012;28(19):2449–57.CrossRefPubMedPubMedCentralGoogle Scholar
- 16.Heffernan R, Yang Y, Paliwal K, Zhou Y. Capturing non-local interactions by long short term memory bidirectional recurrent neural networks for improving prediction of protein secondary structure, backbone angles, contact numbers, and solvent accessibility. Bioinformatics. 2017;33(18):2842–9.CrossRefPubMedGoogle Scholar
- 17.Erb RJ. Introduction to backpropagation neural network computation. Pharm Res. 1993;10(2):165–70.CrossRefPubMedGoogle Scholar
- 18.LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. 2015;521(7553):436–44.CrossRefPubMedGoogle Scholar
- 19.Hansen M, Ostermeier A. Adapting arbitrary normal mutation distributions in evolution strategies: the covariance matrix adaptation. In: Proceedings of IEEE international conference on evolutionary computation, 20 May 1996. IEEE; 1996. p. 312–7.Google Scholar
- 20.Beyer H. Toward a theory of evolution strategies: some asymptotical results from the (1,(+) lambda)-theory. Evolut Comput. 1993;1(2):165–88.CrossRefGoogle Scholar
- 21.Nelder JA, Mead R. A simplex-method for function minimization. Comput J. 1965;7(4):308–13.CrossRefGoogle Scholar
- 22.Andreeva A, Howorth D, Chandonia JM, Brenner SE, Hubbard TJ, Chothia C, Murzin AG. Data growth and its impact on the SCOP database: new developments. Nucleic Acids Res. 2008;36(Database issue):D419–25.PubMedGoogle Scholar
- 23.Chandonia JM, Hon G, Walker NS, Lo Conte L, Koehl P, Levitt M, Brenner SE. The ASTRAL compendium in 2004. Nucleic Acids Res. 2004;32(Database issue):D189–92.CrossRefPubMedPubMedCentralGoogle Scholar
- 24.Boratyn GM, Schaffer AA, Agarwala R, Altschul SF, Lipman DJ, Madden TL. Domain enhanced lookup time accelerated BLAST. Biol Direct. 2012;7:12.CrossRefPubMedPubMedCentralGoogle Scholar
- 25.Zhang Y, Skolnick J. TM-align: a protein structure alignment algorithm based on the TM-score. Nucleic Acids Res. 2005;33(7):2302–9.CrossRefPubMedPubMedCentralGoogle Scholar
- 26.Hahnloser RH, Sarpeshkar R, Mahowald MA, Douglas RJ, Seung HS. Digital selection and analogue amplification coexist in a cortex-inspired silicon circuit. Nature. 2000;405(6789):947–51.CrossRefPubMedGoogle Scholar
- 27.Gotoh O. An improved algorithm for matching biological sequences. J Mol Biol. 1982;162(3):705–8.CrossRefPubMedGoogle Scholar
- 28.Needleman SB, Wunsch CD. A general method applicable to the search for similarities in the amino acid sequence of two proteins. J Mol Biol. 1970;48(3):443–53.CrossRefPubMedGoogle Scholar
- 29.Winn MD, Ballard CC, Cowtan KD, Dodson EJ, Emsley P, Evans PR, Keegan RM, Krissinel EB, Leslie AG, McCoy A, et al. Overview of the CCP4 suite and current developments. Acta Crystallogr D Biol Crystallogr. 2011;67(Pt 4):235–42.CrossRefPubMedPubMedCentralGoogle Scholar
- 30.Levy ED. A simple definition of structural regions in proteins and its use in analyzing interface evolution. J Mol Biol. 2010;403(4):660–70.CrossRefPubMedGoogle Scholar
- 31.Henikoff S, Henikoff JG. Amino acid substitution matrices from protein blocks. Proc Natl Acad Sci USA. 1992;89(22):10915–9.CrossRefPubMedPubMedCentralGoogle Scholar
- 32.Yamada K, Tomii K. Revisiting amino acid substitution matrices for identifying distantly related proteins. Bioinformatics. 2014;30(3):317–25.CrossRefPubMedGoogle Scholar
- 33.Fariselli P, Rossi I, Capriotti E, Casadio R. The WWWH of remote homolog detection: the state of the art. Brief Bioinform. 2007;8(2):78–87.CrossRefPubMedGoogle Scholar
- 34.Olden JD, Joy MK, Death RG. An accurate comparison of methods for quantifying variable importance in artificial neural networks using simulated data. Ecol Model. 2004;178(3–4):389–97.CrossRefGoogle Scholar
- 35.Zhang J, Yang JR. Determinants of the rate of protein sequence evolution. Nat Rev Genet. 2015;16(7):409–20.CrossRefPubMedPubMedCentralGoogle Scholar
- 36.Chakravarty S, Varadarajan R. Residue depth: a novel parameter for the analysis of protein structure and stability. Struct Fold Des. 1999;7(7):723–32.CrossRefGoogle Scholar
- 37.Kinjo AR, Nishikawa K. Eigenvalue analysis of amino acid substitution matrices reveals a sharp transition of the mode of sequence conservation in proteins. Bioinformatics. 2004;20(16):2504–8.CrossRefPubMedGoogle Scholar

## Copyright information

**Open Access**This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.