679
Views
0
CrossRef citations to date
0
Altmetric
Research Paper

Identification of intelligence-related proteins through a robust two-layer predictor

, , ORCID Icon, , & ORCID Icon
Pages 253-264 | Received 13 Oct 2022, Accepted 31 Oct 2022, Published online: 15 Nov 2022

ABSTRACT

In this study, we advance a robust methodology for identifying specific intelligence-related proteins across phyla. Our approach exploits a support vector machine-based classifier capable of predicting intelligence-related proteins based on a pool of meaningful protein features. For the sake of illustration of our proposed general method, we develop a novel computational two-layer predictor, Intell_Pred, to predict query sequences (proteins or transcripts) as intelligence-related or non-intelligence-related proteins or transcripts, subsequently classifying the former sequences into learning and memory-related classes. Based on a five-fold cross-validation and independent blind test, Intell_Pred obtained an average accuracy of 87.48 and 88.89, respectively. Our findings revealed that a score >0.75 (during prediction by Intell_Pred) is a well-grounded choice for predicting intelligence-related candidate proteins in most organisms across biological kingdoms. In particular, we assessed seismonastic movements and associate learning in plants and evaluated the proteins involved using Intell_Pred. Proteins related to seismonastic movement and associate learning showed high percentages of similarities with intelligence-related proteins. Our findings lead us to believe that Intell_Pred can help identify the intelligence-related proteins and their classes using a given protein/transcript sequence.

Introduction

Efforts at defining intelligence have long been attempted, however, there are still challenges in specifying a precise definition. To date, many definitions have been proposed for intelligence which involve a confusing mix of concepts [Citation1]. Intelligence may be defined in terms of speed of learning, the sum of one’s knowledge, and the ability to communicate through language. Thinking and abstract reasoning appear important for intelligence, and one may claim that creative problem solving is at the root of intelligence [Citation2]. However, according to another characterization, learning and memory capacity define intelligence. Accordingly, learning and memory are considered critical components of intelligence [Citation3]. Memory marks a certain experienced event to enable the organism to react faster and more appropriately in similar cases. Learning is tightly associated with memory since it is defined as the capability for more adaptive reactions [Citation4]. Memory and learning are the result of the modification of behavioral patterns and make it possible for an organism to sense, monitor, evaluate and make decisions. Therefore, detailed insight into memory and learning will be useful for studying the various aspects of intelligence.

Proteins are the major constituents of a cell and play vital roles in different parts of biological processes and behavior. Memory which forms the basis of learning and intelligence is among the biological processes that depend on covalent modifications of preexisting proteins (short-term memory) or require regulated gene transcription and translation (long-term memory) [Citation4–8]. Moreover, epigenetic processes resulting from gene expression and thus protein synthesis have been proposed to play a role in cognitive processes [Citation9]

The number of unannotated proteins in databases like UniProtKB/TrEMBL is large; a number that continues to rise daily. Annotation of such proteins not only requires experimental equipment, It is also quite laborious, expensive and time-consuming. Furthermore, in order to investigate intelligence from a holistic perspective, evaluation of a couple of proteins in different organisms cannot deliver the goods. Computational and high-throughput in silico approaches can prove helpful when it comes to studying proteins involved in different biological processes such as memory and learning, critical traits of intelligence [Citation10].

Several homologous detecting-based tools, such as BLAST and PSI-BLAST, are being customarily employed for predicting the potential function of unannotated proteins [Citation11–14]. But there are difficulties in using these methods in terms of time and memory and also in cases where low similarities between input and target sequences exist [Citation15,Citation16]. Therefore, many efforts have been dedicated to developing efficient computational approaches to address these problems. In this regard, machine-learning (ML) based methods have been successfully developed, as alternative computational approaches for the prediction of the potential function of proteins, using various features of proteins as input [Citation11,Citation17–21]. To the best of our knowledge, there is still no implementation of ML methods for identifying intelligence-related proteins. Therefore, it would be highly interesting to develop an algorithm that can directly discriminate between intelligence-related proteins and other protein classes.

Support-vector-machine (SVM) is a supervised learning model and a powerful algorithm in pattern recognition and data classification [Citation22]. The basic idea of SVM for binary classification is firstly to transform the non-linearly separable training data into a higher dimensional feature space by a nonlinear mapping, which is realized by defining proper kernel function (different kernel functions can be used). Then, it constructs a separating hyper-plane as the decision surface in such a way that the margin of separation between two classes is maximized [Citation22]. SVM is one of the widely used ML-based algorithms, based on supervision learning model, which has been successfully applied in various trajectories, including prediction of fertility-related proteins [Citation21,Citation23], lipid metabolism-related proteins [Citation11], protein-protein interactions [Citation24–26], drug discovery and prediction of potentially druggable proteins [Citation20,Citation27], discriminating toxic peptides from nontoxic peptides and proteins [Citation28,Citation29] and prediction of protein structure [Citation30,Citation31]. SVM is more precise than the other ML methods [Citation32] and having the ability to recognize subtle patterns in a variety of complex datasets [Citation33]. Moreover, high accuracy, as well as power of high dimensional data handling, makes SVM the most promising classifier in different realms of science [Citation11,Citation34,Citation35]. Inspired by the wide application of ML methods, in this study a novel predictor, Intell_Pred, was constructed using SVM for identifying proteins related to intelligence. Hence, the main objective of this study was to develop this software, which can discriminate between intelligence-related proteins (in two classes including memory and learning) and non-related intelligence using various features extracted from the already annotated intelligence-related proteins. Based on a series of recent studies [Citation36–38] and our previous studies [Citation11,Citation17,Citation21] Chou’s 5-step rules [Citation39] were applied to develop the software, which make all the processes more understandable and transparent. For the sake of illustration, we make use of the developed software to identify and evaluate specific intelligence-related proteins in plants. Our methodology may nonetheless be applied more broadly.

Material and methods

In order to develop Intell_Pred, Chou’s five-step rules [Citation40] were applied as follows: 1) Creation of an appropriate benchmark dataset for testing and training of the prediction, 2) Usage of an effective formulation for representing the samples that can truly reflect their intrinsic correlation with the target to be predicted, 3) Introduction or development of a powerful algorithm to implement the prediction, 4) Proper interpretation of cross-validation tests to evaluate prediction accuracy; and 5) Establishment of a user-friendly or public available tool for the developed predictor. These steps are presented in .

Figure 1. Pipeline of developing SVM-based models for Intell_Pred software.

Figure 1. Pipeline of developing SVM-based models for Intell_Pred software.

Benchmark datasets

According to previous studies [Citation17,Citation39,Citation41,Citation42] for developing an efficient statistical predictor, it is of crucial importance to construct a high-quality benchmark dataset containing training and testing datasets. To do so, our previous method was applied to construct the datasets [Citation17,Citation21]. Briefly, through searching the reviewed proteins in UniProtKB database with gene ontology terms “learning”, “memory” and “learning or memory”, protein sequences were collected and were considered as positive datasets related to learning, memory and intelligent classes, respectively. To remove the homologous sequences with sequence identity above 50% and avoid overestimation of the prediction accuracy of the developed models, Uniref50 database was used. Then, only protein sequences with the length <6000 or >60 amino acids were kept. Finally, protein sequences containing non-canonical amino acids such as B, X, and Z were removed. By doing so, a total of 598, 302, 296, proteins were obtained for “intelligence”, “learning” and “memory” classes, respectively.

To create the negative datasets, the UniProtKB database was depleted by a comprehensive search of all keywords implying intelligence functionality (including “not learning”, “not behaviour”, “not memory”, “not habituation” and “not cognition” as well as “not signaling”) and the same filtering steps that were used for positive datasets. In the next step, all domains of the proteins in both datasets were extracted through Pfam database and were compared. Proteins in the negative dataset were removed if they had a common domain with the proteins in positive datasets. Reliability of this method has been reported in previous studies [Citation17,Citation21] and ensures that a negative dataset is constructed by non-intelligence-related proteins exclusively. Eventually, a number of 338,139 non-intelligence-related protein sequences were obtained as negative dataset.

Models trained on very skewing datasets (unbalanced datasets) lead to the production of biased results toward the class with larger number of sequences (negative datasets) [Citation43]. In our case, the positive and negative datasets were extremely unbalanced and the negative dataset was substantially greater than that of positive ones. For this reason, and with an eye to minimizing the bias caused by the skewed datasets and handle the class unbalance problem of the positive and negative datasets, a random sampling solution approach was applied. To this end, we created a balanced benchmark dataset by selecting the protein samples randomly from the negative dataset at the same size of the positive datasets (without replacement). On the other hand, by this method, we cannot cover all of the diversity in the negative dataset. In other words, a random sample of the negative dataset may not be enough to assess the generalized predictive ability of the trained model. To overcome the issue several different negative samples were combined with a positive dataset. Previous studies confirmed the reliability of this approach to consider all the diversity in the negative dataset [Citation17,Citation21]. To do this, five samples were randomly obtained from the negative dataset (at the same number of the positive dataset and without replacement). In the next step, the positive dataset was combined with each of the negative samples, separately. In result, five benchmark datasets were created with the same positive proteins and different negative protein sequences. This approach led to the construction of 15 datasets related to intelligence, memory and learning classes, each with five datasets. More details about this approach can be found in [Citation17,Citation21]. Finally, 20% of each dataset was randomly selected as an independent dataset, while the other 80% was used as the training dataset for training SVM models. For readers’ convenience, protein accession numbers of these datasets are provided in Supplementary File S1.

Extraction of protein features

Protein feature extraction is a crucial step of concept learning process in which using mathematical model amino acid sequences are represented into a fixed feature vector. Therefore, in addition to a reliable and stringent benchmark dataset, the perfect formulation of the protein sequences is necessary for the development of a useful statistical predictor. This step enables us to find informative physicochemical features from the primary protein sequences, taking into account both the protein constituents and amino acids relative positions, which can be effectively used in the classification step. In general, protein features can be divided into two categories including structure-based features and sequence-based features. Given the difficulty of solving an experimental protein structure, there are significantly more protein sequences in Uniprot than known structures in the Protein Data Bank (PDB) [Citation44]. Therefore, sequence-based features are more suitable for analyzing the proteins without known structures [Citation45]. In this context, pseudo amino acid composition (PseAAC or Chou’s PseAAC) has been proposed as an effective approach for representing and distinguishing protein sequences of different functional profiles [Citation11,Citation17,Citation21,Citation41,Citation46,Citation47]. In this line, according to the concept of general PseAAC, a python programmed tool, iLearn [Citation48], has been developed to calculate these structural and physicochemical features of proteins and peptides. Here, this software was applied to convert each protein sequence into a vector of 12,669 dimensions. These features were belonged to 15 different features descriptors (). The efficiency of these features for developing the high-quality ML predictors is demonstrated in our previous studies [Citation11,Citation17,Citation21]. Furthermore, inspired by the successes of applying these features to develop the tools for predicting the protein sequences, they have been wildly used in most of the computational proteomics studies [Citation49–53] and many fields of genome analysis [Citation54]. Detailed description of the protein features is provided in Supplementary File S2.

Table 1. Summary of 12,669 extracted protein features that were used in SVM-based model development. More details are provided in Supplementary File S2.

Support vector machine (SVM)

As the third step of the Chou’s five-step rules, an algorithm is required to operate the training and predicting. Being intensively and successfully used in many different areas of bioinformatics [Citation55] and computational biology [Citation11,Citation17,Citation21,Citation56–59], SVM is a supervised learning model and a powerful algorithm in pattern recognition and data classification [Citation22]. The basic idea of SVM for binary classification is firstly to transform the non-linearly separable training data into a higher dimensional feature space by a nonlinear mapping, which is realized by defining proper kernel function (different kernel functions can be used). Then, it constructs a separating hyper-plane as the decision surface in such a way that the margin of separation between two classes is maximized. The more detailed information on SVM has been elaborated in [Citation22]. Moreover, for a brief formulation of SVM and how it works, see the [Citation60,Citation61].

In the present study, the positive and negative training datasets were inputted into LIBSVM package (version 3.22) for building a predictive model. The predictive efficiency of SVM models mainly depends on a proper setting of kernel function that maps the input vectors to a high-dimensional feature space [Citation77]. Here, the often-used radial basis function (RBF) was applied as kernel function to obtain the best classification hyper-plane, owing to its wide and successful applications in most of the exiting studies related to ML methods [Citation11,Citation21,Citation78–80]. Furthermore, the two important uncertain parameters in RBF kernel, the regularization parameter C and the kernel width parameter γ, were optimized via an unbiased parameter selection using the grid search approach. This method can guarantee the unbiased results by selecting SVM parameters in a cross-validating manner (five-fold cross validation) [Citation21]. Here to efficiently develop an SVM model to not only able to predict whether a protein is intelligence-related but also able to predict its class (memory or learning), a two-layer classification framework was constructed. First, intelligence dataset (including memory and learning datasets) was used to train SVM model for the first layer. The trained model in the first layer generally is served to predict a query protein sequence as intelligence or non-intelligence related proteins. In the second layer, the SVM models were trained with memory and learning training datasets, separately as binary predictor. A protein is applied as input to the second layer, if it be recognized as intelligence-related protein in the previous layer. The second layer determines the potential of that protein to be related to memory or learning classes. The final decision of a query sequence which is intelligence-related and determination of its class was made according to the probability obtained by SVM. The software considers a probability score >0.5 to designate putative intelligence-related proteins and their classes (memory or learning).

Model performance evaluation

We need to answer two questions for measuring the prediction quality of the developed models. The first one is: What indices should be used to measure the model’s quality? And the other one is: What test method should be used to score them? The questions are answered as follows, respectively.

The performance of the developed models evaluated by five indices, including accuracy (ability of the predictor to differentiate intelligence from non-intelligence), sensitivity (ability to correctly identify intelligence) and specificity (ability to correctly identify non-intelligence) and Matthew’s correlation coefficient (MCC, a correlation coefficient between the observed and predicted binary classifications). Since the conventional equations of these measures are not quite intuitive, the proposed formulae by Chen et al [Citation81] (based on general formulations of these metrics) are given below:

(1.1) Sn=1N+N+0 Sn 1(1.1)
(1.2) Sp=1N+N0 Sp 1(1.2)
(1.3) Acc=1N++N+N++ N0 Acc 1(1.3)
(1.4) MCC= 1 N+N++N+N 1+ N++N+N+ 1+ N++N+N1 MCC 1(1.4)

In the above formulas, N+ and N denote the total number of the positive and negative protein sequences investigated, whereas N+ and N+ is the number of the intelligence/non-intelligence proteins incorrectly predicted to be non-intelligence /intelligence, respectively.

To examine the ability of the predictive model in identifying intelligence protein sequences, five-fold cross-validation was carried out (as test method) on training datasets. To do so, the training datasets were divided into five equally sized subsets. In the next step, each of the subsets was singled out one by one and tested by the SVM trained with the remaining four subsets. These processes were repeated five times since each subset was used as testing set once. Finally, the five validation results were combined to generate a single accuracy. In addition, independent datasets (or testing datasets, which is definitely blind to the training dataset) were applied to allay the concerns of over-fitting the performance of the constructed model to the training dataset.

Software development

To enhance the practical usability and publicly available software for practical applications, a two-layer classifier software called Intell_Pred (intelligence prediction) has been provided freely at https://github.com/mrb20045/Intell_Pred. The software allows the users to predict intelligence-related proteins along with their potential classes (memory or learning). The software can be used by a wide variety of researchers with limited knowledge of the SVM computing environment, which needs simply upload sequence(s) in FASTA format for prediction. A key point of the Intell_Pred is accepting mRNA transcripts as input, in addition to amino acids sequences. Intell_Pred benefits from TransDecoder tool (version 3.0.1, http://transdecoder.sourceforge.net) to identify candidate open reading frames (ORFs) within mRNA transcripts. Then, the predicted amino acid sequences by TransDecoder, automatically input to the SVM models for predicting their potential as intelligence-related proteins as well as their classes. It is merit to note that the software only needs a FASTA file (DNA/protein) as input and performs all the processes automatically. Finally, it will return the name of the inputted sequence(s), predicted results along with probability scores (0≤ score ≥1) for every sequence.

Results and Discussions

With the explosive growth of biological sequences in the post-genomic era, a growing demand has raised for efficient methods to annotate these sequences. Previous studies declared that forming memory and learning relies on cross talk between genetic and epigenetic variations [Citation82]. Learning involves a series of molecular changes including regulation of gene transcription. The basis of the long-term memory form, when a number of genetic modifications are maintained in post-mitotic cells throughout life [Citation83]. Considering proteins as molecular fossils and the final products of gene regulation [Citation15], in-detail study of proteins involved in intelligent behaviors can light the way to discover the mechanism of forming memory and learning and finally to investigate intelligence in individuals. In this study, for the first time, an SVM-based two-layer classifier, Intell_Pred, designed for predicting the potential of unannotated proteins, which most likely to be suspected in intelligence.

Model development

Here, the benchmark datasets were used with 50% identity based on the previous studies, which noted that datasets with higher identity or high homologous sequences increase overestimation risk of prediction [Citation21]. All accession numbers of the used proteins are represented in Supplementary File S1 (a positive dataset combined with five different negative datasets in each class). To maximize the estimated overall accuracy by SVM, penalty parameter C and γ (width or number of attributes) of RBF were adjusted and computed by grid search strategy based on five-fold cross validation test. As a result, the penalty parameter C was set at 10 for all datasets and the width γ ranged from 0.001 to 0.003 for different datasets. The four metrics achieved by the proposed classifier via the five-fold cross-validation and independent datasets testing along with optimized γ values for each benchmark datasets are given in .

Table 2. Accuracy metrics of train and test datasets for intelligence model. The optimum γ parameter value of kernel function of SVM was chosen using a grid-search technique based on five-fold cross-validation.

Table 3. Accuracy metrics of train and test datasets for memory model. The optimum γ parameter value of kernel function of SVM was chosen using a grid-search technique based on five-fold cross-validation.

Table 4. Accuracy metrics of train and test datasets for learning model. The optimum γ parameter value of kernel function of SVM was chosen using a grid-search technique based on five-fold cross-validation.

SVM model achieved an average prediction accuracy of 89.09% at the first layer, which determines the protein can be related to intelligence or not (intelligence as general class) (). The model maintained its predictive properties (with accuracy of 87.71%), when independent datasets were applied, which shows that Intell_Pred is highly promising. The same situation was observed for the models in the second layer (learning and memory models), which determine the potential of the passed protein from the first layer to be related to memory or learning classes (). From these results, it can be found that sensitivity, specificity and accuracy of different datasets are very close, meaning that there was no bias in classification in all models. In other words, there is an equal chance of identifying the intelligence and non-intelligence-related proteins, correctly. To further emphasize the effectiveness of the proposed models, roughly 20% of each dataset was kept in each class for independent evaluation of the final models (testing datasets). Similar results were obtained based on the independent blind test, as the positive datasets could be discriminated from negative datasets with acceptable accuracy metrics in all classes. This suggests a promising capability and robustness of the proposed method for identifying the intelligence-related proteins and their classes ().

Evaluation of Intell_Pred with new positive and negative proteins

To the best of our knowledge, Intell_Pred is the first predictor ever developed for intelligence and non-intelligence proteins classification. Hence, there is not any possibility to compare its performance with its counterparts for exactly the same purpose. But our method achieved a high accuracy in comparison with those reported in most of the previous studies [Citation11,Citation21,Citation49,Citation51,Citation78–81] in other areas. However, to further challenge the proposed method for predicting intelligence-related proteins, new negative and positive datasets were constructed and evaluated using Intell_Pred. To this end, 10 negative datasets (each dataset contains 1000 protein sequences) were constructed through selecting randomly from negative datasets, as none of these proteins were appeared during training or testing the SVM models. For new positive dataset, UniProtKB database was searched (since after Intell_Pred had been developed) and new annotated proteins as intelligence-related ones were found through comparing with our primary positive dataset. In total, 85 and 84 new proteins were found, which were belonged to memory and learning classes, respectively. Results of evaluation of the new negative and positive datasets using Intell_Pred are displayed in , respectively. Also, complete results of evaluating the new negative and positive proteins using Intell_Pred are provided in Supplementary Files S3 and S4, respectively.

Table 5. Summary results of evaluation of the 10,000 new negative protein sequences by Intell_Pred.

Table 6. Summary results of evaluation of the new annotated protein sequences related to intelligence in Uniprot database by Intell_Pred (169 protein sequences were evaluated including 85 and 84 proteins belong to memory and learning classes, respectively) .

Interesting results were obtained after evaluation of the new negative and positive protein sequences. Out of 10,000 negative protein sequences, only 1,058 (10.58%) sequences were predicted as intelligence related. But, if score >0.90 considered, only 347 (3.47%) sequences were predicted as intelligence related (). Out of 169 positive protein sequences 163 (96.45%) sequences were predicted as intelligence related. Of these, 137 (81.07%) sequences were predicted as intelligence related when score >0.90 was considered (). These findings indicated that larger scores are more reliable, but it may cause to miss some real intelligence related proteins, if score >0.90 be considered. Hence, score >0.75 can be suggested as a reliable choice to balance the sensitivity and specificity for discriminating intelligence and non-intelligence related sequences.

Plant intelligence and candidate proteins

In this study, a molecular approach was adopted to figure out to which extent seismonastic responses to mechanical stimulation relate to memory and learning. In particular, proteins of Arabidopsis thaliana involved in processes related to seismonastic movement mechanisms e.g., proteins involved in pulvinus activation, turgor pressure and leaf movement; cytochalasin b-related protein, the motor cells-related proteins, charge amplification-related proteins, osmotic pressure-related proteins [Citation84] were extracted from the Uniprot database and evaluated using Intell_Pred. As our results reveal, 84.61, 80.76 and 77.07% of the proteins related to seismonastic movement mechanisms were predicted to be involved in intelligence, and learning and memory, respectively (, Supplementary File S5).

Table 7. Summary results of evaluation of the plant candidate proteins using Intell_Pred.

In like vein, proteins related to phototropins were extracted from the Uniprot database, as it was reported that associate learning of the sort that pea plants appear to exhibit [although see Citation85, Citation86, and Citation87, for exchanges on contradictory evidence for the existence of associative learning in plants] [Citation85–88] is related to phototropin function [Citation89]. In total, 18 proteins were found and were evaluated using Intell_Pred. Interestingly, 17 proteins (94.44%) were predicted as intelligence-related proteins, as 16 proteins showed score >0.75. Most of these proteins had score >0.75 in the memory class too (). The results indicated that tropistic responses triggered by an air flow mediated by the function of phototropin, as observed in Gagliano et al.’s pea plant learning experiments, is an intelligent behavior (72). The results also indicate that intelligence can be the result of memory and learning (4). Complete results of evaluating of these proteins using Intell_Pred are provided in Supplementary Files S5.

Our findings are consistent with the working hypothesis that plants are able to learn and develop memory. Learning, memory and intelligent behavior may not require brain cephalization or otherwise neural network of the sort found in animals. The way forward to sift through the complexity of plant intelligence studies is the integration of the information from several areas within the plant sciences, including molecular and cellular studies of signal transduction, plant physiology, ecology, bioinformatics and molecular biology [Citation90].

Conclusion

In this study a two-layer SVM-based classifier (Intell_Pred), developed to identify intelligence-related proteins and their classes (memory or learning), that uses complete structural feature of proteins. Different accuracy metrics demonstrated that Intell_Pred can effectively distinguish between intelligence-related and non-intelligence related proteins. Considering that proteins, as molecular fossils, are the final products of gene regulation, identification of the specific intelligence-related proteins may provide a clue for paving the way to unraveling the plant sort of intelligence. In this way, studying in detail the proteins involved in intelligent behavior can ease the discovery of the mechanism that underlie the formation of memory and learning.

From a computational point of view, it is noteworthy that although the prediction of intelligence-related proteins achieves acceptable results, there is room for improvement, as the number of protein sequences for intelligence and their classes is effectively growing on a daily basis. We envisage that Intell_Pred will be a powerful tool for detecting novel intelligence-related proteins in the future.

Author contributions

Conceptualization: A.Sh, MR.B, MS.VS. Methodology: MR.B, MS.VS, A.Sh. Validation: MR.B, MS.VS. Formal analysis: MR.B, MS.VS. Investigation: A.Sh., MS.V. Data Curation: MR.B. Writing original draft: Ash., MS.VS. Writing, Review & Editing: AT, PC, SA. Project administration: MR.B.

Supplemental material

Supplemental Material

Download Zip (2.6 MB)

Disclosure statement

No potential conflict of interest was reported by the author(s).

Supplementary material

Supplemental data for this article can be accessed online at https://doi.org/10.1080/19420889.2022.2143101

Additional information

Funding

“Research supported by the Office of Naval Research Global (N62909-19-1-2015) to PC.”

References

  • Legg S, Hutter M. A collection of definitions of intelligence. Front Artif Intell Appl. 2007a;157:17.
  • Legg S, Hutter M. Universal intelligence: a definition of machine intelligence. Minds Machines. 2007b;17:391–444.
  • Trewavas A. Green plants as intelligent organisms. Trends Plant Sci. 2005;10:413–419.
  • Witzany G. Memory and Learning as Key Competences of Living Organisms. Memory and Learning in Plants. Springer; 2018. p. 1–16.
  • Byrne JH, Hawkins RD. Nonassociative learning in invertebrates. Cold Spring Harb Perspect Biol. 2015;7:a021675.
  • de Vargas LDS, Sevenster D, Lima KR, et al. Novelty exposure hinders aversive memory generalization and depends on hippocampal protein synthesis. Behav Brain Res. 2019;359:89–94.
  • Lamprecht R, Farb CR, LeDoux JE. Fear memory formation involves p190 RhoGAP and ROCK proteins through a GRB2-mediated complex. Neuron. 2002;36:727–738.
  • Zhang YY, Liu MY, Liu Z, et al. GPR30‐mediated estrogenic regulation of actin polymerization and spatial memory involves SRC‐1 and PI3K‐mTORC2 in the hippocampus of female mice. CNS Neurosci Ther. 2019;25:714–733.
  • Day JJ, Sweatt JD. Cognitive neuroepigenetics: a role for epigenetic mechanisms in learning and memory. Neurobiol Learn Mem. 2011;96:2–12.
  • Calvo P, Baluška F, Trewavas A. Integrated information as a possible basis for plant consciousness. Biochem Biophys Res Commun. 2020.
  • Bakhtiarizadeh MR, Moradi-Shahrbabak M, Ebrahimi M, et al. Neural network and SVM classifiers accurately predict lipid binding proteins, irrespective of sequence homology. J Theor Biol. 2014;356:213–222.
  • Oldfield CJ, Uversky VN, Kurgan L. Predicting functions of disordered proteins with MoRFpred Computational methods in protein evolution. Springer; 2019. p. 337–352.
  • Qin Y, Zheng X, Wang J, et al. Prediction of protein structural class based on Linear Predictive Coding of PSI-BLAST profiles. Open Life Sci. 2015;10.
  • Yao Y, Li M, Xu H, et al. Protein Subcellular localization prediction based on PSI-BLAST profile and principal component analysis. Curr Proteomics. 2019;16:402–414.
  • Xiong J. Essential bioinformatics. Cambridge University Press; 2006.
  • Yang Y, Zheng H, Wang C, et al. Predicting apoptosis protein subcellular locations based on the protein overlapping property matrix and tri-gram encoding. Int J Mol Sci. 2019;20:2344.
  • Bakhtiarizadeh MR, Rahimi M, Mohammadi-Sangcheshmeh A, et al. PrESOgenesis: a two-layer multi-label predictor for identifying fertility-related proteins using support vector machine and pseudo amino acid composition approach. Sci Rep. 2018b;8:9025.
  • Garg A, Raghava GP. A machine learning based method for the prediction of secretory proteins using amino acid composition, their order and similarity-search. In: silico biology. Vol. 8. 2008. p. 129–140.
  • Gromiha MM, Ahmad S, Suwa M, 2008. Neural network based prediction of protein structure and Function: comparison with other machine learning methods. 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence). IEEE, pp. 1739–1744.
  • Jamali AA, Ferdousi R, Razzaghi S, et al. DrugMiner: comparative analysis of machine learning algorithms for prediction of potential druggable proteins. Drug Discov Today. 2016;21:718–724.
  • Rahimi M, Bakhtiarizadeh MR, Mohammadi-Sangcheshmeh A. OOgenesis_Pred: a sequence-based method for predicting oogenesis proteins by six different modes of Chou’s pseudo amino acid composition. J Theor Biol. 2017;414:128–136.
  • Cortes C, Vapnik V. Support-vector networks. Mach Learn. 1995;20:273–297.
  • Bakhtiarizadeh MR, Rahimi M, Mohammadi-Sangcheshmeh A, et al. PrESOgenesis: a two-layer multi-label predictor for identifying fertility-related proteins using support vector machine and pseudo amino acid composition approach. Sci Rep. 2018a;8:1–12.
  • Cui G, Fang C, Han K. Prediction of protein-protein interactions between viruses and human by an SVM model. In: BMC Bioinformatics. Vol. 13. Springer; 2012. p. S5.
  • Romero‐Molina S, Ruiz‐Blanco YB, Harms M, et al. PPI‐Detect: a support vector machine model for sequence‐based prediction of protein–protein interactions. J Comput Chem. 2019;40:1233–1242.
  • You Z, Ming Z, Niu B, et al., 2013. A SVM-based system for predicting protein-protein interactions using a novel representation of protein sequences. International Conference on Intelligent Computing. Springer, pp. 629–637.
  • Houssein EH, Hosney ME, Oliva D, et al. A novel hybrid Harris hawks optimization and support vector machines for drug design and discovery. Comput Chem Eng. 2020;133:106656.
  • Gupta S, Kapoor P, Chaudhary K, et al. In silico approach for predicting toxicity of peptides and proteins. PloS one. 2013;8:e73957.
  • Islam SA, Sajed T, Kearney CM, et al. PredSTP: a highly accurate SVM based model to predict sequential cystine stabilized peptides. BMC Bioinformatics. 2015;16:210.
  • Gubbi J, Lai DT, Palaniswami M, et al. Protein secondary structure prediction using support vector machines and a new feature representation. Int J Comput Intell Appl. 2006;6:551–567.
  • Patel M, Shah H, 2013. Protein secondary structure prediction using support vector machines (SVMs). 2013 International Conference on Machine Intelligence and Research Advancement, Katra, India. IEEE, pp. 594–598.
  • Meng C, Jin S, Wang L, et al. AOPs-SVM: a sequence-based classifier of antioxidant proteins using a support vector machine. Front Bioeng Biotechnol. 2019;7.
  • Aruna S, Rajagopalan S. A novel SVM based CSSFFS feature selection algorithm for detecting breast cancer. Int J Comput Appl. 2011;31.
  • Cinelli M, Sun Y, Best K, et al. Feature selection using a one dimensional naïve Bayes’ classifier increases the accuracy of support vector machine classification of CDR3 repertoires. Bioinformatics. 2017;33:951–955.
  • Guo H, Liu B, Cai D, et al. Predicting protein–protein interaction sites using modified support vector machine. Int J Mach Learn Cybern. 2018;9:393–398.
  • Chen W, Ding H, Feng P, et al. iACP: a sequence-based tool for identifying anticancer peptides. Oncotarget. 2016;7:16895.
  • Liu B, Liu F, Wang X, et al. Pse-in-One: a web server for generating various modes of pseudo components of DNA, RNA, and protein sequences. Nucleic Acids Res. 2015a;43:W65–W71.
  • Xiao X, Wang P, Lin W-Z, et al. iAMP-2L: a two-level multi-label classifier for identifying antimicrobial peptides and their functional types. Anal Biochem. 2013;436:168–177.
  • Chou K-C. Some remarks on protein attribute prediction and pseudo amino acid composition. J Theor Biol. 2011;273:236–247.
  • Chou K-C. Proposing 5-steps rule is a notable milestone for studying molecular biology. Nat Sci. 2020;12:74.
  • Cheng X, Xiao X, Chou K-C. pLoc-mPlant: predict subcellular localization of multi-location plant proteins by incorporating the optimal GO information into general PseAAC. Mol Biosyst. 2017;13:1722–1727.
  • Liu B, Yang F, Chou K-C. 2L-piRNA: a two-layer ensemble classifier for identifying piwi-interacting RNAs and their function. Mol Ther Nucleic Acids. 2017;7:267–277.
  • Eitrich T, Lang B. Efficient optimization of support vector machine learning parameters for unbalanced datasets. J Comput Appl Math. 2006;196:425–436.
  • Berman HM, Westbrook J, Feng Z, et al. The protein data bank. Nucleic Acids Res. 2000;28:235–242.
  • Zhang J, Liu B. A Review on the Recent Developments of Sequence-based Protein Feature Extraction Methods. Curr Bioinf. 2019;14:190–199.
  • Ahmad J, Hayat M. MFSC: multi-voting based feature selection for classification of Golgi proteins by adopting the general form of Chou’s PseAAC components. J Theor Biol. 2019;463:99–109.
  • Butt AH, Rasool N, Khan YD. Prediction of antioxidant proteins by incorporating statistical moments based features into Chou’s PseAAC. J Theor Biol. 2019;473:1–8.
  • Chen Z, Zhao P, Li F, et al., 2019. iLearn: an integrated platform and meta-learner for feature engineering, machine-learning analysis and modeling of DNA, RNA and protein sequence data. Briefings in bioinformatics.
  • Ahmad K, Waris M, Hayat M. Prediction of protein submitochondrial locations by incorporating dipeptide composition into Chou’s general pseudo amino acid composition. J Membr Biol. 2016;249:293–304.
  • Chou K-C, Shen H-B. Recent advances in developing web-servers for predicting protein attributes. Nat Sci. 2009;1:63.
  • Dehzangi A, Heffernan R, Sharma A, et al. Gram-positive and Gram-negative protein subcellular localization by incorporating evolutionary-based descriptors into Chou׳ s general PseAAC. J Theor Biol. 2015;364:284–294.
  • Kabir M, Hayat M. iRSpot-GAEnsC: identifing recombination spots via ensemble classifier and extending the concept of Chou’s PseAAC to formulate DNA samples. Mol Genet Genomics. 2016;291:285–296.
  • Tang H, Chen W, Lin H. Identification of immunoglobulins using Chou’s pseudo amino acid composition with feature selection technique. Mol Biosyst. 2016;12:1269–1275.
  • Chen W, Lin H, Chou K-C. Pseudo nucleotide composition or PseKNC: an effective formulation for analyzing genomic sequences. Mol Biosyst. 2015;11:2620–2634.
  • Liu B, Fang L, Wang S, et al. Identification of microRNA precursor with the degenerate K-tuple or Kmer strategy. J Theor Biol. 2015b;385:153–159.
  • Cai C, Wang W, Sun L, et al. Protein function classification via support vector machine approach. Math Biosci. 2003a;185:111–122.
  • Lin H, Chen W, Ding H. AcalPred: a sequence-based tool for discriminating between acidic and alkaline enzymes. PloS one. 2013;8:e75726.
  • Mondal S, Pai PP. Chou׳ s pseudo amino acid composition improves sequence-based antifreeze protein prediction. J Theor Biol. 2014;356:30–35.
  • Zhu -P-P, Li W-C, Zhong Z-J, et al. Predicting the subcellular localization of mycobacterial proteins by incorporating the optimal tripeptides into the general form of pseudo amino acid composition. Mol Biosyst. 2015;11:558–563.
  • Cai Y-D, Zhou G-P, Chou K-C. Support vector machines for predicting membrane protein types by using functional domain composition. Biophys J. 2003c;84:3257–3263.
  • Chou K-C, Cai Y-D. Using functional domain composition and support vector machines for prediction of protein subcellular location. J Biol Chem. 2002;277:45765–45769.
  • Bhasin M, Raghava GP. Classification of nuclear receptors based on amino acid composition and dipeptide composition. J Biol Chem. 2004;279:23262–23266.
  • Chen K, Jiang Y, Du L, et al. Prediction of integral membrane protein type by collocated hydrophobic amino acid pairs. J Comput Chem. 2009;30:163–172.
  • Lee T-Y, Lin Z-Q, Hsieh S-J, et al. Exploiting maximal dependence decomposition to identify conserved motifs from a group of aligned signal sequences. Bioinformatics. 2011;27:1780–1787.
  • Chen Z, Zhao P, Li F, et al. iFeature: a python package and web server for features extraction and selection from protein and peptide sequences. Bioinformatics. 2018;34:2499–2502.
  • Feng Z-P, Zhang C-T. Prediction of membrane protein types based on the hydrophobic index of amino acids. J Protein Chem. 2000;19:269–275.
  • Horne DS. Prediction of protein helix content from an autocorrelation analysis of sequence hydrophobicities. Biopolymers. 1988;27:451–477.
  • Sokal RR, Thomson BA. Population structure inferred by local spatial autocorrelation: an example from an Amerindian tribal population. Am J Phys Anthropol. 2006;129:121–131.
  • Cai C, Han L, Ji Z, et al. Enzyme family classification by support vector machines. Proteins Struct Funct Bioinf. 2004;55:66–76.
  • Cai C, Han L, Ji ZL, et al. SVM-Prot: web-based support vector machine software for functional classification of a protein from its primary sequence. Nucleic Acids Res. 2003b;31:3692–3697.
  • Dubchak I, Muchnik I, Holbrook SR, et al., 1995. Prediction of protein folding class using global description of amino acid sequence. Proceedings of the National Academy of Sciences 92, 8700–8704.
  • Dubchak I, Muchnik I, Mayor C, et al. Recognition of a protein fold in the context of the SCOP classification. Proteins Struct Funct Bioinf. 1999;35:401–407.
  • Shen J, Zhang J, Luo X, et al., 2007. Predicting protein–protein interactions based only on sequences information. Proceedings of the National Academy of Sciences 104, 4337–4341.
  • Chou K-C. Prediction of protein cellular attributes using pseudo-amino acid composition. Proteins: structure. Funct Bioinf. 2001a;43:246–255.
  • Chou K-C. Using amphiphilic pseudo amino acid composition to predict enzyme subfamily classes. Bioinformatics. 2004;21:10–19.
  • Chou K-C. Using subsite coupling to predict signal peptides. Protein Eng Des Sel. 2001b;14:75–79.
  • Lin S-W, Lee Z-J, Chen S-C, et al. Parameter determination of support vector machine and feature selection using simulated annealing approach. Appl Soft Comput. 2008;8:1505–1512.
  • Lata S, Sharma B, Raghava G. Analysis and prediction of antibacterial peptides. BMC Bioinformatics. 2007;8:263.
  • Ng XY, Rosdi BA, Shahrudin S. Prediction of antimicrobial peptides based on sequence alignment and support vector machine-pairwise algorithm utilizing LZ-complexity. Biomed Res Int. 2015;2015.
  • Thakur N, Qureshi A, Kumar M. AVPpred: collection and prediction of highly effective antiviral peptides. Nucleic Acids Res. 2012;40:W199–W204.
  • Chen W, Feng P-M, Lin H, et al. iRSpot-PseDNC: identify recombination spots with pseudo dinucleotide composition. Nucleic Acids Res. 2013;41:e68–e68.
  • Lauria M, Rossi V. Origin of Epigenetic Variation in Plants: relationship with Genetic Variation and Potential Contribution to Plant Memory. Memory and Learning in Plants. Springer; 2018. p. 111–130.
  • Zovkic IB, Guzman-Karlsson MC, Sweatt JD. Epigenetic regulation of memory formation and maintenance. Learn Memory. 2013;20:61–74.
  • Volkov AG, Foster JC, Baker KD, et al. Mechanical and electrical anisotropy in Mimosa pudica pulvini. Plant Signal Behav. 2010;5:1211–1221.
  • Markel K. Lack of evidence for associative learning in pea plants. Elife. 2020a;9:e57614.
  • Markel K. Response to comment on’Lack of evidence for associative learning in pea plants’. ELife. 2020b;9:e61689.
  • Gagliano M, Vyazovskiy VV, Borbély AA, et al. Comment on’Lack of evidence for associative learning in pea plants’. Elife. 2020;9:e61141.
  • Bhandawat A, Jayaswall K, Sharma H, et al. Sound as a stimulus in associative learning for heat stress in Arabidopsis. Commun Integr Biol. 2020;13:1–5.
  • Gagliano M, Vyazovskiy VV, Borbély AA, et al. Learning by association in plants. Sci Rep. 2016;6:38427.
  • Calvo P. The philosophy of plant neurobiology: a manifesto. Synthese. 2016;193:1323–1343.