Skip to main content
  • Methodology article
  • Open access
  • Published:

Variable selection from a feature representing protein sequences: a case of classification on bacterial type IV secreted effectors

Abstract

Background

Classification of certain proteins with specific functions is momentous for biological research. Encoding approaches of protein sequences for feature extraction play an important role in protein classification. Many computational methods (namely classifiers) are used for classification on protein sequences according to various encoding approaches. Commonly, protein sequences keep certain labels corresponding to different categories of biological functions (e.g., bacterial type IV secreted effectors or not), which makes protein prediction a fantasy. As to protein prediction, a kernel set of protein sequences keeping certain labels certified by biological experiments should be existent in advance. However, it has been hardly ever seen in prevailing researches. Therefore, unsupervised learning rather than supervised learning (e.g. classification) should be considered. As to protein classification, various classifiers may help to evaluate the effectiveness of different encoding approaches. Besides, variable selection from an encoded feature representing protein sequences is an important issue that also needs to be considered.

Results

Focusing on the latter problem, we propose a new method for variable selection from an encoded feature representing protein sequences. Taking a benchmark dataset containing 1947 protein sequences as a case, experiments are made to identify bacterial type IV secreted effectors (T4SE) from protein sequences, which are composed of 399 T4SE and 1548 non-T4SE. Comparable and quantified results are obtained only using certain components of the encoded feature, i.e., position-specific scoring matix, and that indicates the effectiveness of our method.

Conclusions

Certain variables other than an encoded feature they belong to do work for discrimination between different types of proteins. In addition, ensemble classifiers with an automatic assignment of different base classifiers do achieve a better classification result.

Background

Feature extraction from protein sequences plays an important role in protein classification [1,2,3,4] of many areas, such as identification of plant pentatricopeptide repeat coding protein [5], prediction of bacterial type IV secreted effectors [6,7,8,9], identification of heat shock protein [10], prediction of mitochondrial proteins [11], etc. In general, prevailing encoding approaches of protein sequences for feature extraction include pseudo-amino acid composition (PseAAC) [10,11,12,13,14,15,16,17,18,19,20,21,22], position-specific scoring matrix (PSSM) [7, 23,24,25,26,27,28,29,30,31,32], position-specific iterated blast (PSI-BLAST) [33,34,35,36,37] etc.

However, several problems do still exist and are listed as follows. First of all, it needs to be decided which encoding approach is more effective. In fact, this problem can be solved according to the results of classification on each encoding approach using various computational methods (i.e., known as classifiers in the field of machine learning or pattern recognition). In other words, the encoding approach corresponding to the most accurate classification result should be considered. Prevailing classifiers including random forest or decision tree classifier (RF or DTC) [1, 38], gradient boosting machine (GBM) [39, 40], k-nearest-neighbor (kNN) [41, 42], linear discriminant analysis (LDA) [43, 44], logistic regression (LR) [45], multi-layer perceptron (MLP) [46, 47], naive bayesian (NB) [5, 48], support vector machine (SVM) [49, 50] are credible.

Secondly, it needs to be discussed whether protein classification is predictive or not, which is a little confused. Actually, classification labels have commonly been assigned to protein sequences in advance. If these labels are definitive, i.e., having been certified by biological experiments in advance, there won’t be any need to predict the category of a protein sequence again. Conversely, unsupervised learning (e.g. clustering) rather than supervised learning (e.g. classification) should be considered, since these labels are undetermined. And that corresponds to protein prediction. However, prevailing methods are always confusing protein classification an protein prediction.

Thirdly, the extracted feature using an encoding approach is considered to be entirely effective. In fact, there may be only parts of the extracted feature that are effective. However, this phenomenon has been subjectively neglected. As a result, it is an important issue how to select certain components or variables from a feature that really helps to recognize proteins with specific functions. In other words, variable selection from a feature representing protein sequences is a new problem probably not yet avoidable, which may be more helpful to classification of different protein sequences.

In this paper, we propose a new method for variable selection from an encoded feature. The selection of a feature from an encoding approach is excluded from our method. Besides, no prediction work is executed. Focusing on components or variables of an encoded feature, we implement our method at seven steps as shown in Fig. 1. First of all, samples are divided in balance, which constitute a training and testing group. Secondly, a base classifier is automatically assigned in every resampling of the training group. Thirdly, the score of each variable in an encoded feature is accumulated through r rounds of resampling, training and scoring. Fourthly, a scatter plot and corresponding order of variables with their accumulated scores in a descending order are obtained. Fifthly, r rounds of training are made on resampling samples to achieve ensemble classifiers in each dimension (i.e., from one to full dimension of the encoded feature) according to variables incrementally added in the descending order. Sixthly, variable selection is accomplished using a line chart derived from the classification results on the testing group. Seventhly, evaluation metrics are made to estimate the effectiveness of selected variables. Experiments are made on the benchmark dataset [51] to identify bacterial type IV secreted effectors from protein sequences, which indicates the effectiveness of our method. More details can be seen in the following parts of this paper.

Fig. 1
figure 1

A framework of variable selection for identifying different proteins

Results

In this section, we take a benchmark dataset [51] as a case to evaluate the performance of our proposed method. The dataset is composed of 1947 protein sequences across multiple bacterial species, categorized into two groups, i.e., 399 type IV secreted effectors (T4SE) as the positive samples and 1548 non-T4SE as the negative samples. The 1947 protein sequences are randomly divided into two subsets for training and testing, respectively. The training set consists of 973 sequences, among which 199 T4SE and 774 non-T4SE sequences are randomly selected from positive and negative samples, respectively. The left 200 T4SE and 774 non-T4SE samples constitute the testing set. Besides, we choose PSSM, which is composed of 400 variables, as the encoded feature. Following the procedure shown in Fig. 1, the experimental results of score accumulation, ensemble classification, variable selection and the corresponding classification results are listed as follows.

Results of score accumulation

We randomly extract 70% of samples from the training set and choose a classifier with the lowest classification error rate as the base classifier at each round. Meanwhile, scores representing the importance of variables are calculated. After 1000 rounds of resampling, training and scoring (i.e., \(r=1000\)), we obtain the accumulated scores of each variable from PSSM. The corresponding scatter plot is shown in Fig. 2a. Its horizontal and vertical coordinates correspond to 400 variables and their accumulated scores, respectively. In addition, the frequency of each selected base classifier is illustrated in Fig. 2b.

Fig. 2
figure 2

Results of score accumulation

Fig. 3
figure 3

A line chart of ACCs and AUCs corresponding to the incrementally added variables from feature PSSM with their accumulated scores in a descending order

It can be seen in Fig. 2a that the accumulated scores are all relatively low. Since the accumulated scores of the 400 variables have no apparent distinction, all these variables are considered to be enumerated at the following step instead of selecting variables with high accumulated scores, as having been stated in [52].

In Fig. 2b, it can be seen that MLP is automatically assigned as the base classifier for 74.7% of 1000 round resampling. On the contrary, DTC and SVM have never been selected for score accumulation.

Results of ensemble classification on testing group

The ensemble classifiers have been built using 1000 rounds of resampling and training on the training set in each dimension, with 400 variables incrementally added in the descending order according to their accumulated scores. Then, the 400 ensemble classifiers, each of which keeps 1000 base classifiers, are applied to the testing set. As a result, a line chart (see Fig. 3) is obtained with its horizontal corresponding to the dimensions with variables incrementally added in the descending order according to their accumulated scores. The vertical coordinates are referred to the Acc and AUC values in the incremental dimensions. It can be seen in Fig. 3 that it is the first 25 variables with descending accumulated scores that form a 25-D feature from PSSM for effective identification of T4SE.

Table 1 Quantitative results by incrementally adding the variables after the 25th one of 400 variables in a descending order according to their accumulated scores and the comparison results shown in reference [6]

Moreover, some quantitative results are shown to indicate the effectiveness of the obtained 25 variables for classification. Table 1 lists the max, mean and min values of Accs and AUCs by incrementally adding the variables after the 25th one of the 400 variables in a descending order according to their accumulated scores. It can be indicated that features enlarged with higher dimensions can achieve only similar Accs and AUCs as the 25 variables do.

Table 2 Quantitative results among ensemble classification with automatic assignment of base classifiers and the ensemble classifier with single base classifiers

Classification results of the selected variables

In order to show the effectiveness of the selected 25 variables, the confusion matrix, Precision, Recall and \(F1-measure\) are calculated in order to make a quantitative comparison. In addition, ROCs together with AUCs are listed as qualitative results.

Results of ROC and AUC between the first 25 selected variables and all 400 components of PSSM using ensemble classification are shown in Fig. 4. The similar ROC curves and AUC values indicate that the selected 25 variables keep a comparable classification capability with PSSM.

Fig. 4
figure 4

Comparison results of ROC and AUC between all 400 components and the selected 25 variables using ensemble classification

Fig. 5
figure 5

Results of ROC and AUC using different base classifiers with the selected 25 variables

Besides, results of ROC and AUC using the ensemble classifier consisting of 1000 single base classifiers with the selected 25 variables are illustrated in Fig. 5. By making a careful comparison between Figs. 4b and 5, it can be seen that ensemble classification with automatic assignment of base classifier keeps a better ROC curve and AUC value (i.e., 0.9287).

Moreover, quantitative results among ensemble classification with automatic assignment of a base classifier and the ensemble classifier with a single base classifier are listed in Table 2. It can be seen that the ensemble classifier with automatic assignment of a base classifier on the 25 selected variables keeps a high TP (i.e., 157) compared with most of the other classification strategies. Besides, it has better values of Precision, Recall and \(F1-measure\) (i.e., 0.904, 0.903 and 0.903) compared with the other ensemble classifier with a single base classifier on the 25 selected variables. As to the results of the ensemble classifier with automatic assignment of a base classifier on PSSM (i.e., 400 components), the selected 25 variables can also achieve comparable results.

Discussions

Experimental results have indicated the effectiveness of variable selection from the encoded feature PSSM. In this section, we will further discuss the special composition of our variable selection method and the classification results.

Purpose of using base classifier selection

The automatic assignment of a base classifier is creative in this paper. Giving consideration to the sample distribution of resampling, we designed the strategy of automatic assignment. Due to the limited sample size, resampling is only an approximation to the population. In our previous work, it has been pointed out that different base classifiers should be considered according to various sample distributions [52]. However, the base classifier was interactively appointed in [52]. The pan chart in Fig. 2b can also show this phenomenon, which indicates that base classifiers selected at a higher percentage may fit the population better. Besides, quantitative results listed in Table 2 indicate the power of using automatic assignment compared with the interactive appointment of a base classifier.

Purpose of using a line chart for variable selection

In fact, a line chart shown in Fig. 3 is a semi-automatic way for variable selection. Here, it goes against the interactive way that uses a manual selection within a table or on a 2-D scatter plot [52]. Also, it abandons the automatic way of automatic clustering [53] on accumulated scores. Due to the limited distribution of the accumulated scores with relatively low values, the variables have no apparent distinction. It probably means variables are highly correlated. In that case, variable orders instead of values are to be considered.

Comparisons between classification results

We compared the classification results of our method with PredT4SE-Stack [6]. As shown in Table 1, the max, mean and min values of Accs and AUCs on the testing set exhibit a better result than the classification results on its training set using PreT4SE-Stack. The best classification Acc value using PredT4SE-Stack on its training set is 0.911. Most of the other meta-classifier got a lower Acc value and AUC value than the mean Acc and AUC obtained using our method. However, classification results of Pred4SE-Stack on its testing set are better than those using our method. In fact, the classification results on its testing set are even better than those on its training set. By careful observation, it is found that parameters of base classifiers are manually set every time in [6]. That’s why PredT4SE-Stack gets better classification results on its testing set. Anyway, seeking better classification results by setting parameter values doesn’t make any sense for variable selection.

Conclusion

In order to solve the problem of variable selection from an encoded feature representing protein sequences, we propose a variable selection method based on ensemble classification with an automatic assignment of base classifiers. Variable ordering is obtained according to score accumulation on training samples. Then, ensemble classifiers are established from one to all dimensions of the encoded feature according to variables incrementally added in the descending order. Using the ensemble classifiers on testing samples, a line chart is drawn for variable selection. Ultimately, evaluation metrics are made to estimate the effectiveness of selected variables. Taking a dataset containing protein sequences categorized into T4SE and non-T4SE group as a case, the performance of the proposed method is evaluated.

Methods

In this section, we will expound our method in detail. As illustrated in Fig. 1, our method has seven steps, each of which is framed in a dashed box. Each step which keeps its name labeled within the dashed box, corresponds to a following subsection.

Sample division

We make a balanced sample division at the first step. That is, samples derived from a category are divided into two equivalent parts. As a result, half of the samples from different categories form a training group; while, the left ones constitute a testing group. It is noteworthy that this sample division should be performed in a completely random way.

Base classifier selection

A base classifier is automatically assigned at the second step. We make a set of base classifiers including decision tree classifier (DTC), gradient boosting machine (GBM), k-nearest-neighbor (kNN), linear discriminant analysis (LDA), logistic regression (LR), multi-layer perceptron (MLP), naive bayesian (NB) and support vector machine (SVM). Each one is equally assigned in an automatic way corresponding to every round of resampling and training module. In each round j, 70% of training samples are randomly selected in a full dimension of an encoded feature for training each base classifier. The remaining 30% of training samples are regarded as the out-of-bag samples for calculation of the classification error rate \(Err_{j}\), as is expressed in Eq. (1). The base classifier with the lowest classification error rate is automatically assigned in round j.

Score accumulation

Score accumulation is made at the third step. Once a base classifier is automatically assigned according to the classification error rate calculated on the out-of-bag samples in round j, permutations are to be made. As to each variable i of the encoded feature, only one-time permutation of the expression values from the out-of-bag samples is performed. The corresponding classification error rate is denoted as \(Err^{0}_{j}(i)\). Accordingly, a score representing the importance of variable i is expressed as \(score_{j}(i) = Err^{0}_{j}(i)-Err_{j}\). After r rounds of resampling, training and scoring, the accumulated score of variable i is expressed as \(\sum _{j=1}^{r}score_{j}(i)/r\).

Variable ordering

Variables are reordered at the fourth step. A 2-D scatter plot is to be made with its horizontal and vertical coordinates corresponding to the variable indices and the accumulated scores, respectively. Besides, variables are to be sorted in a descending order according to the accumulated scores. If the accumulated scores of the variables have no distinction (i.e., the accumulated scores are all relatively low), all the variables rather than the significant variables selected using previously proposed clustering method [53] are to be enumerated at the following step.

Ensemble classification

Ensemble classifiers are established at the fifth step. Again, r rounds of resampling and training are performed to achieve ensemble classifiers in each dimension according to variables incrementally added in their descending order. As to 1-D space, the variable with the highest accumulated score is considered. At each round of resampling, the base classifier with the lowest classification error rate is trained. Altogether, r base classifiers are selected as the ensemble classifier in 1-D space. This procedure is repeated with a variable keeping a lower accumulated score added, until the full dimension of the encoded feature or the full dimension of significant variables has been considered.

Variable selection

Variable selection is accomplished at the sixth step. In each dimension, the established ensemble classifier is applied to the testing samples. The accuracy (Acc) expressed in Eq. (2) and the area under curve (AUC) of the receiver operating characteristic (ROC) are calculated. Accordingly, a line chart is obtained with its horizontal and vertical coordinates corresponding to the variable indices in their descending order and the corresponding Accs and AUCs in different dimensions. A dimension threshold can be made when Accs and AUCs are keeping almost the same with dimension incrementally increasing. Thus, the variables that really help to recognize proteins with specific functions are selected from the encoded feature.

Measure

Evaluation metrics are made to estimate the effectiveness of selected variables at the seventh step. The classification error rate is expressed as follows,

$$\begin{aligned} Err={{FN+FP} \over {TP+FN+TN+FP}}, \end{aligned}$$
(1)

where TP, TN, FP and FN represent the number of true positive, true negative, false positive and false negative, respectively. On the contrary, Acc is shown as follows,

$$\begin{aligned} Acc={{TN+TP} \over {TP+FN+TN+FP}}. \end{aligned}$$
(2)

Besides, we choose four widely used quantitative measurements. The confusion matrix illustrates TP, TN, FP and FN together. Besides, Precision and Recall are computed as follows,

$$\begin{aligned} Precision= & {} {{TP} \over {TP+FP}}, \end{aligned}$$
(3)
$$\begin{aligned} Recall= & {} {{TP} \over {TP+FN}}. \end{aligned}$$
(4)

In addition, \(F1-measure\) is a harmonic average of Precision and Recall, which is expressed as

$$\begin{aligned} F1-measure = {{2*Precision*Recall} \over {Precision+Recall}}. \end{aligned}$$
(5)

Moreover, the ROC and AUC are also provided here as qualitative measurements.

Availability of data and materials

The public dataset analysed during the current study is available in reference [51], and can be downloaded from the website https://github.com/LoopGan/Effective-prediction-of-bacterial-type-IV-secreted-effectors.

Abbreviations

Acc::

Accuracy;

AUC::

Area under curve;

DTC::

Decision tree classifier;

GBM::

Gradient boosting machine;

kNN::

k-nearest-neighbor;

LDA::

Linear discriminant analysis;

LR::

Logistic regression;

MLP::

Multi-layer perceptron;

NB::

Naive bayesian;

PseAAC::

Pseudo-amino acid composition;

PSI-BLAST::

Position-specific iterated blast;

PSSM::

Position-specific scoring matrix;

RF::

Random forest;

ROC::

Receiver operating characteristic;

SVM::

Support vector machine;

T4SE::

Type IV secreted effectors

References

  1. Lv ZB, Jin SS, Ding H, Zou Q. A random forest sub-Golgi protein classifier optimized via dipeptide and amino acid composition features. Fronti Bioeng Biotechnol. 2019;7:215.

    Article  Google Scholar 

  2. Zhu XJ, Feng CQ, Lai HY, Chen W, Lin H. Predicting protein structural classes for low-similarity sequences by evaluating different features. Knowl-Based Syst. 2019;163:787–93.

    Article  Google Scholar 

  3. Ru XQ, Li LH, Zou Q. Incorporating distance-based top-n-gram and random forest to identify electron transport proteins. J Proteome Res. 2019;18:2931–9.

    Article  CAS  Google Scholar 

  4. Li YJ, Niu MT, Zou Q. ELM-MHC: an improved MHC identification method with extreme learning machine algorithm. J Proteome Res. 2019;18:1392–401.

    Article  CAS  Google Scholar 

  5. Qu K, Wei L, Yu J, Wang C. Identifying plant pentatricopeptide repeat coding gene/protein using mixed feature extraction methods. Front Plant Sci. 2019;9:1–10.

    Article  Google Scholar 

  6. Xiong Y, Wang QK, Yang JC, Zhu XL, Wei DQ. PredT4SE-Stack: prediction of bacterial type IV secreted effectors from protein sequences using a stacked ensemble method. Front Microbiol. 2018;9:2571.

    Article  Google Scholar 

  7. Zou LY, Nan CH, Hu FQ. Accurate prediction of bacterial type IV secreted effectors using amino acid composition and PSSM profiles. Bioinformatics. 2013;29(24):3135–42.

    Article  CAS  Google Scholar 

  8. Ashari ZE, Dasgupta N, Brayton KA, Broschat SL. An optimal set of features for predicting type IV secretion system effector proteins for a subset of species based on a multi-level feature selection approach. PLoS ONE. 2018;13:e0197041.

    Article  CAS  Google Scholar 

  9. Yu LZ, Guo YZ, Li YZ, Li GB, Li ML, Luo JS, Xiong WJ, Qin WL. SecretP: identifying bacterial secreted proteins by fusing new features into Chou’s pseudo-amino acid composition. J Theor Biol. 2010;267:1–6.

    Article  CAS  Google Scholar 

  10. Feng PM, Chen W, Lin H, Chou KC. iHSP-PseRAAAC: identifying the heat shock protein families using pseudo reduced amino acid alphabet composition. Anal Biochem. 2013;442(1):118–25.

    Article  CAS  Google Scholar 

  11. Mirza MT, Khan A, Tahir M, Lee YS. MitProt-Pred: predicting mitochondrial proteins of Plasmodium falciparum parasite using diverse physiochemical properties and ensemble classification. Comput Biol Med. 2013;43(10):1502–11.

    Article  CAS  Google Scholar 

  12. Ahmad J, Hayat M. MFSC: multi-voting based feature selection for classification of Golgi proteins by adopting the general form of Chou’s PseAAC components. J Theor Biol. 2019;463:99–109.

    Article  CAS  Google Scholar 

  13. Zhang SL, Duan X. Prediction of protein subcellular localization with oversampling approach and Chou’s general PseAAC. J Theor Biol. 2018;437:239–50.

    Article  CAS  Google Scholar 

  14. Srivastava A, Kumar R, Kumar M. BlaPred: predicting and classifying beta-lactamase using a 3-tier prediction system via Chou’s general PseAAC. J Theor Biol. 2018;457:29–36.

    Article  CAS  Google Scholar 

  15. Sankari ES, Manimegalai D. Predicting membrane protein types by incorporating a novel feature set into Chou’s general PseAAC. J Theor Biol. 2018;455:319–28.

    Article  CAS  Google Scholar 

  16. Sankari ES, Manimegalai D. Predicting membrane protein types using various decision tree classifiers based on various modes of general PseAAC for imbalanced datasets. J Theor Biol. 2017;435:208–17.

    Article  CAS  Google Scholar 

  17. Liang YY, Zhang SL. Predict protein structural class by incorporating two different modes of evolutionary information into Chou’s general pseudo amino acid composition. J Mol Graph Model. 2017;78:110–7.

    Article  CAS  Google Scholar 

  18. Meher PK, Sahu TK, Banchariya A, Rao AR. DIRProt: a computational approach for discriminating insecticide resistant proteins from non-resistant proteins. BMC Bioinform. 2017;18:190.

    Article  CAS  Google Scholar 

  19. Tiwari AK. Prediction of G-protein coupled receptors and their subfamilies by incorporating various sequence features into Chou’s general PseAAC. Comput Methods Programs Biomed. 2016;134:197–213.

    Article  Google Scholar 

  20. Han GS, Yu ZG, Anh V. A two-stage SVM method to predict membrane protein types by incorporating amino acid classifications and physicochemical properties into a general form of Chou’s PseAAC. J Theor Biol. 2014;344:31–9.

    Article  CAS  Google Scholar 

  21. Chou K. Using amphiphilic pseudo amino acid composition to predict enzyme subfamily classes. Bioinformatics. 2005;21:10–9.

    Article  CAS  Google Scholar 

  22. Chou K. Prediction of protein cellular attrbutes using pseudo-amino acid composition. Proteins. 2001;43:246–55.

    Article  CAS  Google Scholar 

  23. Wang JW, Yang BJ, Revote J, Leier A, Marquez-Lago TT, Webb G, Song JN, Chou KC, Lithgow T. POSSUM: a bioinformatics toolkit for generating numerical sequence feature descriptors based on PSSM profiles. Bioinformatics. 2017;33(17):2756–8.

    Article  CAS  Google Scholar 

  24. Zhang LC, Zhao XQ, Kong L. Predict protein structural class for low-similarity sequences by evolutionary difference information into the general form of Chou’s pseudo amino acid composition. J Theor Biol. 2014;355:105–10.

    Article  CAS  Google Scholar 

  25. Paliwal KK, Sharma A, Lyons J, Dehzangi A. A tri-gram based feature extraction technique using linear probabilities of position specific scoring matrix for protein fold recognition. IEEE Trans Nanobiosci. 2014;13(1):44–50.

    Article  Google Scholar 

  26. Zahiri J, Yaghoubi O, Mohammad-Noori M, Ebrahimpour R, Masoudi-Nejad A. PPIevo: protein–protein interaction prediction from PSSM based evolutionary information. Genomics. 2013;102(4):237–42.

    Article  CAS  Google Scholar 

  27. Zhang SL, Ye F, Yuan XG. Using principal component analysis and support vector machine to predict protein structural class for low-similarity sequences via PSSM. J Biomol Struct Dyn. 2012;29(6):634–42.

    Article  CAS  Google Scholar 

  28. Jeong JC, Lin XT, Chen XW. On position-specific scoring matrix for protein function prediction. IEEE-ACM Trans Comput Biol Bioinform. 2011;8(2):308–15.

    Article  Google Scholar 

  29. Jia CZ, Liu T, Chang AK, Zhai YY. Prediction of mitochondrial proteins of malaria parasite using bi-profile Bayes feature extraction. Biochimie. 2011;93(4):778–82.

    Article  CAS  Google Scholar 

  30. Dong QW, Zhou SG, Guan JH. A new taxonomy-based protein fold recognition approach based on autocross-covariance transformation. Bioinformatics. 2009;25(20):2655–62.

    Article  CAS  Google Scholar 

  31. Cheng CW, Su ECY, Hwang JK, Sung TY, Hsu WL. Predicting RNA-binding sites of proteins using support vector machines and evolutionary information. BMC Bioinform. 2008;9(S12):S6.

    Article  CAS  Google Scholar 

  32. Chou KC, Shen HB. MemType-2L: a web server for predicting membrane proteins and their types by incorporating evolution information through Pse-PSSM. Biochem Biophys Res Commun. 2007;360(2):339–45.

    Article  CAS  Google Scholar 

  33. An JY, You ZH, Chen X, Huang DS, Li ZW, Liu G, Wang Y. Identification of self-interacting proteins by exploring evolutionary information embedded in PSI-BLAST-constructed position specific scoring matrix. Oncotarget. 2016;7(50):82440–9.

    Article  Google Scholar 

  34. Qin YF, Zheng XQ, Wang J, Chen M, Zhou CJ. Prediction of protein structural class based on Linear Predictive Coding of PSI-BLAST profiles. Open Life Sciences. 2015;10(1):529–36.

    Article  CAS  Google Scholar 

  35. Ding SY, Li Y, Shi ZX, Yan SJ. A protein structural classes prediction method based on predicted secondary structure and PSI-BLAST profile. Biochimie. 2014;97:60–5.

    Article  CAS  Google Scholar 

  36. Liu T, Zheng XQ, Wang J. Prediction of protein structural class for low-similarity sequences using support vector machine and PSI-BLAST profile. Biochimie. 2010;92(10):1330–4.

    Article  CAS  Google Scholar 

  37. Kaur H, Raghava GPS. Prediction of alpha-turns in proteins using PSI-BLAST profiles and secondary structure information. Proteins-Struct Funct Bioinform. 2004;55(1):83–90.

    Article  CAS  Google Scholar 

  38. Breiman L. Random forests. Mach Learn. 2001;45:5–32.

    Article  Google Scholar 

  39. Tan CG, Wang T, Yang WY, Deng L. PredPSD: a gradient tree boosting approach for single-stranded and double-stranded DNA binding protein prediction. Molecules. 2020;25(1):98.

    Article  CAS  Google Scholar 

  40. Friedman JH. Greedy function approximation: a gradient boosting machine. Ann Stat. 2001;29(5):1189–232.

    Article  Google Scholar 

  41. He ZY, Liu H, Moch H, Simon H. Machine learning with autophagy-related proteins for discriminating renal cell carcinoma subtypes. Sci Rep. 2020;10(1):720.

    Article  CAS  Google Scholar 

  42. Cover T, Hart P. Nearest neighbor pattern classification. IEEE Trans Inf Theory. 1967;13(1):21–7.

    Article  Google Scholar 

  43. Isopescu RD, Spulber R, Josceanu AM, Mihaiescu DE, Popa O. Romanian bee pollen classification and property modelling. J Apicult Res. 2020.

  44. Belhumeur PN, Hespanha JP, Kriegman DJ. Eigenfaces vs. fisherfaces: recognition using class specific linear projection. IEEE Trans Pattern Anal Mach Intell. 1997;19(7):711–20.

    Article  Google Scholar 

  45. Wachters JE, Kop E, Slagter-Menkema L, Mastik M, van der Wal JE, van der Vegt B. de Bock GH, van der Laan BFAM, Schuuring E. Distinct biomarker profiles and clinical characteristics in T1–T2 glottic and supraglottic carcinomas. The Laryngoscope 2020.

  46. Zhou Y, Li GQ, Li HQ. Automatic cataract classification using deep neural network with discrete state transition. IEEE Trans Med Imaging. 2020;39(2):436–46.

    Article  Google Scholar 

  47. Pal SK, Mitra S. Multilayer perceptron, fuzzy sets, and classification. IEEE Trans Neural Netw. 1992;3(5):683–97.

    Article  CAS  Google Scholar 

  48. Domingos P, Pazzani M. On the optimality of the simple Bayesian classifier under zero-one loss. Mach Learn. 1997;29(2–3):103–30.

    Article  Google Scholar 

  49. Meng CL, Jin SS, Wang L, Guo F, Zou Q. AOPs-SVM: a sequence-based classifier of antioxidant proteins using a support vector machine. Front Bioeng Biotechnol 2019.

  50. Cortes C, Vapnik VN. Support vector networks. Mach Learn. 1995;20(3):273–97.

    Google Scholar 

  51. Wang Y, Guo Y, Pu X, Li M. Effective prediction of bacterial type IV secreted effectors by combined features of both C-termini and N-termini. J Comput Aided Mol Des. 2017;31:1029–38.

    Article  CAS  Google Scholar 

  52. Zhao XD, Jiao Q, Li HY, Wu YM, Wang HX, Huang S, Wang GH. ECFS-DEA: an ensemble classifier-based feature selection for differential expression analysis on expression profiles. BMC Bioinform. 2020;21:43.

    Article  Google Scholar 

  53. Liu T, Li HY, Zhao XD. Clustering by search in descending order and automatic find of density peaks. IEEE Access. 2019;7:133772–80.

    Article  Google Scholar 

Download references

Acknowledgements

This work is derived from Scientific Research Project Supported by Enterprise Suzhou Dachen Medical Technology Co., Ltd.

Funding

This work has been supported by the financial support of This work has been supported by the financial support of Natural Science Foundation of Heilongjiang Province (No. LH2020F002). The funding body of Fundamental Research Funds for Natural Science Foundation of Heilongjiang Province played an important role in the design of the study, collection, analysis and interpretation of data and in writing the manuscript.

Author information

Authors and Affiliations

Authors

Contributions

X.D.Z conceived the general research and supervised it. J.Z performed the research and were the principal developers. D.L.L and L.X.L analyzed the data. D.N.K, M.A.A.A and X.D.Z wrote and revised the manuscript. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Xudong Zhao.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhang, J., Lv, L., Lu, D. et al. Variable selection from a feature representing protein sequences: a case of classification on bacterial type IV secreted effectors. BMC Bioinformatics 21, 480 (2020). https://doi.org/10.1186/s12859-020-03826-6

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12859-020-03826-6

Keywords