 Research
 Open Access
 Published:
Comparison of five supervised feature selection algorithms leading to top features and gene signatures from multiomics data in cancer
BMC Bioinformatics volume 23, Article number: 153 (2022)
Abstract
Background
As many complex omics data have been generated during the last two decades, dimensionality reduction problem has been a challenging issue in better mining such data. The omics data typically consists of many features. Accordingly, many feature selection algorithms have been developed. The performance of those feature selection methods often varies by specific data, making the discovery and interpretation of results challenging.
Methods and results
In this study, we performed a comprehensive comparative study of five widely used supervised feature selection methods (mRMR, INMIFS, DFS, SVMRFECBR and VWMRmR) for multiomics datasets. Specifically, we used five representative datasets: gene expression (Exp), exon expression (ExpExon), DNA methylation (hMethyl27), copy number variation (Gistic2), and pathway activity dataset (Paradigm IPLs) from a multiomics study of acute myeloid leukemia (LAML) from The Cancer Genome Atlas (TCGA). The different feature subsets selected by the aforesaid five different feature selection algorithms are assessed using three evaluation criteria: (1) classification accuracy (Acc), (2) representation entropy (RE) and (3) redundancy rate (RR). Four different classifiers, viz., C4.5, NaiveBayes, KNN, and AdaBoost, were used to measure the classification accuary (Acc) for each selected feature subset. The VWMRmR algorithm obtains the best Acc for three datasets (ExpExon, hMethyl27 and Paradigm IPLs). The VWMRmR algorithm offers the best RR (obtained using normalized mutual information) for three datasets (Exp, Gistic2 and Paradigm IPLs), while it gives the best RR (obtained using Pearson correlation coefficient) for two datasets (Gistic2 and Paradigm IPLs). It also obtains the best RE for three datasets (Exp, Gistic2 and Paradigm IPLs). Overall, the VWMRmR algorithm yields best performance for all three evaluation criteria for majority of the datasets. In addition, we identified signature genes using supervised learning collected from the overlapped top feature set among five feature selection methods. We obtained a 7gene signature (ZMIZ1, ENG, FGFR1, PAWR, KRT17, MPO and LAT2) for EXP, a 9gene signature for ExpExon, a 7gene signature for hMethyl27, one singlegene signature (PIK3CG) for Gistic2 and a 3gene signature for Paradigm IPLs.
Conclusion
We performed a comprehensive comparison of the performance evaluation of five wellknown feature selection methods for mining features from various highdimensional datasets. We identified signature genes using supervised learning for the specific omic data for the disease. The study will help incorporate higher order dependencies among features.
Background
So far, pattern recognition techniques have huge impacts in solving most of the complicated reallife problems such as motif discovery from sequence features, detection of gene signatures for disease status, among others [1,2,3,4]. These methodologies are applied either separately or in conjunction with various softcomputing tools such as neural networks, evolutionary computing, swarm intelligence, etc. to solve various problems. Dealing with biological problems has revealed many challenges from real and complex data [5,6,7].
In general, a pattern recognition scheme involves three stages: data acquisition, dimensionality reduction, and classification or clustering [8,9,10]. In many pattern mining applications, the data acquisition stage yields huge number of features, some of which may be unimportant and redundant. Thus, the dimensionality of the data needs to be reduced by removing those unimportant and redundant features. The subset of selected features thus obtained keeps the optimal salient characteristics of the data as much as possible. In this way, dimensionality reduction accelerates the process of knowledge discovery to achieve better pattern recognition tasks such as classification and clustering [11,12,13], etc.
The dimensionality of data can generally be reduced using two methods: feature selection and feature extraction [8, 14]. The feature selection methodologies selects a small subset of features that are most relevant to the class variable. On the other hand, the feature extraction methodologies acquire a new subset of features by taking various functions on the original features. As the attribute selection methodologies retain the overall originality of attributes when decreasing the number of attributes, they are often preferred in several cases where the chosen attributes require to be interpreted by the domain experts [15]. There are many applications for feature selections such as disease status prediction, microRNA transcription start site (TSS) prediction [16, 17], etc. Analyzing microarray cancer data is one of the most promising applications of feature selection.
Feature selection can be accomplished in three ways, viz., supervised, semisupervised and unsupervised [18]. Supervised approaches evaluate each candidate feature or feature subset by knowing class labels, while the unsupervised approaches derive some alternative criteria without knowing any class labels [19, 20]. The methods of feature selection can also be classified into three categories: filter, wrapper, and embedded, depending upon how the feature evaluation indicators are computed [21]. Filterbased methods assess each candidate subset of features based on some internal data evaluation metrics. These approaches of filterbased feature selection can again be classified into two groups, namely, scalar feature selection and feature vector selection. In a scalar feature selection plan, individual features are assessed based on some statistical criteria and eventually some userspecific number of topranked features are chosen. The features chosen through these approaches are remarkably strong but are less robust because they do not consider the interdependencies between features while choosing the desired number of features. Furthermore, there is no previous information about the accurate number of features that are considered in leading to offer better classification accuracies.
For the purpose of feature selection, collective knowledge is suggested over many alternative criteria such as Pearson correlation coefficient to compute the value (usefulness) of a feature subset due to its capacity of analyzing nonlinear dependency. Among all these measures, mutual information has appeared as one of the most potential criteria to solve the task of feature section. For example, Mutual Information based Feature Selection (MIFS) [22], Mutual Information Feature Selector under Uniform information distribution (MIFSU) [23], minimal Redundancy Maximal Relevance criterionbased feature selection (mRMR) [24], Normalized Mutual Information based Feature Selection (NMIFS) [25], Improved Normalized Informationbased Feature Selection (INMIFS) [26], Variable Weighted Maximal Relevance minimal Redundancy criterionbased feature selection (VWMRmR) [27], are some wellknown algorithms of the mutual informationbased approaches. In this connection, a twostage feature selection method is also used where a reasonably small number of features are first selected from a high dimensional data using the mRMR criterion, and then a better feature subset of features is selected from that candidate sets using a wrapper approach. In contrast to the filterbased approaches, the wrapper approaches employ the classification performance to choose the best subset of the reduced features. The value (usefulness) of feature subsets depends on selected types of classification models are used to measure the classification performances. Besides these, there are several other feature algorithms such as Support Vector Machine Recursive Feature Elimination (SVMRFE) [28], Support Vector Machine Recursive Feature Elimination along incorporating the Correlation Bias Reduction (SVMRFECBR) [29], Discriminative Feature Selection (DFS) [30], etc. However, the pipeline of this study is provided in Fig. 1.
As already mentioned, the feature extraction methods transform all the features of a given data into a new transformed feature space from which the originality of the features cannot be found. On the other hand, the feature selection methods select a small subset of features, where the originality of the selected features remains the same as it was before doing the feature selection. Therefore, the feature selection methods are most suitable for generating gene signatures by selecting a smaller subset of genes, while the feature extraction methods cannot obtain gene signatures.
In this research work, we considered five different stateoftheart methods of feature selection for gene signature detection, such as mRMR, DFS, INMIFS, SVMRFECBR, and VWMRmR. We assessed the their performance on various omics datasets in terms of several evaluation criteria. The experimental results showed that the average classification accuracy rate of the selected feature subset obtained using the VWMRmR method was superior when compared with other feature selection algorithms. The experimental results further showed that the subset of features selected using the VWMRmR method is prominent in terms of two other criteria, namely, average redundancy rate and representation entropy among all methods. Finally, we applied these five feature selection methods to a multiomics dataset in acute myeloid leukemia, and performed the intersection operation among the sets of the selected top statistically significant genes (features) obtained by those methods to identify gene signatures from each omic data. Of note, MIFS (1994), MIFSU (2002) and NMIFS (2009) were fundamental baseline feature selection algorithms that were very old. The mRMR (2005) method is an improved version of MIFS (1994) and MIFSU (2002). On the other hand, another improved latest version of NMIFS called INMIFS (2010) was already developed. Thus, instead of considering the very oldest fundamental methods, we chose the improved version of them for our study.
Methods
Currently, dimensionality reduction techniques are becoming more important due to the advancement of bigdata (i.e., highdimensional datasets). Among various techniques, feature selection has become an emerging tool to identify the most suitable features in big data, both in terms of the number of samples or the number of features. Some of the widely used feature selection techniques are demonstrated below.
MIFS
Battiti (1994) proposed the Mutual InformationBased Feature Selection (MIFS) criterion, which is defined as follows.
The selection criterion is utilized to greedily pick the most notable m attributes from a collection of d attributes [22]. Provided a set of already selected features S, it is used to select a candidate feature \(f_i\) which makes the most of the relevance to the class I(fi; C) at each stage without taking into account the joint mutual information between the selected attribute set and the class variable C. The parameter \(\beta\) has a significant impact on the selected feature set. In the cases where \(\beta\) is too big, it will enlarge the redundancy in the selected features. As a result, the MIFS method may remove redundant features that have high relevance with the class variable. The main disadvantage of the MIFS method is the selection of an ideal value for the parameter \(\beta\).
MIFSU
Kwak and Choi developed an enhanced version of MIFS called as MIFSU, which is based on the assumption that information is handled uniformly with specific features S [23]. The criterion which is used to select the promising feature at each stage is defined as follows.
where \(H (f_s) = \sum _{f_s\epsilon s}{P\ (f_s)}\log {P\ (f_s)}\) is the entropy of \(f_s\). The uniform probability distribution assumption can ensure conditioning by the class C does not alter the proportion of the mutual information between \(f_s\) and \(f_i\), and the entropy of \(f_s\).
The MIFSU method gauges a superior assessment of the MI measure than the MIFS method, however, it additionally needs to pick the parameter \(\beta\) cautiously. With an inappropriate value of \(\beta\), the MIFSU method may provide poor results.
mRMR (minimum redundancy maximum relevance)
Peng et al. proposed a parameterfree feature selection method called minimal Redundancy Maximum Relevancebased feature selection algorithm (mRMR) [24]. The mRMR method replaces the \(\beta\) used in the selection criterion of the earlier methods MIFS and MIFSU by \(\frac{1}{S}\), where S is the size of the set S of already selected features. The selection criterion of the mRMR method is defined as follows.
The main advantage of the mRMR method over the MIFS, MIFSU methods is that it does not require any parameter like \(\beta\) used in their methods.
Feature selection, one of the fundamental issues in pattern recognition as well as machine learning, detects subgroups of data which are important to the parameters used here and is generally denoted as Maximum Relevance. These subsets involve the material that is pertinent but redundant and mRMR attempts to fix its target on this issue through providing those redundant subsets. mRMR has a different variation of forms in many parts alike speech recognition as well as cancer diagnosis. Features can be selected in different ways. One strategy is to select features which correlate strongest to the transportation variable. This has been termed as peakpurpose selection. Many heuristic approaches can be used, such as the sequential backward, forward, or floating selections. On the other scenario, features can be detected to be mutually that is far away from each other when still possessing parallel to the classvariable. This strategy, denoted as Minimum Redundancy Maximum Relevance (mRMR) has been found to be more sophisticated and efficient than the maximum relevance selection. In a specified case, the “correlation” can be updated by the statistical dependency between the variables. Mutual information can be used to estimate the dependency. In that scenario, it is highlighted that mRMR is basically an approximation to maximize the dependency between the joint handling of the chosen features and the distribution variable.
NMIFS (normalized mutual information feature selection)
Estevez et al. enhanced a new version of mRMR, termed as Normalized Mutual Information Feature Selection method (NMIFS). The NMIFS method used the normalized mutual information to compute the redundancy between the selected features. The selection criterion of the NMIFS method is defined as follows [25]:
where \({\hat{I}}\), the normalized mutual information, is demonstrated as
Using the normalized mutual information measure when computing the featurefeature redundancy, the NMIFS method seeks to overcome the problem of imbalance between the relevance and redundancy terms of the selection criteria used in the MIFS, MIFSU and mRMR techniques. However, the NMIFS method has some drawbacks. For binaryclass problems, the entropy of the class variable is one if the distributions of the class variable are the same. But, if the difference between the two terms is large, the entropy value will be less than one, resulting in a discrepancy between these two terms of the selection criterion. For multiclass problems, the value of the left term is usually more than one, while the value of the right hand term is below one. The imbalance between these two terms may sometime force the feature selection algorithm to select a feature based on the maximal value of the lefthand term while neglecting the righthand term. Thus, the NMIFS method may often fail to properly select the desired set of features due to the imbalance problem.
INMIFS (improved normalized mutual information feature selection)
Vinh et al. proposed an improved version of the NMIFS method called Improved Normalized Mutual Information Feature Selection (INMIFS) method [26]. They used normalized mutual information to compute both the classrelevance and featurefeature redundancy. The selection criterion of the INMIFS method is shown below.
where \({\hat{I}}\) denotes the normalized mutual information.
The main advantage of the INMIFS method over the NMIFS method is that it attempts to make both the left and righthand terms of the selection criterion comparable in magnitude. Vinh et al. (2010) obtained better results using the INMIFS method compared to the NMIFS method for six multiclass datasets.
Discriminative feature selection (DFS)
Hong Tao et al. developed a filterbased feature selection technique through combining Linear Discriminant Analysis (LDA) and sparsity regularization [30]. To perform the feature selection, they utilized row sparsity on the LDA transformation matrix by \(l_{2,1}\)norm regularisation, and the outcome formulation optimizes to choose the most discriminating features while simultaneously discarding the redundant ones.
They could use the \(l_{2,1}\)norm centred formulation to the \(l_{2,p}\)norm regularised scenarios, providing them more p options for fitting the wider range of sparsity necessity. They proposed an efficient algorithm to resolve the \(l_{2,p}\)norm minimization problem and highlighted that their proposed technique is able to reduce the target in a monotonic manner prior to convergence while \(0 < p \le 2\). In addition, their method maintains the power for selecting the most discriminative features while also discarding the redundant ones.
SVMRFECBR (support vector machine recursive feature elimination with correlation bias reduction)
Yan et al. proposed an improved version of SVMRFE called as Support Vector Machine Recursive Feature Elimination with Correlation Bias Reduction (SVMRFECBR) by incorporating the correlation bias reduction (CBR) strategy into the feature elimination process [29]. Like SVMRFE, the SVMRFECBR method adopts a backward elimination strategy. And so, it is able to model the dependencies between features. They showed that the SVMRFECBR performed better than SVMRFE and many other common methods. The main advantage of SVMRFECBR over SVMRFE is that the former can handle feature sets having highly correlated features, while the latter one cannot do it. They further suggested that the stability of SVMRFECBR could be improved by applying some ensemble methods.
VWMRmR (variable weighted maximal relevance minimal redundancy)
Bandyopadhyay et al. developed another normalized mutual informationbased feature selection algorithm called Variable Weighted Maximal Relevance minimal Redundancy criterion based feature selection technique (VWMRmR) [27]. Here, features are selected depending upon the weighted maximum relevance minimum redundancy strategy by which the relative weight of the redundancy is linearly increased with concern to the number of already selected features. The selection criterion of this technique is shown below:
where \(\acute{I}(f_i;f_s)\) is the normalized mutual information between two features: \(f_i\) and \(f_s\) (\(\in S\)), w denotes a userspecified positive, realvalued parameter which controls the relative weight of redundancy relative to the relevance, S is the set of already chosen features, and k is the number of the features that needs to be selected.
Comparison of different feature selection algorithms on multiomics data
Here, we used five stateoftheart feature selection algorithms, viz., mRMR, DFS, INMIFS, SVMRFECBR and VWMRmR to select a small subset of relevant features corresponding to multiomics datasets. For each dataset, the topranked 50 features are selected for each of the five feature selection algorithms.
Used datasets
For analysis, five datasets, namely, gene expression (Exp), exon expression (ExpExon), copy number variation (Gistic2), DNA methylation (hMethyl27) and Pathway activity dataset (Paradigm IPLs) from a multiomics dataset of acute myeloid leukemia (LAML) from The Cancer Genome Atlas (TCGA) were acquired from the publicly available repositories [31, 32]. Each of these five datasets has three types of class labels denoting three different stages of diseases, such as favorable (31 samples), intermediate/ normal (96 samples), and poor (34 samples). Each dataset consists of hundreds to thousands continuous features obtained through different biological experiments. There are a total of 161 common samples and among all these five data profiles.
Normalization and statistical test
We first started working with common set of genes and samples among all these five profiles. Each profile consisting of the common gene set was normalized genewise through the zeromean normalization [33, 34] that was formulated below:
where \(\mu\) refers to the mean of the scores of all samples corresponding to the specific gene before normalization and \(\sigma\) denotes the standard deviation of the scores of all samples corresponding to that gene before normalization; while \(x_{i}\) and \(x_{i}^{norm}\) signify the scores of ith sample of that corresponding gene prior to and after normalization, respectively.
We then performed statistical test using Limma within the R statistical package [33, 35], which employs Empirical Bayes test genewise on the respective normalized data. We obtained different sets of statistically significant genes having p value < 0.05 for each profile.
Discretization and computation of mutual information
As all attribute variables contain continuous types of values, each of them was preprocessed by converting into a Zscore (signifying zero mean and units). Each attribute variable is then was then discretized into three states, such as − 1, 0, and 1. If the value of the attribute for a given sample is less than the standard deviation of the attribute, it takes − 1; if the value of the attribute is greater than the standard deviation, it takes 1; and it takes 0 in all other cases. In this way, all the attributes are transformed into a discretized space from a continuous space. Then the mutual information between two features is easily computed in the discretized space.
Supervised feature selection
After discretization, we applied those five earlier mentioned supervised feature selection strategies (viz., mRMR, INMIFS, DFS, SVMRFECBR and VWMRmR) on each of the five data profiles, individually, and extracted 50 top significant features among all the statistically significant genes (features).
Used classifiers
For experimental purposes, we used four classifiers, viz., C4.5, NaïveBayes, KNN and AdaBoost to assess the quality of the reduced feature subsets selected (i.e., 50 top statistically significant features here) using different feature selection algorithms, and then estimated the different evaluation metrics.
Evaluation method
Weka [36] software is employed to make all four classifiers (C4.5, NaïveBayes, KNN and AdaBoost). The third classifier KNN rule has a userdefined parameter K whose value is determined as the rounded value of the square root of the number of objects in the training data. On the contrary, Naive Bayes considers that the features are normally distributed whose means and variances are determined from the data. The worth of these classifiers are assessed by reporting the average classification accuracies obtained by first applying tenfold cross validation to the training data, then by repeating the same process ten times and finally by averaging these ten results to obtain a single estimation.
Evaluation criteria
For each of the above four classifiers, the performance of the corresponding classification model related to a omics datasets is assessed in terms of three evaluation criteria, viz., Accuracy, Redundancy Rate and Representation Entropy. The first measure is a supervised measure, while the remaining two are unsupervised measures.
Accuracy
The accuracy (Acc) is computed as follows:
where TP, TN, FP and FN stands for the number of true positives, the number of true negatives, the number of false positives and the number of false negatives, respectively.
Redundancy rate
For a feature set F having d number of features, the redundancy rate of F (RR) is defined as follows.
where \(Sim(f_i,f_j)\) is the similarity between two features \(f_i, f_j \in F\). To compute the similarity, normalized mutual information, correlation coefficient, etc. can be used as the underlying measures.
Representation entropy
For a feature set F with d number of features, assume, the eigenvalues of the \(d*d\) covariance matrix of F be \(\uplambda _j\), \(j = 1 \ldots d\). The representation entropy of F, denoted by \(R\_Entropy\), is defined as follows.
where \(\widetilde{\uplambda _j}, j=1,\ldots ,s,\) are calculated as follows.
The \(R\_Entropy\) attains its minimum value (zero) when all the eigenvalues but one are equal to zero, that is, when it contains all its information in one coordinate direction. However, if all the eigenvalues are equal, then the information is evenly distributed among all the features; thus, the \(R\_Entropy\) will reach its maximum value. Thus, we have uncertainty in feature reduction. This measure is known as the representation entropy.
Results and discussion
Experimental results
This subsection first describes the overall performance analysis. Subsequently, the comparative performance of five different feature selection algorithms are summarized. Then, the comparative performance of different feature selection algorithms are assessed in terms of redundancy rate and representation entropy.
Differential analysis
Fundamental descriptions (#common samples, #classes, #statistically significant features) of each of these five profiles were described in Table 1.
Comparison of different feature selection algorithms in terms of classification accuracy
We have evaluated the performances of the five existing feature selection algorithms  mRMR, DFS, INMIFS, SVMRFECBR and VWmRMR, on the above mentioned five datasets, viz., Exp, ExpExon, Gistic2, hMethyl27 and Pathway activity. The average scores at the 10fold cross validation with 10 times repetitions obtained by applying different classifiers to the reduced feature subsets selected using the five different feature selection algorithms across LAML datasets are reported in Table 2.
Interestingly, during the utilization of classifiers, we had performed hyper parameter tuning for some possible classifiers estimated classification accuracy for each turning of the parameter. For example, for the KNN classifier, the number of neighbors (K) is tuned from 1 to 15 with step size 1, while for the C4.5 classifier, the confidence parameter (C) is tuned from 0.05 to 0.5 with step size 0.05, but for the Naive Bayes and Adaboost classifiers, there are no hyper parameters that need to be tuned. Finally, the best result among all sets of tuned parameter is considered for that specific classifier. For the first dataset (Exp), the SVMRFECBR algorithm performs better than all other algorithms for all classifiers except KNN. For this dataset, the mRMR algorithm produces the best accuracy (=89.19%) for the KNN classifier. For the second dataset (ExpExon), the VWmRMR algorithm performs better than all other feature selection algorithms for all classifiers except the Naive Bayes classifier. For this dataset, the VWmRMR method achieves the best accuracy (=87.83%) for the KNN classifier. For the third dataset (hMethyl27), the VWmRMR algorithm performs better than all other algorithms except SVMRFECBR for two classifiers, namely, Naive Bayes and AdaBoost. For this dataset, the SVMRFECBR obtains the best accuracy (=84.41%) for the KNN classifier. For the fourth dataset (Gistic2), the mRMR algorithm yields better than all other algorithms for all classifiers other than KNN classifier. For this dataset, the mRMR algorithm obtains the best accuracy (=78.70%) for the AdaBoost classifier. For the last dataset (Pathway activity), the VWmRMR algorithm performs better than other two algorithms, INMIFS and SVMRFECBR for all classifiers. For this dataset, the VWmRMR algorithm obtains the best accuracy (=83.66%) for the Naive Bayes classifier. As an overall, the VWMRmR method offers best accuracy in seven cases. On the other hand, the mRMR and SVMRFECBR methods yield the best accuracy in six cases. Thus, the the VWMRmR method seems to be the best among all five feature selection algorithms in terms of providing the best accuracy in most number of times.
The comparative performances of the VWMRmR method against each of the remaining four feature selection algorithms is summarized in Table 3. For each data, the value of a cell under the column of a given feature selection algorithm specifies the number of times the VWMRmR method wins, draws and looses against the respective feature selection algorithm on that particular data. For the first dataset (Exp), the VWMRmR method wins in all four cases against the DFS method, whereas the VWMRmR method wins against the mRMR method the same number cases (=2) , whatever the mRMR method wins against the other. For this data, both the INMIFS and SVMRFECBR methods wins against the VWMRmR method in more number of cases whatever the VWMRmR method wins against each of the them. For the second dataset ExpExon, the VWMRmR method wins in all four cases against two methods, viz. INMIFS and DFS methods. For this data, the VWMRmR method also wins against each of other two methods: mRMR and SVMRFECBR, in three cases out of four cases. For the third data (hMethyl27), the VWMRmR method wins against the DFS method in all four cases, whereas it wins against each of two methods, namely, mRMR and INMIFS, in three cases. For this data, both the SVMRFECBE and VWMRmR methods win against each other the same number of times (=2). For the fourth data (Gistic2), the VWMRmR against the DSF method in three cases. This is the only one data for which any other method (viz., mRMR and SVMRFECBR) wins against the VWMRmR method in all four cases. Furthermore, for this data, the DFS method wins against the VWMRmR method in three cases. For the fifth data (Pathway activity data), the VWMRmR method wins against each of two methods, viz, INMIFS and SVMRFECBR, in all four cases, whereas it wins against each of other two methods, viz., mRMR and DFS, in two cases. As an overall, the VWMRmR method performs better than other algorithms for most of the datasets.
Comparison of different feature selection algorithms in terms of redundancy rate
In this section, the performance of the aforesaid five feature selection methods is compared in terms of average redundancy rate. For this purpose, two different kinds of similarity measures are used. The first measure is the normalized mutual information and the second measure is the Pearson correlation coefficient. The comparative results of the average redundancy rate, in terms of normalized mutual information, of the topranked 50 features selected using different feature selection methods are shown in Table 4. In the table, for each data, the boldfaced entry under a given feature selection algorithm indicates that it offers the least redundancy rate among all feature selection algorithms on the respective data. The lower the value of redundancy rate, the better the selection of the feature selection algorithm is. As revealed from the table, the feature subset obtained by the VWMRmR approach is observed to be the least redundant in three cases. This shows that the VWMRmR algorithm is able to remove redundant features while selecting a small subset of relevant features.
Similarly, the comparative results of the average redundancy rates, in terms of the Pearson correlation coefficient, of the topranked 50 features obtained using the same five feature selection methods across all datasets are reported in Table 5. Similar to the above case, the lower the value of redundancy rate, the better the feature selection algorithm is. As revealed from Table 5, the VWMRmR method yields a small set of relevant features with the least redundancy for two datasets, viz., Gistic2 and Pathway activity data. On the other hand, the SVMRFECBR method offers the least redundant feature sets for two datasets: Exp and hMethyl27, while the DFS method yields the least redundant feature set for ExpExon data. As an overall, the VWMRmR method is capable of removing redundant features to improve the learning performance.
Comparison of different feature selection algorithms in terms of representation entropy
Table 6 shows the representation entropy of different feature subsets selected using the aforesaid five feature selection algorithms for various datasets. The value of each cell under ith row and jth column represents the representation entropy of the topranked 50 features obtained by applying the jth feature selection algorithm to the ith dataset. In the table, the boldfaced value for each entry indicates that the corresponding feature selection algorithm is the best among all algorithms in terms of providing the maximum representation entropy. The higher the value of representation entropy, the better the feature selection algorithm is. As revealed from Table 6, the VWMRmR algorithm obtains the best representation entropy for three datasets, viz., Exp, Gistic2 and Pathway activity data. On the other hand, the DFS and SVMRFECBR algorithms give the best representation entropy for the remaining two datasets, viz., ExpExon and hMethyl27, respectively. For both these two data, the VWMRmR method offers the second maximum value. As an overall, the VWMRmR method seems to be the best in terms of obtaining the best representation entropy in most cases.
Intersection of features obtained by different feature selection algorithms
In addition, we carried out the intersection of top 50 extracted statistically significant features (genes) among five feature selection algorithms and also represented Venn diagrams for expression data (Fig. 2A), exon expression data (Fig. 2B), methylation data (Fig. 2C), copy number (Gistic2) data (Fig. 2D), and pathway activity data (Fig. 2E).
For gene expression data, we obtained 7 common genes among the five feature selection methods (mRMR, INMIFS, DFS, SVMRFECBR and VWMRmR). Those were ZMIZ1, ENG, FGFR1, PAWR, KRT17, MPO and LAT2. For exon expression data, we obtained 9 common genes among those five feature selection methods (mRMR, INMIFS, DFS, SVMRFECBR and VWMRmR). Those were ZMIZ1, ENG, ALOX15, RTN4R, FGFR1, KRT17, PTPRM, MPO and LAT2. In the analysis of methylation data, 7 overlapped genes (namely, PRF1, FCGR2A, CD3D, MPO, ANGPT1, PREX1 and KRT17) were identified for those feature selection methods. For copy number (Gistic2) dataset, we identified only one gene (PIK3CG), while 3 genes had been found for those feature selection methods for pathway activity data (namely, BMP4, STXBP1 and LEP).
Conclusions
In this paper, we presented a comparative study of five widely used stateofthe art feature selection methods such as mRMR, INMIFS, DFS, SVMRFECBR, and VWMRmR for multiomics datasets. The usefulness of different feature subsets selected using the aforesaid five feature selection algorithms was assessed using three evaluation criteria: classification accuracy, representation entropy and redundancy. For the comparison purpose, four different classification methods such as C4.5, Naive Bayes, KNN and AdaBoost were used to measure the worth of each selected feature subset. Overall, the VWMRmR method offers the best performance for all three evaluation metrics for majority of datasets. The VWMRmR algorithm obtains the best Acc for most of the cases across all datasets except the Exp and Gistic2 data. It also obtains the best RE for three datasets (Exp, ExpExon and Paradigm IPLs); while the DFS and SVMRFECBR algorithms produce the best RE for the remaining two datasets  hMethyl27 and Gistic2, respectively. Additionally, we carried out intersection of top 50 statistically significant features obtained by those five feature selection methods for each omic data and overlapped feature set is defined as signature genes using supervised learning. We obtained a 7gene signature (ZMIZ1, ENG, FGFR1, PAWR, KRT17, MPO and LAT2) for gene expression, a 9gene signature for exon expression data, a 7gene signature for methylation data, a singlegene signature (PIK3CG) for copy number data and a 3gene signature for pathway activity data. In future, we plan to develop sophisticated feature selection algorithms to speed up the process of feature selection for such kinds of datasets with more than ten thousand samples. We have also planned to incorporate higher order dependencies among features.
Availability of data and materials
The following publicly available databases have been used in this study namely xenabrowser database (https://xenabrowser.net/datapages/?cohort=TCGA)
Abbreviations
 INMIFS:

Improved normalized mutual information feature selection
 MIFS:

Mutual informationbased feature selection
 mRMR:

Minimum redundancy maximum relevance
 NMIFS:

Normalized mutual information feature selection
 VWMRmR:

Variable weighted maximal relevance minimal redundancy
References
Maulik U, Bandyopadhyay S, Wang JTL. Computational intelligence and pattern analysis in biological informatics. Singapore: Wiley; 2010.
Aqil M, Naqvi AR, Mallik S, et al. The HIV NEF protein modulates cellular and exosomal mirna profiles in human monocytic cells. J Extracell Vesicles. 2014;3:1–12.
Qin G, Mallik S, Mitra R, et al. Microrna and transcription factor coregulatory networks and subtype classification of seminoma and nonseminoma in testicular germ cell tumors. Nat Sci Rep. 2020;10:1–14.
Mallick K, Mallik S, Bandyopadhyay S, Chakraborty S. A novel graph topology based gosimilarity measure for signature detection from multiomics data and its application to other problems. IEEE/ACM Trans Comput Biol Bioinform. 2020. https://doi.org/10.1109/TCBB.2020.3020537.
Blum AL, Langley P. Selection of relevant features and examples in machine learning. Artif Intell. 1997;97(1–2):245–71.
Mallik S, Zhao Z. Graph and rulebased learning algorithms: a comprehensive review of their applications for cancer type classification and prognosis using genomic data. Brief Bioinform. 2018;21:368–94.
Mallik S, Zhao Z. Identification of gene signatures from RNASEQ data using paretooptimal cluster algorithm. BMC Syst Biol. 2018;12:21–9.
Pal SK, Mitra P. Pattern recognition algorithms for data mining. Boca Raton: CRC Press; 2004.
Mallik S, Zhao Z. Congems: condensed gene coexpression module discovery through rulebased learning and its application to lung squamous cell carcinoma. Genes. 2017;9:1–25.
Bandyopadhyay S, Mallik S. Integrating multiple data sources for combinatorial marker discovery: a study in tumorigenesis. IEEE/ACM Trans Comput Biol Bioinform. 2018;15:673–87.
Maulik U, Bandyopadhyay S, Mukhopadhyay A. Multiobjective genetic algorithms for clustering: applications in data mining and bioinformatics. New York: Springer; 2011.
Mallik S, Bhadra T, Maulik U. Identifying epigenetic biomarkers using maximal relevance and minimal redundancy based feature selection for multiomics data. IEEE Trans Nanobiosci. 2017;16:3–10.
Mallik S, Seth S, Bhadra T, Bandyopadhyay S. A linear regression and deep learning approach for detecting reliable genetic alterations in cancer using DNA methylation and gene expression data. Genes. 2020;11:931.
Jain AK, Duin RPW, Mao J. Statistical pattern recognition: a review. IEEE Trans Pattern Anal Mach Intell. 2000;22(1):4–37.
Saeys Y, Inza I, Larranaga P. A review of feature selection techniques in bioinformatics
Bhattacharyya M, Feuerbach L, Bhadra T, Lengauer T, Bandyopadhyay S. Microrna transcription start site prediction with multiobjective feature selection. Stat Appl Genet Mol Biol. 2012;11(1):1–25.
Bhadra T, Bhattacharyya M, Feuerbach L, Lengauer T, Bandyopadhyay S. Dna methylation patterns facilitate the identification of microrna transcription start sites: a brainspecific study. PLoS ONE. 2013;8(6):66722.
Bhadra T, Bandyopadhyay S. Supervised feature selection using integration of densest subgraph finding with floating forwardbackward search. Inf Sci. 2021;566:1–18.
Bandyopadhyay S, Bhadra T, Maulik U, Mitra P. Integration of dense subgraph finding with feature clustering for unsupervised feature selection. Pattern Recognit Lett. 2014;40:104–12.
Bhadra T, Bandyopadhyay S. Unsupervised feature selection using an improved version of differential evolution. Expert Syst Appl. 2015;42(8):4042–53.
Guyon I, Elisseeff A. An introduction to variable and feature selection. J Mach Learn Res. 2003;3:1157–82.
Battiti R. Using mutual information for selecting features in supervised neural net learning. IEEE Trans Neural Netw. 1994;5(4):537–50.
Kwak N, Choi CH. Input feature selection by mutual information based on Parzen window. IEEE Trans Pattern Anal Mach Intell. 2002;24(12):1667–71.
Peng H, Long F, Ding C. Feature selection based on mutual information: criteria of maxdependency, maxrelevance and minredundancy. IEEE Trans Pattern Anal Mach Intell. 2005;27(8):1226–38.
Estevez PA, Tesmer M, Perez CA, Zurada JM. Normalized mutual information feature selection. IEEE Trans Neural Netw. 2009;20(2):189–201.
Vinh LT, Thang ND, Lee YK. An improved maximum relevance and minimum redundancy feature selection algorithm based on normalized mutual information. In: Tenth IEEE/IPSJ international symposium on applications and the internet (SAINT), 2010. p. 395–98.
Bandyopadhyay S, Bhadra T, Maulik U. Variable weighted maximal relevance minimal redundancy criterion for feature selection using normalized mutual information. J Multvalued Log S. 2015;25:189.
Guyon I, Weston J, Barnhill S, Vapnik V. Gene selection for cancer classification using support vector machines. Mach Learn. 2002;46:389–422.
Yan K, Zhang D. Feature selection and analysis on correlated gas sensor data with recursive feature elimination. Sens Actuators B Chem. 2015;212:353–63.
Tao H, Hou C, Nie F, Jiao Y, Yi D. Effective discriminative feature selection with nontrivial solutions. IEEE Trans Neural Netw Learn Syst. 2016;27(4):796–808.
Goldman MJ, Craft B, Hastie M, et al. Visualizing and interpreting cancer genomics data via the xena platform. Nat Biotechnol. 2020;38:1–4.
The cancer genome atlas (TCGA) acute myeloid leukemia (LAML) dataset. https://xenabrowser.net/datapages/?cohort=TCGA%20Acute%20Myeloid%20Leukemia%20(LAML)&removeHub=https%3A%2F%2Fxena.treehouse.gi.ucsc.edu%3A443. Accessed 25 July 2019.
Bandyopadhyay S, Mallik S, Mukhopadhyay A. A survey and comparative study of statistical tests for identifying differential expression from microarray data. IEEE/ACM Trans Comput Biol Bioinform. 2014;11(1):95–115.
Bhadra T, Mallik S, Bandyopadhyay S. Identification of multiview gene modules using mutual informationbased hypograph mining. IEEE Trans Syst Man Cybern. 2019;49(6):1119–30.
Smyth GK. Linear models and empirical Bayes methods for assessing differential expression in microarray experiments. Stat Appl Genet Mol Biol. 2004;3:1.
Hall M, Frank E, Holmes G, Pfahringer B, Reutemann P, Witten IH. The WEKA data mining software: an update. ACM SIGKDD Explor. 2009;11(1):10–8.
Acknowledgements
We thank the members in Bioinformatics and Systems Medicine Laboratory and Department of Computer Science & Engineering for the useful discussion and valuable suggestions.
About the supplement
This article has been published as part of BMC Bioinformatics Volume 23 Supplement 3, 2022: Selected articles from the International Conference on Intelligent Biology and Medicine (ICIBM 2021): bioinformatics. The full contents of the supplement are available online at https://bmcbioinformatics.biomedcentral.com/articles/supplements/volume23supplement3.
Funding
ZZ was partially supported by the Cancer Prevention and Research Institute of Texas (CPRIT RP170668, RP180734 and RP210045) and National Institutes of Health (R01LM012806). Publications costs are funded by Dr. Zhao’s Professorship Fund. The funder had no role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Author information
Authors and Affiliations
Contributions
TB, SM and NH developed and implemented the proposed methodology, carried out experiments, wrote and revised the manuscript. ZZ conceived the project and participated in manuscript writing and revision. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
About this article
Cite this article
Bhadra, T., Mallik, S., Hasan, N. et al. Comparison of five supervised feature selection algorithms leading to top features and gene signatures from multiomics data in cancer. BMC Bioinformatics 23 (Suppl 3), 153 (2022). https://doi.org/10.1186/s1285902204678y
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s1285902204678y
Keywords
 Feature selection
 Multiomics data
 Classifier
 Representation entropy
 Redundancy rate