Skip to main content

Asymmetric bagging and feature selection for activities prediction of drug molecules



Activities of drug molecules can be predicted by QSAR (quantitative structure activity relationship) models, which overcomes the disadvantages of high cost and long cycle by employing the traditional experimental method. With the fact that the number of drug molecules with positive activity is rather fewer than that of negatives, it is important to predict molecular activities considering such an unbalanced situation.


Here, asymmetric bagging and feature selection are introduced into the problem and asymmetric bagging of support vector machines (asBagging) is proposed on predicting drug activities to treat the unbalanced problem. At the same time, the features extracted from the structures of drug molecules affect prediction accuracy of QSAR models. Therefore, a novel algorithm named PRIFEAB is proposed, which applies an embedded feature selection method to remove redundant and irrelevant features for asBagging. Numerical experimental results on a data set of molecular activities show that asBagging improve the AUC and sensitivity values of molecular activities and PRIFEAB with feature selection further helps to improve the prediction ability.


Asymmetric bagging can help to improve prediction accuracy of activities of drug molecules, which can be furthermore improved by performing feature selection to select relevant features from the drug molecules data sets.


Modeling of quantitative structure activity relationship (QSAR) of drug molecules will help to predict the molecular activities, which reduce the cost of traditional experiments, simultaneously improve the efficiency of drug molecular design [1]. Molecular activity is determined by its structure, so structure parameters are extracted by different methods to build QSAR models. Many machine learning methods have been used to the modeling of QSAR problems, like multiple linear regression, k-nearest neighbor [2], partial least squares [3], Kriging [4], artificial neural networks [5] and support vector machines (SVM), of which SVM is a state-of-arts method and achieved satisfactory results in the previous studies [68].

Nowadays, ensemble learning is becoming a hot topic in the machine learning and bioinformatics communities [9], which has been widely used to improve the generalization performance of single learning machines. For ensemble learning, a good ensemble is one whose individuals are both accurate and make their errors on different parts of the input space [9]. The most popular methods for ensembles creation are Bagging and Boosting [1012]. The effectiveness of such methods comes primarily from the diversity caused by re-sampling the training set. Agrafiotis et al. [13] compared bagging with other single learning machines on handling QSAR problems and found that bagging is not always the best one. Signal was proposed in [14], it created an ensemble of meaningful descriptors chosen from a much larger property space which showed better performance than other methods. Random forest was also used in QSAR problems [15]. Dutta et al. used [16] different learning machines to make an ensemble to build QSAR models, and feature selection is used to produce different subsets for different learning machines.

Although the above learning methods obtained satisfactory results, but most of the previous works ignored a critical problem in the modeling of QSAR that the number of positive examples often greatly fewer than that of negatives. To handle this problem, Hou et al. [17] discussed this problem and assigned different costs for two different classes of SVM and improved the prediction results. Here combing ensemble methods, we propose to use asymmetric bagging of SVM to address the unbalanced problem. Asymmetric bagging of SVM has been used to improve relevance feedback in image retrieval [18]. Instead of re-sampling from the whole data set, asymmetric bagging keeps the positive examples fixed and re-samples only from the negatives to make the data subset of individuals unbalanced. Furthermore, we employ AUC (area under ROC curves) [19] as the measure of predictive results, because only the measure of prediction accuracy of correction can not show the overall performance. We will analysis the experimental results in terms of AUC and other several popular measures like sensitivity and specificity as well as correction. Furthermore, In QSAR problems, many parameters are extracted from the molecular structures as features, but some features are redundant and even irrelevant, these features will hurt the generalization performance of learning machines [20]. For feature selection, different methods can be categorized into the filter model, the wrapper model and the embedded model [2022], where the filter model is independent of the learning machine and both the embedded model and the wrapper model are depending on the learning machine, but the embedded model has lower computation complexity than the wrapper model has. Different methods have been applied to QSAR problems [2325], and shown that proper feature selection of molecular descriptor will help improve the prediction accuracy.

In order to improve the accuracy of asymmetric bagging, we will use the feature selection methods to improve the accuracy of individuals, this is motivated by the work of Li and Liu's work [26], where they found embedded feature selection is effective to improve accuracy of bagging of SVM and proposed an algorithm PRIFEB, which improved generalization performance of ordinary bagging. Here we propose to combine PRIFEB with asymmetric bagging and develop a novel algorithm named PRIFEAB to solve the prediction problem of unbalanced QSAR.

Results and discussion

In order to demonstrate the effect of unbalanced learning methods, we have performed the following series experiments by using support vector machine (SVM) as base classifiers.

1. SVM is a baseline method, which uses a 2-norm soft margin version of SVM.

2. unSVM assigns different C for different classes. The parameter of balanced_bridge is set as the value of the ratio of the number of positive examples to that of negatives which is 0.0188 in this paper.

3. Bagging a commonly used ensemble method, which uses SVM as base learners. The number of individuals is 55.

4. unBagging is also a commonly used bagging method, which uses unSVM as base learners. There are also 55 individuals.

5. asBagging is asymmetric bagging which uses SVM as base learners.

6. PPIFEAB is a bagging method, which employs feature section for asBagging to remove irrelevant and redundant features.

Prediction performance

Experiments are performed to investigate if asymmetric bagging and feature selection help to improve performance of bagging. Support vector machines with C = 100, σ = 0:1 are used as individual classifiers, and the number of individuals is 55 for all bagging methods. For unSVM, balanced_bridge is used to denote the ratio of C+ to C-, which is 0.0188. For ordinary bagging, each individual has one tenth of the training data set, while for asBagging, the size of individual data subset is twice of the positive sample in the whole data set. The 3-fold cross validation scheme is used to validate the results, experiments on each algorithm are repeated 10 times. We test the learning methods on individual molecular descriptors, and there are BCUT, Constitutional, Prop and Topological descriptors, which are represented by BCUT, CONST, PROP and TOPO respectively.

The average BACC values are shown in Figure 1, from which, we can obviously find that:

Figure 1

Performance of different learning algorithms. Both graphs show BACC scores. Top: Results grouped by descriptors. Bottom: Results grouped by different learning algorithm.

(1) unSVM does improve performance of SVM.

(2) Bagging does not reach our expectation, it does not improve performance of SVM, so does unBagging, which has the similar results of Bagging.

(3) asBagging greatly improves performance of SVM, and PRIFEAB slightly improve results of asBagging.

Tables 1, 2, 3, 4, 5, 6, 7 list the results of different measures i.e. AUC, BACC, sensitivity, specificity, PPV, NPV, correction by using the above SVM and bagging methods. We also list the ratio values of the number of features used in PRIFEAB to the total number in Table 8. From tables 1, 2, 3, 4, 5, 6, 7, 8, we can see that:

Table 1 Statistics values of AUC (%).
Table 2 Statistics values of BACC (%).
Table 3 Statistics values of sensitivity (%).
Table 4 Statistics values of specificity (%).
Table 5 Statistics values of positive predictive value (PPV) (%).
Table 6 Statistics values of negative predictive value (NPV) (%).
Table 7 Statistics values of correction (%).
Table 8 Statistics ratio values of the number of features used in PRIFEAB to the total number (%).

(1) unSVM obtains a slight improvement of ordinary SVM on three descriptors in terms of the AUC and BACC measures.

(2) Ordinary Bagging fails to improve single learning methods, not only Bagging but also unBagging get worse results than SVM and unSVM on the measures of AUC, BACC and sensitivity.

(3) asBagging and PRIFEAB obtain 20% better results than SVM, unSVM, Bagging and unBagging on the AUC measure. The sensitivity values of asBagging and PRIFEAB increase by beyond 50% from SVM, unSVM, Bagging and unBagging on average.

(4) PRIFEAB obtains slightly better results than asBagging on both sensitivity and specificity measures. We also observed that only few features are removed by feature selection.

(5) There are several cases, the learning machines fail in prediction and nearly all the examples are classified into negative, i.e. SVM, Bagging, unBagging on CONST and unSVM on TOPO. Only asBagging and PRIFEAB succeed in all predication.


The above results show that asBagging and PRIFEAB perform better than the other several methods of SVM, unSVM, Bagging and unBagging. Here we give some insights on these results:

(1) Though single SVM is not stable, and can not obtain valuable results, in this case of high skew data sets, bagging does not improve its generalization performance in terms of AUC, BACC and sensitivity. Bagging gets a high correction value, which is trivial, because few positive examples are predicted correctly. Especially, when learning machines fail in prediction on some descriptor data sets, all the labels are predicted as negative, a high value of correction is obtained as 98.15%, which is the ratio of negative sample to the whole sample.

(2) Since this is a drug discovery problem, we pay more attention to positives. AUC, BACC and sensitivity are more valuable than correction to measure a classifier. Asymmetric bagging and PRIFEAB improve the AUC values of ordinary bagging. Simultaneously, sensitivity are improved greatly, which shows asymmetric bagging is proper to solve the unbalanced drug discovery problem. Asymmetric bagging wins in two aspects, one is that it make the individual data subsets balanced, the second is that it pay more attention to the positives by leaving the positives always in the data set, which makes sensitivity is higher than ordinary bagging.

(3) PRIFEAB achieves slightly better results than asymmetric bagging does. Feature selection using prediction risk as criteria also make PRIFEAB win in two aspects, one is that embedded feature selection is dependent with the used learning machine, it will select features which benefit the generalization performance of individual classifiers, the second is that different features selected for different individual data subsets, which makes more diversity of bagging and improves their whole performance. The results improved by PRIFEAB than asymmetric are slight, we consider the reason is that few features are removed. Feature selection using prediction risk is dependent on SVM. Here, positives are few, which will hurt generalization performance of SVM, and furthermore hurt effect of feature selection.

(4) The data set used is so skew that the ratio of positives to negatives is only 0.0188, not beyond 2%, which makes SVM, unSVM and Baging, unBagging disable of prediction, they fail on four out of sixteen cases and predict almost all labels to negative, even on other twelve cases, they give low sensitivity. Analysis of high skew data set is still a difficult problem.


To address the unbalanced problem of drug discovery, we propose to apply asymmetric bagging and feature selection to the modeling of QSAR of drug molecules. Asymmetrical bagging of SVM and a novel algorithm PRIFEAB are compared with ordinary bagging of support vector machines on a large drug molecular activities data set, experiments show that asymmetric bagging and feature selection can improve the prediction ability of SVM in terms of AUC and sensitivity. Since this is a drug discovery problem, the positive sample is few but important, AUC and sensitivity is more proper than correction to measure generalization performance of classifiers.

This work introduces asymmetric bagging into prediction of drug activities and furthermore extends feature selection to asymmetric bagging. This work only concerns an embedded feature selection model with the prediction risk criteria, one of the future work will try more efficient and more effective feature selection methods for this task.


Support vector machines

Support vector machines (SVM) [27] proposed by Vapnik and his co-workers in 1990s, have been developed quickly during the last decade [28], and successfully applied to biological data mining [29], drug discovery [6, 8] etc. Denoting the training sample as S = {(x, y)} {n× {-1, 1}}, SVM discriminant hyperplane can be written as

y = sgn(w·x + b)

where w is a weight vector, b is a bias. According to the generalization bound in statistical learning theory [30], we need to minimize the following objective function for a 2-norm soft margin version of SVM

minimize w , b w w + C Σ i = 1 ξ i 2 subject to y i ( w x i + b ) 1 ξ i , i = 1 , ... , , MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqbaeaabiGaaaqaaiabb2gaTjabbMgaPjabb6gaUjabbMgaPjabb2gaTjabbMgaPjabbQha6jabbwgaLnaaBaaaleaaieqacqWF3bWDcqGGSaalcqWGIbGyaeqaaaGcbaGaeyykJeUae83DaCNaeyyXICTae83DaCNaeyOkJeVaey4kaSIaem4qamKaeu4Odm1aa0baaSqaaiabdMgaPjabg2da9iabigdaXaqaaiabloriSbaakiabe67a4naaDaaaleaacqWGPbqAaeaacqaIYaGmaaaakeaacqqGZbWCcqqG1bqDcqqGIbGycqqGQbGAcqqGLbqzcqqGJbWycqqG0baDcqqGGaaicqqG0baDcqqGVbWBaeaacqWG5bqEdaWgaaWcbaGaemyAaKgabeaakiabcIcaOiabgMYiHlab=Dha3jabgwSixlab=Hha4naaBaaaleaacqWGPbqAaeqaaOGaeyOkJeVaey4kaSIaemOyaiMaeiykaKIaeyyzImRaeGymaeJaeyOeI0IaeqOVdG3aaSbaaSqaaiabdMgaPbqabaGccqGGSaalcqWGPbqAcqGH9aqpcqaIXaqmcqGGSaalcqGGUaGlcqGGUaGlcqGGUaGlcqGGSaalcqWItecBcqGGSaalaaaaaa@803A@

in which, slack variable ξ i is introduced when the problem is infeasible. The constant C > 0 is a penalty parameter, a larger C corresponds to assigning a larger penalty to errors.

By building a Lagrangian and using the Karush-Kuhn-Tucker (KKT) complementarity conditions [31, 32], we can obtain the value of optimization problem (1). Because of the KKT conditions, only those Lagrangian multipliers, α i s, which make the constraint active are non zeros, we denote these points corresponding to the non zero α i s as support vectors (sv). Therefore we can describe the classification hyperplane in terms of α and b:

y = sgn ( i s v α i x i x + b ) . MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemyEaKNaeyypa0Jagi4CamNaei4zaCMaeiOBa42aaeWaaeaadaaeqbqaaiabeg7aHnaaBaaaleaacqWGPbqAaeqaaOGaeyykJeocbeGae8hEaG3aaSbaaSqaaiabdMgaPbqabaGccqGHflY1cqWF4baEcqGHQms8cqGHRaWkcqWGIbGyaSqaaiabdMgaPjabgIGiolabdohaZjabdAha2bqab0GaeyyeIuoaaOGaayjkaiaawMcaaiabc6caUaaa@4CDC@

If we replace x i ·x with some function K(x i , x), which satisfies Mercer's condition [33], then the classification function can be written as:

y = sgn ( i s v α i K x i x + b ) . MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemyEaKNaeyypa0Jagi4CamNaei4zaCMaeiOBa42aaeWaaeaadaaeqbqaaiabeg7aHnaaBaaaleaacqWGPbqAaeqaaOGaem4saSKaeyykJeocbeGae8hEaG3aaSbaaSqaaiabdMgaPbqabaGccqGHflY1cqWF4baEcqGHQms8cqGHRaWkcqWGIbGyaSqaaiabdMgaPjabgIGiolabdohaZjabdAha2bqab0GaeyyeIuoaaOGaayjkaiaawMcaaiabc6caUaaa@4DFB@

where K(x, z) is the known kernel function. A commonly used kernel function is the Guass kernel:

K(x, z) = exp(-||x - z||2/σ2),

which is also preferred by C.J. Lin et al. [34].

To address the unbalanced problem, C in Equ.(1) is separated as C+ and C- to adjust the penalties on the false negative vs. false positive, so Equ.(1) becomes:

minimize w , b w w + C + Σ i = 1 ; y i = = 1 ξ i 2 + C Σ i = 1 ; y i = = 1 ξ i 2 subject to y i ( w x i + b ) 1 ξ i , i = 1 , ... , , MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqbaeaabiGaaaqaaiabb2gaTjabbMgaPjabb6gaUjabbMgaPjabb2gaTjabbMgaPjabbQha6jabbwgaLnaaBaaaleaaieqacqWF3bWDcqGGSaalcqWGIbGyaeqaaaGcbaGaeyykJeUae83DaCNaeyyXICTae83DaCNaeyOkJeVaey4kaSIaem4qam0aaSbaaSqaaiabgUcaRaqabaGccqqHJoWudaqhaaWcbaGaemyAaKMaeyypa0JaeGymaeJaei4oaSJaemyEaK3aaSbaaWqaaiabdMgaPbqabaWccqGH9aqpcqGH9aqpcqaIXaqmaeaacqWItecBaaGccqaH+oaEdaqhaaWcbaGaemyAaKgabaGaeGOmaidaaOGaey4kaSIaem4qam0aaSbaaSqaaiabgkHiTaqabaGccqqHJoWudaqhaaWcbaGaemyAaKMaeyypa0JaeGymaeJaei4oaSJaemyEaK3aaSbaaWqaaiabdMgaPbqabaWccqGH9aqpcqGH9aqpcqGHsislcqaIXaqmaeaacqWItecBaaGccqaH+oaEdaqhaaWcbaGaemyAaKgabaGaeGOmaidaaaGcbaGaee4CamNaeeyDauNaeeOyaiMaeeOAaOMaeeyzauMaee4yamMaeeiDaqNaeeiiaaIaeeiDaqNaee4Ba8gabaGaemyEaK3aaSbaaSqaaiabdMgaPbqabaGccqGGOaakcqGHPms4cqWF3bWDcqGHflY1cqWF4baEdaWgaaWcbaGaemyAaKgabeaakiabgQYiXlabgUcaRiabdkgaIjabcMcaPiabgwMiZkabigdaXiabgkHiTiabe67a4naaBaaaleaacqWGPbqAaeqaaOGaeiilaWIaemyAaKMaeyypa0JaeGymaeJaeiilaWIaeiOla4IaeiOla4IaeiOla4IaeiilaWIaeS4eHWMaeiilaWcaaaaa@9DE7@

where C+ is C and C- is balanced_bridge*C, balanced_bridge is a coefficient, it is assigned as the ratio of the number of positive examples to that of negative ones.

The SVM obtained by the above equation is named as unSVM. This is implemented in LibSVM [35], it had been used to analysis AMDE data and is effective for unbalanced data set [17].

Asymmetric bagging

Bagging is one of the traditional ensemble methods, which uses bootstrap to produce the diversity of individuals and uses major voting to obtain the final decision results for classification problems [10]. Figure 2 shows the ordinary bagging approach based on support vector machine. When we use a unSVM to train the individuals of bagging, we obtain the unBagging approach as in Figure 3.

Figure 2

The Bagging approach.

Figure 3

The unBagging approach.

Bagging helps to improve stable of single learning machines, but unbalance also reduce its generalization performance, therefore, we propose to employ asymmetric bagging to handle the unbalanced problem, which only execute the bootstrapping on the negative examples since there are far more negative examples than the positive ones. Tao et al. [18] applied asymmetric bagging to another unbalanced problem of relevance feedback in image retrieval and obtained satisfactory results. This way make individual classifiers of bagging be trained on a balanced number of positive and negative examples, thus solve the problem of unbalanced examples. The asymmetric bagging of SVM (asBagging) is described in Figure 4.

Figure 4

The asBagging approach.

asBagging can solve the unstable problem of SVM classifiers and the unbalance problem in the training set.

However, it can not solve the problem of irrelevant and weak redundant features in the data sets. We will solve it by feature selection embedded in the bagging method as in the next subsection.


Feature selection has been used in ensemble learning and obtained some interesting results, Li and Liu proposed to use the embedded feature selection method with the prediction risk criteria for bagging of SVM, where feature selection can effectively improve the accuracy of bagging methods [36]. As a feature selection method, the prediction risk criteria was proposed by Moody and Utans [37] which evaluates one feature through estimating prediction error of the data sets when the values of all examples of this feature are replaced by their mean value.

S i = AUC AUC ( x ¯ i ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaem4uam1aaSbaaSqaaiabdMgaPbqabaGccqGH9aqpcqqGbbqqcqqGvbqvcqqGdbWqcqGHsislcqqGbbqqcqqGvbqvcqqGdbWqcqGGOaakcuWG4baEgaqeamaaCaaaleqabaGaemyAaKgaaOGaeiykaKcaaa@3C39@

where AUC (Area under ROC) [38] is predicted on the training data set, and AUC( x ¯ i MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGafmiEaGNbaebadaahaaWcbeqaaiabdMgaPbaaaaa@2EEE@ ) is the prediction AUC on the training data set with the mean value of ithfeature. Finally, the feature corresponding with the smallest will be deleted, because this feature causes the smallest error and is the least important one.

Since the asymmetric bagging method can overcome both the problems of unstable and unbalance, and feature selection can overcome the problem of irrelevant features for bagging. So we propose a novel algorithm to combine both two methods. The algorithm is named as PRIFEAB (Prediction rIsk based Feature sElection for Asymmetric Bagging), which is described in Figure 5. The basic idea of PRIFEAB is that we first use bootstrap sampling to generate a negative sample, and combine it with the whole positive sample to obtain a individual training subset. Then, prediction risk based feature selection is used to select optimal features, and we obtain an individual model by training SVM on the optimal training subset. Finally, ensemble the individual SVM classifiers by using majority voting to obtain the final model.

Figure 5

The PRIFEAB approach.

NCI AntiHIV drug screen data set

The NCI AntiHIV Drug Screen data set(NCI) is obtained from It has a categorical response measuring how a compound protects human CEM cells from HIV-1 infection. It has 29374 examples, of which 542 (1.85%) is positive and 28832 (98.15%) is negative. The structure parameters [39] consist 64 BCUT descriptors, 47 Constitutional (CONST) ones, 250 Prop ones and 266 Topological (TOPO) ones. This data set is collected and computed by Young et al., description in detail of the data set can be found in [40, 41]. Here we test the proposed learning methods on individual molecular descriptors. Since the different descriptors have different meaning and few relations with each other, they can be considered as four different data sets.


Since the class distribution of the used data set is unbalanced, only correction of classification accuracy may be misleading. Therefore, AUC (Area Under the Curve of Receiver Operating Characteristic (ROC)) [38] is used to measure the performance.

To furthermore describe the different learning methods, we also define the various measures as below [42], where TP, TN, FP, FN, stand for the number of true positive, true negative, false positive, false negative samples at classification time, respectively.

Sensitivity is defined as T P T P + F N MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqcfa4aaSaaaeaacqWGubavcqWGqbauaeaacqWGubavcqWGqbaucqGHRaWkcqWGgbGrcqWGobGtaaaaaa@3443@ and is also known as Recall.

Specificity is defined as T N T N + F P MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqcfa4aaSaaaeaacqWGubavcqWGobGtaeaacqWGubavcqWGobGtcqGHRaWkcqWGgbGrcqWGqbauaaaaaa@343F@

BACC (Balanced Accuracy) is defined as 1 2 ( T P T P + F N + T N T N + F P ) MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqcfa4aaSaaaeaacqaIXaqmaeaacqaIYaGmaaWaaeWaaeaadaWcaaqaaiabdsfaujabdcfaqbqaaiabdsfaujabdcfaqjabgUcaRiabdAeagjabd6eaobaacqGHRaWkdaWcaaqaaiabdsfaujabd6eaobqaaiabdsfaujabd6eaojabgUcaRiabdAeagjabdcfaqbaaaiaawIcacaGLPaaaaaa@407C@ , which defines the average of sensitivity and specificity.

PPV (Positive Predictive Value) is defined as T P T P + F P MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqcfa4aaSaaaeaacqWGubavcqWGqbauaeaacqWGubavcqWGqbaucqGHRaWkcqWGgbGrcqWGqbauaaaaaa@3447@ and is also known as Precision.

NPV (Negative Predictive Value) is defined as T N T N + F N MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqcfa4aaSaaaeaacqWGubavcqWGobGtaeaacqWGubavcqWGobGtcqGHRaWkcqWGgbGrcqWGobGtaaaaaa@343B@ .

Correction is defined as T P + T N T P + T N + F P + F N MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqcfa4aaSaaaeaacqWGubavcqWGqbaucqGHRaWkcqWGubavcqWGobGtaeaacqWGubavcqWGqbaucqGHRaWkcqWGubavcqWGobGtcqGHRaWkcqWGgbGrcqWGqbaucqGHRaWkcqWGgbGrcqWGobGtaaaaaa@3DD3@ and measures the overall percentage of samples correctly classified.


  1. 1.

    Barrett SJ, Langdon WB: Advances in the Application of Machine Learning Techniques in Drug Discovery, Design and Development. 10th Online World Conference on Soft Computing in Industrial Applications. 2005, Springer

    Google Scholar 

  2. 2.

    Tominaga Y: Comparative Study of Class Data Analysis with PCA-LDA, SIMCA, PLS, ANNs, and K-NN. Chemometrics and Intelligent Laboratory Systems. 1999, 49 (1): 105-115.

    CAS  Article  Google Scholar 

  3. 3.

    Tang K, Li T: Combining PLS with GA-GP for QSAR. Chemometrics and Intelligent Laboratory Systems. 2002, 64: 55-64.

    CAS  Article  Google Scholar 

  4. 4.

    Fang KT, Yin H, Liang YZ: New Approach by Kriging Models to Problems in QSAR. Journal of Chemical Information and Computer Science. 2004, 44: 2106-2113.

    CAS  Article  Google Scholar 

  5. 5.

    Li GZ, Yang J, Song HF, Yang SS, Lu WC, Chen NY: Semiempirical Quantum Chemical Method and Artificial Neural Networks Applied for Max Computation of Some Azo Dyes. Journal of Chemical Information and Computer Science. 2004, 44 (6): 2047-2050.

    CAS  Article  Google Scholar 

  6. 6.

    Xue Y, Li ZR, Yap CW, Sun LZ, Chen X, Chen YZ: Effect of Molecular Descriptor Feature Selection in Support Vector Machine Classification of Pharmacokinetic and Toxicological Properties of Chemical Agents. Journal of Chemical Information & Computer Science. 2004, 44 (5): 1630-1638.

    CAS  Article  Google Scholar 

  7. 7.

    Chen NY, Lu WC, Yang J, Li GZ: Support Vector Machines in Chemistry. 2004, Singapore: World Scientific Publishing Company

    Google Scholar 

  8. 8.

    Bhavani S, Nagargadde A, Thawani A, Sridhar V, Chandra N: Substructure-Based Support Vector Machine Classifiers for Prediction of Adverse Effects in Diverse Classes of Drugs. Journal of Chemical Information and Modeling. 2006, 46 (6): 2478-2486.

    CAS  Article  PubMed  Google Scholar 

  9. 9.

    Dietterich T: Machine-learning research: Four current directions. The AI Magazine. 1998, 18 (4): 97-136.

    Google Scholar 

  10. 10.

    Schapire R: The strength of weak learn ability. Machine learning. 1990, 5 (2): 197-227.

    Google Scholar 

  11. 11.

    Breiman L: Bagging predictors. Machine Learning. Machine learning. 1996, 24 (2): 123-140.

    Google Scholar 

  12. 12.

    Bauer E, Kohavi R: An empirical comparison of voting classification algorithms: Bagging, Boosting, and variants. Machine learning. 1999, 36 (1–2): 105-139.

    Article  Google Scholar 

  13. 13.

    Agrafiotis DK, no WC, Lobanov VS: On the Use of Neural Network Ensembles in QSAR and QSPR. J Chem Inf Comput Sci. 2002, 42: 903-911.

    CAS  Article  PubMed  Google Scholar 

  14. 14.

    Lanctot JK, Putta S, Lemmen C, Greene J: Using Ensembles to Classify Compounds for Drug Discovery. J Chem Inf Comput Sci. 2003, 43: 2163-2169.

    CAS  Article  PubMed  Google Scholar 

  15. 15.

    Guha R, Jurs PC: Development of Linear, Ensemble, and Nonlinear Models for the Prediction andInterpretation of the Biological Activity of a Set of PDGFR Inhibitors. J Chem Inf Comput Sci. 2004, 44: 2179-2189.

    CAS  Article  PubMed  Google Scholar 

  16. 16.

    Dutta D, Guha R, Wild D, Chen T: Ensemble Feature Selection: Consistent Descriptor Subsets for Multiple QSAR Models. Journal of Chemical Information and Modeling. 2007, 47 (3): 989-997.

    CAS  Article  PubMed  Google Scholar 

  17. 17.

    Hou T, Wang J, Li Y: ADME Evaluation in Drug Discovery. 8. The Prediction of Human Intestinal Absorption by a Support Vector Machine. J Chem Inf Model. 2007, 47: 2408-2415.

    CAS  Article  PubMed  Google Scholar 

  18. 18.

    Tao D, Tang X, Li X, Wu X: Asymmetric Bagging and Random Subspace for Support Vector Machines-Based Relevance Feedback in Image Retrieval. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2006, 28 (7): 1088-1099.

    Article  PubMed  Google Scholar 

  19. 19.

    Hand DJ: Construction and Assessment of Classification Rules. 1997, Chichester: John Wiley and Sons

    Google Scholar 

  20. 20.

    Yu L, Liu H: Efficient Feature Selection Via Analysis of Relevance and Redundancy. Journal of Machine Learning Research. 2004, 5 (Oct): 1205-1224.

    Google Scholar 

  21. 21.

    Kohavi R, George JH: Wrappers for Feature Subset Selection. Artificial Intelligence. 1997, 97: 273-324.

    Article  Google Scholar 

  22. 22.

    Guyon I, Elisseeff A: An Introduction to Variable and Feature Selection. Journal of machine learning research. 2003, 3: 1157-1182.

    Google Scholar 

  23. 23.

    Liu Y: A Comparative Study on Feature Selection Methods for Drug Discovery. J Chem Inf Comput Sci. 2004, 44 (5): 1823-1828.

    CAS  Article  PubMed  Google Scholar 

  24. 24.

    Li H, Yap CW, Ung CY, Xue Y, Cao ZW, Chen YZ: Effect of Selection of Molecular Descriptors on the Prediction of Blood-Brain Barrier Penetrating and Nonpenetrating Agents by Statistical Learning Methods. Journal of Chemical Information and Modeling. 2005, 45 (5): 1376-1384.

    CAS  Article  PubMed  Google Scholar 

  25. 25.

    Eitrich T, Kless A, Druska C, Meye W, Grotendorst J: Classification of Highly Unbalanced CYP450 Data of Drugs Using Cost Sensitive Machine Learning Techniques. Journal of Chemical Information and Modeling. 2007, 47 (1): 97-103.

    Article  Google Scholar 

  26. 26.

    Li GZ, Yang J, Liu GP, Xue L: Feature selection for multi-class problems using support vector machines. Lecture Notes on Artificial Intelligence 3173 (PRICAI2004). 2004, Springer, 292-300.

    Google Scholar 

  27. 27.

    Boser B, Guyon L, Vapnik V: A Training Algorithm for Optimal Margin Classifiers. Proceedings of the Fifth Annual Workshop on Computational Learning Theory. 1992, Pittsburgh: ACM, 144-152.

    Google Scholar 

  28. 28.

    Cristianini N, Shawe-Taylor J: An Introduction to Support Vector Machines. 2000, Cambridge: Cambridge University Press

    Google Scholar 

  29. 29.

    Guyon I, Weston J, Barnhill S, Vapnik V: Gene Selection for Cancer Classification Using Support Vector Machines. Machine Learning. 2002, 46: 389-422.

    Article  Google Scholar 

  30. 30.

    Vapnik V: Statistical Learning Theory. 1998, New York: Wiley

    Google Scholar 

  31. 31.

    Karush W: Minima of Functions of Several Variables with Inequalities as Side Constraints. Master's thesis. 1939, Deptment of Mathematics, University of Chicago

    Google Scholar 

  32. 32.

    Kuhn HW, Tucker AW: Nonlinear Programming. Proceeding of the 2nd Berkeley Symposium on Mathematical Statistics and Probabilistic. 1951, Berkeley: University of California Press, 481-492.

    Google Scholar 

  33. 33.

    Mercer J: Functions of Positive and Negative Type and their Connection with the Theory of Integral Equations. Philosophy Transactions on Royal Society in London. 1909, A 209: 415-446.

    Article  Google Scholar 

  34. 34.

    Hsu CW, Chang CC, Lin CJ: A Practical Guide to Support Vector Classification. Tech rep. 2003, Department of Computer Science and Information Engineering of National Taiwan University, [14 August 2003], []

    Google Scholar 

  35. 35.

    Chang CC, Lin CJ: LIBSVM – A Library for Support Vector Machines Version 2.85. 2007, []

    Google Scholar 

  36. 36.

    Li GZ, Liu TY: Feature Selection for Bagging of Support Vector Machines. PRICAI2006 Lecuture Notes in Computer Science 4099. 2006, Springer, 271-277.

    Google Scholar 

  37. 37.

    Moody J, Utans J: Principled Architecture Selection for Neural Networks: Application to Corporate Bond Rating Prediction. Advances in Neural Information Processing Systems. Edited by: Moody JE, Hanson SJ, Lippmann RP. 1992, Morgan Kaufmann Publishers, Inc, 683-690.

    Google Scholar 

  38. 38.

    Duda RO, Hart PE, Stork DG: Pattern Classification. 2000, Wiley Interscience, 2

    Google Scholar 

  39. 39.

    Todeschini R, Consonni V: Handbook of Molecular Descriptors. 2000, Weinheim, Germany: Viley-VCH

    Google Scholar 

  40. 40.

    Young SS, Gombar VK, Emptage MR, Cariello NF, Lambert C: Mixture Deconvolution and Analysis of Ames Mutagenicity Data. Chemometrics and Intelligent Laboratory Systems. 2002, 60: 5-11.

    CAS  Article  Google Scholar 

  41. 41.

    Feng J, Lurati L, Ouyang H, Robinson T, Wang Y, Yuan S, Young SS: Predictive Toxicology: Benchmarking Molecular Descriptors and Statistical Methods. Journal of Chemical Information and Computer Science. 2003, 43: 1463-1470.

    CAS  Article  Google Scholar 

  42. 42.

    Levner I: Feature Selection and Nearest Centroid Classification for Protein Mass Spectrometry. BMC Bioinformatics. 2005, 6 (68):

Download references


This work was supported in part by the Natural Science Foundation of China under grant no. 20503015, the STCSM "Innovation Action Plan" Project of China under grant no. 07DZ19726, Shanghai Leading Academic Discipline Project under no. J50103 and Systems Biology Research Foundation of Shanghai University.

This article has been published as part of BMC Bioinformatics Volume 9 Supplement 6, 2008: Symposium of Computations in Bioinformatics and Bioscience (SCBB07). The full contents of the supplement are available online at

Author information



Corresponding author

Correspondence to Mary Qu Yang.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors' contributions

Guo-Zheng Li proposed the idea, designed the experiments and wrote the paper; Hao-Hua Meng performed experiments; Wen-Cong Lu helped in writing the paper; Mary Qu Yang helped design the experiments; Jack Y. Yang conceived and guided the project.

Rights and permissions

Reprints and Permissions

About this article

Cite this article

Li, G., Meng, H., Lu, W. et al. Asymmetric bagging and feature selection for activities prediction of drug molecules. BMC Bioinformatics 9, S7 (2008).

Download citation


  • Support Vector Machine
  • Feature Selection
  • Quantitative Structure Activity Relationship
  • Feature Selection Method
  • Generalization Performance