Volume 9 Supplement 6

## Symposium of Computations in Bioinformatics and Bioscience (SCBB07)

# Asymmetric bagging and feature selection for activities prediction of drug molecules

- Guo-Zheng Li
^{1, 2}, - Hao-Hua Meng
^{2}, - Wen-Cong Lu
^{3}, - Jack Y Yang
^{4}and - Mary Qu Yang
^{5}Email author

**9(Suppl 6)**:S7

https://doi.org/10.1186/1471-2105-9-S6-S7

© Li et al; licensee BioMed Central Ltd. 2008

**Published: **28 May 2008

## Abstract

### Background

Activities of drug molecules can be predicted by QSAR (quantitative structure activity relationship) models, which overcomes the disadvantages of high cost and long cycle by employing the traditional experimental method. With the fact that the number of drug molecules with positive activity is rather fewer than that of negatives, it is important to predict molecular activities considering such an unbalanced situation.

### Results

Here, asymmetric bagging and feature selection are introduced into the problem and asymmetric bagging of support vector machines (asBagging) is proposed on predicting drug activities to treat the unbalanced problem. At the same time, the features extracted from the structures of drug molecules affect prediction accuracy of QSAR models. Therefore, a novel algorithm named PRIFEAB is proposed, which applies an embedded feature selection method to remove redundant and irrelevant features for asBagging. Numerical experimental results on a data set of molecular activities show that asBagging improve the AUC and sensitivity values of molecular activities and PRIFEAB with feature selection further helps to improve the prediction ability.

### Conclusion

Asymmetric bagging can help to improve prediction accuracy of activities of drug molecules, which can be furthermore improved by performing feature selection to select relevant features from the drug molecules data sets.

## Keywords

## Background

Modeling of quantitative structure activity relationship (QSAR) of drug molecules will help to predict the molecular activities, which reduce the cost of traditional experiments, simultaneously improve the efficiency of drug molecular design [1]. Molecular activity is determined by its structure, so structure parameters are extracted by different methods to build QSAR models. Many machine learning methods have been used to the modeling of QSAR problems, like multiple linear regression, k-nearest neighbor [2], partial least squares [3], Kriging [4], artificial neural networks [5] and support vector machines (SVM), of which SVM is a state-of-arts method and achieved satisfactory results in the previous studies [6–8].

Nowadays, ensemble learning is becoming a hot topic in the machine learning and bioinformatics communities [9], which has been widely used to improve the generalization performance of single learning machines. For ensemble learning, a good ensemble is one whose individuals are both accurate and make their errors on different parts of the input space [9]. The most popular methods for ensembles creation are Bagging and Boosting [10–12]. The effectiveness of such methods comes primarily from the diversity caused by re-sampling the training set. Agrafiotis et al. [13] compared bagging with other single learning machines on handling QSAR problems and found that bagging is not always the best one. Signal was proposed in [14], it created an ensemble of meaningful descriptors chosen from a much larger property space which showed better performance than other methods. Random forest was also used in QSAR problems [15]. Dutta et al. used [16] different learning machines to make an ensemble to build QSAR models, and feature selection is used to produce different subsets for different learning machines.

Although the above learning methods obtained satisfactory results, but most of the previous works ignored a critical problem in the modeling of QSAR that the number of positive examples often greatly fewer than that of negatives. To handle this problem, Hou et al. [17] discussed this problem and assigned different costs for two different classes of SVM and improved the prediction results. Here combing ensemble methods, we propose to use asymmetric bagging of SVM to address the unbalanced problem. Asymmetric bagging of SVM has been used to improve relevance feedback in image retrieval [18]. Instead of re-sampling from the whole data set, asymmetric bagging keeps the positive examples fixed and re-samples only from the negatives to make the data subset of individuals unbalanced. Furthermore, we employ AUC (area under ROC curves) [19] as the measure of predictive results, because only the measure of prediction accuracy of correction can not show the overall performance. We will analysis the experimental results in terms of AUC and other several popular measures like sensitivity and specificity as well as correction. Furthermore, In QSAR problems, many parameters are extracted from the molecular structures as features, but some features are redundant and even irrelevant, these features will hurt the generalization performance of learning machines [20]. For feature selection, different methods can be categorized into the filter model, the wrapper model and the embedded model [20–22], where the filter model is independent of the learning machine and both the embedded model and the wrapper model are depending on the learning machine, but the embedded model has lower computation complexity than the wrapper model has. Different methods have been applied to QSAR problems [23–25], and shown that proper feature selection of molecular descriptor will help improve the prediction accuracy.

In order to improve the accuracy of asymmetric bagging, we will use the feature selection methods to improve the accuracy of individuals, this is motivated by the work of Li and Liu's work [26], where they found embedded feature selection is effective to improve accuracy of bagging of SVM and proposed an algorithm PRIFEB, which improved generalization performance of ordinary bagging. Here we propose to combine PRIFEB with asymmetric bagging and develop a novel algorithm named PRIFEAB to solve the prediction problem of unbalanced QSAR.

## Results and discussion

In order to demonstrate the effect of unbalanced learning methods, we have performed the following series experiments by using support vector machine (SVM) as base classifiers.

1. SVM is a baseline method, which uses a 2-norm soft margin version of SVM.

2. unSVM assigns different *C* for different classes. The parameter of balanced_bridge is set as the value of the ratio of the number of positive examples to that of negatives which is 0.0188 in this paper.

3. Bagging a commonly used ensemble method, which uses SVM as base learners. The number of individuals is 55.

4. unBagging is also a commonly used bagging method, which uses unSVM as base learners. There are also 55 individuals.

5. asBagging is asymmetric bagging which uses SVM as base learners.

6. PPIFEAB is a bagging method, which employs feature section for asBagging to remove irrelevant and redundant features.

### Prediction performance

Experiments are performed to investigate if asymmetric bagging and feature selection help to improve performance of bagging. Support vector machines with *C* = 100, *σ* = 0:1 are used as individual classifiers, and the number of individuals is 55 for all bagging methods. For unSVM, balanced_bridge is used to denote the ratio of *C*_{+} to *C*_{-}, which is 0.0188. For ordinary bagging, each individual has one tenth of the training data set, while for asBagging, the size of individual data subset is twice of the positive sample in the whole data set. The 3-fold cross validation scheme is used to validate the results, experiments on each algorithm are repeated 10 times. We test the learning methods on individual molecular descriptors, and there are BCUT, Constitutional, Prop and Topological descriptors, which are represented by BCUT, CONST, PROP and TOPO respectively.

(1) unSVM does improve performance of SVM.

(2) Bagging does not reach our expectation, it does not improve performance of SVM, so does unBagging, which has the similar results of Bagging.

(3) asBagging greatly improves performance of SVM, and PRIFEAB slightly improve results of asBagging.

Statistics values of AUC (%).

Descriptor | SVM | unSVM | Bagging | unBagging | asBagging | PRIFEAB |
---|---|---|---|---|---|---|

BCUT | 59.4(1.3) | 61.2(1.3) | 55.0(1.0) | 55.2(1.0) | 75.3(0.8) | 75.8(0.6) |

CONST | 50.8(0.8) | 59.3(1.9) | 50.3(1.1) | 50.4(1.1) | 75.0(0.3) | 75.3(0.5) |

PROP | 62.3(1.4) | 63.0(1.2) | 55.4(1.5) | 55.5(1.3) | 78.0(0.9) | 78.3(0.9) |

TOPO | 57.7(1.0) | 50.8(0.8) | 54.0(1.1) | 54.1(2.0) | 73.4(0.5) | 73.6(0.7) |

Average | 57.6(1.1) | 58.6(1.3) | 53.7(1.2) | 53.8(1.4) | 75.4(0.6) | 75.8(0.7) |

Statistics values of BACC (%).

Descriptor | SVM | unSVM | Bagging | unBagging | asBagging | PRIFEAB |
---|---|---|---|---|---|---|

BCUT | 60.0(0.3) | 60.8(0.9) | 54.5(1.0) | 54.2(1.0) | 68.7(0.7) | 67.8(0.5) |

CONST | 50.0(0.1) | 60.1(0.6) | 50.0(0.1) | 50.1(0.3) | 69.1(0.4) | 69.3(0.5) |

PROP | 62.0(1.0) | 62.8(0.4) | 55.6(1.1) | 54.7(1.3) | 70.3(0.1) | 71.0(0.9) |

TOPO | 57.4(0.5) | 50.0(0.0) | 53.2(0.1) | 52.8(1.0) | 66.8(0.4) | 67.3(0.8) |

Average | 57.4(0.5) | 58.4(0.5) | 53.3(0.6) | 53.0(0.9) | 68.7(0.4) | 68.9(0.7) |

Statistics values of sensitivity (%).

Descriptor | SVM | unSVM | Bagging | unBagging | asBagging | PRIFEAB |
---|---|---|---|---|---|---|

BCUT | 20.4(0.7) | 22.3(1.7) | 9.8(1.8) | 8.8(2.0) | 69.1(1.6) | 68.4(1.2) |

CONST | 0.1(0.1) | 20.4(1.2) | 0.1(0.1) | 0.2(0.5) | 73.0(1.0) | 72.5(1.0) |

PROP | 24.5(1.9) | 26.5(1.0) | 12.0(2.3) | 9.9(2.5) | 70.4(1.8) | 71.2(1.9) |

TOPO | 15.4(1.0) | 0.0(0.0) | 7.8(2.3) | 6.3(1.8) | 66.0(0.9) | 66.8(1.6) |

Average | 15.1(0.9) | 17.3(1.0) | 7.4(1.6) | 6.3(1.7) | 69.6(1.3) | 69.7(1.4) |

Statistics values of specificity (%).

Descriptor | SVM | unSVM | Bagging | unBagging | asBagging | PRIFEAB |
---|---|---|---|---|---|---|

BCUT | 99.6(0.1) | 99.3(0.1) | 99.2(0.1) | 99.6(0.1) | 68.3(0.4) | 67.3(0.2) |

CONST | 99.9(0.1) | 99.7(0.1) | 99.9(0.1) | 99.9(0.1) | 65.2(0.5) | 66.1(0.3) |

PROP | 99.5(0.1) | 99.1(0.1) | 99.3(0.1) | 99.4(0.2) | 70.3(0.4) | 70.8(0.2) |

TOPO | 99.3(0.1) | 100.0(0.0) | 98.6(0.3) | 99.2(0.1) | 67.7(0.3) | 67.7(0.3) |

Average | 99.6(0.1) | 99.5(0.1) | 99.3(0.1) | 99.5(0.1) | 67.9(0.4) | 68.0(0.3) |

Statistics values of positive predictive value (PPV) (%).

Descriptor | SVM | unSVM | Bagging | unBagging | asBagging | PRIFEAB |
---|---|---|---|---|---|---|

BCUT | 50.4(2.6) | 38.6(2.8) | 19.1(2.7) | 27.0(4.1) | 3.9(0.1) | 3.8(0.1) |

CONST | NaN(NaN) | 60.7(3.6) | NaN(NaN) | NaN(NaN) | 3.8(0.1) | 3.9(0.1) |

PROP | 4.6(0.2) | 3.6(1.3) | 23.2(2.2) | 25.7(5.4) | 4.3(0.1) | 4.4(0.1) |

TOPO | 3.0(1.8) | NaN(NaN) | 9.7(1.5) | 12.8(4.2) | 3.7(0.1) | 3.7(0.1) |

Average | 19.3(1.5) | 34.3(2.6) | 17.3(2.1) | 21.8(4.6) | 3.9(0.1) | 4.0(0.1) |

Statistics values of negative predictive value (NPV) (%).

Descriptor | SVM | unSVM | Bagging | unBagging | asBagging | PRIFEAB |
---|---|---|---|---|---|---|

BCUT | 98.5(0.1) | 98.6(0.1) | 98.3(0.1) | 98.3(0.1) | 99.2(0.1) | 99.1(0.1) |

CONST | 98.2(0.1) | 98.5(0.1) | 98.2(0.1) | 98.2(0.1) | 99.2(0.1) | 99.2(0.1) |

PROP | 98.5(0.3) | 98.6(0.1) | 98.4(0.1) | 98.3(0.1) | 99.2(0.1) | 99.2(0.1) |

TOPO | 98.4(0.1) | 98.2(0.0) | 98.3(0.1) | 98.3(0.1) | 99.1(0.1) | 99.1(0.1) |

Average | 98.4(0.1) | 98.5(0.1) | 98.3(0.1) | 98.3(0.1) | 99.2(0.1) | 99.2(0.1) |

Statistics values of correction (%).

Descriptor | SVM | unSVM | Bagging | unBagging | asBagging | PRIFEAB |
---|---|---|---|---|---|---|

BCUT | 98.2(0.1) | 97.9(0.1) | 97.6(0.1) | 97.8(0.1) | 68.3(0.4) | 67.3(0.2) |

CONST | 98.2(0.1) | 98.3(0.1) | 98.2(0.1) | 98.2(0.1) | 65.3(0.5) | 66.2(0.3) |

PROP | 98.1(0.1) | 97.8(0.1) | 97.6(0.1) | 97.8(0.1) | 70.3(0.3) | 70.8(0.2) |

TOPO | 97.8(0.1) | 98.2(0.0) | 97.0(0.2) | 97.5(0.1) | 67.6(0.3) | 67.7(0.3) |

Average | 98.1(0.1) | 98.1(0.1) | 97.6(0.1) | 97.8(0.1) | 67.9(0.4) | 68.0(0.3) |

Statistics ratio values of the number of features used in PRIFEAB to the total number (%).

BCUT | CONST | PROP | TOPO | Average |
---|---|---|---|---|

93.3(2.0) | 95.9(2.2) | 98.2(0.5) | 99.0(0.1) | 96.6(1.2) |

(1) unSVM obtains a slight improvement of ordinary SVM on three descriptors in terms of the AUC and BACC measures.

(2) Ordinary Bagging fails to improve single learning methods, not only Bagging but also unBagging get worse results than SVM and unSVM on the measures of AUC, BACC and sensitivity.

(3) asBagging and PRIFEAB obtain 20% better results than SVM, unSVM, Bagging and unBagging on the AUC measure. The sensitivity values of asBagging and PRIFEAB increase by beyond 50% from SVM, unSVM, Bagging and unBagging on average.

(4) PRIFEAB obtains slightly better results than asBagging on both sensitivity and specificity measures. We also observed that only few features are removed by feature selection.

(5) There are several cases, the learning machines fail in prediction and nearly all the examples are classified into negative, i.e. SVM, Bagging, unBagging on CONST and unSVM on TOPO. Only asBagging and PRIFEAB succeed in all predication.

## Discussions

The above results show that asBagging and PRIFEAB perform better than the other several methods of SVM, unSVM, Bagging and unBagging. Here we give some insights on these results:

(1) Though single SVM is not stable, and can not obtain valuable results, in this case of high skew data sets, bagging does not improve its generalization performance in terms of AUC, BACC and sensitivity. Bagging gets a high correction value, which is trivial, because few positive examples are predicted correctly. Especially, when learning machines fail in prediction on some descriptor data sets, all the labels are predicted as negative, a high value of correction is obtained as 98.15%, which is the ratio of negative sample to the whole sample.

(2) Since this is a drug discovery problem, we pay more attention to positives. AUC, BACC and sensitivity are more valuable than correction to measure a classifier. Asymmetric bagging and PRIFEAB improve the AUC values of ordinary bagging. Simultaneously, sensitivity are improved greatly, which shows asymmetric bagging is proper to solve the unbalanced drug discovery problem. Asymmetric bagging wins in two aspects, one is that it make the individual data subsets balanced, the second is that it pay more attention to the positives by leaving the positives always in the data set, which makes sensitivity is higher than ordinary bagging.

(3) PRIFEAB achieves slightly better results than asymmetric bagging does. Feature selection using prediction risk as criteria also make PRIFEAB win in two aspects, one is that embedded feature selection is dependent with the used learning machine, it will select features which benefit the generalization performance of individual classifiers, the second is that different features selected for different individual data subsets, which makes more diversity of bagging and improves their whole performance. The results improved by PRIFEAB than asymmetric are slight, we consider the reason is that few features are removed. Feature selection using prediction risk is dependent on SVM. Here, positives are few, which will hurt generalization performance of SVM, and furthermore hurt effect of feature selection.

(4) The data set used is so skew that the ratio of positives to negatives is only 0.0188, not beyond 2%, which makes SVM, unSVM and Baging, unBagging disable of prediction, they fail on four out of sixteen cases and predict almost all labels to negative, even on other twelve cases, they give low sensitivity. Analysis of high skew data set is still a difficult problem.

## Conclusion

To address the unbalanced problem of drug discovery, we propose to apply asymmetric bagging and feature selection to the modeling of QSAR of drug molecules. Asymmetrical bagging of SVM and a novel algorithm PRIFEAB are compared with ordinary bagging of support vector machines on a large drug molecular activities data set, experiments show that asymmetric bagging and feature selection can improve the prediction ability of SVM in terms of AUC and sensitivity. Since this is a drug discovery problem, the positive sample is few but important, AUC and sensitivity is more proper than correction to measure generalization performance of classifiers.

This work introduces asymmetric bagging into prediction of drug activities and furthermore extends feature selection to asymmetric bagging. This work only concerns an embedded feature selection model with the prediction risk criteria, one of the future work will try more efficient and more effective feature selection methods for this task.

## Methods

### Support vector machines

Support vector machines (SVM) [27] proposed by Vapnik and his co-workers in 1990s, have been developed quickly during the last decade [28], and successfully applied to biological data mining [29], drug discovery [6, 8] etc. Denoting the training sample as *S* = {(**x**, **y**)} ⊆ {ℝ^{
n
}× {-1, 1}}^{ℓ}, SVM discriminant hyperplane can be written as

*y* = sgn(⟨**w**·**x**⟩ + *b*)

**w**is a weight vector,

*b*is a bias. According to the generalization bound in statistical learning theory [30], we need to minimize the following objective function for a 2-norm soft margin version of SVM

in which, slack variable *ξ*_{
i
}is introduced when the problem is infeasible. The constant *C* > 0 is a penalty parameter, a larger *C* corresponds to assigning a larger penalty to errors.

*α*

_{ i }s, which make the constraint active are non zeros, we denote these points corresponding to the non zero

*α*

_{ i }s as

*support vectors*(sv). Therefore we can describe the classification hyperplane in terms of

*α*and

*b*:

**x**

_{ i }·

**x**⟩ with some function

*K*(

**x**

_{ i },

**x**), which satisfies Mercer's condition [33], then the classification function can be written as:

where *K*(**x**, **z**) is the known kernel function. A commonly used kernel function is the Guass kernel:

*K*(**x**, **z**) = exp(-||**x** - **z**||^{2}/*σ*^{2}),

which is also preferred by C.J. Lin et al. [34].

*C*in Equ.(1) is separated as

*C*

_{+}and

*C*

_{-}to adjust the penalties on the false negative vs. false positive, so Equ.(1) becomes:

where *C*_{+} is *C* and *C*_{-} is balanced_bridge**C*, balanced_bridge is a coefficient, it is assigned as the ratio of the number of positive examples to that of negative ones.

The SVM obtained by the above equation is named as unSVM. This is implemented in LibSVM [35], it had been used to analysis AMDE data and is effective for unbalanced data set [17].

### Asymmetric bagging

asBagging can solve the unstable problem of SVM classifiers and the unbalance problem in the training set.

However, it can not solve the problem of irrelevant and weak redundant features in the data sets. We will solve it by feature selection embedded in the bagging method as in the next subsection.

### PRIFEAB

where AUC (Area under ROC) [38] is predicted on the training data set, and AUC(${\overline{x}}^{i}$) is the prediction AUC on the training data set with the mean value of *i*^{
th
}feature. Finally, the feature corresponding with the smallest will be deleted, because this feature causes the smallest error and is the least important one.

### NCI AntiHIV drug screen data set

The NCI AntiHIV Drug Screen data set(NCI) is obtained from http://dtp.nci.nih.gov/docs/aids/aids_data.html. It has a categorical response measuring how a compound protects human CEM cells from HIV-1 infection. It has 29374 examples, of which 542 (1.85%) is positive and 28832 (98.15%) is negative. The structure parameters [39] consist 64 BCUT descriptors, 47 Constitutional (CONST) ones, 250 Prop ones and 266 Topological (TOPO) ones. This data set is collected and computed by Young et al., description in detail of the data set can be found in [40, 41]. Here we test the proposed learning methods on individual molecular descriptors. Since the different descriptors have different meaning and few relations with each other, they can be considered as four different data sets.

### Measures

Since the class distribution of the used data set is unbalanced, only correction of classification accuracy may be misleading. Therefore, AUC (Area Under the Curve of Receiver Operating Characteristic (ROC)) [38] is used to measure the performance.

To furthermore describe the different learning methods, we also define the various measures as below [42], where *TP*, *TN*, *FP*, *FN*, stand for the number of true positive, true negative, false positive, false negative samples at classification time, respectively.

Sensitivity is defined as $\frac{TP}{TP+FN}$ and is also known as Recall.

Specificity is defined as $\frac{TN}{TN+FP}$

BACC (Balanced Accuracy) is defined as $\frac{1}{2}\left(\frac{TP}{TP+FN}+\frac{TN}{TN+FP}\right)$, which defines the average of sensitivity and specificity.

PPV (Positive Predictive Value) is defined as $\frac{TP}{TP+FP}$ and is also known as Precision.

NPV (Negative Predictive Value) is defined as $\frac{TN}{TN+FN}$.

Correction is defined as $\frac{TP+TN}{TP+TN+FP+FN}$ and measures the overall percentage of samples correctly classified.

## Declarations

### Acknowledgements

This work was supported in part by the Natural Science Foundation of China under grant no. 20503015, the STCSM "Innovation Action Plan" Project of China under grant no. 07DZ19726, Shanghai Leading Academic Discipline Project under no. J50103 and Systems Biology Research Foundation of Shanghai University.

This article has been published as part of *BMC Bioinformatics* Volume 9 Supplement 6, 2008: Symposium of Computations in Bioinformatics and Bioscience (SCBB07). The full contents of the supplement are available online at http://www.biomedcentral.com/1471-2105/9?issue=S6.

## Authors’ Affiliations

## References

- Barrett SJ, Langdon WB: Advances in the Application of Machine Learning Techniques in Drug Discovery, Design and Development. 10th Online World Conference on Soft Computing in Industrial Applications. 2005, SpringerGoogle Scholar
- Tominaga Y: Comparative Study of Class Data Analysis with PCA-LDA, SIMCA, PLS, ANNs, and K-NN. Chemometrics and Intelligent Laboratory Systems. 1999, 49 (1): 105-115.View ArticleGoogle Scholar
- Tang K, Li T: Combining PLS with GA-GP for QSAR. Chemometrics and Intelligent Laboratory Systems. 2002, 64: 55-64.View ArticleGoogle Scholar
- Fang KT, Yin H, Liang YZ: New Approach by Kriging Models to Problems in QSAR. Journal of Chemical Information and Computer Science. 2004, 44: 2106-2113.View ArticleGoogle Scholar
- Li GZ, Yang J, Song HF, Yang SS, Lu WC, Chen NY: Semiempirical Quantum Chemical Method and Artificial Neural Networks Applied for Max Computation of Some Azo Dyes. Journal of Chemical Information and Computer Science. 2004, 44 (6): 2047-2050.View ArticleGoogle Scholar
- Xue Y, Li ZR, Yap CW, Sun LZ, Chen X, Chen YZ: Effect of Molecular Descriptor Feature Selection in Support Vector Machine Classification of Pharmacokinetic and Toxicological Properties of Chemical Agents. Journal of Chemical Information & Computer Science. 2004, 44 (5): 1630-1638.View ArticleGoogle Scholar
- Chen NY, Lu WC, Yang J, Li GZ: Support Vector Machines in Chemistry. 2004, Singapore: World Scientific Publishing CompanyView ArticleGoogle Scholar
- Bhavani S, Nagargadde A, Thawani A, Sridhar V, Chandra N: Substructure-Based Support Vector Machine Classifiers for Prediction of Adverse Effects in Diverse Classes of Drugs. Journal of Chemical Information and Modeling. 2006, 46 (6): 2478-2486.View ArticlePubMedGoogle Scholar
- Dietterich T: Machine-learning research: Four current directions. The AI Magazine. 1998, 18 (4): 97-136.Google Scholar
- Schapire R: The strength of weak learn ability. Machine learning. 1990, 5 (2): 197-227.Google Scholar
- Breiman L: Bagging predictors. Machine Learning. Machine learning. 1996, 24 (2): 123-140.Google Scholar
- Bauer E, Kohavi R: An empirical comparison of voting classification algorithms: Bagging, Boosting, and variants. Machine learning. 1999, 36 (1–2): 105-139.View ArticleGoogle Scholar
- Agrafiotis DK, no WC, Lobanov VS: On the Use of Neural Network Ensembles in QSAR and QSPR. J Chem Inf Comput Sci. 2002, 42: 903-911.View ArticlePubMedGoogle Scholar
- Lanctot JK, Putta S, Lemmen C, Greene J: Using Ensembles to Classify Compounds for Drug Discovery. J Chem Inf Comput Sci. 2003, 43: 2163-2169.View ArticlePubMedGoogle Scholar
- Guha R, Jurs PC: Development of Linear, Ensemble, and Nonlinear Models for the Prediction andInterpretation of the Biological Activity of a Set of PDGFR Inhibitors. J Chem Inf Comput Sci. 2004, 44: 2179-2189.View ArticlePubMedGoogle Scholar
- Dutta D, Guha R, Wild D, Chen T: Ensemble Feature Selection: Consistent Descriptor Subsets for Multiple QSAR Models. Journal of Chemical Information and Modeling. 2007, 47 (3): 989-997.View ArticlePubMedGoogle Scholar
- Hou T, Wang J, Li Y: ADME Evaluation in Drug Discovery. 8. The Prediction of Human Intestinal Absorption by a Support Vector Machine. J Chem Inf Model. 2007, 47: 2408-2415.View ArticlePubMedGoogle Scholar
- Tao D, Tang X, Li X, Wu X: Asymmetric Bagging and Random Subspace for Support Vector Machines-Based Relevance Feedback in Image Retrieval. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2006, 28 (7): 1088-1099.View ArticlePubMedGoogle Scholar
- Hand DJ: Construction and Assessment of Classification Rules. 1997, Chichester: John Wiley and SonsGoogle Scholar
- Yu L, Liu H: Efficient Feature Selection Via Analysis of Relevance and Redundancy. Journal of Machine Learning Research. 2004, 5 (Oct): 1205-1224.Google Scholar
- Kohavi R, George JH: Wrappers for Feature Subset Selection. Artificial Intelligence. 1997, 97: 273-324.View ArticleGoogle Scholar
- Guyon I, Elisseeff A: An Introduction to Variable and Feature Selection. Journal of machine learning research. 2003, 3: 1157-1182.Google Scholar
- Liu Y: A Comparative Study on Feature Selection Methods for Drug Discovery. J Chem Inf Comput Sci. 2004, 44 (5): 1823-1828.View ArticlePubMedGoogle Scholar
- Li H, Yap CW, Ung CY, Xue Y, Cao ZW, Chen YZ: Effect of Selection of Molecular Descriptors on the Prediction of Blood-Brain Barrier Penetrating and Nonpenetrating Agents by Statistical Learning Methods. Journal of Chemical Information and Modeling. 2005, 45 (5): 1376-1384.View ArticlePubMedGoogle Scholar
- Eitrich T, Kless A, Druska C, Meye W, Grotendorst J: Classification of Highly Unbalanced CYP450 Data of Drugs Using Cost Sensitive Machine Learning Techniques. Journal of Chemical Information and Modeling. 2007, 47 (1): 97-103.View ArticleGoogle Scholar
- Li GZ, Yang J, Liu GP, Xue L: Feature selection for multi-class problems using support vector machines. Lecture Notes on Artificial Intelligence 3173 (PRICAI2004). 2004, Springer, 292-300.Google Scholar
- Boser B, Guyon L, Vapnik V: A Training Algorithm for Optimal Margin Classifiers. Proceedings of the Fifth Annual Workshop on Computational Learning Theory. 1992, Pittsburgh: ACM, 144-152.View ArticleGoogle Scholar
- Cristianini N, Shawe-Taylor J: An Introduction to Support Vector Machines. 2000, Cambridge: Cambridge University PressGoogle Scholar
- Guyon I, Weston J, Barnhill S, Vapnik V: Gene Selection for Cancer Classification Using Support Vector Machines. Machine Learning. 2002, 46: 389-422.View ArticleGoogle Scholar
- Vapnik V: Statistical Learning Theory. 1998, New York: WileyGoogle Scholar
- Karush W: Minima of Functions of Several Variables with Inequalities as Side Constraints. Master's thesis. 1939, Deptment of Mathematics, University of ChicagoGoogle Scholar
- Kuhn HW, Tucker AW: Nonlinear Programming. Proceeding of the 2nd Berkeley Symposium on Mathematical Statistics and Probabilistic. 1951, Berkeley: University of California Press, 481-492.Google Scholar
- Mercer J: Functions of Positive and Negative Type and their Connection with the Theory of Integral Equations. Philosophy Transactions on Royal Society in London. 1909, A 209: 415-446.View ArticleGoogle Scholar
- Hsu CW, Chang CC, Lin CJ: A Practical Guide to Support Vector Classification. Tech rep. 2003, Department of Computer Science and Information Engineering of National Taiwan University, [14 August 2003], [http://www.csie.ntu.edu.tw/~cjlin/papers/guide/guide.pdf]Google Scholar
- Chang CC, Lin CJ: LIBSVM – A Library for Support Vector Machines Version 2.85. 2007, [http://www.csie.ntu.edu.tw/~cjlin/libsvm/index.html]Google Scholar
- Li GZ, Liu TY: Feature Selection for Bagging of Support Vector Machines. PRICAI2006 Lecuture Notes in Computer Science 4099. 2006, Springer, 271-277.Google Scholar
- Moody J, Utans J: Principled Architecture Selection for Neural Networks: Application to Corporate Bond Rating Prediction. Advances in Neural Information Processing Systems. Edited by: Moody JE, Hanson SJ, Lippmann RP. 1992, Morgan Kaufmann Publishers, Inc, 683-690.Google Scholar
- Duda RO, Hart PE, Stork DG: Pattern Classification. 2000, Wiley Interscience, 2Google Scholar
- Todeschini R, Consonni V: Handbook of Molecular Descriptors. 2000, Weinheim, Germany: Viley-VCHView ArticleGoogle Scholar
- Young SS, Gombar VK, Emptage MR, Cariello NF, Lambert C: Mixture Deconvolution and Analysis of Ames Mutagenicity Data. Chemometrics and Intelligent Laboratory Systems. 2002, 60: 5-11.View ArticleGoogle Scholar
- Feng J, Lurati L, Ouyang H, Robinson T, Wang Y, Yuan S, Young SS: Predictive Toxicology: Benchmarking Molecular Descriptors and Statistical Methods. Journal of Chemical Information and Computer Science. 2003, 43: 1463-1470.View ArticleGoogle Scholar
- Levner I: Feature Selection and Nearest Centroid Classification for Protein Mass Spectrometry. BMC Bioinformatics. 2005, 6 (68):Google Scholar

## Copyright

This article is published under license to BioMed Central Ltd. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.