 Research
 Open Access
 Published:
Integrative disease classification based on crossplatform microarray data
BMC Bioinformatics volume 10, Article number: S25 (2009)
Abstract
Background
Disease classification has been an important application of microarray technology. However, most microarraybased classifiers can only handle data generated within the same study, since microarray data generated by different laboratories or with different platforms can not be compared directly due to systematic variations. This issue has severely limited the practical use of microarraybased disease classification.
Results
In this study, we tested the feasibility of disease classification by integrating the large amount of heterogeneous microarray datasets from the public microarray repositories. Crossplatform data compatibility is created by deriving expression logrank ratios within datasets. One may then compare vectors of logrank ratios across datasets. In addition, we systematically map textual annotations of datasets to concepts in Unified Medical Language System (UMLS), permitting quantitative analysis of the phenotype "distance" between datasets and automated construction of disease classes. We design a new classification approach named ManiSVM, which integrates Manifold data transformation with SVM learning to exploit the data properties. Using the leave one dataset out cross validation, ManiSVM achieved the overall accuracy of 70.7% (68.6% precision and 76.9% recall) with many disease classes achieving the accuracy higher than 80%.
Conclusion
Our results not only demonstrated the feasibility of the integrated disease classification approach, but also showed that the classification accuracy increases with the number of homogenous training datasets. Thus, the power of the integrative approach will increase with the continuous accumulation of microarray data in public repositories. Our study shows that automated disease diagnosis can be an important and promising application of the enormous amount of costly to generate, yet freely available, public microarray data.
Background
Microarray technology provides a revolutionary tool for understanding human diseases. Golub et al. [1] demonstrated that microarray data can be used to classify cancer, e.g. to distinguish between acute myeloid leukemia and acute lymphocytic leukemia. Since then, disease classification has been one of the primary foci of microarray research. For example, microarray technology has been applied to classify cancers as diverse as lung cancer [2], breast cancer [3], and glioma [4]. In principle, a disease classification problem can be solved with a twostep process: (1) build classifiers based on samples with known disease class labels; and (2) classify the unknown samples into known disease classes. In an ideal case, we would hope that the large amount of data generated by different laboratory on various diseases could be integrated into a diagnosis database, such that unknown samples could then be matched to the disease classes in the database. In this way, microarraybased classification could be practical and promising.
Recently, several studies have tested the feasibility of disease classification on crossplatform microarray data [5–9]. Employing different normalization methods, those studies showed promising results. However, all of those studies were based on cancer microarray data with limited scales. Moreover, in some studies, the good performance was biased by correlated training and testing data (samples from the same dataset were distributed into training and testing data) [5, 7]. In addition, the performance evaluations of current studies were mainly focused on precision without considering recall. In this study, we integrated 68 microarray datasets of diverse disease classes to perform a largescale and unbiased evaluation on the classification performance. Furthermore, we design an approach to automatically construct disease classes from microarray data, which is an important step towards automated disease classification by utilizing the enormous amount of public microarray repositories.
Our goal is that given microarray data profiling two samples, one normal condition and another disease condition, the disease condition can be classified based on the phenotype annotations of datasets in the public microarray database. To approach this problem, we need three component tools: (1) a feature vector to describe a microarray profile pair (disease vs. normal) that is comparable among microarray data generated with different platforms; (2) disease classes built from crossplatform microarray data based on their associated phenotype information; and (3) a machine learning approach capable of assigning potential phenotypes to a queried sample pair based on its similarity to profiled pairs in known disease classes.
For the first component, we derive the expression logrank ratio for each gene in each profile pair. By first deriving the expression logrank ratios between a disease and a normal profile as metainformation within the same dataset, and then comparing such ratio profiles across datasets, the results shall be comparable across datasets. Simply speaking, we compare crossdataset signals by emphasizing on differentially expressed genes, which were shown to be relatively robust to platforms or laboratory settings[10]. To complete the second component, we need to systematically annotate the experimental information associated with each microarray dataset. We followed the approach of Butte and Kohane [11] to use the disease concepts in the Unified Medical Language Systems (UMLS) [12] in order to annotate the phenotypes associated with each microarray dataset. Since a disease state is usually defined by several phenotype concepts (e.g. cancer, liver tissue, metastasis), we built disease classes by selecting microarray datasets sharing a common set of UMLS concepts. With respect to the third component, we used Support Vector Machine (SVM) [13, 14] for classification, and further developed a method named ManiSVM by integrating Manifold [15] and SVM where Manifold is employed for nonlinear dimensionality reduction to enhance the performance.
By integrating the microarray data of major platforms in the NCBI Gene Expression Omnibus (GEO) database [16], we constructed 117 classes. Using the leave one dataset out cross validation (LOOCV), ManiSVM and SVM achieved the overall accuracies of 70.7% and 58.8%, respectively. Our result not only demonstrates the feasibility of disease diagnosis by integrating heterogeneous microarray data, but also reveals that the performance of disease classification improves with the number of homogenous training datasets. Thus, the power of the integrative approach can be expected to dramatically increase with the continued accumulation of microarray data in public repositories.
Results
In this section, we will first give a brief introduction to the methods, followed by analysis results. Figure 1 illustrates the three major steps of our analysis: (1) data standardization, (2) construction of disease classes, and (3) classification.
Data standardization for crossdataset comparison
Standardization of expression data
We collected 232 human microarray datasets from three major platforms of the NCBI GEO [16]: U95, U133, and U133 plus 2.0. These platforms contain the majority of GEO human datasets. We only kept the 80 datasets containing both diseased and "normal" (or "control") conditions. The mapping between probe sets and Entrez Gene ID [17] yielded a set of 8229 genes common to all three platforms. For each gene, we calculate the average expression level for probe sets associated to this gene. Within each dataset, any combination of one disease sample and one normal sample is called a profile pair. To avoid systematic bias due to the differences in expression signals measured by different laboratories, we describe the properties of a profile pair in terms of logrank ratios. This is calculated as follows: (1) convert the gene expression values to ranks within each profile, and (2) calculate the log ratio of the two ranks for each gene (disease rank/normal rank). In total, we obtained 12,802 vectors of logrank ratios for an equivalent number of profile pairs. We then filtered out those datasets giving rise to fewer than 5 such vectors. Our final sample consists of 68 datasets and 12,767 logrankratio vectors.
Standardization of dataset annotation
To systematically categorize the phenotype information associated with each microarray dataset, we mapped the MeSH headings and the GEO dataset summary of each dataset to the UMLS concepts. Any UMLS concepts associated with only one microarray dataset were filtered out, resulting in a vocabulary of 185 disease and phenotype concepts. More details on this phase of the analysis are given in the Methods section.
Construction of disease classes
This step groups microarray datasets into disease classes. We first employed the frequent itemset mining (FIM) algorithm [18] to identify candidate disease groups sharing a common set of UMLS concepts. This effort assumes that a particular disease state is usually described by a common group of UMLS concepts (for example, all "breast cancer" datasets match the UMLS concepts "breast" and "neoplasms"). Next, within each group we measured the phenotype distance score among pairs of datasets. This is quantified by the term frequencyinverse document frequency (tfidf) [19]. Only those disease classes with a phenotype distance score whose estimated pvalue is less than 0.05 were kept for further analysis. This cut ensures that each class has a similar level of homogeneity in its associated UMLS concepts. Details of the tfidf calculation and its associated pvalue estimation are described in the Methods section.
In this manner we constructed 117 classes, comprising 68 microarray datasets. Each class contains 3 to 12 datasets. The classes covered a wide spectrum of conditions: cardiovascular/heart diseases, "bacterial infections and mycoses", neoplasms, CNS disorders, skin disorders, and metabolic diseases. Table 1 shows selected disease classes. Note that a given dataset can appear in more than one class, and that many of the classes are interrelated. For example, the disease class described by (neoplasms, "neoplasms, glandular and epithelial", and "neoplasms by histologic type") is the parent class of one characterized by (carcinoma, neoplasms, "neoplasms, glandular and epithelial", and "neoplasms by histologic type").
The datasets within each disease class naturally form a positive set for that class. However, the size of the datasets can vary widely. Large datasets may come to dominate the characteristics of their disease classes. In order to get an unbiased estimator of classification accuracy, we randomly selected 50 logrank ratio vectors from each dataset if its total number of profile pairs is greater than 50. We built the negative set by randomly sampling an equal number of vectors from datasets not in the positive set.
Classification analysis
We applied two approaches to training the disease classifiers. The first was direct application of the SVM algorithm with a linear kernel and Csupport vector classification (CSVC), using the LIBSVM package [20]. Prior to classification, we reduced the number of features by selecting those genes with significantly different logrank ratios (ttest pvalue < 0.05) between the positive and negative training sets. In the second approach, we first constructed a Laplacian matrix to represent the gene expression data [15], thereby transforming the data in a nonlinear fashion into a new and lowerdimensional manifold. We then applied SVM to the transformed data. We call the second approach ManiSVM. The major advantage in integrating graph laplacian with SVM is to transform the data via nonlinear dimension reduction into a new space, where data points close in distance shall share high phenotype similarity based on the chosen similarity metric. Such transformation enhances the separation of data points between positive and negative classes, thus the subsequent application of the linear kernel SVM to the transformed data can achieve better performance than its direct application to the original data. Details of the graph laplacian transformation are described in Methods.
We performed disease classification with SVM and ManiSVM, and evaluated performance with LOOCV (see Methods for details). We performed LOOCV by leaving out one dataset from the positive set (within a disease class) and the equal number of expression logrankratio vectors from the negative set as the testing positive and negative set, respectively. Then the remainder positive and negative set are training positive and negative set, respectively. Even though our classification unit is a single profile, we left out the entire positive dataset to avoid bias caused by replicates in the same dataset. We used the following measures to assess classification performance: precision = TP/(TP+FP); recall = TP/(TP+FN); and accuracy = (TP+TN)/(TP+TN+FP+FN), where TP stands for true positives, TN for true negatives, FP for false positives, and FN for false negatives. The accuracy is actually a way to summarize the precision and recall information. For each LOOCV procedure, we repeated 5 times by random sampling of different negative sets, and we averaged the result to assess the classification performance.
Our results showed that ManiSVM achieved the overall accuracy of 70.7% outperforming SVM (58.8%) by the default hyperplane positions. Although SVM can achieve high classification precision (89.8%), its recall is rather low (19.8%). In the contrast, ManiSVM provides more balanced performance and yielded 68.6% precision and 76.9% recall with 12% disease classes achieving the accuracy higher than 80%. By further shifting the position of hyperplane via adjusting the threshold of SVM decision value, SVM achieved the maximum accuracy of 67.5% (72.0% precision and 57.3% recall) and ManiSVM achieved the maximum accuracy of 75.6% (68.6% precision and 94.4% recall). Again, ManiSVM outperformed SVM. The endowed classification power of ManiSVM is attributed to the fact that the data points can be better separated in the Manifold space in terms of their phenotype similarity based on the chosen data similarity metric.
The performance of individual classes varies greatly. For example, ManiSVM achieved 83.4% accuracy for the disease class described by the UMLS concepts (neoplasms, neoplasms by site, and neoplasms by histologic type), but only 67.2% accuracy for the disease class described by (neoplasms, mammary neoplasms, and neoplasms by site). One reason for this difference is that the former contains 9 datasets, while the latter contains only 3 datasets. In general, a class with more datasets will be easier to classify (or more properly, will have more capacity to train a classifier). Figure 2a shows that the accuracy increases with the number of datasets in a class: as the number of datasets increases from 3 to 7, the accuracy increases from 63.7% to 73.5% and from 55.4% to 58.8% for ManiSVM and SVM respectively. This relationship highlights the advantage of integrating multiple datasets for disease classification.
Discussion
Although the correlation between dataset number and classification performance is strong (Figure 2a), outliers do exist. For example, the disease class characterized by the UMLS concepts (neoplasms, "neoplasms, glandular and epithelial", carcinoma, "head and neck neoplasms", neoplasms by site, and neoplasms by histologic type) contains only 3 datasets, but has a high classification accuracy of 82.8%. In contrast, the disease class (leukocytes, immune system diseases, and blood cells) contains 4 datasets, but is associated with a classification accuracy of 66.0%. These two disease classes differ in terms of withinclass homogeneity. The disease class with 3 datasets benefited from similar dataset annotations, with an average phenotype distance score of 0.59; while the disease class with 4 datasets had an average phenotype distance score of 0.78.
To properly compare the average phenotype distances within disease classes of different sizes (312 datasets), we estimated the statistical significance of phenotype distance scores by random sampling (see the Methods section). In general, more significant pvalues correspond to lower phenotype distance scores and higher degrees of withinclass data homogeneity. Figure 2b shows the significant negative correlation between accuracy and the pvalue of the phenotype distance score. As the pvalue of the phenotype distance score increases from 10^{6} to 10^{2}, the accuracy of the classifier decreases from 74.6% to 63.3% and from 63.7% to 54.0% for ManiSVM and SVM respectively. This analysis demonstrates conclusively that classification power increases with dataset homogeneity. Thus, integrating multiple datasets is only expected to enhance classification performance if they are sufficiently homogenous in the diseases being measured.
In the above analysis, we have assessed the homogeneity of a disease class by comparing dataset annotations that were mapped to the UMLS concepts. Our results show that such assessment generally reflects the true phenotype similarity between datasets. Despite a satisfactory overall performance, we have observed a few exceptions to this rule. For example, the disease class (digestive system disorders and epithelial cells) contained 3 datasets {GDS858, GDS1321, GDS1022} and had a fairly small average phenotype distance score 0.67 (pvalue 0.008), but proved rather difficult to classify. A further investigation of individual datasets revealed that the dataset GDS1022 actually studied lung pneumocytes after infection with Pseudomonas aeruginosa, which is not related to digestive system disorders. The reason that GDS1022 was mapped to this disease class is because its dataset summary mentioned "Pseudomonas aeruginosa causes serious respiratory infections in cystic fibrosis patients," and "cystic fibrosis" was automatically mapped to the "digestive system disorders" by the UMLS system. Thus, although the object of this study (lung pneumocytes) is not related to the digestive system, GDS1022 was nonetheless classed among digestive system disorders by the UMLS concept of "cystic fibrosis." This mapping imprecision resulted in inaccurate phenotype distance scores, and led to a low classification accuracy of 60.2% and 59.1% for this disease class by ManiSVM and SVM, respectively. On the other hand, such failures prove that the performance of our method could be further enhanced by a more advanced UMLS text mining tool. As with all automated text mining methods, however, mapping imprecision cannot be fully avoided. But with the rapid accumulation of microarray data, we should be able to minimize or bypass the influence of UMLS mapping noise by imposing stricter homogeneity requirements on candidate disease classes.
Conclusion
We have proposed a framework for microarraybased molecular diagnosis by combining public microarray repositories with the UMLS knowledge base. We respond to several challenges in integrating crossplatform microarray datasets. In particular, we addressed the issue of data compatibility by expressing the difference in two profiles as the ratio of logarithmic rankings. In addition, we systematically associated each microarray dataset with disease classes by mapping their textual annotations to UMLS concepts. The disease classes were created by comparing phenotype distance scores among pairs of datasets. Although SVM has already been considered one of the best approaches for microarraybased disease classification by several studies [21, 22], we further enhanced its power by using Manifold for nonlinear dimension reduction and data transformation. Our result has not only demonstrated the feasibility of this approach, but also highlighted the fact that classification power increases with the number and homogeneity of training datasets. This work therefore provides a solid foundation to the problem of integrating enormous amounts of microarray data, which are costly to generate yet freely available. The power of our approach will increase dramatically with the continued growth of public microarray repositories. The framework presented here will also benefit from ongoing efforts to develop more advanced UMLS text mining tools.
Methods
Disease annotation with UMLS concepts
To systematically categorize the phenotypes associated with each microarray dataset, we used the UMLS system [12, 23]. For each dataset, we identified its associated publication and downloaded its medical subject headings (MeSH) via NCBI Entrez programming utilities. The MeSH and NCBI GEO summary of a dataset were then parsed with the program MetaMap to find UMLS concepts. To reduce noise we focused on a subset of diseaserelated concepts in UMLS, including all the MeSH vocabulary and terms belonging to the semantic types: pathologic function, "injury or poisoning", anatomical abnormality, "body part, organ, or organ component", tissue, and cell. To infer higherorder links between datasets, all ancestor concepts were included.
Calculation of the phenotype distance scores and the associated pvalues
The calculation procedure is as follows:

1.
For the UMLS concept i in dataset j, we calculated the term frequency tf(i, j) = n_{i, j}/(∑_{ k }n_{k, j}), where n_{i, j}denotes the number of occurrences of UMLS concept i in dataset j. Then by definition, the value of tf_{(i, j)}indicates the level of occurrence frequency of UMLS concept i in dataset j.

2.
We calculated the inverse document frequency idf_{ i }= log(D/D_{ i }), where D denotes the total number of datasets and D_{ i }is the number of datasets containing the UMLS concept i. A smaller idf_{ i }implies a higher popularity of UMLS concept i among the collected microarray datasets.

3.
The tfidf score was defined by tfidf_{(i, j)}= tf_{(i, j)}× idf_{ i }, which adjusted the score of tf_{(i, j)}by taking into account the popularity level of the UMLS concept i. More intuitively, tfidf_{(i, j)}can be considered as a measure of specific relevance of UMLS concept i to dataset j. Let s be the number of UMLS concepts, a dataset j is then associated with a tfidf vector of dimension s, i.e., [tfidf_{(1, j)},..., tfidf_{(s, j)}].

4.
The phenotype similarity between any two datasets was estimated with the cosine between their tfidf vectors. The phenotype distance score of a candidate disease class was calculated as one minus the average phenotype similarity of any dataset pair within the class.

5.
Finally, to evaluate the significance of a phenotype distance score, we estimated its empirical pvalue by bootstrapping all of the datasets. In detail, given a disease class with k datasets, we randomly sampled k datasets from all datasets, and calculated the phenotype distance score, repeated 1,000,000 times, and generated the empirical distribution.
Graph laplacian transformation
In the following, we detail the graph laplacian transformation [14]. Given k expression logrankratio vectors x_{1}, x_{2},⋯,x_{ k }∈ ℜ^{l}where l is the number of selected genes, we assume that the first s <k vectors are in the training set with labels c_{ i }, where c_{ i }= 1 if x_{ i }is in positive set and c_{ i }= 1 otherwise. The rest {x_{s+1}, x_{s+2},⋯,x_{ k }} are in the testing set and unlabeled. The graph laplacian procedures are as follows:

1.
Constructing the adjacency matrix: (1) calculate Pearson correlation for each vector pair x_{ i }and x_{ j }; (2) define adjacency matrix W as w_{ ij }= 1 if the Pearson correlation of the vector x_{ i }and x_{ j }is greater than the threshold γ; and w_{ ij }= 0 otherwise. Here we set γ to be 0.25.

2.
Singular value decomposition (SVD): (1) build laplacian matrix L = D  W where W is the adjacency matrix defined above and D is a diagonal matrix of the same size as W satisfying D_{ ij }= ∑_{ j }w_{ ji }; (2) identify the eigenvalues and eigenvectors by solving the equation L e= λ e for λ and e, and order the obtained k eigenvalues increasingly: λ_{1} ≤ λ_{2} ≤ ⋯ ≤λ_{ k }. The p eigenvectors e_{1}, e_{2}, ⋯ e_{ p }that corresponds to the p smallest eigenvalues λ_{1}, λ_{2}, ⋯, λ_{ p }are then used to represent x_{1}, x_{2},⋯, x_{ k }in the manifold space, where $p=\underset{j=1,\mathrm{..},k}{\mathrm{arg}\mathrm{max}}\{{\displaystyle {\sum}_{i=1}^{j}{\lambda}_{i}};({\displaystyle {\sum}_{i=1}^{j}{\lambda}_{i}})/({\displaystyle {\sum}_{i=1}^{k}{\lambda}_{i}})<0.005\}$. That is that x_{ i }∈ ℜ^{l}(j = 1,...,k) is mapped to (e_{1}(i),...,e_{ p }(i)) with e_{ j }(i) being the i th component of the eigenvector e_{ j }. We note that all the eigenvalues are nonnegative since L is symmetric and positive semidefinite.
Following the notations in Belkin and Niyogi [15], we let E_{ lab }denote the k × p matrix with e_{1}, e_{2}, ⋯, e_{ p }being the column vectors of E_{ lab }(the (i, j)th entry of E_{ lab }is e_{ j }(i)). Then E_{ lab }represents the transformed data in the manifold space with a reduced dimension of p, as described above., SVM analysis is subsequently performed on these transformed data. The motivation for performing the manifold transformation comes from the following important mathematical property of E_{ lab }:
For any linear operation on E_{ lab }, say $({c}_{1}^{\ast},\mathrm{...},{c}_{k}^{\ast})=({f}_{1},\mathrm{...},{f}_{p}){E}_{lab}^{T}$ with ${c}_{i}^{\ast}$ = (f_{1},...,f_{ p })(e_{1}(i),...,e_{ p }(i))^{T}(i = 1,...,k), we have
Applying SVM with the linear kernel to E_{ lab }is essentially to perform a linear operation on E_{ lab }, such as $({f}_{1},\mathrm{...},{f}_{p}){E}_{lab}^{T}$. Then can be naturally considered as the classifiers, and so equation (1) measures the weighted differences among the objects' classification labels with w_{ ij }'s being the weights. We note that the labelling differences associated with larger w_{ ij }'s (corresponding to the pairs of objects with higher similarities) are having more weights. Hence, the smaller S is, the more likely that the objects with high similarities would have the same class label. Furthermore, from equation (3), our manifold data naturally leads to the smallest S for a given linear operation and p due to the use of the smallest eigenvalues/eigenvectors. In brief, the manifold transformation helps better distinguish between the positive and negative sets and thus further improves the classification results.
Abbreviations
 CSVC:

Csupport vector classification
 FIM:

Frequent Itemset Mining
 GEO:

Gene Expression Omnibus
 LOOCV:

Leave One dataset Out Cross Validation
 MeSH:

Medical Subject Heading
 SVD:

Singular Value Decomposition
 SVM:

Support Vector Machine
 tfidf:

term frequencyinverse document frequency
 UMLS:

Unified Medical Language Systems.
References
 1.
Golub TR, Slonim DK, Tamayo P, Huard C, Gaasenbeek M, Mesirov JP, Coller H, Loh ML, Downing JR, Caligiuri MA, et al.: Molecular classification of cancer: class discovery and class prediction by gene expression monitoring. Science 1999, 286(5439):531–537. 10.1126/science.286.5439.531
 2.
Chen HY, Yu SL, Chen CH, Chang GC, Chen CY, Yuan A, Cheng CL, Wang CH, Terng HJ, Kao SF, et al.: A fivegene signature and clinical outcome in nonsmallcell lung cancer. N Engl J Med 2007, 356(1):11–20. 10.1056/NEJMoa060096
 3.
DolledFilhart M, Ryden L, Cregger M, Jirstrom K, Harigopal M, Camp RL, Rimm DL: Classification of breast cancer using genetic algorithms and tissue microarrays. Clin Cancer Res 2006, 12(21):6459–6468. 10.1158/10780432.CCR061383
 4.
Shirahata M, IwaoKoizumi K, Saito S, Ueno N, Oda M, Hashimoto N, Takahashi JA, Kato K: Gene expressionbased molecular diagnostic system for malignant gliomas is superior to histological diagnosis. Clin Cancer Res 2007, 13(24):7341–7356. 10.1158/10780432.CCR062789
 5.
Bloom G, Yang IV, Boulware D, Kwong KY, Coppola D, Eschrich S, Quackenbush J, Yeatman TJ: Multiplatform, multisite, microarraybased human tumor classification. Am J Pathol 2004, 164(1):9–16.
 6.
Liu HC, Chen CY, Liu YT, Chu CB, Liang DC, Shih LY, Lin CJ: Crossgeneration and crosslaboratory predictions of Affymetrix microarrays by rankbased methods. J Biomed Inform 2008, 41(4):570–579. 10.1016/j.jbi.2007.11.005
 7.
Warnat P, Eils R, Brors B: Crossplatform analysis of cancer microarray data improves gene expression based classification of phenotypes. BMC Bioinformatics 2005, 6: 265. 10.1186/147121056265
 8.
Nilsson B, Andersson A, Johansson M, Fioretos T: Crossplatform classification in microarraybased leukemia diagnostics. Haematologica 2006, 91(6):821–824.
 9.
Stec J, Wang J, Coombes K, Ayers M, Hoersch S, Gold DL, Ross JS, Hess KR, Tirrell S, Linette G, et al.: Comparison of the predictive accuracy of DNA arraybased multigene classifiers across cDNA arrays and Affymetrix GeneChips. J Mol Diagn 2005, 7(3):357–367.
 10.
Rhodes DR, Yu J, Shanker K, Deshpande N, Varambally R, Ghosh D, Barrette T, Pandey A, Chinnaiyan AM: Largescale metaanalysis of cancer microarray data identifies common transcriptional profiles of neoplastic transformation and progression. Proc Natl Acad Sci USA 2004, 101(25):9309–9314. 10.1073/pnas.0401994101
 11.
Butte AJ, Kohane IS: Creation and implications of a phenomegenome network. Nat Biotechnol 2006, 24(1):55–62. 10.1038/nbt1150
 12.
Bodenreider O: The Unified Medical Language System (UMLS): integrating biomedical terminology. Nucleic Acids Res 2004, (32 Database):D267–270. 10.1093/nar/gkh061
 13.
Pavlidis P, Wapinski I, Noble WS: Support vector machine classification on the web. Bioinformatics 2004, 20(4):586–587. 10.1093/bioinformatics/btg461
 14.
Cortes C, Vapnik V: Supportvector networks. Machine Learning 1995, 20(3):273–297.
 15.
Belkin M, Niyogi P: Using manifold structure for partially labeled classification. Advances in Neural Information Processing Systems: 2003 2003.
 16.
Barrett T, Troup DB, Wilhite SE, Ledoux P, Rudnev D, Evangelista C, Kim IF, Soboleva A, Tomashevsky M, Edgar R: NCBI GEO: mining tens of millions of expression profiles – database and tools update. Nucleic Acids Res 2007, (35 Database):D760–765. 10.1093/nar/gkl887
 17.
Maglott D, Ostell J, Pruitt KD, Tatusova T: Entrez Gene: genecentered information at NCBI. Nucleic Acids Res 2007, (35 Database):D26–31. 10.1093/nar/gkl993
 18.
Grahne G, Zhu J: Efficiently Using Prefixtrees in Mining Frequent Itemsets. FIMI'03 Workshop on Frequent Itemset Mining Implementations: 2003 2003.
 19.
Ruch P, Baud R, Geissbuhler A: Evaluating and reducing the effect of data corruption when applying bag of words approaches to medical records. Int J Med Inform 2002, 67(1–3):75–83. 10.1016/S13865056(02)000576
 20.
Chang CC, Lin CJ: LIBSVM: a library for support vector machines. Software 2001. [http://www.csie.ntu.edu.tw/~cjlin/libsvm]
 21.
Li T, Zhang C, Ogihara M: A comparative study of feature selection and multiclass classification methods for tissue classification based on gene expression. Bioinformatics 2004, 20(15):2429–2437. 10.1093/bioinformatics/bth267
 22.
Ramaswamy S, Tamayo P, Rifkin R, Mukherjee S, Yeang CH, Angelo M, Ladd C, Reich M, Latulippe E, Mesirov JP, et al.: Multiclass cancer diagnosis using tumor gene expression signatures. Proc Natl Acad Sci USA 2001, 98(26):15149–15154. 10.1073/pnas.211566398
 23.
Aronson AR: Effective mapping of biomedical text to the UMLS Metathesaurus: the MetaMap program. Proc AMIA Symp 2001, 17–21.
Acknowledgements
The authors thank Xuegong Zhang for helpful discussions and suggestions. This work of CCL, JH, MK, HH, and XJZ was supported by the grants NIH R01GM074163, NIH U54CA112952, NSF 0515936, and NSF 0747475.
This article has been published as part of BMC Bioinformatics Volume 10 Supplement 1, 2009: Proceedings of The Seventh Asia Pacific Bioinformatics Conference (APBC) 2009. The full contents of the supplement are available online at http://www.biomedcentral.com/14712105/10?issue=S1
Author information
Affiliations
Corresponding authors
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors' contributions
CCL, JH, HH, and XJZ designed the study. CCL, JH, and MK performed the study. CCL, HH, and XJZ analyzed the result. CCL, HH, and XJZ wrote the paper.
Rights and permissions
Open Access This article is published under license to BioMed Central Ltd. This is an Open Access article is distributed under the terms of the Creative Commons Attribution License ( https://creativecommons.org/licenses/by/2.0 ), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
About this article
Cite this article
Liu, CC., Hu, J., Kalakrishnan, M. et al. Integrative disease classification based on crossplatform microarray data. BMC Bioinformatics 10, S25 (2009). https://doi.org/10.1186/1471210510S1S25
Published:
Keywords
 Support Vector Machine
 Disease Classification
 Microarray Dataset
 Unify Medical Language System
 Disease Class