# PyMix - The Python mixture package - a tool for clustering of heterogeneous biological data

- Benjamin Georgi
^{1, 4}Email author, - Ivan Gesteira Costa
^{2}and - Alexander Schliep
^{1, 3}

**11**:9

https://doi.org/10.1186/1471-2105-11-9

© Georgi et al; licensee BioMed Central Ltd. 2010

**Received: **3 August 2009

**Accepted: **6 January 2010

**Published: **6 January 2010

## Abstract

### Background

Cluster analysis is an important technique for the exploratory analysis of biological data. Such data is often high-dimensional, inherently noisy and contains outliers. This makes clustering challenging. Mixtures are versatile and powerful statistical models which perform robustly for clustering in the presence of noise and have been successfully applied in a wide range of applications.

### Results

PyMix - the Python mixture package implements algorithms and data structures for clustering with basic and advanced mixture models. The advanced models include context-specific independence mixtures, mixtures of dependence trees and semi-supervised learning. PyMix is licenced under the *GNU General Public licence* (GPL). PyMix has been successfully used for the analysis of biological sequence, complex disease and gene expression data.

### Conclusions

PyMix is a useful tool for cluster analysis of biological data. Due to the general nature of the framework, PyMix can be applied to a wide range of applications and data sets.

## Background

### Clustering and biological data

The first step in the analysis of many biological data sets is the detection of mutually similar subgroups of samples by clustering. There are a number of aspects of modern, high-throughput biological data sets which make clustering challenging: The data is often high-dimensional and only a subset of the features can be expected to be informative for the purpose of an analysis. Also, the values of the data points may be distorted by noise and the data set contains a non-negligible number of missing values. In addition, many biological data sets will include outliers due to experimental artifacts. Finally, the data set might incorporate multiple sources of data from different domains (e.g different experimental methods, geno- and phenotypic data, etc.), where the relative relevance for the biological question to be addressed, as well as potential dependencies between the different sources are unknown. In particular, the latter can lead to a clustering which captures regularities not relevant to the specific biological context under consideration. Reflecting its importance in exploratory data analysis, there is a multitude of clustering methods described in the literature (see [1, 2] for reviews). Clustering methods can be divided in two major groups: hierarchical and partitional methods. Hierarchical methods, which transform a distance matrix into a dendogram, have been widely used in bioinformatics, for example in the early gene expression literature, partly due to the appealing visualization of the dendograms [3]. Partitional methods are based on dividing samples into non-overlapping groups by the optimization of an objective function. For example, k-means is a iterative partitional algorithm that minimizes the sum of squared errors between samples and the centroids they have been assigned to [4].

A classical statistical framework for performing partitional clustering, which has attractive properties for biological data, are mixture models [5]. On the clustering level, due to their probabilistic nature, mixture models acknowledge the inherent ambiguity of any group assignment in exploratory biological data analysis, in a structured and theoretically sound way. This leads to a certain robustness towards noise in the data and makes mixtures a superior choice of model for data sets where hard partitioning is inappropriate. On the level of representing individual data points, mixtures are highly flexible and can adapt to a wide range of data sets and applications. Finally, there is a wealth of extensions to the basic mixture framework. For example semi-supervised learning or context-specific independence (see below for details).

In practice, the first big stepping stone for the analysis of any data set by clustering is the choice of model to be used. This can be burdensome, as most available packages are rather narrowly aimed at one specific type of model and re-implementation is time intensive. The PyMix software package aims to provide a general, high-level implementation of mixture models and the most important extensions in an object oriented setup (see additional file 1). This allows for rapid evaluation of modeling choices in an unified framework.

## Implementation

### Mixture models

*X*=

*X*

_{1},...,

*X*

_{ p }denote random variables (RVs) representing the features of a

*p*dimensional data set

*D*with

*N*samples

*x*

_{ i },

*i*= 1,...,

*N*where each

*x*

_{ i }consists of a realization (

*x*

_{i 1},...,

*x*

_{ ip }) of (

*X*

_{1},...,

*X*

_{ p }). A

*K*component mixture distribution is given by

*π*

_{ k }≥ 0 are the mixture coefficients with . Each of the

*K*clusters is identified with one of the components. In the most straightforward case, the component distributions

*P*(

*x*

_{ i }|

*θ*

_{ k }) are given by a product distribution over

*X*

_{1},...,

*X*

_{ p }parameterized by parameters

*θ*

_{ k }= (

*θ*

_{k 1},...,

*θ*

_{ kp }),

This is the well known naïve Bayes model (e.g. [6–8]). However, the formulation can accommodate more complex component distributions, including any multivariate distribution from the exponential family. One advantage of adopting naïve Bayes models as component disributions is that they allow the seamless integration of heterogeneous sources of data (say discrete and continuous features) in a single model. This has been made use of, for instance, for the joined analysis of gene expression and transcription factor binding data [9] or geno- and phenotype data of complex diseases [10].

*Expectation Maximization*(EM) algorithm [11]. The basic idea of the EM procedure is to use the current model parameters to obtain conditional expectations of the component memberships of the data points. These expectations in turn can be used to update the model parameters in a maximum likelihood fashion. Iteration over these two steps can be shown to converge to a local maximum in the likelihood. The central quantity of EM for mixtures is the component membership posterior, i.e. the probability that a data point

*x*

_{ i }was generated by component

*k*. By applying Bayes' rule, this posterior is given by

The final cluster assignments are then also obtained based on this posterior. Each data point is assigned to the component which generated it with the highest probability, i.e. *x*_{
i
}is assigned to *k** = argmax_{
k
}*P* (*k*|*x*_{
i
}, Θ).

### Method Extensions

PyMix includes several theoretical extensions of the basic mixture framework which can be employed to adapt the method for specific applications.

#### Context-specific independence mixtures

In the case of naïve Bayes component distributions, the component parameterization consists of a set of

*θ*

_{ kj }for each component

*k*and feature

*X*

_{ j }. This can be visualized in a matrix as shown in Figure 1a). Here each cell in the matrix represent one of the

*θ*

_{ kj }. The different values of the parameters for each feature and component express the regularities in the data which characterize and distinguish the components. The basic idea of the

*context-specific independence*(CSI) [9] extension to the mixture framework is that very often the regularities found in the data do not require a separate set of parameters for

*all*features in

*every*component. Rather there will be features where several component share a parameterization.

This leads to the CSI parameter matrix shown in Figure 1b). As before each cell in the matrix represents a set of parameters, but now several component might share a parameterization for a feature, as indicated by the cells spanning several rows. The CSI structure conveys a number of desirable properties to the model. First of all, it reduces the number of free parameters which have to be estimated from the data. This leads to more robust parameter estimates and reduces the risk of overfitting. Also, the CSI structure makes explicit which features characterize respectively, discriminate between clusters. In the example in Figure 1b), one can see that for feature *X*_{1} components (*C*_{1}, *C*_{2}) and (*C*_{4}, *C*_{5}) share characteristics and are represented by one set of parameters. On the other hand component *C*_{3} does not share its parameterization for *X*_{1}. Moreover, if components share the same group in the CSI structure for all positions, they can be merged thus reducing the number of components in the model. Therefore learning of a CSI structure can amount to an automatic reduction of the number of components as an integral part of model training. Such a CSI structure can be inferred from the data by an application of the structural EM [12] framework. In terms of the complexity of the mixture distribution, a CSI mixture can be seen as lying in between a conventional naive Bayes mixture (i.e. a CSI structure as shown in Figure 1a) and a single naive Bayes model (i.e. the structure where the parameters of all components are identified for all features). A comparison of the performance of CSI mixtures and these two extreme cases in a classical model selection setup can be found in [13].

The basic idea of the CSI formalism to fit model complexity to the data is also shared by approaches such as *variable order Markov chains* [14] or topology learning for hidden Markov models [15]. The main difference of CSI to these approaches is that CSI identifies parameters across components (i.e. clusters) and the resulting structure therefore carries information about the regularities captured by a clustering. In this CSI mixtures bear some similarity to clustering methods such as the *shrunken nearest centroid* (SNC) algorithm [16]. However the CSI structure allows for a richer, more detailed representation of the regularities characterizing a cluster.

Another important aspect is the relation of the selection of variables implicit in the CSI structure to pure feature selection methods (e.g. [17]). These methods typically make a binary decision about the relevance or irrelevance of variables for the clustering. In contrast to that, CSI allows for more fine grained models where a variable is of relevance to distinguish subsets of clusters.

#### Dependence trees

*X*

_{1},...,

*X*

_{ p }) and a map

*pa*represents the parent relationships between features, the DTree model assumes that the distribution of feature

*X*

_{ j }is conditional on the feature

*X*

_{pa(j)}. For a given tree topology defined by

*pa*, the joint distribution of a DTree is defined as [18],

*P*(·|·,

*θ*) is a conditional distribution, such as conditional Gaussians [19], and

*θ*

_{ jk }are the parameters of the conditional distribution. See Figure 2 for an example of a DTree and its distribution.

One important question is how to obtain the tree structure. For some particular applications, the structure is given by prior knowledge. In the analysis of gene expression of developmental processes for instance, the tree structure is given by the tree of cell development [20]. When the structure is unknown, the tree structure with maximum likelihood can be estimated from the data [18]. The method works by applying a maximum weight spanning tree algorithm on a fully connected, undirected graph, where vertices represent the variables and the weights of edges are equal to the mutual information between variables [18]. When the DTree models are used as components in a mixture model, a tree structure can be inferred for each component model [21]. This can be performed by applying the tree structure estimation method for each model at each EM iteration.

When conditional normals are used in Eq. 4, the DTree can be seen as a particular type of covariance matrix parameterization for a multivariate normal distribution [21]. There, the number of free parameters is linear to the number of variables, as in the case of multivariate normals with diagonal covariance matrices, while it models first-order variable dependencies. In experiments performed in [21], dependence trees compares favorably to Naive Bayes models (multivariate normal with diagonal covariance matrices) and full dependence models (multivariate normal with full covariance matrices) for finding groups of co-expressed genes and even for simulated data arising from variable dependence structures. In particular, dependence trees are not susceptible to over-fitting, which is otherwise a frequent problem in the estimation of mixture models with full diagonal matrices from sparse data.

In summary, the DTree model yields a better approximation of joint distribution in relation to the the simple Naive Bayes Model. Furthermore, the estimated structure can be useful in the indication of important dependencies between features in a cluster [21].

#### Semi-supervised learning

In classical clustering the assignments of samples to clusters is completely unknown and has to be learned from the data. This is referred to as unsupervised learning. However, in practice there is often prior information for at least a subset of samples. This is especially true for biological data, where there is often detailed expert annotation for at least some of the samples in a data set. Integrating this partial prior knowledge into the clustering can potentially greatly increase the performance of the clustering. This leads to a semi-supervised learning setup.

PyMix includes semi-supervised learning with mixtures for both hard labels as well as a formulation of prior knowledge in form of soft pairwise constraints between samples [22, 23]. In this context, in addition to the data *x*_{
i
}, we have a set of positive (and negative) constraints
∈ [0, 1] (
∈ [0, 1]), where *x*_{
i
}, *x*_{
j
}, 1 ≤ *i* <*j* ≤ *N*. The constraints indicate a preference for pair of data points to be clustered together (positive constraints), or not clustered together (negative constraints).

*Y*where the least number of constraints are violated [22] (see Figure 3). This can be achieved by redefining the posterior assignment rule of the EM (Eq. 3), as

where *r*_{
jk
}= *P* (*k*|*x*_{
j
}, Θ, *W*), *W* is the set of all positive and negative constraints and (*λ*^{+}, *λ*^{-}) are Lagrange parameters defining the penalty weights of positive and negative constraint violations.

Hard labels arise as a special case of the above, when the constraints are binary (
∈ {0, 1},
∈ {0, 1}) such that there is no overlap in the constraints, and the penalty parameters are set to high values *λ*^{+} = *λ*^{-}~∞. In this scenario, only solutions in full accordance with the constraints are possible [23].

There are several semi-supervised/constraint-based clustering methods described in the literature [24]. We choose to implement the method proposed in [22], because it can be used within the mixture model framework and supports the use of soft-constraints. Therefore, we can take simultaneous advantage of the characteristics of the probability distribution offered in Pymix and the semi-supervised framework. Moreover, the soft-constraints allow for the inclusion of our prior believes into the constraints, an important aspect in error prone biological data.

### Dependencies

PyMix is written in the *Python* programming language http://www.python.org. It includes a custom written C extension and makes use of the numpy http://numpy.scipy.org array package, the GNU Scientific library (GSL) http://www.gnu.org/software/gsl/ and matplotlib http://matplotlib.sourceforge.net plotting capabilities.

### Overall architecture

LabeledMixtureModel and ConstrainedMixtureModel represent two different approaches to semi-supervised learning. The former implements semi-supervised learning with hard labels, the latter with pairwise constraints on the data points. Full documentation of the objects hierarchy in PyMix (including the dependence trees) can be found on the PyMix website http://www.pymix.org.

Currently, the framework supports mixtures of Gaussians, discrete distributions and exponential distributions. Furthermore, the framework has an extension that allows hidden Markov models as components by the use of the GHMM library http://www.ghmm.org and is also used by the GQL tool [25]. Moreover, due to the object oriented setup, it is easily extendable and more distributions will be added in the future.

### Prior work

Due to the popularity and versatility of the mixture approach, there is a considerable number of software packages available. For instance the R package MCLUST http://www.stat.washington.edu/mclust/ implements algorithms for Gaussian mixture estimation. The MIXMOD http://www-math.univ-fcomte.fr/mixmod/ C++ package contains algorithms for conventional mixtures of Gaussian and multinomial distributions with MATLAB bindings. Another R package for conventional mixture analysis is the MIX http://icarus.math.mcmaster.ca/peter/mix/mix.html package. An example for a rather specialized package would be Mtreemix http://mtreemix.bioinf.mpi-sb.mpg.de/ which allows estimation of mixtures of mutagenic trees. In general it can be said that most of these packages focus rather narrowly on specific model types. The main advantage of PyMix is that the general, object oriented approach allows for a wide variety of mixture variants to be integrated in a single, unified framework. The different advanced models (CSI or semi-supervised) and component distributions (e.g. dependence trees) available in PyMix, make the software applicable for many applications. Also, the object orientation means that the software can be straightforwardly extended with additional model types by advanced users.

## Results and Discussion

### PyMix example session

Assume we have a data set of DNA sequences of length ten and we would like to perform clustering with a standard mixture model. The data is stored in a fasta-file format 'dataset.fa'.

After starting the Python interpreter we first import the required PyMix modules.

>>> import mixture, bioMixture

Here, mixture is the main PyMix module, bioMixture contains convenience functions for the work with DNA and protein sequences.

The next step is to read in the data from the flat file.

>>> data = bioMixture.readFastaSequences('dataset.fa')

The function readFastaSequences parses the sequences in dataset.fa and returns a new DataSet object. Now that the data is available we set up the model and perform parameter estimation using the EM algorithm.

>>> m = bioMixture.getModel(2,10)

>>> m.EM(data,40,0.1)

Parsing data set.done

Step 1: log likelihood: -130.1282276 (diff = -129.1282276)

Step 2: log likelihood: -40.1031405877 (diff = 90.0250870124)

Step 3: log likelihood: -39.1945767199 (diff = 0.908563867739)

Step 4: log likelihood: -37.8237645332 (diff = 1.37081218671)

Step 5: log likelihood: -36.2537338607 (diff = 1.5700306725)

Step 6: log likelihood: -33.9000749475 (diff = 2.35365891327)

Step 7: log likelihood: -31.9680428475 (diff = 1.93203209999)

Step 8: log likelihood: -31.6274670433 (diff = 0.340575804189)

Step 8: log likelihood: -31.6079141237 (diff = 0.0195529196039)

Convergence reached with log_p -31.6079141237 after 8 steps.

The first argument to getModel is the number of clusters (2 in the example), the second gives the length of the sequences (10 in this case). The EM function takes a DataSet as input and performs parameter estimation. The second and third argument are the maximum number of iterations and the convergence criterion respectively. The output shows that in the exmaple the EM took 8 iterations to converge. Note that in practice one would perform multiple EM runs from different starting points to avoid local maxima. This is implemented in the randMaxEM function.

Finally, we perform cluster assignment of the sequences.

>>> c = m.classify(data)

classify loglikelihood: -31.6079141237.

** Clustering **

Cluster 0, size 4

['seq24', 'seq33', 'seq34', 'seq36']

Cluster 1, size 6

['seq10', 'seq21', 'seq28', 'seq29', 'seq30', 'seq31']

Unassigend due to entropy cutoff:

[]

The classify function returns a list with the cluster label of each sequence (and prints out the clustering). This list can now be used to perform subsequent analysis'. PyMix offers various functionalities for visualization or printing of clusterings, ranking of features by relevance for the clustering or cluster validation. This and other examples for working with PyMix can be found in the *examples* directory in the PyMix release. Documentation for the PyMix library and a more detailed tutorials can be found on the PyMix website http://www.pymix.org.

### PyMix applications

PyMix has been applied on clustering problems in a variety of biological settings. In the following we give a short description of some these applications. For more details we refer to the original publications.

#### Context-specific independence mixtures

The CSI structure gives an explicit, high level overview of the regularities that characterize each cluster. This greatly facilitates the interpretation of the clustering results. For instance the CSI formalism has been made use of for the analysis of transcription factor binding sites [13]. There, the underlying biological question under investigation was whether there are transcription factors which have several, distinct patterns of binding in the known binding sites. In this study CSI mixtures were found to outperform conventional mixture models and biological evidence for factors with complex binding behavior were found.

Another application of CSI mixtures was the clustering of protein subfamily sequences with simultaneous prediction of functional residues [27]. Here, the CSI structure was made use of to find residues which differed strongly and consistently between the clusters. Combined with a ranking score for the residues, this allowed the prediction of putative functional residues. The method was evaluated with favorable results on several data sets of protein sequences.

Finally, CSI mixture were brought to bear on the elucidation of complex disease subtypes on a data set of attention deficit hyperactivity (ADHD) phenotypes [10]. The clustering and subsequent analysis of the CSI structure revealed interesting patterns of phenotype characteristics for the different clusters found.

#### Dependence trees

The dependence tree distribution allows modeling of first order dependencies between features. In experiments performed with simulated data [21], dependence trees compared favorably to naive Bayes and full dependence models for finding groups arising from variable dependence structures. In particular, dependence trees are not susceptible to over-fitting, which is otherwise a frequent problem in the estimation of mixture models from sparse data. Thus, it offers a better approximation of the joint distribution in relation to the the simple Naive Bayes Model.

By combining the methods for mixture estimation and for the inference of the DTree structure, it is possible to find DTree dependence structure specific to groups of co-regulated genes (see Figure 6). In the left cluster, genes have a over-expression patterns for T cell developmental stages (in blue); while in the right cluster, we have over-expression of B cell developmental stages (stages in orange). There, the estimated structure indicates important dependencies between features (developmental stages) in a cluster. We also carried out a comparative analysis, where mixtures of DTrees had a higher recovery of groups of genes participating in the same biological pathways than other methods, such as normal mixture models, k-means and SOM.

#### Semi-supervised learning

Semi-supervised learning can improve clustering performance for data sets where there is at least some prior knowledge on either cluster memberships or relations between samples in the data.

An example for the former has been investigated in the context of protein subfamily clustering [28]. There, the impact of differing amounts of labeled samples on the clustering of protein subfamilies is investigated. For several data sets of protein sequences the impact and practical implications of the semi-supervised setup are discussed. In [23], the impact of adding a few high quality constraints for identifying cell cycle genes in data from gene expression time courses with a mixture of hidden Markov models was demonstrated.

## Conclusions

PyMix-the Python mixture package is a powerful tool for the analysis of biological data with basic and advanced mixture models. Due to the general formulation of the framework, it can be readily adapted and extended for a wide variety of applications. This has been demonstrated in multiple publications dealing with a wide variety of biological applications.

## Availability and requirements

**Project name**: PyMix - the Python mixture package

**Project home page**: http://www.pymix.org

**Operating system(s)**: Platform independent

**Programming language**: Python, C

**Other requirements**: GSL, numpy, matplotlib

**License** GPL

## Declarations

## Authors’ Affiliations

## References

- Jain AK, Murty MN, Flynn PJ: Data clustering: a review.
*ACM Comput Surv*1999, 31(3):264–323. 10.1145/331499.331504View ArticleGoogle Scholar - Jain AK: Data clustering: 50 years beyond K-means.
*Pattern Recognition Letters*2009, in press.Google Scholar - Eisen M, Spellman P, Brown P, Botstein D: Cluster analysis and display of genome-wide expression patterns.
*Proc Natl Acad Sci USA*1998, 95: 14863–8. 10.1073/pnas.95.25.14863View ArticlePubMedPubMed CentralGoogle Scholar - McQueen J: Some methods of classification and analysis of multivariate observations.
*5th Berkeley Symposium in Mathematics, Statistics and Probability*1967, 281–297.Google Scholar - McLachlan G, Peel D:
*Finite Mixture Models*. John Wiley & Sons; 2000. full_textView ArticleGoogle Scholar - N S, Lew M, Cohen I, Garg A, TS H: Emotion Recognition Using a Cauchy Naive Bayes Classifier.
*Pattern Recognition, 2002. Proceedings. 16th International Conference on Publication Date*2002, 1: 17–20.Google Scholar - Provost J: Naive-bayes vs. rule-learning in classification of email.
*Technical report, Dept of Computer Sciences at the U of Texas at Austin*1999.Google Scholar - Schneider KM: Techniques for Improving the Performance of Naive Bayes for Text Classification.
*Sixth International Conference on Intelligent Text Processing and Computational Linguistics (CICLing-2005)*2005, 682–693.View ArticleGoogle Scholar - Barash Y, Friedman N: Context-specific Bayesian clustering for gene expression data.
*J Comput Biol*2002, 9(2):169–91. 10.1089/10665270252935403View ArticlePubMedGoogle Scholar - Georgi B, Spence M, Flodman P, Schliep A: Mixture model based group inference in fused genotype and phenotype data. In
*Studies in Classification, Data Analysis, and Knowledge Organization*. Springer; 2007.Google Scholar - Dempster A, Laird N, Rubin D: Maximum likelihood from incomplete data via the EM algorithm.
*Journal of the Royal Statistical Society, Series B*1977, 1–38.Google Scholar - Friedman N: Learning Belief Networks in the Presence of Missing Values and Hidden Variables. In
*ICML '97: Proceedings of the Fourteenth International Conference on Machine Learning*. San Francisco, CA, USA: Morgan Kaufmann Publishers Inc; 1997:125–133.Google Scholar - Georgi B, Schliep A: Context-specific Independence Mixture Modeling for Positional Weight Matrices.
*Bioinformatics*2006, 22(14):166–73. 10.1093/bioinformatics/btl249View ArticleGoogle Scholar - Buhlmann P, Wyner AJ: Variable Length Markov Chains.
*Annals of Statistics*1999, 27: 480–513. 10.1214/aos/1018031204View ArticleGoogle Scholar - Stolcke A, Omohundro SM: Best-first Model Merging for Hidden Markov Model Induction.
*Tech rep*1994.Google Scholar - Tibshirani R, Hastie T, Narasimhan B, Chu G: Diagnosis of multiple cancer types by shrunken centroids of gene expression.
*PNAS*2002, 99(10):6567–6572. 10.1073/pnas.082099299View ArticlePubMedPubMed CentralGoogle Scholar - Maugis C, Celeux G, Martin-Magniette ML: Variable selection in model-based clustering: A general variable role modeling.
*Comput Stat Data Anal*2009, 53(11):3872–3882. 10.1016/j.csda.2009.04.013View ArticleGoogle Scholar - Chow C, Liu C: Approximating discrete probability distributions with dependence trees.
*IEEE Trans Info Theory*1968, 14(3):462–467. 10.1109/TIT.1968.1054142View ArticleGoogle Scholar - Lauritzen SL, Spiegelhalter DJ: Local computations with probabilities on graphical structures and their application to expert systems.
*J Royal Statis Soc B*1988, 50: 157–224.Google Scholar - Costa IG, Roepcke S, Schliep A: Gene expression trees in lymphoid development.
*BMC Immunology*2007, 8: 25. 10.1186/1471-2172-8-25View ArticlePubMedPubMed CentralGoogle Scholar - Costa IG, Roepcke S, Hafemeister C, Schliep A: Inferring differentiation pathways from gene expression.
*Bioinformatics*2008, 24(13):i156-i164. 10.1093/bioinformatics/btn153View ArticlePubMedPubMed CentralGoogle Scholar - Lange T, Law MH, Jain AK, Buhmann JM: Learning with Constrained and Unlabelled Data.
*Computer Vision and Pattern Recognition, IEEE Computer Society Conference*2005, 1: 731–738.Google Scholar - Schliep A, Costa IG, Steinhoff C, Schönhuth A: Analyzing Gene Expression Time-Courses.
*IEEE/ACM Transactions on Computational Biology and Bioinformatics*2005, 2(3):179–193. 10.1109/TCBB.2005.31View ArticlePubMedGoogle Scholar - Chapelle O, Schoelkopf B, Zien A, (Eds):
*Semi-Supervised Learning*. MIT Press; 2006.Google Scholar - Costa IG, Schönhuth A, Schliep A: The Graphical Query Language: a tool for analysis of gene expression time-courses.
*Bioinformatics*2005, 21(10):2544–2545. 10.1093/bioinformatics/bti311View ArticlePubMedGoogle Scholar - Sandelin A, Alkema W, Engstrom P, Wasserman WW, Lenhard B: JASPAR: an open-access database for eukaryotic transcription factor binding profiles. Nucleic Acids Res 2004, (32 Database):91–94. 10.1093/nar/gkh012Google Scholar
- Georgi B, Schultz J, Schliep A: Context-Specific Independence Mixture Modelling for Protein Families. In
*Knowledge Discovery in Databases: PKDD*.*Volume 4702*. Springer Berlin/Heidelberg; 2007:79–90. full_textGoogle Scholar - Georgi B, Schultz J, Schliep A: Partially-supervised protein subclass discovery with simultaneous annotation of functional residues.
*BMC Struct Biol*2009, 9: 68. 10.1186/1472-6807-9-68View ArticlePubMedPubMed CentralGoogle Scholar - Costa IG, Krause R, Optiz L, Schliep A: Semi-supervised learning for the identification of syn-expressed genes from fused microarray and in situ image data.
*BMC Bioinformatics*2007, 8(Suppl 10):S3. 10.1186/1471-2105-8-S10-S3View ArticlePubMedPubMed CentralGoogle Scholar

## Copyright

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.