Supervised inference of gene-regulatory networks
- Cuong C To^{1} and
- Jiri Vohradsky^{1}Email author
https://doi.org/10.1186/1471-2105-9-2
© To and Vohradsky; licensee BioMed Central Ltd. 2008
Received: 13 February 2007
Accepted: 04 January 2008
Published: 04 January 2008
Abstract
Background
Inference of protein interaction networks from various sources of data has become an important topic of both systems and computational biology. Here we present a supervised approach to identification of gene expression regulatory networks.
Results
The method is based on a kernel approach accompanied with genetic programming. As a data source, the method utilizes gene expression time series for prediction of interactions among regulatory proteins and their target genes. The performance of the method was verified using Saccharomyces cerevisiae cell cycle and DNA/RNA/protein biosynthesis gene expression data. The results were compared with independent data sources. Finally, a prediction of novel interactions within yeast gene expression circuits has been performed.
Conclusion
Results show that our algorithm gives, in most cases, results identical with the independent experiments, when compared with the YEASTRACT database. In several cases our algorithm gives predictions of novel interactions which have not been reported.
Keywords
Background
In recent years, the inference of protein interaction networks from various sources of data has become an important topic of both systems and computational biology. Protein networks can be represented as a graph with vertices formed by proteins and edges connecting two proteins representing the relationship between them. The interaction can be either direct, where two or more proteins form a functional complex, or indirect – biochemical or regulatory. The biochemical interaction can be, for example, the participation of two enzymes catalyzing two successive biochemical reactions in a pathway. Regulatory interaction represents binding of a transcription factor to a promoter site, which initiates transcription of a particular gene precursor of a protein.
The diversity of protein "interactions" implies also diverse types of data ranging from literature references and sequence database annotations, through biophysical and biochemical data to the data from microarray and proteomics experiments. The type of data predetermines also the type of interaction studied. In this paper we focus on the gene expression networks, where a regulator protein controls expression of a gene precursor of the corresponding protein.
In the last few years, several approaches to the integration of various data sources into one computational framework for inference of network structures have been reported. Particularly suitable for this approach is the concept of kernels [1]. The concept allows for transformation of various data types into kernel matrices, in which each element represents an interaction between two proteins. As an example of use of kernels in computational biology can serve the methods for the prediction of protein-protein interactions from sequences [2, 3]. The kernels can be weighted and combined according to the kernel rules, in order to integrate various diverse sources of information, which can be used to predict protein interaction networks. In 2004, Yamanishi et al.[4] introduced a supervised approach inspired by spectral clustering for inference of protein networks from multiple data sources, e.g. expression data, protein interaction data, localization data and phylogenetic profiles. They rose an important assumption that interacting proteins (in the general sense mentioned above) share similarities in the data. Based on this assumption, they created a kernel representation of the multisource data of partially known genetic network of Saccharomyces cerevisiae. This representation was used as a training set, which was projected onto a conceptual feature space where interacting proteins were grouped. Unknown interactions of the proteins of the training network with other candidate proteins were inferred using canonical correlation analysis. Thus, the authors were able to make new biological inferences about unknown regulatory interactions; they were also able to predict missing enzymes in biochemical pathways.
Another approach using kernels for protein interaction network inference, also based on a kernel matrix completion problem, was proposed by Kato et al.[5]. Missing entries, i.e. protein regulatory interactions, other than those used as a training set, were predicted according to the rules derived from the known entries of the training set. In addition, they introduced a system for estimation of the weights assigned to the individual datasets, which differentiated among the levels of influence of the different data types. The obvious drawback of any kernel method for the interaction network prediction is the limitation to the prediction of undirected interactions. However, besides others, the greatest advantage of the kernel methods is the possibility of integrating the data sources of different character into one mathematical framework.
Our algorithm focuses on identification of missing interactions in a transcriptional regulatory network. We assume that it is possible to project a protein (vertex of the network) to a feature space where the proteins connected in a network are close to each other. Projecting other proteins into the same space can reveal the unknown searched interactions. We assume that the features of an incomplete network can be extended to other, as yet unknown, interactions among proteins of the incomplete network and other proteins (see Figure 1). For the projection of the incomplete network to the conceptual feature space we use its representation by the diffusion kernel combined with kernel principal component analysis (KPCA). For the projection of proteins with unknown interactions with the sub-network, we use their gene expression profiles over time and a function representing relations within the kernel. The function is derived using genetic programming (GP).
Our method was applied to two already known transcriptional regulatory networks of the Saccharomyces cerevisiae cell cycle [6] using microarray time series data published by Spellman et al.[7]. Other possible interactions, such as protein protein interactions, or indirect biochemical interactions, are in this study ignored; therefore the results obtained here are valid exclusively for regulatory interactions in gene expression.
Results
From Eq. 15, we see that the projection onto the feature space depends deeply on the approximate functions f_{ i }created by GP. Therefore, the control parameters of GP are determined in such way that Eq. 14 is best satisfied using the fitness function defined by Eq. 16.
A list of control parameters of GP for the search of the kernel approximating function.
Population size: | 1000 |
Maximum generation: | 1000 |
Probability of crossover: | 0.90 |
Probability of reproduction: | 0.10 |
Maximum depth for tree created during run: | 10 |
Maximum depth for initial random tree: | 7 |
Terminal set: | {(x_{i 1}, x_{i 2}, ..., x_{ in }), (x_{j 1}, x_{j 2}, ..., x_{ jn })} |
Function set: | +, -, ×, pow2, pow3, ..., pow10 |
Experiments
We used the data from the database of the gene expression profiles of Saccharomyces cerevisiae [8] and two protein networks inferred by Lee et al.[6] to test the algorithm. Lee et al. in their work identified DNA protein interactions for a set of transcriptional regulators in the genome wide location analysis and inferred interactions for five functional gene groups (cell cycle, metabolism, DNA/RNA/protein biosynthesis and environmental response), and, all together, 106 genes.
The Spellman's database was collected using DNA microarrays and samples collected from growing yeast cultures synchronized by three independent methods: alpha factor arrest (18 time points), elutriation (14 time points), and arrest of a cdc 15 temperature-sensitive mutant (24 time points). The database is available at [9]. Although both of these datasets are relatively old, they have been extensively studied in the literature on regulatory networks, thus providing an excellent benchmark for model validation and comparison.
Comparison of predictions of regulatory interactions made by the algorithm presented here, results obtained from the paper of Lee et al. [6] and the data from YEASTRACT database for selected genes (cell cycle – ACE2, SKN7, SWI4, SWI5, DNA/RNA/protein synthesis -, ABF1, RAP1).
Lee et al. | This paper | YEASTRACT | |
---|---|---|---|
ACE2 | FKH2 | FKH2 | FKH2 |
GAT1 | GAT1 | GAT1 | |
NDD1 | NDD1 | - | |
MCM1 | MCM1 | MCM1 | |
SFL1 | SFL1 | SFL1 | |
YAP1 | - | - | |
SKN7 | ROX1 | ROX1 | ROX1 |
NRG1 | NRG1 | NRG1 | |
YAP1 | YAP1 | YAP1 | |
SFL1 | - | SFL1 | |
SOK2 | SOK2 | SOK2 | |
- | FKH1 | FKH1 | |
SWI4 | MBP1 | - | MBP1 |
MCM1 | MCM1 | MCM1 | |
MOT3 | MOT3 | MOT3 | |
NDD1 | - | - | |
SOK2 | SOK2 | SOK2 | |
SWI4 | SWI4 | - | |
SWI6 | SWI6 | - | |
- | CRZ1 | - | |
- | DAL82 | - | |
- | DOT6 | - | |
- | FZF1 | - | |
- | GLN3 | GLN3 | |
- | MAL33 | - | |
- | MSN2 | MSN2 | |
- | RCS1 | - | |
- | RFX1 | RFX1 | |
- | RGT1 | RGT1 | |
- | RTG3 | RTG3 | |
- | SKO1 | SKO1 | |
- | STP2 | - | |
- | THI2 | - | |
- | MDD1 | - | |
SWI5 | ASH1 | ASH1 | ASH1 |
FKH2 | FKH2 | FKH2 | |
GAT1 | - | GAT1 | |
GAT3 | GAT3 | GAT3 | |
MCM1 | MCM1 | MCM1 | |
NDD1 | NDD1 | - | |
SFL1 | - | SFL1 | |
ABF1 | IME4 | IME4 | IME4 |
FHL1 | FHL1 | FHL1 | |
MSN1 | MSN1 | MSN1 | |
DAL81 | DAL81 | DAL81 | |
PHO2 | PHO2 | PHO2 | |
PUT3 | PUT3 | PUT3 | |
STP1 | STP1 | STP1 | |
RIM101 | RIM101 | RIM101 | |
- | FZF1 | - | |
- | HAP2 | HAP2 | |
RAP1 | GAT1 | GAT1 | GAT1 |
RPH1 | RPH1 | RPH1 | |
RCS1 | RCS1 | RCS1 | |
MSN4 | MSN4 | MSN4 | |
SIP4 | SIP4 | SIP4 | |
RAP1 | RAP1 | - | |
HSF1 | HSF1 | - | |
- | SUM1 | - | |
- | RSF2 | - | |
- | HAP4 | - | |
- | GAT3 | - | |
Total | 40 | 56 | - |
Confirmed by YEASTRACT | 32 | 37 | |
Not confirmed by YEASTRACT | 8 | 19 | |
Present in Lee et al. | - | 34 |
For the cell cycle network, the interactions were inferred using the presented algorithm for the following arbitrarily chosen genes – ACE2, SKN7, SWI4, SWI5, while from the DNA/RNA/Protein biosynthesis network, two genes were chosen – ABF1 and RAP1. Table 2 shows the result of application of the algorithm. Figure 7 (see Additional file 1) shows the graph of interactions for the two networks suggested by Lee et al. and a comparison of the Lee et al. and our results for the selected genes. In the majority of cases, both methods gave similar results. However, our algorithm suggested some more interactions and did not infer some which were predicted by Lee et al. 23 interactions which were not predicted by Lee et al. were suggested by our algorithm. Out of them 8 were confirmed by comparison with YEASTRACT database. Remaining 15 are considered here as false positive. These interactions remain to be confirmed or rejected by future studies.
For the ABF1 gene, our algorithm suggested additional interactions with FZF1 and HAP2. A HAP2 interaction was also confirmed by YEASTRACT. For RAP1, new interactions with SUM1, HAP4 and GAT3 were suggested but not confirmed by YEASTRACT. On the other hand, both methods inferred self control of RAP1 which was not found in the YEASTRACT database. For the cell cycle gene ACE2, our algorithm did not find interactions with YAP1 that were suggested by Lee et al., consistently with the YEASTRACT, which also did not record any regulatory interactions between this gene and ACE2. For SKN7, in contrast with Lee et al., an interaction with SFL1 was not identified by our algorithm. An interaction of SKN7 with FKH1, found in YEASTRACT, was also suggested by our algorithm. For SWI4, 15 additional interactions were suggested by our algorithm, out them 6 were confirmed by YEASTRACT. For the other interactions, predicted by our algorithm, comparison with literature did not confirm them. The interaction with NDD1 suggested by Lee et al. was not found, in accordance with the YEASTRACT database, which also did not report any regulatory interactions between these two genes. For SWI5, our algorithm did not identify interaction with GAT1 and SFL1 reported by Lee et al. These interactions were also not found in YEASTRACT. In contrast both methods suggested an interaction between SWI5 and NDD1 which was not confirmed by YEASTRACT.
We run the algorithm with a smaller training network (Fig. 6 and Table 4 (see Additional file 1)) for each of the control networks mentioned above. Even in this case the algorithm found 40% of genes in agreement with the YEASTRACT database. Overall overlap with the results obtained with the original training network was 37%. The results suggest that – 1. That the algorithm can give acceptable results even in the case of relatively small training network (9 genes for cell cycle and 6 genes for DNA/RNA/protein synthesis networks out of 106 total genes investigated) and – 2. In order to improve the reliability of the results it is advisable to run the algorithm several times for different training networks of similar sizes and select genes which are identified in all or most of the runs. Such approach is common in any evolutionary algorithm based methods.
It can be concluded that the two results (Lee's et al. and ours) obtained by principally two different method gave in most cases identical results. Our algorithm provided suggestion of additional interactions which were not found by the Lee et al's. experiments. Suggested interactions which were not confirmed remain for future verification.
Discussion and conclusion
Gene control is a time evolving process initiated by binding of a particular regulator (or regulators) to the promoter region of the regulated gene. After that, transcription is initiated and the particular mRNA is synthesized. Recording of the changes of ideally all mRNA amounts over time therefore encodes the information about the regulatory event and, in principle, allows reverse identification of the interactions. In the literature concerning the regulatory network inference, it is frequently assumed that the dependence between protein and mRNA concentration is linear. Therefore, transcriptional control networks have been inferred from gene expression time series. Although this assumption is coarse, the protein concentrations of transcriptional regulators are difficult to measure, and the microarray data are the best available.
Here, we assume that unknown regulatory interactions of proteins sharing similar function can be deduced from a known sub-network (training network) and the gene expression time series. In order to infer the unknown interactions, the known sub-network, that is a part of a complete hypothetical network, was projected to a feature space where the interactions among the nodes of the network are easier to identify. The projection of the sub-network was made using kernels. Functions describing the relations in the sub network kernel rows were identified using genetic programming. The functions are required for projection of the unknown potential regulators to the space of the sub-network. Interactions of the proteins of the training network with the potential regulators were inferred by application of the trained algorithm to the expression profiles of the potential transcriptional regulators.
Here, we used a set of proteins whose interaction network was independently identified previously [6] and compared our results with the Lee's et al. predictions. Both methods were verified by comparison of the results with independent databases of regulatory interactions YEASTRACT. Results show that our algorithm gives, in most cases, results identical with the experiments made by Lee et al., when compared with the YEASTRACT data, our algorithm suggests additional interactions which were not found in literature. Most of the differences between our results and the results of Lee et al. are concentrated into two genes (SWI4, RAP1), when our algorithm gives predictions of interactions which are neither listed in the work of Lee et al. nor in the YEASTRACT database (if we exclude these genes agreement of our results with YEASTRACT is 92%, 88% for Lee et al.). We have made a Medline search in the attempt to find other possible literature references, but among the thousands of records, we were not able to find those confirming the predicted interactions. Here we consider them as false positives. Nonetheless, as the method proved to be very efficient in the prediction of gene control, they can also be considered as suggestions for further experimental verification. It is necessary to emphasize that the networks reconstructed using kernel methods are in general undirected.
The results of the network reconstruction generally depend on the size of the training network; the bigger it is the more reliable are the predictions. Also, as for the other evolutionary algorithms, the reliability of prediction can be increased by running the algorithm several times for different training networks.
Although this work utilizes the time series and a diffusion kernel, other data sources, such as promoter sequence similarity, literature information and others can be used to create individual kernels and combine them into a single kernel, using the kernel combination rules. This representation can be used further for training of the algorithm and for the inference of additional interactions.
In recent years, a repository of gene expression profiles recorded using either microarrays or the proteomic approach during time evolving processes, has grown rapidly, therefore, large amounts of data containing the information about the regulatory interactions controlling the given process are available. The algorithm suggested here can serve for their identification.
Methods
Kernel representation
Inference of complex regulatory networks from experimental data is a combinatorial problem which has been addressed by various optimization techniques (for review see [10, 11]). Finding the solution to this problem generally is difficult and can lead, for bigger networks, to many equivalent solutions. Kernel representation simplifies the definition of the interaction within the network to a positive definite kernel matrix, with each element proportional to the strength of the interaction between the regulator and a gene precursor of the protein.
The Eq. 6 was used to compute the kernel K_{ β }.
An example of the diffusion kernel for a simple network is given in Figure 1 (see Additional file 1).
Supervised prediction
where f(x) maps vector x into a feature space defined by the l-th solution α of CCA (for details see [4]). K (x_{ k }, x) is a kernel computed from genomic data.
Kernel PCA
The kernel PCA was originally used in spectral clustering [14]. The idea of spectral clustering is a projection of the data matrix onto a feature space where clusters could be identified using a classical clustering algorithm. The feature space is defined by eigenvectors of the kernel principal component analysis (KPCA).
The principle of KPCA is to map an input space to a different high dimensional space. Let X = {x_{1,}..., x_{ m }}, where x_{ k }∈ ℜ^{ n }, is a vectorial representation of k-th protein profile of a total of m proteins. We assume that these vectors are centered, that is. Σ x_{ k }= 0. The basic concept of KPCA is first to map the input data X into a feature space Φ via a nonlinear mapping φ (·) and then perform a linear PCA in Φ [1]. Classical PCA [15] diagonalizes the covariance matrix C byCe = λe
with eigenvectors e and eigenvalues λ. LetΦ = [φ(x_{1})|...|φ(x_{ m })]
be the image of X in the feature space. Assuming that the mapped data are centered,K = Φ^{T} Φ, with K_{ij} = φ (x_{ i })·φ(x_{ j }).
Solving Eq. 8 for K given by Eq. 10, we havee = Φα
If k(x_{i}, x) is a linear kernel, the KPCA converts to a classical PCA.
Choice of appropriate kernels allows for nonlinear mapping of X into the feature space which can be more suitable for the given task than the original space. Moreover the principal components are orthogonal and thus uncorrelated, and the first principal components carry most of the variance of the dataset, as in the classical PCA. Therefore, only the first few principal components can be used for the cluster identification.
The ultimate goal of the network identification algorithm is the identification of a complete regulatory network with all interactions. Unfortunately, we never know in advance all proteins forming the nodes of the network. On the other hand, we usually know some part of this hypothetical network. Therefore, all proteins which can play some role in transcriptional regulation have to be considered as possible members of the network. The basic idea of our algorithm is similar to the spectral clustering i.e. the projection of each protein onto a feature space where the connected proteins have similar coordinates. Conversely, the similarity of the projection of proteins with unknown connections can then be used to deduce the connections among them.
Knowing only part of the network, not all possible network members can readily be projected onto the feature space. However, the known part of the hypothetical network can be represented by the diffusion kernel and projected to the feature space using Eq. 13. Projection of other proteins to the same space can reveal their unknown connections with proteins of the known part of the network (see Figure 3).
Let TP = {x_{ i }∈ R^{ n }, i = 1..m} be a vectorial representation of m proteins forming the known part of the complete hypothetical network. Its graphical representation will be called a training graph. Let UP be a set of n proteins which possibly can be a part of the complete hypothetical network, with z_{i} ∈ R^{ n }, i = 1..n representing their expression profiles. These proteins can be either transcriptional regulators or other proteins which we expect to play a role in transcriptional control. Knowing the connections among the proteins of the training graph, we can directly calculate the diffusion kernel K. Eq 13 is then used to project the proteins of the training graph onto the feature space. Extending the features of the training graph allows the identification of connections for the proteins from UP. The extension of the network requires calculation of the kernel values for the proteins from UP. Due to the principles of the diffusion kernel, the extension of the training graph kernel to the other proteins is not readily possible. The diffusion kernel values k(x, y) can only be computed for two proteins, x and y, with already known connections. We cannot calculate directly k(x, y) for proteins with unknown connections, i.e. the proteins which are not part of the training graph. Consequently, we cannot directly project the other proteins onto the feature space and therefore we cannot identify their connections.
Diffusion kernel values of graph and approximate functions
k(x_{ i }, x_{ j }) | Diffusion kernel values | approximate functions | |||
---|---|---|---|---|---|
x_{1} | x_{2} | ... | x_{ m } | ||
x _{1} | k(x_{1}, x_{1}) | k(x_{1}, x_{2}) | ... | k(x_{1}, x_{ m }) | f_{1}(x_{1}, x_{j}) |
x _{2} | k(x_{2}, x_{1}) | k(x_{2}, x_{2}) | k(x_{2}, x_{ m }) | f_{2}(x_{2}, x_{j}) | |
... | ... | ... | ... | ... | ... |
x _{ m } | k(x_{ m }, x_{1}) | k(x_{ m }, x_{2}) | k(x_{ m }, x_{ m }) | f_{ m }(x_{ m }, x_{j}) |
f_{ i }(x_{ i }, x_{ j }) ≈ k(x_{ i }, x_{ j }) (i, j = 1..m)
for expression profile z of a protein with unknown connection to the training graph. After we have the set of functions approximating the diffusion kernel and the expression profiles z we can calculate the kernel values for the proteins with unknown connections to the known part of the network. After that, we use the KPCA to project them onto the feature space of the training graph. Thus, identification of unknown protein connections converts to an identification of a set of functions approximating the diffusion kernel and performing KPCA for the unknown proteins. The whole scheme is depicted in Figure 5 (see Additional file 1).
Prediction
Proteins of the training graph are projected into the feature space using the diffusion kernel and Eq. 13. Using Eq. 14 and 15 other proteins with unknown but probable interactions with the known part of the network are projected onto the same feature space. The prediction is then computed in the feature space. Let A = {x_{ i }∈ Φ, i = 1..j} be a set of proteins from the training graph which have known direct connections to protein x_{ k }. Let d_{max} = max {d(x_{ i }, x_{ k }), x_{ i }, x_{ k }∈ A} be the maximum distance (Euclidean) from x_{ k }to all proteins of A. If a protein with an expression profile z with unknown connection to the proteins of the training graph has a distance from x_{ k }d(z, x_{ k }) ≤ d_{max}, then z is predicted to have a direct connection to x_{ k }.
Genetic programming
Genetic programming (GP) is an extension of genetic algorithm methods which was thoroughly discussed by Koza [16]. During the training process the GP algorithm discovers relationships between the input variables of the training set using the rules (operators). These rules can contain logical relationships, mathematical operators or any other defined relationships. The rules combine variables to an output, e.g. operation a + b = c is represented by the rule "+" and input variables a and b with the output c. c can be again combined with other variables or results of other operations on variables to a tree-like sequence of operations. Training is performed to find a tree which combines the variables and operators and satisfies best the output conditions which are defined by the so called fitness function. As with any evolutionary computing methods, the optimization of the trees is done by means of operations of reproduction and crossover.
The initial population of trees is created randomly, using the variables and a predefined set of operators. The operations of crossover and reproduction, when parts of the trees are randomly combined or simply copied, create a new generation. Ability of the trees to perform a requested task is evaluated in each generation by means of the fitness function. The control parameters of GP are: the maximum number of generations, probability of crossover, probability of reproduction, maximum depth of the tree created during the run, maximum depth of the initial random tree, and the allowed set of operators (function set).
It has been shown that such a scheme leads to improvement in the fitness value with an increasing number of generations [16]. The iterations are repeated until the criteria for fitness are satisfied or a preset number of iterations is reached. The resulting specific program (tree) is then applied to perform the task coded in the program with a given set of data.
which was used during the GP function search.
Declarations
Acknowledgements
We would like to thank Josef Panek for useful discussions and critical review of the manuscript. The work was supported by grants of the Czech Science foundation No. 310/04/0804, 310/07/1009, Institutional Research Concept AV0Z50200510 and the grant of the EC Integrated Project ActinoGEN, LSHM-CT-2004-005224.
Authors’ Affiliations
References
- Scholkopf B, Smola AJ: . In Learning with kernels. Cambridge, Massachusets , The MIT Press; 2002.Google Scholar
- Wang H, Segal E, Ben-Hur A, Koller D, Brutlag DL: Identifying protein-protein interaction sites on a genome-wide scale. In Advaneces in Neural Information Processing Systems 17. Edited by: LK Saul YWLB. Cambridge, MA , MIT press; 2005:1465–1472.Google Scholar
- Ben-Hur A, Noble WS: Kernel methods for predicting protein-protein interactions. Bioinformatics 2005, 21 Suppl 1: i38–46. 10.1093/bioinformatics/bti1016View ArticlePubMedGoogle Scholar
- Yamanishi Y, Vert JP, Kanehisa M: Protein network inference from multiple genomic data: a supervised approach. Bioinformatics 2004, 20 Suppl 1: I363-I370. 10.1093/bioinformatics/bth910View ArticlePubMedGoogle Scholar
- Kato T, Tsuda K, Asai K: Selective integration of multiple biological data for supervised network inference. Bioinformatics 2005, 21(10):2488–2495. 10.1093/bioinformatics/bti339View ArticlePubMedGoogle Scholar
- Lee TI, Rinaldi NJ, Robert F, Odom DT, Bar-Joseph Z, Gerber GK, Hannett NM, Harbison CT, Thompson CM, Simon I, Zeitlinger J, Jennings EG, Murray HL, Gordon DB, Ren B, Wyrick JJ, Tagne JB, Volkert TL, Fraenkel E, Gifford DK, Young RA: Transcriptional regulatory networks in Saccharomyces cerevisiae. Science 2002, 298(5594):799–804. 10.1126/science.1075090View ArticlePubMedGoogle Scholar
- Yeast Cell Cycle Analysis Project[http://genome-www.stanford.edu/cellcycle/]
- Spellman PT, Sherlock G, Zhang MQ, Iyer VR, Anders K, Eisen MB, Brown PO, Botstein D, Futcher B: Comprehensive identification of cell cycle-regulated genes of the yeast Saccharomyces cerevisiae by microarray hybridization. Mol Biol Cell 1998, 9(12):3273–3297.PubMed CentralView ArticlePubMedGoogle Scholar
- Yeast Search for Transcriptional Regulators And Consensus Tracking[http://www.yeastract.com/]
- Bar-Joseph Z: Analyzing time series gene expression data. Bioinformatics 2004, 20(16):2493–2503. 10.1093/bioinformatics/bth283View ArticlePubMedGoogle Scholar
- de Jong H: Modeling and simulation of genetic regulatory systems: a literature review. J Comput Biol 2002, 9(1):67–103. 10.1089/10665270252833208View ArticlePubMedGoogle Scholar
- Scholkopf B, Tsuda K, Vert JP: Kernel Methods in Computational Biology. Cambridge, MA , MIT Press; 2004.Google Scholar
- Bach FR, Jordan MI: Kernel independent component analysis. J Mach Learning Res 2002, 3: 1–48. 10.1162/153244303768966085Google Scholar
- Ng A, Jordan M, Weiss Y: On spectral clustering: Analysis and an algorithm: Analysis and and algorithm. Advances in Neural Information Processing Systems 2002, 14: 849–856.Google Scholar
- Jollife I: Principal Component Analysis. New York , Springer-Verlag; 1986.View ArticleGoogle Scholar
- Koza JR: Genetic Programming: On the programing of computers by means of natural selection. MA , MIT Press; 1992.Google Scholar
Copyright
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.