 Methodology article
 Open Access
 Published:
Inferring the role of transcription factors in regulatory networks
BMC Bioinformatics volume 9, Article number: 228 (2008)
Abstract
Background
Expression profiles obtained from multiple perturbation experiments are increasingly used to reconstruct transcriptional regulatory networks, from well studied, simple organisms up to higher eukaryotes. Admittedly, a key ingredient in developing a reconstruction method is its ability to integrate heterogeneous sources of information, as well as to comply with practical observability issues: measurements can be scarce or noisy. In this work, we show how to combine a network of genetic regulations with a set of expression profiles, in order to infer the functional effect of the regulations, as inducer or repressor. Our approach is based on a consistency rule between a network and the signs of variation given by expression arrays.
Results
We evaluate our approach in several settings of increasing complexity. First, we generate artificial expression data on a transcriptional network of E. coli extracted from the literature (1529 nodes and 3802 edges), and we estimate that 30% of the regulations can be annotated with about 30 profiles. We additionally prove that at most 40.8% of the network can be inferred using our approach. Second, we use this network in order to validate the predictions obtained with a compendium of real expression profiles. We describe a filtering algorithm that generates particularly reliable predictions. Finally, we apply our inference approach to S. cerevisiae transcriptional network (2419 nodes and 4344 interactions), by combining ChIPchip data and 15 expression profiles. We are able to detect and isolate inconsistencies between the expression profiles and a significant portion of the model (15% of all the interactions). In addition, we report predictions for 14.5% of all interactions.
Conclusion
Our approach does not require accurate expression levels nor times series. Nevertheless, we show on both data, real and artificial, that a relatively small number of perturbation experiments are enough to determine a significant portion of regulatory effects. This is a key practical asset compared to statistical methods for network reconstruction. We demonstrate that our approach is able to provide accurate predictions, even when the network is incomplete and the data is noisy.
Background
A central problem in molecular genetics is to understand the transcriptional regulation of gene expression. A transcription factor (TF) is a protein that binds to a typical domain on the DNA and influences transcription. The effect of this TF can be either a repression or an activation of transcription depending on the type of binding site, the distance to coding regions, or on the presence of other molecules. Finding which gene is controlled by which TF is a reverse engineering problem, usually named network reconstruction. This question has been approached over the past years by various groups.
A first approach to achieve this task is to collect the information spread in the primary literature. Following this idea, a large number of databases that take protein and regulatory interactions from the literature and curate them have been developed [1–5]. For the bacteria E. coli, RegulonDB is a dedicated database that contains experimentally verified regulatory interactions [6]. For the budding yeast (S. cerevisiae), the Yeast Proteome Database contains a large amount of regulatory information [7]. In this latter case, however, the amount of available information is not sufficient to build a reasonably accurate model of transcriptional regulation. Databases with regulatory knowledge extracted from the literature are, nevertheless, an unavoidable starting point for network reconstruction.
The alternative to a literaturecurated approach is a datadriven approach. This approach is supported by the availability of highthroughput experimental data including microarray expression analysis of deletion mutants (simple or more rarely double nonlethal knockouts), over expression of TFencoding genes, proteinprotein interactions, protein localisation, or ChIPchip experiments coupled with promoter sequence analysis. We may cite several classes of methods that use these kinds of data, such as correlation, mutual information or causality studies, Bayesian networks, path analysis, informationtheoretic approaches, and ordinary differential equations [8–10].
In short, most available approaches so far are based on a probabilistic framework which defines a probability distribution over the set of models. The reconstructed network is then defined as the most likely model given the data. Such an optimization problem is usually non convex, and finding a global optimum cannot be guaranteed in practice. Existing algorithms report a local optimum which should be interpreted with care: errors can appear and no consensual model may be produced.
As an illustration, special attention has been paid to the reconstruction of S. cerevisiae network from ChIPchip data and proteinprotein interaction networks [11]. A first regulatory network was obtained with promoter sequence analysis methods [12, 13], yet, some undetected transcriptional regulatory motifs were proposed using nonparametric causality tests [14]. Moreover, Bayesian analysis also identified new regulatory modules for this network [15, 16]. Thus, the results obtained with the different methods do not coincide and a fully datadriven search is in general subject to overfitting and not fully reliable [17].
In regulatory networks an important and nontrivial physiological information is the regulatory role of TFs as inducer or repressor, also called the sign of the interaction. This information is needed if one wants to know, for instance, the physiological effect of a change caused by external conditions or the effect of a perturbation on the TF. While this can be achieved for one gene at a time with (long and expensive) dedicated experiments, probabilistic methods such as Bayesian models [18] or path analysis [19, 20] are capable of proposing models from highthroughput experimental data. However, as for the network reconstruction task, these methods are based on optimization algorithms that compute an optimal solution with respect to an interaction model.
In this paper, we apply formal methods to compute the sign of interactions in networks that have an available topology. By doing so, we also validate the topology of the network. Roughly, we use expression profiles to constrain the possible regulatory roles of TFs, and we report those regulations that are assigned the same role in all feasible models. Thus, we overapproximate the set of feasible models, and then look for invariants in this set. A similar idea was applied in [21] to check the consistency of gene expression assays. However, we use a deeper formalisation and stronger algorithmic methods to achieve the inference task.
Different sources of largescale data are exploited in this study: gene expression arrays, which provide information on the interaction signs; and ChIPchip experiments, which provide the topology of the regulatory network when not available.
The main tasks we address are the following:

1.
Building a formal model of regulation for a set of genes that integrates information from ChIPchip data, sequence analysis, and literature annotations.

2.
Checking its consistency with expression profiles on perturbation assays.

3.
Inferring the regulatory role of TFs as inducer or repressor if the model is consistent with expression profiles.

4.
Isolating ambiguous pieces of information if it is not.
The Results section is organised as follows. We first introduce the mathematical framework which is used to define and to test the consistency between expression profiles and transcriptional networks. Then, we apply our algorithms to address three main issues:

Analysis of the dependence between the number of available observations and the number of inferred regulations. In the case where all genes are observed, we prove that at most 40.8% of E. coli network can be inferred and that 30 perturbation experiments are enough to infer 30% of the network on average. In the case of missing observations, we estimate how the proportion of unobserved genes affects the number of inferred regulations.

Illustration and validation of our method on the transcriptional network of E. coli, obtained from RegulonDB [6], with a compendium of expression profiles [9, 22].

Execution of our inference algorithms over the S. cerevisiae transcriptional network. We inferred, for small scale subnetworks, more than 20% of the roles of regulations. For more complex networks, we detected and isolated inconsistencies (ambiguities) between expression profiles and a significant part of the model (15% of all the interactions).
Results
Detecting the role of a regulation and validating a model
Our goal is to determine the regulatory role of a TF on its target genes by using expression profiles. Let us illustrate our purpose with a simple example.
We suppose that we are given the topology of a network (this topology can be obtained from ChIPchip data or any computational network inference method). In this network, let us consider a node A with a single predecessor. In other words, the model tells us that the protein B acts on the expression of the gene coding for A and no other protein acts on A.
Independently, we suppose that we have several gene expression arrays at our disposal. One of these arrays indicates that A and B simultaneously increase during a steady state shift experiment. Then, common sense tells us that B must have been an activator of A during the experiment. More precisely, protein B cannot have inhibited gene A since they both have increased. Consequently, we say that the model predicts the sign of the interaction from B to A as positive (see Fig. 1).
This naive rule is actually used in a large class of models; we will call it the naive inference rule. When several expression profiles are available, the predictions of the different profiles can be compared. If two expression profiles predict different signs for a given interaction, there is an ambiguity or inconsistency between data and model (see Fig. 2). Then, the ambiguity of the regulatory role can be attributed to three factors: (1) a complex mechanism of regulation, the role of the interaction depends on the state of the system; (2) a missing interaction in the model; (3) an error in the experimental source. This simple strategy is implemented in the Algorithm 1.
Let us consider now the case when A is activated by two proteins B and C. No more natural deduction can be done when A and B increase during an experiment since the influence of C must be taken into account. A model of interactions between A, B, and C has to be proposed. Probabilistic methods estimate the most probable signs of regulations that fit with the theoretical model [18, 23].
Our point of view is different; we introduce a basic rule that shall be checked by each interaction in the model. This rule tells us that any variation of A must be explained by the variation of at least one of its predecessors. In previous papers, we introduced a formal framework to justify this basic rule under some reasonable assumptions. We also tested the consistency between expression profiles and a graphical model of cellular interactions. This formalism will be introduced here in an informal way; its full justification and extensions can be found in the references [24–27].
In our example, the basic rule means that if B and C activate A, and both (B and C) are known to decrease during a steady state experiment, A cannot be observed as increasing. Then A is predicted to decrease (see Fig. 3). More generally, we apply the rule as a constraint for the model, we write constraints for all the nodes of the model, and we use several approaches in order to solve the system of constraints. From the study of the set of solutions, we deduce which signs are surely determined by these rules. Then, we obtain necessary conditions on the signs instead of the most probable signs given by probabilistic methods.
A formal approach
Consider a system of n chemical species {1,...,n}. These species interact with each other and we model these interactions using an interaction graph G = (V, E). The set of nodes is denoted by V = {1,...,n}. There is an edge j → i ∈ E if the level of species j influences the production rate of species i. Edges are labelled by a sign {+, } which indicates whether j activates or represses the production of i.
In a typical stress perturbation experiment a system leaves an initial steady state following a change in control parameters. After waiting long enough, the system may reach a new steady state. In genetic perturbation experiments, a gene of the cell is either knockedout or overexpressed; perturbed cells are then compared to the reference. Our approach relies on the signs of the variations in expression or activity of the species in the network. Let us denote by sign(X_{ i }) ∈ {+, , 0} the sign of the variation of species i during a given perturbation experiment, and by sign(j → i) ∈ {+, } the sign of the edge j → i in the interaction graph.
Let us fix species i such that there is no positive selfregulating action on i. For every predecessor j of i, sign(j → i) * sign(X_{ j }) provides the sign of the influence of j on the species i. Then, we can write a constraint on the variation to interpret the rule that was previously stated: the variation of species i is explained by the variation of at least one of its predecessors in the graph.
When the experiment is a genetic perturbation, the same equation holds for every node that was not genetically perturbed during the experiment and such that all its predecessors were not genetically perturbed. If a predecessor X_{ M }of the node was knockedout, the equation becomes
The same holds with +sign(M → i) when the predecessor X_{ M }was overexpressed. There is no equation for the genetically perturbed node.
The sign algebra is the suitable framework for reading these equations [26]. It is defined as the set {+, , ?, 0}, provided with a sign consistency relation ≈, and arithmetic operations + and ×. The following tables describe this algebra:
For a given interaction graph G, we will refer to the qualitative system associated with G as the set made up by applying constraint (1) for each node in G. We say that node variations X_{ i }∈ {+, , 0} are consistent with the graph G when they satisfy all the constraints associated with G using the sign consistency relation ≈.
With this material at hand, let us come back to our original problem, namely to infer the regulatory role of TFs from the combination of heterogeneous data. In the following we assume that:

The interaction graph is either given by a model to be validated, or built from ChIPchip data and TF binding site search in promoter sequences. Thus, as soon as a TF j binds to the promoter sequence of gene i, j is assumed to regulate i. This is represented by an arrow j → i in the interaction graph.

The regulatory role of a TF j on a gene i (as inducer or repressor) is represented by the variable S_{ ji }, which is constrained by Eqs. (1) or (2).

Expression profiles provide the sign of variation of the gene expression for a set of r steadystate perturbation, mutant, or overexpression experiments. In the following, ${x}_{i}^{k}$ will stand for the sign of the observed variation of gene i in experiment k.
Our inference problem can now be stated as finding values in {+, } for S_{ ji }, subject to the constraints:
Most of the time, this inference problem has a huge number of solutions. However, some variables S_{ ji }may be assigned the same value in all solutions of the system. Then, the recurrent value assigned to S_{ ji }is a logical consequence of the constraints (3), and a prediction of the model. We will refer to these inferred interaction signs as predictions of the qualitative system, that is, sign variables S_{ ji }that have the same value in all solutions of a qualitative system (3). When the inference problem has no solution, we say that the model and the data are inconsistent or ambiguous.
Let us illustrate this formulation with a very simple (yet informative) example. Suppose that we have a system of three genes A, B, C, where B and C influence A, as given in Fig. 4. Let us say that for this interaction graph we obtained six experiments, and in each of them the variation of all products in the graph was observed. Using some or all of the experiments provided will lead us to different qualitative systems, as shown in Table 1, hence to different inference results.
Algorithmic procedure
When the signs on edges of the interaction graph are known (i.e. fixed values of S_{ ji }), finding consistent node variations X_{ i }is a NPcomplete problem [26]. When the node variations are known (i.e. fixed values of X_{ i }), finding the signs of edges S_{ ji }from X_{ i }can be proven NPcomplete in a very similar way. However, we have been able to design algorithms that perform efficiently on a wide class of regulatory networks. These algorithms predict signs of the edges when the network topology and the expression profiles are consistent. In case of inconsistency, though, they identify ambiguous motifs and propose predictions on parts of the network that are not concerned with ambiguities.
The general process flow is as follows (see the Methods section for details):
Step 1 Sign Inference
Divide the graph into motifs (each node with its predecessors). For each motif, find sign valuations (see Algorithm 1 in the Appendix section) that are consistent with all expression profiles. If there are no solutions, call the motif Multiple Behaviours Module (MBM) and remove it from the network.
Solve again the remaining equations and determine the edge signs that are fixed to the same value in all the solutions. These fixed signs are called predicted edge and represent our predictions.
Step 2 Global test/correction of the inferred signs
Solutions at the previous step are not guaranteed to be global. Indeed, two node motifs at step 1 can be consistent separately, but not altogether (with respect to all expression profiles). This step checks global consistency by solving the equations for each expression profile. New Multiple Behaviours Modules can be found and removed from the system.
Step 3 Extending the original set of observations
Once all conflicts have been removed, we get a set of solutions in which signs are assessed to both nodes and edges. Predicted nodes, representing inferred gene variations can be found in the same way as we did for edges. We add the new variations to the set of observations and return to step 1. The algorithm is iterated until no new signs are inferred.
Step 4 Filtering predictions
In the inconsistent case, the validity of the predictions depends on the accuracy of the model and on the correct identification of the MBMs. The model can be incomplete (missing interactions), and MBMs are not always identifiable in a unique way. Thus, it is useful to sort predictions according to their reliability. Our filtering parameter is a positive integer k representing the number of different experiments with which the predicted sign is consistent. For a filtering value k, all the predictions that are consistent with less than k profiles are rejected.
The inference process then generates three results:

1.
A set of MBMs, containing interactions whose role was unclear and generated inconsistencies. We have identified several types of MBMs:

Modules of Type I: are composed of several direct regulations towards the same gene. They are detected in the Step 1 of the algorithm, and most of them are composed of only one edge like illustrated in Fig. 5, but bigger examples exist.

Modules of Type II, III, IV: are detected in Steps 2 or 3, hence they contain either direct regulations coming from the same protein or indirect regulations and/or loops. Each of these regulations represents a consensus of all the experiments, but when we attempt to assess them globally, they lead to contradictions. The indices IIIV have no topological meaning, they label the most frequent situations and are illustrated in Fig. 5.

2.
A set of inferred signs, meaning that the expression profiles fix the signs of certain interactions in a unique way.

3.
A reliability ranking of inferred signs. The filtering parameter k used for ranking is the number of different expression profiles that validate a given sign.
On a computational level, the division between Step 1 (which considers each small motif with all profiles together) and Step 2 (which considers the whole network with each profile separately) is necessary to overcome the memory complexity of the search for solutions. To handle large scale systems we combine decision diagrams and constraint solvers (see details in the Methods section).
Since our basic rule is a weak constraint, we expect it to produce very robust predictions. On the other hand, there are theoretical limits to this approach. For certain interaction graphs, not a single sign may be inferred even with a high number of experiments. In the next paragraphs, we comment on the maximum number of signs that can be inferred from a given graph.
In perturbation experiments, gene responses are observed following changes of external conditions (temperature, nutritional stress, etc.), gene inactivations, knockouts, or overexpression. When one expression profile is available for all the genes in the network we say that we have a complete profile, otherwise the profile is partial (data is missing).
In the following pragraphs we describe the results we obtained. First of all, in order to validate our formal approach, we evaluated the percentage of the E. coli network recovered from a reasonable number of artificial randomly generated perturbation experiments. Secondly, we combined real perturbation experiments with the E. coli network and computed the percentage of the recovered network. Finally, we performed the same previous analysis in a real setting of the S. cerevisiae network obtained from ChIPchip data.
On a computational level, we checked that our algorithms were able to handle large scale data, as produced by highthroughput measurement techniques (expression arrays, ChIPchip data). This is demonstrated in the following by considering networks of thousands of genes.
Stress perturbation experiments: how many do you need?
For any given network topology, even when considering all possible experimental profiles, there are signs that cannot be determined (see Table 1). Sign inference has thus a theoretical limit, referred to here as theoretical percentage of recovered signs, that is unique for a given network topology. If only some perturbation experiments are available, and/or data is missing, the percentage of inferred signs will be lower. For a given number N of available expression profiles, the average percentage of recovered signs is defined over all sets of N different expression profiles consistent with the qualitative constraints Eqs. (1) and (2).
In order to calculate the theoretical and the average percentages of recovered signs for the transcriptional network of E. coli, we modelled the network as an interaction graph using the public database RegulonDB [6]. For each transcriptional regulation A → B we added the corresponding arrow between genes A and B in the interaction graph. This graph will be referred to as the unsigned interaction graph.
From the unsigned interaction graph of E. coli, we build the signed interaction graph by annotating the edges with a sign. Most of the time, the regulatory role of a TF is available in RegulonDB, however, when it is unknown or depends on the TF level, we arbitrarily choose the value + for this regulation. This provides a graph with 1529 nodes and 3802 edges, all signed edges. The signed interaction graph is used to generate complete expression profiles that simulate the effect of perturbations. More precisely, a perturbation experiment is represented by a set of gene expression variations {X_{ i }}_{i = 1,...,n}that are not entirely random, for they are constrained by Eqs.(1) and (2). Then, we forget the signs of the network edges and compute the qualitative system with the signs of regulations as unknown.
The theoretical maximum percentage of inference is given by the number of signs that can be recovered assuming that complete expression profiles of all conceivable perturbation experiments are available. We computed this maximum percentage using constraint solvers (see Algorithm 2 in the Appendix section). We found that at most 40.8% of the signs in the network can be inferred, corresponding to M_{ max }= 1551 edges.
However, this maximum can be obtained only if all conceivable (more than 2^{50}) perturbation experiments are done, which is in practice not possible. We performed computations to understand the influence of the number of experiments (N) on the inference. For each value of N (from 5 to 200), we generated 100 sets of N complete random expression profiles and performed our algorithm for each set. Then, the percentage of inference was calculated as a function of N. The resulting statistics are shown in Fig. 6.
We can obtain a theoretical formula explaining the saturation aspect of the curve in Fig. 6. Let us suppose that the network contains M_{1} single incoming regulations. These can be inferred with probability one from only one experiment, using the naive algorithm (see Algorithm 1). Let us suppose a second category of interactions, whose signs are inferred with probability p (0 <p < 1) on average, per experiment. This implies that the average number of inferred signs for one experiment is M(1) = M_{1} + pM_{2}, where M_{2} is the number of interactions in the second category. Supposing now that inference failures are independent for different experiments, we obtain the average number of inferred signs for N experiments: M(N) = M_{1} + M_{2}(1  (1  p)^{N}). In general, we have M_{1} + M_{2} <E (E is the total number of edges), meaning that there are edges whose signs cannot be inferred.
In our example, the value M_{1} = 609 corresponds to the average number of signs inferred by the naive algorithm. Surprisingly, by using our method we can significantly improve the naive inference with little effort. For the whole E. coli network it appears that a few expression profiles are enough to infer a significant percentage of the network. More precisely, 30 different expression profiles may be enough to infer one third of the network (1267 regulatory roles). Adding more expression profiles continuously increases the percentage of inferred signs. For N > 100 we are practically on the plateau close to 37.3% (this corresponds to M = 1420 signed regulations).
According to our estimates the position of the plateau is M = M_{1} + M_{2} = 1420, which is smaller than the theoretical maximum M <M_{ max }. The difference, although negligible in practice (to obtain M_{ max }one has to perform N > 2^{50} experiments), suggests that the plateau has a very weak slope. This means that contributions of different experiments to sign inference are weakly dependent.
The values of M_{1}, M_{2}, p estimate the efficiency of our method: large p,M_{1},M_{2} mean small number of expression profiles needed for inference.
Inferring the core of the network
Obviously, not all interactions play the same role in the network. The core is a subnetwork that naturally appears for computational purposes and plays an important role in the system. It consists of all oriented loops and of all oriented chains leading to loops. All oriented chains leaving the core without returning are discarded when reducing the network to its core. Acyclic graphs and in particular trees have no core. The main property of the core is that if a system of qualitative equations has no solution, neither has the reduced system built from its core. Hence it corresponds to the most difficult part of the constraints to solve. It is obtained by reduction techniques that are very similar to those used in [28] (see details in the Methods section). As an example, the core of E. coli network (shown in Fig. 7) only has 28 nodes and 57 edges.
In the previous section, we applied the same inference process to this graph. Not surprisingly, we noticed a rather different behaviour when inferring signs on a core graph than on a whole graph as demonstrated in Fig. 6. In the former case, we needed many more experiments for the inference since the sets of expression profiles contained from N = 50 to 2000 random profiles.
Two observations may be concluded. First, a greater number of experiments is required to reach a comparable percentage of inference; the value of p is smaller than for the whole network. This confirms that the core is more difficult to infer than the rest of the network. Second, Fig. 6 displays a much less continuous behaviour for the core. More precisely, when using the core, different perturbation experiments have a strongly variable impact on sign inference. For instance, the experimental maximum percentage of inference (27 signs over 58) can be obtained already from about 400 expression profiles, yet, most of the datasets with 400 profiles infer only 22 signs.
This suggests that not only the core of the network is more difficult to infer, but also that a brute force approach (multiplying the number of experiments) may fail as well. This situation encourages us to apply experiment design and planning, that is, computational methods to minimise the number of perturbation experiments while inferring a maximal number of regulatory roles.
This also illustrates why our approach is complementary to dynamical modelling. In the case of large scale networks, when an interaction stands outside the core of the graph, an inference approach is suitable for inferring the sign of the interaction. However, when an interaction belongs to the core of the network, more complex behaviours occur (e.g. influences that depend on activation thresholds) thus, a precise modelling of the dynamical behaviour of this part of the network should be performed [29].
Influence of missing data
In the previous paragraphs, we assumed that all products in the network were observed. That is, in each experiment each node is assigned a value in {+, 0, }. However, in real measurement devices, such as expression profiles, a part of the values is discarded due to technical reasons. A practical method for network inference should cope with missing data.
We studied the impact of missing values on the percentage of inference. For this, we have considered a fixed number of expression profiles (N = 30 for the whole E. coli network, N = 30 and N = 200 for its core). Then, we have randomly discarded a growing percentage of observed products in the profiles, and computed the percentage of inferred regulations. The resulting statistics are shown in Fig. 8.
In both cases (whole network and core), the dependency between the average percentage of inference and the percentage of missing values is qualitatively linear. Simple arguments allow us to find an analytic dependency. If not observing one node of the network implies losing information on d interaction signs, we are able to obtain the following linear dependency M_{ i }= ${M}_{i}^{max}$  d * f * M_{ total }; where ${M}_{i}^{max}$ is the number of inferred interactions for complete expression profiles (no missing values), f is the fraction of unobserved nodes, and M_{ total }is the total number of nodes. In order to keep M_{ i }non negative, d must decrease with f. Our numerical results imply that the constancy of d and the linearity of the above dependency extend to rather large values of f. This indicates that our qualitative inference method is robust enough for practical use. For the whole network we estimated d = 0.35, meaning that on average we lose one interaction sign for about 2.9 missing values. However, for the same number of expression profiles, the core of the network is more sensitive to missing data (the value of d is larger, it corresponds to losing one sign for about 2.3 missing values). For the core, increasing the number of expression profiles increases d and hence the sensitivity to missing data.
Application to E. coli network with a real compendium of expression profiles
We validated our method on the transcriptional E. coli network using the compendium of expression profiles publicly available in [9] and [22]. This time the network was composed of 1418 nodes and 2888 edges. The difference with the previous model are the sigmafactors – gene interactions.
Several profiles were available, including a reference condition. We grouped together the different profiles corresponding to the same experiment; for each gene we calculated its average variation in the group of profiles. When profiles were time series, we considered that each time series ends with steady state and we used the last state in the time series. Then, we sorted the measured genes in four classes: 2fold upregulated, 2fold downregulated, nonobserved, and zero variation; this last class corresponds to non significantly (2fold) expressed genes. Only the first two classes were used in the algorithm. Therefore, there will be missing data: for some edges, neither the input nor the output are observed. Altogether, we have processed 226 sets of expression profiles corresponding to 61 different experiments (overexpression, genedeletion, and stress perturbation). We verified, for all the experiments, that they correspond to the comparison between one perturbed condition against a control condition with identical levels in all chemical components except for the one altered in the perturbed condition.
We applied our inference algorithm twice: the first time we used the signed network in a preprocessing step, in order to clean the expression data. It appears that the signed network is consistent with only 31 of the 61 selected experiments. After discarding the inconsistent motifs from each experiment (deleting observations that caused conflicts), we stayed with 61 experiments which only contained the data consistent with the signed network. In these 61 experiments, on average 12.62% of the network nodes were observed. When summing up all the observations, we obtained that 6.5% (190) of the edges (input and output) were observed in at least one expression profile; these represent the maximal set of signs that can be inferred at Steps 1 and 2 of our inference algorithm. In order to test our algorithm we wiped out the information on edge signs and then tried to recover it. Since the profiles and network were consistent, our algorithm found no ambiguity and predicted 38 signs, i.e. 20% of the edges observed at least once (input and output). The naive inference algorithm inferred 31 signs. Hence, 18% of the total of our predictions could not be obtained by the naive algorithm.
Afterwards, we tested our algorithm with the full set of observations, no data being discarded. Conflicts appeared and we filtered our inference with different parameters on the full set of 61 experiments including inconsistencies. This time 12.9% of the network products were observed on average. When summing all the observations, 17.2% (497) of the edges (input and output) were observed in at least one expression profile. Several values of the filtering parameter k were used from k = 1 to k = 5. Without filtering we predicted 152 signs of the network (30% of the edges observed at least once), among them, 41.4% were not inferred by the naive algorithm. We compared the predictions to the known interaction signs: 28.3% of the predictions were false predictions. Sources of errors may lie on nonmodelled interactions (possibly effects of sigmafactors), or in using experiments on different E. coli strains. Filtering improves our score allowing us to retain only reliable predictions. Thus, for k = 5, we inferred 41 signs, of them, only 1 was an incorrect prediction (2.5% of false prediction). We conclude that filtering is a good way to strengthen our predictions even when the model is not precise enough. We illustrated the effect of the filtering process in Fig. 9.
It should be noted that we obtained very similar results either by cleaning the data thanks to the signed network, either by using our filtering procedure. This is a particularly clear indication that this filtering procedure is an effective strategy to produce robust predictions.
Our algorithm also detected ambiguous modules in the network. There are seven MBM of Type I (i.e. single incoming interactions); four of them are also stated as ambiguous by the naive algorithm. In addition, there are 4 MBM of Type II that are not detected by the naive inference algorithm. All the ambiguities are shown in Fig. 10. A list of experimental assays that yield ambiguities on each interaction is given in the Supplementary Web site. This analysis shows that there exist nonmodelled interactions that balance the effects on the targets in the MBM detected.
A real case: inference of signs in S. cerevisiae transcriptional regulatory network
We applied our inference algorithm to the transcriptional regulatory network of the budding yeast S. cerevisiae. Let us here briefly review the available sources that can be used to build the unsigned regulatory network. The experimental dataset proposed by Lee et al. [11] is widely used in the network reconstruction literature. It is a study conducted under nutrient rich conditions, and it consists of an extensive ChIPchip screening of 106 TFs. Estimations regarding the number of yeast TFs that are likely to regulate specific groups of genes by direct binding to the DNA vary from 141 to 209, depending on the selection criteria. In followup papers of this work, the ChIPchip analysis was extended to 203 yeast TFs in rich media conditions and 84 of these regulators in at least one environmental perturbation [12]. Analysis methods were refined in 2005 by MacIsaac et al. [13]. Other studies continued to work in this network using different approaches [10, 14–16]. Here we selected two of these sources. All networks are provided in the Supplementary Web site.

(A)
The first network consists of the core of the transcriptional ChIPchip regulatory network produced in [11]. Starting from the full network with a pvalue of 0.005, we reduced it to the set of nodes that have at least one output edge. This network was already studied in [28]. It contains 31 nodes and 52 interactions.

(B)
The second network contains all the transcriptional interactions between TFs shown by [11] with a pvalue below 0.001. It contains 70 nodes and 96 interactions.

(C)
The third network is the set of interactions among TFs as inferred in [13] from sequence comparisons. We have considered the network corresponding to a pvalue of 0.001 and 2 bindings (83 nodes, 131 interactions).

(D)
The last network contains all the transcriptional interactions among genes and regulators shown by [11] with a pvalue below 0.001. It contains 2419 nodes and 4344 interactions.
Inference process with genedeletion expression profiles
We first applied our inference algorithm to the large scale network (D) using a panel of expression profiles for 210 genedeletion experiments [30]. The information given by this panel is quite small, since 1.6% of all the products in the network is on average observed, and 12% of the edges (input and output) of the network are observed in at least one expression profile. Using these data, we inferred 162 regulatory roles.
We validated our prediction with a literaturecurated network on Yeast [31]. We found that among the 162 signpredictions, 12 were referenced with a known interaction in the database, and 9 with a good sign.
Genedeletion expression profiles were used in order to compare our results to path analysis methods [20, 23] since the latter can only be applied to knockout data. Other signregulation inference methods needed either other sources of generegulatory information (promoter binding information, proteinprotein information), or timeseries data to be performed [10, 15, 18].
First, we tested the consistency between the inferred network obtained from path analysis methods with the 210 genedeletion experiments. We obtained that the network was inconsistent with 28 of the 210 experiments. Second, we compared the inference results for both methods, our approach and the path analysis method, obtaining in the latter that 234 roles of widely connected paths were inferred; whereas with our method 162 roles were inferred, mainly localised in the branches of the network. Both results intersected on 17 interactions and no contradiction in the inferred role was reported. An illustration of these results is given in the Supplementary Web site.
This suggests that our approach is complementary to path analysis methods. Our explanation is as follows: in [20, 23], network inference algorithms identify probable paths of physical interactions connecting a gene knockout to genes that are differentially expressed as a result of that knockout. This leads to a search for the smallest number of interactions that carry the largest information in the network. Hence, inferred interactions are located near the core of the network, but not exactly in the core. On the contrary, as we already mentioned, the combinatorics of interactions in the core of the network are too intricate to be determined from a few hundreds of expression profiles with our algorithm, thus, we concentrate on interactions around the core.
Inference with stress perturbation expression profiles
To overcome the problem exposed using the small amount of information contained in [30], we have used stress perturbation experiments. These data correspond to curated information available in SGD (Saccharomyces Genome Database) [32]. When time series profiles were available, we selected the last time expression array. Therefore, we collected and treated 15 experiments described in Table 2. For each expression array, we sorted the measured genes in four classes: 2fold upregulated, 2fold downregulated, nonobserved, and zero variation. Full datasets are available in the Supplementary Web site.
As in the case of E. coli, it appeared that all the networks (A), (B), (C), and (D) were not consistent with the whole set of expression arrays. Thus, when executing our algorithms we identified motifs that held ambiguities, and we marked them as MBM of type IIV (as described in our inference algorithm). We also generated a set of inferred signs and applied the filtered algorithm (with filter k = 3) to the large scale network (D).
We obtained our total inference rate by adding the number of inferred signs fixed in an unique way to the number of nonrepeated interactions in the MBM detected, and dividing it by the total number of edges in the network. In Table 3 we illustrate the inference rate obtained for each of the networks. Depending on the network, the inference rate varies from 19% to 37%; thus, they are similar to the theoretical rates obtained for E. coli network even with a small number of perturbation experiments (14 or 15).
We validated the inferred interactions comparing them to the literaturecurated network published in [31]. We obtained 631 predictions when no filtering is applied. Furthermore, among the 198 interactions predicted with a filter parameter k = 3, 19 were referenced with a known interaction in the database, and only 1 prediction had a wrong sign. As in the case of E. coli, we conclude that filtering is a good way to produce extremely robust predictions. Additionally, we compared our predictions to the naive inference algorithm finding that the naive algorithm usually predicts half of the signs that we obtain. In Fig. 11 we illustrate the inferred interactions for Network (B).
As already mentioned, the algorithm identified a large number of ambiguities. The exhaustive list of MBM is given in the Supplementary Web site and the Type I modules of size 2 found for the networks (A), (B), and (C) are detailed in Table 4. We noticed that the MBM of Type I were detected in the four networks; whereas the MBM of Type IIIV were only detected, in an large number, for Network (D); Type II MBM being the most numerous (85.4%). For each MBM, a precise biological study of the species should enable to understand the origin of the ambiguity: erroneous expression data, missing interactions in the model, or contextdependent regulations.
Contribution of expression profiles to the inference
Analysing only the sign inference process on the global network (D), we wish to estimate how the 14 experiments used influence the unique way {+, } inferred signs. On that account we address the following question: Assuming that all the inferred roles in Step 1 of our inference algorithm are correct, which is the experiment that marks more inferred roles as inconsistent (i.e. that generates more MBM)?
Therefore, we classified the 14 experiments according to the MBM of Type IIIV generated per experiment. MBM of Type I are not included in this computation, for they are inferred in Step 1 of the algorithm. The results of this classification are shown in Fig. 12. The fourth chart illustrates that the real contribution of each expression profile does not depend on the amount of observed genes it contains.
Discussion
Predicting from a "small" number of expression profiles
In principle, inferring the functional effect of regulations could be done using general reconstruction methods. The most outstanding approaches in this domain include Bayesian networks [33], linear ordinary differential equations (ODE) [34, 35] and correlation/causal networks [14, 16, 36] (see [10] for a review, and a comparison on several datasets). These are quantitative methods which are carefully designed to cope with the high level of noise that is generally observed in expression data. They rely either on an explicit parametric modelling of noise distribution (like in Bayesian networks), either on robust statistical estimators for the network and its kinetic parameters. The main limitation of these approaches is the number of independent samples they require in order to be properly used. It is often stated [10, 36] that a minimum of 100 to 300 expression profiles are needed for the estimation procedure. While there exists a couple of datasets of such size, the usual number of available profiles for a given biological system is much smaller. Our approach is meant to be used when the number of profiles ranges from 1 to a couple of hundreds, and should thus be seen as complementary to quantitative methods. Indeed our simulations on E. coli network show that one can characterise about 30% of the regulations from 30 expression profiles. We additionally showed that this is close to the theoretical limit of our approach. This result was confirmed using expression data on the same network: we infer 20% of the regulations whose input and output are simultaneously observed in at least one experiment, using 61 expression profiles.
Generating accurate predictions
The problem of inferring functional effect of transcription factors was specifically addressed by Yeang and colleagues [20, 23], using a probabilistic discrete model. In this approach, one identifies probable paths of physical interactions connecting a gene knockout to genes that are differentially expressed as a result of that knockout. Predictions correspond to the signs found in models of maximum likelihood. More generally, most reconstruction methods are based on computing an "optimal" model with respect to the data. This raises two main issues. First, the underlying optimization problems are often non convex, and finding a global optimum is a very difficult computational task. In practice, most algorithms only guarantee to find a local optimum, which should be cautiously examined before being reported as a prediction. Second, even if a global optimum is found, it is important (but computationally difficult) to check that there is no slightly suboptimal model that yields very different predictions. In other terms, it is necessary to evaluate the robustness of the predictions. In our approach, we describe the (possibly huge) set of models that are consistent with the data, then look for invariants in this set. This means that our predictions are compatible with all feasible models. In order to cope with experimental noise, we combine this strategy with a filtering procedure, which selects predictions that agree with a minimal number of expression profiles. This led us to very accurate predictions, as it was shown on data from E. coli and yeast. We compared our inference approach to the path analysis method by Yeang and colleagues [20, 23]. We found that both algorithms infer a similar number of regulations, and that the predictions coincide. We noticed that the predictions are located in different parts of the network, depending on the algorithm: path analysis tends to infer signs in highly connected regions, while our approach infer signs on regulations acting on small indegree nodes. Another difference is that path analysis requires expression profiles from genedeletion experiments, whereas our method gives better results with stress perturbation experiments (though it can be applied to both types of experiment).
Sign inference and network topology
Using simulations, we evaluated the dependence between the number of available expression profiles and the number of signs that can be inferred from them. Not surprisingly, we noticed that the topology of the regulatory network has a strong influence on the estimated relationship. This was illustrated by computing statistics on both a complete regulatory network and its core. The complete network is characterised by an overrepresentation of feedbackfree regulatory cascades, which are controlled by a small number of TFs. In this setting, the number of inferred signs grows almost continuously with the number of observations. In contrast, the core network does not obey the simple law "the more you observe, the better", some expression profiles being clearly more informative than others. Additionally, in these core networks an unfeasible number of experiments is necessary to infer a small number of signs with high probability. For these core networks, two different strategies may be adopted. First, to build a more accurate model for these restricted subnetworks using dynamic modelling techniques (see [29] for a review). Second, to develop experiment planning in our qualitative framework: given some control parameters, how to find the most informative experiments while keeping their number as low as possible?
Conclusion
In this work we proposed a discrete approach for a particular case of reconstruction problem: given a set of regulations between genes, and a set of expression profiles, determine the functional effect of each regulation, as activation or inhibition. Our approach is based on a qualitative modelling framework, that was initially introduced to check the consistency between a regulatory network and expression data [24, 25]. This framework is based on a rule, which basically says that if the expression of a gene varies between two conditions, then this should be accounted for by the variation of at least one of its predecessors. Here we applied this approach to predict the functional effect of transcription factors on their target genes.
While intuitive and simple, the qualitative rule we propose can be used to infer a significant number of regulatory effects from a reasonable number of expression profiles. As shown using data on E. coli and yeast, the predictions are particularly reliable, especially when they are validated with our filtering procedure. Furthermore, our algorithms can handle datasets of realistic size. It should be noted that computing the predictions presented in this work requires to solve thousands of NPhard problems (more precisely, constraints with variables on a finite domain). Each of these problem has several thousands of variables. Nevertheless, our algorithms are exact and compute the predictions in no more than an hour using a standard desktop PC. This means that they are able to cope with systemwide data in a fairly reasonable amount of time. Due to the structure of the algorithms, we are confident that they can handle even larger datasets in less time, by distributing the computations on several machines.
From our results on yeast, it appears that a significant proportion of the network – as given by ChIPchip data – is not compatible with the available expression profiles. As explained in the Results section, these data is discarded from the analysis, in order to compute safe predictions – but at the expense of a loss of information. The subject of our current work is to develop an improved notion of prediction, that copes better with inconsistent network and data. The goal is to include inconsistent data in the inference process, while preserving the reliability of the predictions.
Methods
Problem statement
We consider the set of equations derived from a given interaction graph G:
where ${X}_{i}^{k}$ stands for the sign of the variation of species i in experiment k, and S_{ ji }the sign of the influence of species j on species i. Recall that the graph G itself comes from chIPchip experiments or sequence analysis. Using expression arrays, we obtain an experimental value for some variables ${X}_{i}^{k}$, which will be denoted ${x}_{i}^{k}$; more generally uppercase (resp. lowercase) letters will stand for variables of the systems (resp. constants +,  or 0).
A single equation in the system (4) can be viewed as a predicate P_{ i,k }(X, S) where i denotes a node in the graph and k one of the r available experiments. If the value for some variables in the equation is known, the predicate resulting from their instantiation will be denoted P_{ i, k }(X, S) [x^{k}, s].
Our problem can now be stated as follows: given a set of expression profiles x^{1},...,x^{r}, decide if the predicate:
can be satisfied. If so, find all variables that take the same value in all admissible valuations (so called hard components of the system).
Decision diagram encoding
In a previous work [26], we showed how the set of solutions of a qualitative system can be computed as a decision diagram [37]. A decision diagram is a data structure meant to represent functions on finite domains; it is widely used for the verification of circuits or network protocols. Using such a compact representation of the set of solutions, we proposed efficient algorithms for computing solutions of the systems, hard components, and other properties of a qualitative system. Back to our problem: in order to predict the regulatory role of TFs on their target genes, it is enough to compute the decision diagram representing the predicate (5), and compute its hard components as proposed in [26]. This approach is suitable for systems of at most a couple of hundred variables. Above this limit, the decision diagram is too large in memory complexity. In our case however, we consider systems of about 4000 variables at most, which is far too large for the above mentioned algorithms.
In order to cope with the size of the problem, we propose to investigate a particular case, when all species are observed, in all experiments. In this case, i ≠ j implies that P_{i, k}(X, S) [x^{k}] and P_{j, k}(X, S) [x^{k}] share no variables. This means that P may be satisfied if and only if each predicate
may be satisfied. As a consequence, a variable S_{ ji }is a hard component of P if and only if it is a hard component of P_{i,..}P_{i,.}correspond to the constraints which relate species i to its predecessors in G for all experiments. The number of variables in P_{i,.}is exactly the indegree of species i in G, which is at most 10–20 in biological networks.
As soon as some species are not observed in some experiment, the predicates P_{i,.}share some variables and it is not guaranteed to find all hard components by studying them separately. A brief investigation showed (data not shown) that due to the topology of the graph, most of the equations are not independent any more, even with few missing nodes. Note however, that any hard component of P_{i,.}is still a hard component of P. The same statement holds for
where P_{.,k}corresponds to the constraints that relate all species in G for a single experiment. Relying on this result, we implemented the following algorithm
In practice, this algorithm is very effective in terms of computation time and number of hard components found. However, as already stated, it is not guaranteed to find all hard components of P. This is what motivates the technique described in the next paragraph.
Solving with Answer Set Programming
In order to solve large qualitative systems, we also tried to encode the problem as a logic program, in the setting of answer set programming (ASP). While decision diagrams represent the set of all solutions, finding a model for a logic program provides one solution. In order to find hard components, it is enough to check for each variable V, if there exists a solution such that V = + and another solution such that V = . The ASP program we used in order to solve the qualitative system is given in supplementary materials. In the following we will denote by asp_solve(P) the call to the ASP solver on the predicate P. The returned value is an admissible valuation if there is one, or ⊥ otherwise. The complete algorithm is reported below
We use clasp for solving ASP programs [38], which performs astonishingly well on our data. The procedure described in Algorithm 3 is particularly efficient in finding non hard components: generating one solution may be enough to prove non hardness of many variables at a time.
To sum up, in order to solve a system of qualitative equations (4) with only partial observations, we use Algorithm 2 first and thus determine most (if not all) hard components. Then, Algorithm 3 is used for the remaining components, which are nearly all non hard.
Reduction technique
As mentioned in the Result section, interaction graphs may be reduced in a way that preserves the satisfiability of the associated qualitative system. Consider a graph G with defined signs on its edges. If some node n has no successor, then deletes it from G. Note then, that any solution of the qualitative system associated to the new graph can be extended in a solution to the system associated to G. The same statement holds if one iteratively delete all nodes in the graph with no successor. The result of this procedure is the subgraph of G such that any node is either on a cycle, or has a cycle downstream. We refer to it as the core of the interaction graph.
The core of an interaction graph corresponds to the most difficult part to solve, because extending a solution for the core to the entire graph can be done in polynomial time, using a breadthfirst traverse.
Diagnosis for noisy data
When working with reallife data, it may happen that the predicate P defined in Eq. (5) cannot be satisfied. This may be due to three (non exclusive) reasons:

a reported expression data is wrong

an arrow (or more generally a subgraph) is missing

the sign on an edge depends on the state of the system
In the third case, the conditions for deriving Eq. (1) are not fulfilled for one node and its qualitative equation should be discarded. This, however, does not affect the validity of the remaining equation.
In all cases, isolating the cause of the problem is a hard task. We propose the following diagnosis approach: as P is a conjunction of smaller predicates, it might happen that some subsets of the predicates are not satisfiable yet. Our strategy is then to find "small" subsets of predicates which cannot be satisfied. A particularly interesting feature of this approach is that by selecting subsets of P_{i,. ,.}predicates, the result might directly be interpreted and visualised as a subgraph of the original model.
How to determine if a sign can be inferred
In the Results section, we have seen some examples showing that even when all feasible observations are available, it might not be possible to infer all signs in the interaction graph. Whether or not a sign can be inferred depends on the topology of the graph, and also on the actual signs on interactions. In practice, it is thus impossible to tell from the unsigned graph only if a sign can be recovered. However, it is still interesting to evaluate on fully signed interaction networks which part can be inferred. A trivial algorithm for this consists in explicitly generating all feasible observations and using the algorithms described above. This is unfeasible due to the number of observations.
With the notations introduced above, consider an observation X and sign variables S for an interaction graph. P_{ i }(X, S) denotes the constraints that link the variation of a node i to that of its predecessors given the signs of the interactions. Moreover, the real signs in the graph are denoted by s. For each node i, we build the predicate giving the feasible observations on node i and its predecessors, given the rest of the graph and the real signs s
Then, the constraint that we can derive on S variables is: for any observation X that is feasible P_{ i }(X, S) should hold. This constraint is more formally defined by
C_{ i }(S) = ∀XO_{ i }(X) ⇒ P_{ i }(X, S)
Finally, the hard components of C_{ i }are exactly the signs that can be inferred using all feasible observations. Let us sum up the procedure:

1.
compute P(X, S) = ∧_{1 ≤ i ≤ n}P_{ i }(X, s)

2.
compute O_{ i }from P and the actual signs s

3.
compute C_{ i }, the constraints of signs given all feasible observations

4.
compute the hard components of C_{ i }, which are exactly the signs that can be inferred.
If it is not possible to compute P(X, S) (mainly because the interaction graph is too large), we use a more sophisticated approach based on a modular decomposition of the interaction graph. The resulting algorithm, as well as all inference algorithms, experimental data, and the results obtained for the S. cerevisiae, and E. coli regulatory networks can be found at: http://www.irisa.fr/symbiose/interactionNetworks/supplementaryInference.html.
Appendix
Algorithm 1
Naive Inference algorithm
Algorithm: Naive Inference algorithm
Input:
a network with its topology
a set of expression profiles
Output:
a set of predicted signs
a set of ambiguous interactions
For all Node A with exactly one predecessor B
if A and B are observed simultaneously then return
prediction sign(B → A) = sign(A) * sign(B)
if sign(B → A) was predicted different by another
expression profile then return Ambiguous arrow B → A
Algorithm 2
Heuristic for finding hard components in large interaction networks with many expression profiles.
Input:
the predicates P_{i,.}and P_{.,k}for all i and k
observed variations x
Output:
a set s of hard components of P
s ← ∅
while True do
s' ← ∪_{ i }hard_components(P_{i,.}[x^{k}, s])
if s' = ∅ then return s
s ← s ∪ s'
x' ← ∪_{ k }hard_components(P_{.,k}[x^{k}, s])
if x' = ∅ then return s
x ← x ∪ x'
end
Algorithm 3
Exact algorithm for finding the set of hard components of P, based on logic programming.
Algorithm: Hard components using ASP
Input:
the predicates P
observed variations x
Output:
a set h of hard components of P
h ← ∅
C ← {S_{ ji }j → i}
s * ← asp_solve(P)
if s* = ⊥ then return ⊥
while C ≠ ∅ do
choose V in C
s ← asp_solve(P [V = ${s}_{V}^{\ast}$])
if s = ⊥ then
h ← {(V, s_{ V })} ∪ h
else
delete from C all W in C s.t. any ${s}_{W}^{\ast}$ ≠ = s_{ W }
end
end
References
 1.
Zanzoni A, MontecchiPalazzi L, Quondam M, Ausiello G, HelmerCitterich M, Cesareni G: MINT: a Molecular INTeraction database. FEBS Lett 2002, 513: 135–40. 10.1016/S00145793(01)032938
 2.
Hermjakob H, MontecchiPalazzi L, Lewington C, Mudali S, Kerrien S, Orchard S, Vingron M, Roechert B, Roepstorff P, Valencia A, Margalit H, Armstrong J, Bairoch A, Cesareni G, Sherman D, Apweiler R: IntAct: an open source molecular interaction database. Nucleic Acids Res 2004, (32 Database):D452–5. 10.1093/nar/gkh052
 3.
Peri S, Navarro JD, Kristiansen TZ, Amanchy R, Surendranath V, Muthusamy B, Gandhi TKB, Chandrika KN, Deshpande N, Suresh S, Rashmi BP, Shanker K, Padma N, Niranjan V, Harsha HC, Talreja N, Vrushabendra BM, Ramya MA, Yatish AJ, Joy M, Shivashankar HN, Kavitha MP, Menezes M, Choudhury DR, Ghosh N, Saravana R, Chandran S, Mohan S, Jonnalagadda CK, Prasad CK, KumarSinha C, Deshpande KS, Pandey A: Human protein reference database as a discovery resource for proteomics. Nucleic Acids Res 2004, (32 Database):D497–501. 10.1093/nar/gkh070
 4.
Kanehisa M, Goto S, Kawashima S, Okuno Y, Hattori M: The KEGG resource for deciphering the genome. Nucleic Acids Res 2004, (32 Database):D277–80. 10.1093/nar/gkh063
 5.
IngenuitySystems: Ingenuity Pathways Knowledge Base.1998. [http://www.ingenuity.com]
 6.
Salgado H, GamaCastro S, PeraltaGil M, DiazPeredo E, SanchezSolano F, SantosZavaleta A, MartinezFlores I, JimenezJacinto V, BonavidesMartinez C, SeguraSalazar J, MartinezAntonio A, ColladoVides J: RegulonDB (version 5.0): Escherichia coli K12 transcriptional regulatory network, operon organization, and growth conditions. Nucleic Acids Res 2006, (34 Database):D394–7. 10.1093/nar/gkj156
 7.
Reguly T, Breitkreutz A, Boucher L, Breitkreutz BJ, Hon GC, Myers CL, Parsons A, Friesen H, Oughtred R, Tong A, Stark C, Ho Y, Botstein D, Andrews B, Boone C, Troyanskya OG, Ideker T, Dolinski K, Batada NN, Tyers M: Comprehensive curation and analysis of global interaction networks in Saccharomyces cerevisiae. J Biol 2006, 5(4):11. 10.1186/jbiol36
 8.
Joyce AR, Palsson BO: The model organism as a system: integrating 'omics' data sets. Nat Rev Mol Cell Biol 2006, 7(3):198–210. 10.1038/nrm1857
 9.
Faith JJ, Hayete B, Thaden JT, Mogno I, Wierzbowski J, Cottarel G, Kasif S, Collins JJ, Gardner TS: LargeScale Mapping and Validation of Escherichia coli Transcriptional Regulation from a Compendium of Expression Profiles. PLoS Biology 2007., 5:
 10.
Bansal M, Belcastro V, AmbesiImpiombato A, di Bernardo D: How to infer gene networks from expression profiles. Mol Syst Biol 2007., 3(78):
 11.
Lee TI, Rinaldi NJ, Robert F, Odom DT, BarJoseph Z, Gerber GK, Hannett NM, Harbison CT, Thompson CM, Simon I, Zeitlinger J, Jennings EG, Murray HL, Gordon DB, Ren B, Wyrick JJ, Tagne JB, Volkert TL, Fraenkel E, Gifford DK, Young RA: Transcriptional regulatory networks in Saccharomyces cerevisiae. Science 2002, 298(5594):799–804. 10.1126/science.1075090
 12.
Harbison CT, Gordon DB, Lee TI, Rinaldi NJ, Macisaac KD, Danford TW, Hannett NM, Tagne JB, Reynolds DB, Yoo J, Jennings EG, Zeitlinger J, Pokholok DK, Kellis M, Rolfe PA, Takusagawa KT, Lander ES, Gifford DK, Fraenkel E, Young RA: Transcriptional regulatory code of a eukaryotic genome. Nature 2004, 431(7004):99–104. 10.1038/nature02800
 13.
MacIsaac KD, Wang T, Gordon DB, Gifford DK, Stormo GD, Fraenkel E: An improved map of conserved regulatory sites for Saccharomyces cerevisiae. BMC Bioinformatics 2006, 7(NIL):113. 10.1186/147121057113
 14.
Xing B, Laan MJ: A causal inference approach for constructing transcriptional regulatory networks. Bioinformatics 2005, 21(21):4007–13. 10.1093/bioinformatics/bti648
 15.
Segal E, Shapira M, Regev A, Pe'er D, Botstein D, Koller D, Friedman N: Module networks: identifying regulatory modules and their conditionspecific regulators from gene expression data. Nat Genet 2003, 34(2):166–76.
 16.
Nariai N, Tamada Y, Imoto S, Miyano S: Estimating gene regulatory networks and proteinprotein interactions of Saccharomyces cerevisiae from multiple genomewide data. Bioinformatics 2005, 21(Suppl 2(NIL)):ii206ii212. 10.1093/bioinformatics/bti1133
 17.
Ferrazzi F, Magni P, Sacchi L, Nuzzo A, Petrovic U, Bellazzi R: Inferring gene regulatory networks by integrating static and dynamic data. Int J Med Inform 2007, in press.
 18.
S B, Eils R: Inferring genetic regulatory logic from expression data. Bioinformatics 2005, 21(11):2706–13. 10.1093/bioinformatics/bti388
 19.
Ideker T: A systems approach to discovering signaling and regulatory pathwaysor, how to digest large interaction networks into relevant pieces. Adv Exp Med Biol 2004, 547(NIL):21–30.
 20.
Yeang CH, Mak HC, McCuine S, Workman C, Jaakkola T, Ideker T: Validation and refinement of generegulatory pathways on a network of physical interactions. Genome Biol 2005, 6(7):R62. 10.1186/gb200567r62
 21.
GutierrezRios RM, Rosenblueth DA, Loza JA, Huerta AM, Glasner JD, Blattner FR, ColladoVides J: Regulatory network of Escherichia coli: consistency between literature knowledge and microarray profiles. Genome Res 2003, 13(11):2435–2443. 10.1101/gr.1387003
 22.
Constantinidou C, Hobman J, Griffiths L, Patel M, Penn C, Cole J, Overton T: A reassessment of the FNR regulon and transcriptomic analysis of the effects of nitrate, nitrite, NarXL, and NarQP as Escherichia coli K12 adapts from aerobic to anaerobic growth. J Biol Chem 2006, 281(8):4802–15. 10.1074/jbc.M512312200
 23.
Yeang CH, Ideker T, Jaakkola T: Physical network models. J Comput Biol 2004, 11(2–3):243–62. 10.1089/1066527041410382
 24.
Radulescu O, Lagarrigue S, Siegel A, Veber P, Borgne ML: Topology and static response of interaction networks in molecular biology. J R Soc Interface 2006, 3(6):185–96. 10.1098/rsif.2005.0092
 25.
Siegel A, Radulescu O, Borgne ML, Veber P, Ouy J, Lagarrigue S: Qualitative analysis of the relation between DNA microarray data and behavioral models of regulation networks. Biosystems 2006, 84(2):153–74. 10.1016/j.biosystems.2005.10.006
 26.
Veber P, Borgne ML, Siegel A, Lagarrique S, Radulescu O: Complex Qualitative Models in Biology: A new approach. Complexus 2005, 2(3–4):140–151.
 27.
Guziolowski C, Veber P, Borgne ML, Radulescu O, Siegel A: Checking Consistency Between Expression Data and Large Scale Regulatory Networks: A Case Study. Journal of Biological Physics and Chemistry 2007, 7: 37–43.
 28.
Kauffman S, Peterson C, Samuelsson B, Troein C: Random Boolean network models and the yeast transcriptional network. Proc Natl Acad Sci USA 2003, 100(25):14796–9. 10.1073/pnas.2036429100
 29.
de Jong H: Modeling and simulation of genetic regulatory systems: a literature review. J Comput Biol 2002, 9: 67–103. 10.1089/10665270252833208
 30.
Hughes T, Marton M, Jones A, Roberts C, Stoughton R, Armour C, Bennett H, Coffey E, Dai H, He Y: Functional Discovery via a Compendium of Expression Profiles. Cell 2000, 102: 109–126. 10.1016/S00928674(00)000155
 31.
Nabil Guelzim N, Bottani S, Bourgine P, 2 Képès F: Topological and causal structure of the yeast transcriptional regulatory network. Nature Genetics 2002, 31: 60–63. 10.1038/ng873
 32.
Hong E, Balakrishnan R, Christie K, Costanzo M, Dwight S, Engel S, Fisk D, Hirschman J, Livstone M, Nash R, Oughtred R, Park J, Skrzypek M, Starr B, Andrada R, Binkley G, Dong Q, Hitz B, Miyasato S, Schroeder M, Weng S, Wong E, Zhu K, Dolinski K, Botstein D, Cherry J: Saccharomyces Genome Database.2001. [http://www.yeastgenome.org/]
 33.
Friedman N, Linial M, Nachman I, Pe'er D: Using Bayesian networks to analyze expression data. J Comput Biol 2000, 7(3–4):601–20. 10.1089/106652700750050961
 34.
Di Bernardo D, Thomson M, Gardner T, Chobot S, Eastwood E, Wojtovich A, Elliot S, Schaus S, Collins J: Chemogenomic profiling on a genomewide scale using reverseengeneered gene networks. Nat Biotechnol 2005, 23: 377–383. 10.1038/nbt1075
 35.
Bansal M, Della Gatta G, di Bernardo D: Inference of gene regulatory networks and compound mode of action from time course gene expression profiles. Bioinformatics 2006, 22: 815–822. 10.1093/bioinformatics/btl003
 36.
Margolin AA, Nemenman I, Basso K, Wiggins C, Stolovitzky G, Favera RD, Califano A: ARACNE: an algorithm for the reconstruction of gene regulatory networks in a mammalian cellular context. BMC Bioinformatics 2006, 7(Suppl 1(NIL)):S7. 10.1186/147121057S1S7
 37.
Bryan R: Graphbased algorithm for boolean function manipulation. IEEE Transactions on Computers 1986, 8: 677–691. 10.1109/TC.1986.1676819
 38.
Gebser M, Kaufmann B, Neumann A, Schaub T: clasp: A ConflictDriven Answer Set Solver. Ninth International Conference on Logic Programming and Nonmonotonic Reasoning, Tempe, AZ, USA 2007.
 39.
Shannon P, Markiel A, Ozier O, Baliga NS, Wang JT, Ramage D, Amin N, Schwikowski B, Ideker T: Cytoscape: a software environment for integrated models of biomolecular interaction networks. Genome Research 2003, 13(11):2498–2504. [http://www.cytoscape.org] 10.1101/gr.1239303
 40.
DeRisi JL, Iyer VR, Brown PO: Exploring the metabolic and genetic control of gene expression on a genomic scale. Science 1997, 278(5338):680–6. 10.1126/science.278.5338.680
 41.
Chu S, DeRisi J, Eisen M, Mulholland J, Botstein D, Brown PO, Herskowitz I: The transcriptional program of sporulation in budding yeast. Science 1998, 282(5389):699–705. 10.1126/science.282.5389.699
 42.
Sudarsanam P, Iyer VR, Brown PO, Winston F: Wholegenome expression analysis of snf/swi mutants of Saccharomyces cerevisiae. Proc Natl Acad Sci USA 2000, 97(7):3364–9. 10.1073/pnas.050407197
 43.
Ogawa N, DeRisi J, Brown PO: New components of a system for phosphate accumulation and polyphosphate metabolism in Saccharomyces cerevisiae revealed by genomic expression analysis. Mol Biol Cell 2000, 11(12):4309–21.
 44.
Gasch AP, Spellman PT, Kao CM, CarmelHarel O, Eisen MB, Storz G, Botstein D, Brown PO: Genomic expression programs in the response of yeast cells to environmental changes. Mol Biol Cell 2000, 11(12):4241–57.
 45.
Gasch AP, Huang M, Metzner S, Botstein D, Elledge SJ, Brown PO: Genomic expression responses to DNAdamaging agents and the regulatory role of the yeast ATR homolog Mec1p. Mol Biol Cell 2001, 12(10):2987–3003.
 46.
Cullen PJ, Sabbagh WJ, Graham E, Irick MM, van Olden EK, Neal C, Delrow J, Bardwell L, Sprague GFJ: A signaling mucin at the head of the Cdc42 and MAPKdependent filamentous growth pathway in yeast. Genes Dev 2004, 18(14):1695–708. 10.1101/gad.1178604
 47.
Roberts GG, Hudson AP: Transcriptome profiling of Saccharomyces cerevisiae during a transition from fermentative to glycerolbased respiratory growth reveals extensive metabolic and structural remodeling. Mol Genet Genomics 2006, 276(2):170–86. 10.1007/s0043800601339
 48.
Mnaimneh S, Davierwala AP, Haynes J, Moffat J, Peng WT, Zhang W, Yang X, Pootoolal J, Chua G, Lopez A, Trochesset M, Morse D, Krogan NJ, Hiley SL, Li Z, Morris Q, Grigull J, Mitsakakis N, Roberts CJ, Greenblatt JF, Boone C, Kaiser CA, Andrews BJ, Hughes TR: Exploration of essential gene functions via titratable promoter alleles. Cell 2004, 118: 31–44. 10.1016/j.cell.2004.06.013
Acknowledgements
The authors are particularly grateful to B. Kauffman, M. Gebser, and T. Schaub from the University of Potsdam for their help on CLASP software. They also wish to thank the referees for their interesting and constructive remarks.
Author information
Additional information
Authors' contributions
PV participated in designing the algorithms described in the Methods section and in performing the simulations. CG designed the algorithms described in the Results section, and performed the analysis on E. coli and yeast data. OR made the statistical modeling in the Results section. MLB participated in implementing the algorithms on decision diagrams. All authors participated in writing the manuscript, read and approved its final state.
Authors’ original submitted files for images
Below are the links to the authors’ original submitted files for images.
Rights and permissions
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
About this article
Cite this article
Veber, P., Guziolowski, C., Le Borgne, M. et al. Inferring the role of transcription factors in regulatory networks. BMC Bioinformatics 9, 228 (2008) doi:10.1186/147121059228
Received
Accepted
Published
DOI
Keywords
 Interaction Graph
 Inference Algorithm
 Transcriptional Regulatory Network
 Network Reconstruction
 Perturbation Experiment