 Methodology Article
 Open Access
 Published:
A shortcut for multiple testing on the directed acyclic graph of gene ontology
BMC Bioinformatics volume 15, Article number: 349 (2014)
Abstract
Background
Gene set testing has become an important analysis technique in high throughput microarray and next generation sequencing studies for uncovering patterns of differential expression of various biological processes. Often, the large number of gene sets that are tested simultaneously require some sort of multiplicity correction to account for the multiplicity effect. This work provides a substantial computational improvement to an existing familywise error rate controlling multiplicity approach (the Focus Level method) for gene set testing in high throughput microarray and next generation sequencing studies using Gene Ontology graphs, which we call the Short Focus Level.
Results
The Short Focus Level procedure, which performs a shortcut of the full Focus Level procedure, is achieved by extending the reach of graphical weighted Bonferroni testing to closed testing situations where restricted hypotheses are present, such as in the Gene Ontology graphs. The Short Focus Level multiplicity adjustment can perform the full topdown approach of the original Focus Level procedure, overcoming a significant disadvantage of the otherwise powerful Focus Level multiplicity adjustment. The computational and power differences of the Short Focus Level procedure as compared to the original Focus Level procedure are demonstrated both through simulation and using real data.
Conclusions
The Short Focus Level procedure shows a significant increase in computation speed over the original Focus Level procedure (as much as ∼15,000 times faster). The Short Focus Level should be used in place of the Focus Level procedure whenever the logical assumptions of the Gene Ontology graph structure are appropriate for the study objectives and when either no a priori focus level of interest can be specified or the focus level is selected at a higher level of the graph, where the Focus Level procedure is computationally intractable.
Background
Microarray technology and next generation sequencing have played an important role in discovering important associations between gene expression patterns and phenotype [1]. Such gene expression technologies have been instrumental in discoveries ranging from the retarding of aging in mice brought about by caloric restrictions in diet [2] to the identification of various types of diffuse large Bcell lymphoma in humans [3]; from characterizing the transcriptomes of in vitro maturated porcine embryos [4] to uncovering the underlying genes and pathways involved in Alzheimer’s disease [5]. While both microarray and next generation sequencing technologies allow researchers to study the differential expression of genes across conditions or treatments, each has their advantages and disadvantages [1]. However, in either case, the resulting increase in genetic knowledge has allowed researchers to group genes with common function into gene sets and test these gene sets for differential expression [6],[7].
One rich source of gene set knowledge is found in the Gene Ontology database [8]. The Gene Ontology (GO) provides a controlled vocabulary that is not specific to any particular species. This vocabulary is divided into three general ontologies, Molecular Function (MF), Cellular Component (CC), and Biological Process (BP). Individual GO Terms form the basis of these vocabularies and are structured through parent child relationships with more general terms as parents and more specific terms as children. Each GO Term typically contains a definition of its biological process (molecular function or cellular component) and other annotation as well as a mapping of all known gene products involved in its specified process (function or component). For convenience in presentation, the remainder of this work will focus on the biological process ontology.
Gene set testing allows for the quantification of the significance of activity level differences between treatment groups for specific biological processes of interest. For example, a recent study on human longevity compared the gene expression profiles corresponding to 1,808 different biological processes for nonagenarians and a control group to identify 73 biological processes associated with longevity [9]. When there are relatively few gene sets (biological processes) of a priori interest (1,808 in [9]), the impact of the multiplicity correction for the tests of differential expression (or differential activity) of the gene sets can be greatly lessened as compared to individually testing all member genes (45,164 in [9]), improving the power of the test. Even when no a priori gene set of interest can be specified, it can still be highly beneficial to test all known gene sets from a biological process database for differential expression, as the number of gene sets is still typically magnitudes smaller than the corresponding number of individual genes [6],[10].
Many methods of gene set testing have been proposed in the literature as reviewed in [11]. These can essentially be divided into two classes of gene set testing, often referred to as competitive tests and self contained tests. The competitive tests compare the expression profiles of the genes in the set to those not in the set. The self contained tests focus only on those genes within the set and compares them to some fixed standard. While the first are more popular [7],[12], the second have been shown to be more powerful [11],[13]. Further, the null hypothesis associated with the self containedtests,
${H}_{0}^{\mathit{\text{self}}}$ : no genes in the gene set are differentially expressed,
has been shown to be the more logical generalization of single gene testing (with other advantages that will be explained later on) as compared to the competitivetest null hypothesis
${H}_{0}^{\mathit{\text{comp}}}$ : the genes in the gene set are at most as often differentially expressed as the genes in the complement of the gene set.
While gene set testing methods are varied in their approach, they are alike in that they test each GO term, i.e. gene set, individually. Thus, when more than one GO term is tested simultaneously (typically hundreds or thousands are tested simultaneously) some sort of multiplicity adjustment is necessary to preserve control over either the familywise error rate (FWER) or the false discovery rate (FDR) or a derivative of these error rates. The FDR is typically the error rate of choice in exploratory studies where follow up confirmatory studies are then conducted [14]. On the other hand, the FWER is typically the suggested error rate for confirmatory studies [15]. We also suggest that the FWER is highly appropriate for exploratory gene set studies as, in our experience, it is seldom more results that are desired, but the most promising real significances that are sought. The FWER offers the best error rate control for such conclusions [15].
The Focus Level method is a powerful method of multiplicity adjustment for self contained gene set testing, which takes into account the structure of the GO graph while controlling (strongly) the FWER [10]. This approach is more powerful than standard FWER controlling methods such as the Bonferroni and uniformly more powerful BonferroniHolm [16] procedures for multiple testing with GO graphs [10]. However, it is important to note that this increase in power comes at the cost of requiring the logical structure of the GO graph to become part of the multiplicity adjustment.
The Focus Level method allows the researcher to select the level of the GO graph in which they are most interested. This is called the focus level. The procedure then applies a topdown and bottomup approach from the specified focus level. First, the terms in the focus level are tested using the BonferroniHolm adjustment [16]. Then, in the bottomup approach, any term above the focus level is declared significant when any of its offspring in the focus level have been declared significant. This inheritance of Pvalues is accomplished through the assumption that a parent term must be differentially expressed if any of its children terms are differentially expressed, a logical assumption for the GO graph structure [10]. In the topdown procedure, significance of the children of the focus level terms is decided through an application of the closed testing procedure of [17].
While the Focus Level method is a powerful approach to adjusting for multiplicity, it quickly becomes computationally infeasible when the selected focus level contains a large number of offspring in the GO graph [10]. This computational limitation makes it essentially impossible to perform the full topdown approach, a rather significant disadvantage [18]. Using the full topdown approach provides researchers the default focus level of the root node (GO:0008150 in the context of the BP GO ontology) whenever they have no a priori interest in a given focus level, a common scenario, see for example [18]. This also allows adjusted Pvalues to be considered apart from their context in the GO graph which is advantageous to reporting on single significant gene sets of interest. Discussions of the significant findings of the Focus Level method are currently restricted to their context within the GO graph [10].
This work proposes an improvement to the topdown portion of the Focus Level method [10] which we call the Short Focus Level as it performs a shortcut of the full Focus Level method. This is accomplished using a novel application of the general graphical Bonferroni adjustment for multiple testing as proposed by [19], which is a generalization of closed testing based on weighted Bonferroni tests [20]. The Short Focus Level procedure shows a significant improvement in computational speed (as much as ∼ 15,000 times faster) while maintaining similar power to that of the original Focus Level procedure and even showing a gain in power over the original Focus Level procedure for certain scenarios while experiencing a loss in power for others. Most importantly, the computational improvements are such that the full topdown method can now be performed on a standard operating system within just a few minutes. The R code [21] for the Short Focus Level procedure is included in the mvGSTmvGST package [22],[23]; see also Additional files 1 and 2.
Methods
The Focus Level procedure [10] adjusts for multiple gene set tests using the structure of the directed acyclic graphs of the Gene Ontology (GO). Two basic assumptions underly the method.

A1.
A nondifferentially expressed parent gene set implies the children gene sets are also nondifferentially expressed.

A2.
If the children gene sets form a partition of the parent gene set and are all nondifferentially expressed, then the parent gene set is also nondifferentially expressed.
These assumptions ensure coherence of the resulting significant subgraph and facilitate interpretations [10]. Note that these assumptions are necessary if the objective is to control the FWER within the structure of the GO graph. If preserving the graph structure in the multiplicity correction is not of interest to the researcher, then the FWER (or even the false discovery rate) could be controlled by the standard Holm’s correction [16] (or any false discovery rate controlling method which allows for arbitrary dependence structures), however such an approach will result in a loss of coherence for the significant subgraph.
Assumptions A1 and A2 require that the null hypothesis for each gene set is that no genes in the gene set are differentially expressed. The alternative in each case being that at least one gene in the set is differentially expressed. Thus, only self contained gene set testing methods (which utilize this hypothesis framework) can be used to test the gene sets of the GO graph if the Focus Level method of multiplicity adjustment is used. This excludes gene set enrichment methods such as those proposed in [12] but supports very well the Global Test of [6], Pvalue combination methods such as Fisher’s and Stouffer’s methods [13],[24], as well as Global Ancova [25], PLAGE [26], and SAMGS [27].
As prescribed by [10] there are two requirements in the selection of the focus level. These requirements are labeled “FL1” and “FL2” here for later reference in subsection “The Shortfocus level procedure”.

FL1.
No offspring of a focus level term be contained in the focus level.

FL2.
All remaining terms are either ancestors or offspring of the focus level terms.
Figure 1 demonstrates on a simplified toy GO graph how the focus level (filled nodes) could be chosen. The full bottomup approach (panel (a) of Figure 1) selects all GO Terms corresponding to terminal nodes as the focus level, in this example, nodes C, D, and E. The full topdown approach (panel (c) of Figure 1) selects the root node, A in this case, as the focus level. Finally, in a typical GO graph there are many (hundreds or thousands) of options for focus levels contained somewhere in the middle of the GO graph. In the simplified example graphs of Figure 1 the most logical intermediate focus level is demonstrated with nodes B and F (panel (b)). It would also be possible to use nodes C, D and F as the focus level but such choices in actual GO graphs do not provide a consistent level of specificity in the graph and would not be as logical a choice. Choosing nodes C, D, E, and F as the focus level would not be allowed as E is a child of F, violating the requirement that the focus level must not contain any offspring of another focus level term (E is an offspring term of F). Choosing only node B as the focus level would also not be allowed as node F is neither an ancestor or offspring term of B, violating the second requirement.
The topdown portion of the Focus Level procedure of [10], which applies the closed testing approach of [17], requires closing the GO graph under all unions from the focus level down. This is done by treating each focus level term, along with all of its offspring terms, as separate graphs which are each closed under all possible unions. As these separate closed graphs will share common elements, the full closed graph $\stackrel{~}{G}$ is obtained by unioning each of the separately closed graphs into a single graph which is also unioned to all ancestor terms of the focus level.
To demonstrate, consider the closures of each of the example GO graphs from Figure 1 as shown in Figure 2. In each case, the nodes above the focus level remain unchanged, while the creation of several sets not present in the original example GO graph (depicted with rounded rectangles) are required in order to close the graph under all possible unions from the focus level down. Since the closing of the graph is only required from the selected focus level down, it is clear from Figure 2 that the more offspring terms the focus level contains, the greater the number of sets that must be created to close the graph. Closing the graph can quickly become computationally infeasible in practice. Importantly, performing the full topdown approach as in panel (c) of Figure 2 is rarely possible in real applications due to the computational burden.
To partially amend the computational difficulties of the Focus Level method, [10] implement a more efficient method of computing the closed graph using what they term “atom sets”. These atom sets form a core collection of gene sets which form a basis for all gene sets in the graph. All other gene sets in the graph (as well as its closure) can be created through unions of the atom sets. This ensures the size of the closed graph is 2^{k}−1, where k is the number of atom sets, which is often smaller (and never larger) than the size of the original closed graph. Further, [10] recommend selecting the focus level so that no more than 912 atom sets are required to recreate the offspring of any single focus level term. They also suggest computing only the smallest few adjusted Pvalues to save computation time in place of computing all adjusted Pvalues.
This work offers an alternative solution to improve on the computational speed of the topdown portion of the Focus Level method through an application of the general graphical Bonferroni adjustment of [19]. This allows for a shortcut of length m in place of the currently applied full closed testing approach of [17]. In the following section, we summarize the general graphical Bonferroni adjustment approach and show how we tailor the method for a powerful application to the topdown portion of the Focus Level method.
The graphical Bonferroni adjustment
A powerful and versatile graphical generalization of weighted Bonferroni based closed testing [17] which provides strong control of the familywise error rate (FWER) at a specified level α was proposed in [19]. Their approach represents all m hypotheses of interest, H_{1},...,H_{ m } as nodes in a directed graph. Each node can be thought of here as a gene set, with a corresponding hypothesis H_{ i } testing for differential expression. Node i, representing hypothesis H_{ i }, is allocated a local threshold α_{ i } for all i=1,...,m. Nodes are joined by edges with weights g_{ ij } dictating the proportion of the local threshold α_{ i } that is allocated to all connected hypotheses (nodes) H_{ j } in the case that hypothesis H_{ i } is rejected. The structure of the graph as well as the size of the local thresholds α_{ i } and edge weights g_{ ij } is dependent on the objectives of the study. The versatility of the method is in the generality of the regularity conditions and updating algorithm for the directed graph. The regularity conditions require the following [19]:

R1.
The local thresholds α_{1},...,α_{ m } satisfy $\sum _{i=1}^{m}{\alpha}_{i}\le \alpha $.

R2.
The edge weights satisfy 0≤g_{ ij }≤1, g_{ ii }=0, and $\sum _{k=1}^{m}{g}_{\mathit{\text{ik}}}\le 1$ for all i,j=1,...,m.
The updating algorithm defines a sequentially rejective test procedure and is given as follows [19]. Note that p_{ i } represents the observed Pvalue for the test of hypothesis H_{ i }.
The proof that Algorithm 1 defines a sequentially rejective closed testing procedure which strongly controls the FWER at level α is found in the Appendix of [19], and depends directly on Theorem 1 from [20]. Both [28] and [29] claim that Theorem 1 from [20] cannot be directly applied to the hypotheses of the GO graph as the hypotheses are nested, creating logical restrictions. In their own words, [28] claim that “the shortcut procedure of [20] cannot be applied to restricted hypotheses”. Similarly, [29] state, “these methods [19] cannot make use of logical relationships between hypotheses and, as such, do not incorporate graphbased methods which exploit such relationships, such as [the Focus Level procedure] of [10]”. However, in the following we present a restricted hypotheses example where the methods of [19] can be applied. The following section sets forward some important notation and vocabulary and then demonstrates that while these claims are technically true, the methods of [19] can be applied to the Focus Level method if one of the assumptions underlying Theorem 1 of [20] is slightly relaxed. We prove this with Theorem 1.
Restricted hypotheses example
Let H_{1},...,H_{ m } denote m hypotheses of interest and call these the elementary hypotheses. Let I denote a nonempty index set such that I⊆{1,...,m} and denote an intersection hypothesis by H_{ I } where H_{ I }=∩_{i∈I}H_{ i }. The closed test procedure [17] utilizes the intersection closed set of hypotheses $\mathcal{\mathscr{H}}=\{{H}_{I}:I\subseteq \{1,...,m\},I=\varnothing \}$. In the case that the hypotheses are unrestricted, $\left\mathcal{\mathscr{H}}\right={2}^{m}1$ and Algorithm 1 of [19] is proven to hold. On the other hand, the hypotheses are restricted if for index sets I and J it is true that I≠J and H_{ I }=H_{ J } so that $\left\mathcal{H}\right<{2}^{m}1$. In this case, Algorithm 1 cannot currently be applied [28],[29].
As the hypotheses corresponding to any GO graph are always restricted, the methods of [19] cannot be applied to the GO graph under the current framework. However, the following closed test example from [28] can be extended to demonstrate how Algorithm 1 can be applied to the case of restricted hypotheses. This example sets the stage for Theorem 1, where we relax the assumptions of [20] to formally establish how the methods of [19] can indeed be applied to restricted hypotheses, and hence, the GO graph.
Consider the partially nested elementary hypotheses H_{1}, H_{2}, H_{3}, and H_{4} defined as follows for the parameters θ_{1} and θ_{2} where θ_{1},θ_{2}>0.
The full closure family of hypotheses of these four elementary hypotheses would contain 2^{4}1=15 distinct intersection hypotheses if they were unrestricted. However, the restrictions stemming from the partial nesting of H_{1} with H_{2} (H_{1}⊂H_{2}) and H_{3} with H_{4} (H_{3}⊂H_{4}) reduce the final closure to just eight distinct intersection hypotheses. For example, H_{12}=H_{1}⊂H_{2}=H_{1} and H_{34}=H_{3}⊂H_{4}=H_{3}. Computing all intersections and retaining only the disctinct intersection hypotheses shows
Each of the null parameter spaces corresponding to the hypotheses in are graphically depicted in panel (a) of Figure 3.
A closed test approach to is given in [28] which begins with the raw pvalues p_{1},p_{2},p_{3}, and p_{4} obtained from testing the original elementary hypotheses H_{1},H_{2},H_{3}, and H_{4}, each with αlevel tests, respectively. To define the closed test approach, they compute the closed test pvalues ${p}_{{H}_{i}}$ for each hypotheses H_{ i } in by the following rules. First, ${p}_{{H}_{1}}={p}_{1}$ and ${p}_{{H}_{3}}={p}_{3}$. Second, ${p}_{{H}_{2}}=max\{{p}_{1},{p}_{2}\}$ and ${p}_{{H}_{4}}=max\{{p}_{3},{p}_{4}\}$. Finally, ${p}_{{H}_{\mathit{\text{ij}}}}=min\{1,2{p}_{{H}_{i}},2{p}_{{H}_{j}}\}$, i=1,2 and j=3,4. The closed test procedure [17] is then applied to as depicted in panel (b) of Figure 3 using the closed test pvalues ${p}_{{H}_{i}}$ as explained in the following paragraph.
The closed test procedure only tests a hypothesis ${H}_{i}\in \mathcal{H}$ if all hypotheses implying H_{ i } are first rejected. For example, H_{1} can only be tested by the closed test procedure if H_{13} and H_{14} are first rejected, see panel (b) of Figure 3. In other words, the hypothesis corresponding to a child node is only tested if its parent node hypothesis is first rejected [28] state that, “this closed test procedure controls the familywise error rate strongly at level α and reflects the logical constraints among the elementary hypotheses”. We show that this closed test approach for these restricted hypotheses can be performed using the directed graph of Figure 4 and Algorithm 1 from [19].
Consider the sequential rejection procedure resulting from the application of Algorithm 1 [19] to the directed graph shown in Figure 4. Initial local thresholds of α/2 are assigned to H_{1} and H_{3} and local thresholds of zero assigned to H_{2} and H_{4} as depicted in Figure 4. The weighted edges provide for the reallocation of the local thresholds in the case of rejection of either H_{1} or H_{3}. If neither H_{1} nor H_{3} can be rejected at the α/2level, then the testing is stopped with no rejections. This corresponds to the first step of the closed test procedure described previously, as proposed in [28]. As can be seen in panel (b) of Figure 3, the closed test requires the rejection of the intersection hypothesis H_{13} before any other rejection can occur. This requires that the previously defined closed test pvalue ${p}_{{H}_{13}}=min\{2{p}_{{H}_{1}},2{p}_{{H}_{3}}\}$ satisfy ${p}_{{H}_{13}}<\alpha $. Since ${p}_{{H}_{1}}$ and ${p}_{{H}_{3}}$ were defined to be p_{1} and p_{3} respectively for this particular example, it follows that ${p}_{{H}_{13}}<\alpha $ implies 2 min{p_{1},p_{3}}<α, witnessing that the methods agree on their starting analysis using only the values of p_{1} and p_{3}. The flow chart in Figure 5 further demonstrates that the two approaches agree for all possible test scenarios and hence, that the shortcut of [19] can successfully be applied to this example of restricted hypotheses.
Definitions and preliminaries to Theorem 1
A deeper inspection of Figure 5 will reveal the reason why the shortcut from [19] can be applied to the example of restricted hypotheses of the previous section. To explain how, we must first define two terms, consonance and natural consonance.
The traditional definition of consonance[30] relies on the idea of maximal hypotheses. It states that consonance is the property of certain closed tests where rejection of an intersection hypothesis ${H}_{i}\in \mathcal{H}$ implies rejection of a maximal hypothesis $H\in \mathcal{H}$. Here, a maximal hypothesis $H\in \mathcal{H}$ is such that there is no ${H}^{\text{'}}\in \mathcal{H}$ with H^{'}⊃H. (When the closed test corresponding to the hypotheses in is depicted graphically, as in panel (b) of Figure 3, in can be seen that maximal hypotheses correspond to the leaf nodes of the graph. Further, in context of the GO graph, maximal hypotheses correspond to the leaf nodes of the graph, while the minimal hypothesis corresponds to the root node of the graph). From the example of the previous section, it can be seen that only H_{2} and H_{4} are maximal. Thus, the closed test of the example is not consonant in the traditional sense as rejection of the intersection hypothesis H_{13} does not imply the rejection of either of the maximal hypotheses H_{2} or H_{4}.
Natural consonance is a similar, but slightly more relaxed property than consonance, and differs in that it implies the rejection of only an elementary hypothesis (not necessarily a leaf node in the closure graph) whenever any other hypothesis ${H}_{i}\in \mathcal{H}$ is first rejected. This relaxed definition is more recent and is due to [28]. Importantly, it is easier for a closed test to satisfy the property of natural consonance than that of consonance. The claims of both [28] and [29] that Algorithm 1 [19] is not applicable to restricted hypotheses rest on the subtle difficulty of how consonance is defined. Note (v) following Theorem 2 of [28] claims that “consonance with respect to the elementary hypotheses [natural consonance] always implies the existence of a nested shortcut of size m”, where m is the number of elementary hypotheses. The natural consonance of the closed test allows for the shortcut from [19] to be applied to the restricted hypothesis example of the previous section, as explained in the following paragraph.
Examining the flow chart of Figure 5 will reveal that the closed test procedure proposed by [28] has this property of consonance with respect to the elementary hypotheses H_{1}, H_{2}, H_{3}, and H_{4}, i.e., the closed test for this example is naturally consonant. This follows from the fact that rejection of the intersection hypothesis H_{13} implies rejection of either of the hypotheses H_{1} or H_{3} which are two of the original four elementary hypotheses. Note as before that rejection of H_{13} requires that either 2p_{1}<α or 2p_{3}<α by the definition of ${p}_{{H}_{13}}$. If say 2p_{1}<α, then H_{13} is rejected. Further, since 2p_{1}<α, H_{14} is also rejected as ${p}_{{H}_{14}}=min\{1,2{p}_{{H}_{1}},2{p}_{{H}_{4}}\}=min\{2{p}_{1},2{p}_{{H}_{4}}\}<\alpha $. Most importantly, 2p_{1}<α provides for H_{1} to be rejected, as the closed test pvalue ${p}_{{H}_{1}}$ requires only p_{1}<α which is certainly satisfied if 2p_{1}<α. Hence, in this case, the rejection of the intersection hypothesis H_{13} implied rejection of the elementary hypothesis H_{1}. A similar scenario holds for the elementary hypothesis H_{3} if 2p_{3}<α instead of (or as well as) 2p_{1}<α. Finally, rejection of H_{24} similarly implies rejection of either H_{2} or H_{4}. Thus, the closed test procedure for these restricted hypotheses admits the shortcut of [19] because of the consonance of the closed test with respect to the elementary hypotheses, i.e. the closed test is naturally consonant.
We now extend Theorem 1 of [20] to restricted hypotheses, and thereby verify the appropriateness of the graphical shortcut of [19] for restricted hypotheses. To this end, let m elementary hypotheses H_{1},...,H_{ m } of interest be given and denote by their closure under intersection. For the purposes of Theorem 1, can be either restricted or unrestricted. Let α_{ i }(I) denote the local significance levels for an intersection hypothesis ${H}_{i}\in \mathcal{H}$ where $\underset{i\in I}{\Sigma}{\alpha}_{i}\le \alpha $ for all nonempty I⊆{1,...,m}.
Theorem1.
(Extension of Theorem 1 from [20] to restricted hypotheses.) If for ∅≠I,J⊆{1,...,m} with ∅≠H_{ I }⊂H_{ J } it holds that α_{ i }(I)≤α_{ i }(J), then the closed test for based on local Bonferroni tests is naturally consonant and a shortcut equivalent to the following procedure is possible (adapted from [19]).

0.
Set M={1,...,m}.

1.
Set I equal to the smallest subset of M such that H _{ I }=H _{ M }.

2.
Reject H _{ j } if there exists j∈I such that p _{ j }≤α _{ j }(I). If no such j exists, then stop.

3.
Set M→M\j.

4.
If M≥1 return to Step 1. Otherwise, stop.
Proof.
First, note that in the case of unrestricted hypotheses, natural consonance and consonance are identical [28] so that the proof is already demonstrated in Theorem 1 of [20]. Consider then the case of restricted hypotheses in the sense that for ∅≠I,J⊆{1,...,m} with I≠J it is true that ∅≠H_{ I }=H_{ J } so that $\left\mathcal{H}\right<{2}^{m}1$. Then, for I,J with ∅≠H_{ I }⊂H_{ J } it follows from α_{ i }(I)≤α_{ i }(J) that p_{ j }≤α_{ j }(I) implies p_{ j }≤α_{ j }(J). Thus, rejection of H_{ I } implies rejection of some elementary hypothesis H_{ j }, witnessing that the closed test for is indeed naturally consonant.
Discussion of Theorem 1
Some comments are in order regarding Theorem 1. First, while an intersection hypothesis H_{ I } may not be unique in , it must not be empty for the nested shortcut of length m to exist. Second, the only difference between the proof here and the proof for unrestricted hypotheses [20] is in the definition of consonance. Here we follow the suggestion in [28] and allow natural consonance, which can be seen as a loosening of the requirements of consonance to include all elementary hypotheses instead of just all maximal hypotheses. The important distinction is that for unrestricted hypotheses, all elementary hypotheses are maximal. The same is not necessarily true for restricted hypotheses. Third, as in the previous restricted hypotheses example, restricted hypotheses are often the result of nested elementary hypotheses. This is certainly the case for the hypotheses attached to the gene sets of the GO graphs. Fourth, the main importance of the extended Theorem 1 rests with its assurance that a naturally consonant closed test based on weighted Bonferroni tests exists so long as the monotonicity condition α_{ i }(I)≤α_{ i }(J) is satisfied for all ∅≠H_{ i }⊂H_{ J } in . Fifth, Theorem 1 does not specify that any graph with local thresholds of α=(α_{1},...,α_{ m }) and edge weights G={g}_{ ij }, denoted by (α,G), can combine with Algorithm 1 and lead to a consonant closed test. It simply specifies the conditions under which a consonant closed test based on local Bonferroni tests can be formed.
One important rule on the graph (α,G) when the hypotheses are restricted is that the local threshold α_{ i } for an elementary hypothesis H_{ i } must remain zero until all elementary hypotheses H_{ j } with H_{ j }⊂H_{ i } are first rejected. This property can be seen to hold for the graph of Figure 4. However, if the graph in Figure 4 allowed for any of H_{1}’s threshold to be passed to H_{4} or similarly, if H_{3} allowed for anything to be passed to H_{2}, this property would no longer hold. So, while Theorem 1 assures that a consonant closed test exists when local Bonferroni tests are used for the testing of each $H\in \mathcal{H}$, not just any graph (α,G) will result in that consonant closed test. In the following section we demonstrate how a graph (α,G) can be applied to the GO graph such that a consonant closed test based on weighted Bonferroni tests is achieved through the application of Algorithm 1.
That Algorithm 1, when applied to a graph (α,G), preserves the monotonic property that α_{ j }(I)α≤_{ j }(J) for I,J such that H_{ I }⊂H_{ J } can be seen by noting that Algorithm 1 only provides for the local thresholds α_{ i } to remain the same size or increase. Never does it allow for them to become smaller. Further, at any point in the iterative process, the local thresholds α_{ i } define the weighted Bonferroni test thresholds α_{ j }(I) for the intersection hypothesis I corresponding to the intersection of the elementary hypotheses with nonzero thresholds (see for example Figure 5). Hence, as H_{ J } will be tested only after H_{ I } is first rejected whenever H_{ I }⊂H_{ J }, it follows that Algorithm 1 will provide α_{ j }(I)≤α_{ j }(J).
The Shortfocus level procedure
We obtain the Short Focus Level procedure by modifying the topdown portion of the Focus Level method. This is done by tailoring the general graphical shortcut [19] to a GO graph as follows. Label the m hypotheses corresponding to the test of significance for each GO term (gene set) as H_{1},...,H_{ m } starting with the root node and proceeding in an organized manner through each level of the GO graph, ending with the terminal nodes. (The precise ordering is not important.) Let F⊂M={1,...,m} denote the index set of the nodes corresponding to the preselected focus level of the GO graph. For all m_{ F } nodes in the focus level, assign local significance levels of α_{ i }=∈/m_{ F } to each hypothesis H_{ i } with i∈F. Assign initial local significance levels of 0 to all children nodes of the focus level. Note that nodes above the focus level will still be tested using the bottomup approach of the Focus Level method and are not considered when applying the topdown portion of the method.
Using the structure of the GO graph, assign to each edge from parent node i to child node j a weight of g_{ ij }=1/m_{ i }, where m_{ i } denotes the number of children nodes of node i. After all edge weights have been assigned for the edges defined by the GO graph, all terminal nodes are individually joined with m_{ F } new edges to each of the m_{ F } focus level nodes. These new edges are given weights of 1/m_{ F }. (In the case that a terminal node is also a focus level node, then edges are made only to all other focus level nodes with weight 1/(m_{ F }1)).
At this point, a modified form of Algorithm 1 of [19] is applied to the resulting directed graph to obtain the final set of significant hypotheses. The modifications ensure that no child node is tested before all parent nodes are first found significant, maintaining the strong control of the FWER under the restricted hypotheses of the GO graph as well as maintaining Property FL2 of the basic assumptions (or requirements) underlying the Focus Level method (see “Methods” section above). Figure 6 demonstrates the application of the described graphical Bonferroni adjustment to the topdown portion of the Example GO graphs of Figure 1. Comparing Figure 2 to Figure 6 provides a heuristic understanding of how the new topdown approach is computationally faster than the original closure approach because no new nodes need to be created.
An algorithm which implements the Short Focus Level procedure is detailed in Algorithm 2. Here, H denotes the index set of testable hypotheses (nodes) and w={w_{ i }}_{i∈H} the corresponding set of weights such that α/w_{ i } provides the local thresholds α_{ i } for each hypothesis H_{ i } indexed by i∈H. As described previously, F⊂{1,...,m} denotes the index set of all preselected focus level nodes. The notation C_{ i } denotes the index set of children nodes of the parent hypothesis H_{ i }. Similarly, the notations P_{ i } and A_{ i } denote the parents and all ancestors, respectively, of the node corresponding to the hypothesis H_{ i }. Finally, we use R and S to denote the index sets of the current and cumulative rejected hypotheses, respectively.
Results and discussion
A natural question at this point concerns the advantages and disadvantages of changing the topdown portion of the Focus Level procedure from the original closed test approach as in [10] to the graphical shortcut of [19] as proposed for the Short Focus Level. If the local tests for each intersection hypothesis were originally performed with weighted Bonferroni tests, then the difference between the methods would be that the first performed the full closure test requiring the testing of somewhere on the order of 2^{m}1 intersection hypotheses, while the second, which applies a shortcut, would test no more than m hypotheses with no reduction in the power of the tests. When using the Global Test for each intersection hypothesis as suggested by [10], the answer to the differences in computation time and power is not as clear. The following simulations demonstrate that neither method is uniformly more powerful than the other, with each having the advantage for certain scenarios. However, as these simulations demonstrate, the newly proposed Short Focus Level procedure is uniformly (and exponentially) computationally faster than the Focus Level method which will hopefully better enable its use by practitioners.
Simulation 1
The following simulation based on the toy GO graph depicted in Figure 7 panel (b) demonstrates the advantages and disadvantages of moving to the newly proposed graphical shortcut of [19] in the topdown portion of the Focus Level procedure. The simulation was performed with the phenotype Y as a dichotomous class variable (say, treatment and control) and the data X representing an RNASeq counts matrix with rows as genes (m) and columns as samples (n). The number of samples belonging to the treatment group was simulated according to a binomial(n, 0.5) distribution, where n is the total number of samples, with the added rule that at least two samples were in each group. This allowed for unbalanced data, with the tendency towards fairly balanced designs. Separate simulations for sample sizes of n=5, 20, and 100 were performed.
The structure of gene assignments to the sets A, B, C, D, E, and F of Figure 4, as well as the total number of genes assigned, was allowed to vary in each simulation according to certain parameters. Genes were first assigned to the leaf node gene sets C, D, and E. This was accomplished by randomly selecting both the number of distinct sets in each of these sets (anywhere from 1 to a maximum specified size of either 10 or 40) as well as the number of genes shared by all possible combinations of the leaf node gene sets. Common genes between all or many gene sets was discouraged with small probabilities of occurrence, while common genes between a few gene sets was allowed to occur more frequently. Following the assignments of genes to leaf nodes, parent nodes were randomly assigned new genes (anywhere from 1 to the maximum specified size) as well as all genes contained by their children nodes. The result was a nested graph with at least some overlap common to many gene sets, as is the case within GO Graphs.
The data counts matrix X was simulated using an actual RNASeq data set as a sampling distribution for the pergene means in the control group. Specifically, the counts k_{ ij } for all samples j assigned to the control group were generated from a NB(${\mu}_{i},{\mu}_{i}+{\mu}_{i}^{2}/d$) distribution, where the means μ_{ i } were randomly sampled from the pergene means of the control group from the actual RNASeq data set. The scaling parameter d was set at 10 for all simulations. Leaf node gene sets (any of nodes C, D, or E in Figure 4) were then selected at random to be significant. Each gene mapping to the selected significant leaf nodes was assigned a treatment mean of where μ_{ i } denotes the mean sampled from the actual RNASeq data for gene i and β_{ i } was an effect size obtained from a Poisson(λ) distribution with the parameter λ set to one of 0, 1, 2, or 3. Thus, not all genes in the significant gene sets necessarily had nonzero effect sizes. The actual counts k_{ ij } for all samples j assigned to the treatment group were obtained from a NB(, ) distribution where, as with the control group, d=10 was constant across all simulations. (See [13] for a similar simulation approach where single gene sets were the object of interest as opposed to an entire GO graphs as in this simulation).
The averaged results of Simulation 1 are presented in Table 1. This example shows greater power for the current implementation of the Focus Level procedure where the Globaltest [6] is used to test all intersection hypotheses and all elementary hypotheses. The greatest power differences of the two methods appear for small sample sizes, n=5 in this simulation, and for nodes with relatively few child nodes. The power of the two methods is comparable otherwise. Importantly, the computation time for the Short Focus Level procedure is significantly faster, even for this extremely small toy GO graph whose closure contains just 14 nodes. Interestingly, the Focus Level procedure as it is currently implemented seems to operate best, computationally speaking, when the sample size is moderate, n=20 in this simulation.
;Simulation 2;
A second simulation study using the toy GO graph of Figure 8 was also used to compare power and computation time of the original Focus Level method to the Short Focus Level. The closure of the toy GO graph in Figure 8 is more complex than that of the previous simulation, containing 574 nodes as compared to the 14 of Figure 7, panel (a). This simulation considered the continuous phenotype Y∼N(0,1) and its correlation with simulated gene expression values X. For this simulation m=100 genes were partitioned to the 14 GO IDs of Figure 8 as specified in Table 2. Expression values X_{ ij } for each sample i=1,...,n and gene j=1,...,m were generated as N(0,1) variates. GO IDs 6, 7, and 13 were designated as significant by adding r Y, r∈[0,1], to the expression values of the corresponding genes (i.e., the columns of X corresponding to genes in GO IDs 6, 7, and 13). Thus, by inherentance, GO IDs 1, 2, 3, 4, 10, and 11 were also significantly associated with the phenotype Y. Values of r close to 1 provided a strong signal and greater power for detection while r near zero resulted in a very weak signal and correspondingly very low power for detection. Goeman’s Globaltest [6] was used to test each GO ID for association with the phenotypic variable Y. Given that Simulation 1 suggested that the current Focus Level procedure performs best at a moderate sample size, n=20 was used for this simulation.
Power and computation time were averaged over 1,000 simulations. Results are presented in Table 3 for the most interesting case of r=0.5. They show the Short Focus Level method having greater power at every GO ID. The computational speed advantage of the improvement is also manifest, showing nearly a 15,000 fold increase in speed over the current Focus Level procedure. This second simulation emphasizes the fact that neither approach to the Focus Level procedure is uniformly more powerful than the other. While it is clear that each has the advantage in certain scenarios, at least theoretically, more work needs to be completed to determine exactly where each is most appropriate. Practically speaking however, the computational advantage and similar statistical power (on average) of the Short Focus Level should solicit its use except perhaps for choices of the focus level near the leaf nodes of the graph where the current Focus Level method is computationally tractable.
Key difference between focus level and short focus level methods
The Focus Level (FL) and Short Focus Level (SFL) methods are the same in the bottomup approach. They differ in the topdown approach. Both are similar in the topdown approach in that they apply the closed testing strategy to the GO graph (from the focus level down to the leaf nodes) by closing the graph under all unions of gene IDs corresponding to the GO IDs (i.e., intersections of the hypotheses corresponding to the testing of each GO ID). The FL method uses the Global Test to test each intersection hypothesis of the closed GO graph[10]. The SFL method would be a direct shortcut of the FL method if the FL method instead used a weighted Bonferroni test to test each intersection hypothesis.
However, it is important to recognize that the FL method could only perform the full closure test, not the shortcut that the SFL method performs, even if the FL method was modified to use the weighted Bonferroni tests. The FL method is more consistent in applying the same Global Test to both original GO ID hypotheses as well as to all intersection hypotheses. However, it is computationally expensive because it performs the full closure test. The SFL method makes a slight shift in allowing any test (not just the Global Test) for the elementary hypotheses (the individual GO ID hypotheses) and then performing weighted Bonferroni tests for all intersection hypotheses. This simplification or generalization allows for the resulting shortcut. Hence, the power comparison between the FL and SFL methods is not obvious, and the simulations above show that neither method is uniformly more powerful than the other.
Real data application 1
A drawback to the otherwise powerful Focus Level method is the computational burden which prohibits the full topdown approach from being applied to real data sets [10]. When no a priori focus level exists, as is often the case [18], the root node of the GO graph is a logical default choice, but requires the full topdown approach. Under the newly proposed Short Focus Level method, this is now a computational possibility. The following application to RNASeq counts data from porcine oocytes demonstrates the performance of the full topdown approach of the Short Focus Level procedure to real data. The Biological Process (BP) root node was selected as the focus level for this study due to there being no other focus level of greater a priori interest.
It is well known that in vivo (naturally) maturated oocytes show far greater developmental competence than do those matured in vitro[31]. Yet, the underlying genetics are still not well understood. To uncover the genetic differences of in vitro matured oocytes as compared to those matured naturally (in vivo), transcript counts for 4 in vivo and 4 in vitro maturated porcine oocytes were obtained using the Illumina RNASeq platform [32]. Lanes were populated as shown in Table 4. These data from the lab of Dr. Clay Isom of the Utah State University Department of Animal, Dairy, and Veterinary Sciences are reported on here with permission. In this oocyte study, all animal procedures were performed with the strictest adherence to animal welfare guidelines and with regulatory oversight by the Institutional Animal Care and Use Committee at Utah State University.
Individual Pvalues testing the differential expression of 12,625 genes were calculated using the DESeq package of Bioconductor [33],[34] with pig mother, as identified in Table 4, included as a covariate. Specifically, these Pvalues were obtained under the null hypotheses that the pergene expression strength of the in vivo maturated oocytes (IVV) is equal to that of the in vitro maturated oocytes (IVV) when accounting for any pig mother effect. This was done through the DESeq package [33] which compares a full model (regressing the RNASeq counts on both the oocyte type and pig mother by a generalized linear model) to a reduced model (regressing only on the pig mother) to determine significance for a given gene.
A gene set analysis using the GO BP ontology was then performed to characterize differentially expressed gene products between the two types of oocytes (IVV and IVM). Pvalues for each of 5,687 BP GO Terms containing at least 5 of the 12,625 Entrez IDs from the single gene (DESeq) analysis were calculated using Stouffer’s Method [24],[35]. The R code [21] for Stouffer’s Method is included in the mvGSTmvGST package [22],[23]; see also Additional file 1. Briefly, Stouffer’s method transforms each of the Pvalues (from the single gene analysis) corresponding to an individual gene in the gene set to a standard normal Zscore. A single Pvalue for the gene set is then obtained from the mean of the Zscores by computing the appropriate tail probability (from a standard normal distribution) beyond the mean Zscore. Stouffer’s Pvalue combination method was applied here as it is more powerful for the consensus alternative than say Fisher’s Pvalue combination test [36] or Goeman’s globaltest [6], see discussions in [24]. Finally, multiplicity adjusted gene set Pvalues for each BP term were calculated using the Short Focus Level procedure, with the root BP GO term (GO:0008510) as the focus level. This adjustment (the full topdown approach) took just 3 minutes and 23 seconds of processing time on an Intel Pentium M 1.86 GHz processor with 1 GB of RAM. The current Focus Level method is computationally intractable for thesedata.
Figure 9 reports the significant subgraph [10] obtained from the Short Focus Level method containing 113 of the original 5,687 BP terms. While a partial legend is included in Figure 9, a full legend identifying all 113 significant BP terms, along with their multiplicityadjusted Pvalues (using the Short Focus Level method at familywise error rate 0.05) is included as a table in Additional file 3. Since the full topdown approach was performed, these GO terms, which are differentially expressed between the two types of oocytes (IVV and IVM), can be discussed either individually or within their context of this significant subgraph. Advantaged by the FWER control of the Short Focus Level procedure, any subset of the significant results can also be reported on (while the others ignored) with the assurance that the FWER remains controlled at the specified level for the selected sets. Possible interpretation discussions of the results include the significant differential activity (between in vivo and in vitro maturated oocytes) of biological processes “response to bacterium” (node 74 in Figure 9), “doublestrand break repair” (node 110), and “ribonucleoside metabolic process” (node 93), among others.
Real data application 2
As a second real data demonstration of the Short Focus Level method, and to further discuss differences between the Short Focus Level and Focus Level methods, we used a subset of the famous Golub data set [37], specifically the 38 training samples publicly available in the R package golubEsets [38]. Briefly, Affymetrix Hgu6800 chips were used to profile gene expression in leukemia patients, and we test for differential biological process activity between 27 patients with acute lymphoblastic leukemia (ALL) and 11 with acute myeloid leukemia (AML). Using the same Focus Level method demonstration as in the vignette of the globaltest package [39], we looked only at biological process “cell cycle” (GO:0007049) and its descendants in the GO graph, with 249 total nodes.
We also applied the Short Focus Level method to the same example (using familywise error rate 0.01), and note that while the Focus Level method (with its default focus level) takes nearly 4 minutes, the Short Focus Level method (using the same focus level as selected by the Focus Level method) takes 1 second. Finally, we applied the Short Focus Level method using the root node as the focus level, which also took 1 second. In stark contrast, the Focus Level method using the root node as the focus level was deemed computationally intractable, as even a run time of two weeks was not sufficient to complete it.
Figure 10 compares the resulting adjusted pvalues for each of the 249 biological processes considered. Figure 10a shows that, when using the same focus level, the Focus Level (FL) and Short Focus Level (SFL) methods can result in different (though largely overlapping in this case) sets of GO terms called significant. This results from the previously discussed key difference between the FL and SFL methods, namely that the SFL method allows any test (not just the Global Test) for the elementary hypotheses (the individual GO ID hypotheses) and then performs weighted Bonferroni tests for all intersection hypotheses.
As discussed in the original Focus Level paper [10], different focus levels provide for different power at differing areas of the GO graph. For this reason, it is difficult to make definitive comparisons using results from different focus levels. While the stronger agreement seen between the FL method (with its default focus level) and SFL method (with the root node focus level) in Figure 10 may seem interesting, the important point is that the FL method is effectively computationally intractable using the root node focus level. The decision to use the FL or SFL method should not be based on power considerations, but rather on computational considerations, especially when no real reason exists to choose the focus level anywhere other than the root node (which will most often be the case), in which case the FL method is computationally intractable. However, the SFL method is computationally efficient and strongly controls the familywise error rate within the structure of the GO graph.
Conclusions
As pointed out in [10], the GO graphs are structured and “it is wasteful not to make use of that structure” in correcting for multiplicity. Further, they stress the importance of not making any assumptions on the joint distribution of the test statistics corresponding to each of the gene sets in the GO graph while correcting for multiplicity. The Focus Level procedure both avoids any such assumptions and capitalizes on the inherent structure of the GO graph to adjust for the multiple tests performed, resulting in a powerful approach. Another advantage of the Focus Level method is the possibility of incorporating biological knowledge into the adjustment approach through the selection of the focus level, where the method has the greatest power.
This work improves upon the Focus Level procedure of [10] by altering the topdown portion of the method to utilize the graphical shortcut of [19] in place of the full closed testing approach of [17] as originally suggested by [10]. This was made possible by extending the result from [20] to restricted hypotheses (Theorem 1) as the hypotheses corresponding to the GO graph are always restricted.
The main advantage of the Short Focus Level procedure proposed in this work is the exponential decrease in computational burden. This provides for the most logical default choice of the root node of the GO graph as the focus level when no other a priori choice can be specified. Another advantage of the improvement is in the ability to consider the adjusted Pvalues apart from their context within the significant subgraph of the full GO graph under the full topdown approach. When the focus level is selected to be anything other than the root node, individual hypotheses must be considered in context of their position within the significant subgraph. However, this is not altogether a disadvantage as “the interpretation of an individual adjusted Pvalue should depend on the location in the graph where it occurs” [10].
It is our hope that this shortcut for the Focus Level procedure, the Short Focus Level, will result in more widespread use of the method. Still, future work remains to be done. The simulations performed within this work demonstrate that each approach appears to be more powerful under different circumstances. Hence, further theoretical work is needed to determine the conditions under which each method is most powerful.
Additional files
References
 1.
Malone JH, Oliver B: Microarrays, deep sequencing and the true measure of the transcriptome. BMC Biol. 2011, 9: 3410.1186/17417007934.
 2.
Lee CK, Klopp RG, Weindruch R, Prolla TA: Gene expression profile of aging and its retardation by caloric restriction. Science. 1999, 285: 13901393. 10.1126/science.285.5432.1390.
 3.
Alizadeh AA, Eisen MB, Davis RE, Ma C, Lossos IS, Rosenwald A, Boldrick JC, Sabet H, Tran T, Yu X, Powell JI, Yang L, Marti GE, Moore T, Lu L, Lewis DB, Tibshirani R, Sherlock G, Chan WC, Greiner TC, Weisenburger DD, Armitage JO, Warnke R, Levy R, Wilson W, Grever MR, Byrd JC, Botstein D, Brown PO, et al: Distinct types of diffuse large Bcell lymphoma identified by gene expression profiling. Nature. 2000, 403 (6769): 503511. 10.1038/35000501.
 4.
Isom SC, Stevens JR, Li R, Spollen WG, Cox L, Spate LD, Murphy CN, Prather RS: Transcriptional profiling by RNAseq of periattachment porcine embryos generated by a variety of assisted reproductive technologies. Physiol Genomics. 2013, 45 (14): 577589. 10.1152/physiolgenomics.00094.2012.
 5.
Miller JA, Woltjer RL, Goodenbour JM, Horvath S, Geschwind GH: Genes and pathways underlying regional and cell type changes in alzheimer’s disease. Genome Med. 2013, 5 (5): 4810.1186/gm452.
 6.
Goeman JJ, van de Geer SA, de Kort F, van Houwelingen HC: A global test for groups of genes: testing association in a clinical outcome. Bioinformatics. 2004, 20: 9399. 10.1093/bioinformatics/btg382.
 7.
Efron B, Tibshirani R: On testing the significance of sets of genes. Ann Appl Stat2007:107129.
 8.
Ashburner M, Ball CA, Blake JA, Botstein D, Butler H, Cherry JM, Davis AP, Dolinski K, Dwight SS, Eppig JT, Harris MA, Hill DP, IsselTarver L, Kasarskis A, Lewis S, Matese JC, Richardson JE, Ringwald M, Rubin GM, Sherlock G: Gene ontology: tool for the unification of biology. Nat Genet. 2000, 25: 2529. 10.1038/75556.
 9.
Passtoors WM, Boer JM, Goeman JJ, Akker EB, Deelen J, Zwaan BJ, Scarborough A, van der Breggen R, Vossen RH, HouwingDuistermaat JJ, Ommen GJ, Westendorp RG, van Heemst D, de Craen AJ, White AJ, Gunn DA, Beekman M, Slagboom PE: Transcriptional profiling of human familial longevity indicates a role for ASF1A and IL7R. PLOS ONE. 2012, 7: [E27759]10.1371/journal.pone.0027759.
 10.
Goeman JJ, Mansmann U: Multiple testing on the directed acyclic graph of gene ontology. Bioinformatics. 2008, 24 (4): 537544. 10.1093/bioinformatics/btm628.
 11.
Goeman JJ, Buhlmann P: Analyzing gene expression data in terms of gene sets: methodological issues. Bioinformatics. 2007, 23 (8): 980987. 10.1093/bioinformatics/btm051.
 12.
Khatri P, Drăghici S: Ontological analysis of gene expression data current tools, limitations, and open problems. Bioinformatics. 2005, 18: 35873595. 10.1093/bioinformatics/bti565.
 13.
Fridley BL, Jenkins GD, Biernacka JM: Selfcontained geneset analysis of expression data: an evaluation of existing and novel methods. PLoS One. 2010, 5 (9): [E12693]10.1371/journal.pone.0012693.
 14.
Nettleton D, Recknor J, Reecy JM: Identification of differentially expressed gene categories in microarray studies using nonparametric multivariate analysis. Bioinformatics. 2008, 24 (2): 192201. 10.1093/bioinformatics/btm583.
 15.
Hochberg Y, Tamhane AC: Multiple Comparison Procedures, 1st edition. 1987, Wiley, New York
 16.
Holm S: A simple sequentially rejective multiple test procedure. Scand J Stat. 1979, 6: 6570.
 17.
Marcus R, Peritz E, Gabriel KR: On closed testing procedures with special reference to ordered analysis of variance. Biometrika. 1976, 63 (3): 655660. 10.1093/biomet/63.3.655.
 18.
Liang K, Nettleton D: A hidden Markov model approach to testing multiple hypotheses on a treetransformed gene ontology graph. J Am Stat Assoc. 2010, 105 (492): 14441454. 10.1198/jasa.2010.tm10195.
 19.
Bretz F, Maurer W, Brannath W, Posch M: A graphical approach to sequentially rejective multiple test procedures. Stat Med. 2009, 28 (4): 586604. 10.1002/sim.3495.
 20.
Hommel G, Bretz F, Maurer W: Powerful shortcuts for multiple testing procedures with special reference to gatekeeping strategies. Stat Med. 2007, 26 (22): 40634073. 10.1002/sim.2873.
 21.
R: A Language and Environment for Statistical Computing. 2013, R Foundation for Statistical Computing, Vienna
 22.
Mecham DS: mvGST: Tools for Multivariate and Directional Gene Set Testing. 2014, Department of Mathematics and Statistics, [MS Report]. Logan: Utah State University
 23.
Stevens JR, Mecham DS: mvGST: Multivariate and directional gene set testing2014. [R package version 1.0.0]. [], [http://www.bioconductor.org/packages/release/bioc/html/mvGST.html]
 24.
Stevens JR, Isom SC: Gene set testing to characterize multivariately differentially expressed genes. Proc Conf Appl Stat Agric. 2012, 24: 125137.
 25.
Hummel M, Meister R, Mansmann U: Global ANCOVA: exploration and assessment of gene group effects. Bioinformatics. 2008, 24: 7885. 10.1093/bioinformatics/btm531.
 26.
Tomfohr J, Lu J, Kepler TB: Pathway level analysis of gene expression using singular value decomposition. BMC Bioinformatics. 2005, 6: 22510.1186/147121056225.
 27.
Dinu I, Potter JD, Mueller T, Liu Q, Adewale AJ, Jhangri GS, Einecke G, Famulski KS, Halloran P, Yasui Y: Improving gene set analysis of microarray data by SAMGS. BMC Bioinformatics. 2007, 8: 24210.1186/147121058242.
 28.
Brannath W, Bretz F: Shortcuts for locally consonant closed test procedures. J Am Stat Assoc. 2010, 105 (490): 660669. 10.1198/jasa.2010.tm08127.
 29.
Goeman JJ, Solari A: The sequential rejection principle of familywise error control. Ann Stat. 2010, 38 (6): 37823810. 10.1214/10AOS829.
 30.
Gabriel KR: Simultaneous test procedures’some theory of multiple comparisons. Ann Math Stat. 1969, 40: 224250. 10.1214/aoms/1177697819.
 31.
Cox L, Ward A, Saunders G, Stevens JR, Isom SC: Gene expression analysis ofin vivoandin vitromatured porcine metaphase II oocytes. Reprod Fertil Dev. 2013, 26 (1): 117117. 10.1071/RDv26n1Ab6.
 32.
Wang Z, Gerstein M, Snyder M: RNASeq: a revolutionary tool for transcriptomics. Nat Rev Genet. 2009, 10: 5763. 10.1038/nrg2484.
 33.
Anders S, Huber W: Differential expression analysis for sequence count data. Genome Biol. 2010, 11 (10): R10610.1186/gb20101110r106.
 34.
Gentleman RC, Carey VJ, Bates DM, Bolstad B, Dettling M, Dudoit S, Ellis B, Gautier L, Ge Y, Gentry J, Hornik K, Hothorn T, Huber W, Iacus S, Irizarry R, Leisch F, Li C, Maechler M, Rossini AJ, Sawitzki G, Smith C, Smyth G, Tierney L, Yang JY, Zhang J: Bioconductor: open software development for computational biology and bioinformatics. Genome Biol. 2004, 10: R8010.1186/gb2004510r80.
 35.
Stouffer S, Suchman E, DeVinney L, Star S, Williams RJ: The American Soldier. 1949, Princeton University Press, Princeton
 36.
Fisher RA: Statistical Methods for Research Workers. 1973, Hafner Publishing Company; (1st ed. in 1925) 14th edition, New York
 37.
Golub TR, Slonim DK, Tamayo P, Huard C, Gaasenbeek M, Mesirov JP, Coller H, Loh ML, Downing JR, Caligiuri MA, Bloomfield CD, Lander ES: Molecular classification of cancer: class discovery and class prediction by gene expression monitoring. Science. 1999, 286: 531537. 10.1126/science.286.5439.531.
 38.
Golub T: golubEsets: exprSets for Golub leukemia data2014. [R package version 1.6.0]., [http://www.bioconductor.org/packages/release/data/experiment/html/golubEsets.html]
 39.
Goeman JJ, Oosting J: globaltest2014. [R package version 5.18.0]. [], [http://www.bioconductor.org/packages/release/bioc/html/globaltest.html]
Acknowledgements
This research was supported by the Utah Agricultural Experiment Station (UAES), Utah State University, and approved as journal paper number 8676, with funding as UAES project number UTA01062, associated with the W2112 multistate project “Reproductive Performance in Domestic Ruminants”.
Author information
Affiliations
Corresponding author
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors’ contributions
GS developed the Short Focus Level procedure and performed the simulation studies and real data analysis as well as the original draft of the manuscript. JRS consulted with GS in the simulation design, real data analysis, and development of the method and helped draft the final manuscript. SCI designed the porcine embryo experiment, collected all corresponding data and provided the biological interpretations as well as helped draft the final manuscript. All authors read and approved the final manuscript.
Electronic supplementary material
Authors’ original submitted files for images
Below are the links to the authors’ original submitted files for images.
Rights and permissions
About this article
Cite this article
Saunders, G., Stevens, J.R. & Isom, S.C. A shortcut for multiple testing on the directed acyclic graph of gene ontology. BMC Bioinformatics 15, 349 (2014). https://doi.org/10.1186/s1285901403493
Received:
Accepted:
Published:
Keywords
 Bonferroni
 Holm
 Gene ontology
 Multiple testing