# Basic properties and information theory of Audic-Claverie statistic for analyzing cDNA arrays

- Peter Tiňo
^{1}Email author

**10**:310

**DOI: **10.1186/1471-2105-10-310

© Tiňo; licensee BioMed Central Ltd. 2009

**Received: **25 March 2009

**Accepted: **23 September 2009

**Published: **23 September 2009

## Abstract

### Background

The Audic-Claverie method [1] has been and still continues to be a popular approach for detection of differentially expressed genes in the SAGE framework. The method is based on the assumption that under the null hypothesis tag counts of the same gene in two libraries come from the same but unknown Poisson distribution. The problem is that each SAGE library represents only a single measurement. We ask: Given that the tag count samples from SAGE libraries are extremely limited, how useful actually is the Audic-Claverie methodology? We rigorously analyze the A-C statistic that forms a backbone of the methodology and represents our knowledge of the underlying tag generating process based on one observation.

### Results

We show that the A-C statistic and the underlying Poisson distribution of the tag counts share the same mode structure. Moreover, the K-L divergence from the true unknown Poisson distribution to the A-C statistic is minimized when the A-C statistic is conditioned on the mode of the Poisson distribution. Most importantly, the expectation of this K-L divergence never exceeds 1/2 bit.

### Conclusion

A rigorous underpinning of the Audic-Claverie methodology has been missing. Our results constitute a rigorous argument supporting the use of Audic-Claverie method even though the SAGE libraries represent very sparse samples.

## Background

It is of utmost importance for biologists to be able to analyze patterns of expression levels of selected genes in different tissues possibly obtained under different conditions or treatment regimes. Even subtle changes in gene expression levels can be indicators of biologically crucial processes such as cell differentiation and cell specialization [2]. Measurement of gene expression levels can be performed either via hybridization to microarrays, or by counting gene tags (signatures) using e.g. Serial Analysis of Gene Expression (SAGE) [3] or Massively Parallel Signature Sequencing (MPSS) [4] methodologies. The SAGE procedure results in a library of short sequence tags, each representing an expressed gene. The key assumption is that every mRNA copy in the tissue has the same chance of ending up as a tag in the library. Selecting a specific tag from the pool of transcripts can be approximately considered as sampling with replacement. The key step in many SAGE studies is identification of "interesting" genes, typically those that are differentially expressed under different conditions/treatments. This is done by comparing the number of specific tags found in the two SAGE libraries corresponding to different conditions or treatments. Several statistical tests have been suggested for identifying differentially expressed genes through comparing such digital expression profiles, e.g. [1, 2, 5, 6].

*N*of clones, the probability of observing

*x*tags of the same gene will be well-approximated by the Poisson distribution parametrized by

*λ*≥ 0.

The unknown parameter *λ* signifies the number of transcripts of the given type (tag) per *N* clones in the cDNA library. When comparing two libraries, it is assumed that under the null hypothesis of not differentially expressed genes the tag count *x* in one library comes from the same underlying Poisson distribution *P*(·|*λ*) as the tag count *y* in the other library. However, each SAGE library represents a single measurement only. From a purely statistical standpoint resolving this issue is potentially quite problematic. One can be excused for being rather skeptical about how much can actually be learned about the underlying unknown Poisson distribution from a single observation.

*y|x*) over tag counts

*y*in one library informed by the tag count

*x*in the other library, under the null hypothesis that the tag counts are generated from the same but unknown Poisson distribution. (

*y|x*) is obtained by Bayesian averaging (infinite mixture) of all possible Poisson distributions

*P*(

*y*|

*λ'*) with mixing proportions equal to the posteriors

*p*(

*λ'*|

*x*) under the flat prior over

*λ*. When the two libraries are of the same size, we obtain [1]:

We will refer to
(*y|x*) as *Audic-Claverie statistic* (A-C statistic) based on counts *x* and *y*. Note that
(*y*|*x*) is symmetric, i.e. for *x, y* ≥ 0,
(*y|x*) =
(*x|y*). Audic and Claverie [1] point out that this is a desirable property, since if the counts *x, y* are related to two libraries of the same size, they should be interchangeable when analyzing whether they come from the same underlying process or not. The A-C statistic
(*y|x*) can be used e.g. for principled inferences, construction of confidence intervals, statistical testing etc. For further details regarding the derivation and mathematical treatment of the A-C statistic see [1].

*how much*and

*in what form*can be actually learned about the distribution in the Audic-Claverie framework, given a single observation provided by a SAGE library. In particular, we ask:

- 1.
How natural is the A-C statistic's representation of the underlying unknown Poisson distribution governing the tag counts?

- 2.
Given that the observed tag count sample is very limited, how well can the Audic-Claverie approach work, i.e. how well does the A-C statistic capture the underlying Poisson distribution?

## Methods

### Basic properties of the A-C statistic

In this section we answer the first question posed above. It turns out that the A-C statistic and the underlying Poisson distribution are quite similar in their nature: for any (integer) mean tag count *λ* ≥ 1, the Poisson distribution *P*(·|*λ*) has two neighboring modes located at *λ* and *λ* - 1, with *P*(*λ*|*λ*) = *P*(*λ* - 1|*λ*). When it comes to the observed tag counts, given a count *x* ≥ 1, the A-C statistic
(*y|x*) has two neighboring modes, one located at *y* = *x*, the other at *y* = *x* - 1, with
(*x|x*) =
(*x* - 1|*x*). As in Poisson distribution, the values of
(*y|x*) decrease as one moves away from the modes in both directions.

**Theorem 1**
*Let x, y and d be integers with ranges specified below. It holds:*

*1*.
(*x|x*) >
(*x* + *d|x*) *for any x* ≥ 0 *and d* ≥ 1.

*2. For x* ≥ 1,
(*x|x*) =
(*x* - 1|*x*).

*3*.
(*x|x*) >
(*x - d|x*) *for any x* ≥ 2 *and* 2 ≤ *d* ≤ *x*.

This follows from 2(*x - d* + *j*) *<* 2*x - d* + *j*, which can be easily verified, since for *j* ∈ {1, 2,..., *d* - 1}, we have (*j - d*) > 2·(*j - d*).

For *j* = *d*, we have the equality (2*x - d* + *j*)/(*x - d* + *j*) = 2.

Q.E.D

We have shown that after observing a count *x*, the A-C statistic expects counts *y* = *x* and *y* = *x* - 1 with the highest and equal probability. The other values of count *y* are, as one would naturally expect, less probable.

*y|x*) and the corresponding Poisson distribution

*P*(

*y|λ*) at

*λ*=

*x*for two values of

*x*,

*x*= 10 and

*x*= 30. As a result of Bayesian averaging in the A-C statistic, (

*y|x*) is less peaked at its modes than the Poisson counterpart

*P*(

*y|x*). However, both (

*y|x*) and

*P*(

*y|x*) have two modes located at

*x*and

*x*- 1.

### Information theory of the A-C statistic

We now answer, in the framework of information theory, the second question posed in the 'Background' section. Assume that there is some "true" underlying Poisson distribution *P*(*y|λ*) (1) over possible counts *y* ≥ 0 with unknown parameter *λ*. In the same process, we first generate a count *x* and then use the A-C statistic
(*y|x*) (3) to define a distribution over *y*, given the already observed count *x*. We ask: How different, in terms of Kullback-Leibler (K-L) divergence, are the two distributions over *y*? For the A-C statistic to work, one would naturally like
(*y|x*) to be sufficiently representative of the true unknown distribution *P*(*y|λ*). In other words, one would expect *P*(*y|λ*) and
(*y|x*) to be close, with the smallest "distance" at
(*y|x* = *λ*) (for *λ* integer), that is, when count *x* is exactly equal to the expected tag count under the Poisson distribution *P*(*y|λ*). In this section we provide a quantitative answer to the above question and show, perhaps surprisingly, that the "statistical distance" between *P*(*y|λ*) and
(*y|x*) is not minimized at *x* = *λ*, but it attains minimum at the mode of *P*(*y|λ*), i.e. when *x* = *λ* - 1.

The divergence *D*(*λ*, *x*) has a nice information-theoretic interpretation: When the log is base 2, *D*(*λ*, *x*) expresses the number of bits of additional information one needs in order to fully specify
(*y|x*), provided one has a perfect knowledge of *P*(*y|λ*). The divergence *D*(*λ*, *x*) is non-negative, with *D*(*λ*, *x*) = 0 if and only if the two distributions
(*y|x*) and *P*(*y|λ*) coincide.

is the entropy of *P*(*y|λ*) and *E*_{Q(y)}[*f* (*y*)] denotes the expectation of the quantity *f* (*y*) under the distribution *Q*(*y*).

As discussed above, one would intuitively expect *D*(*λ*, *x*) to be minimal for *x* = *λ*, as then the conditioning count in the A-C statistic would be the mean of the underlying Poisson distribution. However, the mode of that Poisson distribution, *λ* - 1, is surrounded by enough probability mass to yield the following result:

we have *D*(*λ*, *λ*) *- D*(*λ*, *λ* - 1) > 0, implying *D*(*λ*, *λ*) *> D*(*λ*, *λ* - 1).

Q.E.D

*P*(

*x|λ*), if we repeatedly generated a "representative" count

*x*from

*P*(

*x|λ*), what would be the average divergence of the corresponding A-C statistic (

*y|x*) from the truth

*P*(

*y|λ*)? In other words, we are interested in the quantity

Plugging (14) into (13) we obtain (11).

Q.E.D

We will now show that up to terms of order *O*(*λ*^{-1}), the expected divergence of A-C statistic
(*y|x*)] from the true underlying Poisson distribution *P*(*y|λ*) is equal to (1/2) log 2.

*F*(

*λ*, 0) and

*F*(2λ, 0). To that end, note that the entropy

*H*[

*P*(

*y|λ*)] can be approximated as [18]

Q.E.D

*ε*(

*λ*) by using a higher order entropy expansion (for log base 2):

*F*(

*λ*, 0) and

*F*(2λ, 0) in the style of (18) and (19), respectively, we obtain an expression for the expected divergence measured in bits:

*ε*(

*λ*) (measured in bits) calculated numerically from the definition (9), as well as their analytical approximation calculated from (20). As expected, the two curves are in good correspondence, as our approximation is

*O*(

*λ*

^{-3}).

Results of this section suggest that if the true Poisson source *P*(·|*λ*) is not known, the A-C statistic
(*y|x*), *based on a single observed tag count realization x* from *P*(·|*λ*), is on average not further away from the truth *P*(*y|λ*) than half a bit of additional information. As the mean tag count *λ* increases, so does the uncertainty in the generating Poisson distribution *P*(·|*λ*). As a consequence, the average K-L divergence *ε*(*λ*) from *P*(·|*λ*) to the approximating A-C statistic (based on a single realization from *P*(·|*λ*)) gets larger. The average K-L divergence expressed in bits increases with increasing *λ* from about 0.42 bits to 0.5 bits.

## Results and Discussion

The Audic-Claverie method [1] has been and still continues to be a popular approach for detection of differentially expressed genes in the SAGE framework. The method is based on the assumption that under the null hypothesis the tag counts *x, y* in two libraries come from the same but unknown Poisson distribution *P*(·|*λ*). The problem is that each SAGE library represents only a single measurement. We have rigorously analyzed usefulness of the Audic-Claverie method by investigating the A-C statistic
(*y|x*) that forms a backbone of the method and represents our knowledge of the underlying Poisson distribution *P*(·|*λ*) based on only one tag count *x* drawn from it.

It turns out that the Poisson distribution is rather "rigid" in the sense that it is unimodal and parametrized by a single parameter *λ* representing both its mean and variance. Learning about *P*(·|*λ*) from a very limited sample (as one is effectively bound to do in the SAGE framework) is much less suspicious than one might naively expect.

We have first shown that the A-C statistic
(*y|x*), even though not a Poisson distribution itself, naturally captures the distribution of further tag counts *y*, given a single observation *x* from the unknown *P*(·|*λ*). According to Theorem 1, for integer λ, both
(·|*x*) and *P*(·|*λ*) have two neighboring modes with decreasing probability values as one moves away from the modes in either direction. In particular, *P*(·|*λ*) has the modes located at *λ* and *λ* - 1, with *P*(*λ*|*λ*) = *P*(*λ* - 1|*λ*). Given a tag count *x* ≥ 1,
(*y|x*) has the modes located at *x* and *x* - 1, with
(*x|x*) =
(*x* - 1|*x*).

We then analyzed how 'close' is the A-C statistic
(·|*λ*) (in terms of K-L divergence) to the underlying Poisson distribution *P*(·|*λ*) of tag counts. It turns out that the K-L divergence from *P*(*y|λ*) to
(*y|x*) is minimized at the mode of *P*(*y|λ*), i.e. when *x* = *λ* - 1 (Theorem 2). Most importantly, by Theorem 4, on average, the A-C statistic is never too far from the true underlying distribution. To be precise, up to terms of order *O*(*λ*^{-3}), on average, the A-C statistic is never further away from the truth *P*(·|*λ*) than half-a-bit of additional information. Hence, the Audic-Claverie method can be expected to work well even though the SAGE libraries represent very sparse samples.

So far the Audic-Claverie methodology for detection of differentially expressed genes has been verified only empirically through a series of specific Monte Carlo simulations [1]. It has not been clear how general the apparently stable simulation findings were. Besides detailed explanations of the nature of A-C statistic capturing the unknown Poisson distribution based on single observation only, we showed that the A-C statistic is *universally* applicable in any situation where inferences about the underlying Poisson distribution must be made based on an extremely sparse sample. Such situations are referred to in machine learning as 'one-shot-learning'. In the Monte Carlo simulations of [1] the false alarm rate was small for genes associated with small tag counts and gradually increased for higher tag counts. The false alarm rate, however, never exceeded the significance level of the test. These findings are consistent with the theoretically calculated divergence function *ε*(*λ*) (eq. (20)) illustrated in figure 2. With increasing mean tag count λ, it is more likely that increased counts *x* will be observed. But as *λ* increases, so does the uncertainty in the generating Poisson distribution *P*(·|*λ*). Consequently, the average K-L divergence *ε*(*λ*) from *P*(·|*λ*) to the approximating A-C statistic (based on a single realization *x* from *P*(·|*λ*)) gets larger. For smaller *λ* the underlying Poisson distribution is well captured by the A-C statistic and the test that operates on it will be well behaved. As *λ* grows, the average K-L divergence *ε*(*λ*) saturates at 0.5 bits implying that the test based on the A-C statistic will continue to be well behaved even for large values of the mean tag count *λ*.

The Audic-Claverie method has also been formulated for the case of two cDNA libraries of unequal size. Similar methodologies have been proposed for the case of multiple cDNA libraries (e.g. [7]). Even though developed under the limited assumption of two libraries of the same size, theoretical results obtained in this paper offer deep insights into the workings of the Audic-Claverie approach and provide an information theoretic justification for its use when analyzing expression patterns in cDNA arrays. Of course, when using libraries of unequal size, the A-C statistic will no longer be symmetric, putting more weight on the more populated library. Information theoretic investigation of statistics developed for pattern analysis in the cases of unequal multiple libraries is a matter for our future work.

## Conclusion

- 1.
The A-C statistic and the underlying Poisson distribution share the same mode structure.

- 2.
The K-L divergence from the true unknown Poisson distribution to the A-C statistic is minimized when the A-C statistic is conditioned on the mode (not mean) of the Poisson distribution.

- 3.
The expected K-L divergence from the true unknown Poisson distribution to the A-C statistic is never larger than 1/2 bit, irrespective of the mean of the Poisson distribution.

- 4.
The expected K-L divergence from the true unknown Poisson distribution to the A-C statistic can be approximated up to order

*O*(*λ*^{-3}) by a simple function of the form*a*_{0}+*a*_{1}*λ*^{-1}+*a*_{2}*λ*^{-2}. For the divergence measured in bits,*a*_{0}= 1/2,*a*_{1}= 1/24 and*a*_{2}= 1/32.

Even though the A-C statistic infers the unknown underlying Poisson distribution based on one count observation only, the Audic-Claverie method should work reasonably well in most cases, since under the null hypothesis, the average divergence from the unknown Poisson distribution to the A-C statistic is guaranteed not to exceed 1/2 bit. This constitutes a rigorous quantitative argument, extending the empirical Monte Carlo studies of [1], that supports the wide spread use of Audic-Claverie method, even though by their very nature, the SAGE libraries represent very sparse samples.

## Declarations

### Acknowledgements

I would like to thank Hong Yan for introducing me to the problem of cDNA array analysis and Somak Raychaudhury for inspiring me to study estimation of Poisson processes based on extremely limited samples.

## Authors’ Affiliations

## References

- Audic S, Claverie J: The significance of digital expression profiles.
*Genome Res*1997, 7: 986–995.PubMedGoogle Scholar - Varuzza L, Gruber A, de B Pereira C: Significance tests for comparing digital gene expression profiles.
*Nature Precedings*2008. npre.2008.2002.3 npre.2008.2002.3Google Scholar - Velculescu V, Zhang L, Vogelstein B, Kinzler K: Serial analysis of gene expression.
*Science*1995, 270: 484–487. 10.1126/science.270.5235.484View ArticlePubMedGoogle Scholar - Brenner S, Johnson M, Bridgham J, Golda G, Loyd D, Johnson D, Luo S, McCurdy S, Foy M, Ewan M,
*et al*.: Gene expression analysis by massively parralel signature sequencing on microbead arrays.*Nature Biotechnol*2000, 18: 630–634. 10.1038/76469View ArticleGoogle Scholar - Ruijter J, Kampen AV, Baas F: Statistical evaluation of SAGE libraries: consequences for experimental design.
*Physiol Genomics*2002, 11(2):37–44.View ArticlePubMedGoogle Scholar - Ge N, Epstein C: An empirical Bayesian significance test of cDNA library data.
*Journal of Computational Biology*2004, 11(6):1175–1188. 10.1089/cmb.2004.11.1175View ArticlePubMedGoogle Scholar - Stekel D, Git Y, Falciani F: The comparison of gene expressiom from multiple cDNA libraries.
*Genome Research*2000, 10: 2055–2061. 10.1101/gr.GR-1325RRPubMed CentralView ArticlePubMedGoogle Scholar - Bortoluzzi S, Coppe A, Bisognin A, Pizzi C, Danieli G: A multistep bioinformatic approach detects putative regulatory elements in gene promoters.
*BMC Bioinformatics*2005, 6: 121–136. 10.1186/1471-2105-6-121PubMed CentralView ArticlePubMedGoogle Scholar - Medina C, Rotter B, Horres R, Udupa S, Besser B, Bellarmino L, Baum M, Matsumura H, Terauchi R, Kahl G, Winter P: SuperSAGE: the drought stress-responsive transcriptome of chickpea roots.
*BMC Genomics*2008, 9: 553. 10.1186/1471-2164-9-553View ArticleGoogle Scholar - Kim H, Baek K, Lee S, Kim J, Lee B, Cho H, Kim W, Choi D, Hur C: Pepper EST database: comprehensive in silico tool for analyzing the chili pepper (Capsicum annuum) transcriptome.
*BMC Plant Biology*2008, 8: 101–108. 10.1186/1471-2229-8-101PubMed CentralView ArticlePubMedGoogle Scholar - Zhao Y, Li Q, Yao C, Wang Z, Zhou Y, Wang Y, Liu L, Wang Y, Wang L, Qiao Z: Characterization and quantification of mRNA transcripts in ejaculated spermatozoa of fertile men by serial analysis of gene expression.
*Human Reproduction*2006, 21(6):1583–1590. 10.1093/humrep/del027View ArticlePubMedGoogle Scholar - Metta M, Gudavalli R, Gibert J, Schlötterer C: No Accelerated Rate of Protein Evolution in Male-Biased Drosophila pseudoobscura Genes.
*Genetics*2006, 174: 411–420. 10.1534/genetics.106.057414PubMed CentralView ArticlePubMedGoogle Scholar - Morin R, O'Connor M, Griffith M, Kuchenbauer F, Delaney A, Prabhu A, Zhao Y, McDonald H, Zeng T, Hirst M, Eaves C, Marra M: Application of massively parallel sequencing to microRNA profiling and discovery in human embryonic stem cells.
*Genome Research*2008, 18: 610–621. 10.1101/gr.7179508PubMed CentralView ArticlePubMedGoogle Scholar - Borecký J, Nogueira F, de Oliveira K, Maia I, Vercesi A, Arruda P: The plant energy-dissipating mitochondrial systems: depicting the genomic structure and the expression profiles of the gene families of uncoupling protein and alternative oxidase in monocots and dicots.
*Journal of Experimental Botany*2006, 57(4):849–864. 10.1093/jxb/erj070View ArticlePubMedGoogle Scholar - Lin C, Mueller L, Carthy JM, Crouzillat D, Pétiard V, Tanksley S: Coffee and tomato share common gene repertoires as revealed by deep sequencing of seed and cherry transcripts.
*Theor Appl Genet*2005, 112: 114–130. 10.1007/s00122-005-0112-2PubMed CentralView ArticlePubMedGoogle Scholar - Cervigni G, Paniego N, Pessino S, Selva J, Diaz M, Spangenberg G, Echenique V: Gene expression in diplosporous and sexual Eragrostis curvula genotypes with differing ploidy levels.
*BMC Plant Biology*2008, 67: 11–23.Google Scholar - Miles J, Blomberg A, Krisher R, Everts R, Sonstegard T, Tassell CV, Zeulke K: Comparative Transcriptome Analysis of In Vivoand In Vitro-Produced Porcine Blastocysts by Small Amplified RNA-Serial Analysis of Gene Expression (SAR-SAGE).
*Molecular Reproduction and Development*2008, 75: 976–988. 10.1002/mrd.20844View ArticlePubMedGoogle Scholar - Evans R, Boersma J: The Entropy of a Poisson Distribution.
*SIAM Review*1988, 30(2):314–317. 10.1137/1030059View ArticleGoogle Scholar

## Copyright

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.