- Open Access
PuFFIN - a parameter-free method to build nucleosome maps from paired-end reads
© Polishko et al.; licensee BioMed Central Ltd. 2014
- Published: 10 September 2014
We introduce a novel method, called PuFFIN, that takes advantage of paired-end short reads to build genome-wide nucleosome maps with larger numbers of detected nucleosomes and higher accuracy than existing tools. In contrast to other approaches that require users to optimize several parameters according to their data (e.g., the maximum allowed nucleosome overlap or legal ranges for the fragment sizes) our algorithm can accurately determine a genome-wide set of non-overlapping nucleosomes without any user-defined parameter. This feature makes PuFFIN significantly easier to use and prevents users from choosing the "wrong" parameters and obtain sub-optimal nucleosome maps.
PuFFIN builds genome-wide nucleosome maps using a multi-scale (or multi-resolution) approach. Our algorithm relies on a set of nucleosome "landscape" functions at different resolution levels: each function represents the likelihood of each genomic location to be occupied by a nucleosome for a particular value of the smoothing parameter. After a set of candidate nucleosomes is computed for each function, PuFFIN produces a consensus set that satisfies non-overlapping constraints and maximizes the number of nucleosomes.
We report comprehensive experimental results that compares PuFFIN with recently published tools (NOrMAL, TEMPLATE FILTERING, and NucPosSimulator) on several synthetic datasets as well as real data for S. cerevisiae and P. falciparum. Experimental results show that our approach produces more accurate nucleosome maps with a higher number of non-overlapping nucleosomes than other tools.
- Nucleosome positioning
- genome-wide nucleosome maps
- paired-end reads
One of the central objectives in molecular biology is to characterize all cellular processes controlling gene regulation. The complex interaction between DNA chromatin structure and transcription factors is one of these key processes. The basic unit of chromatin structure is the nucleosome, which is composed of ≈ 146 base pairs of DNA wrapped around a protein complex of eight histones. Loosely speaking, the more compact the chromatin, the harder it is for transcription factors and other DNA binding proteins to access DNA and trigger transcription. Thus, to elucidate the role of interactions between chromatin and transcription factors, it is crucial to determine the location of all nucleosomes along the chromosomes.
Several experimental techniques are available to produce genome-wide nucleosome maps. For instance, one can isolate the portions of DNA that are free of nucleosomes or enrich for genomic regions that are bound to histones. The latter can be achieved via micrococcal nuclease digestion (MNase or MAINE) , which can be combined with chromatin immunoprecipitation (ChIP) to enrich for a particular subset of nucleosomes (e.g., for a particular histone tail mark), typically followed by high-throughput sequencing (MNase-Seq and ChIP-Seq, respectively). In this work, we assume that the sequencing data is either MNase-Seq or ChIP-Seq, which are currently the most popular approaches to study the locations of nucleosomes and histone modifications.
An analysis of the literature reveals that the majority of nucleosome maps have so far been produced from single-end reads (which are less expensive to obtain than paired-end reads). As a consequence, nearly all computational methods available assume that the input data are single-end reads. Nucleosome positioning from single-end reads is, however, more computationally challenging and much less precise than if paired-end data was available. Paired-end reads allow one to determine both ends of nucleosome-enriched DNA fragments, whereas with single-end reads one either obtains one "boundary" or the other. In the latter case, the problem of associating a peak in the forward strand with the correct peak in the negative strand can be difficult, in particular for complex nucleosome configurations.
Existing methods for single-end reads either rely on the assumption that nucleosome-enriched DNA fragments are expected to be of a size compatible with the nucleosome (≈ 146 bp), or use probabilistic models to estimate these sizes from the data. From our experience, the first approach can lead to poor results because there is no fragment size that will work equally well for all nucleosomes in the genome. While one would expect nucleosome-enriched DNA fragments to be about 146 bp, in MNase-Seq the digestion process can either leave nucleosome-free DNA in the sample, or "over-digest" the ends of nucleosome-bound DNA. Furthermore, the rate of digestion is sequence-dependent [2, 3], so nucleosomes in different genomic locations can end up with different DNA fragment sizes.
Despite these challenges, the majority of so-called "peak-calling" approaches usually rely on the assumption that the data is derived from nucleosome-sized DNA fragments and consist of following steps: (1) a nucleosome occupancy score function is obtained from mapping nucleosome-enriched reads to the reference genome, followed by counting, smoothing and normalization; (2) candidate nucleosomes are placed according to the peaks of the score function; (3) the final set of nucleosomes is selected to satisfy additional constraints (which are tool-dependent). To compute the occupancy score, different techniques have been proposed, ranging from simply computing the number of reads covering each genomic location, to sophisticated statistics to estimate the false discovery rate. For instance, nucleR  uses the raw coverage with extensive "profile cleaning" based on the Fourier transform, whereas NSeq  employs a triangle statistic based on read counts within a sliding window.
A second group of methods is based on probabilistic models. Our tool NOrMAL  uses a modified Gaussian mixture model to infer nucleosome-enriched fragment sizes. The parametric probabilistic model allows to deal with the problem of overlapping and complex configurations of nucleosomes. Developed in parallel with NOrMAL, Ping  employs a similar probabilistic model. Both tools provide a clear advantage over algorithms that rely on the user to provide estimated DNA fragment sizes.
Finally, a distinct group of positioning methods depend on the availability of a control track (i.e., "naked" DNA), e.g., NucleoFinder , while others have been designed to perform differential nucleosome positioning, e.g., DANPOS  and DiNuP .
In this work, we focus on the problem of determining nucleosome positions based on the availability of paired-end reads (without a control track). To the best of our knowledge, NucPosSimulator  is the only published tool specifically designed to take advantage of paired-end reads: to place nucleosomes it solves the optimization problem of selecting the subset of peaks which maximizes the total score, under the constraint that these peaks are located at the expected nucleosome distance from each other. Our tool PuFFIN (Positioning for Fuzzy and FIxed Nucleosomes) instead uses a novel multi-resolution approach: while its algorithm is relatively simple, our approach introduces some novel ideas that have the potential to be useful in other domains of genome analysis.
Our method consists of three steps: (A) we build a set of nucleosome profiles and nucleosome "landscapes"; (B) we detect candidate nucleosome locations on each profile; (C) we select a "consensus" set of nucleosomes that satisfies non-overlapping constraints. We discuss these steps in detail in the next subsections.
Computing nucleosome profiles and nucleosome landscapes
We first map sequenced reads to the reference genome and then compute a nucleosome profile that represents the likelihood that a genomic location is occupied by a nucleosome. Candidate nucleosomes are detected at the peaks of the nucleosome profile. In order to reduce false positives, profiles have to be cleaned from their high frequency component. Choosing the best smoothing method (and its parameters) is, however, not easy. For instance, in  the authors show that the kernel density estimation method  works significantly better than moving average-based smoothing. The choice of kernel parameters is also important: too much smoothing can merge adjacent peaks, too little can leave noisy artifacts that can be interpreted as peaks and thus introduce spurious nucleosomes. To address the challenges of choosing the "right" kernel and smoothing parameters, we follow an alternative (novel) procedure to construct nucleosome profiles.
where µ i is the genomic center location of read i, w i is the length of read i (i.e., the distance between the leftmost nucleotide in the left mate and rightmost nucleotide of the right mate), and α is a smoothing parameter. Replacing each mapped read with a Gaussian distribution allows us to model probabilistically the uncertainty in the paired-end mapping. For instance, when the left and right mate are mapped far from each other, the mass of the Gaussian will be distributed on a longer interval because of its large variance. If instead the left and right mate are close to each other, the Gaussian will have its mass concentrated at the center of the read, indicating a higher confidence in the nucleosome position.
where n is the number of mapped reads in input, and β i is the weight of the read i. If we had employed a uniform weighting scheme , paired-end reads with very short insert would dominate the profiles. To reduce the effects of short DNA fragments, we use a non-uniform weighting scheme. For paired-end reads that are shorter than 146bp, we assign a penalty factor γ(w) < 1, such that the shorter the read is, the less the weight is (i.e., ). Additionally, one could use the weights β i to account for sequence quality of individual reads, mappability biases, etc.
As said, parameter α controls the smoothness of function S α . The bigger α is, the smoother S α is, (peaks will be wider), and vice versa. When α is large, we capture nucleosome binding preferences at a lower resolution scale; when α is small we can detect nucleosomes at a high resolution scale (but noisier). In the limit α → 0, function , where is the indicator function. In this case, S0(x) represents how many read centers cover location x in the genome.
One might think that one could obtain the same profiles by computing the read coverage function smoothed by a Gaussian kernel. There is, however, a significant difference: the size of each mapped read independently influences the shape of S α (no matter what smoothing parameter is chosen), while in the case of kernel smoothing the impact of read sizes becomes less and less important as the smoothing strength increases.
Since we do not know the appropriate value for α for the data, in this step we generate a family of functions for several choices of α. Formally, we create a set of m functions , where α1 < α2 < . . . < α m are m distinct choices for α. The value m is hard-coded in our implementation (we used m = 40 for all the experiments).
The set of functions enables our algorithm to detect candidate locations for nucleosomes at different resolution scales, thus eliminating the need to specify in advance the parameters for the range of nucleosome-enriched fragments. In other words, our algorithm can "adapt" to the local properties of the input data by processing the same location at different resolutions (corresponding to the choices of α).
where ε > 0 is a small constant to avoid a division by zero, and A > maxk = 1,2,...,mα i . In our implementation we pick α to range from 0.05 to 0.63 and A = 1.5. Since mappability biases affect each function , we can effectively reduce these biases by taking the log ratio of high-resolution and low-resolution function. Another reason to carry out this normalization step is to reduce the differences in the peak heights.
Detecting candidate nucleosomes
By construction, a nucleosome landscape represents a non-parametric distribution of nucleosomes at resolution α k . The presence of a peak in any nucleosome landscape indicates a candidate nucleosome. The reads that form corresponding peak belong to that candidate.
A peak is defined by a pair (q, s) where q is the center of the peak and s is the width of the peak. We say that (q, s) is a peak for function N when N (q) is local maximum for N and s = min z (|q − z|) where z is any local minimum for function N .
Detecting peaks on each function can be easily computed in linear time along the length of the genome. As a result, for every choice of α k , k = 1, 2, . . . , m we have a set of peaks , where is a pair (center, width) representing the peak, and l is the number of peaks.
Peaks are however not guaranteed to have a symmetric shape. We therefore recompute the location of every nucleosome candidate as the centroid location of its read midpoints. This additional step ensures that candidate nucleosome locations properly represent the corresponding input reads.
Building the final solution
We now explain how to build the final set of non-overlapping nucleosomes from the family of peak sets . We say that two peaks (q1, s1) and (q2, s2) overlap if |q1 − q2| < 146 (the size of a nucleosome). Observe that by construction, the number of peaks detected at lower resolution (i.e., for large α) will be smaller than or equal to the number of peaks detected at higher resolution, i.e., |P α | ≤ |P β | when α > β. As we increase the smoothing parameter α, the total number of peaks decreases: while some peaks are preserved, others are merged. In other words, for every peak in P α we can find at least one corresponding peak in P β if α > β.
Let us consider again our example in Figure 1. Detected peaks are marked with circles in panel C. The algorithm first processes the set of peaks on the blue function (α = 0.07). Since there are no peaks on that curve that are located at a distance greater than 146bp from each other, the final set C remains empty. Next, the algorithm processes the green curve (α = 0.21): here there are three peaks that satisfy the non-overlapping constraint. Thus, the algorithm adds those peaks (marked with solid circles) to C. Then, the algorithm considers the red curve (α = 0.62): all four peaks are non-overlapping with each other, however only one peak (marked with the solid circle) can be added to C. As a result, the final solution C consists a set of four peaks that match the original nucleosomes. Observe that strongly positioned nucleosomes I, III and IV are detected earlier in the algorithm (α = 0.21) than fuzzy nucleosome II (α = 0.62).
To compute a set of profile functions S α we use a precomputed set of curves for every choice α and w in a predefined range. As a result, it takes Θ(nm) operations, where n is the number of reads and m is the number of curves. In our implementation we used m = 40 choices of equally distributed values for α ∈ [0.05, 0.63].
Finding peaks on each curve S α takes Θ(l) time, where l is the length of the processed region. Thus, the total time to find candidate nucleosomes (Figure 2, lines 1-3) is Θ(m(n + l)). Building the resulting set of non-overlapping nucleosomes is determined by the number of candidates that is at most Θ(ml). Given that m is predefined, it follows that the total running time is linear in the region size and number of input reads.
To evaluate the performance of PuFFIN, we performed extensive benchmarking against NucPosSimulator, TEMPLATE FILTERING and NOrMAL. NucPosSimulator is the only published tool designed to deal with paired-end reads . As said, it solves the optimization problem of selecting the subset of peaks which maximizes the total score, under the constraint that these peaks are located at the expected nucleosome distance from each other. TEMPLATE FILTERING is one of the first algorithms developed to infer the size of the fragments from single-end reads . NOrMAL uses a modified Gaussian model to cluster input single-reads such that every cluster represents a nucleosome . Some of the recently published tools that use a control sample to solve the nucleosome positioning problem, e.g., DANPOS and NucleoFinder , are not included in this comparison.
We used default parameters for each tool except for the following provisions. For TEMPLATE FILTERING and NOrMAL we set to zero the allowed overlap between adjacent nucleosomes to allow for a fair comparison with PuFFIN and NucPosSimulator.
Arguably the major challenge for nucleosome position inference is that the true positions of nucleosomes are unknown. The lack of a "ground-truth" makes it very hard to benchmark existing computational methods. For this reason we made extensive use of synthetic data, as explained next.
Results on synthetic data
Next, we performed a more realistic comparison on in silico reads for larger synthetic nucleosome maps. We used the nucleosome map generator syntheticNucMap from nucleR . This tool allows users to specify the number of well-positioned and fuzzy nucleosomes, as well as the variance for the location of synthetic reads and the coverage level. Well-positioned nucleosomes are placed along the chromosome regularly spaced with a fixed linker size (we used linkers of 20bp, which introduces a periodicity of ≈ 167bp). For fuzzy nucleosomes, locations are picked at random and independently from other nucleosomes already on the chromosome. As a consequence, fuzzy nucleosomes can overlap with other nucleosomes. For the variance parameter we choose 30 bases for well-positioned and 50 bases for fuzzy nucleosomes.
To eliminate the effects of over-reporting in NucPosSimulator, we discarded from the counts nucleosomes that are located 148 bases or less from each other, such that every pair of tightly placed nucleosomes is count as one nucleosome. In Figure 5, curves marked "filtered" shows the results of this cleaning step. Observe that the number of nucleosomes reported by NucPosSimulator drops significantly, while only a small number of PuFFIN nucleosomes are affected. In fact, using this cleaning step, PuFFIN reports a larger numbers of nucleosomes than NucPosSimulator. All together, these experimental results on synthetic data show that PuFFIN generates more accurate nucleosome maps, without over-reporting nucleosomes.
Results on real data
For the comparison of nucleosome positioning tools, we used a publicly available dataset for S. cerevisiae (NCBI SRA SRR094649) and our dataset for P. falciparum (NCBI SRA SRS453761). All datasets contain paired-end reads produced by an Illumina sequencing instrument. Reads were mapped to their corresponding reference genomes using Bowtie2  with --very-fast-local --no-discordant flags. We removed reads that were not mapped uniquely or had a distance between the left and right mates smaller than 40bp or bigger than 1,000bp.
Number of reported nucleosomes and execution times on yeast and the human malaria parasite.
S. cerevisiae(W303 contig 7)
P. falciparum(3D7 chr. 2)
Our previous work  has demonstrated that the P. falciparum genome has a greater complexity of nucleosomes configurations. As expected, experimental results show much greater variance in the number of nucleosomes in the malaria dataset reported by the various tools. PuFFIN reports a similar number of nucleosomes compared to NucPosSimulator, but significantly higher numbers than NOrMAL and TEMPLATE FILTERING, indicating that our method is capable to resolve complex configurations of nucleosomes.
The execution time of PuFFIN is higher than NOrMAL and TEMPLATE FILTERING on both datasets, but shorter than NucPosSimulator on P. falciparum and higher on S. cerevisiae datasets. Our implementation of PuFFIN is currently written in Python, while the other tools use either Java or C/C++. We believe that speed of our tool could be easily improved by one order of magnitude by implementing it in C/C++ (work in progress).
We described a novel method to solve the nucleosome positioning problem when paired-end data is available. Our method employs a multi-resolution strategy that circumvents a smoothing step that usually requires user-defined parameters to set the strength of the smoothing and type of kernel to be used. Experimental results show that our method more accurately detects nucleosome positions as compared to existing software tools, in particular when complex nucleosome configurations are present in the data.
This study was financially supported by the Human Frontier Science Program [grant LT000507/2011-L to E.M.B.]; the National Institutes of Health [grant R01 AI85077-01A1 to K.L.R. and S.L.]; and the National Science Foundation [grant IIS-1302134 to S.L. and K.L.R.].
The authors declare that funding for publication of this article was sponsored by National Science Foundation, grant IIS-1302134.
This article has been published as part of BMC Bioinformatics Volume 15 Supplement 9, 2014: Proceedings of the Fourth Annual RECOMB Satellite Workshop on Massively Parallel Sequencing (RECOMB-Seq 2014). The full contents of the supplement are available online at http://www.biomedcentral.com/bmcbioinformatics/supplements/15/S9.
- Zaret Ken: Micrococcal nuclease analysis of chromatin structure. Current protocols in molecular biology. 2005, Chapter 21 FebruaryGoogle Scholar
- Allan James, Fraser Ross M, Owen-Hughes Tom, Keszenman-Pereyra David: High-resolution nucleosome mapping reveals transcription-dependent promoter packaging. Micrococcal nuclease does not substantially bias nucleosome mapping. Journal of Molecular Biology. 2012, 417 (3): 152-64. 10.1016/j.jmb.2012.01.043. JanuaryView ArticlePubMedGoogle Scholar
- Weiner Assaf, Hughes Amanda, Yassour Moran, Rando Oliver J, Friedman Nir: Genome research. 2010, 20 (1): 90-100. 10.1101/gr.098509.109. JanuaryPubMed CentralView ArticlePubMedGoogle Scholar
- Flores Oscar, Orozco Modesto: nucleR:a package for non-parametric nucleosome positioning. Bioinformatics (Oxford, England). 2011, 27 (15): 2149-50. 10.1093/bioinformatics/btr345. AugustView ArticleGoogle Scholar
- Abhinav Nellore, Konstantin Bobkov, Elizabeth Howe, Aleksandr Pankov, Aaron Diaz, Song Jun S: NSeq:a multithreaded Java application for finding positioned nucleosomes from sequencing data. Frontiers in genetics. 2012, 3 (320): 10.3389/fgene.2012.00320.Google Scholar
- Polishko Anton, Ponts Nadia, Le Roch Karine G, Lonardi Stefano: NORMAL: accurate nucleosome positioning using a modified Gaussian mixture model. Bioinformatics (Oxford, England). 2012, 28 (12): i242-9. 10.1093/bioinformatics/bts206. JuneView ArticleGoogle Scholar
- Zhang Xuekui, Robertson Gordon, Woo Sangsoon, Hoffman Brad G, Gottardo Raphael: Probabilistic inference for nucleosome positioning with MNase-based or sonicated short-read data. PloS one. 2012, 7 (2): e32095-10.1371/journal.pone.0032095. JanuaryPubMed CentralView ArticlePubMedGoogle Scholar
- Becker Jeremie, Yau Christopher, Hancock John M, Holmes Christopher C: NucleoFinder: a statistical approach for the detection of nucleosome positions. Bioinformatics. 2013, 29 (6): 711-716. 10.1093/bioinformatics/bts719. MarchPubMed CentralView ArticlePubMedGoogle Scholar
- Kaifu Chen, Yuanxin Xi, Xuewen Pan, Zhaoyu Li, Klaus Kaestner, Jessica Tyler, Sharon Dent, Xiangwei He, Li Wei: DANPOS:dynamic analysis of nucleosome position and occupancy by sequencing. Genome research. 2013, 23 (2): 341-51. 10.1101/gr.142067.112. FebruaryView ArticleGoogle Scholar
- Kai Fu, Qianzi Tang, Jianxing Feng, X Shirley Liu, and Yong Zhang. DiNuP: A Systematic Approach to Identify Regions of Differential Nucleosome Positioning. Bioinformatics (Oxford, England). 2012, 28 (15): 1965-1971. 10.1093/bioinformatics/bts329. JuneView ArticleGoogle Scholar
- Schopflin Robert, Teif Vladimir B, Muller Oliver, Weinberg Christin, Rippe Karsten, Wedemann Gero: Modeling nucleosome position distributions from experimental nucleosome positioning maps. Bioinformatics. 2013, 29 (19): 2380-2386. 10.1093/bioinformatics/btt404.View ArticlePubMedGoogle Scholar
- Parzen E: On estimation of a probability density function and mode. Annals of mathematical statistics. 1962, 33: 1065-1076. 10.1214/aoms/1177704472.View ArticleGoogle Scholar
- Langmead Ben, Salzberg Steven L: Fast gapped-read alignment with Bowtie 2. Nat Meth. 2012, 9 (4): 357-359. 10.1038/nmeth.1923. AprilView ArticleGoogle Scholar
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.