Transcript mapping based on dRNAseq data
 Thorsten Bischler^{1, 2},
 Matthias Kopf^{1} and
 Björn Voß^{1}Email author
DOI: 10.1186/1471210515122
© Bischler et al.; licensee BioMed Central Ltd. 2014
Received: 25 October 2013
Accepted: 24 April 2014
Published: 29 April 2014
Abstract
Background
RNAseq and its variant differential RNAseq (dRNAseq) are today routine methods for transcriptome analysis in bacteria. While expression profiling and transcriptional start site prediction are standard tasks today, the problem of identifying transcriptional units in a genomewide fashion is still not solved for prokaryotic systems.
Results
We present RNAseg, an algorithm for the prediction of transcriptional units based on dRNAseq data. A key feature of the algorithm is that, based on the data, it distinguishes between transcribed and untranscribed genomic segments. Furthermore, the program provides many different predictions in a single run, which can be used to infer the significance of transcriptional units in a consensus procedure. We show the performance of our method based on a wellstudied dRNAseq data set for Helicobacter pylori.
Conclusions
With our algorithm it is possible to identify operons and 5’ and 3’UTRs in an automated fashion. This alleviates the need for labour intensive manual inspection and enables largescale studies in the area of comparative transcriptomics.
Keywords
RNAseq Differential RNAseq Segmentation Transcriptional unit Transcriptome Transcriptional start site Dynamic programmingBackground
The development of RNAseq [1] has boosted research on prokaryotic transcriptomes throughout the last years. It can be used for the detection of novel transcripts, e.g. noncoding RNAs (ncRNAs), analysis of differential expression in response to environmental stimuli and others. Recently, a variant called dRNAseq was introduced [2], which allows the transcriptomewide mapping of transcriptional start sites (TSSs). This provides a means to reliably detect the 5’end of transcripts. But what about the 3’end?
The accurate determination of transcript boundaries is useful for several reasons. It enables the identification of promoter and terminator elements and, thus, UTRs which potentially carry regulatory elements. Furthermore, target prediction for ncRNAs will benefit from the knowledge of exact transcript limits and UTRs. The same holds true for expression analysis, since all reads mapping a transcript can be rigorously accounted for, which improves the precision. Not only expression levels might vary but also the transcript architecture, e.g. by differential processing or alternative TSSs. For these reasons it is of interest to have an automated procedure to detect transcript boundaries. Finally, the organization of genes into operons or transcriptional units can be easily elucidated when the genomic location of transcripts is known.
In the near past, highdensity genomic tiling microarrays were the method of choice for the characterization of complete transcriptomes. For this technology a segmentation method for the hybridization signal along genomic coordinates was proposed [3]. It makes use of a structural change model (SCM) which is fitted to the array data. The goal is to partition the data into blocks with ideally uniform expression and is achieved by computing the set of segments that minimizes the sum of squared residuals. An alternative approach for the analysis of highdensity oligonucleotide tiling arrays makes use of a hidden Markov model [4].
To the best of our knowledge no general method for transcript boundary estimation based on RNAseq data is available. CUFFLINKS [5] and SCRIPTURE [6] perform best with eukaryotic mRNAseq data, which captures polyAtailed transcripts only, and put a focus on the detection of splice variants. Similar considerations hold true for denovo transcript assemblers such as ABYSS [7] and SOAPDENOVO [8]. Prokaryotic transcripts do not have a polyA tail and, thus, bacterial RNAseq provides information on merely any RNA present in the cell. On the one hand, this promises to provide the full picture of a bacterial transcriptome, but on the other hand, this also increases the complexity for its analysis. Nevertheless, we set out to develop a method for transcript boundary determination based on RNAseq data mapped to a reference genome. More precisely, we chose dRNAseq data as the input, since it explicitly provides information on transcriptional start sites.
As a starting point we chose the SCM based segmentation algorithm from [3]. We reimplemented it in C++ and added the ability for parallel computation using openMP (http://openmp.org). The major improvement is achieved by extending the segmentation method to make use of dRNAseq data, especially data from libraries enriched for primary transcripts. For this, we modify the dynamic programming based optimization, such that segments satisfy certain userdefined constraints. This reduces the search space leading to improved speed and accuracy of the algorithm and further allows us to discriminate transcript from nontranscript segments. Finally, we present a method to compute consensus segments, which makes use of the fact that the algorithm intrinsically computes many results. This integrative step results in segments with improved confidence.
Implementation
In the following we will describe the algorithm implemented in RNASEG. Due to the close relation to the algorithm developed in [3] we reuse large parts of their notation.
dRNAseq data
The data provided by a dRNAseq experiment is in the simplest case a set of sequencing reads from two libraries. One library consists of sequencing reads from RNA enriched for primary transcripts, and the second is untreated. Throughout this manuscript we name the reads primary and secondary, respectively. Ideally, sequencing reads from primary libraries start at the native 5’end of a transcript, such that the 5’ends of primary reads represent the 5’ends of native primary transcripts. Sequencing reads from secondary libraries start at the 5’end of secondary transcripts. These may either be degradation or processing products or, as for the data used here, the result of a fragmentation procedure. Note that if the secondary library is treated with tobacco acid pyrophosphatase (TAP) it also contains primary transcripts, which would otherwise not be accessible for ligation due to the pyrophosphate. TAP treatment, although used in [2], is not mandatory and to make RNASEG as widely applicable as possible, we provide means to handle the effects of this treatment (see below).
We store for each position i in the genome the number of primary read starts P_{ i }, the coverage by primary reads ${C}_{i}^{p}$ and the coverage by secondary reads ${C}_{i}^{s}$.
Structural change model segmentation
The data used for modeling consists of the number of primary read starts (P) and the secondary read coverage (C^{ s }) for each position in the genome. The coverage by secondary reads is expected to be uniformly distributed over the fulllength of the transcript and, thus, RNASEG uses this data to compute the sum of squared residuals for a candidate segment. Primary read start information is used as a constraint for start positions of transcript segments and for differentiating between transcript and nontranscript segments. In the following, we recapitulate the definition from [3] and explain our extensions to incorporate dRNAseq data.
Fitting the model
where i = 1,…,n is the genomic position, ${z}_{i}={\mathit{\text{log}}}_{2}({C}_{i}^{s}+1)$, cp_{1},…,cp_{ r } are parameters for segment boundaries also called change points, cp_{1} = 1 and cp_{R+1} = n + 1, R is the maximal number of segments, μ_{ r } is the mean log_{2} normalized C^{ s } for segment r and ε_{ i } are the residuals.
The minimization which leads to the optimal set of parameters ${\stackrel{~}{\mathit{\text{cp}}}}_{1},\dots ,{\stackrel{~}{\mathit{\text{cp}}}}_{R}$ is done by the dynamic programming (DP) algorithm described in the following section.
Optimization
The tool RNASEG implements an extended version of the DP algorithm available from the Bioconductor tilingArray package, which we call the original algorithm from now on. It calculates the minimal sum of squared residuals in the first step and determines the optimal set of change points during backtracing.
The original algorithm starts with computing the cost matrix G which is the main input for the DP procedure that finds the optimal segmentation. The cost matrix contains for each entry G_{i,k} the sum of squared residuals of the segment from i to i + k  1. The calculation of the cost matrix for arbitrary segments would take quadratic time and space with respect to the genome size n. This renders the algorithm inapplicable on a genome wide scale. For our algorithm we reduce the complexity by restricting the segment length to a maximum value $\widehat{k}$, resulting in a complexity of $n\times \widehat{k}$ for G. This strategy was already chosen in the original algorithm but using a fixed, rather than a userdefined $\widehat{k}$.
Just like the original algorithm, our method uses two matrices for calculating the optimal segmentation. The scoring matrix stores in each entry $m{I}_{i}^{\mathit{\text{cp}}}$ the optimal cost for the segmentation from 0 to i with cp change points. The traceback matrix contains in $m{t}_{i}^{\mathit{\text{cp}}1}$ the index of the rightmost change point in the optimal segmentation from the start to i with cp change points. The major difference of our algorithm to the original one is, that in order to decide whether a segment is a transcript or a nontranscript segment our algorithm checks if the current segment suffices some constraints, such as enough primary reads at the segment start or a mean coverage by secondary reads above a threshold. If this is the case, the change point belongs to a transcript segment and is stored as a positive positional value in the mt matrix. Otherwise the change point denotes the start of a nontranscript segment and is assigned a negative positional value. A segment may even be neither a transcript nor a nontranscript, which is for example reasonable for segments with high mean coverage but without a valid TSS. In such a case, the segment is marked invalid and not further considered.
In differentiating between transcript and nontranscript segments, the algorithm allows only for transcript segments to appear one after another, or as an alternating order of transcript and nontranscript segments. The occurrence of two adjacent nontranscript segments is prevented by checking the previous change point in the mt matrix. This restriction is not only biologically reasonable but also results in a speedup of the calculation.
where val(i,j,cp) is a function that checks if the segment from i to j for cp change points suffices the user defined thresholds for transcript or nontranscript segments. The individual checks are described in the following.
Segment constraints
In the following, (i,j) denotes a segment from position i to position j and cp denotes the number of change points of the current recursion.
A constraint that was already introduced above and is essential for the performance of the algorithm is the maximal segment length $\widehat{k}$. In addition, the user can also impose a minimum segment length $\stackrel{\u030c}{k}$. The latter may be useful in cases where the dRNAseq library preparation includes a size selection step, such that only RNAs above a certain length are analyzed. Each segment (i,j) must satisfy $\stackrel{\u030c}{k}\le ji+1\le $$\widehat{k}$.
A transcript segment (i,j) needs to start with a reasonable number of primary reads, say t. Therefore, the number of primary read starts at position i has to exceed t (P_{ i } > t). This is a rather simplistic criterion, and we provide possible alternatives in the discussion, but still the complete method performs well, as will be shown later.
For highly abundant transcripts the enrichment for primary transcripts will not be perfect, thus reads from degradation products will misleadingly show up as primary reads. The user can choose to reduce the resulting increase in potential transcript segment start points, by setting a ratio r (0 ≤ r ≤ 1) between primary read starts and primary read coverage. This changes the way in which the N array is computed, such that N_{ i } stores the position q for which ${P}_{q}>t,q\le i\phantom{\rule{2.77626pt}{0ex}}\text{and}\phantom{\rule{2.77626pt}{0ex}}\frac{{P}_{q}}{{C}_{q}^{p}}\ge r$ holds true. Again the q with maximal P_{ q } is chosen within i,…,i  w.
A segment (i,j) is classified as a transcript segment if i  N_{ i } ≤ w. This remapping of the start position is reflected in the scoring scheme by replacing $m{I}_{jk1}^{\mathit{\text{cp}}1}+{G}_{jk,k}$ with $m{I}_{N(jk)1}^{\mathit{\text{cp}}1}+{G}_{jk,k}$ in (4).
An essential feature of RNASEG is the discrimination between transcript and nontranscript segments. As described before, the mt matrix stores positive and negative positional values for transcript and nontranscript segments, respectively. A nontranscript segment is only allowed to follow a transcript segment, while transcript segments are not constrained, thus $({N}_{i}>w\wedge m{t}_{i1}^{\mathit{\text{cp}}2}>0)\vee {N}_{i}\le w$ must hold true.
Note, that we use the nonnormalized ${C}_{i}^{s}$ here, compared to the log_{2} normalized values for computing the G matrix. We feel that this is more intuitive for the user.
For certain numbers of change points, the imposed restrictions may lead to an invalid segmentation, i.e. for a certain position j in the genome no i can be found, such that (i,j) satisfies all constraints. We mark such instances by setting $m{I}_{i}^{\mathit{\text{cp}}}=1$ and $m{t}_{i}^{\mathit{\text{cp}}1}=2n$. During the recursion, if a candidate segment does not satisfy $m{I}_{i1}^{\mathit{\text{cp}}1}\ne 1$, it is not considered a valid subsolution and, thus, ignored.
The output of the algorithm contains the transcript and nontranscript segments for each number of change points in GFF format. It is generated by parsing the change points stored in the th matrix thereby generating entries for transcript or nontranscript segments in the output file. Change point numbers which have been tagged invalid during backtracing will be ignored and will not appear in the output. For each segment the start position is the current change point i and the end position is located one position in front of the following change point.
Optimal segment number and consensus segments
RNASEG computes for each number of segments, the optimal set of change points. In other words, the algorithm does not provide the overall optimal solution, but rather many solutions which are optimal by themselves, i.e. for the given number of change points. The choice of the optimal number of change points is not trivial, as has already been stated in [3]. One can use information theoretical approaches, such as the AIC (Akaike’s Information Criterion) and BIC (Bayesian Information Criterion), but the authors finally suggest an empirical estimation based on positive control regions. In our opinion, this is not satisfying and we provide two ways to cope with this.
An intrinsic property of the constraints described above is that they limit the maximal number of transcript segments $\widehat{M}$. Since a nontranscript segment has to be followed by a transcript, a maximum of $2\widehat{M}+1$ segments are possible. During the calculation of N RNASEG determines how many positions satisfy the transcript start constraints, which gives an upper bound for $\widehat{M}$. If $\widehat{\mathit{\text{cp}}}>2\widehat{M}+1$RNASEG automatically lowers $\widehat{\mathit{\text{cp}}}$ to $2\widehat{M}+1$, saving computation time and memory. The maximal number of change points for which a valid segmentation could be derived depends on all constraints and is available at the end of the run. This set of segments provides the highest resolution for the given constraints. For high quality data and reasonable constraints, it likely constitutes the final result.
Our second strategy makes use of all segmentations for the different change point numbers. For similar numbers of change points, the segment sets likely share a large number of segments. It is important to note that the computation does not enforce this behavior. In order to take advantage of this information we apply a consensus strategy. This strategy focuses on transcript segments (transcripts for short) and has essentially four steps:

First, transcripts are collected from all numbers of change points, and their occurrence frequencies determined. We use this occurrence frequency as a proxy for the quality of the prediction.

Second, internal TSSs may split a long transcript into two or more short transcripts. Hence, transcripts that together correspond to a longer transcript, i.e., subtranscripts, are chained and their occurrence frequencies added to the long version.

Third, transcripts from different numbers of change points may differ only by a few positions. Thus, we merge transcripts that overlap to 99% or more. For this, we keep the more frequent variant and sum up the occurrence frequencies.

Fourth, for competing (partially overlapping) transcripts we retain the one with higher cumulated occurrence frequency, as this is supported by a larger number of individual segmentations.
As a result, the segment sets for the various change point numbers are merged and provided in a single output file in GFF format, ready for visualization in popular genome browsers. Furthermore, transcript segments are augmented by their occurrence frequencies among all change point numbers, which allows to infer the significance of the actual transcript. As a byproduct this script allows to merge results for the different strands as well as of several partial analyses of adjacent, possibly overlapping, genomic regions. Thus, it is easy to split the analysis of a complete genome into small, overlapping pieces (say 100 kb), do the segmentation piecewise, and merge the individual results. This decreases overall runtime, since for shorter sequences $\widehat{\mathit{\text{cp}}}$ can also be reduced.
Results
We applied RNASEG to the data from [2] for Helicobacter pylori 26695. The individual steps are described in the following.
Data
We downloaded all data for the experiments SRX014058, SRX014056, SRX014054, SRX014051, SRX014034, SRX014033, SRX014031, SRX014018, SRX01401317 from the NCBI Sequence Read Archive (SRA). The samples SRX01401317 represent Solexa sequencing results of untreated RNA, while the other samples correspond to 454 sequences from primary enriched libraries. In total approx. 2 million primary and 83 million secondary reads were obtained. All reads were polyA trimmed at the 3’end and 454 reads were additionally subjected to a 5’ adapter clipping (fixed length clipping using 28 bases).
Read mapping and input file generation
We used SEGEMEHL [9] to map the reads to the genome, requiring a minimum accuracy of 85% and utilizing the cooptimal matching strategy.
In the following a positional coverage file for each strand was generated where the primary read starts and coverage were calculated as the maximum of the primary libraries. The secondary coverage is the mean of the secondary libraries. In both cases the values were normalized beforehand via the number of mapped reads for the respective library. In order to be suitable as input for RNASEG the data was arranged in tabdelimited format as follows: Each row has four values corresponding to primary read starts, primary coverage, secondary read starts and secondary coverage. The genomic position is not represented explicitly, but given implicitly by the position of the row in the file. Thus, values in row 1 correspond to position 1 in the genome, row 2 to position 2, and so on. We selected this format since it can be directly visualized by the Artemis genome browser [10] as a user graph. Together with the output of RNASEG in GFF format this allows a simple and fast visualization of the experimental data together with the prediction. The RNASEG distribution provides a python script to convert mapping files in SAM format to the described format.
Application
In order to speed up the computation, we analyzed the 1,667,867 nt long genome in 17 parts of 103,112 nt length where adjacent parts overlap by 10,000 nt. RNASEG was applied with the following constraints: primary read start threshold t = 10, min./max. segment length $\stackrel{\u030c}{k}=50,\widehat{k}=10,000$, maximal no. of change points $\widehat{\mathit{\text{cp}}}=1,000$, transcript mean coverage cutoff a = 0.5, nontranscript max. mean coverage u = 0.5, and read start ratio r = 0.5. These settings were derived by analyzing small sample regions. The results were combined using our consensus strategy described above and transcripts with an occurrence frequency below $\frac{1}{4}$ discarded. In total, 1,696 transcripts and 2,147 subtranscripts were predicted. We also extracted the transcripts for the maximal number of change points for each analyzed part and joined those that meet without a gap within an annotated gene, resulting in 2,150 transcripts. We term transcripts and subtranscripts derived by summarizing T^{ s } and ${T}_{\mathit{\text{sub}}}^{s}$ and those from the maximum change points T^{ m }.
Figure 1 shows the results for section 563,000 to 583,000. This region comprises the cag pathogenicity island, which was also described in detail in [2]. Overall, the coverage plots give an impression about the complexity of the data. T^{ m } segments (dark grey boxes) or blocks of adjacent T^{ s } segments (grey boxes) nicely reflect the genomic organization in this region. The two alternative operons (cag1825 and cag1822) suggested in [2] can be confirmed when taking into account the ${T}_{\mathit{\text{sub}}}^{s}$ (light grey inlays). In total, 39 and 57 T^{ m } and ${T}^{s}/{T}_{\mathit{\text{sub}}}^{s}$ segments, respectively, were predicted from which 23 and 31, respectively, correspond to the 31 manually selected TSSs from [2]. Among them the acid induced internal TSS in cag23.
Simulated data
RNASEG results on simulated data
a  Fmeasure  Recovered sequencing data [%]  

u  1  10  100  1,000  1  10  100  1,000  
t = 1 (170 h)  1  0.93  0.93  0.93  0.93  100.00  100.00  99.95  97.92  
10    0.93  0.92  0.93    100.00  99.96  98.10  
100      0.86  0.86      99.91  99.12  
1,000        0.67        97.73  
t = 100 (156 h)  1  0.93  0.93  0.93  0.92  100.00  100.00  99.96  98.63  
10    0.93  0.92  0.92    100.00  99.96  98.10  
100      0.86  0.86      99.91  99.12  
1,000        0.67      97.73  
t = 1,000 (110 h)  1  0.87  0.86  0.83  0.81  100.00  99.98  99.83  97.76  
10    0.86  0.83  0.81    99.98  99.83  97.76  
100      0.82  0.80      99.81  97.76  
1,000        0.67        97.75 
Performance
As mentioned before, we restrict the computation allowing only for a maximum segment length $\widehat{k}$. Nevertheless the algorithm is still computationally demanding when applied on a genomic scale. Memory consumption can be estimated as follows. The dominating elements are the matrices G, mI and mt and their sum accounts for over 99% of the memory consumption. Given that each value is stored in 8 bytes (double precision float) the G and mI matrices need $8n\widehat{k}$ and $8n\widehat{\mathit{\text{cp}}}$ bytes of memory, respectively. The mt matrix stores integer values needing only 4 bytes and thus its memory consumption is $4n\widehat{\mathit{\text{cp}}}$, which is half that of mI. In total the memory consumption can be estimated with the equation $8n(\widehat{k}+1.5\widehat{\mathit{\text{cp}}})$ bytes.
The runtime scales linearly with n, $\widehat{k}$ and $\widehat{\mathit{\text{cp}}}$ each. For each cp we compute optimal segmentations for each 0 < i ≤ n. Interestingly, the computation for each i depends solely on results from the previous change point numbers, thus allowing for parallel computation over all i. For this we make use of openMP and the runtime scales nearly reciprocallinear with the number of threads. Computation of our presented results took roughly 8 hours, using 30 CPU cores (AMD Opteron 6282 SE) and a maximum of 12 GB of memory.
RNASEG also checks the values of $\widehat{k}$ and $\widehat{\mathit{\text{cp}}}$ for plausibility. During the computation of the array N, the algorithm counts the number of possible starts π and determines the largest gap between two adjacent starts δ. If $\widehat{\mathit{\text{cp}}}>2\pi +1$ it is reduced to 2π + 1 and if $\widehat{k}<\frac{\delta}{2}$ it is increased to $\frac{2\delta}{3}$. The conservative increase of $\widehat{k}$ is a compromise between increased runtime and the chance to get a valid segmentation. Note that a gap between putative TSSs may be overcome by two segments, one transcript and one nontranscript segments. Thus, in theory gaps of size $2\widehat{k}$ may be segmented correctly.
The parameter t controls the number of putative transcript starts, and thus effects the values of $\widehat{k}$ and $\widehat{\mathit{\text{cp}}}$. As a rule of thumb, the higher t, the lower $\widehat{\mathit{\text{cp}}}$ and the higher $\widehat{k}$. For our simulated data this effect is reflected by decreased runtimes for higher values of t. Memory consumption was more or less constant at 46 Gb because the automatically adjusted $\widehat{\mathit{\text{cp}}}$ numbers (273, 255 and 171 for t = 1, t = 100 and t = 1,000 respectively) were on a relatively low level compared to $\widehat{k}$ and n (20,000 and 302,371, respectively), which dominate memory usage.
Discussion
Using the SCM approach we developed a tool, namely RNASEG, for the mapping of 5’ and 3’ transcript boundaries based on dRNAseq data. Previous dRNAseq based studies on bacteria [2, 13] mainly made use of primary libraries to identify different classes of TSSs, but neglected 3’ends. These are of special interest for cisantisense or transacting sRNAs which lack a coding sequence to determine their approximate range in the genome. Our results show that, despite the partly noisy data, RNASEG performs well and can be used to infer transcriptional units from dRNAseq experiments. Compared with a manually curated operon prediction, our method reconstructs 70% of the known operons and misses many others by only a few genes. This failure can be mainly attributed to the presence of internal TSSs, which result in the prediction of several adjacent transcripts rather than a long one. Furthermore, these alternative transcripts might constitute interesting operon variants. Availability of more robust data together with algorithmic improvements, as described below, will likely yield even better results.
We expect predicted 5’ends of transcripts to be more accurate than their 3’ends for two reasons. First, primary libraries within a dRNAseq experiment provide distinct information on the 5’ends of transcripts and we do not have such data for 3’ends. Second, transcription termination is not as specific as transcription initiation. Especially, Rhoindependent termination does not lead to defined 3’ends since it is a dynamic process guided by the RNA itself ([14], Review). The thermodynamic characteristics of the terminator hairpin and the successive Utail heavily influence termination efficiency [15] and readthrough is a common phenomenon.
A recent study of the transcriptome of the cyanobacterium Synechococcus elongatus PCC 7942 [16] also applies the SCM approach for the identification of noncoding transcripts. Here, for noncoding transcripts the segmentation is applied strandspecific to 15 kb intervals with 5 kb overlaps and 30 change points. Segments with a mean coverage below two reads per nucleotide are classified as nontranscribed regions and removed together with segments overlapping previously defined transcripts for annotated genes. For the remaining segments the 5’ and 3’end are adjusted using a statistical approach, which models the positional drop in read coverage by a binomial test. By design, this test compares positions 1 and 2 nt apart, thus it is susceptible to noise, especially for low coverage transcripts.
The two widely used tools for transcript assembly in eukaryotic studies, CUFFLINKS [17] and SCRIPTURE [6], are tailored to detect transcript isoforms. They are designed for RNAseq of mRNAs, which makes use of the polyAtail for cDNA synthesis, and perform best with pairedend data. In contrast to SCRIPTURE, CUFFLINKS can be applied to nonpairedend data. Although the authors do not recommend CUFFLINKS for the analysis of bacteria [18], we have applied it to our data with default settings and got no reasonable results (data not shown). We think, that the problems mainly originate from the data. RNAseq and also dRNAseq data from bacterial transcriptomes harbour much more noise than polyAguided RNAseq data from eukaryotes. Furthermore, our data does not provide pairedend information.
Currently, TSS identification within RNASEG is mainly based on primary read starts that have to exceed a given threshold. False positives may be ruled out by the fact that a TSS has to be connected to a region satisfying the transcript segment constraints. For low abundant transcripts, a constant threshold may be too simplistic and we may choose a more sophisticated method in a future version. Here, the approach used in [19] seems promising to us, since it explicitly makes use of the two libraries provided by a dRNAseq experiment. Roughly speaking, the read start counts of both the libraries are modeled by a Poisson distribution and the difference of these distributions (which follows a Skellam distribution) is used to compute pvalues, based on which significant TSSs are identified.
The sequencing costs will drop substantially within the next years, thus more sophisticated data sets will become affordable. Especially, data for different growth conditions and with biological replicates will become standard. RNASEG can be easily extended to make use of these. For example, replicate data will contribute equally to the sum of squared residuals, as it is already implemented in [3]. Different growth conditions may be used in such a fashion, that the change of the primary starts at the mapped transcript start should be similar to the change of the mean secondary coverage of the complete transcript. The relation of these two measures is likely not linear and, thus, needs to be carefully analyzed.
Runtime and memory consumption are quite large for the current version of RNASEG. We have several ideas how to improve on this. One solution would be to decrease the resolution. At the moment we work with singlenucleotide resolution, but switching to, e.g., 10 nt resolution would decrease runtime and memory consumption nearly by a factor of 100. Of course, we would loose accuracy but mainly for the 3’ends since the mapping of segment starts to positions with a reasonable number of primary starts can still be done with single nucleotide precision. Memory consumption would benefit in the same way from the reduced resolution.
Conclusions
With RNASEG we provide the first method for the prediction of transcription units tailored for dRNAseq data. It will help in wholetranscriptome characterization and in the identification of operon structures and 5’ and 3’UTRs. The latter are important regions in posttranscriptional gene regulation by ncRNAs and, thus, the results will improve subsequent studies, such as ncRNA target prediction or the identification of cisregulatory elements and transcription termination signals.
Availability and requirements
Project name: RNASEG;
Project home page:http://www.comptrans.unifreiburg.de/Software/RNAseg;
Operating system(s): Platform independent;
Programming language: C++;
Other requirements: Boost libraries >= 1.23;
License: GNU GPLv2
Declarations
Acknowledgements
This work was in part supported by the Deutsche Forschungsgemeinschaft [Vo 1450/21 to B.V.]. The article processing charge was funded by the German Research Foundation (DFG) and the Albert Ludwigs University Freiburg in the funding programme Open Access Publishing.
Authors’ Affiliations
References
 Wang Z, Gerstein M, Snyder M: RNASeq: a revolutionary tool for transcriptomics. Nat Rev Genet. 2009, 10: 5763. 10.1038/nrg2484.View ArticlePubMed CentralPubMedGoogle Scholar
 Sharma CM, Hoffmann S, Darfeuille F, Reignier J, Sittka SFA, Chabas S, Reiche K, Hackermüller J, Reinhardt R, Stadler PF, Vogel J: The primary transcriptome of the major human pathogen Helicobacter pylori. Nature. 2010, 464 (7286): 250255. 10.1038/nature08756.View ArticlePubMedGoogle Scholar
 Huber W, Toedling J, Steinmetz LM: Transcript mapping with highdensity oligonucleotide tiling arrays. Bioinformatics. 2006, 22 (16): 19631970. 10.1093/bioinformatics/btl289.View ArticlePubMedGoogle Scholar
 Nicolas P, Leduc A, Robin S, Rasmussen S, Jarmer H, Bessières P: Transcriptional landscape estimation from tiling array data using a model of signal shift and drift. Bioinformatics. 2009, 25 (18): 23412347. 10.1093/bioinformatics/btp395.View ArticlePubMed CentralPubMedGoogle Scholar
 Roberts A, Pimentel H, Trapnell C, Pachter L: Identification of novel transcripts in annotated genomes using RNASeq. Bioinformatics. 2011, 27 (17): 23252329. 10.1093/bioinformatics/btr355.View ArticlePubMedGoogle Scholar
 Guttman M, Garber M, Levin JZ, Donaghey J, Robinson J, Adiconis X, Fan L, Koziol MJ, Gnirke A, Nusbaum C, Rinn JL, Lander ES, Regev A: Ab initio reconstruction of cell typespecific transcriptomes in mouse reveals the conserved multiexonic structure of lincRNAs. Nat Biotechnol. 2010, 28 (5): 503510. 10.1038/nbt.1633.View ArticlePubMed CentralPubMedGoogle Scholar
 Birol I, Jackman SD, Nielsen CB, Qian JQ, Varhol R, Stazyk G, Morin RD, Zhao Y, Hirst M, Schein JE, Horsman DE, Connors JM, Gascoyne RD, Marra MA, Jones SJM: De novo transcriptome assembly with ABySS. Bioinformatics. 2009, 25 (21): 28722877. 10.1093/bioinformatics/btp367.View ArticlePubMedGoogle Scholar
 Li R, Yu C, Li Y, Lam T, Yiu S, Kristiansen K, Wang J: SOAP2: an improved ultrafast tool for short read alignment. Bioinformatics. 2009, 25 (15): 19661967. 10.1093/bioinformatics/btp336.View ArticlePubMedGoogle Scholar
 Hoffmann S, Otto C, Kurtz S, Sharma CM, Khaitovich P, Vogel J, Stadler PF, Hackermüller J: Fast mapping of short sequences with mismatches, insertions and deletions using index structures. PLoS Comput Biol. 2009, 5 (9): e100050210.1371/journal.pcbi.1000502.View ArticlePubMed CentralPubMedGoogle Scholar
 Rutherford K, Parkhill J, Crook J, Horsnell T, Rice P, Rajandream M, Barrell B: Artemis: sequence visualization and annotation. Bioinformatics. 2000, 16: 944945. 10.1093/bioinformatics/16.10.944.View ArticlePubMedGoogle Scholar
 Mao F, Dam P, Chou J, Olman V, Xu Y: DOOR: a database for prokaryotic operons. Nucl Acids Res. 2009, 37: D459D463. 10.1093/nar/gkn757.View ArticlePubMed CentralPubMedGoogle Scholar
 Li W: RNASeqReadSimulator. 2013, [https://github.com/davidliwei/RNASeqReadSimulator],Google Scholar
 Mitschke J, Georg J, Scholz I, Sharma CM, Dienst D, Bantscheff J, Voß B, Steglich C, Wilde A, Vogel J, Hess WR: An experimentally anchored map of transcriptional start sites in the model cyanobacterium Synechocystis sp. PCC6803. Proc Natl Acad Sci U S A. 2011, 108 (5): 21242129. 10.1073/pnas.1015154108.View ArticlePubMed CentralPubMedGoogle Scholar
 Landick R: RNA polymerase slides home: pause and termination site recognition. Cell. 1997, 88 (6): 741744. 10.1016/S00928674(00)819194.View ArticlePubMedGoogle Scholar
 Lesnik EA, Sampath R, Levene HB, Henderson TJ, McNeil JA, Ecker DJ: Prediction of rhoindependent transcriptional terminators in Escherichia coli. Nucleic Acids Res. 2001, 29 (17): 35833594. 10.1093/nar/29.17.3583.View ArticlePubMed CentralPubMedGoogle Scholar
 Vijayan V, Jain IH, O’Shea EK: A high resolution map of a cyanobacterial transcriptome. Genome Biol. 2011, 12 (5): R4710.1186/gb2011125r47.View ArticlePubMed CentralPubMedGoogle Scholar
 Trapnell C, Williams BA, Pertea G, Mortazavi A, Kwan G, van Baren MJ, Salzberg SL, Wold BJ, Pachter L: Transcript assembly and quantification by RNASeq reveals unannotated transcripts and isoform switching during cell differentiation. Nat Biotechnol. 2010, 28 (5): 511515. 10.1038/nbt.1621.View ArticlePubMed CentralPubMedGoogle Scholar
 Trapnell C: Can i use cufflinks with RNASeq data from bacteria?. 2014, [http://cufflinks.cbcb.umd.edu/faq.html#bact],Google Scholar
 Schmidtke C, Findeiß S, Sharma CM, Kuhfuß J, Hoffmann S, Vogel J, Stadler PF, Bonas U: Genomewide transcriptome analysis of the plant pathogen Xanthomonas identifies sRNAs with putative virulence functions. Nucleic Acids Res. 2012, 40 (5): 20202031. 10.1093/nar/gkr904.View ArticlePubMed CentralPubMedGoogle Scholar
Copyright
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly credited. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.