Modeling the next generation sequencing sample processing pipeline for the purposes of classification
- Noushin Ghaffari^{1}Email author,
- Mohammadmahdi R Yousefi^{2},
- Charles D Johnson^{1},
- Ivan Ivanov^{3} and
- Edward R Dougherty^{4, 5}
https://doi.org/10.1186/1471-2105-14-307
© Ghaffari et al.; licensee BioMed Central Ltd. 2013
Received: 1 March 2013
Accepted: 9 October 2013
Published: 11 October 2013
Abstract
Background
A key goal of systems biology and translational genomics is to utilize high-throughput measurements of cellular states to develop expression-based classifiers for discriminating among different phenotypes. Recent developments of Next Generation Sequencing (NGS) technologies can facilitate classifier design by providing expression measurements for tens of thousands of genes simultaneously via the abundance of their mRNA transcripts. Because NGS technologies result in a nonlinear transformation of the actual expression distributions, their application can result in data that are less discriminative than would be the actual expression levels themselves, were they directly observable.
Results
Using state-of-the-art distributional modeling for the NGS processing pipeline, this paper studies how that pipeline, via the resulting nonlinear transformation, affects classification and feature selection. The effects of different factors are considered and NGS-based classification is compared to SAGE-based classification and classification directly on the raw expression data, which is represented by a very high-dimensional model previously developed for gene expression. As expected, the nonlinear transformation resulting from NGS processing diminishes classification accuracy; however, owing to a larger number of reads, NGS-based classification outperforms SAGE-based classification.
Conclusions
Having high numbers of reads can mitigate the degradation in classification performance resulting from the effects of NGS technologies. Hence, when performing a RNA-Seq analysis, using the highest possible coverage of the genome is recommended for the purposes of classification.
Keywords
Background
In recent years, modern high throughput sequencing technologies have become one of the essential tools in measuring the number of transcripts of each gene in a cell population or even in individual cells. Such information could be used to detect differential gene expression due to different treatment or phenotype. In our case we are interested in using gene-expression measurements to classify phenotypes into one of two classes. The accuracy of classification will depend on the manner in which the phenomena are transformed into data by the measurement technology. We consider the effects of Next-Generation Sequencing (NGS) and Serial Analysis of Gene Expression (SAGE) on gene-expression classification using currently accepted measurement modeling. The accuracy of classification problem has previously been addressed for the LC-MS proteomics pipeline, where state-of-the-art modeling is more refined, the purpose being to characterize the effect of various noise sources on classification accuracy [1].
NGS technology provides a discrete counting measurement for gene-expression levels. In particular, RNA-Seq sequences small RNA fragments (mRNA) to measure gene expression. When a gene is expressed, it produces mRNAs. The RNA-Seq experiment randomly shears and converts the RNA fragments to cDNAs, sequences them, and finally outputs the results in the form of short reads [2, 3]. After obtaining those reads, a typical part of a processing pipeline is to map them back to a reference genome to determine the gene-expression levels. The number of reads mapped to a gene on the reference genome defines the count data, which is a discrete measure of the gene-expression levels. Two popular models for statistical representation of the discrete NGS data are the negative binomial [4, 5] and Poisson [6]. The negative binomial model is more general because it can mitigate over-dispersion issues associated with the Poisson model; however, with the relatively small number of samples available in most current NGS experiments, it is difficult to accurately estimate the dispersion parameter of the negative binomial model. Therefore, in this study we choose to model the NGS data processing pipeline through the transformation via a Poisson model, for the purposes of phenotype classification.
SAGE technology produces short continuous sequences of nucleotides, called tags. After a SAGE experiment is done, one can measure the expression level of a particular region/gene of interest on the genome by counting the number of tags that map to it. SAGE is very similar to RNA-Seq in nature and in terms of statistical modeling. The SAGE data processing pipeline is traditionally modeled as a Poisson random vector [7, 8]. We follow the same approach for synthetically generated SAGE-like data sets.
Our overall methodology is to generate three different types of synthetic data: (1) actual gene expression concentration, called MVN-GC, from a multivariate normal (Gaussian) model formulated to model various aspects of gene expression concentration [9]; (2) Poisson transformed MVN-GC data, called NGS-reads, with specifications that resemble NGS reads; and (3) Poisson transformed MVN-GC data, called SAGE-tags, where the characteristics of the data model SAGE data. The classification results related to these three different types of data sets indicate that MVN-GC misclassification errors are lower compared to data subjected to transformations that produce either NGS-reads or SAGE-tags data. Moreover, classification using RNA-Seq synthetic data outperforms classification using SAGE data when the number of reads is in an acceptable range for an RNA-Seq experiment. The better performance is attributed to the significantly higher genome coverage associated with the RNA-Seq technology.
Next-generation sequencing technologies
Next-Generation Sequencing refers to a class of technologies that sequence millions of short DNA fragments in parallel, with a relatively low cost. The length and number of the reads differ based on the technology. Currently, there are three major commercially available platforms/technologies for NGS: (1) Illumina, (2) Roche, and (3) Life Technologies [10, 11]. The underlying chemistry and technique used in each platform is unique and affects the output. In this paper, we focus on Illumina sequencers and use the NGS term to refer to this platform. High-throughput sequencing and NGS are used interchangeably.
The specific application of NGS for RNA sequencing is called RNA-Seq [2], which is a high-throughput measurement of gene-expression levels of thousands of genes simultaneously as represented by discrete expression values for regions of interest on the genome (e.g. genes). NGS has many advantages when compared to the available microarray expression platforms. NGS does not depend on prior knowledge about regions of expression to measure a gene [11], whereas the microarray probes are designed based on known sequences [12]. The background correction, probe design and spot filtering, which are typical for microarray-based technology, are no longer problematic due to the different nature of NGS technology. RNA-Seq enables scientists to discover new splice junctions [13], due to its flexibility and independence of the pre-designed probes. The prediction of absolute copy-number variation (CNV) is another great advantage of RNA-Seq, which allows scientists to identify large segments of deletions/duplications in a genome with respect to another (a reference) genome [14]. Detecting single-nucleotide polymorphisms (SNP) is yet another application of RNA-Seq [3, 15]. Furthermore, it has been shown that RNA-Seq can detect spike-in RNA controls with good accuracy and reproducibility [16].
The RNA-Seq process starts with samples that are randomly sheared to generate millions of small RNA fragments. These fragments are then converted to cDNA and the adapter sequences are ligated to their ends. The length of a fragment can vary between 30 bp - 110 bp, approximately. The Illumina system provides flowcells for sequencing by synthesis and its reversible terminator chemistry[2, 3, 10]. A flowcell is an eight-channel glass and each channel is commonly referred to as a lane. The size selected fragments with the adapters attach to the flowcell surface inside the lanes and generate clusters of the same fragment through bridge amplification. Following the bending and attaching of both sides of the fragment to the surface, the strand duplicates. This process is repeated many times and results in a cluster of fragments. After the cluster generation step, a pool of floating nucleotides is added to the flowcell along with DNA polymerase to incorporate to the single strand fragments in each cluster. Each nucleotide incorporation makes a unique fluorescent label and the images are captured after the addition. Finally, image processing and base calling determine the base at each cycle of the sequencing. Each cluster produces a read whose length equals the number of cycles. Each RNA-Seq experiment produces millions of reads depending on the input RNA material, length of the reads, desired coverage for the reference genome, number of samples per lane, etc. Following the sequencing experiment, the expression levels of the genes are estimated by mapping the reads to the reference genome. There are many algorithms developed for this task, including: ELAND [3], Bowtie [17], BWA [18], MAQ [19], SHRiMP [20], mrFast [14], mrsFAST [21], SOAP [22], etc. After the gene expressions are determined, they can be used in further analysis, such as SNP detection or detecting differentially expressed genes.
The entire RNA-Seq sample processing pipeline from generating reads to calling gene expressions can involve different sources of error, e.g. the basecalling is a probabilistic procedure and the quality scores assigned to each base of the reads are prone to small likelihoods of being wrong. The certainty of reference genomes and mapping algorithms are additional issues that need attention. Thus, the entire RNA-Seq sample processing pipeline should be considered from a probabilistic point of view. In this study, we model the above mentioned errors as a noise term in the model of the sample processing pipeline.
Discussion
In this section, we consider three specific models for “actual” gene expression data, RNA-Seq count data and SAGE tags data. The models are used to synthetically generate data for the simulation experiments described in this paper. The performances of different classification schemes are analyzed and compared across these three synthetically generated types of data.
A common assumption for modeling of the original mRNA expressions is that they follow a multivariate Gaussian distribution [9, 23, 24]. Starting with this assumption, we model and generate the RNA-Seq and SAGE data by applying a specific nonlinear Poisson transformation to the mRNA expression model. All data are synthetically generated according to a biologically relevant model to emulate the real experimental situations where the number of features/genes is very large, usually tens of thousands, and only a small number of sample points is available. Knowing the full distributional characteristics of the synthetic data makes it possible to measure the classification performance as described by the respective error rates.
MVN-GC model
The model proposed in [9] uses a block-based structure on the covariance matrix, which is a standard tool to model groups of interacting variables where there is negligible interaction between the groups. In genomics, genes within a block represent a distinct pathway and are correlated, whereas genes not in the same group are uncorrelated [25, 26]. Sample points are drawn randomly and independently from two equally likely classes, 0 and 1, each sharing the same D features. There are also c equally likely subclasses in class 1 with different parameters for the probability distribution of the features. The subclasses model scenarios typically seen as different stages or subtypes of a cancer. Each sample point in class 1 belongs to one and only one of these subclasses. Features are categorized into two major groups: markers and non-markers. Markers resemble genes associated with a disease or condition related to the disease and they have different class-conditional distributions for the two classes.
NGS-reads and SAGE-tags models
COV aims to model the unknown errors that can occur during an/a NGS/SAGE experiment, including basecalling, mapping reads to the reference genome, etc. The term E[ X_{i,j}|s_{ i }] serves as the single parameter of a Poisson distribution that models the NGS/SAGE processes by generating random non-negative integers, as the read counts or tag numbers data, having expected value equal to the right-hand side of equation (2).
To generate synthetic datasets for the purposes of our simulation study we proceed as follows: for a sample point i in the experiment, we first randomly generate a number s_{ i } from a uniform distribution, U (α,β), where α > 0 and β > α. Then, a random sample point, λ_{i,j}, is drawn from the MVN-GC model and its value is perturbed with θ_{i,j}, which is drawn randomly according to its distribution defined in (3). Using (2), the mean of the Poisson distribution is calculated for each sample point i and gene j, and a single random realization of X_{i,j} is generated. The processes of generating count data for RNA-Seq reads and SAGE tag numbers are very similar, but the total number of reads per NGS run is significantly more than tags for a SAGE experiment. Therefore, we only change α and β to get the desired number of read counts or tags. We also assume that the SAGE experiments always have a fixed range for the total number of tags, whereas RNA-Seq has a variety of ranges for the total read counts. By having different ranges for the read counts, we can compare the performance of classification resulting from NGS-reads and SAGE-tags models under different experimental settings.
Classification schemes
The setup for the classification problem is determined by a joint feature-label probability distribution F_{ X Y }, where $\mathbf{X}\in {R}^{D}$ is a random feature vector and Y ∈ {0,1} is the unknown class label of X. In the context of genomics, the feature vector is usually the expression levels of many genes and the class labels are different types or stages of disease to which sample points belong to. A classifier rule model is defined as a pair (Ψ, Ξ), where Ψ is a classification rule, possibly including feature selection, and Ξ is a training-data error estimation rule. In a typical classification task, a random training set ${\mathcal{S}}_{n}=\left\{\right({\mathbf{X}}_{1},{Y}_{1}),({\mathbf{X}}_{2},{Y}_{2}),\dots ,({\mathbf{X}}_{n},{Y}_{n}\left)\right\}$ is drawn from F_{ X Y } and the goal is to design a classifier ${\psi}_{n}=\Psi \left({\mathcal{S}}_{n}\right)$, which takes X as the input and outputs a label Y. The true classification error of ψ_{ n } is given by ε_{ n } = P (ψ_{ n } (X) ≠ Y). The error estimation rule Ξ provides an error estimate, ${\widehat{\epsilon}}_{n}=\mathrm{\Xi}\left({\mathcal{S}}_{n}\right)$, for ψ_{ n }.
In this study, we consider linear discriminant analysis (LDA), three nearest neighbors (3NN) and radial basis function support vector machine (RBF-SVM) as the classification rules, and report the true error of the classifiers. We implement t-test feature selection (as a part of the classification rule) before the classifier design procedure to select d ≤ D features with highest t-scores. The training set with d features is then used to design the classifier. The same d features are also used for finding the true error of the designed classifier.
where ${\stackrel{\u0304}{\mu}}_{y}$ is the sample mean for class y ∈ {0,1}, and $\hat{\mathbf{\Sigma}}$ is the pooled sample covariance matrix.
3NN is a special case of kNN rule (with k = 3), which is a nonparametric classification rule based on the training data. The kNN classifier assigns a label, 0 or 1, to a point x according to the majority of the labels of the k nearest training data points to it. To avoid tied votes in binary classification problems, an odd number is usually chosen for k.
A support vector machine finds a maximal margin hyperplane for a given set of training sample points. If it is not possible to linearly separate the data, one can introduce some slack variables in the optimization procedure that allow the mislabeled sample points and solve the dual problem. Alternatively, one can transform the data and project it into a higher-dimensional space, where it becomes linearly separable. The equivalent classifier back in the original feature space will generally be non-linear [33, 34]. If a Gaussian radial basis function used as the kernel function, the corresponding classifier is referred to as RBF-SVM.
Simulation setup and parameters
MVN-GC, SAGE-tags and NGS-reads models parameters
Parameters | Value |
---|---|
Feature size D | 20000 |
Training sample | 60,120,180 |
size n | |
Test sample size n_{ t } | 3000 |
Class 0 (m_{0},σ_{0}) | (0.0, 0.6) |
Class 1 (m_{1},σ_{1}) | (1.0, 0.6) |
Correlation ρ | 0.4, 0.8 |
Block size l | 5 |
Global markers D_{gm} | 10 |
Subclasses c | 2 |
Heterogenous | 50 |
markers per | |
subclass D_{hm} | |
High-variance | 2000 |
non-markers D_{hv} | |
Low-variance | 17890 |
non-markers D_{lv} | |
COV | 0.05, 0.1 |
SAGE-tags range | 50K-100K |
NGS-reads range | 1K-50K, 250K-300K, 500K-550K |
5M-5M+50K, 10M-10M+50K, 15M-15M+50K | |
25M-25M+50K, 32.5M-32.5M+50K, 40M-40M+50K | |
50M-50M+50K, 75M-75M+50K, 100M-100M+50K |
As explained in the previous section, the datasets generated from the MVN-GC model are transformed into the NGS-reads and SAGE-tags datasets through equation (2) and Poisson processes. We only need to properly set the parameters COV, α, and β to get the desired number of read counts or tags. We assume that the parameter COV can take on two values, 0.01 and 0.05, representing two different levels of noise and unknown errors in the experiment. In its current state, RNA-Seq technology can provide different numbers of reads per sample, depending on many factors, such as quality of the sample, the desired coverage, sample multiplexing, etc. In this study, we examine a variety of ranges for RNA-Seq experiments. We start with a very low total number of RNA-Seq reads, which may not match the real-world experiments, however, they are necessary for comparing the SAGE-tags and NGS-reads models with similar coverage. Furthermore, demonstrating the classification results for the NGS-reads model with wide ranges introduces an extra internal variability, which makes interpretations of the results rather difficult. Table 1 lists the NGS-reads ranges we have considered in this study. In a typical SAGE experiment, one expects 50K to 100K tags [35, 36]. Using trial and error, we have found that by choosing the parameters α = 2.0 and β = 3.75 in the distribution of s_{ i }, the observed number of tags usually falls within this range. Similarly, the parameters α and β are chosen to meet the range requirements in the NGS-reads model.
Our goal is to study the performance of the traditional classification schemes on different sources of random samples; thus, we take a Monte-Carlo approach and generate 10,000 random training sets of size n from the MVN-GC model, transform them to the corresponding NGS-reads and SAGE-tags samples, and apply the classification schemes to each training sample. By taking such an approach, we aim to compare the classification performance of the three pipelines, in terms of the true error of the deigned classifiers. We also study the effect of NGS-reads and SAGE-tags transformations on the performance of a simple t-test biomarker discovery method, where we report the probability that global markers are recovered when d ≪ D features are selected after the feature-selection step.
Results
Percentage of the MVN-GC experiments recovering certain number of global markers after feature selection
Correlation | Sample | <8 | 8 | 9 | 10 |
---|---|---|---|---|---|
size | markers | markers | markers | markers | |
n=60 | 5.4733 | 15.8242 | 40.4508 | 38.2517 | |
ρ=0.4 | n=120 | 0.0108 | 0.2375 | 6.1483 | 93.6033 |
n=180 | 0.0000 | 0.0017 | 0.3367 | 99.6617 | |
n=60 | 9.5742 | 9.7425 | 22.0408 | 58.6425 | |
ρ=0.8 | n=120 | 0.1742 | 0.4692 | 2.6925 | 96.6642 |
n=180 | 0.0025 | 0.0175 | 0.1933 | 99.7867 |
Percentage of the NGS-reads experiments recovering certain number of global markers after feature selection for COV=0.05 and ρ =0 . 4
NGS-reads | Sample | <8 | 8 | 9 | 10 |
---|---|---|---|---|---|
range | size | markers | markers | markers | markers |
n=60 | 99.64 | 0.34 | 0.02 | 0.00 | |
1K-50K | n=120 | 69.98 | 20.48 | 8.64 | 0.90 |
n=180 | 27.84 | 33.76 | 31.20 | 7.20 | |
n=60 | 77.30 | 16.54 | 5.60 | 0.56 | |
50K-100K (SAGE-tags) | n=120 | 10.03 | 26.12 | 44.08 | 19.77 |
n=180 | 0.88 | 7.53 | 39.10 | 52.49 | |
n=60 | 41.30 | 32.44 | 22.00 | 4.26 | |
250K-300K | n=120 | 2.10 | 11.82 | 41.56 | 44.52 |
n=180 | 0.10 | 1.74 | 22.88 | 75.28 | |
n=60 | 34.54 | 32.74 | 26.22 | 6.50 | |
500K-550K | n=120 | 1.82 | 11.08 | 41.56 | 45.54 |
n=180 | 0.10 | 1.56 | 20.26 | 78.08 | |
n=60 | 30.98 | 31.36 | 30.04 | 7.62 | |
5M-5M+50K | n=120 | 1.24 | 8.72 | 37.56 | 52.48 |
n=180 | 0.04 | 1.20 | 18.92 | 79.84 | |
n=60 | 29.74 | 32.68 | 29.16 | 8.42 | |
10M-10M+50K | n=120 | 1.22 | 8.28 | 37.36 | 53.14 |
n=180 | 0.04 | 1.00 | 18.84 | 80.12 | |
n=60 | 28.96 | 33.16 | 29.92 | 7.96 | |
15M-15M+50K | n=120 | 1.06 | 8.06 | 38.38 | 52.50 |
n=180 | 0.06 | 1.36 | 18.08 | 80.50 | |
n=60 | 28.90 | 32.36 | 30.38 | 8.36 | |
25M-25M+50K | n=120 | 0.98 | 8.00 | 40.12 | 50.90 |
n=180 | 0.06 | 1.14 | 17.68 | 81.12 | |
n=60 | 28.60 | 32.50 | 30.42 | 8.48 | |
32.5M-32.5M+50K | n=120 | 1.10 | 7.48 | 39.02 | 52.40 |
n=180 | 0.02 | 1.02 | 17.80 | 81.16 | |
n=60 | 30.20 | 31.46 | 29.40 | 8.94 | |
40M-40M+50K | n=120 | 1.34 | 8.50 | 38.12 | 52.04 |
n=180 | 0.08 | 1.02 | 18.62 | 80.28 | |
n=60 | 28.84 | 32.18 | 30.32 | 8.66 | |
50M-50M+50K | n=120 | 1.22 | 8.46 | 39.02 | 51.30 |
n=180 | 0.04 | 1.28 | 17.24 | 81.44 | |
n=60 | 29.12 | 32.50 | 29.72 | 8.66 | |
75M-75M+50K | n=120 | 1.08 | 7.98 | 39.64 | 51.30 |
n=180 | 0.02 | 1.16 | 17.14 | 81.68 | |
n=60 | 29.14 | 32.16 | 30.84 | 7.86 | |
100M-100M+50K | n=120 | 1.22 | 8.78 | 39.10 | 50.90 |
n=180 | 0.08 | 1.22 | 18.32 | 80.38 |
Percentage of the NGS-reads experiments recovering certain number of global markers after feature selection for COV=0.05 and ρ =0 . 8
NGS-reads | Sample | <8 | 8 | 9 | 10 |
---|---|---|---|---|---|
range | size | markers | markers | markers | markers |
n=60 | 99.02 | 0.80 | 0.18 | 0.00 | |
1K-50K | n=120 | 66.16 | 20.68 | 11.20 | 1.96 |
n=180 | 28.54 | 28.38 | 30.84 | 12.24 | |
n=60 | 71.61 | 17.10 | 9.28 | 2.01 | |
50K-100K (SAGE-tags) | n=120 | 12.85 | 20.14 | 36.80 | 30.21 |
n=180 | 2.30 | 7.43 | 28.67 | 61.60 | |
n=60 | 40.94 | 22.42 | 24.12 | 12.52 | |
250K-300K | n=120 | 5.52 | 9.46 | 28.58 | 56.44 |
n=180 | 0.70 | 1.86 | 14.14 | 83.30 | |
n=60 | 34.48 | 21.02 | 27.56 | 16.94 | |
500K-550K | n=120 | 4.84 | 8.16 | 24.20 | 62.80 |
n=180 | 0.78 | 2.16 | 11.38 | 85.68 | |
n=60 | 31.26 | 20.80 | 26.64 | 21.30 | |
5M-5M+50K | n=120 | 3.98 | 6.62 | 21.78 | 67.62 |
n=180 | 0.66 | 2.24 | 10.72 | 86.38 | |
n=60 | 30.04 | 20.86 | 26.84 | 22.26 | |
10M-10M+50K | n=120 | 3.54 | 7.24 | 22.06 | 67.16 |
n=180 | 0.88 | 1.44 | 9.72 | 87.96 | |
n=60 | 30.08 | 20.68 | 27.92 | 21.32 | |
15M-15M+50K | n=120 | 4.32 | 7.10 | 22.42 | 66.16 |
n=180 | 0.60 | 1.72 | 9.62 | 88.06 | |
n=60 | 30.86 | 19.92 | 27.00 | 22.22 | |
25M-25M+50K | n=120 | 3.56 | 7.46 | 21.90 | 67.08 |
n=180 | 0.58 | 1.90 | 8.90 | 88.62 | |
n=60 | 30.50 | 20.20 | 27.50 | 21.80 | |
32.5M-32.5M+50K | n=120 | 3.62 | 7.00 | 21.66 | 67.72 |
n=180 | 0.80 | 1.88 | 10.22 | 87.10 | |
n=60 | 31.14 | 18.40 | 28.14 | 22.32 | |
40M-40M+50K | n=120 | 3.90 | 7.12 | 22.44 | 66.54 |
n=180 | 0.60 | 1.90 | 10.12 | 87.38 | |
n=60 | 30.92 | 19.34 | 27.20 | 22.54 | |
50M-50M+50K | n=120 | 3.58 | 6.68 | 22.24 | 67.50 |
n=180 | 0.68 | 1.66 | 10.22 | 87.44 | |
n=60 | 30.74 | 19.00 | 27.92 | 22.34 | |
75M-75M+50K | n=120 | 3.68 | 7.64 | 23.46 | 65.22 |
n=180 | 0.50 | 1.90 | 10.44 | 87.16 | |
n=60 | 31.06 | 20.32 | 26.70 | 21.92 | |
100M-100M+50K | n=120 | 3.84 | 6.48 | 21.30 | 68.38 |
n=180 | 0.86 | 1.80 | 9.94 | 87.40 |
Percentage of the NGS-reads experiments recovering certain number of global markers after feature selection for COV=0.1 and ρ =0 . 4
NGS-reads | Sample | <8 | 8 | 9 | 10 |
---|---|---|---|---|---|
range | size | markers | markers | markers | markers |
n=60 | 99.62 | 0.36 | 0.02 | 0.00 | |
1K-50K | n=120 | 70.26 | 20.64 | 8.02 | 1.08 |
n=180 | 30.16 | 32.46 | 29.80 | 7.58 | |
n=60 | 78.67 | 15.75 | 5.08 | 0.50 | |
50K-100K (SAGE-tags) | n=120 | 10.75 | 24.12 | 43.45 | 18.68 |
n=180 | 1.01 | 8.32 | 40.50 | 50.17 | |
n=60 | 44.64 | 31.70 | 20.00 | 3.66 | |
250K-300K | n=120 | 2.60 | 13.76 | 44.24 | 39.40 |
n=180 | 0.10 | 2.08 | 23.84 | 73.98 | |
n=60 | 37.16 | 32.54 | 24.48 | 5.82 | |
500K-550K | n=120 | 1.92 | 11.42 | 42.50 | 44.16 |
n=180 | 0.06 | 1.80 | 21.98 | 76.16 | |
n=60 | 32.36 | 31.70 | 28.84 | 7.10 | |
5M-5M+50K | n=120 | 1.46 | 9.38 | 41.20 | 47.96 |
n=180 | 0.08 | 1.68 | 20.06 | 78.18 | |
n=60 | 33.04 | 31.94 | 27.78 | 7.24 | |
10M-10M+50K | n=120 | 1.52 | 9.34 | 41.32 | 47.82 |
n=180 | 0.10 | 1.48 | 19.44 | 78.98 | |
n=60 | 31.92 | 32.64 | 28.00 | 7.44 | |
15M-15M+50K | n=120 | 1.40 | 8.36 | 41.90 | 48.34 |
n=180 | 0.08 | 1.42 | 21.20 | 77.30 | |
n=60 | 31.82 | 32.04 | 28.36 | 7.78 | |
25M-25M+50K | n=120 | 1.48 | 9.32 | 40.86 | 48.34 |
n=180 | 0.02 | 1.38 | 20.40 | 78.20 | |
n=60 | 32.06 | 32.44 | 27.88 | 7.62 | |
32.5M-32.5M+50K | n=120 | 1.42 | 9.44 | 40.50 | 48.64 |
n=180 | 0.06 | 1.32 | 19.90 | 78.72 | |
n=60 | 33.18 | 31.84 | 27.20 | 7.78 | |
40M-40M+50K | n=120 | 1.48 | 9.38 | 41.04 | 48.10 |
n=180 | 0.12 | 1.60 | 20.12 | 78.16 | |
n=60 | 31.90 | 31.94 | 29.42 | 6.74 | |
50M-50M+50K | n=120 | 1.54 | 8.76 | 41.08 | 48.62 |
n=180 | 0.02 | 1.48 | 20.72 | 77.78 | |
n=60 | 32.58 | 31.10 | 28.38 | 7.94 | |
75M-75M+50K | n=120 | 1.22 | 8.78 | 39.94 | 50.06 |
n=180 | 0.02 | 1.34 | 19.84 | 78.80 | |
n=60 | 31.40 | 32.66 | 28.90 | 7.04 | |
100M-100M+50K | n=120 | 1.50 | 9.08 | 40.38 | 49.04 |
n=180 | 0.08 | 1.36 | 20.02 | 78.54 |
Percentage of the NGS-reads experiments recovering certain number of global markers after feature selection for COV=0.1 and ρ =0 . 8
NGS-reads | Sample | <8 | 8 | 9 | 10 |
---|---|---|---|---|---|
range | size | markers | markers | markers | markers |
n=60 | 99.40 | 0.52 | 0.08 | 0.00 | |
1K-50K | n=120 | 68.14 | 19.18 | 10.92 | 1.76 |
n=180 | 29.54 | 28.16 | 30.44 | 11.86 | |
n=60 | 73.14 | 16.55 | 8.57 | 1.74 | |
50K-100K (SAGE-tags) | n=120 | 14.16 | 21.13 | 36.43 | 28.28 |
n=180 | 2.52 | 8.03 | 30.16 | 59.29 | |
n=60 | 44.26 | 21.56 | 23.10 | 11.08 | |
250K-300K | n=120 | 5.54 | 10.34 | 29.88 | 54.24 |
n=180 | 0.84 | 3.06 | 15.92 | 80.18 | |
n=60 | 36.74 | 22.00 | 26.16 | 15.10 | |
500K-550K | n=120 | 4.74 | 9.10 | 28.14 | 58.02 |
n=180 | 0.56 | 2.66 | 13.68 | 83.10 | |
n=60 | 32.76 | 20.90 | 26.96 | 19.38 | |
5M-5M+50K | n=120 | 4.38 | 7.30 | 24.74 | 63.58 |
n=180 | 0.62 | 2.08 | 11.92 | 85.38 | |
n=60 | 32.56 | 20.06 | 26.84 | 20.54 | |
10M-10M+50K | n=120 | 4.38 | 8.72 | 23.72 | 63.18 |
n=180 | 0.70 | 2.18 | 11.62 | 85.50 | |
n=60 | 31.90 | 21.54 | 27.70 | 18.86 | |
15M-15M+50K | n=120 | 4.46 | 8.16 | 23.28 | 64.10 |
n=180 | 0.76 | 2.12 | 12.02 | 85.10 | |
n=60 | 32.56 | 21.02 | 26.90 | 19.52 | |
25M-25M+50K | n=120 | 3.96 | 8.44 | 24.30 | 63.30 |
n=180 | 0.54 | 2.32 | 11.86 | 85.28 | |
n=60 | 33.00 | 21.30 | 26.74 | 18.96 | |
32.5M-32.5M+50K | n=120 | 4.18 | 7.20 | 24.10 | 64.52 |
n=180 | 0.64 | 1.98 | 10.94 | 86.44 | |
n=60 | 32.94 | 21.00 | 27.26 | 18.80 | |
40M-40M+50K | n=120 | 4.36 | 7.48 | 24.22 | 63.94 |
n=180 | 0.80 | 2.24 | 11.16 | 85.80 | |
n=60 | 32.70 | 20.64 | 26.70 | 19.96 | |
50M-50M+50K | n=120 | 4.14 | 8.52 | 23.42 | 63.92 |
n=180 | 0.74 | 2.32 | 11.84 | 85.10 | |
n=60 | 33.38 | 20.34 | 27.02 | 19.26 | |
75M-75M+50K | n=120 | 4.48 | 7.90 | 24.26 | 63.36 |
n=180 | 0.60 | 1.92 | 11.00 | 86.48 | |
n=60 | 32.70 | 21.34 | 26.86 | 19.10 | |
100M-100M+50K | n=120 | 4.54 | 7.88 | 23.64 | 63.94 |
n=180 | 0.80 | 2.58 | 11.64 | 84.98 |
The true error of a designed classifier measures the generalization capability of the classifier on a future sample point. Given a set of training sample points and a classification rule, one needs the full feature-label probability distribution to calculate the true error of the classifier designed on the training set. In this study, we find the true error of a classifier with a large independent sample. Because the training sample is random, the true error is a random variable with its own probability distribution. Therefore, to demonstrate the actual performance of the designed classifiers, the estimated probability density function (PDF) of the true error for each classification rule, distribution model and all data generation models (MVN-GC, NGS-reads and SAGE-tags) is presented.
Figures 3, 4 and 5 show that utilizing the best d features (with or without feature selection) in the model, SAGE-tags and NGS-reads for small ranges of the total reads yield similar results (or even much worse when the range is smaller) in terms of the classification performance, and both are inferior when compared to the ground truth. This may lead to a conclusion that one should not expect improvements in classification performance when gene-expression levels are processed through an NGS-reads pipeline. However, our main objective here is to show that as one increases the total number of reads for the NGS-reads model, improvement can be achieved and the error rates decrease. This conclusion confirms the intuition provided by a simple calculation about the increase of the separability of two Poisson random variables with means proportional to the number of reads. However, notice that the modeling of the sequencing pipeline introduces randomness in the means of the respective Poisson parameters describing the individual gene reads. Moreover, our focus is not that much on the separability of the two classes/phenotypes but rather on the classification performance as measured by the classification error and its dependence on ground truth (MVN-GC model) sample size, sequencing depth, and feature vectors. Our goal for modeling NGS-reads with small ranges is to demonstrate the performance of SAGE and RNA-Seq, when their data is similar. This result suggests that having a larger number of reads for the RNA-Seq experiments could compensate for the errors that can occur during the NGS-reads pipeline sample processing. Here we have shown the results only for COV = 0.05, since in our observations, it appears that changing COV has little effect on the distribution of the true error for both SAGE-tags and NGS-reads models, except that it slightly increases the variance of the PDFs. The effect of feature selection (right columns) on the performance is best shown when the sample size is small. In this case, the variance of the true error is so large that drawing any conclusion regarding the performance on a small sample would be futile. Another interesting observation is that, on average, 3NN and RBF-SVM classification rules outperform LDA for both NGS-reads and SAGE-tags data, supporting the idea that using linear classifiers for these types of data is not the best way to proceed in a highly non-Gaussian model - as is the case after the data have gone through the pipeline. The best rates for the expected true error across all settings are achieved by the RBF-SVM classification rule, even for small samples, especially with small variance.
Conclusions
In this paper, we model gene-expression levels as a multivariate Gaussian distribution that statistically captures the real mRNA levels within the cells. The newly developed technologies of sequencing DNA/RNA, referred to as NGS, and their ascendant SAGE technology generate discrete measures for gene expressions. The count data from these technologies can be modeled with a Poisson distribution. The multivariate Gaussian gene expressions are transformed through a Poisson filter to model NGS and SAGE technologies. The three categories of data are subjected to feature selection and classification. The objective is to evaluate the performance of the NGS technologies in classification. Our simulations show that when the gene expressions are directly used in classification, the best performance in terms of the classification error is achieved. The NGS-reads model generates considerably higher coverage for genes and can outperform SAGE in classification, when the experiment generates large number of reads. Even though NGS still has a variety of error sources involved in its process, its high volume of reads for a specific gene can lower the chance of misclassification. Thus, it is important to use the highest possible coverage for the entire genome while performing a RNA-Seq analysis if the goal of the study is to distinguish the samples of interest. Nevertheless, one must recognize that, as is typical with nonlinear transformations, the NGS pipeline transforms the original Gaussian data in such a way as to increase classification difficulty. As more refined modeling becomes available for the NGS pipeline, further study needs to performed, as has been done in the case of the LC-MS proteomics pipeline [1], to determine which segments of the pipeline are most detrimental to classification and what, if anything, can be done to mitigate classification degradation.
Declarations
Authors’ Affiliations
References
- Sun Y, Braga-Neto UM, Dougherty ER: Modeling and systematic analysis of the LC-MS Proteomics Pipeline. BMC Genomics. 2012, 13 (Supp 6): S2-View ArticleGoogle Scholar
- Mortazavi A, Williams BA, McCue K, Schaeffer L, Wold B: Mapping and quantifying mammalian transcriptomes by RNA-Seq. Nature Methods. 2008, 5 (7): 621-628. 10.1038/nmeth.1226.View ArticlePubMedGoogle Scholar
- Bentley DR, Balasubramanian S, Swerdlow HP, Smith GP, Milton J, Brown CG, Hall KP, Evers DJ, Barnes CL, Bignell HR, Boutell JM, Bryant J: Accurate whole human genome sequencing using reversible terminator chemistry. Nature. 2008, 456 (6): 53-59.PubMed CentralView ArticlePubMedGoogle Scholar
- Anders S, Huber H: Differential expression analysis for sequence count data. Genome Biol. 2010, 11 (10): R106-10.1186/gb-2010-11-10-r106.PubMed CentralView ArticlePubMedGoogle Scholar
- Robinson MD, McCarthy DJ, Smyth GK: edgeR: a Bioconductor package for differential expression analysis of digital gene expression data. Bioinformatics. 2010, 26: 139-140. 10.1093/bioinformatics/btp616.PubMed CentralView ArticlePubMedGoogle Scholar
- Marioni JC, Mason CE, Mane SM, Stephens M, Gilad Y: RNA-seq: An assessment of technical reproducibility and comparison with gene expression arrays. Genome Res. 2008, 18 (9): 1509-1517. 10.1101/gr.079558.108.PubMed CentralView ArticlePubMedGoogle Scholar
- Robinson MD, Smyth GK: Moderated statistical tests for assessing differences in tag abundance. Bioinformatics. 2007, 23 (21): 2881-2887. 10.1093/bioinformatics/btm453.View ArticlePubMedGoogle Scholar
- Robinson MD, Smyth GK: Small sample estimation of negative binomial dispersion, with applications to SAGE data. Biostat. 2008, 9 (2): 321-332.View ArticleGoogle Scholar
- Hua J, Waibhav T, Dougherty ER: Performance of feature selection methods in the classification of high-dimensional data. Pattern Recognit. 2009, 42 (3): 409-424. 10.1016/j.patcog.2008.08.001.View ArticleGoogle Scholar
- Mardis ER: Next-generation DNA sequencing methods. Ann Rev Genomics Human Genet. 2008, 9: 387-402. 10.1146/annurev.genom.9.081307.164359.View ArticleGoogle Scholar
- Auer PL, Doerge RW: Statistical design and analysis of RNA sequencing data. Genet. 2010, 185 (2): 405-416. 10.1534/genetics.110.114983.View ArticleGoogle Scholar
- Sun W: A statistical framework for eQTL mapping using RNA-seq data. Biometrics. 2011, 68: 1-11.PubMed CentralView ArticlePubMedGoogle Scholar
- Sultan M, Schulz MH, Richard H, Magen A, Klingenhoff A, Scherf M, Seifert M, Borodina T, Soldatov A, Parkhomchuk D, Schmidt D, O’Keeffe S, Haas S, Vingron M, Lehrach H, Yaspo ML: A global view of gene activity and alternative splicing by deep sequencing of the human transcriptome. Science. 2008, 321: 956-960. 10.1126/science.1160342.View ArticlePubMedGoogle Scholar
- Alkan C, Kidd JM, Marques-Bonet T, Aksay G, Antonacci F, Hormozdiari F, Kitzman JO, Baker C, Malig M, Mutlu O, Sahinalp SC, Gibbs RA, Eichler EE: Personalized copy number and segmental duplication maps using next-generation sequencing. Nat Genet. 2009, 41: 1061-1067. 10.1038/ng.437.PubMed CentralView ArticlePubMedGoogle Scholar
- Craig DW, Pearson JV, Szelinger S, Sekar A, Redman M, Corneveaux JJ, Pawlowski TL, Laub T, Nunn G, Stephan DA, Homer N, Huentelman MJ: Identification of genetic variants using bar-coded multiplexed sequencing. Nature Methods. 2008, 5: 887-893. 10.1038/nmeth.1251.PubMed CentralView ArticlePubMedGoogle Scholar
- Jiang L, Schlesinger F, Davis CA, Zhang Y, Li R, Salit M, Gingeras TR, Oliver B: Synthetic spike-in standards for RNA-seq experiments. Genome Res. 2011, 21 (9): 1543-1551. 10.1101/gr.121095.111.PubMed CentralView ArticlePubMedGoogle Scholar
- Langmead B, Trapnell C, Pop M, Salzberg SL: Ultrafast and memory-efficient alignment of short DNA sequences to the human genome. Genome Biol. 2009, 10: R25-10.1186/gb-2009-10-3-r25.PubMed CentralView ArticlePubMedGoogle Scholar
- Li H, Durbin R: Fast and accurate short read alignment with Burrows-Wheeler transform. Bioinformatics. 2009, 25: 1754-1760. 10.1093/bioinformatics/btp324.PubMed CentralView ArticlePubMedGoogle Scholar
- Li H, Ruan J, Durbin R: Mapping short DNA sequencing reads and calling variants using mapping quality scores. Genome Res. 2008, 18: 1851-1858. 10.1101/gr.078212.108.PubMed CentralView ArticlePubMedGoogle Scholar
- Rumble SM, Lacroute P, Dalca AV, Fiume M, Sidow A, Brudno M: SHRiMP: accurate mapping of short color-space reads. PLoS Comput Biol. 2009, 5 (5): e1000386-10.1371/journal.pcbi.1000386.PubMed CentralView ArticlePubMedGoogle Scholar
- Hach F, Hormozdiari F, Alkan C, Hormozdiari F, Birol I, Eichler E, Sahinalp SC: mrsFAST: a cache-oblivious algorithm for short-read mapping. Nature Methods. 2010, 7: 576-577. 10.1038/nmeth0810-576.PubMed CentralView ArticlePubMedGoogle Scholar
- Li R, Li Y, Kristiansen K, Wang J: SOAP: short oligonucleotide alignment program. Bioinformatics. 2008, 24 (5): 713-714. 10.1093/bioinformatics/btn025.View ArticlePubMedGoogle Scholar
- Attoor S, Dougherty ER, Chen Y, Bittner ML, Trent JM: Which is better for cDNA-microarray-based classification: ratios or direct intensities. Bioinformatics. 2004, 20 (16): 2513-2520. 10.1093/bioinformatics/bth272.View ArticlePubMedGoogle Scholar
- Dalton LA, Dougherty ER: Application of the Bayesian MMSE error estimator for classification error to gene-expression microarray data. Bioinformatics. 2011, 27 (13): 1822-1831. 10.1093/bioinformatics/btr272.View ArticlePubMedGoogle Scholar
- Dougherty ER: Validation of computational methods in genomics. Curr Genomics. 2007, 8: 1-19. 10.2174/138920207780076956.View ArticleGoogle Scholar
- Shmulevich I, Dougherty ER: Genomic Signal Processing. 2007, Princeton: Princeton University PressView ArticleGoogle Scholar
- Yousefi MR, Hua J, Dougherty ER: Multiple-rule bias in the comparison of classification rules. Bioinformatics. 2011, 27: 1675-1683. 10.1093/bioinformatics/btr262.PubMed CentralView ArticlePubMedGoogle Scholar
- Bullard JH, Purdom E, Hansen KD, Dudoit S: Evaluation of statistical methods for normalization and differential expression in mRNA-Seq experiments. BMC Bioinformatics. 2010, 11 (94): 1471-2105.Google Scholar
- Li J, Wittn DM, Johnstone IM, Tibshirani R: Normalization, testing, and false discovery rate estimation for RNA-sequencing data. Biostat. 2010, 11 (94): 1471-2105.Google Scholar
- Robinson MD, Oshlack A: A scaling normalization method for differential expression analysis of RNA-seq data. Genome Biol. 2010, 11 (R25): 1471-2105.Google Scholar
- Hoyle DC, Rattray M, Jupp R, Brass A: Making sense of microarray data distributions. Bioinformatics. 2002, 18 (4): 576-584. 10.1093/bioinformatics/18.4.576.View ArticlePubMedGoogle Scholar
- Duda RO, Hart PE, Stork DG: Pattern Classification. 2001, New York: WileyGoogle Scholar
- Boser BE, Guyon IE, Vapnik VN: A training algorithm for optimal margin classifiers. Proceedings of the 5th Annual ACM Workshop on Computational Learning Theory. 1992, 144-152. http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.21.3818,Google Scholar
- Cortes C, Vapnik VN: Support-vector networks. Mach Learn. 1995, 20: 273-297.Google Scholar
- Wang SM: Understanding SAGE data. TRENDS Genet. 2006, 23: 42-50.View ArticlePubMedGoogle Scholar
- Bianchetti L, Wu Y, Guerin E, Plewniak F, Poch O: SAGETTARIUS: a program to reduce the number of tags mapped to multiple transcripts and to plan SAGE sequencing stages. Nucleic Acids Res. 2007, 35 (18): e122-10.1093/nar/gkm648.PubMed CentralView ArticlePubMedGoogle Scholar
Copyright
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.