 METHOD
 Open Access
 Published:
DGEclust: differential expression analysis of clustered count data
Genome Biology volume 16, Article number: 39 (2015)
Abstract
We present a statistical methodology, DGEclust, for differential expression analysis of digital expression data. Our method treats differential expression as a form of clustering, thus unifying these two concepts. Furthermore, it simultaneously addresses the problem of how many clusters are supported by the data and uncertainty in parameter estimation. DGEclust successfully identifies differentially expressed genes under a number of different scenarios, maintaining a low error rate and an excellent control of its false discovery rate with reasonable computational requirements. It is formulated to perform particularly well on lowreplicated data and be applicable to multigroup data. DGEclust is available at http://dvav.github.io/dgeclust/.
Background
Nextgeneration sequencing (NGS) and highthroughput sequencing are revolutionary tools for the study of the genome, epigenome and transcriptome in a multitude of organisms (including humans) by allowing the relatively rapid production of millions of short sequence tags, which mirror particular aspects of the molecular state of the biological system of interest. A common application of NGS is the study of the transcriptome, which involves a family of methodologies, such as RNA sequencing (RNAseq) [1], cap analysis of gene expression (CAGE) [2], serial analysis of gene expression (SAGE) [3] and others. Most published studies on the statistical analysis of count data generated by NGS have focused on the themes of experimental design [4], normalisation [5,6] and the development of tests for differential expression [79]. Surprisingly, not much attention has been paid to cluster analysis.
Clustering is considered an important tool in the study of genomic data and it has been used extensively in the analysis of microarrays [1012] (see [13] for a review of different clustering methods). It involves grouping together the expression profiles of different genes across different points in time, treatments and tissues, such that expression profiles in the same group are more similar in some way to each other than to members of other groups. Genes that are clustered together across samples exhibit corelated expression patterns, which might indicate coregulation and involvement of these genes in the same cellular processes [14]. Moreover, whole samples of gene expression profiles can be clustered together, indicating a particular macroscopic phenotype, such as cancer [15].
A large class of clustering methods relies on the definition of a distance metric, which quantifies the similarity between any two gene expression data points. Subsequently, clusters are formed, such that the distance between any two data points in the same cluster is minimised. Typical methods in this category are kmeans clustering and selforganising maps [13]. Another important category includes modelbased clustering algorithms. In this case, the whole gene expression dataset is modelled as a random sample from a finite mixture of probability distributions, where each component of the mixture corresponds to a distinct cluster. The parameters of each component in the mixture (e.g. mean and variance) are usually estimated using an expectationmaximisation algorithm [13]. Hierarchical clustering is yet a third type of clustering methodology, which is particularly suited for modelling genomic (often hierarchically organised) data. It generates a hierarchical series of nested clusters, which can be represented graphically as a dendrogram. This stands in contrast to partitionbased methods (e.g. kmeans or selforganising maps), which decompose the data directly into a finite number of nonoverlapping clusters [13].
In this article, we present a modelbased statistical methodology and associated software (DGEclust) for clustering digital expression data and we show how it can be used in differential expression analysis. From a theoretical standpoint, the significance of the proposed methodology stems from its unification of differential expression and clustering (i.e. it treats differential expression as a particular clustering configuration of the data). This makes it possible to bypass the need for a particular statistical test when inferring differential expression or for multiple hypothesis testing correction, such as the Benjamini–Hochberg procedure. From a statistical point of view, the proposed methodology is important because it simultaneously addresses the problem of model selection (i.e. how many clusters are supported by the data) and uncertainty (i.e. the error associated with estimating the number of clusters and the parameters of each cluster). This is made possible by exploiting a hierarchical Dirichlet process mixture model (HDPMM) [16,17], a statistical framework, which has been applied in the past for clustering microarray data [18,19], for multipopulation haplotype inference [20], for integrating heterogeneous genomic datasets [21,22] and for modelling multiple text corpora [23]. In our version of the HDPMM, individual expression profiles are drawn from the negative binomial distribution (as, for example, in [2427]) and parameter estimation is achieved using a novel, fast blocked Gibbs sampler, which efficiently processes large datasets (e.g. with more than 20K genes). We show how the output of our clustering algorithm can be used in differential expression analysis and, using simulated data and actual experimental data (RNAseq and CAGE) from a range of species, we demonstrate improved performance, compared to popular alternative methods. An early version of the proposed methodology has been presented previously in poster format and in [28].
Results and discussion
Description of the model
Formally, the production of count data using NGS assays can be thought of as random sampling of an underlying population of cDNA fragments. Thus, the counts for each tag describing a class of such fragments can, in principle, be modelled using the Poisson distribution, whose variance is, by definition, equal to its mean. However, it has been shown that, in real count data of gene expression, the variance can be larger than what is predicted by the Poisson distribution [2932]. An approach that accounts for the socalled overdispersion in the data is to adopt quasilikelihood methods, which augment the variance of the Poisson distribution with a scaling factor, thus dropping the assumption of equality between the mean and variance [3336]. An alternative approach is to use the negative binomial distribution, which is derived from the Poisson, assuming a gammadistributed rate parameter. The negative binomial distribution incorporates both a mean and a variance parameter, thus modelling overdispersion in a natural way [2426]. For this reason, in this article we use the negative binomial distribution for modelling count data.
We indicate the number of reads for the ith feature (e.g. gene) at the jth sample/library with the variable y _{ ij }. Each library j is assigned to a group of libraries l=λ(j), based on prior information on tissue, experimental condition, disease state, etc. There are a total of N genes, M libraries and L groups of libraries, where 2≤L≤M. We assume that y _{ ij } is distributed according to a negative binomial distribution, with mean m _{ ij } and dispersion parameter ϕ _{ i }:
where log(m _{ ij })= log(c _{ j })+ log(μ _{ i })+β _{ i λ(j)}. In the previous expression, c _{ j } is a known a priori normalising factor for library j, μ _{ i } is the mean expression level for gene i and β _{ i λ(j)} is the foldchange of the mean expression level of gene i in group λ(j). A foldchange equal to 0 indicates no change in the mean expression level for feature i in group λ(j), while a value larger (smaller) than 0 indicates over(under)expression.
The variance for expression profile y _{ ij }, $\sigma ^{2}_{\textit {ij}}=m_{\textit {ij}}+\phi _{i} m_{\textit {ij}}^{2}$ , is always larger than the mean by the quantity $\phi _{i} m_{\textit {ij}}^{2}$ . Thus, the negative binomial distribution can be thought of as a generalisation of the Poisson distribution, which accounts for overdispersion.
It follows that to specify the above model fully, we need to know the parameters ϕ _{ i }, μ _{ i } and β _{ i λ(j)} for each gene i and group of samples λ(j). We also need to compute the normalising factors c _{ j } from the data (see [37] for a review of available normalisation methods).
Information sharing between genes
A common limitation in experiments using NGS technologies is the low number or even absence of biological replicates, which complicates the statistical analysis of digital expression data. One way to compensate for small sample sizes is to assume that all genes share the same variance [30]. A less restrictive approach is to implement some type of information sharing between genes, which permits the improved estimation of genespecific parameters (e.g. the dispersion parameters) by pooling together genes with similar expression profiles [2426]. In this article, information sharing between genes and between samples is introduced in a natural way through the use of priors for the parameters of the negative binomial distribution. In this and the following section, we use Dirichlet process priors for modelling the foldchanges β _{ i λ(j)}, followed by giving the priors for the ϕ _{ i } and μ _{ i } parameters.
Within each sample group l=λ(j), we assume that the genespecific foldchanges {β _{ il }} are random and distributed according to a prior distribution G _{ l }, i.e.
Furthermore, we assume that G _{ l } is itself randomly sampled from a Dirichlet process with positive concentration parameter γ _{ l } and base probability distribution G _{0} [16]:
Dirichlet process priors are distributions over distributions and they have become a popular choice in Bayesian inference studies, since they provide an elegant solution to the problem of determining the correct number of components in mixture models. Standard theoretical results [38] state that a sample G _{ l } from Equation 3 is a discrete distribution with probability one over a countably infinite set of βs. Large values of γ _{ l } lead to a large number of similarly likely values of β, while small values of this parameter imply a small number of highly probable values of β. This and Equation 2 imply that the foldchanges β _{ il } within the lth group of samples are not all distinct. Different genes may share the same value of β or, in other words, genes are grouped in a (not known in advance) number of clusters, based on the value of β they share. Equivalently, the expression profiles of different groups of genes are drawn from different negative binomial distributions, each characterised by its own unique value of β. This clustering effect is illustarted in Figure 1.
Information sharing between samples
Up to this point, we have considered clustering of genes within the same group of samples, but not across groups of samples (e.g. tissues or conditions). However, in a given dataset, each cluster might include gene expression profiles from the same, as well as from different groups of samples. In other words, clusters are likely shared between samples that belong to different groups. This sharing of information between sample groups can be expressed naturally in the context of HDPMMs [16]. Following directly from the previous section, we assume that the base distribution G _{0} is itself random and sampled from a Dirichlet process with a global scaling parameter δ and a global base distribution H:
The above expression implies that G _{0} is (like each G _{ l }) discrete over a countably infinite set of atoms β k′, which are sampled from H, i.e. β k′∼H. Since G _{0} is the common base distribution of all G _{ j }, the atoms β k′ are shared among all samples, yielding the desired information sharing across samples (see Figure 1).
Generative model
In summary, we have the following hierarchical model for the generation of a matrix of digital gene expression data (see also Figure 1):
where l≡λ(j). Notice that the base distribution H, which provides the global prior for sampling the atoms $\beta ^{\prime }_{k}$ , was assumed normal with mean μ _{ β } and variance $\sigma ^{2}_{\beta }$ . Similarly, the logarithm of the dispersion parameters ϕ _{ i } was assumed normal with mean μ _{ ϕ } and variance $\sigma ^{2}_{\phi }$ (see, for example, [39] for a justification of this choice). The mean expression levels μ _{ i } are modelled as $\mu _{i} = (1  p_{i})p_{i}^{1}\phi _{i}^{1}$ , where ϕ _{ i } is sampled as above and p _{ i }∼Beta(0.5,0.5). This formulation greatly facilitates the posterior inference of μ _{ i } given ϕ _{ i }, since it involves sampling directly from wellknown distributions (see Additional File 1 for more details).
Inference
The definition of the HDPMM in Equations 5 is implicit. To facilitate posterior inference, an equivalent constructive representation of the above model has been introduced in [23], using Sethuraman’s stickbreaking representation of a draw from a Dirichlet process [38]. This representation introduces a matrix of indicator variables z={z _{ il }}, where each element of the matrix, z _{ il }, indicates which cluster the ith logfoldchange in the lth group of samples (i.e. β _{ il }) belongs to. Two different β’s, β _{ il } and $\beta _{i^{\prime }l^{\prime }}$ , belong to the same cluster if and only if their indicator variables, e.g. z _{ il } and $z_{i^{\prime }l^{\prime }}$ , are equal. A major aim of Bayesian inference in the above model, is to calculate the posterior distribution p(zy) of matrix z given the data matrix y.
One approach to estimate this distribution is by using Markov chain Monte Carlo (MCMC) methods, which generate a chain of random samples as a numerical approximation to the desired distribution. We have developed a blocked Gibbs sampler in the software package DGEclust, which efficiently generates new samples from the posterior p(zy). The algorithm is an extension of the method presented in [28,40] for inference in nonHDPMMs and its advantage is that it samples each element of z independently of all others. This not only results in very fast convergence, but it also allows the algorithm to be implemented in vectorised form, which takes advantage of the parallel architecture of modern multicore processors and potentially permits application of the algorithm on very large datasets. Alternative MCMC methods, which are developed on the basis of the popular Chinese restaurant franchise representation of the Hierarchical Dirichlet Process (HDP) [16,41], do not enjoy the same advantage since they are restricted because sampling each indicator variable is conditioned on the remaining ones, thus all of them must be updated in a serial fashion. Details of the algorithm are given as supplementary material (see Additional file 1).
Testing for differential expression
Assuming that the above algorithm has been applied on a digital expression dataset y and a sufficiently large chain of samples $\phantom {\dot {i}\!}z^{(T_{0}+1)},z^{(T_{0}+2)},\ldots,z^{(T_{0}+T)}$ – which approximates the posterior p(zy) – has been generated, we show how these samples can be used for differential expression analysis. We consider two classes of samples, A and B, which might represent, for example, two different tissues or experimental conditions.
A particular gene is said to be not differentially expressed (DE), if its expression measurements in classes A and B belong to the same cluster. In more formal language, we state that the posterior probability π _{ i } that gene i is not DE given data y is equal to the conditional probability p(z _{ iA }=z _{ iB }y) that the indicator variables of feature i in sample classes A and B have the same value. This probability can be approximated as a simple average over the previously generated MCMC samples $\left \{z^{T_{0}+t}\right \}_{t=1}^{T}$ :
where 1 1(·) is equal to 1 if the expression inside the parentheses is true and 0 otherwise. Given a threshold $\tilde \pi $ , we can generate a set of potentially DE features with probabilities less than this threshold, i.e. $\mathcal {D}=\{i: \pi _{i}\le \tilde \pi \}$ , where π _{ i } is calculated as in Equation 6 for all i.
As observed in [42], the quantity π _{ i } measures the conditional probability that including the ith gene in list is a Type I error, i.e. a false discovery (FD). This useful property makes possible the calculation of the conditional false discovery rate (FDR) as follows:
From Equation 7, it can be seen that always has an FDR at most equal to $\tilde \pi $ . Alternatively, one can first set a target FDR, say tFDR, and then find the maximum possible value of $\tilde \pi $ , such that $\text {FDR}(\tilde \pi) \le \text {tFDR}$ .
Notice that, unlike alternative approaches, which make use of genespecific P values, this methodology does not require any correction for multiple hypothesis testing, such as the Benjamini–Hochberg procedure. Although the computation of FDR using Equation 7 is approximate (since it depends on the accuracy of the calculation of π _{ i } using Equation 6), it is reasonable to assume that the error associated with this approximation is minimised, if sufficient care is taken when postprocessing the MCMC samples generated by the Gibbs sampler.
Application to simulated data
To assess the performance of our methodology, we applied it on simulated and actual experimental count data and we compared our results against those obtained from popular software packages, namely DESeq/DESeq2, edgeR and baySeq.
First, we applied our algorithm on simulated data, which provides the advantage that we can control the exact conditions under which the data was generated, including the true differential expression state of each gene. For this purpose, we used the function generateSyntheticData from the independent Bioconductor package compcodeR [43]. Data generated using this approach follow the negative binomial distribution, with mean and dispersion parameters estimated from actual experimental data [44,45]. Each simulated dataset included 10K genes under two different experimental conditions, with the proportion of DE genes set to either 0%, 10% or 30% (0, 1,000 or 3,000 DE genes, respectively). In addition, we considered simulated data where 50% of the genes were generated from the Poisson distribution (and, therefore, were not overdispersed), as well as simulated data which included outliers, i.e. genes with unusually high or low counts. The outliers were introduced by multiplying with probability 5% each observed count independently with a randomly generated factor between 5 and 10. Finally, we evaluated the effect of varying the percondition sample size from 2 to 4 and 8, since small sample sizes reflect the design of most current sequencing experiments. Under each unique simulation setting, we independently generated three different datasets; each evaluated method was applied separately on each dataset and the aggregated results for each simulation setting were reported. All parameters of generateSyntheticData were left as their default values, unless stated otherwise.
Following the prototypical study in [9], we compared the performance of DGEclust and the other aforementioned methods under the experimental conditions encapsulated by the synthetic data using the following criteria: (a) first, we compared the ability of all methods to identify DE genes using as performance measure the area under the receiver operating characteristic (ROC) curve (AUC) for each method, (b) second, we compared the ability of all methods to keep a low number of Type I errors using FD curves as performance measures and also by measuring the number of Type I errors each method returned given a fixed significance threshold, (c) third, we assessed the ability of all methods to keep a low FDR given a relatively large prespecified FDR threshold. Furthermore, we applied all methods on actual RNAseq data from a number of species, as well as on a large CAGE dataset obtained from five different regions of human brains. Details for each category of experiment are given in the following sections.
DGEclust successfully identifies differentially expressed genes under a number of different scenarios
We first evaluated the comparative ability of DGEclust to identify truly DE genes. All examined methods rank each gene by providing P values (edgeR and DESeq/DESeq2) or posterior probabilities (DGEclust and baySeq). Given a threshold score, genes on opposite sides of the threshold are tagged as DE or nonDE, accordingly. In an artificial dataset, the genes that were simulated to be DE are considered to be the true positive group, while the remaining genes are considered the true negative group. By computing the false positive rate (FPR) and the true positive rate (TPR) for all possible score thresholds, we can construct ROC and FD curves for each examined method. The AUC is a measure of the overall discriminative ability of a method (i.e. its ability to classify correctly features as DE or nonDE).
Our results are summarised in Figure 2. As expected, the performance of all methods improves as the sample size in each dataset increases from 2 to 4 to 8. When only overdispersed genes are present and only 10% of them are DE (top panel), DGEclust clearly demonstrates the best performance with an AUC score larger than 80%, followed by DESeq/DESeq2, edgeR and, finally, baySeq. The same trend is observed when we examine datasets with 30% DE genes (second panel), but this time the difference between DGEclust and the other methods is even more prominent. This is because the performance of edgeR, DESeq/DESeq2 and baySeq is negatively affected by the increased proportion of DE genes, while the performance of DGEclust remains essentially the same.
The negative effect of DE gene composition on DESeq, edgeR and baySeq, when datasets with large proportions of DE genes were considered, has already been observed [9] and attributed to an increased proportion of false positives due to the failure of normalisation to capture the effect of the possibly asymmetric distribution of DE genes fully. Here, we provide an alternative explanation that this effect might be an intrinsic characteristic of each classification method, since DGEclust does not seem to be affected by the presence of this asymmetry in the data, although it adopts the same normalisation method as DESeq.
We further evaluated the effect of introducing nonoverdispersed genes or outliers, as outlined above. When the fraction of Poisson distributed genes was increased from 0 to 50% (third panel), we observed a significant increase in the AUC for all methods. DGEclust ranked first in all cases, but DESeq/DESeq2 and edgeR followed closely, particularly at large sample sizes (n=8). The introduction of outliers with abnormally high or low counts (bottom panel), reduced the AUC for all methods, but again DGEclust ranked first among the alternatives for all sample sizes.
DGEclust maintains a low rate of Type I errors
Next, we evaluated the ability of DGEclust and the other methods to control the rate of Type I errors. First, we assessed this ability through the construction of FD curves, which illustrate the number of false positives as a function of the total number of positives (i.e. as the decision threshold increases). Mean FD curves for each of the cases examined in Figure 2 are illustrated in Figure 3. We measure the false positives among the first 1,000 topranked discoveries. Observe that when only overdispersed data with either small (top panel) or large (second panel) numbers of DE genes are considered, DGEclust always keeps the number of false positives among the top discoveries below that of the other evaluated methods. As with the AUC scores, this is more prominent when a large number of DE genes (30%) is considered and it holds at all sample sizes. When examining datasets including nonoverdispersed genes (third panel), DGEclust still performs better than alternative methods, at all sample sizes. For datasets with outliers (bottom panel), the difference between DGEclust and the other methods becomes less prominent and effectively disappears at large sample sizes (n=8), at which point DESeq2 also demonstrates good performance.
In a second stage, we examined the ability of all methods to control the number of Type I errors at a prespecified threshold level in the absence of any truly DE genes (see Figure 4). To make possible a comparison between methods that return P values (DESeq/DESeq2 and edgeR) and methods that return posterior probabilities (DGEclust and baySeq) and since all methods provide an estimation of the FDR, we compared the number of Type I errors (as a proportion of the total number of genes) each method made at a relatively high prespecified FDR = 0.1. In all datasets not including outliers (top two panels), DGEclust had the lowest proportion of Type I errors at small sample sizes (n=2 and n=4), followed closely by baySeq. At large sample sizes (n=8), DGEclust, DESeq and baySeq performed similarly, followed closely by DESeq2 and edgeR. After the inclusion of outliers (bottom panel), DGEclust again had the lowest Type I error rate along with DEseq at all sample sizes.
DGEclust retains excellent control of its false discovery rate
Furthermore, we evaluated the ability of DGEclust and the other methods to control the FDR at a prespecified level. For all methods, we calculated the true FDR as the fraction of discoveries that were false at a prespecified significance threshold equal to 0.1. We labelled as discoveries all features satisfying the condition that the adjusted P values (DESeq/DESeq2) or estimated FDRs (DGEclust, baySeq and edgeR) were less than or equal to the aforementioned significance level. If no discoveries were identified, the corresponding true FDR was assumed undefined.
Our results are summarised in Figure 5. In all cases, DGEclust was the top or among the topperforming methods in terms of controlling the FDR at a prespecified level. As a general observation, increasing the number of replicates per condition from 2 to 4 to 8, clearly increased the ability of all methods to control the FDR. DGEclust demonstrated excellent control of its FDR, keeping it close to or below the prespecified significance threshold, in all examined cases. When no outliers and only overdispersed genes were considered (top two panels), DGEclust was the only method that kept the FDR below the prespecified threshold at low sample sizes (n=2). At larger sample sizes (n=4,8), DESeq appears to reduce this gap and in one case (top panel, n=8), it performs as well as DGEclust. In the presence of nonoverdispersed genes (third panel), DGEclust still retains its FDR at (n=4,8) or slightly above (n=2) the prespecified threshold, although it is now marginally overtaken by baySeq (n=2) or DESeq (n=4,8). When genes with unusually low or high counts were considered (bottom panel), DGEclust managed to retain excellent control of its FDR, even at small sample sizes, and with quite significant differences from all alternative methods. DESeq did not call any DE genes in this case and its FDR was deemed undefined.
DGEclust demonstrates top performance for lowreplicated data from a range of species
In addition to simulated data, we also tested our method on RNAseq data from mouse, rat, worm and fly and on CAGE data from humans. The RNAseq data were obtained from the online resource ReCount [46] and they are briefly described below.
The mouse data [47] consisted of two groups of libraries, controls versus CUGBP1 knockdown myoblasts. Each group included two biological replicates (n=2) and, after filtering out all rows with zero count sum, a total of 9,502 genes. Similarly, the rat data [48] consisted of two groups (controls versus subjects with chronic neuropathic pain induced by spinal nerve ligation of the neighbouring L5 spinal nerve) with two biological replicates per group (n=2) and, after filtering as above, 17,125 genes. The filtered (as above) worm data [49] consisted of two groups (L1LIN351cap1 versus L4MALE6cap2) with two replicates each (n=2) and 19,430 genes. Finally, the fly data [50] consisted of two groups, adult males and females, with each group containing three individuals, 1, 5 and 30 days old (n=3). After removing all rows with zero count sum across all samples, the filtered dataset included 13,188 genes. For more details about how these data were obtained, the reader is referred to [46] and the original publications.
The CAGE dataset was prepared according to the standard Illumina protocol described in [51] and it consisted of 25 libraries isolated from five brain regions (caudate nucleus, frontal lobe, hippocampus, putamen and temporal lobe) from five human donors (n=5) and it included 23,448 features, i.e. tag clusters representing promoter regions (see Materials and methods for more details). In this section, we examine only the libraries from the caudate nucleus and frontal lobe, but we include all libraries in our analysis in the next section.
As in [52], we used the logfoldchange between different groups of samples to establish a ground truth on which we based our subsequent method comparisons. Specifically, a gene was considered DE if the absolute log _{2} foldchange ratio of its normalised mean expression across all replicas between two conditions exceeded a threshold value of 2 (corresponding to a 4× change in expression) and nonDE if this ratio was smaller than 1/2 (corresponding to changes in expression less than approximately 41%). All intermediate cases were considered undefined and they were ignored during computation of ROC curves. For CAGE data, we further validated this approach of approximating the true state of differential expression for each gene by adopting the biological homogeneity index (BHI) [53] as an additional proxy of the unknown ground truth (see details in the next section).
The performance of all tested methods was excellent, reaching a TPR larger than 70% at minimal (≪0.01) FPRs (Figure 6). The two topperforming methods were DGEclust and edgeR (worm, fly and human) or baySeq (mouse and rat). In several cases (mouse, worm, fly and human with n=5), DGEclust achieved almost a perfect classification and in all cases it was the only method that consistently demonstrated top performance. It is interesting to observe that while DGEclust is clearly the best method at small sample sizes (n=2), at larger sample sizes (fly with n=3 and human with n=5) edgeR also performed equally well. Based on this evidence, we conclude that DGEclust has a clear advantage compared to alternative methods at low replication levels.
DGEclust is applicable to multigroup expression data
Up to this point, we have restricted our analysis to datasets consisting of two groups of samples. While pairwise comparisons have been characterised as the bread and butter of differential expression analyses, multigroup datasets are also quite common. In this section, we demonstrate the applicability of DGEclust for such datasets by using the CAGE data from human brains as a test case. DGEclust processed the data for 10K iterations, of which the first 5K were rejected as burnin, while the remaining 5K were used for estimating differential expression, as outlined in a previous section. We also applied edgeR and DESeq2 on the same dataset. baySeq was excluded from this analysis, because this would require examining a large number of possible expression patterns, which we found impractical.
Having established a ground truth as outlined in an earlier section, we constructed ROC curves for all possible pairs of brain regions in the CAGE dataset. It may be observed (Figure 7, upper triangle) that, for all pairs of brain regions (with the exception of the caudate/putamen and frontal/temporal pairs), all three methods demonstrated excellent performance, reaching TPRs larger than 80% at FPRs smaller than 1%. In all cases, DGEclust was the top performer as indicated by the AUC, followed by edgeR and then DESeq2 in all cases. Examination of the Venn diagrams constructed from the DE genes, which were identified by each method at an FDR cutoff of 10%, indicates a significant overlap between all three methods, with at least 1K genes commonly identified as DE in most cases. In terms of novel discoveries (i.e. genes that were not called as DE by alternative methods), DGEclust appears to occupy the middle spot between DESeq2 and edgeR, with DESeq2 calling the largest number of novel discoveries in most cases. Two interesting exceptions were the caudate nucleus/putamen pair and the temporal/frontal lobes, for which DGEclust appears to have called the largest number of novel DE genes.
To validate these results further, we clustered the genes identified as DE by each method between at least one pair of brain regions at an FDR of 10% and we tested the validity of the resulting clusterings using the BHI [53] as a quality measure. BHI exploits gene ontology (GO) annotations to provide a measure of how biologically homogeneous a given clustering partition is. Clusters where many genes share annotations will lead to a high BHI score and vice versa. BHI values range from 0 to 1, with a score of 1 indicating the highly unlikely situation of perfect agreement for all GO terms. If the results in Figure 7 are valid, we expect the BHI score computed for DGEclust to be at least as high as the score computed for the other methods.
For clustering the DE genes called by each method, we used the kmeans algorithm with $\sqrt {N_{\textit {DE}}/2}$ clusters as input, where N _{ DE } is the number of DE genes. We could have used more elaborate methods for choosing the optimal number of clusters, such as Akaike’s information criterion or the Bayesian information criterion, but for our purposes, this simple heuristic suffices. We also used hierarchical clustering with an average linkage and a Euclidean distance metric. An optimal clustering partition was subsequently extracted by cutting the resulting hierarchical clustering (visualised as a dendrogram) at a distance of 0.5, which optimises the expectation of Binder’s loss function [54]. In both cases, the raw counts were logtransformed before clustering using DESeq2’s rlog function.
For DGEclust, we additionally performed hierarchical clustering using a similarity matrix computed internally by our software. Specifically, each element $s_{ii^{\prime }}$ of this matrix measuring the similarity between genes i and i ^{′} is defined as follows:
where L is the number of brain regions in the dataset.
Our results from this analysis are indicated in Table 1. As a general observation, the BHI scores are rather low, with only the overall BHI score reaching values of 0.20 to 0.21, while the BHI scores for each individual GO domain range between 0.05 and 0.09. As expected, DGEclust supports BHI scores at least as good as those computed for the other methods (indicated in bold in Table 1). Specifically, for the biological process and cellular component GO domains, DGEclust has the highest score of 0.07 and 0.09, respectively, while for molecular function and overall BHI, the scores (0.08 and 0.21, respectively) are no worse than those returned from the other methods. This indicates that the clusters supported by the DE genes called by DGEclust are at least as biologically homogeneous as those supported by the other methods, thus increasing our confidence in the ability of DGEclust to identify DE genes correctly as indicated in Figure 7.
Furthermore, we investigated the relation between different brain regions by constructing a similarity matrix, which we used as input to hierarchical clustering routines for the generation of dendrograms and heat maps. Every element $s_{ll^{\prime }}$ of this matrix measures the similarity between brain regions l and l ^{′} and it is computed as follows:
where N is the number of genes in the dataset. The similarity matrix calculated as above was used to construct the dendrograms and heat map in Figure 8, after employing a Euclidean distance metric and average linkage. It may be observed that the resulting hierarchical clustering reflects the evolutionary relations between different brain regions. For example, the temporal and frontal lobe samples, which are both located in the cerebral cortex, are clustered together. The hippocampus, albeit lying beneath the cerebral cortex and much older and more primitive than the surrounding neocortex, is not truly a subcortical structure, but rather a cortical infolding. Thus, it is clustered together with the temporal and frontal samples. The subcortical caudate nucleus and putamen, which form the dorsal striatum – an important part of the basal ganglia – are clustered together and they are maximally distant from the cortical structures.
Collectively, the evidence presented in Figure 8 and in Table 1 is supportive of the improved capacity of DGEclust for calling DE genes, as presented in Figure 7.
DGEclust has reasonable computational requirements
Bayesian approaches employing MCMC sampling methodologies for inference are notorious for their increased computational requirements. Being conscious that this might discourage application of this class of methodologies to large genomic datasets, we investigated how our method scales with increasing number of samples and genes. Specifically, we used IPython’s %memit and %timeit commands to measure peak memory usage and computation time. All simulations were performed on a MacBook Pro with an Intel fourcore i7 processor and 8 Gb of memory.
As indicated in Figure 9, computation time and peak memory usage increase linearly with the number of genes (left top and bottom panels). Operations on genes are performed in parallel by default extensively using NumPy’s vector notation and, for this reason, top performance in terms of speed is expected. Groups of samples can also be processed in parallel using multiple cores, if the user wishes to do so. Peak memory usage and computation time also increase linearly with the number of samples (right top and bottom panels), although peak memory usage appears roughly constant over the examined range for the number of samples in the absence of multiprocessing. There is a clear gain in processing speed from using multiple cores to process samples (top right panel). In conclusion, these data suggest that DGEclust is fundamentally applicable even for large genomic datasets, particularly when multiple cores are used.
Conclusions
Despite the availability of several protocols (e.g. single vs pairedend) and sequencing equipment (e.g. Solexa’s Illumina Genome Analyzer, ABI Solid Sequencing by Life Technologies and Roche’s 454 Sequencing), all NGS technologies follow a common set of experimental steps (see [7] for a review) and, eventually, generate data, which essentially constitute a discrete, or digital measure of gene expression. These data are fundamentally different in nature (and, in general terms, superior in quality) from the continuous fluorescence intensity measurements obtained from the application of microarray technologies. In comparison, NGS methods offer several advantages, including detection of a wider level of expression levels and independence of prior knowledge of the biological system, which is required by hybridisationbased microarrays [7]. Due to their better quality, NGS assays have tended to replace microarrays, despite their higher cost [55].
In this article, we have addressed the important issues of clustering and differential expression analysis of digital expression data and we demonstrate the intimate relation between these two concepts. Most proposals for clustering RNAseq and similar types of data have focused on clustering variables (i.e. biological samples), instead of features (e.g. genes) and they employ distancebased or hierarchical clustering methodologies on appropriately transformed datasets, e.g. [24,56,57]. For example, the authors in [24] calculate a common variance function for all samples in a tagseq dataset of glioblastomaderived and noncancerous neural stem cells using a variancestabilising transformation, followed by hierarchical clustering using a Euclidean distance matrix. In [56], a Pearson correlation dissimilarity metric was used for the hierarchical clustering of RNAseq profiles for 14 different tissues from soybean after these were normalised using a variation of the RPKM method [5,6].
The above approaches, although fast and relatively easy to implement, do not always take into account the discrete nature of digital gene expression data. For this reason, various authors have developed distance metrics based on different parameterisations of the loglinear Poisson model for modelling count data, e.g. [5860]. A more recent class of methods follows a modelbased approach, where the digital dataset is modelled as a random sample from a finite mixture of discrete probability distributions, usually Poisson or negative binomial [6163]. Using a full statistical framework for describing the observed count data, these modelbased approaches often perform better than distancebased algorithms, such as kmeans [61].
Although computationally efficient and attractive due to their relative conceptual simplicity, the utility of both distance and finite modelbased clustering methods has been criticised [19,41]. One particular feature of these methodologies, which compromises their applicability, is that the number of clusters in the data must be known a priori. For example, both the kmeans and the selforganising map algorithms require the number of clusters as input. Similarly, methods that model the data as a finite mixture of Poisson or negative binomial distributions [6163] require prior knowledge of the number of mixture components. Estimating the number of clusters usually makes use of an optimality criterion, such as the Bayesian information criterion or the Akaike information criterion, which requires repeated application of the algorithm on the same dataset with different initial choices of the number of clusters. Thus, the number of clusters and the parameters for each individual cluster are estimated separately, making the algorithm sensitive to the initial model choice. Similarly, hierarchical clustering methods often rely on some arbitrary distance metric (e.g. Euclidean or Pearson correlation) to distinguish between members of different clusters, without providing a criterion for choosing the correct number of clusters or a measure of the uncertainty of a particular clustering, which would serve to assess its quality.
In this article, we have developed a statistical methodology and associated software (DGEclust) for clustering digital gene expression data, which (unlike previously published approaches [5660]) does not require any prior knowledge of the number of clusters, rather it estimates this parameter and its uncertainty simultaneously with the parameters (e.g. location and shape) of each individual cluster. This is achieved by embedding the negative binomial distribution for modelling count data in a hierarchical Dirichlet process mixture framework. Our formulation implies that distributional parameters (i.e. foldchanges) are not all distinct, but they are shared between genes and between groups of samples. This is a form of information sharing between genes and between samples, which is made possible by the particular hierarchical structure of the proposed model. At each level of the hierarchy, the number of mixture components, i.e. the number of clusters, is assumed infinite. This represents a substantial departure from previously proposed finite mixture models and avoids the need for arbitrary prior choices regarding the number of clusters in the data.
Despite the infinite dimension of the mixture model, only the finite number of clusters supported by the data and the associated parameters are estimated. This is achieved by introducing a blocked Gibbs sampler, which permits efficient processing of large datasets containing more than 20K genes. Unlike MCMC inference methods for HDPMM based on the popular Chinese restaurant franchise metaphor [16,41], our algorithm allows all genespecific parameters in each sample to be updated simultaneously and independently from other samples. This allows rapid convergence of the algorithm and permits the development of parallelised implementations of the Gibbs sampler, which enjoy the increased performance offered by modern multicore processors.
The second important contribution of this study is demonstrating how this type of hierarchical clustering can be used for differential expression analysis. We emphasise that differential expression can be thought of as a particular form of clustering. Through comparison with popular alternatives for both simulated and actual experimental data, we demonstrate the applicability of this approach for a wide range of experimental settings and its improved performance, particularly at small sample sizes, which reflects the design of current sequencing experiments.
In conclusion, we have developed a hierarchical, nonparametric Bayesian method for modelling digital expression data. The novelty of our method is simultaneously addressing the problems of model selection and estimation uncertainty and exposing the intimate relation between clustering and differential expression analysis. We expect our work to inspire and support further theoretical research on modelling digital expression data and we believe that our software, DGEclust, will prove to be a useful addition to the existing tools for the statistical analysis of RNAseq and similar types of data.
Materials and methods
We implemented the methodology presented in this article in the software package DGEclust, which is written in Python and uses the SciPy stack. DGEclust expects as input and clusters a matrix of unnormalised count data along with replication information, if this is available. The output of the clustering process is required as input to postprocessing routines, which compute the posterior probabilities that each particular gene between any two groups of samples in the data are DE. Further postprocessing routines can generate gene or librarywise similarity matrices, which can be used as input to hierarchical clustering routines for the generation of heat maps and dendrograms. DGEclust takes advantage of multicore processors to accelerate computations. All analyses in this article were performed using DGEclust and standard Python/SciPy tools, as well as DESeq/DESeq2, edgeR and baySeq for comparison purposes. When using these packages, all parameters were left at their default values. For baySeq, we used 5K samples when estimating the priors using the quasilikelihood approach. To make a comparison possible, we processed all datasets with DGEclust for 10K iterations, rejecting the first 5K of them as burnin.
Normalisation
Internally, DGEclust uses the same normalisation method as DESeq. This behaviour can be overridden by providing a set of library sizes as input. When comparing different software packages, we used the default normalisation method of each package.
CAGE library preparation and data preprocessing
Postmortem human brain tissue from frontal, temporal, hippocampus, caudate and putamen regions from five donors was obtained from the Netherlands Brain Bank (NBB, Amsterdam, Netherlands). All tissue requests received at the NBB are reviewed by the NBB’s scientific committee and all materials and data collected are obtained with written informed consent. The procedures, information and consent forms of the NBB have been approved by the Medical Ethics Committee of the VU Medical Centre (Amsterdam, Netherlands).
Total RNA was extracted and purified using the Trizol tissue kit according to the manufacturer’s instructions (Invitrogen, Waltham, Massachusetts, USA). CAGE libraries were prepared according to the standard Illumina CAGE protocol [51]. Briefly, 5 µg of total RNA was reverse transcribed with reverse transcriptase. Samples were captrapped and a specific linker, containing a 3bp recognition site and the type III restrictionmodification enzyme EcoP15I, was ligated to the singlestrand cDNA. The priming of the second strand was done with specific primers. After synthesis of the second strand and cleavage with EcoP15I, another linker was ligated. Purified cDNA was then amplified with 10 to 12 PCR cycles. PCR products were purified, their concentration was adjusted to 10 nM and they were sequenced on a HiSeq 2000 using the standard protocol for 50bp singleend runs.
Sequenced reads (tags) were filtered for known CAGE artefacts using TagDust [64]. Low quality reads and reads mapping to known rRNA were also removed. The remaining reads were mapped to the human genome (build hg19) using the Burrows–Wheeler aligner for short reads [65]. Mapped reads overlapping or located within 20 bp on the same strand were grouped into tag clusters and tag clusters with low read counts were removed.
URL
The most recent version of DGEclust is available online under the MIT licence [66].
Abbreviations
 AUC:

area under the receiver operating characteristic curve
 BHI:

biological homogeneity index
 bp:

base pair
 CAGE:

cap analysis of gene expression
 DE:

differentially expressed
 FD:

false discovery
 FDR:

false discovery rate
 FPR:

false positive rate
 GO:

gene ontology
 HDP:

hierarchical Dirichlet process
 HDPMM:

hierarchical Dirichlet process mixture model
 MCMC:

Markov chain Monte Carlo
 NBB:

Netherlands Brain Bank
 NGS:

nextgeneration sequencing
 RNAseq:

RNA sequencing
 ROC:

receiver operating characteristic
 SAGE:

serial analysis of gene expression
 TPR:

true positive rate
References
 1
Wang Z, Gerstein M, Snyder M. RNASeq: a revolutionary tool for transcriptomics. Nat Rev Genet. 2009; 10:57–63.
 2
Shiraki T, Kondo S, Katayama S, Waki K, Kasukawa T, Kawaji H, et al. Cap analysis gene expression for highthroughput analysis of transcriptional starting point and identification of promoter usage. Proc Natl Acad Sci USA. 2003; 100(1577):6–81.
 3
Velculescu VE, Zhang L, Vogelstein B, Kinzler KW. Serial analysis of gene expression. Science. 1995; 270:484–7.
 4
Auer PL, Doerge RW. Statistical design and analysis of RNA sequencing data. Genetics. 2010; 185:405–16.
 5
Sun Z, Zhu Y. Systematic comparison of RNASeq normalization methods using measurement error models. Bioinformatics. 2012; 28:2584–91.
 6
Dillies MA, Rau A, Aubert J, HennequetAntier C, Jeanmougin M, Servant N, et al. A comprehensive evaluation of normalization methods for Illumina highthroughput RNA sequencing data analysis. Brief Bioinform. 2013; 14:671–83.
 7
Oshlack A, Robinson MD, Young MD. From RNAseq reads to differential expression results. Genome Biol. 2010; 11:220.
 8
Kvam VM, Liu P, Si Y. A comparison of statistical methods for detecting differentially expressed genes from RNAseq data. Am J Bot. 2012; 99:248–56.
 9
Soneson C, Delorenzi M. A comparison of methods for differential expression analysis of RNAseq data. BMC Bioinf. 2013; 14:91.
 10
Cho RJ, Campbell MJ, Winzeler EA, Steinmetz L, Conway A, Wodicka L, et al. A genomewide transcriptional analysis of the mitotic cell cycle. Mol Cell. 1998; 2:65–73.
 11
Eisen MB, Spellman PT, Brown PO, Botstein D. Cluster analysis and display of genomewide expression patterns. Proc Natl Acad Sci USA. 1998; 95(1486):3–8.
 12
Shannon W, Culverhouse R, Duncan J. Analyzing microarray data using cluster analysis. Pharmacogenomics. 2003; 4:41–52.
 13
Jiang D, Tang C, Zhang A. Cluster analysis for gene expression data: a survey. IEEE Trans Knowl Data Eng. 2004; 16:1370–86.
 14
Yeung K, Medvedovic M, Bumgarner R. From coexpression to coregulation: how many microarray experiments do we need?Genome Biol. 2004; 5:R48.
 15
de Souto MCP, Costa IG, de Araujo DSA, Ludermir TB, Schliep A. Clustering cancer gene expression data: a comparative study. BMC Bioinf. 2008; 9:497.
 16
Teh YW, Jordan MI, Beal MJ, Blei DM. Hierarchical Dirichlet processes. J Am Stat Assoc. 2006; 101:1566–81.
 17
Green PJ. Colouring and breaking sticks: random distributions and heterogeneous clustering In: Bingham N, Goldie C, editors. Probability and mathematical genetics: papers in honour of Sir John Kingman. Cambridge, UK: Cambridge University Press: 2010.
 18
Medvedovic M, Sivaganesan S. Bayesian infinite mixture model based clustering of gene expression profiles. Bioinformatics. 2002; 18:1194–206.
 19
Rasmussen CE, de la Cruz BJ, Ghahramani Z, Wild DL. Modeling and visualizing uncertainty in gene expression clusters using Dirichlet process mixtures. IEEE/ACM Trans Comput Biol Bioinf. 2009; 6:615–28.
 20
Sohn KA, Xing EP. A hierarchical Dirichlet process mixture model for haplotype reconstruction from multipopulation data. Ann Appl Stat. 2009; 3:791–821.
 21
Savage RS, Ghahramani Z, Griffin JE, de la Cruz BJ, Wild DL. Discovering transcriptional modules by Bayesian data integration. Bioinformatics. 2010; 26:i158–67.
 22
Kirk P, Griffin JE, Savage RS, Ghahramani Z, Wild DL. Bayesian correlated clustering to integrate multiple datasets. Bioinformatics. 2012; 28:3290–7.
 23
Wang C, Paisley J, Blei DM. Online variational inference for the hierarchical Dirichlet process In: Gordon G, Dunson D, Dudík M, editors. Proceedings of the 14th, International Conference on Artificial Intelligence and Statistics. vol. 15. JMLR W&CP: 2011.
 24
Anders S, Huber W. Differential expression analysis for sequence count data. Genome Biol. 2010; 11:R106.
 25
Hardcastle TJ, Kelly KA. BaySeq: empirical Bayesian methods for identifying differential expression in sequence count data. BMC Bioinf. 2010; 11:422.
 26
McCarthy DJ, Chen Y, Smyth GK. Differential expression analysis of multifactor RNASeq experiments with respect to biological variation. Nucleic Acids Res. 2012; 40:4288–97.
 27
Love MI, Huber W, Anders S. Moderated estimation of fold change and dispersion for RNAseq data with DESeq2. Genome Biol. 2014; 15:550.
 28
Vavoulis DV, Gough J. Nonparametric Bayesian modelling of digital gene expression data. J Comput Sci Syst Biol. 2013; 7:1–9.
 29
Lu J, Tomfohr JK, Kepler TB. Identifying differential expression in multiple SAGE libraries: an overdispersed loglinear model approach. BMC Bioinf. 2005; 6:165.
 30
Robinson MD, Smyth GK. Moderated statistical tests for assessing differences in tag abundance. Bioinformatics. 2007; 23:2881–7.
 31
Robinson MD, Smyth GK. Smallsample estimation of negative binomial dispersion, with applications to SAGE data. Biostatistics. 2008; 9:321–32.
 32
Nagalakshmi U, Wang Z, Waern K, Shou C, Raha D, Gerstein M, et al. The transcriptional landscape of the yeast genome defined by RNA sequencing. Science. 2008; 320:1344–9.
 33
Auer PL, Doerge RW. A twostage Poisson model for testing RNAseq data. Stat Appl Genet Mol Biol. 2011; 10:26.
 34
Srivastava S, Chen L. A twoparameter generalized Poisson model to improve the analysis of RNAseq data. Nucleic Acids Res. 2010; 38:e170.
 35
Wang L, Feng Z, Wang X, Wang X, Zhang X. DEGseq: an R package for identifying differentially expressed genes from RNAseq data. Bioinformatics. 2010; 26:136–8.
 36
Langmead B, Hansen KD, Leek JT. Cloudscale RNAsequencing differential expression analysis with Myrna. Genome Biol. 2010; 11:R83.
 37
Dillies MA, Rau A, Aubert J, HennequetAntier C, Jeanmougin M, Servant N, et al. French statOmique consortium: a comprehensive evaluation of normalization methods for Illumina highthroughput RNA sequencing data analysis. Brief Bioinform. 2013; 14:671–83.
 38
Sethuraman J. A constructive definition of Dirichlet priors. Statistica Sinica. 1994; 4:639–50.
 39
Wu H, Wang C, Wu Z. A new shrinkage estimator for dispersion improves differential expression detection in RNAseq data. Biostatistics. 2013; 14:232–43.
 40
Ishwaran H, James LF. Gibbs sampling methods for stickbreaking priors. J Am Stat Assoc. 2001; 96:161–73. http://www.jstor.org/stable/2670356.
 41
Wang L, Wang X. Hierarchical Dirichlet process model for gene expression clustering. EURASIP J Bioinform Syst Biol. 2013; 2013:5.
 42
Newton MA, Noueiry A, Sarkar D, Ahlquist P. Detecting differential gene expression with a semiparametric hierarchical mixture method. Biostatistics. 2004; 5:155–76.
 43
Soneson C. CompcodeR: RNAseq data simulation, differential expression analysis and performance comparison of differential expression methods. R package version 1.0.0.
 44
Pickrell JK, Marioni JC, Pai AA, Degner JF, Engelhardt BE, Nkadori E, et al. Understanding mechanisms underlying human gene expression variation with RNA sequencing. Nature. 2010; 464:768–72.
 45
Cheung VG, Nayak RR, Wang I X, Elwyn S, Cousins SM, Morley M, et al. Polymorphic cis and transregulation of human gene expression. PLoS Biol. 2010; 8:e1000480. doi:10.1371/journal.pbio.1000480.
 46
Frazee A, Langmead B, Leek J. ReCount: A multiexperiment resource of analysisready RNAseq gene count datasets. BMC Bioinf. 2011; 12:449. doi:10.1186/1471210512449.
 47
Katz Y, Wang ET, Airoldi EM, Burge CB. Analysis and design of RNA sequencing experiments for identifying isoform regulation. Nat Methods. 2010; 7:1009–15.
 48
Hammer P, Banck MS, Amberg R, Wang C, Petznick G, Luo S, et al. mRNAseq with agnostic splice site discovery for nervous system transcriptomics tested in chronic pain. Genome Res. 2010; 20:847–60.
 49
Hillier LW, Reinke V, Green P, Hirst M, Marra MA, Waterston RH, et al. Massively parallel sequencing of the polyadenylated transcriptome of C. elegans. Genome Res. 2009; 19:657–66.
 50
Graveley B, Brooks A, Carlson JEA. The developmental transcriptome of Drosophila melanogaster. Nature. 2011; 471:473–9.
 51
Takahashi H, Lassmann T, Murata M, Carninci P. 5^{′} endcentered expression profiling using capanalysis gene expression and nextgeneration sequencing. Nat Protoc. 2012; 7:542–61.
 52
Bullard JH, Purdom E, Hansen KD, Dudoit S. Evaluation of statistical methods for normalization and differential expression in mRNASeq experiments. BMC Bioinf. 2010; 11:94.
 53
Datta S, Datta S. Methods for evaluating clustering algorithms for gene expression data using a reference set of functional classes. BMC Bioinf. 2006; 7:397. doi:10.1186/147121057397.
 54
Fritsch A, Ickstadt K. Improved criteria for clustering based on the posterior similarity matrix. Bayesian Anal. 2009; 4:367–92.
 55
Carninci P. Is sequencing enlightenment ending the dark age of the transcriptome?Nat Methods. 2009; 6:711–13.
 56
Severin AJ, Woody JL, Bolon YT, Joseph B, Diers BW, Farmer AD, et al. RNASeq atlas of glycine max: a guide to the soybean transcriptome. BMC Plant Biol. 2010; 10:160.
 57
Li P, Ponnala L, Gandotra N, Wang L, Si Y, Tausta SL, et al. The developmental dynamics of the maize leaf transcriptome. Nat Genet. 2010; 42:1060–7.
 58
Cai L, Huang H, Blackshaw S, Liu J, Cepko C, Wong W, et al. Clustering analysis of SAGE data using a Poisson approach. Genome Biol. 2004; 5:R51.
 59
Kim K, Zhang S, Jiang K, Cai L, Lee IB, Feldman LJ, et al. Measuring similarities between gene expression profiles through new data transformations. BMC Bioinf. 2007; 8:29.
 60
Witten DM. Classification and clustering of sequencing data using a Poisson model. Ann Appl Stat. 2011; 5:2493–518.
 61
Si Y, Liu P, Li P, Brutnell T. Modelbased clustering of RNAseq data 2011. In: JSM Proceedings, Biometrics Section. Miami Beach, Florida: American Statistical Association. p. 1121–9.
 62
Rau A, Celeux G, MartinMagniette ML, MaugisRabusseau C. Clustering highthroughput sequencing data with Poisson mixture models. 2011. Research Report: RR7786, p. 36. https://hal.inria.fr/inria00638082.
 63
Wang N, Wang Y, Hao H, Wang L, Wang Z, Wang J, et al. A biPoisson model for clustering gene expression profiles by RNAseq. Brief Bioinform. 2014; 15:534–41.
 64
Lassmann T, Hayashizaki Y, Daub CO. TagDust – a program to eliminate artifacts from next generation sequencing data. Bioinformatics. 2009; 25:2839–40.
 65
Li H, Durbin R. Fast and accurate short read alignment with Burrows–Wheeler transform. Bioinformatics. 2009; 25:1754–60.
 66
DGEclust: Hierarchical nonparametric Bayesian clustering of digital expression data. http://dvav.github.io/dgeclust/.
Acknowledgements
DVV would like to thank Peter Green, Mark Beaumont and Colin Campbell from the University of Bristol for useful discussions on Dirichlet process mixture models and multiple hypothesis testing. The authors would also like to thank two anonymous reviewers for their constructive criticism. This work was supported by grants EPSRC EP/H032436/1 and BBSRC G022771/1.
Author information
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors’ contributions
DVV and JG developed the method. MF and PH provided the CAGE data. DVV implemented the method. DVV and MF performed the analyses. DVV and JG wrote the article with contributions from all authors. All authors read and approved the final manuscript.
Additional file
Additional file 1
Supplementary Material for DGEclust: differential expression analysis of clustered count data. In this supplementary material we provide a detailed account of posterior inference in the model summarised by Equations 5.
Rights and permissions
About this article
Cite this article
Vavoulis, D.V., Francescatto, M., Heutink, P. et al. DGEclust: differential expression analysis of clustered count data. Genome Biol 16, 39 (2015) doi:10.1186/s1305901506046
Received
Accepted
Published
DOI
Keywords
 False Discovery Rate
 Differentially Express
 Differentially Express Gene
 Negative Binomial Distribution
 Differential Expression Analysis