 Software
 Open Access
 Published:
HiFive: a tool suite for easy and efficient HiC and 5C data analysis
Genome Biology volume 16, Article number: 237 (2015)
Abstract
The chromatin interaction assays 5C and HiC have advanced our understanding of genomic spatial organization, but analysis approaches for these data are limited by usability and flexibility. The HiFive tool suite provides efficient data handling and a variety of normalization approaches for easy, fast analysis and method comparison. Integration of MPIbased parallelization allows scalability and rapid processing time. In addition to singlecommand analysis of an entire experiment from mapped reads to interaction values, HiFive has been integrated into the opensource, webbased platform Galaxy to connect users with computational resources and a graphical interface. HiFive is opensource software available from http://taylorlab.org/software/hifive/.
Background
In the more than a decade since the vast majority of the human genome was first sequenced, it has become clear that sequence alone is insufficient to explain the complex gene and RNA regulatory patterns seen over time and across cell types in eukaryotes. The context of specific sequences – whether from combinations of DNAbinding transcription factors (TFs) [1–3], methylation of the DNA itself [4, 5], or local histone modifications [4, 6] – is integral to how the cell utilizes each sequence element. Although we have known about the potential roles that sequentially distant but spatially proximal sequences and their binding and epigenetic contexts play in regulating expression and function, it has only been over the past decade that new sequencingbased techniques have enabled highthroughput analysis of higherorder structures of chromatin and investigation into how these structures interact among themselves and with other genomic elements to influence cellular function.
Several different sequencing methods for assessing chromatin interactions have been devised, all based on preferentially ligating spatially close DNA sequence fragments. These approaches include ChIAPet [7], tethered chromosome capture [8], and the chromatin conformation capture technologies of 3C, 4C, 5C, and HiC [9–12] (Additional file 1: Figure S1). While these assays have allowed a rapid expansion of our understanding of the nature of genome structure, they also have presented some formidable challenges.
In both HiC and 5C, systematic biases resulting from the nature of the assays have been observed [13, 14], resulting in differential representation of sequences in the resulting datasets. While analyses at a larger scale are not dramatically affected by these biases due to the large number of data points being averaged over, higherresolution approaches must first address these challenges. This is becoming more important as the resolution of experiments is increasing [15]. Several analysis methods have been described in the literature and applied to correcting biases in HiC [14–21] and 5C data [22–24]. There is still room for improving our ability to remove this systematic noise from the data and resolve finerscale features and, perhaps more importantly, for improving the usability and reproducibility of normalization methodologies.
A second challenge posed by data from these types of assays is one of resources. Unlike other nextgeneration sequencing assays where even singlebase resolution is limited to a few billion data points, these assays assess pairwise combinations, potentially increasing the size of the dataset by several orders of magnitude. For a three billion base pair genome cut with a sixbase restriction enzyme (RE), the number of potential interaction pairs is more than half a trillion (if considering both fragment ends) while a fourbase RE can yield more than two and a half quadrillion interaction pairs. Even allowing that the vast majority of those interactions will be absent from the sequencing data, the amount of information that needs to be handled and the complexity of normalizing these data still pose a major computational hurdle, especially for investigators without access to substantial computational resources.
Here we describe HiFive, a suite of tools developed for handling both HiC and 5C data using a combination of empirically determined and probabilistic signal modeling. HiFive has performance on par or better than other available methodologies while showing superior speed and efficient memory usage through parallelization and data management strategies. In addition to providing a simple interface with no preprocessing or reformatting requirements, HiFive offers a variety of normalization approaches including versions of all commonly used algorithmic approaches allowing for straightforward optimization and method comparison within a single application. In addition to its command line interface, HiFive is also available through Galaxy, an opensource webbased platform, connecting users with computational resources and the ability to store and share their 5C and HiC analyses. All of these aspects of HiFive make it simple to use and fast, and make its analyses easily reproducible.
The HiFive analysis suite
HiFive was designed with three goals: first, to provide a simpletouse interface for flexible chromatin interaction data analysis; second, to provide welldocumented support for 5C analysis; and third, to improve performance over existing methodologies while reducing analysis runtimes. These are accomplished through a standalone program built on a Python library designed for customizable analysis and supported under the webbased platform Galaxy.
User interface
HiFive provides three methods of use: the command line; the Internet; or as a development library. The command line interface provides users with the ability to perform analyses as a series of steps or as a single unified analysis. The only inputs that HiFive requires are a description of the genomic partitioning and interaction data, either directly as mapped reads or counts of reads associated with the partitioned genome (for example, fragment pairs and their observed reads). HiFive handles all other formatting and data processing. In addition, HiFive has been bundled as a set of tools available through Galaxy (Fig. 1). This not only provides support with computational resources but also ensures simple installation of all prerequisite libraries and packages. HiFive was also created to allow custom creation of analysis methods as a development library for chromatin interaction analysis through extensive documentation and an efficient datahandling framework.
Organization of HiFive
At its core, HiFive is a series of hierarchical data structures building from general to specific information. There are four primary file types that HiFive creates, all relying on the Numpy scientific computing Python package for efficient data arrays and fast data access. These structures include genomic partitioning, observed interactions, distancedependence relationship and normalization parameters, and heatmaps of observed and expected interactions. By separating these attributes, many datasets can utilize the same genomic partitioning and multiple analyses can be run using the same interaction data without the need to reload or process information.
Data processing and filtering
In order to process 5C or HiC data, the first step after mapping is converting the data into a form compatible with the desired workflow. HiFive appears to be nearly alone in its ability to handle mapped data without additional processing (the only exception is HiCLib [17]). Reads can be read directly from numerous BAMformatted files, and this may be done as an independent step or within the integrated onestep analysis. HiCLib also possesses the ability to input data directly from mapped read files. In all other cases, reads need to be converted to other formats. HiCPipe [14] provides scripts for some but not all of these processes, while HiCNorm [16] relies on prebinned reads. In all cases aside from HiFive, a workflow is required to move from mapped reads to normalization.
Filtering is accomplished in two phases, during the initial processing of reads and during project creation (Additional file 1: Figures S2 and S3). The first phase limits data to acceptable pairedend combinations. For 5C data, this means reads mapping to fragments probed with oppositeorientation primers. HiC data use two criteria, total insert size (a userspecified parameter) and orientation/fragment relationship filtering. In the latter case, reads originating from nonadjacent fragments or from adjacent fragments and in the same orientation are allowed, similar to Jin et al. [19] (Additional file 1: Figure S4). The second phase, common to both 5C and HiC data, is an iterative filtering based on numbers of interactions per fragment or fragment end (fend). Briefly, total numbers of interactions for each fragment are calculated, and fragments with insufficient numbers of interaction partners are removed along with all of their interactions. This is repeated until all fragments interact with a sufficient number of other nonfiltered fragments. This filtering is crucial for any fragment or fendspecific normalization scheme to ensure sufficient interdependency between interaction subsets to avoid convergence issues.
Distancedependence signal estimation
One feature of HiFive that is notably absent from nearly all other available analysis software is the ability to incorporate the effects of sequential distance into the normalization. One exception to this is HiTC [21], which uses a loess regression to approximate the distancedependence relationship of 5C data to genomic distance. This method does not, however, allow for any other normalization of 5C data. Another is FitHiC [25], although this software assigns confidence estimates to midrange contact bins rather than normalizing entire datasets. This feature is of particular importance for analysis of shortrange interactions such as this in 5C data, or for making use of counts data rather than a binary observed/unobserved indicator. For 5C data, HiFive uses a linear regression to estimate parameters for the relationship between the logdistance and logcounts (Additional file 1: Figure S5). HiC data require a more nuanced approximation because of the amount of data involved and the nonlinear relationship over the range of distances queried. To achieve this, HiFive uses a linear piecewise function to approximate the distancedependent portion of the HiC signal, similar but distinct from that used by FitHiC. HiFive partitions the total range of interactions into equally sized logtransformed distance bins with the exception of the smallest bin, whose upper bound is specified by the user. Mean counts and logtransformed distances are calculated for each bin and a line is used to connect each set of adjacent bin points (Additional file 1: Figure S6). For distances extending past the first and last bins, the line segment is simply extended from the last pair of bins on either end. Simultaneously, a similar distancedependence function is constructed using a binary indicator of observed/unobserved instead of read counts for each fend pair. All distances are measured between fragment or fend midpoints.
HiFive normalization algorithms
HiFive offers three different normalization approaches. These include a combinatorial probability model based on HiCPipe’s algorithm called ‘Binning’, a modified matrixbalancing approach called ‘Express’, and a multiplicative probability model called ‘Probability’. In the Binning algorithm, learning is accomplished in an iterative fashion by maximizing each set of characteristic bin combinations independently each round using the Broyden–Fletcher–Goldfarb–Shanno algorithm for maximum likelihood estimation.
The Express algorithm is a generalized version of matrix balancing. While it can use the KnightRuiz algorithm [26] for extremely fast standard matrix balancing (ExpressKR), the Express algorithm also has the ability to take into account differing numbers of possible interactions and find corrections weighted by these numbers of interactions. The set of valid interactions is defined as set A, interactions whose fends have both passed the abovedescribed filtering process and whose interfend distance falls within userspecified limits. In addition, because counts are logtransformed for 5C normalization, only nonzero interactions are included in set A. For each interaction c between fends or fragments i and j for HiC and 5C, respectively, in the set of valid interactions A, correction parameter f _{ i } is updated as in (1) for HiC and (2) for 5C.
The expected value of each HiC interaction is simply product of the exponent of the expected distancedependent signal D(i,j) and the fend corrections (3).
5C interactions have expected values that correspond to the logtransformed count and are the sum of each signal component (4).
By scaling the row sums based on number of interactions, the weighted matrix balancing allows exclusion of interactions based on interaction size criteria without skewing correction factors due to nonuniform restriction site distribution, position along the chromosome, or filtered fragments or fends due to read coverage. Because it can incorporate the distancedependent signal, the Express algorithm can operate on counts data unlike most other matrix balancing approaches, although it also can be performed on binary data (observed vs. unobserved) or logtransformed counts for HiC and 5C, respectively. This algorithm allows for adjustment of counts based on the estimated distancedependence signal prior to normalization in both weighted (1 and 2) and unweighted (KnightRuiz) versions.
The multiplicative Probability algorithm models the data assuming some probability distribution with a prior equal to the estimated distancedependent signal. HiC data can be modeled either with a Poisson or binomial distribution (Additional file 1: Figure S7). In the case of the binomial distribution, counts are transformed into a binary indicator of observed/unobserved and the distancedependence approximation function is based on this same binary data. 5C data are modeled using a lognormal distribution. In both cases only counts in the set of reads A (described above) are modeled.
For both the Express and Probability algorithms, a backtrackingline gradient descent approach is used for learning correction parameters. This allows the learning rate r to be updated each iteration t to satisfy the Armijo criteria (5) based on the cost C, ensuring that parameter updates are appropriate.
Filtering interactions by interaction size
Chromatin topology is organized around highly reproducible regions of frequent local interactions termed ‘topological domains’ [27]. Within these structures it has been observed that specific features can influence the frequency of interactions in a biased and differential way up and downstream of them, such as transcript start sites (TSS) and CTCFbound sites [14]. In order to account for systematic noise and bias without confounding normalization efforts with meaningful biologicalrelevant structures, HiFive allows filtering out of interactions using interaction size cutoffs. In order to assess the effects of filtering out shortersized interactions, we analyzed data both with and without a lower interaction distance cutoff. For HiC data we analyzed two mouse embyronic stem cells (ESC) datasets with no lower limit and with a lower distance limit of 500 Kb using each of the described normalization algorithms. This size was chosen to eliminate all but the weakest interaction effects observed for TSSs and CTCFbound sites [28]. HiC normalization performance was assessed using the interdataset correlations. For 5C data, there is a much smaller range of interactions. In order to handle this, we set a lower interaction size cutoff of 50 Kb. 5C normalization performance was assessed as the correlation between 5C data and HiC data of the same cell type [27] and binned based on probed 5C fragments to create identically partitioned sets of interactions and normalized using HiFive’s Probability algorithm.
The differences in HiC algorithm performances with and without the lower interaction size cutoff were varied, although the largest effects were seen when data were binned in 10 and 50 Kb bins for intrachromosomal interactions and for overall interchromosomal interactions (Additional file 1: Figure S8). Overall, excluding shortrange interactions made little difference for the Express algorithm but did improve the performance of the Probability and Binning algorithms. The 5C algorithms showed an opposite result, with almost universal decrease in performance when shortrange interactions are excluded (Additional file 1: Figure S9). As a result, learning HiC normalization parameters using HiFive algorithms was performed excluding interactions shorter than 500 Kb and 5C analyses were performed using all interaction sizes. All analyses subsequent to normalization (for example, dataset correlations) were performed across all interactions.
Analyzing 5C data
To date, limited work has focused on processing of 5C data to remove technical biases [22–24, 29]. Of that, none has been formalized in published analysis software. In order to assess HiFive’s performance in normalizing 5C data, we used two different 5C mouse ESC datasets [23, 24] and found correlations to HiC data of the same cell type [27] and binned based on probed 5C fragments to create identically partitioned sets of interactions (Fig. 2, Additional file 1: Figures S9 and S10). HiC interactions were normalized using either HiFive’s probability algorithm (Fig. 2) or HiCPipe (Additional file 1: Figure S10) and heatmaps were dynamically binned to account for sparse coverage (see Additional file 1: Methods: 5CHiC data correlations). Correlations were found between all nonzero pairs of bins (fragment level resolution) following logtransformation. All of HiFive’s 5C algorithms showed an improved correlation with HiC data compared to raw data, regardless of HiC normalization approach. The Binning algorithm showed the least improvement, likely due to the limits on the number of bins into which features could be partitioned and characteristics missing from the model, such as primer melting temperature. The standard matrixbalancing approach (ExpressKR) showed good improvement, although not quite as good as the Express and Probability algorithms. All of these normalizations were accomplished in less than one minute proceeding from a BED file and a counts file to heatmaps.
HiC analysis software comparison
Several algorithms have been proposed to handle interaction data normalization (Table 1). These analysis approaches can be divided into two varieties, probabilistic and matrix balancing. The probabilistic approach is further divided into combinatorial and multiplicative corrections. The combinatorial probability model is implemented in HiCPipe [14] and remains one of the most popular approaches. This approach uses one or more restriction fend characteristics partitioned into ranges of values and iteratively learns correction values for each combination of ranges based on a binomial distribution of observed versus unobserved fend interactions. A multiplicative modeling approach is used in the analysis software HiCNorm [16]. HiCNorm uses a Poisson regression model using binned counts instead of binary output and assuming that biases from fend characteristics are multiplicative between bin combinations. A different multiplicative approach is matrix balancing, which finds a value for each row/column of a symmetric matrix (or in this case, heatmap) such that after multiplication of each matrix value by its associated row and column values, the sum of each row and column is one. This has been described with at least four different implementations in the literature [15, 17, 20, 30] although only two software packages making use of it have been published (HiCLib [17], now included in the R package HiTC [21] and HiCorrector [20]). For this paper, we chose to use our own implementation of the algorithm described by Knight and Ruiz [26] for comparison due to speed and ease of use considerations.
Method performances
To assess HiC analysis method performances we used two different pairs of HiC datasets [15, 27, 31], finding interaction correlations across different restriction digests of the same cell type genomes. The Dixon et al. data [27] were produced using mouse ESCs digested with either HindIII or NcoI, yielding approximately 4 Kb fragments. The Selvaraj et al. data [31] were produced from human GM12878 cells using HindIII, while the Rao et al. data [15] were produced from human GM12878 cells using the 4 bp restriction enzyme MboI, producing approximately 250 bp fragments. This allowed assessment of method performance and data handling across a range of experimental resolutions. Correlations were calculated for 10 mutually exclusive intrachromosomal (cis) interaction ranges and across all cis interactions simultaneously for four binning resolutions. Correlations were also calculated for interchromosomal interactions for two resolutions.
HiC analysis methods showed varied performances across interaction size ranges, resolutions, and datasets for intrachromosomal interactions (Fig. 3a and b). For small interaction sizes, HiFive’s Probability and Express algorithms performed consistently well regardless of resolution. At longer interaction distances the Express algorithm typically outperformed the Probability algorithm. HiCNorm showed a nearly opposite performance with poorer interdataset correlations for shorterrange interactions but higher correlations at longer ranges, relative to other methods. HiCPipe’s performance appeared to depend on binning resolution. At higher resolutions (≤50 Kb), HiCPipe performed worse than the majority of methods. However at lower resolutions it tended to outperform other methods, regardless of interaction size range. HiFive’s Binning algorithm had a more consistent performance around the middle of all of the methods across all binning resolutions, with the exception of the 1 Mb resolution for the human data where it performed the worst. Standard matrix balancing consistently performed at or near the bottom of the group regardless of the interaction size range or resolution.
Correlations across all intrachromosomal interactions showed much more consistency between analysis methodologies (Fig. 3c and d). This is primarily due to the fact that the main driver of interdataset correlation, the interaction distancecounts relationship, was present in all of the analyzed data. HiFive’s Probability and Express algorithms were again top performers across almost every intrachromosomal comparison, although the Probability algorithm showed a decreasing advantage with decreasing binning resolution. HiCNorm, HiCPipe, matrix balancing, and HiFive’s Binning algorithm were highly consistent in terms of performance for the mouse datasets. For the human interdataset correlations HiCPipe and matrix balancing showed a slightly better performance than average while HiCNorm faired worse. HiFive’s Express algorithm was still the top performer.
Interchromosomal datasets showed a wider range of performances and were strongly dependent on which datasets were being analyzed (Fig. 3c and d). For mouse interchromosomal interactions, HiFive’s Probability and Express algorithms performed much better than other methods at the 250 Kb binning resolution, but consistent with other methods at the 1 Mb resolution. HiCNorm showed worse performance at both bin sizes for the mouse datasets. HiCPipe showed the best performance at the 1 Mb resolution, slightly above other methods, but the second worst performance at the 250 Kb resolution. Results for the human datasets were more consistent across resolutions. HiCNorm, HiFive’s Express algorithm, and matrix balancing performed best in both cases with Express doing slightly better at the 250 Kb resolution and HiCNorm at the 1 Mb resolution. The remaining methods showed similar performance to each other, although HiFive’s Probability algorithm performed slightly worse than HiFive’s Binning algorithm and HiCPipe.
The inconsistency between results for cis and trans interactions suggests that no approach is ideal for both types of interactions. To further explore this we looked at the effects of pseudocounts in the Binning/HiCPipe normalization scheme and the effects of distancedependence on normalization. Pseudocounts are values added to both expected and observed reads to mitigate the impact of stochastic effects. HiCPipe showed a stronger performance compared to HiFive’s Binning algorithm at longer ranges and at larger bin sizes. We determined that the primary difference was the inclusion of pseudocounts in all feature bins prior to normalization. By progressively adding counts, we found that cis interaction correlations decreased at shorter interaction ranges and overall, although the correlations increased at longer ranges and for trans interactions (Additional file 1: Figure S11).
We also performed parallel analyses using our weighted matrix balancing algorithm, Express, with and without the estimated distancedependence signal removed prior to normalization (Additional file 1: Figure S12). This showed a similar effect to the addition of pseudocounts, such that leaving the distancesignal relationship intact resulted in stronger longrange interaction correlations in larger bin sizes, stronger 1 Mb binned trans correlations, and poorer overall cis interaction correlations across all bin sizes.
Computational requirements
In order to determine the computational requirements of each analysis method, we ran each analysis on an abbreviated dataset consisting of a single chromosome of cis interactions from the mouse NcoI dataset starting from loading data through producing a 10 Kb heatmap. All normalizations were run using a single processor and publicly available scripts/programs. The exception to this was binning the counts and fragment feature data for HiCNorm. No script was provided for this step so one was written in R to supplement HiCNorm’s functionality.
Runtimes varied greatly between normalization methods, ranging from less than 7 min to approximately 12.5 h (Fig. 4). With the exception of HiFive’s Probability algorithm, HiFive performed better in terms of runtime than all other algorithms. HiCPipe and HiCNorm both showed long runtimes at least an order of magnitude above other methods. The slowest approach, though, was HiFive’s Probability algorithm. This was due to its modeling of every interaction combination across the chromosome. HiFive’s implementation of the KnightRuiz matrix balancing algorithm, ExpressKR, showed a dramatically faster runtime than any other approach. This was the result of HiFive’s fast data loading and efficient heatmapping without the need for distancedependence parameter calculations.
Scalability
Because of the everincreasing resolution of experiments and the corresponding size of interaction datasets, scalability is a primary concern for HiC data analysis. Although we compared methods on an even playing field, this does not reflect the complete performance picture when considering finerscale resolution, processing a complete genome interaction dataset, and more available computational resources.
There are two approaches to determining analysis resolution, prior to or after normalization. Of the methods presented, only HiCNorm determines the resolution of analysis prior to normalization. While it performs well, this means that the processing time and memory requirements scale exponentially with resolution. We were unable to perform any analyses at resolutions for bin sizes smaller than 10 Kb using this approach. The remaining methods all find correction values for individual fends, meaning that corrections are performed prior to binning interactions.
The increase in dataset size, either due to genome size itself or a finerscale partitioning of the genome, can be offset by employing more processing power by means of parallelization. HiCLib and HiCNorm do not appear to have any such capability. HiCPipe does have the ability to parallelize calculation of model bin sizes prior to normalization and calculations for heatmap production, although a single processor performs all of the actual normalization calculations. HiFive, on the other hand, has the ability to run in parallel for nearly all phases of an analysis. The two exceptions are loading the initial data and filtering reads, although the latter is very fast already. All normalization algorithms, including the KnightRuiz algorithm implemented in HiFive, have been parallelized for HiC analysis using MPI. The parallelization is implemented in such a way that the additional memory overhead for each new process is minimal.
Conclusions
HiC analysis remains a challenging subject, as demonstrated by the varied performances across all methodologies discussed here. No single approach appears to be ideally suited for all cases, suggesting that the experimental goal should drive the choice of analysis software. It is unclear how best to assess HiC normalization performance as there is no ‘gold standard’ for determining the quality of a HiC dataset or how well systematic noise has been accounted for during an analysis. As seen in the differences in correlation between mouse and human datasets (Fig. 3), factors such as restriction fragment size distributions, cut site density, sequencing depth, and HiC protocol can dramatically impact the similarity of resulting datasets. Further, in order to detect biologically relevant features against the background of the distancesignal relationship, the data need to be transformed, typically using a logtransformation. This skews the resulting comparison by ignoring interactions for which no reads have been observed, an increasing problem as binning size decreases or interaction size increases. At longer ranges, nonzero bins are sparse and dominated by macro features (such as AB compartments), a situation that can result in increasing correlations (Fig. 3a and b). Two observations suggest this is not an artifact. First, the longrange interaction correlation increase is seen in the human but not mice data, reflecting differences in genome organization. Second, the correlation increases are seen across all methodologies and algorithms.
Normalization software attempts to account for many of these confounding factors and allow direct comparison between datasets produced by different labs, protocols, and even across species although what can reasonably be expected in terms of this normalization process is unclear. This question depends on many factors and we may not have sufficient understanding of chromatin architecture variability across a cell population to answer it accurately. The resolution (bin size), similarity of datasets in terms of sequencing depth, restriction fragment size distributions, and protocol, as well as cell population size and population similarity from which the HiC libraries were made will all influence the correlation. At a low resolution, say 1 Mb, we should expect nearly a perfect correlation. However, at much higher resolution differences in mappability and RE cutsite frequency will strongly influence the correlation. Further, we need to consider the distance dependence of the signal as this is the strongest driver of the correlation and can give a false impression of comparability between datasets.
To address these normalization challenges, we have created HiFive, an easytouse, fast, and efficient framework for working with a variety of chromatin conformation data types. Because of the modular storage scheme, reanalysis and downstream analysis is made easier without additional storage or processing requirements. We have included several different normalization approaches and made nearly all aspects of each algorithm adjustable, allowing users to tune parameters for a wide range of analysis goals. HiFive is parallelized via MPI, making it highly scalable for nearly every step of HiC data processing.
For 5C data, HiFive is the only analysis program available for normalization and allows easy management of 5C data. We have demonstrated that 5C data normalizations performed by HiFive greatly improve consistency between 5C data and corresponding HiC data across multiple datasets.
We have also shown HiFive’s performance in handling HiC data. HiFive is consistently performing at or above other available methods as measured by interdataset correlations for cis interactions. In addition, we have demonstrated that HiFive is tunable to achieve superior trans performance if desired, albeit at the expense of performance across cis interactions. HiFive has also proved capable of handling very highresolution data, making it useful for the next generation of HiC experimental data.
In terms of performance considerations, our analysis suggests that, out of all of the methods considered, the balance between speed and accuracy is best achieved by HiFiveExpress or HiFiveExpressKR. This appears to be true regardless of resolution or dataset size. In order to get this performance, it is crucial to use the distancedependence adjustment prior to normalizing, necessitating the need to precalculate the distancedependence function. Because this requires iterating over every possible interaction, using multiple processors is highly recommended. If not possible, HiFiveExpressKR without distance correction is a robust fallback method. If computational resources are not a limiting factor, we recommend HiFiveProbability. With approximately 100 CPUs, the highresolution human data were processed in about a day. At finescale binning, this approach yields the best results of all methods.
While HiFive allows for superior normalization of data compared to other available software under many conditions, it also provides users with alternative options for fast analysis with minimal computational requirements at only a slight accuracy cost, opening highresolution HiC and 5C analysis to a much larger portion of the scientific community. HiFive is available at http://taylorlab.org/software/hifive/. Source code is provided under an MIT license and at https://github.com/bxlab/hifive or installed using pip from http://pypi.python.org.
Abbreviations
 3C:

chromosome conformation capture
 5C:

3C carbon copy
 bp:

base pair
 cis:

intrachromosomal
 CPU:

central processing unit
 ESC:

Embryonic stem cell
 Fend:

fragmentend
 Kb:

kilobase
 Mb:

megabase
 MPI:

message passing interface
 RE:

restriction enzyme
 trans:

interchromosomal
References
 1.
Arnone MI, Davidson EH. The hardwiring of development: organization and function of genomic regulatory systems. Development. 1997;124:1851–64.
 2.
Zinzen RP, Girardot C, Gagneur J, Braun M, Furlong EE. Combinatorial binding predicts spatiotemporal cisregulatory activity. Nature. 2009;462:65–70.
 3.
He A, Kong SW, Ma Q, Pu WT. Cooccupancy by multiple cardiac transcription factors identifies transcriptional enhancers active in heart. Proc Natl Acad Sci U S A. 2011;108:5632–7.
 4.
Cantone I, Fisher AG. Epigenetic programming and reprogramming during development. Nat Struct Mol Biol. 2013;20:282–9.
 5.
Varriale A. DNA Methylation, epigenetics, and evolution in vertebrates: facts and challenges. Int J Evol Biol. 2014;2014:475981.
 6.
Kimura H. Histone modifications for human epigenome analysis. J Hum Genet. 2013;58:439–45.
 7.
Fullwood MJ, Han Y, Wei CL, Ruan X, Ruan Y. Chromatin interaction analysis using pairedend tag sequencing. Curr Protoc Mol Biol. 2010;Chapter 21:Unit 21.15.21–25.
 8.
Kalhor R, Tjong H, Jayathilaka N, Alber F, Chen L. Genome architectures revealed by tethered chromosome conformation capture and populationbased modeling. Nat Biotechnol. 2012;30:90–8.
 9.
Dekker J, Rippe K, Dekker M, Kleckner N. Capturing chromosome conformation. Science. 2002;295:1306–11.
 10.
Dostie J, Richmond TA, Arnaout RA, Selzer RR, Lee WL, Honan TA, et al. Chromosome Conformation Capture Carbon Copy (5C): a massively parallel solution for mapping interactions between genomic elements. Genome Res. 2006;16:1299–309.
 11.
Zhao Z, Tavoosidana G, Sjolinder M, Gondor A, Mariano P, Wang S, et al. Circular chromosome conformation capture (4C) uncovers extensive networks of epigenetically regulated intra and interchromosomal interactions. Nat Genet. 2006;38:1341–7.
 12.
LiebermanAiden E, van Berkum NL, Williams L, Imakaev M, Ragoczy T, Telling A, et al. Comprehensive mapping of longrange interactions reveals folding principles of the human genome. Science. 2009;326:289–93.
 13.
van Berkum NL, Dekker J. Determining spatial chromatin organization of large genomic regions using 5C technology. Methods Mol Biol. 2009;567:189–213.
 14.
Yaffe E, Tanay A. Probabilistic modeling of HiC contact maps eliminates systematic biases to characterize global chromosomal architecture. Nat Genet. 2011;43:1059–65.
 15.
Rao SS, Huntley MH, Durand NC, Stamenova EK, Bochkov ID, Robinson JT, et al. A 3D map of the human genome at kilobase resolution reveals principles of chromatin looping. Cell. 2014;159:1665–80.
 16.
Hu M, Deng K, Selvaraj S, Qin Z, Ren B, Liu JS. HiCNorm: removing biases in HiC data via Poisson regression. Bioinformatics. 2012;28:3131–3.
 17.
Imakaev M, Fudenberg G, McCord RP, Naumova N, Goloborodko A, Lajoie BR, et al. Iterative correction of HiC data reveals hallmarks of chromosome organization. Nat Methods. 2012;9:999–1003.
 18.
Hu M, Deng K, Qin Z, Dixon J, Selvaraj S, Fang J, et al. Bayesian inference of spatial organizations of chromosomes. PLoS Comput Biol. 2013;9:e1002893.
 19.
Jin F, Li Y, Dixon JR, Selvaraj S, Ye Z, Lee AY, et al. A highresolution map of the threedimensional chromatin interactome in human cells. Nature. 2013;503:290–4.
 20.
Li W, Gong K, Li Q, Alber F, Zhou XJ. HiCorrector: a fast, scalable and memoryefficient package for normalizing largescale HiC data. Bioinformatics. 2015;31:960–2.
 21.
Servant N, Lajoie BR, Nora EP, Giorgetti L, Chen CJ, Heard E, et al. HiTC: exploration of highthroughput ‘C’ experiments. Bioinformatics. 2012;28:2843–4.
 22.
Rousseau M, Fraser J, Ferraiuolo MA, Dostie J, Blanchette M. Threedimensional modeling of chromatin structure from interaction frequency data using Markov chain Monte Carlo sampling. BMC Bioinformatics. 2011;12:414.
 23.
Nora EP, Lajoie BR, Schulz EG, Giorgetti L, Okamoto I, Servant N, et al. Spatial partitioning of the regulatory landscape of the Xinactivation centre. Nature. 2012;485:381–5.
 24.
PhillipsCremins JE, Sauria ME, Sanyal A, Gerasimova TI, Lajoie BR, Bell JS, et al. Architectural protein subclasses shape 3D organization of genomes during lineage commitment. Cell. 2013;153:1281–95.
 25.
Ay F, Bailey TL, Noble WS. Statistical confidence estimation for HiC data reveals regulatory chromatin contacts. Genome Res. 2014;24:999–1011.
 26.
Knight PA, Ruiz D. A fast algorithm for matrix balancing. IMA J Numerical Anal. 2013;33:1029–47.
 27.
Dixon JR, Selvaraj S, Yue F, Kim A, Li Y, Shen Y, et al. Topological domains in mammalian genomes identified by analysis of chromatin interactions. Nature. 2012;485:376–80.
 28.
Floyd SR, Pacold ME, Huang Q, Clarke SM, Lam FC, Cannell IG, et al. The bromodomain protein Brd4 insulates chromatin from DNA damage signalling. Nature. 2013;498:246–50.
 29.
Naumova N, Imakaev M, Fudenberg G, Zhan Y, Lajoie BR, Mirny LA, et al. Organization of the mitotic chromosome. Science. 2013;342:948–53.
 30.
Cournac A, MarieNelly H, Marbouty M, Koszul R, Mozziconacci J. Normalization of a chromosomal contact map. BMC Genomics. 2012;13:436.
 31.
Selvaraj S, RD J, Bansal V, Ren B. Wholegenome haplotype reconstruction using proximityligation and shotgun sequencing. Nat Biotechnol. 2013;31:8.
Acknowledgements
Research reported in this publication was supported by the National Institutes of Health under awards R01GM035463 to VC and R01DK065806 to JT and by American Recovery and Reinvestment Act (ARRA) funds through grant number RC2HG005542 to JT. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.
Author information
Affiliations
Corresponding author
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors’ contributions
MEGS, JEPC, VGC, and JT conceived the project and developed feature requirements. JEPC made significant contributions to the design of the 5C tools. MEGS developed all algorithms, designed and wrote all software, and wrote the manuscript. JT contributed to the manuscript and supported the project. All authors read and approved the final manuscript.
Additional files
Additional file 1:
A detailed methods section, a table listing the sources of datasets used, and Supplemental figures. (PDF 2245 kb)
Additional file 2:
A tar archive containing the HiFive software library. (BZ2 578 kb)
Additional file 3:
A tar archive containing all scripts using to generate the data, analyses, and figures presented in this paper. (BZ2 20009 kb)
Additional file 4:
The software documentation. (PDF 530 kb)
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
About this article
Cite this article
Sauria, M.E., PhillipsCremins, J.E., Corces, V.G. et al. HiFive: a tool suite for easy and efficient HiC and 5C data analysis. Genome Biol 16, 237 (2015). https://doi.org/10.1186/s130590150806y
Received:
Accepted:
Published:
Keywords
 Chromatin conformation
 Normalization
 Software
 Spatial organization
 Scalable
 Galaxy