Skip to main content

BayesHammer: Bayesian clustering for error correction in single-cell sequencing

Abstract

Error correction of sequenced reads remains a difficult task, especially in single-cell sequencing projects with extremely non-uniform coverage. While existing error correction tools designed for standard (multi-cell) sequencing data usually come up short in single-cell sequencing projects, algorithms actually used for single-cell error correction have been so far very simplistic.

We introduce several novel algorithms based on Hamming graphs and Bayesian subclustering in our new error correction tool BAYES HAMMER. While BAYES HAMMER was designed for single-cell sequencing, we demonstrate that it also improves on existing error correction tools for multi-cell sequencing data while working much faster on real-life datasets. We benchmark BAYES HAMMER on both k-mer counts and actual assembly results with the SPADES genome assembler.

Background

Single-cell sequencing [1, 2] based on the Multiple Displacement Amplification (MDA) technology [1, 3] allows one to sequence genomes of important uncultivated bacteria that until recently had been viewed as unamenable to genome sequencing. Existing metagenomic approaches (aimed at genes rather than genomes) are clearly limited for studies of such bacteria despite the fact that they represent the majority of species in such important studies as the Human Microbiome Project [4, 5] or discovery of new antibiotics-producing bacteria [6].

Single-cell sequencing datasets have extremely non-uniform coverage that may vary from ones to thousands along a single genome (Figure 1). For many existing error correction tools, most notably QUAKE [7], uniform coverage is a prerequisite: in the case of non-uniform coverage they either do not work or produce poor results.

Figure 1
figure 1

Logarithmic coverage plot for the single-cell E. coli dataset. Logarithmic coverage plot for the single-cell E. coli dataset (similar plot is also given in [2]).

Error correction tools usually attempt to correct the set of k-character substrings of reads called k-mers and then propagate corrections to whole reads which are important to have for many assemblers. Error correction tools often employ a simple idea of discarding rare k-mers, which obviously does not work in the case of non-uniform coverage.

Medvedev et al. [8] recently presented a new approach to error correction for datasets with non-uniform coverage. Their algorithm HAMMER makes use of the Hamming graph (hence the name) on k-mers (vertices of the graph correspond to k-mers and edges connect pairs of k-mers with Hamming distance not exceeding a certain threshold). HAMMER employs a simple and fast clustering technique based on selecting a central k-mer in each connected component of the Hamming graph. Such central k-mers are assumed to be error-free (i.e., they are assumed to actually appear in the genome), while the other k-mers from connected components are assumed to be erroneous instances of the corresponding central k-mers. However, HAMMER may be overly simplistic: in connected components of large diameter or connected components with several k-mers of large multiplicities, it is more reasonable to assume that there are two or more central k-mers (rather than one as in HAMMER). Biologically, such connected components may correspond to either (1) repeated regions with similar but not identical genomic sequences (repeats) which would be bundled together by existing error correction tools (including HAMMER); or (2) artificially united k-mers from distinct parts of the genome that just happen to be connected by a path in the Hamming graph (characteristic to HAMMER).

In this paper, we introduce the BAYES HAMMER error correction tool that does not rely on uniform coverage. BAYES HAMMER uses the clustering algorithm of HAMMER as a first step and then refines the constructed clusters by further subclustering them with a procedure that takes into account reads quality values (e.g., provided by Illumina sequencing machines) and introduces Bayesian (BIC) penalties for extra subclustering parameters. BAYES HAMMER subclustering aims to capture the complex structure of repeats (possibly of varying coverage) in the genome by separating even very similar k-mers that come from different instances of a repeat. BAYES HAMMER also uses a new approach for propagating corrections in k-mers to corrections in the reads. All algorithms in BAYES HAMMER are heavily parallelized whenever possible; as a result, BAYES HAMMER gains a significant speedup with more processing cores available. These features make BAYES HAMMER a perfect error correction tool for single-cell sequencing.

We remark that HAMMER produces only a set of central k-mers but does not correct reads, making it incompatible with most genome assemblers. QUAKE does correct reads but has severe memory limitations for large k and assumes uniform coverage. In contrast, EULER-SR [9] and CAMEL [2] correct reads and do not make strong assumptions on coverage (both tools have been used for single-cell assembly projects [2]) which makes these tools suitable for comparison to BAYES HAMMER. Our benchmarks show that BAYES HAMMER outperforms these tools in both single-cell and standard (multi-cell) modes. We further couple BAYES HAMMER with a recently developed genome assembler SPADES [10] and demonstrate that assembly of BAYES HAMMER-corrected reads significantly improves upon assembly with reads corrected by other tools for the same datasets, while the total running time also improves significantly.

BAYES HAMMER is freely available for download as part of the SPADES genome assembler at http://bioinf.spbau.ru/spades/.

Methods

Notation and outline

Let ∑ = {A, C, G, T} be the alphabet of nucleotides (BAYES HAMMER discards k-mers with uncertain bases denoted N). A k-mer is an element of ∑k, i.e., a string of k nucleotides. We denote the ith letter (nucleotide) of a k-mer x by x[i], indexing them from zero: 0 ≤ ik - 1. A subsequence of x corresponding to a set of indices I is denoted by x[I]. We use interval notation [i, j] for intervals of integers {i, i + 1,..., j} and further abbreviate x[i, j] = x [{i, i + 1,..., j}]; thus, x = x[0, k - 1]. Input reads are represented as a set of strings R Σ* along with their quality values ( q r [ i ] ) i = 0 | r | - 1 for each r R. We assume that q r [i] estimates the probability that there has been an error in position i of read r. Notice that in practice, the fastq file format [11] contains characters that encode probabilities on a logarithmic scale (in particular, products of probabilities used below correspond to sums of actual quality values).

Below we give an overview of BAYES HAMMER workflow (Figure 2) and refer to subsequent sections for further details. On Step (1), k-mers in the reads are counted, producing a triple statistics(x) = (count x , quality x , error x ) for each k-mer x. Here, count x is the number of times x appears as a substring in the reads, quality x is its total quality expressed as a probability of sequencing error in x, and error x is a k-dimensional vector that contains products of error probabilities (sums of quality values) for individual nucleotides of x across all its occurrences in the reads. On Step (2), we find connected components of the Hamming graph constructed from this set of k-mers. On Step (3), the connected components become subject to Bayesian subclustering; as a result, for each k-mer we know the center of its subcluster. On Step (4), we filter subcluster centers according to their total quality and form a set of solid k-mers which is then iteratively expanded on Step (5) by mapping them back to the reads. Step (6) deals with reads correction by counting the majority vote of solid k-mers in each read. In the iterative version, if there has been a substantial amount of changes in the reads, we run the next iteration of error correction; otherwise, output the corrected reads. Below we describe specific algorithms employed in the BAYES HAMMER pipeline.

Figure 2
figure 2

B AYES H AMMER workflow.

Algorithms

Step (1): computing k-mer statistics

To collect k-mer statistics, we use a straightforward hash map approach [12] that does not require storing instances of all k-mers in memory (as excessive amount of RAM might be needed otherwise). For a certain positive integer N (the number of auxiliary files), we use a hash function h: ∑k N that maps k-mers over the alphabet Σ to integers from 0 to N - 1.

Algorithm 1 Count k-mers

for each k-mer x from the reads R: do

compute h(x) and write x to Fileh(x).

for i [0, N - 1]: do

sort File i with respect to the lexicographic order;

reading File i sequentially, compute statistics(s) for each k-mer s from File i .

Step (2): constructing connected components of Hamming graph

Step (2) is the essence of the HAMMER approach [8]. The Hamming distance between k-mers x, y k is the number of nucleotides in which they differ:

d ( x , y ) = { i [ 0 , k - 1 ] : x [ i ] y [ i ] } .

For a set of k-mers X, the Hamming graph HG τ (X) is an undirected graph with the set of vertices X and edges corresponding to pairs of k-mers from X with Hamming distance at most τ, i.e., x, y X are connected by an edge in HG τ (X) iff d(x, y) ≤ τ (Figure 3). To construct HG τ (X) efficiently, we notice that if two k-mers are at Hamming distance at most τ, and we partition the set of indices [0,k - 1] into τ + 1 parts, then at least one part corresponds to the same subsequence in both k-mers. Below we assume with little loss of generality that τ + 1 divides k, i.e., k = σ (τ + 1) for some integer σ.

Figure 3
figure 3

Hamming graphs HG 1 ( X ) and HG 2 ( X ). Hamming graphs HG1(X) and HG2(X) for X being the set of 4-mers {ACGTG, CGTGT, GTGTG, ACATG, CATGT, ATGTG, ACCTG, CCTGT, CTGTC} of the reads ACGTGTG, ACATGTG, ACCTGTC. Blue edges denote Hamming distance 2.

For a subset of indices I [0, k - 1], we define a partial lexicographic ordering I as follows: x I y iff x[I] y[I], where is the lexicographic ordering on Σ*. Similarly, we define a partial equality = I such that x = I y iff x[I] = y[I]. We partition the set of indices [0, k - 1] into τ + 1 parts of size σ and for each part I, sort a separate copy of X with respect to I . As noticed above, for every two k-mers x, y X with d(x, y) ≤ τ, there exists a part I such that x = I y. It therefore suffices to separately consider blocks of equivalent k-mers with respect to = I for each part I. If a block is small (i.e., of size smaller than a certain threshold), we go over the pairs of k-mers in this block to find those with Hamming distance at most τ. If a block is large, we recursively apply to it the same procedure with a different partition of the indices. In practice, we use two different partitions of [0, k - 1]: the first corresponds to contigious subsets of indices (recall that σ = k τ + 1 ):

Algorithm 2 Hamming graph processing

procedure HGPROCESS(X, max_quadratic)

Init components with singletons X = { { x } : x X } .

for all ϒ FindBlocks ( X , { I s cnt } s = 0 τ ) do

if |ϒ| > max_quadratic then

   for all Z FindBlocks ( ϒ , { I s str } s = 0 τ ) do

      ProcessExhaustively ( Z , X )

else

   ProcessExhaustively ( ϒ , X ) .

function FindBlocks ( X , { I s } s = 0 τ )

for s = 0,...,τ do

sort a copy of X with respect to I s , getting X s .

for s = 0,...,τ do

output the set of equiv. blocks { ϒ } w . r . t . = I s .

procedure PROCESS EXHAUSTIVELY ( ϒ , X )

for each pair x, y ϒ do

if d(x, y) ≤ τ then join their sets in X :

   for all x Z x X , y Z y X do

       X : = X { Z x Z y } \ { Z x , Z y } .

I s cnt = { s σ , s σ + 1 , , s σ + σ - 1 } , s = 0 , , τ ,

while the second corresponds to strided subsets of indices:

I s str = { s , s + τ + 1 , s + 2 ( τ + 1 ) , , s + ( σ - 1 ) ( τ + 1 ) } , s = 0 , , τ .

BAYES HAMMER uses a two-step procedure, first splitting with respect to { I s cnt } s = 0 τ (Figure 4) and then, if an equivalence block is large, with respect to { I s str } s = 0 τ . On the block processing step, we use the disjoint set data structure [12] to maintain the set of connected components. Step (2) is summarized in Algorithm 2.

Figure 4
figure 4

Partial lexicographic orderings. Partial lexicographic orderings of a set X of 9-mers with respect to the index sets I 0 cnt = { 0 , 1 , 2 } , I 1 cnt = { 3 , 4 , 5 } , and I 2 cnt = { 6 , 7 , 8 } . Red dotted lines indicate equivalence blocks.

Step (3): Bayesian subclustering

In HAMMER's generative model [8], it is assumed that errors in each position of a k-mer are independent and occur with the same probability ε, which is a fixed global parameter (HAMMER used ε = 0.01). Thus, the likelihood that a k-mer x was generated from a k-mer y under HAMMER's model equals

L H AMMER ( x | y ) = ( 1 - ε ) k - d ( x , y ) ε d ( x , y ) .

Under this model, the maximum likelihood center of a cluster is simply its consensus string [8].

In BAYES HAMMER, we further elaborate upon HAMMER's model. Instead of a fixed ε, we use reads quality values that approximate probabilities q x [i] of a nucleotide at position i in the k-mer x being erroneous. We combine quality values from identical k-mers in the reads: for a multiset of k-mers X that agree on the jth nucleotide, it is erroneous with probability ΠxXq x [j].

The likelihood that a k-mer x has been generated from another k-mer c (under the independent errors assumption) is given by

L ( x | c ) = j : x [ j ] c [ j ] q x [ j ] j : x [ j ] = c [ j ] ( 1 - q x [ j ] ) ,

and the likelihood of a specific subclustering C = C1 ... C m is

L m ( C 1 , , C m ) = i = 1 m x C i L ( x | c i )

where c i is the center (consensus string) of the subcluster C i .

In the subclustering procedure (see Algorithm 3), we sequentially subcluster each connected component of the Hamming graph into more and more clusters with the classical k-means clustering algorithm (denoted m-means since k has different meaning). For the objective function, we use the likelihood as above penalized for overfitting with the Bayesian information criterion (BIC) [13]. In this case, there are |C| observations in the dataset, and the total number of parameters is 3 km + m - 1:

  • m - 1 for probabilities of subclusters,

  • km for cluster centers, and

  • 2 km for error probabilities in each letter: there are 3 possible errors for each letter, and the probabilities should sum up to one. Here error probabilities are conditioned on the fact that an error has occurred (alternatively, we could consider the entire distribution, including the correct letter, and get 3 km parameters for probabilities but then there would be no need to specify cluster centers, so the total number is the same).

Algorithm 3 Bayesian subclustering

for all connected components C of the Hamming graph do

m := 1

1 := 2 log L1(C) (likelihood of the cluster generated by the consensus)

repeat

m := m + 1

do m-means clustering of C = C1 ... C m w.r.t. the Hamming distance; the initial approximation to the centers is given by k-mers that have the least error probability

m := 2 · log L m (C1,...,C m ) (3 km + m - 1) · log |C|

until m m-1

output the best found clustering C = C1 ... Cm-1

Therefore, the resulting objective function is

m : = 2 log L m ( C 1 , , C m ) - ( 3 k m + m - 1 ) log | C |

for subclustering into m clusters; we stop as soon as m ceases to increase.

Steps (4) and (5): selecting solid k-mers and expanding the set of solid k-mers

We define the quality of a k-mer x as the probability that it is error-free: p x = j = 0 k - 1 ( 1 - q x [ j ] ) . The k-mer qualities are computed on Step (1) along with computing k-mer statistics. Next, we (generously) define the quality of a cluster C as the probability that at least one k-mer in C is correct:

p C = 1 - x C ( 1 - p x ) .

In contrast to HAMMER, we do not distinguish whether the cluster is a singleton (i.e., |C| = 1); there may be plenty of superfluous clusters with several k-mers obtained by chance (actually, it is more likely to obtain a cluster of several k-mers by chance than a singleton of the same total multiplicity).

Initially we mark as solid the centers of the clusters whose total quality exceeds a predefined threshold (a global parameter for BAYES HAMMER, set to be rather strict). Then we expand the set of solid k-mers iteratively: if a read is completely covered by solid k-mers we conclude that it actually comes from the genome and mark all other k-mers in this read as solid, too (Algorithm 4).

Step (6): reads correction

After Steps (1)-(5), we have constructed the set of solid k-mers that are presumably error-free. To construct corrected reads from the set of solid k-mers, for each base of every read, we compute the consensus of all solid k-mers and solid centers of clusters of all non-solid k-mers covering this base (Figure 5). This step is formally described as Algorithm 5.

Figure 5
figure 5

Read correction. Reads correction. Grey k-mers indicate non-solid k-mers. Red k-mers are the centers of the corresponding clusters (two grey k-mers striked through on the right are non-solid singletons). As a result, one nucleotide is changed.

Algorithm 4 Solid k-mers expansion

procedure ITERATIVE EXPANSION(R, X)

while ExpansionStep(R, X) do

function EXPANSION STEP(R, X)

for all reads r R do

if r is completely covered by solid k-mers then

   mark all k-mers in r as solid

Return TRUE if X has increased and FALSE otherwise.

Algorithm 5 Reads correction

Input: reads R, solid k-mers X, clusters C.

for all reads r R do

init consensus array υ: [0, |r| - 1] × {A, C, G, T} → with zeros: υ(j, x[i]):= 0 for all i = 0,...,|r| - 1 and j = 0,...,k - 1

for i = 0,...,|r| - k do

if r[i, i + k - 1] X (it is solid) then

   for j [i, i + k - 1] do

      υ(j, r[i]):= υ(j, r[i]) + 1

if r[i, i + k - 1] C for some C Cthen

   let x be the center of C

   if x X (r belongs to a cluster with solid center) then

      for j [i, i + k - 1] do

         υ(j, x[i]):= υ(j, x[i]) + 1

for i [0, |r| - 1] do

r[i]:= arg maxaΣυ(i, a).

Results and discussion

Datasets

In our experiments, we used three datasets from [2]: a single-cell E. coli, a single-cell S. aureus, and a standard (multicell) E. coli dataset. Paired-end libraries were generated by an Illumina Genome Analyzer IIx from MDA-amplified single-cell DNA and from multicell genomic DNA prepared from cultured E. coli, respectively These datasets consist of 100 bp paired-end reads with insert size 220; both E. coli datasets have average coverage ≈ 600×, although the coverage is highly non-uniform in the single-cell case.

In all experiments, BAYES HAMMER used k = 21 (we observed no improvements for higher values of k).

k-mer counts

Table 1 shows error correction statistics produced by di erent tools on all three datasets. For a comparison with HAMMER, we have emulated HAMMER with read correction by turning off Bayesian subclustering (HammerExpanded in the table) and both Bayesian subclustering and read expansion, another new idea of BAYES HAMMER (HammerNoExpansion in the table). Note that despite its more complex processing, BAYES HAMMER is significantly faster than other error correction tools (except, of course, for HAMMER which is a strict subset of BAYES HAMMER processing in our experiments and is run on BAYES HAMMER code). BAYES HAMMER also produces, in the single-cell case, a much smaller set of k-mers in the resulting reads which leads to smaller de Bruijn graphs and thus reduces the total assembly running time. Since BAYES HAMMER trims only bad quality bases and does not, like QUAKE, trim bases that it has not been able to correct (it has been proven detrimental for single-cell assembly in our experiments), it does produce a much larger set of k-mers than Quake on a multi-cell dataset.

Table 1 k-mer statistics.

For a comparison of BAYES HAMMER with other tools in terms of error rate reduction across an average read, see the logarithmic error rate graphs on Figure 6. Note that we are able to count errors only for the reads that actually aligned to the genome, so the graphs are biased in this way. Note how the first 21 bases are corrected better than others in BAYES HAMMER and both versions of HAMMER since we have run it with k = 21; still, other values of k did not show a significant improvement in either k-mer statistics or, more importantly, assembly results.

Figure 6
figure 6

Error reduction. Error reduction by read position on logarithmic scale for the single-cell E. coli, single-cell S. aureus, and multi-cell E. coli datasets.

Assembly results

Tables 2 and 3 shows assembly results by the recently developed SPAdes assembler [10]; SPAdes was designed specifically for single-cell assembly, but has by now demonstrated state-of-the-art results on multi-cell datasets as well.

Table 2 Assembly results, single-cell E.coli and S. aureus datasets (contigs of length ≥ 200 are used).
Table 3 Assembly results, multi-cell E.coli dataset (contigs of length ≥ 200 are used).

In the tables, N50 is such length that contigs of that length or longer comprise 1 2 of the assembly; NG50 is a metric similar to N50 but only taking into account contigs comprising (and aligning to) the reference genome; NA50 is a metric similar to N50 after breaking up misassembled contigs by their misassemblies. NGx and NAx metrics have a more direct relevance to assembly quality than regular Nx metrics; our result tables have been produced by the recently developed tool QUAST [14].

All assemblies have been done with SPADES. The results show that after BAYES HAMMER correction, assembly results improve significantly, especially in the single-cell E. coli case; it is especially interesting to note that even in the multi-cell case, where BAYES HAMMER loses to QUAKE by k-mer statistics, assembly results actually improve over assemblies produced from QUAKE-corrected reads (including genome coverage and the number of genes).

Conclusions

Single-cell sequencing presents novel challenges to error correction tools. In contrast to multi-cell datasets, for single-cell datasets, there is no pretty distribution of k-mer multiplicities; one therefore has to work with k-mers on a one-by-one basis, considering each cluster of k-mers separately. In this work, we further developed the ideas of HAMMER from a Bayesian clustering perspective and presented a new tool BAYES HAMMER that makes them practical and yields significant improvements over existing error correction tools.

There is further work to be done to make our underlying models closer to real life; for instance, one could learn a non-uniform distribution of single nucleotide errors and plug it in our likelihood formulas. Another natural improvement would be to try and rid the results of contamination by either human or some other DNA material; we observed significant human DNA contamination in our single-cell dataset, so weeding it out might yield a significant improvement. Finally, a new general approach that we are going to try in our further work deals with the technique of minimizers introduced by Roberts et al. [15]. It may provide significant reduction in memory requirements and a possible approach to dealing with paired information.

Declarations

The publication costs for this article were funded by the Government of the Russian Federation, grant 11.G34.31.0018.

This article has been published as part of BMC Genomics Volume 14 Supplement 1, 2013: Selected articles from the Eleventh Asia Pacific Bioinformatics Conference (APBC 2013): Genomics. The full contents of the supplement are available online at http://www.biomedcentral.com/bmcgenomics/supplements/14/S1.

References

  1. Grindberg R, Ishoey T, Brinza D, Esquenazi E, Coates R, Liu W, Gerwick L, Dorrestein P, Pevzner P, Lasken R, Gerwick W: Single cell genome amplification accelerates identification of the apratoxin biosynthetic pathway from a complex microbial assemblage. PLOS One. 2011, 6 (4): e18565-10.1371/journal.pone.0018565.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  2. Chitsaz H, Yee-Greenbaum JL, Tesler G, Lombardo MJ, Dupont CL, Badger JH, Novotny M, Rusch DB, Fraser LJ, Gormley NA, Schulz-Trieglaff O, Smith GP, Evers DJ, Pevzner PA, Lasken RS: Efficient de novo assembly of single-cell bacterial genomes from short-read data sets. Nat Biotechnol. 2011, 29: 915-921. 10.1038/nbt.1966.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  3. Ishoey T, Woyke T, Stepanauskas R, Novotny M, Lasken R: Genomic sequencing of single microbial cells from environmental samples. Current Opinion in Microbiology. 2008, 11 (3): 198-204. 10.1016/j.mib.2008.05.006.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  4. Gill S, Pop M, Deboy R, Eckburg P, Turnbaugh P, Samuel B, Gordon J, Relman D, Fraser-Liggett C, Nelson K: Metagenomic analysis of the human distal gut microbiome. Science. 2006, 312 (5778): 1355-1359. 10.1126/science.1124234.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  5. Hamady M, Knight R: Microbial community profiling for human microbiome projects: tools, techniques, and challenges. Genome Res. 2009, 19 (7): 1141-1152. 10.1101/gr.085464.108.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  6. Li J, Vederas J: Drug discovery and natural products: end of an era or an endless frontier?. Science. 2009, 325 (5937): 161-165. 10.1126/science.1168243.

    Article  PubMed  Google Scholar 

  7. Kelley DR, Schatz MC, Salzberg SL: Quake: quality-aware detection and correction of sequencing errors. Genome Biology. 2010, 11 (11): R116-10.1186/gb-2010-11-11-r116.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  8. Medvedev P, Scott E, Kakaradov B, Pevzner P: Error correction of high-throughput sequencing datasets with non-uniform coverage. Bioinformatics. 2011, 27 (13): i137-41. 10.1093/bioinformatics/btr208.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  9. Chaisson MJ, Pevzner P: Short read fragment assembly of bacterial genomes. Genome Research. 2008, 18: 324-330. 10.1101/gr.7088808.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  10. Bankevich A, Nurk S, Antipov D, Gurevich A, Dvorkin M, Kulikov A, Lesin V, Nikolenko S, Pham S, Prjibelski A, Pyshkin A, Sirotkin A, Vyahhi N, Tesler G, Alekseyev M, Pevzner P: SPAdes: a new genome assembler and its applications to single cell sequencing. Journal of Computational Biology. 2012, 19 (5): 455-477. 10.1089/cmb.2012.0021.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  11. Cock P, Fields C, Goto N, Heuer M, Rice P: The Sanger FASTQ file format for sequences with quality scores, and the Solexa/Illumina FASTQ variants. Nucleic Acids Res. 2010, 38 (6): 1767-1771. 10.1093/nar/gkp1137.

    Article  PubMed Central  CAS  PubMed  Google Scholar 

  12. Cormen TH, Leiserson CE, Rivest R: Introduction to Algorithms. 2009, MIT Press

    Google Scholar 

  13. Schwarz G: Estimating the dimension of a model. Annals of Statistics. 1978, 6: 461-464. 10.1214/aos/1176344136.

    Article  Google Scholar 

  14. Gurevich A, Saveliev V, Vyahhi N, Tesler G: QUAST: Quality Assessment for Genome Assemblies. 2012, [Submitted]

    Google Scholar 

  15. Roberts M, Hayes W, Hunt BR, Mount SM, Yorke JA: Reducing storage requirements for biological sequence comparison. Bioinformatics. 2004, 20 (18): 3363-3369. 10.1093/bioinformatics/bth408.

    Article  CAS  PubMed  Google Scholar 

Download references

Acknowledgements

We thank Pavel Pevzner for many fruitful discussions on all stages of the project. We are also grateful to Andrei Prjibelski and Alexei Gurevich for help with the experiments and to the anonymous referees whose comments have benefited the paper greatly. This work was supported the Government of the Russian Federation, grant 11.G34.31.0018. Work of the first author was also supported by the Russian Fund for Basic Research grant 12-01-00450-a and the Russian Presidential Grant MK-6628.2012.1. Work of the second author was additionally supported by the Russian Fund for Basic Research grant 12-01-00747-a.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Sergey I Nikolenko.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors' contributions

All authors contributed extensively to the work presented in this paper.

Rights and permissions

This article is published under license to BioMed Central Ltd. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and permissions

About this article

Cite this article

Nikolenko, S.I., Korobeynikov, A.I. & Alekseyev, M.A. BayesHammer: Bayesian clustering for error correction in single-cell sequencing. BMC Genomics 14 (Suppl 1), S7 (2013). https://doi.org/10.1186/1471-2164-14-S1-S7

Download citation

  • Published:

  • DOI: https://doi.org/10.1186/1471-2164-14-S1-S7

Keywords