Review of massively parallel DNA sequencing technologies
- First Online:
- 1.3k Downloads
Since the development of technologies that can determine the base-pair sequence of DNA, the ability to sequence genes has contributed much to science and medicine. However, it has remained a relatively costly and laborious process, hindering its use as a routine biomedical tool. Recent times are seeing rapid developments in this field, both in the availability of novel sequencing platforms, as well as supporting technologies involved in processes such as targeting and data analysis. This is leading to significant reductions in the cost of sequencing a human genome and the potential for its use as a routine biomedical tool. This review is a snapshot of this rapidly moving field examining the current state of the art, forthcoming developments and some of the issues still to be resolved prior to the use of new sequencing technologies in routine clinical diagnosis.
KeywordsNext generation sequencing Targeting Massively parallel
The basic principles of DNA sequencing have remained constant since the development of the first practical method by Sanger et al. (1977). Even so, the classic Sanger method has undergone various modifications and refinements in the intervening years, most recently driven by the requirements of the Human Genome Project to facilitate automation and to increase throughput. Perhaps most significantly, contemporary methods use four different base-specific fluorescent dyes (Smith et al. 1986) instead of radioactive labels, and cumbersome gel electrophoresis has been replaced by automated capillary electrophoresis (Luckey et al. 1990). These developments have dramatically increased the efficiency of Sanger sequencing, which is now widely considered the gold standard for clinical diagnostic use. However, the technique remains too laborious and expensive for routine sequencing of anything more than a few genes. In an attempt to address this short-coming, a diverse array of new sequencing technologies have been developed and are currently in development. Although the widely perceived aim of practical and affordable whole genome sequencing is ambitious, requiring major improvements in run capacity, speed of processing and cost, progress to date has been remarkable (see reviews Mardis 2011; Metzker 2010; Pettersson et al. 2009; Tucker et al. 2009; Voelkerding et al. 2009).
Classification of technologies
The terminology surrounding the new sequencing technologies is diverse and often confusing with terms such as ‘next generation’, ‘massively parallel’ and ‘clonal’ sequencing being used as global classifiers for, what is, essentially the same thing. In an attempt to bring some clarity to classification we have divided DNA sequencing technologies into three generations (Pettersson et al. 2009).
The first generation is synonymous with Sanger sequencing, which has been predominant since the 1970s. The defining characteristics of this technology are that each sequencing reaction represents a single, predefined target (up to about 1 kb) and this represents all copies of that target present in the original sample and thus its allelic content. The underlying principle of all post-Sanger DNA sequencing technologies, which is enabling the explosion in capacity and exponentially decreasing costs, is massive parallelisation. A fragmented input sample is captured on an array in such a way that each spatially identifiable location or feature is populated by a single target molecule. Depending on the technology a single sequencing array may comprise many millions or even billions of features, which are all sequenced in parallel in a single run, each feature generating a single sequencing ‘read’. This is fundamentally different from Sanger sequencing in two key respects: first, the specific location of the reads is not pre-determined and so must be computationally determined (referred to as mapping or alignment); and second, because each read represents a single starting molecule, multiple coverage is required to analyse the full allelic content of the sample.
With second generation sequencing, widely referred to as ‘next generation sequencing (NGS)’ it is necessary to clonally amplify the isolated targets in order to generate sufficient signal for detection during the sequencing run. This process is usually performed in situ on a solid substrate and generates clusters of many thousands of identical DNA targets (sometimes called polonies) at each feature. With these technologies sequencing is performed through stepwise incorporation of suitably modified subunits. Generally speaking read lengths are shorter than those achieved by Sanger sequencing, although they are rapidly improving. This is an important consideration since short read length can make accurate assembly and alignment computationally challenging (Flicek et al. 2011; Li et al. 2008; Li and Durbin 2009). The key difference with third generation, or ‘next next generation’, sequencing is that the chemistry and/or detection has been refined so that no clonal amplification of the target is required before the run. These technologies are predominantly still in development and use a wide range of different detection methodologies.
Below we review the principles behind these alternative technologies, compare and contrast their characteristics, and provide an overview of some of the targeting techniques and bioinformatics tools that have been developed alongside them.
Second generation sequencing
Summary of existing NGS platforms
This method closely resembles the Sanger sequencing-by-synthesis method, but uses special fluorescently labelled terminator nucleotides, which allow the chain termination process to be reversed (Bentley et al. 2008). It was originally developed by Solexa and is now commercialised by Illumina through the Genome Analyser and HiSeq systems; a further addition to the range will be the MiSeq, a lower capacity instrument due for release in mid-2011.
Template DNA molecules are generated by fragmentation of the sample followed by ligation of end specific universal adaptors. These fragments are then hybridised to a dense ‘lawn’ of universal probes immobilised to a glass surface known as a flow cell upon which both amplification and sequencing take place. Clonal amplification is performed using a process termed ‘bridge amplification’; a surface PCR which uses two tethered universal primers to create dense clusters of identical DNA across the plate. The sequencing reaction begins with the addition of a universal sequencing primer, which hybridises to the adaptor sequences added in the first stage. The sequencing chemistry involves three stages. First, chain extension is performed using DNA polymerase and the four reversible nucleotide terminators, each labelled with a different fluorescent dye. Incorporation of a complementary nucleotide results in termination of polymerisation—this process is allowed to run to completion to ensure all templates on the flow cell are extended by a single base. Next, unincorporated nucleotides are washed off and the incorporated base on each cluster is identified by colour imaging. Finally, the dye and the terminating group are chemically cleaved to prepare the templates for the next round of incorporation and imaging. These three stages are repeated over several hundred cycles generating a temporal series of colour images, which can be computationally converted into sequence reads each corresponding to a feature on the array.
Pyrosequencing is also based on a sequencing-by-synthesis technique, but rather than measuring fluorescence associated with specific nucleotides, it relies on indirect detection of incorporation events (Margulies et al. 2005). This technology was available in individual reaction form (Qiagen) before a massively parallelised version was commercialised by Roche/454. Two platforms supporting this chemistry are now available: the Genome Sequencer FLX and the GS Junior, a low capacity version.
Template DNA molecules can be generated either by fragmentation or standard PCR. If fragmentation is used, universal adaptors are ligated to the fragment end, similar to the method used by Illumina. In the case of PCR, the adaptors can be built into the primers. The prepared fragments are hybridised to special beads upon which both amplification and sequencing takes place; the beads are used in excess to ensure that each bead binds a maximum of one template molecule. A mix of these beads, PCR reagents and oil is then agitated to form an emulsion of tiny oil reaction chambers, each containing a single bead with a single molecule attached and all the components of a PCR. This is subjected to thermal cycling to clonally amplify the DNA template on the surface of each bead (known as emulsion PCR or emPCR). The sequencing reaction is performed on a specially fabricated ‘PicoTitre Plate (PTP)’—this comprises millions of microscopic wells, each just big enough to contain a single template bead. After breaking the emulsion the beads are loaded onto the PTP along with other, much smaller beads that contain all the reagents necessary for the sequencing reaction except the nucleotides. Sequencing proceeds with the sequential addition of each individual nucleotide in turn (i.e. A, then C, then G, then T). If a nucleotide is incorporated by DNA polymerase into the growing DNA strand, an inorganic phosphate ion is released. This initiates an enzyme cascade resulting in the release of a flash of light. Since no terminators are used in this chemistry, incorporation of nucleotides into homopolymer stretches continues until a different base is encountered and the associated light flash is proportionally brighter. The location and intensity of light emitted is detected by a camera across the whole plate. Excess nucleotides are then washed off in preparation for the next cycle. This process is repeated several hundred times to build the temporal image sequence. Unlike other chemistries the number of incorporation cycles required to reach a particular read length is dependent on the sequence composition of the template. On average read length is expected to be ~2.5× cycle number but this could be less with more homopolymers. As before the temporal series of images can be computationally converted into sequence reads.
Sequencing by ligation
Unlike the previous techniques, this method does not involve polymerase based DNA synthesis, but instead uses ligation of fluorescently labelled hybridisation probes to determine the sequence of a template DNA strand two bases at a time (Shendure et al. 2005). It has been commercialised by Life Technologies/Applied Biosystems through the SOLiD (Sequencing by Oligonucleotide Ligation and Detection) system.1
Template DNA molecules are prepared by fragmentation, adaptor ligation, hybridisation to beads and emPCR in a similar fashion to that described for the Roche/454 system above. After breaking the emulsion the beads are immobilised at high density on a glass slide. Sequencing proceeds with the addition of a universal primer, followed by fluorescently labelled oligonucleotide probes. Each probe comprises eight bases, of which only the first two define the probe whilst the following six are degenerate (i.e. able to pair with any nucleotide sequence on the template strand). After the complementary probe hybridises to the template DNA, it is chemically linked to the growing strand by the enzyme DNA ligase. The flow cell is then washed to remove excess probes and imaged to record the ligation cycle. Then, the three terminal degenerate bases, along with the fluorescent dye, are cleaved from the bound probes and the flow cell is washed again (this is known as a ligation cycle). This process is repeated a number of times, after which the newly synthesised strand is entirely denatured and removed from the template. At this point a new primer, which is one base shorter than that previously used (i.e. n − 1), is hybridised to the template and a new round of ligation cycles performed. In all, five rounds of ligation cycles are performed, each one using a primer one base shorter than the last. By this process, all bases on the template strand are interrogated twice. Counter-intuitively, although there are 16 possible permutations of the first two bases in the probe, only four coloured dyes are used; thus each colour represents four possible different two-base permutations. The arrangement of colours is such that if the first base is known, the second can be inferred. Since the first base in the sequence belongs to the universal primer added initially, the rest of the sequence can be sequentially inferred from the raw colour data, which is called colour space, by applying logical rules. This system, known as 2-base encoding, enables miss-called bases to be distinguished from true sequence variants as the former lead to logical impossibilities.
Third generation sequencing
A large number of companies are involved in developing much faster and higher throughput third or next–next generation DNA sequencing systems, some of which have already launched and some of which are still in stealth mode. Most of these are focussed on sequencing single molecules of DNA in real-time, and although many are based on sequencing-by-synthesis, there are several novel methodologies such as monitoring the passage of DNA through nanopores. A key advantage of single molecule sequencing is that no clonal amplification is required. This not only reduces preparation time, but effectively eliminates biases and errors introduced at this stage. In addition it is generally expected that these methods will generate much longer reads (potentially tens to hundreds of kilobases) which will enable much more accurate mapping, particularly in repetitive regions, and facilitate haplotyping. Moreover, some technologies have been demonstrated to be capable of distinguishing methylated cytosine bases, which could open the door to direct epigenetic analysis.
There are numerous third generation sequencing platforms at very different stages of development—ranging from basic research through to a launched product—which may be destined for different applications based on the precise idiosyncrasies of the sequencing chemistry and resultant performance metrics (e.g. error rate, read length, yield per run, cost per base, etc.—see later). The third generation technologies can be divided into categories based on the method they use to detect the DNA sequence:
Most of the third generation sequencing platforms under development using fluorescence detection are based on the standard sequencing-by-synthesis method. The first single molecule sequencing platform to market was the Heliscope from Helicos Bioscience (Harris et al. 2008), launched in 2009, which is based on a similar methodology to that described for the Solexa/Illumina second generation platform. However this is a single molecule method and all the nucleotides have the same fluorescent label which acts as the terminating moiety. This means that the nucleotides need to be added individually and sequentially in order to identify which base is added when (Bowers et al. 2009). Following washing of excess nucleotides and polymerase, the slide is imaged to identify where bases were incorporated. The dye is then cleaved in preparation for the next nucleotide addition and the process repeated for each nucleotide, with each cycle extending the DNA strand by a single base. The data are then analysed to build up a sequence read for each location.
Another single molecule approach is the SMRT™ chemistry which has been made commercially available on the PACBIO RS platform from Pacific Biosciences. One of the main problems with single molecule sequencing is that the incorporation event that needs to be detected is so small it is difficult to detect above background noise. This is the main reason for the use of clonal amplification in second generation systems. Pacific Biosciences have solved this problem by performing the sequencing in specially designed wells called zero mode wave guides which effectively eliminate the background noise (Eid et al. 2009). Template DNA forms a complex with the polymerase and nucleotide incorporation is detected by laser excitation and fluorescence monitoring in each well. The difference between this method and others that use fluorophores is that the dye is attached to the phosphate of the nucleotide rather than the base itself. Thus it is cleaved and released as a natural part of DNA synthesis, resulting in release of the dye without interruption to the sequencing process. The sequencing is therefore both single molecule and real-time.
Life Technologies have developed an approach which uses a DNA polymerase modified by the addition a quantum dot [Qdot® (Karow 2010b)]. This is a tiny nanocrystal that absorbs photons of light, then re-emits photons at a different wavelength. Template DNA is immobilised to the surface of a glass slide, and sequencing is initiated by the addition of a primer, the modified DNA polymerase and nucleotides with base specific fluorescent labels. As bases are incorporated the nucleotide labels are energised by the Qdot on the polymerase in a process known as fluorescence resonance energy transfer (FRET), which generates a very strong localised fluorescence signal (around 100-fold greater than standard dyes). FRET can only occur when the two fluorescent moieties (polymerase and nucleotide) are in close proximity to each other i.e. at incorporation thus elegantly eliminating interfering background fluorescence in the reaction chamber. At the end of a sequencing run both the polymerase and newly synthesised DNA strand can be removed, allowing the immobilised template DNA to be sequenced repeatedly. Life technologies claim that this system allows the read length and sequencing accuracy to be tailored to the application by adjusting the mode of repetitive sequencing.
There are also numerous other smaller companies developing third generation DNA sequencing platforms based on fluorescence detection, such as GnuBio, which is developing a microfluidics device that uses microdroplets as miniature reaction vessels, thus vastly reducing the cost of the reagents.
Various third generation DNA sequencing platforms are being developed that are capable of converting a DNA sequence directly into an electrical signal. This essentially amounts to direct generation of digital information promising the enticing prospect of label-free sequencing. Platforms using such technology would be extremely cheap to produce and be both fast and scalable.
The system recently released by Ion Torrent/Life Technologies uses a sequencing-by-synthesis method almost identical to pyrosequencing. The key difference is that incorporations are detected by monitoring the release of H+ ions (protons) which are also released as a by-product of nucleotide incorporation (Karow 2010a). A proprietary semi-conductor chip, which is essentially a miniature pH meter, is divided into wells in which the sequencing reactions take place. If a nucleotide is incorporated in a particular well, a single H+ is release into solution and a concomitant change in acidity (pH) is detected as a voltage shift by sensors. The magnitude of the pH change can be related to the number of molecules of a particular base incorporated. Currently this system does not detect single molecules and amplification is required prior to sequencing, but the synthesis reaction is detected in real-time and no modified reagents are required.
A majority of other platforms currently under development for using electronic detection are not based on the sequencing-by-synthesis method, but on an entirely new method using either biological or solid state nanopores. These technologies monitor changes in electrical current as DNA strands or individual bases pass through a nanopore. The sequencing chamber is divided into two sub-chambers by a synthetic membrane or some other septa. Each sequencing chamber contains a single nanopore penetrating the septum providing a single channel between the two chambers. The nanopores themselves can either be small holes in an inorganic membrane (solid-state nanopores), such as silicon nitride (Aksimentiev et al. 2004) or grapheme (Garaj et al. 2010), or modified natural channel proteins like α-haemolysin (Howorka et al. 2001; Olasagasti et al. 2010) embedded in a lipid bilayer or synthetic membrane. Nanopore sequencing technologies are based on one of two approaches—either the DNA strand itself passes through the nanopore (strand sequencing), or individual bases are cleaved from the target DNA and fed sequentially through the nanopore. A voltage is placed across the membrane to drive the translocation of negatively charged DNA molecules through the pore. As DNA bases pass through the pore, the current is blocked and since each base blocks the current by a different amount the strand composition can be determined.
Numerous companies are currently developing nanopore-based DNA sequencing platforms, including Oxford Nanopore Technologies, NABSys, base4innovation, and IBM/Roche. Whilst this technique is extremely promising, there are still challenges to be overcome both technical, such as controlling the passage of bases though the nanopore to allow sequencing of consecutive bases, and those related to system performance, such as pore shelf life and parallelisation (Branton et al. 2008; Kircher and Kelso 2010). Perhaps the most advanced to date is the platform under development by Oxford Nanopore Technologies, which uses α-haemolysin nanopores modified with a cyclodextrin ring covalently bound in the barrel. The DNA is digested by an exonuclease and the individual bases are drawn through the pore one at a time driven by an electrical potential (Astier et al. 2006; Wu et al. 2007). It has been demonstrated that this system can also distinguish 5-methyl cytosine thus enabling direct methylation analysis (Clarke et al. 2009).
Recently an alternative method for detection and identification of nucleotides has been described: here the transverse conductivity of the molecule is measured as it passes between two electrodes embedded in a solid state nanopore (Tsutsui et al. 2010, 2011). The authors suggest alternative methods for translocation of the DNA through the nanopore such as ‘magnetic tweezers’ (Peng and Ling 2009).
Another novel technique for DNA uses transmission electron microscopy to directly visualise strands of DNA that have been suitably modified with heavy metal atoms to distinguish the bases (Krivanek et al. 2010). This method is being developed and commercialised by several companies, including Halcyon Molecular and ZSGenetics. The use of scanning tunnelling microscopy to sequence DNA molecules has also been described (Tanaka and Kawai 2009).
Advantages and disadvantages of different chemical targeting approaches
High sensitivity, specificity, reproducibility and uniformity
High cost, low throughput, and cannot be used for large regions or a very large number of genes
Low cost (if many samples), easy to use, high sensitivity and specificity
Uniformity and sensitivity depends on design of probes. Cannot be used for a very large number of genes
Medium cost, easy to use, high sensitivity and specificity. Can target large sections of DNA and large numbers of genes
Uniformity and sensitivity depends on design of probes. Array design may be rather inflexible
The current method of targeting for capillary sequencing is the polymerase chain reaction (PCR) (Saiki et al. 1985). This can equally be used for preparation of targets for NGS but owing to the massive capacity of these platforms very large numbers of PCRs are required to fill a run. The processing required can be limited by utilising multiplex PCR or long range PCR (Fredriksson et al. 2007; Varley and Mitra 2008). Commercial solutions to this problem include the RainStorm technology from RainDance Technologies which uses an emPCR approach to simultaneously amplify up to 4,000 short DNA sequences in separate microdroplets (Tewhey et al. 2009), and the Access Array from Fluidigm which uses proprietary microfluidics to setup an array of 2,304 PCRs (48 samples × 48 assays). It should be noted that neither of these systems is actually multiplex as the individual reactions are separated—this enables much higher levels of parallelisation than are achievable in a single reaction.
Circularisation methods are designed to resolve the interference issues that limit the level of multiplexing achievable by standard PCR and are suitable for targeting small to medium sized regions of interest. Several approaches have been demonstrated including gene collector (Fredriksson et al. 2007) gene selector (Dahl et al. 2005, 2007) and connector inversion probes (Akhras et al. 2007) but all are essentially based on padlock and molecular inversion probes (Krishnakumar et al. 2008; Li et al. 2009). The basic principle is that large panels of target molecules can be selected and circularised in a single reaction using specially designed probes containing universal sequences. The reaction is then subjected to exonuclease digestion which degrades all the unwanted DNA but leaves the targets untouched since, being circles, they have no ends. The target sequences can then be amplified using the universal sequences to generate suitable target material for sequencing.
The final method of targeting is hybridisation, which is based on the same principle as DNA microarray technology. Oligonucleotide probes are used to pull-down sequences of interest from whole, fragmented genomic DNA. Unwanted DNA is then washed off, and the captured material eluted and prepared for sequencing. The capture capacity ranges from a few Mb up to the entire exome (Hodges et al. 2007; Porreca et al. 2007) using two general methodologies: conventional solid state arrays (on-array capture) and paramagnetic beads (in-solution capture) (Albert et al. 2007; Chou et al. 2010; Gnirke et al. 2009). A number of custom hybridisation platforms are available including Agilent, Roche Nimblegen and Illumina.
The method of choice is dependent on application; in particular target size, type of target and sample number, required performance, ease of use and costs (Albert et al. 2007; Mamanova et al. 2010). Ideally the targeting method should allow enrichment of multiple different loci independent of their size, sequence composition or spatial distribution, and should be amenable to automation so that it can match the sequencing capacity. However, the current approaches have their own biases (see Table 2), which relate both to the types of sequences that they are able to capture and their ease of use. Key issues, which apply to all methods with varying degrees are uniformity, efficiency of coverage and off-target capture. Whilst these methods are continually being improved, it may ultimately be more cost-effective to sequence a whole genome, computationally masking regions of the genome that are irrelevant to a particular clinical question, and target analysis only to regions with proven clinical significance.
Rather than review the current performance of each platform (which can be found on each of the manufacturer’s websites and at http://www.molecularecologist.com/next-gen-fieldguide/), we outline some of the factors that affect performance and influence the utility of all whole genome sequencing technologies.
Analytical accuracy, systematic errors and quality of base calls
In addition to amplification errors (which will be eliminated by single molecule sequencing), all sequencing methodologies suffer from both random and systematic errors. The raw accuracy of the sequencing process and quality of base calling are critically important factors, particularly for clinical diagnostics. A quality score representing of the probability that the base is called correctly is assigned to each base (These are generally given on a logarithmic scale so that Q10 would be 10% probability of miscall Q20 is 1% probability, Q30 is 0.1% probability etc). Factors affecting the quality score include signal intensity, background noise in the reaction itself or generated by the instrument and crosstalk between clusters. Errors can include overcalls and undercalls (insertions or deletions of bases from the sequence) as well as miscalls (incorrect base assigned) (Albert et al. 2007; Brockman et al. 2008). Different sequencing technologies are prone to different systematic errors, which influence their utility for different applications; for example, accurately sequencing homopolymeric regions can be difficult using pyrosequencing due to intermediate fluorescence signal intensities resulting from the incorporation of n identical nucleotides.
Sanger sequencing has a low (but non-zero) error rate of around 10−4 to 10−5 for single calls (one error per 1,000–10,000 bases), but the accuracy for detecting heterozygous variants is much more difficult to assess and is almost certainly context dependent to some extent. When it comes to detection of a low level variant, for example mosaic or somatic mutation, the limit of detection in terms of minor allele representation is only around 20% for Sanger sequencing. Current NGS platforms have a somewhat higher raw error rate of around 10−2 to 10−3 (depending on read length), but this is easily offset by increasing read depth (i.e. consensus accuracy). In fact the desired accuracy can essentially be determined by altering the read depth appropriately. In contrast to Sanger sequencing detection of low level variants is basically limited by the raw error rate and would be substantially below 0.1%.
Read depth, genome coverage and uniformity
The read depth or depth of coverage refers simply to the number of times a base is sequenced in a single run of the machine. The required read depth varies depending upon the specific application and level of certainty required for the result. However, coverage of the genome is non-uniform due to factors such as repetitive elements, non-uniform targeting and variable GC content, which affects both amplification and sequencing efficiency (Dohm et al. 2008). For diagnostic purposes it is necessary to increase overall coverage to ensure the regions with least coverage meet the desired standard; any that do not should be failed. This can be very costly in terms of capacity, particularly where coverage is very variable. The theoretical depth required to detect a heterozygous variation with particular probability of success can be calculated. For re-sequencing applications mapping the reads is guided by a reference sequence and theoretically requires much lower coverage (8–12x), than assembling genomes de novo (25–70x) (Schuster 2008). However, in practice, a depth of coverage of around 20× at each base is required for confident variant calling (Bentley et al. 2008).
Read length (number of bases per read)
Read length is an important factor in certain applications, such as sequencing through repetitive regions, identifying genomic rearrangements and getting short range haplotype information. In addition, longer reads make alignment to a reference sequence substantially easier by reducing the number of potential matches. Current second generation NGS platforms achieve reads length of 35–400 bases (Metzker 2010; ten Bosch and Grody 2008), but this is rapidly improving. It is anticipated that many the third generation platforms will have substantially longer read lengths. For example Oxford Nanopore Technologies, Pacific Biosciences and Life Technologies claim that their new platforms will have read lengths in excess of 1 kb and claims beyond this are not infrequent.
Factors such as run capacity, sample multiplexing, run time and cost all have a major impact on the suitability of a particular platform for a particular application or laboratory. These factors vary substantially between machines, applications and chosen sequencing protocol. Whilst NGS has significantly reduced the per-base cost of sequencing, cost per test savings will only be realised if the capacity of the instrument is effectively used. In many cases the format of the experiment does not require the full capacity of a run for a single sample so methods for analysing multiple samples in a single run are important. Several methods are available to achieve such sample multiplexing (ten Bosch and Grody 2008). For targeted sequencing, it is possible to mix multiple different tests so that results from each specific test relate to only one individual patient. However, this is only effective if the reads are mapped to specific regions of interest and in many cases it is preferable to use the whole genome as a reference as this guards against non-specific targeting. In addition, many sequencing platforms allow physical separation between samples, for example by dividing the flow cell in to a number of channels. Finally, DNA ‘barcode tags’ can be added to the ends of DNA fragments during initial preparation. These are sequenced along with the fragment during the run and serve to identify the source of each sequence read during analysis (Binladen et al. 2007; Meyer et al. 2007).
Reagent and instrument cost
Reagent cost for sequencing has plummeted over the last decade, from a cost of around $500/Mb for Sanger sequencing reagents, to less than $0.50/Mb for reagents on the newest NGS platforms (Wetterstrand 2011). However, the sequencing machines themselves are often fairly expensive ranging from US$ 0.2–1 million. With ever increasing capacity the output from sequencing runs becomes greater and greater. The cost of handling and storing all this data should not be treated lightly—ultimately this is likely to be a more significant cost than generating the data.
Data analysis and interpretation
Primary analysis: base calling—converting light signal intensities into a sequence of nucleotides. This is generally performed automatically by software on the sequencing machine itself and each call is associated with a raw quality score.
Secondary analysis: alignment and variant calling—mapping DNA reads to an annotated reference sequence and determining the extent of variation from the reference.2 Because it is often not possible to unambiguously align a read to a unique position in the reference genome, particularly allowing for variation between the reference and the sample genome, a mapping quality score may be used to measure the likelihood that a read is mapped correctly. Numerous algorithms and software packages have been developed for this process (Flicek and Birney 2009; Magi et al. 2010) which is becoming increasingly automated. Various dedicated software packages have also been developed specifically for cancer genome assembly and variant calling, which take into account factors such as genetic heterogeneity in the sample (Ding et al. 2010; Magi et al. 2010). In the final stage of the alignment phase, sequence data are annotated with structural and functional biological information and visualised through a graphical interface or genome browser.
Tertiary analysis: interpretation—analysing and filtering variants to assess their inheritance, uniqueness, and likely functional impact (Kuhlenbñumer et al. 2010). This process requires comparison against databases of genomic variation (Kuntzer et al. 2010) (including both normal and pathogenic variants) and algorithms for evaluating the likely pathogenicity of a particular mutation [e.g. by assessing haploinsufficiency (Huang et al. 2010) of large deletions or loss of function variants (MacArthur and Tyler-Smith 2010), predicting the effect of amino acid substitutions caused by non-synonymous coding variants (Ng and Henikoff 2006)]. Although many software packages already exist for this process, accurate interpretation of the effect of genomic variants in an individual is still in its infancy, and more purpose-build packages will need to be developed to allow clinical diagnostic use. Meaningful clinical interpretation is likely to remain a major challenge for the foreseeable future.
A vast array of software packages are now available, both commercial and open source (i.e. freely available). Whilst open source options may offer a more flexible final solution, building a pipeline involves linking multiple software units, each performing a specific task, and remains the domain of dedicated bioinformaticians. The scope and applicability of integrated commercial packages is now enabling laboratory scientists to directly analyse their own data although it should be noted that significant time and effort will be necessary in order to acquire an appropriate level of understanding of the processes and how they affect the final results. A useful reference for available packages can be found at: http://seqanswers.com/forums/showthread.php?t=43.
A number of national and international centres offer both data production and analysis services. In the UK the MRC has funded the establishment of four regional sequencing hubs which are primarily intended to support small and medium sized research projects, and the Wellcome Trust Sanger Institute continues to undertake large-scale sequencing research projects. In addition to research facilities in numerous countries international providers include Complete Genomics, a US company established in 2005 with the specific aim of providing a comprehensive human DNA service for pharmaceutical and academic research, and BGI (formally Beijing Genomics Institute), the first citizen-managed, non-profit research institution in China with probably the largest sequencing capacity in the world. It is unclear what effect these integrated service providers will have on the future of human whole genome sequencing, and currently most research and diagnostic laboratories both produce and analyse their own data.
Multi-gene diagnostic panels (Morgan et al. 2010)
However, numerous barriers to clinical translation still exist, including: validation of the technology; standardisation of the analysis pipeline; integration of information from the numerous databases of genomic variation; building a robust evidence base to allow clinical interpretation of novel variants; developing a service delivery infrastructure that can capitalise upon the high-throughput advantages of new sequencing technologies; providing an appropriately skilled health care workforce to deal with genomic medicine; and addressing the numerous ethical, legal and social implications of sequencing, storing and accessing whole genomes. These issues will need to be addressed before human whole genome resequencing can be used routinely in the clinic.
An open platform based on sequencing-by-ligation has also been developed by George Church/Dover Systems, known as the Polonator (Shendure et al. 2005), but is not described further as it had a substantially smaller impact than the other technologies.
The process of alignment is substantially harder for de novo genome assembly, as there is no reference sequence against which DNA reads can be compared and mapped. Therefore specialist genome assembly methodologies and software have been developed, which are no longer directly relevant to human genome sequencing.
The content of this paper forms part of a PHG Foundation Report on the implications of whole genome sequencing for health (downloadable at www.phgfoundation.org). The authors wish to thank the project team at the PHG Foundation for their invaluable feedback on this work. The PHG Foundation is the working name of the Foundation for Genomics and Population Health, a charitable company registered in England and Wales, charity No. 1118664 company No. 5823194.