Several very different technologies constitute next generation sequencing (NGS), each of which has its own set of characteristics (Table 1). NGS rapidly generates huge amounts of sequence data in a very cost-effective way, and molecular ecologists are now starting to take advantage of this democratization of sequencing and embracing the discipline of ‘ecological genomics’ (Gilad et al., 2009). By shifting the realms of genomics from laboratory-based studies of model-species towards studies of natural populations of ecologically well-characterized organisms, researchers can now start to address important ecological and evolutionary questions on a scale and precision that was unrealistic only a few years ago. We will not go into any technical details of NGS because this has already been extensively reviewed elsewhere (Hudson, 2008; Morozova and Marra, 2008; Shendure and Ji, 2008).

Table 1 Currently available NGS technologies and their characteristics

NGS can follow either a genomic or a transcriptomic route (Figure 1). In the latter, complementary DNA (cDNA) (see Glossary) is produced from the mRNA of a specific tissue or life stage. By this approach, data will be obtained on nucleotide variation as well as transcriptome (see Glossary) characteristics and gene expression levels. NGS allows for nucleotide variation profiling and large-scale discovery of genetic markers, which in turn will aid in the pursuit of the genetic basis of ecologically important phenotypic variation through quantitative trait loci (QTL) mapping (see Glossary) or genome-wide association studies (GWAS). Studies of population history and demography, genetic structure and inference of relatedness will also be greatly improved. Genome-wide scans (for example, outlier analysis; see Glossary) and comparative genomics using NGS data will provide better chances of identifying loci under selection. Finally studies of gene regulation, DNA–protein interactions and epigenetics are also facilitated by the use of NGS, enabling molecular ecologists to venture into these areas of growing interest.

Figure 1
figure 1

Scheme showing the workflow from sample to applications of NGS in molecular ecology. We considered three different sources of genetic variation, shown in the central circle of the diagram and indicated by different background shades: genome (light grey), transcriptome (white) and epigenome (grey). Genomic DNA (gDNA), mitochondrial or chloroplast DNA (mtDNA, cpDNA), vectors (for example, BAC), mRNA, non-coding small RNAs or target regions of the genome (targeted DNA) are the samples regarded as starting material (marked with thick circles). Different steps that can be performed before or during sequencing are shown within the grey dotted square representing NGS. gDNA samples, for example, can be used to produce reduced representation libraries (RRLs) and these can be either sequenced directly or used to generate S-RAD markers (sequenced restriction-site associated DNA). Targeted DNA can be generated in several different ways: PCR through amplicon sequencing, microdroplet PCR with RainStorm (RainDance Technologies, Lexington, MA, USA) or through sequence capture (NimbleGen, Roche; SureSelect, Agilent Technologies). mRNA can be used in a deep-serial analysis of gene expression (SAGE) approach or RNA-seq can be performed. gDNA can be also used in epigenetic studies of DNA–protein interactions through ChIP-seq or for studying methylation patterns through BS-seq (high-throughput bisulfite sequencing). The main applications (placed outside of the grey dotted ‘NGS’ square) of NGS are gene regulation, expression, transcriptome characterization, development of molecular markers (SNPs, microsatellites, InDels), nucleotide profiling and genome assembly. Each one of these main applications can then be divided further into many interesting areas of molecular ecology (kinship analysis, QTL mapping, and so on). An exceptional case is the use of S-RAD markers that can be used directly in mapping studies without the need of genotyping new developed molecular markers.

Genomic surveys in non-model species are much aided if there are genomic resources available for a related species (referred to as genomic reference species; see Glossary), for example, for assembly and functional annotation purposes (Wheat, 2010). With a growing number of species with sequenced genomes, this approach will be feasible for many non-model species. Such ‘genome-enabled taxa’ include a large number of species with ecological and/or conservation relevance (Kohn et al., 2006). The increased number of whole genome sequencing (WGS) projects means that more and more ecological model organisms, for example, Daphnia (Eads et al., 2007), Mimulus (Wu et al., 2007) and sticklebacks (Gasterosteus; Hohenlohe et al., 2010) are becoming important genetic models (Mitchell-Olds et al., 2008). At the same time, more molecular ecology studies are focusing on natural variation and adaptation in classical genetic model species, or close relatives of these, like Drosophila (Nolte and Schlötterer, 2008), Mus (Teeter et al., 2008) and Arabidopsis (Metcalf and Mitchell-Olds, 2009), thus closing the gap between model and non-model organisms from this end as well. In the following paragraphs, we will outline various applications of NGS that will aid molecular ecologists and evolutionary biologists in their research. We will briefly summarize recent studies in non-model organisms (Table 2) that have taken advantage of NGS to answer important ecological and evolutionary questions and also briefly discuss various future prospects of this ‘genomic revolution’ in molecular ecology.

Table 2 Some early case studies for various applications of NGS in non-model organisms

Transcriptome characterization

Currently the most common application of NGS in non-model species is transcriptome characterization (see Table 2). By this we mean generally describing what genes are expressed in a certain tissue, life stage or organism as well as functional characterization of these. Here, cDNA is synthesized by reverse transcription of mRNA and then sequenced. The first study describing the transcriptome in a non-model species through NGS was performed on the wasp Polistes metricus (Toth et al., 2007). Here, genome information from the related honey bee was used as template (see Glossary) for mapping the reads (see Glossary) from 454 sequencing and for downstream analysis. In the Glanville fritillary butterfly (Melitaea cinxia), however, the transcriptome was assembled de novo (see Glossary) without the help of a closely related reference genome (Vera et al., 2008). The fritillary butterfly is a text-book example of a species with complex meta-population dynamics (Saccheri et al., 1998), and the aim of the genomics approaches recently employed in this system is to understand the genetics behind the variation in dispersal and colonization abilities seen between individuals. Since these first ground-breaking studies, sequencing and successful assembly of the transcriptome via 454 sequencing has been performed in a number of non-model organisms (Table 2). The first (to the best of our knowledge) study to use Illumina/Solexa sequencing data in a non-model species used a combination of de novo assembly and genomic reference species mapped assembly to study the transcriptome of the polyploid plant Pachycladon enysii (Collins et al., 2008). However, since then, complete de novo assembly of Illumina/Solexa data has also been accomplished (Birol et al., 2009).

Most studies characterizing transcriptomes so far have been very descriptive by nature, but they provide an important starting point and a valuable resource for further analysis and ecological applications (Ellegren, 2008). For example, these sequences may be used as an assembly template (reference sequence) for further in-depth transcriptome re-sequencing and surveys of genetic variation. They may also be used to develop molecular markers, create targeted sequencing assays or to construct microarrays (see Glossary) for gene expression profiling, and to study alternative splicing (Harr and Turner, 2010), a phenomenon likely to be involved in processes of adaptation and speciation.

After the novel transcriptome has been annotated using a genomic reference species or publicly available sequence databases (for example, Genbank, Ensembl and UniProt), it can be used as a starting point for more detailed functional characterization, such as annotation using gene ontology (see Glossary) databases. As an example of this, (Dassanayake et al., 2009) used transcriptome characterization of two species of mangrove to investigate convergent evolution of gene expression. Shared transcriptomic profiles between species may of course be a result of a common evolutionary origin or a joint distribution (Nuzhdin et al., 2004). However, in the case of the two mangrove species investigated by Dassanayake et al., this is not likely to explain the pattern seen. These two mangroves are not evolutionary related to each other but belong to very different lineages (one of the species is more related to Arabidopsis and the other more related to Populus than they are to each other). They also inhabit distinctly different areas, one being neotropical and the other Indo-West Pacific in the distribution. Despite of these differences between the two mangroves studied, they shared many functional characteristics of their transcriptomes (but differed substantially from Arabidopsis and Populus), probably resulting from parallel adaptation to a similar environment. Researchers might also be interested in functional comparisons between different sexes, life stages or tissues within the same species. For example, studying which transcripts are tissue specific or if certain pathways are overrepresented in a specific development stage can have important evolutionary implications. As a result of the cDNA library construction methods and the currently available read lengths, there is a bias of the transcripts towards the 3′-end and full-length transcripts are difficult to reach. This is currently a problem when studying gene structure or when trying to sequence a complete gene (for example, candidate genes).

For many transcriptome characterization studies, it is preferable to have as broad a representation of the transcriptome as possible. This can be accomplished by pooling of several tissues, life stages and/or individuals when producing the cDNA library (Hahn et al., 2009). A complementary way of increasing the breadth of the transcriptome coverage is normalization of the cDNA library (Figure 2a) and the depletion of ribosomal RNA (Figure 2b). The effect of this is to reduce the representation of very common transcripts (Zhulidov et al., 2004). Although gene discovery may still only be marginally more efficient than for non-normalized libraries (Hale et al., 2009), normalization will increase the depth of coverage for most transcripts, which is very valuable for nucleotide variation profiling and single-nucleotide polymorphism (SNP) discovery. However, normalization procedure will introduce biases in the relative representation of different transcripts, making estimates of gene expression and allele frequencies from such data less reliable. Nevertheless, normalized cDNA libraries have still been useful in some cases to this end (for example, Schwarz et al., 2009; Ekblom et al., 2010). When planning the experiment it is also valuable to note that there is a trade off between the cost of normalization and the cost of sequencing (Wall et al., 2009). As sequencing technology keeps improving and the quantity of sequence yield increases, cDNA normalization may soon be superfluous for transcriptome characterization, as a single sequencing run will be able to cover the complete transcriptome of the sampled tissue, regardless of the distribution of transcript abundances.

Figure 2
figure 2

cDNA library normalization and ribosomal RNA (rRNA) depletion from total RNA. (a) Agarose gel showing the same double-strand cDNA sample before and after normalization. Normalization decreases the prevalence of highly abundant transcripts (seen as distinctive bands in the un-normalized sample) and equalizes mRNA concentrations in the cDNA library. It will increase the number of genes covered, but unless the sequencing quantity is high, fewer genes will be fully covered. Normalization consequently increases the coverage of most of the sequenced transcript. There is also an increase in the gene discovery rate in a normalized cDNA library, enhancing the identification and analysis of rare transcripts (Cheung et al., 2006). Several normalization methods are reviewed in Bogdanova et al. (2009) but duplex-specific nuclease (DSN) normalization (Zhulidov et al., 2004) has been widely used in recent years. (b) Analysis of RNA with an Agilent 2100 Bioanalyzer (Agilent Technologies) using a pooled RNA sample and the same sample after ribosomal RNA (rRNA) depletion with RiboMinus RNA-seq kit (Invitrogen). Concentrations are showed in fluorescence units (FU) and size in basepairs (nucleotides, nt). Notice the dramatic change of scale of the y axis, which is due to the reduced amount of 18s rRNA in the depleted sample.

Gene expression profiling

In gene expression profiling, the aim is not only to characterize what genes are expressed but also to investigate the specific level (absolute or relative) of gene expression. Traditionally this has been accomplished using microarrays (Kammenga et al., 2007), and has thus mainly been restricted to model species with previous genome information. However, as mentioned above, microarrays can now be constructed for non-model species using data from NGS transcriptome profiling (Garcia-Reyero et al., 2008), or used in a cross-specific way if developed in related genome reference species (Bar-Or et al., 2007). Development of microarrays through NGS considerably reduces both the cost and the effort involved.

An attractive alternative to microarrays for gene expression studies in non-model organisms is sequencing-based expression profiling, or digital transcriptomics (Murray et al., 2007). The rationale behind this approach is that the representation of specific sequences derived from deep cDNA sequencing is proportional to the amount of RNA from the gene in the original sample (t Hoen et al., 2008). Basically, there are two different versions of this kind of analysis. Either more or less random parts of whole transcripts are sequenced directly (RNA-Seq; see Glossary), or specific parts of transcripts are cut out using restriction enzymes and then sequenced. Restriction enzyme-based methods, for example deep serial analysis of gene expression (see Glossary) (Nielsen et al., 2006) always outputs the same sequence tag for a given transcript, facilitating data analysis. The utility of this approach for surveying gene expression in a non-model organism was recently verified in a study of drought stress responses of chick pea (Cicer arietinum) roots (Molina et al., 2008). The main drawback is that the tags need to be mapped to a reference sequence. In contrast, RNA-seq methods produce novel cDNA sequences across the whole range (or more random parts) of the transcript (Nagalakshmi et al., 2008; Wilhelm and Landry, 2009; Wang et al., 2009b). They have the benefit of determining gene expression patterns and characterizing the transcriptome at the same time. Thus, this approach also enables gene characterization, molecular marker finding and detection of splice variants among other applications (Simon et al., 2009).

Digital transcriptomic approaches typically do not suffer from high background noise and cross-hybridizations as are common for microarrays. Furthermore these approaches are more efficient at detecting very rare transcripts and variation in highly expressed genes (because of higher resolution) and the analyses therefore have a greater dynamic range. Another advantage of digital transcriptomics over microarrays is the ability to detect expression levels in previously unknown genes (Nielsen et al., 2006). In a recent study, there was a very high correlation between the number of 454 reads mapping to a gene and microarray determined gene expression of it (Kristiansson et al., 2009), thus validating the RNA-Seq methodology for non-model organisms. The correlation in this study was surprisingly strong considering that the cDNA library used for the digital transcriptomic analysis was normalized. Gene expression profiling using an RNA-Seq approach was also performed in a study of blight resistance in different chestnut species (Castanea spp.). Several different genes were found to be differentially expressed in infected versus healthy tissues indicating a function in pathogen response (Barakat et al., 2009). This study shows the strength of genomic approaches for addressing questions related to pathogen resistance and host–parasite co-evolution. Another study investigated the genomics of speciation by comparing expression profiles between samples from a hybrid zone of two subspecies of the crow (Corvus corone) with samples from a pure population of one of these subspecies (Wolf et al., 2010). The investigators concluded that there is strong divergence between these in terms of gene expression levels, although there seem to be very little genetic sequence divergence in this system, and suggest that variation in expression may prove to be an important factor in the early stages of speciation processes.

NGS allows for gene expression studies in species without previous genomic resources, thus RNA-Seq is, for example, a very promising application for the study of adaptation. Determining which set of genes are differentially expressed because of adaptation to certain environmental conditions represents an important question in molecular ecology and represents a first step to better understand the genetics behind the regulatory mechanisms involved. Although RNA-Seq represents a favourable strategy in many studies of gene expression it becomes an expensive route when many different sets of conditions and/or populations need to be tested for gene expression (for example, gene expression for different temperature conditions). In these cases, a more economic strategy would be using transcriptome information from NGS to design a microarray and then use these microarrays to study gene expression. It is, however, likely that in the future individual sample-tagging approaches and increased efficiencies of RNA-Seq can overcome this downside.

Candidate gene finding

Information about specific genes of interest can be mined from NGS transcriptome or genome data of non-model organisms using coding nucleotide or protein sequence information from genomic reference species. For example, (Toth et al., 2007) focused on candidate genes for food provisioning and foraging behaviour in their study of a primitively eusocial wasp (Polistes metricus). By 454 sequencing of the transcriptome and aligning the reads to the honey bee genome, they were able to annotate these candidate genes and perform follow up real-time quantitative PCR experiments to study differential gene expression in different social casts of the wasp species. The results supported the hypothesis that provisioning behaviour is linked to the evolution of eusociality. In another early study, several hundred genes related to immunity and defence responses were identified in the tobacco hornworm (Manduca sexta). This species is a well-known model for insect physiological processes but before this 454 based sequencing study there were very limited genomic resources. This data thus represent an important stepping stone in understanding the functional basis of insect pathogen resistance (Zou et al., 2008).

The candidate gene approach has also been widely used in conservation genetics (Höglund, 2009; Primmer, 2009). With the goal of conserving functionally important genetic information, studies have used genetic structure in ecologically relevant loci to identify taxonomic units of conservation interest (Hedrick, 2004; Ekblom et al., 2007). NGS has great potential to open up this approach to conservation genetics to more species and include analyses of a larger number of potentially important genes. The eelpout (Zoarces viviparous) is a fish species commonly used in environmental monitoring and ecotoxicological studies. It recently had its transcriptome characterized using 454 sequencing and a number of biomarker genes for ecotoxicology were specifically identified (Kristiansson et al., 2009), thus providing an important tool for future studies of the genetic basis for physiological responses to pollutant exposures.

When performing transcriptome sequencing with the aim of detecting candidate genes it is important to make sure that RNA from the right tissues and/or life stages are used. Genes under positive selection (therefore likely to be genes of interest) are expressed in a more tissue-specific manner compared with evolutionary conserved genes (Zhang and Li, 2004; Ekblom et al., 2010) and could thus easily be missed if coverage is too low.

Whole genome sequencing (WGS)

A vast majority of studies using NGS are re-sequencing already fully described genomes (Wheeler et al., 2008). As this is not possible for non-model organisms, it will not be discussed further here. However, as was recently demonstrated by work on the giant panda (Ailuropoda melanoleuca), NGS may also be utilized for de novo WGS of large and complex genomes (Li et al., 2010). Although more genomes will be sequenced through NGS and released in the near future, the costs, expertise and infrastructure required for data collection, analysis and output handling for this kind of application are still beyond reach for most molecular ecology research groups. However, large research centres will be carrying out an impressive amount of WGS projects in the next few years. One example is the 1000 plant and animal reference genomes project (, which aims to sequence 1000 economical and scientific important species in 2 years.

An alternative approach to WGS in species without a characterized reference genome may be to use NGS to generate a large amount of sequence data, and to analyse this without attempting a full genome assembly. This approach was taken in a study of mammoth genomics, in which previously generated Sanger sequence data and the sequenced genome of the African Elephant (Loxodonta africana) was used in annotation and analysis of 454 sequence reads (Miller et al., 2008). Recently, a large amount of genome sequence data was produced by Illumina/Solexa sequencing of the great tit (Parus major). The investigators used this to assemble about 2.5% of the genome, after using a reduced representation library (see Glossary) strategy to increase the coverage of the sequenced fraction of the genome, which was then used for downstream SNP discovery (van Bers et al., 2010). NGS can also be used to sequence bits of genomes after cloning these into bacterial artificial chromosomes (BACs). However, the assembly of 454 sequencing data of eight different BACs from Atlantic salmon (Salmo salar) was found to be problematic (Quinn et al., 2008). The increase in the length of the sequences obtained with the current technologies in combination with the use of paired end sequencing will improve the quality of this type of assemblies in the future (DiGuistini et al., 2009; Rounsley et al., 2009). Targeted NGS of BACs will be a valuable resource for studying chromosomal regions involved in adaptation (Baxter et al., 2010).

Targeted sequencing

The NGS applications reviewed so far generate an impressive amount of sequence data, but in terms of population genetics (for example, variation between individuals and populations) the amount of information is limited. For such applications, it is more informative to use NGS to sequence a limited number of targeted loci. By decreasing the number of targets, the coverage is considerably increased, and consequently more valuable information for population analyses is obtained. Targets for sequencing can be obtained either using PCR or genetic capture techniques before sequencing (Mamanova et al., 2010). The targeted regions can represent individual independent loci (for example, exons) or a long stretch of genomic DNA. In order to apply either of these approaches, the sequence of the target regions needs to be known.

NGS of PCR products to target specific loci is generally referred to as amplicon sequencing (see Glossary) (Peng and Zhang, 2009). This approach has been very useful when addressing population genetic and evolutionary questions for large functional groups of genes in model-organisms like Drosophila (Obbard et al., 2009). By specifically tagging the PCR primers for each individual, this methodology can be used for high coverage genotyping of the loci of interest in a large number of samples (Binladen et al., 2007). Another possibility is pooling of individuals before the sequencing step. A case study with high ecological relevance on a non-model organism is the recent survey of major histocompatibility complex class II (see Glossary) variation in the bank vole (Myodes glareolus). Using only a fraction of a 454 sequencing run, the investigators were able to genotype 96 individuals for this complicated multi locus gene (Babik et al., 2009). As indicated by this study, great care needs to be taken to reduce problems with sequencing errors producing artificial alleles. But after dealing with these, it was shown that the NGS approach is able to detect alleles that are present at low frequency in the PCR product (and therefore could not be detected using conventional genotyping). Variation in these major histocompatibility complex loci was found to be related to prevalence of specific parasitic nematodes. These relationships were also population specific (Kloch et al., 2010), thus providing a potential genetic basis for local adaptation. One of the problems of amplicon sequencing is that PCR products need to be carefully standardized to the same concentrations to avoid overrepresentation of a certain locus or a certain population during the sequencing.

Another popular application of amplicon NGS is so called ‘barcoding’ studies. Here, a small variable part of the genome (usually from mitochondria or chloroplasts) is amplified from unidentified or complex samples, and sequenced using NGS. The sequence information is then used to identify the species present in the sample (Valentini et al., 2009). Studies using this approach have, for example, investigated the diet of a variety of animals by sequencing faecal samples (Deagle et al., 2009), and characterized the extinct mammalian fauna using ancient DNA from frozen tundra sediment (Haile et al., 2009). NGS barcoding approaches have also been extensively utilized to study meta-genomics of micro organism communities (Alvarez et al., 2009; Buée et al., 2009; Andersson et al., 2010).

Finally, sequence capture methods such as NimbleGen arrays (Roche) (Hodges et al., 2007; Okou et al., 2007) and the SureSelect platform (Agilent Technologies, Santa Clara, CA, USA) (Gnirke et al., 2009) are starting to become increasingly demanded by molecular ecologists. These capture techniques coupled with NGS generate high coverage sequence data from targeted DNA (for example, many independent fragments or a sequence of DNA of tens of kb) in several individuals or populations (that is, pool of individuals). These methodologies will allow molecular ecologists to study the sequence divergence between populations, morphs or species for hundreds of genes and gene families simultaneously, at a very reasonable cost and effort compared with previously available techniques. A drawback of the sequence capture approach is the need for a genome reference to mask repeated regions. Also, each individual or population will be hybridized independently, considerably increasing the cost of the methodology when many samples are studied. A future improvement of these approaches will be the use of tagged samples, wherein multiple individuals/populations can by hybridized simultaneously. These approaches are predicted to change the way we think about phylogeography, demography and conservation genetics, by massively increasing the number of loci studied. This improvement will require the development of new software tools to make the analyses feasible from a computational point of view.

Large-scale identification and development of molecular markers

One of the most important application of NGS within ecological and population genetics is the development of molecular markers on a large scale. NGS generally allows for cost-effective and rapid identification of hundreds of microsatellite loci and thousands of SNPs, even if only a fraction of a sequencing run is used. This will, for example, facilitate QTL mapping studies and will increase the quality of outlier- and structure analyses. Massively increasing the number of markers will enable researchers not only to get better precision in population genetic studies (Novembre et al., 2008), QTL and linkage disequilibrium (LD) mapping projects (Slate et al., 2009) and kinship assignments (Santure et al., 2010), but also to pursue topics such as historical demographic patterns, introgression and admixture (Jakobsson et al., 2008).

As demonstrated by several recent studies (see Table 2) molecular markers can be developed on a large scale, in almost any organism from either transcriptome or genome sequences. One advantage of developing markers from transcriptome sequences is that those will be associated with functional genes, thus increasing the interest of these for studies of adaptation (Imelfort et al., 2009). For example, expressed sequence tag (EST)-linked microsatellites typically occur in the untranslated regions (5′- and 3′-UTRs; Primmer, 2009). These loci are predicted to have a higher probability (compared with neutral genetic markers) of showing signatures of selection (Vasemagi et al., 2005), and they can also have a functional significance in regulating gene expression and function. Microsatellite loci can be detected within NGS data sets with programs like msatcommander (Faircloth, 2008) and MSatFinder (Thurston and Field, 2005). SNPs (located in the coding region or in the UTRs) can also be discovered using many different approaches (for an updated list of specific software available for various applications see: Both data from individually sequenced samples or samples of pooled individuals are useful to this end (Futschik and Schlötterer, 2010).

One obvious problem with SNP detection from NGS data is that sequencing errors will show a very similar signature to low-frequency SNP alleles. In most SNP discovery algorithms, this drawback is dealt with by including a minimum depth of sequencing at the place of interest to call a SNP as well as a minimum number of reads with the minor allele that is >1 (often 2 or 3). New theoretical work is also aiming to statistically distinguishing true polymorphisms from sequencing errors (see, for example, Lynch, 2009). GigaBayes (Hillier et al., 2008) and VarScan (Koboldt et al., 2009) represent examples of programs for SNP detection. GigaBayes calculates the probability that a polymorphism represent a true SNP or a sequencing error, for this calculation the program uses a Bayesian approach, taking into account the alignment depth, the base call in each sequence, the base composition in the region and the expected a priori polymorphism rate. With the advent of NGS, the use of SNPs in molecular ecology studies is predicted to expand dramatically. Once identified these markers can be typed in a large number of individuals using a wide variety of platforms (reviewed in Slate et al., 2009). InDel (insertion–deletion) markers can be detected with similar software as the SNPs and they will probably become a popular kind of molecular marker in future population genetic studies, as they are easy and cheap to identify and score (Väli et al., 2008).

Genomic DNA sequencing can also be used to generate genetic markers on a large scale. Sometimes low-coverage 454 genome sequencing is enough to identify several thousand microsatellite loci, like in the copperhead snake (Agkistrodon contortrix; Castoe et al., 2009). Microsatellites have also been identified from museum samples of extinct moa species (Dinornithiformes) using a similar approach (Allentoft et al., 2009). However, in order to increase the coverage of sequencing (and the efficiency of polymorphism detection), it may be very helpful to employ a genome complexity reduction technique (Santana et al., 2009). Reduced representation libraries are obtained through restriction digestion of the DNA followed by an electrophoresis of the digestion and a size selection step (Van Tassell et al., 2008). Another complexity reduction technique is the S-RAD markers (sequenced restriction-site-associated DNA; Baird et al., 2008) (see Glossary), which also represent a very promising tool for SNP discovery from DNA samples (Hohenlohe et al., 2010). S-RAD libraries are prepared by digesting genomic DNA with a restriction enzyme. Individually tagged adaptors are then ligated to the fragments and the samples are pooled. After this, physical shearing of the ligated DNA and a size selection are performed. The resulting library is sequenced through Illumina/Solexa paired-end sequencing and the tags are used to identify each sample after the sequencing. Many molecular ecology research groups are now investing their efforts in obtaining S-RAD markers in their study species. Library preparation is relatively easy and the reagents (for example, adaptors) can be used, a posteriori, for other organisms by the same research group. The data analysis involved might be the most difficult step, but once it is up and running in a research group it can be applied to almost any species.

Nucleotide variation profiling

Identifying the genes involved in ecologically important phenotypic variation is a major goal in ecological genomics (Feder and Mitchell-Olds, 2003). This has previously been accomplished by screening a large number of molecular markers (such as, SNPs or amplified fragment length polymorphisms; see Glossary) for outliers. As outlined above, such markers can now be developed on a large scale using NGS. Importantly, the deep coverage provided in pools of samples (genomic DNA or cDNA) by NGS can also be used for directly screening genomic variation, bypassing the SNP genotyping step. This kind of approach may thus provide a shortcut in studies such as those trying to integrate population genomic and quantitative genetic approaches (Stinchcombe and Hoekstra, 2007). Such methodology has proven very efficient for outlier type analyses (see below) but may also be applicable in other kinds of population genetic studies. Tagging of samples, which is necessary for investigating genetic variation on an individual level, will be facilitated by the use of a recently launched highly automated procedure (Lennon et al., 2010). The trade off between sequence depth of individual samples and number of samples sequenced also needs to be considered, as the calculations of allele frequencies and detection of low-frequency SNPs will be severely hampered if there is insufficient coverage of individual SNPs. It should also be noted that for SNP detection and estimation of population biology parameters, sequencing pools of individuals may be more effective than individual tagging of sequences (Futschik and Schlötterer, 2010).

A recent study used transcriptome 454 sequencing estimated SNP allele frequencies in two sympatric whitefish species (Coregonus spp.) to identify candidate markers for follow-up studies trying to determine the genetic basis of speciation and adaptation (Renaut et al., 2010). Until recently, such studies were performed through QTL mapping, genome scans and microarrays for gene expression. Renaut et al. were able to corroborate the success of the methodology because results from these three analyses, which were previously performed in the same populations, were concordant with the new NGS study. Another study using a similar method to study the genetic basis for ecological speciation and adaptation was performed on two different host races of the apple maggot fly, Rhagoletis pomonella (Schwarz et al., 2009). After performing SNP detection in their contigs, (see Glossary) they determined allele frequencies for each host race and those SNPs presenting significant differences were claimed as candidates for being involved in speciation. A slightly different approach was recently used to analyse the transcriptomes of two divergent ecotypes of the marine snail Littorina saxatilis undergoing ecological speciation (Galindo et al., 2010). Here, allele frequencies for both ecotypes were calculated and neutral simulations were used to detect outlier SNPs. Some of these SNPs were found in genes related with shell formation and energetic metabolism, both functions that are very important in the adaptation of these ecotypes. At this stage, the main drawbacks of this type of analysis is the variance in coverage between SNPs, especially the low coverage of many of the SNPs. Additional problems are pooling of the RNA samples because of overrepresentation of some samples over the others (for example, not accurate normalization of the concentrations) or of some transcripts (for example, high expression levels of one transcript in certain samples).

Another recent study tried to determine the genetic basis of local adaptation in Arabidopsis lyrata using whole genome resequencing on the Illumina/Solexa platform. Divergent nucleotide polymorphisms between soil types were detected from allele frequency differences, and a sliding window approach was used to identify outlier FST values between populations. Genes responsible for local adaptation to serpentine soils were detected after functional annotation and loci involved in heavy metal detoxification as well as calcium and magnesium transport pathways were overrepresented in markers with high divergence between soil types (Turner et al., 2010). This kind of analysis will represent the outlier analysis of the future. As mentioned in the previous section, S-RAD markers can also be used to determine the SNP allele frequencies without additional genotyping because samples can be tagged individually. This strategy is thus also suitable for outlier analysis and genetic mapping (Baird et al., 2008). A recent study using S-RAD markers in sticklebacks (Hohenlohe et al., 2010) has identified regions across the genome showing signatures of selection between oceanic and freshwater populations.

When generating genomic data from a number of different species, NGS may be used to study molecular evolution (such as dN/dS ratios; see Glossary) on a large scale, an approach that has hereto been restricted to model organisms. Such a comparative genomics analysis has recently been performed in a study of avian genome evolution using 454 sequence data from 10 different bird species (Künstner et al., 2010). Another interesting example studied two sympatric species of crater lake cichlids (Elmer et al., 2010). After 454 sequencing of the transcriptome, dN/dS ratios were inferred to detect genes with signatures of divergent selection. Follow-up studies on these genes might reveal new insights about ecological speciation and adaptive radiation.


Epigenetics is generally defined as the study of trait variation that does not come from changes in the DNA sequence, but rather involve other kinds of genetic modifications (such as patterns of DNA methylation and histone posttranslational modifications). Epigenetic modifications are primarily important because of their role in regulation of gene expression (Simon et al., 2009). Traditionally, these phenomena have mainly been studied because of their importance in cancer biology and in regulation of development. As many ecologically important traits are also likely to be influenced by epigenetic variation (Bossdorf et al., 2008), we include a short discussion of this topic here as a likely direction for future research on non-model organisms. Epigenetic modifications have mainly been studied on a large scale using microarray-based approaches. The one approach that has been applied so far in non-model organisms is methylation-dependent amplified fragment length polymorphism (Salmon et al., 2005). NGS has, however, opened up a large range of different high-throughput methods for epigenetic surveys and some of these are also likely to be applicable to ecologically relevant systems (Hurd and Nelson, 2009; Simon et al., 2009).

One major type of epigenetic modification that has been an important research focus for many years is the methylation of specific cytidine residues in the DNA. Generally, methylation has been studied using bisulphite treatment. Un-methylated cytidine residues in the DNA are converted to uracil after treatment with bisulphite, whereas methylated cytidine are protected from this conversion. Sequencing of these regions can then pinpoint the specific methylated nucleotides. In recent years, NGS has been utilized to characterize DNA methylation patterns genome-wide, using an application known as ultra-deep bisulphite DNA sequencing (BS-Seq) (Taylor et al., 2007; Cokus et al., 2008). Methylated fractions of genomes can also be sequenced using a combination of NGS and methyl-DNA immunoprecipitation (MeDIP). By using this approach, detailed methylation maps of the genome will become available (Pomraning et al., 2009), and by comparing these for different samples we will be able to address the importance of methylation for a large range of ecologically and evolutionary important questions, like the genetic architecture behind differential gene expression because of natural selection.

Another important determinant of transcription levels of genes is the structure of DNA packing, together with histone proteins, into nucleosomes (chromatin packing). This chromatin structure can also be studied using a high-throughput sequencing approach (Johnson et al., 2006). The histone proteins themselves, particularly the N-terminal tails, are also subject to a large number of posttranslational modifications, such as methylation and acetylation of specific amino acid residues (Kouzarides, 2007). These can be studied on a large scale using a ChIP-Seq approach with Illumina/Solexa sequencing (Barski et al., 2007). ChIP-Seq technology can also be used to study a large range of other DNA–protein interactions (Hurd and Nelson, 2009) including the identification of binding sites for transcription factors (Bhinge et al., 2007). To the best of our knowledge, neither of these interesting approaches has yet been applied to non-model organisms, but such studies are undoubtedly under way.

Some general considerations regarding planning and data analysis

NGS provides a very cost-effective way to generate large amounts of sequence data, but a single sequencing run is still a considerable expense for most small labs working with non-model organisms. It is thus crucial to carefully evaluate whether the methodology will be able to answer the relevant biological questions asked. It is also important to consider expenses, skills and infra-structure needed for sample preparation and data analysis. The sheer volume of sequences produced by these new technologies constitutes a genuine challenge for data storage and analysis. For many applications, such as de novo assembly and downstream analysis, computing power may also be an issue. Also, as there are currently major advances in algorithm and software development for NGS analysis, and it may be a good idea for molecular ecologists to liaise with a bioinformatics research group. An alternative is to outsource parts of the data analysis to the sequencing facility.

There are several important considerations during the analysis and data handling steps of NGS data. This is currently a very active field of research and software are being developed and refined to deal with the special problems faced in NGS analyses (Pepke et al., 2009). The first crucial steps in the data analyses are trimming (see Glossary) and assembly of the reads. Several software exist (both freely available open source software and commercial programs) to perform de novo assembly of sequence reads, some of which also perform the pre-assembly trimming ( If there is already a reference genome (or transcriptome) sequence available, the reads may be mapped directly onto this, without the need for previous assembly (Trapnell and Salzberg, 2009). Much care is needed to avoid or remove mis-assemblies as these will introduce bias to downstream analyses and applications. A completely novel strategy for handling of NGS data was recently introduced by Cannon et al. (2010). They analysed the occurrence of specific complex short read motifs produced by Illumina/Solexa sequencing directly, without any previous assembly of the reads. Using this approach, they were able to study comparative genomics of several non-model tree species. Arguably, this approach will greatly facilitate short read sequence analyses in non-model organisms.

It is important to note that NGS technologies will not always be the optimal way to generate the sequence data of interest to molecular ecologists. If a particular candidate gene or gene family is of interest, then specific methods targeting these may be more cost effective than NGS (but see also discussion above for ways to utilize NGS in targeting specific genes). There is no guarantee that these genes will be present in the NGS data set even if suitable tissues are sampled for RNA extraction. Also, even if traces of the gene of interest are found it may not be straight forward to utilize this to get the full-length sequence, as there are no clones to go back to and sequence separately, as in traditional sequencing approaches (Wheat, 2010). Also, for many population genetic surveys, more traditional methods like marker typing or amplified fragment length polymorphism genome scans may prove to be sufficient to answer the questions of interest (at a significantly smaller cost than a large NGS-based study). For small labs with limited funding it may prove efficient to use only a small fraction of an NGS run to generate a small amount of background data and use this for more targeted studies of genes or regions of interest.

Finally, an important consideration is which sequencing platform to use (see Table 1). This will depend on the questions addressed and on the available genomic resources in the study species or in related species. Choosing a single platform is always a trade off between read length and sequence output (number of bases). Short read platforms (Illumina/Solexa and ABI SOLiD) provide more data than Roche 454, but at lower length of individual reads (although these are now increasing the read length, thus enabling de novo assembly). Until now, the 454 has been the most extensive method used (Table 2) because of the advantage of the relatively long reads produced with this technology. However, 454 sequencing presents slightly higher rates of sequencing errors compared with competing technologies (see Table 1), especially in homopolymers (see Glossary) (Huse et al., 2007). Simulation experiments have shown that when there is no reference genome available, a combination of different NGS technologies may be most cost effective for transcriptome characterization (Wall et al., 2009).

Future prospects

Given the current rate of technological advances in this field it is difficult to speculate very far into the future. Nevertheless, we will try to outline some improvements that are likely to be waiting round the corner. One major advantage of NGS in ecological studies is the small amounts of genetic material needed for analysis, making these technologies suitable for analyses of endangered species wherein non-invasive sampling is needed and for studies of ancient DNA. Future developments are aiming at decreasing this amount even further, down to the possibility of conducting sequence-based expression analysis on the scale of individual cells (Simon et al., 2009). Also, as currently available technologies will continue to improve, these will be able to produce more and longer sequence reads, as well as decrease the sequence error rates. Together with improvements in data analysis algorithms, this will give higher quality assemblies of both genome and transcriptome data.

Within the applications of NGS that we have reviewed here, transcriptome characterization and gene expression profiling are most widely used until now (see Table 2). These approaches represent the first steps in more complex studies in which the availability of hundreds of genetic markers allows for QTL mapping or genome scans. In the near future, we expect that the number of studies taking advantage of genomic resources generated in this initial ‘boom’ of NGS will dominate over, for example, RNA-Seq. We anticipate the prices for sequencing and tagging might drop considerably enabling more population genomic type analyses with multiple samples and more precise and unbiased estimates of for example demographic parameters. Longer sequence reads will also improve downstream analysis applications because large haplotype blocks including several linked polymorphisms will become available. With time and with an increase in NGS studies, the goal of the research projects will not only be to detect signatures of selection but also to focus on the genetic architecture, the regulation and the history of selection. These topics are also directly linked to conservation biology. Until recently, conservation genetics projects generally studied a small number of neutral molecular markers to show the variation and/or heterozygosity of a population. NGS enables the shift to ‘conservation genomics’ (Ouborg et al., 2010) wherein hundreds of genes can be study simultaneously. Some of these may be involved in important phenotypic variation and this is relevant from the conservation point of view, because such variation may be important to maintain within the population.

The most important emerging NGS technique might be single-molecule sequencing (Gupta, 2008). Applying this so called ‘Next-next generation sequencing’ (or ‘Third generation sequencing’) will eliminate the need of amplification during the sequencing reaction. This will not only be more cost effective and remove sequencing errors produced in the amplification step, but will also reduce bias in detecting expression levels of individual genes or alleles. Several platforms are currently being developed for this kind of analysis. At the time of writing only the Helicos tSMS system (Harris et al., 2008) is commercially available but others (like the SMRT technology of Pacific Biosciences) are due to be launched later this year. Taking things even one step further is the recently introduced method of direct sequencing of RNA through a modification of the Helicos tSMS protocol (Ozsolak et al., 2009). This eliminates errors and biases in cDNA synthesis and thus provides a very accurate representation of gene expression levels.

As the cost of sequencing drops even further and the amount of data produced increases, there will be new demands for novel analysis methods and infrastructure. The future bottlenecks are more likely to be at the bioinformatics end rather than in producing the sequences (Schuster, 2008). Furthermore, there will likely be a large demand for molecular ecologists trying to make biological sense of all the gathered genomics data. It is probable that we will need radically new approaches for data storage and sharing as currently available databases might be unable to cope with the rapid generation of new sequencing data. We predict that keeping researchers from drowning in this data flood will be one of the major challenges in years to come. By bringing the realm of genomics into reach for studies of non-model organisms, NGS is currently radically changing our way of conducting genetic research, and it will continue to do so in the foreseeable future. As has been outlined in this review, this revolution is enhancing the scale that population genetic research can be conducted and bringing new objectives into reach of molecular ecologists.