skip to main content


Title: Population Genomics Training for the Next Generation of Conservation Geneticists: ConGen 2018 Workshop
Abstract The increasing availability and complexity of next-generation sequencing (NGS) data sets make ongoing training an essential component of conservation and population genetics research. A workshop entitled “ConGen 2018” was recently held to train researchers in conceptual and practical aspects of NGS data production and analysis for conservation and ecological applications. Sixteen instructors provided helpful lectures, discussions, and hands-on exercises regarding how to plan, produce, and analyze data for many important research questions. Lecture topics ranged from understanding probabilistic (e.g., Bayesian) genotype calling to the detection of local adaptation signatures from genomic, transcriptomic, and epigenomic data. We report on progress in addressing central questions of conservation genomics, advances in NGS data analysis, the potential for genomic tools to assess adaptive capacity, and strategies for training the next generation of conservation genomicists.  more » « less
Award ID(s):
1655809 1639014
NSF-PAR ID:
10185993
Author(s) / Creator(s):
; ; ; ; ; ; ; ; ; ; ;
Date Published:
Journal Name:
Journal of Heredity
Volume:
111
Issue:
2
ISSN:
0022-1503
Page Range / eLocation ID:
227 to 236
Format(s):
Medium: X
Sponsoring Org:
National Science Foundation
More Like this
  1. Abstract

    New computational methods and next‐generation sequencing (NGS) approaches have enabled the use of thousands or hundreds of thousands of genetic markers to address previously intractable questions. The methods and massive marker sets present both new data analysis challenges and opportunities to visualize, understand, and apply population and conservation genomic data in novel ways. The large scale and complexity of NGS data also increases the expertise and effort required to thoroughly and thoughtfully analyze and interpret data. To aid in this endeavor, a recent workshop entitled “Population Genomic Data Analysis,” also known as “ConGen 2017,” was held at the University of Montana. The ConGen workshop brought 15 instructors together with knowledge in a wide range of topics including NGS data filtering, genome assembly, genomic monitoring of effective population size, migration modeling, detecting adaptive genomic variation, genomewide association analysis, inbreeding depression, and landscape genomics. Here, we summarize the major themes of the workshop and the important take‐home points that were offered to students throughout. We emphasize increasing participation by women in population and conservation genomics as a vital step for the advancement of science. Some important themes that emerged during the workshop included the need for data visualization and its importance in finding problematic data, the effects of data filtering choices on downstream population genomic analyses, the increasing availability of whole‐genome sequencing, and the new challenges it presents. Our goal here is to help motivate and educate a worldwide audience to improve population genomic data analysis and interpretation, and thereby advance the contribution of genomics to molecular ecology, evolutionary biology, and especially to the conservation of biodiversity.

     
    more » « less
  2. Abstract

    The development of next-generation sequencing (NGS) enabled a shift from array-based genotyping to directly sequencing genomic libraries for high-throughput genotyping. Even though whole-genome sequencing was initially too costly for routine analysis in large populations such as breeding or genetic studies, continued advancements in genome sequencing and bioinformatics have provided the opportunity to capitalize on whole-genome information. As new sequencing platforms can routinely provide high-quality sequencing data for sufficient genome coverage to genotype various breeding populations, a limitation comes in the time and cost of library construction when multiplexing a large number of samples. Here we describe a high-throughput whole-genome skim-sequencing (skim-seq) approach that can be utilized for a broad range of genotyping and genomic characterization. Using optimized low-volume Illumina Nextera chemistry, we developed a skim-seq method and combined up to 960 samples in one multiplex library using dual index barcoding. With the dual-index barcoding, the number of samples for multiplexing can be adjusted depending on the amount of data required, and could be extended to 3,072 samples or more. Panels of doubled haploid wheat lines (Triticum aestivum, CDC Stanley x CDC Landmark), wheat-barley (T.aestivumxHordeum vulgare) and wheat-wheatgrass (Triticum durum x Thinopyrum intermedium) introgression lines as well as known monosomic wheat stocks were genotyped using the skim-seq approach. Bioinformatics pipelines were developed for various applications where sequencing coverage ranged from 1 × down to 0.01 × per sample. Using reference genomes, we detected chromosome dosage, identified aneuploidy, and karyotyped introgression lines from the skim-seq data. Leveraging the recent advancements in genome sequencing, skim-seq provides an effective and low-cost tool for routine genotyping and genetic analysis, which can track and identify introgressions and genomic regions of interest in genetics research and applied breeding programs.

     
    more » « less
  3. Abstract

    Next Generation Sequencing (NGS) has become an important tool in the biological sciences and has a growing number of applications across medical fields. Currently, few undergraduate programs provide training in the design and implementation of NGS applications. Here, we describe an inquiry‐based laboratory exercise for a college‐level molecular biology laboratory course that uses real‐time MinION deep sequencing and bioinformatics to investigate characteristic genetic variants found in cancer cell‐lines. The overall goal for students was to identify non‐small cell lung cancer (NSCLC) cell‐lines based on their unique genomic profiles. The units described in this laboratory highlight core principles in multiplex PCR primer design, real‐time deep sequencing, and bioinformatics analysis for genetic variants. We found that the MinION device is an appropriate, feasible tool that provides a comprehensive, hands‐on NGS experience for undergraduates. Student evaluations demonstrated increased confidence in using molecular techniques and enhanced understanding of NGS concepts. Overall, this exercise provides a pedagogical tool for incorporating NGS approaches in the teaching laboratory as way of enhancing students' comprehension of genomic sequence analysis. Further, this NGS lab module can easily be added to a variety of lab‐based courses to help undergraduate students learn current DNA sequencing methods with limited effort and cost.

     
    more » « less
  4. null (Ed.)
    Abstract Background Significant progress has been made in advancing and standardizing tools for human genomic and biomedical research. Yet, the field of next-generation sequencing (NGS) analysis for microorganisms (including multiple pathogens) remains fragmented, lacks accessible and reusable tools, is hindered by local computational resource limitations, and does not offer widely accepted standards. One such “problem areas” is the analysis of Transposon Insertion Sequencing (TIS) data. TIS allows probing of almost the entire genome of a microorganism by introducing random insertions of transposon-derived constructs. The impact of the insertions on the survival and growth under specific conditions provides precise information about genes affecting specific phenotypic characteristics. A wide array of tools has been developed to analyze TIS data. Among the variety of options available, it is often difficult to identify which one can provide a reliable and reproducible analysis. Results Here we sought to understand the challenges and propose reliable practices for the analysis of TIS experiments. Using data from two recent TIS studies, we have developed a series of workflows that include multiple tools for data de-multiplexing, promoter sequence identification, transposon flank alignment, and read count repartition across the genome. Particular attention was paid to quality control procedures, such as determining the optimal tool parameters for the analysis and removal of contamination. Conclusions Our work provides an assessment of the currently available tools for TIS data analysis. It offers ready to use workflows that can be invoked by anyone in the world using our public Galaxy platform ( https://usegalaxy.org ). To lower the entry barriers, we have also developed interactive tutorials explaining details of TIS data analysis procedures at https://bit.ly/gxy-tis . 
    more » « less
  5. Abstract

    Genome editing technologies have revolutionized genetic studies in the life sciences community in recent years. The application of these technologies allows researchers to conveniently generate mutations in almost any gene of interest. This is very useful for species such as maize that have complex genomes and lack comprehensive mutant collections. With the improvement of genome editing tools and transformation methods, these technologies are also widely used to assist breeding research and implementation in maize. However, the detection and genotyping of genomic edits rely on low‐throughput, high‐cost methods, such as traditional agarose gel electrophoresis and Sanger sequencing. This article describes a method to barcode the target regions of genomic edits from many individuals by low‐cost polymerase chain reaction (PCR) amplification. It also employs next‐generation sequencing (NGS) to genotype the genome‐edited plants at high throughput and low cost. This protocol can be used for initial screening of genomic edits as well as derived population genotyping on a small or large scale, at high efficiency and low cost. © 2021 Wiley Periodicals LLC.

    Basic Protocol 1: A fast genomic DNA preparation method from genome edited plants

    Basic Protocol 2: Barcoding the amplicons of edited regions from each individual by two rounds of PCR

    Basic Protocol 3: Bioinformatics analysis

     
    more » « less