Odyssey: a Semi-Automated Pipeline for Phasing, Imputation, and Analysis of Genome-Wide Genetic Data Ryan J
Total Page:16
File Type:pdf, Size:1020Kb
Eller et al. BMC Bioinformatics (2019) 20:364 https://doi.org/10.1186/s12859-019-2964-5 METHODOLOGY ARTICLE Open Access Odyssey: a semi-automated pipeline for phasing, imputation, and analysis of genome-wide genetic data Ryan J. Eller1* , Sarath C. Janga2,3 and Susan Walsh1 Abstract Background: Genome imputation, admixture resolution and genome-wide association analyses are timely and computationally intensive processes with many composite and requisite steps. Analysis time increases further when building and installing the run programs required for these analyses. For scientists that may not be as versed in programing language, but want to perform these operations hands on, there is a lengthy learning curve to utilize the vast number of programs available for these analyses. Results: In an effort to streamline the entire process with easy-to-use steps for scientists working with big data, the Odyssey pipeline was developed. Odyssey is a simplified, efficient, semi-automated genome-wide imputation and analysis pipeline, which prepares raw genetic data, performs pre-imputation quality control, phasing, imputation, post-imputation quality control, population stratification analysis, and genome-wide association with statistical data analysis, including result visualization. Odyssey is a pipeline that integrates programs such as PLINK, SHAPEIT, Eagle, IMPUTE, Minimac, and several R packages, to create a seamless, easy-to-use, and modular workflow controlled via a single user-friendly configuration file. Odyssey was built with compatibility in mind, and thus utilizes the Singularity container solution, which can be run on Linux, MacOS, and Windows platforms. It is also easily scalable from a simple desktop to a High-Performance System (HPS). Conclusion: Odyssey facilitates efficient and fast genome-wide association analysis automation and can go from raw genetic data to genome: phenome association visualization and analyses results in 3–8h on average, depending on the input data, choice of programs within the pipeline and available computer resources. Odyssey was built to be flexible, portable, compatible, scalable, and easy to setup. Biologists less familiar with programing can now work hands on with their own big data using this easy-to-use pipeline. Keywords: Imputation, Phasing, Pipeline, Genome-wide-association study, Admixture, Odyssey Background datasets so they can be combined with others that Genome-wide association studies (GWAS) have grown are genotyped with different arrays [4]. Over the last in popularity thanks to the increased availability of few decades several reference datasets have become genome-wide data and sequence information. GWAS, available to use for imputation, such as the while successful at identifying candidate variants, has international HapMap Project in 2003 [5], the 1000 also been aided by imputation methods [1–3]thatin- Genome Project in 2015 [6], the Haplotype Reference crease coverage and allow for increased sensitivity. Consortium (HRC) in 2016 [7], and the more recently Imputation is often performed to fill in the genomic announced All of US Research Program currently be- gaps, increase statistical power, and to “standardize” ing conducted by the NIH in 2018 [8]. Increasing the number and diversity of reference panels allow for in- * Correspondence: [email protected] creased flexibility in how imputation is performed for 1Department of Biology, Indiana University-Purdue University Indianapolis, 723 W. Michigan Street, Indianapolis, IN, USA a particular sample set. Van Rheenen et al., 2016 has Full list of author information is available at the end of the article shown that custom reference panels that combine an © The Author(s). 2019 Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated. Eller et al. BMC Bioinformatics (2019) 20:364 Page 2 of 8 existing reference panel with sequence data collected integrity, but most of the data cleanup is left to the user. from a subset of individuals within their analysis co- While this data prep must be done offline, most of these hort may also increase imputation accuracy [9]. platforms provide a thorough walk-through on how to Current imputation options include the popular free- implement these steps. It is also worth noting that the to-use imputation servers such as the Michigan Imput- offline docker solution, which is similar to the Michigan ation Server (https://imputationserver.sph.umich.edu/) Imputation Server, provides guidance with quality con- and the Sanger Imputation Server (https://imputation. trol but like its online counterpart does not perform it sanger.ac.uk/), which provide an online solution to im- automatically. Thus, the responsibility of proper data putation. There are also offline solutions such as the preparation falls largely on the user and their ability to Michigan Imputation Server Docker [10], and imput- find acceptable imputation quality control thresholds, ation packages such as Python’s Genipe [11]. The such as those found in Manolio et al., 2007 [13]. In strengths of online solutions are that they normally re- addition to cleaning the data, the user is expected to quire no setup and are easy to use. However, a major provide data in a compatible format for the imputation drawback is that they require data to be sent off-site (al- workflow, which is normally a VCF (for the Sanger and beit via a secure SFTP or SCP connection in the cases of Michigan Imputation Servers and docker image) or a the Sanger and Michigan Imputation Servers), which PLINK .bed/.bim/.fam (for the Genipe package). While may or may not be possible for a researcher due to eth- most commercial genome array software, such as Illumi- ical or legal constraints. As with most online servers na’s Bead Studio or Affymetrix’s Power Tools, perform users may need to sit in a queue before their job is run, these conversions, the user must still rectify any genome and users are often restricted by analysis options, such array compatibility issues, such as remapping incompat- as the choice of phasing/imputation programs as well as ible sample data to the same genome build used by the the reference panels the sites support. At the time of imputation reference panel. writing, both the Sanger and Michigan Imputation Genome-wide association studies that use probabilistic Servers support three main panels: the HRC, 1000 Ge- imputation data or dosage data require a considerable nomes (Phase 3), and the CAAPA African American number of programs for the analysis to be run. Cur- Panel [12]. Sanger also provides access to the UK10K rently, only PLINK 2.0 [14]andSNPTEST[3, 4, 15] are dataset, which is currently unsupported by the Michigan capable of performing such analyses, short of writing a Imputation Server. It is important to note that apart custom script. It is important to note that additional from Sanger’s option of imputing with a combined programs accept dosage data, but subsequently hard-call UK10K dataset and 1000 Genomes Reference panel, the (i.e. probabilistically round) genotypes that are used in online solutions do not give much flexibility if the user downstream analysis. While analyzing hard-called data is wishes to combine several reference panels or integrate a valid strategy, data is ultimately being altered and may collected data into a custom reference set to enrich the alter study outcomes. In addition to having few analysis imputation. Users must then opt for offline solutions, programs that can analyze dosage data, it is often cum- such as the Michigan Imputation Server Docker image bersome and time consuming to input data into these and Python Packages such as Genipe, that do not re- programs. Dosage data often needs to be concatenated quire data to be sent offsite and provide considerably or merged (as imputation is normally done in segments) more flexibility in the imputation analysis as they allow and then converted into a format accepted by PLINK 2.0 custom reference datasets to be installed. However, an or SNPTEST in a manner that does not alter the data as issue of using offline solutions is that they need to be previously described. Further complicating the matter of configured by the user, which may not be straightfor- compatibility is the continual evolution of dosage data ward due to the many programs these pipelines require formats, such as Oxford’s .bgen and PLINK’s .pgen, since as well as their interconnected library dependencies. programs may not accept both file formats or even cer- While imputation is the main goal of all these plat- tain version iterations of a particular filetype. Due to the forms, it is imperative that data must be formatted prop- aforementioned issues, the transition between imput- erly before submitting it through phasing and ation and data analysis is the largest hurdle to analyzing imputation. Furthermore, quality control measures imputation data and is probably an area in the largest should be enacted to achieve the highest possible imput- need of improvement in imputation analysis workflows. ation accuracy. While a researcher should know what Admixture considerations while performing a GWAS quality control measures they would like to use for im- lie primarily in performing a separate stratified analysis putation, there are inconsistencies between different using ancestry informative programs such as the model- programs and their default settings. The established on- based Admixture [16] or PCA-based Eigensoft [17] pro- line imputation servers perform some filters for minor grams, and therefore require knowledge of these add- allele frequency, duplicated variant detection, and VCF itional programs to account for population stratification Eller et al.