Bismark Bisulfite Mapper – User Guide - V0.15.0
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Gene Discovery and Annotation Using LCM-454 Transcriptome Sequencing Scott J
Downloaded from genome.cshlp.org on September 23, 2021 - Published by Cold Spring Harbor Laboratory Press Methods Gene discovery and annotation using LCM-454 transcriptome sequencing Scott J. Emrich,1,2,6 W. Brad Barbazuk,3,6 Li Li,4 and Patrick S. Schnable1,4,5,7 1Bioinformatics and Computational Biology Graduate Program, Iowa State University, Ames, Iowa 50010, USA; 2Department of Electrical and Computer Engineering, Iowa State University, Ames, Iowa 50010, USA; 3Donald Danforth Plant Science Center, St. Louis, Missouri 63132, USA; 4Interdepartmental Plant Physiology Graduate Major and Department of Genetics, Development, and Cell Biology, Iowa State University, Ames, Iowa 50010, USA; 5Department of Agronomy and Center for Plant Genomics, Iowa State University, Ames, Iowa 50010, USA 454 DNA sequencing technology achieves significant throughput relative to traditional approaches. More than 261,000 ESTs were generated by 454 Life Sciences from cDNA isolated using laser capture microdissection (LCM) from the developmentally important shoot apical meristem (SAM) of maize (Zea mays L.). This single sequencing run annotated >25,000 maize genomic sequences and also captured ∼400 expressed transcripts for which homologous sequences have not yet been identified in other species. Approximately 70% of the ESTs generated in this study had not been captured during a previous EST project conducted using a cDNA library constructed from hand-dissected apex tissue that is highly enriched for SAMs. In addition, at least 30% of the 454-ESTs do not align to any of the ∼648,000 extant maize ESTs using conservative alignment criteria. These results indicate that the combination of LCM and the deep sequencing possible with 454 technology enriches for SAM transcripts not present in current EST collections. -
EMBL-EBI Powerpoint Presentation
Processing data from high-throughput sequencing experiments Simon Anders Use-cases for HTS · de-novo sequencing and assembly of small genomes · transcriptome analysis (RNA-Seq, sRNA-Seq, ...) • identifying transcripted regions • expression profiling · Resequencing to find genetic polymorphisms: • SNPs, micro-indels • CNVs · ChIP-Seq, nucleosome positions, etc. · DNA methylation studies (after bisulfite treatment) · environmental sampling (metagenomics) · reading bar codes Use cases for HTS: Bioinformatics challenges Established procedures may not be suitable. New algorithms are required for · assembly · alignment · statistical tests (counting statistics) · visualization · segmentation · ... Where does Bioconductor come in? Several steps: · Processing of the images and determining of the read sequencest • typically done by core facility with software from the manufacturer of the sequencing machine · Aligning the reads to a reference genome (or assembling the reads into a new genome) • Done with community-developed stand-alone tools. · Downstream statistical analyis. • Write your own scripts with the help of Bioconductor infrastructure. Solexa standard workflow SolexaPipeline · "Firecrest": Identifying clusters ⇨ typically 15..20 mio good clusters per lane · "Bustard": Base calling ⇨ sequence for each cluster, with Phred-like scores · "Eland": Aligning to reference Firecrest output Large tab-separated text files with one row per identified cluster, specifying · lane index and tile index · x and y coordinates of cluster on tile · for each -
The Encodedb Portal: Simplified Access to ENCODE Consortium Data Laura L
Downloaded from genome.cshlp.org on September 30, 2021 - Published by Cold Spring Harbor Laboratory Press Resource The ENCODEdb portal: Simplified access to ENCODE Consortium data Laura L. Elnitski, Prachi Shah, R. Travis Moreland, Lowell Umayam, Tyra G. Wolfsberg, and Andreas D. Baxevanis1 Genome Technology Branch, National Human Genome Research Institute, National Institutes of Health, Bethesda, Maryland 20892, USA The Encyclopedia of DNA Elements (ENCODE) project aims to identify and characterize all functional elements in a representative chromosomal sample comprising 1% of the human genome. Data generated by members of The ENCODE Project Consortium are housed in a number of public databases, such as the UCSC Genome Browser, NCBI’s Gene Expression Omnibus (GEO), and EBI’s ArrayExpress. As such, it is often difficult for biologists to gather all of the ENCODE data from a particular genomic region of interest and integrate them with relevant information found in other public databases. The ENCODEdb portal was developed to address this problem. ENCODEdb provides a unified, single point-of-access to data generated by the ENCODE Consortium, as well as to data from other source databases that lie within ENCODE regions; this provides the user a complete view of all known data in a particular region of interest. ENCODEdb Genomic Context searches allow for the retrieval of information on functional elements annotated within ENCODE regions, including mRNA, EST, and STS sequences; single nucleotide polymorphisms, and UniGene clusters. Information is also retrieved from GEO, OMIM, and major genome sequence browsers. ENCODEdb Consortium Data searches allow users to perform compound queries on array-based ENCODE data available both from GEO and from the UCSC Genome Browser. -
An Open-Sourced Bioinformatic Pipeline for the Processing of Next-Generation Sequencing Derived Nucleotide Reads
bioRxiv preprint doi: https://doi.org/10.1101/2020.04.20.050369; this version posted May 28, 2020. The copyright holder for this preprint (which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under aCC-BY 4.0 International license. An open-sourced bioinformatic pipeline for the processing of Next-Generation Sequencing derived nucleotide reads: Identification and authentication of ancient metagenomic DNA Thomas C. Collin1, *, Konstantina Drosou2, 3, Jeremiah Daniel O’Riordan4, Tengiz Meshveliani5, Ron Pinhasi6, and Robin N. M. Feeney1 1School of Medicine, University College Dublin, Ireland 2Division of Cell Matrix Biology Regenerative Medicine, University of Manchester, United Kingdom 3Manchester Institute of Biotechnology, School of Earth and Environmental Sciences, University of Manchester, United Kingdom [email protected] 5Institute of Paleobiology and Paleoanthropology, National Museum of Georgia, Tbilisi, Georgia 6Department of Evolutionary Anthropology, University of Vienna, Austria *Corresponding Author Abstract The emerging field of ancient metagenomics adds to these Bioinformatic pipelines optimised for the processing and as- processing complexities with the need for additional steps sessment of metagenomic ancient DNA (aDNA) are needed in the separation and authentication of ancient sequences from modern sequences. Currently, there are few pipelines for studies that do not make use of high yielding DNA cap- available for the analysis of ancient metagenomic DNA ture techniques. These bioinformatic pipelines are tradition- 1 4 ally optimised for broad aDNA purposes, are contingent on (aDNA) ≠ The limited number of bioinformatic pipelines selection biases and are associated with high costs. -
Sequence Alignment/Map Format Specification
Sequence Alignment/Map Format Specification The SAM/BAM Format Specification Working Group 3 Jun 2021 The master version of this document can be found at https://github.com/samtools/hts-specs. This printing is version 53752fa from that repository, last modified on the date shown above. 1 The SAM Format Specification SAM stands for Sequence Alignment/Map format. It is a TAB-delimited text format consisting of a header section, which is optional, and an alignment section. If present, the header must be prior to the alignments. Header lines start with `@', while alignment lines do not. Each alignment line has 11 mandatory fields for essential alignment information such as mapping position, and variable number of optional fields for flexible or aligner specific information. This specification is for version 1.6 of the SAM and BAM formats. Each SAM and BAMfilemay optionally specify the version being used via the @HD VN tag. For full version history see Appendix B. Unless explicitly specified elsewhere, all fields are encoded using 7-bit US-ASCII 1 in using the POSIX / C locale. Regular expressions listed use the POSIX / IEEE Std 1003.1 extended syntax. 1.1 An example Suppose we have the following alignment with bases in lowercase clipped from the alignment. Read r001/1 and r001/2 constitute a read pair; r003 is a chimeric read; r004 represents a split alignment. Coor 12345678901234 5678901234567890123456789012345 ref AGCATGTTAGATAA**GATAGCTGTGCTAGTAGGCAGTCAGCGCCAT +r001/1 TTAGATAAAGGATA*CTG +r002 aaaAGATAA*GGATA +r003 gcctaAGCTAA +r004 ATAGCT..............TCAGC -r003 ttagctTAGGC -r001/2 CAGCGGCAT The corresponding SAM format is:2 1Charset ANSI X3.4-1968 as defined in RFC1345. -
Differential Gene Expression Profiling in Bed Bug (Cimex Lectularius L.) Fed on Ibuprofen and Caffeine in Reconstituted Human Blood Ralph B
Herpe y & tolo og g l y: o C th i u Narain et al., Entomol Ornithol Herpetol 2015, 4:3 n r r r e O n , t y R g DOI: 10.4172/2161-0983.1000160 e o l s o e a m r o c t h n E Entomology, Ornithology & Herpetology ISSN: 2161-0983 ResearchResearch Article Article OpenOpen Access Access Differential Gene Expression Profiling in Bed Bug (Cimex Lectularius L.) Fed on Ibuprofen and Caffeine in Reconstituted Human Blood Ralph B. Narain, Haichuan Wang and Shripat T. Kamble* Department of Entomology, University of Nebraska, Lincoln, NE 68583, USA Abstract The recent resurgence of the common bed bug (Cimex lectularius L.) infestations worldwide has created a need for renewed research on biology, behavior, population genetics and management practices. Humans serve as exclusive hosts to bed bugs in urban environments. Since a majority of humans consume Ibuprofen (as pain medication) and caffeine (in coffee and other soft drinks) so bug bugs subsequently acquire Ibuprofen and caffeine through blood feeding. However, the effect of these chemicals at genetic level in bed bug is unknown. Therefore, this research was conducted to determine differential gene expression in bed bugs using RNA-Seq analysis at dosages of 200 ppm Ibuprofen and 40 ppm caffeine incorporated into reconstituted human blood and compared against the control. Total RNA was extracted from a single bed bug per replication per treatment and sequenced. Read counts obtained were analyzed using Bioconductor software programs to identify differentially expressed genes, which were then searched against the non-redundant (nr) protein database of National Center for Biotechnology Information (NCBI). -
Multi-Scale Analysis and Clustering of Co-Expression Networks
Multi-scale analysis and clustering of co-expression networks Nuno R. Nen´e∗1 1Department of Genetics, University of Cambridge, Cambridge, UK September 14, 2018 Abstract 1 Introduction The increasing capacity of high-throughput genomic With the advent of technologies allowing the collec- technologies for generating time-course data has tion of time-series in cell biology, the rich structure stimulated a rich debate on the most appropriate of the paths that cells take in expression space be- methods to highlight crucial aspects of data struc- came amenable to processing. Several methodologies ture. In this work, we address the problem of sparse have been crucial to carefully organize the wealth of co-expression network representation of several time- information generated by experiments [1], including course stress responses in Saccharomyces cerevisiae. network-based approaches which constitute an excel- We quantify the information preserved from the origi- lent and flexible option for systems-level understand- nal datasets under a graph-theoretical framework and ing [1,2,3]. The use of graphs for expression anal- evaluate how cross-stress features can be identified. ysis is inherently an attractive proposition for rea- This is performed both from a node and a network sons related to sparsity, which simplifies the cumber- community organization point of view. Cluster anal- some analysis of large datasets, in addition to being ysis, here viewed as a problem of network partition- mathematically convenient (see examples in Fig.1). ing, is achieved under state-of-the-art algorithms re- The properties of co-expression networks might re- lying on the properties of stochastic processes on the veal a myriad of aspects pertaining to the impact of constructed graphs. -
Bedtools Documentation Release 2.30.0
Bedtools Documentation Release 2.30.0 Quinlan lab @ Univ. of Utah Jan 23, 2021 Contents 1 Tutorial 3 2 Important notes 5 3 Interesting Usage Examples 7 4 Table of contents 9 5 Performance 169 6 Brief example 173 7 License 175 8 Acknowledgments 177 9 Mailing list 179 i ii Bedtools Documentation, Release 2.30.0 Collectively, the bedtools utilities are a swiss-army knife of tools for a wide-range of genomics analysis tasks. The most widely-used tools enable genome arithmetic: that is, set theory on the genome. For example, bedtools allows one to intersect, merge, count, complement, and shuffle genomic intervals from multiple files in widely-used genomic file formats such as BAM, BED, GFF/GTF, VCF. While each individual tool is designed to do a relatively simple task (e.g., intersect two interval files), quite sophisticated analyses can be conducted by combining multiple bedtools operations on the UNIX command line. bedtools is developed in the Quinlan laboratory at the University of Utah and benefits from fantastic contributions made by scientists worldwide. Contents 1 Bedtools Documentation, Release 2.30.0 2 Contents CHAPTER 1 Tutorial We have developed a fairly comprehensive tutorial that demonstrates both the basics, as well as some more advanced examples of how bedtools can help you in your research. Please have a look. 3 Bedtools Documentation, Release 2.30.0 4 Chapter 1. Tutorial CHAPTER 2 Important notes • As of version 2.28.0, bedtools now supports the CRAM format via the use of htslib. Specify the reference genome associated with your CRAM file via the CRAM_REFERENCE environment variable. -
Quantification of Experimentally Induced Nucleotide Conversions in High-Throughput Sequencing Datasets Tobias Neumann1* , Veronika A
Neumann et al. BMC Bioinformatics (2019) 20:258 https://doi.org/10.1186/s12859-019-2849-7 RESEARCH ARTICLE Open Access Quantification of experimentally induced nucleotide conversions in high-throughput sequencing datasets Tobias Neumann1* , Veronika A. Herzog2, Matthias Muhar1, Arndt von Haeseler3,4, Johannes Zuber1,5, Stefan L. Ameres2 and Philipp Rescheneder3* Abstract Background: Methods to read out naturally occurring or experimentally introduced nucleic acid modifications are emerging as powerful tools to study dynamic cellular processes. The recovery, quantification and interpretation of such events in high-throughput sequencing datasets demands specialized bioinformatics approaches. Results: Here, we present Digital Unmasking of Nucleotide conversions in K-mers (DUNK), a data analysis pipeline enabling the quantification of nucleotide conversions in high-throughput sequencing datasets. We demonstrate using experimentally generated and simulated datasets that DUNK allows constant mapping rates irrespective of nucleotide-conversion rates, promotes the recovery of multimapping reads and employs Single Nucleotide Polymorphism (SNP) masking to uncouple true SNPs from nucleotide conversions to facilitate a robust and sensitive quantification of nucleotide-conversions. As a first application, we implement this strategy as SLAM-DUNK for the analysis of SLAMseq profiles, in which 4-thiouridine-labeled transcripts are detected based on T > C conversions. SLAM-DUNK provides both raw counts of nucleotide-conversion containing reads as well as a base-content and read coverage normalized approach for estimating the fractions of labeled transcripts as readout. Conclusion: Beyond providing a readily accessible tool for analyzing SLAMseq and related time-resolved RNA sequencing methods (TimeLapse-seq, TUC-seq), DUNK establishes a broadly applicable strategy for quantifying nucleotide conversions. -
Penguin: a Tool for Predicting Pseudouridine Sites in Direct RNA Nanopore Sequencing Data
bioRxiv preprint doi: https://doi.org/10.1101/2021.03.31.437901; this version posted June 9, 2021. The copyright holder for this preprint (which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under aCC-BY-NC 4.0 International license. Penguin: A Tool for Predicting Pseudouridine Sites in Direct RNA Nanopore Sequencing Data Doaa Hassan1,4, Daniel Acevedo1,5, Swapna Vidhur Daulatabad1, Quoseena Mir1, Sarath Chandra Janga1,2,3 1. Department of BioHealth Informatics, School of Informatics and Computing, Indiana University Purdue University, 535 West Michigan Street, Indianapolis, Indiana 46202 2. Department of Medical and Molecular Genetics, Indiana University School of Medicine, Medical Research and Library Building, 975 West Walnut Street, Indianapolis, Indiana, 46202 3. Centre for Computational Biology and Bioinformatics, Indiana University School of Medicine, 5021 Health Information and Translational Sciences (HITS), 410 West 10th Street, Indianapolis, Indiana, 46202 4. Computers and Systems Department, National Telecommunication Institute, Cairo, Egypt. 5. Computer Science Department, University of Texas Rio Grande Valley Keywords: RNA modifications, Pseudouridine, Nanopore *Correspondence should be addressed to: Sarath Chandra Janga ([email protected]) Informatics and Communications Technology Complex, IT475H 535 West Michigan Street Indianapolis, IN 46202 317 278 4147 bioRxiv preprint doi: https://doi.org/10.1101/2021.03.31.437901; this version posted June 9, 2021. The copyright holder for this preprint (which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under aCC-BY-NC 4.0 International license. -
NGS Raw Data Analysis
NGS raw data analysis Introduc3on to Next-Generaon Sequencing (NGS)-data Analysis Laura Riva & Maa Pelizzola Outline 1. Descrip7on of sequencing plaorms 2. Alignments, QC, file formats, data processing, and tools of general use 3. Prac7cal lesson Illumina Sequencing Single base extension with incorporaon of fluorescently labeled nucleodes Library preparaon Automated cluster genera3on DNA fragmenta3on and adapter ligaon Aachment to the flow cell Cluster generaon by bridge amplifica3on of DNA fragments Sequencing Illumina Output Quality Scores Sequence Files Comparison of some exisng plaGorms 454 Ti RocheT Illumina HiSeqTM ABI 5500 (SOLiD) 2000 Amplificaon Emulsion PCR Bridge PCR Emulsion PCR Sequencing Pyrosequencing Reversible Ligaon-based reac7on terminators sequencing Paired ends/sep Yes/3kb Yes/200 bp Yes/3 kb Read length 400 bp 100 bp 75 bp Advantages Short run 7mes. The most popular Good base call Longer reads plaorm accuracy. Good improve mapping in mulplexing repe7ve regions. capability Ability to detect large structural variaons Disadvantages High reagent cost. Higher error rates in repeat sequences The Illumina HiSeq! Stacey Gabriel Libraries, lanes, and flowcells Flowcell Lanes Illumina Each reaction produces a unique Each NGS machine processes a library of DNA fragments for single flowcell containing several sequencing. independent lanes during a single sequencing run Applications of Next-Generation Sequencing Basic data analysis concepts: • Raw data analysis= image processing and base calling • Understanding Fastq • Understanding Quality scores • Quality control and read cleaning • Alignment to the reference genome • Understanding SAM/BAM formats • Samtools • Bedtools Understanding FASTQ file format FASTQ format is a text-based format for storing a biological sequence and its corresponding quality scores. -
Reference Genomes and Common File Formats Overview
Reference genomes and common file formats Overview ● Reference genomes and GRC ● Fasta and FastQ (unaligned sequences) ● SAM/BAM (aligned sequences) ● Summarized genomic features ○ BED (genomic intervals) ○ GFF/GTF (gene annotation) ○ Wiggle files, BEDgraphs, BigWigs (genomic scores) Why do we need to know about reference genomes? ● Allows for genes and genomic features to be evaluated in their genomic context. ○ Gene A is close to gene B ○ Gene A and gene B are within feature C ● Can be used to align shallow targeted high-throughput sequencing to a pre-built map of an organism Genome Reference Consortium (GRC) ● Most model organism reference genomes are being regularly updated ● Reference genomes consist of a mixture of known chromosomes and unplaced contigs called as Genome Reference Assembly ● Genome Reference Consortium: ○ A collaboration of institutes which curate and maintain the reference genomes of 4 model organisms: ■ Human - GRCh38.p9 (26 Sept 2016) ■ Mouse - GRCm38.p5 (29 June 2016) ■ Zebrafish - GRCz10 (12 Sept 2014) ■ Chicken - Gallus_gallus-5.0 (16 Dec 2015) ○ Latest human assembly is GRCh38, patches add information to the assembly without disrupting the chromosome coordinates ● Other model organisms are maintained separately, like: ○ Drosophila - Berkeley Drosophila Genome Project Overview ● Reference genomes and GRC ● Fasta and FastQ (unaligned sequences) ● SAM/BAM (aligned sequences) ● Summarized genomic features ○ BED (genomic intervals) ○ GFF/GTF (gene annotation) ○ Wiggle files, BEDgraphs, BigWigs (genomic scores) The