Parallel Decompression of Gzip-Compressed Files and Random

Total Page:16

File Type:pdf, Size:1020Kb

Parallel Decompression of Gzip-Compressed Files and Random Parallel decompression of gzip-compressed files and random access to DNA sequences Mael¨ Kerbiriou Rayan Chikhi INRIA Lille - Nord Europe C3BI USR 3756 CRIStAL, Universite´ de Lille Institut Pasteur & CNRS Lille, France Paris, France [email protected] [email protected] Abstract—Decompressing a file made by the gzip program slowdown of 1-2 orders of magnitudes exists at the beginning at an arbitrary location is in principle impossible, due to the of many tools, and in particular bioinformatics pipelines that nature of the DEFLATE compression algorithm. Consequently, take compressed files as input. no existing program can take advantage of parallelism to rapidly decompress large gzip-compressed files. This is an unsatisfac- If one could efficiently perform random accesses to content tory bottleneck, especially for the analysis of large sequencing in gzipped files, it would unlock multi-core decompression. As data experiments. Here we propose a parallel algorithm and an a result, data would be streamed significantly faster than if it implementation, pugz, that performs fast and exact decompres- was read from an uncompressed file. However, the DEFLATE sion of any text file. We show that pugz is an order of magnitude compression algorithm in gzip works in such a way that faster than gunzip, and 5x faster than a highly-optimized sequential implementation (libdeflate). We also study the accessing a random location does in principle require to read related problem of random access to compressed data. We give the whole file up to the desired location. This is because simple models and experimental results that shed light on the (1) variable-length blocks are created without any index nor structure of gzip-compressed files containing DNA sequences. explicit block starts/ends, and (2) back-references are made Preliminary results show that random access to sequences within to strings that appear previously in the decompressed file, a gzip-compressed FASTQ file is almost always feasible at low compression levels, yet is approximate at higher compression which is the essence of Lempel-Ziv parsing. A more precise levels. description of DEFLATE is provided later in the article. Index Terms—compression, bioinformatics, DNA sequences, To build some intuition on why random access is hard, let parallel algorithms an example file contain n identical repetitions of a sentence then followed by m identical repetitions of a completely I. INTRODUCTION different sentence. Suppose that neither the sentences lengths, The compression of genomic sequencing data continues to n nor m are known. The compression algorithm in gzip uses be a busy research area, using fundamental techniques such heuristics yet is very likely to always encode repetitions of the as Lempel-Ziv parsing [16], Burrows-Wheeler transform [5], sentences using back-references and not in clear form, except or de Bruijn graphs [2]. Despite many promising advances, at two locations: (i) the beginning of the file, and (ii) the the most popular and practical program for compressing raw DEFLATE block(s) where the second sentence appears for sequencing data nowadays remains the Lempel-Ziv based the first time. A random access towards the end of the file gzip program, arguably due to its speed and availability on all would require to somehow obtain information that is present systems. There exist parallel programs for speeding-up gzip only at location (ii). Yet, since this location cannot be easily compression, e.g. pigz1. The underlying compression algo- guessed, it is challenging to design a better strategy than to arXiv:1905.07224v1 [cs.DS] 17 May 2019 rithm of gzip, DEFLATE, easily lends itself to processing of fully decompress the file. blocks of data concurrently. There exist techniques to create so-called ”blocked gzip” However, the decompression aspects are more challenging files, i.e. files that are compressed as a sequence of small and and have been less studied, leaving us with only sequential al- independent blocks. Such files can also contain an additional gorithms. Investigating decompression would not be a pressing index that enables random access and parallel decompression. matter if it was already an IO-bound operation, but actually However as we will see in Section 2, these techniques are i) only around 30-50 MB of compressed data per second can be not so wide-spread and ii) yield worse compression ratios. processed by gunzip on a modern processor. This is much In this work, we will focus on the parallel decompression below the 500 (resp. 100–200) MB/sec read throughput of of files compressed by the gzip program. In bioinformatics, SATA solid-state (resp. mechanical) drives. In the near future, virtually every tool that processes large amounts of raw the wide availability of NVMe solid-state drives will allow sequencing data begins by reading large .fastq.gz file(s). for even faster reads, up to 3 GB/sec. Therefore a potential We first present simple models to provide a general under- standing of gzip compression. Parallel decompression would 1github.com/madler/pigz be straightforward if one could perform decompression at random locations. We therefore start by studying the feasibility will be omitted [7]. Input data is partitioned into blocks, of random access in compressed files, with a focus on files that may or may not be compressed, depending on whether containing DNA sequences. Random access turns out to be the data is ccompressible. To simplify the exposition we challenging for files with high compression ratio. We thus will consider (in this section only) the creation of a single propose an alternative strategy via the design of a new, general- compressed block, which is performed in two stages. The purpose, exact and parallel gzip decompression algorithm first stage is LZ77-style parsing, which encodes the input that can process any arbitrary ASCII file. We demonstrate data into a succession of literals (i.e. unmodified bytes) and the usefulness of this algorithm by showing significant speed- offset/length integer pairs, later denoted as matches, referring ups compared to the baseline method (gunzip) and a highly to previous occurrences of decompressed characters. In the optimized implementation (libdeflate). original definition of LZ77 [16], the parsing has to be of the form ”match-literal-match-literal-. ”, as follows. II. RELATED WORKS Several algorithms have been proposed for the compression Definition 1 (LZ77 parsing). Let s be a string. A LZ77 parsing of genomic data (for a survey, see [6]). Yet, the most popular of s is a sequence of phrases p1; : : : ; pz, where each pi is solution arguably remains gzip. It achieves roughly a 3- a tuple consisting of two integers (offset and length) and a fold size reduction on files in the FASTQ format. While character. The first phrase p1 encodes an empty match (0; 0) files produced by gzip are not designed to make random along with the first character of s. If p1; : : : ; pi−1 represent s access efficient, there exists variants that permit this operation. up to position j, then pi records the offset and length of the For instance, a modification of Lempel-Ziv parsing has been longest prefix of s[j + 1 :::] that occurs in s[: : : j], along with proposed to support random access at the expense of slightly the character that immediately follows the prefix. lower compression ratio [10], but the gzip program does We now define a more general flavor that better abstracts not implement it. In the same spirit, the tabix/bgzip pro- the product of DEFLATE. grams from the SAMtools/HTSlib project [12] create so-called blocked files that are indexed and gzip-compatible2. One Definition 2 (mixed LZ77-style parsing). Let s be a string, a 0 0 can also create an index of blocks within a classical gzip- mixed LZ77-style parsing (mLZ) of s is a sequence p1; : : : ; pz, 0 compressed file for later faster access to random locations [11]. where each pi is either a single character of s or a pair of The index is created during an initial sequential decompression offset/length integers that refers to a substring of s within 0 0 of the whole file. This essentially solves random access, except p1; : : : ; pi−1. that the technique is not so widespread as it requires a separate LZ77 parsing can be seen as a special case of mLZ. In file or a different file format, and does not apply when one DEFLATE, matches are of length between 3 and 258 bytes, only needs to read a given compressed file once. Anecdotally, and are made within a sliding window of length 32 KB, i.e. a majority of compressed files hosted in the Sequence Read offsets are never larger than 32 KB. Archive repository and uploaded in 2018 are not compressed The second stage of DEFLATE uses Huffman coding to in blocks. entropy-compress the output of the first stage. In practice, Previous efforts have aimed to improve compression via DEFLATE uses two Huffman code trees, but going into more better algorithms or hardware acceleration, e.g. in [1], [14]. Ef- details will not be necessary here. An important point however ficient decompression has been relatively less studied, except is that Huffman compression is reset at the beginning of each in the following works. A fine-grained hardware-accelerated block. To summarize, we can abstract the compression process decompression algorithm achieves a 2x speedup compared as follows. to gunzip [15]. A recent unpublished highly-optimized se- quential implementation (libdeflate) achieves roughly a Algorithm 1 (DEFLATE compression). Let s be a string, 3x speedup3. Perhaps the closest work to ours is an attempt DEFLATE compression produces a Huffman-compressed rep- to recover corrupted DEFLATE-compressed files [3], [4]. It resentation C from a mixed LZ77-style parsing of s.
Recommended publications
  • Data Compression: Dictionary-Based Coding 2 / 37 Dictionary-Based Coding Dictionary-Based Coding
    Dictionary-based Coding already coded not yet coded search buffer look-ahead buffer cursor (N symbols) (L symbols) We know the past but cannot control it. We control the future but... Last Lecture Last Lecture: Predictive Lossless Coding Predictive Lossless Coding Simple and effective way to exploit dependencies between neighboring symbols / samples Optimal predictor: Conditional mean (requires storage of large tables) Affine and Linear Prediction Simple structure, low-complex implementation possible Optimal prediction parameters are given by solution of Yule-Walker equations Works very well for real signals (e.g., audio, images, ...) Efficient Lossless Coding for Real-World Signals Affine/linear prediction (often: block-adaptive choice of prediction parameters) Entropy coding of prediction errors (e.g., arithmetic coding) Using marginal pmf often already yields good results Can be improved by using conditional pmfs (with simple conditions) Heiko Schwarz (Freie Universität Berlin) — Data Compression: Dictionary-based Coding 2 / 37 Dictionary-based Coding Dictionary-Based Coding Coding of Text Files Very high amount of dependencies Affine prediction does not work (requires linear dependencies) Higher-order conditional coding should work well, but is way to complex (memory) Alternative: Do not code single characters, but words or phrases Example: English Texts Oxford English Dictionary lists less than 230 000 words (including obsolete words) On average, a word contains about 6 characters Average codeword length per character would be limited by 1
    [Show full text]
  • Contrasting the Performance of Compression Algorithms on Genomic Data
    Contrasting the Performance of Compression Algorithms on Genomic Data Cornel Constantinescu, IBM Research Almaden Outline of the Talk: • Introduction / Motivation • Data used in experiments • General purpose compressors comparison • Simple Improvements • Special purpose compression • Transparent compression – working on compressed data (prototype) • Parallelism / Multithreading • Conclusion Introduction / Motivation • Despite the large number of research papers and compression algorithms proposed for compressing genomic data generated by sequencing machines, by far the most commonly used compression algorithm in the industry for FASTQ data is gzip. • The main drawbacks of the proposed alternative special-purpose compression algorithms are: • slow speed of either compression or decompression or both, and also their • brittleness by making various limiting assumptions about the input FASTQ format (for example, the structure of the headers or fixed lengths of the records [1]) in order to further improve their specialized compression. 1. Ibrahim Numanagic, James K Bonfield, Faraz Hach, Jan Voges, Jorn Ostermann, Claudio Alberti, Marco Mattavelli, and S Cenk Sahinalp. Comparison of high-throughput sequencing data compression tools. Nature Methods, 13(12):1005–1008, October 2016. Fast and Efficient Compression of Next Generation Sequencing Data 2 2 General Purpose Compression of Genomic Data As stated earlier, gzip/zlib compression is the method of choice by the industry for FASTQ genomic data. FASTQ genomic data is a text-based format (ASCII readable text) for storing a biological sequence and the corresponding quality scores. Each sequence letter and quality score is encoded with a single ASCII character. FASTQ data is structured in four fields per record (a “read”). The first field is the SEQUENCE ID or the header of the read.
    [Show full text]
  • The Basic Principles of Data Compression
    The Basic Principles of Data Compression Author: Conrad Chung, 2BrightSparks Introduction Internet users who download or upload files from/to the web, or use email to send or receive attachments will most likely have encountered files in compressed format. In this topic we will cover how compression works, the advantages and disadvantages of compression, as well as types of compression. What is Compression? Compression is the process of encoding data more efficiently to achieve a reduction in file size. One type of compression available is referred to as lossless compression. This means the compressed file will be restored exactly to its original state with no loss of data during the decompression process. This is essential to data compression as the file would be corrupted and unusable should data be lost. Another compression category which will not be covered in this article is “lossy” compression often used in multimedia files for music and images and where data is discarded. Lossless compression algorithms use statistic modeling techniques to reduce repetitive information in a file. Some of the methods may include removal of spacing characters, representing a string of repeated characters with a single character or replacing recurring characters with smaller bit sequences. Advantages/Disadvantages of Compression Compression of files offer many advantages. When compressed, the quantity of bits used to store the information is reduced. Files that are smaller in size will result in shorter transmission times when they are transferred on the Internet. Compressed files also take up less storage space. File compression can zip up several small files into a single file for more convenient email transmission.
    [Show full text]
  • Lzw Compression and Decompression
    LZW COMPRESSION AND DECOMPRESSION December 4, 2015 1 Contents 1 INTRODUCTION 3 2 CONCEPT 3 3 COMPRESSION 3 4 DECOMPRESSION: 4 5 ADVANTAGES OF LZW: 6 6 DISADVANTAGES OF LZW: 6 2 1 INTRODUCTION LZW stands for Lempel-Ziv-Welch. This algorithm was created in 1984 by these people namely Abraham Lempel, Jacob Ziv, and Terry Welch. This algorithm is very simple to implement. In 1977, Lempel and Ziv published a paper on the \sliding-window" compression followed by the \dictionary" based compression which were named LZ77 and LZ78, respectively. later, Welch made a contri- bution to LZ78 algorithm, which was then renamed to be LZW Compression algorithm. 2 CONCEPT Many files in real time, especially text files, have certain set of strings that repeat very often, for example " The ","of","on"etc., . With the spaces, any string takes 5 bytes, or 40 bits to encode. But what if we need to add the whole string to the list of characters after the last one, at 256. Then every time we came across the string like" the ", we could send the code 256 instead of 32,116,104 etc.,. This would take 9 bits instead of 40bits. This is the algorithm of LZW compression. It starts with a "dictionary" of all the single character with indexes from 0 to 255. It then starts to expand the dictionary as information gets sent through. Pretty soon, all the strings will be encoded as a single bit, and compression would have occurred. LZW compression replaces strings of characters with single codes. It does not analyze the input text.
    [Show full text]
  • Pack, Encrypt, Authenticate Document Revision: 2021 05 02
    PEA Pack, Encrypt, Authenticate Document revision: 2021 05 02 Author: Giorgio Tani Translation: Giorgio Tani This document refers to: PEA file format specification version 1 revision 3 (1.3); PEA file format specification version 2.0; PEA 1.01 executable implementation; Present documentation is released under GNU GFDL License. PEA executable implementation is released under GNU LGPL License; please note that all units provided by the Author are released under LGPL, while Wolfgang Ehrhardt’s crypto library units used in PEA are released under zlib/libpng License. PEA file format and PCOMPRESS specifications are hereby released under PUBLIC DOMAIN: the Author neither has, nor is aware of, any patents or pending patents relevant to this technology and do not intend to apply for any patents covering it. As far as the Author knows, PEA file format in all of it’s parts is free and unencumbered for all uses. Pea is on PeaZip project official site: https://peazip.github.io , https://peazip.org , and https://peazip.sourceforge.io For more information about the licenses: GNU GFDL License, see http://www.gnu.org/licenses/fdl.txt GNU LGPL License, see http://www.gnu.org/licenses/lgpl.txt 1 Content: Section 1: PEA file format ..3 Description ..3 PEA 1.3 file format details ..5 Differences between 1.3 and older revisions ..5 PEA 2.0 file format details ..7 PEA file format’s and implementation’s limitations ..8 PCOMPRESS compression scheme ..9 Algorithms used in PEA format ..9 PEA security model .10 Cryptanalysis of PEA format .12 Data recovery from
    [Show full text]
  • Improved Neural Network Based General-Purpose Lossless Compression Mohit Goyal, Kedar Tatwawadi, Shubham Chandak, Idoia Ochoa
    JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, AUGUST 2015 1 DZip: improved neural network based general-purpose lossless compression Mohit Goyal, Kedar Tatwawadi, Shubham Chandak, Idoia Ochoa Abstract—We consider lossless compression based on statistical [4], [5] and generative modeling [6]). Neural network based data modeling followed by prediction-based encoding, where an models can typically learn highly complex patterns in the data accurate statistical model for the input data leads to substantial much better than traditional finite context and Markov models, improvements in compression. We propose DZip, a general- purpose compressor for sequential data that exploits the well- leading to significantly lower prediction error (measured as known modeling capabilities of neural networks (NNs) for pre- log-loss or perplexity [4]). This has led to the development of diction, followed by arithmetic coding. DZip uses a novel hybrid several compressors using neural networks as predictors [7]– architecture based on adaptive and semi-adaptive training. Unlike [9], including the recently proposed LSTM-Compress [10], most NN based compressors, DZip does not require additional NNCP [11] and DecMac [12]. Most of the previous works, training data and is not restricted to specific data types. The proposed compressor outperforms general-purpose compressors however, have been tailored for compression of certain data such as Gzip (29% size reduction on average) and 7zip (12% size types (e.g., text [12] [13] or images [14], [15]), where the reduction on average) on a variety of real datasets, achieves near- prediction model is trained in a supervised framework on optimal compression on synthetic datasets, and performs close to separate training data or the model architecture is tuned for specialized compressors for large sequence lengths, without any the specific data type.
    [Show full text]
  • Randomized Lempel-Ziv Compression for Anti-Compression Side-Channel Attacks
    Randomized Lempel-Ziv Compression for Anti-Compression Side-Channel Attacks by Meng Yang A thesis presented to the University of Waterloo in fulfillment of the thesis requirement for the degree of Master of Applied Science in Electrical and Computer Engineering Waterloo, Ontario, Canada, 2018 c Meng Yang 2018 I hereby declare that I am the sole author of this thesis. This is a true copy of the thesis, including any required final revisions, as accepted by my examiners. I understand that my thesis may be made electronically available to the public. ii Abstract Security experts confront new attacks on TLS/SSL every year. Ever since the compres- sion side-channel attacks CRIME and BREACH were presented during security conferences in 2012 and 2013, online users connecting to HTTP servers that run TLS version 1.2 are susceptible of being impersonated. We set up three Randomized Lempel-Ziv Models, which are built on Lempel-Ziv77, to confront this attack. Our three models change the determin- istic characteristic of the compression algorithm: each compression with the same input gives output of different lengths. We implemented SSL/TLS protocol and the Lempel- Ziv77 compression algorithm, and used them as a base for our simulations of compression side-channel attack. After performing the simulations, all three models successfully pre- vented the attack. However, we demonstrate that our randomized models can still be broken by a stronger version of compression side-channel attack that we created. But this latter attack has a greater time complexity and is easily detectable. Finally, from the results, we conclude that our models couldn't compress as well as Lempel-Ziv77, but they can be used against compression side-channel attacks.
    [Show full text]
  • User Commands GZIP ( 1 ) Gzip, Gunzip, Gzcat – Compress Or Expand Files Gzip [ –Acdfhllnnrtvv19 ] [–S Suffix] [ Name ... ]
    User Commands GZIP ( 1 ) NAME gzip, gunzip, gzcat – compress or expand files SYNOPSIS gzip [–acdfhlLnNrtvV19 ] [– S suffix] [ name ... ] gunzip [–acfhlLnNrtvV ] [– S suffix] [ name ... ] gzcat [–fhLV ] [ name ... ] DESCRIPTION Gzip reduces the size of the named files using Lempel-Ziv coding (LZ77). Whenever possible, each file is replaced by one with the extension .gz, while keeping the same ownership modes, access and modification times. (The default extension is – gz for VMS, z for MSDOS, OS/2 FAT, Windows NT FAT and Atari.) If no files are specified, or if a file name is "-", the standard input is compressed to the standard output. Gzip will only attempt to compress regular files. In particular, it will ignore symbolic links. If the compressed file name is too long for its file system, gzip truncates it. Gzip attempts to truncate only the parts of the file name longer than 3 characters. (A part is delimited by dots.) If the name con- sists of small parts only, the longest parts are truncated. For example, if file names are limited to 14 characters, gzip.msdos.exe is compressed to gzi.msd.exe.gz. Names are not truncated on systems which do not have a limit on file name length. By default, gzip keeps the original file name and timestamp in the compressed file. These are used when decompressing the file with the – N option. This is useful when the compressed file name was truncated or when the time stamp was not preserved after a file transfer. Compressed files can be restored to their original form using gzip -d or gunzip or gzcat.
    [Show full text]
  • A Software Tool for Data Compression Using the LZ77 ("Sliding Window") Algorithm Student Authors: Vladan R
    A Software Tool for Data Compression Using the LZ77 ("Sliding Window") Algorithm Student authors: Vladan R. Djokić1, Miodrag G. Vidojković1 Mentors: Radomir S. Stanković2, Dušan B. Gajić2 Abstract – Data compression is a field of computer science that close the paper with some conclusions in the final section. is always in need of fast algorithms and their efficient implementations. Lempel-Ziv algorithm is the first which used a dictionary method for data compression. In this paper, we II. LEMPEL-ZIV ALGORITHM present the software implementation of this so-called "sliding window" method for compression and decompression. We also include experimental results considering data compression rate A. Theoretical basis and running time. This software tool includes a graphical user interface and is meant for use in educational purposes. The Lempel-Ziv algorithm [1] is an algorithm for lossless data compression. It is actually a whole family of algorithms, Keywords – Lempel-Ziv algorithm, C# programming solution, (see Figure 1) stemming from the two original algorithms that text compression. were first proposed by Jacob Ziv and Abraham Lempel in their landmark papers in 1977. [1] and 1978. [2]. LZ77 and I. INTRODUCTION LZ78 got their name by year of publishing. The Lempel-Ziv algorithms belong to adaptive dictionary While reading a book it is noticeable that some words are coders [1]. On start of encoding process, dictionary does not repeating very often. In computer world, textual files exist. The dictionary is created during encoding. There is no represent those books and same logic may be applied there. final state of dictionary and it does not have fixed size.
    [Show full text]
  • The Ark Handbook
    The Ark Handbook Matt Johnston Henrique Pinto Ragnar Thomsen The Ark Handbook 2 Contents 1 Introduction 5 2 Using Ark 6 2.1 Opening Archives . .6 2.1.1 Archive Operations . .6 2.1.2 Archive Comments . .6 2.2 Working with Files . .7 2.2.1 Editing Files . .7 2.3 Extracting Files . .7 2.3.1 The Extract dialog . .8 2.4 Creating Archives and Adding Files . .8 2.4.1 Compression . .9 2.4.2 Password Protection . .9 2.4.3 Multi-volume Archive . 10 3 Using Ark in the Filemanager 11 4 Advanced Batch Mode 12 5 Credits and License 13 Abstract Ark is an archive manager by KDE. The Ark Handbook Chapter 1 Introduction Ark is a program for viewing, extracting, creating and modifying archives. Ark can handle vari- ous archive formats such as tar, gzip, bzip2, zip, rar, 7zip, xz, rpm, cab, deb, xar and AppImage (support for certain archive formats depends on the appropriate command-line programs being installed). In order to successfully use Ark, you need KDE Frameworks 5. The library libarchive version 3.1 or above is needed to handle most archive types, including tar, compressed tar, rpm, deb and cab archives. To handle other file formats, you need the appropriate command line programs, such as zipinfo, zip, unzip, rar, unrar, 7z, lsar, unar and lrzip. 5 The Ark Handbook Chapter 2 Using Ark 2.1 Opening Archives To open an archive in Ark, choose Open... (Ctrl+O) from the Archive menu. You can also open archive files by dragging and dropping from Dolphin.
    [Show full text]
  • Effective Variations on Opened GIF Format Images
    70 IJCSNS International Journal of Computer Science and Network Security, VOL.8 No.5, May 2008 Effective Variations on Opened GIF Format Images Hamza A. Ali1† and Bashar M. Ne’ma2††, Isra Private University, Amman, JORDAN Picture Format (Pict), Portable Network Graphic Summary (PNG), Photoshop native file (PSD), PCX from Zsoft, The CompuServe GIF format and the LZW compression method Kodac PCD. Usually scanners and digital cameras used to compress image data in this format is investigated in this acquire images in bitmapped formats. paper. Because of its better compression and greater color depth, JPEG has generally replaced GIF for photographic images. - Vector formats: images are a series of pixels that are Thorough study and discussion on GIF format images is carried "turned on" based on a mathematical formula. They are out in details in this work. Although, opening the header of GIF format images is difficult to achieve, we opened this format and basically defined by shapes and lines with no more studied all acceptable variations which may be have influence on than 256 colors. They are not resolution dependent, the viewing of any GIF format images. To get appropriate results, infinitely scalable and not appropriate for photo all practical is carried out via both Object Oriented Programming realistic images. Examples of vector file formats are (OOP) and concepts of software engineering tools. Windows Metafile Format (MWF), PostScript Format, Portable Document Format (PDF), and Computer Key words: Graphic Metafile (CGM). GIF format, Image Header, Image processing, LZW, Compression, Software Engineering Tools, and Security. In the rest of this section, a short definition is included for the most widely used computer graphic formats.
    [Show full text]
  • Arrow: Integration to 'Apache' 'Arrow'
    Package ‘arrow’ September 5, 2021 Title Integration to 'Apache' 'Arrow' Version 5.0.0.2 Description 'Apache' 'Arrow' <https://arrow.apache.org/> is a cross-language development platform for in-memory data. It specifies a standardized language-independent columnar memory format for flat and hierarchical data, organized for efficient analytic operations on modern hardware. This package provides an interface to the 'Arrow C++' library. Depends R (>= 3.3) License Apache License (>= 2.0) URL https://github.com/apache/arrow/, https://arrow.apache.org/docs/r/ BugReports https://issues.apache.org/jira/projects/ARROW/issues Encoding UTF-8 Language en-US SystemRequirements C++11; for AWS S3 support on Linux, libcurl and openssl (optional) Biarch true Imports assertthat, bit64 (>= 0.9-7), methods, purrr, R6, rlang, stats, tidyselect, utils, vctrs RoxygenNote 7.1.1.9001 VignetteBuilder knitr Suggests decor, distro, dplyr, hms, knitr, lubridate, pkgload, reticulate, rmarkdown, stringi, stringr, testthat, tibble, withr Collate 'arrowExports.R' 'enums.R' 'arrow-package.R' 'type.R' 'array-data.R' 'arrow-datum.R' 'array.R' 'arrow-tabular.R' 'buffer.R' 'chunked-array.R' 'io.R' 'compression.R' 'scalar.R' 'compute.R' 'config.R' 'csv.R' 'dataset.R' 'dataset-factory.R' 'dataset-format.R' 'dataset-partition.R' 'dataset-scan.R' 'dataset-write.R' 'deprecated.R' 'dictionary.R' 'dplyr-arrange.R' 'dplyr-collect.R' 'dplyr-eval.R' 'dplyr-filter.R' 'expression.R' 'dplyr-functions.R' 1 2 R topics documented: 'dplyr-group-by.R' 'dplyr-mutate.R' 'dplyr-select.R' 'dplyr-summarize.R'
    [Show full text]