Comparison of Spark Resource Managers and Distributed File Systems

Total Page:16

File Type:pdf, Size:1020Kb

Comparison of Spark Resource Managers and Distributed File Systems 2016 IEEE International Conferences on Big Data and Cloud Computing (BDCloud), Social Computing and Networking (SocialCom), Sustainable Computing and Communications (SustainCom) Comparison of Spark Resource Managers and Distributed File Systems Solmaz Salehian Yonghong Yan Department of Computer Science & Engineering, Department of Computer Science & Engineering, Oakland University, Oakland University, Rochester, MI 48309-4401, United States Rochester, MI 48309-4401, United States [email protected] [email protected] Abstract— The rapid growth in volume, velocity, and variety Section 4 provides information of different resource of data produced by applications of scientific computing, management subsystems of Spark. In Section 5, Four commercial workloads and cloud has led to Big Data. Traditional Distributed File Systems (DFSs) are reviewed. Then, solutions of data storage, management and processing cannot conclusion is included in Section 6. meet demands of this distributed data, so new execution models, data models and software systems have been developed to address II. CHALLENGES IN BIG DATA the challenges of storing data in heterogeneous form, e.g. HDFS, NoSQL database, and for processing data in parallel and Although Big Data provides users with lot of opportunities, distributed fashion, e.g. MapReduce, Hadoop and Spark creating an efficient software framework for Big Data has frameworks. This work comparatively studies Apache Spark many challenges in terms of networking, storage, management, distributed data processing framework. Our study first discusses analytics, and ethics [5, 6]. the resource management subsystems of Spark, and then reviews Previous studies [6-11] have been carried out to review open several of the distributed data storage options available to Spark. Keywords—Big Data, Distributed File System, Resource challenges in Big Data. [6] proposes an overview of Big Data Management, Apache Spark. initiatives and technologies. It also discusses the challenges and potential solutions in industries and academics from I. INTRODUCTION different perspectives such as engineers, computer scientists and statisticians. In [7], the authors provide a survey of diverse ig Data refers to different forms of large data sets which B demand special computational platforms in order to hardware and software platforms available for Big Data discover valuable knowledge. Big Data has revolutionized analytics such as Hadoop, MapReduce, High Performance all aspects of human’s lives and many industries including Computing (HPC), Graphics Processing Unit (GPU) and Field medicine, business, education, and science and engineering. Programmable Gate Arrays (FPGA). The hardware platforms Big Data environment are software tools to acquire, are then compared based on various metrics including organize and analyze heterogeneous large data with different scalability, data I/O rate, fault tolerance, real-time processing. velocities. There are two ways to import data: batch data, The software frameworks used with each of these hardware where a dataset is loading all at once, and streaming data where platforms are discussed with regards to their strengths and stream data is continuously imported. Big Data can be in drawbacks. different forms of structured and unstructured data, of arbitrary A comprehensive survey on different technologies, size. These features of Big Data refer to 3Vs (volume, variety opportunities and challenges in Big Data has been provided by and velocity) and define properties or dimensions of Big Data. [8]. This paper also classified the various attributes of Big Data Because this type of data is different from the traditional data including definitions, volume management, analysis, rapid and is not practical to be stored in a single machine, so special growth rate and security. tools and technologies are demanded to store, access, and Indexing technique for efficient retrieval and management of analyze them. These tools need to process this large amount of data is considered as another open challenge in Big Data, data in a timely and cost-effective manner. which has been investigated in [9]. There are different data processing frameworks proposed [10] illustrates the importance of Big Data and discusses for Big Data including Apache Hadoop [1], Dyrad [2], various areas where Big Data is being used. The background Yahoo!S4 [3], Apache Spark [4], etc. Each of which has its and state-of-the-art of Big Data have been reviewed by [11]. own advantages and disadvantages. In this paper, discussion concentrates on Spark, which is one of the most recent and Related technologies to Big Data including cloud computing, popular interactive data processing frameworks. Because IoT, data centers, and Hadoop have been discussed in the Spark is inspired by Hadoop, we first review Google paper. Then, four phases of the value chain of Big Data, e.g. MapReduce and Hadoop. Then, Spark resource managers are data generation, data acquisition, data storage, and data discussed in details. Four data storage options used in Spark analysis have been proposed. For each phase, the general are also discussed in this work. background, the technical challenges, and the latest advances The rest of the paper is organized as follow. In next section, have been introduced and reviewed. challenges in Big Data are summarized. In Section 3, Big Data Currently, software tools for managing Big Data resources processing frameworks are discussed. which have been developed for data acquisition, storage, 978-1-5090-3936-4/16 $31.00 © 2016 IEEE 567 DOI 10.1109/BDCloud-SocialCom-SustainCom.2016.88 management and analysis, are considered as the initial is slow for interactive data exploration [14], and is not efficient challenges in Big Data. In this paper we review Apache Spark, for real time jobs. Because, Hadoop makes heavy use of which is an efficient data processing framework supporting intermediate files in every step of Map and Reduce functions, real time and interactive job analysis. We also compare the and in every step it needs to read and write frequent used data four data storage options available for Big Data. from and to HDFS. This demands a lot of disk accesses and incurs large I/O overheads. The creation of Spark was III. BIG DATA PROCESSING FRAMEWORKS motivated by this challenges in Hadoop by offering in-memory computation feature. A. Google MapReduce MapReduce is a distributed parallel computing framework C. Spark designed by Google to process large amounts of data, and Apache Spark distributed data processing framework was Google File System (GFS) is its storage system [12]. Google originally developed in 2009 in UC Berkeley’s AMPLab, and MapReduce provides an interface for users to process a huge open sourced in 2010 as an Apache project. Spark provides in- amount of data while also allowing searching for data by memory computation, which leads to significant increase of Google search engine. data processing performance over Hadoop MapReduce [4], MapReduce has three main components: Master, Map because sharing data in memory is 10x to 100x times faster function and Reduce function [12]. The Master is responsible than sharing through network and disk. Spark in comparison for keeping the status of Map and Reduce servers and for with other Big Data technologies has several advantages. Most feeding data to and launching procedures for the Map and MapReduce solutions are efficient for one-pass process but Reduce functions. The Map function processes data and they do not work well for multi-pass computations. Spark with generates intermediate data. The Reduce function combines the in-memory data storage feature is able to handle multi process intermediate data and generates data for the next stage of job faster than other Big Data technologies, since it is able to processing. The assumption made by MapReduce is that all hold dataset in memory that is needed multiple times. Some data can be expressed in a form of key and value pairs [12]. studies compare the performance of Hadoop and Spark [14- The key is the correlation between the data, and the value is the 15], and demonstrate the advantage of Spark over Hadoop. content of data. Map and reduce functions are written by user. Spark also introduces a new data model, Resilient The input file is divided into M chunks with size ranges from Distributer Dataset (RDD), the main abstraction in Spark for 16 MB to 64 MB. Master finds the available server to do representing a distributed collection of elements. All data in computation and assigns servers for the Map function. Those Spark are expressed in form of RDD. There are two types of servers that are responsible for the Map function, retrieve the operation on RDD, transformation and action operations. input data and process it, and then store the Map results in the Transformation operations create a new RDD from existing file system. Map function takes a key and value pair as input RDD, while the action operations return final value and the and creates the intermediate data. The intermediate data is then result is not in form of RDD. Spark provides lazy evaluation, cut into multiple pieces and the Master receives their locations. which postpones the processing until it reaches to the action Servers that are responsible for the Reduce function retrieve operations. By this feature, Spark reduces the number of passes the data and sort and group the value with the same key. The of data by grouping operations together. In systems like data with the same key are processed by the Reduce function Hadoop MapReduce, developers often need to spend a lot of and a new set of key and value is generated. The final results time considering how to group operations in order to minimize will be available to users after processes are completed. the number of MapReduce passes [16]. Spark is written in Scala programming language [17] and it B. Hadoop runs on Java Virtual Machine (JVM) environment. It supports Apache Software Foundation developed Hadoop which was Scala, Python and Java programming languages and also offers inspired by Google MapReduce.
Recommended publications
  • Is 'Distributed' Worth It? Benchmarking Apache Spark with Mesos
    Is `Distributed' worth it? Benchmarking Apache Spark with Mesos N. Satra (ns532) January 13, 2015 Abstract A lot of research focus lately has been on building bigger dis- tributed systems to handle `Big Data' problems. This paper exam- ines whether typical problems for web-scale companies really benefits from the parallelism offered by these systems, or can be handled by a single machine without synchronisation overheads. Repeated runs of a movie review sentiment analysis task in Apache Spark were carried out using Apache Mesos and Mesosphere, on Google Compute Engine clusters of various sizes. Only a marginal improvement in run-time was observed on a distributed system as opposed to a single node. 1 Introduction Research in high performance computing and `big data' has recently focussed on distributed computing. The reason is three-pronged: • The rapidly decreasing cost of commodity machines, compared to the slower decline in prices of high performance or supercomputers. • The availability of cloud environments like Amazon EC2 or Google Compute Engine that let users access large numbers of computers on- demand, without upfront investment. • The development of frameworks and tools that provide easy-to-use id- ioms for distributed computing and managing such large clusters of machines. Large corporations like Google or Yahoo are working on petabyte-scale problems which simply can't be handled by single computers [9]. However, 1 smaller companies and research teams with much more manageable sizes of data have jumped on the bandwagon, using the tools built by the larger com- panies, without always analysing the performance tradeoffs. It has reached the stage where researchers are suggesting using the same MapReduce idiom hammer on all problems, whether they are nails or not [7].
    [Show full text]
  • Tachyon: Reliable, Memory Speed Storage for Cluster Computing Frameworks
    Tachyon: Reliable, Memory Speed Storage for Cluster Computing Frameworks The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation Haoyuan Li, Ali Ghodsi, Matei Zaharia, Scott Shenker, and Ion Stoica. 2014. Tachyon: Reliable, Memory Speed Storage for Cluster Computing Frameworks. In Proceedings of the ACM Symposium on Cloud Computing (SOCC '14). ACM, New York, NY, USA, Article 6 , 15 pages. As Published http://dx.doi.org/10.1145/2670979.2670985 Publisher Association for Computing Machinery (ACM) Version Author's final manuscript Citable link http://hdl.handle.net/1721.1/101090 Terms of Use Creative Commons Attribution-Noncommercial-Share Alike Detailed Terms http://creativecommons.org/licenses/by-nc-sa/4.0/ Tachyon: Reliable, Memory Speed Storage for Cluster Computing Frameworks Haoyuan Li Ali Ghodsi Matei Zaharia Scott Shenker Ion Stoica University of California, Berkeley MIT, Databricks University of California, Berkeley fhaoyuan,[email protected] [email protected] fshenker,[email protected] Abstract Even replicating the data in memory can lead to a signifi- cant drop in the write performance, as both the latency and Tachyon is a distributed file system enabling reliable data throughput of the network are typically much worse than sharing at memory speed across cluster computing frame- that of local memory. works. While caching today improves read workloads, Slow writes can significantly hurt the performance of job writes are either network or disk bound, as replication is pipelines, where one job consumes the output of another. used for fault-tolerance. Tachyon eliminates this bottleneck These pipelines are regularly produced by workflow man- by pushing lineage, a well-known technique, into the storage agers such as Oozie [4] and Luigi [7], e.g., to perform data layer.
    [Show full text]
  • Alluxio: a Virtual Distributed File System by Haoyuan Li A
    Alluxio: A Virtual Distributed File System by Haoyuan Li A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in Computer Science in the Graduate Division of the University of California, Berkeley Committee in charge: Professor Ion Stoica, Co-chair Professor Scott Shenker, Co-chair Professor John Chuang Spring 2018 Alluxio: A Virtual Distributed File System Copyright 2018 by Haoyuan Li 1 Abstract Alluxio: A Virtual Distributed File System by Haoyuan Li Doctor of Philosophy in Computer Science University of California, Berkeley Professor Ion Stoica, Co-chair Professor Scott Shenker, Co-chair The world is entering the data revolution era. Along with the latest advancements of the Inter- net, Artificial Intelligence (AI), mobile devices, autonomous driving, and Internet of Things (IoT), the amount of data we are generating, collecting, storing, managing, and analyzing is growing ex- ponentially. To store and process these data has exposed tremendous challenges and opportunities. Over the past two decades, we have seen significant innovation in the data stack. For exam- ple, in the computation layer, the ecosystem started from the MapReduce framework, and grew to many different general and specialized systems such as Apache Spark for general data processing, Apache Storm, Apache Samza for stream processing, Apache Mahout for machine learning, Ten- sorflow, Caffe for deep learning, Presto, Apache Drill for SQL workloads. There are more than a hundred popular frameworks for various workloads and the number is growing. Similarly, the storage layer of the ecosystem grew from the Apache Hadoop Distributed File System (HDFS) to a variety of choices as well, such as file systems, object stores, blob stores, key-value systems, and NoSQL databases to realize different tradeoffs in cost, speed and semantics.
    [Show full text]
  • Crystal: a Unified Cache Storage System for Analytical Databases
    Crystal: A Unified Cache Storage System for Analytical Databases Dominik Durner∗ Badrish Chandramouli Yinan Li Technische Universität München Microsoft Research Microsoft Research [email protected] [email protected] [email protected] ABSTRACT 1.1 Challenges Cloud analytical databases employ a disaggregated storage model, These caching solutions usually operate as a black-box at the file or where the elastic compute layer accesses data persisted on remote block level for simplicity, employing standard cache replacement cloud storage in block-oriented columnar formats. Given the high policies such as LRU to manage the cache. In spite of their sim- latency and low bandwidth to remote storage and the limited size of plicity, these solutions have not solved several architectural and fast local storage, caching data at the compute node is important and performance challenges for cloud databases: has resulted in a renewed interest in caching for analytics. Today, • Every DBMS today implements its own caching layer tailored to each DBMS builds its own caching solution, usually based on file- its specific requirements, resulting in a lot of work duplication or block-level LRU. In this paper, we advocate a new architecture of across systems, reinventing choices such as what to cache, where a smart cache storage system called Crystal, that is co-located with to cache, when to cache, and how to cache. compute. Crystal’s clients are DBMS-specific “data sources” with • Databases increasingly support analytics over raw data formats push-down predicates. Similar in spirit to a DBMS, Crystal incorpo- such as CSV and JSON, and row-oriented binary formats such as rates query processing and optimization components focusing on Apache Avro [6] – all very popular in the data lake [16].
    [Show full text]
  • Scalable Collaborative Caching and Storage Platform for Data Analytics
    Scalable Collaborative Caching and Storage Platform for Data Analytics by Timur Malgazhdarov A thesis submitted in conformity with the requirements for the degree of Master of Applied Science Edward S. Rogers Sr. Department of Electrical and Computer Engineering University of Toronto c Copyright 2018 by Timur Malgazhdarov Abstract Scalable Collaborative Caching and Storage Platform for Data Analytics Timur Malgazhdarov Master of Applied Science Edward S. Rogers Sr. Department of Electrical and Computer Engineering University of Toronto 2018 The emerging Big Data ecosystem has brought about dramatic proliferation of paradigms for analytics. In the race for the best performance, each new engine enforces tight cou- pling of analytics execution with caching and storage functionalities. This one-for-all ap- proach has led to either oversimplifications where traditional functionality was dropped or more configuration options that created more confusion about optimal settings. We avoid user confusion by following an integrated multi-service approach where we assign responsibilities to decoupled services. In our solution, called Gluon, we build a collab- orative cache tier that connects state-of-art analytics engines with a variety of storage systems. We use both open-source and proprietary technologies to implement our archi- tecture. We show that Gluon caching can achieve 2.5x-3x speedup when compared to uncustomized Spark caching while displaying higher resource utilization efficiency. Fi- nally, we show how Gluon can integrate traditional storage back-ends without significant performance loss when compared to vanilla analytics setups. ii Acknowledgements I would like to thank my supervisor, Professor Cristiana Amza, for her knowledge, guidance and support. It was my privilege and honor to work under Professor Amza's supervision.
    [Show full text]
  • An Architecture for Fast and General Data Processing on Large Clusters
    An Architecture for Fast and General Data Processing on Large Clusters Matei Zaharia Electrical Engineering and Computer Sciences University of California at Berkeley Technical Report No. UCB/EECS-2014-12 http://www.eecs.berkeley.edu/Pubs/TechRpts/2014/EECS-2014-12.html February 3, 2014 Copyright © 2014, by the author(s). All rights reserved. Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, to republish, to post on servers or to redistribute to lists, requires prior specific permission. An Architecture for Fast and General Data Processing on Large Clusters by Matei Alexandru Zaharia A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in Computer Science in the GRADUATE DIVISION of the UNIVERSITY OF CALIFORNIA, BERKELEY Committee in charge: Professor Scott Shenker, Chair Professor Ion Stoica Professor Alexandre Bayen Professor Joshua Bloom Fall 2013 An Architecture for Fast and General Data Processing on Large Clusters Copyright c 2013 by Matei Alexandru Zaharia Abstract An Architecture for Fast and General Data Processing on Large Clusters by Matei Alexandru Zaharia Doctor of Philosophy in Computer Science University of California, Berkeley Professor Scott Shenker, Chair The past few years have seen a major change in computing systems, as growing data volumes and stalling processor speeds require more and more applications to scale out to distributed systems.
    [Show full text]
  • Thesis Enabling Autoscaling for In-Memory Storage In
    THESIS ENABLING AUTOSCALING FOR IN-MEMORY STORAGE IN CLUSTER COMPUTING FRAMEWORK Submitted by Bibek Raj Shrestha Department of Computer Science In partial fulfillment of the requirements For the Degree of Master of Science Colorado State University Fort Collins, Colorado Spring 2019 Master’s Committee: Advisor: Sangmi Lee Pallickara Shrideep Pallickara Stephen C. Hayne Copyright by Bibek Raj Shrestha 2019 All Rights Reserved ABSTRACT ENABLING AUTOSCALING FOR IN-MEMORY STORAGE IN CLUSTER COMPUTING FRAMEWORK IoT enabled devices and observational instruments continuously generate voluminous data. A large portion of these datasets are delivered with the associated geospatial locations. The increased volumes of geospatial data, alongside the emerging geospatial services, pose computational chal- lenges for large-scale geospatial analytics. We have designed and implemented STRETCH, an in-memory distributed geospatial storage that preserves spatial proximity and enables proactive autoscaling for frequently accessed data. STRETCH stores data with a delayed data dispersion scheme that incrementally adds data nodes to the storage system. We have devised an autoscaling feature that proactively repartitions data to alleviate computational hotspots before they occur. We compared the performance of STRETCH with Apache Ignite and the results show that STRETCH provides up to 3 times the throughput when the system encounters hotspots. STRETCH is built on Apache Spark and Ignite and interacts with them at runtime. ii ACKNOWLEDGEMENTS I would like to thank my advisor, Dr. Sangmi Lee Pallickara, for her enthusiastic support and sincere guidance that have resulted in driving this research project to completion. I would also like to thank my thesis committee members, Dr. Shrideep Pallickara and Dr.
    [Show full text]
  • Alluxio: a Virtual Distributed File System
    Alluxio: A Virtual Distributed File System Haoyuan Li Electrical Engineering and Computer Sciences University of California at Berkeley Technical Report No. UCB/EECS-2018-29 http://www2.eecs.berkeley.edu/Pubs/TechRpts/2018/EECS-2018-29.html May 7, 2018 Copyright © 2018, by the author(s). All rights reserved. Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, to republish, to post on servers or to redistribute to lists, requires prior specific permission. Acknowledgement I would like to thank my advisors, Ion Stoica and Scott Shenker, for their support and guidance throughout this journey, and for providing me with great opportunities. They are both inspiring mentors, and continue to lead by example. They always challenge me to push ideas and work further, and share their advice and experience on life and research. Being able to learn from both of them has been my great fortune. Alluxio: A Virtual Distributed File System by Haoyuan Li A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in Computer Science in the Graduate Division of the University of California, Berkeley Committee in charge: Professor Ion Stoica, Co-chair Professor Scott Shenker, Co-chair Professor John Chuang Spring 2018 Alluxio: A Virtual Distributed File System Copyright 2018 by Haoyuan Li 1 Abstract Alluxio: A Virtual Distributed File System by Haoyuan Li Doctor of Philosophy in Computer Science University of California, Berkeley Professor Ion Stoica, Co-chair Professor Scott Shenker, Co-chair The world is entering the data revolution era.
    [Show full text]
  • Thesis, We Describe Two Main Areas That Lack Support for Long-Running Applications in Traditional System Designs
    Supporting Long-Running Applications in Shared Compute Clusters Panagiotis Garefalakis Department of Computing Imperial College London This dissertation is submitted for the degree of Doctor of Philosophy June 2020 Abstract Clusters that handle data-intensive workloads at a data-centre scale have become commonplace. In this setting, clusters are typically shared across several users and applications, and consolidate workloads that range from traditional analytics applications to critical services, stream processing, machine learning, and complex data processing applications. This constitutes a growing class of applications called long-running, consisting of containers that are used for durations ranging from hours to months. Even though long-running applications occupy a significant amount of resources in shared compute clusters today, there is currently rudimentary systems support that not only hinders application performance and resilience but also decreases cluster resource efficiency i.e., the effective utility extracted from cluster resources. In this thesis, we describe two main areas that lack support for long-running applications in traditional system designs. First, the way modern data processing frameworks execute complex computation tasks as part of shared long-running containers is broken. Even though these frameworks enable users to combine different types of computation as part of the same application using high-level programming interfaces, they ignore their diverse latency and throughput requirements during execution. Second, existing systems that allocate resources for long-running applications in the form of containers lack an expressive interface that can capture their placement requirements in shared compute clusters. Such placements can be expressed by means of complex constraints and are critical for both the performance and resilience of long-running applications.
    [Show full text]
  • Performance Tuning and Evaluation of Iterative Algorithms in Spark
    Performance Tuning and Evaluation of Iterative Algorithms in Spark Janani Gururam Department of Computer Science University of Maryland College Park, MD 20742 [email protected] Abstract. Spark is a widely used distributed, open-source framework for machine learning, relational queries, graph analytics and stream process- ing. This paper discusses key features of Spark that differentiate it from other MapReduce frameworks, focusing specifically on how iterative algo- rithms are implemented and optimized. We then analyze the impact of Resilient Distributed Dataset (RDD) caching, Serialization, Disk I/O, and other parameter considerations to improve performance in Spark. Finally, we discuss benchmarking tools currently available to Spark users. While there are several general purpose tools available, we do note that only a few Spark-specific solutions exist, and do not fully bridge the gap between workload characterization and performance implications. 1 Introduction Spark is an open-source, in-memory distributed framework for big data processing. The core engine of Spark underlays several modules for specialized analytical computation, including MLlib for machine learning applications, GraphX for graph analytics, SparkSQL for queries on structured, relational data, and Spark Streaming [25]. Spark provides several APIs for languages like Java, Scala, R, and Python. The base data structure in Spark is a Resilient Distributed Dataset(RDD), an immutable, dis- tributed collection of records. There are two types of operations in Spark: transformations, which perform operations on an input RDD to produce one or more new RDDs, and actions, which launch the series of transformations to return the final result RDD. Operations in Spark are lazily evaluated.
    [Show full text]
  • A Machine Learning Data Processing Framework
    tf.data: A Machine Learning Data Processing Framework Derek G. Murray Jiří Šimša Lacework* Google [email protected] [email protected] Ana Klimovic Ihor Indyk ETH Zurich* Google [email protected] [email protected] ABSTRACT Training machine learning models requires feeding input data for models to ingest. Input pipelines for machine learning jobs are often challenging to implement efficiently as they require reading large volumes of data, applying complex transformations, and transfer- ring data to hardware accelerators while overlapping computation and communication to achieve optimal performance. We present tf.data, a framework for building and executing efficient input pipelines for machine learning jobs. The tf.data API provides op- erators that can be parameterized with user-defined computation, composed, and reused across different machine learning domains. Figure 1: CDF showing the fraction of compute time that These abstractions enable users to focus on the application logic millions of ML training jobs executed in our fleet over one of data processing, while tf.data’s runtime ensures that pipelines month spend in the input pipeline. 20% of jobs spend more run efficiently. than a third of their compute time ingesting data. We demonstrate that input pipeline performance is critical to the end-to-end training time of state-of-the-art machine learning models. tf.data delivers the high performance required, while 1 INTRODUCTION avoiding the need for manual tuning of performance knobs. We Data is the lifeblood of machine learning (ML). Training ML models show that tf.data features, such as parallelism, caching, static op- requires steadily pumping examples for models to ingest and learn timizations, and optional non-deterministic execution are essential from.
    [Show full text]
  • Scaling Data Mining in Massively Parallel Dataflow Systems
    Scaling Data Mining in Massively Parallel Dataflow Systems vorgelegt von Dipl.-Inf. Sebastian Schelter aus Selb der Fakult¨atIV - Elektrotechnik und Informatik der Technischen Universit¨atBerlin zur Erlangung des akademischen Grades Doktor der Ingenieurwissenschaften - Dr. Ing. - genehmigte Dissertation Promotionsausschuss: Gutachter: Prof. Dr. Volker Markl Database Systems and Information Management Group, TU Berlin Prof. Dr. Klaus Robert M¨uller Machine Learning and Intelligent Data Analysis Group, TU Berlin Prof. Reza Zadeh, PhD Institute for Computational & Mathematical Engineering, Stanford University Berlin 2015 D 83 Acknowledgements I would like to express my gratitude to the people who accompanied me in the course of my studies and helped me shape this thesis. I'd like to first and foremost thank my advisor Volker Markl. He introduced me to the academic world, gave me freedom in choosing my research topics, and motivated me many times to leave my comfort zone and take on the challenges involved in conducting a PhD. Furthermore, I want to express my appreciation to Klaus-Robert M¨ullerand Reza Zadeh for agreeing to review this thesis. Very importantly, I'd like to thank my colleagues Christoph Boden, Max Heimel, Alan Akbik, Stephan Ewen, Fabian H¨uske, Kostas Tzoumas, Asterios Katsifodimos and Isabel Drost at the Database and Information Systems Group of TU Berlin, who helped me advance my research through collaboration, advice and discussions. I also enjoyed working with the DIMA students Christoph Nagel, Janani Chakkaradhari, and Till Rohrmann who conducted excellent master thesis in the context of my line of research. Furthermore, I'd like to thank my colleagues Mikio Braun, J´er^omeKunegis, J¨orgFliege and Mathias Peters for ideas and discussions in the research projects in which we worked together.
    [Show full text]