9781782161097 Preview.Pdf

Total Page:16

File Type:pdf, Size:1020Kb

9781782161097 Preview.Pdf [ 1 ] Apache Hive Cookbook Easy, hands-on recipes to help you understand Hive and its integration with frameworks that are used widely in today's big data world Hanish Bansal Saurabh Chauhan Shrey Mehrotra BIRMINGHAM - MUMBAI Apache Hive Cookbook Copyright © 2016 Packt Publishing All rights reserved. No part of this book may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, without the prior written permission of the publisher, except in the case of brief quotations embedded in critical articles or reviews. Every effort has been made in the preparation of this book to ensure the accuracy of the information presented. However, the information contained in this book is sold without warranty, either express or implied. Neither the authors, nor Packt Publishing, and its dealers and distributors will be held liable for any damages caused or alleged to be caused directly or indirectly by this book. Packt Publishing has endeavored to provide trademark information about all of the companies and products mentioned in this book by the appropriate use of capitals. However, Packt Publishing cannot guarantee the accuracy of this information. First published: April 2016 Production reference: 1260416 Published by Packt Publishing Ltd. Livery Place 35 Livery Street Birmingham B3 2PB, UK. ISBN 978-1-78216-108-0 www.packtpub.com Credits Authors Project Coordinator Hanish Bansal Bijal Patel Saurabh Chauhan Shrey Mehrotra Proofreader Safs Editing Reviewer Aristides Villarreal Bravo Indexer Priya Sane Commissioning Editor Wilson D'souza Graphics Kirk D'Penha Acquisition Editor Tushar Gupta Production Coordinator Shantanu N. Zagade Content Development Editor Anish Dhurat Cover Work Shantanu N. Zagade Technical Editor Vishal K. Mewada Copy Editor Dipti Mankame About the Authors Hanish Bansal is a software engineer with over 4 years of experience in developing big data applications. He loves to study emerging solutions and applications mainly related to big data processing, NoSQL, natural language processing, and neural networks. He has worked on various technologies such as Spring Framework, Hibernate, Hadoop, Hive, Flume, Kafka, Storm, and NoSQL databases, which include HBase, Cassandra, MongoDB, and search engines such as Elasticsearch. In 2012, he completed his graduation in Information Technology stream from Jaipur Engineering College and Research Center, Jaipur, India. He was also the technical reviewer of the book Apache Zookeeper Essentials. In his spare time, he loves to travel and listen to music. You can read his blog at http://hanishblogger.blogspot.in/ and follow him on Twitter at https://twitter.com/hanishbansal786. I would like to thank my parents for their love, support, encouragement and the amazing chances they've given me over the years. Saurabh Chauhan is a module lead with close to 8 years of experience in data warehousing and big data applications. He has worked on multiple Extract, Transform and Load tools, such as Oracle Data Integrator and Informatica as well as on big data technologies such as Hadoop, Hive, Pig, Sqoop, and Flume. He completed his bachelor of technology in 2007 from Vishveshwarya Institute of Engineering and Technology. In his spare time, he loves to travel and discover new places. He also has a keen interest in sports. I would like to thank everyone who has supported me throughout my life. Shrey Mehrotra has 6 years of IT experience and, since the past 4 years, in designing and architecting cloud and big data solutions for the governmental and fnancial domains. Having worked with big data R&D Labs and Global Data and Analytical Capabilities, he has gained insights into Hadoop, focusing on HDFS, MapReduce, and YARN. His technical strengths also include Hive, Pig, Spark, Elasticsearch, Sqoop, Flume, Kafka, and Java. He likes spending time performing R&D on different big data technologies. He is the co- author of the book Learning YARN, a certifed Hadoop developer, and has also written various technical papers. In his free time, he listens to music, watches movies, and spending time with friends. I would like to thank my mom and dad for giving me support to accomplish anything I wanted. Also, I would like to thank my friends, who bear with me while I am busy writing. About the Reviewer Aristides Villarreal Bravo is a Java developers, a member of the NetBeans Dream Team, and a Java User Groups leader. He has organized and participated in various conferences and seminars related to Java, JavaEE, NetBeans, NetBeans Platform, free software, and mobile devices, nationally and internationally. He has written tutorials and blogs about Java, NetBeans, and web development. He has participated in several interviews on sites such as NetBeans, NetBeans Dzone, and JavaHispano. He has developed plugins for NetBeans. He has been a technical reviewer for the book PrimeFaces Blueprints. Aristides is the CEO of Javscaz Software Developers. He lives in Panamá To my mother, father, and all family and friends. www.PacktPub.com eBooks, discount offers, and more Did you know that Packt offers eBook versions of every book published, with PDF and ePub fles available? You can upgrade to the eBook version at www.PacktPub.com and as a print book customer, you are entitled to a discount on the eBook copy. Get in touch with us at [email protected] for more details. At www.PacktPub.com, you can also read a collection of free technical articles, sign up for a range of free newsletters and receive exclusive discounts and offers on Packt books and eBooks. TM https://www2.packtpub.com/books/subscription/packtlib Do you need instant solutions to your IT questions? PacktLib is Packt's online digital book library. Here, you can search, access, and read Packt's entire library of books. Why Subscribe? f Fully searchable across every book published by Packt f Copy and paste, print, and bookmark content f On demand and accessible via a web browser Table of Contents Preface v Chapter 1: Developing Hive 1 Introduction 1 Deploying Hive on a Hadoop cluster 2 Deploying Hive Metastore 3 Installing Hive 6 Confguring HCatalog 10 Understanding different components of Hive 11 Compiling Hive from source 13 Hive packages 15 Debugging Hive 16 Running Hive 17 Changing confgurations at runtime 18 Chapter 2: Services in Hive 19 Introducing HiveServer2 19 Understanding HiveServer2 properties 21 Confguring HiveServer2 high availability 22 Using HiveServer2 Clients 24 Introducing the Hive metastore service 34 Confguring high availability of metastore service 36 Introducing Hue 36 Chapter 3: Understanding the Hive Data Model 43 Introduction 43 Using numeric data types 45 Using string data types 46 Using Date/Time data types 47 Using miscellaneous data types 48 Using complex data types 48 i Table of Contents Using operators 50 Partitioning 57 Partitioning a managed table 58 Partitioning an external table 65 Bucketing 65 Chapter 4: Hive Data Defnition Language 69 Introduction 70 Creating a database schema 70 Dropping a database schema 72 Altering a database schema 73 Using a database schema 74 Showing database schemas 74 Describing a database schema 75 Creating tables 76 Dropping tables 78 Truncating tables 79 Renaming tables 80 Altering table properties 80 Creating views 81 Dropping views 82 Altering the view properties 83 Altering the view as select 83 Showing tables 84 Showing partitions 85 Show the table properties 85 Showing create table 86 HCatalog 87 WebHCat 88 Chapter 5: Hive Data Manipulation Language 89 Introduction 89 Loading fles into tables 90 Inserting data into Hive tables from queries 93 Inserting data into dynamic partitions 96 Writing data into fles from queries 98 Enabling transactions in Hive 99 Inserting values into tables from SQL 101 Updating data 104 Deleting data 105 ii Table of Contents Chapter 6: Hive Extensibility Features 107 Introduction 108 Serialization and deserialization formats and data types 108 Exploring views 112 Exploring indexes 114 Hive partitioning 115 Creating buckets in Hive 119 Analytics functions in Hive 121 Windowing in Hive 125 File formats 132 Chapter 7: Joins and Join Optimization 137 Understanding the joins concept 137 Using a left/right/full outer join 141 Using a left semi join 145 Using a cross join 147 Using a map-side join 149 Using a bucket map join 152 Using a bucket sort merge map join 154 Using a skew join 155 Chapter 8: Statistics in Hive 157 Bringing statistics in to Hive 157 Table and partition statistics in Hive 158 Column statistics in Hive 163 Top K statistics in Hive 165 Chapter 9: Functions in Hive 167 Using built-in functions 167 Using the built-in User-defned Aggregation Function (UDAF) 184 Using the built-in User Defned Table Function (UDTF) 188 Creating custom User-Defned Functions (UDF) 191 Chapter 10: Hive Tuning 193 Enabling predicate pushdown optimizations in Hive 193 Optimizations to reduce the number of map 198 Sampling 200 Chapter 11: Hive Security 207 Securing Hadoop 207 Authorizing Hive 212 Confguring the SQL standards-based authorization 215 Authenticating Hive 222 iii Table of Contents Chapter 12: Hive Integration with Other Frameworks 229 Working with Apache Spark 229 Working with Accumulo 233 Working with HBase 234 Working with Google Drill 240 Index 243 iv Preface Hive is an open source big data framework in the Hadoop ecosystem. It provides an SQL-like interface to query data stored in HDFS. Underlying it runs MapReduce programs corresponding to the SQL query. Hive was initially developed by Facebook and later added to the Hadoop ecosystem. Hive is currently the most preferred framework to query data in Hadoop. Because most of the historical data is stored in RDBMS data stores, including Oracle and Teradata. It is convenient for the developers to run similar SQL statements in Hive to query data. Along with simple SQL statements, Hive supports wide variety of windowing and analytical functions, including rank, row num, dense rank, lead, and lag.
Recommended publications
  • E6895 Advanced Big Data Analytics Lecture 4: Data Store
    E6895 Advanced Big Data Analytics Lecture 4: Data Store Ching-Yung Lin, Ph.D. Adjunct Professor, Dept. of Electrical Engineering and Computer Science Chief Scientist, Graph Computing, IBM Watson Research Center E6895 Advanced Big Data Analytics — Lecture 4 © CY Lin, 2016 Columbia University Reference 2 E6895 Advanced Big Data Analytics – Lecture 4: Data Store © 2015 CY Lin, Columbia University Spark SQL 3 E6895 Advanced Big Data Analytics – Lecture 4: Data Store © 2015 CY Lin, Columbia University Spark SQL 4 E6895 Advanced Big Data Analytics – Lecture 4: Data Store © 2015 CY Lin, Columbia University Apache Hive 5 E6895 Advanced Big Data Analytics – Lecture 4: Data Store © 2015 CY Lin, Columbia University Using Hive to Create a Table 6 E6895 Advanced Big Data Analytics – Lecture 4: Data Store © 2015 CY Lin, Columbia University Creating, Dropping, and Altering DBs in Apache Hive 7 E6895 Advanced Big Data Analytics – Lecture 4: Data Store © 2015 CY Lin, Columbia University Another Hive Example 8 E6895 Advanced Big Data Analytics – Lecture 4: Data Store © 2015 CY Lin, Columbia University Hive’s operation modes 9 E6895 Advanced Big Data Analytics – Lecture 4: Data Store © 2015 CY Lin, Columbia University Using HiveQL for Spark SQL 10 E6895 Advanced Big Data Analytics – Lecture 4: Data Store © 2015 CY Lin, Columbia University Hive Language Manual 11 E6895 Advanced Big Data Analytics – Lecture 4: Data Store © 2015 CY Lin, Columbia University Using Spark SQL — Steps and Example 12 E6895 Advanced Big Data Analytics – Lecture 4: Data Store © 2015 CY Lin, Columbia University Query testtweet.json Get it from Learning Spark Github ==> https://github.com/databricks/learning-spark/tree/master/files 13 E6895 Advanced Big Data Analytics – Lecture 4: Data Store © 2015 CY Lin, Columbia University SchemaRDD 14 E6895 Advanced Big Data Analytics – Lecture 4: Data Store © 2015 CY Lin, Columbia University Row Objects Row objects represent records inside SchemaRDDs, and are simply fixed-length arrays of fields.
    [Show full text]
  • Unravel Data Systems Version 4.5
    UNRAVEL DATA SYSTEMS VERSION 4.5 Component name Component version name License names jQuery 1.8.2 MIT License Apache Tomcat 5.5.23 Apache License 2.0 Tachyon Project POM 0.8.2 Apache License 2.0 Apache Directory LDAP API Model 1.0.0-M20 Apache License 2.0 apache/incubator-heron 0.16.5.1 Apache License 2.0 Maven Plugin API 3.0.4 Apache License 2.0 ApacheDS Authentication Interceptor 2.0.0-M15 Apache License 2.0 Apache Directory LDAP API Extras ACI 1.0.0-M20 Apache License 2.0 Apache HttpComponents Core 4.3.3 Apache License 2.0 Spark Project Tags 2.0.0-preview Apache License 2.0 Curator Testing 3.3.0 Apache License 2.0 Apache HttpComponents Core 4.4.5 Apache License 2.0 Apache Commons Daemon 1.0.15 Apache License 2.0 classworlds 2.4 Apache License 2.0 abego TreeLayout Core 1.0.1 BSD 3-clause "New" or "Revised" License jackson-core 2.8.6 Apache License 2.0 Lucene Join 6.6.1 Apache License 2.0 Apache Commons CLI 1.3-cloudera-pre-r1439998 Apache License 2.0 hive-apache 0.5 Apache License 2.0 scala-parser-combinators 1.0.4 BSD 3-clause "New" or "Revised" License com.springsource.javax.xml.bind 2.1.7 Common Development and Distribution License 1.0 SnakeYAML 1.15 Apache License 2.0 JUnit 4.12 Common Public License 1.0 ApacheDS Protocol Kerberos 2.0.0-M12 Apache License 2.0 Apache Groovy 2.4.6 Apache License 2.0 JGraphT - Core 1.2.0 (GNU Lesser General Public License v2.1 or later AND Eclipse Public License 1.0) chill-java 0.5.0 Apache License 2.0 Apache Commons Logging 1.2 Apache License 2.0 OpenCensus 0.12.3 Apache License 2.0 ApacheDS Protocol
    [Show full text]
  • Assessment of Multiple Ingest Strategies for Accumulo Key-Value Store
    Assessment of Multiple Ingest Strategies for Accumulo Key-Value Store by Hai Pham A thesis submitted to the Graduate Faculty of Auburn University in partial fulfillment of the requirements for the Degree of Master of Science Auburn, Alabama May 7, 2016 Keywords: Accumulo, noSQL, ingest Copyright 2016 by Hai Pham Approved by Weikuan Yu, Co-Chair, Associate Professor of Computer Science, Florida State University Saad Biaz, Co-Chair, Professor of Computer Science and Software Engineering, Auburn University Sanjeev Baskiyar, Associate Professor of Computer Science and Software Engineering, Auburn University Abstract In recent years, the emergence of heterogeneous data, especially of the unstructured type, has been extremely rapid. The data growth happens concurrently in 3 dimensions: volume (size), velocity (growth rate) and variety (many types). This emerging trend has opened a new broad area of research, widely accepted as Big Data, which focuses on how to acquire, organize and manage huge amount of data effectively and efficiently. When coping with such Big Data, the traditional approach using RDBMS has been inefficient; because of this problem, a more efficient system named noSQL had to be created. This thesis will give an overview knowledge on the aforementioned noSQL systems and will then delve into a more specific instance of them which is Accumulo key-value store. Furthermore, since Accumulo is not designed with an ingest interface for users, this thesis focuses on investigating various methods for ingesting data, improving the performance and dealing with numerous parameters affecting this process. ii Acknowledgments First and foremost, I would like to express my profound gratitude to Professor Yu who with great kindness and patience has guided me through not only every aspect of computer science research but also many great directions towards my personal issues.
    [Show full text]
  • Hive, Spark, Presto for Interactive Queries on Big Data
    DEGREE PROJECT IN INFORMATION AND COMMUNICATION TECHNOLOGY, SECOND CYCLE, 30 CREDITS STOCKHOLM, SWEDEN 2018 Hive, Spark, Presto for Interactive Queries on Big Data NIKITA GUREEV KTH ROYAL INSTITUTE OF TECHNOLOGY SCHOOL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCE TRITA TRITA-EECS-EX-2018:468 www.kth.se Abstract Traditional relational database systems can not be efficiently used to analyze data with large volume and different formats, i.e. big data. Apache Hadoop is one of the first open-source tools that provides a dis- tributed data storage system and resource manager. The space of big data processing has been growing fast over the past years and many tech- nologies have been introduced in the big data ecosystem to address the problem of processing large volumes of data, and some of the early tools have become widely adopted, with Apache Hive being one of them. How- ever, with the recent advances in technology, there are other tools better suited for interactive analytics of big data, such as Apache Spark and Presto. In this thesis these technologies are examined and benchmarked in or- der to determine their performance for the task of interactive business in- telligence queries. The benchmark is representative of interactive business intelligence queries, and uses a star-shaped schema. The performance Hive Tez, Hive LLAP, Spark SQL, and Presto is examined with text, ORC, Par- quet data on different volume and concurrency. A short analysis and con- clusions are presented with the reasoning about the choice of framework and data format for a system that would run interactive queries on big data.
    [Show full text]
  • HDP 3.1.4 Release Notes Date of Publish: 2019-08-26
    Release Notes 3 HDP 3.1.4 Release Notes Date of Publish: 2019-08-26 https://docs.hortonworks.com Release Notes | Contents | ii Contents HDP 3.1.4 Release Notes..........................................................................................4 Component Versions.................................................................................................4 Descriptions of New Features..................................................................................5 Deprecation Notices.................................................................................................. 6 Terminology.......................................................................................................................................................... 6 Removed Components and Product Capabilities.................................................................................................6 Testing Unsupported Features................................................................................ 6 Descriptions of the Latest Technical Preview Features.......................................................................................7 Upgrading to HDP 3.1.4...........................................................................................7 Behavioral Changes.................................................................................................. 7 Apache Patch Information.....................................................................................11 Accumulo...........................................................................................................................................................
    [Show full text]
  • Apache Hadoop Goes Realtime at Facebook
    Apache Hadoop Goes Realtime at Facebook Dhruba Borthakur Joydeep Sen Sarma Jonathan Gray Kannan Muthukkaruppan Nicolas Spiegelberg Hairong Kuang Karthik Ranganathan Dmytro Molkov Aravind Menon Samuel Rash Rodrigo Schmidt Amitanand Aiyer Facebook {dhruba,jssarma,jgray,kannan, nicolas,hairong,kranganathan,dms, aravind.menon,rash,rodrigo, amitanand.s}@fb.com ABSTRACT 1. INTRODUCTION Facebook recently deployed Facebook Messages, its first ever Apache Hadoop [1] is a top-level Apache project that includes user-facing application built on the Apache Hadoop platform. open source implementations of a distributed file system [2] and Apache HBase is a database-like layer built on Hadoop designed MapReduce that were inspired by Googles GFS [5] and to support billions of messages per day. This paper describes the MapReduce [6] projects. The Hadoop ecosystem also includes reasons why Facebook chose Hadoop and HBase over other projects like Apache HBase [4] which is inspired by Googles systems such as Apache Cassandra and Voldemort and discusses BigTable, Apache Hive [3], a data warehouse built on top of the applications requirements for consistency, availability, Hadoop, and Apache ZooKeeper [8], a coordination service for partition tolerance, data model and scalability. We explore the distributed systems. enhancements made to Hadoop to make it a more effective realtime system, the tradeoffs we made while configuring the At Facebook, Hadoop has traditionally been used in conjunction system, and how this solution has significant advantages over the with Hive for storage and analysis of large data sets. Most of this sharded MySQL database scheme used in other applications at analysis occurs in offline batch jobs and the emphasis has been on Facebook and many other web-scale companies.
    [Show full text]
  • Apache Ranger 0.5.0 Installation
    Apache Ranger 0.5.0 Installation 1 Overview 2 Prerequisites 3 Installation Instructions 3.1 Preparing to install 3.1.1 Install Maven 3.1.2 Install git 3.1.3 Install gcc 3.1.4 Install MySQL 3.2 Build Ranger Admin from source 3.2.1 Clone the ranger source code 3.2.2 Build the source 3.2.3 Install Steps for Ranger Policy Admin on RHEL/CentOS 3.2.3.1 Lay down the build into appropriate places. 3.2.3.2 Install and configure Solr or SolrCloud 3.3 Installing the Ranger UserSync Process 3.4 Installing Apache Hadoop 3.4.1 Enabling Ranger HDFS Plugins 3.5 Installing Apache Hive(1.2.0) 3.5.1 Enabling Ranger Hive Plugin 3.5.2 Helpful info Ranger Hive Plugin / Hive Services 3.6 Installing Apache HBase (1.1.0.1) 3.6.1 Enabling Ranger HBase Plugins 3.6.2 Helpful info Ranger HBase Plugin / HBase Services 3.7 Installing Apache Knox Gateway 3.7.1 Enabling Ranger Knox Plugins 3.7.2 Helpful info Ranger Knox Plugin / Knox Services 3.7.3 Trusting Self Signed Knox Certificate 3.8 Enabling Ranger Solr Plugin 3.8.1 Install and configure Apache Solr in SolrCloud mode 3.8.2 Solr Service in Ranger Admin 3.8.3 Install and Enable Solr Plugin 3.8.4 Configuring Solr for Ranger 3.9 Installing Apache Storm (0.10.0) 3.9.1 Enabling Ranger Storm Plugins 3.9.2 Enabling Ranger Yarn Plugin 4 Installing Ranger KMS (0.5.0) 5 Enabling Audit Logging To HDFS 6 Enabling Audit Logging To SOLR 7 Configuring Kerberos Authentication 7.1 Installing and Configuring the KDC 7.2 Creating the Kerberos Database 7.3 Installing and Configuring the Kerberos Clients 7.4 Configure Storm to Work with a Secured Zookeeper Cluster 7.5 Configure Kerberos Authentication for Storm 7.6 Ranger UI setup Overview This document details the steps involved in Installing latest version of Apache Incubator Ranger independently on RHEL / Ubuntu / SUSE / Debian OS.
    [Show full text]
  • International Journal of Advanced Research in Computer Science And
    Volume 5, Issue 9, September 2015 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Effective Image Analysis on Twitter Streaming using Hadoop Eco System on Amazon Web Service EC2 Gautam Goswami Irisidea Technologies Pvt. Limited, India Abstract: Twitter is becoming the most popular online micro blogging network of real time post that enables users to send and read short 140-character messages called "tweets". Registered users can read and post tweets, but unregistered users can only read them. Today's Twitter is now less focused on what are you doing but has emerged as a source for discovery, with a focus on sharing relevant information and engaging in conversation. Sharing various visual information in the form of images/photos are becoming very popular where all the follower can see what images/photos have been posted/twitted instantly. In this paper I am going to explain how effectively registered users shares/uploads images among the followers. This information/statistics would be of great value for any organization/company when they launch their new product in market. If a particular image/photo sharing is high among tweeter community, organization/company can be assured that their product is penetrating more in the market. Here I have analyzed the momentum of visual information propagation . So that followers can be aware of that something new have been lunched in market and subsequently will have the curiosity to dig more on it. In this paper, the collected twitter steaming which has been (rich amount of data in semi structure format JSON for an interval of time) referred to as big data are processed efficiently to achieve mentioned output.
    [Show full text]
  • Cloudera JDBC Driver for Hive Installation and Configuration Guide
    Cloudera JDBC Driver for Apache Hive Important Notice © 2010-2021 Cloudera, Inc. All rights reserved. Cloudera, the Cloudera logo, and any other product or service names or slogans contained in this document, except as otherwise disclaimed, are trademarks of Cloudera and its suppliers or licensors, and may not be copied, imitated or used, in whole or in part, without the prior written permission of Cloudera or the applicable trademark holder. Hadoop and the Hadoop elephant logo are trademarks of the Apache Software Foundation. All other trademarks, registered trademarks, product names and company names or logos mentioned in this document are the property of their respective owners. Reference to any products, services, processes or other information, by trade name, trademark, manufacturer, supplier or otherwise does not constitute or imply endorsement, sponsorship or recommendation thereof by us. Complying with all applicable copyright laws is the responsibility of the user. Without limiting the rights under copyright, no part of this document may be reproduced, stored in or introduced into a retrieval system, or transmitted in any form or by any means (electronic, mechanical, photocopying, recording, or otherwise), or for any purpose, without the express written permission of Cloudera. Cloudera may have patents, patent applications, trademarks, copyrights, or other intellectual property rights covering subject matter in this document. Except as expressly provided in any written license agreement from Cloudera, the furnishing of this document does not give you any license to these patents, trademarks copyrights, or other intellectual property. The information in this document is subject to change without notice. Cloudera shall not be liable for any damages resulting from technical errors or omissions which may be present in this document, or from use of this document.
    [Show full text]
  • TR-4744: Secure Hadoop Using Apache Ranger with Netapp In
    Technical Report Secure Hadoop using Apache Ranger with NetApp In-Place Analytics Module Deployment Guide Karthikeyan Nagalingam, NetApp February 2019 | TR-4744 Abstract This document introduces the NetApp® In-Place Analytics Module for Apache Hadoop and Spark with Ranger. The topics covered in this report include the Ranger configuration, underlying architecture, integration with Hadoop, and benefits of Ranger with NetApp In-Place Analytics Module using Hadoop with NetApp ONTAP® data management software. TABLE OF CONTENTS 1 Introduction ........................................................................................................................................... 4 1.1 Overview .........................................................................................................................................................4 1.2 Deployment Options .......................................................................................................................................5 1.3 NetApp In-Place Analytics Module 3.0.1 Features ..........................................................................................5 2 Ranger ................................................................................................................................................... 6 2.1 Components Validated with Ranger ................................................................................................................6 3 NetApp In-Place Analytics Module Design with Ranger..................................................................
    [Show full text]
  • Customer Behavior Analysis of Web Server Logs Using Hive in Hadoop Framework Lavanya KS, Srinivasa R Dept
    International Journal of Advanced Networking & Applications (IJANA) ISSN: 0975-0282 Customer behavior analysis of web server logs using Hive in Hadoop Framework Lavanya KS, Srinivasa R Dept. of Studies in CSE, RRCE College of Engineering, Bangalore, Karnataka, India Email: [email protected] , [email protected] Abstract: Web log file is a log file created and stored by a web server automatically. Analyzing such web server access logs files will provide us various insights about website usage. Due to high usage of web, the log files are growing at much faster rate with increase in size. Processing this fast growing log files using relational database technology has been a challenging task these days. Therefore to analyze such large datasets we need a parallel processing system and a reliable data storage mechanism (Hadoop). Hadoop runs the big data where a massive quantity of information is processed via cluster of commodity hardware. In this paper we present the methodology used in pre-processing of high volume web log files, studying the statics of website and learning the user behavior using the architecture of Hadoop MapReduce framework, Hadoop Distributed File System, and HiveQL query language. Keywords: big data, customer behavior analysis, hadoop, log analysis, web server logs Foundation. Hadoop enables applications to work with 1. INTRODUCTION thousands of nodes with petabytes of data. While it can be used on a single machine, its true capability lies in scaling to In today's competitive environment, manufacturers/service hundreds or thousands of computers. Tom White describes providers are keen to know whether they provide the best Hadoop is specially designed to work on large volume of service/product to customers or whether customers look data using commodity hardware in parallel.
    [Show full text]
  • Licensing Information User Manual Release 4 (4.9) E87604-01
    Oracle® Big Data Appliance Licensing Information User Manual Release 4 (4.9) E87604-01 May 2017 Describes licensing and support for Oracle Big Data Appliance. Oracle Big Data Appliance Licensing Information User Manual, Release 4 (4.9) E87604-01 Copyright © 2011, 2017, Oracle and/or its affiliates. All rights reserved. Primary Author: Frederick Kush This software and related documentation are provided under a license agreement containing restrictions on use and disclosure and are protected by intellectual property laws. Except as expressly permitted in your license agreement or allowed by law, you may not use, copy, reproduce, translate, broadcast, modify, license, transmit, distribute, exhibit, perform, publish, or display any part, in any form, or by any means. Reverse engineering, disassembly, or decompilation of this software, unless required by law for interoperability, is prohibited. The information contained herein is subject to change without notice and is not warranted to be error-free. If you find any errors, please report them to us in writing. If this is software or related documentation that is delivered to the U.S. Government or anyone licensing it on behalf of the U.S. Government, then the following notice is applicable: U.S. GOVERNMENT END USERS: Oracle programs, including any operating system, integrated software, any programs installed on the hardware, and/or documentation, delivered to U.S. Government end users are "commercial computer software" pursuant to the applicable Federal Acquisition Regulation and agency- specific supplemental regulations. As such, use, duplication, disclosure, modification, and adaptation of the programs, including any operating system, integrated software, any programs installed on the hardware, and/or documentation, shall be subject to license terms and license restrictions applicable to the programs.
    [Show full text]