Apache Hbase. | 1

Total Page:16

File Type:pdf, Size:1020Kb

Apache Hbase. | 1 apache hbase. | 1 how hbase works *this is a study guide that was created from lecture videos and is used to help you gain an understanding of how hbase works. HBase Foundations Yahoo released the Hadoop data storage system and Google added HDFS programming interface. HDFS stands for Hadoop Distributed File System and it spreads data across what are called nodes in it’s cluster. The data does not have a schema as it is just document/files. HDFS is schemaless, distributed and fault tolerant. MapReduce is focused on data processing and jobs to write to MapReduce are in Java. The operations of a MapReduce job is to find the data and list tasks that it needs to execute and then execute them. The action of executing is called reducers. A downside is that it is batch oriented, which means you would have to read the entire file of data even if you would like to read a small portion of data. Batch oriented is slow. Hadoop is semistructured data and unstructured data, there is no random access for Hadoop and no transaction support. HBase is also called the Hadoop database and unlike Hadoop or HDFS, it has a schema. There is an in-memory feature that gives you the ability to read information quickly. You can isolate the data you want to analyze. HBase is random access. HBase allows for CRUD, which is Creating a new document, Reading the information into an application or process, Update which will allow you to change the value and Delete where mind movement machine. apache hbase. | 2 you can delete data from the system. NoSQL Databases CAP Theorem- Consistency: every read from a database receives the most recent data, Availability: every request receives a non-error- response, Partition tolerance: the system will continue to operate despite the number of arbitrary messages begin dropped between the network between the nodes in a database cluster Each database platform can ONLY have 2 out of the 3. Relational database are consistent and available, Cassandra or CouchDB example databases are available and have built in partition tolerance, Hbase or MongoDB databases are consistent and partition tolerant Hbase can support real time applications like messaging or transactions because it is consistent Visit infoworld to learn more about NoSQL databases. NoSQL Database Structure The data structure is made up of two columns: Keys, which identify the element in the row and Value, which is one or an array of items. Document stores help with reliability because if one database or node in the cluster goes down, there are other copies of it used in the operations that have been stored. mind movement machine. apache hbase. | 3 Graph databases are made above nodes, edges and properties. Neo4j is an example of a graph database. Column families is how Hbase stores its data. Every row has a key and the columns are grouped together, which is called column families. Relational Database Management Systems RDBMS are made up of tables and views, rows and columns, primary keys and foreign keys, data types and using SQL to develop and analyze the database. RDBMS are rigid and you must define a schema that defines the structure of the columns, data types and relationships before using the database. This can result in a slow application development as it is time-consuming. RDBMS store data to eliminate redundancies to prevent anomalies or data that doesn’t match between copies of the record. This is the definition of normalization. This makes the data become slow to analyze as the data has been spread out across many different tables. The solution to this was using a centralized location to store the data, known as a data warehouse which brings all the data into a denormalized format. RDBMS does not scale well because the data does not have the ability to be distributed across multiple data centers or different regions If you do need the ability to scale or support hundreds of users, RBMS are a good choice. NoSQL databases differ as they do not have any rigid system that has to define everything. mind movement machine. apache hbase. | 4 NoSQL are schemaless, distributed and fault tolerant. Visit Codeacademy for a further understanding of Relational Database Management Systems. HBase Interfaces HBase Shell allows you to perform CRUD operations or you can store CRUD commands in a file to read later. HBase Shell has JRuby style references to tables where you can store variables and other operations using JRuby to interact with data in HBase Java API HBase Interface is best used for more complex things than the HBase Shell and if your application is built on Java. Apache Phoenix is an interface that will allow you to enable OLTP and operational analytics on top of HBase. It offers SQL and JDBC interfaces and full acid transaction. It can give you a regular database running on top of HBase and can work with other Apache Hadoop platforms like Spark, Hive, Pig, Flume and MapReduce Column Table Layout Columnar Tables allow you to scan a smaller dataset for faster results. It is optimized for storage and speed and flexibility towards application development. Sparsely populated tables- if you do not have a value for an attribute, then there is nothing stored in the database HBase Terminology A table is a collection of rows, a row is identified by what is called a row key, and inside a row there are different column families. mind movement machine. apache hbase. | 5 In HBase, a namespace are logical groupings of tables and you can refer to them in a similar way you would a relational database. Versioning occurs at the individual value level and HBase has built-in versions, which allow you to look back in time. How is data stored in HBase? HBase Namespaces Run commands from shell: create_namespace ‘people’ to create a namespace for people. Run create ‘people:employees’, ‘work’, ‘demo’ to create employees in people namespace Run drop_namespace ‘people’ to remove a namespace from HBase Run alter_namespace ‘people’, {METHOD => ‘set’, ‘PROPERTY_NAME’ => ‘PROPERTY_VALUE’} to change the namespace There is a default namespace if you do not create one within HBase. Data Model Operations in HBase Get command is used to get a specific row from an HBase table Ex. get ‘tablename’, ‘1’ to get row 1 from table called tablename Scan command will get data from the entire table Ex. scan ‘tablename’ to get all rows from tablename mind movement machine. apache hbase. | 6 Put command will insert data if it does not exist, or update the rowx if it does exist Ex. put ‘tablename’, ‘row:rowID’,’values-to-put-in-row’ Delete command to remove data within a table Ex. delete ‘<table name>’, ‘<row>’, ‘<column name>’, ‘<time stamp>’ Versioning for HBase Versioning is used for historical tracking of the changes made to your data. Each value within a column has a version associated with it. The timestamps within HBase use Unix Timestamps which is the difference in milliseconds between the current time and midnight of January 1 1970. HBase Architecture The HBase architecture is split between the Master and the Slave. Master has the working functions embedded and Slaves respond to the request from the master parts of the system. The Name Node is a part of the Master and is used to maintain information about where the data lives within the HBase cluster. Zookeeper (is not specific to HBase) is within the Master and is used to coordinate the other parts of the system. Zookeeper manages the server state of the cluster and keeps it all connected. It is a coordinator that talks with the HMaster component. mind movement machine. apache hbase. | 7 The HMaster handles all of the region servers and can reassign regions and load-balancing of the cluster. Unless you are using Apache Phoenix, it is likely you are using HMaster to interact with the HBase interface. Slave holds region servers, which is where the regions are and has the HDFS data nodes. All data in HBase is stored on HDFS data nodes. Ensure the HDFS data nodes are in the same place as the region servers. Region and Region Servers Tables are inside of a region, regions are inside of region servers and are scaled horizontally based on key ranges. Every row has a unique row key inside of HBase. All the other values are within columns contained within column families. Region servers can store about 1000 regions. They are the admins of regions. They can be used to split regions or issue compaction, which is used to combine data into the least amount of files as needed to save time when retrieving data. Regions are the basic element of organization for HBase. You want to have 20 to 200 large regions instead of a bunch of small regions, because every region requires memory and having a lot of smaller regions can add overhead just for keeping track of where the data is. One region is from 5 to 20 GB. Regions are used to load-balance data across different region servers after major compaction. Ensure you have a good region design from the start. Coprocessors Coprocessors allow you to make changes and update data as the events occur rather than mind movement machine. apache hbase. | 8 having to pull the data across different nodes in a cluster. You can place custom code on a region server. Coprocessors are an efficient way of processing data in a HBase cluster. Observer coprocessors look for specific changes to the system and executes a command when the change occurs. Endpoint processors are related to endpoint rather than certain specified events.
Recommended publications
  • Administration and Configuration Guide
    Red Hat JBoss Data Virtualization 6.4 Administration and Configuration Guide This guide is for administrators. Last Updated: 2018-09-26 Red Hat JBoss Data Virtualization 6.4 Administration and Configuration Guide This guide is for administrators. Red Hat Customer Content Services Legal Notice Copyright © 2018 Red Hat, Inc. This document is licensed by Red Hat under the Creative Commons Attribution-ShareAlike 3.0 Unported License. If you distribute this document, or a modified version of it, you must provide attribution to Red Hat, Inc. and provide a link to the original. If the document is modified, all Red Hat trademarks must be removed. Red Hat, as the licensor of this document, waives the right to enforce, and agrees not to assert, Section 4d of CC-BY-SA to the fullest extent permitted by applicable law. Red Hat, Red Hat Enterprise Linux, the Shadowman logo, JBoss, OpenShift, Fedora, the Infinity logo, and RHCE are trademarks of Red Hat, Inc., registered in the United States and other countries. Linux ® is the registered trademark of Linus Torvalds in the United States and other countries. Java ® is a registered trademark of Oracle and/or its affiliates. XFS ® is a trademark of Silicon Graphics International Corp. or its subsidiaries in the United States and/or other countries. MySQL ® is a registered trademark of MySQL AB in the United States, the European Union and other countries. Node.js ® is an official trademark of Joyent. Red Hat Software Collections is not formally related to or endorsed by the official Joyent Node.js open source or commercial project.
    [Show full text]
  • Synthesis and Development of a Big Data Architecture for the Management of Radar Measurement Data
    1 Faculty of Electrical Engineering, Mathematics & Computer Science Synthesis and Development of a Big Data architecture for the management of radar measurement data Alex Aalbertsberg Master of Science Thesis November 2018 Supervisors: dr. ir. Maurice van Keulen (University of Twente) prof. dr. ir. Mehmet Aks¸it (University of Twente) dr. Doina Bucur (University of Twente) ir. Ronny Harmanny (Thales) University of Twente P.O. Box 217 7500 AE Enschede The Netherlands Approval Internship report/Thesis of: …………………………………………………………………………………………………………Alexander P. Aalbertsberg Title: …………………………………………………………………………………………Synthesis and Development of a Big Data architecture for the management of radar measurement data Educational institution: ………………………………………………………………………………..University of Twente Internship/Graduation period:…………………………………………………………………………..2017-2018 Location/Department:.…………………………………………………………………………………435 Advanced Development, Delft Thales Supervisor:……………………………………………………………………………R. I. A. Harmanny This report (both the paper and electronic version) has been read and commented on by the supervisor of Thales Netherlands B.V. In doing so, the supervisor has reviewed the contents and considering their sensitivity, also information included therein such as floor plans, technical specifications, commercial confidential information and organizational charts that contain names. Based on this, the supervisor has decided the following: o This report is publicly available (Open). Any defence may take place publicly and the report may be included in public libraries and/or published in knowledge bases. • o This report and/or a summary thereof is publicly available to a limited extent (Thales Group Internal). tors . It will be read and reviewed exclusively by teachers and if necessary by members of the examination board or review ? committee. The content will be kept confidential and not disseminated through publication or inclusion in public libraries and/or knowledge bases.
    [Show full text]
  • Using Apache Phoenix to Store and Access Data Date Published: 2020-02-29 Date Modified: 2020-07-28
    Cloudera Runtime 7.2.1 Using Apache Phoenix to Store and Access Data Date published: 2020-02-29 Date modified: 2020-07-28 https://docs.cloudera.com/ Legal Notice © Cloudera Inc. 2021. All rights reserved. The documentation is and contains Cloudera proprietary information protected by copyright and other intellectual property rights. No license under copyright or any other intellectual property right is granted herein. Copyright information for Cloudera software may be found within the documentation accompanying each component in a particular release. Cloudera software includes software from various open source or other third party projects, and may be released under the Apache Software License 2.0 (“ASLv2”), the Affero General Public License version 3 (AGPLv3), or other license terms. Other software included may be released under the terms of alternative open source licenses. Please review the license and notice files accompanying the software for additional licensing information. Please visit the Cloudera software product page for more information on Cloudera software. For more information on Cloudera support services, please visit either the Support or Sales page. Feel free to contact us directly to discuss your specific needs. Cloudera reserves the right to change any products at any time, and without notice. Cloudera assumes no responsibility nor liability arising from the use of products, except as expressly agreed to in writing by Cloudera. Cloudera, Cloudera Altus, HUE, Impala, Cloudera Impala, and other Cloudera marks are registered or unregistered trademarks in the United States and other countries. All other trademarks are the property of their respective owners. Disclaimer: EXCEPT AS EXPRESSLY PROVIDED IN A WRITTEN AGREEMENT WITH CLOUDERA, CLOUDERA DOES NOT MAKE NOR GIVE ANY REPRESENTATION, WARRANTY, NOR COVENANT OF ANY KIND, WHETHER EXPRESS OR IMPLIED, IN CONNECTION WITH CLOUDERA TECHNOLOGY OR RELATED SUPPORT PROVIDED IN CONNECTION THEREWITH.
    [Show full text]
  • HDP 3.1.4 Release Notes Date of Publish: 2019-08-26
    Release Notes 3 HDP 3.1.4 Release Notes Date of Publish: 2019-08-26 https://docs.hortonworks.com Release Notes | Contents | ii Contents HDP 3.1.4 Release Notes..........................................................................................4 Component Versions.................................................................................................4 Descriptions of New Features..................................................................................5 Deprecation Notices.................................................................................................. 6 Terminology.......................................................................................................................................................... 6 Removed Components and Product Capabilities.................................................................................................6 Testing Unsupported Features................................................................................ 6 Descriptions of the Latest Technical Preview Features.......................................................................................7 Upgrading to HDP 3.1.4...........................................................................................7 Behavioral Changes.................................................................................................. 7 Apache Patch Information.....................................................................................11 Accumulo...........................................................................................................................................................
    [Show full text]
  • Hortonworks Data Platform for Enterprise Data Lakes Delivers Robust, Big Data Analytics That Accelerate Decision Making and Innovation
    IBM United States Software Announcement 218-187, dated March 20, 2018 Hortonworks Data Platform for Enterprise Data Lakes delivers robust, big data analytics that accelerate decision making and innovation Table of contents 1 Overview 5 Publications 2 Key prerequisites 5 Technical information 2 Planned availability date 6 Ordering information 2 Description 7 Terms and conditions 5 Program number 9 Prices 10 Corrections Overview Hortonworks Data Platform is an enterprise ready open source Apache Hadoop distribution based on a centralized architecture supported by YARN. Hortonworks Data Platform is designed to address the needs of data at rest, power real-time customer applications, and deliver big data analytics that can help accelerate decision making and innovation. The official Apache versions for Hortonworks Data Platform V2.6.4 include: • Apache Accumulo 1.7.0 • Apache Atlas 0.8.0 • Apache Calcite 1.2.0 • Apache DataFu 1.3.0 • Apache Falcon 0.10.0 • Apache Flume 1.5.2 • Apache Hadoop 2.7.3 • Apache HBase 1.1.2 • Apache Hive 1.2.1 • Apache Hive 2.1.0 • Apache Kafka 0.10.1 • Apache Knox 0.12.0 • Apache Mahout 0.9.0 • Apache Oozie 4.2.0 • Apache Phoenix 4.7.0 • Apache Pig 0.16.0 • Apache Ranger 0.7.0 • Apache Slider 0.92.0 • Apache Spark 1.6.3 • Apache Spark 2.2.0 • Apache Sqoop 1.4.6 • Apache Storm 1.1.0 • Apache TEZ 0.7.0 • Apache Zeppelin 0.7.3 IBM United States Software Announcement 218-187 IBM is a registered trademark of International Business Machines Corporation 1 • Apache ZooKeeper 3.4.6 IBM(R) clients can download this new offering from Passport Advantage(R).
    [Show full text]
  • Presto: the Definitive Guide
    Presto The Definitive Guide SQL at Any Scale, on Any Storage, in Any Environment Compliments of Matt Fuller, Manfred Moser & Martin Traverso Virtual Book Tour Starburst presents Presto: The Definitive Guide Register Now! Starburst is hosting a virtual book tour series where attendees will: Meet the authors: • Meet the authors from the comfort of your own home Matt Fuller • Meet the Presto creators and participate in an Ask Me Anything (AMA) session with the book Manfred Moser authors + Presto creators • Meet special guest speakers from Martin your favorite podcasts who will Traverso moderate the AMA Register here to save your spot. Praise for Presto: The Definitive Guide This book provides a great introduction to Presto and teaches you everything you need to know to start your successful usage of Presto. —Dain Sundstrom and David Phillips, Creators of the Presto Projects and Founders of the Presto Software Foundation Presto plays a key role in enabling analysis at Pinterest. This book covers the Presto essentials, from use cases through how to run Presto at massive scale. —Ashish Kumar Singh, Tech Lead, Bigdata Query Processing Platform, Pinterest Presto has set the bar in both community-building and technical excellence for lightning- fast analytical processing on stored data in modern cloud architectures. This book is a must-read for companies looking to modernize their analytics stack. —Jay Kreps, Cocreator of Apache Kafka, Cofounder and CEO of Confluent Presto has saved us all—both in academia and industry—countless hours of work, allowing us all to avoid having to write code to manage distributed query processing.
    [Show full text]
  • Research Article Using Distributed Data Over Hbase in Big Data Analytics Platform for Clinical Services
    Hindawi Computational and Mathematical Methods in Medicine Volume 2017, Article ID 6120820, 16 pages https://doi.org/10.1155/2017/6120820 Research Article Using Distributed Data over HBase in Big Data Analytics Platform for Clinical Services Dillon Chrimes1 and Hamid Zamani2 1 Database Integration and Management, IMIT Quality Systems, Vancouver Island Health Authority, Vancouver, BC, Canada V8R 1J8 2School of Health Information Science, Faculty of Human and Social Development, University of Victoria, Victoria, BC,CanadaV8P5C2 Correspondence should be addressed to Dillon Chrimes; [email protected] Received 1 March 2017; Accepted 1 November 2017; Published 11 December 2017 Academic Editor: Yu Xue Copyright © 2017 Dillon Chrimes and Hamid Zamani. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Big data analytics (BDA) is important to reduce healthcare costs. However, there are many challenges of data aggregation, maintenance, integration, translation, analysis, and security/privacy. The study objective to establish an interactive BDA platform with simulated patient data using open-source software technologies was achieved by construction of a platform framework with Hadoop Distributed File System (HDFS) using HBase (key-value NoSQL database). Distributed data structures were generated from benchmarked hospital-specific metadata of nine billion patient records. At optimized iteration, HDFS ingestion of HFiles to HBase store files revealed sustained availability over hundreds of iterations; however, to complete MapReduce to HBase required a week (for 10 TB) and a month for three billion (30 TB) indexed patient records, respectively. Found inconsistencies of MapReduce limited the capacity to generate and replicate data efficiently.
    [Show full text]
  • Hortonworks Data Platform Date of Publish: 2018-09-21
    Release Notes 3 Hortonworks Data Platform Date of Publish: 2018-09-21 http://docs.hortonworks.com Contents HDP 3.0.1 Release Notes..........................................................................................3 Component Versions.............................................................................................................................................3 New Features........................................................................................................................................................ 3 Deprecation Notices..............................................................................................................................................4 Terminology.............................................................................................................................................. 4 Removed Components and Product Capabilities.....................................................................................4 Unsupported Features........................................................................................................................................... 4 Technical Preview Features......................................................................................................................4 Upgrading to HDP 3.0.1...................................................................................................................................... 5 Before you begin.....................................................................................................................................
    [Show full text]
  • Using Apache Phoenix to Store and Access Data 3
    Using Apache Phoenix to store and access data 3 Using Apache Phoenix to store and access data Date of Publish: 2018-08-30 http://docs.hortonworks.com Using Apache Phoenix to store and access data | Contents | ii Contents What's New in Apache Phoenix..............................................................................4 Orchestrating SQL and APIs with Apache Phoenix.............................................4 Enable Phoenix and interdependent components.................................................................................................4 Thin Client connectivity with Phoenix Query Server..........................................................................................5 Secure authentication on the Phoenix Query Server............................................................................... 5 Options to obtain a client driver.......................................................................................................................... 5 Obtaining a driver for application development...................................................................................... 6 Creating and using User-Defined functions (UDFs) in Phoenix.......................... 6 Overview of mapping Phoenix schemas to HBase namespaces............................7 Enable namespace mapping................................................................................................................................. 7 Namespace mapping properties in the hbase-site.xml file.......................................................................7
    [Show full text]
  • Pro Apache Phoenix
    Pro Apache Phoenix An SQL Driver for HBase First Edition Shakil Akhtar Ravi Magham Pro Apache Phoenix: An SQL Driver for HBase Shakil Akhtar Ravi Magham Bangalore, Karnataka Santa Clara, California India USA ISBN-13 (pbk): 978-1-4842-2369-7 ISBN-13 (electronic): 978-1-4842-2370-3 DOI 10.1007/978-1-4842-2370-3 Library of Congress Control Number: 2016961814 Copyright © 2017 by Shakil Akhtar and Ravi Magham This work is subject to copyright. All rights are reserved by the Publisher, whether the whole or part of the material is concerned, specifically the rights of translation, reprinting, reuse of illustrations, recitation, broadcasting, reproduction on microfilms or in any other physical way, and transmission or information storage and retrieval, electronic adaptation, computer software, or by similar or dissimilar methodology now known or hereafter developed. Trademarked names, logos, and images may appear in this book. Rather than use a trademark symbol with every occurrence of a trademarked name, logo, or image we use the names, logos, and images only in an editorial fashion and to the benefit of the trademark owner, with no intention of infringement of the trademark. The use in this publication of trade names, trademarks, service marks, and similar terms, even if they are not identified as such, is not to be taken as an expression of opinion as to whether or not they are subject to proprietary rights. While the advice and information in this book are believed to be true and accurate at the date of publication, neither the authors nor the editors nor the publisher can accept any legal responsibility for any errors or omissions that may be made.
    [Show full text]
  • Hortonworks Data Platform Data Access (August 29, 2016)
    Hortonworks Data Platform Data Access (August 29, 2016) docs.cloudera.com HDP Data Access Guide August 29, 2016 Hortonworks Data Platform: Data Access Copyright © 2012-2016 Hortonworks, Inc. Some rights reserved. The Hortonworks Data Platform, powered by Apache Hadoop, is a massively scalable and 100% open source platform for storing, processing and analyzing large volumes of data. It is designed to deal with data from many sources and formats in a very quick, easy and cost-effective manner. The Hortonworks Data Platform consists of the essential set of Apache Hadoop projects including YARN, Hadoop Distributed File System (HDFS), HCatalog, Pig, Hive, HBase, ZooKeeper and Ambari. Hortonworks is the major contributor of code and patches to many of these projects. These projects have been integrated and tested as part of the Hortonworks Data Platform release process and installation and configuration tools have also been included. Unlike other providers of platforms built using Apache Hadoop, Hortonworks contributes 100% of our code back to the Apache Software Foundation. The Hortonworks Data Platform is Apache-licensed and completely open source. We sell only expert technical support, training and partner-enablement services. All of our technology is, and will remain, free and open source. Please visit the Hortonworks Data Platform page for more information on Hortonworks technology. For more information on Hortonworks services, please visit either the Support or Training page. Feel free to Contact Us directly to discuss your specific needs. Except where otherwise noted, this document is licensed under Creative Commons Attribution ShareAlike 4.0 License. http://creativecommons.org/licenses/by-sa/4.0/legalcode ii HDP Data Access Guide August 29, 2016 Table of Contents 1.
    [Show full text]
  • HPC-ABDS High Performance Computing Enhanced Apache Big Data Stack
    HPC-ABDS High Performance Computing Enhanced Apache Big Data Stack Geoffrey C. Fox, Judy Qiu, Supun Kamburugamuve Shantenu Jha, Andre Luckow School of Informatics and Computing RADICAL Indiana University Rutgers University Bloomington, IN 47408, USA Piscataway, NJ 08854, USA fgcf, xqiu, [email protected] [email protected], [email protected] Abstract—We review the High Performance Computing En- systems as they illustrate key capabilities and often motivate hanced Apache Big Data Stack HPC-ABDS and summarize open source equivalents. the capabilities in 21 identified architecture layers. These The software is broken up into layers so that one can dis- cover Message and Data Protocols, Distributed Coordination, Security & Privacy, Monitoring, Infrastructure Management, cuss software systems in smaller groups. The layers where DevOps, Interoperability, File Systems, Cluster & Resource there is especial opportunity to integrate HPC are colored management, Data Transport, File management, NoSQL, SQL green in figure. We note that data systems that we construct (NewSQL), Extraction Tools, Object-relational mapping, In- from this software can run interoperably on virtualized or memory caching and databases, Inter-process Communication, non-virtualized environments aimed at key scientific data Batch Programming model and Runtime, Stream Processing, High-level Programming, Application Hosting and PaaS, Li- analysis problems. Most of ABDS emphasizes scalability braries and Applications, Workflow and Orchestration. We but not performance and one of our goals is to produce summarize status of these layers focusing on issues of impor- high performance environments. Here there is clear need tance for data analytics. We highlight areas where HPC and for better node performance and support of accelerators like ABDS have good opportunities for integration.
    [Show full text]