Machine Learning with WEKA

Total Page:16

File Type:pdf, Size:1020Kb

Machine Learning with WEKA WEKA: A Machine Machine Learning with Learning Toolkit WEKA The Explorer • Classification and Regression • Clustering Bernhard Pfahringer • Association Rules (based on material by Eibe Frank, Mark • Attribute Selection Hall, and Peter Reutemann) • Data Visualization Department of Computer Science, The Experimenter University of Waikato, New Zealand The Knowledge Flow GUI Other Utilities Conclusions WEKA: the bird The Weka or woodhen (Gallirallus australis) is an endemic bird of New Zealand. (Source: WikiPedia) Copyright: Martin Kramer ([email protected]) 5/14/2007 University of Waikato 2 WEKA: the software Machine learning/data mining software written in Java (distributed under the GNU Public License) Used for research, education, and applications Complements “Data Mining” by Witten & Frank Main features: Comprehensive set of data pre-processing tools, learning algorithms and evaluation methods Graphical user interfaces (incl. data visualization) Environment for comparing learning algorithms 5/14/2007 University of Waikato 3 History Project funded by the NZ government since 1993 Develop state-of-the art workbench of data mining tools Explore fielded applications Develop new fundamental methods 5/14/2007 University of Waikato 4 History (2) Late 1992 - funding was applied for by Ian Witten 1993 - development of the interface and infrastructure WEKA acronym coined by Geoff Holmes WEKA’s file format “ARFF” was created by Andrew Donkin ARFF was rumored to stand for Andrew’s Ridiculous File Format Sometime in 1994 - first internal release of WEKA TCL/TK user interface + learning algorithms written mostly in C Very much beta software Changes for the b1 release included (among others): “Ambiguous and Unsupported menu commands removed.” “Crashing processes handled (in most cases :-)” October 1996 - first public release: WEKA 2.1 5/14/2007 University of Waikato 5 History (3) July 1997 - WEKA 2.2 Schemes: 1R, T2, K*, M5, M5Class, IB1-4, FOIL, PEBLS, support for C5 Included a facility (based on Unix makefiles) for configuring and running large scale experiments Early 1997 - decision was made to rewrite WEKA in Java Originated from code written by Eibe Frank for his PhD Originally codenamed JAWS (JAva Weka System) May 1998 - WEKA 2.3 Last release of the TCL/TK-based system Mid 1999 - WEKA 3 (100% Java) released Version to complement the Data Mining book Development version (including GUI) 5/14/2007 University of Waikato 6 The GUI back then… TCL/TK interface of Weka 2.1 5/14/2007 University of Waikato 7 WEKA: versions There are several versions of WEKA: WEKA 3.4: “book version” compatible with description in data mining book WEKA 3.5.5: “development version” with lots of improvements This talk is based on a nightly snapshot of WEKA 3.5.5 (12-Feb-2007) 5/14/2007 University of Waikato 8 WEKA only deals with “flat” files @relation heart-disease-simplified @attribute age numeric @attribute sex { female, male} @attribute chest_pain_type { typ_angina, asympt, non_anginal, atyp_angina} @attribute cholesterol numeric @attribute exercise_induced_angina { no, yes} @attribute class { present, not_present} @data 63,male,typ_angina,233,no,not_present 67,male,asympt,286,yes,present 67,male,asympt,229,yes,present 38,female,non_anginal,?,no,not_present ... 5/14/2007 University of Waikato 9 WEKA only deals with “flat” files @relation heart-disease-simplified @attribute age numeric @attribute sex { female, male} @attribute chest_pain_type { typ_angina, asympt, non_anginal, atyp_angina} @attribute cholesterol numeric @attribute exercise_induced_angina { no, yes} @attribute class { present, not_present} @data 63,male,typ_angina,233,no,not_present 67,male,asympt,286,yes,present 67,male,asympt,229,yes,present 38,female,non_anginal,?,no,not_present ... 5/14/2007 University of Waikato 10 java weka.gui.GUIChooser 5/14/2007 University of Waikato 11 5/14/2007 University of Waikato 12 5/14/2007 University of Waikato 13 java -jar weka.jar 5/14/2007 University of Waikato 14 Explorer: pre-processing the data Data can be imported from a file in various formats: ARFF, CSV, C4.5, binary Data can also be read from a URL or from an SQL database (using JDBC) Pre-processing tools in WEKA are called “filters” WEKA contains filters for: Discretization, normalization, resampling, attribute selection, transforming and combining attributes, … 5/14/2007 University of Waikato 15 5/14/2007 University of Waikato 16 5/14/2007 University of Waikato 17 5/14/2007 University of Waikato 18 5/14/2007 University of Waikato 19 5/14/2007 University of Waikato 20 5/14/2007 University of Waikato 21 5/14/2007 University of Waikato 22 5/14/2007 University of Waikato 23 5/14/2007 University of Waikato 24 5/14/2007 University of Waikato 25 5/14/2007 University of Waikato 26 5/14/2007 University of Waikato 27 5/14/2007 University of Waikato 28 5/14/2007 University of Waikato 29 5/14/2007 University of Waikato 30 5/14/2007 University of Waikato 31 5/14/2007 University of Waikato 32 5/14/2007 University of Waikato 33 5/14/2007 University of Waikato 34 5/14/2007 University of Waikato 35 5/14/2007 University of Waikato 36 Explorer: building “classifiers” Classifiers in WEKA are models for predicting nominal or numeric quantities Implemented learning schemes include: Decision trees and lists, instance-based classifiers, support vector machines, multi-layer perceptrons, logistic regression, Bayes’ nets, … “Meta”-classifiers include: Bagging, boosting, stacking, error-correcting output codes, locally weighted learning, … 5/14/2007 University of Waikato 37 5/14/2007 University of Waikato 38 5/14/2007 University of Waikato 39 5/14/2007 University of Waikato 40 5/14/2007 University of Waikato 41 5/14/2007 University of Waikato 42 5/14/2007 University of Waikato 43 5/14/2007 University of Waikato 44 5/14/2007 University of Waikato 45 5/14/2007 University of Waikato 46 5/14/2007 University of Waikato 47 5/14/2007 University of Waikato 48 5/14/2007 University of Waikato 49 5/14/2007 University of Waikato 50 5/14/2007 University of Waikato 51 5/14/2007 University of Waikato 52 5/14/2007 University of Waikato 53 5/14/2007 University of Waikato 54 5/14/2007 University of Waikato 55 5/14/2007 University of Waikato 56 5/14/2007 University of Waikato 57 5/14/2007 University of Waikato 58 5/14/2007 University of Waikato 59 5/14/2007 University of Waikato 60 5/14/2007 University of Waikato 61 5/14/2007 University of Waikato 62 5/14/2007 University of Waikato 63 5/14/2007 University of Waikato 64 5/14/2007 University of Waikato 65 5/14/2007 University of Waikato 66 5/14/2007 University of Waikato 67 5/14/2007 University of Waikato 68 5/14/2007 University of Waikato 69 5/14/2007 University of Waikato 70 5/14/2007 University of Waikato 71 QuickTime?and5/14/2007 a TIFF (LZW) decompressor are needed to see this picture. University of Waikato 72 QuickTime?and5/14/2007 a TIFF (LZW) decompressor are needed to see this picture. University of Waikato 73 5/14/2007 University of Waikato 74 5/14/2007 University of Waikato 75 5/14/2007 University of Waikato 76 5/14/2007 University of Waikato 77 QuickTime?and a TIFF (LZW) decompressor are needed to see this pictu 5/14/2007 University of Waikato 78 5/14/2007 University of Waikato 79 5/14/2007 University of Waikato 80 5/14/2007 University of Waikato 81 5/14/2007 University of Waikato 82 5/14/2007 University of Waikato 83 5/14/2007 University of Waikato 84 5/14/2007 University of Waikato 85 5/14/2007 University of Waikato 86 5/14/2007 University of Waikato 87 5/14/2007 University of Waikato 88 5/14/2007 University of Waikato 89 5/14/2007 University of Waikato 90 5/14/2007 University of Waikato 91 5/14/2007 University of Waikato 92 Explorer: clustering data WEKA contains “clusterers” for finding groups of similar instances in a dataset Some implemented schemes are: k-Means, EM, Cobweb, X-means, FarthestFirst Clusters can be visualized and compared to “true” clusters (if given) Evaluation based on loglikelihood if clustering scheme produces a probability distribution 5/14/2007 University of Waikato 93 5/14/2007 University of Waikato 94 5/14/2007 University of Waikato 95 5/14/2007 University of Waikato 96 5/14/2007 University of Waikato 97 5/14/2007 University of Waikato 98 5/14/2007 University of Waikato 99 5/14/2007 University of Waikato 100 5/14/2007 University of Waikato 101 5/14/2007 University of Waikato 102 Explorer: finding associations WEKA contains the Apriori algorithm (among others) for learning association rules Works only with discrete data Can identify statistical dependencies between groups of attributes: milk, butter ⇒ bread, eggs (with confidence 0.9 and support 2000) Apriori can compute all rules that have a given minimum support and exceed a given confidence 5/14/2007 University of Waikato 103 5/14/2007 University of Waikato 104 5/14/2007 University of Waikato 105 5/14/2007 University of Waikato 106 Explorer: attribute selection Panel that can be used to investigate which (subsets of) attributes are the most predictive ones Attribute selection methods contain two parts: A search method: best-first, forward selection, random, exhaustive, genetic algorithm, ranking An evaluation method: correlation-based, wrapper, information gain, chi-squared, … Very flexible: WEKA allows (almost) arbitrary combinations of these two 5/14/2007 University of Waikato 107 5/14/2007 University of Waikato 108 5/14/2007 University of Waikato 109 5/14/2007 University of Waikato 110 5/14/2007 University of Waikato 111 5/14/2007 University of Waikato 112 5/14/2007 University of Waikato 113 Explorer: data visualization Visualization very useful in practice: e.g.
Recommended publications
  • Data Mining – Intro
    Data warehouse& Data Mining UNIT-3 Syllabus • UNIT 3 • Classification: Introduction, decision tree, tree induction algorithm – split algorithm based on information theory, split algorithm based on Gini index; naïve Bayes method; estimating predictive accuracy of classification method; classification software, software for association rule mining; case study; KDD Insurance Risk Assessment What is Data Mining? • Data Mining is: (1) The efficient discovery of previously unknown, valid, potentially useful, understandable patterns in large datasets (2) Data mining is the analysis step of the "knowledge discovery in databases" process, or KDD (2) The analysis of (often large) observational data sets to find unsuspected relationships and to summarize the data in novel ways that are both understandable and useful to the data owner Knowledge Discovery Examples of Large Datasets • Government: IRS, NGA, … • Large corporations • WALMART: 20M transactions per day • MOBIL: 100 TB geological databases • AT&T 300 M calls per day • Credit card companies • Scientific • NASA, EOS project: 50 GB per hour • Environmental datasets KDD The Knowledge Discovery in Databases (KDD) process is commonly defined with the stages: (1) Selection (2) Pre-processing (3) Transformation (4) Data Mining (5) Interpretation/Evaluation Data Mining Methods 1. Decision Tree Classifiers: Used for modeling, classification 2. Association Rules: Used to find associations between sets of attributes 3. Sequential patterns: Used to find temporal associations in time series 4. Hierarchical
    [Show full text]
  • The Pentaho Big Data Guide This Document Supports Pentaho Business Analytics Suite 4.8 GA and Pentaho Data Integration 4.4 GA, Documentation Revision October 31, 2012
    The Pentaho Big Data Guide This document supports Pentaho Business Analytics Suite 4.8 GA and Pentaho Data Integration 4.4 GA, documentation revision October 31, 2012. This document is copyright © 2012 Pentaho Corporation. No part may be reprinted without written permission from Pentaho Corporation. All trademarks are the property of their respective owners. Help and Support Resources If you have questions that are not covered in this guide, or if you would like to report errors in the documentation, please contact your Pentaho technical support representative. Support-related questions should be submitted through the Pentaho Customer Support Portal at http://support.pentaho.com. For information about how to purchase support or enable an additional named support contact, please contact your sales representative, or send an email to [email protected]. For information about instructor-led training on the topics covered in this guide, visit http://www.pentaho.com/training. Limits of Liability and Disclaimer of Warranty The author(s) of this document have used their best efforts in preparing the content and the programs contained in it. These efforts include the development, research, and testing of the theories and programs to determine their effectiveness. The author and publisher make no warranty of any kind, express or implied, with regard to these programs or the documentation contained in this book. The author(s) and Pentaho shall not be liable in the event of incidental or consequential damages in connection with, or arising out of, the furnishing, performance, or use of the programs, associated instructions, and/or claims. Trademarks Pentaho (TM) and the Pentaho logo are registered trademarks of Pentaho Corporation.
    [Show full text]
  • Star Schema Modeling with Pentaho Data Integration
    Star Schema Modeling With Pentaho Data Integration Saurischian and erratic Salomo underworked her accomplishment deplumes while Phil roping some diamonds believingly. Torrence elasticize his umbrageousness parsed anachronously or cheaply after Rand pensions and darn postally, canalicular and papillate. Tymon trodden shrinkingly as electropositive Horatius cumulates her salpingectomies moat anaerobiotically. The email providers have a look at pentaho user console files from a collection, an individual industries such processes within an embedded saiku report manager. The database connections in data modeling with schema. Entity Relationship Diagram ERD star schema Data original database creation. For more details, the proposed DW system ran on a Windowsbased server; therefore, it responds very slowly to new analytical requirements. In this section we'll introduce modeling via cubes and children at place these models are derived. The data presentation level is the interface between the system and the end user. Star Schema Modeling with Pentaho Data Integration Tutorial Details In order first write to XML file we pass be using the XML Output quality This is. The class must implement themondrian. Modeling approach using the dimension tables and fact tables 1 Introduction The use. Data Warehouse Dimensional Model Star Schema OLAP Cube 5. So that will not create a lot when it into. But it will create transformations on inventory transaction concepts, integrated into several study, you will likely send me? Thoughts on open Vault vs Star Schemas the bi backend. Table elements is data integration tool which are created all the design to the farm before with delivering aggregated data quality and data is preventing you.
    [Show full text]
  • A Plan for an Early Childhood Integrated Data System in Oklahoma
    A PLAN FOR AN EARLY CHILDHOOD INTEGRATED DATA SYSTEM IN OKLAHOMA: DATA INVENTORY, DATA INTEGRATION PLAN, AND DATA GOVERNANCE PLAN January 31, 2020 The Oklahoma Partnership for School Readiness would like to acknowledge the Oklahoma Early Childhood Integrated Data System (ECIDS) Project Oversight Committee for advising and supporting development of this plan: Steve Buck, Cabinet Secretary of Human Services and Early Childhood Initiatives Jennifer Dalton, Oklahoma Department of Human Services Erik Friend, Oklahoma State Department of Education Becki Moore, Oklahoma State Department of Health Funding for development of this plan was provided by the Preschool Development Grant Birth through Five (Grant Number 90TP0037), a grantmaking program of the U.S. Department of Health and Human Services, Administration for Children and Families, Office of Child Care. 2 Contents Glossary ......................................................................................................................................................... 6 Image Credits .............................................................................................................................................. 14 1. Executive Summary ............................................................................................................................. 15 1.1. Uses of an ECIDS ......................................................................................................................... 15 1.2. About this ECIDS Plan .................................................................................................................
    [Show full text]
  • Base Handbook Copyright
    Version 4.0 Base Handbook Copyright This document is Copyright © 2013 by its contributors as listed below. You may distribute it and/or modify it under the terms of either the GNU General Public License (http://www.gnu.org/licenses/gpl.html), version 3 or later, or the Creative Commons Attribution License (http://creativecommons.org/licenses/by/3.0/), version 3.0 or later. All trademarks within this guide belong to their legitimate owners. Contributors Jochen Schiffers Robert Großkopf Jost Lange Hazel Russman Martin Fox Andrew Pitonyak Dan Lewis Jean Hollis Weber Acknowledgments This book is based on an original German document, which was translated by Hazel Russman and Martin Fox. Feedback Please direct any comments or suggestions about this document to: [email protected] Publication date and software version Published 3 July 2013. Based on LibreOffice 4.0. Documentation for LibreOffice is available at http://www.libreoffice.org/get-help/documentation Contents Copyright..................................................................................................................................... 2 Contributors.............................................................................................................................2 Feedback................................................................................................................................ 2 Acknowledgments................................................................................................................... 2 Publication
    [Show full text]
  • Open Source ETL on the Mainframe
    2011 JPMorgan Chase ROBERT ZWINK , VP Implementation Services, Chief Development Office [RUNNING OPEN SOURCE ETL ON A MAINFRAME] Pentaho is an open source framework written in Java which includes a full featured Extract Transform Load (ETL) tool called Pentaho Data Integration (PDI). Programmers leverage PDI to create custom transformations which can be a direct 1:1 translation of existing COBOL. A rich palette of out of the box components allows the transformation to be assembled visually. Once finished, the transformation is a completely portable Java application, written in a visual programming language, which runs fully within a java virtual machine (JVM). Java programs created by PDI are 100% zAAP eligible. Contents ABSTRACT ........................................................................................................................................ 3 GENERAL TERMS ............................................................................................................................. 3 INTRODUCTION ............................................................................................................................... 3 BACKGROUND ................................................................................................................................. 4 Assumptions and Requirements ................................................................................................. 4 Chargeback Model .....................................................................................................................
    [Show full text]
  • Chapter 6 Reports Copyright
    Base Handbook Chapter 6 Reports Copyright This document is Copyright © 2013 by its contributors as listed below. You may distribute it and/or modify it under the terms of either the GNU General Public License (http://www.gnu.org/licenses/gpl.html), version 3 or later, or the Creative Commons Attribution License (http://creativecommons.org/licenses/by/3.0/), version 3.0 or later. All trademarks within this guide belong to their legitimate owners. Contributors Robert Großkopf Jost Lange Jochen Schiffers Hazel Russman Jean Hollis Weber Feedback Please direct any comments or suggestions about this document to: [email protected]. Caution Everything you send to a mailing list, including your email address and any other personal information that is in the mail, is publicly archived and can not be deleted. Acknowledgments This chapter is based on an original German document and was translated by Hazel Russman. Publication date and software version Published 22 April 2013. Based on LibreOffice 3.5. Note for Mac users Some keystrokes and menu items are different on a Mac from those used in Windows and Linux. The table below gives some common substitutions for the instructions in this chapter. For a more detailed list, see the application Help. Windows or Linux Mac equivalent Effect Tools > Options menu LibreOffice > Preferences Access setup options selection Right-click Control+click Opens a context menu Ctrl (Control) z (Command) Used with other keys F5 Shift+z+F5 Opens the Navigator F11 z+T Opens the Styles and Formatting window Documentation for LibreOffice is available at http://www.libreoffice.org/get-help/documentation Contents Copyright ...........................................................................................................................
    [Show full text]
  • Frequent Item Set Mining Using INC MINE in Massive Online Analysis Frame Work
    Available online at www.sciencedirect.com ScienceDirect Procedia Computer Science 45 ( 2015 ) 133 – 142 International Conference on Advanced Computing Technologies and Applications (ICACTA- 2015) Frequent Item set Mining using INC_MINE in Massive Online Analysis Frame work Prof.Dr.P.K.Srimania, Mrs. Malini M. Patilb* aFormer Chairman and Director, R & D, Bangalore University, Karnataka, India bAssistant Professor , Dept of ISE , J.S.S. Academy of Technical Education, Bangalore-560060, Karnataka, India Research Scholar, Bharthiar University, Coimbatore, Tamilnadu Abstract Frequent Pattern Mining is one of the major data mining techniques, which is exhaustively studied in the past decade. The technological advancements have resulted in huge data generation, having increased rate of data distribution. The generated data is called as a 'data stream'. Data streams can be mined only by using sophisticated techniques. The paper aims at carrying out frequent pattern mining on data streams. Stream mining has great challenges due to high memory usage and computational costs. Massive online analysis frame work is a software environment used to perform frequent pattern mining using INC_MINE algorithm. The algorithm uses the method of closed frequent mining. The data sets used in the analysis are Electricity data set and Airline data set. The authors also generated their own data set, OUR-GENERATOR for the purpose of analysis and the results are found interesting. In the experiments five samples of instance sizes (10000, 15000, 25000, 35000, 50000) are used with varying minimum support and window sizes for determining frequent closed itemsets and semi frequent closed itemsets respectively. The present work establishes that association rule mining could be performed even in the case of data stream mining by INC_MINE algorithm by generating closed frequent itemsets which is first of its kind in the literature.
    [Show full text]
  • Pentaho Machine Learning Orchestration
    Pentaho Machine Learning Orchestration DATASHEET Pentaho from Hitachi Vantara streamlines the entire machine learning workflow and enables teams of data scientists, engineers and analysts to train, tune, test and deploy predictive models. Pentaho Data Integration and analytics platform ends the ‘gridlock’ 2 Train, Tune and Test Models associated with machine learning by enabling smooth team col- Data scientists often apply trial and error to strike the right balance laboration, maximizing limited data science resources and putting of complexity, performance and accuracy in their models. With predictive models to work on big data faster — regardless of use integrations for languages like R and Python, and for machine case, industry, or language — whether models were built in R, learning libraries like Spark MLlib and Weka, Pentaho allows data Python, Scala or Weka. scientists to seamlessly train, tune, build and test models faster. Streamline Four Areas of the Machine 3 Deploy and Operationalize Models Learning Workflow Pentaho allows data professionals to easily embed models devel- Most enterprises struggle to put models to work because data oped by a data scientist directly in an operational workflow. They professionals often operate in silos and create bottlenecks in can leverage existing data and feature engineering efforts, sig- the data preparation to model updates workflow. The Pentaho nificantly reducing time-to-deployment. With embeddable APIs, platform enables collaboration and removes bottlenecks in four organizations can also include the full power of Pentaho within key areas: existing applications. 1 Prepare Data and Engineer New Features 4 Update Models Regularly Pentaho makes it easy to prepare and blend traditional sources Ventana Research finds that less than a third (31%) of organizations like ERP and CRM with big data sources like sensors and social use an automated process to update their models.
    [Show full text]
  • Performance Analysis of Hoeffding Trees in Data Streams by Using Massive Online Analysis Framework
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by ePrints@Bangalore University Int. J. Data Mining, Modelling and Management, Vol. 7, No. 4, 2015 293 Performance analysis of Hoeffding trees in data streams by using massive online analysis framework P.K. Srimani R & D Division, Bangalore University Jnana Bharathi, Mysore Road, Bangalore-560056, Karnataka, India Email: [email protected] Malini M. Patil* Department of Information Science and Engineering, J.S.S. Academy of Technical Education, Uttaralli-Kengeri Main Road, Mylasandra, Bangalore-560060, Karnataka, India Email: [email protected] *Corresponding author Abstract: Present work is mainly concerned with the understanding of the problem of classification from the data stream perspective on evolving streams using massive online analysis framework with regard to different Hoeffding trees. Advancement of the technology both in the area of hardware and software has led to the rapid storage of data in huge volumes. Such data is referred to as a data stream. Traditional data mining methods are not capable of handling data streams because of the ubiquitous nature of data streams. The challenging task is how to store, analyse and visualise such large volumes of data. Massive data mining is a solution for these challenges. In the present analysis five different Hoeffding trees are used on the available eight dataset generators of massive online analysis framework and the results predict that stagger generator happens to be the best performer for different classifiers. Keywords: data mining; data streams; static streams; evolving streams; Hoeffding trees; classification; supervised learning; massive online analysis; MOA; framework; massive data mining; MDM; dataset generators.
    [Show full text]
  • Massive Online Analysis, a Framework for Stream Classification and Clustering
    MOA: Massive Online Analysis, a Framework for Stream Classification and Clustering. Albert Bifet1, Geoff Holmes1, Bernhard Pfahringer1, Philipp Kranen2, Hardy Kremer2, Timm Jansen2, and Thomas Seidl2 1 Department of Computer Science, University of Waikato, Hamilton, New Zealand fabifet, geoff, [email protected] 2 Data Management and Exploration Group, RWTH Aachen University, Germany fkranen, kremer, jansen, [email protected] Abstract. In today's applications, massive, evolving data streams are ubiquitous. Massive Online Analysis (MOA) is a software environment for implementing algorithms and running experiments for online learn- ing from evolving data streams. MOA is designed to deal with the chal- lenging problems of scaling up the implementation of state of the art algorithms to real world dataset sizes and of making algorithms compa- rable in benchmark streaming settings. It contains a collection of offline and online algorithms for both classification and clustering as well as tools for evaluation. Researchers benefit from MOA by getting insights into workings and problems of different approaches, practitioners can easily compare several algorithms and apply them to real world data sets and settings. MOA supports bi-directional interaction with WEKA, the Waikato Environment for Knowledge Analysis, and is released under the GNU GPL license. Besides providing algorithms and measures for evaluation and comparison, MOA is easily extensible with new contri- butions and allows the creation of benchmark scenarios through storing and sharing setting files. 1 Introduction Nowadays data is generated at an increasing rate from sensor applications, mea- surements in network monitoring and traffic management, log records or click- streams in web exploring, manufacturing processes, call detail records, email, blogging, twitter posts and others.
    [Show full text]
  • Pentaho MAPR510 SHIM 7.1.0.0 Open Source Software Packages
    Pentaho MAPR510 SHIM 7.1.0.0 Open Source Software Packages Contact Information: Project Manager Pentaho MAPR519 SHIM Hitachi Vantara Corporation 2535 Augustine Drive Santa Clara, California 95054 Name of Product/Product Version License Component Apache Thrift 0.9.2 Apache License Version 2.0 Automaton 1.11-8 automation.bsd.2.0 hbase-client-1.1.1-mapr-1602 for 1.1.1-mapr-1602 Apache License Version 2.0 MapR 5.1 shim hbase-common-1.1.1-mapr-1602 for 1.1.1-mapr-1602 Apache License Version 2.0 MapR 5.1 shim hbase-hadoop-compat-1.1.1-mapr- 1.1.1-mapr-1602 Apache License Version 2.0 1602 for MapR 5.1 shim hbase-protocol-1.1.1-mapr-1602 for 1.1.1-mapr-1602 Apache License Version 2.0 MapR 5.1 shim hbase-server-1.1.1-mapr-1602 for 1.1.1-mapr-1602 Apache License Version 2.0 MapR 5.1 shim hive-common-1.2.0-mapr-1605 for 1.2.0-mapr-1605 Apache License Version 2.0 MapR 5.1 shim hive-exec-1.2.0-mapr-1605 for MapR 1.2.0-mapr-1605 Apache License Version 2.0 5.1 shim hive-jdbc-1.2.0-mapr-1605 for MapR 1.2.0-mapr-1605 Apache License Version 2.0 5.1 shim hive-metastore-1.2.0-mapr-1605 for 1.2.0-mapr-1605 Apache License Version 2.0 MapR 5.1 shim Name of Product/Product Version License Component hive-service-1.2.0-mapr-1605 for 1.2.0-mapr-1605 Apache License Version 2.0 MapR 5.1 shim hive-shims-0.23-1.2.0-mapr-1605 for 1.2.0-mapr-1605 Apache License Version 2.0 MapR 5.1 shim hive-shims-common-1.2.0-mapr-1605 1.2.0-mapr-1605 Apache License Version 2.0 for MapR 5.1 shim htrace-core 3.1.0-incubating Apache License Version 2.0 Metrics Core Library 2.2.0 Apache
    [Show full text]