DATA STREAM MINING a Practical Approach

Total Page:16

File Type:pdf, Size:1020Kb

DATA STREAM MINING a Practical Approach DATA STREAM MINING A Practical Approach Albert Bifet and Richard Kirkby August 2009 Contents I Introduction and Preliminaries3 1 Preliminaries5 1.1 MOA Stream Mining........................5 1.2 Assumptions.............................7 1.3 Requirements............................8 1.4 Mining Strategies.......................... 10 1.5 Change Detection Strategies.................... 13 2 MOA Experimental Setting 17 2.1 Previous Evaluation Practices................... 19 2.1.1 Batch Setting........................ 19 2.1.2 Data Stream Setting.................... 21 2.2 Evaluation Procedures for Data Streams............. 25 2.2.1 Holdout........................... 26 2.2.2 Interleaved Test-Then-Train or Prequential....... 26 2.2.3 Comparison......................... 27 2.3 Testing Framework......................... 28 2.4 Environments............................ 30 2.4.1 Sensor Network....................... 31 2.4.2 Handheld Computer.................... 31 2.4.3 Server............................ 31 2.5 Data Sources............................. 32 2.5.1 Random Tree Generator.................. 33 2.5.2 Random RBF Generator.................. 34 2.5.3 LED Generator....................... 34 2.5.4 Waveform Generator.................... 34 2.5.5 Function Generator..................... 34 2.6 Generation Speed and Data Size................. 35 2.7 Evolving Stream Experimental Setting.............. 39 2.7.1 Concept Drift Framework................. 40 2.7.2 Datasets for concept drift................. 42 II Stationary Data Stream Learning 45 3 Hoeffding Trees 47 3.1 The Hoeffding Bound for Tree Induction............. 48 3.2 The Basic Algorithm........................ 50 3.2.1 Split Confidence...................... 51 3.2.2 Sufficient Statistics..................... 51 3.2.3 Grace Period........................ 52 3.2.4 Pre-pruning......................... 53 i CONTENTS 3.2.5 Tie-breaking......................... 54 3.2.6 Skewed Split Prevention.................. 55 3.3 Memory Management....................... 55 3.3.1 Poor Attribute Removal.................. 60 3.4 MOA Java Implementation Details................ 60 3.4.1 Fast Size Estimates..................... 63 4 Numeric Attributes 65 4.1 Batch Setting Approaches..................... 66 4.1.1 Equal Width......................... 67 4.1.2 Equal Frequency...................... 67 4.1.3 k-means Clustering..................... 67 4.1.4 Fayyad and Irani...................... 68 4.1.5 C4.5.............................. 68 4.2 Data Stream Approaches...................... 69 4.2.1 VFML Implementation................... 70 4.2.2 Exhaustive Binary Tree................... 71 4.2.3 Quantile Summaries.................... 72 4.2.4 Gaussian Approximation................. 73 4.2.5 Numerical Interval Pruning................ 77 5 Prediction Strategies 79 5.1 Majority Class............................ 79 5.2 Naive Bayes Leaves......................... 79 5.3 Adaptive Hybrid.......................... 82 6 Hoeffding Tree Ensembles 85 6.1 Batch Setting............................. 89 6.1.1 Bagging........................... 89 6.1.2 Boosting........................... 90 6.1.3 Option Trees......................... 95 6.2 Data Stream Setting......................... 97 6.2.1 Bagging........................... 97 6.2.2 Boosting........................... 98 6.2.3 Option Trees......................... 101 6.3 Realistic Ensemble Sizes...................... 106 III Evolving Data Stream Learning 109 7 Evolving data streams 111 7.1 Algorithms for mining with change............... 112 7.1.1 OLIN: Last.......................... 112 7.1.2 CVFDT: Domingos..................... 112 7.1.3 UFFT: Gama......................... 113 7.2 A Methodology for Adaptive Stream Mining.......... 115 ii CONTENTS 7.2.1 Time Change Detectors and Predictors: A General Frame- work............................. 116 7.2.2 Window Management Models.............. 118 7.3 Optimal Change Detector and Predictor............. 120 8 Adaptive Sliding Windows 121 8.1 Introduction............................. 121 8.2 Maintaining Updated Windows of Varying Length....... 122 8.2.1 Setting............................ 122 8.2.2 First algorithm: ADWIN0 ................. 122 8.2.3 ADWIN0 for Poisson processes.............. 127 8.2.4 Improving time and memory requirements....... 128 8.3 K-ADWIN = ADWIN + Kalman Filtering.............. 132 9 Adaptive Hoeffding Trees 135 9.1 Introduction............................. 135 9.2 Decision Trees on Sliding Windows................ 136 9.2.1 HWT-ADWIN : Hoeffding Window Tree using ADWIN . 136 9.2.2 CVFDT............................ 139 9.3 Hoeffding Adaptive Trees..................... 140 9.3.1 Example of performance Guarantee........... 141 9.3.2 Memory Complexity Analysis.............. 142 10 Adaptive Ensemble Methods 143 10.1 New method of Bagging using trees of different size...... 143 10.2 New method of Bagging using ADWIN .............. 146 10.3 Adaptive Hoeffding Option Trees................. 147 10.4 Method performance........................ 147 Bibliography 149 iii Introduction Massive Online Analysis (MOA) is a software environment for implement- ing algorithms and running experiments for online learning from evolving data streams. MOA includes a collection of offline and online methods as well as tools for evaluation. In particular, it implements boosting, bagging, and Hoeffd- ing Trees, all with and without Na¨ıve Bayes classifiers at the leaves. MOA is related to WEKA, the Waikato Environment for Knowledge Anal- ysis, which is an award-winning open-source workbench containing imple- mentations of a wide range of batch machine learning methods. WEKA is also written in Java. The main benefits of Java are portability, where applica- tions can be run on any platform with an appropriate Java virtual machine, and the strong and well-developed support libraries. Use of the language is widespread, and features such as the automatic garbage collection help to reduce programmer burden and error. This text explains the theoretical and practical foundations of the meth- ods and streams available in MOA. The moa and the weka are both birds na- tive to New Zealand. The weka is a cheeky bird of similar size to a chicken. The moa was a large ostrich-like bird, an order of magnitude larger than a weka, that was hunted to extinction. Part I Introduction and Preliminaries 3 1 Preliminaries In today’s information society, extraction of knowledge is becoming a very important task for many people. We live in an age of knowledge revolution. Peter Drucker [42], an influential management expert, writes “From now on, the key is knowledge. The world is not becoming labor intensive, not material intensive, not energy intensive, but knowledge intensive”. This knowledge revolution is based in an economic change from adding value by producing things which is, ultimately limited, to adding value by creating and using knowledge which can grow indefinitely. The digital universe in 2007 was estimated in [64] to be 281 exabytes or 281 billion gigabytes, and by 2011, the digital universe will be 10 times the size it was 5 years before. The amount of information created, or captured exceeded available storage for the first time in 2007. To deal with these huge amount of data in a responsible way, green com- puting is becoming a necessity. Green computing is the study and practice of using computing resources efficiently. A main approach to green comput- ing is based on algorithmic efficiency. The amount of computer resources required for any given computing function depends on the efficiency of the algorithms used. As the cost of hardware has declined relative to the cost of energy, the energy efficiency and environmental impact of computing sys- tems and programs are receiving increased attention. 1.1 MOA Stream Mining A largely untested hypothesis of modern society is that it is important to record data as it may contain valuable information. This occurs in almost all facets of life from supermarket checkouts to the movements of cows in a paddock. To support the hypothesis, engineers and scientists have pro- duced a raft of ingenious schemes and devices from loyalty programs to RFID tags. Little thought however, has gone into how this quantity of data might be analyzed. Machine learning, the field for finding ways to automatically extract infor- mation from data, was once considered the solution to this problem. Histor- ically it has concentrated on learning from small numbers of examples, be- cause only limited amounts of data were available when the field emerged. Some very sophisticated algorithms have resulted from the research that 5 CHAPTER 1. PRELIMINARIES can learn highly accurate models from limited training examples. It is com- monly assumed that the entire set of training data can be stored in working memory. More recently the need to process larger amounts of data has motivated the field of data mining. Ways are investigated to reduce the computation time and memory needed to process large but static data sets. If the data cannot fit into memory, it may be necessary to sample a smaller training set. Alternatively, algorithms may resort to temporary external storage, or only process subsets of data at a time. Commonly the goal is to create a learning process that is linear in the number of examples. The essential learning procedure is treated like a scaled up version of classic machine learning, where learning is considered a single, possibly expensive, operation—a set of training examples are processed to output a final static model. The data
Recommended publications
  • Concept Drift Adaptation Techniques in Distributed Environment for Real-World Data Streams
    smart cities Article Concept Drift Adaptation Techniques in Distributed Environment for Real-World Data Streams Hassan Mehmood 1,* , Panos Kostakos 1, Marta Cortes 1, Theodoros Anagnostopoulos 2 , Susanna Pirttikangas 1 and Ekaterina Gilman 1 1 Center for Ubiquitous Computing, University of Oulu, Pentti Kaiteran katu 1, 90570 Oulu, Finland; Panos.Kostakos@oulu.fi (P.K.); Marta.Cortes@oulu.fi (M.C.); Susanna.Pirttikangas@oulu.fi (S.P.); Ekaterina.Gilman@oulu.fi (E.G.) 2 DigiT.DSS.Lab, Department of Business Administration, University of West Attica, P. Ralli & Thivon 250, Aigaleo, 122 44 Athens, Greece; [email protected] * Correspondence: Hassan.Mehmood@oulu.fi Abstract: Real-world data streams pose a unique challenge to the implementation of machine learning (ML) models and data analysis. A notable problem that has been introduced by the growth of Internet of Things (IoT) deployments across the smart city ecosystem is that the statistical properties of data streams can change over time, resulting in poor prediction performance and ineffective decisions. While concept drift detection methods aim to patch this problem, emerging communication and sensing technologies are generating a massive amount of data, requiring distributed environments to perform computation tasks across smart city administrative domains. In this article, we implement and test a number of state-of-the-art active concept drift detection algorithms for time series analysis within a distributed environment. We use real-world data streams and provide critical analysis of results retrieved. The challenges of implementing concept drift adaptation algorithms, along with their applications in smart cities, are also discussed. Citation: Mehmood, H.; Kostakos, P.; Keywords: Cortes, M.; Anagnostopoulos, T.; concept drift; machine learning; smart cities; edge computing; time series analysis; Pirttikangas, S.; Gilman, E.
    [Show full text]
  • Adaptation Strategies for Automated Machine Learning on Evolving Data
    1 Adaptation Strategies for Automated Machine Learning on Evolving Data Bilge Celik and Joaquin Vanschoren Abstract—Automated Machine Learning (AutoML) systems have been shown to efficiently build good models for new datasets. However, it is often not clear how well they can adapt when the data evolves over time. The main goal of this study is to understand the effect of concept drift on the performance of AutoML methods, and which adaptation strategies can be employed to make them more robust to changes in the underlying data. To that end, we propose 6 concept drift adaptation strategies and evaluate their effectiveness on a variety of AutoML approaches for building machine learning pipelines, including Bayesian optimization, genetic programming, and random search with automated stacking. These are evaluated empirically on real-world and synthetic data streams with different types of concept drift. Based on this analysis, we propose ways to develop more sophisticated and robust AutoML techniques. Index Terms—AutoML, data streams, concept drift, adaptation strategies F 1 INTRODUCTION HE field of automated machine learning (AutoML) aims concept drift. We propose six different adaptation strategies T to automatically design and build machine learning to cope with concept drift and implement these in open- systems, replacing manual trial-and-error with systematic, source AutoML libraries. We find that these can indeed be data-driven decision making [43]. This makes robust, state- effectively used, paving the way towards novel AutoML of-the-art machine learning accessible to a much broader techniques that are robust against evolving data. range of practitioners and domain scientists. The online learning setting under consideration is one Although AutoML has been shown to be very effective where data can be buffered in batches using a moving win- and even rival human machine learning experts [13], [24], dow.
    [Show full text]
  • Anytime Algorithms for Stream Data Mining
    Anytime Algorithms for Stream Data Mining Von der Fakultat¨ fur¨ Mathematik, Informatik und Naturwissenschaften der RWTH Aachen University zur Erlangung des akademischen Grades eines Doktors der Naturwissenschaften genehmigte Dissertation vorgelegt von Diplom-Informatiker Philipp Kranen aus Willich, Deutschland Berichter: Universitatsprofessor¨ Dr. rer. nat. Thomas Seidl Visiting Professor Michael E. Houle, PhD Tag der mundlichen¨ Prufung:¨ 14.09.2011 Diese Dissertation ist auf den Internetseiten der Hochschulbibliothek online verfugbar.¨ Contents Abstract / Zusammenfassung1 I Introduction5 1 The Need for Anytime Algorithms7 1.1 Thesis structure......................... 16 2 Knowledge Discovery from Data 17 2.1 The KDD process and data mining tasks ........... 17 2.2 Classification .......................... 25 2.3 Clustering............................ 36 3 Stream Data Mining 43 3.1 General Tools and Techniques................. 43 3.2 Stream Classification...................... 52 3.3 Stream Clustering........................ 59 II Anytime Stream Classification 69 4 The Bayes Tree 71 4.1 Introduction and Preliminaries................. 72 4.2 Indexing density models.................... 76 4.3 Experiments........................... 87 4.4 Conclusion............................ 98 i ii CONTENTS 5 The MC-Tree 99 5.1 Combining Multiple Classes.................. 100 5.2 Experiments........................... 111 5.3 Conclusion............................ 116 6 Bulk Loading the Bayes Tree 117 6.1 Bulk loading mixture densities . 117 6.2 Experiments..........................
    [Show full text]
  • Improving Iot Data Stream Analytics Using Summarization Techniques Maroua Bahri
    Improving IoT data stream analytics using summarization techniques Maroua Bahri To cite this version: Maroua Bahri. Improving IoT data stream analytics using summarization techniques. Machine Learn- ing [cs.LG]. Institut Polytechnique de Paris, 2020. English. NNT : 2020IPPAT017. tel-02865982 HAL Id: tel-02865982 https://tel.archives-ouvertes.fr/tel-02865982 Submitted on 12 Jun 2020 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Improving IoT Data Stream Analytics Using Summarization Techniques These` de doctorat de l’Institut Polytechnique de Paris prepar´ ee´ a` Tel´ ecom´ Paris Ecole´ doctorale n◦626 Denomination´ (Sigle) Specialit´ e´ de doctorat : Informatique NNT : 2020IPPAT017 These` present´ ee´ et soutenue a` Palaiseau, le 5 juin 2020, par MAROUA BAHRI Composition du Jury : Albert Bifet Professor, Tel´ ecom´ Paris Co-directeur de these` Silviu Maniu Associate Professor, Universite´ Paris-Sud Co-directeur de these` Joao˜ Gama Professor, University of Porto President´ Cedric´ Gouy-Pailler Engineer-Researcher, CEA-LIST Examinateur Ons Jelassi
    [Show full text]
  • Data Stream Clustering Techniques, Applications, and Models: Comparative Analysis and Discussion
    big data and cognitive computing Review Data Stream Clustering Techniques, Applications, and Models: Comparative Analysis and Discussion Umesh Kokate 1,*, Arvind Deshpande 1, Parikshit Mahalle 1 and Pramod Patil 2 1 Department of Computer Engineering, SKNCoE, Vadgaon, SPPU, Pune 411 007 India; [email protected] (A.D.); [email protected] (P.M.) 2 Department of Computer Engineering, D.Y. Patil CoE, Pimpri, SPPU, Pune 411 007 India; [email protected] * Correspondence: [email protected]; Tel.: +91-989-023-9995 Received: 16 July 2018; Accepted: 10 October 2018; Published: 17 October 2018 Abstract: Data growth in today’s world is exponential, many applications generate huge amount of data streams at very high speed such as smart grids, sensor networks, video surveillance, financial systems, medical science data, web click streams, network data, etc. In the case of traditional data mining, the data set is generally static in nature and available many times for processing and analysis. However, data stream mining has to satisfy constraints related to real-time response, bounded and limited memory, single-pass, and concept-drift detection. The main problem is identifying the hidden pattern and knowledge for understanding the context for identifying trends from continuous data streams. In this paper, various data stream methods and algorithms are reviewed and evaluated on standard synthetic data streams and real-life data streams. Density-micro clustering and density-grid-based clustering algorithms are discussed and comparative analysis in terms of various internal and external clustering evaluation methods is performed. It was observed that a single algorithm cannot satisfy all the performance measures.
    [Show full text]
  • Massive Online Analysis, a Framework for Stream Classification and Clustering
    MOA: Massive Online Analysis, a Framework for Stream Classification and Clustering. Albert Bifet1, Geoff Holmes1, Bernhard Pfahringer1, Philipp Kranen2, Hardy Kremer2, Timm Jansen2, and Thomas Seidl2 1 Department of Computer Science, University of Waikato, Hamilton, New Zealand fabifet, geoff, [email protected] 2 Data Management and Exploration Group, RWTH Aachen University, Germany fkranen, kremer, jansen, [email protected] Abstract. In today's applications, massive, evolving data streams are ubiquitous. Massive Online Analysis (MOA) is a software environment for implementing algorithms and running experiments for online learn- ing from evolving data streams. MOA is designed to deal with the chal- lenging problems of scaling up the implementation of state of the art algorithms to real world dataset sizes and of making algorithms compa- rable in benchmark streaming settings. It contains a collection of offline and online algorithms for both classification and clustering as well as tools for evaluation. Researchers benefit from MOA by getting insights into workings and problems of different approaches, practitioners can easily compare several algorithms and apply them to real world data sets and settings. MOA supports bi-directional interaction with WEKA, the Waikato Environment for Knowledge Analysis, and is released under the GNU GPL license. Besides providing algorithms and measures for evaluation and comparison, MOA is easily extensible with new contri- butions and allows the creation of benchmark scenarios through storing and sharing setting files. 1 Introduction Nowadays data is generated at an increasing rate from sensor applications, mea- surements in network monitoring and traffic management, log records or click- streams in web exploring, manufacturing processes, call detail records, email, blogging, twitter posts and others.
    [Show full text]
  • An Extensible Framework for Data Stream Clustering Research with R
    JSS Journal of Statistical Software February 2017, Volume 76, Issue 14. doi: 10.18637/jss.v076.i14 Introduction to stream: An Extensible Framework for Data Stream Clustering Research with R Michael Hahsler Matthew Bolaños John Forrest Southern Methodist University Microsoft Corporation Microsoft Corporation Abstract In recent years, data streams have become an increasingly important area of research for the computer science, database and statistics communities. Data streams are ordered and potentially unbounded sequences of data points created by a typically non-stationary data generating process. Common data mining tasks associated with data streams include clustering, classification and frequent pattern mining. New algorithms for these types of data are proposed regularly and it is important to evaluate them thoroughly under standardized conditions. In this paper we introduce stream, a research tool that includes modeling and simu- lating data streams as well as an extensible framework for implementing, interfacing and experimenting with algorithms for various data stream mining tasks. The main advantage of stream is that it seamlessly integrates with the large existing infrastructure provided by R. In addition to data handling, plotting and easy scripting capabilities, R also provides many existing algorithms and enables users to interface code written in many program- ming languages popular among data mining researchers (e.g., C/C++, Java and Python). In this paper we describe the architecture of stream and focus on its use for data stream clustering research. stream was implemented with extensibility in mind and will be ex- tended in the future to cover additional data stream mining tasks like classification and frequent pattern mining.
    [Show full text]
  • Semi-Supervised Hybrid Windowing Ensembles for Learning from Evolving Streams
    Semi-Supervised Hybrid Windowing Ensembles for Learning from Evolving Streams Sean Louis Alan FLOYD Thesis submitted in partial fulfilment of the requirements for the Master of Science in Computer Science School of Electrical Engineering and Computer Science Faculty of Engineering University of Ottawa c Sean Louis Alan FLOYD, Ottawa, Canada, 2019 ii Abstract In this thesis, learning refers to the intelligent computational extraction of knowl- edge from data. Supervised learning tasks require data to be annotated with labels, whereas for unsupervised learning, data is not labelled. Semi-supervised learning deals with data sets that are partially labelled. A major issue with su- pervised and semi-supervised learning of data streams is late-arriving or miss- ing class labels. Assuming that correctly labelled data will always be available and timely is often unfeasible, and, as such, supervised methods are not directly applicable in the real world. Therefore, real-world problems usually require the use of semi-supervised or unsupervised learning techniques. For instance, when considering a spam detection task, it is not reasonable to assume that all spam will be identified (correctly labelled) prior to learning. Additionally, in semi- supervised learning, "the instances having the highest [predictive] confidence are not necessarily the most useful ones" [41]. We investigate how self-training performs without its selective heuristic in a streaming setting. This leads us to our contributions. We extend an existing concept drift detec- tor to operate without any labelled data, by using a sliding window of our en- semble’s prediction confidence, instead of a boolean indicating whether the en- semble’s predictions are correct.
    [Show full text]
  • Mining Data Streams: a Review
    Mining Data Streams: A Review Mohamed Medhat Gaber, Arkady Zaslavsky and Shonali Krishnaswamy Centre for Distributed Systems and Software Engineering, Monash University 900 Dandenong Rd, Caulfield East, VIC3145, Australia {Mohamed.Medhat.Gaber, Arkady.Zaslavsky, Shonali.Krishnaswamy} @infotech.monash.edu.au Abstract and parallel data mining. The goal was how to extract knowledge from different subsets of a dataset and The recent advances in hardware and software have integrate these generated knowledge structures in order enabled the capture of different measurements of data in to gain a global model of the whole dataset. a wide range of fields. These measurements are Client/server, mobile agent based and hybrid models generated continuously and in a very high fluctuating have been proposed to address the communication data rates. Examples include sensor networks, web logs, overhead issue. Different variations of algorithms have and computer network traffic. The storage, querying and been developed in order to increase the accuracy of the mining of such data sets are highly computationally generated global model. More details about distributed challenging tasks. Mining data streams is concerned data mining could be found in [47]. with extracting knowledge structures represented in Recently, the data generation rates in some models and patterns in non stopping streams of data sources become faster than ever before. This rapid information. The research in data stream mining has generation of continuous streams of information has gained a high attraction due to the importance of its challenged our storage, computation and applications and the increasing generation of streaming communication capabilities in computing systems. information. Applications of data stream analysis can Systems, models and techniques have been proposed vary from critical scientific and astronomical and developed over the past few years to address these applications to important business and financial ones.
    [Show full text]
  • An Analytical Framework for Data Stream Mining Techniques Based on Challenges and Requirements
    Mahnoosh Kholghi et al. / International Journal of Engineering Science and Technology (IJEST) AN ANALYTICAL FRAMEWORK FOR DATA STREAM MINING TECHNIQUES BASED ON CHALLENGES AND REQUIREMENTS MAHNOOSH KHOLGHI Department of Electronic, Computer and IT, Islamic Azad University, Qazvin Branch, Qazvin, Iran and member of Young Researchers Club [email protected] MOHAMMADREZA KEYVANPOUR Department of Computer Engineering Alzahra University Tehran, Iran [email protected] Abstract: A growing number of applications that generate massive streams of data need intelligent data processing and online analysis. Real-time surveillance systems, telecommunication systems, sensor networks and other dynamic environments are such examples. The imminent need for turning such data into useful information and knowledge augments the development of systems, algorithms and frameworks that address streaming challenges. The storage, querying and mining of such data sets are highly computationally challenging tasks. Mining data streams is concerned with extracting knowledge structures represented in models and patterns in non stopping streams of information. Generally, two main challenges are designing fast mining methods for data streams and need to promptly detect changing concepts and data distribution because of highly dynamic nature of data streams. The goal of this article is to analyze and classify the application of diverse data mining techniques in different challenges of data stream mining. In this paper, we present the theoretical foundations of data stream analysis and propose an analytical framework for data stream mining techniques. Keywords: Data Stream, Data Stream Mining, Stream Preprocessing. 1. Introduction Data mining techniques are suitable for simple and structured data sets like relational databases, transactional databases and data warehouses.
    [Show full text]
  • A Review of Meta-Level Learning in the Context of Multi-Component
    AREVIEW OF META-LEVEL LEARNING IN THE CONTEXT OF MULTI-COMPONENT,MULTI-LEVEL EVOLVING PREDICTION SYSTEMS Abbas Raza Ali Marcin Budka Bogdan Gabrys Faculty of Science and Technology Faculty of Science and Technology Advanced Analytics Institute Bournemouth University Bournemouth University University Technology Sydney Poole BH12 5BB Poole BH12 5BB Ultimo NSW 2007 United Kingdom United Kingdom Australia [email protected] [email protected] [email protected] ABSTRACT The exponential growth of volume, variety and velocity of data is raising the need for investigations of automated or semi-automated ways to extract useful patterns from the data. It requires deep expert knowledge and extensive computational resources to find the most appropriate mapping of learning methods for a given problem. It becomes a challenge in the presence of numerous configurations of learning algorithms on massive amounts of data. So there is a need for an intelligent recommendation engine that can advise what is the best learning algorithm for a dataset. The techniques that are commonly used by experts are based on a trial and error approach evaluating and comparing a number of possible solutions against each other, using their prior experience on a specific domain, etc. The trial and error approach combined with the expert’s prior knowledge, though computationally and time expensive, have been often shown to work for stationary problems where the processing is usually performed off-line. However, this approach would not normally be feasible to apply to non-stationary problems where streams of data are continuously arriving. Furthermore, in a non- stationary environment, the manual analysis of data and testing of various methods whenever there is a change in the underlying data distribution would be very difficult or simply infeasible.
    [Show full text]
  • SAMOA: Scalable Advanced Massive Online Analysis
    Journal of Machine Learning Research 16 (2015) 149-153 Submitted 6/14; Revised 9/14; Published 1/15 SAMOA: Scalable Advanced Massive Online Analysis Gianmarco De Francisci Morales [email protected] Albert Bifet [email protected] Yahoo Labs Av. Diagonal 177, 8th floor, 08018, Barcelona, Spain Editor: Geoff Holmes Abstract samoa (Scalable Advanced Massive Online Analysis) is a platform for mining big data streams. It provides a collection of distributed streaming algorithms for the most common data mining and machine learning tasks such as classification, clustering, and regression, as well as programming abstractions to develop new algorithms. It features a pluggable architecture that allows it to run on several distributed stream processing engines such as Storm, S4, and Samza. samoa is written in Java, is open source, and is available at http://samoa-project.net under the Apache Software License version 2.0. Keywords: data streams, distributed systems, classification, clustering, regression, tool- box, machine learning 1. Introduction Big data is “data whose characteristics forces us to look beyond the traditional methods that are prevalent at the time” (Jacobs, 2009). Currently, there are two main ways to deal with big data: streaming algorithms and distributed computing (e.g., MapReduce). samoa aims at satisfying the future needs for big data stream mining by combining the two approaches in a single platform under an open source umbrella. Data mining and machine learning are well established techniques among web companies and startups. Spam detection, personalization, and recommendation are just a few of the applications made possible by mining the huge quantity of data available nowadays.
    [Show full text]