KDD-SC: Subspace Clustering Extensions for Knowledge Discovery Frameworks

Total Page:16

File Type:pdf, Size:1020Kb

KDD-SC: Subspace Clustering Extensions for Knowledge Discovery Frameworks KDD-SC: Subspace Clustering Extensions for Knowledge Discovery Frameworks Stephan Günnemann◦• Hardy Kremer◦ Matthias Hannen◦ Thomas Seidl◦ ◦RWTH Aachen University, Germany •Carnegie Mellon University, USA {lastname}@cs.rwth-aachen.de [email protected] ABSTRACT Analyzing high dimensional data is a challenging task. For these data it is known that traditional clustering algorithms fail to detect meaningful patterns. As a solution, subspace clustering techniques have been introduced. They analyze arbitrary subspace projections of the data to detect cluster- ing structures. In this paper, we present our subspace clustering exten- sion for KDD frameworks, termed KDD-SC. In contrast to existing subspace clustering toolkits, our solution neither is Figure 1: Subspace clusters hidden in locally rele- a standalone product nor is it tightly coupled to a specific vant subspace projections KDD framework. Our extension is realized by a common codebase and easy-to-use plugins for three of the most pop- ular KDD frameworks, namely KNIME, RapidMiner, and Existing systems: Today, general data mining functional- WEKA. KDD-SC extends these frameworks such that they ity is provided to the end-user in a convenient and intuitive offer a wide range of different subspace clustering functional- way by established knowledge discovery frameworks as KN- ities. It provides a multitude of algorithms, data generators, IME (Konstanz Information Miner, [6]), RapidMiner [18], evaluation measures, and visualization techniques specifi- and WEKA (Waikato Environment for Knowledge Analy- cally designed for subspace clustering. These functionalities sis, [14]). These systems are succesfully and frequently used integrate seamlessly with the frameworks' existing features in research and practice. The applicability of subspace clus- such that they can be flexibly combined. KDD-SC is pub- tering, in contrast, is still limited. licly available on our website. So far, there are two systems that support the user in the task of subspace clustering, namely OpenSubspace [20] and ELKI [1]. Both systems are milestones in the process of 1. INTRODUCTION providing subspace clustering functionality to the end-user, Clustering is one of the core data mining tasks. The goal but have severe limitations concerning their integration into of clustering is to automatically group similar objects while established data mining workflows. While ELKI, as a stand- separating dissimilar ones. Traditional clustering methods alone java framework, does not offer any integration into consider all dimensions of the dataspace to measure the sim- existing data mining toolkits, OpenSubspace is highly cou- ilarity between objects. For today's high dimensional data, pled and its current form only applicable within the WEKA however, these full-space clustering approaches fail to detect framework. Due to this strong coupling, it is difficult to in- meaningful patterns since irrelevant dimensions obfuscate tegrate new algorithms and to (re)use already implemented the clustering structure [7, 16]. Using global dimensionality subspace clustering functionality in other KDD frameworks. arXiv:1407.3850v1 [cs.DB] 15 Jul 2014 reduction techniques such as principle components analysis Accordingly, for end-users running their established KDD is not sufficient to solve this problem: by definition, all ob- workflows in other frameworks than WEKA or ELKI, the jects are projected to the same lower dimensional subspace. integration of subspace clustering into these workflows is a However, as Figure 1 illustrates, each cluster might have hard and time-consuming challenge. locally relevant dimensions and objects can be part of mul- Our contribution: In this paper, we propose a new sys- tiple clusters in different subspaces. These effects cannot be tem for subspace clustering, which is seamlessly integrated captured by global dimensionality reduction approaches. into KNIME, RapidMiner, and WEKA. By covering this To tackle this challenge, subspace clustering techniques broad spectrum of knowledge discovery frameworks, many have been introduced, aiming at detecting locally relevant researchers and practitioners can benefit from our system. dimensions per cluster [16, 21]. They analyze arbitrary sub- It is based on a common code basis across all KDD frame- space projections of the data to detect the hidden clusters. works. Thus, it is possible to quickly deploy new subspace Typical applications for subspace clustering include gene ex- clustering methods in multiple frameworks at the same time. pression analysis, customer profiling, and sensor network By integrating our system into these established knowl- analysis. In each of these scenarios, subsets of the objects edge discovery frameworks, the user can easily use subspace (e.g., genes) are similar regarding subsets of the dimensions clustering functionality within the whole KDD process. Our (e.g., different experimental conditions). methods can be combined with the existing algorithms, data transformations techniques, and visualization tools of these By using a common codebase, i.e. the CoreSC package, frameworks. Overall, our system offers it is easy to integrate new subspace clustering techniques • a seamless integration of subspace clustering functional- for each of the knowledge discovery frameworks. The actual ity into KNIME, RapidMiner, and WEKA. Accordingly, subspace clustering algorithm has only to be implemented in many researchers and practitioners can use their estab- the core package. Additionally, one can easily support other lished KDD workflows without any loss in productivity. (e.g., R) or even new data mining frameworks by simply providing a new adapter package. • a common code basis for subspace clustering algorithms, evaluation measures, and synthetic data generators. It 2.1 Subspace Clustering Algorithms is independent of the chosen data mining framework and realizes easy extensibility and reusability of all compo- The first component of the CoreSC package contains the nents. actual subspace clustering algorithms. In our extension, the user can select among a multitude of different algorithms. • visualization and interaction principles for subspace clus- These algorithms include grid based clustering techniques tering exploiting the capabilities of the different data (CLIQUE [3], DOC/FastDOC [23], MineClus [25], SCHISM mining toolkits, which support the user in the interpre- [24]), DBSCAN-based techniques (FIRES [15], INSCY [5], tation of the obtained results. SUBLCU [17]) and optimization-based techniques for sub- space clustering (PROCLUS [2], STATPC [19]). 2. GENERAL ARCHITECTURE Each algorithm implements the interface SubspaceAlgo- In this section we describe the general architecture and rithm, which defines the input and output of the algorithms. functionality of our subspace clustering extension. The us- The input corresponds to a database of objects described by age of our extension within the different knowledge discovery numerical features, i.e. each algorithm needs to be provided d frameworks is described in the Sections 3-5. with a list of objects ho1; : : : ; oni where oi 2 R . The output For reusability and easy portability of the developed meth- of each algorithm is a list of subspace clusters hC1;:::;Cki. ods, our KDD-SC framework is separated into a core pack- Each subspace cluster Ci represents the objects and rele- age (CoreSC) and packages realizing the integration into the vant dimensions belonging to this clusters. Note that in different KDD frameworks (KnimeSC, RapidSC, WekaSC). subspace clustering each cluster has its individual set of rel- Figure 2 shows an overview of this design. evant dimensions (cf. Figure 1). Thus, each subspace cluster corresponds to a tuple Ci = (Oi;Si) where Oi represents the clustered objects by their objects ids, i.e. Oi ⊆ f1; : : : ; ng, and S represents the relevant dimensions of the cluster, i.e. KnimeSC RapidSC WekaSC i Si ⊆ f1; : : : ; dg. It is worth mentioning that subspace clustering in general CoreSC is not restricted to disjoint clusters; thus, the result set might - subspace clustering algorithms contain clusters Ci and Cj (with i 6= j) where Oi \ Oj 6= - data generators ; or Si \ Sj 6= ;. Additionally, dependent on the chosen - evaluation measures algorithm, not necessarily each object or dimension needs - visualization Sk to be part of some cluster, i.e. it might hold i=1 Oi 6= DB Sk or i=1 Si 6= f1; : : : ; dg. Figure 2: General Architecture of KDD-SC 2.2 Data Generators In the core package, the actual functionality of our sys- The second component of the core package contains a tem is implemented. This functionality is independent of a flexible data generator first introduced in [13], which gen- specific system. The core package is divided into four major erates synthetic data with hidden subspace clusters. These components: subspace clustering algorithms, data genera- datasets can be used to evaluate the correctness of subspace tors, evaluation measures, and visualization tools. A de- clustering algorithms and to assess the methods' scalability. tailed description of these components is provided in the The data generator implements the interface SubspaceData- following sections. Generator which defines the two outputs of the generator. The WekaSC user interface as well as parts of the core The first output corresponds to the generated data, i.e. as above it corresponds to a list of objects ho1; : : : ; oni with package (algorithms & evaluation measures) have been ex- d tracted from the OpenSubspace project [20].
Recommended publications
  • Text Mining Course for KNIME Analytics Platform
    Text Mining Course for KNIME Analytics Platform KNIME AG Copyright © 2018 KNIME AG Table of Contents 1. The Open Analytics Platform 2. The Text Processing Extension 3. Importing Text 4. Enrichment 5. Preprocessing 6. Transformation 7. Classification 8. Visualization 9. Clustering 10. Supplementary Workflows Licensed under a Creative Commons Attribution- ® Copyright © 2018 KNIME AG 2 Noncommercial-Share Alike license 1 https://creativecommons.org/licenses/by-nc-sa/4.0/ Overview KNIME Analytics Platform Licensed under a Creative Commons Attribution- ® Copyright © 2018 KNIME AG 3 Noncommercial-Share Alike license 1 https://creativecommons.org/licenses/by-nc-sa/4.0/ What is KNIME Analytics Platform? • A tool for data analysis, manipulation, visualization, and reporting • Based on the graphical programming paradigm • Provides a diverse array of extensions: • Text Mining • Network Mining • Cheminformatics • Many integrations, such as Java, R, Python, Weka, H2O, etc. Licensed under a Creative Commons Attribution- ® Copyright © 2018 KNIME AG 4 Noncommercial-Share Alike license 2 https://creativecommons.org/licenses/by-nc-sa/4.0/ Visual KNIME Workflows NODES perform tasks on data Not Configured Configured Outputs Inputs Executed Status Error Nodes are combined to create WORKFLOWS Licensed under a Creative Commons Attribution- ® Copyright © 2018 KNIME AG 5 Noncommercial-Share Alike license 3 https://creativecommons.org/licenses/by-nc-sa/4.0/ Data Access • Databases • MySQL, MS SQL Server, PostgreSQL • any JDBC (Oracle, DB2, …) • Files • CSV, txt
    [Show full text]
  • Imagej2-Allow the Users to Use Directly Use/Update Imagej2 Plugins Inside KNIME As Well As Recording and Running KNIME Workflows in Imagej2
    The KNIME Image Processing Extension for Biomedical Image Analysis Andries Zijlstra (Vanderbilt University Medical Center The need for image processing in medicine Kevin Eliceiri (University of Wisconsin-Madison) KNIME Image Processing and ImageJ Ecosystem [email protected] [email protected] The need for precision oncology 36% of newly diagnosed cancers, and 10% of all cancer deaths in men Out of every 100 men... 16 will be diagnosed with prostate cancer in their lifetime In reality, up to 80 will have prostate cancer by age 70 And 3 will die from it. But which 3 ? In the meantime, we The goal: Diagnose patients that have over-treat many aggressive disease through Precision Medicine patients Objectives of Approach to Modern Medicine Precision Medicine • Measure many things (data density) • Improved outcome through • Make very accurate measurements (fidelity) personalized/precision medicine • Consider multiple perspectives (differential) • Reduced expense/resource allocation through • Achieve confidence in the diagnosis improved diagnosis, prognosis, treatment • Match patients with a treatment they are most • Maximize quality of life by “targeted” therapy likely to respond to. Objectives of Approach to Modern Medicine Precision Medicine • Measure many things (data density) • Improved outcome through • Make very accurate measurements (fidelity) personalized/precision medicine • Consider multiple perspectives (differential) • Reduced expense/resource allocation through • Achieve confidence in the diagnosis improved diagnosis,
    [Show full text]
  • KNIME Workbench Guide
    KNIME Workbench Guide KNIME AG, Zurich, Switzerland Version 4.4 (last updated on 2021-06-08) Table of Contents Workspaces . 1 KNIME Workbench . 2 Welcome page . 4 Workflow editor & nodes . 5 KNIME Explorer . 13 Workflow Coach . 35 Node repository . 37 KNIME Hub view . 38 Description. 40 Node Monitor. 40 Outline. 41 Console. 41 Customizing the KNIME Workbench . 42 Reset and logging . 42 Show heap status . 42 Configuring KNIME Analytics Platform . 43 Preferences . 43 Setting up knime.ini. 47 KNIME runtime options . 49 KNIME tables . 55 Data table . 55 Column types. 56 Sorting . 59 Column rendering . 59 Table storage. 61 KNIME Workbench Guide This guide describes the first steps to take after starting KNIME Analytics Platform and points you to the resources available in the KNIME Workbench for building workflows. It also explains how to customize the workbench and configure KNIME Analytics Platform to best suit specific needs. In the last part of this guide we introduce data tables. Workspaces When you start KNIME Analytics Platform, the KNIME Analytics Platform launcher window appears and you are asked to define the KNIME workspace, as shown in Figure 1. The KNIME workspace is a folder on the local computer to store KNIME workflows, node settings, and data produced by the workflow. Figure 1. KNIME Analytics Platform launcher The workflows and data stored in the workspace are available through the KNIME Explorer in the upper left corner of the KNIME Workbench. © 2021 KNIME AG. All rights reserved. 1 KNIME Workbench Guide KNIME Workbench After selecting a workspace for the current project, click Launch. The KNIME Analytics Platform user interface - the KNIME Workbench - opens.
    [Show full text]
  • Data Analytics with Knime
    DATA ANALYTICS WITH KNIME v.3.4.0 QUALIFICATIONS & EXPERIENCE ▶ 38 years of providing professional services to state and local taxing officials ▶ TMA works exclusively with government partners WHO ▶ TMA is composed of 150+ WE ARE employees in five main offices across the United States Tax Management Associates is a professional services firm that has ▶ Our main focus is on revenue served the interests of state and local enhancement services for state government since 1979. and local jurisdictions and property tax compliance efforts KNIME POWERED CUSTOM ANALYTICS ▶ TMA is a proud KNIME Trusted Consulting Partner. Visit: www.knime.org/knime-trusted-partners ▶ Successful analytics solutions: ○ Fraud Detection (Michigan Department of Treasury) ○ Entity Discovery (multiple counties) ○ Data Aggregation (Louisiana State Tax Commission) KNIME POWERED CUSTOM ANALYTICS ▶ KNIME is an open source data toolkit ▶ Active development community and core team ▶ GUI based with scripting integration ○ Easy adoption, integration, and training ▶ Data ingestion, transformation, analytics, and reporting FEATURES & TERMINOLOGY KNIME WORKBENCH TAX MANAGEMENT ASSOCIATES, INC. KNIME WORKFLOW TAX MANAGEMENT ASSOCIATES, INC. KNIME NODES TAX MANAGEMENT ASSOCIATES, INC. DATA TYPES & SOURCES DATA AGNOSTIC ▶ Flat Files ▶ Shapefiles ▶ Xls/x Reader ▶ HTTP Requests ▶ Fixed Width ▶ RSS Feeds ▶ Text Files ▶ Custom API’s/Curl ▶ Image Files ▶ Standard API’s ▶ XML ▶ JSON TAX MANAGEMENT ASSOCIATES, INC. KNIME DATA NODES TAX MANAGEMENT ASSOCIATES, INC. DATABASE AGNOSTIC ▶ Microsoft SQL ▶ Oracle ▶ MySQL ▶ IBM DB2 ▶ Postgres ▶ Hadoop ▶ SQLite ▶ Any JDBC driver TAX MANAGEMENT ASSOCIATES, INC. KNIME DATABASE NODES TAX MANAGEMENT ASSOCIATES, INC. CORE DATA ANALYTICS FEATURES KNIME DATA ANALYTICS LIFECYCLE Read Data Extract, Data Analytics Reporting or Predictive Read Transform, and/or Load (ETL) Analysis Injection Data Read Data TAX MANAGEMENT ASSOCIATES, INC.
    [Show full text]
  • Effect of Distance Measures on Partitional Clustering Algorithms
    Sesham Anand et al, / (IJCSIT) International Journal of Computer Science and Information Technologies, Vol. 6 (6) , 2015, 5308-5312 Effect of Distance measures on Partitional Clustering Algorithms using Transportation Data Sesham Anand#1, P Padmanabham*2, A Govardhan#3 #1Dept of CSE,M.V.S.R Engg College, Hyderabad, India *2Director, Bharath Group of Institutions, BIET, Hyderabad, India #3Dept of CSE, SIT, JNTU Hyderabad, India Abstract— Similarity/dissimilarity measures in clustering research of algorithms, with an emphasis on unsupervised algorithms play an important role in grouping data and methods in cluster analysis and outlier detection[1]. High finding out how well the data differ with each other. The performance is achieved by using many data index importance of clustering algorithms in transportation data has structures such as the R*-trees.ELKI is designed to be easy been illustrated in previous research. This paper compares the to extend for researchers in data mining particularly in effect of different distance/similarity measures on a partitional clustering algorithm kmedoid(PAM) using transportation clustering domain. ELKI provides a large collection of dataset. A recently developed data mining open source highly parameterizable algorithms, in order to allow easy software ELKI has been used and results illustrated. and fair evaluation and benchmarking of algorithms[1]. Data mining research usually leads to many algorithms Keywords— clustering, transportation Data, partitional for similar kind of tasks. If a comparison is to be made algorithms, cluster validity, distance measures between these algorithms.In ELKI, data mining algorithms and data management tasks are separated and allow for an I. INTRODUCTION independent evaluation.
    [Show full text]
  • Sheffield HPC Documentation
    Sheffield HPC Documentation Release November 14, 2016 Contents 1 Research Computing Team 3 2 Research Software Engineering Team5 i ii Sheffield HPC Documentation, Release The current High Performance Computing (HPC) system at Sheffield, is the Iceberg cluster. A new system, ShARC (Sheffield Advanced Research Computer), is currently under development. It is not yet ready for use. Contents 1 Sheffield HPC Documentation, Release 2 Contents CHAPTER 1 Research Computing Team The research computing team are the team responsible for the iceberg service, as well as all other aspects of research computing. If you require support with iceberg, training or software for your workstations, the research computing team would be happy to help. Take a look at the Research Computing website or email research-it@sheffield.ac.uk. 3 Sheffield HPC Documentation, Release 4 Chapter 1. Research Computing Team CHAPTER 2 Research Software Engineering Team The Sheffield Research Software Engineering Team is an academically led group that collaborates closely with CiCS. They can assist with code optimisation, training and all aspects of High Performance Computing including GPU computing along with local, national, regional and cloud computing services. Take a look at the Research Software Engineering website or email rse@sheffield.ac.uk 2.1 Using the HPC Systems 2.1.1 Getting Started If you have not used a High Performance Computing (HPC) cluster, Linux or even a command line before this is the place to start. This guide will get you set up using iceberg in the easiest way that fits your requirements. Getting an Account Before you can start using iceberg you need to register for an account.
    [Show full text]
  • Mathematica Document
    Mathematica Project: Exploratory Data Analysis on ‘Data Scientists’ A big picture view of the state of data scientists and machine learning engineers. ����� ���� ��������� ��� ������ ���� ������ ������ ���� ������/ ������ � ���������� ���� ��� ������ ��� ���������������� �������� ������/ ����� ��������� ��� ���� ���������������� ����� ��������������� ��������� � ������(�������� ���������� ���������) ������ ��������� ����� ������� �������� ����� ������� ��� ������ ����������(���� �������) ��������� ����� ���� ������ ����� (���������� �������) ����������(���������� ������� ���������� ��� ���� ���� �����/ ��� �������������� � ����� ���� �� �������� � ��� ����/���������� ��������������� ������� ������������� ��� ���������� ����� �����(���� �������) ����������� ����� / ����� ��� ������ ��������������� ���������� ����������/�++ ������/������������/����/������ ���� ������� ����� ������� ������� ����������������� ������� ������� ����/����/�������/��/��� ����������(�����/����-�������� ��������) ������������ In this Mathematica project, we will explore the capabilities of Mathematica to better understand the state of data science enthusiasts. The dataset consisting of more than 10,000 rows is obtained from Kaggle, which is a result of ‘Kaggle Survey 2017’. We will explore various capabilities of Mathematica in Data Analysis and Data Visualizations. Further, we will utilize Machine Learning techniques to train models and Classify features with several algorithms, such as Nearest Neighbors, Random Forest. Dataset : https : // www.kaggle.com/kaggle/kaggle
    [Show full text]
  • Direct Submission Or Co-Submission Direct Submission
    Z-Matrix template-based substitution approach Title for enumeration of 3D molecular structures Authors Wanutcha Lorpaiboon and Taweetham Limpanuparb* Science Division, Mahidol University International College, Affiliations Mahidol University, Salaya, Nakhon Pathom 73170, Thailand Corresponding Author’s email address [email protected] • Chemical structures • Education Keywords • Molecular generator • Structure generator • Z-matrix Direct Submission or Co-Submission Direct Submission ABSTRACT The exhaustive enumeration of 3D chemical structures based on Z-matrix templates has recently been used in the quantum chemical investigation of constitutional isomers, diastereomers and 5 rotamers. This simple yet powerful initial structure generation approach can apply beyond the investigation of compounds of identical formula by quantum chemical methods. This paper aims to provide a short description of the overall concept followed by a practical tutorial to the approach. • The four steps required for Z-matrix template-based substitution are template construction, generation of tuples for substitution sites, removal of duplicate tuples and 10 substitution on the template. • The generated tuples can be used to create chemical identifiers to query compound properties from chemical databases. • All of these steps are demonstrated in this paper by common model compounds and are very straightforward for an undergraduate audience to reproduce. A comparison of the 15 approach in this tutorial and other options is also discussed. SPECIFICATIONS TABLE Subject Area Chemistry More specific subject area Cheminformatics Method name Z-matrix template-based substitution Name and reference of original method N/A Source codes are available as supplementary information in this Resource availability paper. 2 of 10 Method details 20 1. Introduction Initial structures (Z-matrix or Cartesian coordinate) are important starting points for the in silico investigation of chemical species.
    [Show full text]
  • Role of Materials Data Science and Informatics in Accelerated Materials Innovation Surya R
    Role of materials data science and informatics in accelerated materials innovation Surya R. Kalidindi , David B. Brough , Shengyen Li , Ahmet Cecen , Aleksandr L. Blekh , Faical Yannick P. Congo , and Carelyn Campbell The goal of the Materials Genome Initiative is to substantially reduce the time and cost of materials design and deployment. Achieving this goal requires taking advantage of the recent advances in data and information sciences. This critical need has impelled the emergence of a new discipline, called materials data science and informatics. This emerging new discipline not only has to address the core scientifi c/technological challenges related to datafi cation of materials science and engineering, but also, a number of equally important challenges around data-driven transformation of the current culture, practices, and workfl ows employed for materials innovation. A comprehensive effort that addresses both of these aspects in a synergistic manner is likely to succeed in realizing the vision of scaled-up materials innovation. Key toolsets needed for the successful adoption of materials data science and informatics in materials innovation are identifi ed and discussed in this article. Prototypical examples of emerging novel toolsets and their functionality are described along with select case studies. Introduction goal of reducing the time and cost of materials development Materials innovation initiatives and deployment by 50%. 1 Essential to achieving this goal is A number of US-based, 1 – 3 as well as international, 4 , 5 efforts are the development and deployment of a supporting infrastruc- now focused on accelerated deployment of advanced materials ture that integrates a wide range of data, experimental, and in commercial products.
    [Show full text]
  • Titel Untertitel
    KNIME Image Processing Nycomed Chair for Bioinformatics and Information Mining Department of Computer and Information Science Konstanz University, Germany Why Image Processing with KNIME? KNIME UGM 2013 2 The “Zoo” of Image Processing Tools Development Processing UI Handling ImgLib2 ImageJ OMERO OpenCV ImageJ2 BioFormats MatLab Fiji … NumPy CellProfiler VTK Ilastik VIGRA CellCognition … Icy Photoshop … = Single, individual, case specific, incompatible solutions KNIME UGM 2013 3 The “Zoo” of Image Processing Tools Development Processing UI Handling ImgLib2 ImageJ OMERO OpenCV ImageJ2 BioFormats MatLab Fiji … NumPy CellProfiler VTK Ilastik VIGRA CellCognition … Icy Photoshop … → Integration! KNIME UGM 2013 4 KNIME as integration platform KNIME UGM 2013 5 Integration: What and How? KNIME UGM 2013 6 Integration ImgLib2 • Developed at MPI-CBG Dresden • Generic framework for data (image) processing algoritms and data-structures • Generic design of algorithms for n-dimensional images and labelings • http://fiji.sc/wiki/index.php/ImgLib2 → KNIME: used as image representation (within the data cells); basis for algorithms KNIME UGM 2013 7 Integration ImageJ/Fiji • Popular, highly interactive image processing tool • Huge base of available plugins • Fiji: Extension of ImageJ1 with plugin-update mechanism and plugins • http://rsb.info.nih.gov/ij/ & http://fiji.sc/ → KNIME: ImageJ Macro Node KNIME UGM 2013 8 Integration ImageJ2 • Next-generation version of ImageJ • Complete re-design of ImageJ while maintaining backwards compatibility • Based on ImgLib2
    [Show full text]
  • Leveraging SAS with KNIME
    Leveraging SAS with KNIME Thomas Gabriel and Phil Winters Copyright © 2013 by KNIME.com AG. All rights reserved. SAS and all other SAS Institute Inc. product or service names are registered trademarks or trademarks of SAS Institute Inc. in the USA and other countries. ® indicates USA registration. Copyright © 2013 by KNIME.com AG. All rights reserved. 2 A bit of History: The Original SAS Concept Access Manage Present Analyse Copyright © 2013 by KNIME.com AG. All rights reserved. 4 More Products to meet new requirements EG DI RTD DB2 Macro Oracle PC FILIE Connect Access Manage MM SCL Present Analyse EM ® AF DVD QC JMP OR IML IDP ETS STAT Graph Copyright © 2013 by KNIME.com AG. All rights reserved. 5 Even More Products…. Teradata Hadoop EG DI RTD DB2 Macro Oracle PC FILIE Connect Access Manage MM SCL Present Analyse EM PMML …Model Manager …Model PMML High Performance Analytics Performance High ® AF DVD QC JMP OR IML IDP ETS STAT Graph Text Mining Social Media Analytics R…. in IML in R…. Copyright © 2013 by KNIME.com AG. All rights reserved. 6 Our new reality: You must have Choice and Control New Infrastructures New Data New Other Science Applications New New Users Methods New Business Copyright © 2013 by KNIME.com AG. All rights reserved. Challenges 7 The KNIME Platform Open, Open Source, Free on the Desktop Copyright © 2013 by KNIME.com AG. All rights reserved. 8 The question is not “which is better”…. Copyright © 2013 by KNIME.com AG. All rights reserved. 9 The question is: What’s the Big Difference? SAS KNIME A script-oriented 4GL programming language A Script-free environment in four major parts: comprised of • The DATA step • Nodes and Connectors • Procedure steps • A macro language, • Metanodes and Flow variables for packaging a metaprogramming language • ODS statements • GUIs: are most often front-ends • GUI is the interface to facilitate SAS Program script generation • Through Proprietary Technology “Provide it all” • Through Open Source “Provide it all” Copyright © 2013 by KNIME.com AG.
    [Show full text]
  • BETULA: Numerically Stable CF-Trees for BIRCH Clustering?
    BETULA: Numerically Stable CF-Trees for BIRCH Clustering? Andreas Lang[0000−0003−3212−5548] and Erich Schubert[0000−0001−9143−4880] TU Dortmund University, Dortmund, Germany fandreas.lang,[email protected] Abstract. BIRCH clustering is a widely known approach for clustering, that has influenced much subsequent research and commercial products. The key contribution of BIRCH is the Clustering Feature tree (CF-Tree), which is a compressed representation of the input data. As new data arrives, the tree is eventually rebuilt to increase the compression. After- ward, the leaves of the tree are used for clustering. Because of the data compression, this method is very scalable. The idea has been adopted for example for k-means, data stream, and density-based clustering. Clustering features used by BIRCH are simple summary statistics that can easily be updated with new data: the number of points, the linear sums, and the sum of squared values. Unfortunately, how the sum of squares is then used in BIRCH is prone to catastrophic cancellation. We introduce a replacement cluster feature that does not have this nu- meric problem, that is not much more expensive to maintain, and which makes many computations simpler and hence more efficient. These clus- ter features can also easily be used in other work derived from BIRCH, such as algorithms for streaming data. In the experiments, we demon- strate the numerical problem and compare the performance of the orig- inal algorithm compared to the improved cluster features. 1 Introduction The BIRCH algorithm [23,24,22] is a widely known cluster analysis approach, that won the 2006 SIGMOD Test of Time Award.
    [Show full text]