About This Template

Total Page:16

File Type:pdf, Size:1020Kb

About This Template Real-Time Actionable Insights with IBM Streams — Roger Rea IBM Streams Senior Offering Manager [email protected] Data and AI Forum 2019 Session 309 Data and AI Forum / © 2019 IBM Corporation Real-time processing needs are growing Data and AI Forum / © 2019 IBM Corporation Continuous intelligence – What if you could analyze data as it’s created? – What if you could visualize your business? – What if you could better predict your customers’ needs? – What if you could gain insights from unstructured data like audio, text or video? – What if could automate immediate actions? – What if you always knew where your assets were, and where they would be? – What if you could update machine learning models continuously? And do it all in real time? IBM Data & AI / © 2019 IBM Corporation “Continuous intelligence is a design pattern in which real-time analytics are integrated into a business operation, processing current and historical data to prescribe actions in response to business moments and other events.” Gartner: Innovation Insight for Continuous Intelligence Data and AI Forum / © 2019 IBM Corporation 4 Continuous intelligence Stream – Engage data from inside and outside processing of applications or business operations – Enable fast-paced digital business decisions Artificial Machine and process optimization intelligence learning Continuous – Leverage AI, ML, data analytics, real-time intelligence analytics and streaming event data to deliver business optimized solutions – Ensure you can take advantage of the Business rules Deep “perfect storm” of rising supply and management and decision learning demand for real-time situation awareness systems and responsiveness IBM Data & AI / © 2019 IBM Corporation Streams Real-time Analytics Platform TCP TCP SOURCE Recommendation Customer Signature detection Engine affinity UDP UDP Consumption Engine TCP Consumer ingest ingest apps Consumer SINK Edge adapters Edge MQ adapters Edge Systemic memory profile and analytics scoring Final customer MQ store HTTP product affinity Click HTTP Product Owners Stream Big Big data sources HDFS HDFS Purchases Recommendation DM models API ODBC ODBC Mobility & Landline Context Affinity Business rules filtering Continuous Activity FILES Files Product TV affinity Content Mgmt. intelligence Activity Consumer Dynamic Viewing Campaign Mgmt. SPSS C&DS example YP.COM Dashboard CE Profile Mgmt. URL Classificati running over SPSS Modeling on Development only DataRepositories STREAMS FILE LANDING ZONE (I/O) User 1700 models Data Reference Identification → Iterative analytical model in production Catalogs Data Lake deployment Production Feeds Demograp →Real-time customer and DEV T.Data Data Warehouse hics clickstream activity Customer DataSource Samples Analytics Schema & Account Consumer Data Consumer Info Systemic Memory → Real-time Content/ Profile Reporting Views Campaign ingestion Modeling Dev. Data Warehouse Data Warehouse →Real-time recommendations Production IBM Streams to act on all your data in real time Market leader in Enterprise Ready: streaming analytics included with Cloud Pak for Data – Machine Learning – Visual development – Model Scoring – Web console – Geospatial – Management – Video/Image – Enterprise – Text, Speech to connectors like Text, Predictive, JSON, JMS, MQ, Descriptive MQTT Data and AI Forum / © 2019 IBM Corporation 7 Streams to act on all your data in real time Filter /Sample Transform Correlate Classify, Annotate Data and AI Forum / © 2019 IBM Corporation 8 IBM Streams offerings Free Lite Plan – 50 Available 1Q19 hours a month Streams v4.3.1 July 2019 Streams for IBM Cloud Pak Streaming Analytics for Data Streams runtime Streams runtime Streams runtime Containers Baremetal/VM Containers SaaS Common runtime: develop anywhere and deployment everywhere Moved from VM to Containers in 2018 Data and AI Forum / © 2019 IBM Corporation *Moved from VM to Containers in 2018 9 IBM Streams development options Streams Developer IBM Cloud Pak Watson Studio Plus IBM Runner Edition, Streams for Data Streams Flows for Apache Beam, Quick Start Edition – Integrated data in IBM Cloud Java, Scala, Python – Dedicated Streams and AI project – Integrated data and beta plug-ins development IDE experience and AI project for VSCode and Atom and tools – Containers* experience – Local machine – SaaS Python Notebooks Quick Start free Available 4Q18 for non-production *Moved from VM to Containers in 2018 Data and AI Forum / © 2019 IBM Corporation 10 IBM Streams at a glance And even more at github.com/IBMStreams Out of the box with over 200 operators with 1300 functions Communications IBM Streams Hadoop Hadoop data sources Scale-out Runtime HDFS, GPFS, Hive, Hbase, BigSQL, – TCP/IP Parquet, Thrift, Avro – UDP/IP Primary analytics – HTTP – Filter – FTP – Enrich Data RDBMS – RSS – Normalize warehouse IBM Db2, IBM Db2 Parallel writer, – Messaging Toolkit (Kafka, – Windowed Aggregations IBM Db2 Event Store, IBM Informix, XMS, IBM MQ, Apache – Machine Learning IBM BigSQL, IBM Netezza, IBM ActiveMQ, RabbitMQ, MQTT) Netezza NZLoad, Oracle, Microsoft – IBM Data Replication – Scoring (SPSS, Python, R, MLlib) SQL Server, MySQL, Teradata, HP – IP Packet ingest Aster, HP Vertica – Signal Processing Application – CEP & Pattern Matching development – xDR Mediation – Geospatial NoSQL – Java NoSQL – Video/Image – Scala Key Value Stores (Memcached, Redis, – Text Analytics (AQL, UIMA) – Python Redis-Cluster, Aerospike), Column – Speech to Text Oriented Stores (Cassandra, Hbase), – Drag/Drop Streams Processing Document Oriented Stores (IBM Language (SPL) – Rules Cloudant, Mongo, Couchbase, Elastic – – Deep Packet Inspection VS Code, Atom for SPL Search), Object Store Data and AI Forum / © 2019 IBM Corporation 11 Machine learning and real time analytics with “The science of getting computers to act without IBM streams being explicitly programmed”1 Many different approaches to Machine Learning: Use machine learning models created with Watson – Mechanisms: Supervised and Unsupervised Studio to score live data on Streaming Analytics – Algorithms: Decision Trees, Regressions, Classification, using Watson Studio Streams Flows. Clustering, etc. – Inputs: Single and multi-variant, rich feature vectors – Streams has 20 ML algorithms that learn as you go in real time – Streams scores models created offline from popular tools including IBM SPSS, Watson Machine Learning, SparkMLLib, Python, R & PMML libraries – Native ML and Model scoring can all be integrated within a Streams application. With this approach, real time scores can be generated on the incoming data Data and AI Forum / © 2019 IBM Corporation 12 Streams: Use the language of choice Streams Processing Language stream<rstring item> Sale = Join(Bid; Ask) { window Bid: sliding, time(30); Streams Processing Language (SPL) Ask: sliding, count(50); param match : Bid.item == Ask.item – Tailored to stream processing && Bid.price >= Ask.price; output Sale: item = Bid.item; – High-level, declarative composition language Java Topology } – Graphical editor support /* * Declare a source stream (hw) with String as tuples that sends * two tuples "Hello" and "World!", and prints them to output. Create topologies in Java */ Topology = new Topology("HelloWorld"); – Indirect support for Scala TStream<String> hw = topology.strings("Hello", "World!"); hw.print(); Python topologies and operators StreamsContextFactory.getEmbedded().submit(topology).get(); – Integration with Jupyter notebook Python Topology – Integration with IBM Watson Studio – Add Python functions in line with SPL code Publish/subscribe data exchange – Between applications written in any language Data and AI Forum / © 2019 IBM Corporation 13 Advantage: Scale out with ease Create logical application flow – Without concern for throughput limitations As needed, add parallel paths – Based on runtime performance profiling User-Defined Parallelism (UDP) – Simple change in code or graphical editor • @parallel annotation Data and AI Forum / © 2019 IBM Corporation 14 Static vs. dynamic composition Static connections – Specified at application development-time and do not change at run-time Dynamic connections – Partially specified at application development-time (Name or properties) – Established at run-time, as new jobs come and go • Specifications can also be updated at run-time Dynamic application composition – Incremental deployment of applications – Dynamic adaptation of applications Data and AI Forum / © 2019 IBM Corporation 15 Application scenarios and real-world use cases IBM Streams is being applied in many industries – Market and Customer Watch the Video Watch the video Watch the Video Watch the Video Watch the Video Intelligence Read the Case Study – Call Center Customer Care – Manufacturing – Personalized Customer Watch the video Watch the video Read the Case Study Read the Case Study Watch the video Experience – Network Analytics – IoT, Connected Car Watch the Video Watch the Video Watch the Video Watch the Video Watch the Video and Telematics – Cyber Security – Health and Improved Patient Outcomes Watch the Video Watch the Video See the Presentation Read the Case Study Read the Case Study – Operational Optimization Data and AI Forum / © 2019 IBM Corporation 16 Medtronic Sugar.IQ IBM Technology Medtronic Sugar.IQ The Results: Sugar.IQ with Watson Watson Platform Past–Present– 90% 36 minutes for Health Future AUC accuracy level when Average more in range – Mobile app management How have I done? predicting the risk of per day hypoglycemia two to four – Data management – Important glucose hours
Recommended publications
  • IBM Software Support
    Version 5.0.0 What’s New . Phone Contacts We have eliminated the phone number section from this document and direct you to IBM Planetwide for contact information. I would appreciate your feedback on what you like and what you think should be improved about this document. e-mail me at mcknight@ us.ibm.com Cover design by Rick Nelson, IBM Raleigh 2 June, 2014 Contents What’s New ................................................................................... 2 Overview of IBM Support .............................................................. 4 IBM Commitment......................................................................... 4 Software Support Organization ...................................................... 4 Software Support Portfolio ................................................................. 5 IBM Software Support ...................................................... 5 Support Foundation .......................................................................... 5 Passport Advantage & IBM Software Maintenance ................................ 6 System z (S/390) ......................................................................... 7 Support Line and SoftwareXcel ..................................................... 7 IBM Selected Support Offering ....................................................... 8 Premium Support ............................................................................ 8 Enhanced Technical Support ........................................................ 8 IBM Software Accelerated Value
    [Show full text]
  • IBM Cloud Unit 2016 IBM Cloud Unit Leadership Organization
    IBM Cloud Technical Academy IBM Cloud Unit 2016 IBM Cloud Unit Leadership Organization SVP IBM Cloud Robert LeBlanc GM Cloud Platform GM Cloud GM Cloud Managed GM Cloud GM Cloud Object Integration Services Video Storage Offering Bill Karpovich Mike Valente Braxton Jarratt Line Execs Line Execs Marie Wieck John Morris GM Strategy, GM Client Technical VP Development VP Service Delivery Business Dev Engagement Don Rippert Steve Robinson Harish Grama Janice Fischer J. Comfort (GM & CTO) J. Considine (Innovation Lab) Function Function Leadership Leadership VP Marketing GM WW Sales & VP Finance VP Human Quincy Allen Channels Resources Steve Cowley Steve Lasher Sam Ladah S. Carter (GM EcoD) GM Design VP Enterprise Mobile GM Digital Phil Gilbert Phil Buckellew Kevin Eagan Missions Missions Enterprise IBM Confidential IBM Hybrid Cloud Guiding Principles Choice with! Hybrid ! DevOps! Cognitive Powerful, Consistency! Integration! Productivity! Solutions! Accessible Data and Analytics! The right Unlock existing Automation, tooling Applications and Connect and extract workload in the IT investments and composable systems that insight from all types right place and Intellectual services to increase have the ability to of data Property speed learn Three entry points 1. Create! 2. Connect! 3. Optimize! new cloud apps! existing apps and data! any app! 2016 IBM Cloud Offerings aligned to the Enterprise’s hybrid cloud needs IBM Cloud Platform IBM Cloud Integration IBM Cloud Managed Offerings Offerings Services Offerings Mission: Build true cloud platform
    [Show full text]
  • Michael Mcgeein, CPA, CA September 2018
    IBM Cognos Analytics Data Prep. and Modelling Working with your data Michael McGeein, CPA, CA September 2018 IBM Business Analytics :: IBM Confidential :: © 2016 IBM Corporation Notices and disclaimers Copyright © 2018 by International Business Machines Corporation (IBM). as illustrations of how those customers have used IBM products and No part of this document may be reproduced or transmitted in any form without the results they may have achieved. Actual performance, cost, savings or other written permission from IBM. results in other operating environments may vary. U.S. Government Users Restricted Rights — use, duplication or disclosure References in this document to IBM products, programs, or services does not restricted by GSA ADP Schedule Contract with IBM. imply that IBM intends to make such products, programs or services available in all countries in which IBM operates or does business. Information in these presentations (including information relating to products that have not yet been announced by IBM) has been reviewed for accuracy as Workshops, sessions and associated materials may have been prepared by of the date of initial publication and could include unintentional technical or independent session speakers, and do not necessarily reflect the typographical errors. IBM shall have no responsibility to update this information. views of IBM. All materials and discussions are provided for informational This document is distributed “as is” without any warranty, either express or purposes only, and are neither intended to, nor shall constitute legal or other implied. In no event shall IBM be liable for any damage arising from the use of guidance or advice to any individual participant or their specific situation.
    [Show full text]
  • Oral History Interview with Sam Wyly
    An Interview with SAM WYLY OH 374 Conducted by David Allison on 6 December 2002 Smithsonian National Museum of American History Washington, D.C. Charles Babbage Institute Center for the History of Information Technology University of Minnesota, Minneapolis Copyright 2004, Charles Babbage Institute Sam Wyly Interview 6 December 2002 Abstract Wyly begins by recounting his childhood, and education prior to going to work for IBM’s Service Bureau Corporation, and then joining Honeywell as an area sales manager. He discusses how he left Honeywell to form University Computer Corporation (UCC), a firm that sold computer time, but transitioned into a software services business. Wyly explains his growing focus on computing and telecommunications, his formation of Datran, and his unsuccessful attempt to acquire Western Union. Much of the interview focuses on ongoing developments at UCC, the eventual sale of this firm to Computer Associates, his formation of Sterling Software, its acquisition of Informatics, the sale of Sterling, and his ideas on the future of information technology. Throughout Wyly’s discussion of UCC and Sterling, he elucidates upon his leadership philosophy, and the strategic, technical, operational, and financial management of these firms. This oral history was sponsored by the Software History Center in conjunction with the Center's ADAPSO reunion (3 May 2002). Preface As part of its preservation activities, the Software History Center (SHC) worked with Dr. David Allison of the Smithsonian Institution’s National Museum of American History and Dr. Jeffrey Yost of the Charles Babbage Institute to plan and conduct a number of oral history interviews of early software company founders and other key industry contributors.
    [Show full text]
  • Regeldokument
    Master’s degree project Source code quality in connection to self-admitted technical debt Author: Alina Hrynko Supervisor: Morgan Ericsson Semester: VT20 Subject: Computer Science Abstract The importance of software code quality is increasing rapidly. With more code being written every day, its maintenance and support are becoming harder and more expensive. New automatic code review tools are developed to reach quality goals. One of these tools is SonarQube. However, people keep their leading role in the development process. Sometimes they sacrifice quality in order to speed up the development. This is called Technical Debt. In some particular cases, this process can be admitted by the developer. This is called Self-Admitted Technical Debt (SATD). Code quality can also be measured by such static code analysis tools as SonarQube. On this occasion, different issues can be detected. The purpose of this study is to find a connection between code quality issues, found by SonarQube and those marked as SATD. The research questions include: 1) Is there a connection between the size of the project and the SATD percentage? 2) Which types of issues are the most widespread in the code, marked by SATD? 3) Did the introduction of SATD influence the bug fixing time? As a result of research, a certain percentage of SATD was found. It is between 0%–20.83%. No connection between the size of the project and the percentage of SATD was found. There are certain issues that seem to relate to the SATD, such as “Duplicated code”, “Unused method parameters should be removed”, “Cognitive Complexity of methods should not be too high”, etc.
    [Show full text]
  • WMS User Guide
    Sterling Warehouse Management System: User Guide Release 8.5 October 2009 Copyright Notice Copyright © 1999 - 2009 Sterling Commerce, Inc. ALL RIGHTS RESERVED STERLING COMMERCE SOFTWARE ***TRADE SECRET NOTICE*** THE STERLING COMMERCE SOFTWARE DESCRIBED BY THIS DOCUMENTATION ("STERLING COMMERCE SOFTWARE") IS THE CONFIDENTIAL AND TRADE SECRET PROPERTY OF STERLING COMMERCE, INC., ITS AFFILIATED COMPANIES OR ITS OR THEIR LICENSORS, AND IS PROVIDED UNDER THE TERMS OF A LICENSE AGREEMENT. NO DUPLICATION OR DISCLOSURE WITHOUT PRIOR WRITTEN PERMISSION. RESTRICTED RIGHTS. This documentation, the Sterling Commerce Software it describes, and the information and know-how they contain constitute the proprietary, confidential and valuable trade secret information of Sterling Commerce, Inc., its affiliated companies or its or their licensors, and may not be used for any unauthorized purpose, or disclosed to others without the prior written permission of the applicable Sterling Commerce entity. This documentation and the Sterling Commerce Software that it describes have been provided pursuant to a license agreement that contains prohibitions against and/or restrictions on their copying, modification and use. Duplication, in whole or in part, if and when permitted, shall bear this notice and the Sterling Commerce, Inc. copyright notice. Commerce, Inc. copyright notice. U.S. GOVERNMENT RESTRICTED RIGHTS. This documentation and the Sterling Commerce Software it describes are "commercial items" as defined in 48 C.F.R. 2.101. As and when provided to any agency or instrumentality of the U.S. Government or to a U.S. Government prime contractor or a subcontractor at any tier ("Government Licensee"), the terms and conditions of the customary Sterling Commerce commercial license agreement are imposed on Government Licensees per 48 C.F.R.
    [Show full text]
  • Trifacta Data Preparation for Amazon Redshift and S3 Must Be Deployed Into an Existing Virtual Private Cloud (VPC)
    Install Guide for Data Preparation for Amazon Redshift and S3 Version: 7.1 Doc Build Date: 05/26/2020 Copyright © Trifacta Inc. 2020 - All Rights Reserved. CONFIDENTIAL These materials (the “Documentation”) are the confidential and proprietary information of Trifacta Inc. and may not be reproduced, modified, or distributed without the prior written permission of Trifacta Inc. EXCEPT AS OTHERWISE PROVIDED IN AN EXPRESS WRITTEN AGREEMENT, TRIFACTA INC. PROVIDES THIS DOCUMENTATION AS-IS AND WITHOUT WARRANTY AND TRIFACTA INC. DISCLAIMS ALL EXPRESS AND IMPLIED WARRANTIES TO THE EXTENT PERMITTED, INCLUDING WITHOUT LIMITATION THE IMPLIED WARRANTIES OF MERCHANTABILITY, NON-INFRINGEMENT AND FITNESS FOR A PARTICULAR PURPOSE AND UNDER NO CIRCUMSTANCES WILL TRIFACTA INC. BE LIABLE FOR ANY AMOUNT GREATER THAN ONE HUNDRED DOLLARS ($100) BASED ON ANY USE OF THE DOCUMENTATION. For third-party license information, please select About Trifacta from the Help menu. 1. Quick Start . 4 1.1 Install from AWS Marketplace . 4 1.2 Upgrade for AWS Marketplace . 7 2. Configure . 8 2.1 Configure for AWS . 8 2.1.1 Configure for EC2 Role-Based Authentication . 14 2.1.2 Enable S3 Access . 16 2.1.2.1 Create Redshift Connections 28 3. Contact Support . 30 4. Legal 31 4.1 Third-Party License Information . 31 Page #3 Quick Start Install from AWS Marketplace Contents: Product Limitations Internet access Install Desktop Requirements Pre-requisites Install Steps - CloudFormation template SSH Access Troubleshooting SELinux Upgrade Documentation Related Topics This guide steps through the requirements and process for installing Trifacta® Data Preparation for Amazon Redshift and S3 through the AWS Marketplace.
    [Show full text]
  • Portable Stateful Big Data Processing in Apache Beam
    Portable stateful big data processing in Apache Beam Kenneth Knowles Apache Beam PMC Software Engineer @ Google https://s.apache.org/ffsf-2017-beam-state [email protected] / @KennKnowles Flink Forward San Francisco 2017 Agenda 1. What is Apache Beam? 2. State 3. Timers 4. Example & Little Demo What is Apache Beam? TL;DR (Flink draws it more like this) 4 DAGs, DAGs, DAGs Apache Beam Apache Flink Apache Cloud Hadoop Apache Apache Dataflow Spark Samza MapReduce Apache Apache Apache (paper) Storm Gearpump Apex (incubating) FlumeJava (paper) Heron MillWheel (paper) Dataflow Model (paper) 2004 2005 2006 2007 2008 2009 2010 2011 2012 2013 2014 2015 2016 Apache Flink local, on-prem, The Beam Vision cloud Cloud Dataflow: Java fully managed input.apply( Apache Spark Sum.integersPerKey()) local, on-prem, cloud Sum Per Key Apache Apex Python local, on-prem, cloud input | Sum.PerKey() Apache Gearpump (incubating) ⋮ ⋮ 6 Apache Flink local, on-prem, The Beam Vision cloud Cloud Dataflow: Python fully managed input | KakaIO.read() Apache Spark local, on-prem, cloud KafkaIO Apache Apex ⋮ local, on-prem, cloud Apache Java Gearpump (incubating) class KafkaIO extends UnboundedSource { … } ⋮ 7 The Beam Model PTransform Pipeline PCollection (bounded or unbounded) 8 The Beam Model What are you computing? (read, map, reduce) Where in event time? (event time windowing) When in processing time are results produced? (triggers) How do refinements relate? (accumulation mode) 9 What are you computing? Read ParDo Grouping Composite Parallel connectors to Per element Group
    [Show full text]
  • The Forrester Wave™: Streaming Analytics, Q3 2019 the 11 Providers That Matter Most and How They Stack up by Mike Gualtieri September 23, 2019
    LICENSED FOR INDIVIDUAL USE ONLY The Forrester Wave™: Streaming Analytics, Q3 2019 The 11 Providers That Matter Most And How They Stack Up by Mike Gualtieri September 23, 2019 Why Read This Report Key Takeaways In our 26-criterion evaluation of streaming Software AG, IBM, Microsoft, Google, And analytics providers, we identified the 11 most TIBCO Software Lead The Pack significant ones — Alibaba, Amazon Web Forrester’s research uncovered a market in which Services, Cloudera, EsperTech, Google, IBM, Software AG, IBM, Microsoft, Google, and TIBCO Impetus, Microsoft, SAS, Software AG, and Software are Leaders; Cloudera, SAS, Amazon TIBCO Software — and researched, analyzed, Web Services, and Impetus are Strong Performers; and scored them. This report shows how each and EsperTech and Alibaba are Contenders. provider measures up and helps application Analytics Prowess, Scalability, And development and delivery (AD&D) professionals Deployment Freedom Are Key Differentiators select the right one for their needs. Depth and breadth of analytics types on streaming data are critical. But that is all for naught if streaming analytics vendors cannot also scale to handle potentially huge volumes of streaming data. Also, it’s critical that streaming analytics can be deployed where it is most needed, such as on-premises, in the cloud, and/ or at the edge. This PDF is only licensed for individual use when downloaded from forrester.com or reprints.forrester.com. All other distribution prohibited. FORRESTER.COM FOR APPLICATION DEVELOPMENT & DELIVERY PROFESSIONALS The Forrester Wave™: Streaming Analytics, Q3 2019 The 11 Providers That Matter Most And How They Stack Up by Mike Gualtieri with Srividya Sridharan and Robert Perdoni September 23, 2019 Table Of Contents Related Research Documents 2 Enterprises Must Take A Streaming-First The Future Of Machine Learning Is Unstoppable Approach To Analytics Predictions 2019: Artificial Intelligence 3 Evaluation Summary Predictions 2019: Business Insights 6 Vendor Offerings 6 Vendor Profiles Leaders Share reports with colleagues.
    [Show full text]
  • Technology Data Card
    FountMedia Delivers the highest quality business information FountMedia- is a leading provider of Targeted contact lists, with experience in Direct and Interactive Marketing Services for our valued clients. Our database consists of over 150 Million prospects from 200+ highly targeted business contact lists with the ability to provide detailed segmentation to isolate your target audience. EACH RECORD INCLUDES: Company Name, Address, Address1, City, State, Zip Code, Country, Phone Number, Fax Number, Employees, Revenue, SIC Code, Industry, Web Address, Contact Name, First Name, Middle Name, Last Name, Job Title, and Email Address. GUARANTEE: 90%+ on email deliverability and 97%+ guarantee on other data fields. Replacements: If you encounter faulty contacts more than the above parameters would be replaced with fresh contacts or we would provide you a credit balance. Update: We constantly update and verify our information to ensure that your business mailing lists always contain the most current, most deliverable addresses and responsive phone numbers available. We update our list every 60-90 days to make sure it delivers to the recipients. Our accuracy standards would have lower bounce rates, and better deliverability to protect email sender reputation Ownership: List will have 100% verified emails and are provided for your complete ownership i.e. for your unlimited Multi-Channel marketing usage. ADDITIONAL BENEFITS: We will provide 3 updates in the next 12 months. Along with each update, you will receive 6% to 8% new contacts from the same segment. Eventually, you will have 20%+ new contacts end of 12 months and existing names will also been cleansed and fresh so you don’t have to spend another budget next year.
    [Show full text]
  • Scalable and Flexible Middleware for Dynamic Data Flows
    SCALABLEANDFLEXIBLEMIDDLEWAREFORDYNAMIC DATAFLOWS stephan boomker Master thesis Computing Science Software Engineering and Distributed Systems Primaray RuG supervisor : Prof. Dr. M. Aiello Secondary RuG supervisor : Prof. Dr. P. Avgeriou Primary TNO supervisor : MSc. E. Harmsma Secondary TNO supervisor : MSc. E. Lazovik August 16, 2016 – version 1.0 [ August 16, 2016 at 11:47 – classicthesis version 1.0 ] Stephan Boomker: Scalable and flexible middleware for dynamic data flows, Master thesis Computing Science, © August 16, 2016 [ August 16, 2016 at 11:47 – classicthesis version 1.0 ] ABSTRACT Due to the concepts of Internet of Things and Big data, the traditional client-server architecture is not sufficient any more. One of the main reasons is wide range of expanding heterogeneous applications, data sources and environments. New forms of data processing require new architectures and techniques in order to be scalable, flexible and able to handle fast dynamic data flows. The backbone of all those objects, applications and users is called the middleware. This research goes about designing and implementing a middle- ware by taking into account different state of the art tools and tech- niques. To come up to a solution which is able to handle a flexible set of sources and models across organizational borders. At the same time it is de-centralized, distributed and, although de-central able to perform semantic based system integration centrally. This is accom- plished by introducing of an architecture containing a combination of data integration patterns, semantic storage and stream processing patterns. A reference implementation is presented of the proposed architec- ture based on Apache Camel framework. This prototype provides the ability to dynamically create and change flexible and distributed data flows during runtime.
    [Show full text]
  • Big Data Analysis Using Hadoop Lecture 4 Hadoop Ecosystem
    1/16/18 Big Data Analysis using Hadoop Lecture 4 Hadoop EcoSystem Hadoop Ecosytems 1 1/16/18 Overview • Hive • HBase • Sqoop • Pig • Mahoot / Spark / Flink / Storm • Hadoop & Data Management Architectures Hive 2 1/16/18 Hive • Data Warehousing Solution built on top of Hadoop • Provides SQL-like query language named HiveQL • Minimal learning curve for people with SQL expertise • Data analysts are target audience • Ability to bring structure to various data formats • Simple interface for ad hoc querying, analyzing and summarizing large amountsof data • Access to files on various data stores such as HDFS, etc Website : http://hive.apache.org/ Download : http://hive.apache.org/downloads.html Documentation : https://cwiki.apache.org/confluence/display/Hive/LanguageManual Hive • Hive does NOT provide low latency or real-time queries • Even querying small amounts of data may take minutes • Designed for scalability and ease-of-use rather than low latency responses • Translates HiveQL statements into a set of MapReduce Jobs which are then executed on a Hadoop Cluster • This is changing 3 1/16/18 Hive • Hive Concepts • Re-used from Relational Databases • Database: Set of Tables, used for name conflicts resolution • Table: Set of Rows that have the same schema (same columns) • Row: A single record; a set of columns • Column: provides value and type for a single value • Can can be dived up based on • Partitions • Buckets Hive – Let’s work through a simple example 1. Create a Table 2. Load Data into a Table 3. Query Data 4. Drop a Table 4 1/16/18
    [Show full text]