Oracle® Exadata and IBM® Netezza® Data Warehouse Appliance

Total Page:16

File Type:pdf, Size:1020Kb

Oracle® Exadata and IBM® Netezza® Data Warehouse Appliance IBM Software Information Management Oracle® Exadata® and IBM® Netezza® Data Warehouse Appliance compared by Phil Francisco, Vice President, Product Management and Product Marketing, IBM and Mike Kearney, Senior Director, Product Marketing, IBM Information Management Oracle Exadata and IBM Netezza Data Warehouse Appliance compared Contents Introduction Introduction 2 IBM Netezza data warehouse appliances focus on technology designed to query and analyze big data. Online Transaction Processing (OLTP) “ Netezza was part of the inspiration for and data warehousing 3 IBM Netezza data warehouse appliances are disrupting the market. Wishing to exploit data at Query performance 5 Exadata. Teradata was part of the lower costs of operation and ownership, many of our Simplicity of operation 10 inspiration for Exadata. We’d like to customers have moved their data warehouses from Value 14 Oracle. Oracle has now brought Exadata to market, thank them for forcing our hand Conclusion 16 a machine which apparently does everything an IBM and forcing us to go into the Netezza data warehouse appliance does, and also hardware business.” processes online transactions. This examination of Exadata and IBM Netezza as data warehouse — Larry Ellison, January 2010 platforms is written from an unashamedly IBM viewpoint, however to ensure credibility we have taken advice from Philip Howard, Research Director of Bloor Research and Curt Monash, warehouses. More important, our customers create President, Monash Research. new business value by deploying analytic applications which they previously considered To innovate requires us to think and do things beyond their reach. differently, solving a problem using new approaches. IBM Netezza data warehouse appliances deliver “Netezza was part of the inspiration for Exadata. excellent performance for our customers’ Teradata was part of the inspiration for Exadata,” warehouse queries. IBM Netezza data warehouse acknowledged Larry Ellison on January 27, 2010. appliances offer customers simplicity; anyone with “We’d like to thank them for forcing our hand and basic knowledge of SQL and Linux has the skills forcing us go into the hardware business.”1 While needed to perform the few administrative tasks delivered with Larry Ellison’s customary pizzazz, required to maintain consistent service levels through dynamically changing workloads. IBM Netezza data warehouse appliances’ performance with simplicity 1 See http://oracle.com.edgesuite.net/ivt/4000/8104/9238/12652/ reduces their costs of owning and running their data lobby_external_flash_clean_480x360/default.htm 2 Information Management Oracle Exadata and IBM Netezza Data Warehouse Appliance compared there is a serious point to his comment: only the best All we ask of readers is that they do as our Online Transaction Processing catch Oracle’s attention. Exadata represents a customers and partners have done: put aside strategic direction for Oracle; adapting their OLTP notions of how a database management system (OLTP) and data warehousing database management system, partnering it with a should work, be open to new ways of thinking OLTP systems execute many short transactions. massively parallel storage system from Sun. Oracle and be prepared to do less, not more, to achieve Each transaction’s scope is small, limited to one or a launched Exadata V2 with the promise of extreme a better result. small number of records and is so predictable that performance for processing both online transactions often times data is cached. Although OLTP systems and analytic queries. Therefore, Oracle Exadata V2 is One caveat: The IBM Netezza data warehouse process large volumes of database queries, their a general purpose platform for managing mixed appliance team has no direct access to an Exadata focus is writing (UPDATE, INSERT and DELETE) to a workloads. Oracle Database was designed for OLTP. machine. We are fortunate in the detailed feedback current data set. These systems are typically specific But data warehousing and analytics make very we receive from many organizations that have to a business process or function, for example different demands of their software and hardware evaluated both technologies and selected IBM managing the current balance of a checking than OLTP. Quite simply, some workloads for data Netezza data warehouse appliances. Given Oracle’s account. Their data is commonly structured in third warehousing perform much better and are more cost size and their focus on Exadata, publicly available normal form (3NF). Transaction types of OLTP effective on a system that is purpose-built for information on Exadata is surprisingly scarce. systems are stable and their data requirements are analytics. Exadata’s data warehousing credentials The use cases quoted by Oracle provide little input well-understood, so secondary data structures such demand scrutiny, particularly with respect to to the discussion, which in itself is of concern to as indices can usefully locate records on disk, prior simplicity and value. several industry followers, e.g., Information Week.2 to their transfer to memory for processing. The information shared in this paper is made This eBook opens by reviewing differences between available in the spirit of openness. Any inaccuracies In comparison, data warehouse systems are processing online transactions and processing result from our mistakes, not an intent to mislead. characterized by predominantly heavy database read queries and analyses in a data warehouse. It then (SELECT) operations against a current and historical discusses Exadata and the IBM Netezza data data set. Whereas an OLTP operation accesses a warehouse appliance from perspectives of their small number of records, a data warehouse query query performance, simplicity of operation and value. might scan a table of billions of rows and join its records with those from multiple other tables. Furthermore, queries in a data warehouse are often so unpredictable in nature, it is difficult to exploit caching and indexing strategies. Choices for 2 See http://www.informationweek.com/news/business_intelligence/ warehouses/showArticle.jhtml? articleID=225702836&cid=RSSfeed_ structuring data in the warehouse range from IWK_News 3NF to dimensional models such as star and 3 Information Management Oracle Exadata and IBM Netezza Data Warehouse Appliance compared snowflake schemas. Data within each system In this century, corporations and public sector Organizations are redefining how they need and feeding a typical warehouse is structured to reflect agencies accept growth rates for data of 30-50 want to exploit their data; this eBook refers to this the needs of a specific business process. Before percent per year as normal. Technologies and development as the second-generation data data is loaded to the warehouse it is cleansed, practices successful in the world of OLTP prove less warehouse. These new warehouses, managing de-duplicated and integrated. and less applicable to data warehousing; the index massive data sets with ease, serve as the corporate as aid to data retrieval is a case in point. As the memory. When interrogated, they recall events This eBook divides data warehouses into either first- database system processes jobs to load data, it is recorded years previously; these distant memories or second-generation. While this classification may also busy updating its multiple indices. With large increase the accuracy of predictive analytic not stand the deepest scrutiny, it reflects how many data volumes this becomes a very slow process, applications. Constant trickle feeds are replacing of our customers talk about their evolutionary path to causing load jobs to overrun their allotted processing overnight batch loads, reducing latency between the generating greater and greater value from their data. window. Despite working long hours, the technical recording of an event and its analysis. Beyond the team misses service levels negotiated with the simple SQL used to populate reports and First-generation data warehouses are typically business. Productivity suffers as business units wait dashboards, the warehouse processes linear loaded overnight. They provide information to their for reports and data to become available. regressions, Naive Bayes and other mathematical business via a stable body of slowly evolving SQL- algorithms of advanced analytics. Noticing a sudden based reports and dashboards. As these simple spike in sales of a high-margin product at just five warehouses somewhat resemble OLTP systems – stores drives a retailer to understand what happened their workload and data requirements are Technologies and practices successful in the and why. This knowledge informs strategies to understood and stable – organizations often adopt promote similar sales activity at all 150 store the same database management products they use world of OLTP prove less and less locations. The computing system underpinning the for OLTP. With the product comes the practice: applicable to data warehousing… warehouse must be capable of managing these database administrators analyze each report’s data sudden surges in demand without disrupting regular requirements and build indices to accelerate data reports and dashboards. The business users are retrieval. Creep of OLTP’s technology and techniques demanding the freedom to exploit their data at appears a success, until data volumes in the the time and in the manner of their choosing. warehouse outstrip those commonly managed in Their appetite for immediacy leaves
Recommended publications
  • Mcafee Database Security Installation Guide
    CONFIDENTIAL McAfee Database Security Installation Guide McAfee Integrity Monitor for Databases McAfee Database Activity Monitoring McAfee Vulnerability Manager for Databases McAfee Vulnerability Manager for Databases End User License Agreement NOTICE TO ALL USERS: PLEASE READ THIS CONTRACT CAREFULLY. BY CLICKING THE ACCEPT BUTTON OR INSTALLING THE SOFTWARE, YOU (EITHER AN INDIVIDUAL OR A SINGLE ENTITY) AGREE THAT THIS AGREEMENT IS ENFORCEABLE LIKE ANY WRITTEN CONTRACT SIGNED BY YOU. IF YOU DO NOT AGREE TO ALL THE TERMS OF THIS AGREEMENT, CLICK ON THE BUTTON THAT INDICATES THAT YOU DO NOT ACCEPT THE TERMS OF THIS CONTRACT AND DO NOT INSTALL THE SOFTWARE. 1. Definitions. a. “Software” means (a) all of the contents of the files, disk(s), CD-ROM(s) or other media (including electronic media) with which this Agreement is provided or such contents as are hosted by McAfee or its distributors, resellers, OEM/MSP partners, or other business partners (collectively “Authorized Partner(s)”), including but not limited to (i) McAfee or third party computer information or software; (ii) related explanatory materials in printed, electronic, or online form (“Documentation”); and (b) upgrades, modified or subsequent versions and updates including any virus or vulnerability updates (collectively “Updates”), and Software, if any, licensed to you by McAfee or an Authorized Partner as part of a maintenance contract or service subscription. b. “Use” or “Using” means to access, install, download, copy or otherwise benefit from using the Software. c. “Permitted Number” means one (1) unless otherwise indicated under a valid license (e.g., volume license) granted by McAfee. d. “Computer” means a device that accepts information in digital or similar form and manipulates it for a specific result based upon a sequence of instructions.
    [Show full text]
  • Query Optimizing for On-Line Analytical Processing Adventures in the Land of Heuristics
    Aalto University School of Science Master's Programme in Computer, Communication and Information Sciences Jonas Berg Query Optimizing for On-line Analytical Processing Adventures in the land of heuristics Master's Thesis Espoo, May 22, 2017 Supervisor: Professor Eljas Soisalon-Soininen Advisors: Jarkko Miettinen M.Sc. (Tech.) Marko Nikula M.Sc. (Tech) Aalto University School of Science Master's Programme in Computer, Communication and In- ABSTRACT OF formation Sciences MASTER'S THESIS Author: Jonas Berg Title: Query Optimizing for On-line Analytical Processing { Adventures in the land of heuristics Date: May 22, 2017 Pages: vii + 71 Major: Computer Science Code: SCI3042 Supervisor: Professor Eljas Soisalon-Soininen Advisors: Jarkko Miettinen M.Sc. (Tech.) Marko Nikula M.Sc. (Tech) Newer database technologies, such as in-memory databases, have largely forgone query optimization. In this thesis, we presented a use case for query optimization for an in-memory column-store database management system used for both on- line analytical processing and on-line transaction processing. To date, the system in question has used a na¨ıve query optimizer for deciding on join order. We went through related literature on the history and evolution of database technology, focusing on query optimization. Based on this, we analyzed the current system and presented improvements for its query processing abilities. We implemented a new query optimizer and experimented with it, seeing how it performed on several queries, concluding that it is a successful improvement
    [Show full text]
  • Mcafee Real-Time Database Monitoring, Auditing, and Intrusion
    White Paper McAfee Real-Time Database Monitoring, Auditing, and Intrusion Prevention Table of Contents Introduction 3 Existing Components of Database Security 3 Authentication and access control 3 Native database audit tools 4 The Network Approach to Database Activity Monitoring 4 The local access challenge 4 The encryption challenge 6 McAfee Database Activity Monitoring: Real-Time Database Monitoring 7 and Intrusion Prevention Unique advantages 11 2 McAfee Real-Time Database Monitoring, Auditing, and Intrusion Prevention Databases store an enterprise’s most valuable information assets, but in most organizations today, they are poorly protected. It’s good common sense to secure databases as well as or better than other systems in the enterprise. But it’s not that simple. In this white paper, we take a look at the shortcomings of the database security tools used by enterprises worldwide and the technology and approach required to provide real-time, comprehensive protection for all types of databases. Components of Database Security A wide array of technologies and tools are currently used to provide various aspects of database security. As with other areas of IT security, no single tool can provide an ironclad defense against all threats and abuses. It’s best to implement a combination of tools to achieve adequate security. The two main categories of tools in use across enterprises large and small are authentication/access “I hack, I ruin, I make piles of control and native database audit tools. money.” Authentication and access control —Andrew Auernheimer The ability to designate roles, logins, and passwords is the most basic and most widely used level of (Charged in US Federal Court with database security.
    [Show full text]
  • The Gamma Database Machine Project
    The Gamma Database Machine Project David J. DeWitt Shahram Ghandeharizadeh Donovan Schneider Allan Bricker Hui-I Hsiao Rick Rasmussen Computer Sciences Department University of Wisconsin This research was partially supported by the Defense Advanced Research Projects Agency under contract N00039- 86-C-0578, by the National Science Foundation under grant DCR-8512862, by a DARPA/NASA sponsored Gradu- ate Research Assistantship in Parallel Processing, and by research grants from Intel Scientific Computers, Tandem Computers, and Digital Equipment Corporation. ABSTRACT This paper describes the design of the Gamma database machine and the techniques employed in its imple- mentation. Gamma is a relational database machine currently operating on an Intel iPSC/2 hypercube with 32 pro- cessors and 32 disk drives. Gamma employs three key technical ideas which enable the architecture to be scaled to 100s of processors. First, all relations are horizontally partitioned across multiple disk drives enabling relations to be scanned in parallel. Second, novel parallel algorithms based on hashing are used to implement the complex rela- tional operators such as join and aggregate functions. Third, dataflow scheduling techniques are used to coordinate multioperator queries. By using these techniques it is possible to control the execution of very complex queries with minimal coordination - a necessity for configurations involving a very large number of processors. In addition to describing the design of the Gamma software, a thorough performance evaluation of the iPSC/2 hypercube version of Gamma is also presented. In addition to measuring the effect of relation size and indices on the response time for selection, join, aggregation, and update queries, we also analyze the performance of Gamma relative to the number of processors employed when the sizes of the input relations are kept constant (speedup) and when the sizes of the input relations are increased proportionally to the number of processors (scaleup).
    [Show full text]
  • Security Guide for Oracle Exadata Database Machine
    Oracle® Exadata Database Machine Security Guide for Oracle Exadata Database Machine 21.2 F29252-09 September 2021 Oracle Exadata Database Machine Security Guide for Oracle Exadata Database Machine, 21.2 F29252-09 Copyright © 2008, 2021, Oracle and/or its affiliates. Primary Authors: Peter Fusek, Glenn Maxey Contributing Authors: Craig Brown, Dan Norris, James Spiller Contributors: Yang Liu, Philip Newlan, Tina Rose, Kevin Simmons, Zheren Zhang This software and related documentation are provided under a license agreement containing restrictions on use and disclosure and are protected by intellectual property laws. Except as expressly permitted in your license agreement or allowed by law, you may not use, copy, reproduce, translate, broadcast, modify, license, transmit, distribute, exhibit, perform, publish, or display any part, in any form, or by any means. Reverse engineering, disassembly, or decompilation of this software, unless required by law for interoperability, is prohibited. The information contained herein is subject to change without notice and is not warranted to be error-free. If you find any errors, please report them to us in writing. If this is software or related documentation that is delivered to the U.S. Government or anyone licensing it on behalf of the U.S. Government, then the following notice is applicable: U.S. GOVERNMENT END USERS: Oracle programs (including any operating system, integrated software, any programs embedded, installed or activated on delivered hardware, and modifications of such programs) and Oracle computer documentation or other Oracle data delivered to or accessed by U.S. Government end users are "commercial computer software" or "commercial computer software documentation" pursuant to the applicable Federal Acquisition Regulation and agency-specific supplemental regulations.
    [Show full text]
  • Foglight Performance Tuning Field Guide Updated - January 2018 Software Version - 5.9.2 Overview
    Foglight® 5.9.2 Performance Tuning Field Guide © 2018 Quest Software Inc. ALL RIGHTS RESERVED. This guide contains proprietary information protected by copyright. The software described in this guide is furnished under a software license or nondisclosure agreement. This software may be used or copied only in accordance with the terms of the applicable agreement. No part of this guide may be reproduced or transmitted in any form or by any means, electronic or mechanical, including photocopying and recording for any purpose other than the purchaser’s personal use without the written permission of Quest Software Inc. The information in this document is provided in connection with Quest Software products. No license, express or implied, by estoppel or otherwise, to any intellectual property right is granted by this document or in connection with the sale of Quest Software products. EXCEPT AS SET FORTH IN THE TERMS AND CONDITIONS AS SPECIFIED IN THE LICENSE AGREEMENT FOR THIS PRODUCT, QUEST SOFTWARE ASSUMES NO LIABILITY WHATSOEVER AND DISCLAIMS ANY EXPRESS, IMPLIED OR STATUTORY WARRANTY RELATING TO ITS PRODUCTS INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTY OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE, OR NON-INFRINGEMENT. IN NO EVENT SHALL QUEST SOFTWARE BE LIABLE FOR ANY DIRECT, INDIRECT, CONSEQUENTIAL, PUNITIVE, SPECIAL OR INCIDENTAL DAMAGES (INCLUDING, WITHOUT LIMITATION, DAMAGES FOR LOSS OF PROFITS, BUSINESS INTERRUPTION OR LOSS OF INFORMATION) ARISING OUT OF THE USE OR INABILITY TO USE THIS DOCUMENT, EVEN IF QUEST SOFTWARE HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH DAMAGES. Quest Software makes no representations or warranties with respect to the accuracy or completeness of the contents of this document and reserves the right to make changes to specifications and product descriptions at any time without notice.
    [Show full text]
  • Administering JDBC Data Sources for Oracle Weblogic Server 12C (12.2.1.2.0) E77969-03
    Oracle® Fusion Middleware Administering JDBC Data Sources for Oracle WebLogic Server 12c (12.2.1.2.0) E77969-03 December 2016 This book contains JDBC data source configuration and administration information. Oracle Fusion Middleware Administering JDBC Data Sources for Oracle WebLogic Server, 12c (12.2.1.2.0) E77969-03 Copyright © 2015, 2016, Oracle and/or its affiliates. All rights reserved. This software and related documentation are provided under a license agreement containing restrictions on use and disclosure and are protected by intellectual property laws. Except as expressly permitted in your license agreement or allowed by law, you may not use, copy, reproduce, translate, broadcast, modify, license, transmit, distribute, exhibit, perform, publish, or display any part, in any form, or by any means. Reverse engineering, disassembly, or decompilation of this software, unless required by law for interoperability, is prohibited. The information contained herein is subject to change without notice and is not warranted to be error-free. If you find any errors, please report them to us in writing. If this is software or related documentation that is delivered to the U.S. Government or anyone licensing it on behalf of the U.S. Government, then the following notice is applicable: U.S. GOVERNMENT END USERS: Oracle programs, including any operating system, integrated software, any programs installed on the hardware, and/or documentation, delivered to U.S. Government end users are "commercial computer software" pursuant to the applicable Federal Acquisition Regulation and agency- specific supplemental regulations. As such, use, duplication, disclosure, modification, and adaptation of the programs, including any operating system, integrated software, any programs installed on the hardware, and/or documentation, shall be subject to license terms and license restrictions applicable to the programs.
    [Show full text]
  • Cost/Benefit Case for IBM Db2 11.1 Compared to Oracle Database 12C for OLAP Deployments Sponsored By: IBM
    WHITEPAPER / JULY 14, 2017 / QL218356 Cost/Benefit Case for IBM Db2 11.1 Compared to Oracle Database 12c for OLAP Deployments Sponsored by: IBM Market Situation Today, organizations looking to grow their businesses have the opportunity to adopt innovative analytics applications to facilitate the process. New generation applications, coupled with modern database technologies and cutting-edge hardware, will enable organizations to rapidly generate actionable business insights to increase competitive advantage. At the same time, organizations must not only evaluate the performance of these business intelligence (BI) and analytics applications and the systems required to run them, but also their cost-effectiveness, compatibility with existing IT assets, capacity to meet growing organizational requirements, and ability to ensure business continuity. An increasing number of businesses are investing in cost-effective emerging technologies as a way to boost competitive advantage. However, despite the spotlight placed on BI and analytics offerings, the database foundation remains the crucial engine driving businesses processes and critical decision making for most organizations. A database platform that efficiently manages data and allows users to extract insights with ease will improve an organization’s efficiently and agility in confronting the challenges faced by modern businesses. IBM Db2, Oracle Database, MySQL, and Microsoft SQL Server continue to be dominant enterprise-ready databases, each supported by various hardware from established vendors. This paper presents a cost/benefit case for two industry-leading database platforms for analytics workloads—IBM Power Systems with AIX running IBM Db2 11.1 with BLU Acceleration, and Oracle Exadata Database Machine configured with Oracle Linux and Oracle Database 12c Release 2.
    [Show full text]
  • Risparmiare Consolidando Con Sun Oracle Database Machine
    <Insert Picture Here> Risparmiare consolidando con Sun Oracle Database Machine Alessandro Bracchini Sales Consulting – Oracle Italia Infrastruttura tradizionale: costosa ed inefficiente App A App B App C App D App E • Architettura a “Silos” lascia sottoutilizzate per diverso tempo le risorse Middleware HW disponibili Server 1 Server 2 Server 3 Server 4 Server 5 Scarso • utilizzo medio Alta Disponibilita’ ottenuta raddoppiando DB A DB B DB C DB D DB E risorse Server – Alti costi Database • Non ecologica: richiede piu’ elettricita’, spazio, Server 6 Server 7 Server 8 Server 9 Server 10 condizionamento, … Storage Consolidamento con il Grid Computing • E’ possibile far coesistere su risorse condivise virtuali Utilizzo medio Server carichi di lavoro Middleware Consolidated Apps A-E 75% Riutilizzabili complementari Server 1 Server 2 Server 3 Server 4 Server 5 • Si ottimizza il livello di utilizzo medio dei Server risparmiando Server sugli acquisti HW Database Cluster Database Riutilizzabili Utilizzo medio 70% ABCDE Server 6 Server 7 Server 8 Server 9 Server 10 • L’architettura e’ intrinsecamente affidabile Storage Storage Shared, Consolidated Storage Riutilizzabile • Maggiore efficienza = Green Computing Grid Computing Consolidamento e Virtualizzazione risorse IT •Dimensionati per picchi di carico •Gruppi di risorse condivise •Rigida •Ridistribuzione risorse se necessario •Complessa da gestire •Gestione centralizzata Best Machine for Consolidating Databases ● ERP Più database su uno stesso server ● Un unico storage per tutti i database CRM
    [Show full text]
  • Fortidb 5.1.13 Admin Guide 00-400-000000-20181031 TABLE of CONTENTS
    FortiDB - Admin Guide Version 5.1.13 FORTINET DOCUMENT LIBRARY https://docs.fortinet.com FORTINET VIDEO GUIDE https://video.fortinet.com FORTINET BLOG https://blog.fortinet.com CUSTOMER SERVICE & SUPPORT https://support.fortinet.com FORTINET TRAINING & CERTIFICATION PROGRAM https://www.fortinet.com/support-and-training/training.html NSE INSTITUTE https://training.fortinet.com FORTIGUARD CENTER https://fortiguard.com/ END USER LICENSE AGREEMENT https://www.fortinet.com/doc/legal/EULA.pdf FEEDBACK Email: [email protected] October 31, 2018 FortiDB 5.1.13 Admin Guide 00-400-000000-20181031 TABLE OF CONTENTS What’s new 11 Introduction 15 FortiDB tutorials 16 Tutorial: Generating a vulnerability assessment (VA) report 16 Tutorial: Monitoring a database table using the TCP/IP sniffer 19 Tutorial: Monitoring a database table using the native auditing feature 23 Tutorial: Monitoring changes to metadata 26 Tutorial: Generating PCI, SOX, and HIPAA compliance reports 28 Installation (software-only) 31 System requirements 31 Preparing to install 32 Configuring the FortiDB repository database 33 Configuring a PostgreSQL repository 33 Configuring an Oracle repository 34 Configuring an Microsoft SQL Server repository 35 UNIX/Linux installation 36 Windows installation 37 Confirming the installation 37 Starting or stopping FortiDB 38 Installing a new license 38 Managing disk space 39 Useful directories, files, and folders 39 Log files for troubleshooting 40 General logs 40 Tomcat logs 41 Upgrading FortiDB 41 How to set up your FortiDB 42 Registering your FortiDB 42 Planning the network topology for database activity monitoring (DAM) 42 Connecting to the web UI and CLI 43 Updating the firmware 43 Upgrading the firmware 44 Installing FortiDB firmware 45 Changing the "admin" account password 46 Setting the system time 47 Configuring the network settings 49 Configuring network settings using the web UI 49 Configuring network settings using the CLI 51 Backups 52 Administrators 53 Configuring permissions 54 FortiDB 5.1.13 Admin Guide 3 Fortinet Inc.
    [Show full text]
  • Datafeed Loader User Guide V2.10.1.0
    DataFeed Loader User Guide V2.10.1.0 Last Edited: August 2018 Table of Contents Updates ............................................................................................................................................................. 6 Overview ........................................................................................................................................................... 8 Product Offering ................................................................................................................................................ 8 Documentation ................................................................................................................................................. 8 Product Components ........................................................................................................................................ 8 Data Modeling .............................................................................................................................................. 8 Data Processing ............................................................................................................................................ 9 Data Integrity and Error Recovery ................................................................................................................. 9 Performance ................................................................................................................................................. 9 Disk Space ....................................................................................................................................................
    [Show full text]
  • Database Meets AI: a Survey
    1 Database Meets AI: A Survey Xuanhe Zhou, Chengliang Chai, Guoliang Li, Ji Sun Abstract—Database and Artificial Intelligence (AI) can benefit from each other. On one hand, AI can make database more intelligent (AI4DB). For example, traditional empirical database optimization techniques (e.g., cost estimation, join order selection, knob tuning, index and view selection) cannot meet the high-performance requirement for large-scale database instances, various applications and diversified users, especially on the cloud. Fortunately, learning-based techniques can alleviate this problem. On the other hand, database techniques can optimize AI models (DB4AI). For example, AI is hard to deploy in real applications, because it requires developers to write complex codes and train complicated models. Database techniques can be used to reduce the complexity of using AI models, accelerate AI algorithms and provide AI capability inside databases. Thus both DB4AI and AI4DB have been extensively studied recently. In this paper, we review existing studies on AI4DB and DB4AI. For AI4DB, we review the techniques on learning-based configuration tuning, optimizer, index/view advisor, and security. For DB4AI, we review AI-oriented declarative language, AI-oriented data governance, training acceleration, and inference acceleration. Finally, we provide research challenges and future directions. Index Terms—Database, Artificial Intelligence, DB4AI, AI4DB F 1 INTRODUCTION Artificial intelligence (AI) and database (DB) have been Learning-based Database Configuration. (1) Knob tuning. extensively studied over the last five decades. First, database Databases have hundreds of knobs and it requires DBAs systems have been widely used in many applications, be- to tune the knobs so as to adapt to different scenarios.
    [Show full text]