SQL Server Analysis Services (SSAS)?

Total Page:16

File Type:pdf, Size:1020Kb

SQL Server Analysis Services (SSAS)? What is SQL Server Analysis Services (SSAS)? SQL Server Analysis Services (SSAS) is the On-Line Analytical Processing (OLAP) Component of SQL Server. SSAS allows you to build multidimensional structures called Cubes to pre-calculate and store complex aggregations, and also to build mining models to perform data analysis to identify valuable information like trends, patterns, relationships etc. within the data using Data Mining capabilities of SSAS, which otherwise could be really difficult to determine without Data Mining capabilities. SSAS comes bundled with SQL Server and you get to choose whether or not to install this component as part of the SQL Server Installation. What is OLAP? How is it different from OLTP? OLAP stands for On-Line Analytical Processing. It is a capability or a set of tools which enables the end users to easily and effectively access the data warehouse data using a wide range of tools like Microsoft Excel, Reporting Services, and many other 3rd party business intelligence tools. OLAP is used for analysis purposes to support day-to-day business decisions and is characterized by less frequent data updates and contains historical data. Whereas, OLTP (On- Line Transactional Processing) is used to support day-to-day business operations and is characterized by frequent data updates and contains the most recent data along with limited historical data based on the retention policy driven by business needs. What is a Data Source? What are the different data sources supported by SSAS? A Data Source contains the connection information used by SSAS to connect to the underlying database to load the data into SSAS during processing. A Data Source primarily contains the following information (apart from various other properties like Query timeout, Isolation etc.): o Provider o Server Name o Database Name o Impersonation Information SSAS Supports both .Net and OLE DB Providers. Following are some of the major sources supported by SSAS: SQL Server, MS Access, Oracle, Teradata, IBM DB2, and other relational databases with the appropriate OLE DB provider. What is Impersonation? What are the different impersonation options available in SSAS? Impersonation allows SSAS to assume the identity/security context of the client application which is used by SSAS to perform the server side data operations like data access, processing etc. As part of impersonation, the following options are available in SSAS: o Use a specific Windows user name and password: This option lets you to specify Windows account credentials which will be used by SSAS to perform operations like source data access, processing etc. o Use the service account: When this option is selected, SSAS uses the credentials of the service account under which the Analysis Services service is configured/running for source data access, processing etc. o Use the credentials of the current user: When this option is set, SSAS uses the credentials of the current user for performing operations like DMX Open Queries, Local cubes etc. This option cannot be used for performing server side operations like source data access, processing etc. o Inherit: This option let's the SSAS server decide which impersonation mode is suitable for each type of operation. When this option is set, by default SSAS will use the service account for operations like processing and the credentials of the current user for operations like Local cubes, querying the data mining models, etc... What is a Data Source View? A Data Source View (DSV) is a logical view of the underlying database schema and offers a layer of abstraction for the underlying database schema. This layer acts as a source for SSAS and captures the schema related information from the underlying database. The schematic information present in DSV includes the following: o Underlying database Table(s)/View(s) metadata o Primary Key & Foreign Key relationships between the underlying database Table(s) o Additional columns in the form of Named Calculations o Complex logic on the underlying Table(s)/View(s) in the form of Named Queries SSAS can only see the schematic information present in the DSV and it cannot see the schematic information from the underlying database. What is a Named Calculation? In what scenarios do you use it? A Named Calculation is a new column added to a Table in DSV and is based on an expression. This capability allows you to add an extra column into your DSV which is based on one or more columns from underlying data source Table(s)/View(s) combined using an expression without requiring the addition of a physical column in the underlying database Table(s)/View(s). The expression used in the Named Calculation should conform to the underlying data source dialect. For example, if the underlying data source is SQL Server, then it should conform to T- SQL, If is it Oracle, then it should conform to PL/SQL, etc... Named Calculations can be used in many scenarios, following are some of the common scenarios: o For creating Derived Columns. Say you have First Name and Last Name in the underlying data source Table/View and you want to get the Full Name as "First Name + space + Last Name". Such things can be added as a Named Calculation. o For performing Lookup Operations. Say you have an Employee table which has AddressID and an Address table in which AddressID is the Primary Key. Now, to get the address information (say Country) into the Employee table in DSV, a Named Calculation can be added to the Employee table with the following expression: ( SELECT Country FROM Address WHERE AddressID = Employee.AddressID ) What is a Named Query? In what scenarios do you use it? A Named Query is a SQL query/expression in your DSV which acts as a Table. It is used to combine data from one or more Table(s)/View(s) from the underlying data source without requiring any schematic changes to the underlying data source Table(s)/View(s). The SQL Query used in the Named Query should conform to the underlying data source dialect. For example, if the underlying data source is SQL Server, then it should conform to T-SQL, If is it Oracle, then it should conform to PL/SQL, etc... Named Queries are used in various scenarios, following are some of the common scenarios: o Combining data from multiple Tables/Views from the underlying data source by using either simple or complex join operations. o Adding filter conditions for filtering out unwanted data or selecting the required data (limiting the data). Pretty much everything that can be done using a Named Calculation can also be done using a Named Query. What are the pros and cons of using Tables and Named Queries in DSV? Following are some of the pros and cons of using Tables and Named Queries in DSV. Tables in the below comparison refers to the Table in DSV which references a single Table or a View in the underlying source database. Tables Named Queries Named Calculations cannot be added to Named Calculations can be added to Tables in DSV. Named Queries in DSV. Keys and Relationships are automatically set (by Keys and Relationships have to be set the wizard) based on the Keys and Relationships in explicitly in the DSV. the underlying database schema. More than one Table/View from the Only one Table/View from the underlying data underlying data source can be referenced source can be referenced in DSV. using a SQL Expression in the DSV. Filter/limiting conditions can be applied as Any filter/limiting conditions cannot be applied on part of the SQL expression in the Named a table in DSV. Query in the DSV. Although Named Calculations and Named Queries can be used to extend the functionality of SSAS to address the evolving business needs, it is always a good practice to first build a good Dimensional Model in the beginning of a Data Warehousing/SSAS project. What is the purpose of setting Logical Keys and Relationships in DSV? Many of the user interfaces/designers/wizards in BIDS which are part of a SSAS project depend on the Primary Key and Relationships between Fact and Dimension tables. Hence it is important to define the Primary Key and Relationships in DSV. By default, the Data Source View Wizard detects the Physical Primary Keys and Relationships between the tables in the underlying source database and applies the same Keys and Relationships in DSV layer. However, Logical Keys and Relationships need to be defined explicitly in the following scenarios: o If the DSV table is referring to an underlying database View. o If the DSV table is created as a Named Query. o If any additional relationships need to be defined in the DSV layer apart from the ones that are physically defined in the underlying source database. Is it possible to combine data from multiple data sources in SSAS? If yes, how do you accomplish it? SSAS allows combining data from multiple underlying data sources into a single DSV. To be able to add Table(s)/View(s) from multiple data sources, first you need to create a DSV using your first source and this source acts as the primary data source. Now after the initial DSV is created, you can add one or more data sources into DSV which will act as secondary data sources and you can choose additional Table(s)/View(s) from the secondary data sources which you want to include in your DSV. The key thing while combining data from multiple data sources is that the Primary Data Source must support OPENROWSET queries. Hence in most cases, SQL Server is used as the Primary Data Source.
Recommended publications
  • MCSA SQL Server 2016
    Microsoft MCSA: SQL Server Pre-course Reading R-481-01 Version 1.0 https://firebrand.training Pre-course Reading So you are taking a course about SQL from Firebrand. As you may already know there are three different tracks for studying SQL 2016 offered at Firebrand: . MCSASQLDD - a track for database developers . MCSASQLDA - a track for database administrators . MCSASQLBID - a track for business intelligence developers The track for Database Development will entail writing code in Transact SQL ranging in complexity from a simple Select statement to retrieve the data stored in a table or tables to creating more complex programming objects like Stored Procedures, Triggers, Functions and the like. The Database Administration track covers items like how to create logins and users, do backup and restore as well as more complex items like index management and creating an Azure database out in the cloud. The Business Intelligence track includes developing ways to move data into a data warehouse and then using that data warehouse as a data source for models that your business users can run reports against. In this document there will be sections for each of the tracks including some links to some helpful websites that can help you prepare for your upcoming Firebrand course. Before we break down into the different subsections for the different tracks, let’s start off with a little history about Microsoft SQL Server. Many years ago Microsoft bought a product that was called Sybase and it became Microsoft SQL Server. It was a multiuser relational database which means it provided a way to hold data in tables that were related to each other and allowed multiple users to access that data simultaneously.
    [Show full text]
  • Oracle White Paper June 2009
    An Oracle White Paper June 2009 Oracle Data Mining 11g Competing on In-Database Analytics Oracle White Paper— Oracle Data Mining 11g: Competing on In-Database Analytics Disclaimer The following is intended to outline our general product direction. It is intended for information purposes only, and may not be incorporated into any contract. It is not a commitment to deliver any material, code, or functionality, and should not be relied upon in making purchasing decisions. The development, release, and timing of any features or functionality described for Oracle’s products remains at the sole discretion of Oracle. Oracle White Paper— Oracle Data Mining 11g: Competing on In-Database Analytics Executive Overview............................................................................. 1 In-Database Data Mining .................................................................... 1 Key Benefits .................................................................................... 3 Introduction ......................................................................................... 4 Oracle Data Mining ......................................................................... 4 Data Mining Deep Dive ....................................................................... 6 Oracle Data Mining for Data Analysts ............................................... 15 Oracle Data Mining for Applications Developers............................... 16 Competing on In-Database Analytics................................................ 18 Beyond a Tool; Enabling
    [Show full text]
  • Data Mining with Microsoft SQL Server 2008 / Jamie Maclennan, Bogdan Crivat, Zhaohui Tang
    Maclennan ffirs.tex V3 - 10/04/2008 3:27am Page ii Maclennan ffirs.tex V3 - 10/04/2008 3:27am Page i Data Mining with Microsoft SQL Server2008 Maclennan ffirs.tex V3 - 10/04/2008 3:27am Page ii Maclennan ffirs.tex V3 - 10/04/2008 3:27am Page iii Data Mining with Microsoft SQL Server2008 Jamie MacLennan ZhaoHui Tang Bogdan Crivat Wiley Publishing, Inc. Maclennan ffirs.tex V3 - 10/04/2008 3:27am Page iv Data Mining with MicrosoftSQL Server2008 Published by Wiley Publishing, Inc. 10475 Crosspoint Boulevard Indianapolis, IN 46256 www.wiley.com Copyright 2009 by Wiley Publishing, Inc., Indianapolis, Indiana Published by Wiley Publishing, Inc., Indianapolis, Indiana Published simultaneously in Canada ISBN: 978-0-470-27774-4 Manufactured in the United States of America 10987654321 No part of this publication may be reproduced, stored in a retrieval system or transmitted in any form or by any means, electronic, mechanical, photocopying, recording, scanning or otherwise, except as permitted under Sections 107 or 108 of the 1976 United States Copyright Act, without either the prior written permission of the Publisher, or authorization through payment of the appropriate per-copy fee to the Copyright Clearance Center, 222 Rosewood Drive, Danvers, MA 01923, (978) 750-8400, fax (978) 646-8600. Requests to the Publisher for permission should be addressed to the Legal Department, Wiley Publishing, Inc., 10475 Crosspoint Blvd., Indianapolis, IN 46256, (317) 572-3447, fax (317) 572-4355, or online at www.wiley.com/go/permissions. Limit of Liability/Disclaimer of Warranty: The publisher and the author make no representations or warranties with respect to the accuracy or completeness of the contents of this work and specifically disclaim all warranties, including without limitation warranties of fitness for a particular purpose.
    [Show full text]
  • Modelo Tese MGI / MEGI
    MODELO ZeEN Uma abordagem minimalista para o desenho de data warehouses Miguel Nuno da Silva Gomes Rodrigues Gago Dissertação apresentada como requisito parcial para obtenção do grau de Mestre em Estatística e Gestão de Informação Dissertation presented as partial requirement for obtaining the Master’s degree in Statistics and Information Management ii TÍTULOTÍTULO Subtítulo Subtítulo Nome completo do Candidato Nome completo do Candidato Dissertação / Trabalho de Projeto / Relatório de Dissertação / Trabalho de Projeto / Relatório de Estágio apresentada(o)Estágio apresentada como requisito(o) como parcial requisito para obtenção parcial do para grauobtenção de Mestre do emgrau Gestão de Mestre de Informação em Estatística e Gestão de Informação Instituto Superior de Estatística e Gestão de Informação Universidade Nova de Lisboa MODELO ZeEN Uma abordagem minimalista para o desenho de data warehouses por Miguel Nuno da Silva Gomes Rodrigues Gago Dissertação apresentada como requisito parcial para a obtenção do grau de Mestre em Estatística e Gestão de Informação, Especialização em Gestão dos Sistemas e Tecnologias de Informação Orientador: Prof. Dr. Miguel de Castro Neto Março 2013 iii Ao meu Pai, o Engenheiro Armando Rodrigues Gago, que me ensinou a procurar sempre mais além. iv Agradecimentos À minha Mãe Maria Ondina, À minha Mulher Luísa, pelo tempo que lhes subtraí e por acreditarem sempre em mim. Ao Prof. Dr. Miguel de Castro Neto, por me ter incutido confiança em desenvolver esta dissertação na área da Business Intelligence. v Il semble que la perfection soit atteinte, non quand il n'y a plus rien à ajouter mais quand il n'y a plus rien à retrancher.
    [Show full text]
  • SQL Server 2012 Tutorials – Analysis Services Data Mining
    SQL Server 2012 Tutorials: Analysis Services - Data Mining SQL Server 2012 Books Online Summary: Microsoft SQL Server Analysis Services makes it easy to create sophisticated data mining solutions. The step-by-step tutorials in the following list will help you learn how to get the most out of Analysis Services, so that you can perform advanced analysis to solve business problems that are beyond the reach of traditional business intelligence methods. Category: Step-by-Step Applies to: SQL Server 2012 Source: SQL Server Books Online (link to source content) E-book publication date: June 2012 Copyright © 2012 by Microsoft Corporation All rights reserved. No part of the contents of this book may be reproduced or transmitted in any form or by any means without the written permission of the publisher. Microsoft and the trademarks listed at http://www.microsoft.com/about/legal/en/us/IntellectualProperty/Trademarks/EN-US.aspx are trademarks of the Microsoft group of companies. All other marks are property of their respective owners. The example companies, organizations, products, domain names, email addresses, logos, people, places, and events depicted herein are fictitious. No association with any real company, organization, product, domain name, email address, logo, person, place, or event is intended or should be inferred. This book expresses the author’s views and opinions. The information contained in this book is provided without any express, statutory, or implied warranties. Neither the authors, Microsoft Corporation, nor its resellers, or distributors will be held liable for any damages caused or alleged to be caused either directly or indirectly by this book.
    [Show full text]
  • Building a Data Mining Model Using Data Warehouse and OLAP Cubes IST 734 SS Chung
    Cleveland State University Building a Data Mining Model using Data Warehouse and OLAP Cubes IST 734 SS Chung 14 Sunnie S Chung IST 734 Build a Data Mining Model using Data Warehouse and OLAP cubes Contents 1. Abstract: ................................................................................................................................................ 3 2. Introduction: .......................................................................................................................................... 3 3. Adventure works database: ................................................................................................................... 4 4. Getting familiar with Sql Server Analysis Services (SSAS) tools and various datamining algorithms 5 4.1. Microsoft Association Algorithm .................................................................................................. 6 4.2. Microsoft Clustering Algorithm .................................................................................................... 9 4.3. Microsoft Time Series Algorithm ................................................................................................ 12 4.4. Microsoft Decision Trees Algorithm .......................................................................................... 16 5. Star schema: ........................................................................................................................................ 19 5.1. Fact Tables and Dimension Tables .................................................................................................
    [Show full text]
  • Predictive Analysis in Microsoft SQL Server 2012 Gain Intuitive and Comprehensive Predictive Insight
    Predictive Analysis in Microsoft SQL Server 2012 Gain intuitive and comprehensive predictive insight automatically grouping similar Top Features Complete customers together. Test multiple data-mining models Inform decisions with intuitive and simultaneously with statistical scores comprehensive predictive insight Forecasting of error and accuracy and confirm available to all users. Predict sales and inventory amounts their stability with cross-validation. and learn how they are interrelated Rich and Innovative Algorithms to foresee bottlenecks and improve Build multiple, incompatible mining Benefit from many rich and performance. models within a single structure; innovative data-mining algorithms apply model analysis over filtered to support common business Data Exploration data; query against structure data to problems promptly and accurately. Analyze profitability across present complete information, all customers or compare customers enabled by enhanced mining Market Basket Analysis who prefer different brands of the structures. Discover which items tend to be same product to discover new Combine the best of both worlds by bought together to create opportunities. blending optimized near-term recommendations on-the-fly and to predictions (ARTXP) and stable long- determine how product placement Unsupervised Learning term predictions (ARIMA) with can directly contribute to your Identify previously unknown Better Time Series Support. bottom line. relationships between various elements of your business to better Discover the relationship between Churn Analysis inform your decisions. items that are frequently purchased Anticipate customers who may be together by using Shopping Basket considering canceling their service Website Analysis Analysis and generate interactive Understand how people use your forms for scoring new cases by and identify benefits that will keep them from leaving.
    [Show full text]
  • A Query Language for Analyzing Networks
    A Query Language for Analyzing Networks Anton Dries Siegfried Nijssen Luc De Raedt K.U.Leuven, Celestijnenlaan 200A, Leuven, Belgium {anton.dries,siegfried.nijssen,luc.deraedt}@cs.kuleuven.be ABSTRACT to analyze the network in order to discover new knowledge With more and more large networks becoming available, and use that knowledge to improve the network. mining and querying such networks are increasingly impor- Discovering new knowledge in databases (also known as tant tasks which are not being supported by database models KDD) typically involves a process in which multiple oper- and querying languages. This paper wants to alleviate this ations are repeatedly performed on the data, and as new situation by proposing a data model and a query language insights are gained, the data is being transformed and ex- for facilitating the analysis of networks. Key features in- tended. As one example consider a bibliographical network clude support for executing external tools on the networks, with authors, papers, and citations such as that gathered by flexible contexts on the network each resulting in a different Citeseer or Google Scholar. Important problems in such bib- graph, primitives for querying subgraphs (including paths) liographical networks include: entity resolution [3], which is and transforming graphs. The data model provides for a concerned with detecting which nodes (authors or papers) closure property, in which the output of every query can be refer to the same entity, and collective classification [18], stored in the database and used for further querying. where the task is to categorize papers according to their subject. This type of analysis typically requires one to call multiple tools and to perform a lot of transformations on Categories and Subject Descriptors the data.
    [Show full text]
  • Inductive Databases and Constraint-Based Data Mining Sašo Džeroski • Bart Goethals • Panþe Panov Editors
    Inductive Databases and Constraint-Based Data Mining Sašo Džeroski • Bart Goethals • Panþe Panov Editors Inductive Databases and Constraint-Based Data Mining 1 C Editors Sašo Džeroski Panče Panov Jožef Stefan Institute Jožef Stefan Institute Dept. of Knowledge Technologies Dept. of Knowledge Technologies Jamova cesta 39 Jamova cesta 39 SI-1000 Ljubljana SI-1000 Ljubljana Slovenia Slovenia [email protected] [email protected] Bart Goethals University of Antwerp Mathematics and Computer Science Dept. Middelheimlaan 1 B-2020 Antwerpen Belgium [email protected] ISBN 978-1-4419-7737-3 e-ISBN 978-1-4419-7738-0 DOI 10.1007/978-1-4419-7738-0 Springer New York Dordrecht Heidelberg London Library of Congress Control Number: 2010938297 © Springer Science+Business Media, LLC 2010 All rights reserved. This work may not be translated or copied in whole or in part without the written permission of the publisher (Springer Science+Business Media, LLC, 233 Spring Street, New York, NY 10013, USA), except for brief excerpts in connection with reviews or scholarly analysis. Use in connec- tion with any form of information storage and retrieval, electronic adaptation, computer software, or by similar or dissimilar methodology now known or hereafter developed is forbidden. The use in this publication of trade names, trademarks, service marks, and similar terms, even if they are not identified as such, is not to be taken as an expression of opinion as to whether or not they are subject to proprietary rights. Printed on acid-free paper Springer is part of Springer Science+Business Media (www.springer.com) Preface This book is about inductive databases and constraint-based data mining, emerging research topics lying at the intersection of data mining and database research.
    [Show full text]
  • SPARQL-ML: Knowledge Discovery for the Semantic Web University Of
    University of Zurich Department of Informatics SPARQL-ML: Diploma Thesis December 18, 2007 Knowledge Discovery for the Semantic Web Andre´ Locher of Bratsch VS, Switzerland Student-ID: 03-706-405 [email protected] Advisor: Christoph Kiefer Prof. Abraham Bernstein, PhD Department of Informatics University of Zurich http://www.ifi.uzh.ch/ddis Acknowledgements I would like to thank Christoph Kiefer and Prof. Abraham Bernstein for giving me the oppor- tunity to write this thesis and for their valuable input, which always made me strive for more. Additionally I want to give a big thank you to Claudia von Bastian for her moral support during the last six months and of course for proofreading. I would also like to use this opportunity to thank my parents for supporting my studies and allowing me to go all the way. Abstract Machine learning as well as data mining has been successfully applied to automatically or semi- automatically create Semantic Web data from plain data. Only little work has been done so far to explore the possibilities of machine learning to induce models from existing Semantic Web data. The interlinked structure of Semantic Web data allows to include relations between entities in addition to attributes of entities of propositional data mining techniques. It is, therefore, a perfect match for Statistical Relational Learning methods (SRL), which combine relational learning with statistics and probability theory. This thesis presents SPARQL-ML, a novel approach to perform data mining tasks for knowl- edge discovery in the Semantic Web. Our approach is based on SPARQL and allows the use of sta- tistical relational learning methods, such as Relational Probability Trees and Relational Bayesian Classifiers, as well as traditional propositional learning methods.
    [Show full text]
  • Mining Model Content
    © 2014 IJIRT | Volume 1 Issue 4 | ISSN : 2349-6002 MINING MODEL CONTENT Aastha Trehan, Ritika Grover, Prateek Puri Dronacharya College Of Engineering, Gurgaon Abstract- This paper highlights about the mining model content used in data Microsoft Generic Content Tree Viewer, provided in SQL Server Data mining. The mining model is complete after you have designed and processed a Tools (SSDT), and then switch to one of the custom viewers to see how the mining model using data from the underlying mining structure and information is interpreted and displayed graphically for each model contains mining model content. You can use this content to make predictions or analyze your data. This paper describes in-depth the structure of mining model type. You can also create queries against the mining model content by using content, nodes in the model content, mining model content by algorithm type any client that supports the MINING_MODEL_CONTENT schema rowset. and tools for viewing and querying mining model content. Index Terms- Mining Model Conent, Mining Model, Mining Content Nodes, II. STRUCTURE OF MINING MODEL CONTENT Tools The content of each model is presented as a series of nodes. A node is an object within a mining model that contains metadata and information about I. INTRODUCTION a portion of the model. Nodes are arranged in a hierarchy. The exact Mining model content includes metadata about the model, statistics about arrangement of nodes in the hierarchy, and the meaning of the hierarchy, the data, and patterns discovered by the mining algorithm. Depending on depends on the algorithm that you used. For example, if you create a the algorithm that was used, the model content may include regression decision trees model, the model can contain multiple trees, all connected to formulas, the definitions of rules and itemsets, or weights and other the model root; if you create a neural network model, the model may statistics.
    [Show full text]
  • Normalization Rules
    APPENDIX Normalization Rules Normalization is the process of removing data redundancy by implementing normalization rules. There are five degrees of normal forms, from the first normal form through the fifth normal form, as described in this appendix. First Normal Form The following are the characteristics of first normal form (1NF): • There must not be any repeating columns or groups of columns. An example of a repeating column is a customer table with Phone Number 1 and Phone Number 2 columns. Using “table (column, column)” notation, an example of a repeating group of columns is Order Table (Order ID, Order Date, Product ID, Price, Quantity, Product ID, Price, Quantity). Product ID, Price, and Quantity are the repeating group of columns. • Each table must have a primary key (PK) that uniquely identifies each row. The PK can be a composite, that is, can consist of several columns, for example, Order Table (Order ID, Order Date, Customer ID, Product ID, Product Name, Price, Quantity). In this notation, the underlined columns are the PKs; in this case, Order ID and Product ID are a composite PK. Second Normal Form The following are the characteristics of second normal form (2NF): • It must be in 1NF. • When each value in column 1 is associated with a value in column 2, we say that column 2 is dependant on column 1, for example, Customer (Customer ID, Customer Name). Customer Name is dependant on Customer ID, noted as Customer ID ➤ Customer Name. 505 506 APPENDIX ■ NORMALIZATION RULES • In 2NF, all non-PK columns must be dependent on the entire PK, not just on part of it, for example, Order Table (Order ID, Order Date, Product ID, Price, Quantity).
    [Show full text]