From Data Mining to Knowledge Discovery in Databases

Total Page:16

File Type:pdf, Size:1020Kb

From Data Mining to Knowledge Discovery in Databases AI Magazine Volume 17 Number 3 (1996) (© AAAI) Articles From Data Mining to Knowledge Discovery in Databases Usama Fayyad, Gregory Piatetsky-Shapiro, and Padhraic Smyth ■ Data mining and knowledge discovery in This article begins by discussing the histori- databases have been attracting a significant cal context of KDD and data mining and their amount of research, industry, and media atten- intersection with other related fields. A brief tion of late. What is all the excitement about? summary of recent KDD real-world applica- This article provides an overview of this emerging tions is provided. Definitions of KDD and da- field, clarifying how data mining and knowledge ta mining are provided, and the general mul- discovery in databases are related both to each other and to related fields, such as machine tistep KDD process is outlined. This multistep learning, statistics, and databases. The article process has the application of data-mining al- mentions particular real-world applications, gorithms as one particular step in the process. specific data-mining techniques, challenges in- The data-mining step is discussed in more de- volved in real-world applications of knowledge tail in the context of specific data-mining al- discovery, and current and future research direc- gorithms and their application. Real-world tions in the field. practical application issues are also outlined. Finally, the article enumerates challenges for future research and development and in par- cross a wide variety of fields, data are ticular discusses potential opportunities for AI being collected and accumulated at a technology in KDD systems. Adramatic pace. There is an urgent need for a new generation of computational theo- ries and tools to assist humans in extracting useful information (knowledge) from the Why Do We Need KDD? rapidly growing volumes of digital data. The traditional method of turning data into These theories and tools are the subject of the knowledge relies on manual analysis and in- emerging field of knowledge discovery in terpretation. For example, in the health-care databases (KDD). industry, it is common for specialists to peri- At an abstract level, the KDD field is con- odically analyze current trends and changes cerned with the development of methods and in health-care data, say, on a quarterly basis. techniques for making sense of data. The basic The specialists then provide a report detailing problem addressed by the KDD process is one the analysis to the sponsoring health-care or- of mapping low-level data (which are typically ganization; this report becomes the basis for too voluminous to understand and digest easi- future decision making and planning for ly) into other forms that might be more com- health-care management. In a totally differ- pact (for example, a short report), more ab- ent type of application, planetary geologists stract (for example, a descriptive sift through remotely sensed images of plan- approximation or model of the process that ets and asteroids, carefully locating and cata- generated the data), or more useful (for exam- loging such geologic objects of interest as im- ple, a predictive model for estimating the val- pact craters. Be it science, marketing, finance, ue of future cases). At the core of the process is health care, retail, or any other field, the clas- the application of specific data-mining meth- sical approach to data analysis relies funda- ods for pattern discovery and extraction.1 mentally on one or more analysts becoming Copyright © 1996, American Association for Artificial Intelligence. All rights reserved. 0738-4602-1996 / $2.00 FALL 1996 37 Articles intimately familiar with the data and serving areas is astronomy. Here, a notable success as an interface between the data and the users was achieved by SKICAT, a system used by as- and products. tronomers to perform image analysis, For these (and many other) applications, classification, and cataloging of sky objects this form of manual probing of a data set is from sky-survey images (Fayyad, Djorgovski, slow, expensive, and highly subjective. In and Weir 1996). In its first application, the fact, as data volumes grow dramatically, this system was used to process the 3 terabytes type of manual data analysis is becoming (1012 bytes) of image data resulting from the completely impractical in many domains. Second Palomar Observatory Sky Survey, Databases are increasing in size in two ways: where it is estimated that on the order of 109 (1) the number N of records or objects in the sky objects are detectable. SKICAT can outper- database and (2) the number d of fields or at- form humans and traditional computational tributes to an object. Databases containing on techniques in classifying faint sky objects. See the order of N = 109 objects are becoming in- Fayyad, Haussler, and Stolorz (1996) for a sur- creasingly common, for example, in the as- vey of scientific applications. tronomical sciences. Similarly, the number of In business, main KDD application areas There is an fields d can easily be on the order of 102 or includes marketing, finance (especially in- even 103, for example, in medical diagnostic vestment), fraud detection, manufacturing, urgent need applications. Who could be expected to di- telecommunications, and Internet agents. for a new gest millions of records, each having tens or Marketing: In marketing, the primary ap- hundreds of fields? We believe that this job is plication is database marketing systems, generation of certainly not one for humans; hence, analysis which analyze customer databases to identify computation- work needs to be automated, at least partially. different customer groups and forecast their al theories The need to scale up human analysis capa- behavior. Business Week (Berry 1994) estimat- bilities to handling the large number of bytes ed that over half of all retailers are using or and tools to that we can collect is both economic and sci- planning to use database marketing, and assist entific. Businesses use data to gain competi- those who do use it have good results; for ex- tive advantage, increase efficiency, and pro- ample, American Express reports a 10- to 15- humans in vide more valuable services to customers. percent increase in credit-card use. Another extracting Data we capture about our environment are notable marketing application is market-bas- the basic evidence we use to build theories ket analysis (Agrawal et al. 1996) systems, useful and models of the universe we live in. Be- which find patterns such as, “If customer information cause computers have enabled humans to bought X, he/she is also likely to buy Y and (knowledge) gather more data than we can digest, it is on- Z.” Such patterns are valuable to retailers. ly natural to turn to computational tech- Investment: Numerous companies use da- from the niques to help us unearth meaningful pat- ta mining for investment, but most do not rapidly terns and structures from the massive describe their systems. One exception is LBS volumes of data. Hence, KDD is an attempt to Capital Management. Its system uses expert growing address a problem that the digital informa- systems, neural nets, and genetic algorithms volumes of tion era made a fact of life for all of us: data to manage portfolios totaling $600 million; overload. since its start in 1993, the system has outper- digital formed the broad stock market (Hall, Mani, data. Data Mining and Knowledge and Barr 1996). Discovery in the Real World Fraud detection: HNC Falcon and Nestor PRISM systems are used for monitoring credit- A large degree of the current interest in KDD card fraud, watching over millions of ac- is the result of the media interest surrounding counts. The FAIS system (Senator et al. 1995), successful KDD applications, for example, the from the U.S. Treasury Financial Crimes En- focus articles within the last two years in forcement Network, is used to identify finan- Business Week, Newsweek, Byte, PC Week, and cial transactions that might indicate money- other large-circulation periodicals. Unfortu- laundering activity. nately, it is not always easy to separate fact Manufacturing: The CASSIOPEE trou- from media hype. Nonetheless, several well- bleshooting system, developed as part of a documented examples of successful systems joint venture between General Electric and can rightly be referred to as KDD applications SNECMA, was applied by three major Euro- and have been deployed in operational use pean airlines to diagnose and predict prob- on large-scale real-world problems in science lems for the Boeing 737. To derive families of and in business. faults, clustering methods are used. CASSIOPEE In science, one of the primary application received the European first prize for innova- 38 AI MAGAZINE Articles tive applications (Manago and Auriol 1996). Data Mining and KDD Telecommunications: The telecommuni- cations alarm-sequence analyzer (TASA) was Historically, the notion of finding useful pat- built in cooperation with a manufacturer of terns in data has been given a variety of telecommunications equipment and three names, including data mining, knowledge ex- telephone networks (Mannila, Toivonen, and traction, information discovery, information Verkamo 1995). The system uses a novel harvesting, data archaeology, and data pattern processing. The term data mining has mostly framework for locating frequently occurring been used by statisticians, data analysts, and alarm episodes from the alarm stream and the management information systems (MIS) presenting them as rules. Large sets of discov- communities. It has also gained popularity in ered rules can be explored with flexible infor- the database field. The phrase knowledge dis- mation-retrieval tools supporting interactivity covery in databases was coined at the first KDD and iteration. In this way, TASA offers pruning, workshop in 1989 (Piatetsky-Shapiro 1991) to grouping, and ordering tools to refine the re- emphasize that knowledge is the end product sults of a basic brute-force search for rules.
Recommended publications
  • An Overview of the 50 Most Common Web Scraping Tools
    AN OVERVIEW OF THE 50 MOST COMMON WEB SCRAPING TOOLS WEB SCRAPING IS THE PROCESS OF USING BOTS TO EXTRACT CONTENT AND DATA FROM A WEBSITE. UNLIKE SCREEN SCRAPING, WHICH ONLY COPIES PIXELS DISPLAYED ON SCREEN, WEB SCRAPING EXTRACTS UNDERLYING CODE — AND WITH IT, STORED DATA — AND OUTPUTS THAT INFORMATION INTO A DESIGNATED FILE FORMAT. While legitimate uses cases exist for data harvesting, illegal purposes exist as well, including undercutting prices and theft of copyrighted content. Understanding web scraping bots starts with understanding the diverse and assorted array of web scraping tools and existing platforms. Following is a high-level overview of the 50 most common web scraping tools and platforms currently available. PAGE 1 50 OF THE MOST COMMON WEB SCRAPING TOOLS NAME DESCRIPTION 1 Apache Nutch Apache Nutch is an extensible and scalable open-source web crawler software project. A-Parser is a multithreaded parser of search engines, site assessment services, keywords 2 A-Parser and content. 3 Apify Apify is a Node.js library similar to Scrapy and can be used for scraping libraries in JavaScript. Artoo.js provides script that can be run from your browser’s bookmark bar to scrape a website 4 Artoo.js and return the data in JSON format. Blockspring lets users build visualizations from the most innovative blocks developed 5 Blockspring by engineers within your organization. BotScraper is a tool for advanced web scraping and data extraction services that helps 6 BotScraper organizations from small and medium-sized businesses. Cheerio is a library that parses HTML and XML documents and allows use of jQuery syntax while 7 Cheerio working with the downloaded data.
    [Show full text]
  • A Study on Social Network Analysis Through Data Mining Techniques – a Detailed Survey
    ISSN (Online) 2278-1021 ISSN (Print) 2319-5940 IJARCCE International Journal of Advanced Research in Computer and Communication Engineering ICRITCSA M S Ramaiah Institute of Technology, Bangalore Vol. 5, Special Issue 2, October 2016 A Study on Social Network Analysis through Data Mining Techniques – A detailed Survey Annie Syrien1, M. Hanumanthappa2 Research Scholar, Department of Computer Science and Applications, Bangalore University, India1 Professor, Department of Computer Science and Applications, Bangalore University, India2 Abstract: The paper aims to have a detailed study on data collection, data preprocessing and various methods used in developing a useful algorithms or methodologies on social network analysis in social media. The recent trends and advancements in the big data have led many researchers to focus on social media. Web enabled devices is an another important reason for this advancement, electronic media such as tablets, mobile phones, desktops, laptops and notepads enable the users to actively participate in different social networking systems. Many research has also significantly shows the advantages and challenges that social media has posed to the research world. The principal objective of this paper is to provide an overview of social media research carried out in recent years. Keywords: data mining, social media, big data. I. INTRODUCTION Data mining is an important technique in social media in scrutiny and analysis. In any industry data mining recent years as it is used to help in extraction of lot of technique based reports become the base line for making useful information, whereas this information turns to be an decisions and formulating the decisions. This report important asset to both academia and industries.
    [Show full text]
  • Data and Computer Communications (Eighth Edition)
    DATA AND COMPUTER COMMUNICATIONS Eighth Edition William Stallings Upper Saddle River, New Jersey 07458 Library of Congress Cataloging-in-Publication Data on File Vice President and Editorial Director, ECS: Art Editor: Gregory Dulles Marcia J. Horton Director, Image Resource Center: Melinda Reo Executive Editor: Tracy Dunkelberger Manager, Rights and Permissions: Zina Arabia Assistant Editor: Carole Snyder Manager,Visual Research: Beth Brenzel Editorial Assistant: Christianna Lee Manager, Cover Visual Research and Permissions: Executive Managing Editor: Vince O’Brien Karen Sanatar Managing Editor: Camille Trentacoste Manufacturing Manager, ESM: Alexis Heydt-Long Production Editor: Rose Kernan Manufacturing Buyer: Lisa McDowell Director of Creative Services: Paul Belfanti Executive Marketing Manager: Robin O’Brien Creative Director: Juan Lopez Marketing Assistant: Mack Patterson Cover Designer: Bruce Kenselaar Managing Editor,AV Management and Production: Patricia Burns ©2007 Pearson Education, Inc. Pearson Prentice Hall Pearson Education, Inc. Upper Saddle River, NJ 07458 All rights reserved. No part of this book may be reproduced in any form or by any means, without permission in writing from the publisher. Pearson Prentice Hall™ is a trademark of Pearson Education, Inc. All other tradmarks or product names are the property of their respective owners. The author and publisher of this book have used their best efforts in preparing this book.These efforts include the development, research, and testing of the theories and programs to determine their effectiveness.The author and publisher make no warranty of any kind, expressed or implied, with regard to these programs or the documentation contained in this book.The author and publisher shall not be liable in any event for incidental or consequential damages in connection with, or arising out of, the furnishing, performance, or use of these programs.
    [Show full text]
  • Geoseq2seq: Information Geometric Sequence-To-Sequence Networks
    GEOSEQ2SEQ:INFORMATION GEOMETRIC SEQUENCE-TO-SEQUENCE NETWORKS Alessandro Bay Biswa Sengupta Cortexica Vision Systems Ltd. Noah’s Ark Lab (Huawei Technologies UK) London, UK Imperial College London, London, UK [email protected] ABSTRACT The Fisher information metric is an important foundation of information geometry, wherein it allows us to approximate the local geometry of a probability distribu- tion. Recurrent neural networks such as the Sequence-to-Sequence (Seq2Seq) networks that have lately been used to yield state-of-the-art performance on speech translation or image captioning have so far ignored the geometry of the latent embedding, that they iteratively learn. We propose the information geometric Seq2Seq (GeoSeq2Seq) network which abridges the gap between deep recurrent neural networks and information geometry. Specifically, the latent embedding offered by a recurrent network is encoded as a Fisher kernel of a parametric Gaus- sian Mixture Model, a formalism common in computer vision. We utilise such a network to predict the shortest routes between two nodes of a graph by learning the adjacency matrix using the GeoSeq2Seq formalism; our results show that for such a problem the probabilistic representation of the latent embedding supersedes the non-probabilistic embedding by 10-15%. 1 INTRODUCTION Information geometry situates itself in the intersection of probability theory and differential geometry, wherein it has found utility in understanding the geometry of a wide variety of probability models (Amari & Nagaoka, 2000). By virtue of Cencov’s characterisation theorem, the metric on such probability manifolds is described by a unique kernel known as the Fisher information metric. In statistics, the Fisher information is simply the variance of the score function.
    [Show full text]
  • Data Mining and Soft Computing
    © 2002 WIT Press, Ashurst Lodge, Southampton, SO40 7AA, UK. All rights reserved. Web: www.witpress.com Email [email protected] Paper from: Data Mining III, A Zanasi, CA Brebbia, NFF Ebecken & P Melli (Editors). ISBN 1-85312-925-9 Data mining and soft computing O. Ciftcioglu Faculty of Architecture, Deljl University of Technology Delft, The Netherlands Abstract The term data mining refers to information elicitation. On the other hand, soft computing deals with information processing. If these two key properties can be combined in a constructive way, then this formation can effectively be used for knowledge discovery in large databases. Referring to this synergetic combination, the basic merits of data mining and soft computing paradigms are pointed out and novel data mining implementation coupled to a soft computing approach for knowledge discovery is presented. Knowledge modeling by machine learning together with the computer experiments is described and the effectiveness of the machine learning approach employed is demonstrated. 1 Introduction Although the concept of data mining can be defined in several ways, it is clear enough to understand that it is related to information extraction especially in large databases. The methods of data mining are essentially borrowed from exact sciences among which statistics play the essential role. By contrast, soft computing is essentially used for information processing by employing methods, which are capable to deal with imprecision and uncertainty especially needed in ill-defined problem areas. In the former case, the outcomes are exact within the error bounds estimated. In the latter, the outcomes are approximate and in some cases they may be interpreted as outcomes from an intelligent behavior.
    [Show full text]
  • Data Mining with Cloud Computing: - an Overview
    International Journal of Advanced Research in Computer Engineering & Technology (IJARCET) Volume 5 Issue 1, January 2016 Data Mining with Cloud Computing: - An Overview Miss. Rohini A. Dhote Dr. S. P. Deshpande Department of Computer Science & Information Technology HOD at Computer Science Technology Department, HVPM, COET Amravati, Maharashtra, India. HVPM, COET Amravati, Maharashtra, India Abstract: Data mining is a process of extracting competitive environment for data analysis. The cloud potentially useful information from data, so as to makes it possible for you to access your information improve the quality of information service. The from anywhere at any time. The cloud removes the integration of data mining techniques with Cloud need for you to be in the same physical location as computing allows the users to extract useful the hardware that stores your data. The use of cloud information from a data warehouse that reduces the computing is gaining popularity due to its mobility, costs of infrastructure and storage. Data mining has attracted a great deal of attention in the information huge availability and low cost. industry and in society as a whole in recent Years Wide availability of huge amounts of data and the imminent 2. DATA MINING need for turning such data into useful information and Data Mining involves the use of sophisticated data knowledge Market analysis, fraud detection, and and tool to discover previously unknown, valid customer retention, production control and science patterns and Relationship in large data sets. These exploration. Data mining can be viewed as a result of tools can include statistical models, mathematical the natural evolution of information technology.
    [Show full text]
  • Challenges in Bioinformatics for Statistical Data Miners
    This article appeared in the October 2003 issue of the “Bulletin of the Swiss Statistical Society” (46; 10-17), and is reprinted by permission from its editor. Challenges in Bioinformatics for Statistical Data Miners Dr. Diego Kuonen Statoo Consulting, PSE-B, 1015 Lausanne 15, Switzerland [email protected] Abstract Starting with possible definitions of statistical data mining and bioinformatics, this article will give a general side-by-side view of both fields, emphasising on the statistical data mining part and its techniques, illustrate possible synergies and discuss how statistical data miners may collaborate in bioinformatics’ challenges in order to unlock the secrets of the cell. What is statistics and why is it needed? Statistics is the science of “learning from data”. It includes everything from planning for the collection of data and subsequent data management to end-of-the-line activities, such as drawing inferences from numerical facts called data and presentation of results. Statistics is concerned with one of the most basic of human needs: the need to find out more about the world and how it operates in face of variation and uncertainty. Because of the increasing use of statistics, it has become very important to understand and practise statistical thinking. Or, in the words of H. G. Wells: “Statistical thinking will one day be as necessary for efficient citizenship as the ability to read and write”. But, why is statistics needed? Knowledge is what we know. Information is the communication of knowledge. Data are known to be crude information and not knowledge by itself. The way leading from data to knowledge is as follows: from data to information (data become information when they become relevant to the decision problem); from information to facts (information becomes facts when the data can support it); and finally, from facts to knowledge (facts become knowledge when they are used in the successful completion of the decision process).
    [Show full text]
  • The Theoretical Framework of Data Mining & Its
    International Journal of Social Science & Interdisciplinary Research__________________________________ ISSN 2277- 3630 IJSSIR, Vol.2 (1), January (2013) Online available at indianresearchjournals.com THE THEORETICAL FRAMEWORK OF DATA MINING & ITS TECHNIQUES SAYYAD RASHEEDUDDIN Associate Professor – HOD CSE dept Pulla Reddy Engineering College Wargal, Medak(Dist), A.P. ABSTRACT Data mining is a process which finds useful patterns from large amount of data. The research in databases and information technology has given rise to an approach to store and manipulate this precious data for further decision making. Data mining is a process of extraction of useful information and patterns from huge data. It is also called as knowledge discovery process, knowledge mining from data, knowledge extraction or data pattern analysis. The current paper discusses the following Objectives: 1. To discuss briefly about the concept of Data Mining 2. To explain about the Data Mining Algorithms & Techniques 3. To know about the Data Mining Applications Keywords: Concept of Data mining, Data mining Techniques, Data mining algorithms, Data mining applications. Introduction to Data Mining The development of Information Technology has generated large amount of databases and huge data in various areas. The research in databases and information technology has given rise to an approach to store and manipulate this precious data for further decision making. Data mining is a process of extraction of useful information and patterns from huge data. It is also called as knowledge discovery process, knowledge mining from data, knowledge extraction or data /pattern analysis. The current paper discuss about the following Objectives: 1. To discuss briefly about the concept of Data Mining 2.
    [Show full text]
  • Reinforcement Learning for Data Mining Based Evolution Algorithm
    Reinforcement Learning for Data Mining Based Evolution Algorithm for TSK-type Neural Fuzzy Systems Design Sheng-Fuu Lin, Yi-Chang Cheng, Jyun-Wei Chang, and Yung-Chi Hsu Department of Electrical and Control Engineering, National Chiao Tung University Hsinchu, Taiwan, R.O.C. ABSTRACT membership functions of a fuzzy controller, with the fuzzy rule This paper proposed a TSK-type neural fuzzy system set assigned in advance. Juang [5] proposed the CQGAF to (TNFS) with reinforcement learning for data mining based simultaneously design the number of fuzzy rules and free evolution algorithm (R-DMEA). In R-DMEA, the group-based parameters in a fuzzy system. In [6], the group-based symbiotic symbiotic evolution (GSE) is adopted that each group represents evolution (GSE) was proposed to solve the problem that all the the set of single fuzzy rule. The R-DMEA contains both fuzzy rules were encoded into one chromosome in the structure learning and parameter learning. In the structure traditional genetic algorithm. In the GSE, each chromosome learning, the proposed R-DMEA uses the two-step self-adaptive represents a single rule of fuzzy system. The GSE uses algorithm to decide the suitable number of rules in a TNFS. In group-based population to evaluate the fuzzy rule locally. parameter learning, the R-DMEA uses the mining-based Although GSE has a better performance when compared with selection strategy to decide the selected groups in GSE by the tradition symbiotic evolution, the combination of groups is data mining method called frequent pattern growth. Illustrative fixed. example is conducted to show the performance and applicability Although the above evolution learning algorithms [4]-[6] of R-DMEA.
    [Show full text]
  • Research Data Management Best Practices
    Research Data Management Best Practices Introduction ............................................................................................................................................................................ 2 Planning & Data Management Plans ...................................................................................................................................... 3 Naming and Organizing Your Files .......................................................................................................................................... 6 Choosing File Formats ............................................................................................................................................................. 9 Working with Tabular Data ................................................................................................................................................... 10 Describing Your Data: Data Dictionaries ............................................................................................................................... 12 Describing Your Project: Citation Metadata ......................................................................................................................... 15 Preparing for Storage and Preservation ............................................................................................................................... 17 Choosing a Repository .........................................................................................................................................................
    [Show full text]
  • Data Management, Analysis Tools, and Analysis Mechanics
    Chapter 2 Data Management, Analysis Tools, and Analysis Mechanics This chapter explores different tools and techniques for handling data for research purposes. This chapter assumes that a research problem statement has been formulated, research hypotheses have been stated, data collection planning has been conducted, and data have been collected from various sources (see Volume I for information and details on these phases of research). This chapter discusses how to combine and manage data streams, and how to use data management tools to produce analytical results that are error free and reproducible, once useful data have been obtained to accomplish the overall research goals and objectives. Purpose of Data Management Proper data handling and management is crucial to the success and reproducibility of a statistical analysis. Selection of the appropriate tools and efficient use of these tools can save the researcher numerous hours, and allow other researchers to leverage the products of their work. In addition, as the size of databases in transportation continue to grow, it is becoming increasingly important to invest resources into the management of these data. There are a number of ancillary steps that need to be performed both before and after statistical analysis of data. For example, a database composed of different data streams needs to be matched and integrated into a single database for analysis. In addition, in some cases data must be transformed into the preferred electronic format for a variety of statistical packages. Sometimes, data obtained from “the field” must be cleaned and debugged for input and measurement errors, and reformatted. The following sections discuss considerations for developing an overall data collection, handling, and management plan, and tools necessary for successful implementation of that plan.
    [Show full text]
  • Computer Files & Data Storage
    STORAGE & FILE CONCEPTS, UTILITIES (Pages 6, 150-158 - Discovering Computers & Microsoft Office 2010) I. Computer files – data, information or instructions residing on secondary storage are stored in the form of a file. A. Software files are also called program files. Program files (instructions) are created by a computer programmer and generally cannot be modified by a user. It’s important that we not move or delete program files because your computer requires them to perform operations. Program files are also referred to as “executables”. 1. You can identify a program file by its extension:“.EXE”, “.COM”, “.BAT”, “.DLL”, “.SYS”, or “.INI” (there are others) or a distinct program icon. B. Data files - when you select a “save” option while using an application program, you are in essence creating a data file. Users create data files. 1. File naming conventions refer to the guidelines followed while assigning file names and will vary with the operating system and application in use (see figure 4-1). File names in Windows 7 may be up to 255 characters, you're not allowed to use reserved characters or certain reserved words. File extensions are used to identify the application that was used to create the file and format data in a manner recognized by the source application used to create it. FALL 2012 1 II. Selecting secondary storage media A. There are three type of technologies for storage devices: magnetic, optical, & solid state, there are advantages & disadvantages between them. When selecting a secondary storage device, certain factors should be considered: 1. Capacity - the capacity of computer storage is expressed in bytes.
    [Show full text]