The Commercialization of Database Management Systems, 1969–1983

Total Page:16

File Type:pdf, Size:1020Kb

The Commercialization of Database Management Systems, 1969–1983 The Commercialization of Database Management Systems, 1969–1983 Thomas J. Bergin American University Thomas Haigh University of Wisconsin, Milwaukee Database management systems were the most important commercial software packages of the 1970s. The authors reconstruct their early history by examining the evolution of their capabilities and installed base. They also document early user experiences, including the sources from which potential users learned about these new tech- nologies, new roles such as the database administrator, and new concepts such as the data dictionary. One approach to examining the early days of characteristics and exploring the changing commercial database management systems dynamics of the DBMS marketplace into the (DBMSs) is to study the evolution of the lead- early 1980s. This includes a brief discussion ing DBMS packages of the 1970s. Doing so, of the commercial threat posed by relational we can reconstruct the technical issues and database systems to established products and management decisions facing information efforts made to add relational capabilities to technology managers as they investigated traditional products such as IDMS. Our focus the new world of database software in the is on database management packages for early 1970s. This includes not only the IBM machines, which dominated the main- DBMS as a productivity aid for application frame market at that time. We rely here on programmers but also the new jobs, such as contemporary market-analysis reports, partic- database administrator, created by the tech- ularly Datapro’s ‘‘Buyer’s Guide to Data Base nology and the managerial challenges of cre- Management Systems’’ starting in 1974.1 ating a shared and consistent database to be used across different departments. Indeed, Database beginnings during the 1970s, the DBMS existed both as Life in the data processing departments of a tangible technology (with considerable the 1960s revolved around applications strengths and pronounced weaknesses) and requiring the storage, updating, and retrieval as the symbol of a movement to raise the sta- of large volumes of data. The arrival of disk tus of computing within the managerial technology in the mid-1960s posed a huge world and establish the idea of data as a cor- challenge to application developers.3 Organi- porate resource. (See the ‘‘Data Management zations struggled to exploit the potential of Definitions’’ sidebar for a brief explanation random access data storage and interactive of related terminology.1) query. As historian Martin Campbell-Kelly noted, Corporate data processing budgets mush- finding good data on software adoptions and roomed in the 1960s, driven in large part by usage is difficult.2 Our research utilizes vari- the rapidly growing cost of maintaining sys- ous contemporary journal and newsletter tems and application software. Organizations articles, consulting reports, and advisory not wanting to continue to create and main- guides to document early DBMS user experi- tain large applications had two choices: hire a ences and the processes by which they eval- contractor to develop the software to the uated different packages. We also examine organization’s specifications or purchase a some of the major DBMS products of standard software package from an indepen- the 1970s, documenting their technical dent vendor (and in many cases, pay them 26 IEEE Annals of the History of Computing Published by the IEEE Computer Society 1058-6180/09/$26.00 c 2009 IEEE Data Management Definitions To frame our discussion on specific software pack- and centralize the data storage, the system’s principal ages, let’s review a pair of definitions from the period: intent is to perform such functions as information re- trieval, report generation, and inquiry for single DBMSs can be defined as those that are intended to applications.1 manage and maintain data in a non-redundant struc- ture for the purpose of being processed by multiple Data-management systems were also known as file- applications. A DBMS organizes data elements in management systems. Although we provide a short dis- some predefined structure and retains relationships cussion of file-management and report-generation soft- between different data elements within the database. ware, largely to clarify the continued success these A data management system, on the other hand, is in- products enjoyed throughout the 1970s, this article fo- tended primarily to permit access to and retrieval cuses on early mainframe database management soft- from pre-existing files (usually for a single applica- ware. We believe the history of data management tion). Although a data management system might systems is sufficiently important to warrant a separate provide the capability to minimize data redundancy investigation. to tailor the package to the buyer’s environ- its applications, and save money. Infosystems, ment). In the mid-1960s, however, the soft- one of the leading management-oriented ware services industry was in its infancy, data-processing publications, assured its read- and almost no commercial market existed ers in 1973 that database systems were like for standard software packages such as sort the aeronautical efforts of the Wright broth- utilities, report writers, and information- ers; although carefully planned early efforts retrieval packages. When vendors and user had ‘‘never developed much lift when ap- groups did create such software, it was plied to the practical realities of processing given away without additional charge. large files that had to be stored, indexed Two factors changed this state of affairs. and sorted with live data,’’ they were now On 7 April 1964, IBM announced the System/ poised to rise majestically into the air.9 The 360, a family of compatible computers and problem was that no one knew much about operating systems (OSs) that let users mi- them, even the organization that had grate to larger computers without throwing attempted to write generalized data access out their existing programs.4 With these, software.10 organizations relied heavily on standard Most organizations wanting to explore OSs provided by computer vendors. Then DBMSs needed training for their technical on 6 December 1968, IBM announced that staff and consultants to guide them. Initially, it was unbundling software products, systems there was not much help available, so con- engineering services, and training costs rather sulting firms and software vendors realized than supplying them free of charge to its cus- that educating potential clients was an im- tomers with the hardware lease or purchase.5 portant way to spread their message, sell Together these decisions provided the products, and build market share. Technical foundation for the software products and ser- training, off-the-shelf software, and various vices industry’s growth. The creation of com- types of support services became part of the patible computer systems families gave growing software industry. prospective software developers a large and Industry periodicals of the early 1970s growing market in which to sell their wares. contained numerous articles introducing the IBM’s unbundling meant that early software notion of DBMSs. Datamation and Computer- vendors would not have to compete with world printed tutorials, articles detailing free software from IBM.6 This environment early user experiences, and articles focusing not only motivated software entrepreneurs on specific software packages.11 Some brave to create software, it allowed the fledgling authors (and editors) published short case industry to flourish. By 1972, there were studies as well as articles comparing individ- 82 vendors offering 275 packages for the life ual software packages and introduced readers insurance industry.7 Indeed, Cincom (Total), to advanced concepts.12 We say ‘‘brave’’ be- Informatics (Mark IV), and CIM (Datacom/ cause vendors were constantly changing DB) began as software services firms.8 their offerings while working with existing DBMS software promised to reduce the customers to plan future improvements and organization’s hardware burden, speed up additions. EDP Analyzer brought a new October–December 2009 27 History of DBMS management and organizational viewpoint than rely on (out-of-date) monthly or quar- to the dialogue.13 Most users, however, terly hardcopy reports? Identifying these depended on their hardware vendors. Of always-changing requirements was not a triv- course, industry periodicals, consultants, ial process and required the attention of and reports were not the only ways for poten- senior personnel. tial users to learn about advances in database After defining the organization’s require- technologies and share practical experiences ments, the selection team had to compare with its implementation. Computer user the hardware and software needed to run groups, particularly SHARE and GUIDE, con- the various DBMS packages with their cur- tinued to play an important role, with rent hardware configuration and operating many sessions organized to discuss current system and estimate the cost of additional and forthcoming software. hardware, software, media, training, and per- As databases emerged as a distinct research sonnel. Although all vendors were happy to area, new specialized interest groups, jour- demonstrate their products, some vendors nals, and professional meetings were created. encouraged users to benchmark their prod- For example, the ACM created a Special Inter- ucts against their competitors, and this est Group on File
Recommended publications
  • Oracle Database Gateway for Adabas User's Guide, 11G Release 2 (11.2) E12074-01
    Oracle® Database Gateway for Adabas User’s Guide 11g Release 2 (11.2) E12074-01 July 2009 Oracle Database Gateway for Adabas User's Guide, 11g Release 2 (11.2) E12074-01 Copyright © 2008, 2009, Oracle and/or its affiliates. All rights reserved. Primary Author: Jeanne Wiegelmann Contributing Author: Maitreyee Chaliha, Sami Zeitoun, Oussama Mkaabal Contributor: Vira Goorah, Peter Wong This software and related documentation are provided under a license agreement containing restrictions on use and disclosure and are protected by intellectual property laws. Except as expressly permitted in your license agreement or allowed by law, you may not use, copy, reproduce, translate, broadcast, modify, license, transmit, distribute, exhibit, perform, publish, or display any part, in any form, or by any means. Reverse engineering, disassembly, or decompilation of this software, unless required by law for interoperability, is prohibited. The information contained herein is subject to change without notice and is not warranted to be error-free. If you find any errors, please report them to us in writing. If this software or related documentation is delivered to the U.S. Government or anyone licensing it on behalf of the U.S. Government, the following notice is applicable: U.S. GOVERNMENT RIGHTS Programs, software, databases, and related documentation and technical data delivered to U.S. Government customers are "commercial computer software" or "commercial technical data" pursuant to the applicable Federal Acquisition Regulation and agency-specific supplemental regulations. As such, the use, duplication, disclosure, modification, and adaptation shall be subject to the restrictions and license terms set forth in the applicable Government contract, and, to the extent applicable by the terms of the Government contract, the additional rights set forth in FAR 52.227-19, Commercial Computer Software License (December 2007).
    [Show full text]
  • Mapping out the DBMS Territory
    DATA BASE Mapping out the DBMS territory by Dr. George Schussel, stant both in terms of new product environment and should include the COP announcements and users since following: 1975. Of companies already using a • A DBMS offering data indepen­ Data base and data management DBMS product, 85 percent of these dence products have changed dramatically are not CODASYL, so it seems clear that CODASYL DDL and DML • A data dictionary defining and in the last several years. New hard­ controlling the data environment ware and software technologies standards will not be adopted by the marketplace, at least in their cur­ • A query language allowing user have greatly changed the character personnel access to the data base and number of products available. rent form. Just as the oil crisis hit the auto Mainframe DBMS • A report-generating system al­ makers, the "productivity crisis" The market for full-function trans­ lowing simplified programming has hit the software suppliers mar­ action processing DBMS, with their generation of reports ket. required high-level technical data • A screen mapper allowing gener­ base administration gurus and large ation of data entry screens by a In the 1970s there were no more simple example process than two dozen widely marketed mainframes, is still active but ex­ DBMS product lines. Non-IBM DP tremely competitive. Successful • A user language that is high -level, easy to use and an alternative to shops, using equipment such as vendors of this type of product have procedural languages such as Univac, Honeywell or Burroughs, COBOL and PL/l. simply took the DBMS offered by the hardware vendor.
    [Show full text]
  • ARCAD-Transformer RPG Version 10.06.Xx
    User Guide ARCAD-Transformer RPG Version 10.06.xx Publication Date: May, 2016 Prepared by the ARCAD Software Documentation Team FRANCE (HQ) 55 Rue Adrastée 74650 Annecy/Chavanod Tel. +33 4 50 57 83 96 GERMANY Richardstr. 84 22089 Hamburg Tel. +49 40 357 09 10-2 SWEDEN Prostvägen 36 141 43 HUDDINGE Tel. +46(0) 70-793 6570 USA 1 Phoenix Mill Lane, Suite 203 Peterborough, NH 03458 Tel. +1 (603) 371-9074 HONG KONG Room 22, Smart-Space 3F, 100 Cyberport Road Tel. +852 3618 6118 ARCAD-Transformer RPG User Guide | Copyright © 1992, 2016 by ARCAD Software All rights reserved. The following terms are names owned by International Business Machines Corporation in the United States, other countries, or both: AS/400®, ClearCase, ClearQuest®, DB2, DB2 Connect™, DB2 Universal Database™, ibm.com, IBM i, iSeries, System i, OS/400, Rational®, SP2, Service Pack, WebSphere. Java and all names based on Java are owned by Oracle Corp. in the United States, other countries, or both. Eclipse is a registered trademark of Eclipse Foundation, Inc. Other names of companies, products or services are the property of their respective owners. Page 2 / 108 ARCAD Software • Copyright © 2016 • All Rights reserved. • arcadsoftware.com ARCAD-Transformer RPG Contact ARCAD Software | User Guide Contact ARCAD Software arcadsoftware.com Headquartered in France at the foot of the Alps, ARCAD Software offers global services and has offices on three continents. Country Address Telephone Arcad Software Tel. +33 4 50 57 83 96 55 Rue Adrastée Fax +33 4 50 57 52 79 74650 Annecy/Chavanod [email protected] FRANCE (HQ) Arcad Software 67 Rue du Bourbonnais Tel.
    [Show full text]
  • Paper Mainframe
    Club des Responsables d’Infrastructures et de Production IT Infrastructure and Operations IT INFRAsTRucTuRE & Operations MANAgEMENT Best Practices PAPER MAINFRAME IBM zSeries Mainframe Cost Control te i Authors Laurent Buscaylet, Frédéric Didier Bernard Dietisheim, Bruno Koch, Fabrice Vallet sponsored by september 2012 Wh PAgE 2 table of contents INTRODucTION 5 1. “z” POsITIONINg AND Strategy wIThIN A Company 6 1.1 Role of the Mainframe in the eyes of cRIP’s Members 6 1.2 The Position of system z vs. Distributed systems (windows, uNIX, Linux) 6 1.3 system z - Banking/Finance/Insurance vs. general Industry sectors 7 1.4 Evolution strategies for system z 9 2. z Platform OVERVIEw 11 2.1 Technical terms of reference 11 2.1.1 Servers 11 2.1.2 SAN & Networks 12 2.1.3 Storage 13 2.1.4 Backups 15 2.1.5 technical Architecture & Organization 16 2.1.6 Business Continuity 19 2.2 Mainframe software vendors (IsV’s) analysis & positioning 20 2.2.1 the Main Suppliers (ISV’s) 22 2.2.2 the Secondary Suppliers (ISV’s) 22 2.3 survey Results 3. z PLATFORM cOsT cONTROL 28 3.1 controling software cost 28 3.1.1 IBM Contractual Arrangements 28 3.1.2 ISV negotiation strategies: holistic/generic or piecemeal/ad-hoc 31 3.1.3 ISV contractual arrangements: licensing modes & billing 35 3.1.4 MLC: Controlling the billing level 38 3.1.4.1 The IBM billing method 38 3.1.4.2 The problem of smoothing (consolidating) peak loads 41 3.1.4.3 The methods and controls for controlling the software invoice 42 3.1.4.4 Survey feedback and recommendations 43 3.1.4.5 Other Cost Saving Options 48 3.2 Infrastructure cost Reduction 50 3.2.1 Grouping and Sharing Infrastructure 50 3.2.2 Use of specialty engines: zIIP, zAAP, IFL 54 3.3 controlling operational costs 55 3.3.1 Optimization, performance & technical/application component quality 55 3.3.2 Capacity Management, tools & methods 58 4.
    [Show full text]
  • Proposal for a Ph.D. Degree in Informatics
    Proposal for a Ph.D. Degree in Informatics Contacts Judith S. Olson Bren Professor of Information and Computer Sciences Department of Informatics Donald Bren School of Information and Computer Sciences 949-824-0080 [email protected] Gillian R. Hayes Assistant Professor Department of Informatics Donald Bren School of Information and Computer Sciences 949-824-1483 [email protected] November 19, 2012 1 Introduction 1.1 Aims and Objectives This proposal represents a natural evolution in the field of Informatics (elsewhere called Information or Information Science), a broad term covering a host of subspecialties, including human-computer interaction, social informatics, information economics, information policy, and library and archival science. Complementing our existing offerings, we propose to offer a Ph.D. in Informatics in the Department of Informatics to signify the recognizable “flavor” of the field of study we offer. Until recently, the department offered four tracks within the Informatics concentration of the ICS Ph.D.: • Informatics – General • Informatics – Track in Interactive and Collaborative Technology • Informatics – Track in Ubiquitous Computing • Informatics – Track in Software Engineering A Ph.D. degree in Software Engineering has now been added, replacing the degree currently called Informatics—Track in Software Engineering. This proposal specifies the Ph.D. in Informatics, replacing the three remaining tracks (Informatics- General, Informatics-ICT and Informatics-Ubiquitous Computing). The new arrangement (two PhDs, one in Software Engineering and one in Informatics) mirrors our undergraduate offerings (two B.S. degrees, one in Software Engineering and one in Informatics.) In contrast to our current Ph.D. offering in Information and Computer Science, we believe that the naming of the Ph.D.
    [Show full text]
  • Software Inspections at Nasa Ames
    N86-.. 30363 Oualitv Assurance Software Insoections at NASA Ames Metrics for Feedback and Modification Greg Wenneson, Informatics General Corporation Software Inspections are a set of formal technical review procedures held at selected key points during software development for the purpose of finding defects in software documents. Inspections are a Quality Assurance tool and a Management tool. Their primary purposes are to improve overall software system quality while reducing lifecycle costs and to improve management control over the software development cycle. The Inspections process can be customized to specific project and development type requirements and are specialized for each stage of the development cycle. For each type of Inspection, materials to be inspected are prepared to predefined levels. The Inspection team follows defined roles and procedures and uses a specialized checklist of common problems in reviewing the materials. The materials and results from the Inspection have to meet explicit completion criteria before the Inspection is finished and the next stage of development proceeds. Statistics, primarily time and error data, from each Inspection are captured and maintained in a historical database. These statistics provide feedback and feedforward to the developer and manager and longer term feedback for modification and control of the development process for most effective application of design and quality assurance efforts. HISTORY Software Inspections were developed in the early mid-1970s at IBM by Dr. Mike Fagan, who was subsequently named software innovator of the year. Fagan also credits IBM members O.R.Kohli, R.A.Radice and R.R.Larson for their contributions to the development of Inspections.
    [Show full text]
  • Objex, Inc. Product Overview April 2015
    WHERE COMMON SENSE MEETS UNCOMMON INGENUITY ObjEx, Inc. Product Overview April 2015 ObjEx INC PO Box 25608 Scottsdale, AZ 85255-5608 Tel: (480) 588-7776 http://www.obj-ex.com ObjEx, Inc. Product Overview Table of Contents Contents TABLE OF CONTENTS I OVERVIEW 1 PACKAGED SOFTWARE 1 SERVICES 2 OBJEX XMLQUERY 4 OBJEX XMLPUBLISHER 5 OBJEX PROGRAM CALL 6 ABOUT OBJEX 7 2015 ObjEx, Inc. Page i ObjEx, Inc. Product Overview Overview ObjEx was founded in 1996 by former Cullinet and CA executives and employees who have more than 65 years of IDMS experience and 20 years as IDMS Integrators. We provide software and services to our clients which include many large organizations. Packaged Software Our packaged software is conceived and developed as a result of our integration experiences related to Service Oriented Architecture, cloud computing calls to and from IDMS, trigger- based data publishing, and replication. ObjEx products are focused on Service Oriented Architecture (SOA) and are designed to participate within the latest n-tier architectures. ObjEx has been used to: Extend existing IDMS systems with new modern application components either direct client/server or through SOA, all in one database. Allow existing IDMS applications to call out to internal and external web services. Provide modern applications the ability to use existing IDMS DC-Cobol and ADSO programs as web services. This allows the reuse of business rules without the hassles of screen scraping. Push IDMS data to non-mainframe applications without altering IDMS programs. The products are useful for ERP and warehouse integration, data replication to operational data stores, and data warehousing feeds.
    [Show full text]
  • Installation for Z/OS
    Natural Installation for z/OS Version 8.2.3 für Großrechner November 2012 Dieses Dokument gilt für Natural ab Version 8.2.3 für Großrechner. Hierin enthaltene Beschreibungen unterliegen Änderungen und Ergänzungen, die in nachfolgenden Release Notes oder Neuausgaben bekanntgegeben werden. Copyright © 1979-2012 Software AG, Darmstadt, Deutschland und/oder Software AG USA, Inc., Reston, VA, Vereinigte Staaten von Amerika, und/oder ihre Lizenzgeber.. Nähere Informationen zu den Patenten und Marken der Software AG und ihrer Tochtergesellschaften befinden sich unter http://documentation.softwareag.com/legal/. Die Nutzung dieser Software unterliegt den Lizenzbedingungen der Software AG. Diese Bedingungen sind Bestandteil der Produkt- dokumentation und befinden sich unter http://documentation.softwareag.com/legal/ und/oder im Wurzelverzeichnis des lizensierten Produkts. Diese Software kann Teile von Drittanbieterprodukten enthalten. Die Hinweise zu den Urheberrechten und Lizenzbedingungen der Drittanbieter entnehmen Sie bitte den "License Texts, Copyright Notices and Disclaimers of Third Party Products". Dieses Dokument ist Bestandteil der Produktdokumentation und befindet sich unter http://documentation.softwareag.com/legal/ und/oder im Wurzelverzeichnis des lizensierten Produkts. Dokument-ID: NATMF-INSTALL-ZOS-823-20121108 Table of Contents Preface ............................................................................................................................... ix I Installation Process and Major Natural Features on z/OS ..............................................
    [Show full text]
  • Name Synopsis Description
    DECKMATE(1) DECKMATE(1) NAME deckmate − simulate line printer paper output in PostScript SYNOPSIS deckmate [−i input_file.txt][−o output_file.ps][−C[C] [c1[,c2[,...[,c12]]]]] [−l {f |c|r}] [−c R,G,B][−F list- ing_font][−f nob[order]] [−f noc[olumns]] [−f nop[agenumbers]] [−t n1,n2,...] [−w] [DD=[step- name.]ddname] DESCRIPTION deckmate(1) reads a plain text program file and generates a PostScript output that simulates 80- and 132-column line printer paper,with alternating sets of three lines of white paper and three lines of colored paper.Output pages show66lines of input text per sheet as on line printer paper.Bydefault, each page prints with column column headings numbered from 01 through 80 or 132, and with a border around the listing area. The −f noborder option turns offborder printing, and the −f nocolumns option turns offthe column number heading. The page number prints at the bottom of each page unless disabled with the -f nopagenumbers option. The default output is 80 columns in portrait mode. The "−w"(wide) option selects 132 column output in landscape mode. Pages are scaled to fit on U.S. letter and A4 paper. Special cases for FORTRAN, COBOL, and RPG reproduce vertical lines between columns corresponding to fields on programming template forms for early fixed-format versions of those languages. These lines can be especially advantageous when examining RPG, which still uses fixed-form statements. The −t option allows adding dashed gray lines at additional, user-selectable column boundaries. Among other uses, these customized column markings can delineate fields of fixed-format data records.
    [Show full text]
  • Database Machines in Support of Very Large Databases
    Rochester Institute of Technology RIT Scholar Works Theses 1-1-1988 Database machines in support of very large databases Mary Ann Kuntz Follow this and additional works at: https://scholarworks.rit.edu/theses Recommended Citation Kuntz, Mary Ann, "Database machines in support of very large databases" (1988). Thesis. Rochester Institute of Technology. Accessed from This Thesis is brought to you for free and open access by RIT Scholar Works. It has been accepted for inclusion in Theses by an authorized administrator of RIT Scholar Works. For more information, please contact [email protected]. Rochester Institute of Technology School of Computer Science Database Machines in Support of Very large Databases by Mary Ann Kuntz A thesis. submitted to The Faculty of the School of Computer Science. in partial fulfillment of the requirements for the degree of Master of Science in Computer Systems Management Approved by: Professor Henry A. Etlinger Professor Peter G. Anderson A thesis. submitted to The Faculty of the School of Computer Science. in partial fulfillment of the requirements for the degree of Master of Science in Computer Systems Management Approved by: Professor Henry A. Etlinger Professor Peter G. Anderson Professor Jeffrey Lasky Title of Thesis: Database Machines In Support of Very Large Databases I Mary Ann Kuntz hereby deny permission to reproduce my thesis in whole or in part. Date: October 14, 1988 Mary Ann Kuntz Abstract Software database management systems were developed in response to the needs of early data processing applications. Database machine research developed as a result of certain performance deficiencies of these software systems.
    [Show full text]
  • Comparison of Static Analysis Tools for Quality Measurement of RPG Programs
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by SZTE Publicatio Repozitórium - SZTE - Repository of Publications Comparison of Static Analysis Tools for Quality Measurement of RPG Programs Zolt´anT´oth1, L´aszl´oVid´acs2, and Rudolf Ferenc1 1 Department of Software Engineering, University of Szeged, Hungary [email protected], [email protected] 2 MTA-SZTE Research Group on Artificial Intelligence, Hungary [email protected] Abstract. The RPG programming language is a popular language em- ployed widely in IBM i mainframes nowadays. Legacy mainframe systems that evolved and survived the past decades usually data intensive and even business critical applications. Recent, state of the art quality assur- ance tools are mostly focused on popular languages like Java, C++ or Python. In this work we compare two source code based quality man- agement tools for the RPG language. The study is focused on the data obtained using static analysis, which is then aggregated to higher level quality attributes. SourceMeter is a command line tool-chain capable to measure various source attributes like metrics and coding rule viola- tions. SonarQube is a quality management platform with RPG language support. To facilitate the objective comparison, we used the SourceMe- ter for RPG plugin for SonarQube, which seamlessly integrates into the framework extending its capabilities. The evaluation is built on analysis success and depth, source code metrics, coding rules and code duplica- tions. We found that SourceMeter is more advanced in analysis depth, product metrics and finding duplications, while their performance of cod- ing rules and analysis success is rather balanced.
    [Show full text]
  • Tcvision IBM Mainframe Integration Through Change Data Capture Fact Sheet
    tcVISION IBM Mainframe Integration Through Change Data Capture Fact Sheet Mainframe data integration has taken on more urgency in recent years The tcVISION Solution as organizations seek to relocate mainframe workloads to lower-cost tcVISION is ready to meet new technologies and challenges. Thanks to platforms, modernize mainframe applications and leverage analytics for tcVISION’s flexible architecture, support for new targets—including AWS, customer insight and competitive advantage. These factors are driving specialty, NoSQL and analytic databases such as Exasol, IBM DB2 BLU adoption of Cloud (e.g., Amazon Web Services [AWS]) and Big Data as and MongoDB—, transport layers and protocols is being continuously strategic components in corporate technology architecture. added, quickly and with minimal effort. With tcVISION, real-time Cloud and Big Data integration can embrace both mainframe (IBM DB2, IMS/ tcVISION’s support for Cloud and Big Data as targets is fully integrated DB, DL/1, Software AG Adabas, CA IDMS, CA Datacom and alongside traditional Linux/Unix/Windows (LUW) targets such as Oracle VSAM), Cloud, and LUW (IBM DB2 LUW, Oracle, IBM Informix, Database, IBM DB2 LUW, Software AG Adabas LUW, IBM Informix, Sybase, Microsoft SQL Server, PostgreSQL, Software AG Adabas Sybase, Microsoft SQL Server, PostgreSQL and ODBC. LUW) sources. tcVISION can deliver replicated data to Cloud and Big Data targets Why tcVISION? through a variety of means: creating files, writing directly into Hadoop • Increasing number of enterprise applications that utilize their own HDFS, and via streaming using Apache Kafka as the transport layer. Data databases. can be packaged using standard JSON and CSV protocols. • Requirement for up-to-date information demands real-time, bi- directional data synchronization between mainframe and open systems.
    [Show full text]