Theory of Files Non-Arithmetic Data Processing, Data Flow, Algebraic

Total Page:16

File Type:pdf, Size:1020Kb

Theory of Files Non-Arithmetic Data Processing, Data Flow, Algebraic absorbed by IFIPS. Also, the British Standards Institution re- (e) Development of other languages, e.g., the Algebraic Busi- cently circulated to many individuals in the U.S. a "Draft British ness Language. The v. Neumann-type data processing languages Standard Glossary of Tei'ms Used in Data Processing" for com- (i.e., languages in which procedures are represented as images of ment. flow charts) do not satisfy the needs of most non-arithmetic All these groups are working toward the same goal. They hope procedures. to clarify and establish the meaning of certain technical terms REFERENCES : used in the computing literature so that these words convey the GOLDSTINE, H.; NEUMANN, J. v. Planning and coding for an same meaning to the readers. electronic digital computer. The Institute for Advanced Study Publications, 1947. LOMBAROI, L. Theory of files. Proc. 1960 Eastern Joint Comput. Conf., Paper 3.3. Theory of Files , System handling of functional operators. J. ACM 8, (1961), Numerical Analysis Research Project, University of Cali- 168. fornia at Los Angeles, Calif. , Mathematical structure of non-arithmetic data processing Reported by: Lionello Lombardi (May 1961) procedures. Submitted for publication. , Inexpensive punched card equipment. In press, Journal of Descriptors: theory of files, flow control expressions, Machine Accounting. non-arithmetic data processing, data , Logic of automation of system communications. In prepa- flow, algebraic business language, yon ration. Neumann languages, sorting, Boolean , Coding in storage ~nd se~rching systems. In P. Garvin algebra (Editor), Natural Language and the Computer, ch. 14; in press, McGraw-Hill Book Co. In hardware design, the theory of files is used as a tool to analyze the features of data flow of the systems for which the equipment is designed. As a result, it is proved possible to formu- A Class of Search-Models for Machine Retrieval late a punched card data processing system considerably less expensive than any presently available. The logical features of Information Retrieval Project, IBM Research Center, this system, based on a novel organization of data flow, is de- Yorktown Heights, N. Y. termined by synthesis. Reported by: Eugene Wont (May 1961) Currently, the theory of files is being applied to the study of Descriptors: search, search-strategy, optimization, the flow of information through random-access fixed-plus-erasable access to mechanical storage, informa- memory systems in order to select a compact set of parameters which characterize the flow involved in any specific application, tion retrieval, stochastic models and then to formulate relations between these system parameters, A class of theoretical models for studying optimum search- the minimum storage capacity requirements of the systems, and strategies in one dimension, has been constructed with a view the optimum storage capacity distribution among the components. towards the eventual application of these strategies to machine Development and applications of the theory of files are carried retrieval. The basic assumption underlying these models can be out under the sponsorship of the Office of Naval Research. The briefly stated as follows: following results have been obtained: (1) The probability that tile object of search lies in the interval (a) Determination of a common pattern to which the coordi- (x, x+dx) is p(x)dx, where the density function p(x) is known a nated data flow conforms of any non-arithlnetic data processing priori. procedure involving files, such as machine accounting, access to (2) The search is conducted with constant speed v. stored information, dictionary analysis, etc. In particular, it is (3) It is possible to skip, i.e., to move from one point to another shown that the data flow configuration of any procedure involving without searching, with constant speed s. In general it can be n files can be fully represented at any time by means of only 5n+4 assumed that s > v. boolean variables (the indicators). The major feature of this model is a balance between the ad- (b) Development of a pattern of system language (the Algebraic vantage of always searching where the probability of success is Business Language) which allows for the use of logico-mathe- the highest against the disadvantage of frequent skipping that matical techniques to describe and control the data flow by means this may incur. This feature exists in many practical storage of specially designed boolean expressions (the Flow Control Ex- devices, e.g., magnetic tape. pressions). Without loss of generality any search procedure can be ex- (c) Description of processes in rigorous mathematical terms pressed as a sequence of alternately searching and skipping where the relevance of arithmetic operations is low and the main operations. Using the letters v for searching s for skipping, the problem is logical input-output coordination. Such procedures, search-sequence can be expressed as where a small amount of processing is performed on relatively ~) 8 V S V large amounts of information, include most business data handling Xo ) Xl ) :g2 ) X3 ) X4 ) etc. and document processing procedures. Now it is possible to repre- sent such processes as sets of operations defined in boolean alge- The object of the analysis is to find the sequence {x,~} which bras whose elements are files. optimizes the search in stone sense. For example, a complete k-way v. Neumann sorting procedure The model, that has been described, with p(x) = (a/2)e -"l~ is represented by the formula, and the criterion of minimum mean search-time T has been k, ( i+1 ) studied intensively with considerable success. Both parametric F~"+') = ~ F~m) , (i = O, 1,2 ... ; m = O, 1,2... ) optimization, e.g., let x~ = (-1)n(x0+n~) and find x0 and ~ which k*(i+l)] minimize T, and non-paremetric optimization have been achieved. where F~t) represents the sth ordered sequence of records avail- Analysis with other forms of p(x) and other criteria for optimi- able after t runs. zation are being undertaken. Modifications of the basic model to (d) Representation of procedm'es as sets of equations relating adalSt more closely to practical storage arrangements are also the input data to the output results. being considered. 324 Communications of the ACM .
Recommended publications
  • An Operational Data Processing System for Natural History Specimens
    profiles to be plotted on lIlaps—a technique particu- larly helpful in reconnaissance surveys (see, for An Operational example, Herron, in press). Data Processing System for The production of charts merits further discussion because of the potential application of this tech- Natural History Specimens nique to other investigations. For instance, final ad- justed track charts could be produced on board ship on a- day to day basis. The programs incorporate B. J. LANDRUM subprogFams which provide scaling, grid drawing, 2 Smithsonian Oceanographic Sorting Center and annotating for a Mercator Projection of vari- Smithsonian Institution able scale. Functions are employed which convert latitude and longitude to the proper x-y coordinates and for plotting within a specified area. These programs K. G. SANDVED utilize tables (within the program) which are effec- tively tables of meridional parts and allow plotting Office of Antarctic Programs of all areas of the world from 80°N. to 80°S. The National Science Foundation size of a specific plot is limited only by the scale chosen and the physical limitations of the plotter. In this fast-moving world where advances in It is the intention of the Lainont-Doherty investi- technology often seem to outpace the progress in gators eventually to do all basic data processing and science, there are some areas where the phase dis- display on board Eltanin within a few hours or a placement is particularly evident. A good example day of the actual data collecting. There are many is museums, which, although faced with formidable advantages to this approach: Errors in data collec- problems in inventorying and cataloging specimens tion will be caught early and corrected, more intelli- collected over decades and even centuries, only re- gent decisions regarding the day-to-day and cruise- cently have begun to apply computers to the task to-cruise schedule and track of the ship can be made, of inventorying and retrieving their ever-increasing and time in the laboratory can be devoted entirely holdings.
    [Show full text]
  • IBM 709 ...A Powerful New Data Processing System, 1957
    .. a powerful new data processing system The IBM 709-outstanding new features: Increased input-output flexibility with the new Data Synchronizer Greater input-output capacity-up to 48 tape units Automatic checking while writing information on tape Increased storage capacity-up to a half billion decimal digits Fast, simplified conversion with new "convert" instructions Simplified data handling with "indirect addressing" An advanced combination of new features enables the 709 to handle commercial, scientific, engineering, and management science problems-all with outstanding efficiency. Here for the first time is a large-scale, high-speed electronic computer designed to solve problems of increased complexity and magnitude which confront both modern business and science. The 709 revolutionizes the basic elements of data processing.. INPUT-OUTPUT taneously. Forexatnple,three tapes may tions. 32,768 "words" of magnetic core The new Data Synchronizer of the 709 be operating concurrently with a card storage, equivalent to over 327,000 is a powerful link between input- reader, a card punch, and a printer. decimal digits of storage, enable in- output facilities and magnetic core To assure the maximum effective- ternal high-speed handling of the most storage. Simultaneous reading and ness of this flexibility, a large input- complex and voluminous problems. writing of many combinations of mul- output capacity is pr0vided.A~many as Random access to any location in tiple input-output units is made 48 magnetic tape units,3 card readers, memory, and the parallel transfer of possible, each unit operating inde- 3 card punches, and 3 printers are avail- information to and from magnetic core pendently of the other and independ- able to the 709 at any one time.
    [Show full text]
  • Research; *Electronic Data Processing; Information Processing; Information Systems; *Manpower Utilization; *Personnel Data; Personnel Management; Questionnaires
    DOCUMENT RESUME ED 084 382 CE 000 533 AUTHOR Bayhylle, J. E., Ed.; Hersleb, A., Ed. TITLE The Development of Electronic Data Processing in Manpower Areas; A Comparative Study of Personnel Information Systems and their Implications in Six European Countries. INSTITUTION Organisation for Economic Cooperation and Development, Paris (France) . PUB DATE 73 NOTE 366p. AVAILABLE FROMOECD Publications Center, Suite 1207, 1750 Pennsylvania Ave., N.W., Washington, DC 20006 ($8.50) EDRS PRICE MF-$0.65 HC- $13.16 DESCRIPTORS Case Studies; Data Bases; Data Processing; *Ecqnomic .Research; *Electronic Data Processing; Information Processing; Information Systems; *Manpower Utilization; *Personnel Data; Personnel Management; Questionnaires ABSTRACT Usable data about human resources in the ecqnqmy are fundamentally important, but present methods of obtaining, organizing, and using such information are increasingly unsatisfactory. The application of electronic data processing to manpower and social concerns should be improved and increased. This could stimulate closer exchange between government, institutions, and business enterprises in the collection and dissemination of manpower data. The first phase of the study was a broad-scope questionnaire of 38 companies in Austria, France, Germany, Great Britain, Sweden, and Switzerland which have developed, or are planning to develop, computer-based personnel information, systems. The second phase consisted of depth study of eight of the companies. Case study investigations were concerned with the content and function of the system and with the administrative and personnel consequences of the introduction of such systems. A set of key.points which emerged from the study is developed. A model demonstrates the need for coordination to increase communication vertically and horizontally.
    [Show full text]
  • Applied Accounting and Automatic Data-Processing Systems." (1962)
    Louisiana State University LSU Digital Commons LSU Historical Dissertations and Theses Graduate School 1962 Applied Accounting and Automatic Data- Processing Systems. William Elwood Swyers Louisiana State University and Agricultural & Mechanical College Follow this and additional works at: https://digitalcommons.lsu.edu/gradschool_disstheses Recommended Citation Swyers, William Elwood, "Applied Accounting and Automatic Data-Processing Systems." (1962). LSU Historical Dissertations and Theses. 745. https://digitalcommons.lsu.edu/gradschool_disstheses/745 This Dissertation is brought to you for free and open access by the Graduate School at LSU Digital Commons. It has been accepted for inclusion in LSU Historical Dissertations and Theses by an authorized administrator of LSU Digital Commons. For more information, please contact [email protected]. This dissertation has been 62—3671 microfilmed exactly as received SWYERS, William Elwood, 1922- APPLIED ACCOUNTING AND AUTOMATIC DATA-PRO CESSING SYSTEMS. Louisiana State University, Ph.D., 1962 Economics, commerce—business University Microfilms, Inc., Ann Arbor, Michigan APPLIED ACCOUNTING AND AUTOMATIC DATA-PROCESSING SYSTEMS A Dissertation Submitted to the Graduate Faculty of the Louisiana State University and Agricultural and Mechanical College in partial fulfillment of the requirements for the degree of Doctor of Philosophy in The Department of Accounting by William Eiu<Swyers B.S., Centenary College, 1942 M.B.A. t Louisiana State University, 1948 January, 1962 ACKNOWLEDGMENTS The writer wishes to express appreciation to Dr. Robert H. Van Voorhis, Professor of Accounting, Louisiana State University, for his valuable assistance and guidance in the preparation of this dissertation. The writer wishes to acknowledge, also, the helpful suggestions made by Dr. W. D. Ross, Dean of the College of Business Administration, Dr.
    [Show full text]
  • Course Title: Fundermentals of Data Processing
    NATIONAL OPEN UNIVERSITY OF NIGERIA SCHOOL OF EDUCATION COURSE CODE: BED 212: COURSE TITLE: FUNDERMENTALS OF DATA PROCESSING BED 212: FUNDERMENTALS OF DATA PROCESSING COURSE DEVELOPER & WRITER: DR. OPATEYE JOHNSON AYODELE School of Education National Open University of Nigeria 14/16 Ahmadu Bello Way, Victoria Island Lagos. Nigeria. COURSE CORDINATOR: DR. INEGBEDION, JULIET O. School of Education National Open University of Nigeria 14/16 Ahmadu Bello Way, Victoria Island Lagos. Nigeria. INTRODUCTION BED 212: FUNDERMENTALS OF DATA PROCESSING This course is designed to equip the students with knowledge of the data processing in technical and vocational education (TVE) especially Business Education research. WHAT YOU WILL LEARN You will learn the components of data processing, hardware and software components, file management and organization in Business Education- based data. It will also interest you to learn about basics of research, approaches and designs with data gathering techniques. Relevant statistical tools you can use to analyse your data would be across and how to write your research reports. COURSE AIMS This course aims at producing competent business educators who will be versed in organizational and research data processing in order to foster time and effective information that will be used in decision making. In order to enable you meet the above aims, modules constituting of units have been produced for your study. Apart from meeting the aims of the course as a whole, each course unit consists of learning objectives which are intended to ensure your learning effectiveness. COURSE OBJECTIVES The course objectives are meant to enable you achieve/acquire the following: 1) Gain in-depth knowledge of data processing and its functions in business organisations and educational settings.
    [Show full text]
  • IBM 1440: New Low Cost Data Processing System, 1962
    New low cost A manaaement tool, a profit tool for YOW business L IBM announces new low cost 1440-problem solver for all types of businesses It's an accounting system, and a business infor- 1 mation system. And its cost makes it practical '11 and profitable for smaller volume businesses. In fact, IBM'S new 1440 Data Processing Sys- tem is specifically designed to meet the needs, ' and solve the problems, of smaller volume businesses. Whether you're producing parts, forwarding freight, lending money or caring for patients, the 1440 can help make your job easier. The 1440 handles payrolls, accounts receiv- ' able, inventory, all your day-to-day business , operations. It keeps records always up-to-date, ,: always immediately available, always useful. I But even more important, the 1440 gives you the up-to-the-minute control and supervision . you need for your business. With the 1440 YOU i get the kind of business information you need, I on your desk when you need it, in a form you can use. These reports, schedules and analyses eliminate much of the guess work in day-to-day business decisions-those decisions that spell success or failure, profit or loss. And the 1440 does all this at a price smaller volume business can easily afford. The 1440 provides the tools for modern business management. When you use the 1440 as a management tool, you make it a profit tool as well. i . I What's the 1440 like? These instructions, called a "stored pro- gram," are stored in the system.
    [Show full text]
  • Optimizing Performance of Data Processing
    SPECIAL TOPIC: DATA PROCESSING Optimizing performance of data processing Samuel Brown1* and Tony Martin1 describe the process of optimizing data throughput, to reduce turnaround and enable both full automation and interactivity of geophysical tasks in an architecture agnostic environment. Introduction transcribed many times over. While optimization of single pro- The industry associates data processing with the application of cesses, using advanced algorithms, and access to almost limitless geophysical algorithms to seismic data, but this definition is too resources is important, it is only as important as the ability to both narrow. We should consider a broader meaning particularly in effectively apply the algorithm, and manipulate the underlying these times of increased automation, and time and cost conscious- data. Long and Martin (2019) outline a method of data mining ness. If processing is a method to produce information from the for parameter selection to reduce the impact of manual processing manipulation of data, then we should consider the framework of data. To fulfil their work, they require dynamic manipulation that enables this process. A system must exist to facilitate the of the data and a productive implementation of the geophysical work. Generally, systems are a collection of computers, networks, tasks. They also need this to happen in a programme and platform processes and to a lesser extent people. For any given input, they independent way, where more resources mean ever more fruitful create an output, which in seismic data processing can be the returns on turnaround. application of a geophysical algorithm, or it can be reordering, In this paper we describe a data processing system that is reducing or aggregating data.
    [Show full text]
  • Metadata Driven Integrated Statistical Data Processing and Dissemination System
    METADATA DRIVEN INTEGRATED STATISTICAL DATA PROCESSING AND DISSEMINATION SYSTEM By Karlis Zeila Central Statistical Bureau of Latvia Abstract The aim of this report is to introduce participants with the experience gained within the process of the development and implementation of the new generation statistical data processing system, which integrates several subsystems, and is metadata driven. The new system in accordance with Central Statistical Bureau (CSB) IT Strategy is developed as centralized metadata driven data processing system, where all data and metadata is stored in corporate data warehouse. The new approach is to unite what logically belongs together by using advanced IT tools, to ensure the rationalizing, standardization and integration of the statistical data production processes. This in practice means movement from classic stovepipe to process oriented data processing approach. System integration is based on direct use of the Business Register data in case of Business statistics surveys data processing. Registers module of the system foreseen exploitation of several different statistical registers. Most of the system software modules are connected with the Registers module. All System software modules are connected with the Core Metadata module. System is modular and consists of 10 software modules, which are covering and to supporting all phases of the statistical data processing. The main conclusions gained within the system development and implementation process are listed at the end of this report. 1. INTRODUCTION Production of high quality statistics is impossible without wide use of newest solutions of information technologies. Therefore CSB of Latvia has set the goal to create a modern, effectively working IT infrastructure which provides convenient, fast and safe collection, processing, storage, analysis and dissemination of statistical data taking into account the requirements of information society.
    [Show full text]
  • IBM 1401 Data Processing System
    Jrrn~ General Information Manual ® 1401 Data Processing System MINOR REVISION (February, 1960) This edition D24-1401-1 is a minor revision of the preceding edition but does not obsolete D24-1401-0. The principal changes in this edition are: PAGE SUBJECT 20 Input-Output Storage Assignments 23,24 Input-Output Branch Codes 28 B (I) Unconditional Branch, Figure 24 30 Z (A) (B) Move and Zero Suppress 32 • (I) Stop and Branch 35 Figure 30 Cycles 1 through 8 41 Auxiliary Console 42 Figure 37 43 Printer Controls 51 A and B Auxiliary Registers 55 Figures 54 and 55 MUltiply Subroutine 57 Figure 57 Divide-Subroutine 58,59 Character Code Chart 60 1401 Timing Formulas 62 Magnetic Tape Timings © 1959 by International Business Machines Corporation Contents INTRODUCTION 5 LOGIC OPERATIONS .. 28 IBM 1401 DATA PROCESSING SYSTEM 5 Logic Operation Codes 28 Move and Load Codes 29 THE PHILOSOPHY OF THE IBM 1401 6 Miscellaneous Operation Codes 31 The Stored Program Concept 6 EDITING 33 Magnetic-Core Storage 7 Magnetic-Tape Storage 8 Expanded Print Edit 36 -----~ Language .. 8 OPERATING FEATURES 38 Processing 9 Solid State Circuitry 9 Console Keys, Lights, and Switches 38 Advanced Design 10 Auxiliary Console 40 IBM 1402 Card Read-Punch Operating Keys IBM 1401 CARD SYSTEM 11 and Lights . .. 42 IBM 1403 Pri~ter Operating Keys, Lights 42 Physical Features 12 Data Flow 16 IBM 1401 MAGNETIC TAPE SYSTEMS 44 Checking 18 Data Flow 46 Word Mark 18 Magnetic Tape 46 T~eSillti~ 50 STORED PROGRAM INSTRUCTIONS 19 Console Keys, Lights, and Switches 51 Instruction Format 19 COLUMN BINARY DEVICE (Optional) 52 ADDRESSING 20 PROGRAM LOADING ROUTINE 54 Input-Output Storage Assignments 20 CLEAR ROUTINE 55 Address Registers 21 MUL TIPLICA TION AND DIVISION SUBROUTINES 55 Chaining Instructions 21 Loading Instructions 22 Multiplication 55 Division 56 INPUT-OUTPUT OPERATIONS 23 IBM 1401 TIMINGS 60 Input-Output Codes 23 Card Systems 60 Magnetic Tape 62 ARITHMETIC OPERATIONS 25 Arithmetic Operation Codes 25 INDEX 63 FIGURE 1.
    [Show full text]
  • Introduction to Mm Data Processing Systems Llrn~ General Information Manual ® Introduction to IBM Data Processing Systems Preface
    Introduction to mM Data Processing Systems llrn~ General Information Manual ® Introduction to IBM Data Processing Systems Preface All IBM Data Processing Systems, regardless of size, type, or basic use, have certain common fundamental concepts and operational principles. This manual presents these concepts and principles as an aid in developing a basic knowledge of computers. The manual is designed for use in training programs where a basic knowledge of computers is the end objective or is a prerequisite to the detailed study of a particular IBM system. Each section is organized to present a logical association of related concepts and operational principles. The sections may be used in a progressive sequence to develop a concept of the computer system, or they may be used independently as refer­ ence material. The subject matter has been generalized and refers to actual machines and systems as little as possible. Specific systems are mentioned only to illustrate a general principle, not to compare one system with another. The appendix contains supplementary reference information which, because of its format or its logical use, is not related directly to any specific section of the manual. @ 1960 by International Business Machines Corporation Contents INTRODUCTION TO IBM DATA PROCESSING SYSTEMS ..... 5 The Data Processing System ............................ 12 DATA REPRESENTATION ..................................................................................... 17 Computer Data Representation .......................................................
    [Show full text]
  • Registration of Personal Data Processing Systems
    The NPC is an attached agency of the Department of Information and Communications Technology Registration of Personal Data Processing Systems A data processing system refers to the structure and procedure by which personal data is collected and processed in an information and communications system, or any other relevant filing system. It includes the purpose and intended output of the processing. What are the contents of the application for registration? An application for registration shall include the following information: 1. Name and address of the personal information controller Who are required to register or personal information processor, and of its representative, their personal data processing systems? if any, including their contact details; 2. Purpose/s of the processing, and whether processing is being done under an outsourcing or subcontracting agreement; 3. Description of the category or categories of data subjects, and of the data or categories of data relating to them; 4. Recipients or categories of recipients to whom the data might be disclosed; The registration of personal data processing systems operating in the 5. Proposed transfers of personal data outside the Philippines is required for PIC and PIPs, whether in the government or Philippines, if applicable; in the private sector, if: 6. General description of privacy and security measures for The processing involves sensitive personal information of data protection; at least one thousand (1,000) individuals; 7. Brief description of the data processing system; The PIC or PIP employs at least two hundred fifty (250) persons; or 8. Copy of all policies relating to data governance, data privacy, and information security; The PIC or PIP employs less than two hundred fifty (250) 9.
    [Show full text]
  • Data Processing, Analysis, and Dissemination*
    UNITED NATIONS SECRETARIAT ESA/STAT/AC.93/8 Statistics Division 03 November 2003 English only Expert Group Meeting to Review the Draft Handbook on Designing of Household Sample Surveys 3-5 December 2003 D R A F T Data processing, analysis, and dissemination* by Maphion Mungofa Jambwa ** * This document is being issued without formal editing. ** The views expressed in this paper are those of the author and do not imply the expression of any opinion on the part of the United Nations Secretariat. Table of contents Chapter Eight: Data processing, analysis, and dissemination....................................................... 3 8.1. Introduction..................................................................................................................... 3 8.2. A visual summary of the phases and processes of a household survey.......................... 3 8.3. Survey planning .............................................................................................................. 4 8.3.1. Survey objectives .......................................................................................................... 4 8.3.2. The tabulation and analysis plan................................................................................... 5 8.3.3. Design of the survey questionnaire............................................................................... 6 8.3.4. The conceptualization and general considerations for systems design for household surveys..................................................................................................................
    [Show full text]