R Data Import/Export

Total Page:16

File Type:pdf, Size:1020Kb

R Data Import/Export R Data Import/Export Version 4.1.1 Patched (2021-09-22) R Core Team This manual is for R, version 4.1.1 Patched (2021-09-22). Copyright c 2000{2021 R Core Team Permission is granted to make and distribute verbatim copies of this manual provided the copyright notice and this permission notice are preserved on all copies. Permission is granted to copy and distribute modified versions of this manual under the conditions for verbatim copying, provided that the entire resulting derived work is distributed under the terms of a permission notice identical to this one. Permission is granted to copy and distribute translations of this manual into an- other language, under the above conditions for modified versions, except that this permission notice may be stated in a translation approved by the R Core Team. i Table of Contents Acknowledgements ::::::::::::::::::::::::::::::::::::::::::::::::: 1 1 Introduction :::::::::::::::::::::::::::::::::::::::::::::::::::: 2 1.1 Imports:::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 2 1.1.1 Encodings :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 3 1.2 Export to text files::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 3 1.3 XML :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 5 2 Spreadsheet-like data :::::::::::::::::::::::::::::::::::::::::: 6 2.1 Variations on read.table:::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 6 2.2 Fixed-width-format files:::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 8 2.3 Data Interchange Format (DIF) :::::::::::::::::::::::::::::::::::::::::::::::::::: 8 2.4 Using scan directly :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 9 2.5 Re-shaping data :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 10 2.6 Flat contingency tables ::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 11 3 Importing from other statistical systems:::::::::::::::::::: 12 3.1 EpiInfo, Minitab, S-PLUS, SAS, SPSS, Stata, Systat :::::::::::::::::::::::::::::: 12 3.2 Octave ::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 13 4 Relational databases :::::::::::::::::::::::::::::::::::::::::: 14 4.1 Why use a database? ::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 14 4.2 Overview of RDBMSs::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 14 4.2.1 SQL queries :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 15 4.2.2 Data types::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 16 4.3 R interface packages :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 16 4.3.1 Packages using DBI :::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 17 4.3.2 Package RODBC::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 18 5 Binary files :::::::::::::::::::::::::::::::::::::::::::::::::::: 20 5.1 Binary data formats :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 20 5.2 dBase files (DBF)::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 20 6 Image files ::::::::::::::::::::::::::::::::::::::::::::::::::::: 21 7 Connections ::::::::::::::::::::::::::::::::::::::::::::::::::: 22 7.1 Types of connections:::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 22 7.2 Output to connections :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 23 7.3 Input from connections ::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 23 7.3.1 Pushback :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 24 7.4 Listing and manipulating connections ::::::::::::::::::::::::::::::::::::::::::::: 24 7.5 Binary connections ::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 24 7.5.1 Special values :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 25 ii 8 Network interfaces :::::::::::::::::::::::::::::::::::::::::::: 26 8.1 Reading from sockets ::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 26 8.2 Using download.file::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: 26 9 Reading Excel spreadsheets :::::::::::::::::::::::::::::::::: 27 Appendix A References ::::::::::::::::::::::::::::::::::::::::: 28 Function and variable index:::::::::::::::::::::::::::::::::::::: 29 Concept index::::::::::::::::::::::::::::::::::::::::::::::::::::: 31 1 Acknowledgements The relational databases part of this manual is based in part on an earlier manual by Douglas Bates and Saikat DebRoy. The principal author of this manual was Brian Ripley. Many volunteers have contributed to the packages used here. The principal authors of the packages mentioned are DBI (https://CRAN.R-project.org/package=DBI): David A. James dataframes2xls (https://CRAN.R-project.org/package=dataframes2xls): Guido van Steen foreign (https://CRAN.R-project.org/package=foreign): Thomas Lumley, Saikat DebRoy, Douglas Bates, Duncan Murdoch and Roger Bivand gdata (https://CRAN.R-project.org/package=gdata): Gregory R. Warnes ncdf4 (https://CRAN.R-project.org/package=ncdf4): David Pierce rJava (https://CRAN.R-project.org/package=rJava): Simon Urbanek RJDBC (https://CRAN.R-project.org/package=RJDBC): Simon Urbanek RMySQL (https://CRAN.R-project.org/package=RMySQL): David James and Saikat DebRoy RNetCDF (https://CRAN.R-project.org/package=RNetCDF): Pavel Michna RODBC (https://CRAN.R-project.org/package=RODBC): Michael Lapsley and Brian Ripley ROracle (https://CRAN.R-project.org/package=ROracle): David A. James RPostgreSQL (https://CRAN.R-project.org/package=RPostgreSQL): Sameer Kumar Prayaga and Tomoaki Nishiyama RSPerl: Duncan Temple Lang RSPython: Duncan Temple Lang RSQLite (https://CRAN.R-project.org/package=RSQLite): David A. James SJava: John Chambers and Duncan Temple Lang WriteXLS (https://CRAN.R-project.org/package=WriteXLS): Marc Schwartz XLConnect (https://CRAN.R-project.org/package=XLConnect): Mirai Solutions GmbH XML (https://CRAN.R-project.org/package=XML): Duncan Temple Lang Brian Ripley is the author of the support for connections. 2 1 Introduction Reading data into a statistical system for analysis and exporting the results to some other system for report writing can be frustrating tasks that can take far more time than the statistical analysis itself, even though most readers will find the latter far more appealing. This manual describes the import and export facilities available either in R itself or via packages which are available from CRAN or elsewhere. Unless otherwise stated, everything described in this manual is (at least in principle) available on all platforms running R. In general, statistical systems like R are not particularly well suited to manipulations of large-scale data. Some other systems are better than R at this, and part of the thrust of this manual is to suggest that rather than duplicating functionality in R we can make another system do the work! (For example Therneau & Grambsch (2000) commented that they preferred to do data manipulation in SAS and then use package survival (https://CRAN.R-project. org/package=survival) in S for the analysis.) Database manipulation systems are often very suitable for manipulating and extracting data: several packages to interact with DBMSs are discussed here. There are packages to allow functionality developed in languages such as Java, perl and python to be directly integrated with R code, making the use of facilities in these languages even more appropriate. (See the rJava (https://CRAN.R-project.org/package=rJava) package from CRAN.) It is also worth remembering that R like S comes from the Unix tradition of small re-usable tools, and it can be rewarding to use tools such as awk and perl to manipulate data before import or after export. The case study in Becker, Chambers & Wilks (1988, Chapter 9) is an example of this, where Unix tools were used to check and manipulate the data before input to S. The traditional Unix tools are now much more widely available, including for Windows. This manual was first written in 2000, and the number of scope of R packages has increased a hundredfold since. For specialist data formats it is worth searching to see if a suitable package already exists. 1.1 Imports The easiest form of data to import into R is a simple text file, and this will often be acceptable for problems of small or medium scale. The primary function to import from a text file is scan, and this underlies most of the more convenient functions discussed in Chapter 2 [Spreadsheet-like data], page 6. However, all statistical consultants are familiar with being presented by a client with a memory stick (formerly, a floppy disc or CD-R) of data in some proprietary binary format, for example `an Excel spreadsheet' or `an SPSS file’. Often the simplest thing to do is touse the originating application to export the data as a text file (and statistical consultants will have copies of the most common applications on their computers for that purpose). However, this is not always possible, and Chapter 3 [Importing from other statistical systems], page 12, discusses what facilities are available to access such files directly from R. For Excel spreadsheets, the available methods are summarized in Chapter 9 [Reading Excel spreadsheets], page 27. In a few cases, data have been stored in a binary form for compactness and speed of access. One application of this that we have seen several times is imaging data, which is normally stored as a stream of bytes as represented in memory, possibly preceded by a header. Such data formats
Recommended publications
  • A Definition of Database Design Standards for Human Rights Agencies
    A Definition of Database Design Standards for Human Rights Agencies by Patrick Ball, et al. AMERICAN ASSOCIATION FOR THE ADVANCEMENT OF SCIENCE 1994 TABLE OF CONTENTS Introduction Definition of an Entity Entity versus Role versus Link What is a Link? Presentation of a Data Model Rule-types with Rule Examples and Instance Examples . 6.1 Type of Rule: Act . 6.2 Type of Rule: Relationship . 6.3 Rule-type Name: Biography Rule Parsimony: the Trade between Precision and Simplicity xBase Table Design Implementation Example . 8.1 xBase Implementation Example . 8.2 Abstract Fields & Data Types SQL Database Design Implementation Example . 9.1 Querying & Performance . 9.2 Main Difference . 9.3 Storage of the Rules in SQL Model 1. Introduction In 1993, HURIDOCS published their Standard Human Rights Event Formats (Dueck et al. 1993a) which describe standards for the collection and exchange of information on human rights abuses. These formats represent a major step forward in enabling human rights organizations to develop manual and computerized systems for collecting and exchanging data. The formats define common fields for collection of human rights event, victim, source, perpetrator and agency intervention data, and a common vocabulary for many of the fields in the formats, for example occupation, type of event and geographical location. The formats are designed as a tool leading toward both manual and computerized systems of human rights violation documentation. Before organizations implement documentation systems which will meet their needs, a wide range of issues must be considered. One of these problems is the structural problems of some data having complex relations to other data.
    [Show full text]
  • Xbase++ Language Concepts for Newbies Geek Gatherings Roger Donnay
    Xbase++ Language Concepts for Newbies Geek Gatherings Roger Donnay Introduction Xbase++ has extended the capabilities of the language beyond what is available in FoxPro and Clipper. For FoxPro developers and Clipper developers who are new to Xbase++, there are new variable types and language concepts that can enhance the programmer's ability to create more powerful and more supportable applications. The flexibility of the Xbase++ language is what makes it possible to create libraries of functions that can be used dynamically across multiple applications. The preprocessor, code blocks, ragged arrays and objects combine to give the programmer the ability to create their own language of commands and functions and all the advantages of a 4th generation language. This session will also show how these language concepts can be employed to create 3rd party add-on products to Xbase++ that will integrate seamlessly into Xbase++ applications. The Xbase++ language in incredibly robust and it could take years to understand most of its capabilities, however when migrating Clipper and FoxPro applications, it is not necessary to know all of this. I have aided many Clipper and FoxPro developers with the migration process over the years and I have found that only a basic introduction to the following concepts are necessary to get off to a great start: * The Xbase++ Project. Creation of EXEs and DLLs. * The compiler, linker and project builder . * Console mode for quick migration of Clipper and Fox 2.6 apps. * INIT and EXIT procedures, DBESYS, APPSYS and MAIN. * The DBE (Database engine) * LOCALS, STATICS, PRIVATE and PUBLIC variables. * STATIC functions.
    [Show full text]
  • Copy on Write Based File Systems Performance Analysis and Implementation
    Copy On Write Based File Systems Performance Analysis And Implementation Sakis Kasampalis Kongens Lyngby 2010 IMM-MSC-2010-63 Technical University of Denmark Department Of Informatics Building 321, DK-2800 Kongens Lyngby, Denmark Phone +45 45253351, Fax +45 45882673 [email protected] www.imm.dtu.dk Abstract In this work I am focusing on Copy On Write based file systems. Copy On Write is used on modern file systems for providing (1) metadata and data consistency using transactional semantics, (2) cheap and instant backups using snapshots and clones. This thesis is divided into two main parts. The first part focuses on the design and performance of Copy On Write based file systems. Recent efforts aiming at creating a Copy On Write based file system are ZFS, Btrfs, ext3cow, Hammer, and LLFS. My work focuses only on ZFS and Btrfs, since they support the most advanced features. The main goals of ZFS and Btrfs are to offer a scalable, fault tolerant, and easy to administrate file system. I evaluate the performance and scalability of ZFS and Btrfs. The evaluation includes studying their design and testing their performance and scalability against a set of recommended file system benchmarks. Most computers are already based on multi-core and multiple processor architec- tures. Because of that, the need for using concurrent programming models has increased. Transactions can be very helpful for supporting concurrent program- ming models, which ensure that system updates are consistent. Unfortunately, the majority of operating systems and file systems either do not support trans- actions at all, or they simply do not expose them to the users.
    [Show full text]
  • The Linux Kernel Module Programming Guide
    The Linux Kernel Module Programming Guide Peter Jay Salzman Michael Burian Ori Pomerantz Copyright © 2001 Peter Jay Salzman 2007−05−18 ver 2.6.4 The Linux Kernel Module Programming Guide is a free book; you may reproduce and/or modify it under the terms of the Open Software License, version 1.1. You can obtain a copy of this license at http://opensource.org/licenses/osl.php. This book is distributed in the hope it will be useful, but without any warranty, without even the implied warranty of merchantability or fitness for a particular purpose. The author encourages wide distribution of this book for personal or commercial use, provided the above copyright notice remains intact and the method adheres to the provisions of the Open Software License. In summary, you may copy and distribute this book free of charge or for a profit. No explicit permission is required from the author for reproduction of this book in any medium, physical or electronic. Derivative works and translations of this document must be placed under the Open Software License, and the original copyright notice must remain intact. If you have contributed new material to this book, you must make the material and source code available for your revisions. Please make revisions and updates available directly to the document maintainer, Peter Jay Salzman <[email protected]>. This will allow for the merging of updates and provide consistent revisions to the Linux community. If you publish or distribute this book commercially, donations, royalties, and/or printed copies are greatly appreciated by the author and the Linux Documentation Project (LDP).
    [Show full text]
  • Use of Seek When Writing Or Reading Binary Files
    Title stata.com file — Read and write text and binary files Description Syntax Options Remarks and examples Stored results Reference Also see Description file is a programmer’s command and should not be confused with import delimited (see [D] import delimited), infile (see[ D] infile (free format) or[ D] infile (fixed format)), and infix (see[ D] infix (fixed format)), which are the usual ways that data are brought into Stata. file allows programmers to read and write both text and binary files, so file could be used to write a program to input data in some complicated situation, but that would be an arduous undertaking. Files are referred to by a file handle. When you open a file, you specify the file handle that you want to use; for example, in . file open myfile using example.txt, write myfile is the file handle for the file named example.txt. From that point on, you refer to the file by its handle. Thus . file write myfile "this is a test" _n would write the line “this is a test” (without the quotes) followed by a new line into the file, and . file close myfile would then close the file. You may have multiple files open at the same time, and you may access them in any order. 1 2 file — Read and write text and binary files Syntax Open file file open handle using filename , read j write j read write text j binary replace j append all Read file file read handle specs Write to file file write handle specs Change current location in file file seek handle query j tof j eof j # Set byte order of binary file file set handle byteorder hilo j lohi j 1 j 2 Close
    [Show full text]
  • File Storage Techniques in Labview
    File Storage Techniques in LabVIEW Starting with a set of data as if it were generated by a daq card reading two channels and 10 samples per channel, we end up with the following array: Note that the first radix is the channel increment, and the second radix is the sample number. We will use this data set for all the following examples. The first option is to send it directly to a spreadsheet file. spreadsheet save of 2D data.png The data is wired to the 2D array input and all the defaults are taken. This will ask for a file name when the program block is run, and create a file with data values, separated by tab characters, as follows: 1.000 2.000 3.000 4.000 5.000 6.000 7.000 8.000 9.000 10.000 10.000 20.000 30.000 40.000 50.000 60.000 70.000 80.000 90.000 100.000 Note that each value is in the format x.yyy, with the y's being zeros. The default format for the write to spreadsheet VI is "%.3f" which will generate a floating point number with 3 decimal places. If a number with higher decimal places is entered in the array, it would be truncated to three. Since the data file is created in row format, and what you really need if you are going to import it into excel, is column format. There are two ways to resolve this, the first is to set the transpose bit on the write function, and the second, is to add an array transpose, located in the array pallet.
    [Show full text]
  • System Calls System Calls
    System calls We will investigate several issues related to system calls. Read chapter 12 of the book Linux system call categories file management process management error handling note that these categories are loosely defined and much is behind included, e.g. communication. Why? 1 System calls File management system call hierarchy you may not see some topics as part of “file management”, e.g., sockets 2 System calls Process management system call hierarchy 3 System calls Error handling hierarchy 4 Error Handling Anything can fail! System calls are no exception Try to read a file that does not exist! Error number: errno every process contains a global variable errno errno is set to 0 when process is created when error occurs errno is set to a specific code associated with the error cause trying to open file that does not exist sets errno to 2 5 Error Handling error constants are defined in errno.h here are the first few of errno.h on OS X 10.6.4 #define EPERM 1 /* Operation not permitted */ #define ENOENT 2 /* No such file or directory */ #define ESRCH 3 /* No such process */ #define EINTR 4 /* Interrupted system call */ #define EIO 5 /* Input/output error */ #define ENXIO 6 /* Device not configured */ #define E2BIG 7 /* Argument list too long */ #define ENOEXEC 8 /* Exec format error */ #define EBADF 9 /* Bad file descriptor */ #define ECHILD 10 /* No child processes */ #define EDEADLK 11 /* Resource deadlock avoided */ 6 Error Handling common mistake for displaying errno from Linux errno man page: 7 Error Handling Description of the perror () system call.
    [Show full text]
  • Course Description
    6/20/2018 Syllabus Syllabus This is a single, concatenated file, suitable for printing or saving as a PDF for offline viewing. Please note that some animations or images may not work. Course Description This module is also available as a concatenated page, suitable for printing or saving as a PDF for offline viewing. MET CS669 Database Design and Implementation for Business This course uses the latest database tools and techniques for persistent data and object-modeling and management. Students gain extensive hands-on experience with exercises and a term project using Oracle, SQL Server, and other leading database management systems. Students learn to model persistent data using the standard Entity-Relationship model (ERM) and how to diagram those models using Entity-Relationship Diagrams (ERDs), Extended Entity-Relationship Diagrams (EERDs), and UML diagrams. Students learn the standards-based Structured Query Language (SQL) and the extensions to the SQL standards implemented in Oracle and SQL Server. Students learn the basics of database programming, and write simple stored procedures and triggers. The Role of this Course in the MSCIS Online Curriculum This is a core course in the MSCIS online curriculum. It provides students with an understanding and experience with database technology, database design, SQL, and the roles of databases in enterprises. This course is a prerequisite for the three additional database courses in the MSCIS online curriculum, which are CS674 Database Security, CS699 Data Mining and Business Intelligence and CS779 Advanced Database Management. By taking these three courses you can obtain the Concentration in Database Management and Business Intelligence. CS674 Database Security also satisfies an elective requirement for the Concentration in Security.
    [Show full text]
  • Xbase PDF Class
    Xbase++ PDF Class User guide 2020 Created by Softsupply Informatica Rua Alagoas, 48 01242-000 São Paulo, SP Brazil Tel (5511) 3159-1997 Email : [email protected] Contact : Edgar Borger Table of Contents Overview ....................................................................................................................................................4 Installing ...............................................................................................................................................5 Changes ...............................................................................................................................................6 Acknowledgements ............................................................................................................................10 Demo ..................................................................................................................................................11 GraDemo ............................................................................................................................................13 Charts .................................................................................................................................................15 Notes ..................................................................................................................................................16 Class Methods .........................................................................................................................................17
    [Show full text]
  • Design and Development of a Thermodynamic Properties Database Using the Relational Data Model
    DESIGN AND DEVELOPMENT OF A THERMODYNAMIC PROPERTIES DATABASE USING THE RELATIONAL DATA MODEL By PARIKSHIT SANGHAVI Bachelor ofEngineering Birla Institute ofTechnology and Science, Pilani, India 1992 Submitted to the Faculty ofthe Graduate College ofthe Oklahoma State University in partial fulfillment of the requirements for the Degree of MASTER OF SCIENCE December, 1995 DESIGN AND DEVELOPMENT OF A THERMODYNAMIC PROPERTIES DATABASE USING THE RELATIONAL DATA MODEL Thesis Approved: ~ (}.J -IJ trT,J T~ K t 4J~ _ K1t'Il1·B~~ Dean ofthe Graduate College ii ACKNOWLEDGMENTS I would like to thank my adviser Dr. Jan Wagner, for his expert guidance and criticism towards the design and implementation ofthe GPA Database. I would like to express my appreciation to Dr. Martin S. High, for his encouragement and support on the GPA Database project. I am grateful to Dr. Khaled Gasem, for helping me gain a better understanding of the aspects pertaining to thermodynamics in.the GPA Database project. I would also like to thank Dr. James R. Whiteley for taking the time to provide constructive criticism for this thesis as a member ofmy thesis committee. I would like to thank the Enthalpy and Phase Equilibria Steering Committees of the Gas Processors Association for their support. The GPA Database project would not have been possible without their confidence in the faculty and graduate research assistants at Oklahoma State University. I wish to acknowledge Abhishek Rastogi, Heather Collins, C. S. Krishnan, Srikant, Nhi Lu, and Eric Maase for working with me on the GPA Database. I am grateful for the emotional support provided by my family.
    [Show full text]
  • Ext4 File System and Crash Consistency
    1 Ext4 file system and crash consistency Changwoo Min 2 Summary of last lectures • Tools: building, exploring, and debugging Linux kernel • Core kernel infrastructure • Process management & scheduling • Interrupt & interrupt handler • Kernel synchronization • Memory management • Virtual file system • Page cache and page fault 3 Today: ext4 file system and crash consistency • File system in Linux kernel • Design considerations of a file system • History of file system • On-disk structure of Ext4 • File operations • Crash consistency 4 File system in Linux kernel User space application (ex: cp) User-space Syscalls: open, read, write, etc. Kernel-space VFS: Virtual File System Filesystems ext4 FAT32 JFFS2 Block layer Hardware Embedded Hard disk USB drive flash 5 What is a file system fundamentally? int main(int argc, char *argv[]) { int fd; char buffer[4096]; struct stat_buf; DIR *dir; struct dirent *entry; /* 1. Path name -> inode mapping */ fd = open("/home/lkp/hello.c" , O_RDONLY); /* 2. File offset -> disk block address mapping */ pread(fd, buffer, sizeof(buffer), 0); /* 3. File meta data operation */ fstat(fd, &stat_buf); printf("file size = %d\n", stat_buf.st_size); /* 4. Directory operation */ dir = opendir("/home"); entry = readdir(dir); printf("dir = %s\n", entry->d_name); return 0; } 6 Why do we care EXT4 file system? • Most widely-deployed file system • Default file system of major Linux distributions • File system used in Google data center • Default file system of Android kernel • Follows the traditional file system design 7 History of file system design 8 UFS (Unix File System) • The original UNIX file system • Design by Dennis Ritche and Ken Thompson (1974) • The first Linux file system (ext) and Minix FS has a similar layout 9 UFS (Unix File System) • Performance problem of UFS (and the first Linux file system) • Especially, long seek time between an inode and data block 10 FFS (Fast File System) • The file system of BSD UNIX • Designed by Marshall Kirk McKusick, et al.
    [Show full text]
  • File Handling in Python
    hapter C File Handling in 2 Python There are many ways of trying to understand programs. People often rely too much on one way, which is called "debugging" and consists of running a partly- understood program to see if it does what you expected. Another way, which ML advocates, is to install some means of understanding in the very programs themselves. — Robin Milner In this Chapter » Introduction to Files » Types of Files » Opening and Closing a 2.1 INTRODUCTION TO FILES Text File We have so far created programs in Python that » Writing to a Text File accept the input, manipulate it and display the » Reading from a Text File output. But that output is available only during » Setting Offsets in a File execution of the program and input is to be entered through the keyboard. This is because the » Creating and Traversing a variables used in a program have a lifetime that Text File lasts till the time the program is under execution. » The Pickle Module What if we want to store the data that were input as well as the generated output permanently so that we can reuse it later? Usually, organisations would want to permanently store information about employees, inventory, sales, etc. to avoid repetitive tasks of entering the same data. Hence, data are stored permanently on secondary storage devices for reusability. We store Python programs written in script mode with a .py extension. Each program is stored on the secondary device as a file. Likewise, the data entered, and the output can be stored permanently into a file.
    [Show full text]