Cleaning Data A

Total Page:16

File Type:pdf, Size:1020Kb

Cleaning Data A Dealing with messy data Cleaning data A. The data cleaning process It is mandatory for the overall quality of an Data cleaning deals mainly with data problems assessment to ensure that its primary and once they have occurred. Error-prevention secondary data be of sufficient quality. “Messy strategies (see data quality control procedures data” refers to data that is riddled with later in the document) can reduce many problems inconsistencies, because of human error, poorly but cannot eliminate them. Many data errors are designed recording systems, or simply because detected incidentally during activities other than there is incomplete control over the format and data cleaning, i.e.: type of data imported from external data sources, When collecting or entering data such as a database, text file, or a Web page. So, When transforming/extracting/transferring a column that contains country names may data contain “Burma”, “Myanmar” or “Myanma”. When exploring or analysing data When submitting the draft report to peer Such inconsistencies will impede the data review processing. Care should be taken to ensure data is as accurate and consistent (i.e. spellings, to It is more efficient to detect errors by actively allow aggregation) as possible. Inconsistencies searching for them in a planned way. Data can wreak havoc when trying to perform analysis cleaning involves repeated cycles of screening, with the data, so they have to be addressed diagnosing, and treatment. before starting the analysis. Used mainly when dealing with large volumes of • Lack of data Diagnosis • Leave unchanged • Excess of data • Correction • Missing data data stored in a database, the terms data • Outliers or • Deletion insconsistencies • Errors • Strange patterns • True extremes cleansing, data cleaning or data scrubbing refer • Suspect analysis results • True normal to the process of detecting, correcting, replacing, • No diagnosis, still suspect Treatment modifying or removing incomplete, incorrect, Screening irrelevant, corrupt or inaccurate records from a record set, table, or database. Adapted from Van den Broeck J, Argeseanu Cunningham S, Eeckels R, Herbst K (2005) This document provides guidance for data analysts to find the right data cleaning strategy Screening involves systematically looking for when dealing with needs assessment data, either suspect features in assessment questionnaires, primary or secondary. It covers situations where: databases, or analysis datasets (in small Raw data is being produced by assessment assessments, with the analysts closely involved teams using a questionnaire and is entered at all stages, there may be little or no distinction into a centralized database. between a database and an analysis dataset). Data is obtained from secondary sources (displacement monitoring system, food The diagnostic (identifying the nature of the security data, census data, etc.) and is defective data) and treatment (deleting, editing or integrated, compared or merged with the data leaving the data as it is) phases of data cleaning obtained from field assessment to require insight into the sources and types of complement the final analysis. errors at all stages of the assessment. After measurement, data are the object of a sequence This document complements the ACAPS of typical activities: they are entered into technical note on How to approach a dataset databases, extracted, transferred to other tables, which specifically details data cleaning edited, selected, transformed, summarized, and operations for primary data entered into an Excel presented. It is important to realize that errors can spreadsheet during rapid assessments. occur at any stage of the data flow, including during data cleaning itself. 1 Dealing with messy data B. Sources of errors An erroneous entry happens if, e.g., age is mistyped as 26 instead of 25. Many of the sources of error in databases fall into Extraneous entries add correct, but unwanted one or more of the following categories: information, e.g. name and title in a name-only field. Measurement errors: Data is generally intended Incorrectly derived value occurs when a to measure some physical process, subjects or function was incorrectly calculated for a objects, i.e. the waiting time at the water point, the derived field (i.e. error in the age derived from size of a population, the incidence of diseases, the date of birth). etc. In some cases these measurements are Inconsistencies across tables or files occur undertaken by human processes that can have e.g. when the number of visited sites in the systematic or random errors in their design (i.e., province table and the number of visited sites improper sampling strategies) and execution (i.e., in the total sample table do not match. misuse of instruments, bias, etc.). Identifying and solving such inconsistencies goes beyond the Processing errors: In many settings, raw data are scope of this document. It is recommended to pre-processed before they are entered into a refer to the ACAPS Technical Brief How sure are database. This data processing is done for a you? To get an empirical understanding of how to variety of reasons: to reduce the complexity or deal with measurement errors in general. noise in the raw data, to emphasize aggregate properties of the raw data (often with some Data entry error: "Data entry" is the process of editorial bias), and in some cases simply to transferring information from the medium that reduce the volume of data being stored. All these records the response (traditionally answers processes have the potential to produce errors. written on printed questionnaires) to a computer application. Data entry is generally done by Data integration errors: It is actually quite rare for humans, who typically extract information from a database of significant size and age to contain speech (i.e., key informant interviews) or by using data from a single source, collected and entered secondary data from written or printed sources in the same way over time. Very often, a (i.e. health statistics from health centres). Under database contains information collected from time pressure, or for lack of proper supervision, multiple sources via multiple methods over time data is often corrupted at entry time. Main errors (i.e. tracking of affected population numbers over type include: the crisis, where the definition of “affected” is being refined or changed over time). Moreover, in practice, many databases evolve by merging in other pre-existing databases; this merging task almost always requires some attempt to resolve inconsistencies across the databases involving different data units, measurement periods, formats, and so on. Any procedure that integrates data from multiple sources can lead to errors. The merging of two or more databases will both identify errors (where there are differences between the two databases) and create new errors (i.e. duplicate records). Table 1 below illustrates some of the possible sources and types of errors in a large assessment, at three basic levels: When filling the questionnaire, when entering data into the Adapted from Kim et Al, 2003; Aldo Benini 2013 database and when performing the analysis. 2 Dealing with messy data Table 1: Sources of data error C. First things first Sources of problems Stage Lack or excess of Outliers and The first thing to do is to make a copy of the data inconsistencies original data in a separate workbook and name Questionnaire Form missing Correct value filled out the sheets appropriately, or save in a new file. Form double, in the wrong box ALWAYS keep the source files in a separate collected repeatedly Not readable Answering box or Writing error folder and change its attribute to READ-ONLY, to options left blank Answer given is out of avoid modifying any of those files, even if it is More than one expected (conditional) option selected range opened for reference. when not allowed Database Lack or excess of Outliers and data transferred from inconsistencies carried D. Screening data the questionnaire over from Form of field not questionnaire entered Value incorrectly No matter how data are collected (in face-to-face Value entered in entered, misspelling wrong field Value incorrectly interviews, telephone interviews, self- Inadvertent deletion changed during administered questionnaires, etc.), there will be and duplication previous data cleaning during database Transformation some level of error, including a number of handling (programming) error inconsistencies. While some of these will be Analysis Lack or excess of Outliers and legitimate, reflecting variation in the context, data extracted from inconsistencies carried the database over from the others will likely reflect a data collection error. Data extraction, database coding or transfer Data extraction, coding error or transfer error Examine data for the following possible errors: Deletions or Sorting errors Spelling and formatting irregularities: are they duplications by (spreadsheets) analyst Data-cleaning errors categorical variables written incorrectly? Are date format consistent? Etc. Adapted from Van den Broeck J, Argeseanu Cunningham S, Eeckels R, Herbst K (2005) Lack of data: Do some questions have far fewer answers than surrounding questions? Inaccuracy of a single measurement and data Excess of data: Are there duplicate entries? point may be acceptable, and related to the Are there more answers than originally inherent technical error of the measurement allowed? instrument. Hence, data cleaning
Recommended publications
  • An Analysis of Data Corruption in the Storage Stack
    An Analysis of Data Corruption in the Storage Stack Lakshmi N. Bairavasundaram∗, Garth R. Goodson†, Bianca Schroeder‡ Andrea C. Arpaci-Dusseau∗, Remzi H. Arpaci-Dusseau∗ ∗University of Wisconsin-Madison †Network Appliance, Inc. ‡University of Toronto {laksh, dusseau, remzi}@cs.wisc.edu, [email protected], [email protected] Abstract latent sector errors, within disk drives [18]. Latent sector errors are detected by a drive’s internal error-correcting An important threat to reliable storage of data is silent codes (ECC) and are reported to the storage system. data corruption. In order to develop suitable protection Less well-known, however, is that current hard drives mechanisms against data corruption, it is essential to un- and controllers consist of hundreds-of-thousandsof lines derstand its characteristics. In this paper, we present the of low-level firmware code. This firmware code, along first large-scale study of data corruption. We analyze cor- with higher-level system software, has the potential for ruption instances recorded in production storage systems harboring bugs that can cause a more insidious type of containing a total of 1.53 million disk drives, over a pe- disk error – silent data corruption, where the data is riod of 41 months. We study three classes of corruption: silently corrupted with no indication from the drive that checksum mismatches, identity discrepancies, and par- an error has occurred. ity inconsistencies. We focus on checksum mismatches since they occur the most. Silent data corruptionscould lead to data loss more of- We find more than 400,000 instances of checksum ten than latent sector errors, since, unlike latent sector er- mismatches over the 41-month period.
    [Show full text]
  • Chapter 4 Research Methodology
    The impact of E-marketing practices on market performance of small business enterprises. An empirical investigation. Item Type Thesis Authors El-Gohary, Hatem O.A.S. Rights <a rel="license" href="http://creativecommons.org/licenses/ by-nc-nd/3.0/"><img alt="Creative Commons License" style="border-width:0" src="http://i.creativecommons.org/l/by- nc-nd/3.0/88x31.png" /></a><br />The University of Bradford theses are licenced under a <a rel="license" href="http:// creativecommons.org/licenses/by-nc-nd/3.0/">Creative Commons Licence</a>. Download date 26/09/2021 11:58:39 Link to Item http://hdl.handle.net/10454/4308 Chapter 4 Research Methodology 4.1 Introduction:- This chapter discusses the methodology used to conduct the research study as well as the issues related to the chosen research methodology when investigating the different factors affecting the adoption of E-Marketing by industrial and trading UK SBEs, the different forms, tools and levels of implementation of E-Marketing by these enterprises as well as the relationship between E-Marketing adoption and marketing performance of these SBEs. Through this discussion the steps that were taken to address the research design, the data collection and analysis methods used to conduct the research study will be illustrated. These issues are addressed in light of the basic research objectives and the relevant research questions discussed in chapter one and the research framework discussed in chapter seven. Firstly it discusses some possible research designs and methodologies and provides a brief review of the literature of research methodologies, research strategies and research methods implemented in the fields of E-Marketing and SBEs.
    [Show full text]
  • Evaluation Design Report for the National Cancer Institute's
    February 2009 Evaluation Design Report for the National Cancer Institute’s Community Cancer Centers Program (NCCCP) Final Report Prepared for Steve Clauser, PhD Chief, Outcomes Research Branch Applied Research Program Division of Cancer Control and Population Sciences National Cancer Institute Executive Plaza North, Room 4086 6130 Executive Boulevard, (MSC 7344) Bethesda, MD 20892-7344 Prepared by Debra J. Holden, PhD Kelly J. Devers, PhD Lauren McCormack, PhD Kathleen Dalton, PhD Sonya Green, MPH Katherine Treiman, PhD RTI International 3040 Cornwallis Road Research Triangle Park, NC 27709 RTI Project Number 0210903.000.005 RTI Project Number 0210903.0001.005 Evaluation Design Report for the National Cancer Institute’s Community Cancer Centers Program (NCCCP) Final Report February 2009 Prepared for Steve Clauser, PhD Chief, Outcomes Research Branch Applied Research Program Division of Cancer Control and Population Sciences National Cancer Institute Executive Plaza North, Room 4086 6130 Executive Boulevard, (MSC 7344) Bethesda, MD 20892-7344 Prepared by Debra J. Holden, PhD Kelly J. Devers, PhD Lauren McCormack, PhD Kathleen Dalton, PhD Sonya Green, MPH Katherine Treiman, PhD RTI International 3040 Cornwallis Road Research Triangle Park, NC 27709 Contents Section Page Acronyms Error! Bookmark not defined. Executive Summary ES-Error! Bookmark not defined. 1. Introduction Error! Bookmark not defined. 1.1 Overview of the NCCCP Error! Bookmark not defined. 1.2 Summary of the NCCCP Sites Error! Bookmark not defined. 1.3 Process Completed for an Evaluability Assessment Error! Bookmark not defined. 1.3.1 Step 1: Pilot Site Evaluability Assessment (September 2007–June 2008) .................................................. Error! Bookmark not defined. 1.3.2 Step 2: Engage Stakeholders (September 2007–September 2008)Error! Bookmark not defined.
    [Show full text]
  • Methods Used in the Development of the Global Roads Open Access Data Set (Groads), Version 1
    Methods Used in the Development of the Global Roads Open Access Data Set (gROADS), Version 1 Table of Contents Introduction ................................................................................................................................................................... 2 Data Assessment and Selection ..................................................................................................................................... 2 Root Mean Square Error (RMSE) ............................................................................................................................... 2 Validation against Google Earth (GE) imagery .......................................................................................................... 4 Calculating total length of roads ............................................................................................................................... 6 Choosing a data set for further edits ........................................................................................................................ 8 Data ingest: Creating and populating geodatabases ..................................................................................................... 8 Data editing ................................................................................................................................................................... 9 Creating a topology ..................................................................................................................................................
    [Show full text]
  • DHS Data Editing and Imputation Trevor Croft
    DHS Data Editing and Imputation Trevor Croft, Demographic and Health Surveys Abstract Producing good quality demographic and health data and making it accessible to data users worldwide is one of the main aims of the Demographic and Health Surveys program. However, large scale surveys in developing countries, particularly those collecting retrospective data, are prone to poor reporting. Survey data have suffered traditionally from incomplete and inconsistent reporting, To handle these problems, DHS performed extensive data editing operations. In addition, imputation procedures were established to deal with partial reporting of dates of key events in the respondent's life. General techniques for handling incomplete and inconsistent data are considered. The paper then presents the DHS approach to data editing. The major focus is on the editing of dates of events and the intervals between events. The editing and imputation process starts with the calculation of initial logical ranges for each date, and gradually constrains these ranges to produce final logical ranges. Inconsistent data are reported in error listings during this process. Dates are imputed for events with incomplete reporting within these final logical ranges. The levels of imputation required in the DHS-I surveys are presented. Various problem areas involved with the imputation of incomplete dates are explained. These include biases caused by questionnaire design, miscalculation of dates by interviewers and ancillary data biases. Problems relating to fine temporal variables and to unconstrained ranges for dates are also reviewed. Finally, the changes introduced as part of the editing and imputation procedures for DHS- II to resolve some of the problem areas are presented and ideas for further improvements are discussed.
    [Show full text]
  • Data Management, Analysis Tools, and Analysis Mechanics
    Chapter 2 Data Management, Analysis Tools, and Analysis Mechanics This chapter explores different tools and techniques for handling data for research purposes. This chapter assumes that a research problem statement has been formulated, research hypotheses have been stated, data collection planning has been conducted, and data have been collected from various sources (see Volume I for information and details on these phases of research). This chapter discusses how to combine and manage data streams, and how to use data management tools to produce analytical results that are error free and reproducible, once useful data have been obtained to accomplish the overall research goals and objectives. Purpose of Data Management Proper data handling and management is crucial to the success and reproducibility of a statistical analysis. Selection of the appropriate tools and efficient use of these tools can save the researcher numerous hours, and allow other researchers to leverage the products of their work. In addition, as the size of databases in transportation continue to grow, it is becoming increasingly important to invest resources into the management of these data. There are a number of ancillary steps that need to be performed both before and after statistical analysis of data. For example, a database composed of different data streams needs to be matched and integrated into a single database for analysis. In addition, in some cases data must be transformed into the preferred electronic format for a variety of statistical packages. Sometimes, data obtained from “the field” must be cleaned and debugged for input and measurement errors, and reformatted. The following sections discuss considerations for developing an overall data collection, handling, and management plan, and tools necessary for successful implementation of that plan.
    [Show full text]
  • An Introduction to the Data Editing Process
    1 AN INTRODUCTION TO THE DATA EDITING PROCESS by Dania P. Ferguson United States Department of Agriculture National Agricultural Statistics Service Abstract: A primer on the various data editing methodologies, the impact of their usage, available supporting software, and considerations when developing new software. I. INTRODUCTION The intention of this paper is to promote a better understanding of the various data editing methods and the impact of their usage as well as the software available to support the use of the various methodologies. It is hoped that this paper will serve as: - a brief overview of the most commonly used editing methodologies, - an aid to organize further reading about data editing systems, - a general description of the data editing process for use by designers and developers of generalized data editing systems. II. REVIEW OF THE PROCESS Data editing is defined as the process involving the review and adjustment of collected survey data. The purpose is to control the quality of the collected data. This process is divided into four (4) major sub-process areas. These areas are: - Survey Management - Data Capture - Data Review - Data Adjustment The rest of this section will describe each of the four (4) sub-processes. 1. Survey Management The survey management functions are: a) completeness checking, and b) quality control including audit trails and the gathering of cost data. These functions are administrative in nature. Completeness checking occurs at both survey and questionnaire levels. 2 At survey level, completeness checking ensures that all survey data have been collected. It is vitally important to account for all samples because sample counts are used in the data expansion procedures that take place during Summary.
    [Show full text]
  • Active Storage Activeraid ES Data Sheet
    RAID Storage Evolved Welcome to RAID Storage Evolved Key Features Introducing the ActiveRAID ES — the affordable high performance RAID storage solution Optimized RAID performance in Apple Mac for Apple® users. The ActiveRAID ES was designed using the same principles of the OS X® Applications original ActiveRAID, but represents a previously unobtainable level of value by providing the capabilities of the ActiveRAID in a lower cost configuration that does not sacrifice Modern Linux RAID kernel and RAID the performance, reliability or ease of management of the original ActiveRAID. The controller design ActiveRAID ES was carefully designed for use in business critical applications where Proprietary self-optimizing architecture organizations require a high level of data integrity with ease of installation and management, but not the 24/7 365 day duty cycles of a large enterprise or Post-Production or Broadcast facility. Native Mac OS X Storage management suite The ActiveRAID ES provides redundant power, redundant cooling and redundant Fibre Dashboard Widget monitoring Channel connections, and the same powerful, yet easy to use management suite of the original ActiveRAID. The ES differs from the original ActiveRAID in that it has a single high Enterprise performance and reliability performance RAID controller and uses business class hard drives with a smaller cache size. without Enterprise complexity This results in a very robust RAID solution ideal for business class applications. Performance No long Quick Start guides to memorize of the ActiveRAID ES is impressive at up to 774 MB/s*† throughput and will easily handle Performance profiling and statistical most Mac OS X Server applications. The ActiveRAID ES is the ideal storage for business planning tools class applications such as File serving, database services, web hosting, calendar, Wiki, Mail, iChat and much more.
    [Show full text]
  • Handbook on Precision Requirements and Variance Estimation for ESS Households Surveys
    ISSN 1977-0375 Methodologies and Working papers Handbook on precision requirements and variance estimation for ESS households surveys 2013 edition Methodologies and Working papers Handbook on precision requirements and variance estimation for ESS household surveys 2013 edition Europe Direct is a service to help you find answers to your questions about the European Union. Freephone number (*): 00 800 6 7 8 9 10 11 (*)0H The information given is free, as are most calls (though some operators, phone boxes or hotels may charge you). More information on the European Union is available on the Internet (http://europa.eu). Cataloguing data can be found at the end of this publication. Luxembourg: Publications Office of the European Union, 2013 ISBN 978-92-79-31197-0 ISSN 1977-0375 doi:10.2785/13579 Cat. No: KS-RA-13-029-EN-N Theme: General and regional statistics Collection: Methodologies & Working papers © European Union, 2013 Reproduction is authorised provided the source is acknowledged. Acknowledgments Acknowledgments The European Commission expresses its gratitude and appreciation to the following members of the ESS1 Task Force, for their work on Precision Requirements and Variance Estimation for Household Surveys: Experts from the European Statistical System (ESS): Martin Axelson Sweden — Statistics Sweden Loredana Di Consiglio Italy — ISTAT Kari Djerf Finland — Statistics Finland Stefano Falorsi Italy — ISTAT Alexander Kowarik Austria — Statistics Austria Mārtiņš Liberts Latvia — CSB Ioannis Nikolaidis Greece — EL.STAT Experts from European
    [Show full text]
  • Algorithms for Data Cleaning in Knowledge Bases
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by VFAST - Virtual Foundation for Advancement of Science and Technology (Pakistan) VFAST Transactions on Software Engineering http://vfast.org/journals/index.php/VTSE@ 2018, ISSN(e): 2309-6519; ISSN(p): 2411-6327 Volume 15, Number 2, May-August, 2018 pp. 78-83 ALGORITHMS FOR DATA CLEANING IN KNOWLEDGE BASES 1 1 1 ADEEL ASHRAF , SARAH ILYAS , KHAWAJA UBAID UR REHMAN AND SHAKEEL AHMAD2 ` 1Department of Computer Science, University of Management and Technology, Punjab, Pakistan 2Department of Computer Sciences, Faculty of Computing and Information Technology in rabigh, King Abdulaziz University Jeddah Email: [email protected] ABSTRACT. Data cleaning is an action which includes a process of correcting and identifying the inconsistencies and errors in data warehouse. Different terms are uses in these papers like data cleaning also called data scrubbing. Using data scrubbing to get high quality data and this is one the data ETL (extraction transformation and loading tools). Now a day there is a need of authentic information for better decision-making. So we conduct a review paper in which six papers are reviewed related to data cleaning. Relating papers discussed different algorithms, methods, problems, their solutions and approaches etc. Each paper has their own methods to solve a problem in efficient way, but all the paper have common problem of data cleaning and inconsistencies. In these papers data inconsistencies, identification of the errors, conflicting, duplicates records etc problems are discussed in detail and also provided the solutions. These algorithms increase the quality of data.
    [Show full text]
  • A Study Over Importance of Data Cleansing in Data Warehouse Shivangi Rana Er
    Volume 6, Issue 4, April 2016 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com A Study over Importance of Data Cleansing in Data Warehouse Shivangi Rana Er. Gagan Prakesh Negi Kapil Kapoor Research Scholar Assistant Professor Associate Professor CSE Department CSE Department ECE Department Abhilashi Group of Institutions Abhilashi Group of Institutions Abhilashi Group of Institutions (School of Engg.) (School of Engg.) (School of Engg.) Chail Chowk, Mandi, India Chail Chowk, Mandi, India Chail Chowk Mandi, India Abstract: Cleansing data from impurities is an integral part of data processing and maintenance. This has lead to the development of a broad range of methods intending to enhance the accuracy and thereby the usability of existing data. In these days, many organizations tend to use a Data Warehouse to meet the requirements to develop decision-making processes and achieve their goals better and satisfy their customers. It enables Executives to access the information they need in a timely manner for making the right decision for any work. Decision Support System (DSS) is one of the means that applied in data mining. Its robust and better decision depends on an important and conclusive factor called Data Quality (DQ), to obtain a high data quality using Data Scrubbing (DS) which is one of data Extraction Transformation and Loading (ETL) tools. Data Scrubbing is very important and necessary in the Data Warehouse (DW). This paper presents a survey of sources of error in data, data quality challenges and approaches, data cleaning types and techniques and an overview of ETL Process.
    [Show full text]
  • Design Tradeoffs for SSD Reliability Bryan S
    Design Tradeoffs for SSD Reliability Bryan S. Kim, Seoul National University; Jongmoo Choi, Dankook University; Sang Lyul Min, Seoul National University https://www.usenix.org/conference/fast19/presentation/kim-bryan This paper is included in the Proceedings of the 17th USENIX Conference on File and Storage Technologies (FAST ’19). February 25–28, 2019 • Boston, MA, USA 978-1-939133-09-0 Open access to the Proceedings of the 17th USENIX Conference on File and Storage Technologies (FAST ’19) is sponsored by Design Tradeoffs for SSD Reliability Bryan S. Kim Jongmoo Choi Sang Lyul Min Seoul National University Dankook University Seoul National University Abstract SSD Redun- Data dancy mgmt. Flash memory-based SSDs are popular across a wide range scheme of data storage markets, while the underlying storage medium—flash memory—is becoming increasingly unreli- SSD reliability Flash Data Data Host enhancement able. As a result, modern SSDs employ a number of in- memory re-read scrubbing system techniques device reliability enhancement techniques, but none of them RBER of UBER offers a one size fits all solution when considering the multi- Thres. Error underlying requirement for voltage correction dimensional requirements for SSDs: performance, reliabil- memories: SSDs: tuning code ity, and lifetime. 10-4 ~ 10-2 10-17 ~ 10-15 In this paper, we examine the design tradeoffs of exist- Figure 1: SSD reliability enhancement techniques. ing reliability enhancement techniques such as data re-read, intra-SSD redundancy, and data scrubbing. We observe that an uncoordinated use of these techniques adversely affects ory [10]. The high error rates in today’s flash memory are the performance of the SSD, and careful management of the caused by various reasons, from wear-and-tear [11,19,27], to techniques is necessary for a graceful performance degrada- gradual charge leakage [11,14,51] and data disturbance [13, tion while maintaining a high reliability standard.
    [Show full text]