Lab 6: Importing Data Into a Geodatabase

Total Page:16

File Type:pdf, Size:1020Kb

Lab 6: Importing Data Into a Geodatabase Lab 6: Importing Data into a Geodatabase Overview & Background Information In the first lab, we asked you to explore the availability of GIS data on the Internet. Your task was to identify data sets that could be used in support of specific research problems or questions. This week, you will practice importing the data that you identified in the first lab, or new data sets that you have identified since then into a geodatabase. This geodatabase can then be used for your research or project. Although it is possible to perform many operations and analyses using data in the format you acquire them in, converting to a geodatabase is a skill that will help you organize your project data and make it more universally useful within ArcGIS. We want you to include at least three thematic layers into your map. Learning Objectives: To understand the organization of geodatabases To understand the process of importing data into a geodatabase To revisit the procedures for projecting data and map creation for delivery of final products To be submitted: 1. (15 pts) A write-up answering the questions at the end of the lab, including a screenshot of your final geodatabase in ArcCatalog, expanded so all the feature datasets and feature classes can be seen. 2. (5 pts) A thematic map of your downloaded data that communicates a particular message to a general audience. Procedure: 1. Download and standardize your data Locate the data layers you want to import; these could be the ones you explored to answer your own research question in Lab 1 or new data you identify as your questions or ability to find data have improved. Download and unzip them (if necessary) to your Lab06 folder. If you are unable to download your data from Lab 1 you must find spatial data that you can use. See lecture Powerpoints for spatial data sources. Be sure to form a question before acquiring your data. Unzipping data Much of the GIS data that is available for download comes compressed in a .zip file (like our lab data). However, you may also encounter RAR (.rar) files and gzip (.gz) files. These are also compressed files. Like .zip files, they may be unzipped using WinZip or 7-Zip. **Note: If you downloaded any TIGER census files, you must first unzip the file that you downloaded, then open the folder in Windows Explorer. Navigate through all the nested folders, and you will reach another zip file that contains your data. Be sure to unzip this folder before you try to import any data.** 1 Lab 6: Importing Data into a Geodatabase Supported Data After you have unzipped your data, you will be able to see the format in which it is stored. We provide instructions for how to import the following data formats into geodatabases. Other file types may be more complicated to import or contain data that are not simply translatable into geodatabases. To simplify your tasks, if you find data files that are not among those listed here, we suggest that you look for files that are. Ask your GSI if you are unsure about a file type. Vector data model Raster data model ♦ Shapefiles (.shp) ♦ ARC/INFO interchange files (.e00) (containing ♦ Coverages ESRI GRID rasters) ♦ ARC/INFO interchange files (.e00) ♦ ERDAS IMAGINE rasters (.img) (containing coverage files) ♦ MrSID rasters (.sid) ♦ JPEG rasters (.jpg) ♦ JPEG 2000 rasters (.jp2) ♦ ESRI Band Interleaved by Line rasters (.bil) ♦ Tagged Image File Format (TIFF) rasters (.tif) ♦ GeoTIFF rasters (.tif) ♦ Graphic Interchange Format (GIF) rasters (.gif) ♦ ESRI GRID rasters ♦ Portable Network Graphics rasters (.png) If you downloaded any files ending with the .e00 extension, follow these next steps. If not, you may skip to the next section. Converting .e00 files Files with the extension .e00 are ARC/INFO interchange files. These files are created to transport coverages as well ESRI GRID and TIN files across different networks and between different types of computers. They need to be converted back to coverages before they can be used. ♦ In ArcCatalog, open the toolbox and go to Conversion tools │ To Coverage │ Import from E00. ♦ A new window will open. Click the browse folder to navigate to your Input interchange file. Then click the browse folder to set an Output folder. Finally, in the Output name field, provide a proper name for your file. You will not be able to save it in your geodatabase just yet. Click Save and OK to convert the file. 2 Lab 6: Importing Data into a Geodatabase Standardizing projections Before you add your files to a feature dataset, you must make sure they are all in the same projection. Feature classes cannot be stored within the same feature dataset if they do not share a projection. If you found your data from different sources, it is likely that the data sets will be in different coordinate systems with different projections. If you got all your data from the same source, it is still a good idea to verify that the projections match. Follow the steps in Lab 4 to check for the projection of your data and refer to your answer for Question 5 in Lab 4 to determine what the ideal projection for your data should be. If you need to change the projection, please refer to the steps indicated in Lab 4. Remember that you can import projections from an existing file in addition to selecting them from a list. This ensures that the layer you are working with has exactly the same projection as the layer whose projection you import. 2. Create an empty geodatabase Follow the steps you learned in Lab 05 to create an empty geodatabase, and a feature dataset within it that has the projection set to the projection of the data. As with reprojecting data, you can select the projection from a list or import one from an existing dataset. You do not need to define a vertical coordinate system. You may create as many feature datasets as you deem appropriate. Just ensure that the projection you set for the dataset matches the projection of the layers that will go into it. 3. Import data into your geodatabase 3 Lab 6: Importing Data into a Geodatabase Now you will import your downloaded data into your geodatabase. Double-check to make sure your data are all in the same projection before proceeding. This section includes instructions for importing the many types of files you may have downloaded. To import a shapefile (including TIGER data): ♦ Within ArcCatalog, right-click on your feature dataset and go to Import | Feature Class (single). ♦ Click on browse folder next to the Input Features box and navigate to your shapefile. Click Add. ♦ In the Output Feature Class box, type a name you want to give your feature class. ♦ Click OK. When the process is complete, click Close in the status window. To import a coverage: ♦ Importing a coverage is similar to importing a shapefile. Right-click on your feature dataset and go to Import/ Feature Class (single). ♦ Click on the browse folder next to the Input Features box and navigate to your coverage. Double click on it to see the arcs, polygons, and tics it contains. Select the feature you’d like to import and click Add. ♦ In the Output Feature Class box, type a name for your feature data class. ♦ Click OK. When the process is complete, click Close in the status window. To import a raster dataset (.img, .bil, .sid, .tif, GRID, GeoTIFF, .gif, jpeg2000): ♦ Right-click on your geodatabase (not your feature dataset) and go to Import / Raster Datasets… ♦ In the new window, in the Input Raster field click on the to browse to your directory and select the raster data ♦ In the Output Geodatabase field, the software will automatically be set to the geodatabase you created. Click OK. ♦ When the import operation is complete, click Close in the status window. Some of you might need to perform some extra tasks (e.g. if you have a table with information that needs to be joined to a feature class). Instructions in Appendix I will guide you in how to join tables and how to clip data to your extent of interest (vector and raster). 4. Create a thematic map 4 Lab 6: Importing Data into a Geodatabase Now that you have your data organized into a geodatabase, add it to a map document to create a thematic map that helps answer your research question. Be sure to include the projection used for your map along with source data. QUESTIONS TO ANSWER: Question 1 (5 points): Give some reasons why you might want to have multiple feature datasets in your geodatabase. That is, what are the potential advantages to storing feature classes in different feature datasets of a geodatabase? Similarly, under what circumstances would it make sense to start a new geodatabase? Question 2 (5 points): When you have imported all of your data, expand your geodatabase and all of its feature datasets in the ArcCatalog file tree by clicking the (+) buttons next to them. Make sure all your imported files are visible. Take a screen shot of your expanded geodatabase by pressing ALT+Print Screen, or use the snipping tool, and paste this image into your lab write-up. Question 3 (Map Reasoning) (5 points): What are you trying to show in your map? What do you want the viewer to get out of it, and how did you use different design elements to communicate your message? How does your map help you answer your research question? What other data do you wish you had? Map (5 points): Create a thematic map of your data, including at least three different layers and all essential map elements.
Recommended publications
  • Data Preparation & Descriptive Statistics
    Data Preparation & Descriptive Statistics (ver. 2.4) Oscar Torres-Reyna Data Consultant [email protected] PU/DSS/OTR http://dss.princeton.edu/training/ Basic definitions… For statistical analysis we think of data as a collection of different pieces of information or facts. These pieces of information are called variables. A variable is an identifiable piece of data containing one or more values. Those values can take the form of a number or text (which could be converted into number) In the table below variables var1 thru var5 are a collection of seven values, ‘id’ is the identifier for each observation. This dataset has information for seven cases (in this case people, but could also be states, countries, etc) grouped into five variables. id var1 var2 var3 var4 var5 1 7.3 32.27 0.1 Yes Male 2 8.28 40.68 0.56 No Female 3 3.35 5.62 0.55 Yes Female 4 4.08 62.8 0.83 Yes Male 5 9.09 22.76 0.26 No Female 6 8.15 90.85 0.23 Yes Female 7 7.59 54.94 0.42 Yes Male PU/DSS/OTR Data structure… For data analysis your data should have variables as columns and observations as rows. The first row should have the column headings. Make sure your dataset has at least one identifier (for example, individual id, family id, etc.) id var1 var2 var3 var4 var5 First row should have the variable names 1 7.3 32.27 0.1 Yes Male 2 8.28 40.68 0.56 No Female Cross-sectional data 3 3.35 5.62 0.55 Yes Female 4 4.08 62.8 0.83 Yes Male 5 9.09 22.76 0.26 No Female 6 8.15 90.85 0.23 Yes Female 7 7.59 54.94 0.42 Yes Male id year var1 var2 var3 1 2000 7 74.03 0.55 Group 1 1 2001 2 4.6 0.44 At least one identifier 1 2002 2 25.56 0.77 2 2000 7 59.52 0.05 Cross-sectional time series data Group 2 2 2001 2 16.95 0.94 or panel data 2 2002 9 1.2 0.08 3 2000 9 85.85 0.5 Group 3 3 2001 3 98.85 0.32 3 2002 3 69.2 0.76 PU/DSS/OTR NOTE: See: http://www.statistics.com/resources/glossary/c/crossdat.php Data format (ASCII)… ASCII (American Standard Code for Information Interchange).
    [Show full text]
  • Full Document
    R&D Centre for Mobile Applications (RDC) FEE, Dept of Telecommunications Engineering Czech Technical University in Prague RDC Technical Report TR-13-4 Internship report Evaluation of Compressibility of the Output of the Information-Concealing Algorithm Julien Mamelli, [email protected] 2nd year student at the Ecole´ des Mines d'Al`es (N^ımes,France) Internship supervisor: Luk´aˇsKencl, [email protected] August 2013 Abstract Compression is a key element to exchange files over the Internet. By generating re- dundancies, the concealing algorithm proposed by Kencl and Loebl [?], appears at first glance to be particularly designed to be combined with a compression scheme [?]. Is the output of the concealing algorithm actually compressible? We have tried 16 compression techniques on 1 120 files, and the result is that we have not found a solution which could advantageously use repetitions of the concealing method. Acknowledgments I would like to express my gratitude to my supervisor, Dr Luk´aˇsKencl, for his guidance and expertise throughout the course of this work. I would like to thank Prof. Robert Beˇst´akand Mr Pierre Runtz, for giving me the opportunity to carry out my internship at the Czech Technical University in Prague. I would also like to thank all the members of the Research and Development Center for Mobile Applications as well as my colleagues for the assistance they have given me during this period. 1 Contents 1 Introduction 3 2 Related Work 4 2.1 Information concealing method . 4 2.2 Archive formats . 5 2.3 Compression algorithms . 5 2.3.1 Lempel-Ziv algorithm .
    [Show full text]
  • ARC File Revision 3.0 Proposal
    ARC file Revision 3.0 Proposal Steen Christensen, Det Kongelige Bibliotek <ssc at kb dot dk> Michael Stack, Internet Archive <stack at archive dot org> Edited by Michael Stack Revision History Revision 1 09/09/2004 Initial conversion of wiki working doc. [http://crawler.archive.org/cgi-bin/wiki.pl?ArcRevisionProposal] to docbook. Added suggested edits suggested by Gordon Mohr (Others made are still up for consideration). This revision is what is being submitted to the IIPC Framework Group for review at their London, 09/20/2004 meeting. Table of Contents 1. Introduction ............................................................................................................................2 1.1. IIPC Archival Data Format Requirements .......................................................................... 2 1.2. Input ...........................................................................................................................2 1.3. Scope ..........................................................................................................................3 1.4. Acronyms, Abbreviations and Definitions .......................................................................... 3 2. ARC Record Addressing ........................................................................................................... 4 2.1. Reference ....................................................................................................................4 2.2. The ari Scheme ............................................................................................................
    [Show full text]
  • Pack, Encrypt, Authenticate Document Revision: 2021 05 02
    PEA Pack, Encrypt, Authenticate Document revision: 2021 05 02 Author: Giorgio Tani Translation: Giorgio Tani This document refers to: PEA file format specification version 1 revision 3 (1.3); PEA file format specification version 2.0; PEA 1.01 executable implementation; Present documentation is released under GNU GFDL License. PEA executable implementation is released under GNU LGPL License; please note that all units provided by the Author are released under LGPL, while Wolfgang Ehrhardt’s crypto library units used in PEA are released under zlib/libpng License. PEA file format and PCOMPRESS specifications are hereby released under PUBLIC DOMAIN: the Author neither has, nor is aware of, any patents or pending patents relevant to this technology and do not intend to apply for any patents covering it. As far as the Author knows, PEA file format in all of it’s parts is free and unencumbered for all uses. Pea is on PeaZip project official site: https://peazip.github.io , https://peazip.org , and https://peazip.sourceforge.io For more information about the licenses: GNU GFDL License, see http://www.gnu.org/licenses/fdl.txt GNU LGPL License, see http://www.gnu.org/licenses/lgpl.txt 1 Content: Section 1: PEA file format ..3 Description ..3 PEA 1.3 file format details ..5 Differences between 1.3 and older revisions ..5 PEA 2.0 file format details ..7 PEA file format’s and implementation’s limitations ..8 PCOMPRESS compression scheme ..9 Algorithms used in PEA format ..9 PEA security model .10 Cryptanalysis of PEA format .12 Data recovery from
    [Show full text]
  • Steganography and Vulnerabilities in Popular Archives Formats.| Nyxengine Nyx.Reversinglabs.Com
    Hiding in the Familiar: Steganography and Vulnerabilities in Popular Archives Formats.| NyxEngine nyx.reversinglabs.com Contents Introduction to NyxEngine ............................................................................................................................ 3 Introduction to ZIP file format ...................................................................................................................... 4 Introduction to steganography in ZIP archives ............................................................................................. 5 Steganography and file malformation security impacts ............................................................................... 8 References and tools .................................................................................................................................... 9 2 Introduction to NyxEngine Steganography1 is the art and science of writing hidden messages in such a way that no one, apart from the sender and intended recipient, suspects the existence of the message, a form of security through obscurity. When it comes to digital steganography no stone should be left unturned in the search for viable hidden data. Although digital steganography is commonly used to hide data inside multimedia files, a similar approach can be used to hide data in archives as well. Steganography imposes the following data hiding rule: Data must be hidden in such a fashion that the user has no clue about the hidden message or file's existence. This can be achieved by
    [Show full text]
  • Winzip 12 Reviewer's Guide
    Introducing WinZip® 12 WinZip® is the most trusted way to work with compressed files. No other compression utility is as easy to use or offers the comprehensive and productivity-enhancing approach that has made WinZip the gold standard for file-compression tools. With the new WinZip 12, you can quickly and securely zip and unzip files to conserve storage space, speed up e-mail transmission, and reduce download times. State-of-the-art file compression, strong AES encryption, compatibility with more compression formats, and new intuitive photo compression, make WinZip 12 the complete compression and archiving solution. Building on the favorite features of a worldwide base of several million users, WinZip 12 adds new features for image compression and management, support for new compression methods, improved compression performance, support for additional archive formats, and more. Users can work smarter, faster, and safer with WinZip 12. Who will benefit from WinZip® 12? The simple answer is anyone who uses a PC. Any PC user can benefit from the compression and encryption features in WinZip to protect data, save space, and reduce the time to transfer files on the Internet. There are, however, some PC users to whom WinZip is an even more valuable and essential tool. Digital photo enthusiasts: As the average file size of their digital photos increases, people are looking for ways to preserve storage space on their PCs. They have lots of photos, so they are always seeking better ways to manage them. Sharing their photos is also important, so they strive to simplify the process and reduce the time of e-mailing large numbers of images.
    [Show full text]
  • The Ark Handbook
    The Ark Handbook Matt Johnston Henrique Pinto Ragnar Thomsen The Ark Handbook 2 Contents 1 Introduction 5 2 Using Ark 6 2.1 Opening Archives . .6 2.1.1 Archive Operations . .6 2.1.2 Archive Comments . .6 2.2 Working with Files . .7 2.2.1 Editing Files . .7 2.3 Extracting Files . .7 2.3.1 The Extract dialog . .8 2.4 Creating Archives and Adding Files . .8 2.4.1 Compression . .9 2.4.2 Password Protection . .9 2.4.3 Multi-volume Archive . 10 3 Using Ark in the Filemanager 11 4 Advanced Batch Mode 12 5 Credits and License 13 Abstract Ark is an archive manager by KDE. The Ark Handbook Chapter 1 Introduction Ark is a program for viewing, extracting, creating and modifying archives. Ark can handle vari- ous archive formats such as tar, gzip, bzip2, zip, rar, 7zip, xz, rpm, cab, deb, xar and AppImage (support for certain archive formats depends on the appropriate command-line programs being installed). In order to successfully use Ark, you need KDE Frameworks 5. The library libarchive version 3.1 or above is needed to handle most archive types, including tar, compressed tar, rpm, deb and cab archives. To handle other file formats, you need the appropriate command line programs, such as zipinfo, zip, unzip, rar, unrar, 7z, lsar, unar and lrzip. 5 The Ark Handbook Chapter 2 Using Ark 2.1 Opening Archives To open an archive in Ark, choose Open... (Ctrl+O) from the Archive menu. You can also open archive files by dragging and dropping from Dolphin.
    [Show full text]
  • How to 'Zip and Unzip' Files
    How to 'zip and unzip' files The Windows 10 operating system provides a very easy way to zip-up any file (or folder) you want by using a utility program called 7-zip (Seven Zip). The program is integrated in the context menu which you get when you right-click on anything selected. Here are the basic steps you need to take in order to: Zip a file and create an archive 1. Open a folder in your machine and select any file (by clicking on it once). I selected a single file called 'how-to send an email.docx' 2. Now right click the mouse to have the context menu appear 3. In the context menu you should be able to see some commands like the capture below 4. Since we want to zip up the file you need to select one of the bottom two options a. 'Add to archive' will actually open up a dialog of the 7-zip application and will give you the opportunity to customise the archive. b. 'Add to how-to send an email.zip' is actually the quickest way possible to create an archive. The software uses the name of the file and selects a default compression scheme (.zip) so that you can, with two clicks, create a zip archive containing the original file. 5. Now you can use the 'how-to send an email.zip' file and send it as a 'smaller' email attachment. Now consider that you have just received (via email) the 'how-to send an email.zip' file and you need to get (the correct phrase is extract) the file it contains.
    [Show full text]
  • Rapc) 97 6.1 Compression Scheme
    UCC Library and UCC researchers have made this item openly available. Please let us know how this has helped you. Thanks! Title Content-aware compression for big textual data analysis Author(s) Dong, Dapeng Publication date 2016 Original citation Dong, D. 2016. Content-aware compression for big textual data analysis. PhD Thesis, University College Cork. Type of publication Doctoral thesis Rights © 2016, Dapeng Dong. http://creativecommons.org/licenses/by-nc-nd/3.0/ Embargo information No embargo required Item downloaded http://hdl.handle.net/10468/2697 from Downloaded on 2021-10-11T16:19:07Z Content-aware Compression for Big Textual Data Analysis Dapeng Dong MSC Thesis submitted for the degree of Doctor of Philosophy ¡ NATIONAL UNIVERSITY OF IRELAND, CORK FACULTY OF SCIENCE DEPARTMENT OF COMPUTER SCIENCE May 2016 Head of Department: Prof. Cormac J. Sreenan Supervisors: Dr. John Herbert Prof. Cormac J. Sreenan Contents Contents List of Figures . iv List of Tables . vii Notation . viii Acknowledgements . .x Abstract . xi 1 Introduction 1 1.1 The State of Big Data . .1 1.2 Big Data Development . .2 1.3 Our Approach . .4 1.4 Thesis Structure . .8 1.5 General Convention . .9 1.6 Publications . 10 2 Background Research 11 2.1 Big Data Organization in Hadoop . 11 2.2 Conventional Compression Schemes . 15 2.2.1 Transformation . 15 2.2.2 Modeling . 17 2.2.3 Encoding . 22 2.3 Random Access Compression Scheme . 23 2.3.1 Context-free Compression . 23 2.3.2 Self-synchronizing Codes . 25 2.3.3 Indexing . 26 2.4 Emerging Techniques for Big Data Compression .
    [Show full text]
  • Deduplicating Compressed Contents in Cloud Storage Environment
    Deduplicating Compressed Contents in Cloud Storage Environment Zhichao Yan, Hong Jiang Yujuan Tan* Hao Luo University of Texas Arlington Chongqing University University of Nebraska Lincoln [email protected] [email protected] [email protected] [email protected] Corresponding Author Abstract Data compression and deduplication are two common approaches to increasing storage efficiency in the cloud environment. Both users and cloud service providers have economic incentives to compress their data before storing it in the cloud. However, our analysis indicates that compressed packages of different data and differ- ently compressed packages of the same data are usual- ly fundamentally different from one another even when they share a large amount of redundant data. Existing data deduplication systems cannot detect redundant data among them. We propose the X-Ray Dedup approach to extract from these packages the unique metadata, such as the “checksum” and “file length” information, and use it as the compressed file’s content signature to help detect and remove file level data redundancy. X-Ray Dedup is shown by our evaluations to be capable of breaking in the boundaries of compressed packages and significantly Figure 1: A user scenario on cloud storage environment reducing compressed packages’ size requirements, thus further optimizing storage space in the cloud. will generate different compressed data of the same con- tents that render fingerprint-based redundancy identifi- cation difficult. Third, very similar but different digital 1 Introduction contents (e.g., files or data streams), which would other- wise present excellent deduplication opportunities, will Due to the information explosion [1, 3], data reduc- become fundamentally distinct compressed packages af- tion technologies such as compression and deduplica- ter applying even the same compression algorithm.
    [Show full text]
  • Installing Your Cinesamples Product - Cinebells
    Installing Your Cinesamples Product - CineBells Step 1. In your Downloads folder, or in the location you have told your browser to send your downloads, you will find a file called“CineBells.zip”along with the five .rar files shown to the left. First Unzip the zip file to create your main CineBells folder. Then you will need a utility that can extract the .rar files, like RarMarchine or UnRarX for Mac, or WinRar for Windows. No matter which you use, you will only need to select and unarchive the first file (part1). The utility will automatically create a Step 2. CineBells_Samples folder, and decompress all five .rar files into it in one step. Please DO NOT use Stuffit for this - it will not work correctly. Make sure to check that all five rars downloaded completely - notice the sizes to the left. After the rars are extracted, go into the CineBells_Samples folder, and drag the Samples Step 3. folder inside it into the main CineBells folder that was created from the zip file. Next drag your final CineBells folder to your sample hard drive. It should look like the picture below afterwards. If this is your first Cinesamples product, you may want to create a Cinesamples folder beforehand. The “CineBells_Samples”folder should now be empty and can be deleted. Step 4. Next, open the full version of Kontakt 4 (at least 4.2.3) or Kontakt 5, select the Files tab, and navigate to your CineBells folder. Double click or drag nki files into the main window to load the different instruments.
    [Show full text]
  • Chapter 5 Record-Aware Compression
    UCC Library and UCC researchers have made this item openly available. Please let us know how this has helped you. Thanks! Title Content-aware compression for big textual data analysis Author(s) Dong, Dapeng Publication date 2016 Original citation Dong, D. 2016. Content-aware compression for big textual data analysis. PhD Thesis, University College Cork. Type of publication Doctoral thesis Rights © 2016, Dapeng Dong. http://creativecommons.org/licenses/by-nc-nd/3.0/ Embargo information No embargo required Item downloaded http://hdl.handle.net/10468/2697 from Downloaded on 2021-10-10T20:32:08Z Content-aware Compression for Big Textual Data Analysis Dapeng Dong MSC Thesis submitted for the degree of Doctor of Philosophy ¡ NATIONAL UNIVERSITY OF IRELAND, CORK FACULTY OF SCIENCE DEPARTMENT OF COMPUTER SCIENCE May 2016 Head of Department: Prof. Cormac J. Sreenan Supervisors: Dr. John Herbert Prof. Cormac J. Sreenan Contents Contents List of Figures . iv List of Tables . vii Notation . viii Acknowledgements . .x Abstract . xi 1 Introduction 1 1.1 The State of Big Data . .1 1.2 Big Data Development . .2 1.3 Our Approach . .4 1.4 Thesis Structure . .8 1.5 General Convention . .9 1.6 Publications . 10 2 Background Research 11 2.1 Big Data Organization in Hadoop . 11 2.2 Conventional Compression Schemes . 15 2.2.1 Transformation . 15 2.2.2 Modeling . 17 2.2.3 Encoding . 22 2.3 Random Access Compression Scheme . 23 2.3.1 Context-free Compression . 23 2.3.2 Self-synchronizing Codes . 25 2.3.3 Indexing . 26 2.4 Emerging Techniques for Big Data Compression .
    [Show full text]