The Google File System

Total Page:16

File Type:pdf, Size:1020Kb

The Google File System TheThe GoogleGoogle FileFile SystemSystem ByBy GhemawatGhemawat,, GobioffGobioff andand LeungLeung OutlineOutline ¾¾ OverviewOverview ¾¾ AssumptionAssumption ¾¾ DesignDesign ofof GFSGFS ¾¾ SystemSystem InteractionsInteractions ¾¾ MasterMaster OperationsOperations ¾¾ FaultFault ToleranceTolerance ¾¾ MeasurementsMeasurements OverviewOverview ¾¾ GFS:GFS: ScalableScalable distributeddistributed filefile systemsystem forfor largelarge distributeddistributed datadata--intensiveintensive applicationsapplications AssumptionAssumption ¾¾ BuiltBuilt fromfrom inexpensiveinexpensive commoditycommodity componentscomponents thatthat oftenoften failfail ¾¾ SystemSystem storesstores modestmodest numbernumber ofof largelarge files.files. z FewFew millionmillion files,files, eacheach 100MB+.100MB+. MultiMulti--GBGB common.common. SmallSmall filesfiles supportedsupported butbut needneed notnot optimizeoptimize for.for. AssumptionAssumption ¾¾ TypicalTypical workloadworkload ReadRead patternpattern z LargeLarge streamingstreaming readsreads • Individual operations read hundreds of KBs, more commonly 1MB or more. • Succesive operations from same client often read through contiguous region of file. z SmallSmall randomrandom readsreads • Read a few KB at some random offset. • Performance-conscious apps often batch small reads. AssumptionAssumption ¾¾ TypicalTypical workloadworkload WriteWrite patternpattern z ManyMany largelarge sequentialsequential writeswrites thatthat appendappend datadata toto files.files. • Once written, files seldom modified again. • Small writes at arbitrary positions supported, not efficient. (not often in typical workload) DesignDesign ofof GFSGFS ¾ Multiple standalone and independent “clusters” ¾ Each cluster includes a master and multiple chunkservers ¾ Cluster accessed by multiple clients. ¾ Can run chunkserver and client on same machine. ¾ Files divided into fixed-size chunks. z Each chunk is 64MB big ¾ Each file has 64 bit chunk handle assigned by master at time of chunk creation. z The addressable range of a GFS cluster is • 264 * 64MB = 1125899906842624 TB DesignDesign ofof GFSGFS ¾ Master z maintains all file system metadata. • Namespace, access control info, files Æ chunks mapping, locations of chunks. z Manages chunk lease management, garbage collection of orphaned chunks, chunk migration. z Periodically communicates with each chunkserver in HeartBeat messages to give it instructions and collect its state. ¾ Chunkservers z Repositories of chunks in GFS z Chunks stored on each chunkserver’s local disks as LIINUX files z Chunks replicated (default factor of 3). DesignDesign ofof GFSGFS ¾ GFSGFS clientclient codecode linkedlinked intointo eacheach applicationapplication ¾ ClientClient doesdoes notnot cachecache filefile datadata z Most applications stream through huge files or have working sets too large to be cached z No need to worry about Cache coherence issues ¾ ChunkserversChunkservers dodo notnot cachecache filefile data.data. z Most applications stream through huge files or have working sets too large to be cached z Chunkservers could benefit from some LINUX buffer caches DesignDesign ofof GFSGFS ¾¾ FaultFault ToleanceToleance MeasuresMeasures z (Dynamic)(Dynamic) ReplicationReplication z HeartbeatHeartbeat messagesmessages z LoggingLogging z CheckpointingCheckpointing // RecoveryRecovery DesignDesign ofof GFSGFS DesignDesign ofof GFSGFS ¾ ConsistencyConsistency ModelModel z File regions: “consistent”, “defined” • Consistent: All clients will always see same data regardless of which replicas. • Defined: Consistent AND clients will see what a mutation writes in its entirety z Guarantees by GFS • File namespace mutations are atomic (e.g. creation). Handled exclusively by master. • Successful mutation without interference → defined • Concurrent successful mutations → undefined but consistent • Failed mutation → inconsistent (and undefined) DesignDesign ofof GFSGFS ¾¾ ImplicationsImplications ofof thethe consistencyconsistency modelmodel onon thethe applicationsapplications z ApplicationsApplications needneed toto distinguishdistinguish betweenbetween defineddefined andand undefinedundefined regionsregions • Rely on appends not overwrites => avoid undefined regions • Application-level Checkpointing • Write self-validating self-identifying records SystemSystem InteractionsInteractions ¾¾ InteractionsInteractions includeinclude z FileFile Read,Read, FileFile Mutations,Mutations, andand SnapshotsSnapshots ¾¾ HasHas toto ensureensure thethe consistencyconsistency guaranteesguarantees ¾¾ HasHas toto ensureensure thethe throughputthroughput z MinimizeMinimize thethe involvementinvolvement ofof mastermaster inin thethe interactionsinteractions z DelegateDelegate operationsoperations ontoonto chunkserverschunkservers/clients/clients SystemSystem InteractionsInteractions ¾¾ ReadRead accessaccess SystemSystem InteractionsInteractions ¾¾ FileFile Mutations:Mutations: Write/RecordWrite/Record Append/SnapshotsAppend/Snapshots z WriteWrite:: causescauses datadata toto bebe writtenwritten atat anan applicationapplication--specifiedspecified filefile offset.offset. z RecordRecord appendappend:: causescauses datadata toto bebe appendedappended atomicallyatomically atat leastleast onceonce eveneven inin thethe presencepresence ofof concurrentconcurrent mutations.mutations. z SnapshotSnapshot:: MakesMakes aa copycopy ofof aa filefile oror directorydirectory treetree z NeedNeed toto ensureensure consistencyconsistency modelmodel SystemSystem InteractionsInteractions ¾ EachEach mutationmutation performedperformed atat allall thethe chunkchunk’’ss replicasreplicas ¾ UseUse leasesleases toto maintainmaintain aa consistentconsistent mutationmutation orderorder acrossacross replicas.replicas. z Master grants chunk lease to one of the replicas – the primary. z Primary picks a serial order for all mutations to the “chunk”. All replicas follow this order when applying mutations. z THUS global mutation order defined first by lease grant order chosen by master and then by serial numbers in the lease SystemSystem InteractionsInteractions ¾¾ LeaseLease z DelegateDelegate authorizationauthorization toto thethe primaryprimary replicareplica inin datadata mutationsmutations z MinimizeMinimize managementmanagement overheadoverhead atat thethe mastermaster z LeaseLease expirationexpiration time:time: 6060 seconds.seconds. CanCan getget extended.extended. SystemSystem InteractionsInteractions ¾¾ WriteWrite accessaccess 1. Client asks master which chunkserver holds current lease and locations of other replicas. If no lease, master grants one to a replica it chooses. 2. Master replies with primary and secondaries. Client caches this for future mutations; contacts master again only when primary unreachable or does not have a lease 3. Client pushes data to all replicas. Client can do so in any order. Each chunkserver stores data in internal LRU buffer cache until data used or aged out. 4. Once all replicas have received data, client sends write to the primary. Primary assigns consecutive serial numbers to mutations it receives, possibly from multiple clients, which provides necessary serialization. Applies mutation to its own local state in serial number order SystemSystem InteractionsInteractions ¾¾ WriteWrite accessaccess 5. Primary forwards write request to all secondary replicas. Each secondary replica applies mutations in the same serial number order assigned by the primary 6. Secondaries all reply to primary – “finished the operation” 7. Primary replies to the client • Success of operation • Operation failed SystemSystem InteractionsInteractions ¾¾ WriteWrite accessaccess z IfIf aa writewrite byby thethe applicationapplication isis largelarge andand straddlesstraddles aa chunkchunk boundary,boundary, GFSGFS clientclient codecode breaksbreaks itit downdown intointo multiplemultiple writewrite operationsoperations • Concurrent writes on the same chunk could result in “undefined” region • SystemSystem InteractionsInteractions ¾¾ RecordRecord AppendAppend z AnAn atomicatomic operationoperation whichwhich appendsappends datadata toto thethe endend ofof aa filefile • The appended data is guaranteed to be “defined” and “consistent” • GFS appends it to the file at least once automatically at an offset of GFS’s choosing and returns that offset to the client • Control flow similar to a write with some small changes z The length of the appended data is limited to 16 MB (1/4 of the maximum chunk size) SystemSystem InteractionsInteractions ¾¾ RecordRecord AppendAppend (steps)(steps) z FollowFollow similarsimilar stepssteps asas inin thethe WriteWrite accessaccess • Client pushes data to all replicas • Client sends “Append” request to the primary replica • Primary checks if the data being appended will straddle the boundary of the last chunk. (Note: the appended data size <= 16MB) z Pads the chunk to the maximum size (on all replicas) z Replies to the client indicating the operation should be retired on the next chunk z This is a uncommon case SystemSystem InteractionsInteractions ¾¾ RecordRecord appendappend (steps)(steps) • If the data is within the chunk’s boundary z The primary appends the data z Tells the secondaries to append the data at the same offset z Reply success to the client z This is the common case ! SystemSystem InteractionsInteractions ¾¾ SnapshotsSnapshots z MakesMakes aa copycopy ofof aa filefile oror directorydirectory treetree (the(the ““sourcesource””)) reallyreally fastfast whilewhile minimizingminimizing anyany interruptionsinterruptions
Recommended publications
  • The Google File System (GFS)
    The Google File System (GFS), as described by Ghemwat, Gobioff, and Leung in 2003, provided the architecture for scalable, fault-tolerant data management within the context of Google. These architectural choices, however, resulted in sub-optimal performance in regard to data trustworthiness (security) and simplicity. Additionally, the application- specific nature of GFS limits the general scalability of the system outside of the specific design considerations within the context of Google. SYSTEM DESIGN: The authors enumerate their specific considerations as: (1) commodity components with high expectation of failure, (2) a system optimized to handle relatively large files, particularly multi-GB files, (3) most writes to the system are concurrent append operations, rather than internally modifying the extant files, and (4) a high rate of sustained bandwidth (Section 1, 2.1). Utilizing these considerations, this paper analyzes the success of GFS’s design in achieving a fault-tolerant, scalable system while also considering the faults of the system with regards to data-trustworthiness and simplicity. Data-trustworthiness: In designing the GFS system, the authors made the conscious decision not prioritize data trustworthiness by allowing applications to access ‘stale’, or not up-to-date, data. In particular, although the system does inform applications of the chunk-server version number, the designers of GFS encouraged user applications to cache chunkserver information, allowing stale data accesses (Section 2.7.2, 4.5). Although possibly troubling
    [Show full text]
  • A Review on GOOGLE File System
    International Journal of Computer Science Trends and Technology (IJCST) – Volume 4 Issue 4, Jul - Aug 2016 RESEARCH ARTICLE OPEN ACCESS A Review on Google File System Richa Pandey [1], S.P Sah [2] Department of Computer Science Graphic Era Hill University Uttarakhand – India ABSTRACT Google is an American multinational technology company specializing in Internet-related services and products. A Google file system help in managing the large amount of data which is spread in various databases. A good Google file system is that which have the capability to handle the fault, the replication of data, make the data efficient, memory storage. The large data which is big data must be in the form that it can be managed easily. Many applications like Gmail, Facebook etc. have the file systems which organize the data in relevant format. In conclusion the paper introduces new approaches in distributed file system like spreading file’s data across storage, single master and appends writes etc. Keywords:- GFS, NFS, AFS, HDFS I. INTRODUCTION III. KEY IDEAS The Google file system is designed in such a way that 3.1. Design and Architecture: GFS cluster consist the data which is spread in database must be saved in of single master and multiple chunk servers used by a arrange manner. multiple clients. Since files to be stored in GFS are The arrangement of data is in the way that it can large, processing and transferring such huge files can reduce the overhead load on the server, the consume a lot of bandwidth. To efficiently utilize availability must be increased, throughput should be bandwidth files are divided into large 64 MB size highly aggregated and much more services to make chunks which are identified by unique 64-bit chunk the available data more accurate and reliable.
    [Show full text]
  • F1 Query: Declarative Querying at Scale
    F1 Query: Declarative Querying at Scale Bart Samwel John Cieslewicz Ben Handy Jason Govig Petros Venetis Chanjun Yang Keith Peters Jeff Shute Daniel Tenedorio Himani Apte Felix Weigel David Wilhite Jiacheng Yang Jun Xu Jiexing Li Zhan Yuan Craig Chasseur Qiang Zeng Ian Rae Anurag Biyani Andrew Harn Yang Xia Andrey Gubichev Amr El-Helw Orri Erling Zhepeng Yan Mohan Yang Yiqun Wei Thanh Do Colin Zheng Goetz Graefe Somayeh Sardashti Ahmed M. Aly Divy Agrawal Ashish Gupta Shiv Venkataraman Google LLC [email protected] ABSTRACT 1. INTRODUCTION F1 Query is a stand-alone, federated query processing platform The data processing and analysis use cases in large organiza- that executes SQL queries against data stored in different file- tions like Google exhibit diverse requirements in data sizes, la- based formats as well as different storage systems at Google (e.g., tency, data sources and sinks, freshness, and the need for custom Bigtable, Spanner, Google Spreadsheets, etc.). F1 Query elimi- business logic. As a result, many data processing systems focus on nates the need to maintain the traditional distinction between dif- a particular slice of this requirements space, for instance on either ferent types of data processing workloads by simultaneously sup- transactional-style queries, medium-sized OLAP queries, or huge porting: (i) OLTP-style point queries that affect only a few records; Extract-Transform-Load (ETL) pipelines. Some systems are highly (ii) low-latency OLAP querying of large amounts of data; and (iii) extensible, while others are not. Some systems function mostly as a large ETL pipelines. F1 Query has also significantly reduced the closed silo, while others can easily pull in data from other sources.
    [Show full text]
  • A Study of Cryptographic File Systems in Userspace
    Turkish Journal of Computer and Mathematics Education Vol.12 No.10 (2021), 4507-4513 Research Article A study of cryptographic file systems in userspace a b c d e f Sahil Naphade , Ajinkya Kulkarni Yash Kulkarni , Yash Patil , Kaushik Lathiya , Sachin Pande a Department of Information Technology PICT, Pune, India [email protected] b Department of Information Technology PICT, Pune, India [email protected] c Department of Information Technology PICT, Pune, India [email protected] d Department of Information Technology PICT, Pune, India [email protected] e Veritas Technologies Pune, India, [email protected] f Department of Information Technology PICT, Pune, India [email protected] Article History: Received: 10 January 2021; Revised: 12 February 2021; Accepted: 27 March 2021; Published online: 28 April 2021 Abstract: With the advancements in technology and digitization, the data storage needs are expanding; along with the data breaches which can expose sensitive data to the world. Thus, the security of the stored data is extremely important. Conventionally, there are two methods of storage of the data, the first being hiding the data and the second being encryption of the data. However, finding out hidden data is simple, and thus, is very unreliable. The second method, which is encryption, allows for accessing the data by only the person who encrypted the data using his passkey, thus allowing for higher security. Typically, a file system is implemented in the kernel of the operating systems. However, with an increase in the complexity of the traditional file systems like ext3 and ext4, the ones that are based in the userspace of the OS are now allowing for additional features on top of them, such as encryption-decryption and compression.
    [Show full text]
  • Cloud Computing Bible Is a Wide-Ranging and Complete Reference
    A thorough, down-to-earth look Barrie Sosinsky Cloud Computing Barrie Sosinsky is a veteran computer book writer at cloud computing specializing in network systems, databases, design, development, The chance to lower IT costs makes cloud computing a and testing. Among his 35 technical books have been Wiley’s Networking hot topic, and it’s getting hotter all the time. If you want Bible and many others on operating a terra firma take on everything you should know about systems, Web topics, storage, and the cloud, this book is it. Starting with a clear definition of application software. He has written nearly 500 articles for computer what cloud computing is, why it is, and its pros and cons, magazines and Web sites. Cloud Cloud Computing Bible is a wide-ranging and complete reference. You’ll get thoroughly up to speed on cloud platforms, infrastructure, services and applications, security, and much more. Computing • Learn what cloud computing is and what it is not • Assess the value of cloud computing, including licensing models, ROI, and more • Understand abstraction, partitioning, virtualization, capacity planning, and various programming solutions • See how to use Google®, Amazon®, and Microsoft® Web services effectively ® ™ • Explore cloud communication methods — IM, Twitter , Google Buzz , Explore the cloud with Facebook®, and others • Discover how cloud services are changing mobile phones — and vice versa this complete guide Understand all platforms and technologies www.wiley.com/compbooks Shelving Category: Use Google, Amazon, or
    [Show full text]
  • Mapreduce: Simplified Data Processing On
    MapReduce: Simplified Data Processing on Large Clusters Jeffrey Dean and Sanjay Ghemawat [email protected], [email protected] Google, Inc. Abstract given day, etc. Most such computations are conceptu- ally straightforward. However, the input data is usually MapReduce is a programming model and an associ- large and the computations have to be distributed across ated implementation for processing and generating large hundreds or thousands of machines in order to finish in data sets. Users specify a map function that processes a a reasonable amount of time. The issues of how to par- key/value pair to generate a set of intermediate key/value allelize the computation, distribute the data, and handle pairs, and a reduce function that merges all intermediate failures conspire to obscure the original simple compu- values associated with the same intermediate key. Many tation with large amounts of complex code to deal with real world tasks are expressible in this model, as shown these issues. in the paper. As a reaction to this complexity, we designed a new Programs written in this functional style are automati- abstraction that allows us to express the simple computa- cally parallelized and executed on a large cluster of com- tions we were trying to perform but hides the messy de- modity machines. The run-time system takes care of the tails of parallelization, fault-tolerance, data distribution details of partitioning the input data, scheduling the pro- and load balancing in a library. Our abstraction is in- gram's execution across a set of machines, handling ma- spired by the map and reduce primitives present in Lisp chine failures, and managing the required inter-machine and many other functional languages.
    [Show full text]
  • Google File System 2.0: a Modern Design and Implementation
    Google File System 2.0: A Modern Design and Implementation Babatunde Micheal Okutubo Gan Tu Xi Cheng Stanford University Stanford University Stanford University [email protected] [email protected] [email protected] Abstract build a GFS-2.0 from scratch, with modern software design principles and implementation patterns, in order to survey The Google File System (GFS) presented an elegant de- ways to not only effectively achieve various system require- sign and has shown tremendous capability to scale and to ments, but also with great support for high-concurrency tolerate fault. However, there is no concrete implementa- workloads and software extensibility. tion of GFS being open sourced, even after nearly 20 years A second motivation of this project is that GFS has a of its initial development. In this project, we attempt to build single master, which served well back in 2003 when this a GFS-2.0 in C++ from scratch1 by applying modern soft- work was published. However, as data volume, network ware design discipline. A fully functional GFS has been de- capacity as well as hardware capability all have been dras- veloped to support concurrent file creation, read and paral- tically improved from that date, our hypothesis is that a lel writes. The system is capable of surviving chunk servers multi-master version of GFS is advantageous as it provides going down during both read and write operations. Various better fault tolerance (especially from the master servers’ crucial implementation details have been discussed, and a point of view), availability and throughput. Although the micro-benchmark has been conducted and presented.
    [Show full text]
  • A Survey on Cloud Storage
    1764 JOURNAL OF COMPUTERS, VOL. 6, NO. 8, AUGUST 2011 A Survey on Cloud Storage Jiehui JU1,4 1.School of Information and Electronic Engineering, Zhejiang University of Science and Technology,Hangzhou,China Email: [email protected] Jiyi WU2,3, Jianqing FU3, Zhijie LIN1,3, Jianlin ZHANG2 2. Key Lab of E-Business and Information Security, Hangzhou Normal University, Hangzhou,China 3.School of Computer Science and Technology, Zhejiang University,Hangzhou,China 4.Key Lab of E-business Market Application Technology of Guangdong Province, Guangdong University of Business Studies, Guangzhou 510320, China Email: [email protected]; [email protected]; [email protected]; [email protected] Abstract— Cloud storage is a new concept come into being simultaneously with cloud computing, and can be divided into public cloud storage, private cloud storage and hybrid cloud storage. This article gives a quick introduction to cloud storage. It covers the key technologies in Cloud Computing and Cloud Storage. Google GFS massive data storage system and the popular open source Hadoop HDFS were detailed introduced to analyze the principle of Cloud Storage technology. As an important technology area and research direction, cloud storage is becoming a hot research for both academia and industry session. The future valuable research works were summarized at the end. Index Terms—Cloud Storage, Distributed File System, Research Status, survey I. INTRODUCTION As we all known disk storage is one of the largest expenditure in IT projects. ComputerWorld estimates that storage is responsible for almost 30% of capital expenditures as the average growth of data approaches Figure 1.
    [Show full text]
  • Comparative Analysis of Distributed and Parallel File Systems' Internal Techniques
    Comparative Analysis of Distributed and Parallel File Systems’ Internal Techniques Viacheslav Dubeyko Content 1 TERMINOLOGY AND ABBREVIATIONS ................................................................................ 4 2 INTRODUCTION......................................................................................................................... 5 3 COMPARATIVE ANALYSIS METHODOLOGY ....................................................................... 5 4 FILE SYSTEM FEATURES CLASSIFICATION ........................................................................ 5 4.1 Distributed File Systems ............................................................................................................................ 6 4.1.1 HDFS ..................................................................................................................................................... 6 4.1.2 GFS (Google File System) ....................................................................................................................... 7 4.1.3 InterMezzo ............................................................................................................................................ 9 4.1.4 CodA .................................................................................................................................................... 10 4.1.5 Ceph.................................................................................................................................................... 12 4.1.6 DDFS ..................................................................................................................................................
    [Show full text]
  • Qos-Enabled Distributed Mutual Exclusion in Public Clouds
    QoS-enabled Distributed Mutual Exclusion in Public Clouds James Edmondson1, Doug Schmidt1, and Aniruddha Gokhale1 Dept. of EECS, Vanderbilt University, Nashville, TN 37212, USA Email: fjames.r.edmondson,doug.schmidt,[email protected] Abstract. Popular public cloud infrastructures tend to feature central- ized, mutual exclusion models for distributed resources, such as file sys- tems. The result of using such centralized solutions in the Google File System (GFS), for instance, reduces scalability, increases latency, creates a single point of failure, and tightly couples applications with the under- lying services. In addition to these quality-of-service (QoS) and design problems, the GFS methodology does not support generic priority pref- erence or pay-differentiated services for cloud applications, which public cloud providers may require under heavy loads. This paper presents a distributed mutual exclusion algorithm called Pri- oritizable Adaptive Distributed Mutual Exclusion (PADME) that we de- signed to meet the need for differentiated services between applications for file systems and other shared resources in a public cloud. We analyze the fault tolerance and performance of PADME and show how it helps cloud infrastructure providers expose differentiated, reliable services that scale. Results of experiments with a prototype of PADME indicate that it supports service differentiation by providing priority preference to cloud applications, while also ensuring high throughput. keywords: mutual exclusion, public cloud, QoS, file systems 1 Introduction The Google File System (GFS) was designed to support the sustained file through- put capacities of the Google search engine [1{3]. GFS provides high throughput in a single cluster of thousands of computers, each servicing the Google search engine.
    [Show full text]
  • Google: the World's First Information Utility?
    BISE – STATE OF THE ART Google: The World’s First Information Utility? The idea of information utilities came to naught in the 70s and 80s, but the world wide web, the Internet, mobile access, and a plethora of access devices have now made information utilities within reach. Google’s business model, IT infrastructure, and innovation strategies have created the capabilities needed for Google to become the world’s first global information utility. But, the difficulty of monetizing utility services and concerns about privacy, which could bring government regulation, might stymie Google’s growth plans. DOI 10.1007/s12599-008-0011-6 a national scale (Sackmann and Nie 1970; the number of users to the point where it The Authors Sackman and Boehm 1972). Not to be left is the dominant search player in both the out, the idea was also promoted by the U.S. and the world. This IT-based strategy Rex Chen Computer Usage Development Institute has allowed Google to reach more than Prof. Kenneth L. Kraemer, PhD in Japan, the British Post Office in the UK, $16 billion in annual revenue, a stock price Prakul Sharma and Bell Canada and the Telecommunica- over $600 per share and a market capital- University of California, Irvine tions Board in Canada (Press 1974). ization of nearly $200 billion. Center for Research on Information The idea was so revolutionary at the Over the next decade, Google plans to Technology and Organization (CRITO) time that at least one critic called for a extend its existing services to the wire- 5251 California Ave., Ste.
    [Show full text]
  • Scaling to Build the Consolidated Audit Trail
    Scaling to Build the Consolidated Audit Trail: A Financial Services Application of Google Cloud Bigtable Neil Palmer, Michael Sherman, Yingxia Wang, Sebastian Just (neil.palmer;michael.sherman;yingxia.wang;sebastian.just)@fisglobal.com FIS Abstract Google Cloud Bigtable is a fully managed, high-performance, extremely scalable NoSQL database service offered through the industry-standard, open-source Apache HBase API, powered by Bigtable. The Consolidated Audit Trail (CAT) is a massive, government-mandated database that will track every equities and options market event in the US financial industry over a six-year period. We consider Google Cloud Bigtable in the context of CAT, and perform a series of experiments measuring the speed and scalability of Cloud Bigtable. We find that Google Cloud Bigtable scales well and will be able to meet the requirements of CAT. Additionally, we discuss how Google Cloud Bigtable fits into the larger full technology stack fo the CAT platform. 1 Background 1.1 Introduction Google Cloud Bigtable1 is the latest formerly-internal Google But the step in between is the most challenging aspect of building technology to be opened to the public through Google Cloud Platform CAT: taking a huge amount of raw data from broker-dealers and stock [1] [2]. We undertook an analysis of the capabilities of Google Cloud exchanges and turning it into something useful for market regulators. Bigtable, specifically to see if Bigtable would fit into a proof-of-concept system for the Consolidated Audit Trail (CAT) [3]. In this whitepaper, The core part of converting this mass of raw market events into we provide some context, explain the experiments we ran on Bigtable, something that regulators can leverage is presenting it in a logical, and share the results of our experiments.
    [Show full text]