Translattice Elastic Database (TED) a Modern Database for Proactive Data Management: Translattice Elastic Database (TED) White Paper

Total Page:16

File Type:pdf, Size:1020Kb

Translattice Elastic Database (TED) a Modern Database for Proactive Data Management: Translattice Elastic Database (TED) White Paper WHITE PAPER A Modern Database for Proactive Data Management: TransLattice Elastic Database (TED) A Modern Database for Proactive Data Management: TransLattice Elastic Database (TED) White Paper Introduction Businesses today face more challenges than ever before in maintaining correct business and customer information. Mergers and acquisitions make maintaining accurate and complete data an expensive and time-consuming activity. Today’s emphasis on regulatory compliance forces many global businesses to maintain data silos in various jurisdictions. Data virtualization, data federation, or other means are then employed to get a 360-degree view of their business. When changes occur, they must ripple through the various layers of data infrastructure. Master data management solutions are often employed to help ensure data accuracy. All this is time-consuming, can lead to costly mistakes, and customer dissatisfaction. And scaling these systems can be a daunting task. New approaches in data management can greatly simplify and streamline these challenges. A Revolutionary Approach TransLattice Elastic Database (TED) provides a dramatically different approach to data management. TED is the world’s !rst geographically distributed relational database management system (RDBMS) that incorporates data compliance capabilities directly in the database. TED is designed to support enterprise business applications such as ERP, SCM, and CRM systems. It provides all the transactional and ACID guarantees expected in a traditional SQL database. However, nodes can be placed thousands of miles apart, over standard Internet connections, anywhere you have business presence. These nodes form a high-performance computing fabric built from all-active commodity servers, cloud instances, and/or virtual machines, called a cluster. Policy-driven Data Placement Data policies can be de!ned to ensure that sensitive data or other data subject to local governance restrictions is stored only in permissible locations. Data on any node can be visible from all nodes. Updates to information made in India are immediately visible in New York. Personally Identi!able Information (PII) data safely stored on nodes in the European Union is continually available to your billing systems in Atlanta. A A C C B C Policy PII data cannot Publicly available Certain data on A leave the EU B data OK for cloud C 3 continents TED ensures that data remains available and organizations remain in compliance with data jurisdiction regulations. 2 A Modern Database for Proactive Data Management: TransLattice Elastic Database (TED) White Paper TED is designed to give organizations peace of mind. Data is split into small segments and automatically stored redundantly across multiple nodes in the system. If a node is lost due to hardware or network failure, the data remains available from other nodes within the cluster. Administrators can control the level of redundancy to ensure availability. Policy controls also allow administrators to direct data placement to improve performance. For example, policies can dictate that records associated with customers on the West Coast be stored on nodes on the West Coast. This reduces network traf!c and improves the end user experience for users accessing the data. Additional copies can be maintained automatically in Atlanta nodes to ensure the billing systems run at peak performance. Changes made on any node are immediately re"ected and available by querying any node in the cluster. There is no need to reconcile data differences between East Coast systems and West Coast systems because the system appears to all users as a single system. Proactive Cloud Data Location Control Policy controls also give peace of mind in the cloud. Leveraging the cloud is a great way to get burst capacity fast. But with traditional systems, it’s not that easy. TransLattice changes that. Cloud nodes can be added to an existing cluster in minutes, providing additional processing capacity and immediate visibility to existing data sets. Policy controls can be used to control just which data can or cannot be stored in cloud nodes giving business the con!dence to move boldly into the cloud. The ability to add nodes in minutes, whether cloud, virtual machine, or appliances, makes scaling systems a turnkey operation. Conclusion Both large and small organizations must face the complex challenges of regulatory compliance, mergers and acquisitions, and scaling data systems in order to keep pace with their growth. The TransLattice Elastic Database is a new and innovative approach to proactive data management. TED is designed to help organizations simplify compliance issues, eliminate costly master data management and data virtualization processes, take advantage of cloud computing, and transform the scaling of business systems from torture to a turn-key operation. Corporate Headquarters: TransLattice, Inc. | 2900 Gordon Avenue, Santa Clara, CA | phone: (408) 749-8478 email: [email protected] | translattice.com © 2013 TransLattice, Inc. All Rights Reserved. TransLattice and the TransLattice logo are property of TransLattice, Inc. in the United States and other countries. Part # 9800-0101-01 3.
Recommended publications
  • Towards Full ACID Distributed Transactions Support with Foreign Data Wrapper
    Event / Conference name Location, Date Towards full ACID Distributed Transactions Support with Foreign Data Wrapper Nikhil Sontakke https://www.2ndQuadrant.com PGConf India 2020 WhoAmI? • Working on PostgreSQL code since 2006 • Part of a few startups - OyeStyle - StormDB (acquired by Translattice, Inc.) - SecureDB •Co-Organizer of PGConf India :-) • PostgreSQL Consultant, 2ndQuadrant https://www.2ndQuadrant.com PGConf India 2020 What is Foreign Data Wrapper? • Implementation of the SQL/MED • Access data that resides outside PostgreSQL using regular SQL queries • Foreign tables • Pluggable architecture https://www.2ndQuadrant.com PGConf India 2020 Foreign Data Wrapper PostgreSQL postgres_fdw mysql_fdw Client PostgreSQL MySQL oracle_fdw Oracle Database https://www.2ndQuadrant.com PGConf India 2020 Foreign Data Wrapper in details PostgreSQL PostgreSQL SELECT * SELECT id, name Users FROM users Users (foreign table) FROM users; ID Name Client ID Name 100 Alice 200 Bob 300 Charlie 400 Dave https://www.2ndQuadrant.com PGConf India 2020 Foreign Data Wrapper with Reads • Most of the operations can be pushed down to the remote server • Joins • Aggregations • Sorts etc https://www.2ndQuadrant.com PGConf India 2020 Foreign Data Wrapper with Writes • Writable foreign tables • PostgreSQL version 9.3 and later https://www.2ndQuadrant.com PGConf India 2020 Transactions: Anomalies The phenomena due to concurrent transactions which are prohibited at various levels are: ● Dirty Read A transaction reads data written by a concurrent uncommitted transaction. ● Non-repeatable Read A transaction re-reads data it has previously read and finds that data has been modified by another transaction (that committed since the initial read). ● Phantom Read A transaction re-executes a query returning a set of rows that satisfy a search condition and finds that the set of rows satisfying the condition has changed due to another recently-committed transaction.
    [Show full text]
  • PGCONF.EU an Illustrated History of Postgresql
    AN ILLUSTRATED HISTORY OF POSTGRES(QL) WHO AM I? PETER VAN HARDENBERG ▸ Formerly Heroku Postgres, where we ran >1M Postgres databases. ▸ Alright at writing SQL. ▸ Started with PostgreSQL in 2001 for arctic oceanography data. ▸ Born in Canada, live in San Francisco. ▸ @pvh / [email protected] INTRODUCTION MEET POSTGRES AN APOLOGY THERE ARE TOO MANY STORIES FOR ONE TALK CHAPTER ONE STONEBRAKER'S POSTGRES BERKELEY, CALIFORNIA, 1976 MICHAEL STONEBRAKER IS WORRIED ONCE YOU GET IN, YOU'VE GOT FIVE YEARS TO GET FAMOUS, THAT'S IT. Michael Stonebraker EDGAR F. CODD RELATIONAL ALGEBRA INTERACTIVE GRAPHICS RETRIEVAL SYSTEM INGRES range of E is EMPLOYEE retrieve into W (COMP = E.Salary / (E.Age - 18)) where E.Name = "Jones" QUEL (QUERY LANGUAGE) SOLD TO COMPUTER ASSOCIATES INGRES STONEBRAKER'S POSTGRES BERKELEY RESEARCH LABS ▸ Interdisciplinary groups ▸ Lots of collaboration ▸ Limited duration ▸ Real-world applications ▸ Path to commercialization ▸ BSD license (usually) ▸ Examples: Ingres, Postgres, RISC, Spark, and many more. WHAT COMES POST INGRES? POST-GRES. STAKE OUT A CONTROVERSIAL POSITION, AND PROVE IT CORRECT. Michael Stonebraker STONEBRAKER AND ROWE THE DESIGN OF POSTGRES STONEBRAKER'S POSTGRES THE DESIGN OF POSTGRES, 1986 ▸ Designed to combine applications & traditional business data ▸ Response to Ingres' shortcomings ▸ Lots of extensibility ▸ Adds a "rules engine" ▸ (predates modern constraints) THE UNIVERSITY ERA 1987-1994 WE WOULD RECRUIT THE SMARTEST [STUDENTS] WE COULD FIND, GIVE THEM WONDERFUL EQUIPMENT, AND THEY WOULD BASICALLY DIE WRITING CODE FOR US. Michael Stonebraker POSTGRES 1.0 IMPLEMENTED 1987-1989 POSTGRES 2.0 REWRITES RULES 1990 POSTGRES 3.0 REWRITES RULES AGAIN 1991 STONEBRAKER FOUNDS ILLUSTRA TO COMMERCIALIZE POSTGRES 1992 POSTGRES 4.2 BERKELEY POSTGRES LAB ENDS 1994 POSTGRES RESEARCH PROJECT OVER THE END ..
    [Show full text]
  • Recent Trends in Data Persistence and Analysis
    SQL Strikes Back Recent Trends in Data Persistence and Analysis Codemesh 2014 November 4, 2014 Dean Wampler, Ph.D [email protected] ©Typesafe 2014 – All Rights Reserved Tuesday, November 4, 14 All photos are copyright (C) 2008 - 2014, Dean Wampler. All Rights Reserved. Photo: Before dawn above the Western USA [email protected] polyglotprogramming.com/talks @deanwampler ©Typesafe 2014 – All Rights Reserved 2 Tuesday, November 4, 14 Typesafe provides products and services for building Reactive, Big Data applications typesafe.com/reactive-big-data ©Typesafe 2014 – All Rights Reserved 3 Tuesday, November 4, 14 For @coderoshi... Cat meme! ©Typesafe 2014 – All Rights Reserved 4 Tuesday, November 4, 14 If you were at Eric Raymond’s (@coderoshi’s) talk... My cat Oberon. Three Trends ©Typesafe 2014 – All Rights Reserved 5 Tuesday, November 4, 14 Three trends to organizer our thinking… Photo: Dusk over the American Midwest, in Winter Data Size ⬆ ©Typesafe 2014 – All Rights Reserved 6 Tuesday, November 4, 14 Data volumes are obviously growing… rapidly. Facebook now has over 600PB (Petabytes) of data in Hadoop clusters! Formal Schemas ⬇ ©Typesafe 2014 – All Rights Reserved 7 Tuesday, November 4, 14 There is less emphasis on “formal” schemas and domain models, i.e., both relaonal models of data and OO models, because data schemas and sources change rapidly, and we need to integrate so many disparate sources of data. So, using relavely-agnos^c so_ware, e.g., collec^ons of things where the so_ware is more agnos^c about the structure of the data and the domain, tends to be faster to develop, test, and deploy.
    [Show full text]
  • © 2011 by the 451 Group. All Rights Reserved Nosql, Newsql and Beyond Open Source Driving Innovation in Distributed Data Management
    © 2011 by The 451 Group. All rights reserved NoSQL, NewSQL and Beyond Open Source Driving Innovation in Distributed Data Management Matt hew As lett, The 451 Group [email protected] © 2011 by The 451 Group. All rights reserved Overview – Open source impact in the database market – NoSQL and NewSQL databases • Adoption/development drivers • Use cases – And beyond • Data grid/cache • Cloud databases • Total data © 2011 by The 451 Group. All rights reserved 451 Research is focused on the business of enterprise IT innovation. The company’s analysts provide critical and timely insight into the competitive dynamics of innovation in emerging technology segments. The 451 Group Tier1 Research is a single‐source research and advisory firm covering the multi‐tenant datacenter, hosting, IT and cloud‐computing sectors, blending the best of industry and financial research. The Uptime Institute is ‘The Global Data Center Authority’ and a pioneer in the creation and facilitation of end‐user knowledge communities to improve reliability and uninterruptible availability in datacenter facilities. TheInfoPro is a leading IT advisory and research firm that provides real‐world ppperspectives on the customer and market dynamics of the TOTAL DATA enterprise information technology landscape, harnessing the collective knowledge and insight of leading IT organizations worldwide. ChangeWave Research is a research firm that identifies and quantifies ‘change’ in consumer spending behavior, corporate purchasing, and industry, company and technology trends. © 2011 by The 451 Group. All rights reserved Coverage areas – Senior analyst, enterprise software – Commercial Adoption of Open Source (CAOS) • Adoption by enterprises • Adoption by vendors – Information Management • Databases • Data warehousing • Data caching © 2011 by The 451 Group.
    [Show full text]
  • PGXC CTL Primer Configuring and Operating Postgres-XC Database Cluster May 7Th, 2014 Koichi Suzuki
    PGXC_CTL Primer Configuring and operating Postgres-XC database cluster May 7th, 2014 Koichi Suzuki Change History: May 7th, 2014: Initial version. Oct 2nd, 2014: Added license condition. This article is licensed under Creative Commons Attribution­NonCommercial­ShareAlike 4.0 International License. 1. Introduction This document is an outline of Postgres­XC database cluster configuration and operation using pgxc_ctl. Pgxc_ctl will be found in the contrib module of postgres­xc distribution. This module helps to configure and operate your Postgres­XC database cluster easily. This saves much of the small but important things. 2. Plannings you Postgres-XC cluster 2.1 Postgres-XC overview Before going into Postgres­XC configuration, please take a bit to consider if Postgres­XC cluster is useful to your application. Unlike any other PostgreSQL replication cluster, Postgres­XC is write­scalable PostgreSQL cluster. More servers you have, more throughput you get, both in read and write. Of course, you can handle bigger databsae if you have more servers and storages. To achieve this, you should consider each table to be distributed (sharded) or replicated. If tables are updated very frequently, you should consider them to be distributed. If they are relatively static and are refered by distributed table or other replicated tables, you should consider them to be replicated. This mixture will make most of the update local to one of the datanode. Because each datanode operates in paralle, you can get both read and write scalability. -1/59 - You can connect to any of the coordinators1 , each of which provides full transaction ACID capabiliy and transparent view to the database.
    [Show full text]
  • Scaling Postgresql on Postgres-XL
    Scaling PostgreSQL on Postgres-XL February, 2015 Mason Sharp PGConf.ru 1 Agenda • Intro • Postgres-XL Background • Architecture Overview • Distributing Tables • Configuring a Local Cluster • Differences to PostgreSQL • Community 2 whoami • Co-organizer of NYC PUG • I like database clusters • GridSQL • Postgres-XC • Postgres-XL • Work: • TransLattice • StormDB • EnterpriseDB 3 PGConf.US • March 25–27, 2015 • New York City • 44 Talks • 4 Tracks! 4 Postgres-XL n Open Source n Scale-out RDBMS – MPP for OLAP – OLTP write and read scalability – Custer-wide ACID properties – Multi-tenant security – Can also act as distributed key-value store 7 Postgres-XL n Hybrid Transactional and Analytical Processing (HTAP) n Can replace multiple systems 8 Postgres-XL n Rich SQL support – Correlated Subqueries 9 10 Postgres-XL Features • Large subset of PostgreSQL SQL • Distributed Multi-version Concurrency Control • Contrib-friendly • Full Text Search • JSON and XML Support • Distributed Key-Value Store 13 Postgres-XL Missing Features • Built-in High Availability • Use external like Corosync/Pacemaker • Area for future improvement • “Easy” to re-shard / add nodes • Some pieces there • Tables locked during redistribution • Can however “pre-shard” multiple nodes on the same server or VM • Some FK and UNIQUE constraints • Recursive CTEs (WITH) 14 Performance Transactional work loads Dell DVD Benchmark 18000 16000 14000 12000 StormDB 10000 Amazon RDS 8000 6000 4000 2000 0 10 50 100 200 300 400 15 OLTP Characteristics n Coordinator Layer adds about 30% overhead
    [Show full text]
  • Database Management Systems Ebooks for All Edition (
    Database Management Systems eBooks For All Edition (www.ebooks-for-all.com) PDF generated using the open source mwlib toolkit. See http://code.pediapress.com/ for more information. PDF generated at: Sun, 20 Oct 2013 01:48:50 UTC Contents Articles Database 1 Database model 16 Database normalization 23 Database storage structures 31 Distributed database 33 Federated database system 36 Referential integrity 40 Relational algebra 41 Relational calculus 53 Relational database 53 Relational database management system 57 Relational model 59 Object-relational database 69 Transaction processing 72 Concepts 76 ACID 76 Create, read, update and delete 79 Null (SQL) 80 Candidate key 96 Foreign key 98 Unique key 102 Superkey 105 Surrogate key 107 Armstrong's axioms 111 Objects 113 Relation (database) 113 Table (database) 115 Column (database) 116 Row (database) 117 View (SQL) 118 Database transaction 120 Transaction log 123 Database trigger 124 Database index 130 Stored procedure 135 Cursor (databases) 138 Partition (database) 143 Components 145 Concurrency control 145 Data dictionary 152 Java Database Connectivity 154 XQuery API for Java 157 ODBC 163 Query language 169 Query optimization 170 Query plan 173 Functions 175 Database administration and automation 175 Replication (computing) 177 Database Products 183 Comparison of object database management systems 183 Comparison of object-relational database management systems 185 List of relational database management systems 187 Comparison of relational database management systems 190 Document-oriented database 213 Graph database 217 NoSQL 226 NewSQL 232 References Article Sources and Contributors 234 Image Sources, Licenses and Contributors 240 Article Licenses License 241 Database 1 Database A database is an organized collection of data.
    [Show full text]
  • Postgresql 9 High Availability with Linux-HA Agenda
    PostgreSQL 9 High Availability With Linux-HA Agenda • Introduction • HA considerations • PostgreSQL HA evolution • Linux HA – components • PostgreSQL streaming replication + Linux HA recipe • DEMO! • Summary 2 Who am I? • Nikhil Sontakke • Responsible for the HA aspects of the Postgres-XL product • PostgreSQL community contributor • Co-organizer Indian PUG Meetup • Stints earlier at EnterpriseDB, StormDB, Translattice Inc. 3 HA - Definition • What is High Availability (HA): • HA is a “concept” • A percentage of time that a given system is providing service since it has been deployed • For example: A system is 99% available if the downtime is 4 days in a year • Everyone craves for the five 9s (downtime of less than 5 minutes in a year – 99.999%) • HA is NOT designed for high performance • HA is NOT designed for high throughput (aka load balancing) 4 HA – Why does it matter? • Why do we bother with HA: • Downtime is expensive • You miss out on earnings due to the downtime • You bother because your boss might complain ;) • Users might not return! 5 HA – wish our sysadmin is like this ;) xkcd.com/705 6 PostgreSQL – HA evolution • Log Shipping and Point In Time Recovery • PostgreSQL 8.1 • Base backup of the database • Write Ahead Logs (WAL) sent to the standby • Warm Standby • PostgreSQL 8.2 • Continuously apply WAL on the standby 7 PostgreSQL – HA evolution (contd...) • HA using Logical Replication • Slony (PG 7.3 onwards), Londiste, Bucardo • Trigger based replication systems • HA using statement based replication • Pgpool-II (PG 6.4 onwards) • Intercept SQL queries and send to multiple servers 8 PostgreSQL – HA evolution (contd...) • HA using Shared Storage • Sharing disk array between two servers • SAN environment needed (very expensive) • HA using Block-Device replication • All changes to a filesystem residing on a block device are replicated to a block device on another system • DRBD pretty popular 9 PostgreSQL – HA latest..
    [Show full text]
  • Databases Theoretical Introduction Contents
    Databases Theoretical Introduction Contents 1 Databases 1 1.1 Database ................................................ 1 1.1.1 Terminology and overview .................................. 1 1.1.2 Applications .......................................... 2 1.1.3 General-purpose and special-purpose DBMSs ........................ 2 1.1.4 History ............................................ 2 1.1.5 Research ........................................... 6 1.1.6 Examples ........................................... 6 1.1.7 Design and modeling ..................................... 7 1.1.8 Languages ........................................... 9 1.1.9 Performance, security, and availability ............................ 10 1.1.10 See also ............................................ 12 1.1.11 References .......................................... 12 1.1.12 Further reading ........................................ 13 1.1.13 External links ......................................... 14 1.2 Schema migration ........................................... 14 1.2.1 Risks and Benefits ...................................... 14 1.2.2 Schema migration in agile software development ...................... 14 1.2.3 Available Tools ........................................ 15 1.2.4 References .......................................... 15 1.3 Star schema .............................................. 16 1.3.1 Model ............................................ 16 1.3.2 Benefits ............................................ 16 1.3.3 Disadvantages .......................................
    [Show full text]
  • The Steadily Growing Database Market Is Increasing Enterprises
    For: Application The Steadily Growing Database Market Is Development & Delivery Increasing Enterprises’ Choices Professionals by Noel Yuhanna, June 7, 2013 KEY TAKEAWAYS Information Management Challenges Create New Database Market Opportunities Data volume, data velocity, mobility, cloud, globalization, and increased compliance requirements demand new features, functionality, and innovation. A recent Forrester survey found that performance, integration, security, unpredictable workloads, and high availability are companies’ top data management challenges. Look Beyond Traditional Database Solutions And Architectures The database market is transforming. Relational DBs are OK for established OLTP and decision support apps, but new business requirements, dynamic workloads, globalization, and cost-cutting mean that firms need a new database strategy. Devs facing issues with existing DB implementations or designing new web-scale apps should look beyond relational. Invest In New Database Technologies To Succeed Business data comprises an increasing share of the value of your systems. Past investments in relational DBs are a solid basis, but you need more to stay competitive amid the many innovative new ways to use information. Next-gen DB strategies require you to invest anew so you can deliver the speed, agility, and insights critical to business growth. Forrester Research, Inc., 60 Acorn Park Drive, Cambridge, MA 02140 USA Tel: +1 617.613.6000 | Fax: +1 617.613.5000 | www.forrester.com FOR APPLICATION DEVELOPMENT & DELIVERY PROFESSIONALS JUNE 7, 2013 The Steadily Growing Database Market Is Increasing Enterprises’ Choices An Overview Of The Enterprise Database Management System Market In 2013 by Noel Yuhanna with Mike Gilpin and Vivian Brown WHY READ THIS REPort Business demand for more applications, larger databases, and new analytics initiatives is driving innovation and growth in the database market, with new options for cloud, in-memory, mobility, predictive analytics, and real-time data management.
    [Show full text]
  • Citus Data Prepares Citusdb 4.0, Now a Massively Parallel Postgresql Analytic Database
    Citus Data prepares CitusDB 4.0, now a massively parallel PostgreSQL analytic database Analyst: Matt Aslett 9 Mar, 2015 Citus Data has changed its positioning since our last update, evolving CitusDB from being a scalable analytics database predominantly designed to bring SQL analytics to Hadoop to offering a stand-alone massively parallel columnar analytics database that is PostgreSQL-compatible. The 451 Take We noted that Citus Data was entering a crowded market in 2013, and therefore see the change of direction as a good thing. While the MPP analytic-database market is no less crowded, Citus Data is differentiated by its focus on extending, rather than forking, PostgreSQL. Making the cstore_fdw and pg-shard projects open source should grow the company's profile in the PostgreSQL user community and lay the foundation for potential CitusDB adoption. The competitive situation is likely to heat up, given that Pivotal's open source Greenplum strategy appears to be dependent on making it the default MPP choice for PostgreSQL, but we agree with Citus Data that Greenplum, having forked from PostgreSQL several years ago, will be a challenge. Either way, the PostgreSQL community will decide. Context The first time we encountered Citus Data, almost two years ago, the company had just launched CitusDB 2.0, bringing real-time SQL analytics to the Apache Hadoop data-processing framework. The SQL-on-Hadoop party was already in full swing, and got very crowded very quickly. While Citus Data had planned to differentiate itself by bringing SQL-based analytics to other nonrelational data Copyright 2015 - The 451 Group 1 platforms, including NoSQL databases, the company instead took a more radical change of direction, turning CitusDB into a stand-alone PostgreSQL-compatible massively parallel columnar analytics database.
    [Show full text]
  • Research Article the Art of Handling Databases by Cloud Computing
    Research Journal of Applied Sciences, Engineering and Technology 7(12): 2539-2544, 2014 DOI:10.19026/rjaset.7.564 ISSN: 2040-7459; e-ISSN: 2040-7467 © 2014 Maxwell Scientific Publication Corp. Submitted: August 19, 2013 Accepted: September 16, 2013 Published: March 29, 2014 Research Article The Art of Handling Databases by Cloud Computing 1R. Anandhi and 2K. Chitra 1Department of Computer Science, SCSVMV University, Kanchipuram, 2 Govt. Arts College, Melur, Madurai, Tamil Nadu, India Abstract: It is obvious that there are tremendous inventions and technological growth in IT industry. All upcoming technologies are almost trying to deal with only one concept “DATA” -how the data can be effectively stored, easily and accurately retrieved, efficiently distributed and queried etc. Hence one of the new such computing facility is Cloud Computing-accessing various computing resources that are geographically apart. It surely gets a significant attraction over the IT industry. This study explains and analyses the capabilities of Cloud Computing while dealing with databases such as Scalability, Availability, Consistency and Elasticity. Keywords: Cloud computing, cloud databases, database-as-a-service, elasticity, scalability INTRODUCTION traditional database system in cloud, since traditional databases are usually optimized to all sorts of queries Cloud computing provides accession to remotely and they obey consistency (Anandhi and Chitra, available computing resources over a network (the 2012b). Unfortunately, the assumptions of availability, Internet) (Maggiani, 2009). Users can buy these scalability and flexibility do not correspond to the cloud computing resources, such as storage and computing model since resources are allocated dynamically and power, as a utility on demand. The name ‘Cloud there exist a loose coupling between data management Computing’ is derived from the cloud-shaped symbol and systems (Chitra and Jeevarani, 2013).
    [Show full text]