Oracle Nosql Database Compared to Cassandra and Hbase

Total Page:16

File Type:pdf, Size:1020Kb

Oracle Nosql Database Compared to Cassandra and Hbase Oracle NoSQL Database Compared to Cassandra and HBase Overview . Oracle NoSQL Database is licensed under AGPL while Cassandra and HBase are Apache 2.0 licensed. Oracle NoSQL Database is in many respects, as a NoSQL Database implementation leveraging BerkeleyDB in its storage layer, a commercialization of the early NoSQL implementations which lead to the adoption of this category of technology. Several of the earliest NoSQL solutions were based on BerkeleyDB and some are still to this day e.g. LinkedIn’s Voldemort. The Oracle NoSQL Database is a Java based key-value store implementation that supports a value abstraction layer currently implementing Binary and JSON types. Its key structure is designed in such a way as to facilitate large scale distribution and storage locality with range based search and retrieval. The implementation uniquely supports built in cluster load balancing and a full range of transaction semantics from ACID to relaxed eventually consistent. In addition, the technology is integrated with important open source technologies like Hadoop / MapReduce, an increasing number of Oracle software solutions and tools and can be found on Oracle Engineered Systems. Cassandra is a key-value store that supports a single value abstraction known as table-structure. It uses partition based hashing over a ring based architecture where every node in the system can handle any read-write request, so nodes become coordinators of requests when they do not actually hold the data involved in the request operation. HBase is a key-value store that supports a single value abstraction known as table-structure ( popularly referred to as column family ). It is based on the Google Big Table design and is written entirely in Java. HBase is designed to work on top of the HDFS file system. Unlike Hive, HBase does not use MapReduce in its implementation, but accesses HDFS storage blocks directly and storing a natively managed file type. The physical storage is similar to a column oriented database and as such works particularly well for queries involving aggregations, similar to the shared nothing analytic databases AsterData, GreenPlum, etc. Comparison The table below gives a high level comparison of Oracle NoSQL Database and Cassandra features/capabilities. Low level details are found in links to Oracle and Cassandra online documentation. Point HBase Cassandra ONDB Cassandra is based on DynamoDB (Amazon). Initially developed at ONDB is based Oracle Facebook by former Amazon HBase is based Berkeley DB Java Edition a Foundatio engineers. This is one reason why on BigTable mature log-structured, high ns Cassandra supports multi data (Google) performance, transactional center. Rackspace is a big database. contributor to Cassandra due to multi data center support. HBase uses the Hadoop Infrastructure (Zookeeper, Cassandra started and evolved NameNode, separate from Hadoop and its HDFS). infrastructure and Operational ONDB has simple infrastructure Organizations knowledge requirements are requirements and does not use Infrastruct that will different than Hadoop. However, Zookeeper. Hadoop based ure deploy Hadoop for analytics, many Cassandra analytics are supported via a anyway may deployments use Cassandra + ONDB/Hadoop connector. be comfortable Storm (which uses Zookeeper), with leveraging and/or Cassandra + Hadoop. Hadoop knowledge by using HBase The HBase- ONDB uses a single node type Hadoop to store data and satisfy read Infrastructure Cassandra uses a a single Node- requests. Any node can accept a has several type. All nodes are equal and request and forward it if Infrastruct "moving parts" perform all functions. Any Node necessary. There is no SPOF. In ure consisting of can act as a coordinator, ensuring addition, there is a simple Simplicity Zookeeper, no SPOF. Adding Storm or watchdog process (the Storage and SPOF Name Node, Hadoop, of course, adds Node Agent or SNA for short) Hbase Master, complexity to the infrastructure. on each machine to ensure high and Data availability and automatically Nodes, restart any data storage node in Zookeeper is case of process level failures. clustered and The SNA also helps with naturally fault administration of the store. tolerant. Name Node needs to be clustered to be fault tolerant. HBase is optimized for reads, Cassandra has excellent single- supported by row read performance as long as ONDB provides: 1) Strict single-write eventual consistency semantics are consistency reads at the master master, and sufficient for the use-case. 2) eventual consistency reads, resulting strict Cassandra quorum reads, which with optional time constraints consistency are required for strict consistency Read on the recency of data and 3) model, as well will naturally be slower than Intensive application level Read your as use of Hbase reads. Cassandra does not Use Cases writes consistency. All reads Ordered support Range based row-scans contact just a single storage Partitioning which may be limiting in certain node making read operations which supports use-cases. Cassandra is well suited very efficient. ONDB also row-scans. for supporting single-row queries, supports range based scans. HBase is well or selecting multiple rows based suited for on a Column-Value index. doing Range based scans. HBase provides for asynchronous replication of Cassandra Random Partitioning an HBase provides for row-replication of a Cluster across single row across a WAN, either a WAN. HBase Multi-Data asynchronous (write.ONE, clusters cannot Center write.LOCAL_QUORUM), or be set up to [ Release 3.0 provides for Support synchronous (write.QUORUM, achieve zero asynchronous cascaded and write.ALL). Cassandra clusters RPO, but in replication across data centers. ] Disaster can therefore be set up to achieve steady-state Recovery zero RPO, but each write will HBase should require at least one wan-ACK be roughly back to the coordinator to achieve failover- this capability. equivalent to any other DBMS that relies on asynchronous replication over a WAN. Fall-back processes and procedures (e.g. after failover) are TBD. Writes are replicated in a pipeline fashion: the first-data-node for the region persists the write, and then sends the write to the next Natural ONDB considers a request with Cassandra's coordinators will send Endpoint, and ReplicaAckPolicy.NONE (the parallel write-requests to all so-on in a ONDB equivalent of Natural Endpoints, The pipeline Write.ONE) as having coordinator will "ack" the write Write.ON fashion. completed after the change has after exactly one Natural Endpoint E HBase’s been written to the master's log has "acked" the write, which Durability commit log buffer; the change is propagated means that node has also persisted "acks" a write to the other members of the the write to its WAL. The writes only after *all* replication group, via an may or may not have committed to of the nodes in efficient asynchronous stream- any other Natural Endpoint. the pipeline based protcol. have written the data to their OS buffers. The first Region Server in the pipeline must also have persisted the write to its WAL. HBase only Cassandra officially supports ONDB only supports random Ordered supports Ordered Partitioning, but no partitioning. Prevailing Partitionin Ordered production user of Cassandra uses experience indicates that other g Partitoning. Ordered Partitioning due to the forms of partioning are really This means "hot spots" it creates and the hard to administer in practice. that Rows for a operational difficulties such hot- CF are stored spots cause. Random Partitioning in RowKey is the only recommended order in Cassandra partitioning scheme, HFiles, where and rows are distributed across all each Hfile nodes in the cluster. contains a "block" or "shard" of all the rows in a CF. HFiles are distributed across all data- nodes in the Cluster Because of ordered partitioning, HBase queries can be formulated Because of random partitioning, with partial partial rowkeys cannot be used start and end with Cassandra. RowKeys must be ONDB range requests can be RowKey row-keys, and known exactly. Counting rows in a defined with partial start and Range can locate rows CF is complicated. It is highly end row-keys. The start and end Scans inclusive-of, or recommended that for these types row-keys in a range-scan need exclusive of of use-cases, data should be stored not exist in the store. these partial- in columns in Cassandra, not in rowkeys. The rows. start and end row-keys in a range-scan need not even exist in Hbase. There are no limits on range Due to Ordered scans across major or minor Partitioning, If data is stored in columns in keys. Range scans across major Linear HBase will Cassandra to support range scans, keys require access to each Scalability easily scale the practical limitation of a row shard in the store. Release 3 will for large horizontally size in Cassandra is 10's of support major key and index tables and while still Megabytes. Rows larger than that range scans that are parallelized range supporting causes problems with compaction across all the nodes in the store. scans rowkey range overhead and time. Minor key scans are serviced by scans. the single shard that contains the data associated with the minor key range. Cassandra does not support Atomic Compare and Set. HBase Counters require dedicated counter supports column-families which because of ONDB supports atomic Atomic eventual-consistency requires that compare and set, making it Atomic Compare and all replicas in all natural end- simple to implement counters. Compare Set. HBase points be read and updated with ONDB also supports atomic and Set supports ACK. However, hinted-handoff modification of multiple minor supports mechanisms can make even these key/value pairs under the same transaction built-in counters suspect for major key. within a Row. accuracy. FIFO queues are difficult (if not impossible) to implement with Cassandra. Hbase does not support Read Load Balancing against a single row.
Recommended publications
  • Combined Documents V2
    Outline: Combining Brainstorming Deliverables Table of Contents 1. Introduction and Definition 2. Reference Architecture and Taxonomy 3. Requirements, Gap Analysis, and Suggested Best Practices 4. Future Directions and Roadmap 5. Security and Privacy - 10 Top Challenges 6. Conclusions and General Advice Appendix A. Terminology Glossary Appendix B. Solutions Glossary Appendix C. Use Case Examples Appendix D. Actors and Roles 1. Introduction and Definition The purpose of this outline is to illustrate how some initial brainstorming documents might be pulled together into an integrated deliverable. The outline will follow the diagram below. Section 1 introduces a definition of Big Data. An extended terminology Glossary is found in Appendix A. In section 2, a Reference Architecture diagram is presented followed by a taxonomy describing and extending the elements of the Reference Architecture. Section 3 maps requirements from use case building blocks to the Reference Architecture. A description of the requirement, a gap analysis, and suggested best practice is included with each mapping. In Section 4 future improvements in Big Data technology are mapped to the Reference Architecture. An initial Technology Roadmap is created on the requirements and gap analysis in Section 3 and the expected future improvements from Section 4. Section 5 is a placeholder for an extended discussion of Security and Privacy. Section 6 gives an example of some general advice. The Appendices provide Big Data terminology and solutions glossaries, Use Case Examples, and some possible Actors and Roles. Big Data Definition - “Big Data refers to the new technologies and applications introduced to handle increasing Volumes of data while enhancing data utilization capabilities such as Variety, Velocity, Variability, Veracity, and Value.” The key attribute is the large Volume of data available that forces horizontal scalability of storage and processing and has implications for all the other V-attributes.
    [Show full text]
  • Oracle Nosql Database
    An Oracle White Paper November 2012 Oracle NoSQL Database Oracle NoSQL Database Table of Contents Introduction ........................................................................................ 2 Technical Overview ............................................................................ 4 Data Model ..................................................................................... 4 API ................................................................................................. 5 Create, Remove, Update, and Delete..................................................... 5 Iteration ................................................................................................... 6 Bulk Operation API ................................................................................. 7 Administration .................................................................................... 7 Architecture ........................................................................................ 8 Implementation ................................................................................... 9 Storage Nodes ............................................................................... 9 Client Driver ................................................................................. 10 Performance ..................................................................................... 11 Conclusion ....................................................................................... 12 1 Oracle NoSQL Database Introduction NoSQL databases
    [Show full text]
  • Oracle® Nosql Database Changelog
    Oracle® NoSQL Database Changelog Release 20.1 E91819-17 July 2020 Oracle NoSQL Database Changelog, Release 20.1 E91819-17 Copyright © 2011, 2020, Oracle and/or its affiliates. This software and related documentation are provided under a license agreement containing restrictions on use and disclosure and are protected by intellectual property laws. Except as expressly permitted in your license agreement or allowed by law, you may not use, copy, reproduce, translate, broadcast, modify, license, transmit, distribute, exhibit, perform, publish, or display any part, in any form, or by any means. Reverse engineering, disassembly, or decompilation of this software, unless required by law for interoperability, is prohibited. The information contained herein is subject to change without notice and is not warranted to be error-free. If you find any errors, please report them to us in writing. If this is software or related documentation that is delivered to the U.S. Government or anyone licensing it on behalf of the U.S. Government, then the following notice is applicable: U.S. GOVERNMENT END USERS: Oracle programs (including any operating system, integrated software, any programs embedded, installed or activated on delivered hardware, and modifications of such programs) and Oracle computer documentation or other Oracle data delivered to or accessed by U.S. Government end users are "commercial computer software" or “commercial computer software documentation” pursuant to the applicable Federal Acquisition Regulation and agency-specific supplemental regulations. As such, the use, reproduction, duplication, release, display, disclosure, modification, preparation of derivative works, and/or adaptation of i) Oracle programs (including any operating system, integrated software, any programs embedded, installed or activated on delivered hardware, and modifications of such programs), ii) Oracle computer documentation and/or iii) other Oracle data, is subject to the rights and limitations specified in the license contained in the applicable contract.
    [Show full text]
  • Oracle Nosql Database and Cisco- Collaboration That Produces Results
    Oracle NoSQL Database and Cisco- Collaboration that produces results 1 Copyright © 2011, Oracle and/or its affiliates. All rights reserved. What is Big Data? SOCIAL BLOG SMART METER VOLUME VELOCITY VARIETY VALUE 2 Copyright © 2011, Oracle and/or its affiliates. All rights reserved. Why Is It Important? US HEALTH CARE US RETAIL MANUFACTURING GLOBAL PERSONAL EUROPE PUBLIC LOCATION DATA SECTOR ADMIN Increase industry Increase net Decrease dev., Increase service Increase industry value per year by margin by assembly costs by provider revenue by value per year by $300 B 60+% –50% $100 B €250 B “In a big data world, a competitor that fails to sufficiently develop its capabilities will be left behind.” 3 Copyright © 2011, Oracle and/or its affiliates. All rights reserved. Source: * McKinsey Global Institute: Big Data – The next frontier for innovation, competition and productivity (May 2011) Big Data in Action DECIDE ACQUIRE Make Better Decisions Using Big Data ANALYZE ORGANIZE 4 Copyright © 2011, Oracle and/or its affiliates. All rights reserved. Oracle Integrated Solution Stack DATA VARIETY HDFS HADOOP (MapReduce) In-DB Oracle Loader Mining for HADOOP Oracle NoSQL DB Oracle In-DB Exadata ‘R’ In-DB MapReduce OBIEE Oracle Data Analytics Advanced Oracle Database Integrator INFORMATION DENSITY ACQUIRE ORGANIZE ANALYZE DECIDE 5 Copyright © 2011, Oracle and/or its affiliates. All rights reserved. Big Data in Action DECIDE ACQUIRE Acquire all available, schema-based and non- relational data ANALYZE ORGANIZE 6 Copyright © 2011, Oracle and/or its affiliates. All rights reserved. Acquiring Big Data Challenge Process high volume, low- Application changes With sub-millisecond density information frequently Velocity from various data-sets 7 Copyright © 2011, Oracle and/or its affiliates.
    [Show full text]
  • Oracle Big Data SQL Release 4.1
    ORACLE DATA SHEET Oracle Big Data SQL Release 4.1 The unprecedented explosion in data that can be made useful to enterprises – from the Internet of Things, to the social streams of global customer bases – has created a tremendous opportunity for businesses. However, with the enormous possibilities of Big Data, there can also be enormous complexity. Integrating Big Data systems to leverage these vast new data resources with existing information estates can be challenging. Valuable data may be stored in a system separate from where the majority of business-critical operations take place. Moreover, accessing this data may require significant investment in re-developing code for analysis and reporting - delaying access to data as well as reducing the ultimate value of the data to the business. Oracle Big Data SQL enables organizations to immediately analyze data across Apache Hadoop, Apache Kafka, NoSQL, object stores and Oracle Database leveraging their existing SQL skills, security policies and applications with extreme performance. From simplifying data science efforts to unlocking data lakes, Big Data SQL makes the benefits of Big Data available to the largest group of end users possible. KEY FEATURES Rich SQL Processing on All Data • Seamlessly query data across Oracle Oracle Big Data SQL is a data virtualization innovation from Oracle. It is a new Database, Hadoop, object stores, architecture and solution for SQL and other data APIs (such as REST and Node.js) on Kafka and NoSQL sources disparate data sets, seamlessly integrating data in Apache Hadoop, Apache Kafka, • Runs all Oracle SQL queries without modification – preserving application object stores and a number of NoSQL databases with data stored in Oracle Database.
    [Show full text]
  • Oracle Nosql Database EE Data Sheet
    Oracle NoSQL Database 21.1 Enterprise Edition (EE) Oracle NoSQL Database is a multi-model, multi-region, multi-cloud, active-active KEY BUSINESS BENEFITS database, designed to provide a highly-available, scalable, performant, flexible, High throughput and reliable data management solution to meet today’s most demanding Bounded latency workloads. It can be deployed in on-premise data centers and cloud. It is well- Linear scalability suited for high volume and velocity workloads, like Internet of Things, 360- High availability degree customer view, online contextual advertising, fraud detection, mobile Fast and easy deployment application, user personalization, and online gaming. Developers can use a single Smart topology management application interface to quickly build applications that run in on-premise and Online elastic configuration cloud environments. Multi-region data replication Enterprise grade software Applications send network requests against an Oracle NoSQL data store to and support perform database operations. With multi-region tables, data can be globally distributed and automatically replicated in real-time across different regions. Data can be modeled as fixed-schema tables, documents, key-value pairs, and large objects. Different data models interoperate with each other through a single programming interface. Oracle NoSQL Database is a sharded, shared-nothing system which distributes data uniformly across multiple shards in a NoSQL database cluster, based on the hashed value of the primary keys. An Oracle NoSQL Database data store is a collection of storage nodes, each of which hosts one or more replication nodes. Data is automatically populated across these replication nodes by internal replication mechanisms to ensure high availability and rapid failover in the event of a storage node failure.
    [Show full text]
  • An Intelligent Approach for Handling Complexity by Migrating from Conventional Databases to Big Data
    S S symmetry Article An Intelligent Approach for Handling Complexity by Migrating from Conventional Databases to Big Data Shabana Ramzan 1, Imran Sarwar Bajwa 1,* and Rafaqut Kazmi 2 1 Department of Computer Science & IT, Islamia University of Bahawalpur, Bahawalpur 63100, Pakistan; [email protected] 2 School of Computing, University of Technology Malaysia, Johor 81310, Malaysia; [email protected] * Correspondence: [email protected] Received: 26 October 2018; Accepted: 14 November 2018; Published: 3 December 2018 Abstract: Handling complexity in the data of information systems has emerged into a serious challenge in recent times. The typical relational databases have limited ability to manage the discrete and heterogenous nature of modern data. Additionally, the complexity of data in relational databases is so high that the efficient retrieval of information has become a bottleneck in traditional information systems. On the side, Big Data has emerged into a decent solution for heterogenous and complex data (structured, semi-structured and unstructured data) by providing architectural support to handle complex data and by providing a tool-kit for efficient analysis of complex data. For the organizations that are sticking to relational databases and are facing the challenge of handling complex data, they need to migrate their data to a Big Data solution to get benefits such as horizontal scalability, real-time interaction, handling high volume data, etc. However, such migration from relational databases to Big Data is in itself a challenge due to the complexity of data. In this paper, we introduce a novel approach that handles complexity of automatic transformation of existing relational database (MySQL) into a Big data solution (Oracle NoSQL).
    [Show full text]
  • Oracle Big Data Appliance X8-2
    ORACLE DATA SHEET Oracle Big Data Appliance X8-2 Oracle Big Data Appliance is a flexible, high-performance, secure platform for running diverse workloads on Hadoop, Kafka and Spark. With Oracle Big Data SQL, Oracle Big Data Appliance extends Oracle’s industry-leading implementation of SQL to Hadoop/NoSQL and Kafka systems. By combining the newest technologies from the Hadoop ecosystem and powerful Oracle SQL capabilities together on a single pre-configured platform, Oracle Big Data Appliance is uniquely capable to support rapid development of new Big Data applications and tight integration with existing relational data. Oracle Big Data Appliance X8-2 Oracle Big Data Appliance is an open, multi-purpose engineered system for Hadoop and Spark workloads and streaming data processing. Big Data Appliance is designed to run diverse workloads – from Hadoop-only workloads (Yarn, Spark, Hive etc.) to interactive, all-encompassing interactive SQL queries using Oracle Big Data SQL across Apache Kafka, Hadoop and NoSQL databases. Big Data Appliance, is a Cloudera Certified platform and supports both Cloudera 5.x and Cloudera 6.x. Big Data Appliance provides an open environment for innovation while maintaining tight integration and enterprise-level support. Organizations can deploy external software to KEY FEATURES support new functionality – such as graph analytics, natural language processing and • Massively scalable, open infrastructure to store, analyze and fraud detection. Support for non-Oracle components is delivered by their respective manage big data support channels and not by Oracle. • Industry-leading security, performance and the most Lower TCO and Faster Time to Value comprehensive big data tool set on the market all bundled in an easy to Big Data Appliance provides unique pricing to offer both a lower initial deployment cost deploy appliance as well as a dramatically reduced three and four-year TCO when compared to a Do-It- • Flexible configuration and elastic Yourself Hadoop, Spark or Kafka system.
    [Show full text]
  • Multimodel Database with Ora
    Disclaimer The following is intended to outline our general product direction. It is intended for information purposes only, and may not be incorporated into any contract. It is not a commitment to deliver any material, code, or functionality, and should not be relied upon in making purchasing decisions. The development, release, and timing of any features or functionality described for Oracle’s products remains at the sole discretion of Oracle. MULTIMODEL DATABASE WITH ORACLE DATABASE 18C Table of Contents Introduction 1 Multimodel Database Architecture 2 Multimodel Database Features in Oracle 18c 3 JSON in Oracle Database 5 Graph Database and Analytics in Oracle Spatial and Graph 6 Property Graph Features in Oracle Spatial and Graph 6 RDF Semantic Graph Triple Store Features in Oracle Spatial and Graph 7 Spatial Database and Analytics in Oracle Spatial and Graph 7 Sharded Database Model 8 Oracle XML DB 9 Oracle Text 10 Oracle SecureFiles 10 Storage Optimization in SecureFiles 10 SecureFiles Features in Oracle Database 18c 11 Conclusion 12 0 | MULTIMODEL DATABASE WITH ORACLE DATABASE 18C Introduction Over the nearly 40 years in the evolution of commercial relational database management systems, a consistent pattern has emerged as the capabilities, data types, analytics, and data models have been developed and adopted. With each new generation of computing architecture – from centralized mainframe, to client server, to internet computing, to the Cloud – new generations of data management systems have been developed to address new applications, workloads and workflows. Today, the successful operation of corporations, enterprises, and other organizations relies on the management, understanding and efficient use of vast amounts of unstructured Big Data that may come from social media, web content, sensors and machine output, and documents.
    [Show full text]
  • Scalable Storage: the Drive for Web-Scale Data Management
    Scalable Storage: The drive for web-scale data management Bryan Rosander University of Central Florida [email protected] March 28, 2012 Abstract Data-intensive applications have become prevalent in todays information econ- omy. The sheer amount of data stored and utilized by todays web services presents unique challenges in the areas of scalability, security, and availability. This has opened new possibilities in data mining, allowing for more tightly integrated, in- formative services. It has also created new challenges. Traditional, monolithic, relational databases are inherently limited in terms of scalability. This has caused many leading companies to abandon traditional databases in favor of horizontally scalable data stores. This paper will evaluate the state of the art in data stor- age and retrieval, covering the history of the database and moving on to newer database technologies such as Googles Bigtable, Apache Cassandra, and Amazons DynamoDB. 1 Introduction Data storage and retrieval has become a central part of many popular web applications. As the amount of data available increases, the database capacity must scale up to meet it. Traditional methods of scaling up database capacity focus mainly on increasing the computing power of the single server on which the database resides. This strategy has been sufficient for many applications but has become infeasible for those that need to store more data than can be efficiently processed by one machine. Newer database paradigms emphasizing horizontal scalability, the ability to add as many nodes as are necessary and redistribute the data between all active nodes, have been growing in popularity. This increase in scalability does come at a cost.
    [Show full text]
  • Oracle Paas and Iaas Universal Credits Service Descriptions
    Oracle PaaS and IaaS Universal Credits Service Descriptions Effective Date: 10-September-2021 Oracle UCM 091021 Page 1 of 202 Table of Contents metrics 6 Oracle PaaS and IaaS Universal Credit 20 1. AVAILABLE SERVICES 20 a. Eligible Oracle PaaS Cloud Services 20 b. Eligible Oracle IaaS Cloud Services 20 c. Additional Services 20 d. Always Free Cloud Services 21 Always Free Cloud Services 22 2. ACTIVATION USAGE AND BILLING 23 a. Introduction 23 i. Annual Universal Credit 24 Overage 24 Replenishment of Account at End of Services Period 25 Additional Services 25 ii. Monthly Universal Credit (subject to Oracle approval) 25 Overage 26 Orders Placed via a Partner 26 Replenishment of Account at End of Services Period 26 iii. Pay as You Go 26 iv. Funded Allocation Model 27 Overage 27 Additional Services 28 Replenishment of Account at End of Services Period 28 3. INCLUDED SERVICES 28 i. Developer Cloud Service 28 ii. Oracle Identity Foundation Cloud Service 29 b. Additional Licenses and Oracle Linux Technical Support 29 c. Oracle Cloud Infrastructure Data Catalog 30 d. Oracle Cloud Infrastructure Data Transfer Disk 30 Your Obligations/Responsibilities and Project Assumptions 30 Your Obligations/Responsibilities 31 Project Assumptions 31 Export 32 Oracle Cloud Infrastructure - Application Migration 32 f. Oracle Cloud Infrastructure Console 33 g. Oracle Cloud Infrastructure Cloud Shell 33 Access and Usage 33 4. SERVICES AVAILABLE VIA THE ORACLE CLOUD MARKETPLACE 33 a. Oracle Cloud Services Delivered via the Oracle Cloud Marketplace 33 b. Third Party
    [Show full text]
  • Data Sheet: Berkeley Database Products
    ORACLE DATA SHEET Oracle Berkeley Database Products The Oracle Berkeley DB product family consists of Berkeley DB, Berkeley DB Java Edition and Berkeley DB XML. All three are high performance, self- contained, software libraries which provide data storage services for applications, devices, and appliances. They deliver superior performance, scalability and availability for applications that must run unattended without administration. Overview The Oracle Berkeley DB family of high performance, self-contained databases provides CONSIDER BERKELEY DB WHEN: developers with a fast, transactional database solution with a track record of reliability, • You need an SQLite API compatible unmatched scalability and five-nines (99.999%) or better availability. Oracle Berkeley database within mobile, handheld or DB is well suited to Independent Software Vendors, device and equipment other hardware devices. manufacturers, and enterprises or software companies building solutions which need a • Performance, scalability, concurrency data management component. The Oracle Berkeley DB family of products provides fast, is important. local persistence with zero oversight administration. • Zero oversight administration in deployment is required Customers and end-users will experience an application that simply works, reliably • Flexibility to choose SQL, XQuery, manages data, scales under extreme load, and requires zero oversight in deployment. Java Object, or Key/Value data Your development team can focus on your application and be confident that Berkeley management. DB will manage your application’s data. • Mobile data synchronization with Oracle Database is a requirement. Storage Engine Design • Failure from recovery must be automatic and reliable The Berkeley DB products are self-contained software components which support your • High availability and fault tolerance are application.
    [Show full text]