Software As a Service for Data Scientists

Total Page:16

File Type:pdf, Size:1020Kb

Software As a Service for Data Scientists Doi:10.1145/2076450.2076468 Globus Online manages fire-and-forget file transfers for big-data, high-performance scientific collaborations. By BRyce Allen, John BResnahan, Lisa childeRs, ian fosTeR, GoPi KanDasWaMy, RaJ KettiMuThu, JacK Kordas, MiKe LinK, Stuart Martin, Karl PicKett, anD SteVen TuecKe software as a service for Data scientists As Big Data emerges as a force in science,2,3 so, too, do new, onerous tasks for researchers. Data from specialized instrumentation, numerical simulations, and downstream manipulations must be collected, indexed, archived, shared, replicated, and analyzed. These tasks are not new, but the complexities involved in performing them for terabyte or when data volumes were measured larger datasets (increasingly common in kilobytes. The result is a computa- across scientific disciplines) are quite tional crisis in many laboratories and different from those that applied a growing need for far more powerful data-management tools, yet the typi- key insights cal researcher lacks the resources and expertise to operate these tools. The costs of research data life-cycle The answer may be to deliver re- management are growing dramatically as search data-management capabili- data becomes larger and more complex. ties to users as hosted “software as a saas approaches are a promising service,” or SaaS,18 a software-delivery solution, outsourcing time-consuming model in which software is hosted research data management tasks to third-party services. centrally and accessed by users using a thin client (such as a Web browser) Globus online demonstrates the potential over the Internet. As demonstrated in of saas for research data management, simplifying data movement for research- many business and consumer tools, ers and research facilities alike. SaaS leverages intuitive Web 2.0 in- February 2012 | VoL. 55 | No. 2 | communicaTions of The acM 81 contributedarticles terfaces, deep domain knowledge, mand is issued; that is, third-party and economies of scale to deliver ca- transfers may be (indeed, frequently pabilities that are easier to use, more are) involved. Our goal in designing capable, and/or more cost-effective GO is a solution that provides extreme than software accessed through other ease-of-use without compromising means. The opportunity for continu- a common question reliability, speed, or security. A 2008 ous improvement via dynamic deploy- about Go is whether report by Childers et al.5 of Argonne ment of new features and bug fixes National Laboratory makes clear the is also significant, as is the potential data can be moved importance of usability. In particu- for expert operators to intervene and more effectively lar, failure recovery is often a human- troubleshoot on the user’s behalf. intensive process, as reported by We report here on a research data- through the Childers et al.5: “The tools that we use management system called Globus to move files typically are the standard Online, or GO, that adopts this ap- physical movement Unix tools included with ssh… it’s just proach, focusing on GO’s data-move- of media rather painful. Painful in the sense of having ment functions (“GO-Transfer”). We to manage the transfers by hand, re- describe how GO leverages modern than through starting transfers when they fail—all Web 2.0 technologies to provide in- communication of this is done by hand.” tuitive interfaces for fire-and-forget In addition, datasets may have file transfers between GridFTP end- over networks. nested structures and contain many points1 while leveraging hosting for files of different size (see the side- automatic fault recovery, high per- bar “Other Approaches”). Source and formance, simplified security con- destination may have different se- figuration, and no client software in- curity requirements and authentica- stallation. We also describe a novel tion interfaces. Networks and storage approach for providing a command- servers may suffer transient failures. line interface (CLI) to SaaS without Transfers must be tuned to exploit distributing client software and Glo- high-speed research networks. Direc- bus Connect to simplify installation tories may have to be mirrored across of a personal GridFTP server for use multiple sites, but only some files dif- with GO. Our experiments show low fer between source and destination. overhead for small transfers and high Firewalls, Network Address Transla- performance for large transfers, rela- tion, and other network complexities tive to conventional tools. may have to be addressed. For these Adoption of this new service has and other reasons, it is not unusual to been notable. One year after product hear of even modest-scale wide-area launch, November 2010, more than data transfers requiring days of care- 3,000 registered users had in aggre- ful “babysitting” or of being aban- gate moved more than two petabytes doned for high-bandwidth but high- (2×1015B) and 150 million files; nu- latency (frequently labor-intensive merous high-performance comput- and error-prone) “sneakernet.”11 ing (HPC) facilities and experimental facilities recommend GO to their us- Why Move Data at all? ers; and several “science gateways” Why not just leave data where it is are integrating GO as a data upload/ created? Such an option is certainly download solution. GO has also been preferred when possible, and we may adopted as a foundational element hope that over time moving computa- of the National Science Foundation’s tion to data rather than the other way new (as of 2011) Extreme Science and round will be more common. How- Engineering Discovery Environment ever, in practice, data scientists often (XSEDE) supercomputer network find data is “in the wrong place” and (http://www.xsede.org/). thus must be moved for a variety of reasons. Data may be produced at a Data Movement location (such as a telescope or sen- Researchers often must copy many sor array) where large-scale storage files with potentially large aggregate cannot be located easily. It may be size among two or more network- desirable to collocate data from many connected locations, or “endpoints,” sources to facilitate analysis—a com- that may or may not include the com- mon requirement in, say, genomics. puter from which the transfer com- Remote copies may be required for 82 communicaTions of The acM | February 2012 | VoL. 55 | No. 2 contributedarticles disaster recovery. Data analysis may require computing power or special- ized computing systems not avail- able locally. Policy or sociology may Other Approaches require replication of data sets in One alternative for data movement involves running tools on the user’s computer; for example, rsync,20 scp, file transfer program (FtP), secure FtP, and bbftp13 are all distinct geographical regions; for ex- used to move data between a client computer and a remote location. Other software ample, in high-energy physics, all data (such as globus-url-copy, reliable File transfer, File transfer Service, and Lightweight produced at the Large Hadron Col- Data replicator) can each manage large numbers of transfers. however, the need to lider, Geneva, Switzerland, must es- download, install, and run software is a significant barrier to use. Users spend much time configuring, operating, and updating such tools though rarely have theIT and networking sentially be replicated in the U.S. and knowledge necessary to fix things when they do not “just work,” which is all too often. elsewhere for independent analysis. Some big-science projects have developed specialized solutions to the problem; for It is also frequently the case that the example, the PheDex high-throughput data-transfer-management system9 manages data movement among sites participating in the Compact Muon Solenoid experiment aggregate data-analysis requirements at Cern, and the Laser interferometer gravitational wave Observatory (LigO) project of a community exceed the analysis developed the LigO Data replicator.4 these centrally managed systems allow users capacity of a data provider, in which to hand off data-movement tasks to a third-party service that performs them on their case data must be downloaded for lo- behalf. however, these services require professional operators functioning only among carefully controlled endpoints within these communities. cal analysis. This is the case in, for ex- Managed services (such as YouSendit and DropBox) also provide data-management ample, the Earth System Grid, which solutions but do not address researchers’ need for high-performance movement of delivers climate simulation output to large quantities of data. Bittorrent8 and Content Distribution networks21 are good at distributing a relatively stable set of large files (such as movies) but do not address data its 25,000 users worldwide. scientists’ need for many frequently updated files managed in directory hierarchies. Another common question about the integrated rule-Oriented Data System17 is often run in hosted configurations, GO is whether data can be moved but, though it performs some data-transfer operations (such as for data import), data more effectively through the physi- transfer is not its primary function or focus. the Kangaroo,19 Stork,14 and CATCh15 systems all manage data movement over cal movement of media rather than wide-area networks using intermediate storage systems where appropriate to optimize through communication over net- end-to-end reliability and/or performance. they are not designed as SaaS data- works. After all, no network can ex- movement solutions, but their methods could be incorporated into gO. ceed the bandwidth of a FedEx truck. web and reSt interfaces to centrally operated services are conventional in business, underpinning such services as Salesforce.com (customer relationship The answer is, again, that while physi- management), google Docs, Facebook, and twitter—an approach not yet common in cal shipment has its place (and may science. two exceptions are the PheDex Data Service,9 with both reSt and CLis, and be much cheaper if the alternative is the national energy research Supercomputing Center, or nerSC, web toolkit called newt7 that enables reStful operations against hPC center resources.
Recommended publications
  • BEST PRACTICE GUIDE for CLOUD and AS-A-SERVICE PROCUREMENTS Executive Summary 1 Introduction
    BEST PRACTICE GUIDE FOR CLOUD AND AS-A-SERVICE PROCUREMENTS Executive Summary 1 Introduction Specific Models and Understanding Cloud Procurement Service Models Data EXECUTIVE SUMMARY Breach Notification Personnel Security While private companies rapidly move systems and Vendors share blame, too. Lots of cloud providers are new to Encryption applications to the cloud, public agencies still struggle to adopt public sector business, having grown up selling to consumers Audits Operations hosted services that could save money and provide better value. and private firms. These companies don’t always understand Hybrid Cloud Environments legitimate demands that make government contracting Preparation for Migrating Yet states and localities have much to gain from the different from selling to other markets. Failure to accommodate Workloads to the Cloud technology industry’s “as-a-service” revolution. Many unique government requirements can be a deal-breaker for jurisdictions face huge legacy system replacement challenges. agencies charged with protecting the public’s interests. Conclusion They’re also under pressure to provide new classes of digital services. The cloud can offer a better path toward All too often, government and industry aren’t on the same page Workgroup Members modernization — there’s no hardware to buy, you’re always when it comes to cloud services. They may not even speak the and Contributors on the latest version of the software and system capacity same language. can be adjusted almost instantly based on your needs. Appendix 1 Bridging the Gap Model Terms and Conditions Templates So why is government lagging behind? The fact is that These pressures led us to release the first version of this guide Software-as-a-Service governments often struggle to buy cloud-based services because two years ago.
    [Show full text]
  • Software As a Service
    Software as a Service Haojie Hang Ogheneovo Dibie Executive Summary • In this presentation, we go through the Software as a Service Methodology, examine its benefits and drawbacks and talk about two state-of-art SaaS systems– Amazon Web Service and Google App Engine • We also look into Service Oriented Architecture powering SaaS applications and its impact on modern web 2.0 applications • Finally, we examine hybrids of traditional and SaaS applications Overview • What is Software as a Service (SaaS) • Background o Brief history o Concept o Big picture o Related terms • Computing Today o SasS is everywhere o The SaaS Market • Benefits of SaaS • Drawbacks of SaaS o Robustness o Privacy o Security o Reliability • Service Oriented Architectures (SOA) o Guiding principles of SOA • Case studies o Amazon Web Services (AWS) o Google App Engine • Influence of SOA on Web 2.0 development o Zend Framework • Hybrids of Traditional and SaaS applications o Dropbox o Microsoft Office • Summary • References What is SaaS? • Definition: Software as a Service (SaaS), a.k.a. on- demand software, is a software delivery model in which software and its associated data are hosted centrally and accessed using a thin-client, usually a web browser over the internet. – Wikipedia • Simply put, SaaS is a method for delivering software that provides remote access to software as a web- based service. The software service can be purchased with a monthly fee and pay as you go. What is SaaS? • Where does the term SaaS come from? o The SAAS acronym allegedly first appeared
    [Show full text]
  • Openstack: the Path to Cloud
    OpenStack: The Path to Cloud Considerations and recommendations for businesses adopting cloud technology openstack.org Table of Contents Executive Overview 1 Enterprise Cloud Strategy 2 Approaches to an OpenStack Private Cloud 5 Forming the OpenStack Team 9 Organization and Process Considerations 13 Choosing Workloads for Your Cloud 16 Implementation Phases 22 Post-deployment 30 Summary 32 References 33 Glossary 34 *Underlined gray bold words and concepts are defined in the Glossary at the end. CONTRIBUTORS Carol Barrett, Cloud Software Planner, Intel Corporation Tyler Britten, Technical Advocate, Blue Box, an IBM company Kathy Cacciatore, Consulting Marketing Manager, OpenStack Foundation Pete Chadwick, Senior Product Manager, SUSE Paula Phipps, Senior Manager, Infrastructure Software Marketing, Hitachi Data Systems Gerd Prüßmann, Director Cloud Solutions, Mirantis Megan Rossetti, Cloud Infrastructure Operations, Walmart Yih Leong Sun, PhD, Senior Software Cloud Architect, Intel Corporation Shamail Tahir, Offering Manager, IBM Heidi Joy Tretheway, Senior Marketing Manager, OpenStack Foundation Susan Wu, Director of Technical Marketing, Midokura Executive Overview This book is written to help enterprise architects implement an OpenStack® cloud. With architects with one foot in information technology and the other in business operations in mind, we want to offer insights and best practices to help you achieve multiple (and sometimes competing) goals. If you’re looking for vendor-neutral answers about planning your path to an OpenStack cloud, you’re in the right place. Members of the OpenStack community—technologists, business leaders and product managers—collaborated on this book to explain how to get started with an OpenStack cloud. We’ve included pros and cons to help you make better choices when setting up your cloud, along with anticipated investments of both time and money.
    [Show full text]
  • Cloud Solutions – Infrastructure, Platform Or Software: Where Should You Go? Arlene F Minkiewicz PRICE Systems, LLC [email protected]
    Cloud Solutions – Infrastructure, Platform or Software: Where should you go? Arlene F Minkiewicz PRICE Systems, LLC [email protected] © 2016 PRICE Systems, LLC All Rights RReeserrvveedd | DecaDecaddeess ooff CoCostst MMaannaaggemeemenntt ExcellExcellenencece Agenda . Introduction . Cloud Computing . Picking the right ‘as a Service’ . Case Study . Discussion and Final Thoughts © 2016 PRICE Systems, LLC All Rights Reserved | Decades of Cost Management Excellence 2 Introduction . Cloud Computing as defined by National Institute of Standards and Technology (NIST “Cloud computing is a model for enabling ubiquitous, convenient, on- demand network access to a shared pool of configurable computing resources (e.g. networks, servers, storage, applications and services) that can be rapidly provisioned and release with minimal management effort or service provider interaction” . PRNewswire reports that 90% of medium to large enterprises plan to increase or maintain annual spend on cloud for 2016 . According to CIO Magazine, the battle of the infrastructure is over – organizations have embrace outsourcing their hardware . The new battle will be in the application space © 2016 PRICE Systems, LLC All Rights Reserved | Decades of Cost Management Excellence 3 Application Migration comes with Management and Planning Challenges . How does an organization determine the right solutions to migrate to (or host in) the cloud? . How do they identify the right platform for migration? . What challenges do the various cloud solutions present? –
    [Show full text]
  • Understanding the Cloud Computing Landscape
    Chapter 1 Understanding the Cloud Computing Landscape Lamia Youseff, Dilma M. Da Silva, Maria Butrico, and Jonathan Appavoo Contents 1.1 Introduction .................................................................................................2 1.2 Cloud Systems Classifications ......................................................................2 1.3 SPI Cloud Classification ...............................................................................2 1.3.1 Cloud Software Systems ...................................................................3 1.3.2 Cloud Platform Systems ....................................................................3 1.3.3 Cloud Infrastructure Systems ...........................................................4 1.4 UCSB-IBM Cloud Ontology .......................................................................4 1.4.1 Applications (SaaS) ...........................................................................5 1.4.2 Cloud Software Environment (PaaS) ................................................7 1.4.3 Cloud Software Infrastructure ..........................................................8 1.4.4 Software Kernel Layer .......................................................................9 1.4.5 Cloud Hardware/Firmware ...............................................................9 1.5 Jackson’s Expansion on the UCSB-IBM Ontology .....................................10 1.6 Hoff’s Cloud Model ...................................................................................11 1.7 Discussion ..................................................................................................13
    [Show full text]
  • The Reality of Cloud Computing Chris Rose, Walden University, USA
    International Journal of Management & Information Systems – Fourth Quarter 2011 Volume 15, Number 4 A Break In The Cloud? The Reality Of Cloud Computing Chris Rose, Walden University, USA ABSTRACT Cloud computing is on the forefront of the technological landscape with Google, Microsoft and Amazon, among others, building up their server capacity to handle the next perceived leap in technological innovation. However, not only does the security of cloud computing have to be seriously addressed, but SOA governance and the capacity in the core, connectivity and fiber layers of the Internet to absorb the increased bandwidth will also have to be considered. INTRODUCTION he term ―cloud computing‖ has many different meanings. A discussion at the 2008 IEEE International Conference on Web Services (ICWS), in Beijing, concluded that the definition depends T on whom you ask. For application and IT users, it’s IT as a service (ITaaS)—that is, delivery of computing, storage, and applications over the Internet from centralized data centers. For Internet application developers, it’s an Internet-scale software development platform and runtime environment. For infrastructure providers and administrators, it’s the massive, distributed data center infrastructure connected by IP networks (Lin et al, 2009). Everyone sees the cloud and uses the cloud differently. Some people offload ordinary desktop software and instead use web-based applications, others want on-demand computing with instant availability of extra processing power and extra storage. Nevertheless, the inescapable fact is that companies of all sizes are increasingly using the cloud to increase their businesses, perhaps by enabling remote working or simply to cut costs or even to create entirely new business models (Marshall, 2009).
    [Show full text]
  • Cisco Fog Computing Solutions: Unleash the Power of the Internet of Things
    White Paper Cisco Fog Computing Solutions: Unleash the Power of the Internet of Things Connect things. Analyze and act on the data they produce in milliseconds. Then send the right data to the cloud for big-data analytics and storage. Benefits The Internet of Things (IoT) speeds up awareness and Only Cisco Fog computing solutions provide the response to events. It’s transforming whole industries, including following benefits: manufacturing, oil and gas, utilities, transportation, public safety, ● Data privacy: When necessary, analyze sensitive data within the building instead of and local government. sending it to a remote data center. Cisco Fog Data Services applies your policy to determine But the IoT requires a new kind of infrastructure. The cloud by the best place for analysis. ● Comprehensive security: Protect fog nodes itself can’t connect and analyze data from thousands and millions using the same Cisco physical security and of different kinds of things spread out over large areas. Capturing cybersecurity solutions you use to protect other IT assets. These solutions provide the the power of the IoT requires a solution that can: capabilities you need before, during, and after an attack. ● Connect new kinds of things to your network. Some of them ● Business agility: Gain business insights more quickly by hosting applications closer to might be in harsh environments. Others might communicate your IoT devices and analyzing data closer to the source. using industrial protocols, not IP. ● Rapid innovation: Create and deploy new applications more quickly by using our ● Secure the things that produce data. And secure the data as it infrastructure-as-a-service (IaaS) and travels from the network edge to the cloud.
    [Show full text]
  • Evolution of As-A-Service Era in Cloud
    Evolution of as-a-Service Era in Cloud Sugam Sharma Center for Survey Statistics and Methodology, Iowa State University, Ames, Iowa, USA Email: [email protected] Abstract. Today, a paradigm shift is being observed in science, where the focus is gradually shifting toward the cloud environments to obtain appropriate, robust and affordable services to deal with Big Data challenges (Sharma et al. 2014, 2015a, 2015b). Cloud computing avoids any need to locally maintain the overly scaled computing infrastructure that include not only dedicated space, but the expensive hardware and software also. In this paper, we study the evolution of as-a-Service modalities, stimulated by cloud computing, and explore the most complete inventory of new members beyond traditional cloud computing stack. Keywords. Cloud, as-a-Service 1. Introduction Today, the cloud computing also has emerged as one of the major shifts in recent information and communication age that promises an affordable and robust computational architecture for large-scale and even for overly complex enterprise applications (Fenn et al. 2008). It is a powerful and revolutionary paradigm that offers service-oriented computing and abstracts the software-equipped hardware infrastructure from the clients or users. Although, the concept of cloud computing is mainly popular in three praxises- 1) IaaS (Infrastructure-as-a-Service ), 2) PaaS (Platform-as-a-Service ), and 3) SaaS (Software-as-a-Service ), but in this data science age, should be equally expandable to DBaaS (Database- as-a-Service ) (Curino, et al., 2011; Seibold et al., 2012). Over the last few year, the cloud computing has evolved as scalable, secure and cost-effective solutions to an overwhelming number of applications from diversified areas.
    [Show full text]
  • How to Become a Google Workspace Partner
    How to become a Google Workspace partner Partnering with Tech Data and Google Workspace is an exciting opportunity to differentiate your business and open up new revenue streams. Grow your business by unlocking Opportunities to win new customers in exciting markets. This quick guide outlines the requirements to becoming a Google Workspace partner, so you can get selling fast. 1 Get ready... 3 Get registered… • Do you have an established or growing Cloud • Do you have a Tech Data trading account? business with the capability to design and If not, click here to get set up manage customer solutions or are you looking • Log in to StreamOne: Click on Tech Data’s to develop a Cloud Story? StreamOne Cloud Marketplace • Do you have experience of implementing • Register as a Google partner in Partner Collaboration and Productivity projects or are Advantage looking to develop this? • Accept the 2-way agreement with Google • Do you offer professional, managed and • Create your Cloud ID and verify your domain support services or are you looking to start offering these? • Do you have customers looking to change the way they communicate and collaborate? 4 Get inspired...and Go… • Are you able to identify key target verticals / industries or would you like to understand more? • Access Partner Advantage for all partner info • Are your sales force cloud-ready with a • Align to the correct Engagement level knowledge of SaaS (Software-as-a-Service)? • Take the training courses, get certified to become Google Cloud-ready and start selling and growing • Suggested
    [Show full text]
  • What Every CEO Needs to Know About the Cloud by Andrew Mcafee
    What Every CEO Needs to Know About The Cloud by Andrew McAfee The true benefits of cloud computing will surprise you. But you’ll need the right people to lead the transformation. This article originally appeared in Harvard Business Review (reprint #R1111J). sponsored by Copyright © 2012 Harvard Business School Publishing Corporation. All rights reserved. WHAT EVERY CEO NEEDS TO KNOW ABOUT THE CLOUD In 2010 an IBM survey factory owners, so too will the cloud confer advan- of more than 1,500 CEOs worldwide revealed tages on its adopters. a troubling gap: Close to 80% of them believed their At present, there’s a lot of uncertainty and skepti- environment would grow much more complex in cism around the cloud, particularly among technol- the coming years, but fewer than half thought their ogy professionals who have deep expertise with, or companies were well equipped to deal with this shift. attachment to, on-premise computing. Companies The survey team called it “the largest leadership shouldn’t give such people too much influence over challenge identified in eight years of research.” plans to move into the cloud; that would be like put- Unfortunately, the information technology in- ting the crew that ran the boiler and steam turbine in frastructure at many large companies only makes charge of electrifying a factory. The CEO and other this challenge more difficult. Their technology en- senior business executives need to take responsi- vironments actually impede their ability to sense bility for bringing their organizations into the era of change and respond quickly. While there is no sim- cloud computing.
    [Show full text]
  • Delivering IT As a Service with a Software-Defined Data Center
    Your ClouD: ITaaS white paper | itaas Delivering IT as a Service With a Software-Defined Data Center This white paper is Executive Summary based on research conducted by IDG The pace of IT change is increasing remarkably. Mobility, cloud computing, social research Services media, consumerization, big data and analytics have all become deeply interwoven on behalf of VMware. within today’s enterprise. The result is a paradigm shift in the way organizations use The June 2012 “Cloud and consume IT resources. Today, new and dynamic service delivery models promise Innovation Study: IT as a Service” was to revolutionize business and transform IT. conducted across North America, One of the most disruptive elements of the new order more holistic and transparent way. EMEA, APAC and is IT as a Service (ITaaS). ITaaS starts at the top with latin America. All a crystal-clear understanding of business needs; For many organizations, this approach creates a 650 respondents it also starts at the bottom with a set of virtualized dramatically different outside-in view of IT and busi- were from non-uS resources and pre-configured and certified building ness resources. When ITaaS is used effectively, the companies with blocks that can be combined and deployed at the resulting solution drives transformation through the 500+ employees or click of a button. enterprise and helps IT evolve from a tactical role to uS companies with strategic one that potentially leads to a competitive 1,000+ employees. With a top-down view and bottom-up technology advantage. capability, IT can quickly and reliably respond to the changing needs of the business with optimized yet This paper explores the concept of ITaaS and the highly standardized solutions.
    [Show full text]
  • Open Source Solution for Cloud Computing Platform Using Openstack
    See discussions, stats, and author profiles for this publication at: http://www.researchgate.net/publication/263581733 Open Source Solution for Cloud Computing Platform Using OpenStack CONFERENCE PAPER · MAY 2014 DOI: 10.13140/2.1.1695.9043 CITATIONS DOWNLOADS VIEWS 8 826 204 5 AUTHORS, INCLUDING: Rakesh Kumar Neha Gupta JECRC Foundation University of Adelaide 20 PUBLICATIONS 69 CITATIONS 8 PUBLICATIONS 36 CITATIONS SEE PROFILE SEE PROFILE Shilpi Charu Kanishk Jain Rajasthan Technical University JECRC Foundation 10 PUBLICATIONS 33 CITATIONS 2 PUBLICATIONS 14 CITATIONS SEE PROFILE SEE PROFILE Available from: Rakesh Kumar Retrieved on: 10 August 2015 Rakesh Kumar et al, International Journal of Computer Science and Mobile Computing, Vol.3 Issue.5, May- 2014, pg. 89-98 Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology ISSN 2320–088X IJCSMC, Vol. 3, Issue. 5, May 2014, pg.89 – 98 RESEARCH ARTICLE Open Source Solution for Cloud Computing Platform Using OpenStack Rakesh Kumar1, Neha Gupta2, Shilpi Charu3, Kanishk Jain4, Sunil Kumar Jangir5 1,2,3,4,5Department of Information Technology, JECRC, Jaipur, India 1 [email protected], 2 [email protected], 3 [email protected] 4 [email protected], 5 [email protected] Abstract— OpenStack is a massively scalable open source cloud operating system that is a global collaboration of developers and cloud computing technologists producing the ubiquitous open source cloud computing platform for public and private clouds. OpenStack provides series of interrelated projects delivering various components for a cloud infrastructure solution as well as controls large pools of storage, compute and networking resources throughout a datacenter that all managed through a dashboard(Horizon) that gives administrators control while empowering their users to provision resources through a web interface.
    [Show full text]