Download Transcript

Total Page:16

File Type:pdf, Size:1020Kb

Download Transcript KubeAcademy Transcript Course: Building Images Lesson: Building Images with Buildpacks: The Cloud Native Buildpacks Project Instructor: Cora Iberkleid URL: https://kube.academy/courses/building-images/lessons/building-images-with-buildpacks-the- cloud-native-buildpacks-project In this lesson, we'll introduce the Cloud Native Buildpacks project. We'll talk about the evolution of buildpacks and the motivation behind the project. We'll explain at a high level how it works and some of the choices in the ecosystem for using it. Buildpacks were first introduced by Heroku in 2011. They're simply components that prepare an application for launch in the cloud. Cloud Foundry embraced the concept of buildpacks early on and over the following years, both Heroku and Cloud Foundry focused on providing buildpacks to support popular programming languages. The community also saw broad adoption, as an example, there are nearly 6,500 build packs listed on the Heroku marketplace alone. The developer experience with buildpacks is to simply push source code or an application artifact, and the cloud platform takes care of the rest. The platform detects which build pack to use and provides a staging environment. The build pack produces a staged application in the form of a tarball containing the compiled source and any dependencies. The platform launches the staged application, and you've got your application running in the cloud. This model is very easy to use and it provides enterprises the consistency, governance, security, and manageability that they need to operate at scale. But this design also has some limitations because it's tightly coupled and it lacks flexibility. By and large, if you're not using Heroku or Cloud Foundry, you can't use buildpacks or the staged application that they create. Buildpacks are also difficult to modularize because they play both the role of orchestrator of the overall build process and worker providing the language specific dependencies. Now, this design predates Docker, the open container initiative and Kubernetes. It made adjustments over the years, but in order to really bring it strengths to the modern container ecosystem and enable more advanced features, it needed to be redesigned. So Heroku and Pivotal now VMware, and the original incubator of Cloud Foundry jointly established the Cloud Native Buildpacks project in 2018 for this purpose. To alleviate the problems of coupling and flexibility, the project provides a well-defined contract between platforms and buildpacks. It includes a platform API so that new platforms can be created each with a freedom to offer a unique user experience. It also separates the orchestration of the build process from the buildpacks themselves and provides a buildpack API that enables the creation and composability of modularized buildpacks. In addition, the end result of a build is an OCI image, which means it can be stored in any image registry and run on any container runtime system. This makes it possible to implement new capabilities and optimizations. We'll talk more about this shortly. So a platform is a tool that you use as an end user or as part of your build pipeline in order to take advantage of buildpacks. Buildpacks are components that provide runtime support for applications. They enable us to add content or behavior to an image in a way that we can easily operationalize and scale. The interaction between the two, which would be teased out by implementing both APIs, constitutes the generic operations that any platform needs to carry out in order to leverage an arbitrary set of buildpacks. Because it is common behavior that all platforms can share, the Cloud Native Buildpacks project makes it easy by providing a reference implementation. This implementation is called the lifecycle and it comprises a set of well-defined phases. Platforms only need to orchestrate the life cycle rather than implement their own. Let's walk through the phases for the first build of an application. The detect phase determines which buildpacks to run and in what order. For example, for a Java application, it may identify a JVM buildpack, a [inaudible 00:04:12] buildpack, and an Apache Tomcat buildpack. The build phase runs the buildpacks, each buildpack prepares layers of assets to add to the final image. Each layer includes metadata about the contents, such as the JVM version, for example. Now, the lifecycle is executing on a build image so much like in the multi-stage Docker file example that we saw earlier in this course, the export phase copies the layers from the build image to a slimmer run image. If the platform provides storage, then the export can also cache layers to optimize subsequent builds. One interesting feature is that for each layer, a buildpack can specify whether or not to export and cache it. For example, a JVM buildpack may create JDK and JRE layers cache both of them, but only export the JRE to the run image. On a rebuild, the detect phase once again determines the set of ordered buildpacks to run. The analyze phase retrieves metadata about image layers and cache layers, so that buildpacks can decide which layers need to be recreated. The fact that each build pack can make decisions independently about reusing layers means that cache is not as sensitive to order as it is with Docker file and Gibb. If cache is available, the restore phase makes it accessible by loading it into the build image file system. The build phase, once again, executes all of the buildpacks, but the rebuild is optimized because it has the local cash, as well as the necessary information to minimize the number of layers to recreate. The export pushes only new layers to the registry, along with a new manifest that can point to unchanged layers from the previous image as well as the new layers. It also updates the cache as necessary. The lifecycle also provides the rebase phase that can be invoked independently to swap out the runtime base image. This is intended for cases where an operating system vulnerability is discovered and a patch is made available in the form of a new run image. The rebase is invoked and a new application image is created that points to the digests of the run image layers that changed. There's no need to recreate or copy the other layers. This operation can complete within milliseconds on a registry and the new base image is guaranteed to be compatible by compliance with what's called an application binary interface. So using the right platform, this rebase can be easily applied to hundreds or thousands of images in a matter of seconds, which makes it a powerful security tool. Let's recap by way of an example, using a CLI called pack as our platform to build an image. We use the command pack build and provide a name for our image. We need to provide the path to our source code, a build image, which is comparable to the front statement in the first stage of the multi-stage build that we saw in our Docker file examples. The build image should have the lifecycle installed, then a run image, which is comparable to the front statement in the second stage from our Docker file examples, and then the list of build packs to apply. You can see that the build and run images as well as the build packs are all provided by the build packs implementation, in this case, an implementation called paketo. Since we know the lifecycle can auto detect which buildpacks to use, it makes sense to bundle the buildpacks into the build image as well, which is exactly how buildpacks are distributed. The build image serves as a builder with all of the components and intelligence needed for the build including a reference to the proper run image to use. Now that's still a lot to type, so you can set your builder of choice as the default and CD into your source code directory, which brings us to the very simple command pack, build, and the image name. And this will work for applications written in Java, Node, Ruby, .NET, and more. Very simple and extremely powerful. So we understand the building blocks that we need to build images with Cloud Native Buildpacks. In addition to the pack, CLI there's a growing ecosystem of platforms. Pack is in fact, the reference implementation from the Cloud Native Buildpacks project, and it optimizes for the developer experience based on imperative commands. Another example is Kpack, which is a service that can be hosted in a Kubernetes cluster and configured declaratively and Spring Boot offers Cloud Native Buildpacks support and provides integrations with Maven and Gretel workflows, and so on. You can also mix and match as long as you use the same base images and build pack versions. Two different platforms will create the same image from the same source. Now, the cloud native buildpacks project doesn't provide reference implementations for builders or buildpacks because it only seeks to be opinionated about how to orchestrate a build not about how to run an application. However, Heroku and Cloud Foundry both provide versions of their buildpacks including base and run images that are compatible with Cloud Native Buildpacks. Paketo is in fact, the evolution of cloud Foundry build packs. VMware provides a commercial super set of paketo that includes buildpacks for integrations with security scanners, application performance monitoring tools, and so on, and this ecosystem is growing as well. In the next lesson, we'll go hands-on with cloud native buildpacks using pack Kpack, Spring Boot, and Paketo buildpacks.
Recommended publications
  • WEIYANG (STEPHEN) YUAN [email protected] | Chicago | 608-504-0649 | Stephenyuan.Urspace.Io Education University of Wisconsin-Madison B.S
    WEIYANG (STEPHEN) YUAN [email protected] | Chicago | 608-504-0649 | stephenyuan.urspace.io Education University of Wisconsin-Madison B.S. in Computer Engineering May 2020 ● GPA: 3.83/4.0 ​ ● Related Coursework: Operating Systems • A​ rtificial Intelligence • Computer Networks and Communication • Databases • Information Security • Big Data Systems • Android Mobile Development Skills ● Programming Languages: Java • Golang • C++ • Scala • MATLAB • SQL • Julia • C • Python ​ ● Technologies: Git, Linux, Java Spring, Amazon Web Services (AWS), MongoDB, Postgres, React, Node.js, Docker, Jenkins, ​ Play Framework, Hadoop, Spark, Wireshark, Visual Studio Experience Enfusion, Chicago Java Software Developer July 2020 - Current ● Develop the portfolio management software system used by over 500 clients that supports a variety of financial calculation and valuation over 20 financial derivatives as well as back office general ledger and cash flow with more than 10,000 daily positions on average ● Take responsibility in the whole development lifecycle from designing (10%), implementing (40%), running regression & unit testing (40%) to supporting internal and production issues (10%) ● Apply experience of Object-Oriented design patterns and best practices to creating a robust and reliable infrastructure for the system with knowledge of Java SE, Hibernate, JMS, JVM and MySQL and deliver constant results in weekly production ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ release ● Automate development and testing frameworks by writing python and shell scripts to improve overall
    [Show full text]
  • Q& a – Wavemaker Demo/Training Webinar – March 15, 2016
    Q& A – WaveMaker Demo/Training Webinar – March 15, 2016 Q: Is there a workflow engine ie BPMN v2 A: WaveMaker has strong integrations with APIs, web services and SDKs. You can integrate with rules engine platforms like JBPM and Drools using their Java SDKs (jars) or through their ReST APIs. You need to do this only if the workflow requires very complex workflows. Otherwise, the native capabilities of the WaveMaker should be enough to take care of your app requirements. Q: on oracle db's there are schemas, which should be considered...what is important to consider regarding to that (schema)? do I have to import the tables over each schema or is there a way how to import the whole tables (over all schemas)...? A: WaveMaker allows you to import tables from multiple schemas. See the image below. This is the second step of importing an Oracle DB. [Steps: Import > Database > Select Oracle > ….] Q: Can we provide read only access to a user group? A: For instance if you are dealing with an editable grid, which has an add/save button, all you to do is to hide the button for users with a specific role. If you are dealing with a specific widget then you can use the conditional read-only option as shown below. Here you need to write a one-line javascript as shown below, where the users with the role “rolename” will be presented a read-only birthdate. Q: Can we integrate the application for SSO ? A: You can configure SSO easily through the following approach.
    [Show full text]
  • Salesforce Heroku Enterprise: a Cloud Security Overview June 2016
    Salesforce Heroku Enterprise: A Cloud Security Overview June 2016 1 Contents INTRODUCTION 3 CASESTUDY:GLIBC 19 Heroku behind the Curtain: Patching SALESFORCETRUSTMODEL 4 the glibc Security Hole What Do We Do When a Security CLOUD COMPUTING AND Vulnerability Lands? THESHAREDSECURITYMODEL 5 Provider Responsibilities How Do We Do This with Minimum Downtime? Tenant Responsibilities What about Data? INFRASTRUCTURE AND What about Heroku Itself? APPLICATIONSECURITY 8 Keep Calm, Carry On Server Hardening Customer Applications BUSINESSCONTINUITY 23 Heroku Platform: High Availability Container Hardening and Disaster Recovery Application Security Customer Applications Heroku Flow Postgres Databases Identity and Access Management Customer Configuration and Identity Federation via Single Sign-On Meta-Information Organizations, Roles, and Permissions Service Resiliency and Availability Business Continuity and Emergency NETWORKSECURITY 14 Preparedness Secure Network Architecture Secure Access Points INCIDENTRESPONSE 27 Data in Motion ELEMENTSMARKETPLACE 28 Private Spaces App Permissions Building Secure Applications with Add-Ons DATASECURITY 16 Heroku Postgres PHYSICALSECURITY 29 Encryption Data Center Access Customer Data Retention and Destruction Environmental Controls Management SECURITYMONITORING 17 Storage Device Decommissioning Logging and Network Monitoring DDoS COMPLIANCEANDAUDIT 31 Man in the Middle and IP Spoofing SUMMARY 33 Patch Management HEROKU ENTERPRISE SECURITY WHITE PAPER 2 Introduction eroku Enterprise, a key component of the Salesforce Platform, is a cloud application platform used by organizations of all sizes to deploy and operate applications throughout Hthe world. The Heroku platform is one of the first cloud application platforms delivered entirely as a service, allowing organizations to focus on application development and business strategy while Salesforce and the Heroku division of Salesforce focus on infrastructure management, scaling, and security.
    [Show full text]
  • Cloud Computing: a Taxonomy of Platform and Infrastructure-Level Offerings David Hilley College of Computing Georgia Institute of Technology
    Cloud Computing: A Taxonomy of Platform and Infrastructure-level Offerings David Hilley College of Computing Georgia Institute of Technology April 2009 Cloud Computing: A Taxonomy of Platform and Infrastructure-level Offerings David Hilley 1 Introduction Cloud computing is a buzzword and umbrella term applied to several nascent trends in the turbulent landscape of information technology. Computing in the “cloud” alludes to ubiquitous and inexhaustible on-demand IT resources accessible through the Internet. Practically every new Internet-based service from Gmail [1] to Amazon Web Services [2] to Microsoft Online Services [3] to even Facebook [4] have been labeled “cloud” offerings, either officially or externally. Although cloud computing has garnered significant interest, factors such as unclear terminology, non-existent product “paper launches”, and opportunistic marketing have led to a significant lack of clarity surrounding discussions of cloud computing technology and products. The need for clarity is well-recognized within the industry [5] and by industry observers [6]. Perhaps more importantly, due to the relative infancy of the industry, currently-available product offerings are not standardized. Neither providers nor potential consumers really know what a “good” cloud computing product offering should look like and what classes of products are appropriate. Consequently, products are not easily comparable. The scope of various product offerings differ and overlap in complicated ways – for example, Ama- zon’s EC2 service [7] and Google’s App Engine [8] partially overlap in scope and applicability. EC2 is more flexible but also lower-level, while App Engine subsumes some functionality in Amazon Web Services suite of offerings [2] external to EC2.
    [Show full text]
  • Cloud Computing Parallel Session Cloud Computing
    Cloud Computing Parallel Session Jean-Pierre Laisné Open Source Strategy Bull OW2 Open Source Cloudware Initiative Cloud computing -Which context? -Which road map? -Is it so cloudy? -Openness vs. freedom? -Opportunity for Europe? Cloud in formation Source: http://fr.wikipedia.org/wiki/Fichier:Clouds_edited.jpg ©Bull, 2 ITEA2 - Artemis: Cloud Computing 2010 1 Context 1: Software commoditization Common Specifications Not process specific •Marginal product •Economies of scope differentiation Offshore •Input in many different •Recognized quality end-products or usage standards •Added value is created •Substituable goods downstream Open source •Minimize addition to end-user cost Mature products Volume trading •Marginal innovation Cloud •Economies of scale •Well known production computing •Industry-wide price process levelling •Multiple alternative •Additional margins providers through additional volume Commoditized IT & Internet-based IT usage ©Bull, 3 ITEA2 - Artemis: Cloud Computing 2010 Context 2: The Internet is evolving ©Bull, 4 ITEA2 - Artemis: Cloud Computing 2010 2 New trends, new usages, new business -Apps vs. web pages - Specialized apps vs. HTML5 - Segmentation vs. Uniformity -User “friendly” - Pay for convenience -New devices - Phones, TV, appliances, etc. - Global economic benefits of the Internet - 2010: $1.5 Trillion - 2020: $3.8 Trillion Information Technology and Innovation Foundation (ITIF) Long live the Internet ©Bull, 5 ITEA2 - Artemis: Cloud Computing 2010 Context 3: Cloud on peak of inflated expectations According to
    [Show full text]
  • The Line Download Torrent Spec Ops: the Line Torrent PC Game Free Download
    the line download torrent Spec Ops: The Line Torrent PC Game Free Download. Spec Ops: The Line Download For PC. Spec Ops: The Line Download For PC is an action shooter game, That you will play from the point of view of a third person. It is the 10th game in the series of Spec Ops games. The game is set in a virtual open world. And is based on war battels. This game has both single-player and multiplayer modes. Furthermore, it also has an online multiplayer mode. The main character of the game is Captain Martin. In the game, the captain was given a mission after the end of the war. The mission is called the Recon mission. Captain Martin creates his force and goes for the mission. Gameplay Of Spec Ops: The Line Highly Compressed. Gameplay Of Spec Ops: The Line Highly Compressed is like war battels gameplay. In which players take control of the main character Captian Martin. The gameplay is set on different levels. The main four difficult missions of the game are Fubar, Combat Op, Walk on the beach, and a Suicide mission. At the start of the game, the player can select only one mission to play. As he passes on a mission then he can unlock other missions. This game also includes various types of powerful weapons. That also helps the player to fight against enemies. The weapons include pistols, Machine guns firearms, and many more. But the player can take two weapons at a time. And can quickly change them during the fight.
    [Show full text]
  • Annual Report 2018
    Annual Report 2018 January 2019 In 2018, the Foundation expanded its definition of Cloud Foundry, shifting away from prioritizing the Application Runtime in order to spread awareness of the many projects that comprise Cloud Foundry technologies. This new messaging aligned with the 2018 vision for interoperability, a theme that underpinned all Foundation content. The interoperability of Cloud Foundry was evident across the ecosystem -- in the technologies integrated into the platform itself to the technologies with which Cloud Foundry integrates to form a multi-platform strategy. This year, the Foundation launched a Certified Systems Integrators program, announced new Certified Providers Cloud.gov and SUSE, accepted interoperable projects Eirini and CF Containerization into the Project Management Council and published four research reports -- while planning and hosting two major Summits and six Cloud Foundry Days. Like the platform itself, the Cloud Foundry Foundation has reached a new level of maturity and continues to evolve. Cloud Foundry Foundation Annual Report 2018 2 2018 Highlights TECHNICAL & COMMUNITY • Interoperability: Two new projects were accepted by the Project Management Committees in order to further integrate Kubernetes with Cloud Foundry technologies. CF Containerization, initially developed and donated to the Foundation by SUSE, is designed to package Cloud Foundry BOSH releases into containers and deploy those containers into Kubernetes. Eirini, proposed by IBM and seeing contributions from IBM, SUSE and SAP, is working towards allowing operators and product vendors to use Kubernetes as the underlying container scheduler for the Cloud Foundry Application Runtime. • 2018 Certified Providers: In 2018, Cloud.gov and SUSE joined the list of certified providers of Cloud Foundry, bringing the total to eight -- the other six being Atos, Huawei, IBM, Pivotal, SAP and Swisscom.
    [Show full text]
  • Holistic Configuration Management at Facebook
    Holistic Configuration Management at Facebook Chunqiang Tang, Thawan Kooburat, Pradeep Venkatachalam, Akshay Chander, Zhe Wen, Aravind Narayanan, Patrick Dowell, and Robert Karl Facebook Inc. ftang, thawan, pradvenkat, akshay, wenzhe, aravindn, pdowell, [email protected] Abstract the many challenges. This paper presents Facebook’s holistic Facebook’s web site and mobile apps are very dynamic. configuration management solution. Facebook uses Chef [7] Every day, they undergo thousands of online configuration to manage OS settings and software deployment [11], which changes, and execute trillions of configuration checks to is not the focus of this paper. Instead, we focus on the home- personalize the product features experienced by hundreds grown tools for managing applications’ dynamic runtime of million of daily active users. For example, configuration configurations that may be updated live multiple times a changes help manage the rollouts of new product features, day, without application redeployment or restart. Examples perform A/B testing experiments on mobile devices to iden- include gating product rollouts, managing application-level tify the best echo-canceling parameters for VoIP, rebalance traffic, and running A/B testing experiments. the load across global regions, and deploy the latest machine Below, we outline the key challenges in configuration learning models to improve News Feed ranking. This paper management for an Internet service and our solutions. gives a comprehensive description of the use cases, design, Configuration sprawl. Facebook internally has a large implementation, and usage statistics of a suite of tools that number of systems, including frontend products, backend manage Facebook’s configuration end-to-end, including the services, mobile apps, data stores, etc.
    [Show full text]
  • Platform As a Service (Paas)
    Cloud Computing – Lecture 3 Platform as a Service (PaaS) Pelle Jakovits [email protected] Outline • Introduction to Platform as a Service (PaaS) Cloud model • Different types of PaaS • Google App Engine • Heroku PaaS • Advantages & disadvantages of PaaS 2 Cloud Models http://nolegendhere.blogspot.com.ee/2012/06/presentation-4-5-7.html 3 Background • Previous lecture introduced the different cloud computing models (IaaS, PaaS, SaaS) • IaaS provides computing resources – Virtual machines, storage, network. • User do not need to purchase hardware themselves • IaaS can utilize resources more efficiently in comparison to each customer using own hardware • You have worked with OpenStack instances in the lab 4 Issues with using IaaS • To deploy applications in IaaS, need to choose and set up: – Computing infrastructure – Software environment • User is responsible for: – System administration, backups – Monitoring, log analysis – Managing software updates – Stability & scalability of the software environment 5 Cloud Model complexity 6 Platform as a Service - PaaS • Complete platform for hosting applications in Cloud • The underlying infrastructure & software environment is managed for you • Enables businesses to build and run web-based, custom applications in an on-demand fashion • Eliminates the complexity of selecting, purchasing, configuring, and managing hardware and software • Dramatically decreases upfront costs 7 PaaS Characteristics • Multi-tenant architecture • Built-in scalability of deployed software • Integrated with cloud services and databases • Simplifies prototyping and deploying startup solutions • More fine-grained cost model – Generally do not pay for unused resources – Users only pay for services they use • Typically introduces vendor lock-in 8 Different types of PaaS 1. Fully managed cloud platforms for web applications – Google App Engine, AWS BeanStalk, Windows Azure 2.
    [Show full text]
  • Developing a Static Website and Deploying It to Heroku
    Sankalpa Neupane DEVELOPING A STATIC WEBSITE AND DEPLOYING IT TO HEROKU Thesis CENTRIA UNIVERSITY OF APPLIED SCIENCES Information Technology November 2020 ABSTRACT Centria University Date Author of Applied Sciences November 2020 Sankalpa Neupane Degree programme Information Technology Name of thesis DEVELOPING A STATIC WEBSITE AND DEPLOYING IT TO HEROKU Instructor Pages 33 + 2 Supervisor Jari Isohanni A website is a most useful tool for all kind of business around the globe and the easiest way to reach out to the people. So, a good user satisfying website is a must in today’s business. A website should be deployed to make it available for all the users around the world. So, a website deployment can be done using a cloud platform which allows developers to maintain, scale and test the website. This thesis is implemented in such a way that it demonstrates how a website is developed and can be deployed in the cloud platform. The thesis aims to present the process of building and deploying a static website on a cloud platform. The website was made up of a single page. The website was focused on front-end development. The process included designing and developing by using HTML5, CSS 3, JavaScript, and Bootstrap 4.5. The process also included the deployment of a static website to Heroku. Heroku is a cloud platform which allows users to build, test, maintain and scale the web applications. It also allows the user to deploy the web application from a local environment to the server. It supports many programming languages like PHP, Java, Node.js, Python, and Go.
    [Show full text]
  • ENHANCED DATA SECURITY on CLOUD BASED USING ENCRYPTION ALGORITHM, ELLIPTIC CURVE CRYPTOGRAPHY & BLOWFISH ALGORITHM Neetu Sharma1, Monika Kansal2 1Asstt
    ENHANCED DATA SECURITY ON CLOUD BASED USING ENCRYPTION ALGORITHM, ELLIPTIC CURVE CRYPTOGRAPHY & BLOWFISH ALGORITHM 1 2 Neetu Sharma , Monika Kansal 1Asstt. Prof. , ABES Engg. College GZB (India) 2 Asstt. Prof. , ITS College GZB (India) ABSTRACT With the advent internet in the 1990s to the present day facilities of ubiquitous computing, the internet has changed the computing world in a drastic way. It has traveled from the concept of parallel computing to distribute computing to grid computing and recently to cloud computing. Although the idea of cloud computing has been around for quite some time, it is an emerging field of computer science. Some of the major firms like Amazon, Microsoft and Google have implemented the “CLOUD” and have been using it to speed up their business. In this paper we will discuss Distributed scheme and Different algorithm to provide security of the data in cloud to prevent Data access from unauthorized access. Keywords: Cloud Computing, Security Algorithm, For Security I. INTRODUCTION Cloud computing can be defined as a computing environment where computing needs by one party can be outsourced to another party and when need be arise to use the computing power or resources like database or emails, they can access them via internet. Cloud computing [1, 2] is a recent trend in IT that moves computing and data away from desktop and portable PCs into large data centers. The main advantage of cloud computing is that customers do not have to pay for infrastructure, its installation, required man power to handle such infrastructure and maintenance. The rapid deployment [1] of cloud computing promises network users with elastic, abundant, and on-demand cloud services.
    [Show full text]
  • Lecture 5: Cloud Computing
    UNIVERSITY OF JYVÄSKYLÄ Lecture 5: Cloud Computing TIES4560 SOA and Cloud Computing Autumn 2021 University of Jyväskylä Khriyenko Oleksiy UNIVERSITY OF JYVÄSKYLÄ Cloud Computing 05/10/2021 TIES4560 - Lecture 5 2 UNIVERSITY OF JYVÄSKYLÄ Cloud Computing https://medium.com/edureka/on-premise-vs-cloud-computing-f9aee3b05f50 05/10/2021 TIES4560 - Lecture 5 3 UNIVERSITY OF JYVÄSKYLÄ Cloud Computing Internet WEB 05/10/2021 TIES4560 - Lecture 5 4 UNIVERSITY OF JYVÄSKYLÄ Cloud Computing Cloud DATA DATA DATA DATA DATA 05/10/2021 TIES4560 - Lecture 5 5 UNIVERSITY OF JYVÄSKYLÄ Cloud Computing Cloud 05/10/2021 TIES4560 - Lecture 5 6 UNIVERSITY OF JYVÄSKYLÄ Cloud Computing Cloud 05/10/2021 TIES4560 - Lecture 5 7 UNIVERSITY OF JYVÄSKYLÄ Cloud Computing Cloud Computing is a model for enabling convenient, on-demand network access to a shared pool of configurable computing resources (e.g., networks, servers, storage, applications, and services) that can be rapidly provisioned and released with minimal management effort or service provider interaction. Characteristics for a service to be considered “Cloud” are: o On-demand self-service. The ability for an end user to sign up and receive services without the long delays that have characterized traditional IT o Broad network access. Ability to access the service via standard platforms (desktop, laptop, mobile, etc.) o Resource pooling. Resources are pooled across multiple customers using virtualization (the ability to increase computing efficiency). Services that can apply for resource pooling: data storage services, processing services and bandwidth provided services. o Rapid elasticity. Scalability and fast provisioning a cloud service or application isn't limited to what a particular server can cope with; it can automatically expand or contract its capacity as needed.
    [Show full text]