Autonomic Computing in Virtualized Environment

Total Page:16

File Type:pdf, Size:1020Kb

Autonomic Computing in Virtualized Environment International Journal of Information Technology Convergence and Services (IJITCS) Vol.2, No.1, February 2012 AUTONOMIC COMPUTING IN VIRTUALIZED ENVIRONMENT Anala M R 1 and Dr.Shobha G 2 1Department of CSE, RVCE, Bangalore [email protected] 2Dean PG Studies (CSE & ISE), RVCE, Bangalore [email protected] ABSTRACT The infrastructure of IT-sector is heterogeneous and complex. It is difficult for a system administrator to manage these complex systems. This led to complex, unmanageable and insecure systems. A new approach to solve this problem is the IT infrastructure automation. Autonomic computing is the solution to manage complex infrastructure which provides "must have" solutions for the problems above. The increased resource utilization in IT made the enterprises to use the costly physical servers. These physical servers were under-utilized. Virtualization is a technology with efficient and better utilization of resources. Virtualization and autonomic computing are of major importance for the IT- sector and could be successfully combined to reach the ideals of the autonomic computing. This paper presents a framework for automating live migration of virtual machines in virtualized environment using autonomic computing. This paper also addresses the types of attacks encountered during live migration of virtual machines. KEYWORDS Virtualization, autonomic computing, Self-management and live migration. 1. INTRODUCTION 1.1 Autonomic computing The heterogeneous infrastructures such as hardware and software applications, services and networks, directed towards the complex, unmanageable and insecure systems. The system administrator will not be able to manage alone such a complex computing system. Therefore these complex computing systems are managed by autonomic computing. Autonomic computing is the technique of managing the computing environment by computing systems themselves based on the policies decided by the system administrator. Self-management is the core of autonomic computing [1] and represents the process by which computing systems manage themselves. There are four aspects of self-management: • Self-configuration : The property in which the computing systems will automatically configure itself by sparing the system administrator with respect to high-level policies. Any new components added to the system are incorporated seamlessly. It can dynamically adapt to changing environments. Such changes could include the deployment of new components or the removal of existing ones, or dramatic changes in the system characteristics • Self-optimization : The computing system will continuously try to improve its performance by dynamically changing its parameter at run-time. It can monitor, allocate and tune resources DOI : 10.5121/ijitcs.2012.2108 67 International Journal of Information Technology Convergence and Services (IJITCS) Vol.2, No.1, February 2012 automatically. The tuning of resources means reallocation of resources in response to dynamically changing workloads in order to improve overall utilization. Without self- optimizing functions it is difficult to effectively use the computing resources when an application does not completely use its allocated resources. • Self-healing : the property of a computing system to detect, diagnose and repair local problems caused either by software of by hardware failures. It can discover, diagnose and react to disruptions. Self-healing components can detect system malfunctions and initiate policy-based corrective action without disrupting the IT environment. • Self-protecting : the ability of a computing system to secure itself from a malicious attack. The self-protected computing system is capable of predicting possible problems based on logs and reports. It can anticipate, detect, identify and protect against threats from anywhere. 1.2Virtualization Virtualization [2] is a technique of separating the resources of computer into multiple execution environments. Virtualization techniques create multiple partitions which are isolated with each other called virtual machines. A virtual machine is a software implementation of a physical machine where an isolated operating system is installed within the operating system of a physical machine. The desire to run multiple operating systems was the original motivation for virtual machines. Virtualization is a technology with efficient and better utilization of resources. Techniques of virtualization: There are three different approaches to achieve server virtualization: the virtual machine model, the paravirtual machine model, and virtualization at the operating system (OS) layer. Architecture of Virtual machine model: The first technique, virtual machine model uses the host/guest paradigm as shown in figure 1. Each guest runs on a virtual imitation of the hardware layer. This approach does not require any alteration to the guest operating system to run. The administrator can create guests that use different operating systems. The guest has no knowledge of the host's operating system because it is not aware that it's not running on real hardware. The guests require real computing resources from the host therefore it uses a hypervisor [3] to coordinate instructions to the CPU. The hypervisor is called a virtual machine monitor (VMM) [4]. It validates all the guest-issued CPU instructions and manages any executed code that requires addition privileges. VMware, Virtualbox and Microsoft Virtual Server both use the virtual machine model. In the bottom of the stack the Host OS manages the physical computer. The VMM manages all VMs running on this physical computer. The Guest OS and the application run on these virtual machines. 68 International Journal of Information Technology Convergence and Services (IJITCS) Vol.2, No.1, February 2012 Architecture of para-virtual machine model: As shown in figure 2 the paravirtual machine (PVM) model is also based on the host/guest paradigm. This technique also uses a virtual machine monitor. Unlike virtual machine model the para-virtual machine model the VMM actually modifies the guest operating system's code. As guest OS is to be modified, this is possible for only open source OS such as Linux and BSD and wnkdows cannot. The Intel(Intel VT) and AMD (AMD-V) provide functionality that enables unmodified OS to be hosted by a para-virtualized hypervisor. Like virtual machines, para-virtual machines are capable of running multiple operating systems. Xen[2,5] and UML both use the para-virtual machine model. Architecture of Virtualization at the OS level: This technique is different from virtual machine models and paravirtual machines. It isn't based on the host/guest paradigm. In the OS level model, the host runs a single OS kernel as its core and exports operating system functionality to each of the guests. Guests must use the same operating system as the host. Each virtual environment has its own file system, process table, networking configuration and system libraries as shown in figure 3. OS virtualization provides software emulation. This distributed architecture [6] eliminates system calls between layers, which reduces CPU usage overhead. Virtuozzo and Solaris Zones both use OS-level virtualization. The major limitation of this technique is the selection of OS since each container must be of same type, version and patch level. Comparison: The three techniques available for virtualization differ in implementation complexity, the extent of OS support, performance in comparison with standalone server, the level of access to common resources. The virtual machine models has wider scope of usage, but exhibits poor performance. Para-VMs have better performance, but they support fewer OSs because they modify OS. Therefore para-virtual machine models preferably use open source OS like linux. Migrating operating system instances across distinct physical hosts is a useful tool for administrators of data centers and clusters. It allows a clean separation between hardware and software, and facilitates fault management, load balancing, and low-level system maintenance. By carrying out the majority of migration while OSes continue to run, we achieve impressive 69 International Journal of Information Technology Convergence and Services (IJITCS) Vol.2, No.1, February 2012 performance with minimal service downtimes. Migration can be conducted offline (where the guest is suspended and then moved) or live (where a guest is moved without suspending). An offline migration suspends the guest then moves an image of the guest’s memory to the destination host. The guest is resumed on the destination host and the memory the guest used on the source host is freed. The time an offline migration depends on the network bandwidth and latency. A guest with 2GB of memory should take several seconds on a 1G bit Ethernet link. A live migration keeps the guest running on the source host and begins moving the memory without stopping the guest. Migration is useful for • Load balancing – Guests running in source host can be moved to some other host when guest doesn’t get sufficient resources at source host. • Hardware failover – When source machine’s hardware fails, for the continued services guests can be safely moved to new host. • Energy saving – During lower usage of a host system in order to save power guests can be reallocated to other hosts and host systems are powered off. • Geographic migration – To minimize the latency guests can be moved to another location in serious circumstances. During live migration of the virtual machine
Recommended publications
  • Personal Autonomic Computing Reflex Reactions and Self-Healing 305
    304 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS—PART C: APPLICATIONS AND REVIEWS, VOL. 36, NO. 3, MAY 2006 Personal Autonomic Computing Reflex Reactions and Self-Healing Roy Sterritt, Member, IEEE, and David F. Bantz, Member, IEEE Abstract—The overall goal of this research is to improve the Personal computing is an area that can benefit substantially self-awareness and environment-awareness aspect of personal au- from autonomic principles. Examples of current difficult expe- tonomic computing (PAC) to facilitate self-managing capabilities riences that can be overcome by such an approach include [2]: such as self-healing. Personal computing offers unique challenges for self-management due to its multiequipment, multisituation, and 1) trouble connecting to a wired or a wireless network at a multiuser nature. The aim is to develop a support architecture for conference, hotel, or other work location; 2) switching between multiplatform working, based on autonomic computing concepts home and work; 3) losing a working connection (and shouting and techniques. Of particular interest is collaboration among per- across the office to see if anyone else has had the same prob- sonal systems to take a shared responsibility for self-awareness and lem!); 4) going into the IP settings area in Windows and being environment awareness. Concepts mirroring human mechanisms, such as reflex reactions and the use of vital signs to assess oper- unsure about the correct values to use; 5) having a PC which ational health, are used in designing and implementing the PAC stops booting and needs major repair or reinstallation of the architecture. As proof of concept, this was implemented as a self- operating system; 6) recovering from a hard-disk crash; and 7) healing tool utilizing a pulse monitor and a vital signs health moni- migrating efficiently to a new PC.
    [Show full text]
  • Self-Adaptive Systems
    1 Self-Adaptive Systems: A Systematic Literature Review Across Categories and Domains Terence Wong, Markus Wagner, and Christoph Treude Abstract—Championed by IBM’s vision of autonomic computing paper in 2003, the autonomic computing research field has seen increased research activity over the last 20 years. Several conferences (SEAMS, SASO, ICAC) and workshops (SISSY) have been established and have contributed to the autonomic computing knowledge base in search of a new kind of system – a self-adaptive system (SAS). These systems are characterized by being context-aware and can act on that awareness. The actions carried out could be on the system or on the context (or environment). The underlying goal of a SAS is the sustained achievement of its goals despite changes in its environment. Despite a number of literature reviews on specific aspects of SASs ranging from their requirements to quality attributes, we lack a systematic understanding of the current state of the art. This paper contributes a systematic literature review into self-adaptive systems using the dblp computer science bibliography as a database. We filtered the records systematically in successive steps to arrive at 293 relevant papers. Each paper was critically analyzed and categorized into an attribute matrix. This matrix consisted of five categories, with each category having multiple attributes. The attributes of each paper, along with the summary of its contents formed the basis of the literature review that spanned 30 years (1990-2020). We characterize the maturation process of the research area from theoretical papers over practical implementations to more holistic and generic approaches, frameworks, and exemplars, applied to areas such as networking, web services, and robotics, with much of the recent work focusing on IoT and IaaS.
    [Show full text]
  • Introduction to Autonomic Concepts Applied to Future Self-Managed Networks
    Chapter 1 Introduction to Autonomic Concepts Applied to Future Self-Managed Networks Nazim Agoulmine DEFINITION AND SCOPE Network area has seen tremendous changes during the last decade through several waves. The first wave has been the Internet, which has completely changed the way network services are provided and helped global usage net- worked applications worldwide. The second wave came with the cellular and wireless technologies that have allowed the provisioning of telephony and data services anyplace, anytime. Cellular technologies have also drastically changed our behavior, allowing any person to be able to make phone calls from anywhere. They have also helped poor countries to develop efficient and cost- effective telephony infrastructure quickly, which was not possible with wire technologies. The third wave has surprisingly come from the POTS (Plain Old Telephony System) last mile access. Indeed, development of DSL (Digital Sub- scriber Line) technologies has enabled operators to provide high-speed IP access through a telephone line without having to pay for optical fiber installation and by taking advantage of existing telephone lines. This has allowed the emergence of the so-called Triple Play Services to the home (IP data access, Telephony of IP, and TV over IP). The fourth ware is probably under way with the con- vergence of the services Triple Play plus Mobile (sometimes called Quadruple Play) but also All-in-One emerging services such as P2P, social networking, and presence. In this ideal picture of technology development, there are many behind-the- scenes issues. Indeed, slowly but steadily, the users’ focus has changed from the high-speed network to value-added services.
    [Show full text]
  • A Concise Introduction to Autonomic Computing
    Advanced Engineering Informatics 19 (2005) 181–187 www.elsevier.com/locate/aei A concise introduction to autonomic computing Roy Sterritta,*, Manish Parasharb,1, Huaglory Tianfieldc,2, Rainer Unlandd,3 aSchool of Computing and Mathematics, Faculty of Engineering, University of Ulster at Jordanstown, Shore Road, Newtownabbey, County Antrim BT37 0QB, Northern Ireland, UK bThe Applied Software Systems Laboratory (TASSL), Electrical and Computer Engineering, Rutgers, The State University of New Jersey, 94 Brett Road, Piscataway, NJ 08854-8058, USA cSchool of Computing and Mathematical Sciences, The SRIF/SHEFC Centre for Virtual Organization Technology Enabling Research (VOTER), Glasgow Caledonian University, 70 Cowcaddens Road, Glasgow G4 0BA, UK dInstitute for Computer Science and Business Information Systems (ICB), University of Duisburg-Essen, Schu¨tzenbahn 70, 45117 Essen, Germany 1. Introduction specifically, an autonomic system is a self-managing, autonomous and ubiquitous computing environment that The advances in computing and communication tech- completely hides its complexity, thus providing the user nologies and software have resulted in an explosive growth with an interface that exactly meets her/his needs. The in computing systems and applications that impact all system will always decide on its own, using high-level aspects of our life. However, as the scale and complexity of guidance from humans, what needs to be done to keep it these systems and applications grow, their development, stable. It will constantly check and optimize its status, configuration and management challenges are beginning to and automatically adapt itself to changing conditions. break current paradigms, overwhelm the capabilities of Self-management is achieved through key aspects such as existing tools and methodologies, and rapidly render the self-governing, self-adaptation, self-organization, self- systems and applications brittle, unmanageable and optimization, self-configuration, self-diagnosis of faults, insecure.
    [Show full text]
  • Application of Autonomic Computing Principles in Virtualized Environment
    APPLICATION OF AUTONOMIC COMPUTING PRINCIPLES IN VIRTUALIZED ENVIRONMENT Anala M R 1 and Dr.Shobha G 2 1Department of CSE, RVCE, Bangalore [email protected] 2Dean PG Studies (CSE & ISE), RVCE, Bangalore [email protected] ABSTRACT The infrastructure of IT-sector is heterogeneous and complex. It is difficult for a system administrator to manage these complex systems. This led to complex, unmanageable and insecure systems. A new approach to solve this problem is the IT infrastructure automation. Autonomic computing is the solution to manage complex infrastructure which provides "must have" solutions for the problems above. The increased resource utilization in IT made the enterprises to use the costly physical servers. These physical servers were under-utilized. Virtualization is a technology with efficient and better utilization of resources. Virtualization and autonomic computing are of major importance for the IT-sector and could be successfully combined to reach the ideals of the autonomic computing. This paper addresses the design issues in automating live migration of virtual machines in virtualized environment using autonomic computing. KEYWORDS Virtualization, autonomic computing, Self-management and live migration. 1. INTRODUCTION 1.1 AUTONOMIC COMPUTING The heterogeneous infrastructures such as hardware and software applications, services and networks, directed towards the complex, unmanageable and insecure systems. The system administrator will not be able to manage alone such a complex computing system. Therefore these complex computing systems are managed by autonomic computing. Autonomic computing is the technique of managing the computing environment by computing systems themselves based on the policies decided by the system administrator. Self-management is the core of autonomic computing [1] and represents the process by which computing systems manage themselves.
    [Show full text]
  • Toward a Secure Autonomic Mobile Ad Hoc Cloud Computing System Devu Manikantan Shila, Wenlong Shen, Yu Cheng, Xiaohua Tian, and Xuemin (Sherman) Shen
    This article has been accepted for inclusion in a future issue of this magazine. Content is final as presented, with the exception of pagination. ACCEPTED FROM OPEN CALL AMCLOUD: TOWARD A SECURE AUTONOMIC MOBILE AD HOC CLOUD COMPUTING SYSTEM DEVU MANIKANTAN SHILA, WENLONG SHEN, YU CHENG, XIAOHUA TIAN, AND XUEMIN (SHERMAN) SHEN ABSTRACT improved manageability, and reduced mainte- nance. Furthermore, it could also significantly Cloud computing is a revolutionary paradigm boost partnership, agility, and scalability, thus to deliver computing resources, ranging from facilitating a truly global computing model. data storage/processing to software, as a service Nevertheless, existing archetypical cloud over the network, with the benefits of efficient computing models (public, private, or partner) resource utilization and improved manageability. encompass a cluster of expensive and dedicated The current popular cloud computing models machines to run those cloud computing resourc- encompass a cluster of expensive and dedicated es, leading to significant investment in capital machines to provide cloud computing services, outlay and ongoing costs. It is further observed incurring significant investment in capital outlay that the resources inside the data centers often and ongoing costs. A more cost effective solution operate at “low utilization” due to resource would be to exploit the capabilities of an ad hoc stranding and fragmentation. Moreover, the cloud which consists of a cloud of distributed and energy consumed by machines housed in data dynamically untapped local resources. The ad centers also embodies a financial burden on the hoc cloud can be further classified into static and organizations that operate them as well as an mobile clouds: an ad hoc static cloud harnesses infrastructure burden on power utilities.
    [Show full text]
  • Survey of Autonomic Computing and Experiments on JMX-Based Autonomic Features
    University of New Orleans ScholarWorks@UNO University of New Orleans Theses and Dissertations Dissertations and Theses Spring 5-13-2016 Survey of Autonomic Computing and Experiments on JMX-based Autonomic Features Adel R. Azzam The University of New Orleans, [email protected] Follow this and additional works at: https://scholarworks.uno.edu/td Part of the Systems Architecture Commons Recommended Citation Azzam, Adel R., "Survey of Autonomic Computing and Experiments on JMX-based Autonomic Features" (2016). University of New Orleans Theses and Dissertations. 2123. https://scholarworks.uno.edu/td/2123 This Thesis-Restricted is protected by copyright and/or related rights. It has been brought to you by ScholarWorks@UNO with permission from the rights-holder(s). You are free to use this Thesis-Restricted in any way that is permitted by the copyright and related rights legislation that applies to your use. For other uses you need to obtain permission from the rights-holder(s) directly, unless additional rights are indicated by a Creative Commons license in the record and/or on the work itself. This Thesis-Restricted has been accepted for inclusion in University of New Orleans Theses and Dissertations by an authorized administrator of ScholarWorks@UNO. For more information, please contact [email protected]. Survey of Autonomic Computing and Experiments on JMX-Based Autonomic Features A Thesis Submitted to the Graduate Faculty of the University of New Orleans in partial fulfillment of the requirements for the degree of Master of Science in Computer Science by Adel Azzam B.S. University of New Orleans, December 2010 May 2016 Dedication I dedicate this to my family, for all their love and support.
    [Show full text]
  • A Survey on Autonomic Computing: Inspired by the Human Body's Autonomic Nervous System
    Radha R, International Journal of Computer Science and Mobile Computing, Vol.10 Issue.5, May- 2021, pg. 81-86 Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology ISSN 2320–088X IMPACT FACTOR: 7.056 IJCSMC, Vol. 10, Issue. 5, May 2021, pg.81 – 86 A Survey on Autonomic Computing: Inspired by the Human Body's Autonomic Nervous System Radha R Assistant Professor, Department of Computer Science and Engineering Alliance University, Bangalore, India [email protected] DOI: 10.47760/ijcsmc.2019.v08i12.001 Abstract— Autonomic Computing is a concept that stands for self-management features of distributed computing agents being adaptive to unexpected changes and hiding the complication to users and operators. Introduced by IBM in 2001, this inventiveness ultimately meant to develop computer systems proficient of self-management, to overcome the fast growing intricacy of computing systems management, and to lessen the barrier that complexity poses to added growth. The progress of networks and the Internet, which have presented high accessible and obtainable services, have made atmospheres more multifaceted. The growing intricacy, rate, and heterogenic in distributed computing systems have interested researchers to explore a novel idea to manage with the supervision of intricacy in IT business. Autonomic Computing Systems (ACSs) have been presented for this reason. This paper consists of features of Autonomic Computing Systems, their effects, architecture and challenges. KEYWORDS: Autonomic Computing Systems, Self-Management Systems, Multi-agent System, Self-healing Systems. I. INTRODUCTION Autonomic computing technology is an emerging research topic to solve the software complexity crisis.
    [Show full text]
  • Why Computer-Based Systems Should Be Autonomic
    Why Computer-Based Systems Should be Autonomic Roy Sterritt Mike Hinchey School of Computing and Mathematics, NASA Goddard Space Flight Center Faculty of Engineering Software Engineering Laboratoly University of Ulster Greenbelt, MD 20771 Northern Ireland USA r.sterritt@,ulster.ac.uk [email protected] Abstract we need to reach out to other research communities to make this a reality. The objective of this paper is to discuss why The purpose of this paper is to consider why systems computer-based systems should be autonomic, where should be autonomic, how many more than we might autonomicity implies self-managing, often first think are already autonomic, and to look to the conceptualized in terms of being self-configwing, self- fiture and see what more we need to consider. healing, self-optimising, se2f-protecting and sew-aware. We look at motivations for autonomicity, examine how more and more systems are exhibiting autonomic 2. Facing Life As It Is.. behavior, and finally look atfuture directions. Upon launchmg the Autonomic Computing initiative, Keywords: Autonomic Computing, Autonomic IBM called on the industry ta face up to the ever- Systems, Self-Managing Systems, Complexity, Total increasing complexity and total cost of ownership. Cost of Ownership, Future Computing Paradigms. 2.1. Business Realities 1. Introduction 0 It is estimated that companies now spend between Autonomic Computing and other self-managing 33% and 50% of their total cost of ownership initiatives have emerged as a significant vision for the recovering from or preparing against failures [6]. design of computing based systems. Their goals are the 0 Many of these outages, with some estimates at as development of system that are self-configwing, self- high as 40%, are caused by operators themselves healing, self-protecting and self-optimizing, among other ~71.
    [Show full text]
  • Autonomic Cloud Computing: a Review
    International Journal of Computer (IJC) ISSN 2307-4523 (Print & Online) © Global Society of Scientific Research and Researchers http://ijcjournal.org/ Autonomic Cloud Computing: A Review Yazid Ado Ibrahima*, Alhassan Adamub, Salisu Mamman Abdulrahmanc, Akilu d Rilwan Muhammad a,b,cDepartment of Computer Science, Kano University of Science and Technology, Wudil, Wudil, Kano, 3244, Nigeria dDepartment of Computer Science, Federal University Dutse, Dutse, Nigeria aEmail: [email protected] bEmail: [email protected] cEmail: [email protected] dEmail: [email protected] Abstract The wide acceptability of cloud computing and its adoption, though remarkable, also gave the technology its greatest challenge in 'user expectation'. These challenges include; Reliability and availability, integration and interoperability, scalability, virtual machine migration policies, failure prediction, and resource management. In this paper, a general review of cloud computing was done highlighting its challenges. Autonomic Cloud computing was also reviewed. Future research areas were identified. Keywords: Cloud Computing; Autonomic Cloud. 1. Introduction Cloud computing is a large scale distributed network system which is implemented on servers in a data centre(s). That is, it refers to the applications that are delivered over the internet and the hardware and system software that provide it in data centres. This is largely viewed as a layer of services that are offered to users. The data centre(s) form a network of servers which serve as the main backbone to the cloud infrastructure. The Infrastructure as a service layer (IaaS) provides the necessary infrastructure for storage, servers, and networking components. Platform as a service (Paas) provides the necessary environment for building, deploying and testing custom applications.
    [Show full text]
  • Evaluation Issues in Autonomic Computing
    Evaluation issues in Autonomic Computing Julie A. McCann, Markus Huebscher Department Of Computing, Imperial College London {jamm, mch1} @doc.ic.ac.uk Abstract • Self-optimization: an autonomic comput- ing system optimises its use of resources. Autonomic Computing is a concept that brings It may decide to initiate a change to the together many fields of computing with the system proactively (as opposed to reactive purpose of creating computing systems that are behaviour) in an attempt to improve per- reflective and self-adaptive. In this paper we formance. draw upon our experience of this field to dis- • Self-healing: an autonomic computing cuss how we can attempt to evaluate auto- system detects and diagnoses problems. nomic systems. By looking at the diverse sys- What kinds of problems are detected can tems that describe themselves as autonomic, be interpreted broadly: they can be as low- we provide an introduction to the concepts of level as a bit-error in a memory chip Autonomic Computing and describe some (hardware failure) or as high-level as an achievements that have already been made. We erroneous entry in a directory service then discuss this work in terms of what is nec- (software problem) [3]. If possible, it essary to evaluate and compare such systems. should attempt to fix the problem, for ex- We conclude with a definitive set of metrics, ample by switching to a redundant compo- which we believe are useful to evaluate nent or by downloading and installing soft- autonomicity. ware updates. However, it is important that as a result of the healing process the system is not further harmed, for example 1.
    [Show full text]
  • A Review of the Autonomic Computing Models and Applications
    International Journal of Computer Applications (0975 – 8887) Volume 94 – No 4, May 2014 A Review of the Autonomic Computing Models and Applications Leandro Zambrano Alejandro Rosete Suárez Humberto Díaz Pando Méndez Facultad Ingeniería Facultad Ingeniería Facultad Ingeniería Informática. ISPJAE, Cuba Informática. ISPJAE, Cuba Informática. ISPJAE, Cuba ABSTRACT alternative paradigm for the system and application design, Todaymore and morebusiness processesof companiesrely on which is based on strategies used by biological systems to theirinfrastructureICT (Information Technology and deal with challenges similar scale, complexity, heterogeneity Communications).Along with thisbusiness evolution andthe and uncertainty, a view that has been referred to as autonomic broader economicenvironment, these computing computing [2]. infrastructureshave grown exponentiallyinthe The autonomic computing paradigm[1] has been inspired by storagecapacityandeaseand accessibilityfromanywhere in the the autonomy of the human nervous system.Its main objective worldusing multiple devices. However, thisdevelopment has is to detect computer systems and applications that can be notbeen accomplished bysimilar advancesin the operationof handled themselves in accordance with high level guidance these infrastructures, on the contrary way; infrastructures have from humans.The fulfillment of the great challenges of becomeincreasingly complexto manage.The complexitythat autonomic computing requires scientific and technological has achieved the systems and communicationsinfrastructureof
    [Show full text]