Declarative Task Delegation in Openmole

Total Page:16

File Type:pdf, Size:1020Kb

Declarative Task Delegation in Openmole Declarative Task Delegation in OpenMOLE Romain Florent Mathieu Thierry Nicolas David Reuillon Chuffart Leclaire Faure Dumoulin Hill ISC-PIF LIP ISC-PIF LISC LISC LIMOS CNRS INRIA, ENS CNRS Cemagref Cemagref CNRS, UBP [email protected] ABSTRACT cessor architectures and available applications. Even if grid infrastructures are able to take advantage of stor- In this paper we present OpenMOLE, a scientific frame- age and computational power through the grid mid- work providing a virtualized runtime environment for dis- dleware, a fine-grain analysis is still necessary at the tributed computing. Current distributed execution systems user level to achieve a successful execution on such an do not hide the hardware and software heterogeneity of environment. computing and data resources whereas OpenMOLE pro- vides generic services to develop distributed scientific al- The main object of this paper, is to describe gorithms independently from the execution environment ar- distributed computing facilities within OpenMOLE. chitecture. OpenMOLE uses abstraction layers to dele- OpenMOLE is a scientific framework, which aims at gate computing tasks with the same high level interface providing a virtualized runtime environment. In this for the major underlying architectures: local processors, approach the user can be confident that an experi- batch systems, computational grids, Internet computing ment successfully executed on a local computer will be and cloud computing. The file access abstraction layer is successfully executed on other execution environments another key feature helping a generic usage of the compu- without preliminary tests or software installation. tation power provided by grids and clusters. The Open- MOLE framework has been tested with the exploration This paper is organized as follows. Section one de- of a bacterial biofilm simulation with an individual-based scribes existing tools to execute jobs on large scale model. grid computing and clusters. The lack of existing tools compared to our main goal are pointed out. Sec- tion two presents novel methods implemented in Open- KEYWORDS: Task Delegation, Grid Computing, MOLE providing a transparent access to various dis- High Perfomance Computing, Distributed Simulation tributed computational environments. Exploration, Design of Computer Experiments. 2. RELATED WORK IN DISTRIBUTED EX- 1. INTRODUCTION ECUTION MANAGEMENT TOOLS The computational grid concept was originally pre- 1 2 sented as a metaphor of the US power grid [1]: a global Batch systems such as openPBS , Condor or Sun Grid 3 infrastructure where end-users can access computing Engine (SGE) and grid middleware like globus [2], power on demand. But this concept clashes with re- gLite [3] or DIET [4] provide integrated commands ality. Indeed, end-users are faced with heterogeneity and/or specific Application Programming Interfaces of the execution environments. Power grid has to deal (API) to submit and manage jobs as well as transfer with potential difference and electricity flow, whereas data. computational grid has to deal with not only network 1http://www.pbsgridworks.com/ and computing power availability considerations, but 2http://www.cs.wisc.edu/condor/ also higher level concepts like operating systems, pro- 3http://gridengine.sunsource.net/ 978-1-4244-6830-0/10/$26.00 ©2010 IEEE 55 Even though these distributed execution systems share tributed execution environments. Currently, the g- many common concepts, the set of commands provided Eclipse project supports implementations for two grid by each of them is specific. In this context, using ab- middleware systems: gLite7 [3] and gria [9] and one straction layers is a key issue to easily benefit from cloud computing service: Amazon EC2 (Elastic Com- the distributed computing power independently of the pute Cloud)8. These three environments are based on underlying architecture. A well designed generic ab- similar concepts but on radically dissimilar architec- straction layer for computational resources should al- tures, illustrating the genericity of the g-Eclipse mid- low the use of local processors, batch systems, compu- dleware abstraction layer. tational grids, Internet computing and cloud comput- ing through the same high level interface. Besides the generic access to several grid systems, g- Eclipse provides an implementation of the Eclipse file- 2.1. Ganga system (efs) to access grid storages. efs is an abstrac- tion layer in which files are referenced by Uniform Ganga [5] is a Python application for job management. Resource Identifier (uri)s 9. Actual file accesses are Ganga allows local and distributed executions on many achieved using dynamically registered modules con- grid middleware and batch systems based on: openpbs, taining implementations of the file system access pro- lsf (distributed by Platform4), dirac [6] and gLite [3]. tocols. Various implementations of efs are available, among them: SFTP, HTTP, CVS. Implementations Even though Ganga provides a generic job manage- for gLite [3] file access protocols have been developed ment on many cluster and grid environments, it does by the g-Eclipse project, including the following proto- not provide a full support for accessing grid storage cols: GridFTP (the grid file transfer protocol) and srm elements and file servers. Ganga enables the user to (Storage Resource Management). An implementation define input and output files for jobs. This function- for the on-line storage Amazon S3 is also available. ality fits the \bag-of-task" [7] paradigm, in which a job is transfered jointly with data it refers to in an G-Eclipse provides very useful classes to interact with object called input sandbox. However in many real various execution environments in a generic way. How- distributed simulation cases, jobs need to access large ever its development has been more focused on build- files. In this case, this voluminous file should not be ing a nice graphical interface interacting with the grid mentioned in the input sandbox. Otherwise the file than a nice application programming interface (API). would be uploaded whenever the job is launched in- A consequence for developers is that further develop- stead of being stored once and for all on a storage ment based on it, might be tricky. element. 2.3. JavaGAT and JSAGA In our experience, this missing functionality prevents designing generic jobs. Indeed, the lack of generic Java Grid Application Toolkit (JavaGAT) [10] is a file access interface necessitates writing specific (for Java API designed to abstract various distributed ex- a given execution environment) file access instructions ecution environments. It is designed as an extensible in the job executable code. Therefore, the file access environment in which modules for accessing computing abstraction layer is another key feature for generic us- platforms can be plugged. Modules, also called \adap- age of the computation power provided by grids and tors", for execution systems (such as gLite, Globus, clusters. SGE, SSH) and for remote file system access (such as HTTP, FTP, SFTP or GridFTP) are currently avail- 2.2. G-Eclipse able. On this work basis, the Simple API for Grid Ap- plications (SAGA) [11] generalizes the JavaGAT archi- G-Eclipse5 [8] is an integrated workspace environment tecture by providing a language-independent specifica- for accessing grid computing power. It is based on tion for accessing distributed execution environments. Eclipse6. G-Eclipse provides high-level abstractions Several implementations of SAGA have been achieved for grid resources such as job management services, in C++, Python and Java. This approach, endorsed authentication services and information services. This by the Open Grid Forum (OGF), is the most promis- abstraction layer is adaptable for various kinds of dis- 7http://glite.web.cern.ch/glite/ 4http://www.platform.com/ 8http://aws.amazon.com/ec2/ 5http://www.g-eclipse.eu/ 9URIs are normalized in the RFC3986, they allow identifying resources 6http://www.eclipse.org/ on a network such as servers, web services or files 56 ing we have seen for a transparent access to remote 3. OpenMOLE should be autonomous in accessing computing and storage resources. various remote execution environments, without requiring third party servers, 3. OPENMOLE 4. the end-user's desktop computer can be con- nected to the Internet without being necessarily Tools such as G-Eclipse, Ganga or JSAGA simplify the the owner of a public IP address (Internet access access to distributed execution environments, however via a Network Address Translation (NAT) system they do not hide the hardware and software hetero- is enough). geneity of computing resources. From our previous work on grid computing [12, 13], this drawback pre- vents developing distributed scientific algorithms in- Constraints 1 and 2 have been established so that the dependently from the execution environment architec- user can work on his own algorithms, without knowing ture. In order to tackle this problem, OpenMOLE how to deploy or execute them on various execution takes advantage from generic interface of JSAGA and environments. OpenMOLE should provide means to provides, on top of that, a virtualized runtime environ- integrate external programs, which can then be exe- ment as shown on Figure 1. Transparent delegation of cuted directly and with confidence on targeted execu- scientific algorithms to remote environment becomes tion environments despite the hardware and software possible within OpenMOLE. Indeed, virtualizing guar- heterogeneity. antees a successful
Recommended publications
  • Going Large-Scale in P2P Experiments Using the JXTA Distributed Framework Gabriel Antoniu, Luc Bougé, Mathieu Jan, Sébastien Monnet
    Going Large-scale in P2P Experiments Using the JXTA Distributed Framework Gabriel Antoniu, Luc Bougé, Mathieu Jan, Sébastien Monnet To cite this version: Gabriel Antoniu, Luc Bougé, Mathieu Jan, Sébastien Monnet. Going Large-scale in P2P Experiments Using the JXTA Distributed Framework. [Research Report] RR-5151, INRIA. 2004. inria-00071432 HAL Id: inria-00071432 https://hal.inria.fr/inria-00071432 Submitted on 23 May 2006 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. INSTITUT NATIONAL DE RECHERCHE EN INFORMATIQUE ET EN AUTOMATIQUE Going Large-scale in P2P Experiments Using the JXTA Distributed Framework Gabriel Antoniu, Luc Bougé, Mathieu Jan, Sébastien Monnet N˚5151 March 2004 THÈME 1 apport de recherche ISRN INRIA/RR--5151--FR+ENG ISSN 0249-6399 Going Large-scale in P2P Experiments Using the JXTA Distributed Framework Gabriel Antoniu, Luc Bougé, Mathieu Jan, Sébastien Monnet Thème 1 — Réseaux et systèmes Projet Paris Rapport de recherche n˚5151 — March 2004 — 12 pages Abstract: The interesting properties of P2P systems (high availability despite node volatil- ity, support for heterogeneous architectures, high scalability, etc.) make them attractive for distributed computing. However, conducting large-scale experiments with these systems arise as a major challenge.
    [Show full text]
  • Cloud Based Scaling of Grid Resources Through Grid Middleware
    International Journal of Computer Applications (0975 – 8887) Volume 64– No.15, February 2013 Cloud based Scaling of Grid Resources through Grid Middleware Mary Sumi Kurian S.P.Jeno Lovesum Post Graduate student, Department of Computer Assistant professor (SG), Department of Computer Science and Engineering, Karunya University, India Science and Engineering, Karunya University, India ABSTRACT The paper is organized as follows: Section 1 provides and In Grid computing, a common task is performed by combining introduction and background to cloud and grid computing. the resources from different locations or domains. It will use Section 2 explains the related works present in the field of grid- only the fixed number of resources available in different cloud integration. Section 3 explains the overall architecture locations. It doesn’t scale the resources according to the users’ and working of the system. Section 4 contains the details of the demand. On the other hand cloud computing provides resources experiments conducted and their analysis results. Finally to users according to their demand. Grid computing uses the section 6 discusses the conclusion and the future enhancement autoscaling capacity of cloud in resource scaling. That means, of the work presented in this paper. cloud provides resources to grid and thereby achieving resource autoscaling in grid computing. In this paper, grid resource 2. RELATED WORKS scaling using cloud is addressed. In this paper the combination Different Grid-Cloud integration is proposed in this section. of grid and cloud is achieved thorough a middleware called as I.M. Llorente et al. [3] proposed a system in which grid directly DIET.
    [Show full text]
  • Comparison on Openstack and Opennebula Performance
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by HAL-ENS-LYON Comparison on OpenStack and OpenNebula performance to improve multi-Cloud architecture on cosmological simulation use case Eddy Caron, Lamiel Toch, Jonathan Rouzaud-Cornabas To cite this version: Eddy Caron, Lamiel Toch, Jonathan Rouzaud-Cornabas. Comparison on OpenStack and Open- Nebula performance to improve multi-Cloud architecture on cosmological simulation use case. [Research Report] RR-8421, INRIA. 2013, pp.23. <hal-00916908> HAL Id: hal-00916908 https://hal.inria.fr/hal-00916908 Submitted on 10 Dec 2013 HAL is a multi-disciplinary open access L'archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destin´eeau d´ep^otet `ala diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publi´esou non, lished or not. The documents may come from ´emanant des ´etablissements d'enseignement et de teaching and research institutions in France or recherche fran¸caisou ´etrangers,des laboratoires abroad, or from public or private research centers. publics ou priv´es. Comparaison de performance entre OpenStack et OpenNebula et les architectures multi-Cloud: Application à la cosmologie. E. Caron, L. Toch, J. Rouzaud-Cornabas RESEARCH REPORT N° 8421 December 2013 Project-Team Avalon ISSN 0249-6399 ISRN INRIA/RR--8421--FR+ENG Comparaison de performance entre OpenStack et OpenNebula et les architectures multi-Cloud: Application la cosmologie. E. Caron∗†, L. Toch‡, J. Rouzaud-Cornabas‡ Project-Team Avalon Research Report n° 8421 — December 2013 — 20 pages Abstract: With the increasing numbers of Cloud Service Providers and the migration of the Grids to the Cloud paradigm, it is necessary to be able to leverage these new resources.
    [Show full text]
  • DIET Users Manual
    USER’S MANUAL VERSION 2.8 DATE November 2011 PROJECT MANAGER Fr´ed´eric Desprez. EDITORIAL STAFF Yves Caniou, Eddy Caron and David Loureiro. AUTHORS STAFF Abdelkader Amar, Rapha¨el Bolze, Eric´ Boix, Yves Caniou, Eddy Caron, Pushpinder Kaur Chouhan, Philippe Combes, Sylvain Dahan, Holly Dail, Bruno Delfabro, Benjamin Depardon, Pe- ter Frauenkron, Georg Hoesch, Benjamin Isnard, Mathieu Jan, Jean-Yves L’Excellent, Gal Le Mahec, Christophe Pera, Cyrille Pontvieux, Alan Su, C´edric Tedeschi, and Antoine Vernois. Copyright INRIA, ENS-Lyon, UCBL, SysFera DIET User’s Manual Contents Introduction 8 1 A Diet platform 11 1.1 Diet components .................................... 12 1.2 Communications layer ................................. 12 1.3 Diet initialization ................................... 13 1.4 Solving a problem ................................... 13 1.5 Diet Extensions .................................... 14 1.5.1 Multi-MA .................................... 14 1.5.2 CoRI ....................................... 14 2 Diet installation 15 2.1 Dependencies ...................................... 15 2.1.1 General remarks on Diet platform dependencies .............. 15 2.1.2 Hardware dependencies ............................ 15 2.1.3 Supported compilers .............................. 15 2.1.4 Operating system dependencies ........................ 16 2.1.5 Software dependencies ............................. 16 2.2 Compiling the platform ................................ 16 2.2.1 Obtaining and installing cmake per se .................... 16 2.2.2 Configuring Diet’s compilation: cmake quick introduction ......... 16 2.2.3 A ccmake walk-through for the impatients .................. 18 2.2.4 Diet’s main configuration flags ........................ 18 2.2.5 Diet’s extensions configuration flags ..................... 19 2.2.6 Diet’s advanced configuration flags ..................... 19 2.2.7 Compiling and installing ............................ 20 2.3 Diet client/server examples .............................. 21 2.3.1 Compiling the examples ...........................
    [Show full text]
  • Resource Localization Using Peer-To-Peer Technology for Network Enabled Servers*
    Resource Localization Using Peer-To-Peer Technology for Network Enabled Servers? Eddy Caron1, Fr´ed´eric Desprez1, Franck Petit2, and C´edric Tedeschi1 1 [Eddy.Caron,Frederic.Desprez,Cedric.Tedeschi]@ens-lyon.fr LIP Laboratory / GRAAL Project UMR CNRS, ENS Lyon, INRIA, Univ. Claude Bernard Lyon 1 46 All´ee d’Italie, F-69364 Lyon Cedex 07 2 [email protected] LaRIA Laboratory University of Picardie - Jules Vernes 5 rue du Moulin Neuf, F-80000 Amiens Abstract. DIET (Distributed Interactive Engineering Toolbox) is a set of hierarchical components to design Network Enabled Server systems. These systems are built upon servers managed through distributed scheduling agents for a better scalability. Clients ask to these scheduling components to find servers available using some performance metrics and information about the location of data already on the network. Our target architecture is the grid which is highly heterogeneous and dynamic. Clients, servers, and schedulers are thus better connected in a dynamic (or peer-to-peer) fashion. One critical issue to be solved is the localization of resources on the grid. In this paper, we present the use of an asynchronous version of the Propagate Information with Feedback algorithm to discover computation resources in Network Enabled Servers with distributed schedulers and arbitrary networks. Resource discovery uses peer-to-peer connections between components. Our implementation is based on JXTA from Sun Microsystems and DIET developed in the GRAAL team from INRIA. The algorithm and its implementation are discussed and performance results are given from experiments over the VTHD network which connects several supercomputers in different research institutes through a high-speed network.
    [Show full text]
  • DIET: a Scalable Toolbox to Build Network Enabled Servers on the Grid Eddy Caron, Fr´Ed´Ericdesprez
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by HAL-ENS-LYON DIET: A Scalable Toolbox to Build Network Enabled Servers on the Grid Eddy Caron, Fr´ed´ericDesprez To cite this version: Eddy Caron, Fr´ed´ericDesprez. DIET: A Scalable Toolbox to Build Network Enabled Servers on the Grid. RR-5601, INRIA. 2005, pp.28. <inria-00070406> HAL Id: inria-00070406 https://hal.inria.fr/inria-00070406 Submitted on 19 May 2006 HAL is a multi-disciplinary open access L'archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destin´eeau d´ep^otet `ala diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publi´esou non, lished or not. The documents may come from ´emanant des ´etablissements d'enseignement et de teaching and research institutions in France or recherche fran¸caisou ´etrangers,des laboratoires abroad, or from public or private research centers. publics ou priv´es. INSTITUT NATIONAL DE RECHERCHE EN INFORMATIQUE ET EN AUTOMATIQUE DIET: A Scalable Toolbox to Build Network Enabled Servers on the Grid Eddy Caron — Frédéric Desprez N° 5601 June 2005 Thème NUM apport de recherche ISRN INRIA/RR--5601--FR+ENG ISSN 0249-6399 DIET: A Scalable Toolbox to Build Network Enabled Servers on the Grid Eddy Caron , Fr¶ed¶eric Desprez Theme NUM | Systemes num¶eriques Projet GRAAL Rapport de recherche n 5601 | June 2005 | 28 pages Abstract: Among existing grid middleware approaches, one simple, powerful, and exible approach consists of using servers available in di®erent administrative domains through the classical client-server or Remote Procedure Call (RPC) paradigm.
    [Show full text]
  • Cloud Computing Resource Management Through a Grid Middleware: a Case Study with DIET and Eucalyptus Eddy Caron, Frédéric Desprez, David Loureiro, Adrian Muresan
    Cloud Computing Resource Management through a Grid Middleware: A Case Study with DIET and Eucalyptus Eddy Caron, Frédéric Desprez, David Loureiro, Adrian Muresan To cite this version: Eddy Caron, Frédéric Desprez, David Loureiro, Adrian Muresan. Cloud Computing Resource Man- agement through a Grid Middleware: A Case Study with DIET and Eucalyptus. [Research Report] RR-7096, INRIA. 2009, pp.12. inria-00435785 HAL Id: inria-00435785 https://hal.inria.fr/inria-00435785 Submitted on 26 Nov 2009 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. INSTITUT NATIONAL DE RECHERCHE EN INFORMATIQUE ET EN AUTOMATIQUE Cloud Computing Resource Management through a Grid Middleware: A Case Study with DIET and Eucalyptus Eddy Caron — Frédéric Desprez — David Loureiro — Adrian Muresan N° 7096 October 2009 Domaine 3 apport de recherche ISSN 0249-6399 ISRN INRIA/RR--7096--FR+ENG Cloud Computing Resource Management through a Grid Middleware: A Case Study with DIET and Eucalyptus Eddy Caron , Frédéric Desprez , David Loureiro , Adrian Muresan Domaine : Réseaux, systèmes et services, calcul distribué Équipe-Projet GRAAL Rapport de recherche n° 7096 — October 2009 — 9 pages Abstract: The Cloud phenomenon is quickly growing towards becoming the de facto standard of Internet Computing, storage and hosting both in industry and academia.
    [Show full text]
  • DIET Users Manual
    USER'S MANUAL VERSION 2.4 DATE April 2010 PROJECT MANAGER Fr´ed´eric Desprez. EDITORIAL STAFF Yves Caniou, Eddy Caron and David Loureiro. AUTHORS STAFF Abdelkader Amar, Rapha¨el Bolze, Eric´ Boix, Yves Caniou, Eddy Caron, Pushpinder Kaur Chouhan, Philippe Combes, Sylvain Dahan, Holly Dail, Bruno Delfabro, Peter Frauenkron, Georg Hoesch, Benjamin Isnard, Mathieu Jan, Jean-Yves L'Excellent, Gal Le Mahec, Christophe Pera, Cyrille Pontvieux, Alan Su, C´edric Tedeschi, and Antoine Vernois. Copyright INRIA, ENS-Lyon, UCBL DIET User's Manual Contents Introduction 8 1 A Diet platform 11 1.1 Diet components.................................... 12 1.2 Communications layer................................. 12 1.3 Diet initialization................................... 13 1.4 Solving a problem................................... 13 1.5 Diet Extensions.................................... 14 1.5.1 Multi-MA.................................... 14 1.5.2 FAST...................................... 14 1.5.3 CoRI....................................... 14 2 Diet installation 15 2.1 Dependencies...................................... 15 2.1.1 General remarks on Diet platform dependencies.............. 15 2.1.2 Hardware dependencies............................ 15 2.1.3 Supported compilers.............................. 15 2.1.4 Operating system dependencies........................ 16 2.1.5 Software dependencies............................. 16 2.2 Compiling the platform................................ 16 2.2.1 Obtaining and installing cmake per se.................... 16 2.2.2 Configuring Diet's compilation: cmake quick introduction......... 16 2.2.3 A ccmake walk-through for the impatients.................. 18 2.2.4 Diet's main configuration flags........................ 18 2.2.5 Diet's extensions configuration flags..................... 19 2.2.6 Diet's advanced configuration flags..................... 20 2.2.7 Compiling and installing............................ 21 2.3 Diet client/server examples.............................. 22 2.3.1 Compiling the examples...........................
    [Show full text]
  • Comparison on Openstack and Opennebula
    Comparison on OpenStack and OpenNebula performance to improve multi-Cloud architecture on cosmological simulation use case Eddy Caron, Lamiel Toch, Jonathan Rouzaud-Cornabas To cite this version: Eddy Caron, Lamiel Toch, Jonathan Rouzaud-Cornabas. Comparison on OpenStack and OpenNebula performance to improve multi-Cloud architecture on cosmological simulation use case. [Research Report] RR-8421, INRIA. 2013, pp.23. hal-00916908 HAL Id: hal-00916908 https://hal.inria.fr/hal-00916908 Submitted on 10 Dec 2013 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Comparaison de performance entre OpenStack et OpenNebula et les architectures multi-Cloud: Application à la cosmologie. E. Caron, L. Toch, J. Rouzaud-Cornabas RESEARCH REPORT N° 8421 December 2013 Project-Team Avalon ISSN 0249-6399 ISRN INRIA/RR--8421--FR+ENG Comparaison de performance entre OpenStack et OpenNebula et les architectures multi-Cloud: Application la cosmologie. E. Caron∗†, L. Toch‡, J. Rouzaud-Cornabas‡ Project-Team Avalon Research Report n° 8421 — December 2013 — 20 pages Abstract: With the increasing numbers of Cloud Service Providers and the migration of the Grids to the Cloud paradigm, it is necessary to be able to leverage these new resources.
    [Show full text]
  • Econstor Wirtschaft Leibniz Information Centre Make Your Publications Visible
    A Service of Leibniz-Informationszentrum econstor Wirtschaft Leibniz Information Centre Make Your Publications Visible. zbw for Economics Chacin, Pablo; Chao, Isaac; Freitag, Felix Working Paper Analysis of current middleware used in peer-to- peer and grid implementations for enhancement by catallactic mechanisms Bayreuther Arbeitspapiere zur Wirtschaftsinformatik, No. 3 Provided in Cooperation with: University of Bayreuth, Chair of Information Systems Management Suggested Citation: Chacin, Pablo; Chao, Isaac; Freitag, Felix (2005) : Analysis of current middleware used in peer-to-peer and grid implementations for enhancement by catallactic mechanisms, Bayreuther Arbeitspapiere zur Wirtschaftsinformatik, No. 3, Universität Bayreuth, Lehrstuhl für Wirtschaftsinformatik, Bayreuth, http://nbn-resolving.de/urn:nbn:de:bvb:703-opus-3078 This Version is available at: http://hdl.handle.net/10419/52626 Standard-Nutzungsbedingungen: Terms of use: Die Dokumente auf EconStor dürfen zu eigenen wissenschaftlichen Documents in EconStor may be saved and copied for your Zwecken und zum Privatgebrauch gespeichert und kopiert werden. personal and scholarly purposes. Sie dürfen die Dokumente nicht für öffentliche oder kommerzielle You are not to copy documents for public or commercial Zwecke vervielfältigen, öffentlich ausstellen, öffentlich zugänglich purposes, to exhibit the documents publicly, to make them machen, vertreiben oder anderweitig nutzen. publicly available on the internet, or to distribute or otherwise use the documents in public. Sofern die Verfasser die Dokumente unter Open-Content-Lizenzen (insbesondere CC-Lizenzen) zur Verfügung gestellt haben sollten, If the documents have been made available under an Open gelten abweichend von diesen Nutzungsbedingungen die in der dort Content Licence (especially Creative Commons Licences), you genannten Lizenz gewährten Nutzungsrechte. may exercise further usage rights as specified in the indicated licence.
    [Show full text]
  • (Iaas) with Amazon EC2/Eucalyptus
    Infrastructure-as-a-Service (IaaS) with Amazon EC2/Eucalyptus Florian Fittkau Christian-Albrechts-University Kiel Department of Computer Science 24098 Kiel, Germany Abstract. Cloud computing, for instance provided by Eucalyptus, be- comes more and more popular because it facilitates the construction of scalable applications. However, the horizontal scaling is not solely ex- hausted by web application in a cloud. Additional techniques for hori- zontal scaling are required. This paper describes Eucalyptus and Amazon EC2, and investigates the performance of the online-adaptation frame- work SLAstic utilizing Eucalyptus for automatic scaling of an example web application. The paper presents a performance evaluation of SLAs- tic utilizing Eucalyptus incorporating a xed number of nodes scenario and a varying number of nodes scenario. It shows that web applications running in Eucalyptus can automatically scale up and down with the usage of SLAstic. The results are demonstrated on the basis of JPetstore with a simulated typical day-night-cycle load for web applications. 1 Introduction Nowadays, many software service providers have to buy server hardware to serve the demand of their customers. This hardware has to be laid out for the peak rates of the workload. At times where the peak rate is not reached, the server hardware is idle, which is not economic. Cloud computing can help to solve this problem. When the user demand for the service increases, the cloud can provide more server instances and when it drops, some server instances can be terminated for saving money. Thus, the cloud provides the possibility for appli- cations to dynamically scale up and down.
    [Show full text]
  • DIET: a Scalable Toolbox to Build Network Enabled Servers on the Grid Eddy Caron, Frédéric Desprez
    DIET: A Scalable Toolbox to Build Network Enabled Servers on the Grid Eddy Caron, Frédéric Desprez To cite this version: Eddy Caron, Frédéric Desprez. DIET: A Scalable Toolbox to Build Network Enabled Servers on the Grid. [Research Report] LIP RR-2005-23, Laboratoire de l’informatique du parallélisme. 2005, 2+20p. hal-02102562 HAL Id: hal-02102562 https://hal-lara.archives-ouvertes.fr/hal-02102562 Submitted on 17 Apr 2019 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Laboratoire de l’Informatique du Parallélisme École Normale Supérieure de Lyon Unité Mixte de Recherche CNRS-INRIA-ENS LYON-UCBL no 5668 DIET: A Scalable Toolbox to Build Network Enabled Servers on the Grid Eddy Caron , Fr´ed´eric Desprez June 2005 Research Report No 2005-23 École Normale Supérieure de Lyon 46 Allée d’Italie, 69364 Lyon Cedex 07, France Téléphone : +33(0)4.72.72.80.37 Télécopieur : +33(0)4.72.72.80.80 Adresse électronique : [email protected] DIET: A Scalable Toolbox to Build Network Enabled Servers on the Grid Eddy Caron , Fr´ed´eric Desprez June 2005 Abstract Among existing grid middleware approaches, one simple, powerful, and flexible approach consists of using servers available in different administrative domains through the classical client-server or Remote Procedure Call (RPC) paradigm.
    [Show full text]