Laboratoire d’Informatique Fondamentale de Lille

Parallel : from Design to Implementation

Prof. E-G. Talbi

Parallel Cooperative Optimization Research Group Motivations

ƒ High-dimensional and complex optimization problems in many areas of industrial concern → Telecommunication, genomics, transportation, engineering design, …

ƒ Metaheuristics reduce the computational complexity of search BUT Parallel issue remains important: ƒ More and more complex metaheuristics ƒ Huge search space ƒ CPU or memory intensive of the objective function Motivations

ƒ Rapid development of technology ƒ Processors: Multi-core processors, … ƒ Networks (LAN & WAN): Myrinet, Infiniband, Optical networks, … ƒ Data storage

ƒ Increasing ratio performance / cost Goals

ƒ Speedup the search Æ real-time and interactive optimization methods, dynamic optimisation, …

ƒ Improve quality of solutions Æ cooperation, better even on a single processor

ƒ Improve robustness (different instances, design space)

ƒ Solving large problems Æ instances, accuracy of mathematics models Taxonomy of metaheuristics

Metaheuristics

Single Population solution

Hill Simulated Tabu Evol. Scatter Climbing annealing search algorithms search

ƒ Solution-based metaheuristics: Hill Climbing, Simulated Annealing, Tabu Search, VNS, …

ƒ Population-based metaheuristics: Evolutionary Algorithms, Scatter Search, Swarm optimization, … Outline

ƒ Parallel Metaheuristics: Design issues ƒ Parallel Metaheuristics: Implementation issues ƒ Hardware platforms, Programming models ƒ Performance evaluation ƒ Adaptation to Multi-objective Optimization ƒ Software frameworks ƒ IIlustration : Network design problem Parallel Models for Metaheuristics

ƒ An unified view for single-based metaheuristics and population based metaheuristics

ƒ Three major hierarchical models:

ƒ Algorithm-Level: Independent/Cooperative self-contained metaheuristics

ƒ Iteration-Level: parallelization of a single step of the (based on distribution of the handled solutions)

ƒ Solution-Level: parallelization of the processing of a single solution Parallel Models for Metaheuristics

Solution-level: Processing of a single solution (Objective / Data partitioning)

Algorithm-level: Independent walks, Multi-start model, Hybridization/Cooperation of metaheuristics

Iteration-level: Parallel evaluation of the neighborhood/population

Scalability ×× SPH

Heuristic Population / Neighborhood Solution Algorithm-level Parallel Model Meta. ƒ Independent on the problem ƒ Alter the behavior of the Meta metaheuristic Cooperation ƒ Design questions: ƒ When ? Migration decision: ƒ Blind: Periodic or Probabilistic Meta ƒ « Intelligent »: state dependent Meta. ƒ Where ? Exchange topology: complete graph, ring, torus, random, … Simulated Annealing ƒ Which ? Information : elite solutions, Genetic Programming search memory, … Evolution Strategy Tabu search ƒ How ? Integration Ant Colonies Scatter search, … Ex : Evolutionary Algorithms (Island Model ƒ Distribution of the population in a set of islands in which semi-isolated EAs are executed GA GA

GA GA ƒ Sparse individual exchanges are performed among these islands with GA GA the goal of introducing more diversity GA GA into the target populations

ƒ Improvement of robustness and quality Ex : The multi-start model in LS

ƒ Independent set of Local Searches ƒ Improve robustness and quality of solutions ƒ May start from the same or different initial solution, population ƒ Different parameters ƒ encoding, operators, memory sizes (tabu list, …), etc. Flat classification

Algorithm-Level Parallel Model

Homogeneous Heterogenenous

Partial Global

Specialist General

E-G. Talbi, «A taxonomy of hybrid metaheuristics », Journal of Heuristics, Vol.8, No2, 2002. Homogeneous / Heterogeneous

ƒ Homogeneous hybrids = Same Metaheuristic ƒ Heterogeneous = Different Metaheuristics

ƒ Example Evolutionary Algorithm Simulated Annealing Communication medium

Tabu Search

Several different metaheuristics cooperate and co-evolve some solutions Global / Partial ƒ Partial : Problem is decomposed in sub-problems. Each algorithm is dedicated to solve one sub- problem ƒ Combinatorial / Continuous : Decomposition of the problem, decision space, … ƒ Problem dependant: Vehicle routing, Scheduling, …

Problem

Sub-problem Sub-problem Sub-problem

Meta Meta Meta EAs, Tabu search Simulated annealing

Synchronization : Build a global viable solution Specialist / General ƒ Specialist: Algorithms which solve different problems (ex: Co-evolutionary) Diversifying Agent Intensifying Agent

Solutions Promising from solutions unexplored regions Refers to Refers to Adaptative Memory Explored regions Promising regions

explored Initial solutions Good space solutions

Local Search Agent

S. Cahon, N. Melab and E-G. Talbi. COSEARCH: A parallel cooperativ metaheuristic. Journal of Mathematical Modeling and algorithms JMMA, Vol.5(2), 2006. Iteration-level Parallel Model

ƒ Independent on the problem ƒ Complex objective functions (non linear, simulations, …)

ƒ Don’t alter the behavior of the Meta metaheuristic Æ Speedup the search Solution(s) Full ƒ Design questions: fitness ƒ Single-solution based algorithms: decomposition of the neighborhood ƒ Population based algorithms: decomposition of the population Ex : Single-Solution Based Metaheuristic

LS

Solution ƒ Evaluation of the neighborhood: computationally Partition of the intensive step neighborhood ƒ Decomposition of the neighborhood : ƒ Asynchronous for SA (# behavior) ƒ Synchronous for deterministic algorithms Ex : Population Based Metaheuristic

Selection Replacement ƒ Decomposition of the E.A. population: Couple of ƒ Individuals (EAs), Ants (AS), individuals Particles (PSO), etc. Evaluated offsprings ƒ Sharing Information: Pheromone matrix, etc. ƒ Asynchronous for steady state GAs (# behavior) Recombination, ƒ Synchronous for generational Mutation, GAs evaluation

E-G. Talbi, O. Roux, C. Fonlupt, D. Robillard, «Parallel ant colonies for the quadratic assignment problem», Future Generation Computer Systems, 17(4), 2001. Solution-level Parallel Model

Aggregation of ƒ Dependent on the problem partial fitnesses ƒ Don’t alter the behavior of the metaheuristic Æ Speedup the Solution Partial search (CPU or I/O intensive) fitness ƒ Synchronous ƒ Design questions: ƒ Data / Task Decomposition CPU #1 CPU #3 ƒ Data : database, geographical area, structure, … ƒ Task : sub-functions, solvers, …

CPU #2 CPU #4 Outline

ƒ Parallel Metaheuristics: Design issues ƒ Parallel Metaheuristics: Implementation issues ƒ Hardware platforms, Programming models ƒ Performance evaluation ƒ Adaptation to Multi-objective Optimization ƒ Software frameworks ƒ IIlustration : Network design problem Parallel Metaheuristics: Implementation issues

Design of Parallel Metaheuristics

Programming Paradigms

ParallelParallel Programm Programminging Execution Support EnvironmentsEnvironments Parallel Architecture Hardware P P P P P .. P

Main criteria : Memory sharing, Homogeneity, Dedicated, Scalability, Volatility

PP Processor Thread Process Shared Memory Machine

Memory ƒ Easy to program: conventional OS network and programming paradigms

CPU CPU CPU ƒ Poor Scalability: Increase in processors leads memory contention interconnection network: bus, crossbar, multistage ƒ Ex. : Dual-core (Intel, AMD), Origin crossbar (Silicon Graphics), .... Distributed Memory Architectures

ƒ Good Scalability : Several hundred CPU Memory nodes with a high speed interconnection network/switch network ƒ Harder to program Memory CPU Memory CPU ƒ Communication cost Interconnection schema : ƒ Ex. : Clusters hypercube, (2D or 3D) torus, fat-tree, multistage crossbars Clusters & NOWs ƒ Clusters: A collections of PCs interconnected through high speed network, ƒ Low cost ƒ Standard components

ƒ NOWs: Network Of Workstations ƒ take advantage of unused computing power ccNUMA architectures

ƒ Mixing SMP and DSM ƒ Small number of processors (up to 16) clustered in SMP nodes (fast connection, crossbar for instance) ƒ SMPs are connected through a less costly network with poorer performance

Memory CPU Memory CPU Memory CPU network network network CPUCPUCPU CPUCPUCPU CPUCPUCPU Interconnection network Periph. Grid Computing “Coordinated resource sharing and problem solving in dynamic, multi-institutional virtual organizations”

High-Performance High-Throughput Computing GRID Computing GRID

• Offer a virtual supercomputer ƒ Billions of idle PCs … ƒ Stealing unused CPU cycles of processors (a mean of 47%) ƒ Inexpensive, potentially very powerful but more difficult to program than traditional parallel computers

N. Melab, S. Cahon, E-G. Talbi, « Grid computing for parallel bioinspired algorithms», Journal of Parallel and Distributed Computing (JDPC), 66(8), 2006. GRID Platforms Lille

Orsay Nancy Rennes

Lyon

Grenoble Bordeaux Sophia Toulouse

HPC Grid: GRID’5000: 9 sites distributed in France and HTC Grid: PlanetLab: 711 inter-connected by Renater nodes on 338 sites over 25 5000 proc: between 500 and 1000 CPUs on each site countries

R. Bolze, …, E-G. Talbi, …, «GRID’5000: A large scale and highly reconfigurable Grid», International Journal of High Performance Computing Applications (IJHPCA), Vol.20(4), 2006. Main criteria for an efficient implementation

Criteria Shared Homogenous Dedicated Local Volatility Memory / / / network / & Fault Distributed Heterogenous Non Dedi. Large Tolerance Architecture network SMP SM Hom Dedi Local No

COW DM Hom Dedi Local No

NOW DM Het Non Local Yes

HP Grid DM Het Dedi Large No

HT Grid DM Het Non Large Yes Parallel Programming Environments

ƒ Shared-memory:

System Category Language binding PThreads C Java threads Programming language Java

OpenMP Compiler directives Fortran, C, C++

ƒ Distributed-memory: Message RPC or Object- Grid passing based systems computing Sockets Java RMI Globus

ƒ Hybrid model PVM CORBA Condor

MPI Algorithm-Level Parallel Model

ƒ Granularity = Execution cost between exchanges / Information exchanged ƒ Large Granularity Æ Well suited to large scale architectures ƒ Frequency of migration, Size of information exchanged ƒ Asynchronous exchange more efficient than Synchronous exchange for Heterogeneous or Non-dedicated Architectures ƒ Fault tolerance for Volatile architectures Æ Checkpointing (reduced cost) Iteration-Level Parallel Model

ƒ Granularity = Evaluation of a partition / Partition communication cost Æ ƒ Adapt granularity to target architecture (size of partitions) ƒ Asynchronous evaluation is more efficient: ƒ Heterogeneous or Non Dedicated or Volatile platforms ƒ Heterogeneous computation of the objective function (#solutions #costs) ƒ Scale: Limited by ƒ Size of population ƒ Size of Neighborhood. Solution-Level Parallel Model ƒ Granularity = Evaluation of sub-function / Cost of solution communication Æ Not well suited to Large scale and distributed memory architectures

ƒ Limited scalability (number of sub- functions or data partitions)

ƒ Synchronous Fault-tolerance issues

ƒ Important in volatile, non-dedicated, large-scale platforms ƒ Checkpointing – Recovery mechanism ƒ Application-oriented and NOT System-oriented (reduced complexity in time and space) ƒ Algorithmic-level: Current solution or population, iteration number, … ƒ Iteration-level: Which partition of population or neighborhood + fitnesses ƒ Solution-level: Solution + partial fitnesses

E-G. Talbi, Z. Hafidi, D. Kebbal, J-M. Geib, "A fault-tolerant parallel heuristic for assignment problems", Future Generation Computer Systems, 14(5), 1998. Load Balancing, Security, …

ƒ Static load balancing important for heterogeneous architectures ƒ Dynamic load balancing important for non dedicated, volatile architectures. ƒ Security issues important for large-scale architectures (multi-domain administration, firewalls, …) and some applications (medical and bio research, industrial, …) Performance evaluation

ƒ Speedup ?

SN = T1 / TN Ti : Execution time using i processors (wall clock time) N : number of used processors

ƒ If S > N Î Sublinear speedup ƒ If S = N Î Linear speedup ƒ If S > N Î Superlinear speedup

ƒ Absolute speedup

ƒ Strong: T1 = Best known sequential algorithm ?? ƒ Weak : - Population of N individuals compared to K islands of N/K individuals - Single-solution metaheuristic with N iterations WITH K SSMs with N/K iterations

E-G. Talbi, P.Bessière, "Superlinear speedup of a parallel genetic algorithm on the SuperNode", SIAM News, Vol.24, No.4, pp.12-27, Juil 1991. Performance evaluation ƒ Relative speedup ƒ Fixed number of generations ƒ Interesting for evaluating the efficiency of the implementation ƒ Superlinear speedup possible: architecture source (memory, cache, …) ƒ Convergence to a solution with a given quality ƒ Interesting for evaluation the efficiency of the parallel design (Algorithm-level parallel model) ƒ Superlinear speedup possible: search source (such as in branch and bound) ƒ Heterogenous or Non Dedicated architectures:

ƒ Efficiency: EN = SN * 100% / N (fraction of time processors are conducting work) ƒ 100% efficiency means linear speedup )( jt )( ∑ ∈ Jj ƒ t(j): time for task i, U(i): availaibility of worker i E = iU )( ƒ Stochastic : Mean, … ∑ ∈ Ii Outline

ƒ Parallel Metaheuristics: Design issues ƒ Parallel Metaheuristics: Implementation issues ƒ Hardware platforms, Programming models ƒ Adaptation to Multi-objective Optimization ƒ Software frameworks ƒ IIlustration : Network design problem Multi-Objective Optimization

= f f f xxxxf )(),...,(),()(min n≥2 (MOP) ()1 2 n s.c. ∈ Sx

ƒ Dominance f 2 ƒ y dominates z if and only if

∀i∈[1, …, n], yi ≤ zi

and ∃i∈[1, …, n], yi < zi ƒ Pareto solution Feasible solutions ƒ A solution x is Pareto if a solution which dominates x does not exist

Î Goal: Find a good quality and well Pareto front Solution found diversified set of Pareto solutions f 1

E-G. Talbi, et al., “Parallel multi-objective optimization”. In Multi-criteria Optimization …, Springer, 2007. Algorithm-Level Parallel Model

ƒ General and Global Cooperative Models: ƒ Which Information to exchange ? % Pareto archive or current populations (random, elite, uniform, …), … ƒ Any replacement strategy for the current population ƒ Partial: ƒ Decompose the Pareto front (objective space) ƒ # dominance criteria ƒ Specialist : ƒ Solve # problems (# objectives subsets) Iteration-Level Parallel Model

ƒ Many MOP applications with complex objectives (CFD – computational fluid dynamics, CEM – computations electromagnetics, FEM - finite element method, …) ƒ Pareto ranking : complex procedure to parallelize ƒ Pareto archiving : complex procedure to parallelize Solution-Level Parallel Model

ƒ Decomposition of the n objectives ƒ Multi-disciplinary Design Optimization ƒ Many engineering domains with different models (# disciplines, #solvers) ƒ Ex: Car design ƒ Optimize the air flow around a car Æ computational fluid dynamics (CFD) solver ƒ Optimize the toughness of materials Æ finite element method (FEM) solver Outline

ƒ Parallel Metaheuristics: Design issues ƒ Parallel Metaheuristics: Implementation issues ƒ Hardware platforms, Programming models ƒ Adaptation to Multi-objective Optimization ƒ Software Frameworks for Parallel Metaheuristics ƒ IIlustration : Network design problem Why ?

ƒ From scratch: high development cost, error prone, difficult to maintain, … ƒ Code reuse: difficult to reuse, adaptation cost, … ƒ Design and code reuse – software components: Hollywood principle « Don’t call us, we call you »

Combinatorial Optimization Problems (COPs) in practice: ƒ Diversity ƒ Continual evolution of the modeling (regards needs, objectives, constraints, …) ƒ Need to experiment many solving methods, techniques of parallelization, hybridization, parameters, … Design Objectives

ƒ Maximal Reuse of code and design ƒ Separation between resolution methods and target problems ƒ Invariant part given ƒ Problem specific part specified but to implement ƒ Flexibility et Adaptability ƒ Adding and updating other optimization methods, search mechanisms, operators, encoding, ... ƒ … to solve new problems ƒ Utility ƒ Large panel of methods, hybrids, parallel strategies, … ƒ Portability ƒ Deployment on different platforms (Standard ) ƒ Transparent access to performance and robustness ƒ Parallel implementation is tranparent to the target hardware platform ƒ Open source, Efficiency, Easy to use, … Examples

Metaheuristics Parallelism Parall. and dist. support at design at implementation ECJ E.A. Island cooperation Threads / Sockets D. BEAGLE E.A. Centralized model / Sockets Island cooperation. J-DEAL E.A. Centralized model Sockets DREAM E.A. Island cooperation Sockets / P2P MALLBA L.S. / E.A. All MPI, Netstream

PARADISEO LS / EA All PVM, MPI, Condor, PThreads, Globus

N. Melab, E-G. Talbi, S. Cahon, E. Alba, G. Luque, “Parallel metaheuristics: algorithms and frameworks”. In Parallel Combinatorial Optimization, Wiley, 2006. PARADISEO (PARAllel and DIStributed Evolving Objects)

„ PARADISEO in some words … http://paradiseo.gforge.inria.fr

„ An Open Source C++ framework (STL-Template) „ Paradigm-free, unifying metaheuristics „ Flexible regards the tackled problem „ Generic and reusable components (operators of variation, selection, replacement, criterion of termination, …) „ Many services (visualization, management of command line parameters, check-pointing, …) PARADISEO : http://paradiseo.gforge.inria.fr

PEO

MO MOEO

EO

ƒ Evolving Objects (EO) for the design of population-based metaheuristics, ƒ Moving Objects (MO) for the design of solution-based metaheuristics, ƒ Multi-Objective EO (MOEO) embedding features and techniques related to multi-objective optimization, ƒ PEO for the parallelization and hybridization of metaheuristics

S. Cahon, N. Melab and E-G. Talbi. ParadisEO: A framework for the reusable design of parallel and distributed metaheuristics. Journal of Heuristics, Vol.10(3), 2004. Architecture (level of execution) Parallel and distributed platforms

PEO

MO MOEO LAM-MPI PVM PThreads

EO

ƒ Parallelism and distribution ƒ Communication libraries (MPI LAM, PVM) → Deployment on networks/clusters of stations (COWs, NOWs) ƒ Multi-threading layer (Posix threads) → multi-core, multi-processors with shared memory (SMPs) ƒ Support of parallel and distributed environments → Clusters of SMPs (CLUMPS) ƒ Transparent to the user

S. Cahon, N. Melab and E-G. Talbi. Building with ParadisEO reusible parallel and distributed evolutionary Algorithms, Journal of Parallel Computing, , Vol.30(5), 2004. Architecture (level of execution) Grid computing

PEO

MO MOEO Master/Worker

EO Condor (HTC) Globus (HPC)

ƒ Gridification ƒ Re-visit parallel models taken into account the characterisitics of Grids ƒ Coupling of ParadisEO with a Grid middleware (Condor-MW and Globus)

ƒ Transparent volatility & checkpointing ƒ Ex : Definition in ParadisEO-CMW of the memory of each metaheuristic and the associated parallel models

N. Melab, S. Cahon and E-G. Talbi. Grid Computing for Parallel Bioinspired Algorithms. Journal of Parallel and Distributed Computing (JPDC), 2006. Illustration: Core classes of the Hill Climbing

UML notation (Unified Modeling Language) Illustration: Core classes of the Evolutionary Algorithm Illustration: The cooperative island model of E.A.s

/* To enable migrations (i.e. exchanges of individuals) */ eoPopChan pop_chan ; /* Migrations will occur periodically */ eoFreqContinue mig_cont (FREQ_MIG) ; /* Migrations are composed of random individuals */ eoRandomSelect mig_select_one ; /* Selector of NUM_EMIG emigrants*/ eoSelectNumber mig_select (mig_select_one, NUM_EMIG) ; /* Emigrants replace the worst individuals*/ eoPlusReplacement mig_replace ; /* The ring topology */ eoRingTopology topo (naming_chan) ; E.A. /* Building a manager of migrations */ eoDistAsyncIslandMig island_mig (naming_chan, pop_chan, mig_cont, E.A. mig_select, Migration mig_replace, pop, pop, topo) ; Illustration: The parallelization of the evaluation step

E.A.

Solution Full fitness Illustration: The parallelization of the objective function

Aggregation of partial fitnesses

Solution

Partial fitness Outline

ƒ Parallel Metaheuristics: Design issues ƒ Parallel Metaheuristics: Implementation issues ƒ Hardware platforms, Programming models ƒ Adaptation to Multi-objective Optimization ƒ Software frameworks ƒ IIlustration : Network design problem ƒ Combined use of the 3 parallel models ƒ Multi-objective problem ƒ Implemented on COWs, NOWs, HPC Grids and HTC Grids ƒ Using of PARADISEO – EO & MO & MOEO & PEO Design of radio networks in mobile telecommunication

„ Financial context (cost of the network)

„ Number of sites

„ Quality of Service

„ Network design

„ Positioning sites

„ Fixing a set of parameters for each antenna

„ NP-hard problem with:

„ Huge search space

„ High cost (CPU and memory ) objective functions, constraints. A brief description Working area

ƒ A set of base stations that satisfy Parameters of the following constraints … antennas ƒ Cover Données bornes ƒ Handover Ps [26, 55] dBm ƒ … and optimizes the following Diagram 3 types Hauteur [30, 50] m criterion Azim ut [0, 359] ° ƒ Min. the number of sites Inclinaison [-15, 0] ° ƒ Min. interferences TRX [1, 7] ƒ Max. yield traffic

Propagation model Used sites (Free spaceOkumura-Hata) Useless sites Handover area Cover area A multi-layer hierarchical parallel/hybrid metaheuristic

1. Cooperative island EAs

1. Deployment of 2. Distribution incremental of networks Local Searches

3. Parallel evaluation (data partitioning)

E-G. Talbi, H. Meunier, « Hierarchical parallel approach for GSM mobile network design», Journal of Parallel and Distributed Computing (JDPC), 66(2), 2006 Iteration-level Parallel Model (Homogeneous and Dedicated Cluster) ƒ Synchronous/Asynchronous ƒ Deploying irregular tasks → The computation time is dependent of the number of activated sites of the network ƒ Limited scalability of the synchronous model (size of the pop., e.g. 100) A.E.

Network Full fitness Solution-level Parallel Model

ƒ Synchronous, fine-grained ! Agregation ƒ More larger instance Solution Partial ƒ Poor scalability fitness

Partitioned data

Partitioning of the geographical data

(f1.3,f2.3, f3.3) (f1.1,f2.1, f3.1) (f1.2,f2.2, f3.2) Experimentation under PARADISEO (non-dedicated cluster of PCs)

Superlinear speedup models 2&3

model 2 alone

Parallel evaluation of the Parallel Evaluation of a solution population (model 2) (model 3) Synchronous vs. Asynchronous Influence of the granularity on the efficiency (synchronous) High-Throughput Computing Grid: Campus of Lille (3 # administrative domains)

Platform HTC Grid (Polytech, IUT, LIFL) Prog. Environment Condor Number of proc. 100 (heterog. and non dedicated) Cumulative wall clock time 30681 h. Wall clock time Almost 15 days Parallel efficiency 0.98

Highway Urban

E-G. Talbi, S. Cahon and N. Melab. Designing cellular networks using a parallel hybrid metaheuristic on the Grid. Computer Communications Journal, 30(2), 2007. High-Performance Computing Grid: GRID’5000 under Globus

ƒ 400 CPUs on 6 sites: Lille, Nice-Sophia Antipolis, Lyon, Nancy, Rennes

ƒ Parallel efficiency = 0.92 ƒ Best results obtained ƒ More than 22 years of cumulative wall clock time (other benchmark on 2465 GRID’5000: A fully reconfigurable grid! : processors) « images » having Globus and MPICH-G2 already installed. Conclusions

ƒ Unifying Parallel Models for Metaheuristics [Parallel] Metaheuristics

ƒ Clear separation between parallel design and Software framework for [Parallel] metaheuristics parallel implementation

ƒ Encourage the use of Different Architectures: software framework for Sequential, Cluster, NOW, [parallel] metaheuristics Grid, SMP, … Perspectives

ƒ Parallel models combining Metaheuristics & Exact methods (Algorithms, Coupling of Software, …) ƒ Parallel models for dynamic and robust optimization problems ƒ Parallel models for optimization problems with uncertainty ƒ Need a multiple evaluation of a solution ƒ Solving challenging problems on Grids (Ex: Molecular biology, Engineering design, …) Perpectives: Towards an international Grid

DAS

March 23th, 2006 Grid’5000 NAREGI

ƒ NAtional REsearch Grid Intiative (NAREGI) – Japan (11 sites) ƒ Distributed ASCII Supercomputer 2 (DAS2) – Netherlands (5 sites) • E-G. Talbi, « Parallel Combinatorial Optimization », Wiley, USA, 2006

• E-G. Talbi, A. Zomaya, « Grid computing for Bioinformatics and Computational Biology », Wiley, USA, 2007 Iteration-Level Parallel Model

ƒ Large-Scale to improve robustness : multiple evaluation Fitness

x x Design x−dx x+dx x−dx x+dx Illustration with a non-dedicated NOWs

Number of used processors (full day, 2 # administrative domains )

Condor : • Management and exploitation of distributed computational resources that are dynamic in space and time • Main features: Check-pointing, Security