KMD, PL-GRID PLATON Grid Infrastructure

Total Page:16

File Type:pdf, Size:1020Kb

KMD, PL-GRID PLATON Grid Infrastructure IntegrationIntegration ofof nationalnational andand EuropeanEuropean ee--InfrastructureInfrastructure Norbert Meyer On behalf of PIONIER Consortium Grids and e-Science workshop June 17th, 2009, Santander 15th years of Polish e‐Infrastructure for Science • 1993 –starting academic MANs (FDDI) • 1995 – MANs’ transition to ATM • 1997 – POL‐34, 155, 622 • 2000 –PIONIER take‐of • 2001 – dark fiber deployment • 2003 – 10GE • 2004 –multi‐lambda • 2006 – PIONIER2 strategy • 2008 – PLATON, PL‐GRID 21 ACADEMIC METROPOLITAN AREA NETWORKS • Area 320k sq km • Population 38M • Main academic centers 21 • State universities 120+ • Students 2M+ • R&D institutions and Univ. interconnected via PIONIER network 700+ MAN MAN & HPC Center What PIONIER is all about • developing country wide optical infrastructure based on the academic community ownership model • setting up national optical networks interconnecting on separate AS: 21 MANs, 5 HPCCs • extending optical reach to expensive and uniq national labs • grid technology as a main tool to integrate distributed R&D resources • portal as a main service delivery platform • broadband IP based services development • international co‐operation enabling and stimulating This was not that obvious in 1999, as is today… How PIONIER is organized: • PIONIER is a consortium of 22 academic MANs and HPCCs • PIONIER is supervised by PIONIER Board consisting of 22 representatives • PIONIER is managed by PIONIER Executive consisting of 4 people • PIONIER network is financed from the member fee • member fee is based on the cost sharing model • each year PIONIER Board take a decision about the framework and parameters of a cost sharing model • one member is selected to play the role of the PIONIER network operator (PSNC) KALININGRAD 2x10 Gb/s LITHUANIA 2x10 Gb/s GDAŃSK KOSZALIN PIONIER ELBLĄG SUWAŁKI 2Q2009 OLSZTYN SZCZECIN BYDGOSZCZ BELARUS 2x10 Gb/s BIAŁYSTOK GORZÓW TORUŃ 2 x 10 Gb/s /s DFN 2x10 Gb (2 λ) GÉANT2 10+10 Gb/s POZNAŃ CBDF 10Gb/s (2 λ) WARSZAWA GÉANT2/Internet ZIELONA 7,5 Gb/s GÓRA SDH 2,5 Gb/s ŁÓDŹ Internet 7,5 Gb/s ETH 1Gb/s RADOM WROCŁAW CZĘSTOCHOWA LUBLIN KIELCE MAN OPOLE PUŁAWY KATOWICE UKRAINE 2x10 Gb/s ZAMOŚĆ RZESZÓW BIELSKO-BIAŁA KRAKÓW CESNET; SANET 2x10 Gb/s PIONIER infrastructure is an integral part of ERA Network Infrastructure Phosphorus Porta Optica FEDERICA GÉANT/ GÉANT2/ GÉANT3 EMANICS MUPBED 6NET SEQUIN Atrium PIONIER Cross Border Fiber development directions Grid Infrastructure DORII Chemomentum RINgrid GridLab CrossGrid EUFORIA PRACE HPC Europa I/II EGEE I/II/III BalticGrid int.eu.grid InteliGrid FP5, FP6, FP7 OMII_Europe PROGRESS, g-Eclipse SGI GRID, CLUSTERIX, QosCos Grid ACGT KMD, PL-GRID ViroLab PLATON EuroGrid Scientific Data Layer Infrastructure sustained managed protected trusted repositories infrastructure concern for discoverable quality organisational selected contents context EuropeanaLocal ENRICH IMPACT, NMDB, METAFOR, CACAO EuroVO-AIDA, DRIVER/DRIVER2 Digital Libraries Federation GENESI-DR, DRIVER II Virtual Library of Science PIONIER2PIONIER2 Networks Grids Security Data HPC Common Services Added Services Intelligent Infrastructure for e‐Science A Concept of Platform Services Current national activities in Poland PLATON – Platform of Integrated Services PL-GRID – the Polish Grid Correlated projects KMD – National Data Storage R&D project PKI – Public Key Infrastructure Preparatory study, certificate service VoIP VoIP service PL‐Grid Foundations – Summary Polish Infrastructure for Supporting Computational Science in the European Research Space Response to the needs of Polish scientists and ongoing Grid activities in Poland, other European countries and all over the world •Motivation –E‐Science approach to research – EGI initiative ongoing in collaboration with NGIs •Creationof PolishGrid(PL‐Grid) Consortium: http://plgrid.pl – Consortium Agreement signed in January 2007 • PL‐Grid Project (2009‐2012) – Application in Operational Programme Innovative Economy, Activity 2.3 (in Sept. 2008) –Get funded March 2, 2009 (via European Structural Funds) •Consortiummadeupof five Polish supercomputing and networking centres (founders) –ACK CYFRONET AGH (Cracow) – Coordinator PL‐Grid Consortium Founders • Academic Computer Center Cyfronet AGH (ACK CYFRONET AGH) Coordinator • Poznań Supercomputing and Networking Center (PCSS) • Wrocław Centre for Networking and Supercomputing (WCSS) • Academic Computer Center in Gdańsk (TASK) • Interdisciplinary Center for Math. and Computat. Modelling, Warsaw University (ICM) PL‐Grid Infrastructure Polish Grid is going to have a common base infrastructure – similar to solutions adopted in other countries. Specialized, domain Grid systems – including services and tools focused on specific types of applications – will be built upon this infrastructure. These domain Grid systems can be further developed and maintained in the framework of separate projects. Such an approach should enable efficient use of available financial resources. Creation of a Grid infrastructure fully compatible and interoperable with European and World Grids thanks to cooperation with teams involved in the development of European Grid systems (EGEE, DEISA, OMII, C‐OMEGA, ESFRI). PL‐Grid Architecture Users Grid Application Programming Grid portals, development tools Interface Virtual organizations and security systems Other Grid LCG/gLite UNICORE Grids (EGEE) services systems Basic Grid services Distributed Grid Distributed National data computer computational resources resources repositories network The PL‐Grid Project is split into several workpackages P1 PROJECT MANAGEMENT Structure Coordination Dissemination Operation Rules P2 P6 PLANNING AND P3 DEVELOPMENT OPERATIONS CENTER SECURITY CENTER OF INFRASTRUCTURE P4 P5 GRID SOFTWARE SUPPORT FOR VARIOUS AND USERS gLite Unicore …. DOMAIN GRIDS TOOLS DEVELOPMENT Training Main Project Indicators: • Peak Perf.: 215 Tflops • Disk Storage: 2500 TB Challenges • Short term –To start – establishing PL‐Grid VO using Partners’ local and EGEE resources –To provide resources for covering operational costs –To select computational/storage resources for PL‐Grid infrastructure •Long term – continously –To provide the necessary infrastructure (!!) • Computer rooms, electrical power, many organizational issues –Be prepared / work on approaching paradigms and integration development • Clouds (internal‐external, computing clouds, data clouds) •SOA paradigm, knowledge usage … PLATON Integrated services platform for science – HD Videoconferencing services – EDUROAM service – Campus service – Archive service – Interactive research HDTV service VoIP service U6 U1 HD Videoconferencing services EDUROAM service PKI –global certifcates service U7 U2 U3 Campus services U4 Archive service U5 Interactive research HDTV service Management Platform Archive service 12,5 PB tapes 1 PB disc arrays Campus service in PLATON Project (1) Campus service in PLATON Project (2) Global management layer (open-source) Local resource managers On-demand Microsoft Graphical Other jobs virtual environment applications (batch, etc.) (Windows HPC machines Server 2008) Videoconferencing Portal Video Resources Reservation EDUROAM service Access Point Institution’s A Users’ Institution’s B Users’ Radius Database Radius Database Guest Internet id@inst‐b.pl VLAN VLAN Staff Guests VLAN Intermediate Students Radius Interactive research HDTV service (1) Interactive research HDTV service (2) Production studios & camera teams Camera Team Production Studio PLATON timetable – Take off 3Q2009 • EduROAM 4Q2009 • Videoconferencing 1Q2010 • Archive 2Q2010 • Campus services 3Q2010 • HDTV 1Q2011 – Harmonization / integration 2Q2012 Summary • Pionier integrates services on different levels – upto the application (proposition of domain platforms) • PL‐GRID and PLATON allow to deploy the national computing and data infrastructures • .. But the e‐Science requires additional services, like • Videoconferencing • HDTV Thank You for Your kind attention. May I answer Your questions, please ?.
Recommended publications
  • PIONIER – the National Fibre Optic Network for New Generation Services
    PIONIER – the national fibre optic network for new generation services Artur Binczewski, Maciej Stroiński Poznań Supercomputing and Networking Center e-IRG Workshop October 12 - 13, 2011; Poznań, Poland 18th years of Polish e-Infrastructure for Science • 1993 – starting academic MANs based on own fibers (FDDI) • 1995 – MANs’ transition to ATM • 1997 – POL-34, 155, 622 • 1999 – joining to TEN-155 (precursor of GÉANT) network • 2000 – PIONIER take-of • 2001 – dark fiber deployment & 1st EU funded project: ATRIUM • 2003 – 10GE • 2004 – multi-lambda • 2006 – PIONIER2 strategy • 2008 – PLATON project • 2009 – NEWMAN project • 2011 – 100NET project National Research and Education Network - PIONIER (topology) • Area 312k sq km • Population 38M • Main academic centers 21 • State universities 165+ • Researchers & Students 2M+ • R&D institutions and Univ. interconnected via PIONIER network 700+ 6494 km of fiber infrastructure in Poland 763 km of fiber in Germany (IRU) 7257 km of fiber in total 21 MANs and 5 HPC Centers in PIONIER Consortium with PSNC as Operator National Research and Education Network - PIONIER (3Q2011) RUSSIA (Kaliningrad) 1 x 10 Gb/s GDAŃSK LITHUANIA 1 x 10 Gb/s Elbląg Suwałki KOSZALIN OLSZTYN SZCZECIN BIELARUS 1 x 10 Gb/s BYDGOSZCZ BIAŁYSTOK Gorzów TORUŃ Hamburg (GLIF, Surfnet, Nordunet) 4x10 Gb/s POZNAŃ Sochaczew GÉANT2 10+10 Gb/s WARSZAWA ZIELONA GÓRA ŁÓDŹ PUŁAWY LUBLIN RADOM WROCŁAW CZĘSTOCHOWA KIELCE 2 x 10 Gb/s (2 lambdas) OPOLE Zamość CBDF 10Gb/s KATOWICE UKRAINE 1x10 Gb/s Colorless, directionless and contentionless Bielsko-Biała
    [Show full text]
  • The Polish Road to Optical Networks
    The Polish Road to Optical Networks Artur Binczewski, Maciej Stroiński artur | [email protected] Pisa, 10-12.05.2005 National programs of development IT infrastructure for academic community • State Committee of Scientific Research created the national program of development IT infrastructure in 1992-2000: 421 Metropolitan Area Networks 45 High Performance Computing Centers 4National Network (NASK, POL-34/155/622) 4LANs and campus networks in research institutions 4Resource servers and databases in MANs POZMAN – Poznan Metropolitan Area Network • Own fiber network (under 220 km) • 105 connections to research institutions • Simple migration to new technologies • 1992 - FDDI • 1995 - ATM • 2000 - Gigabit Ethernet • 2005 - 10 Gigabit Ethernet POZMAN Metropolitan Area Network 1Gb M10 ps 1G bps PSNC M10 Resources Network 10 G Juniper T320 bp s s p BD6808 Catalyst 6509 b PIONIER G 10 Catalyst 6503 Catalyst 6503 Catalyst 6506 Catalyst 6503 10GigabitEthernet Layer 2 Catalyst 6503 Catalyst 6503 Catalyst 6503 Core Network Catalyst 6503 1Gbps 1Gbps 1Gbps 1Gbps AT-8024GB 1GbpsAT-8326GB 1Gbps 1Gbps AT-8024GB AT-8326GB AT-8024GB Access Layer HPC centers in Poland PSNC: SGI Origin3800 (MIPS R12k/160), Cray SV1/8, Cluster PC (IA-64/250), SunFire 6800 (UltraSPARCIII/48), SGI Onyx2 (MIPS R16k/8, Infinity Reality Engine), SGI Altix ACK Cyfronet: SGI Origin2000 (MIPS R14k/128), HP Integrity Superdome (IA-64/8), SunFire 6800 (UltraSPARCIII/20), Cluster PC (Intel PentiumIII, Xeon/80), Cluster PC (IA-64/40). TASK: Cluster PC (IA-64/256), Cluster PC (IA-32/Xeon/128). ICM: Cray SV1ex (32), SGI Origin2000 (MIPS R14k/16), SUN E10K (UltraSPARC/16) WCSS: SGI Origin2400 (MIPS R14k/32), Cluster PC (Xeon/58).
    [Show full text]
  • PIONIER-Research and Education Networking in Poland
    PIONIER - Research and Education Networking in Poland Bartosz Belter, Poznan Supercomputing and Networking Center, Poland SEAIP2014, December 2014, Tainan, Taiwan Poznan Supercomputing and Networking Center Center of e-Infrastructure • National Research and Education Network PIONIER • Research Metropolitan Area Network - POZMAN • HPC Center • Data repositories and Digital Libraries Federation Center for R & D • New Generation Networks • HPC, Grids & Clouds • Grand challenge applications • New media and visualization technologies • Knowledge Platforms • Future Internet - Technology, Applications and Services for IS • Cyber Security PIONIER in a nutshell The PIONIER Consortium: 21 MANs and 5 HPC Centers (PSNC as the operator) • Area 312k sq km • Population 38M • Main academic centers 21 • State universities 165+ • Students 2M+ • R&D institutions and Univ. interconnected via PIONIER network 700+ 6479 km of fiber infrastructure in Poland 2359 km of fiber in Europe (IRU) 8838Source km of- http fiber://www.glif.is/publications/maps in total PIONIER: international collaboration GÉANT GLIF PIONIER EU PIONIER @GLIF PIONIER network infrastructure Regional Networks PIONIER National Infrastructure NEWMAN project • Nation scale deployment of MPLS/IP platform for all MANs and PIONIER – Last mile is provided for most of users – Full portoflio of offered services: IPV4, IPv6, MPLS, VPLS, multicasts, QoS,… • New generation DWDM with automatic restoration (GMPLS) – 100G ready – Colorless, Directionless and Contentionless functions implemented – All nodes
    [Show full text]
  • PIONIER Services
    and then things speeded up… • 1993 – starting academic MANs (FDDI) • 1995 – MANs’ transition to ATM • 1997 – POL-34, 155, 622 • 2000 – PIONIER take-of • 2003 – 1 lambda - 10GE • 2004 – multi-lambda 21 ACADEMIC METROPOLITAN AREA NETWORKS • Population 38M • Main academic centers 21 • State universities 120+ • Students 2M+ • R&D institutions and Univ. interconnected via PIONIER network 700+ MAN MAN & HPC Center What PIONIER is all about • closing the gap to EC academic communities an more… • developing country wide optical infrastructure based on the academic community ownership model • setting up national optical networks interconnecting on separate lambdas: 21 MANs, 5 HPCCs and to-do-research network • extending optical reach to expensive national laboratories • grid technology as a main tool to integrate distributed R&D resources • portal as a main service delivery platform • broadband IP based services development • international co-operation enabling and stimulating This was not that obvious in 1999, as is today… How PIONIER is organized: • PIONIER is a consortium of 22 academic MANs and HPCCs • PIONIER is supervised by PIONIER Board consisting of 22 representatives • PIONIER is managed by PIONIER Executive consisting of 4 people • PIONIER network is financed from the member fee • member fee is based on the cost sharing model • each year PIONIER Board take a decision about the framework and parameters of a cost sharing model • one member is selected to play the role of the PIONIER network operator (PSNC) PIONIER dark fiber topology Operational
    [Show full text]
  • Optical Fiber Communications in Poland
    More light in Polish optical fibres Ryszard S. Romaniuk [email protected] Institute of Electronic Systems, Warsaw University of Technology ABSTRACT Optical communications infrastructure is undergoing an intense development in his country now. A number of international investors and domestic operators are building from the beginning, modernizing or developing proprietary network or leasing wide area systems on a large scale. The aggregate level of these processes is of the order of bil. $. Despite of this the network is not homogeneous, has not satisfactory bandwidth, lacks the QoS, has inadequate international connections, and (according to prevailing opinions) the prices are too high for corporate as well as private users. The intense development of the optical infrastructure is governed by two dominant tendencies: burying new large, fat, optical pipes – cables containing even as much fibers as 500 (for C, L and XL optical bands) and investments in DWDM for main traffic directions (previously working in 1300 nm band). Keywords: Optical fibre technology, optical fibers, optical communications, optoelectronics, Internet 1. INTRODUCTION The economic crisis of the last decade has touched also certain sectors of the communication field, in particular cable carriers, and to a lesser degree wireless and mobile carriers. This enforces strong consolidation activities in the global scale. The telecom market is segmented and re-grouped. Big operators aim frequently at new investments in such conditions. These processes can be observed on domestic telecom market in this country. The last years are a period of very intense investments in optical fibre communications infrastructure in this country, especially in backbone networks. Several operators carry out the investments: Telia, Energis, TPSA, Netia, Telbank, Tel-Energo, Railway-Telecom and others.
    [Show full text]
  • Prezentacja Programu Powerpoint
    Poland - e-Infrastructure ecosystem and relation to EOSC Norbert Meyer [email protected] in co-operation with Ministry of Science and Higher Education Polish research infrastructure stakeholders ✓ Owners of the infrastructure ✓ MAN institutions + HPC centres ✓ Service providers ✓ MAN + HPC and universities ✓ Users ✓ Universities, R&D institutions, Polish Academy of Science institutes ✓ Digital libraries, hospitals ✓ Founders ✓ Ministry of Science and Higher Education ✓ Ministry of Regional Development, Marshals of the regions ✓ Inkind contribution of MAN and HPC centres Partners: PIONIER consortium – easier together ! 1. Institute of Bioorganic Chemistry of the Polish Academy of Science – Poznań Supercomputing and Networking Center 2. University of Technology and Life Sciences in Bydgoszcz 3. AGH University of Science and Technology – Academic Computer Centre CYFRONET 4. Institute of Soil Science and Plant Cultivation – State Research Institute 5. Maria Curie – Skłodowska University in Lublin LUBMAN UMCS 6. Bialystok University of Technology 7. Czestochowa University of Technology 8. Gdansk University of Technology Academic Computer Centre TASK 9. Koszalin University of Technology 10. Technical University of Lodz 11. Technical University of Radom 12. Rzeszow University of Technology 13. West Pomeranian University of Technology Szczecin 14. Silesian University of Technology – Computer Centre 15. Kielce University of Technology 16. Wrocław University of Technology 17. Nicolaus Copernicus University 18. Opole University 19. University of
    [Show full text]
  • POLFAR Consortium Activities
    Current POLFAR Consortium Activities Andrzej Krankowski Leszek Blaszkiewicz Chair of POLFARO Consortium Space Radio-Diagnostics Research Centre (SRRC/UWM), University of Warmia and Mazury in Olsztyn, Poland BALTICS SCIENTIFIC CONFERENCE Latvia, Jūrmala, 5 December 2018 POLFARO August, 2015 University of Warmia and Mazury in Olsztyn, the leader of the POLFARO Consortium - (coordinator: prof. Andrzej Krankowski) Jagiellonian University, Krakow - (dr hab. Marian Soida, prof. UJ) Space Research Centre of PAS, Warsaw - (dr hab. Hanna Rothkaehl, prof. CBK) PCSS/PIONIER- (Robert Pekal) ż Nicolaus Copernicus Astronomical Center of PAS in Warsaw, Torun (dr hab. Jarosław Dyks, prof. CAMK) The Nicolaus Copernicus University in Torun (NCU) (dr hab. Magdalena Kunert-Bajraszewska) Szczecin University (dr hab. Ewa Szuszkiewicz, prof.US) University of Zielona Góra (dr hab. Jarosław Kijak, prof.UZ) Wrocław University of Environmental and Life Sciences (prof. Bernard Kontny) LOFAR- The Key Science Projects Epoch of Reionisation Surveys Transients ż Cosmic Rays Magnetism Sun, Space Weather Baldy, Borowiec, Lazy LOFAR stations ż POLFARO PIONIER GLIF GÉANT POLFARO PIONIER EU /PL Pulsars • Big help from the GLOW (D. Schwarz and Bielefeld group mainly) (We are using GLOW observational and raw data reduction tools and them well known software like DPSR and PSARCHIVE) ż • Close cooperation with Zielona Góra pulsar’s team POLFARO Sun and scintillations • Big help in observations initiating and support in solar LOFAR research from Richard Fallows (ASTRON), Derek McKay (Sodankyla Geophysical Observatory and STFC Rutherford Appleton Laboratory, Didcot, UK) and Gottfried Mann (Leibniz-Institut fur Astrophysik Potsdam) from KSP team. ż • Close cooperation on data analysis with University of Wroclaw solar team (prof.
    [Show full text]
  • PIONIER Organizational Model of E-Infrastructure for Research and New Services Development
    PIONIER Organizational model of e-infrastructure for research and new services development Robert Pękal [email protected] . 20th years of Polish e-Infrastructure for Science • 1993 – starting academic MANs based on own fibers (FDDI) • 1995 – MANs’ transition to ATM • 1997 – POL-34, 155, 622 • 1999 – joining to TEN-155 (precursor of GÉANT) network • 2000 – PIONIER take-of • 2001 – dark fiber deployment & 1st EU funded project: ATRIUM • 2003 – 10GE • 2004 – multi-lambda • 2006 – PIONIER2 strategy • 2008 – PLATON project • 2009 – NEWMAN project • 2011 – 100NET project • 2013 – CBPIO project Polish e-Infrastructure • 21 Academic Optical Based MAN (Metropolitean Area Network) • 5 HPC (High Performance Computing) Centers • Connected with owned fiber infrastructure – PIONIER network (Polish Optical Internet) • Science Services Platform - PLATON • Digital Libraries Federation • National Data Storage European & Polish Models of e-Infrastructure European Model Polish Model Responsible Intitutions MAN activity • Local dark fibre investment and ownership • Delivering IP based & VPN services • Cost sharing (recovery) model for academic community • Testbeds in new technologies • Taking part in Regional Development Plan setup • R&D regional IT infrastructure development coordination Every academic MAN: • Is acting under the umbrella of the legal entity of university or R&D institution • Is obliged to service regional academic community • Has a legal permission to act as IT operator or ISP • Is authorized to apply for Ministry of Science grants to develop new services • Is co-operating closely with local community • Is a member of the Polish national PIONIER Consortium Example of MAN – POZMAN 220 km of own Fiber Optic Cables in the city of Poznań • 600k city population • 105 connections to R&D institutions & Univ.
    [Show full text]
  • PDF Presentation
    TELEMEDICINE in POLAND AntoniAntoni NowakowskiNowakowski Department of Biomedical Engineering Gdansk University of Technology Gdańsk, POLAND Ministry of Science and Higher Education Committee of Information Infrastructure CODATA Poland Computer networks in Poland: KASK till 1990 (practically not existing); beginning of TCP/IP - EARN 1991 - CIUW INTERNET - NASK - 1993 (6000 hosts in Poland) HPCC - 1994 CI TASK - http://www.task.gda.pl/ GDANSK ICM - http://www.icm.edu.pl/ WARSZAWA CYFRONET - http://www.cyfronet.krakow.pl/ KRAKOW PCSS - http://rose.man.poznan.pl/ POZNAN WCSS - http://www.wcss.wroc.pl/ WROCLAW 1995 - 2000 - Programme of information infrastructure for science development (full accessibility of any scientist in Poland to the INTERNET & HPC) 2000 - e-Poland programme 2001 - 2005 - PIONIER: Polish Optical Internet - Advanced Applications, Services and Technologies for Information Society 2006 - 2012 - PIONIER 2 Development of MAN in 21 regions II is a property of regional scientific institutions! & since 2000 development of national inter-MAN optical backbone “PIONIER” - as property of MAN consortium NASK - 1998 2003 April POL 34 & MAN - 1998 PIONIERPIONIER • > 5000 km inter-metropolitan fiber-optic network (NZDS G.655 + SM G.652) • System DWDM with lambda (to 40Gb/s) • Optical switches - GMPLS • Optical VPN ¾ Internet ¾ HPC network ¾ administration network ¾ dedicated virtual networks for specific projects as ATLAS, VLBI Gdańsk Słupsk Elbląg Koszalin Olsztyn s / b M Bydgoszcz Białystok Szczecin 4 3 Toruń /s b M 5 5 1 Poznań
    [Show full text]
  • The New Challenges for PIONIER Infrastructure Artur Binczewski
    The New Challenges for PIONIER Infrastructure 10th CEF Networks Workshop 2019 Artur Binczewski Network Technology Division Director EU Policies for e-Infrastructures for research in IT area STRESSES the importance of PRACE, a world-class European High Performance Computing (HPC) infrastructure for research that provides access to computing resources and services for large-scale scientific and engineering applications; ACKNOWLEDGES the need to develop the new generation of HPC technologies and CALLS for the reinforcement of the interconnected network of data processing facilities GEANT. In this respect, INVITES ESFRI to explore MechanisMs for better coordination of MeMber States’ investMent strategies in e-infrastructures, covering also HPC, distributed coMputing, scientific data and networks; http://data.consilium.europa.eu/doc/document/ST-9360-2015-INIT/en/pdf The European Strategy ForuM on Research Infrastructures – Report 2018 http://roadmap2018.esfri.eu/media/1066/esfri-roadmap-2018.pdf Future visions: HPC Europe & EOSC but it is subject for another presentation Source: https://ec.europa.eu/newsroom/dae/document.cfm?doc_id=60156 https://publications.europa.eu/en/publication-detail/-/publication/78ae5276-ae8e-11e9-9d01-01aa75ed71a1/language-en Direction: national strategies and e-infrastructure initiatives: Polish RoadMap of Research Infrastructure (PRRI) • Launched in 2009 • Updated in 2013 • The current PRRI contains 53 project proposals in many broadly understood fields of science: • physical and mathematical sciences - 14, • technical
    [Show full text]
  • Why Is Ipv6 Deployment Important for the Internet Evolution?
    Paper Why is IPv6 Deployment Important for the Internet Evolution? Jordi Mongay Batallaa, Artur Binczewskib, Wojciech Burakowskic, Krzysztof Chudzikd, Bartosz Gajdab, Mariusz Gajewskia, Adam Grzechd, Piotr Krawiecc, Jan Kwiatkowskid, Tomasz Mrugalskie, Krzysztof Nowickie, Wiktor Procykb, Konrad Sienkiewicza, Robert Szumanb, Jarosław Śliwińskic, Jacek Światowiake, Piotr Wiśniewskic, Józef Woźniake a National Institute of Telecommunications, Warsaw, Poland b Poznań Supercomputing and Networking Center, Poznań, Poland c Institute of Telecommunications, Warsaw University of Technology, Warsaw, Poland d Institute of Informatics, Wrocław University of Technology, Wrocław, Poland e Faculty of Electronics, Telecommunications and Informatics, Gdańsk Univeristy of Technology, Gdańsk, Poland Abstract—Replacing the IPv4 protocol with IPv6 on the Inter- pool will be exhausted before 2013. Methods: polynomial net is currently one of the aims of the European Union policy. and exponential estimate the depletion of addresses during The main reason for this replacement is the effeteness of the the year 2012. Regardless of the method of research, the addresses pool in the IPv4 protocol, which can cause serious prospect of exhaustion of IPv4 addresses is so close that complications in the evolution of the Internet and its adapta- the urgent implementation of IPv6 has become a necessity. tion in new areas, e.g., in next generation mobile telephony or the so called Internet of Things. Simultaneously, the address- ing capabilities of the IPv6 protocol are practically unlimited and its new functionalities increase the attractiveness of its usage. The article discusses the problems connected with the IPv6 deployment on the Internet. Especially, the rules for re- alization of the IPv6 deployment and rules for cooperation of IPv4 with IPv6 (including cooperation tests) in network in- frastructure and in applications are presented.
    [Show full text]
  • Concept and Implementation of CLUSTERIX: National Cluster of Linux Systems
    Concept and Implementation of CLUSTERIX: National Cluster of Linux Systems Roman Wyrzykowski1, Norbert Meyer2, and Maciej Stroinski2 1 Czestochowa University of Technology Institute of Computer & Information Sciences Dabrowskiego 73, 42-200 Czestochowa, Poland [email protected] http://icis.pcz.pl 2 Poznan Supercomputing and Networking Center Noskowskiego 10, 61-704 Poznan, Poland {meyer, stroins}@man.poznan.pl http://www.man.poznan.pl Abstract. This paper presents the concept and implementation of the National Cluster of Linux Systems (CLUSTERIX) - a distributed PC- cluster (or metacluster) of a new generation, based on the Polish Optical Network PIONIER. Its implementation makes it possible to deploy a production Grid environment, which consists of local PC- clusters with 64- and 32-bit Linux machines, located in independent centers across Poland. The management software developed as Open Source allows for dynamic changes in the metacluster configuration. The resulting system will be tested on a set of pilot distributed applications developed as a part of the project. The project is implemented by 12 Polish supercomputing centers and metropolitan area networks. 1 Introduction PC-clusters using Open Source software such as Linux are the most common and available parallel systems now. At the same time, the capability of Gigabit/s wide area networks are increasing rapidly, to the point when it becomes feasible and indeed interesting to think of the high-end integrated metacluster environment rather than a set of disjoint local clusters. Such metaclusters [3,17,18] can be viewed as key elements of the modern Grid infrastructure, and used by scientists and engineers to solve computationally and data demanding problems.
    [Show full text]