Porting the DFT Code CASTEP to Gpgpus
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
GPAW, Gpus, and LUMI
GPAW, GPUs, and LUMI Martti Louhivuori, CSC - IT Center for Science Jussi Enkovaara GPAW 2021: Users and Developers Meeting, 2021-06-01 Outline LUMI supercomputer Brief history of GPAW with GPUs GPUs and DFT Current status Roadmap LUMI - EuroHPC system of the North Pre-exascale system with AMD CPUs and GPUs ~ 550 Pflop/s performance Half of the resources dedicated to consortium members Programming for LUMI Finland, Belgium, Czechia, MPI between nodes / GPUs Denmark, Estonia, Iceland, HIP and OpenMP for GPUs Norway, Poland, Sweden, and how to use Python with AMD Switzerland GPUs? https://www.lumi-supercomputer.eu GPAW and GPUs: history (1/2) Early proof-of-concept implementation for NVIDIA GPUs in 2012 ground state DFT and real-time TD-DFT with finite-difference basis separate version for RPA with plane-waves Hakala et al. in "Electronic Structure Calculations on Graphics Processing Units", Wiley (2016), https://doi.org/10.1002/9781118670712 PyCUDA, cuBLAS, cuFFT, custom CUDA kernels Promising performance with factor of 4-8 speedup in best cases (CPU node vs. GPU node) GPAW and GPUs: history (2/2) Code base diverged from the main branch quite a bit proof-of-concept implementation had lots of quick and dirty hacks fixes and features were pulled from other branches and patches no proper unit tests for GPU functionality active development stopped soon after publications Before development re-started, code didn't even work anymore on modern GPUs without applying a few small patches Lesson learned: try to always get new functionality to the -
Free and Open Source Software for Computational Chemistry Education
Free and Open Source Software for Computational Chemistry Education Susi Lehtola∗,y and Antti J. Karttunenz yMolecular Sciences Software Institute, Blacksburg, Virginia 24061, United States zDepartment of Chemistry and Materials Science, Aalto University, Espoo, Finland E-mail: [email protected].fi Abstract Long in the making, computational chemistry for the masses [J. Chem. Educ. 1996, 73, 104] is finally here. We point out the existence of a variety of free and open source software (FOSS) packages for computational chemistry that offer a wide range of functionality all the way from approximate semiempirical calculations with tight- binding density functional theory to sophisticated ab initio wave function methods such as coupled-cluster theory, both for molecular and for solid-state systems. By their very definition, FOSS packages allow usage for whatever purpose by anyone, meaning they can also be used in industrial applications without limitation. Also, FOSS software has no limitations to redistribution in source or binary form, allowing their easy distribution and installation by third parties. Many FOSS scientific software packages are available as part of popular Linux distributions, and other package managers such as pip and conda. Combined with the remarkable increase in the power of personal devices—which rival that of the fastest supercomputers in the world of the 1990s—a decentralized model for teaching computational chemistry is now possible, enabling students to perform reasonable modeling on their own computing devices, in the bring your own device 1 (BYOD) scheme. In addition to the programs’ use for various applications, open access to the programs’ source code also enables comprehensive teaching strategies, as actual algorithms’ implementations can be used in teaching. -
D6.1 Report on the Deployment of the Max Demonstrators and Feedback to WP1-5
Ref. Ares(2020)2820381 - 31/05/2020 HORIZON2020 European Centre of Excellence Deliverable D6.1 Report on the deployment of the MaX Demonstrators and feedback to WP1-5 D6.1 Report on the deployment of the MaX Demonstrators and feedback to WP1-5 Pablo Ordejón, Uliana Alekseeva, Stefano Baroni, Riccardo Bertossa, Miki Bonacci, Pietro Bonfà, Claudia Cardoso, Carlo Cavazzoni, Vladimir Dikan, Stefano de Gironcoli, Andrea Ferretti, Alberto García, Luigi Genovese, Federico Grasselli, Anton Kozhevnikov, Deborah Prezzi, Davide Sangalli, Joost VandeVondele, Daniele Varsano, Daniel Wortmann Due date of deliverable: 31/05/2020 Actual submission date: 31/05/2020 Final version: 31/05/2020 Lead beneficiary: ICN2 (participant number 3) Dissemination level: PU - Public www.max-centre.eu 1 HORIZON2020 European Centre of Excellence Deliverable D6.1 Report on the deployment of the MaX Demonstrators and feedback to WP1-5 Document information Project acronym: MaX Project full title: Materials Design at the Exascale Research Action Project type: European Centre of Excellence in materials modelling, simulations and design EC Grant agreement no.: 824143 Project starting / end date: 01/12/2018 (month 1) / 30/11/2021 (month 36) Website: www.max-centre.eu Deliverable No.: D6.1 Authors: P. Ordejón, U. Alekseeva, S. Baroni, R. Bertossa, M. Bonacci, P. Bonfà, C. Cardoso, C. Cavazzoni, V. Dikan, S. de Gironcoli, A. Ferretti, A. García, L. Genovese, F. Grasselli, A. Kozhevnikov, D. Prezzi, D. Sangalli, J. VandeVondele, D. Varsano, D. Wortmann To be cited as: Ordejón, et al., (2020): Report on the deployment of the MaX Demonstrators and feedback to WP1-5. Deliverable D6.1 of the H2020 project MaX (final version as of 31/05/2020). -
Introducing ONETEP: Linear-Scaling Density Functional Simulations on Parallel Computers Chris-Kriton Skylaris,A) Peter D
THE JOURNAL OF CHEMICAL PHYSICS 122, 084119 ͑2005͒ Introducing ONETEP: Linear-scaling density functional simulations on parallel computers Chris-Kriton Skylaris,a) Peter D. Haynes, Arash A. Mostofi, and Mike C. Payne Theory of Condensed Matter, Cavendish Laboratory, Madingley Road, Cambridge CB3 0HE, United Kingdom ͑Received 29 September 2004; accepted 4 November 2004; published online 23 February 2005͒ We present ONETEP ͑order-N electronic total energy package͒, a density functional program for parallel computers whose computational cost scales linearly with the number of atoms and the number of processors. ONETEP is based on our reformulation of the plane wave pseudopotential method which exploits the electronic localization that is inherent in systems with a nonvanishing band gap. We summarize the theoretical developments that enable the direct optimization of strictly localized quantities expressed in terms of a delocalized plane wave basis. These same localized quantities lead us to a physical way of dividing the computational effort among many processors to allow calculations to be performed efficiently on parallel supercomputers. We show with examples that ONETEP achieves excellent speedups with increasing numbers of processors and confirm that the time taken by ONETEP as a function of increasing number of atoms for a given number of processors is indeed linear. What distinguishes our approach is that the localization is achieved in a controlled and mathematically consistent manner so that ONETEP obtains the same accuracy as conventional cubic-scaling plane wave approaches and offers fast and stable convergence. We expect that calculations with ONETEP have the potential to provide quantitative theoretical predictions for problems involving thousands of atoms such as those often encountered in nanoscience and biophysics. -
Supporting Information
Electronic Supplementary Material (ESI) for RSC Advances. This journal is © The Royal Society of Chemistry 2020 Supporting Information How to Select Ionic Liquids as Extracting Agent Systematically? Special Case Study for Extractive Denitrification Process Shurong Gaoa,b,c,*, Jiaxin Jina,b, Masroor Abroc, Ruozhen Songc, Miao Hed, Xiaochun Chenc,* a State Key Laboratory of Alternate Electrical Power System with Renewable Energy Sources, North China Electric Power University, Beijing, 102206, China b Research Center of Engineering Thermophysics, North China Electric Power University, Beijing, 102206, China c Beijing Key Laboratory of Membrane Science and Technology & College of Chemical Engineering, Beijing University of Chemical Technology, Beijing 100029, PR China d Office of Laboratory Safety Administration, Beijing University of Technology, Beijing 100124, China * Corresponding author, Tel./Fax: +86-10-6443-3570, E-mail: [email protected], [email protected] 1 COSMO-RS Computation COSMOtherm allows for simple and efficient processing of large numbers of compounds, i.e., a database of molecular COSMO files; e.g. the COSMObase database. COSMObase is a database of molecular COSMO files available from COSMOlogic GmbH & Co KG. Currently COSMObase consists of over 2000 compounds including a large number of industrial solvents plus a wide variety of common organic compounds. All compounds in COSMObase are indexed by their Chemical Abstracts / Registry Number (CAS/RN), by a trivial name and additionally by their sum formula and molecular weight, allowing a simple identification of the compounds. We obtained the anions and cations of different ILs and the molecular structure of typical N-compounds directly from the COSMObase database in this manuscript. -
GROMACS: Fast, Flexible, and Free
GROMACS: Fast, Flexible, and Free DAVID VAN DER SPOEL,1 ERIK LINDAHL,2 BERK HESS,3 GERRIT GROENHOF,4 ALAN E. MARK,4 HERMAN J. C. BERENDSEN4 1Department of Cell and Molecular Biology, Uppsala University, Husargatan 3, Box 596, S-75124 Uppsala, Sweden 2Stockholm Bioinformatics Center, SCFAB, Stockholm University, SE-10691 Stockholm, Sweden 3Max-Planck Institut fu¨r Polymerforschung, Ackermannweg 10, D-55128 Mainz, Germany 4Groningen Biomolecular Sciences and Biotechnology Institute, University of Groningen, Nijenborgh 4, NL-9747 AG Groningen, The Netherlands Received 12 February 2005; Accepted 18 March 2005 DOI 10.1002/jcc.20291 Published online in Wiley InterScience (www.interscience.wiley.com). Abstract: This article describes the software suite GROMACS (Groningen MAchine for Chemical Simulation) that was developed at the University of Groningen, The Netherlands, in the early 1990s. The software, written in ANSI C, originates from a parallel hardware project, and is well suited for parallelization on processor clusters. By careful optimization of neighbor searching and of inner loop performance, GROMACS is a very fast program for molecular dynamics simulation. It does not have a force field of its own, but is compatible with GROMOS, OPLS, AMBER, and ENCAD force fields. In addition, it can handle polarizable shell models and flexible constraints. The program is versatile, as force routines can be added by the user, tabulated functions can be specified, and analyses can be easily customized. Nonequilibrium dynamics and free energy determinations are incorporated. Interfaces with popular quantum-chemical packages (MOPAC, GAMES-UK, GAUSSIAN) are provided to perform mixed MM/QM simula- tions. The package includes about 100 utility and analysis programs. -
An Ab Initio Materials Simulation Code
CP2K: An ab initio materials simulation code Lianheng Tong Physics on Boat Tutorial , Helsinki, Finland 2015-06-08 Faculty of Natural & Mathematical Sciences Department of Physics Brief Overview • Overview of CP2K - General information about the package - QUICKSTEP: DFT engine • Practical example of using CP2K to generate simulated STM images - Terminal states in AGNR segments with 1H and 2H termination groups What is CP2K? Swiss army knife of molecular simulation www.cp2k.org • Geometry and cell optimisation Energy and • Molecular dynamics (NVE, Force Engine NVT, NPT, Langevin) • STM Images • Sampling energy surfaces (metadynamics) • DFT (LDA, GGA, vdW, • Finding transition states Hybrid) (Nudged Elastic Band) • Quantum Chemistry (MP2, • Path integral molecular RPA) dynamics • Semi-Empirical (DFTB) • Monte Carlo • Classical Force Fields (FIST) • And many more… • Combinations (QM/MM) Development • Freely available, open source, GNU Public License • www.cp2k.org • FORTRAN 95, > 1,000,000 lines of code, very active development (daily commits) • Currently being developed and maintained by community of developers: • Switzerland: Paul Scherrer Institute Switzerland (PSI), Swiss Federal Institute of Technology in Zurich (ETHZ), Universität Zürich (UZH) • USA: IBM Research, Lawrence Livermore National Laboratory (LLNL), Pacific Northwest National Laboratory (PNL) • UK: Edinburgh Parallel Computing Centre (EPCC), King’s College London (KCL), University College London (UCL) • Germany: Ruhr-University Bochum • Others: We welcome contributions from -
Automated Construction of Quantum–Classical Hybrid Models Arxiv:2102.09355V1 [Physics.Chem-Ph] 18 Feb 2021
Automated construction of quantum{classical hybrid models Christoph Brunken and Markus Reiher∗ ETH Z¨urich, Laboratorium f¨urPhysikalische Chemie, Vladimir-Prelog-Weg 2, 8093 Z¨urich, Switzerland February 18, 2021 Abstract We present a protocol for the fully automated construction of quantum mechanical-(QM){ classical hybrid models by extending our previously reported approach on self-parametri- zing system-focused atomistic models (SFAM) [J. Chem. Theory Comput. 2020, 16 (3), 1646{1665]. In this QM/SFAM approach, the size and composition of the QM region is evaluated in an automated manner based on first principles so that the hybrid model describes the atomic forces in the center of the QM region accurately. This entails the au- tomated construction and evaluation of differently sized QM regions with a bearable com- putational overhead that needs to be paid for automated validation procedures. Applying SFAM for the classical part of the model eliminates any dependence on pre-existing pa- rameters due to its system-focused quantum mechanically derived parametrization. Hence, QM/SFAM is capable of delivering a high fidelity and complete automation. Furthermore, since SFAM parameters are generated for the whole system, our ansatz allows for a con- venient re-definition of the QM region during a molecular exploration. For this purpose, a local re-parametrization scheme is introduced, which efficiently generates additional clas- sical parameters on the fly when new covalent bonds are formed (or broken) and moved to the classical region. arXiv:2102.09355v1 [physics.chem-ph] 18 Feb 2021 ∗Corresponding author; e-mail: [email protected] 1 1 Introduction In contrast to most protocols of computational quantum chemistry that consider isolated molecules, chemical processes can take place in a vast variety of complex environments. -
Chem Compute Quickstart
Chem Compute Quickstart Chem Compute is maintained by Mark Perri at Sonoma State University and hosted on Jetstream at Indiana University. The Chem Compute URL is https://chemcompute.org/. We will use Chem Compute as the frontend for running electronic structure calculations with The General Atomic and Molecular Electronic Structure System, GAMESS (http://www.msg.ameslab.gov/gamess/). Chem Compute also provides access to other computational chemistry resources including PSI4 and the molecular dynamics packages TINKER and NAMD, though we will not be using those resource at this time. Follow this link, https://chemcompute.org/gamess/submit, to directly access the Chem Compute GAMESS guided submission interface. If you are a returning Chem Computer user, please log in now. If your University is part of the InCommon Federation you can log in without registering by clicking Login then "Log In with Google or your University" – select your University from the dropdown list. Otherwise if this is your first time working with Chem Compute, please register as a Chem Compute user by clicking the “Register” link in the top-right corner of the page. This gives you access to all of the computational resources available at ChemCompute.org and will allow you to maintain copies of your calculations in your user “Dashboard” that you can refer to later. Registering also helps track usage and obtain the resources needed to continue providing its service. When logged in with the GAMESS-Submit tabs selected, an instruction section appears on the left side of the page with instructions for several different kinds of calculations. -
Popular GPU-Accelerated Applications
LIFE & MATERIALS SCIENCES GPU-ACCELERATED APPLICATIONS | CATALOG | AUG 12 LIFE & MATERIALS SCIENCES APPLICATIONS CATALOG Application Description Supported Features Expected Multi-GPU Release Status Speed Up* Support Bioinformatics BarraCUDA Sequence mapping software Alignment of short sequencing 6-10x Yes Available now reads Version 0.6.2 CUDASW++ Open source software for Smith-Waterman Parallel search of Smith- 10-50x Yes Available now protein database searches on GPUs Waterman database Version 2.0.8 CUSHAW Parallelized short read aligner Parallel, accurate long read 10x Yes Available now aligner - gapped alignments to Version 1.0.40 large genomes CATALOG GPU-BLAST Local search with fast k-tuple heuristic Protein alignment according to 3-4x Single Only Available now blastp, multi cpu threads Version 2.2.26 GPU-HMMER Parallelized local and global search with Parallel local and global search 60-100x Yes Available now profile Hidden Markov models of Hidden Markov Models Version 2.3.2 mCUDA-MEME Ultrafast scalable motif discovery algorithm Scalable motif discovery 4-10x Yes Available now based on MEME algorithm based on MEME Version 3.0.12 MUMmerGPU A high-throughput DNA sequence alignment Aligns multiple query sequences 3-10x Yes Available now LIFE & MATERIALS& LIFE SCIENCES APPLICATIONS program against reference sequence in Version 2 parallel SeqNFind A GPU Accelerated Sequence Analysis Toolset HW & SW for reference 400x Yes Available now assembly, blast, SW, HMM, de novo assembly UGENE Opensource Smith-Waterman for SSE/CUDA, Fast short -
Natural Bond Orbital Analysis in the ONETEP Code: Applications to Large Protein Systems Louis P
WWW.C-CHEM.ORG FULL PAPER Natural Bond Orbital Analysis in the ONETEP Code: Applications to Large Protein Systems Louis P. Lee,*[a] Daniel J. Cole,[a] Mike C. Payne,[a] and Chris-Kriton Skylaris[b] First principles electronic structure calculations are typically Generalized Wannier Functions of ONETEP to natural atomic performed in terms of molecular orbitals (or bands), providing a orbitals, NBO analysis can be performed within a localized straightforward theoretical avenue for approximations of region in such a way that ensures the results are identical to an increasing sophistication, but do not usually provide any analysis on the full system. We demonstrate the capabilities of qualitative chemical information about the system. We can this approach by performing illustrative studies of large derive such information via post-processing using natural bond proteins—namely, investigating changes in charge transfer orbital (NBO) analysis, which produces a chemical picture of between the heme group of myoglobin and its ligands with bonding in terms of localized Lewis-type bond and lone pair increasing system size and between a protein and its explicit orbitals that we can use to understand molecular structure and solvent, estimating the contribution of electronic delocalization interactions. We present NBO analysis of large-scale calculations to the stabilization of hydrogen bonds in the binding pocket of with the ONETEP linear-scaling density functional theory package, a drug-receptor complex, and observing, in situ, the n ! p* which we have interfaced with the NBO 5 analysis program. In hyperconjugative interactions between carbonyl groups that ONETEP calculations involving thousands of atoms, one is typically stabilize protein backbones. -
Starting SCF Calculations by Superposition of Atomic Densities
Starting SCF Calculations by Superposition of Atomic Densities J. H. VAN LENTHE,1 R. ZWAANS,1 H. J. J. VAN DAM,2 M. F. GUEST2 1Theoretical Chemistry Group (Associated with the Department of Organic Chemistry and Catalysis), Debye Institute, Utrecht University, Padualaan 8, 3584 CH Utrecht, The Netherlands 2CCLRC Daresbury Laboratory, Daresbury WA4 4AD, United Kingdom Received 5 July 2005; Accepted 20 December 2005 DOI 10.1002/jcc.20393 Published online in Wiley InterScience (www.interscience.wiley.com). Abstract: We describe the procedure to start an SCF calculation of the general type from a sum of atomic electron densities, as implemented in GAMESS-UK. Although the procedure is well known for closed-shell calculations and was already suggested when the Direct SCF procedure was proposed, the general procedure is less obvious. For instance, there is no need to converge the corresponding closed-shell Hartree–Fock calculation when dealing with an open-shell species. We describe the various choices and illustrate them with test calculations, showing that the procedure is easier, and on average better, than starting from a converged minimal basis calculation and much better than using a bare nucleus Hamiltonian. © 2006 Wiley Periodicals, Inc. J Comput Chem 27: 926–932, 2006 Key words: SCF calculations; atomic densities Introduction hrstuhl fur Theoretische Chemie, University of Kahrlsruhe, Tur- bomole; http://www.chem-bio.uni-karlsruhe.de/TheoChem/turbo- Any quantum chemical calculation requires properly defined one- mole/),12 GAMESS(US) (Gordon Research Group, GAMESS, electron orbitals. These orbitals are in general determined through http://www.msg.ameslab.gov/GAMESS/GAMESS.html, 2005),13 an iterative Hartree–Fock (HF) or Density Functional (DFT) pro- Spartan (Wavefunction Inc., SPARTAN: http://www.wavefun.