A Cpu-Gpu Framework for Astronomical Data Reduction and Analysis

Total Page:16

File Type:pdf, Size:1020Kb

A Cpu-Gpu Framework for Astronomical Data Reduction and Analysis UNIVERSIDAD DE CHILE FACULTAD DE CIENCIAS F´ISICAS Y MATEMATICAS´ DEPARTAMENTO DE CIENCIAS DE LA COMPUTACION´ FADRA: A CPU-GPU FRAMEWORK FOR ASTRONOMICAL DATA REDUCTION AND ANALYSIS TESIS PARA OPTAR AL GRADO DE MAG´ISTER EN CIENCIAS, MENCION´ COMPUTACION´ FRANCISCA ANDREA CONCHA RAM´IREZ PROFESOR GU´IA: MAR´IA CECILIA RIVARA ZU´NIGA~ PROFESOR CO-GU´IA: PATRICIO ROJO RUBKE MIEMBROS DE LA COMISION:´ ALEXANDRE BERGEL JOHAN FABRY GONZALO ACUNA~ LEIVA Este trabajo ha sido parcialmente financiado por Proyecto FONDECYT 1120299 SANTIAGO DE CHILE 2016 Resumen Esta tesis establece las bases de FADRA: Framework for Astronomical Data Reduction and Analysis. El framework FADRA fue dise~nadopara ser eficiente, simple de usar, modular, expandible, y open source. Hoy en d´ıa,la astronom´ıaes inseparable de la computaci´on,pero algunos de los software m´asusados en la actualidad fueron desarrollados tres d´ecadasatr´asy no est´andise~nadospara enfrentar los actuales paradigmas de big data. El mundo del software astron´omicodebe evolucionar no solo hacia pr´acticasque comprendan y adopten la era del big data, sino tambi´enque est´enenfocadas en el trabajo colaborativo de la comunidad. El trabajo desarollado consisti´oen el dise~no e implementaci´onde los algoritmos b´asicos para el an´alisisde datos astron´omicos, dando inicio al desarrollo del framework. Esto con- sider´ola implementaci´onde estructuras de datos eficientes al trabajar con un gran n´umero de im´agenes,la implementaci´onde algoritmos para el proceso de calibraci´ono reducci´onde im´agenesastron´omicas,y el dise~noy desarrollo de algoritmos para el c´alculode fotometr´ıay la obtenci´onde curvas de luz. Tanto los algoritmos de reducci´oncomo de obtenci´onde curvas de luz fueron implementados en versiones CPU y GPU. Para las implementaciones en GPU, se dise~naronalgoritmos que minimizan la cantidad de datos a ser procesados de manera de reducir la transferencia de datos entre CPU y GPU, proceso lento que muchas veces eclipsa las ganancias en tiempo de ejecuci´onque se pueden obtener gracias a la paralelizaci´on. A pesar de que FADRA fue dise~nadocon la idea de utilizar sus algoritmos dentro de scripts, un m´odulo wrapper para interactuar a trav´esde interfaces gr´aficastambi´enfue implementado. Una de las principales metas de esta tesis consisti´oen la validaci´onde los resultados obtenidos con FADRA. Para esto, resultados de la reducci´ony curvas de luz fueron compara- dos con resultados de AstroPy, paquete de Python con distintas utilidades para astr´onomos. Los experimentos se realizaron sobre seis datasets de im´agenesastron´omicasreales. En el caso de reducci´onde im´agenesastron´omicas,el Normalized Root Mean Squared Error (NRMSE) fue utilizado como m´etricade similaridad entre las im´agenes. Para las curvas de luz, se prob´o que las formas de las curvas eran iguales a trav´esde la determinaci´onde offsets constantes entre los valores num´ericosde cada uno de los puntos pertenecientes a las distintas curvas. En t´erminosde la validez de los resultados, tanto la reducci´oncomo la obtenci´onde curvas de luz, en sus implementaciones CPU y GPU, generaron resultados correctos al ser comparados con los de AstroPy, lo que significa que los desarrollos y aproximaciones dise~nados para FADRA otorgan resultados que pueden ser utilizados con seguridad para el an´alisis cient´ıficode im´agenesastron´omicas. En t´erminosde tiempos de ejecuci´on,la naturaleza intensiva en uso de datos propia del proceso de reducci´onhace que la versi´onGPU sea incluso m´aslenta que la versi´onCPU. Sin embargo, en el caso de la obtenci´onde curvas de luz, el algoritmo GPU presenta una disminuci´onimportante en tiempo de ejecuci´oncomparado con su contraparte en CPU. i Abstract This thesis sets the bases for FADRA: Framework for Astronomical Data Reduction and Analysis. The FADRA framework is designed to be efficient and easy to use, modular, expandable, and open source. Nowadays, astronomy is inseparable from computer science, but some of the software still widely used today was developed three decades ago and is not up to date with the current data paradigms. The world of astronomical software development must start evolving not only towards practices that comprehend and embrace the big data era, but also that lead to collaborative work in the community. The work carried out in this thesis consisted in the design and implementation of basic algorithms for astronomical data analysis, to set the beginning of the FADRA framework. This encompassed the implementation of data structures that are efficient when working with a large number of astronomical images, the implementation of algorithms for astronomical data calibration or reduction, and the design and development of automated photometry and light curve obtention algorithms. Both the reduction and the light curve obtention algorithms were implemented on CPU and GPU versions. For the GPU implementations, the algorithms were designed considering the minimization of the amount of data to be processed, as a means to reduce the data transfer between CPU and GPU, a slow process which in many cases can even overshadow the gains in execution time obtatined through parallelization. Even though the main idea is for the FADRA algorithms to be run within scripts, a wrapper module to run Graphical User Interfaces (GUIs) for the code was also implemented. One of the most important steps of this thesis was validating the correctness of the results obtained with FADRA algorithms. For this, the results from the reduction and the light curve obtention processes were compared against results obtained using AstroPy, a Python package with different utilities for astronomers. The experiments were carried out over six datasets of real astronomical images. For the case of astronomical data reduction, the Normalized Root Mean Squared Error (NRMSE) was calculated between the images to measure their similarity. In the case of light curves, the shapes of the curves were proved to be equal by finding constant offsets between the numerical values for each data point belonging to a curve. In terms of correctness of results, both the reduction and light curve obtention algorithms, in their CPU and GPU implementations, proved to be correct when compared to AstroPy's results, meaning that the implementations and approximations designed for the FADRA framework provide correct results that can be confidently used in scientific analysis of as- tronomical images. Regarding execution times, the intensive data aspect of the reduction algorithm makes the GPU implementation even slower than the CPU implementation. How- ever, for the case of light curve obtention, the GPU algorithm presents an important speedup compared to its CPU counterpart. ii Acknowledgements First I would like to thank my family for always supporting me and helping me follow my dreams. This work and everything else I've accomplished so far would not have been possible without their love and encouragement. I would also like to thank Fernando Caro for his support and company, not only through the development of this thesis but in life. I would like to thank my friends for being the best company I could ask for, and for putting up with my long disappearances because \I have to work on my thesis tonight", many nights. Thank you for being so patient and for always cheering for me and supporting me. I also want to thank Professor Patricio Rojo for all these many years of friendly work and advice. This thesis would not have happened if it wasn't for him and his insistence on making better astronomical software. I would also like to thank Professor Maria Cecilia Rivara for her great support through my years as a student and all through this thesis, which I probably wouldn't have finished already if it wasn't for her relevant advice and comments. Both of my advisors were a fundamental part of my student years and of this work and I would not have made it this far if it wasn't for them. Finally I would like to express my thanks to the members of the revision committee, Professors Alexandre Bergel, Johan Fabry, and Gonzalo Acu~na,for their careful reviews of my thesis and for their relevant comments to improve it. Last but definitely not least I want to thank Ren Cerro for kindly taking the time to proof-read this text. iii Contents List of Tables vi List of Figures vii 1 Introduction 1 1.1 Astronomical data analysis . .1 1.2 Astronomical software development . .2 1.3 Thesis description . .3 1.3.1 Goals and objectives . .3 1.3.2 Research questions . .4 1.3.3 Software architecture . .4 1.3.4 Use of previous work . .5 1.3.5 Programming languages . .5 1.3.6 Validation of results . .5 2 Literature revision 8 2.1 Existing software . .9 2.2 Criticism of existing solutions . 15 3 Astronomical data and analysis 17 3.1 Astronomical data . 17 3.1.1 Astronomical images . 17 3.1.2 Astronomical spectra . 18 3.2 Astronomical image acquisition . 19 3.3 Astronomical image reduction . 21 3.4 Astronomical image processing . 22 3.4.1 Image arithmetic and combining . 23 3.4.2 Filter application . 24 3.4.3 Photometry . 25 3.4.4 Light curve or time series generation . 29 4 Introduction to General-Purpose Graphics Processing Unit (GPGPU) computing 30 4.1 What is the Graphics Processing Unit (GPU)? . 30 4.2 General-Purpose GPU computing (GPGPU) . 33 4.3 GPGPU use in astronomy . 34 4.3.1 GPGPU use for astronomical data analysis in this thesis . 35 iv 5 Software design and implementation 36 5.1 Data handling: AstroFile and AstroDir classes . 36 5.2 Calibration image combination and obtention of Master fields .
Recommended publications
  • CFITSIO User's Reference Guide
    CFITSIO User’s Reference Guide An Interface to FITS Format Files for C Programmers Version 3.3 HEASARC Code 662 Goddard Space Flight Center Greenbelt, MD 20771 USA February 2013 ii Contents 1 Introduction 1 1.1 ABriefOverview .................................. ... 1 1.2 Sources of FITS Software and Information . ........... 1 1.3 Acknowledgments................................. ..... 2 1.4 LegalStuff ....................................... 4 2 Creating the CFITSIO Library 5 2.1 BuildingtheLibrary .............................. ...... 5 2.1.1 UnixSystems................................... 5 2.1.2 VMS......................................... 7 2.1.3 WindowsPCs.................................... 7 2.1.4 MacintoshPCs .................................. 7 2.2 TestingtheLibrary............................... ...... 8 2.3 LinkingProgramswithCFITSIO . ....... 9 2.4 Using CFITSIO in Multi-threaded Environments . ............ 9 2.5 GettingStartedwithCFITSIO . ....... 10 2.6 ExampleProgram .................................. 10 3 A FITS Primer 13 4 Programming Guidelines 15 4.1 CFITSIODefinitions............................... ..... 15 4.2 Current Header Data Unit (CHDU) . ...... 17 4.3 Function Names and Variable Datatypes . .......... 18 4.4 Support for Unsigned Integers and Signed Bytes . ............ 20 4.5 Dealing with Character Strings . ......... 22 iii iv CONTENTS 4.6 ImplicitDataTypeConversion . ........ 23 4.7 DataScaling ..................................... 23 4.8 SupportforIEEESpecialValues . ........ 24 4.9 Error Status Values and the
    [Show full text]
  • The Starlink Build System SSN/78.1 —Abstract I
    SSN/78.1 Starlink Project Starlink System Note 78.1 Norman Gray, Peter W Draper, Mark B Taylor, Steven E Rankin 11 April 2005 Copyright 2004-5, Council for the Central Laboratory of the Research Councils Copyright 2007, Particle Physics and Astronomy Research Council Copyright 2007, Science and Technology Facilities Council The Starlink Build System SSN/78.1 —Abstract i Abstract This document provides an introduction to the Starlink build system. It describes how to use the Starlink versions of the GNU autotools (autoconf, automake and libtool), how to build the software set from a checkout, how to add and configure new components, and acts as a reference manual for the Starlink-specific autoconf macros and Starlink automake features. It does not describe the management of the CVS repository in detail, nor any other source maintainance patterns. It should be read in conjunction with the detailed build instructions in the README file at the top of the source tree (which takes precedence over any instructions in this document, though there should be no major disagreements), and with sun248, which additionally includes platform-specific notes. Copyright 2004-5, Council for the Central Laboratory of the Research Councils Copyright 2007, Particle Physics and Astronomy Research Council Copyright 2007, Science and Technology Facilities Council ii SSN/78.1—Contents Contents 1 Introduction 1 1.1 Quick entry-points . 2 2 Tools 3 2.1 Overview of the Autotools . 3 2.1.1 Autoconf . 5 2.1.2 Automake . 9 2.1.3 Libtool . 13 2.1.4 Autoreconf: why you don’t need to know about aclocal .
    [Show full text]
  • ORAC-DR: Overview and General Introduction 4.1-0 SUN/230.6 —Abstract Ii
    SUN/230.6 Starlink Project Starlink User Note 230.6 Frossie Economou, Tim Jenness, Malcolm Currie, Andy Adamson, Alasdair Allan, Brad Cavanagh Joint Astronomy Centre, Hilo, Hawaii June 2004 Copyright c 1997-2004 Particle Physics and Astronomy Research Council ORAC-DR: Overview and General Introduction 4.1-0 SUN/230.6 —Abstract ii Abstract ORAC-DR is a general purpose automatic data reduction pipeline environment. It currently supports data reduction for the United Kingdom Infrared Telescope (UKIRT) instruments UFTI, IRCAM, UIST and CGS4, for the James Clerk Maxwell Telescope (JCMT) instrument SCUBA, for the William Herschel Telescope (WHT) instrument INGRID, for the European Southern Observatory (ESO) instrument ISAAC and for the Anglo-Australian Telescope (AAT) instrument IRIS-2. This document describes the general pipeline environment. For specific information on how to reduce the data for a particular instrument, please consult the appropriate ORAC-DR instrument guide. Copyright c 1997-2004 Particle Physics and Astronomy Research Council iii SUN/230.6—Contents Contents 1 Introduction to ORAC-DR 1 2 ORAC-DR 2 3 Setting up to run oracdr 4 4 ORAC-DR Components 5 5 Xoracdr 6 6 oracdr 12 7 oracdr_monitor 18 8 ORAC-DR 20 9 Release Notes 21 A The ORAC-DR Data Loops 22 B The ORAC-DR Display System 23 C The ORAC-DR Calibration Selection 28 D Shell Variables 30 E oracdisp 32 F oracdr_nuke 33 1 SUN/230.6 —Introduction to ORAC-DR 1 Introduction to ORAC-DR An ORAC-DR HowTo. Description This document gives a general introduction to the pipeline, what it does and what it will not do.
    [Show full text]
  • ORAC-DR: a Generic Data Reduction Pipeline Infrastructure✩
    Astronomy and Computing 9 (2015) 40–48 Contents lists available at ScienceDirect Astronomy and Computing journal homepage: www.elsevier.com/locate/ascom Full length article ORAC-DR: A generic data reduction pipeline infrastructureI Tim Jenness ∗, Frossie Economou 1 Joint Astronomy Centre, 660 N. A`ohok¯ u¯ Place, Hilo, HI 96720, USA article info a b s t r a c t Article history: ORAC-DR is a general purpose data reduction pipeline system designed to be instrument and observatory Received 5 September 2014 agnostic. The pipeline works with instruments as varied as infrared integral field units, imaging arrays Accepted 27 October 2014 and spectrographs, and sub-millimeter heterodyne arrays and continuum cameras. This paper describes Available online 1 November 2014 the architecture of the pipeline system and the implementation of the core infrastructure. We finish by discussing the lessons learned since the initial deployment of the pipeline system in the late 1990s. Keywords: ' 2014 The Authors. Published by Elsevier B.V. Data reduction pipelines This is an open access article under the CC BY license Techniques: miscellaneous Methods: data analysis (http://creativecommons.org/licenses/by/3.0/). 1. Introduction The Observatory Reduction and Acquisition Control Data Re- duction pipeline (orac-dr; Cavanagh et al., 2008; Economou et al., In the early 1990s each instrument delivered to the United King- 1999; ascl:1310.001) was the resulting system. In the sections that dom Infrared Telescope (UKIRT) and the James Clerk Maxwell Tele- follow we present an overview of the architectural design and scope (JCMT) came with its own distinct data reduction system that then describe the pipeline implementation.
    [Show full text]
  • The Significant Properties of Software: a Study
    The Significant Properties of Software: A Study Brian Matthews, Brian McIlwrath, David Giaretta, Esther Conway STFC Rutherford Appleton Laboratory Chilton OX11 0QX UK December 2008 Significant Properties of Software Revision History: Version Date Authors Sections Affected / Comments 0.1 18/02/2008 BMM Outline 0.5 05/03/2008 BMM First Draft 0.6 09/03/2008 DG DG added section on OAIS/CASPAR 0.7 11/03/2008 BMM Added section on StarLink + revisions. 0.8 23/03/2008 BMcI, EC, BMM Expanded use cases 1.0 28/03/2008 BMM, BMcI First Complete release version 1.1 23/12/2008 BMM Final Revision 2 Significant Properties of Software Executive Summary ...................................................................................................... 5 Recommendations ..................................................................................................................... 6 1 Background to the Study ....................................................................................... 9 1.1 Introduction ................................................................................................................... 9 1.2 Significant Properties .................................................................................................. 10 2 Scope of Study .................................................................................................... 12 2.1 Definition of Software ................................................................................................ 12 2.2 Diversity of Software .................................................................................................
    [Show full text]
  • Book of Abstracts 2009 European Week of Astronomy and Space
    rs uvvwxyuzyws { yz|z|} rsz}~suzywsu}u~w vz~wsw 456789@A C 99D 7EFGH67A7I P @AQ R8@S9 RST9AS9 UVWUX `abcdUVVe fATg96GTHP7Eh96HE76QGiT69pf q rAS76876@HTAs tFR u Fv wxxy @AQ 4FR 4u Fv wxxy UVVe abbc d dbdc e f gc hi` ij ad bch dgcadabdddc c d ac k lgbc bcgb dmg agd g` kg bdcd dW dd k bg c ngddbaadgc gabmob nb boglWad g kdcoddog kedgcW pd gc bcogbpd kb obpcggc dd kfq` UVVe c iba ! " #$%& $' ())01023 Book of Abstracts – Table of Contents Welcome to the European Week of Astronomy & Space Science ...................................................... iii How space, and a few stars, came to Hatfield ............................................................................... v Plenary I: UK Solar Physics (UKSP) and Magnetosphere, Ionosphere and Solar Terrestrial (MIST) ....... 1 Plenary II: European Organisation for Astronomical Research in the Southern Hemisphere (ESO) ....... 2 Plenary III: European Space Agency (ESA) .................................................................................. 3 Plenary IV: Square Kilometre Array (SKA), High-Energy Astrophysics, Asteroseismology ................... 4 Symposia (1) The next era in radio astronomy: the pathway to SKA .............................................................. 5 (2) The standard cosmological models - successes and challenges .................................................. 17 (3) Understanding substellar populations and atmospheres: from brown dwarfs to exo-planets .......... 28 (4) The life cycle of dust ...........................................................................................................
    [Show full text]
  • SLALIB — Positional Astronomy Library Programmer's Manual
    CCLRC / Rutherford Appleton Laboratory SUN/67.63 Particle Physics & Astronomy Research Council Starlink Project Starlink User Note 67.63 P. T. Wallace 21 October 2003 SLALIB — Positional Astronomy Library 2.4-13 Programmer’s Manual Abstract SLALIB is a library used by writers of positional-astronomy applications. Most of the 187 routines are concerned with astronomical position and time, but a number have wider trigono- metrical, numerical or general applications. ii Contents 1 INTRODUCTION 1 1.1 Purpose . 1 1.2 Example Application . 1 1.3 Scope . 2 1.4 Objectives . 3 1.5 Fortran Version . 4 1.6 CVersion ........................................ 4 1.7 Future Versions . 4 1.8 New Functions . 5 1.9 Acknowledgements . 5 2 LINKING 5 3 SUBPROGRAM SPECIFICATIONS 6 4 EXPLANATION AND EXAMPLES 169 4.1 Spherical Trigonometry . 169 4.2 Vectors and Matrices . 172 4.3 Celestial Coordinate Systems . 174 4.4 Precession and Nutation . 177 4.5 Mean Places . 179 4.6 Epoch . 179 4.7 Proper Motion . 180 4.8 Parallax and Radial Velocity . 180 4.9 Aberration . 181 4.10 Different Sorts of Mean Place . 182 4.11 Mean Place Transformations . 183 4.12 Mean Place to Apparent Place . 185 4.13 Apparent Place to Observed Place . 186 4.14 The Hipparcos Catalogue and the ICRS . 189 4.15 Timescales . 189 4.16 Calendars . 193 4.17 Geocentric Coordinates . 193 4.18 Ephemerides . 194 4.19 Radial Velocity and Light-Time Corrections . 203 4.20 Focal-Plane Astrometry . 204 4.21 Numerical Methods . 205 5 SUMMARY OF CALLS 208 SUN/67.63 1 1 INTRODUCTION 1.1 Purpose SLALIB1 is a library of routines intended to make accurate and reliable positional-astronomy applications easier to write.
    [Show full text]
  • Arxiv:1410.7513V1 [Astro-Ph.IM] 28 Oct 2014
    Learning from 25 years of the extensible N-Dimensional Data Format Tim Jennessa,∗, David S. Berryb, Malcolm J. Currieb, Peter W. Draperc, Frossie Economoud, Norman Graye, Brian McIlwrathf, Keith Shortridgeg, Mark B. Taylorh, Patrick T. Wallacef, Rodney F. Warren-Smithf aDepartment of Astronomy, Cornell University, Ithaca, NY 14853, USA bJoint Astronomy Centre, 660 N. A‘oh¯ok¯uPlace, Hilo, HI 96720, USA cDepartment of Physics, Institute for Computational Cosmology, University of Durham, South Road, Durham DH1 3LE, UK dLSST Project Office, 933 N. Cherry Ave, Tucson, AZ 85721, USA eSUPA School of Physics & Astronomy, University of Glasgow, Glasgow G12 8QQ, UK fRAL Space, STFC Rutherford Appleton Laboratory, Harwell Oxford, Didcot, Oxfordshire OX11 0QX, UK gAustralian Astronomical Observatory, 105 Delhi Rd, North Ryde, NSW 2113, Australia hH. H. Wills Physics Laboratory, Bristol University, Tyndall Avenue, Bristol, UK Abstract The extensible N-Dimensional Data Format (NDF) was designed and developed in the late 1980s to provide a data model suitable for use in a variety of astronomy data processing applications supported by the UK Starlink Project. Starlink applications were used extensively, primarily in the UK astronomical community, and form the basis of a number of advanced data reduction pipelines today. This paper provides an overview of the historical drivers for the development of NDF and the lessons learned from using a defined hierarchical data model for many years in data reduction software, data pipelines and in data acquisition systems. Keywords: data formats, data models, Starlink, History of computing 1. Introduction In this paper the term “data model” refers to the organization, naming and semantics of components in a hierarchy.
    [Show full text]
  • Starlinksoftware Collection
    SUN/1.24 Starlink Project Starlink User Note 1.24 ed. S. E. Rankin 5 August 2003 STARLINK Software Collection SUN/1.24 —Abstract ii Abstract The Starlink Software Collection is a set of software which is managed and distributed by the Starlink Project. Some of the software was written by members of the Project, but some of it comes from outside the Project. This note describes the functions of the individual items in the Collection and provides an overview of the software so that readers can identify the items they need. The software is classified into four main divisions: • Packages – are large collections of programs for people who want to analyse, convert, and display data. They are subdivided into eleven classes to help you find what you want. • Utilities – are small programs devoted to a specific purpose. For example, they help you prepare for observations, write documents, and write programs. • Subroutine Libraries – are for programmers writing astronomical software. They provide facilities such as astronomical calculations, data management and graphics. • Infrastructure – are items which are mainly of interest to people writing programs within the Starlink Software Environment. They are included for completeness. Each item is described in sufficient detail for you to decide whether or not to investigate it further. If you want to find out more about an item, follow the document references given opposite the item name. If you are using the hypertext version of this document, the most up-to-date document references can be found by following the link from the software item name. iii SUN/1.24—Contents Contents 1 Introduction 1 2 Changes since the last issue 2 2.1 New packages .
    [Show full text]
  • FITSIO User's Guide
    FITSIO User’s Guide A Subroutine Interface to FITS Format Files for Fortran Programmers Version 3.0 HEASARC Code 662 Goddard Space Flight Center Greenbelt, MD 20771 USA April 2009 ii Contents 1 Introduction 1 2 Creating FITSIO/CFITSIO 3 2.1 BuildingtheLibrary .............................. ...... 3 2.2 TestingtheLibrary............................... ...... 6 2.3 LinkingProgramswithFITSIO . ....... 7 2.4 GettingStartedwithFITSIO . ....... 8 2.5 ExampleProgram .................................. ... 8 2.6 LegalStuff ....................................... 9 2.7 Acknowledgments................................. ..... 10 3 A FITS Primer 13 4 FITSIO Conventions and Guidelines 15 4.1 CFITSIOSizeLimitations. ....... 15 4.2 Multiple Access to the Same FITS File . ......... 16 4.3 Current Header Data Unit (CHDU) . ...... 16 4.4 SubroutineNames ................................. 16 4.5 Subroutine Families and Datatypes . .......... 17 4.6 ImplicitDataTypeConversion . ........ 17 4.7 DataScaling ..................................... 18 4.8 Error Status Values and the Error Message Stack . ............ 18 4.9 Variable-Length Array Facility in Binary Tables . ............... 19 4.10 SupportforIEEESpecialValues . ......... 20 4.11 When the Final Size of the FITS HDU is Unknown . ......... 21 4.12 Local FITS Conventions supported by FITSIO . ........... 21 iii iv CONTENTS 4.12.1 Support for Long String Keyword Values. ......... 21 4.12.2 Arrays of Fixed-Length Strings in Binary Tables . ............ 22 4.12.3 KeywordUnitsStrings. ..... 23 4.12.4 HIERARCH Convention for Extended Keyword Names . ........ 23 4.13 Optimizing Code for Maximum Processing Speed . ............ 24 4.13.1 Background Information: How CFITSIO Manages Data I/O ......... 25 5 Basic Interface Routines 29 5.1 FITSIOErrorStatusRoutines . ....... 29 5.2 FileI/ORoutines................................. ..... 30 5.3 KeywordI/ORoutines.............................. ..... 32 5.4 DataI/ORoutines ................................. 33 6 Advanced Interface Subroutines 35 6.1 FITSFileOpenandCloseSubroutines: .
    [Show full text]
  • Learning from 25 Years of the Extensible N-Dimensional Data Format
    Elsevier Editorial System(tm) for Astronomy and Computing Manuscript Draft Manuscript Number: ASCOM-D-14-00017 Title: Learning from 25 years of the extensible N-Dimensional Data Format Article Type: Full Length Article Section/Category: Data management, archives, and virtual observatory Keywords: data formats Starlink Corresponding Author: Dr. Tim Jenness, Corresponding Author's Institution: Cornell University First Author: Tim Jenness Order of Authors: Tim Jenness; David S Berry; Malcolm J Currie; Peter W Draper; Frossie Economou; Norman Gray; Brian McIlwrath; Keith Shortridge; Mark B Taylor; Patrick T Wallace; Rodney F Warren-Smith Abstract: The extensible N-Dimensional Data Format (NDF) was designed and developed in the late 1980s to provide a data model suitable for use in a variety of astronomy data processing applications supported by the UK Starlink Project. Starlink applications were used extensively, primarily in the UK astronomical community, and form the basis of a number of advanced data reduction pipelines today. This paper provides an overview of the historical drivers for the development of NDF and the lessons learned from using a defined hierarchical data model for many years in data reduction software, data pipelines and in data acquisition systems. Manuscript Click here to view linked References Learning from 25 years of the extensible N-Dimensional Data Format Tim Jennessa,∗, David S. Berryb, Malcolm J. Currieb, Peter W. Draperc, Frossie Economoud, Norman Graye, Brian McIlwrathf, Keith Shortridgeg, Mark B. Taylorh,
    [Show full text]
  • Paper Writing
    Data Management at the UKIRT and JCMT Tim Jennessa and Frossie Economoua aJoint Astronomy Centre, 660 N. A‘ohok¯ u¯ Place, Hilo, HI, 96720, U.S.A. ABSTRACT For more than a decade the Joint Astronomy Centre has been developing software tools to simplify observing and make it possible to use the telescopes in many different operational modes. In order to support remote operations the data handling systems need to be in place to allow observation preparation, flexible queue scheduling, data quality pipelines and science archives all to be connected in a data-driven environment. We discuss the history of these developments at UKIRT and JCMT and how the decision to combine software development at both telescopes led each to get features that they could not have justified if they were treated independently. Keywords: JCMT, UKIRT, eSTAR, Data Handling, Pipelines 1. INTRODUCTION The Joint Astronomy Centre runs two distinctly different telescopes. The James Clerk Maxwell Telescope (JCMT) is the world’s largest sub-millimetre telescope with a 15-m primary mirror. The United Kingdom Infrared Telescope (UKIRT) is a 3.8-m infrared telescope. Both are sited on Mauna Kea in Hawaii and both telescopes have been operating for many years (UKIRT since 1979 and JCMT since 1987) and developed many legacy systems. In the late 1990s the software groups were merged and code re-use and shared support became an important driver for the future. 2. DATA PIPELINES In the mid-1990s data reduction software was written directly for the instrument with no intent to re-use when the instru- ment was retired.
    [Show full text]