Lightning Talk Presentations
Total Page:16
File Type:pdf, Size:1020Kb
Automated Kaguya TC and MRO CTX Stereo DEM Generation 5th Planetary Data Workshop and 2nd Planetary Science Informatics & Data Analytics Meeting Processing Session 1: June 28, 2021 Lauren A. Adoram-Kershner, [email protected] Ben H. Wheeler, [email protected] Jason R. Laura, [email protected] Robin L. Fergason, [email protected] David P. Mayer, [email protected] Motivation • The volume of stereographic images coupled with manual generation standard means a lot of DEM potential. Figure 2: Distribution of • Targeted generation; not covering full datasets. Kaguya TC stereopairs. • Served across various research institutes applying their own processing. • Difficulty tracing provenance of processing typically contained in separate paper(s). Method: Serving LOLA RDR ASAP-Stereo, Ames Stereo Automated Pipeline A. M. Annex1, K. W. Lewis1. 1Department of Earth and Planetary Sciences, Johns Hopkins University, Baltimore, MD 21218, USA ([email protected]) 1 @AndrewAnnex [email protected] www.andrewannex.com What is ASAP-Stereo? 1. Python Library and CLI, higher-level wrapper for Ames Stereo Pipeline 2. Framework to implement ASP workflows, currently implements “asp_scripts” for HiRISE & CTX with enhancements Why use ASAP-Stereo? 1. Reproducibility with Jupyter Notebooks, avoid ad hoc processing! 2. Speedups over asp_scripts 3. Ease of use, less monitoring 2 @AndrewAnnex [email protected] www.andrewannex.com Try ASAP-Stereo! Welcoming contributions for other instruments! Links: https://asap-stereo.readthedocs.io https://asap-stereo.readthedocs.io/en/main/install.html https://github.com/AndrewAnnex/asap_stereo https://doi.org/10.5281/zenodo.4171570 3 @AndrewAnnex [email protected] www.andrewannex.com High-Throughput Processing of Diviner Data from the Planetary Data System (PDS) Norbert Schorghofer ([email protected]) Planetary Science Institute, Arizona & Hawaii June 2021, Lightning Presentation, Abstract #7004 I The PDS currently holds about 2 PB of data; High-throughput data processing is increasingly important I Case Study: Diviner, onboard LRO, 160 TB of numerical data I Benchmarked various compression file formats for these data I Benchmarked decompression | awk I Benchmarked parallel processing from single drive (HDD + SSD) I How can I get my hardware's worth of throughput, or more? File Size and Decompression Speed 20 lzma br zst-19 15 rzip bz2 10 lzma-0 br-9 rar zst zip gz Compression Ratio Compression lz4-9 5 lz4 txt 0 0 0.2 0.4 0.6 0.8 Decompression Speed (GB/s) Parallel Throughput, Solid State Drive (SSD) 1.2 txt gzip lz4-9 1 lzma zip 0.8 zstd-19 0.6 Throughput (GB/s)Throughput 0.4 0.2 0 1 2 3 4 5 6 # Parallel Processes Throughput relative to uncompressed file size..