Proceedings of the 16Th Python in Science Conference

Total Page:16

File Type:pdf, Size:1020Kb

Proceedings of the 16Th Python in Science Conference Proceedings of the 16th Python in Science Conference July 10 - July 16 Austin, Texas • Katy Huff David Lippa Dillon Niederhut M Pacer PROCEEDINGS OF THE 16TH PYTHON IN SCIENCE CONFERENCE Edited by Katy Huff, David Lippa, Dillon Niederhut, and M Pacer. SciPy 2017 Austin, Texas July 10 - July 16, 2017 Copyright c 2017. The articles in the Proceedings of the Python in Science Conference are copyrighted and owned by their original authors This is an open-access publication and is distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. For more information, please see: http://creativecommons.org/licenses/by/3.0/ ISSN:2575-9752 https://doi.org/10.25080/Majora-c9725a0-011 ORGANIZATION Conference Chairs PRABHU RAMACHANDRAN, Enthought Inc. & IIT Bombay SERGE REY, Arizona State University Program Chairs LORENA BARBA, George Washington University GIL FORSYTH, George Washington University Communications JUAN SHISHIDO, Twitter Birds of a Feather AASHISH CHAUDHARY, Kitware MATT MCCORMICK, Kitware Proceedings KATY HUFF, University of Illinois DAVID LIPPA, Amazon DILLON NIEDERHUT, Enthought MPACER, Berkeley Institute of Data Science Financial Aid SCOTT COLLIS, Argonne National Laboratory ERIC MA, MIT Tutorials ALEXANDRE CHABOT-LECLERC, Enthought MIKE HEARNE, USGS BEN ROOT, Atmospheric and Environmental Research, Inc. Sprints JONATHAN ROCHER, KBI Biopharma CHARLYE TRAN, Lewis University and The Recurse Center Diversity JULIE KRUGLER HOLLEK, Twitter Activities CHRIS NIEMIRA, AOL RANDY PAFFENROTH, Worcester Polytechnic Institute Sponsors JILL COWAN, Enthought Financial BILL COWAN, Enthought JODI HAVRANEK, Enthought Logistics JILL COWAN, Enthought LEAH JONES, Enthought Proceedings Reviewers DAVID LIPPA KATY HUFF MPACER DILLON NIEDERHUT CHIEN-HSIU LEE PRIYA SAWANT GUY TEL-ZUR NICK WAN ALEJANDRO DE LA VEGA GEORGE MARKOMANOLIS MATTHEW ROCKLIN OLEKSANDR PAVLYK JEAN BILHEUX BRIAN MCFEE ALEJANDRO WEINSTEIN STEFAN VAN DER WALT LI WANG SERGE GUELTON KAPIL SARASWAT THOMAS ARILDSEN PAUL CELICOURT CYRUS HARRISON KYLE NIEMEYER SARTAJ SINGH ANKUR ANKAN NATHAN GOLDBAUM ANGELOS KRYPOTOS JAIME ARIAS JONATHAN J. HELMUS FERNANDO CHIRIGATI CHRIS CALLOWAY NICHOLAS MALAYA DANIEL WHEELER JEREMIAH JOHNSON YU FENG YINGWEI YU MANOJ PANDEY JAMES A. BEDNAR ADRIAN HEILBUT JASON VERTREES HORACIO VARGAS FATIH AKICI SEBASTIAN BENTHALL NIRMAL SAHUJI SCHOLARSHIP RECIPIENTS WILL BARNES, Rice University SCOTT COLE, University of California, San Diego ROBERTO COLISTETE JR, UFES BJORN DAHLGREN, KTH Royal Institute of Technology FILIPE FERNANDES, SECOORA ERIC MA, Massachusetts Institute of Technology JEREMY MCGIBBON, University of Washington HIMANSHU MISHRA, NetworkX and Timelab DANIIL PAKHOMOV, Johns Hopkins University ZACH SAILER, University of Oregon CARL SAVAGE, School District 69 (Qualicum) TOBIAS SCHRAINK, New York University SCOTT SIEVERT, UWisconsin Madison SARTAJ SINGH, India Institute of Technology, BHU JUMP TRADING AND NUMFOCUS DIVERSITY SCHOLARSHIP RECIPIENTS SELENA ALEMAN, University of Texas at Austin CELIA CINTAS, CENPAT-CONICET NATALIA CLEMENTI, The George Washington University JOHANNA COHOON, The University of Texas at Austin CHARLYE DELGADO, Lewis University ALEXANDER IVANOV, The Institute for Information Transmission Problems of Russian Academy of Sciences DANA MILLER, University of California, Berkeley REBECCA MINICH, None JULIETTE SEIVE, University of Texas at Austin MALVIKA SHARAN, European Molecular Biology Laboratory (EMBL), Heidelberg CHAYA STERN, Weill Cornell Graduate School PAMELA WU, New York University ISRAEL ZU´ NIGA˜ DE LA MORA, Alturin CONTENTS SPORCO: A Python package for standard and convolutional sparse representations 1 Brendt Wohlberg Software Transactional Memory in Pure Python 9 Dillon Niederhut BespON: Extensible config files with multiline strings, lossless round-tripping, and hex floats 12 Geoffrey M. Poore LabbookDB: A Wet-Work-Tracking Database Application Framework 20 Horea-Ioan Ioanas, Bechara John Saab, Markus Rudin pyMolDyn: Identification, structure, and properties of cavities in condensed matter and molecules 28 Ingo Heimbach, Florian Rhiem, Fabian Beule, David Knodt, Josef Heinen, Robert O. Jones PyHRF: A Python Library for the Analysis of fMRI Data Based on Local Estimation of the Hemodynamic Response Function 34 Jaime Arias, Philippe Ciuciu, Michel Dojat, Florence Forbes, Aina Frau-Pascual, Thomas Perret, Jan M. Warnking SciSheets: Providing the Power of Programming With The Simplicity of Spreadsheets 41 Alicia Clark, Joseph L. Hellerstein The Sacred Infrastructure for Computational Research 49 Klaus Greff, Aaron Klein, Martin Chovanec, Frank Hutter, Jurgen¨ Schmidhuber FigureFirst: A Layout-first Approach for Scientific Figures 57 Theodore Lindsay, Peter T. Weir, Floris van Breugel Parallel Analysis in MDAnalysis using the Dask Parallel Computing Library 64 Mahzad Khoshlessan, Ioannis Paraskevakos, Shantenu Jha, Oliver Beckstein MatchPy: A Pattern Matching Library 73 Manuel Krebber, Henrik Bartels, Paolo Bientinesi pulse2percept: A Python-based simulation framework for bionic vision 81 Michael Beyeler, Geoffrey M. Boynton, Ione Fine, Ariel Rokem Optimised finite difference computation from symbolic equations 89 Michael Lange, Navjot Kukreja, Fabio Luporini, Mathias Louboutin, Charles Yount, Jan Huckelheim,¨ Gerard J. Gorman Python meets systems neuroscience: affordable, scalable and open-source electrophysiology in awake, behaving rodents 98 Narendra Mukherjee, Joseph Wachutka, Donald B Katz Accelerating Scientific Python with Intel Optimizations 106 Oleksandr Pavlyk, Denis Nagorny, Andres Guzman-Ballen, Anton Malakhov, Hai Liu, Ehsan Totoni, Todd A. Anderson, Sergey Maidanov NEXT: A system to easily connect crowdsourcing and adaptive data collection 113 Scott Sievert, Daniel Ross, Lalit Jain, Kevin Jamieson, Rob Nowak, Robert Mankoff ChiantiPy: a Python package for Astrophysical Spectroscopy 120 Will T. Barnes, Kenneth P. Dere PROC. OF THE 16th PYTHON IN SCIENCE CONF. (SCIPY 2017) 1 SPORCO: A Python package for standard and convolutional sparse representations ‡ Brendt Wohlberg ∗ ! Abstract—SParse Optimization Research COde (SPORCO) is an open-source Python package for solving optimization problems with sparsity-inducing regu- = larization, consisting primarily of sparse coding and dictionary learning, for both standard and convolutional forms of sparse representation. In the current ver- D s0 sion, all optimization problems are solved within the Alternating Direction Method x0 of Multipliers (ADMM) framework. SPORCO was developed for applications in signal and image processing, but is also expected to be useful for problems in computer vision, statistics, and machine learning. = s D Index Terms—sparse representations, convolutional sparse representations, s1 x sparse coding, convolutional sparse coding, dictionary learning, convolutional 1 dictionary learning, alternating direction method of multipliers Fig. 1: Independent sparse coding of overlapping blocks Introduction SPORCO is an open-source Python package for solving inverse problems with sparsity-inducing regularization [MBP14]. This D = type of regularization has become one of the leading techniques D in signal and image processing, with applications including image D denoising, inpainting, deconvolution, superresolution, and com- pressed sensing, to name only a few. It is also a prominent s method in machine learning and computer vision, with applica- tions including image classification, video background modeling, + + ∗ ∗ ··· collaborative filtering, and genomic data analysis, and is widely used in statistics as a regression technique. d0 d1 SPORCO was initially a Matlab library, but the implementa- tion language was switched to Python for a number of reasons, x0 x1 x including (i) the substantial cost of Matlab licenses (particularly in an environment that does not qualify for an academic discount), Fig. 2: Convolutional sparse coding of an entire signal and the difficulty of running large scale experiments on multiple hosts with a limited supply of toolbox licenses, (ii) the greater maintainability and flexibility of the object-oriented design possi- of sparse representations are supported. In the standard form the ble in Python, (iii) the flexibility provided by NumPy in indexing dictionary is a matrix, which limits the sizes of signals, images, arrays of arbitrary numbers of dimensions (essentially impossible etc. that can be directly represented; the usual strategy is to in Matlab), and (iv) the superiority of Python as a general-purpose compute independent representations for a set of overlapping programming language. blocks, as illustrated in Figure1. In the convolutional form SPORCO supports a variety of inverse problems, including [LS99][ZKTF10][Woh16d] the dictionary is a set of linear filters, Total Variation [ROF92][All92] denoising and deconvolution, making it feasible to directly represent an entire signal or image. and Robust PCA [CCS10], but the primary focus is on sparse The convolutional form is equivalent to sparse coding with a coding and dictionary learning, for solving problems with sparse structured dictionary constructed from translations of a smaller representations [MBP14]. Both standard and convolutional forms generating dictionary, as illustrated in Figure2. The support for the convolutional form is one of the major strengths of SPORCO * Corresponding author: [email protected] ‡ Theoretical Division,
Recommended publications
  • RAPIDS Scaling on Dell EMC Poweredge Servers
    Whitepaper RAPIDS Scaling on Dell EMC PowerEdge Servers Revision: 1.1 Issue Date: 2/4/2020 Issue Date: 2/4/2020 Abstract In this project we tested Dell EMC PowerEdge Servers with end to end (E2E) workflow using New York City - Taxi notebook to accelerate large scale workloads in scale-up and scale-out solutions using RAPID library from NVIDIA™ and DASK-CUDA for parallel computing with XGBoost. February 2020 1 RAPIDS Scaling on Dell EMC PowerEdge Servers Revisions Date Description 2/4/2020 Initial release Authors This paper was produced by the following: Name Vilmara Sanchez Dell EMC, Advanced Engineering Bhavesh Patel Dell EMC, Advanced Engineering Acknowledgements This paper was supported by the following: Name Josh Anderson Dell EMC Robert Crovella NVIDIA NVIDIA Account Team NVIDIA NVIDIA Developer Forum NVIDIA NVIDIA RAPIDS Dev Team NVIDIA Dask Development Team Library for dynamic task scheduling https://dask.org The information in this publication is provided “as is.” Dell Inc. makes no representations or warranties of any kind with respect to the information in this publication, and specifically disclaims implied warranties of merchantability or fitness for a particular purpose. Use, copying, and distribution of any software described in this publication requires an applicable software license. © February, 2020 Dell Inc. or its subsidiaries. All Rights Reserved. Dell, EMC, Dell EMC and other trademarks are trademarks of Dell Inc. or its subsidiaries. Other trademarks may be trademarks of their respective owners. Dell believes the information in this document is accurate as of its publication date. The information is subject to change without notice. 2 RAPIDS Scaling on Dell EMC PowerEdge Servers Table of Contents 1 RAPIDS Overview ...........................................................................................................................
    [Show full text]
  • Analytical and Other Software on the Secure Research Environment
    Analytical and other software on the Secure Research Environment The Research Environment runs with the operating system: Microsoft Windows Server 2016 DataCenter edition. The environment is based on the Microsoft Data Science virtual machine template and includes the following software: • R Version 4.0.2 (2020-06-22), as part of Microsoft R Open • R Studio Desktop 1.3.1093 working with R 4.0.2 • Anaconda 3, including an environment for Python 3.8.5 • Python, 3.8.5 as part of the Anaconda base environment • Jupyter Notebook, as part of the Anaconda3 environment • Microsoft Office 2016 Standard edition, including Word, Excel, PowerPoint, and OneNote (Access not included) • JuliaPro 0.5.1.1 and the Juno IDE for Julia • PyCharm Community Edition, 2020.3 • PLINK • JAGS • WinBUGS • OpenBUGS • stan and rstan • Apache Spark 2.2.0 • SparkML and pySpark • Apache Drill 1.11.0 • MAPR Drill driver • VIM 8.0.606 • TensorFlow • MXNet, MXNet Model Server • Microsoft Cognitive Toolkit (CNTK) • Weka • Vowpal Wabbit • xgboost • Team Data Science Process (TDSP) Utilities • VOTT (Visual Object Tagging Tool) 1.6.11 • Microsoft Machine Learning Server • PowerBI • Docker version 10.03.5, build 2ee0c57608 • SQL Server Developer Edition (2017), including Management Studio and SQL Server Integration Services (SSIS) • Visual Studio Code 1.17.1 • Nodejs • 7-zip • Evince PDF Viewer • Acrobat Reader • Microsoft Photo Viewer • PowerShell 6 March 2021 Version 1.2 And in the Premium research environments: • STATA 16.1 • SAS 9.4, m4 (academic license) Users also have the ability to bring in additional software if the software was specified in the data request, the software runs in the operating system described above, and the user can provide Vivli with any necessary licensing keys.
    [Show full text]
  • Intel® Distribution for Python* High Performance Python for Data Analytics and More Asma Farjallah
    Intel® Distribution for Python* High Performance Python for Data Analytics and More Asma Farjallah Courtesy of Frank Schlimbach 1 Legal Disclaimer & Optimization Notice INFORMATION IN THIS DOCUMENT IS PROVIDED “AS IS”. NO LICENSE, EXPRESS OR IMPLIED, BY ESTOPPEL OR OTHERWISE, TO ANY INTELLECTUAL PROPERTY RIGHTS IS GRANTED BY THIS DOCUMENT. INTEL ASSUMES NO LIABILITY WHATSOEVER AND INTEL DISCLAIMS ANY EXPRESS OR IMPLIED WARRANTY, RELATING TO THIS INFORMATION INCLUDING LIABILITY OR WARRANTIES RELATING TO FITNESS FOR A PARTICULAR PURPOSE, MERCHANTABILITY, OR INFRINGEMENT OF ANY PATENT, COPYRIGHT OR OTHER INTELLECTUAL PROPERTY RIGHT. Software and workloads used in performance tests may have been optimized for performance only on Intel microprocessors. Performance tests, such as SYSmark and MobileMark, are measured using specific computer systems, components, software, operations and functions. Any change to any of those factors may cause the results to vary. You should consult other information and performance tests to assist you in fully evaluating your contemplated purchases, including the performance of that product when combined with other products. Tests document performance of components on a particular test, in specific systems. Differences in hardware, software, or configuration will affect actual performance. Consult other sources of information to evaluate performance as you consider your purchase. Results have been simulated and are provided for informational purposes only. Results were derived using simulations run on an architecture simulator or model. Any difference in system hardware or software design or configuration may affect actual performance. For more complete information about performance and benchmark results, visit http://www.intel.com/performance. Intel does not control or audit the design or implementation of third party benchmark data or Web sites referenced in this document.
    [Show full text]
  • Latest Version of Dask.Distributed from the Conda-Forge Repository Using Conda: Conda Install Dask Distributed-C Conda-Forge
    Dask.distributed Documentation Release 2021.09.1+8.g672217ce Matthew Rocklin Sep 27, 2021 GETTING STARTED 1 Motivation 3 2 Architecture 5 3 Contents 7 Index 235 i ii Dask.distributed Documentation, Release 2021.09.1+8.g672217ce Dask.distributed is a lightweight library for distributed computing in Python. It extends both the concurrent. futures and dask APIs to moderate sized clusters. See the quickstart to get started. GETTING STARTED 1 Dask.distributed Documentation, Release 2021.09.1+8.g672217ce 2 GETTING STARTED CHAPTER ONE MOTIVATION Distributed serves to complement the existing PyData analysis stack. In particular it meets the following needs: • Low latency: Each task suffers about 1ms of overhead. A small computation and network roundtrip can complete in less than 10ms. • Peer-to-peer data sharing: Workers communicate with each other to share data. This removes central bottle- necks for data transfer. • Complex Scheduling: Supports complex workflows (not just map/filter/reduce) which are necessary for sophis- ticated algorithms used in nd-arrays, machine learning, image processing, and statistics. • Pure Python: Built in Python using well-known technologies. This eases installation, improves efficiency (for Python users), and simplifies debugging. • Data Locality: Scheduling algorithms cleverly execute computations where data lives. This minimizes network traffic and improves efficiency. • Familiar APIs: Compatible with the concurrent.futures API in the Python standard library. Compatible with dask API for parallel algorithms • Easy Setup: As a Pure Python package distributed is pip installable and easy to set up on your own cluster. 3 Dask.distributed Documentation, Release 2021.09.1+8.g672217ce 4 Chapter 1.
    [Show full text]
  • Scaling Data Science in Python with Dask
    Scaling Data Science in Python with Dask Accelerate your existing workflow with little to no code changes for data science that is screaming fast. Using data to innovate, deliver, or disrupt in today’s marketplace requires the right software tools for the job. For decades, it has been difficult to process data that is large or complex on a single machine. Open source solutions have emerged to ease the processing of big data, and the Python and the PyData ecosystem offer an incredibly rich, mature and well-supported set of solutions for data science. If you’re using Python and your team is like most, you are patching together disparate software and workloads, manually building and deploying images to run in containers, and creating ad hoc clusters. You may be working with Python packages like pandas or NumPy, which are extremely useful for data science but struggle with big data that exceeds the processing power of your machine. Are you looking for a better way? Dask is a free and open source library that provides a framework for distributed computing using Python syntax. With minimal code changes, you can process your data in parallel, which means it will take you less time to execute and give you more time for analysis. We created this guide for data scientists, software engineers, and DevOps engineers. It will define parallel and distributed computing and the challenges of creating these environments for data science at scale. We will discuss the benefits of the Pythonic ecosystem and share how scientists and innovators use Dask to scale data science.
    [Show full text]
  • Ipython Parallel Documentation Release 6.0.2
    IPython Parallel Documentation Release 6.0.2 The IPython Development Team Feb 23, 2017 Contents 1 Installing IPython Parallel 3 2 Contents 5 3 ipyparallel API 93 4 Indices and tables 95 Bibliography 97 Python Module Index 99 i ii IPython Parallel Documentation, Release 6.0.2 Release 6.0.2 Date Feb 23, 2017 Contents 1 IPython Parallel Documentation, Release 6.0.2 2 Contents CHAPTER 1 Installing IPython Parallel As of 4.0, IPython parallel is now a standalone package called ipyparallel. You can install it with: pip install ipyparallel or: conda install ipyparallel And if you want the IPython clusters tab extension in your Jupyter Notebook dashboard: ipcluster nbextension enable 3 IPython Parallel Documentation, Release 6.0.2 4 Chapter 1. Installing IPython Parallel CHAPTER 2 Contents Changes in IPython Parallel 6.0.2 Upload fixed sdist for 6.0.1. 6.0.1 Small encoding fix for Python 2. 6.0 Due to a compatibility change and semver, this is a major release. However, it is not a big release. The main compat- ibility change is that all timestamps are now timezone-aware UTC timestamps. This means you may see comparison errors if you have code that uses datetime objects without timezone info (so-called naïve datetime objects). Other fixes: • Rename Client.become_distributed() to Client.become_dask(). become_distributed() remains as an alias. • import joblib from a public API instead of a private one when using IPython Parallel as a joblib backend. • Compatibility fix in extensions for security changes in notebook 4.3 5.2 • Fix compatibility with changes in ipykernel 4.3, 4.4 • Improve inspection of @remote decorated functions 5 IPython Parallel Documentation, Release 6.0.2 • Client.wait() accepts any Future.
    [Show full text]
  • Dask Stories Documentation
    Dask Stories Documentation Dask Community Mar 10, 2021 Contents 1 Sidewalk Labs: Civic Modeling3 2 Genome Sequencing for Mosquitos5 3 Full Spectrum: Credit and Banking9 4 IceCube: Detecting Cosmic Rays 11 5 Pangeo: Earth Science 13 6 NCAR: Hydrological Modeling 17 7 Mobile Network Modeling 19 8 Satellite Imagery Processing 23 9 Prefect: Production Workflows 27 10 Discovering biomarkers for rare diseases in blood samples 31 11 Overview 35 12 Contributing 37 i ii Dask Stories Documentation Dask is a versatile tool that supports a variety of workloads. This page contains brief and illustrative examples of how people use Dask in practice. These emphasize breadth and hopefully inspire readers to find new ways that Dask can serve them beyond their original intent. Contents 1 Dask Stories Documentation 2 Contents CHAPTER 1 Sidewalk Labs: Civic Modeling 1.1 Who am I? I’m Brett Naul. I work at Sidewalk Labs. 1.2 What problem am I trying to solve? My team @ Sidewalk (“Model Lab”) uses machine learning models to study human travel behavior in cities and produce high-fidelity simulations of the travel patterns/volumes in a metro area. Our process has three main steps: • Construct a “synthetic population” from census data and other sources of demographic information; this popu- lation is statistically representative of the true population but contains no actual identifiable individuals. • Train machine learning models on anonymous mobile location data to understand behavioral patterns in the region (what times do people go to lunch, what factors affect an individual’s likelihood to use public transit, etc.). • For each person in the synthetic population, generate predictions from these models and combine the resulting into activities into a single model of all the activity in a region.
    [Show full text]
  • Dask.Distributed Documentation Release 2021.09.1
    Dask.distributed Documentation Release 2021.09.1 Matthew Rocklin Sep 21, 2021 GETTING STARTED 1 Motivation 3 2 Architecture 5 3 Contents 7 Index 235 i ii Dask.distributed Documentation, Release 2021.09.1 Dask.distributed is a lightweight library for distributed computing in Python. It extends both the concurrent. futures and dask APIs to moderate sized clusters. See the quickstart to get started. GETTING STARTED 1 Dask.distributed Documentation, Release 2021.09.1 2 GETTING STARTED CHAPTER ONE MOTIVATION Distributed serves to complement the existing PyData analysis stack. In particular it meets the following needs: • Low latency: Each task suffers about 1ms of overhead. A small computation and network roundtrip can complete in less than 10ms. • Peer-to-peer data sharing: Workers communicate with each other to share data. This removes central bottle- necks for data transfer. • Complex Scheduling: Supports complex workflows (not just map/filter/reduce) which are necessary for sophis- ticated algorithms used in nd-arrays, machine learning, image processing, and statistics. • Pure Python: Built in Python using well-known technologies. This eases installation, improves efficiency (for Python users), and simplifies debugging. • Data Locality: Scheduling algorithms cleverly execute computations where data lives. This minimizes network traffic and improves efficiency. • Familiar APIs: Compatible with the concurrent.futures API in the Python standard library. Compatible with dask API for parallel algorithms • Easy Setup: As a Pure Python package distributed is pip installable and easy to set up on your own cluster. 3 Dask.distributed Documentation, Release 2021.09.1 4 Chapter 1. Motivation CHAPTER TWO ARCHITECTURE Dask.distributed is a centrally managed, distributed, dynamic task scheduler.
    [Show full text]
  • Ipython Parallel Documentation Release 6.2.1
    IPython Parallel Documentation Release 6.2.1 The IPython Development Team Jun 05, 2018 Contents 1 Installing IPython Parallel 3 2 Contents 5 3 ipyparallel API 95 4 Indices and tables 109 Bibliography 111 Python Module Index 113 i ii IPython Parallel Documentation, Release 6.2.1 Release 6.2.1 Date Jun 05, 2018 Contents 1 IPython Parallel Documentation, Release 6.2.1 2 Contents CHAPTER 1 Installing IPython Parallel As of 4.0, IPython parallel is now a standalone package called ipyparallel. You can install it with: pip install ipyparallel or: conda install ipyparallel As of IPython Parallel 6.2, this will additionally install and enable the IPython Clusters tab in the Jupyter Notebook dashboard. You can also enable/install the clusters tab yourself after the fact with: ipcluster nbextension enable Or the individual Jupyter commands: jupyter serverextension install [–sys-prefix] –py ipyparallel jupyter nbextension install [–sys-prefix] –py ipyparallel jupyter nbextension enable [–sys-prefix] –py ipyparallel 3 IPython Parallel Documentation, Release 6.2.1 4 Chapter 1. Installing IPython Parallel CHAPTER 2 Contents 2.1 Changes in IPython Parallel 2.1.1 6.2.1 • Workaround a setuptools issue preventing installation from sdist on Windows 2.1.2 6.2.0 • Drop support for Python 3.3. IPython parallel now requires Python 2.7 or >= 3.4. • Further fixes for compatibility with tornado 5 when run with asyncio (Python 3) • Fix for enabling clusters tab via nbextension • Multiple fixes for handling when engines stop unexpectedly • Installing IPython Parallel enables the Clusters tab extension by default, without any additional commands.
    [Show full text]