ISBN # 1-60132-514-2; American Council on Science & Education / CSCE 2021
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
ACM SIGACT News Distributed Computing Column 28
ACM SIGACT News Distributed Computing Column 28 Idit Keidar Dept. of Electrical Engineering, Technion Haifa, 32000, Israel [email protected] Sergio Rajsbaum, who edited this column for seven years and established it as a relevant and popular venue, is stepping down. This issue is my first step in the big shoes he vacated. I would like to take this opportunity to thank Sergio for providing us with seven years’ worth of interesting columns. In producing these columns, Sergio has enjoyed the support of the community at-large and obtained material from many authors, who greatly contributed to the column’s success. I hope to enjoy a similar level of support; I warmly welcome your feedback and suggestions for material to include in this column! The main two conferences in the area of principles of distributed computing, PODC and DISC, took place this summer. This issue is centered around these conferences, and more broadly, distributed computing research as reflected therein, ranging from reviews of this year’s instantiations, through influential papers in past instantiations, to examining PODC’s place within the realm of computer science. I begin with a short review of PODC’07, and highlight some “hot” trends that have taken root in PODC, as reflected in this year’s program. Some of the forthcoming columns will be dedicated to these up-and- coming research topics. This is followed by a review of this year’s DISC, by Edward (Eddie) Bortnikov. For some perspective on long-running trends in the field, I next include the announcement of this year’s Edsger W. -
A Study on Asynchronous Randomized Consensus Title Algorithms for Byzantine Fault Tolerant Replication
A Study on Asynchronous Randomized Consensus Title Algorithms for Byzantine Fault Tolerant Replication Author(s) 中村, 純哉 Citation Issue Date Text Version ETD URL https://doi.org/10.18910/34568 DOI 10.18910/34568 rights Note Osaka University Knowledge Archive : OUKA https://ir.library.osaka-u.ac.jp/ Osaka University A Study on Asynchronous Randomized Consensus Algorithms for Byzantine Fault Tolerant Replication Submitted to Graduate School of Information Science and Technology Osaka University January 2014 Junya NAKAMURA iii List of Major Publications Journal Papers 1. Junya Nakamura, Tadashi Araragi, Toshimitsu Masuzawa, and Shigeru Masuyama, \A method of parallelizing consensuses for accelerating byzantine fault tolerance," IEICE Trans- actions on Information and Systems, vol. E97-D, no. 1, 2014. (to appear). 2. Junya Nakamura, Tadashi Araragi, Shigeru Masuyama, and Toshimitsu Masuzawa, “Effi- cient randomized byzantine fault-tolerant replication based on special valued coin tossing," IEICE Transactions on Information and Systems, vol. E97-D, no. 2, 2014. (to appear). Conference Papers 3. Junya Nakamura, Tadashi Araragi, and Shigeru Masuyama, \Asynchronous byzantine request- set agreement algorithm for replication," in Proceedings of the 1st AAAC Annual Meeting, p. 35, 2008. 4. Junya Nakamura, Tadashi Araragi, and Shigeru Masuyama, \Acceleration of byzantine fault tolerance by parallelizing consensuses," in Proceedings of the 10th International Conference on Parallel and Distributed Computing, Applications and Technologies, PDCAT '09, pp. 80{ 87, Dec. 2009. Technical Reports 5. Junya Nakamura, Tadashi Araragi, and Shigeru Masuyama, \Byzantine agreement on the order of processing received requests is solvable deterministically in asynchronous systems," in IEICE Technical Report, vol. 106 of COMP2006-35, pp. 33{40, Oct. -
A Tensor Compiler for Unified Machine Learning Prediction Serving
A Tensor Compiler for Unified Machine Learning Prediction Serving Supun Nakandala, UC San Diego; Karla Saur, Microsoft; Gyeong-In Yu, Seoul National University; Konstantinos Karanasos, Carlo Curino, Markus Weimer, and Matteo Interlandi, Microsoft https://www.usenix.org/conference/osdi20/presentation/nakandala This paper is included in the Proceedings of the 14th USENIX Symposium on Operating Systems Design and Implementation November 4–6, 2020 978-1-939133-19-9 Open access to the Proceedings of the 14th USENIX Symposium on Operating Systems Design and Implementation is sponsored by USENIX A Tensor Compiler for Unified Machine Learning Prediction Serving Supun Nakandalac,∗, Karla Saurm, Gyeong-In Yus,∗, Konstantinos Karanasosm, Carlo Curinom, Markus Weimerm, Matteo Interlandim mMicrosoft, cUC San Diego, sSeoul National University {<name>.<surname>}@microsoft.com,[email protected], [email protected] Abstract TensorFlow [13] combined, and growing faster than both. Machine Learning (ML) adoption in the enterprise requires Acknowledging this trend, traditional ML capabilities have simpler and more efficient software infrastructure—the be- been recently added to DNN frameworks, such as the ONNX- spoke solutions typical in large web companies are simply ML [4] flavor in ONNX [25] and TensorFlow’s TFX [39]. untenable. Model scoring, the process of obtaining predic- When it comes to owning and operating ML solutions, en- tions from a trained model over new data, is a primary con- terprises differ from early adopters in their focus on long-term tributor to infrastructure complexity and cost as models are costs of ownership and amortized return on investments [68]. trained once but used many times. In this paper we propose As such, enterprises are highly sensitive to: (1) complexity, HUMMINGBIRD, a novel approach to model scoring, which (2) performance, and (3) overall operational efficiency of their compiles featurization operators and traditional ML models software infrastructure [14]. -
Auto-Stabilisation : Solution Elégante Pour Lutter Contre Les Fautes Sylvie
Auto-stabilisation : Solution Elégante pour Lutter contre Les Fautes Sylvie Delaët Rapport scientifique présenté en vue de l’obtention de l’Habilitation à Diriger les Recherches. Soutenu le 8 novembre 2013 devant le jury composé de : Alain Denise, Professeur, Université Paris Sud, France Shlomi Dolev, Professeur, Ben Gurion University of the Negev, Israël Pierre Fraigniaud, Directeur de recherche CNRS, France Ted Herman, Professeur, University of Iowa, Etats Unis d’Amérique Nicolas Thiéry, Professeur, l’Université Paris Sud, France Franck Petit, Professeur, Université Paris Pierre et Marie Curie, France Après avis des rapporteurs : Ted Herman, Professeur, University of Iowa, Etats Unis d’Amérique Rachid Guerraoui, Professeur, Ecole Polytechnique Fédérale de Lausanne, Suisse Franck Petit, Université Paris Pierre et Marie Curie, France ii Sommaire Remerciements iii 1 Ma vision de ce document 1 1.1 Organisation du document . .1 2 Un peu de cuisine(s) 3 2.1 L'auberge espagnole de l'algorithmique r´epartie. .4 2.2 Exemples de probl`emes`ar´esoudreet quelques solutions . .8 3 Comprendre le monde 15 3.1 Comprendre l'auto-stabilisation . 15 3.2 Comprendre l'algorithmique r´epartie . 30 3.3 Comprendre mes contributions . 34 4 Construire l'avenir 43 4.1 Perspectives d'am´eliorationdes techniques . 43 4.2 Perspective d'ouverture `ala diff´erence . 45 4.3 Perspective de diffusion plus large . 47 4.4 Perspective de l'ordinateur auto-stabilisant . 48 A Agrafage 51 A.1 Cerner les probl`emes . 51 A.2 D´emasquerles planqu´es . 59 A.3 Recenser les pr´esents . 67 A.4 Obtenir l'auto-stabilisation gratuitement . -
Benchmarking and Optimization of Gradient Boosting Decision Tree Algorithms
Benchmarking and Optimization of Gradient Boosting Decision Tree Algorithms Andreea Anghel, Nikolaos Papandreou, Thomas Parnell Alessandro de Palma, Haralampos Pozidis IBM Research – Zurich, Rüschlikon, Switzerland {aan,npo,tpa,les,hap}@zurich.ibm.com Abstract Gradient boosting decision trees (GBDTs) have seen widespread adoption in academia, industry and competitive data science due to their state-of-the-art perfor- mance in many machine learning tasks. One relative downside to these models is the large number of hyper-parameters that they expose to the end-user. To max- imize the predictive power of GBDT models, one must either manually tune the hyper-parameters, or utilize automated techniques such as those based on Bayesian optimization. Both of these approaches are time-consuming since they involve repeatably training the model for different sets of hyper-parameters. A number of software GBDT packages have started to offer GPU acceleration which can help to alleviate this problem. In this paper, we consider three such packages: XG- Boost, LightGBM and Catboost. Firstly, we evaluate the performance of the GPU acceleration provided by these packages using large-scale datasets with varying shapes, sparsities and learning tasks. Then, we compare the packages in the con- text of hyper-parameter optimization, both in terms of how quickly each package converges to a good validation score, and in terms of generalization performance. 1 Introduction Many powerful techniques in machine learning construct a strong learner from a number of weak learners. Bagging combines the predictions of the weak learners, each using a different bootstrap sample of the training data set [1]. Boosting, an alternative approach, iteratively trains a sequence of weak learners, whereby the training examples for the next learner are weighted according to the success of the previously-constructed learners. -
Population Protocols: Expressiveness, Succinctness and Automatic Verification
Fakultat¨ fur¨ Informatik Technische Universitat¨ Munchen¨ Population Protocols: Expressiveness, Succinctness and Automatic Verification. Stefan Jaax Vollst¨andigerAbdruck der von der Fakult¨atf¨urInformatik der Technischen Universit¨at M¨unchen zur Erlangung des akademischen Grades eines Doktor der Naturwissenschaften (Dr. rer. nat.) genehmigten Dissertation. Vorsitzender: Prof. Dr. Helmut Seidl Pr¨ufendeder Dissertation: 1. Prof. Dr. Francisco Javier Esparza Estaun 2. Prof. Rupak Majumdar, Ph.D., Technische Universit¨atKaiserslautern Die Dissertation wurde am 03.03.2020 bei der Technischen Universit¨atM¨unchen eingereicht und durch die Fakult¨atf¨urInformatik am 04.06.2020 angenommen. Abstract Population protocols (Angluin et al., PODC, 2004) are a model of distributed computa- tion in which identical, finite-state, passively mobile agents interact in pairs to achieve a common goal. In the basic model of population protocols, agents compute number predicates by reaching a stable consensus. It is well known that population protocols compute precisely the semilinear predicates, or, equivalently, the predicates definable in Presburger arithmetic, the first-order theory of the natural numbers equipped with addition and the standard linear order. This thesis investigates three fundamental questions of the theory of population pro- tocols: Space complexity, verification complexity, and expressiveness of reasonable ex- tensions. Space Complexity. We show that every quantifier-free Presburger predicate ' aug- mented with remainder predicates is computable by a population protocol with poly(j'j) states, where j'j denotes the size of ' in binary encoding. Further, the protocol can be constructed in polynomial time. This is a major improvement to the previously known construction, which requires 2poly(j'j) states. As a special case, we consider predicates of the form 'c(x) = x ≥ c, where c is a positive integer constant. -
Lightgbm Release 3.2.1.99
LightGBM Release 3.2.1.99 Microsoft Corporation Sep 26, 2021 CONTENTS: 1 Installation Guide 3 2 Quick Start 21 3 Python-package Introduction 23 4 Features 29 5 Experiments 37 6 Parameters 43 7 Parameters Tuning 65 8 C API 71 9 Python API 99 10 Distributed Learning Guide 175 11 LightGBM GPU Tutorial 185 12 Advanced Topics 189 13 LightGBM FAQ 191 14 Development Guide 199 15 GPU Tuning Guide and Performance Comparison 201 16 GPU SDK Correspondence and Device Targeting Table 205 17 GPU Windows Compilation 209 18 Recommendations When Using gcc 229 19 Documentation 231 20 Indices and Tables 233 Index 235 i ii LightGBM, Release 3.2.1.99 LightGBM is a gradient boosting framework that uses tree based learning algorithms. It is designed to be distributed and efficient with the following advantages: • Faster training speed and higher efficiency. • Lower memory usage. • Better accuracy. • Support of parallel, distributed, and GPU learning. • Capable of handling large-scale data. For more details, please refer to Features. CONTENTS: 1 LightGBM, Release 3.2.1.99 2 CONTENTS: CHAPTER ONE INSTALLATION GUIDE This is a guide for building the LightGBM Command Line Interface (CLI). If you want to build the Python-package or R-package please refer to Python-package and R-package folders respectively. All instructions below are aimed at compiling the 64-bit version of LightGBM. It is worth compiling the 32-bit version only in very rare special cases involving environmental limitations. The 32-bit version is slow and untested, so use it at your own risk and don’t forget to adjust some of the commands below when installing. -
A Hybrid Machine Learning/Deep Learning COVID-19 Severity Predictive Model from CT Images and Clinical Data
A hybrid machine learning/deep learning COVID-19 severity predictive model from CT images and clinical data Matteo Chieregato1,*, Fabio Frangiamore2, Mauro Morassi3, Claudia Baresi4, Stefania Nici1, Chiara Bassetti1, Claudio Bna` 2, and Marco Galelli1 1Fondazione Poliambulanza Istituto Ospedaliero, Unit of Medical Physics, Brescia, 25124, Italy 2Tattile s.r.l., Mairano (BS), 25030 Italy 3Fondazione Poliambulanza Istituto Ospedaliero, Department of Diagnostic Imaging, Unit of Radiology, Brescia, 25124, Italy 4Fondazione Poliambulanza Istituto Ospedaliero, Information and Communications Technology, Unit of Lean Managing, Brescia, 25124, Italy *[email protected] ABSTRACT COVID-19 clinical presentation and prognosis are highly variable, ranging from asymptomatic and paucisymptomatic cases to acute respiratory distress syndrome and multi-organ involvement. We developed a hybrid machine learning/deep learning model to classify patients in two outcome categories, non-ICU and ICU (intensive care admission or death), using 558 patients admitted in a northern Italy hospital in February/May of 2020. A fully 3D patient-level CNN classifier on baseline CT images is used as feature extractor. Features extracted, alongside with laboratory and clinical data, are fed for selection in a Boruta algorithm with SHAP game theoretical values. A classifier is built on the reduced feature space using CatBoost gradient boosting algorithm and reaching a probabilistic AUC of 0.949 on holdout test set. The model aims to provide clinical decision support to medical doctors, with the probability score of belonging to an outcome class and with case-based SHAP interpretation of features importance. Introduction To date (May 2021), more than one hundred millions of individuals have been reported as affected by COVID-19. -
Population Protocols Are Fast
Population Protocols Are Fast Adrian Kosowski1 and Przemysław Uznanski´ 2 1Inria Paris, France 2 ETH Zürich, Switzerland Abstract A population protocol describes a set of state change rules for a population of n indistinguish- able finite-state agents (automata), undergoing random pairwise interactions. Within this very basic framework, it is possible to resolve a number of fundamental tasks in distributed computing, includ- ing: leader election, aggregate and threshold functions on the population, such as majority compu- tation, and plurality consensus. For the first time, we show that solutions to all of these problems can be obtained quickly using finite-state protocols. For any input, the designed finite-state proto- cols converge under a fair random scheduler to an output which is correct with high probability in expected O(poly log n) parallel time. In the same setting, we also show protocols which always reach a valid solution, in expected parallel time O(nε), where the number of states of the interacting automata depends only on the choice of ε > 0. The stated time bounds hold for any semi-linear predicate computable in the population protocol framework. The key ingredient of our result is the decentralized design of a hierarchy of phase-clocks, which tick at different rates, with the rates of adjacent clocks separated by a factor of Θ(log n). The construction of this clock hierarchy relies on a new protocol composition technique, combined with an adapted analysis of a self-organizing process of oscillatory dynamics. This clock hierarchy is used to provide nested synchronization primitives, which allow us to view the population in a global manner and design protocols using a high-level imperative programming language with a (limited) capacity for loops and branching instructions. -
New Directions in Automated Traffic Analysis
New Directions in Automated Traffic Analysis Jordan Holland1, Paul Schmitt1, Nick Feamster2, Prateek Mittal1 1 Princeton University 2 University of Chicago https://nprint.github.io/nprint ABSTRACT This paper reconsiders long-held norms in applying machine Despite the use of machine learning for many network traffic anal- learning to network traffic analysis; namely, we seek to reduce ysis tasks in security, from application identification to intrusion reliance upon human-driven feature engineering. To do so, we detection, the aspects of the machine learning pipeline that ulti- explore whether and how a single, standard representation of a mately determine the performance of the model—feature selection network packet can serve as a building block for the automation of and representation, model selection, and parameter tuning—remain many common traffic analysis tasks. Our goal is not to retread any manual and painstaking. This paper presents a method to auto- specific network classification problem, but rather to argue that mate many aspects of traffic analysis, making it easier to apply many of these problems can be made easier—and in some cases, machine learning techniques to a wider variety of traffic analysis completely automated—with a unified representation of traffic that tasks. We introduce nPrint, a tool that generates a unified packet is amenable for input to existing automated machine learning (Au- representation that is amenable for representation learning and toML) pipelines [14]. To demonstrate this capability, we designed a model training. We integrate nPrint with automated machine learn- standard packet representation, nPrint, that encodes each packet in ing (AutoML), resulting in nPrintML, a public system that largely an inherently normalized, binary representation while preserving eliminates feature extraction and model tuning for a wide variety the underlying semantics of each packet. -
2020 Sut20 CASC-J10.Pdf
CASC-J10 CASC-J10 CASC-J10 CASC-J10 Proceedings of the 10th IJCAR ATP System Competition (CASC-J10) Geo↵Sutcli↵e University of Miami, USA Abstract The CADE ATP System Competition (CASC) evaluates the performance of sound, fully automatic, classical logic, ATP systems. The evaluation is in terms of the number of problems solved, the number of acceptable proofs and models produced, and the average runtime for problems solved, in the context of a bounded number of eligible problems chosen from the TPTP problem library and other useful sources of test problems, and specified time limits on solution attempts. The 10th IJCAR ATP System Competition (CASC-J10) was held on 2th July 2020. The design of the competition and its rules, and information regarding the competing systems, are provided in this report. 1 Introduction The CADE and IJCAR conferences are the major forum for the presentation of new research in all aspects of automated deduction. In order to stimulate ATP research and system de- velopment, and to expose ATP systems within and beyond the ATP community, the CADE ATP System Competition (CASC) is held at each CADE and IJCAR conference. CASC-J10 was held on 2nd July 2020, as part of the 10th International Joint Conference on Automated Reasoning (IJCAR 2020)1, Online, Earth. It was the twenty-fifth competition in the CASC series [139, 145, 142, 95, 97, 138, 136, 137, 102, 104, 106, 108, 111, 113, 115, 117, 119, 121, 123, 144, 125, 127, 130, 132]. CASC evaluates the performance of sound, fully automatic, classical logic, ATP systems. -
WRMSDC 2018 Annual Report
WRMSDC 2018 Annual Report National Supplier Diversity leverages Kaiser Permanente's buying power to make a sustainable impact on the total health of the communities we serve. Our mission is to ensure the dollars spent by Kaiser Permanente contribute to economic and environmental health and reflect the diversity of the communities we serve. Table of Contents About Us ............................................................................. 2 Message From the Council ....................................................... 3 Highlights, Awards and Achievements ........................................ 4 Our Minority Business Enterprises ............................................12 Advisory Committee ............................................................. 15 Board of Directors ............................................................... 16 Committees .........................................................................17 Economic Impact ................................................................. 19 Financials .......................................................................... 20 20 Top Bay Area Companies for Supplier Diversity .................... 23 Members & Supporters ......................................................... 24 Testimonials ........................................................................ 29 Events ............................................................................... 30 Signature Events .................................................................. 36 Message of Thanks