Open Source Copyrights

Total Page:16

File Type:pdf, Size:1020Kb

Open Source Copyrights Kuri App - Open Source Copyrights: 001_talker_listener-master_2015-03-02 ===================================== Source Code can be found at: https://github.com/awesomebytes/python_profiling_tutorial_with_ros 001_talker_listener-master_2016-03-22 ===================================== Source Code can be found at: https://github.com/ashfaqfarooqui/ROSTutorials acl_2.2.52-1_amd64.deb ====================== Licensed under GPL 2.0 License terms can be found at: http://savannah.nongnu.org/projects/acl/ acl_2.2.52-1_i386.deb ===================== Licensed under LGPL 2.1 License terms can be found at: http://metadata.ftp- master.debian.org/changelogs/main/a/acl/acl_2.2.51-8_copyright actionlib-1.11.2 ================ Licensed under BSD Source Code can be found at: https://github.com/ros/actionlib License terms can be found at: http://wiki.ros.org/actionlib actionlib-common-1.5.4 ====================== Licensed under BSD Source Code can be found at: https://github.com/ros-windows/actionlib License terms can be found at: http://wiki.ros.org/actionlib adduser_3.113+nmu3ubuntu3_all.deb ================================= Licensed under GPL 2.0 License terms can be found at: http://mirrors.kernel.org/ubuntu/pool/main/a/adduser/adduser_3.113+nmu3ubuntu3_all. deb alsa-base_1.0.25+dfsg-0ubuntu4_all.deb ====================================== Licensed under GPL 2.0 License terms can be found at: http://mirrors.kernel.org/ubuntu/pool/main/a/alsa- driver/alsa-base_1.0.25+dfsg-0ubuntu4_all.deb alsa-utils_1.0.27.2-1ubuntu2_amd64.deb ====================================== Licensed under GPL 2.0 License terms can be found at: http://mirrors.kernel.org/ubuntu/pool/main/a/alsa- utils/alsa-utils_1.0.27.2-1ubuntu2_amd64.deb alsa-utils_1.0.27.2-1ubuntu2_i386.deb ===================================== Licensed under GPL 2.0 License terms can be found at: http://mirrors.kernel.org/ubuntu/pool/main/a/alsa- utils/alsa-utils_1.0.27.2-1ubuntu2_i386.deb alufr-ros-pkg-stacks ==================== Licensed under BSD Source Code can be found at: http://alufr-ros-pkg.googlecode.com/svn/ License terms can be found at: http://library.isr.ist.utl.pt/docs/roswiki/alufr(2d)ros(2d)pkg.html anacron_2.3-20ubuntu1_amd64.deb =============================== Licensed under GPL 2.0 License terms can be found at: http://mirrors.kernel.org/ubuntu/pool/main/a/anacron/anacron_2.3- 20ubuntu1_amd64.deb anacron_2.3-20ubuntu1_i386.deb ============================== Licensed under GPL 2.0 License terms can be found at: http://mirrors.kernel.org/ubuntu/pool/main/a/anacron/anacron_2.3-20ubuntu1_i386.deb ansible_1.5.4+dfsg-1_all.deb ============================ Licensed under GPL 3.0, Apache 2.0, GPL 2.0, BSD 2 License terms can be found at: http://mirrors.kernel.org/ubuntu/pool/universe/a/ansible/ansible_1.5.4+dfsg- 1_all.debLicense terms can be found at: http://mirrors.kernel.org/ubuntu/pool/universe/a/ansible/ansible_1.5.4+dfsg- 1_all.debLicense terms can be found at: http://mirrors.kernel.org/ubuntu/pool/universe/a/ansible/ansible_1.5.4+dfsg- 1_all.debLicense terms can be found at: http://mirrors.kernel.org/ubuntu/pool/universe/a/ansible/ansible_1.5.4+dfsg-1_all.deb apt-utils_1.0.1ubuntu2.17_amd64.deb =================================== Licensed under GPL 2.0 License terms can be found at: http://security.ubuntu.com/ubuntu/pool/main/a/apt/apt- utils_1.0.1ubuntu2.17_amd64.deb apt-utils_1.0.1ubuntu2.17_i386.deb ================================== Licensed under GPL 2.0 License terms can be found at: http://security.ubuntu.com/ubuntu/pool/main/a/apt/apt- utils_1.0.1ubuntu2.17_i386.deb apt_1.0.1ubuntu2.17_amd64.deb ============================= Licensed under GPL 2.0 License terms can be found at: http://security.ubuntu.com/ubuntu/pool/main/a/apt/apt_1.0.1ubuntu2.17_amd64.deb apt_1.0.1ubuntu2.17_i386.deb ============================ Licensed under GPL 2.0 License terms can be found at: http://security.ubuntu.com/ubuntu/pool/main/a/apt/apt_1.0.1ubuntu2.17_i386.deb ardrone_swarm-master_2013-05-02 =============================== Licensed under Assorted Source Code can be found at: https://github.com/futuhal-arifin/ardrone_swarm License terms can be found at: https://github.com/daraosn/ardrone- swarm/blob/master/LICENSE at-spi2-core_2.10.2.is.2.10.1-0ubuntu1_amd64.deb ================================================ Licensed under Public Domain, LGPL 2.0 License terms can be found at: http://mirrors.kernel.org/ubuntu/pool/main/a/at-spi2- core/at-spi2-core_2.10.2.is.2.10.1-0ubuntu1_amd64.debLicense terms can be found at: http://mirrors.kernel.org/ubuntu/pool/main/a/at-spi2-core/at-spi2-core_2.10.2.is.2.10.1- 0ubuntu1_amd64.deb at-spi2-core_2.10.2.is.2.10.1-0ubuntu1_i386.deb =============================================== Licensed under Public Domain, LGPL 2.0 License terms can be found at: http://mirrors.kernel.org/ubuntu/pool/main/a/at-spi2- core/at-spi2-core_2.10.2.is.2.10.1-0ubuntu1_i386.debLicense terms can be found at: http://mirrors.kernel.org/ubuntu/pool/main/a/at-spi2-core/at-spi2-core_2.10.2.is.2.10.1- 0ubuntu1_i386.deb avahi-daemon_0.6.31-4ubuntu1.1_amd64.deb ======================================== Licensed under LGPL 2.1, GPL 2.0, BSD 2 License terms can be found at: http://security.ubuntu.com/ubuntu/pool/main/a/avahi/avahi-daemon_0.6.31- 4ubuntu1.1_amd64.debLicense terms can be found at: http://security.ubuntu.com/ubuntu/pool/main/a/avahi/avahi-daemon_0.6.31- 4ubuntu1.1_amd64.debLicense terms can be found at: http://security.ubuntu.com/ubuntu/pool/main/a/avahi/avahi-daemon_0.6.31- 4ubuntu1.1_amd64.deb avahi-daemon_0.6.31-4ubuntu1.1_i386.deb ======================================= Licensed under LGPL 2.1, GPL 2.0, BSD 2 License terms can be found at: http://security.ubuntu.com/ubuntu/pool/main/a/avahi/avahi-daemon_0.6.31- 4ubuntu1.1_i386.debLicense terms can be found at: http://security.ubuntu.com/ubuntu/pool/main/a/avahi/avahi-daemon_0.6.31- 4ubuntu1.1_i386.debLicense terms can be found at: http://security.ubuntu.com/ubuntu/pool/main/a/avahi/avahi-daemon_0.6.31- 4ubuntu1.1_i386.deb avahi-utils_0.6.31-4ubuntu1.1_amd64.deb ======================================= Licensed under LGPL 2.1, GPL 2.0, BSD 2 License terms can be found at: http://security.ubuntu.com/ubuntu/pool/main/a/avahi/avahi-utils_0.6.31- 4ubuntu1.1_amd64.debLicense terms can be found at: http://security.ubuntu.com/ubuntu/pool/main/a/avahi/avahi-utils_0.6.31- 4ubuntu1.1_amd64.debLicense terms can be found at: http://security.ubuntu.com/ubuntu/pool/main/a/avahi/avahi-utils_0.6.31- 4ubuntu1.1_amd64.deb avahi-utils_0.6.31-4ubuntu1.1_i386.deb ====================================== Licensed under LGPL 2.1, GPL 2.0, BSD 2 License terms can be found at: http://security.ubuntu.com/ubuntu/pool/main/a/avahi/avahi-utils_0.6.31- 4ubuntu1.1_i386.debLicense terms can be found at: http://security.ubuntu.com/ubuntu/pool/main/a/avahi/avahi-utils_0.6.31- 4ubuntu1.1_i386.debLicense terms can be found at: http://security.ubuntu.com/ubuntu/pool/main/a/avahi/avahi-utils_0.6.31- 4ubuntu1.1_i386.deb avr-libc_1.8.0-4.1_all.deb ========================== Licensed under GPL License terms can be found at: http://www.nongnu.org/avr-libc/LICENSE.txt avr-libc_1.8.1-0mayfield_all.deb ================================ Licensed under GPL License terms can be found at: http://www.nongnu.org/avr-libc/LICENSE.txt avrdude_6.1_amd64.deb ===================== Licensed under GPL 2.0 License terms can be found at: http://metadata.ftp- master.debian.org/changelogs/main/a/avrdude/avrdude_6.1-2_copyright base-files_7.2ubuntu5.5_amd64.deb ================================= Licensed under GPL 2.0 License terms can be found at: http://security.ubuntu.com/ubuntu/pool/main/b/base- files/base-files_7.2ubuntu5.5_amd64.deb base-files_7.2ubuntu5.5_i386.deb ================================ Licensed under GPL 2.0 License terms can be found at: http://security.ubuntu.com/ubuntu/pool/main/b/base- files/base-files_7.2ubuntu5.5_i386.deb base-passwd_3.5.33_amd64.deb ============================ Licensed under GPL 2.0 License terms can be found at: http://mirrors.kernel.org/ubuntu/pool/main/b/base- passwd/base-passwd_3.5.33_amd64.deb base-passwd_3.5.33_i386.deb =========================== Licensed under GPL 2.0 License terms can be found at: http://mirrors.kernel.org/ubuntu/pool/main/b/base- passwd/base-passwd_3.5.33_i386.deb bash-completion_2.1-4ubuntu0.2_all.deb ====================================== Licensed under GPL 2.0 License terms can be found at: http://security.ubuntu.com/ubuntu/pool/main/b/bash- completion/bash-completion_2.1-4ubuntu0.2_all.deb bash_4.3-7ubuntu1.5_amd64.deb ============================= Licensed under GPL 3.0, BSD 4 License terms can be found at: http://mirrors.kernel.org/ubuntu/pool/main/b/bash/bash_4.3- 7ubuntu1.5_amd64.debLicense terms can be found at: http://mirrors.kernel.org/ubuntu/pool/main/b/bash/bash_4.3-7ubuntu1.5_amd64.deb bash_4.3-7ubuntu1.5_i386.deb ============================ Licensed under GPL 3.0, BSD 4 License terms can be found at: http://mirrors.kernel.org/ubuntu/pool/main/b/bash/bash_4.3- 7ubuntu1.5_i386.debLicense terms can be found at: http://mirrors.kernel.org/ubuntu/pool/main/b/bash/bash_4.3-7ubuntu1.5_i386.deb bind9-host_9.9.5.dfsg-3ubuntu0.11_amd64.deb =========================================== Licensed under GPL License terms can be found at: http://changelogs.ubuntu.com/changelogs/pool/main/b/bind9/bind9_9.9.5.dfsg- 3ubuntu0.16/copyright bind9-host_9.9.5.dfsg-3ubuntu0.11_i386.deb ========================================== Licensed under GPL License terms can be found at: http://changelogs.ubuntu.com/changelogs/pool/main/b/bind9/bind9_9.9.5.dfsg-
Recommended publications
  • What If What I Need Is Not in Powerai (Yet)? What You Need to Know to Build from Scratch?
    IBM Systems What if what I need is not in PowerAI (yet)? What you need to know to build from scratch? Jean-Armand Broyelle June 2018 IBM Systems – Cognitive Era Things to consider when you have to rebuild a framework © 2017 International Business Machines Corporation 2 IBM Systems – Cognitive Era CUDA Downloads © 2017 International Business Machines Corporation 3 IBM Systems – Cognitive Era CUDA 8 – under Legacy Releases © 2017 International Business Machines Corporation 4 IBM Systems – Cognitive Era CUDA 8 Install Steps © 2017 International Business Machines Corporation 5 IBM Systems – Cognitive Era cuDNN and NVIDIA drivers © 2017 International Business Machines Corporation 6 IBM Systems – Cognitive Era cuDNN v6.0 for CUDA 8.0 © 2017 International Business Machines Corporation 7 IBM Systems – Cognitive Era cuDNN and NVIDIA drivers © 2017 International Business Machines Corporation 8 IBM Systems – Cognitive Era © 2017 International Business Machines Corporation 9 IBM Systems – Cognitive Era © 2017 International Business Machines Corporation 10 IBM Systems – Cognitive Era cuDNN and NVIDIA drivers © 2017 International Business Machines Corporation 11 IBM Systems – Cognitive Era Prepare your environment • When something goes wrong it’s better to Remove local anaconda installation $ cd ~; rm –rf anaconda2 .conda • Reinstall anaconda $ cd /tmp; wget https://repo.anaconda.com/archive/Anaconda2-5.1.0-Linux- ppc64le.sh $ bash /tmp/Anaconda2-5.1.0-Linux-ppc64le.sh • Activate PowerAI $ source /opt/DL/tensorflow/bin/tensorflow-activate • When you
    [Show full text]
  • Nosql Databases
    Query & Exploration SQL, Search, Cypher, … Stream Processing Platforms Data Storm, Spark, .. Data Ingestion Serving ETL, Distcp, Batch Processing Platforms BI, Cubes, Kafka, MapReduce, SparkSQL, BigQuery, Hive, Cypher, ... RDBMS, Key- OpenRefine, value Stores, … Data Definition Tableau, … SQL DDL, Avro, Protobuf, CSV Storage Systems HDFS, RDBMS, Column Stores, Graph Databases Computing Platforms Distributed Commodity, Clustered High-Performance, Single Node Query & Exploration SQL, Search, Cypher, … Stream Processing Platforms Data Storm, Spark, .. Data Ingestion Serving ETL, Distcp, Batch Processing Platforms BI, Cubes, Kafka, MapReduce, SparkSQL, BigQuery, Hive, Cypher, ... RDBMS, Key- OpenRefine, value Stores, … Data Definition Tableau, … SQL DDL, Avro, Protobuf, CSV Storage Systems HDFS, RDBMS, Column Stores, Graph Databases Computing Platforms Distributed Commodity, Clustered High-Performance, Single Node Computing Single Node Parallel Distributed Computing Computing Computing CPU GPU Grid Cluster Computing Computing A single node (usually multiple cores) Attached to a data store (Disc, SSD, …) One process with potentially multiple threads R: All processing is done on one computer BidMat: All processing is done on one computer with specialized HW Single Node In memory Retrieve/Stores from Disc Pros Simple to program and debug Cons Can only scale-up Does not deal with large data sets Single Node solution for large scale exploratory analysis Specialized HW and SW for efficient Matrix operations Elements: Data engine software for
    [Show full text]
  • Easybuild Documentation Release 20210907.0
    EasyBuild Documentation Release 20210907.0 Ghent University Tue, 07 Sep 2021 08:55:41 Contents 1 What is EasyBuild? 3 2 Concepts and terminology 5 2.1 EasyBuild framework..........................................5 2.2 Easyblocks................................................6 2.3 Toolchains................................................7 2.3.1 system toolchain.......................................7 2.3.2 dummy toolchain (DEPRECATED) ..............................7 2.3.3 Common toolchains.......................................7 2.4 Easyconfig files..............................................7 2.5 Extensions................................................8 3 Typical workflow example: building and installing WRF9 3.1 Searching for available easyconfigs files.................................9 3.2 Getting an overview of planned installations.............................. 10 3.3 Installing a software stack........................................ 11 4 Getting started 13 4.1 Installing EasyBuild........................................... 13 4.1.1 Requirements.......................................... 14 4.1.2 Using pip to Install EasyBuild................................. 14 4.1.3 Installing EasyBuild with EasyBuild.............................. 17 4.1.4 Dependencies.......................................... 19 4.1.5 Sources............................................. 21 4.1.6 In case of installation issues. .................................. 22 4.2 Configuring EasyBuild.......................................... 22 4.2.1 Supported configuration
    [Show full text]
  • Mochi-JCST-01-20.Pdf
    Ross R, Amvrosiadis G, Carns P et al. Mochi: Composing data services for high-performance computing environments. JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY 35(1): 121–144 Jan. 2020. DOI 10.1007/s11390-020-9802-0 Mochi: Composing Data Services for High-Performance Computing Environments Robert B. Ross1, George Amvrosiadis2, Philip Carns1, Charles D. Cranor2, Matthieu Dorier1, Kevin Harms1 Greg Ganger2, Garth Gibson3, Samuel K. Gutierrez4, Robert Latham1, Bob Robey4, Dana Robinson5 Bradley Settlemyer4, Galen Shipman4, Shane Snyder1, Jerome Soumagne5, and Qing Zheng2 1Argonne National Laboratory, Lemont, IL 60439, U.S.A. 2Parallel Data Laboratory, Carnegie Mellon University, Pittsburgh, PA 15213, U.S.A. 3Vector Institute for Artificial Intelligence, Toronto, Ontario, Canada 4Los Alamos National Laboratory, Los Alamos NM, U.S.A. 5The HDF Group, Champaign IL, U.S.A. E-mail: [email protected]; [email protected]; [email protected]; [email protected]; [email protected] E-mail: [email protected]; [email protected]; [email protected]; [email protected] E-mail: [email protected]; [email protected]; [email protected]; {bws, gshipman}@lanl.gov E-mail: [email protected]; [email protected]; [email protected] Received July 1, 2019; revised November 2, 2019. Abstract Technology enhancements and the growing breadth of application workflows running on high-performance computing (HPC) platforms drive the development of new data services that provide high performance on these new platforms, provide capable and productive interfaces and abstractions for a variety of applications, and are readily adapted when new technologies are deployed. The Mochi framework enables composition of specialized distributed data services from a collection of connectable modules and subservices.
    [Show full text]
  • Towards a Fully Automated Extraction and Interpretation of Tabular Data Using Machine Learning
    UPTEC F 19050 Examensarbete 30 hp August 2019 Towards a fully automated extraction and interpretation of tabular data using machine learning Per Hedbrant Per Hedbrant Master Thesis in Engineering Physics Department of Engineering Sciences Uppsala University Sweden Abstract Towards a fully automated extraction and interpretation of tabular data using machine learning Per Hedbrant Teknisk- naturvetenskaplig fakultet UTH-enheten Motivation A challenge for researchers at CBCS is the ability to efficiently manage the Besöksadress: different data formats that frequently are changed. Significant amount of time is Ångströmlaboratoriet Lägerhyddsvägen 1 spent on manual pre-processing, converting from one format to another. There are Hus 4, Plan 0 currently no solutions that uses pattern recognition to locate and automatically recognise data structures in a spreadsheet. Postadress: Box 536 751 21 Uppsala Problem Definition The desired solution is to build a self-learning Software as-a-Service (SaaS) for Telefon: automated recognition and loading of data stored in arbitrary formats. The aim of 018 – 471 30 03 this study is three-folded: A) Investigate if unsupervised machine learning Telefax: methods can be used to label different types of cells in spreadsheets. B) 018 – 471 30 00 Investigate if a hypothesis-generating algorithm can be used to label different types of cells in spreadsheets. C) Advise on choices of architecture and Hemsida: technologies for the SaaS solution. http://www.teknat.uu.se/student Method A pre-processing framework is built that can read and pre-process any type of spreadsheet into a feature matrix. Different datasets are read and clustered. An investigation on the usefulness of reducing the dimensionality is also done.
    [Show full text]
  • Enforcing Abstract Immutability
    Enforcing Abstract Immutability by Jonathan Eyolfson A thesis presented to the University of Waterloo in fulfillment of the thesis requirement for the degree of Doctor of Philosophy in Electrical and Computer Engineering Waterloo, Ontario, Canada, 2018 © Jonathan Eyolfson 2018 Examining Committee Membership The following served on the Examining Committee for this thesis. The decision of the Examining Committee is by majority vote. External Examiner Ana Milanova Associate Professor Rensselaer Polytechnic Institute Supervisor Patrick Lam Associate Professor University of Waterloo Internal Member Lin Tan Associate Professor University of Waterloo Internal Member Werner Dietl Assistant Professor University of Waterloo Internal-external Member Gregor Richards Assistant Professor University of Waterloo ii I hereby declare that I am the sole author of this thesis. This is a true copy of the thesis, including any required final revisions, as accepted by my examiners. I understand that my thesis may be made electronically available to the public. iii Abstract Researchers have recently proposed a number of systems for expressing, verifying, and inferring immutability declarations. These systems are often rigid, and do not support “abstract immutability”. An abstractly immutable object is an object o which is immutable from the point of view of any external methods. The C++ programming language is not rigid—it allows developers to express intent by adding immutability declarations to methods. Abstract immutability allows for performance improvements such as caching, even in the presence of writes to object fields. This dissertation presents a system to enforce abstract immutability. First, we explore abstract immutability in real-world systems. We found that developers often incorrectly use abstract immutability, perhaps because no programming language helps developers correctly implement abstract immutability.
    [Show full text]
  • A Dataset for Github Repository Deduplication
    A Dataset for GitHub Repository Deduplication Diomidis Spinellis Audris Mockus Zoe Kotti [email protected] {dds,zoekotti}@aueb.gr University of Tennessee Athens University of Economics and Business ABSTRACT select distinct p1, p2 from( select project_commits.project_id as p2, GitHub projects can be easily replicated through the site’s fork first_value(project_commits.project_id) over( process or through a Git clone-push sequence. This is a problem for partition by commit_id empirical software engineering, because it can lead to skewed re- order by mean_metric desc) as p1 sults or mistrained machine learning models. We provide a dataset from project_commits of 10.6 million GitHub projects that are copies of others, and link inner join forkproj.all_project_mean_metric each record with the project’s ultimate parent. The ultimate par- on all_project_mean_metric.project_id = ents were derived from a ranking along six metrics. The related project_commits.project_id) as shared_commits projects were calculated as the connected components of an 18.2 where p1 != p2; million node and 12 million edge denoised graph created by direct- Listing 1: Identification of projects with common commits ing edges to ultimate parents. The graph was created by filtering out more than 30 hand-picked and 2.3 million pattern-matched GitHub contains many millions of copied projects. This is a prob- clumping projects. Projects that introduced unwanted clumping lem for empirical software engineering. First, when data contain- were identified by repeatedly visualizing shortest path distances ing multiple copies of a repository are analyzed, the results can between unrelated important projects. Our dataset identified 30 end up skewed [27]. Second, when such data are used to train thousand duplicate projects in an existing popular reference dataset machine learning models, the corresponding models can behave of 1.8 million projects.
    [Show full text]
  • A Multilingual Information Extraction Pipeline for Investigative Journalism
    A Multilingual Information Extraction Pipeline for Investigative Journalism Gregor Wiedemann Seid Muhie Yimam Chris Biemann Language Technology Group Department of Informatics Universita¨t Hamburg, Germany gwiedemann, yimam, biemann @informatik.uni-hamburg.de { } Abstract 2) court-ordered revelation of internal communi- cation, 3) answers to requests based on Freedom We introduce an advanced information extrac- tion pipeline to automatically process very of Information (FoI) acts, and 4) unofficial leaks large collections of unstructured textual data of confidential information. Well-known exam- for the purpose of investigative journalism. ples of such disclosed or leaked datasets are the The pipeline serves as a new input proces- Enron email dataset (Keila and Skillicorn, 2005) sor for the upcoming major release of our or the Panama Papers (O’Donovan et al., 2016). New/s/leak 2.0 software, which we develop in To support investigative journalism in their cooperation with a large German news organi- zation. The use case is that journalists receive work, we have developed New/s/leak (Yimam a large collection of files up to several Giga- et al., 2016), a software implemented by experts bytes containing unknown contents. Collec- from natural language processing and visualiza- tions may originate either from official disclo- tion in computer science in cooperation with jour- sures of documents, e.g. Freedom of Informa- nalists from Der Spiegel, a large German news or- tion Act requests, or unofficial data leaks. Our ganization. Due to its successful application in the software prepares a visually-aided exploration investigative research as well as continued feed- of the collection to quickly learn about poten- tial stories contained in the data.
    [Show full text]
  • Introduction to Openrefine
    Introduction to OpenRefine 2018 Bibliometrics and Research Assessment Symposium Candace Norton, MLS © Library Carpentry: "OpenRefine Lessons for Librarians." Housekeeping • This session will be recorded and distributed after editing and captioning. • Slides will be available following the Symposium. • Questions are welcome throughout the presentation; please use the microphone. • Due to the set up of the breakout room, this is a lecture style course and will not be hands on. Course Objectives • Explain what OpenRefine does and why to use it. • Recognize when using OpenRefine is appropriate. • Understand how to import, edit, transform, and export data using OpenRefine. Introduction to OpenRefine • OpenRefine is a powerful tool for working with messy data • OpenRefine can clean data, transform data from one format to another, and can extend data via web services and external data • OpenRefine is available for download online at http://openrefine.org/download.html Why use OpenRefine? • Get an overview of a data set • Resolve inconsistencies in a data set – standardizing date formatting • Helps split data up into more granular parts – splitting up cells with multiple authors into separate cells • Match local data up to other data sets – matching local subjects against the Library of Congress Subject Headings • Enhance a data set with data from other sources Common Scenarios: Dates Data as entered Desired data 1st January 2014 2014-01-01 01/01/2014 2014-01-01 Jan 1 2014 2014-01-01 2014-01-01 2014-01-01 Common Scenarios: Names Data as entered
    [Show full text]
  • Issue Editor
    Bulletin of the Technical Committee on Data Engineering June 2018 Vol. 41 No. 2 IEEE Computer Society Letters Letter from the Editor-in-Chief . David Lomet 1 Letter from the Special Issue Editor . Guoliang Li 2 Special Issue on Large-scale Data Integration Data Integration: The Current Status and the Way Forward . Michael Stonebraker, Ihab F. Ilyas 3 BigGorilla: An Open-Source Ecosystem for Data Preparation and Integration. ........................... Chen Chen, Behzad Golshan, Alon Halevy, Wang-Chiew Tan, AnHai Doan 10 Self-Service Data Preparation: Research to Practice . Joseph M. Hellerstein, Jeffrey Heer, Sean Kandel 23 Toward a System Building Agenda for Data Integration (and Data Science) . ..........................AnHai Doan, Pradap Konda, Paul Suganthan G.C., Adel Ardalan, Jeffrey R. Ballard, Sanjib Das,Yash Govind, Han Li, Philip Martinkus, Sidharth Mudgal, Erik Paulson, Haojun Zhang 35 Explaining Data Integration . Xiaolan Wang, Laura Haas, Alexandra Meliou 47 Making Open Data Transparent: Data Discovery on Open Data . Renee´ J. Miller, Fatemeh Nargesian, Erkang Zhu, Christina Christodoulakis, Ken Q. Pu, Periklis Andritsos 59 Big Data Integration for Product Specifications . Luciano Barbosa, Valter Crescenzi, Xin Luna Dong, Paolo Merialdo, Federico Piai, Disheng Qiu, Yanyan Shen, Divesh Srivastava 71 Integrated Querying of SQL database data and S3 data in Amazon Redshift . Mengchu Cai, Martin Grund, Anurag Gupta, Fabian Nagel, Ippokratis Pandis, Yannis Papakonstantinou, Michalis Petropoulos 82 Robust Entity Resolution Using a CrowdOracle . .............................. Donatella Firmani, Sainyam Galhotra, Barna Saha, Divesh Srivastava 91 Human-in-the-loop Rule Learning for Data Integration . Ju Fan, Guoliang Li 104 Conference and Journal Notices ICDE 2019 Conference. 116 TCDE Membership Form . back cover Editorial Board TCDE Executive Committee Editor-in-Chief Chair David B.
    [Show full text]
  • Please Stand by for Realtime Captions. My Name Is Jamie Hayes I Am Your
    Please stand by for realtime captions. My name is Jamie Hayes I am your host for today's webinar. We also have hash Lee on tech support so if you're having any technical technical difficulties plea email Ashley. Are presented as Amy Solis and she is the librarian for social sciences data. Government documents and public administration at the University of Southern California USC. In Los Angeles California. She also serves as USC's official representative to the inter-Consortium for political and social research or IC PFR. She is responsible for overseeing data services in the USC libraries and works closely with university faculty, subject librarians and the referenced team to build and promote numeric data collections in support of data literacy, research and curriculum. In the social sciences. Before we get started I will walk you through a few housekeeping reminders. If you have any questions you would like to ask the presenter or if you have any technical issues please feel free to use the chat box. It is located in the bottom right-hand corner of your screen and I will keep track of all the questions that come in and at the end of the presentation Amy will respond to each of them. We are recording today's session and will email you a link of the recording and slides to everyone who registered for this webinar. We are also going to be sending you a certificate of participation using the email you used to register for today's webinar. If anyone needs additional certificates because multiple people are watching the webinar with you please email STL P [email protected] and include the title of today's webinar along with the names and email addresses of those needing certificates.
    [Show full text]
  • Mlsm: Making Authenticated Storage Faster in Ethereum
    mLSM: Making Authenticated Storage Faster in Ethereum Pandian Raju1 Soujanya Ponnapalli1 Evan Kaminsky1 Gilad Oved1 Zachary Keener1 Vijay Chidambaram1;2 Ittai Abraham2 1University of Texas at Austin 2VMware Research Abstract the LevelDB [15] key-value store. We show that reading a single key (e.g., the amount of ether in a given account) Ethereum provides authenticated storage: each read can result in 64 LevelDB reads, while writing a single returns a value and a proof that allows the client to verify key can lead to a similar number of LevelDB writes. In- the value returned is correct. We experimentally show ternally, LevelDB induces extra write amplification [23], that such authentication leads to high read and write am- further increasing overall amplification. Such write and × plification (64 in the worst case). We present a novel read amplification reduces throughput (storage band- data structure, Merkelized LSM (mLSM), that signifi- width is wasted by the amplification), and write ampli- cantly reduces the read and write amplification while still fication in particular significantly reduces the lifetime of allowing client verification of reads. mLSM significantly devices such as Solid State Drives (SSDs) which wear increases the performance of the storage subsystem in out after a limited number of write cycles [1, 16, 20]. Ethereum, thereby increasing the performance of a wide Thus, reducing the read and write amplification can both range of Ethereum applications. increase Ethereum throughput and reduce hardware re- 1 Introduction placement costs. Modern crypto-currencies such as Bitcoin [21] and We trace the read and write amplification in Ethereum Ethereum [26] seek to provide a decentralized, to the fact that it provides authenticated storage.
    [Show full text]