PMML: an Open Standard for Sharing Models

Total Page:16

File Type:pdf, Size:1020Kb

PMML: an Open Standard for Sharing Models 60 CONTRIBUTED RESEARCH ARTICLES PMML: An Open Standard for Sharing Models by Alex Guazzelli, Michael Zeller, Wen-Ching Lin and series, as well as different ways to explain models, Graham Williams including statistical and graphical measures. Ver- sion 4.0 also expands on existing PMML elements and their capabilities to further represent the diverse Introduction world of predictive analytics. The PMML package exports a variety of predic- tive and descriptive models from R to the Predic- tive Model Markup Language (Data Mining Group, PMML Structure 2008). PMML is an XML-based language and has become the de-facto standard to represent not only PMML follows a very intuitive structure to describe predictive and descriptive models, but also data pre- a data mining model. As depicted in Figure1, it and post-processing. In so doing, it allows for the is composed of many elements which encapsulate interchange of models among different tools and en- different functionality as it relates to the input data, vironments, mostly avoiding proprietary issues and model, and outputs. incompatibilities. The PMML package itself (Williams et al., 2009) was conceived at first as part of Togaware’s data min- ing toolkit Rattle, the R Analytical Tool To Learn Eas- ily (Williams, 2009). Although it can easily be ac- cessed through Rattle’s GUI, it has been separated from Rattle so that it can also be accessed directly in R. In the next section, we describe PMML and its overall structure. This is followed by a description of the functionality supported by the PMML pack- age and how this can be used in R. We then discuss the importance of working with a valid PMML file and finish by highlighting some of the debate sur- rounding the adoption of PMML by the data mining community at large. A PMML primer Developed by the Data Mining Group, an indepen- dent, vendor-led committee (http://www.dmg.org), PMML provides an open standard for representing data mining models. In this way, models can easily Figure 1: PMML overall structure described sequen- be shared between different applications, avoiding tially from top to bottom. proprietary issues and incompatibilities. Not only can PMML represent a wide range of statistical tech- niques, but it can also be used to represent their input Sequentially, PMML can be described by the fol- data as well as the data transformations necessary to lowing elements: transform these into meaningful features. PMML has established itself as the lingua franca for the sharing of predictive analytics solutions be- Header tween applications. This enables data mining scien- tists to use different statistical packages, including R, The header element contains general information to build, visualize, and execute their solutions. about the PMML document, such as copyright in- PMML is an XML-based language. Its current 3.2 formation for the model, its description, and infor- version was released in 2007. Version 4.0, the next mation about the application used to generate the version, which is to be released in 2009, will expand model such as name and version. It also contains an the list of available techniques covered by the stan- attribute for a timestamp which can be used to spec- dard. For example, it will offer support for time ify the date of model creation. The R Journal Vol. 1/1, May 2009 ISSN 2073-4859 CONTRIBUTED RESEARCH ARTICLES 61 Data dictionary – Outlier Treatment: Defines the outlier treat- ment to be used. In PMML, outliers can be The data dictionary element contains definitions for treated as missing values, as extreme val- all the possible fields used by the model. It is in the ues (based on the definition of high and data dictionary that a field is defined as continuous, low values for a particular field), or as is. categorical, or ordinal. Depending on this definition, the appropriate value ranges are then defined as well – Missing Value Replacement Policy: If this at- as the data type (such as string or double). The data tribute is specified then a missing value is dictionary is also used for describing the list of valid, automatically replaced by the given val- invalid, and missing values relating to every single ues. input data. – Missing Value Treatment: Indicates how the missing value replacement was derived Data transformations (e.g. as value, mean or median). Transformations allow for the mapping of user data • Targets: The targets element allows for the scal- into a more desirable form to be used by the mining ing of predicted variables. model. PMML defines several kinds of data transfor- mation: • Model Specifics: Once we have the data schema in place we can specify the details of the actual • Normalization: Maps values to numbers, the in- model. put can be continuous or discrete. A multi-layered feed-forward neural network, for example, is represented in PMML by its • Discretization: Maps continuous values to dis- activation function and number of layers, fol- crete values. lowed by the description of all the network • Value mapping: Maps discrete values to discrete components, such as connectivity and connec- values. tion weights. A decision tree is represented in PMML, recur- • Functions: Derive a value by applying a func- sively, by the nodes in the tree. For each node, tion to one or more parameters. a test on one variable is recorded and the sub- nodes then correspond to the result of the test. • Aggregation: Summarizes or collects groups of Each sub-node contains another test, or the fi- values. nal decision. The ability to represent data transformations (as well Besides neural networks and decision trees, as outlier and missing value treatment methods) in PMML allows for the representation of many conjunction with the parameters that define the mod- other data mining models, including linear re- els themselves is a key concept of PMML. gression, generalised linear models, random forests and other ensembles, association rules, Model cluster models, näive Bayes models, support vector machines, and more. The model element contains the definition of the data mining model. Models usually have a model PMML also offers many other elements such as name, function name (classification or regression) built-in functions, statistics, model composition and and technique-specific attributes. verification, etc. The model representation begins with a mining schema and then continues with the actual represen- tation of the model: Exporting PMML from R • Mining Schema: The mining schema (which is The PMML package in R provides a generic pmml embedded in the model element) lists all fields function to generate PMML 3.2 for an object. Using used in the model. This can be a subset of the an S3 generic function, the appropriate method for fields defined in the data dictionary. It contains the class of the supplied object is dispatched. specific information about each field, such as name and usage type. Usage type defines the An example way a field is to be used in the model. Typ- ical values are: active, predicted, and supple- A simple example illustrates the steps involved in mentary. Predicted fields are those whose val- generating PMML. We will build a decision tree, us- ues are predicted by the model. It is also in the ing rpart, to illustrate. With a standard installation mining schema that special values are treated. of R with the PMML package installed, the follow- These involve: ing should be easily repeatable. The R Journal Vol. 1/1, May 2009 ISSN 2073-4859 62 CONTRIBUTED RESEARCH ARTICLES First, we load the appropriate packages and <DataDictionary numberOfFields="5"> dataset. We will use the well known Iris dataset that <DataField name="Species" ... records sepal and petal characteristics for different <Value value="setosa"/> varieties of iris. <Value value="versicolor"/> <Value value="virginica"/> > library(pmml) <DataField name="Sepal.Length" > library(rpart) optype="continuous" > data(iris) dataType="double"/> </DataField> We can build a simple decision tree to classify ex- ... amples into iris varieties. We see the resulting struc- ture of the tree below. The root node test is on the The header and the data dictionary are common variable Petal.Length against a value of 2.45. The to all PMML, irrespective of the model. “left” branch tests for Petal.Length < 2.45 and deliv- Next we record the details of the model itself. In ers a decision of setosa. The “right” branch splits on a this case we have a tree model, and so a ‘TreeModel’ further variable, Petal.Width, to distinguish between element is defined. Within the tree model we begin versicolor and virginica. with the mining schema. > my.rpart <- rpart(Species ~ ., data=iris) <TreeModel modelName="RPart_Model" > my.rpart functionName="classification" algorithmName="rpart" n= 150 ...> <MiningSchema> node), split, n, loss, yval, (yprob) <MiningField name="Species" * denotes terminal node usageType="predicted"/> <MiningField name="Sepal.Length" 1) root 150 100 setosa ... usageType="active"/> 2) Petal.Length< 2.45 50 0 setosa (0.33 ... ... 3) Petal.Length>=2.45 100 50 versicolor ... This is followed by the actual nodes of the tree. 6) Petal.Width< 1.75 54 5 versicolor ... We can see that the first test involves a test on the 7) Petal.Width>=1.75 46 1 virginica ... Petal.Length. Once again, it is testing the value against 2.45. To convert this model to PMML we simply call the pmml function. Below we see some of the output. <Node id="1" score="setosa" The exported PMML begins with a header that recordCount="150" defaultChild="3"> contains information about the model. As indicated <True/> above, this will contain general information about <ScoreDistribution value="setosa" the model, including copyright, the application used recordCount="50" confidence="0.33"/> to build the model, and a timestamp. Other informa- <ScoreDistribution value="versicolor" tion might also be included.
Recommended publications
  • Licensing Open Government Data Jyh-An Lee
    Hastings Business Law Journal Volume 13 Article 2 Number 2 Winter 2017 Winter 2017 Licensing Open Government Data Jyh-An Lee Follow this and additional works at: https://repository.uchastings.edu/ hastings_business_law_journal Part of the Business Organizations Law Commons Recommended Citation Jyh-An Lee, Licensing Open Government Data, 13 Hastings Bus. L.J. 207 (2017). Available at: https://repository.uchastings.edu/hastings_business_law_journal/vol13/iss2/2 This Article is brought to you for free and open access by the Law Journals at UC Hastings Scholarship Repository. It has been accepted for inclusion in Hastings Business Law Journal by an authorized editor of UC Hastings Scholarship Repository. For more information, please contact [email protected]. 2 - LEE MACROED.DOCX (DO NOT DELETE) 5/5/2017 11:09 AM Licensing Open Government Data Jyh-An Lee* Governments around the world create and collect an enormous amount of data that covers important environmental, educational, geographical, meteorological, scientific, demographic, transport, tourism, health insurance, crime, occupational safety, product safety, and many other types of information.1 This data is generated as part of a government’s daily functions.2 Given government data’s exceptional social and economic value, former U.S. President Barack Obama described it as a “national asset.”3 For various policy reasons, open government data (“OGD”) has become a popular governmental practice and international * Assistant Professor at the Faculty of Law in the Chinese University
    [Show full text]
  • The Principles of Open Standards
    11/13/13 World Standards Day 1998 A newer version of this paper was published in The International Journal of IT Standards and Standardization Research, Vol. 4 No. 1, January - June 2006. This paper is the property of Standards Engineering Society. It was submitted to the World Standards Day paper competition for 1998, and won second place. It was published in Standards Engineering, Vol. 50, No. 6, November/December 1998, p. 1- 6. It is reprinted here with permission. The Principles of Open Standards by Ken Krechmer Communications Standards Review 757 Greer Road Palo Alto, CA USA 94303-3024 +1 650 856-8836 http://www.csrstds.com [email protected] "Who ever knew Truth put to the worse, in a free and open encounter?" 1 The values underlying Open Standards are impressive. In the USA the Open Standards concept has been applied to technical standards 2 since 1918 by the American National Standards Institute (ANSI) 3 as the administrator and coordinator of the United States private sector voluntary standardization system. A major goal of ANSI is to ensure that its guiding principles of openness, consensus and due process - the cornerstones of Open Standards - are followed. The personal computer revolution and the following Internet explosion have resulted in a large influx of new technical standards stakeholders.4 These new stakeholders are making new demands on the Open Standards processes. Today, considering the rapid changes in technology and the expanding importance of technical standards, ANSI and its Standards Development Organizations (SDOs) need to reexamine what constitutes the "free and open encounter" that Open Standards strive to achieve.
    [Show full text]
  • Applying Library Values to Emerging Technology Decision-Making in the Age of Open Access, Maker Spaces, and the Ever-Changing Library
    ACRL Publications in Librarianship No. 72 Applying Library Values to Emerging Technology Decision-Making in the Age of Open Access, Maker Spaces, and the Ever-Changing Library Editors Peter D. Fernandez and Kelly Tilton Association of College and Research Libraries A division of the American Library Association Chicago, Illinois 2018 The paper used in this publication meets the minimum requirements of Ameri- can National Standard for Information Sciences–Permanence of Paper for Print- ed Library Materials, ANSI Z39.48-1992. ∞ Cataloging-in-Publication data is on file with the Library of Congress. Copyright ©2018 by the Association of College and Research Libraries. All rights reserved except those which may be granted by Sections 107 and 108 of the Copyright Revision Act of 1976. Printed in the United States of America. 22 21 20 19 18 5 4 3 2 1 Contents Contents Introduction .......................................................................................................ix Peter Fernandez, Head, LRE Liaison Programs, University of Tennessee Libraries Kelly Tilton, Information Literacy Instruction Librarian, University of Tennessee Libraries Part I Contemplating Library Values Chapter 1. ..........................................................................................................1 The New Technocracy: Positioning Librarianship’s Core Values in Relationship to Technology Is a Much Taller Order Than We Think John Buschman, Dean of University Libraries, Seton Hall University Chapter 2. ........................................................................................................27
    [Show full text]
  • FOSS Philosophy 6 the FOSS Development Method 7
    1 Published by the United Nations Development Programme’s Asia-Pacific Development Information Programme (UNDP-APDIP) Kuala Lumpur, Malaysia www.apdip.net Email: [email protected] © UNDP-APDIP 2004 The material in this book may be reproduced, republished and incorporated into further works provided acknowledgement is given to UNDP-APDIP. For full details on the license governing this publication, please see the relevant Annex. ISBN: 983-3094-00-7 Design, layout and cover illustrations by: Rezonanze www.rezonanze.com PREFACE 6 INTRODUCTION 6 What is Free/Open Source Software? 6 The FOSS philosophy 6 The FOSS development method 7 What is the history of FOSS? 8 A Brief History of Free/Open Source Software Movement 8 WHY FOSS? 10 Is FOSS free? 10 How large are the savings from FOSS? 10 Direct Cost Savings - An Example 11 What are the benefits of using FOSS? 12 Security 13 Reliability/Stability 14 Open standards and vendor independence 14 Reduced reliance on imports 15 Developing local software capacity 15 Piracy, IPR, and the WTO 16 Localization 16 What are the shortcomings of FOSS? 17 Lack of business applications 17 Interoperability with proprietary systems 17 Documentation and “polish” 18 FOSS SUCCESS STORIES 19 What are governments doing with FOSS? 19 Europe 19 Americas 20 Brazil 21 Asia Pacific 22 Other Regions 24 What are some successful FOSS projects? 25 BIND (DNS Server) 25 Apache (Web Server) 25 Sendmail (Email Server) 25 OpenSSH (Secure Network Administration Tool) 26 Open Office (Office Productivity Suite) 26 LINUX 27 What is Linux?
    [Show full text]
  • Open Data Policy Purpose Definitions
    Open Data Policy OD-1 V1.0 February 1, 2016 City of Seattle Purpose The City of Seattle is committed to expanding both the data it makes available to the public and tools for understanding this data, and the Seattle Open Data Program has been created to realize these commitments. This Open Data Policy defines the principles governing City of Seattle Open Data and describes the expectations for department participation and governance of the Seattle Open Data Program. Definitions City of Seattle Data All data created, collected and/or maintained by the City of Seattle or by contractors or agencies on the City’s behalf. Open Data Specific datasets that are made available to the public by the City. Machine-Readable Any widely-accepted, nonproprietary, platform-independent, machine- readable method for formatting data (such as JSON, XML, and API’s) which permits automated processing of such data and facilitates search capabilities. Open Standard A technical standard developed and maintained by a voluntary consensus standards body that is available to the public without royalty or fee. Open Data Program dedicated to making City of Seattle data available to the public Program and engaging civic technologists, the research community, and other partners to make use of Open Data in support of the Program’s goals. Open Data Guide defining strategies City departments and offices can implement to Playbook making their data open, encourage public use consistent with the City’s privacy and security policies, and realize benefits for their departments. Open Data Portal Data.seattle.gov, the City’s catalog and primary repository for Open Data, created and maintained by the City for the express purpose of ensuring permanent, lasting open access to public information and Open Data Policy V1.0 February 1, 2016 1 enabling the development of innovative solutions that exemplify the goals of the Open Data Program.
    [Show full text]
  • Open Educational Resources (Oers)
    Open Educational Resources: Introduction Booklet and Webinar sensitize everyone. Open Educational Resources (OERs) Introduction Booklet and Webinar In this context, the Development Gateway Open Education Resources http://topics.developmentgateway.org/openeducation (OER) dgCommunity drives a Community Awareness & Sensitization project which includes a free Booklet and Webinar introducing the concept of Open Educational Resources and its potential to serve Education and Development programmes and practitioners, with a special emphasis on Developing Countries. The Booklet and the Webinar both provide: - a definition of Open Educational Resources, - an introduction to Open licenses and Standards, - an introduction to the Production and Distribution of OER, - a list of OER content repositories, search engines and projects. This project directly serves the OER dgCommunity main objective, which is to sensitize Development practitioners and citizens to Open Educational Resources, for the benefit of all. The Booklet and the Webinar are also to be considered as gateways for beginners to major OER projects, content repositories and search engines. An important part of its content introduces third part initiatives and portals in the domain of Open Drawings by Andrzej Krause. © 2004 BioMed Central, reproduced with permission. Education. Context The Webinar uses real-time Web conferencing tool, enabling trainees to During the past five years, thanks to innovative universities and interact with the trainer and to discuss together at the end of the session. educational projects from around the world, the Open Educational The associated Booklet is published under Creative Commons Attribution Resources (OER) movement has appeared and grown rapidly, addressing 3.0 license and can be used freely. major educational issues such as the “knowledge divide” and access for all to relevant information.
    [Show full text]
  • MASTER SPECIFICATIONS Division 25 – INTEGRATED AUTOMATION
    MASTER SPECIFICATIONS Division 25 – INTEGRATED AUTOMATION AND HVAC CONTROL Release 1.0 [March] 2017 Released by: Northwestern University Facilities Management Operations 2020 Ridge Avenue, Suite 200 Evanston, IL 60208-4301 All information within this Document is considered CONFIDENTIAL and PROPRIETARY. By receipt and use of this Document, the recipient agrees not to divulge any of the information herein and attached hereto to persons other than those within the recipients’ organization that have specific need to know for the purposes of reviewing and referencing this information. Recipient also agrees not to use this information in any manner detrimental to the interests of Northwestern University. Northwestern University Master Specifications Copyright © 2017 By Northwestern University These Specifications, or parts thereof, may not be reproduced in any form without the permission of the Northwestern University. NORTHWESTERN UNIVERSITY PROJECT NAME ____________ FOR: ___________ JOB # ________ ISSUED: 03/29/2017 MASTER SPECIFICATIONS: DIVISION 25 – INTEGRATED AUTOMATION AND HVAC CONTROL SECTION # TITLE 25 0000 INTEGRATED AUTOMATION AND HVAC TEMPERATURE CONTROL 25 0800 COMMISSIONING OF INTEGRATED AUTOMATION ** End of List ** NORTHWESTERN UNIVERSITY PROJECT NAME ____________ FOR: ___________ JOB # ________ ISSUED: 03/29/2017 SECTION 25 0000 - INTEGRATED AUTOMATION AND HVAC TEMPERATURE CONTROL PART 1 - GENERAL 1.1 SUMMARY AND RESPONSIBILITIES A. The work covered under this Division 25 of the project specifications is basically for: 1. Providing fully functioning facility/project HVAC system DDC controls. 2. Controlling systems locally, with communication up to the Tridium integration platform for remote monitoring, trending, control, troubleshooting and adjustment by NU Engineering and Maintenance personnel. 3. Tying specified systems to the Northwestern University (NU) integration platform for remote monitoring, trending, control, management, troubleshooting and adjustment by NU Engineering and Maintenance personnel.
    [Show full text]
  • Open Standards - Definitions and Background
    Open Standards - Definitions and Background Open Standards Standards (in the context of technology) are a set of definitions, rules, or specifications which allow interoperability. Standards can be open or closed. Open standards are distinguished from closed standards in that anyone is allowed to know and use an open standard while closed standards are kept secret and/or have significant encumbrances upon their use. Open standards are typically arrived at through negotiation and consensus within the com- munity affected. The Wikipedia article on open standards contains additional accurate information: http://en.wikipedia.org/wiki/Open standards “Open standards are publicly available specifications for achieving a specific task. By allow- ing anyone to use the standard, they increase compatibility between various hardware and software components, since anyone with the technical know-how and the necessary equipment to implement solutions can build something that works together with those of other vendors.” Occassions when a choice between open/closed standard occurs You wish to distribute a document to a large group of collegues. You select the format for this file. If you select an open standard (eg ascii/txt, rtf, pdf, odf), it is likely that all users will have an appropriate viewing application. If you choose a closed standard (eg MS doc), it is likely that a fraction of your collegues will not be able to view your file. In particular, Microsoft does not have a version of Word for Linux and has gone back and forth on whether it will provide a version for Macs (the current plan is to support macs).
    [Show full text]
  • Memory-Driven Computing Kimberly Keeton Distinguished Technologist
    Memory-Driven Computing Kimberly Keeton Distinguished Technologist Non-Volatile Memories Workshop (NVMW) 2019 – March 2019 Need answers quickly and on bigger data Data growth 180 175 160 140 130 120 100 100 Projected 80 80 Historical 60 63 50 40 33 41 25 19 15 20 9 12 of analyzed data ($) Value Global datasphere (zettabytes) 5 0.3 1 3 0 2005 2010 2015 2020 2025 10-2 100 102 104 106 Data nearly doubles every two years (2013-25) Time to result (seconds) Source: IDC Data Age 2025 study, sponsored by Seagate, Nov 2018 ©Copyright 2019 Hewlett Packard Enterprise Company 2 What’s driving the data explosion? Record Electronic record of event Ex: banking Mediated by people Structured data ©Copyright 2019 Hewlett Packard Enterprise Company 3 What’s driving the data explosion? Record Engage Electronic record of event Interactive apps for humans Ex: banking Ex: social media Mediated by people Interactive Structured data Unstructured data ©Copyright 2019 Hewlett Packard Enterprise Company 4 What’s driving the data explosion? Record Engage Act Electronic record of event Interactive apps for humans Machines making decisions Ex: banking Ex: social media Ex: smart and self-driving cars Mediated by people Interactive Real time, low latency Structured data Unstructured data Structured and unstructured data ©Copyright 2019 Hewlett Packard Enterprise Company 5 More data sources and more data Record Engage Act 40 petabytes 4 petabytes a day 40,000 petabytes a day* 200B rows of recent Posted daily by Facebook’s 4TB daily per self-driving car transactions for
    [Show full text]
  • Open Standards Principles
    Open Standards Principles: For software interoperability, data and document formats in government IT specifications ii OPEN STANDARDS PRINCIPLES: For software interoperability, data and document formats in government IT specifications #opengovIT iii iv Foreword Government IT must be open - open to the people and organisations that use our services and open to any provider, regardless of their size. We currently have many small, separate platforms operating across disconnected departments and IT that is tied into monolithic contracts. We need to have a platform for government that allows us to share appropriate data effectively and that gives us flexibility and choice. Already we have started to make progress and there is activity right across government that demonstrates our commitment to achieving the change we need. The Government Digital Service has launched GOV.UK - a simpler, clearer, faster and common platform for government's digital services, based on open source software. The whole site is designed around the needs of the user. GDS has also developed the online petitions system for the UK Parliament to enable people to raise, sign and track petitions. It was developed by SMEs and is built on open standards. Another site legislation.gov.uk – the online home of legislation – was designed around open standards so more participants could get involved. It connects with inhouse, public and private sector systems and has allowed The National Archives to develop an entirely new operating model for revising legislation. The private sector is investing effort in improving the underlying data. Companies then re- use the source code for processing the data and are including this in their own commercial products and services.
    [Show full text]
  • Open Content Licensing
    Open Content Licensing Open Content Licensing From Theory to Practice Edited by Lucie Guibault & Christina Angelopoulos Amsterdam University Press Cover design: Kok Korpershoek bno, Amsterdam Lay-out: JAPES, Amsterdam ISBN 978 90 8964 307 0 e-ISBN 978 90 4851 408 3 NUR 820 Creative Commons License CC BY NC (http://creativecommons.org/licenses/by-nc/3.0) L. Guibault & C. Angelopoulos / Amsterdam University Press, Amsterdam, 2011 Some rights reversed. Without limiting the rights under copyright reserved above, any part of this book may be reproduced, stored in or introduced into a retrieval system, or transmitted, in any form or by any means (electronic, mechanical, photocopying, recording or otherwise). Contents 1. Open Content Licensing: From Theory to Practice – An Introduction 7 Lucie Guibault, Institute for Information Law, University of Amsterdam 2. Towards a New Social Contract: Free-Licensing into the Knowledge Commons 21 Volker Grassmuck, Humboldt University Berlin and University of Sao Paulo 3. Is Open Content a Victim of its Own Success? Some Economic Thoughts on the Standardization of Licenses 51 Gerald Spindler and Philipp Zimbehl, University of Göttingen 4. (Re)introducing Formalities in Copyright as a Strategy for the Public Domain 75 Séverine Dusollier, Centre de Recherche Informatique et Droit, Université Notre- Dame de la Paix (Namur) 5. User-Related Assets and Drawbacks of Open Content Licensing 107 Till Kreutzer, Institute for legal questions on Free and Open Source Software (ifrOSS) 6. Owning the Right to Open Up Access to Scientific Publications 137 Lucie Guibault, Institute for Information Law, University of Amsterdam 7. Friends or Foes? Creative Commons, Freedom of Information Law and the European Union Framework for Reuse of Public Sector Information 169 Mireille van Eechoud, Institute for Information Law, University of Amsterdam 8.
    [Show full text]
  • An Economic Basis for Open Standards Flosspols.Org
    An Economic Basis for Open Standards flosspols.org An Economic Basis for Open Standards Rishab A. Ghosh Maastricht, December 2005 This report was supported by the FLOSSPOLS project, funded under the Sixth Framework Programme of the European Union, managed by the eGovernment Unit of the European Commission's DG Information Society. Rishab Ghosh ([email protected]) is Programme Leader, FLOSS and Senior Researcher at the United Nations University Maastricht Economic and social Research and training centre on Innovation and Technology (UNU- MERIT), located at the University of Maastricht, The Netherlands Copyright © MERIT, University of Maastricht 1 An Economic Basis for Open Standards flosspols.org An Economic Basis for Open Standards Executive Summary This paper provides an overview of standards and standard-setting processes. It describes the economic effect of technology standards – de facto as well as de jure – and differentiates between the impact on competition and welfare that various levels of standards have. It argues that most of what is claimed for “open standards” in recent policy debates was already well encompassed by the term “standards”; a different term is needed only if it is defined clearly in order to provide a distinct economic effect. This paper argues that open standards, properly defined, can have the particular economic effect of allowing “natural” monopolies to form in a given technology, while ensuring full competition among suppliers of that technology. This is a distinct economic effect that deserves to be distinguished by the use of a separate term, hence “open” rather than “ordinary” standards – referred to as “semi-open” in this paper.
    [Show full text]