Implementation of a Linked Open Data Solution for the Statistics Agency of Cantabria's Metadata and Data Bank

Total Page:16

File Type:pdf, Size:1020Kb

Implementation of a Linked Open Data Solution for the Statistics Agency of Cantabria's Metadata and Data Bank Proc. Int’l Conf. on Dublin Core and Metadata Applications 2013 Implementation of a Linked Open Data Solution for the Statistics Agency of Cantabria's Metadata and Data Bank Alejandro Villar Fernández Ana Santurtún Zarrabeitia Spain University of Cantabria, contacto@alejandro- Spain villar.es [email protected] Abstract Statistics is a fundamental piece inside the Open Government philosophy, being a basic tool for citizens to know and make informed decisions about the society in which they participate. Due to the great number of organizations and agencies that collect, process and publish statistical data all over the world, several standards and methodologies for information exchange have been created in recent years in order to improve interoperability between data producers and consumers, of which SDMX is one of the most renowned examples. Despite having been developed independently of this, the global Semantic Web effort (backed mainly by the W3C-driven Linked Open Data initiatives) presents itself as an extremely useful tool for publishing both completely contextualized metadata and data, therefore making them easily understandable and processable by third parties. This report details the changes made to the IT systems of the Statistical Agency of Cantabria (Instituto Cántabro de Estadística, ICANE) with the purpose of implementing a Linked Open Data solution for its website and statistical data bank, making all data and metadata published by this Agency available not only to humans, but to automatized consumers, too. Multiple standards, recommendations and vocabularies were used for this task, ranging from Dublin Core metadata RDFa tagging, through the creation of several SKOS concept schemes, to providing statistical data using the RDF Data Cube vocabulary. Keywords: Linked Open Data; statistics; ICANE; Spain; Cantabria; SKOS; RDF; RDFa; Semantic Web. 1. Background Legally created in 1998, but not officially operative until 2004, the Statistics Agency of Cantabria (Instituto Cántabro de Estadística, ICANE) is the official public organization in Cantabria (Northern Spain) in charge of the production and dissemination of statistical data regarding every aspect of Cantabria's economy and society. Despite its young age, ICANE's face to the public has been in constant evolution, from its inception as a static web site populated with Microsoft Excel files, to its current state of a machine-navigable, metadata-annotated portal serving on-the-fly generated data in 7 different formats. 1.1 Initial Situation and Motivation ICANE's web portal and data bank underwent a major visual and functional overhaul in 2011, 1 2 replacing a legacy OpenCMS installation with Liferay Enterprise Portal 5.2, and an in-house- developed web application for data selection and display with a Pentaho Mondrian3 3.1 + 1 http://www.opencms.org/ 2 http://www.liferay.com/es/ 3 http://mondrian.pentaho.com/ 1 Proc. Int’l Conf. on Dublin Core and Metadata Applications 2013 JPivot41.8 combination, which also involved converting its previous data warehouse to an OLAP5-compatible setup. On the second half of 2012, ICANE decided to implement a solution for improving its data and metadata availability and to publish and link its resources on the Semantic Web. At the time, all internal, structural metadata about data series (OLAP cubes, dimensions, measures, etc.) was stored in XML Schema files and deployed onto the data bank application servers, and all descriptive metadata (including hierarchical organization into folders and sections) resided in a relational database, which was used by the web portal and by the data bank application to present a hierarchically navigable view to the end user. This setup relied on the following domain model classes to represent navigational and descriptive metadata: • Area: subject area for a group of series. • Type: type of series (time series, historical data or municipal data). • Group: series group (subject area subdivision), belonging to an area and a type. • Nodes: hierarchical layout of folders and time series. Includes all descriptive metadata for every series and folder. Most series metadata (modification dates, periodicity, data sources, etc.) had not been yet normalized. In addition, no URL path hierarchy was used when accessing the different sections of the web portal or data bank application (with the latter relying on user-provided HTTP GET parameters for time series selection and descriptive metadata retrieval). 2. Metadata Reorganization, Normalization and Centralization A decision was made to reorganize and normalize all database-stored metadata. The first step was dropping the old area/group/type schema in favor of a section/subsection/category one so that: • Four main sections (economy, population, society, territory and environment) would exist. • Each section would contain several, more specific subsections, which would not be dependent on type or category. • Inside each subsection the end user would find up to three different series/folder trees, one for each category. Moreover, the following efforts were made, with regards to metadata normalization and preparation for the Linked Open Data implementation: • URI tags (friendly URL components) were defined for all user-referenceable classes. A concrete URL hierarchy based on these tags was laid out to make it easier for users to access and understand the web portal and data bank application layouts. More information on the use of URI tags can be found in section 3.2. • Folder and series metadata were normalized whenever possible. This included series periodicities (which were linked to their DC CDF Vocabulary, 2007, equivalent) and initial and final periods, data sources, reference areas, etc. • A link repository was created to store all entities' semantic relationships towards external resources. Finally, ICANE's IT department developed a Web Service to provide centralized access to this newly-created metadata database through a REST API, and avoid directly querying the database server containing it, thus protecting the clients from the specific implementation details. This Web Service was to be used by all internal applications, if possible. 4 http://jpivot.sourceforge.net/ 5 Online Analytical Processing. An introduction to OLAP can be found in Pedersen, T.B (2001). 2 Proc. Int’l Conf. on Dublin Core and Metadata Applications 2013 3. Proposed Solution Alejandro Villar’s proposal consisted of 5 different, but interdependent, phases: 1. Definition of a RDF model for all publicly accessible entities, centered around the DCMI model. 2. Implementation of a Linked Open Data strategy to serve entity metadata. 3. Development of an RDF export filter for the data bank application. 4. Implementation of a SPARQL endpoint to access metadata. 5. Semantic linking of entities to external resources, in order to provide context to automatized data consumers. 3.1. RDF Model In addition to property mapping, the definition of the RDF model had to take into consideration the hierarchical nature of the time series and folders tree. Therefore, a SKOS ontology was conceived, comprising several Concept Schemes (one for each section and for each subsection), which would in turn contain a hierarchical array of folders (represented by SKOS Concepts). Time series and folders would be linked using the DCMI subject property. Given that SKOS does not provide an explicit mechanism to define a hierarchy between Concept Schemes, its inScheme property (whose domain is effectively any RDF resource) was used to this effect. A property summary of this model, using CURIE syntax, can be seen in Table 1, while table 2 contains a list of all prefixes used throughout this project report. In addition, a special ICANE vocabulary6 was created to enhance entity classification and to provide additional means of entity interlinking. TABLE 1: Property summary of the RDF model Entity Property Description Class of this resource. Section rdf:type Values: icane:Section and A section has many subsections. skos:ConceptScheme. skos:prefLabel and rdfs:label Resource label. rdf:type Class of this resource. Subsection Values: icane:Subsection and A subsection belongs to a single skos:ConceptScheme. section and has many folders and skos:prefLabel and rdfs:label Resource label. time series. icane:section and skos:inScheme Section that this subsection belongs to. rdf:type Class of this resource. Category Value: icane:Category. A category has many folders and rdfs:label Resource label. time series. icane:acronym Acronym for this category. Class of this resource. rdf:type Value: skos:Concept. Folder skos:prefLabel and rdfs:label Resource label. A folder is a branch node in the icane:section Section that this folder belongs to. subject matter tree. It has many icane:subsection Subsection that this folder belongs to. child folders and time series, and it Concept Schemes that this folder belongs to one subsection (and skos:inScheme therefore to one section) and to one belongs to (Its section and subsection). category. icane:category Category that this folder belongs to. Used for creating a hierarchical structure skos:broader / skos:narrower of folders. Table continued 6 http://www.icane.es/opendata/vocab 3 Proc. Int’l Conf. on Dublin Core and Metadata Applications 2013 Entity Property Description Class of this resource. rdf:type Values: icane:TimeSeries and qb:DataSet. rdfs:label and dcterms:title Resource label. Subject of this resource, which will be its dcterms:subject parent folder in the hierarchy. icane:section
Recommended publications
  • Metadata for Semantic and Social Applications
    etadata is a key aspect of our evolving infrastructure for information management, social computing, and scientific collaboration. DC-2008M will focus on metadata challenges, solutions, and innovation in initiatives and activities underlying semantic and social applications. Metadata is part of the fabric of social computing, which includes the use of wikis, blogs, and tagging for collaboration and participation. Metadata also underlies the development of semantic applications, and the Semantic Web — the representation and integration of multimedia knowledge structures on the basis of semantic models. These two trends flow together in applications such as Wikipedia, where authors collectively create structured information that can be extracted and used to enhance access to and use of information sources. Recent discussion has focused on how existing bibliographic standards can be expressed as Semantic Metadata for Web vocabularies to facilitate the ingration of library and cultural heritage data with other types of data. Harnessing the efforts of content providers and end-users to link, tag, edit, and describe their Semantic and information in interoperable ways (”participatory metadata”) is a key step towards providing knowledge environments that are scalable, self-correcting, and evolvable. Social Applications DC-2008 will explore conceptual and practical issues in the development and deployment of semantic and social applications to meet the needs of specific communities of practice. Edited by Jane Greenberg and Wolfgang Klas DC-2008
    [Show full text]
  • Provenance and Annotations for Linked Data
    Proc. Int’l Conf. on Dublin Core and Metadata Applications 2013 Provenance and Annotations for Linked Data Kai Eckert University of Mannheim, Germany [email protected] Abstract Provenance tracking for Linked Data requires the identification of Linked Data resources. Annotating Linked Data on the level of single statements requires the identification of these statements. The concept of a Provenance Context is introduced as the basis for a consistent data model for Linked Data that incorporates current best-practices and creates identity for every published Linked Dataset. A comparison of this model with the Dublin Core Abstract Model is provided to gain further understanding, how Linked Data affects the traditional view on metadata and to what extent our approach could help to mediate. Finally, a linking mechanism based on RDF reification is developed to annotate single statements within a Provenance Context. Keywords: Provenance; Annotations; RDF; Linked Data; DCAM; DM2E; 1. Introduction This paper addresses two challenges faced by many Linked Data applications: How to provide, access, and use provenance information about the data; and how to enable data annotations, i.e., further statements about the data, subsets of the data, or even single statements. Both challenges are related as both require the existence of identifiers for the data. We use the Linked Data infrastructure that is currently developed in the DM2E project as an example with typical use- cases and resulting requirements. 1.1. Foundations Linked Data, the publication of data on the Web, enables easy access to data and supports the reuse of data. The Hypertext Transfer Protocol (HTTP) is used to access a Uniform Resource Identifier (URI) and to retrieve data about the resource.
    [Show full text]
  • Rdfa in XHTML: Syntax and Processing Rdfa in XHTML: Syntax and Processing
    RDFa in XHTML: Syntax and Processing RDFa in XHTML: Syntax and Processing RDFa in XHTML: Syntax and Processing A collection of attributes and processing rules for extending XHTML to support RDF W3C Recommendation 14 October 2008 This version: http://www.w3.org/TR/2008/REC-rdfa-syntax-20081014 Latest version: http://www.w3.org/TR/rdfa-syntax Previous version: http://www.w3.org/TR/2008/PR-rdfa-syntax-20080904 Diff from previous version: rdfa-syntax-diff.html Editors: Ben Adida, Creative Commons [email protected] Mark Birbeck, webBackplane [email protected] Shane McCarron, Applied Testing and Technology, Inc. [email protected] Steven Pemberton, CWI Please refer to the errata for this document, which may include some normative corrections. This document is also available in these non-normative formats: PostScript version, PDF version, ZIP archive, and Gzip’d TAR archive. The English version of this specification is the only normative version. Non-normative translations may also be available. Copyright © 2007-2008 W3C® (MIT, ERCIM, Keio), All Rights Reserved. W3C liability, trademark and document use rules apply. Abstract The current Web is primarily made up of an enormous number of documents that have been created using HTML. These documents contain significant amounts of structured data, which is largely unavailable to tools and applications. When publishers can express this data more completely, and when tools can read it, a new world of user functionality becomes available, letting users transfer structured data between applications and web sites, and allowing browsing applications to improve the user experience: an event on a web page can be directly imported - 1 - How to Read this Document RDFa in XHTML: Syntax and Processing into a user’s desktop calendar; a license on a document can be detected so that users can be informed of their rights automatically; a photo’s creator, camera setting information, resolution, location and topic can be published as easily as the original photo itself, enabling structured search and sharing.
    [Show full text]
  • CODATA Workshop on Big Data Programme Book
    Sponsor I CODACODATA S UU Co-Sponsors Organizer Workshop on Big Data for International Scientific Programmes CONTENTS I. Sponsoring Organizations International Council for Science (ICSU) I. Sponsoring Organizations 2 The International Council for Science (ICSU) is a the international scientific community to II. Programme 7 non-governmental organization with a global strengthen international science for the benefit of membership of national scientific bodies society. (121members, representing 141 countries) and international scientific unions (31 members). ICSU: www.icsu.org III. Remarks and Abstracts 13 ICSU mobilizes the knowledge and resources of Committee on Data for Science and Technology (CODATA) IV. Short Biography of Speakers 28 I CODACODATA S UU V. Conference Venue Layout 41 CODATA, the ICSU Committee on Data for Science challenges and ‘hot topics’ at the frontiers and Technology, was established in 1966 to meet of data science (through CODATA Task a need for an international coordinating body to Groups and Working Groups and other improve the management and preservation of initiatives). VI. General Information 43 scientific data. CODATA has been at the forefront 3. Developing data strategies for international of data science and data policy issues since that science programmes and supporting ICSU date. activities such as Future Earth and Integrated About Beijing 43 Research on Disaster Risk (IRDR) to address CODATA supports ICSU’s mission of ‘strengthening data management needs. international science for the benefit of society’ by ‘promoting improved scientific and technical data Through events like the Workshop on Big Data for About the Workshop Venue 43 management and use’. CODATA achieves this International Scientific Programmes and mission through three strands of activity: SciDataCon 2014, CODATA collaborates with 1.
    [Show full text]
  • Introduction to Ontology- Based Semantics Goals Service
    Goals • To provide some insight into the usefulness of ontologies Introduction to Ontology- • To provide an understanding of the based Semantics features of RDFS and OWL and their use in automated reasoning Semantic Web, ontologies, RDF, OWL, • To provide an outline of the N3 syntax N3 • Use N3 to express RDFS and OWL – sufficient for later examples and exercises on service semantics With thanks to Declan O’Sullivan @ David Lewis @ David Lewis Service Semantics Functional Semantics • WSDL provides the syntax we need for interoperating with a service, but little in the way of semantics • What do ‘origin’ and ‘destination’ strings • Examining this example raises many questions about represent? functional and non-functional semantics – Country, city, airport, restrictions (airline, national, <message name=“getcheapestFlightRequest"> regional)? <part name=“origin" type="xsd:string"/> <part name=“destination" type="xsd:string"/> • What does ‘flight’ string represent? <part name=“date" type="xsd:date"/> – Airline, flight number, plane type? </message> <message name=“getcheapestFlightResponse"> • What does ‘time’ string represent? <part name=“flight" type="xsd:string"/> <part name=“time" type="xsd:time"/> – Departure time? <part name=“cost” type=“xsd:float”/> </message> – Note xsd:time probably is adequate - supports time- <portType name=“cheapestFlight"> zone information <operation name="getCheapestFlight"> <input message=“getcheapestFlightRequest"/> • What does ‘cost’ float represent? <output message=“getcheapestFlightResponse"/> </operation>
    [Show full text]
  • Folksonomies - Cooperative Classification and Communication Through Shared Metadata
    Folksonomies - Cooperative Classification and Communication Through Shared Metadata Adam Mathes Computer Mediated Communication - LIS590CMC Graduate School of Library and Information Science University of Illinois Urbana-Champaign December 2004 Abstract This paper examines user-generated metadata as implemented and applied in two web services designed to share and organize digital me- dia to better understand grassroots classification. Metadata - data about data - allows systems to collocate related information, and helps users find relevant information. The creation of metadata has generally been approached in two ways: professional creation and author creation. In li- braries and other organizations, creating metadata, primarily in the form of catalog records, has traditionally been the domain of dedicated profes- sionals working with complex, detailed rule sets and vocabularies. The primary problem with this approach is scalability and its impracticality for the vast amounts of content being produced and used, especially on the World Wide Web. The apparatus and tools built around professional cataloging systems are generally too complicated for anyone without spe- cialized training and knowledge. A second approach is for metadata to be created by authors. The movement towards creator described docu- ments was heralded by SGML, the WWW, and the Dublin Core Metadata Initiative. There are problems with this approach as well - often due to inadequate or inaccurate description, or outright deception. This paper examines a third approach: user-created metadata, where users of the documents and media create metadata for their own individual use that is also shared throughout a community. 1 The Creation of Metadata: Professionals, Con- tent Creators, Users Metadata is often characterized as “data about data.” Metadata is information, often highly structured, about documents, books, articles, photographs, or other items that is designed to support specific functions.
    [Show full text]
  • Extending the Role of Metadata in a Digital Library System*
    Extending the Role of Metadata in a Digital Library System* Alexa T. McCray, Marie E. Gallagher, Michael A. Flannick National Library of Medicine Bethesda, MD, USA {mccray,gallagher,flannick}@nlm.nih.gov Abstract Metadata efforts often fall into the trap of We describe an approach to the development of a digital trying to create a universal metadata schema. library system that is founded on a number of basic Such efforts fail to recognize the basic nature principles. In particular, we discuss the critical role of of metadata: namely, that it is far too diverse metadata in all aspects of the system design. We begin by to fit into one useful taxonomy...the creation, describing how the notion of metadata is sometimes administration, and enhancement of individual interpreted and go on to discuss some of our early metadata forms should be left to the relevant experiences in a digital conversion project. We report on communities of expertise. Ideally this would the Profiles in Science project, which is making the occur within a framework that will support archival collections of prominent biomedical scientists interoperability across data and domains. available on the World Wide Web. We discuss the [4:277] principles that are used in our system design, illustrating Roszkowski and Lukas describe an approach for these throughout the discussion. Our approach has linking distributed collections of metadata so that they are involved interpreting metadata in its broadest sense. We searchable as a single collection [5], and Baldonado et al. capture data about the items in our digital collection for a describe an architecture that facilitates metadata a variety of purposes and use those data to drive the compatibility and interoperability [6].
    [Show full text]
  • A Blockchain-Based Application to Protect Minor Artworks
    A Blockchain-based Application to Protect Minor Artworks Clara Bacciu, Angelica Lo Duca, Andrea Marchetti IIT-CNR - via G. Moruzzi 1, 56124 Pisa (Italy) [name].[surname]@iit.cnr.it Keywords: blockchain, Ethereum, recordkeeping, IPFS, Cultural Heritage, minor artwork Abstract: A new emerging trend concerns the implementation of services and distributed applications through the blockchain technology. A blockchain is an append-only database, which guarantees security, transparency and immutability of records. Blockchains can be used in the field of Cultural Heritage to protect minor artworks, i.e. artistic relevant works not as famous as masterpieces. Minor artworks are subjected to counterfeiting, thefts and natural disasters because they are not well protected as famous artworks. This paper describes a blockchain-based application, called MApp (Minor Artworks application), which lets authenticated users (pri- vate people or organizations), store the information about their artworks in a secure way. The use of blockchain produces three main advantages. Firstly, artworks cannot be deleted from the register thus preventing thieves to remove records associated stolen objects. Secondly, artworks can be added and updated only by authorized users, thus preventing counterfeiting in objects descriptions. Finally, records can be used to keep artworks memory in case of destruction caused by a natural disaster. 1 INTRODUCTION cessible. To complicate things, in the last decades, Italy (as many other places in the world) has suffered Minor artworks are works that are artistically rel- a series of catastrophic events that largely affected evant but not as well-known as famous masterpieces, cultural heritage: earthquakes, landslides, floods, col- or belonging to the so-called minor arts, such as books lapses in important archaeological sites such as Pom- and manuscripts, pottery, lacquerware, furniture, jew- pei.
    [Show full text]
  • Integrating Dublin Core Metadata for Cultural Heritage Collections Using Ontologies**
    2007 Proc. Int’l Conf. on Dublin Core and Metadata Applications Integrating Dublin Core metadata for cultural heritage ** collections using ontologies Constantia Kakali Irene Lourdi Thomais Stasinopoulou Department of Archive and Department of Archive and Department of Archive and Library Sciences, Ionian Library Sciences, Ionian Library Sciences, Ionian University, Greece University, Greece University, Greece [email protected] [email protected] [email protected] Lina Bountouri Christos Papatheodorou Martin Doerr Department of Archive and Department of Archive and Institute of Computer Library Sciences, Ionian Library Sciences, Ionian Science, Foundation for University, Greece University, Greece Research and Technology, [email protected] [email protected] Greece [email protected] Manolis Gergatsoulis Department of Archive and Library Sciences, Ionian University, Greece [email protected] Abstract Metadata interoperability is an active research area, especially for cultural heritage collections, which consist of heterogeneous objects described by a variety of metadata schemas. In this paper we propose an ontology-based metadata interoperability approach, which exploits, in an optimal way, the semantics of metadata schemas. In particular, we propose the use of CIDOC/CRM ontology as a mediating schema and present a methodology for mapping DC Type Vocabulary to CIDOC/CRM, demonstrating a real-world effort for ontology-based metadata integration. Keywords: ontology-based integration; metadata interoperability; CIDOC/CRM; DC-Type. 1. Introduction Heterogeneity is one of the main characteristics of cultural heritage collections. Such collections may be composed of text, written on different materials, paintings, photographs, 3D objects, sound recordings, maps or even digital objects. Furthermore, the objects are strongly related with the social and historical events that take place over time.
    [Show full text]
  • Publishing E-Resources of Digital Institutional Repository As Linked Open Data: an Experimental Study
    University of Nebraska - Lincoln DigitalCommons@University of Nebraska - Lincoln Library Philosophy and Practice (e-journal) Libraries at University of Nebraska-Lincoln 11-30-2020 Publishing E-resources of Digital Institutional Repository as Linked Open Data: an experimental study Subhendu Kar Vivekananda Satavarshiki Mahavidyalaya, [email protected] Dr. Rajesh Das The University of Burdwan, [email protected] Follow this and additional works at: https://digitalcommons.unl.edu/libphilprac Part of the Library and Information Science Commons Kar, Subhendu and Das, Dr. Rajesh, "Publishing E-resources of Digital Institutional Repository as Linked Open Data: an experimental study" (2020). Library Philosophy and Practice (e-journal). 4699. https://digitalcommons.unl.edu/libphilprac/4699 Publishing E-resources of Digital Institutional Repository as Linked Open Data: an experimental study Subhendu Kar, Librarian, Vivekananda Satavarshiki Mahavidyalaya, Manikpara, Dist, Jhargram, West Bengal, India Pin-721513 And Dr. Rajesh Das, Assistant Professor, Department of Library and Information Science, The University of Burdwan, Burdwan, West Bengal, India, Pin-713104 Abstract Linked open data (LOD) is an essential component in semantic web architecture and is becoming increasingly important over time due to its ability to share and re-use structured data which is both human and computer readable over the web. Currently, many libraries, archives, museums etc. are using open source digital library software to manage and preserve their digital collections. They may also intend to publish their e-resources as “Linked Open Datasets” for further usage. LOD enables the libraries or information centers to publish and share the structured metadata that is generated and maintained with their own bibliographic and authority data in such a way that the other libraries and general community across the world can consume, interact, enrich and share.
    [Show full text]
  • Developing Cultural Heritage Preservation Databases Based on Dublin Core Data Elements
    Developing cultural heritage preservation databases based on Dublin Core data elements Fenella G. France Art Preservation Services Tel: +1 212 722 6300 x4 [email protected] Michael B. Toth R.B. Toth Associates Tel: +1 703 938 4499 [email protected] Abstract: Preserving our cultural heritage – from fragile historic textiles such as national flags to heavy and seemingly solid artifacts recovered from 9/11 – requires careful monitoring of the state of the artifact and environmental conditions. The standard for a web-accessible textile fiber database is established with Dublin Core elements to address the needs of textile conservation. Dynamic metadata and classification standards are also incorporated to allow flexibility in recording changing conditions and deterioration over the life of an object. Dublin core serves as the basis for data sets of information about the changing state of artifacts and environmental conditions. With common metadata standards, such as Dublin Core, this critical preservation knowledge can be utilized by a range of scientists and conservators to determine optimum conditions for slowing the rate of deterioration, as well as comparative use in the preservation of other artifacts. Keywords: Preservation, cultural heritage, Dublin Core, conservation, deterioration, environmental parameters, preservation vocabularies, textiles, fiber, metadata standards, digital library. 1. Introduction Preserving our cultural heritage for future generations is a critical aspect of the stewardship of historical objects. This requires careful monitoring of the state of the artifact and environmental conditions, and standardized storage of this information to support a range of professionals who play a role in the preservation, storage and display of the artifact – including conservators, scientists, curators, and other cultural heritage professionals.
    [Show full text]
  • Choosing a Metadata Type on Metadata Types This Is Our Future; We Can No Longer Rely on Only One Record Structure
    Choosing a Metadata Type On Metadata Types This is our future; we can no longer rely on only one record structure. We must be able to accept many different kinds of bibliographic record structures, from ONIX to Dublin Core to whatever else comes along that contains useful information. To use these record formats we will need rules and guidelines to follow in their application. We need both general rules and schema-specific rules, similar to the way we have used AACR2 to define what information we capture in MARC. – Roy Tennant. “Building a New Bibliographic Infrastructure.”1 Considerations in choosing a metadata type The following questions to consider were put forth by Dorothea Salo, Digital Repository Services Librarian at George Mason University: • What is the problem domain? • Is the choice baked into the system? For example, if you want to use OAI-PMH, Dublin Core is a natural choice. • What are similar projects using? • What else do you have to interoperate with? • What kind of usage infrastructure is there? Is it open or proprietary? Consider the size of the community around this metadata type. • What will this metadata do? • Is it a good standard that encourages good metadata? • Don’t panic; a crosswalk can be used to convert your data to a new metadata standard later.2 Defining the problem domain at the outset is one of the most important considerations. What is the purpose of the system? Is it data retrieval, resource identification, applying access rights to certain metadata, or a combination of these? 3 Defining these requirements up front will make the choice of metadata type more obvious.
    [Show full text]