Knowledge Management and Dublin Core

Total Page:16

File Type:pdf, Size:1020Kb

Knowledge Management and Dublin Core Proc. Int’l Conf. on Dublin Core and Metadata Applications 2009 Knowledge Management and Dublin Core Jon Mason Queensland University of Technology, Australia [email protected] Abstract This paper is concerned with investigating existing and potential scope of Dublin Core metadata in Knowledge Management contexts. Modelling knowledge is identified as a conceptual prerequisite in this investigation, principally for the purpose of clarifying scope prior to identifying the range of tasks associated with organising knowledge. A variety of models is presented and relationships between data, information, and knowledge discussed. It is argued that the two most common modes of organisation, hierarchies and networks, influence the effectiveness and flow of knowledge. Practical perspective is provided by reference to implementations and projects providing evidence of how DC metadata is applied in such contexts. A sense-making model is introduced that can be used as a shorthand reference for identifying useful facets of knowledge that might be described using metadata. Discussion is aimed at presenting this model in a way that both validates current applications and points to potential novel applications. Keywords: KM; knowledge; sense-making; DC; metadata; knowledge management; models; KOS; knowledge organization systems. 1. Introduction The application of Dublin Core (DC) metadata for information management purposes has been taking place ever since the 15 simple DC elements were developed. Its application for industry- specific purposes such as education or government service identification has also been widely adopted, though not without challenges along the way. Knowledge Management (KM) presents an interesting set of challenges for those interested in utilising DC metadata because it can be perceived as a number of different things – as an academic discourse, an organisational intervention, a set of activities that a community of practice might undertake to ensure optimum knowledge flows, or even what an individual might do to maximise the reuse and retrievability of their own knowledge. To underscore this challenge, it is a very interesting exercise just to find a broadly accepted definition of KM that might be both flexible and comprehensive enough to deal with all these scenarios. Although KM first emerged in the 1980s it only began to seriously establish in the mid 1990s when the impact of the Web was just beginning to be felt upon the business world. As such, its early character was biased toward business process improvement from a managerial perspective. The influence of networks upon the way KM has been understood or implemented is something that has only emerged in later discourse (Beerli, et al., 2003; Back, et al., 2006). It has been argued that the fundamental KM problem is all about changing organisational ‘silos’ of activity (manifest in organisational divisions, hierarchical management structures, projects, work teams, documents, and individual workspaces) so that knowledge flows more readily, is shared and leveraged for maximum benefit, and is not pigeonholed nor rendered inaccessible through poor information management practices (Xu and Quaddus, 2005: 382). Such a problem will likely resonate for most people who are employed by organisations. However, while this ‘silo’ problem is clearly evident within hierarchically structured organisations it is also manifest within networks, and it is commonplace for potentially synergistic communities of practice to actually exist more as disconnected ‘islands’. The problem is in fact a deeper one, and 41 2009 Proc. Int’l Conf. on Dublin Core and Metadata Applications is to do with the nature of knowledge and organization—hierarchies and networks are the two most effective forms of organisation that human beings have yet created; both have their place, but it will always be the situational context that will suggest the most effective way to act. Another challenge concerning KM for the DCMI community is that it represents a domain that is arguably more expansive than all other communities currently developing application profiles. Thus, a definition of Knowledge Management that might be useful for the DCMI KM community to consider is presented here: Knowledge Management (KM) finds expression as both an organisational intervention aimed at delivering better efficiencies in the handling of knowledge, and an academic discourse that develops theoretical frameworks and practical techniques for managing the entire knowledge lifecycle from a variety of perspectives: individual, community, and organisational. It can, though doesn’t need to, involve a multiplicity of considerations and tasks and is always influenced by context. 1.1. DCMI KM Community The DCMI Knowledge Management Community was established in mid 2007 as “a forum for individuals and organisations with an interest in the application and use of the Dublin Core standard in knowledge management.” (DCMI, 2007) To some extent this new community provides continuity with issues addressed by the DCMI Global Corporate Circle, which was deactivated in 2007. To date, while there is clearly an interest in this area with over 120 subscribers to the email listserv there is little serious documented discourse. This paper therefore aims to make a contribution by presenting some theoretical framework for consideration, providing examples of how DC metadata is being effectively applied in some Knowledge Management (KM) contexts, and pointing to current limitations of DC in such contexts. 1.2. Overview The following discussion deals with topics on modelling and organising knowledge from a theoretical perspective. A number of scenarios are then presented, indicating how DC metadata has been effectively utilised for KM purposes. A sense-making abstraction is presented as a reference model for identifying prominent facets or pathways of knowledge that will typically need to be considered in KM contexts. This model aims to validate current applications while also pointing to potential novel applications, thereby indicating any limitations with currently available schemas that may need to be overcome. 2. Representing, Modelling, and Managing Knowledge For knowledge to be organised and managed it is necessary to first establish the scope of such an undertaking. This task has been approached by various practitioners and communities of practice in a broad variety of ways. The proceeding discussion is an attempt to summarise some of the more prominent approaches that have relevance for the application of DC metadata. The field of Computer Science provides at least two key (mutually informing) approaches – through formal knowledge representation languages such as Prolog, the Resource Description Framework (RDF), the Web Ontology Language (OWL), and Attempto Controlled English (ACE); or, through conceptual classification into three categories: declarative, procedural, and conditional knowledge. Declarative knowledge is expressed by explicit statements of the kind ‘it is known that’ and typically involves facts or ‘objective’ information; procedural knowledge is typically expressed by statements that represent ‘know-how’; and, conditional knowledge is represented by statements that represent ‘knowing-if’ and/or ‘knowing-why’ (Murphy, 2008). Through recent years of work on developing a robust ‘abstract model’ that can inform future extensibility and application of DC metadata, the Dublin Core Metadata Initiative is now strongly aligned with the knowledge representation capabilities of RDF (Powell, et al., 2007). It is 42 Proc. Int’l Conf. on Dublin Core and Metadata Applications 2009 noteworthy that this work has taken both considerable effort and time and has required ongoing explication through other documents such as the Singapore Framework (Nilsson, et al., 2008). This work has also demonstrated that the goal of defining and using shared semantics is not sufficient for sustainable knowledge sharing in Web environments—for this to occur, semantics must be associated with statements that make clear certain relationships, thereby establishing context through syntax and structural relations. Another approach is to situate knowledge within its relationship to data, information, and wisdom as a value hierarchy and is often depicted as a pyramid (see Figure 1). FIG. 1. Traditional value pyramid representing knowledge However, just as the Internet has rendered geographical boundaries and legal jurisdictions as debatable constructs when it comes to information flows, the boundaries that separate data, information, and knowledge can be very fuzzy and depend upon context. Thus, in the context of the Internet, Figure 1 becomes a very poor representation for the simple reason that data, information, and knowledge become intermeshed. In this environment value can be created through rendering information and knowledge as data and many datasets can either comprise or be extracted from a knowledge-base. In short, this inverts the value-chain depicted by the pyramid (Mason, et al., 2003). Capturing explicit knowledge and organising it as structured information for sharing and reuse is thus one of the powerful features and potentials of the Web. This same recursive property can also be seen with metadata. While it is pragmatic that simple models identify ‘digital assets’ or content as one entity and ‘metadata’ as another (that describes the content) such models can mask deeper complexity. For example, in the case of a repository designed to broker resource discovery,
Recommended publications
  • Metadata for Semantic and Social Applications
    etadata is a key aspect of our evolving infrastructure for information management, social computing, and scientific collaboration. DC-2008M will focus on metadata challenges, solutions, and innovation in initiatives and activities underlying semantic and social applications. Metadata is part of the fabric of social computing, which includes the use of wikis, blogs, and tagging for collaboration and participation. Metadata also underlies the development of semantic applications, and the Semantic Web — the representation and integration of multimedia knowledge structures on the basis of semantic models. These two trends flow together in applications such as Wikipedia, where authors collectively create structured information that can be extracted and used to enhance access to and use of information sources. Recent discussion has focused on how existing bibliographic standards can be expressed as Semantic Metadata for Web vocabularies to facilitate the ingration of library and cultural heritage data with other types of data. Harnessing the efforts of content providers and end-users to link, tag, edit, and describe their Semantic and information in interoperable ways (”participatory metadata”) is a key step towards providing knowledge environments that are scalable, self-correcting, and evolvable. Social Applications DC-2008 will explore conceptual and practical issues in the development and deployment of semantic and social applications to meet the needs of specific communities of practice. Edited by Jane Greenberg and Wolfgang Klas DC-2008
    [Show full text]
  • Provenance and Annotations for Linked Data
    Proc. Int’l Conf. on Dublin Core and Metadata Applications 2013 Provenance and Annotations for Linked Data Kai Eckert University of Mannheim, Germany [email protected] Abstract Provenance tracking for Linked Data requires the identification of Linked Data resources. Annotating Linked Data on the level of single statements requires the identification of these statements. The concept of a Provenance Context is introduced as the basis for a consistent data model for Linked Data that incorporates current best-practices and creates identity for every published Linked Dataset. A comparison of this model with the Dublin Core Abstract Model is provided to gain further understanding, how Linked Data affects the traditional view on metadata and to what extent our approach could help to mediate. Finally, a linking mechanism based on RDF reification is developed to annotate single statements within a Provenance Context. Keywords: Provenance; Annotations; RDF; Linked Data; DCAM; DM2E; 1. Introduction This paper addresses two challenges faced by many Linked Data applications: How to provide, access, and use provenance information about the data; and how to enable data annotations, i.e., further statements about the data, subsets of the data, or even single statements. Both challenges are related as both require the existence of identifiers for the data. We use the Linked Data infrastructure that is currently developed in the DM2E project as an example with typical use- cases and resulting requirements. 1.1. Foundations Linked Data, the publication of data on the Web, enables easy access to data and supports the reuse of data. The Hypertext Transfer Protocol (HTTP) is used to access a Uniform Resource Identifier (URI) and to retrieve data about the resource.
    [Show full text]
  • Rdfa in XHTML: Syntax and Processing Rdfa in XHTML: Syntax and Processing
    RDFa in XHTML: Syntax and Processing RDFa in XHTML: Syntax and Processing RDFa in XHTML: Syntax and Processing A collection of attributes and processing rules for extending XHTML to support RDF W3C Recommendation 14 October 2008 This version: http://www.w3.org/TR/2008/REC-rdfa-syntax-20081014 Latest version: http://www.w3.org/TR/rdfa-syntax Previous version: http://www.w3.org/TR/2008/PR-rdfa-syntax-20080904 Diff from previous version: rdfa-syntax-diff.html Editors: Ben Adida, Creative Commons [email protected] Mark Birbeck, webBackplane [email protected] Shane McCarron, Applied Testing and Technology, Inc. [email protected] Steven Pemberton, CWI Please refer to the errata for this document, which may include some normative corrections. This document is also available in these non-normative formats: PostScript version, PDF version, ZIP archive, and Gzip’d TAR archive. The English version of this specification is the only normative version. Non-normative translations may also be available. Copyright © 2007-2008 W3C® (MIT, ERCIM, Keio), All Rights Reserved. W3C liability, trademark and document use rules apply. Abstract The current Web is primarily made up of an enormous number of documents that have been created using HTML. These documents contain significant amounts of structured data, which is largely unavailable to tools and applications. When publishers can express this data more completely, and when tools can read it, a new world of user functionality becomes available, letting users transfer structured data between applications and web sites, and allowing browsing applications to improve the user experience: an event on a web page can be directly imported - 1 - How to Read this Document RDFa in XHTML: Syntax and Processing into a user’s desktop calendar; a license on a document can be detected so that users can be informed of their rights automatically; a photo’s creator, camera setting information, resolution, location and topic can be published as easily as the original photo itself, enabling structured search and sharing.
    [Show full text]
  • CODATA Workshop on Big Data Programme Book
    Sponsor I CODACODATA S UU Co-Sponsors Organizer Workshop on Big Data for International Scientific Programmes CONTENTS I. Sponsoring Organizations International Council for Science (ICSU) I. Sponsoring Organizations 2 The International Council for Science (ICSU) is a the international scientific community to II. Programme 7 non-governmental organization with a global strengthen international science for the benefit of membership of national scientific bodies society. (121members, representing 141 countries) and international scientific unions (31 members). ICSU: www.icsu.org III. Remarks and Abstracts 13 ICSU mobilizes the knowledge and resources of Committee on Data for Science and Technology (CODATA) IV. Short Biography of Speakers 28 I CODACODATA S UU V. Conference Venue Layout 41 CODATA, the ICSU Committee on Data for Science challenges and ‘hot topics’ at the frontiers and Technology, was established in 1966 to meet of data science (through CODATA Task a need for an international coordinating body to Groups and Working Groups and other improve the management and preservation of initiatives). VI. General Information 43 scientific data. CODATA has been at the forefront 3. Developing data strategies for international of data science and data policy issues since that science programmes and supporting ICSU date. activities such as Future Earth and Integrated About Beijing 43 Research on Disaster Risk (IRDR) to address CODATA supports ICSU’s mission of ‘strengthening data management needs. international science for the benefit of society’ by ‘promoting improved scientific and technical data Through events like the Workshop on Big Data for About the Workshop Venue 43 management and use’. CODATA achieves this International Scientific Programmes and mission through three strands of activity: SciDataCon 2014, CODATA collaborates with 1.
    [Show full text]
  • Introduction to Ontology- Based Semantics Goals Service
    Goals • To provide some insight into the usefulness of ontologies Introduction to Ontology- • To provide an understanding of the based Semantics features of RDFS and OWL and their use in automated reasoning Semantic Web, ontologies, RDF, OWL, • To provide an outline of the N3 syntax N3 • Use N3 to express RDFS and OWL – sufficient for later examples and exercises on service semantics With thanks to Declan O’Sullivan @ David Lewis @ David Lewis Service Semantics Functional Semantics • WSDL provides the syntax we need for interoperating with a service, but little in the way of semantics • What do ‘origin’ and ‘destination’ strings • Examining this example raises many questions about represent? functional and non-functional semantics – Country, city, airport, restrictions (airline, national, <message name=“getcheapestFlightRequest"> regional)? <part name=“origin" type="xsd:string"/> <part name=“destination" type="xsd:string"/> • What does ‘flight’ string represent? <part name=“date" type="xsd:date"/> – Airline, flight number, plane type? </message> <message name=“getcheapestFlightResponse"> • What does ‘time’ string represent? <part name=“flight" type="xsd:string"/> <part name=“time" type="xsd:time"/> – Departure time? <part name=“cost” type=“xsd:float”/> </message> – Note xsd:time probably is adequate - supports time- <portType name=“cheapestFlight"> zone information <operation name="getCheapestFlight"> <input message=“getcheapestFlightRequest"/> • What does ‘cost’ float represent? <output message=“getcheapestFlightResponse"/> </operation>
    [Show full text]
  • Folksonomies - Cooperative Classification and Communication Through Shared Metadata
    Folksonomies - Cooperative Classification and Communication Through Shared Metadata Adam Mathes Computer Mediated Communication - LIS590CMC Graduate School of Library and Information Science University of Illinois Urbana-Champaign December 2004 Abstract This paper examines user-generated metadata as implemented and applied in two web services designed to share and organize digital me- dia to better understand grassroots classification. Metadata - data about data - allows systems to collocate related information, and helps users find relevant information. The creation of metadata has generally been approached in two ways: professional creation and author creation. In li- braries and other organizations, creating metadata, primarily in the form of catalog records, has traditionally been the domain of dedicated profes- sionals working with complex, detailed rule sets and vocabularies. The primary problem with this approach is scalability and its impracticality for the vast amounts of content being produced and used, especially on the World Wide Web. The apparatus and tools built around professional cataloging systems are generally too complicated for anyone without spe- cialized training and knowledge. A second approach is for metadata to be created by authors. The movement towards creator described docu- ments was heralded by SGML, the WWW, and the Dublin Core Metadata Initiative. There are problems with this approach as well - often due to inadequate or inaccurate description, or outright deception. This paper examines a third approach: user-created metadata, where users of the documents and media create metadata for their own individual use that is also shared throughout a community. 1 The Creation of Metadata: Professionals, Con- tent Creators, Users Metadata is often characterized as “data about data.” Metadata is information, often highly structured, about documents, books, articles, photographs, or other items that is designed to support specific functions.
    [Show full text]
  • Extending the Role of Metadata in a Digital Library System*
    Extending the Role of Metadata in a Digital Library System* Alexa T. McCray, Marie E. Gallagher, Michael A. Flannick National Library of Medicine Bethesda, MD, USA {mccray,gallagher,flannick}@nlm.nih.gov Abstract Metadata efforts often fall into the trap of We describe an approach to the development of a digital trying to create a universal metadata schema. library system that is founded on a number of basic Such efforts fail to recognize the basic nature principles. In particular, we discuss the critical role of of metadata: namely, that it is far too diverse metadata in all aspects of the system design. We begin by to fit into one useful taxonomy...the creation, describing how the notion of metadata is sometimes administration, and enhancement of individual interpreted and go on to discuss some of our early metadata forms should be left to the relevant experiences in a digital conversion project. We report on communities of expertise. Ideally this would the Profiles in Science project, which is making the occur within a framework that will support archival collections of prominent biomedical scientists interoperability across data and domains. available on the World Wide Web. We discuss the [4:277] principles that are used in our system design, illustrating Roszkowski and Lukas describe an approach for these throughout the discussion. Our approach has linking distributed collections of metadata so that they are involved interpreting metadata in its broadest sense. We searchable as a single collection [5], and Baldonado et al. capture data about the items in our digital collection for a describe an architecture that facilitates metadata a variety of purposes and use those data to drive the compatibility and interoperability [6].
    [Show full text]
  • A Blockchain-Based Application to Protect Minor Artworks
    A Blockchain-based Application to Protect Minor Artworks Clara Bacciu, Angelica Lo Duca, Andrea Marchetti IIT-CNR - via G. Moruzzi 1, 56124 Pisa (Italy) [name].[surname]@iit.cnr.it Keywords: blockchain, Ethereum, recordkeeping, IPFS, Cultural Heritage, minor artwork Abstract: A new emerging trend concerns the implementation of services and distributed applications through the blockchain technology. A blockchain is an append-only database, which guarantees security, transparency and immutability of records. Blockchains can be used in the field of Cultural Heritage to protect minor artworks, i.e. artistic relevant works not as famous as masterpieces. Minor artworks are subjected to counterfeiting, thefts and natural disasters because they are not well protected as famous artworks. This paper describes a blockchain-based application, called MApp (Minor Artworks application), which lets authenticated users (pri- vate people or organizations), store the information about their artworks in a secure way. The use of blockchain produces three main advantages. Firstly, artworks cannot be deleted from the register thus preventing thieves to remove records associated stolen objects. Secondly, artworks can be added and updated only by authorized users, thus preventing counterfeiting in objects descriptions. Finally, records can be used to keep artworks memory in case of destruction caused by a natural disaster. 1 INTRODUCTION cessible. To complicate things, in the last decades, Italy (as many other places in the world) has suffered Minor artworks are works that are artistically rel- a series of catastrophic events that largely affected evant but not as well-known as famous masterpieces, cultural heritage: earthquakes, landslides, floods, col- or belonging to the so-called minor arts, such as books lapses in important archaeological sites such as Pom- and manuscripts, pottery, lacquerware, furniture, jew- pei.
    [Show full text]
  • Integrating Dublin Core Metadata for Cultural Heritage Collections Using Ontologies**
    2007 Proc. Int’l Conf. on Dublin Core and Metadata Applications Integrating Dublin Core metadata for cultural heritage ** collections using ontologies Constantia Kakali Irene Lourdi Thomais Stasinopoulou Department of Archive and Department of Archive and Department of Archive and Library Sciences, Ionian Library Sciences, Ionian Library Sciences, Ionian University, Greece University, Greece University, Greece [email protected] [email protected] [email protected] Lina Bountouri Christos Papatheodorou Martin Doerr Department of Archive and Department of Archive and Institute of Computer Library Sciences, Ionian Library Sciences, Ionian Science, Foundation for University, Greece University, Greece Research and Technology, [email protected] [email protected] Greece [email protected] Manolis Gergatsoulis Department of Archive and Library Sciences, Ionian University, Greece [email protected] Abstract Metadata interoperability is an active research area, especially for cultural heritage collections, which consist of heterogeneous objects described by a variety of metadata schemas. In this paper we propose an ontology-based metadata interoperability approach, which exploits, in an optimal way, the semantics of metadata schemas. In particular, we propose the use of CIDOC/CRM ontology as a mediating schema and present a methodology for mapping DC Type Vocabulary to CIDOC/CRM, demonstrating a real-world effort for ontology-based metadata integration. Keywords: ontology-based integration; metadata interoperability; CIDOC/CRM; DC-Type. 1. Introduction Heterogeneity is one of the main characteristics of cultural heritage collections. Such collections may be composed of text, written on different materials, paintings, photographs, 3D objects, sound recordings, maps or even digital objects. Furthermore, the objects are strongly related with the social and historical events that take place over time.
    [Show full text]
  • Publishing E-Resources of Digital Institutional Repository As Linked Open Data: an Experimental Study
    University of Nebraska - Lincoln DigitalCommons@University of Nebraska - Lincoln Library Philosophy and Practice (e-journal) Libraries at University of Nebraska-Lincoln 11-30-2020 Publishing E-resources of Digital Institutional Repository as Linked Open Data: an experimental study Subhendu Kar Vivekananda Satavarshiki Mahavidyalaya, [email protected] Dr. Rajesh Das The University of Burdwan, [email protected] Follow this and additional works at: https://digitalcommons.unl.edu/libphilprac Part of the Library and Information Science Commons Kar, Subhendu and Das, Dr. Rajesh, "Publishing E-resources of Digital Institutional Repository as Linked Open Data: an experimental study" (2020). Library Philosophy and Practice (e-journal). 4699. https://digitalcommons.unl.edu/libphilprac/4699 Publishing E-resources of Digital Institutional Repository as Linked Open Data: an experimental study Subhendu Kar, Librarian, Vivekananda Satavarshiki Mahavidyalaya, Manikpara, Dist, Jhargram, West Bengal, India Pin-721513 And Dr. Rajesh Das, Assistant Professor, Department of Library and Information Science, The University of Burdwan, Burdwan, West Bengal, India, Pin-713104 Abstract Linked open data (LOD) is an essential component in semantic web architecture and is becoming increasingly important over time due to its ability to share and re-use structured data which is both human and computer readable over the web. Currently, many libraries, archives, museums etc. are using open source digital library software to manage and preserve their digital collections. They may also intend to publish their e-resources as “Linked Open Datasets” for further usage. LOD enables the libraries or information centers to publish and share the structured metadata that is generated and maintained with their own bibliographic and authority data in such a way that the other libraries and general community across the world can consume, interact, enrich and share.
    [Show full text]
  • Developing Cultural Heritage Preservation Databases Based on Dublin Core Data Elements
    Developing cultural heritage preservation databases based on Dublin Core data elements Fenella G. France Art Preservation Services Tel: +1 212 722 6300 x4 [email protected] Michael B. Toth R.B. Toth Associates Tel: +1 703 938 4499 [email protected] Abstract: Preserving our cultural heritage – from fragile historic textiles such as national flags to heavy and seemingly solid artifacts recovered from 9/11 – requires careful monitoring of the state of the artifact and environmental conditions. The standard for a web-accessible textile fiber database is established with Dublin Core elements to address the needs of textile conservation. Dynamic metadata and classification standards are also incorporated to allow flexibility in recording changing conditions and deterioration over the life of an object. Dublin core serves as the basis for data sets of information about the changing state of artifacts and environmental conditions. With common metadata standards, such as Dublin Core, this critical preservation knowledge can be utilized by a range of scientists and conservators to determine optimum conditions for slowing the rate of deterioration, as well as comparative use in the preservation of other artifacts. Keywords: Preservation, cultural heritage, Dublin Core, conservation, deterioration, environmental parameters, preservation vocabularies, textiles, fiber, metadata standards, digital library. 1. Introduction Preserving our cultural heritage for future generations is a critical aspect of the stewardship of historical objects. This requires careful monitoring of the state of the artifact and environmental conditions, and standardized storage of this information to support a range of professionals who play a role in the preservation, storage and display of the artifact – including conservators, scientists, curators, and other cultural heritage professionals.
    [Show full text]
  • Choosing a Metadata Type on Metadata Types This Is Our Future; We Can No Longer Rely on Only One Record Structure
    Choosing a Metadata Type On Metadata Types This is our future; we can no longer rely on only one record structure. We must be able to accept many different kinds of bibliographic record structures, from ONIX to Dublin Core to whatever else comes along that contains useful information. To use these record formats we will need rules and guidelines to follow in their application. We need both general rules and schema-specific rules, similar to the way we have used AACR2 to define what information we capture in MARC. – Roy Tennant. “Building a New Bibliographic Infrastructure.”1 Considerations in choosing a metadata type The following questions to consider were put forth by Dorothea Salo, Digital Repository Services Librarian at George Mason University: • What is the problem domain? • Is the choice baked into the system? For example, if you want to use OAI-PMH, Dublin Core is a natural choice. • What are similar projects using? • What else do you have to interoperate with? • What kind of usage infrastructure is there? Is it open or proprietary? Consider the size of the community around this metadata type. • What will this metadata do? • Is it a good standard that encourages good metadata? • Don’t panic; a crosswalk can be used to convert your data to a new metadata standard later.2 Defining the problem domain at the outset is one of the most important considerations. What is the purpose of the system? Is it data retrieval, resource identification, applying access rights to certain metadata, or a combination of these? 3 Defining these requirements up front will make the choice of metadata type more obvious.
    [Show full text]