The Continuous Media Web: a Distributed Multimedia Information Retrieval Architecture Extending the World Wide Web

Total Page:16

File Type:pdf, Size:1020Kb

The Continuous Media Web: a Distributed Multimedia Information Retrieval Architecture Extending the World Wide Web Multimedia Systems (2005) 10(6): 544–558 DOI 10.1007/s00530-005-0181-8 REGULAR PAPER Silvia Pfeiffer · Conrad Parker · André Pang The Continuous Media Web: a distributed multimedia information retrieval architecture extending the World Wide Web Published online: 8 August 2005 c Springer-Verlag 2005 Abstract The World Wide Web, with its paradigms of surf- given in the URI (Unified Resource Identifier) format (see ing and searching for information, has become the predom- http://www.w3.org/Addressing/). inant system for computer-based information retrieval. Me- The World Wide Web is currently the predominant dis- dia resources, however information-rich, only play a minor tributed information retrieval architecture on the Internet [2]. role in providing information to Web users. While band- However, it is not yet a distributed information retrieval sys- width (or the lack thereof) may be an excuse for this situ- tem for time-continuously sampled data such as audio and ation, the lack of surfing and searching capabilities on me- video. It is only set up well for information retrieval on dia resources are the real issue. We present an architecture HTML resources. Although the number of audio and video that extends the Web to media, enabling existing Web in- data on the Internet is increasing, these documents cannot be frastructures to provide seamless search and hyperlink ca- used as intensively as HTML content. pabilities for time-continuous Web resources, with only mi- While URIs provide a uniform means to access and ex- nor extensions. This makes the Web a true distributed infor- plore information, they can only point to a complete audio mation system for multimedia data. The article provides an or video file, not into a specific temporal offset or a spe- overview of the specifications that have been developed and cific segment of interest. Also, when viewing an audio or submitted to the IETF for standardization. It also presents video resource, the hyperlinking functionality of the Web is experimental results with prototype applications. “left behind” and a dead end is reached. Audio or video re- sources cannot typically hyperlink to further Web resources, · · · Keywords Continuous Media Web Annodex CMML thus interrupting the uniform means of information explo- · Markup of media Open media metadata standards ration provided by the Web. Web information search is also deficient for media re- sources. While it is nowadays possible to search the content 1 Introduction of binary text documents, such as Postscript files or word processor documents, and identify whether they contain in- Since the Web’s very foundation by Tim Berners-Lee, aca- formation relevant to a particular query, this is not possible demics, entrepreneurs, and users have dreamt of ways to in- for media resources. Media resources do not currently have tegrate the Web’s various multimedia contents into one dis- a textual representation that fits into the text-based index- tributed system for storage and retrieval. The power of the ing paradigm of existing Web search engines, and therefore World Wide Web stems from being a world-wide distributed their content, however information-rich, cannot be searched information storage system architected to be scalable, simple uniformly on the Web. to author for, and simple to use. Information is found mostly Realizing this situation, we have developed an exten- through Web search engines such as Google or through Web sion to the World Wide Web that addresses these issues. We portals, both of which commonly just provide starting points are proposing a new standard for the way time-continuously for hyperlinking (surfing) to other related documents. In- sampled data such as audio and video files are placed on formation nowadays encompasses heterogeneous Web re- Web sites to make media content just as “surfable” as ordi- sources of all sorts, including media. The uniform means nary text files. In this new standard, users can link not only of accessing information on the Web is through hyperlinks from a text passage to e.g., a video, but into a specific time interval containing the information being sought [21]. The S. Pfeiffer (B) · C. Parker · A. Pang time-continuous resource itself is annotated with HTML- CSIRO-ICT Centre PO Box 76 Epping, NSW 1710, Australia like markup and has an XML representation of its content, E-mail: {Silvia.Pfeiffer, Conrad.Parker, Andre.Pang}@csiro.au enabling Web search engines to index it. The markup may The continuous media web 545 contain hyperlinks to other Web resources, which enables Decompression will lead to the higher temporal resolu- search engines to crawl time-continuous resources in the tion that the resource was sampled at in the first place. same way as they crawl HTML pages. These hyperlinks also This is the case for all common audio and video compres- enable the Web user to hyperlink through a time-continuous sion formats. Other types of time-continuously sampled resource, thus integrating these resources into the “surfing” data may, for example, be physical measurements of nat- paradigm of the Web. ural phenomena, such as the acidity (pH) value of a lake, At CSIRO we call the Web that is extended to time- or the temporally changing value of a stock market. Time- continuous resources the Continuous Media Web (CMWeb) continuously sampled data are also sometimes just called [24]. The technology is based upon the observation that the time-continuous data or, in the case of audio or video, just data stream of existing time-continuous data might be tem- media. porally subdivided based on a semantic concept, and that • Time-continuous (Web) resource is a time-continuous data this structure enables access to interesting subparts, known stream that can be distributed by a Web server. The re- as clips, of the stream. This structure is captured in an source may exist as a file or may be created on the fly HTML-like markup language, which we have named Con- (live or composed). To retain the ability to handle and dis- tinuous Media Markup Language (CMML) [23]. CMML play the time-continuous data progressively during deliv- allows annotations and hyperlinks to be attached to me- ery over HTTP, a time-continuous Web resource must fur- dia resources. For synchronized delivery of the markup and ther be streamable, i.e., it must be possible to decode it in the time-continuous data over the Web, we have developed a single linear pass using a fixed amount of memory. a streamable container format called Annodex [22], which • Clip is an arbitrary temporal section of a time-continuous encapsulates CMML together with the time-continuous re- Web resource as defined by a user. A clip makes sense source. When consuming (i.e., listening, viewing, reading) with respect to some semantic measure of the user. A such an annotated and indexed resource, the user experience whole time-continuous Web resource may be regarded as is such that while, e.g., watching a video, the annotations a clip, though the more common case is to describe such and links change over time and enable browsing of collec- a resource as a combination of several clips. tions simply by following a link to another Web resource. • Annotation is a fee-text, unstructured description of a clip. The remainder of this article is organized as follows. • Metadata are a set of name–value pairs that provide Section 2 provides some definitions for the key concepts dis- databaselike structured descriptions, e.g., in HTML meta- cussed in this article, and Sect. 3 discusses some related re- data are represented in meta elements within the head tag search work. Section 4 describes the specifications of the [29]. An existing metadata scheme that is independent of CMWeb architecture and technologies, introducing tempo- the format of data that it can describe is the Dublin Core ral URI addressing, the Continuous Media Markup Lan- [7]). guage (CMML), and the Annodex format. Switching the fo- • Hyperlink is a Unified Resource Identifier that can point cus to the actual handling of Annodex resources, Sects. 5 to or into any Web resource. In this context, the term me- and 6 detail how an Annodex resource is being produced dia browsing is then mainly used for following hyperlinks fromanexistingmediafilesuchasanMPEG-1 encoded into and out of media files rather than looking through a video and how it is distributed over the Internet. Section 7 collection of media files, as is the more common use of looks at information retrieval through Web search engines. the term in media management systems. Section 8 provides some experimental results of first imple- • Meta-information is a collection of information about a mentations. Section 9 compares the CMWeb to other media time-continuous data stream, which may include annota- standards with respect to their capabilities of extending the tions, metadata, and hyperlinks. World Wide Web to a distributed multimedia information • Markup is the collection of tags that are used to provide retrieval infrastructure. Finally, Sect. 10 concludes the paper metainformation to a Web resource. The most common with a summary and outlook for future work. language for providing markup is XML [33]. 2 Definitions 3 Related research Before we can discuss work in this field, we need to estab- There have been several approaches to a better integration lish that the terms we use have a well-understood meaning. of media documents into the World Wide Web that have not Therefore, we now introduce some fundamental definitions. been standards based. These will be briefly discussed here. A standards-based approach is superior as it enables interop- • Time-continuously sampled data are any sequence of erability between different proprietary solutions and scales numbers that represents an analog-time signal sampled in over the whole Internet rather than providing a solution lim- discrete time steps.
Recommended publications
  • Xincluding Plain-Text Fragments for Symmetry and Profit Piotr Bański, University of Warsaw, [email protected]
    XIncluding plain-text fragments for symmetry and profit Piotr Bański, University of Warsaw, [email protected] <xi:include href=”resource.xml” parse=”text|xml” Summary xpointer=”fragment_identifier” /> XML for the annotation of images, audio or video is by necessity of the stand-off kind: the annotations are virtually anchored to the parse=“xml” parse=“text” binary stream (typically by x,y coordinates or time offsets). Work on language resource annotation has extended this approach to plain text, where annotations are anchored by character offsets. This is referred to as extreme stand-off markup. No generic XML tools or standards support it, but... they could, fairly easily. inclusion of inclusion of inclusion of inclusion of Rationale the entire the entire a fragment a fragment ISO Language Annotation Framework (e.g. Ide & Romary, 2007) re- XML text of an XML of a text commends that data be kept as read-only plain text, with one or mo- resource resource resource resource re layers of XML annotation identifying individual spans of charac- ters and adding metadata to them (identifying e.g. segments of va- rious levels of granularity, parts of speech, syntactic or discourse @xpointer present functions, etc.) no @xpointer However, the only way to handle this (as evidenced by the American National Corpus), is to use in-house tools working on correspon- dence semantics for hyperlinks (see Bański, 2010, for terminology What makes it impossible right now? and more remarks). 1. XInclude Ban: XInclude Rec, 3.1: “The xpointer attribute must not be present when parse="text".” But character streams can do better! Because text is what XML is built out of and what it contains, instead of correspondence 2.
    [Show full text]
  • W3C Working Draft: Scalable Vector
    next contents properties index Scalable Vector Graphics (SVG) 1.0 Specification W3C Working Draft 03 March 2000 This version: http://www.w3.org/TR/2000/03/WD-SVG-20000303 (Available as: PDF, zip archive of HTML) Previous public version: http://www.w3.org/TR/1999/WD-SVG-19991203/ Latest public version: http://www.w3.org/TR/SVG/ Editor: Jon Ferraiolo <[email protected]> Authors: See author list Copyright ©1998, 1999, 2000 W3C® (MIT, INRIA, Keio), All Rights Reserved. W3C liability, trademark, document use and software licensing rules apply. Abstract This specification defines the features and syntax for Scalable Vector Graphics (SVG), a language for describing two-dimensional vector and mixed vector/raster graphics in XML. Status of this document This document is a public review draft version of the SVG specification. This working draft attempts to address review comments that were received during the initial Last Call period, which started 12 August 1999, and also incorporates other modifications resulting from continuing collaboration with other working groups and continuing work within the SVG working group. With the publication of this draft, the SVG specification enters a second "Last Call". The second Last Call period will end on 31 March, 2000. The SVG specification is going through a second Last Call review process to provide the public and other working groups an opportunity to review the changes to the specification since the initial Last Call period. A complete list of all changes since the initial Last Call version of the specification is available in Appendix L: Change History. Last call comments should be sent to [email protected].
    [Show full text]
  • Understanding JSON Schema Release 2020-12
    Understanding JSON Schema Release 2020-12 Michael Droettboom, et al Space Telescope Science Institute Sep 14, 2021 Contents 1 Conventions used in this book3 1.1 Language-specific notes.........................................3 1.2 Draft-specific notes............................................4 1.3 Examples.................................................4 2 What is a schema? 7 3 The basics 11 3.1 Hello, World!............................................... 11 3.2 The type keyword............................................ 12 3.3 Declaring a JSON Schema........................................ 13 3.4 Declaring a unique identifier....................................... 13 4 JSON Schema Reference 15 4.1 Type-specific keywords......................................... 15 4.2 string................................................... 17 4.2.1 Length.............................................. 19 4.2.2 Regular Expressions...................................... 19 4.2.3 Format.............................................. 20 4.3 Regular Expressions........................................... 22 4.3.1 Example............................................. 23 4.4 Numeric types.............................................. 23 4.4.1 integer.............................................. 24 4.4.2 number............................................. 25 4.4.3 Multiples............................................ 26 4.4.4 Range.............................................. 26 4.5 object................................................... 29 4.5.1 Properties...........................................
    [Show full text]
  • Iso/Iec 21000-17:2006(E)
    This is a preview - click here to buy the full publication INTERNATIONAL ISO/IEC STANDARD 21000-17 First edition 2006-09-15 Information technology — Multimedia framework (MPEG-21) — Part 17: Fragment Identification of MPEG Resources Technologies de l'information — Cadre multimédia (MPEG-21) — Partie 17: Identification de fragments de ressources MPEG Reference number ISO/IEC 21000-17:2006(E) © ISO/IEC 2006 ISO/IEC 21000-17:2006(E) This is a preview - click here to buy the full publication PDF disclaimer This PDF file may contain embedded typefaces. In accordance with Adobe's licensing policy, this file may be printed or viewed but shall not be edited unless the typefaces which are embedded are licensed to and installed on the computer performing the editing. In downloading this file, parties accept therein the responsibility of not infringing Adobe's licensing policy. The ISO Central Secretariat accepts no liability in this area. Adobe is a trademark of Adobe Systems Incorporated. Details of the software products used to create this PDF file can be found in the General Info relative to the file; the PDF-creation parameters were optimized for printing. Every care has been taken to ensure that the file is suitable for use by ISO member bodies. In the unlikely event that a problem relating to it is found, please inform the Central Secretariat at the address given below. © ISO/IEC 2006 All rights reserved. Unless otherwise specified, no part of this publication may be reproduced or utilized in any form or by any means, electronic or mechanical, including photocopying and microfilm, without permission in writing from either ISO at the address below or ISO's member body in the country of the requester.
    [Show full text]
  • Pipenightdreams Osgcal-Doc Mumudvb Mpg123-Alsa Tbb
    pipenightdreams osgcal-doc mumudvb mpg123-alsa tbb-examples libgammu4-dbg gcc-4.1-doc snort-rules-default davical cutmp3 libevolution5.0-cil aspell-am python-gobject-doc openoffice.org-l10n-mn libc6-xen xserver-xorg trophy-data t38modem pioneers-console libnb-platform10-java libgtkglext1-ruby libboost-wave1.39-dev drgenius bfbtester libchromexvmcpro1 isdnutils-xtools ubuntuone-client openoffice.org2-math openoffice.org-l10n-lt lsb-cxx-ia32 kdeartwork-emoticons-kde4 wmpuzzle trafshow python-plplot lx-gdb link-monitor-applet libscm-dev liblog-agent-logger-perl libccrtp-doc libclass-throwable-perl kde-i18n-csb jack-jconv hamradio-menus coinor-libvol-doc msx-emulator bitbake nabi language-pack-gnome-zh libpaperg popularity-contest xracer-tools xfont-nexus opendrim-lmp-baseserver libvorbisfile-ruby liblinebreak-doc libgfcui-2.0-0c2a-dbg libblacs-mpi-dev dict-freedict-spa-eng blender-ogrexml aspell-da x11-apps openoffice.org-l10n-lv openoffice.org-l10n-nl pnmtopng libodbcinstq1 libhsqldb-java-doc libmono-addins-gui0.2-cil sg3-utils linux-backports-modules-alsa-2.6.31-19-generic yorick-yeti-gsl python-pymssql plasma-widget-cpuload mcpp gpsim-lcd cl-csv libhtml-clean-perl asterisk-dbg apt-dater-dbg libgnome-mag1-dev language-pack-gnome-yo python-crypto svn-autoreleasedeb sugar-terminal-activity mii-diag maria-doc libplexus-component-api-java-doc libhugs-hgl-bundled libchipcard-libgwenhywfar47-plugins libghc6-random-dev freefem3d ezmlm cakephp-scripts aspell-ar ara-byte not+sparc openoffice.org-l10n-nn linux-backports-modules-karmic-generic-pae
    [Show full text]
  • Curriculum Vitae
    Curriculum Vitae Dr. Silvia Pfeiffer PERSONAL DETAILS Address: 1 Colston Street Ryde NSW 2112 Australia Phone: +61 2 8012 0937 (home) +61 401 38 40 41 (mobile) Email: [email protected] PROFICIENCY IN LANGUAGES German Mother Tongue English Fluent, 7 years at school, 6 months au-pair, 10 years in Australia French Acceptable, 7 years at school, no practice since EDUCATION 1987-1993 “Wirtschaftsinformatik” studies (business management and computer science) at the University of Mannheim, Germany short-term jobs coding in SME companies Masters Degree (double degree) with excellent marks 1994-1999 PhD in Computer Science at the University of Mannheim at the Chair of Computer Networks and Multimedia Technology, Prof. Dr. W. Effelsberg PhD thesis on “Information Retrieval from the Digital Audio Tracks of Video” as Doctor of Science (Dr. rer. nat.), mark magna cum laude PROFESSIONAL CAREER July – Dec 2009 Mozilla Corporation Contract Contract to add basic subtitle/caption support for <video> into Firefox for both out-of-band (i.e. separate text file) and in-line (i.e. a track inside Ogg Theora) captions, as well as work with the W3C to define extensions to HTML5 for video accessibility. Oct – Dec 2008 Mozilla Foundation Grant Grant to produce a study and recommendation on the current state of video accessibility for the Web. July 2006 – now Vquence Pty Ltd CEO since July 2007: refocused business to providing competitive intelligence on video through video metrics and online engagement platform. CTO and Co-Founder: Start-up company in digital online media with technology in video search, and a mash-up application using either flash format video or Ogg Theora/Annodex.
    [Show full text]
  • 10FH2006.Pdf (PDF, 1.19MB)
    EOPAS, THE ETHNOER ONLINE REPRESENTATION OF INTERLINEAR TEXT1 Ronald Schroeter School of Information Technology & Electrical Engineering University of Queensland Nick Thieberger School of Linguistics & Applied Linguistics University of Melbourne One of the goals of the Australian Research Council-funded E-research project called Sharing access and analytical tools for ethnographic digital media using high speed networks, or simply EthnoER, is to take outputs of normal linguistic analytical processes and present them online in a system we have called the EthnoER online presentation and annotation system, or EOPAS. EthnoER has twenty-three chief investigators from ten organisations, both Australian and international, at Universities and other agencies, including the Commonwealth Scientific and Industrial Research Organisation (CSIRO), language centres and language archives. The project includes a set of testbed projects with varying requirements for online annotation and presentation of data. We aim to provide online mechanisms for annotating data in order that, for example, archival media files can be annotated by experts with local knowledge and that samples of that media, perhaps a three minute chunk, can be presented so that users can get a sense of the quality of the recording and other information that may influence their decision to download the whole file. We were inspired by Michel Jacobson's work (CNRS, n.d.) in presenting interlinear text for the LACITO group in Paris, and wanted to make a similar tool available more generally. In this paper we will focus on the linguistic goals of the project using interlinear text and transcribed media as the inputs to an online delivery and querying system, and on the archival data curated by the Pacific And Regional Archive for Digital Sources in Endangered Cultures (PARADISEC).
    [Show full text]
  • 3 Normative References
    3 Normative References NR1 W3C/TR ldp Linked Data Platform1.0. W3C Recommendation, 26 February 2015. http://www.w3.org/TR/ldp/ NR2 W3C/TR owl2-syntax OWL 2 Web Ontology Language: Structural Specification and Functional-Style Syntax(Second Edition). W3C Recommendation, 11 December 2012. http://www.w3.org/TR/owl2-syntax/ NR3 ISO/IEC 14977:1996 Information technology – Syntactic metalanguage – Extended BNF. NR4 W3C/TR xml Extensible Markup Language (XML)1.0 (Fifth Edition). W3C Recommendation, 26 November 2008. http://www.w3.org/TR/xml/ NR5 W3C/TR owl2-primer OWL 2 Web Ontology Language: Primer(Second Edition). W3C Recommendation, 11 December 2012. http://www.w3.org/TR/owl2-primer/ NR6 W3C/TR owl2-profiles OWL 2 Web Ontology Language: Profiles(Second Edition). W3C Recommendation, 11 Decem- ber 2012. http://www.w3.org/TR/owl2-profiles/ NR7 ISO/IEC 24707:2007 Information technology – Common Logic (CL): a framework for a family of logic-based languages. : NR8 OMG Document ptc/2013-09-05 OMG Unified Modeling Language (OMG UML). http://www.omg.org/spec/UML/Current NR9 IETF/RFC 2046 Multipurpose Internet Mail Extensions (MIME) Part Two: Media Types. https://www.ietf.org/rfc/rfc2046.txt NR10 IETF/RFC 3986 Uniform Resource Identifier (URI): Generic Syntax. January 2005. http://tools.ietf.org/html/rfc3986 NR11 IETF/RFC 3987 Internationalized Resource Identifiers (IRIs). January 2005. http://tools.ietf.org/html/rfc3987 NR12 IETF/RFC 5147 URI Fragment Identifiers for the text/plain Media Type. April 2008. http://tools.ietf.org/html/rfc5147 NR13 W3C/TR xptr-framework XPointer Framework. W3C Recommendation, 25 March 2003.
    [Show full text]
  • Deep Linking Desktop Resources
    Deep Linking Desktop Resources Markus Schr¨oder,Christian Jilek, and Andreas Dengel 1 Smart Data & Knowledge Services Dept., DFKI GmbH, Kaiserslautern, Germany 2 Computer Science Dept., TU Kaiserslautern, Germany fmarkus.schroeder, christian.jilek, [email protected] Abstract. Deep Linking is the process of referring to a specific piece of web content. Although users can browse their files in desktop environ- ments, they are unable to directly traverse deeper into their content using deep links. In order to solve this issue, we demonstrate \DeepLinker", a tool which generates and interprets deep links to desktop resources, thus enabling the reference to a certain location within a file using a simple hyperlink. By default, the service responds with an HTML representation of the resource along with further links to follow. Additionally, we allow the use of RDF to interlink our deep links with other resources. Keywords: Deep Link, Desktop, URI, Fragment Identifier, RDF 1 Introduction Internet resources are commonly identified using URIs3 which contain the in- formation to look them up. In this context, \deep linking"4 is the process of using a hyperlink to link to a specific piece of web content. For example, https://en.wikipedia.org/wiki/Deep_linking#Example refers to the Exam- ple section of Wikipedia's Deep Linking page. Similarly, desktop resources, like files (in various formats5) and folders, are typically addressed with their absolute paths or file URI schemata6. Although users can browse a file in this way, they are unable to directly descend deeper into its content. As a result, users cannot refer to a certain location deep inside their desktop resources, for example, refer- ring to a shape in a slide of a presentation.
    [Show full text]
  • Open Modeling and Exchange (OMEX) Metadata Specification Version
    Journal of Integrative Bioinformatics 2020; 17(2–3): 20200020 Maxwell L. Neal*, John H. Gennari, Dagmar Waltemath, David P. Nickerson and Matthias König Open modeling and exchange (OMEX) metadata specification version 1.0 https://doi.org/10.1515/jib-2020-0020 Received April 16, 2020; accepted April 26, 2020; published online June 25, 2020 Abstract: A standardized approach to annotating computational biomedical models and their associated files can facilitate model reuse and reproducibility among research groups, enhance search and retrieval of models and data, and enable semantic comparisons between models. Motivated by these potential benefits and guided by consensus across the COmputational Modeling in BIology NEtwork (COMBINE) community, we have developed a specification for encoding annotations in Open Modeling and EXchange (OMEX)-formatted ar- chives. Distributing modeling projects within these archives is a best practice established by COMBINE, and the OMEX metadata specification presented here provides a harmonized, community-driven approach for anno- tating a variety of standardized model and data representation formats within an archive. The specification primarily includes technical guidelines for encoding archive metadata, so that software tools can more easily utilize and exchange it, thereby spurring broad advancements in model reuse, discovery, and semantic analyses. Keywords: data integration; metadata; semantics; simulation. *Corresponding author: Maxwell L. Neal, Seattle Children’s Research Institute, Seattle, USA, E-mail: [email protected]. https://orcid.org/0000-0002-2390-6572 John H. Gennari: University of Washington, Seattle, USA. https://orcid.org/0000-0001-8254-4957 Dagmar Waltemath: University Medicine Greifswald, Greifswald, Germany. https://orcid.org/0000-0002-5886-5563 David P. Nickerson: University of Auckland, Auckland, New Zealand.
    [Show full text]
  • Mediacentral | Asset Management Swodl Reference, Created 9/20/2020
    MediaCentral® | Asset Management SWoDL Reference Release 2020.9 Contents Using This Reference ....................................................................... 6 Symbols and Conventions .......................................................................................... 6 If You Need Help ......................................................................................................... 7 Avid Training Services ................................................................................................ 7 1 Introduction ....................................................................................... 8 2 Basic Language Elements ................................................................ 9 Command Separator, Identifiers ................................................................................. 9 Comments ................................................................................................................. 10 Status Definitions ...................................................................................................... 10 Literals ....................................................................................................................... 11 String Literals ....................................................................................................... 11 Integer Literals ..................................................................................................... 12 Floating-Point Literals .........................................................................................
    [Show full text]
  • A URI-Based Approach for Addressing Fragments of Media Resources on the Web
    Noname manuscript No. (will be inserted by the editor) A URI-Based Approach for Addressing Fragments of Media Resources on the Web Erik Mannens · Davy Van Deursen · Rapha¨elTroncy · Silvia Pfeiffer · Conrad Parker · Yves Lafon · Jack Jansen · Michael Hausenblas · Rik Van de Walle Received: date / Accepted: date E. Mannens Ghent University - IBBT ELIS - Multimedia Lab Ghent, Belgium E-mail: [email protected] D. Van Deursen E-mail: [email protected] Ghent University - IBBT ELIS - Multimedia Lab Ghent, Belgium E-mail: [email protected] R. Troncy EURECOM, Multimedia Communications Department Sophia Antipolis, France E-mail: [email protected] S. Pfeiffer Vquence Sydney, Australia E-mail: silviapfeiff[email protected] C. Parker Kyoto University Kyoto, Japan E-mail: [email protected] Y. Lafon W3C/ERCIM Sophia Antipolis, France E-mail: [email protected] J. Jansen CWI, Distributed Multimedia Languages and Infrastructures Amsterdam, Netherlands E-mail: [email protected] M. Hausenblas National University of Ireland, Digital Enterprise Research Institute - LiDRC Galway, Ireland E-mail: [email protected] R. Van de Walle E-mail: [email protected] Ghent University - IBBT 2 Abstract To make media resources a prime citizen on the Web, we have to go beyond simply replicating digital media files. The Web is based on hyperlinks between Web resources, and that includes hyperlinking out of resources (e.g. from a word or an image within a Web page) as well as hyperlinking into resources (e.g. fragment URIs into Web pages). To turn video and audio into hypervideo and hyperaudio, we need to enable hyperlinking into and out of them.
    [Show full text]