<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>April</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>How to Apply Linked Data Principles to Multimedia Fragments</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Michael Hausenblas</string-name>
          <email>michael.hausenblas@deri.org</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tobias Bürger</string-name>
          <email>tobias.buerger@sti2.at</email>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Raphaël Troncy</string-name>
          <email>raphael.troncy@cwi.nl</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yves Raimond</string-name>
          <email>yves.raimond@bbc.co.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>BBC Audio &amp; Music interactive</institution>
          ,
          <country country="UK">United Kingdom</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>CWI Amsterdam</institution>
          ,
          <country country="NL">The Netherlands</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>DERI, National University of</institution>
          ,
          <addr-line>Ireland, Galway</addr-line>
          ,
          <country country="IE">Ireland</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>STI Innsbruck</institution>
          ,
          <country country="AT">Austria</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2009</year>
      </pub-date>
      <volume>20</volume>
      <issue>2009</issue>
      <abstract>
        <p>In this paper, we introduce interlinking multimedia (iM), a pragmatic way to apply the linked data principles to fragments of multimedia items. We report on use cases showing the need for retrieving and describing multimedia fragments. We then introduce the principles for interlinking multimedia in the Web of Data, discussing potential solutions which sometimes highlight controversial debates regarding what the various representations of a Web resource span. We finally present methods for enabling a widespread use of interlinking multimedia.</p>
      </abstract>
      <kwd-group>
        <kwd>Linked Data</kwd>
        <kwd>Media Fragments</kwd>
        <kwd>Media Annotations</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>MOTIVATION</title>
      <p>Multimedia content is easy to produce but rather hard
to find and to reuse on the Web. Digital photographs can
be easily uploaded, communicated and shared in
community portals such as Flickr, Picasa and Riya, while video are
available on portals such as YouTube, DailyMotion,
Metacafe or Vimeo to name a few. These systems allow their users
to manually tag, comment and annotate the digital content,
but they lack a general support for fine-grained semantic
descriptions and look-up, especially when talking about things
“inside” multimedia content, such as an object in a video or
a person depicted in a still image.</p>
      <p>Figure 1 illustrates the problem. A photo is host and
shared on a well-known photo portal. One can draw and tag
a particular region (“notes” in Flickr) in a picture to state
that this region actually depicts a certain person. Both the
region definition and the annotation are represented in a
proprietary format and locked-into the system. It can not
be used by other applications across the Web.</p>
      <p>
        For addressing this issue, the multimedia semantics
community has focused on developing so-called multimedia
ontologies [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] where fragments are firstly defined before being
used in semantic annotations. This approach has not yet
proved it can scale the Web. Worst, it also means an
indirection since the aboutness of a semantic description is a
piece of XML document defining a multimedia fragment and
not the fragment itself. On the other hand, the linked data
community has successfully applied the linked data
principles to publicly available datasets offering a light-weight,
scalable solution for annotating web resources.
      </p>
      <p>The aim of this paper is to discuss how to implement the
linked data principles along with media fragments, yielding
what we call “interlinking Multimedia” (iM)1. The
contribution of this paper is a theoretical and practical framework
of the required technologies and discussions of their
suitability in order to be compliant with the Web architecture.
We present a technology that enables to address multimedia
fragments in URIs and we critically review some interlinking
methods.</p>
      <p>The remainder of the paper is structured as follows. In
the Section 2, we discuss use cases and requirements for
addressing and describing multimedia fragments. Based on
these use cases, we state the interlinking multimedia
principles as follows:
1. Apply linked data principles for fine-grained
identification and description of spatio-temporal multimedia
fragments (Section 3);
2. Deploy legacy multimedia metadata formats such
as EXIF, ID3, XMP on the Web of Data (Section 4);
3. Discuss a set of specialized interlinking methods
for multimedia (Section 5).</p>
      <p>We reflect our work in the light of the existing work in
Section 6 and finally conclude in Section 7.
2.</p>
    </sec>
    <sec id="sec-2">
      <title>USE CASES AND REQUIREMENTS</title>
      <p>
        In the previous section, we have made a case for
addressing and describing a particular region of an image (anchor
value in [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]). The current Web architecture [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] provides a
means for uniquely identifying sub-parts of resources using
URI fragment identifiers (e.g. for referring to a part of an
HTML or XML document). However, for almost any other
media type, the semantics of the fragment identifier has
either not been specified or is not commonly accepted.
Providing an agreed upon way to localize sub-parts of multimedia
objects (e.g. sub-regions of images, temporal sequences of
videos or tracking moving objects in space and in time) is
fundamental [
        <xref ref-type="bibr" rid="ref30">30</xref>
        ]. In this section, we report on use cases
for interlinking multimedia on the Web (Section 2.1) and we
extract the requirements for iM (Section 2.2).
2.1
      </p>
    </sec>
    <sec id="sec-3">
      <title>Use Cases</title>
      <p>
        We observe a general demand in several communities for
annotation tools enabling to specify links between compound
objects or parts within these objects as well as the type of
these relationships. For example, media researchers want to
annotate and interrelate segments between books,
screenplays or different film versions [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ]. The following four use
cases further demonstrate the need for interlinking
multimedia in the Web of Data.
2.1.1
      </p>
      <sec id="sec-3-1">
        <title>Sharing video clips on Twitter</title>
        <p>Silvia is a big fan of Tim’s research keynotes. She used
to watch numerous videos starring Tim for following his
research activities and often would like to share the highlight
announcements with her collaborators. The company that
she runs uses the Twitter2 service to communicate and stay</p>
        <sec id="sec-3-1-1">
          <title>1http://www.interlinkingmultimedia.info/</title>
        </sec>
        <sec id="sec-3-1-2">
          <title>2http://twitter.com/</title>
          <p>connected among business partners. Silvia is interested in
TweeTube3 that will allow her to share video directly on
Twitter but she would like to point and reference only small
temporal sequences of these longer videos. She would like
to have a simple interface, similar to VideoSurf4, to edit the
start and end time points delimiting a particular sequence,
and get back in return the media fragment URI to share
with the rest of the world. She would also like to send her
comments and (semantic) annotations about this particular
video fragment.
2.1.2</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>Annotating faces in my personal photos</title>
        <p>Michael is really enthusiastic with the new features of the
iPhoto 2009 suite5. He went through the guide and he is
happy to see that annotating people on his personal
photos will be easier than ever. As soon as Michael uses the
software, it learns who Michael’s friends and relatives are,
and suggests annotations as faces are automatically
recognized in his pictures using some visual processing techniques.
Bounding boxes around faces are therefore drawn on the
photos and can be exported on Flickr for sharing. Michael
is also a Linked Data guru and would like to tag his
photos not with the name of his friends, but with the URI that
identifies them on the Web of Data. Michael stores RDF
annotations of all these spatial fragment URIs and hopes to
create an artistic collage of his family.
2.1.3</p>
      </sec>
      <sec id="sec-3-3">
        <title>Tracking your favorite artists on BBC Music</title>
        <p>BBC Music6 aims to provide a comprehensive guide to
music content across the BBC. The service provides
information about artists who appear on BBC programmes7 or
who have been covered in one of their reviews. Each artist is
interlinked with biographical information (where available)
supplied by Wikipedia, and with BBC programmes in which
it has been played. Frank loves this service so he can quickly
and easily find the kind of shows that might suit his taste.
He issues a query for his two favorite German composers,
Wagner and Strauss, and gets a list of media fragment URIs
pointing to segments from tracklists of various performances
broadcasted this week.
2.1.4</p>
      </sec>
      <sec id="sec-3-4">
        <title>Watching named video clips on a mobile phone</title>
        <p>Yves is a busy person. He doesn’t have time to attend all
meetings that he is supposed to. These are generally video
recorded and podcasted on the internal Web site of his
company together with a full text speech transcription aligned
with the video. Yves often uses his mobile smart phone
for accessing Web resources while traveling. He receives a
daily digest email from the system containing a list of media
fragment URIs pointing to video clips where his name has
been pronounced, together with the term ’ACTION’,
during meetings. While on his next trip, Yves goes though his
email backlog and watches the video clips by simply clicking
on the links. The media server of his company dynamically
composes a valid media resource from the URIs that Yves
is requesting, so that Yves’ video player just plays the right
sequences where an action has been given to him.</p>
        <sec id="sec-3-4-1">
          <title>3http://www.tweetube.com/</title>
        </sec>
        <sec id="sec-3-4-2">
          <title>4http://www.videosurf.com/</title>
        </sec>
        <sec id="sec-3-4-3">
          <title>5http://www.apple.com/ilife/iphoto/</title>
        </sec>
        <sec id="sec-3-4-4">
          <title>6http://www.bbc.co.uk/music/beta</title>
        </sec>
        <sec id="sec-3-4-5">
          <title>7http://www.bbc.co.uk/programmes</title>
          <p>2.2</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Requirements</title>
      <p>
        Several aspects and requirements for iM have already been
discussed [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. In this section, we argue that interlinking
multimedia resources at a fine-grained level requires to deal with
the addressing and the description of multimedia fragments.
2.2.1
      </p>
      <sec id="sec-4-1">
        <title>Addressing Multimedia Fragments</title>
        <p>We are working within the W3C Media Fragments
Working Group8 to provide a URI-based mechanism to address
fragments of image, audio and video resources. Our first
assumption is that an audio or a video resource has a single
unified timeline.</p>
        <p>The use cases described in the Section 2.1 highlight the
need for being able to address parts of multimedia resources.
By parts, we mean playable resources that can be extracted
from the parent resources according to a number of
dimensions, namely: time, space, track and names. The time
dimension allows to address a temporal sequence of an audio
or a video resource. The space dimension allows to address a
rectangle bounding box of a frame or still image. The track
dimension allows to address a particular track of multimodal
resource if the container format exposes such a notion.
Finally, the name dimension is a convenient shortcut for a
combination of any of the three other dimensions that can
be further referred by a name under the condition that the
container format allows such a markup (e.g. a chapter name
in a DVD).</p>
        <p>Numerous codec and container formats are used on the
Web. A URI denoting a media fragment should be agnostic
on these formats but bound to what they can expose in the
compressed domain9 (i.e. without any further transcoding
operation). The HTTP protocol should at least be
supported while we observe that any solution should be
compatible with the notion of media fragment defined in the
RTSP protocol10.</p>
        <p>
          Media fragments are really parts of a parent resource. The
use of URI fragment [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] seems therefore appropriate to
specify these media fragments. As for any URI fragment, access
to the parent URI shall be possible in order to inspect its
context.
2.2.2
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>Describing Multimedia Fragments</title>
        <p>
          In order to be become part of the LOD cloud, iM must
follow the linked data principles (see Section 3.1). Metadata
descriptions have to be interoperable in order to reference
and integrate parts of the described resources. The diversity
of media content types, application scenarios and domains
directly translates to the existence of a huge number of
(partially) diverse metadata formats [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ].
        </p>
        <p>
          The integration of these formats, though often desirable [
          <xref ref-type="bibr" rid="ref31">31</xref>
          ]
is difficult due to syntactic and semantic interoperability
problems. Solutions should further take into account the
characteristics of multimedia whose semantics – when
interpreted by a user – are typically derived based on his/her
experiences, culture and knowledge. Thus, solutions should
consider provenance information and contextual
information (e.g. who says what and when) when describing
fragment of multimedia resources. These issues are in
particu
        </p>
        <sec id="sec-4-2-1">
          <title>8http://www.w3.org/2008/WebVideo/Fragments/</title>
          <p>9The abilities and limitations of most of the multimedia
formats are described in http://www.w3.org/2008/WebVideo/
Fragments/wiki/Types_of_Fragment_Addressing
10http://www.ietf.org/rfc/rfc2326.txt
lar addressed within the W3C Media Annotations Working
Group11 and in other related forums.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>3. IM PRINCIPLES</title>
      <p>As discussed above, the motivation for introducing iM
stems from the fact that we currently do not have proper
means to address and describe fragments of multimedia
resources in the Web of data. We believe that we can
overcome these limitations by defining a URI-based mechanism
to address media fragments and by applying the linked data
principles to those fragments. Our ultimate goal is to derive
both semantic and audio-visual representations from
multimedia resources on the Web.</p>
      <p>
        In the context of the Web of Data, we deal with
documents (e.g. a JPG file) and things (e.g. a person). The
former is generally called an Information Resource while the
later will be referred as a Non-Information Resource12. The
W3C’s “Architecture of the World Wide Web, Volume One”
(AWW) [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] specifies that globally unique identifiers (URIs)
are used to denote both things and documents. In the
following, we first present the linked data principles (Section 3.1)
before describing how iM can be deployed for both
information resources (Section 3.2) and non-information resources
(Section 3.3).
      </p>
      <p>In both cases, we consider a related issue regarding the
addressing versus the description of fragments of multimedia
resources. The former case will refer to the ability of getting
only the multimedia fragment served using the Web
architecture while the latter will refer to the ability of getting
semantic metadata about the media fragment.
3.1</p>
    </sec>
    <sec id="sec-6">
      <title>Linked Data Principles</title>
      <p>
        The basic idea of linked data was outlined by Tim
BernersLee [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]:
1. All resources should be identified using URIs;
2. All URIs should be dereferenceable, that is HTTP URIs,
as it allows looking up the resources identified;
3. When looking up an URI, it leads to more (useful)
data about that resource;
4. Links to other URIs should be included in order to
enable the discovery of more data.
      </p>
      <p>We note that these four linked data principles are agnostic
regarding the type of the resource. In the following, we
revisit these four principles for interlinking multimedia. In
particular, we discuss how looking up an URI to lead to
more data can be realized.</p>
    </sec>
    <sec id="sec-7">
      <title>3.2 iM for Information Resources</title>
      <p>Let’s imagine that Silvia would like to share with her
colleagues a specific part of a recent podcast from Tim Berners
Lee on the BBC. More particularly, she is interested in
sending a link pointing to the sequence comprised between the
seconds 15 and 45 of this podcast. Following the temporary
Media Fragments URI syntax13, Silvia will build the URI
http://www.example.org/myPodcast.mp3#t=15,45.
11http://www.w3.org/2008/WebVideo/Annotations/
12An ontology implementing the concepts discussed in the
Generic URIs “Design Issues” note is available at http://
www.w3.org/2006/gen/ont.rdf
13http://www.w3.org/2008/WebVideo/Fragments/wiki/
Syntax
3.2.1</p>
      <sec id="sec-7-1">
        <title>Retrieving Media Fragments</title>
        <p>Conrad is working with Silvia. He is interested in listening
to just this sequence, and would prefer to not download the
one hour podcast. The following interaction could happen
between his browser (the user agent) and the server. Conrad
has a smart user agent that can further process requests
containing a media fragment by stripping out the fragment
part, but encoding it into a range header. The following
GET command will therefore be issued:
GET http://www.example.org/myPodcast.mp3
Accept: application/mp3
Range: seconds=15-45</p>
        <p>The server has a module for slicing on demand
multimedia resources, that is, establishing the relationship between
seconds and bytes, extract the bytes corresponding to the
requested fragment, and add the new container headers in
order to serve a playable resource. The server will then reply
with the closest inclusive range in a 206 HTTP response:
HTTP/1.1 206 Partial Content
Accept-Ranges: bytes, seconds
Content-Length: 1201290
Content-Type: audio/mpeg
Content-Range: seconds 14.875-45.01/321</p>
        <p>The user agent will then have to skip 0.125s to start
playing the multimedia fragment as 15s. We observe that the
relationship between bytes and seconds is in the general case
unknown. The Media Fragment WG is considering a
technical solution involving a second rountrip between the user
agent and the server for establishing this mapping.
3.2.2</p>
      </sec>
      <sec id="sec-7-2">
        <title>Describing Media Fragments</title>
        <p>Conrad is also interested in retrieving the semantic
description of this fragment to feed his semantic web agent.</p>
        <p>He could issue a similar request changing simply the accept
header:
GET http://www.example.org/myPodcast.mp3
Accept: application/rdf+xml
Range: seconds=15-45</p>
        <p>Providing an adequate configuration, the server could
return an RDF file containing the semantic annotations of this
media fragment. The additional “HTTP Link: header”
proposal14 could further establish the relationship between the
mp3 file and the RDF file using the rdfs:seeAlso property,
the resource referenced by the request URI being then the
subject of the assertion.
3.2.3</p>
        <p>
          In this scenario, we have considered that the podcast
resource is available in several different representations.
Therefore, content negotiation can be used to serve alternatively
an excerpt of the audio file, or a semantic description of this
fragment depending on the HTTP accept header. We
observe for example that this is how the jigsaw web server [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ]
is configured. As benefits, it works right away with all
textbased browsers (lynx, emacs with emacsspeak, etc.) and the
output can be rendered directly by selecting, e.g., the
transcript of the audio file contained in the description from the
RDF file. An RDF crawler will be able to get all the
descriptions of a collection of audio files to create a knowledge
database, just by asking for the right MIME type.
        </p>
        <p>However, this solution boils down to say that the RDF
description of the podcast and the podcast itself are both
representations of the same information resource. On one
hand, we observe that the Web Accessibility Guidelines15
define two equivalent content when both fulfill essentially
the same function or purpose upon presentation to the user.</p>
        <p>For example, the text “Tim Berners Lee promoting the Web
of Data” might convey the same information as an excerpt
of an audio podcast when presented to (deaf) users. On the
other hand, some voices16 in the Web of Data community
consider that a description of a multimedia resource is not
the same as the resource itself since it cannot convey all
its perceptual and cognitive effect. Consequently, following
this way of thought, content negotiating between these two
resources would be just plain wrong. Embedding the
multimedia resource within an HTML document on one side,
and providing RDF metadata within this document on the
other side would, however, work since metadata about the
resource would convey the same information as the resource
and thus can be subject to content negotiation. We
conclude that discovery and HTTP17 is a controversial issue for
which we advocate a TAG resolution.</p>
        <p>
          The use of a URI fragment for addressing a media
fragment is also problematic. The URI RFC states [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]:
“The semantics of a fragment identifier are
defined by the set of representations that might
result from a retrieval action on the primary
resource. The fragment’s format and resolution is
therefore dependent on the media type [RFC2046] [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]
of a potentially retrieved representation, even though
such a retrieval is only performed if the URI is
dereferenced. If no such representation exists,
then the semantics of the fragment are
considered unknown and are effectively unconstrained.
        </p>
        <p>Fragment identifier semantics are independent of
the URI scheme and thus cannot be redefined by
scheme specifications.”
14http://tools.ietf.org/html/
draft-nottingham-http-link-header-04
HTTP/1.1 200 OK
Accept-Ranges: bytes, seconds Therefore, it might be necessary to register new media-types
Content-Length: 1088 defining the semantics of a fragment for each media formats
Content-Location: http://www.example.org/myPodcast.rdf using for example sub-class/class hierarchies provided by the
Content-Type: application/rdf+xml IANA registry.</p>
        <p>Link: &lt;http://www.example.org/myPodcast.mp3&gt;;
rel="http://www.w3.org/2000/01/rdf-schema#seeAlso";
Vary: accept
15http://www.w3.org/TR/WAI-WEBCONTENT-TECHS/
#glossary
16http://chatlogs.planetrdf.com/swig/2009-02-09.
html#T15-09-20
17http://www.hueniverse.com/hueniverse/2008/09/
discovery-and-h.html</p>
      </sec>
    </sec>
    <sec id="sec-8">
      <title>3.3 iM for General Resources</title>
      <p>Let’s now imagine that Michael would like to annotate
some specific parts of his personal photos. More precisely,
he would like to highlight the face of his new born daughter
and send a media fragment URI to his family and relatives
(Figure 2). Following the temporary Media Fragments URI
syntax, Michael will build the URI http://www.example.
org/children#xywh=40,10,100,100. He can then further
produce the annotation depicted in the listing 1:
1 &lt; http :// www . example . org / children # xywh =40 ,10 ,100 ,100 &gt;
foaf : depicts : Saphira .</p>
      <p>Listing 1: Description of a spatial region depicting
a person.</p>
      <p>The media fragment URI denotes here a part of a thing.
Let’s have a look at the communication between a (smart)
user agent and a server with such a URI request. In case
the accept header is image/jpeg:
GET http://www.example.org/children
Accept: image/jpeg
Range: pixels=40,10,100,100</p>
      <p>The server will answer with a 307 Temporary Redirect18
response indicating the location of the image file.
HTTP/1.1 307 Temporary Redirect
Location: http://www.example.org/children.jpg
Content-Type: image/jpeg</p>
      <p>This response code requires that the request should be
repeated with another URI, for example the one specified in
the Location header. Similarly, a GET with a different
accept header (e.g. application/rdf+xml) could return a 307
response code with the Location header pointing to the
location of the RDF file. We conclude that apply iM for general
resources might imply one more round-trip due to the fact
that the URI denotes a non-information resource.
18The Web of Data community tends to use the 303 See Other
response code. However, this HTTP response has originally
been defined for changing the HTTP (verb) method, and
the 307 code seems to be more appropriate in our case.
3.4</p>
    </sec>
    <sec id="sec-9">
      <title>Summary</title>
      <p>Let us review the adequacy of the four general linked data
principles against the iM principles stated above. We
acknowledge that the first and the second principle has been
respected as, for example, http://example.org/children.
jpg#xywh=40,10,100,100 is a valid HTTP URI and the
second principle can easily be done substituting e.g., :Saphira
with a DBpedia resource.</p>
      <p>The problem comes down to the third principle: “When
looking up an URI, it leads to more (useful) data about that
resource”. Hence, when dereferencing http://www.example.
org/myPodcast.mp3#t=15,45, what we want to have is both
an audio and a semantic representation of the resource, i.e.
the bytes corresponding to a particular playable sequence of
an audio file and the RDF triples that might describe its
transcript. We have argued that content negotiation could
technically be used for this purpose but that it is unclear
whether it is appropriate with the spirit of the Web
architecture. We have also argued that the link header proposal
is a viable alternative.</p>
      <p>We observe also that there are cases where multimedia
resources “embed” semantic annotations. For example, an
image can have some XMP metadata represented in RDF
stored in its header. We finally note that the whole RDF
file is systematically returned when the user agent requests
it. The problem is that it is hard to map the fragment of
a multimedia resource with the corresponding piece in the
RDF description of this resource.
4.</p>
    </sec>
    <sec id="sec-10">
      <title>LEGACY MULTIMEDIA METADATA</title>
      <p>
        We have described in the previous section a technology
that realizes the first iM principle: how to identify in a URI
a fragment of a multimedia resource. The second principle
states that legacy multimedia metadata formats should be
deployed. Actually, descriptions about multimedia resources
must be interoperable in order to enable the interlinking of
the described resources. Semantic technologies have already
been considered as a viable solution to leverage these
interoperability issues [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ]. For the description of multimedia
resources, a plethora of metadata formats are in use, causing
interoperability issues. Further, a multitude of multimedia
ontologies have been identified [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
      </p>
      <p>Describing content using multimedia ontologies, however,
also causes some problems as different ontologies are most
often not aligned with each others. For example, different
names are used for its elements which again hinders their
integration. This problem of capturing the semantics of the
various multimedia formats and of aligning them is currently
tackled by the W3C Media Annotations Working Group19.</p>
      <p>
        Alternatively, we have recently proposed ramm.x20 (“RDFa
enabled multimedia metadata”), a proposal to integrate
various descriptions based on different metadata standards [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
The basic idea of ramm.x is to define a light-weight
deployment description vocabulary allowing—deployed with
RDFa [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]—a Semantic Web agent to determine the
formalisation steps in order to process the native multimedia
metadata format. This, in turn, allows a Semantic Web agent to
determine what a multimedia object is about which enables
19A mapping table between numerous multimedia formats
is available at http://dev.w3.org/2008/video/mediaann/
mediaont-1.0/mapping_table_090223_common.htm
20http://sw.joanneum.at/rammx/spec/
him to set links at least semi-automatically.
      </p>
    </sec>
    <sec id="sec-11">
      <title>METHODS FOR INTERLINKING MUL</title>
    </sec>
    <sec id="sec-12">
      <title>TIMEDIA</title>
      <p>Finally, the third iM principle states that specialized
interlinking methods can be used to effectively create interlinking
between multimedia resources at a fine-grained level. In this
section, we critically review various methods and tools
generally used for interlinking resources in the Web of Data.
5.1</p>
    </sec>
    <sec id="sec-13">
      <title>Manual Methods</title>
      <p>
        We have recently introduced User Contributed
Interlinking (UCI) [
        <xref ref-type="bibr" rid="ref14 ref8">8, 14</xref>
        ], a manual interlinking methodology which
relies on the end user as a source of qualitative information.
UCI has been applied to enrich the Eurostat dataset [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
A recent proposal, called CaMiCatzee [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] implements UCI
for multimedia. CaMiCatzee allows people to semantically
annotate picture on Flickr and to query for person’s using
their FOAF documents, URIs or person names.
      </p>
      <p>
        Manual method for interlinking multimedia could be
combined with incentives such as Game Based Interlinking (GBI),
following the principles set forward by Louis van Ahn with
his games with a purpose21 [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ]. One approach is to make
the interlinking of resources fun and to hide the task of
interlinking behind games. This is related to UCI but with
the main difference that the user is not aware of him
contributing links as his task is hidden behind a game.
      </p>
      <p>GBI seems to be a promising direction for multimedia
interlinking. The most interesting examples to build on are
Ahn’s ESP games in which users are asked to describe
images, or Squigl22 in which users are asked to trace objects in
pictures. Another interesting approach is followed by
OntoGame whose general aim is to find shared
conceptualizations of a domain. OntoGame players are asked to describe
images, audio or video files. Users are awarded if they
describe content in the same way. Further exemplary games
are OntoTube, Peekaboom, or ListenGame which hide the
complexity of the annotation process of videos, images or
audio files respectively, behind entertaining games. These
approaches together with appropriate browsing interfaces
for multimedia assets could be a promising starting point
to let users draw meaningful relations between objects and
their parts.
5.2</p>
    </sec>
    <sec id="sec-14">
      <title>Collaborative Interlinking</title>
      <p>
        Collaborative approach to interlinking of resources could
be followed using Semantic Wikis. Semantic Wikis extend
the principles of traditional Wikis such as collaboration, easy
use, linking and versioning with means to type links and
articles via semantic annotations [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ]. Some of the systems
support the annotation of multimedia objects including
Semantic Wikis with dedicated multimedia support such as
Ylvi [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ], MultiMakna [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]. Most of these systems however
treat a multimedia object as part of an article in which they
appear. Thus, they do not allow specific annotations of it
or treat them in the same manner like articles which can be
only annotated globally. MultiMakna allows to assign
annotations to temporal segments in videos through the use of an
appliesT o-relation. While annotations may be constrained
21http://www.gwap.com/
22http://www.gwap.com/gwap/gamesPreview/squigl/
to its temporal context, to the best of our knowledge, links
can only be established between articles and not segments.
      </p>
      <p>
        Another Semantic Wiki with multimedia support is
MetaVidWiki (MVW)23 which enables community engagement
with audio/visual media assets and associative temporal
metadata. MVW extends the popular Semantic MediaWiki [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]
with media specific features such as streaming, temporal
metadata, and viewing and editing of video sequences. MVW
supports the addressing and linking between temporal
fragments. Segments of videos can be treated like “articles”,
referenced via URIs which support time intervals according
to the temporalURI specification [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ] and metadata about
them can be exported in CMML [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ].
5.3
      </p>
    </sec>
    <sec id="sec-15">
      <title>Semi-Automatic Methods</title>
      <p>Semi-automatic interlinking methods consist in combining
multimedia-analysis techniques with human feedback.
Analysis techniques can process the content itself or the context
surrounding the content such as the user profile in order
to suggest potential interlinking. The user would need to
accept, reject, modify or ignore those suggestions.
Inspiration for this type of approach can be found in the area of
semi-automatic multimedia annotation.</p>
      <p>
        Emergent Interlinking (EI) is another approach based on
the principles of Emergent Semantics whose aim is to
discover semantics through observing how multimedia
information is used [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. This can be essentially accomplished by
putting multimedia resources in context-rich environments
being able to monitor the user and his behavior. In these
environments, two different types of context are present: (i)
static or structural context, which is derived from the way
how the content is placed in the environment (e.g. a Web
page) and (ii) dynamic context, which is derived from the
interactions of the user in the environment (e.g. his
browsing behavior, which links he follows, or on which object he
zooms). The assumption is that in appropriate
environments, the browsing path of a user is semantically coherent
and thus allows to derive links between objects which are
semantically close to each other.
5.4
      </p>
    </sec>
    <sec id="sec-16">
      <title>Automatic Methods</title>
      <p>Finally, automatic interlinking of fragments of a
multimedia resource can be achieved by purely analyzing its content.
For example, in the case of such a musical audio content,
the audio signal can be analyzed in order to derive a
temporal segmentation. The resulting segments can be
automatically linked to musically relevant concepts, e.g. keys,
chords, beats or notes. The media fragment URI
specification described in the Section 3.2 can then be used to relate
these different segments to fragments of audio content.</p>
      <p>
        The Music Ontology [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ] provides a framework for the
temporal annotation of audio signals. An audio file encodes
an audio signal. This audio signal is linked to its timeline24,
i.e. its temporal backbone. The Event ontology25 can then
be used to classify particular regions of such timelines. For
example, we define here two classifiers, capturing two chorus
and two verses.
@prefix ps: &lt;http://purl.org/ontology/pop-structure/&gt;.
@prefix event: &lt;http://purl.org/NET/c4dm/event.owl#&gt;.
23http://metavid.org/wiki/
24http://purl.org/NET/c4dm/timeline.owl
25http://purl.org/NET/c4dm/event.owl
ps:Chorus rdfs:subClassOf event:Event .
ps:Verse rdfs:subClassOf event:Event .
_:evt a af:KeyChange;
event:time [tl:at ?start; tl:timeline ?tl] ;
af:new_key ?key }
5.5
      </p>
    </sec>
    <sec id="sec-17">
      <title>Discussion</title>
      <p>We can now use these two classifiers for annotating
fragments of an audio signal. In the following, we describe the
first chorus and the first verse in the the Beatles “Can’t buy
me love”. We use two events, classifying two regions of the
corresponding audio signal’s timeline.</p>
      <p>
        Currently, most interlinks between web datasets are
generated entirely automatically, using heuristics to determine
when two resources in two web datasets identify the same
object [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ]. However, this interlinking is done on the basis of
existing RDF descriptions, and cannot directly be applied on
@prefix mo: &lt;http://purl.org/ontology/mo/&gt;. multimedia objects. Such automated interlinking algorithms
@prefix event: &lt;http://purl.org/NET/c4dm/event.owl#&gt;. can only be applied if some RDF statements about the
mul@prefix tl: &lt;http://purl.org/NET/c4dm/timeline.owl#&gt;. timedia object (e.g. the performer involved in a particular
@prefix ps: &lt;http://purl.org/ontology/pop-structure/&gt;. recording) exist. These automated interlinking algorithms
@prefix rdfs: &lt;http://www.w3.org/2000/01/rdf-schema#&gt;. can be adapted to include an analysis step, extracting some
@prefix : &lt;#&gt;. information from the content of a multimedia object before
deriving interlinks. The Henry software described in
Sec:signal owl:sameAs tion 5.4 is one example of automated interlinking performed
&lt;http://zitgist.com/music/signal/eb20ee61-414f-4eee- by analyzing the multimedia object itself. However, the
ac8dce-190db516a466&gt;. curacy of the derived links is debatable, and heavily depends
on the underlying analysis algorithm.
:signal mo:time [ Another large source of interlinks on the current data Web
tl:duration "PT2M14S"; comes from the work of large communities. For example, the
tl:timeline :tl; Musicbrainz community created the links between artists
]. in Musicbrainz and the corresponding artists in DBpedia.
:chorus1 a ps:Chorus; However, manual interlinking of media fragments is tedious,
rdfs:label "First chorus"; as lots of different annotations can be done. For example, a
event:time [ musical track could be described by its structural segments,
tl:timeline :tl; by notes being played, by performers playing, by beats, etc.
tl:start "PT0S"; Therefore, large communities need to be involved in that
tl:duration "PT9S"; process for it to be successful, perhaps using the
collabo]. rative annotation frameworks mentioned in Section 5.2 and
:verse1 a ps:Verse; the emergent interlinking mentioned in Section 5.3.
rdfs:label "First verse"; A possible solution to make such annotations scale would
event:time [ be to combine both approaches. Automated interlinking
altl:timeline :tl; gorithms could post the resulting interlinks to a Semantic
tl:start "PT9S"; Wiki, where these links could be reviewed and modified by
tl:duration "PT33S"; the community. Automated interlinking algorithm would
]. then kickstart the interlinking process, and the resulting
interlinks would gradually become more accurate.
      </p>
    </sec>
    <sec id="sec-18">
      <title>RELATED WORK</title>
      <p>
        An application of automatic interlinking of media
fragments in the music domain is Henry26 [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ]. Henry
aggregates music processing workflows available on the Web and
applies them on audio signals to dynamically derive
temporal segmentations and interlink these different segments
with Web identifiers for music-related concepts.
      </p>
      <p>For example, the following SPARQL query issued to Henry
will dynamically generate annotations corresponding to changes
of musical keys , described in the same way as in the above
example. When processing the query, Henry accesses the
audio file
http://dbtune.org/audio/Both-Axel.ogg and applies a key
extraction workflow to derive the temporal annotations. Each
annotation is linked to a Web identifier corresponding to a
musical key. Henry then binds the times at which the key
change to the variable ?start, and the Web identifiers for
musical keys to the variable ?key.</p>
      <p>Related work can be traced back to hypermedia research.</p>
      <p>
        An hypermedia document [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] refers to a collection of
information units including information about synchronization
between these units and about references between them.
      </p>
      <p>
        Temporal and spatial dimensions are typically included,
whereas references can be made between parts in both
dimensions. The issue of linking in hypermedia is discussed in [
        <xref ref-type="bibr" rid="ref10 ref11 ref19">10,
19, 11</xref>
        ]. Linking within multimedia presentations, within
and among linear and non-linear multimedia presentations
is discussed in [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] discusses links in time-based
preselect ?start ?key where { sentations and proposes a method to specify the context of
&lt;http://dbtune.org/audio/Both-Axel.ogg&gt; mo:encodes ?sig.links, i.e., what happens with the source or destination
pre?sig mo:time ?time. sentation of a link when it is traversed.
?time tl:timeline ?tl. Hypermedia presentations consist of both static and
dynamic media objects which are grouped together in so-called
composite entities. Parts of these entities, identified via
an
      </p>
      <p>In this section, we discuss related work and previous
attempts for defining URI-based mechanisms for defining
media fragments.
6.1</p>
    </sec>
    <sec id="sec-19">
      <title>Hypermedia Links</title>
      <p>chors that provide hooks for links, can be linked with each
other and the behaviour of source and destination entities
can be defined (e.g. shall the source video be paused or
replaced). The ideas discussed above were implemented in
the “Synchronized Internet Markup Langauge” (SMIL), a
W3C recommendation that enables the integration of
independent multimedia objects such as text, audio, graphics or
videos into a synchronized multimedia presentation. Within
this presentation, an author can specify the temporal
coordination of the playback, the layout of the presentation
and hyperlinks for single multimedia components. The
latest version of SMIL provides a rich MetaInformation
module which allows the description of all elements of a SMIL
document using RDF. Media Fragments URI as defined in
the previous section can be legally used in conjunction with
SMIL documents.
6.2</p>
    </sec>
    <sec id="sec-20">
      <title>Addressing Multimedia Fragments</title>
      <p>
        Providing a standardized way to localize spatial and
temporal sub-parts of any non-textual media content has been
recognized as urgently needed to make video a first class
citizen on the Web [
        <xref ref-type="bibr" rid="ref30">30</xref>
        ].
      </p>
      <p>Previous attempts include non-URI based mechanisms.
For images, one can use either MPEG-7 or SVG snippet
code to define the bounding box coordinates of specific
regions. Assuming a simple multimedia ontology available
(designated with “mm:”) the following listing 2 provides a
semantic annotation of a region within an image: However,
1 &lt; http :// example . org / myRegion &gt; foaf : depicts : Saphira
;
2 rdf : type mm : ImageFragment ;
3 mm : topX " 40 px " ;
4 mm : topY " 10 px " ;
5 mm : width " 100 px " ;
6 mm : height " 100 px " ;
7 mm : hasSource &lt; http :// example . org / children . jpg &gt; .
Listing 2: Description of a spatial region depicting
a person using a dedicated multimedia ontology.
the identification and the description of the region is
intertwined and one needs to parse and understand the
multimedia ontology in order to access the multimedia fragment.</p>
      <p>
        URI-based mechanisms for addressing media fragments
have also been proposed. MPEG-21 specifies a normative
syntax to be used in URIs for addressing parts of any
resource but whose media type is restricted to MPEG [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ].
The temporalURI RFC 27 defines fragment of multimedia
resources using the query parameter (’ ?’) thus creating a
new resource. YouTube launched a first facility28 to
annotate parts of videos spatio-temporally and to link to
particular time points in videos. It uses the URI fragment (’#’)
but the whole resource is still sent to the user agent that just
perform a seek in the media file. In contrast, the solution we
are advocating allows to send only the bytes corresponding
to media fragments while being still able to cache them.
      </p>
    </sec>
    <sec id="sec-21">
      <title>CONCLUSION AND FUTURE WORK</title>
      <p>In this paper, we have described the principles behind
interlinking multimedia (iM), a pragmatic way to apply the
27http://www.annodex.net/TR/URI_fragments.html
28http://youtube.com/watch?v=UxnopxbOdic
linked data principles to fragments of multimedia resources.
We have presented a URI-based mechanism for addressing
parts of a multimedia resources following four dimensions
(time, space, track and name). Furthermore, we have shown
how these URIs can be used in the linked data context.
We have pointed out that the use of content negotiation
to serve alternatively media resource or description of these
resources is debatable, though some implementation exists.
We have stressed the importance of having mechanism to
deploy multimedia metadata using light-weight approach such
as ramm.x. Finally, we have presented various methods that
can be used to actually generate interlinks between
multimedia resources.</p>
      <p>The presentation of these technologies left a number of
challenging problems unsolved. It is unclear what content
negotiation (in spirit if not technically) should do in this
context. The semantics of media fragments is currently
undefined. Retrieving partial content from a video resource
given the definition of a temporal media fragment makes
sense while it is hard to find use cases for its spatial
counterpart. The Media Fragments WG is currently tackling these
issues. We finally plan to work on a general framework for
establishing the mapping between a media fragment and its
RDF description in the general case.</p>
    </sec>
    <sec id="sec-22">
      <title>Acknowledgements</title>
      <p>The authors would like to thank Richard Cyganiak, Yves
Lafon, Ivan Herman, Silvia Pfeiffer, Tim Berners Lee and
all the participants of the W3C Media Fragments Working
Group for their willingness to discuss the linked data
principles, the definition of media fragments and more generally
their adequacy within the Web architecture.</p>
      <p>The work of Michael Hausenblas has partly been
supported by the European Commission under Grant No. 231335,
FP7/ICT-2007.4.4 project “Intelligent metadata-driven
processing and distribution of audiovisual media” (iMP).
8.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>B.</given-names>
            <surname>Adida</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Birbeck</surname>
          </string-name>
          .
          <source>RDFa Primer 1</source>
          .0, W3C Working Group Note. http://www.w3.org/TR/xhtml-rdfa-primer/, 14
          <string-name>
            <surname>October</surname>
          </string-name>
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>R.</given-names>
            <surname>Arndt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Troncy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Staab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Hardman</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Vacura</surname>
          </string-name>
          . COMM:
          <article-title>Designing a Well-Founded Multimedia Ontology for the Web</article-title>
          .
          <source>In 6th International Semantic Web Conference (ISWC'07)</source>
          , pages
          <fpage>30</fpage>
          -
          <lpage>43</lpage>
          , Busan, Korea,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>T.</given-names>
            <surname>Berners-Lee</surname>
          </string-name>
          .
          <article-title>Linked Data, Design Issues</article-title>
          . http://www.w3.org/DesignIssues/LinkedData.html,
          <issue>27</issue>
          <year>July 2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>T.</given-names>
            <surname>Berners-Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Fielding</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L.</given-names>
            <surname>Masinter. Uniform Resource</surname>
          </string-name>
          <article-title>Identifier (URI): Generic Syntax, RFC3986</article-title>
          . IETF Network Working Group,
          <year>2005</year>
          . http://www.ietf.org/rfc/rfc3986.txt.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>T.</given-names>
            <surname>Bu</surname>
          </string-name>
          <article-title>¨rger and M. Hausenblas. Interlinking Multimedia - Principles and Requirements</article-title>
          . In International Workshop on Interacting with
          <article-title>Multimedia Content in the Social Semantic Web (IMC-</article-title>
          <source>SSW'08)</source>
          , pages
          <fpage>31</fpage>
          -
          <lpage>36</lpage>
          , Koblenz, Germany,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>N.</given-names>
            <surname>Freed</surname>
          </string-name>
          and
          <string-name>
            <given-names>N.</given-names>
            <surname>Borenstein</surname>
          </string-name>
          .
          <article-title>Multipurpose Internet Mail Extensions (MIME) Part Two: Media Types, RFC2046</article-title>
          . IETF Network Working Group,
          <year>1996</year>
          . http://www.ietf.org/rfc/rfc2046.txt.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>F.</given-names>
            <surname>Halasz</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Schwartz</surname>
          </string-name>
          .
          <article-title>The Dexter Hypertext Reference Model</article-title>
          .
          <source>Communications of the ACM</source>
          ,
          <volume>37</volume>
          (
          <issue>2</issue>
          ):
          <fpage>30</fpage>
          -
          <lpage>39</lpage>
          ,
          <year>1994</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>W.</given-names>
            <surname>Halb</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Raimond</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Hausenblas</surname>
          </string-name>
          .
          <article-title>Building Linked Data For Both Humans and Machines</article-title>
          .
          <source>In International Workshop on Linked Data on the Web (LDOW'08)</source>
          , Beijing, China,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>L.</given-names>
            <surname>Hardman</surname>
          </string-name>
          .
          <article-title>Modelling and Authoring Hypermedia Documents</article-title>
          .
          <source>PhD thesis</source>
          , CWI, Amsterdam,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>L.</given-names>
            <surname>Hardman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. C. A.</given-names>
            <surname>Bulterman</surname>
          </string-name>
          , and
          <string-name>
            <surname>G. van Rossum.</surname>
          </string-name>
          <article-title>Links in hypermedia: the requirement for context</article-title>
          .
          <source>In HYPERTEXT'93: 4th ACM Conference on Hypertext and Hypermedia</source>
          , pages
          <fpage>183</fpage>
          -
          <lpage>191</lpage>
          , Seattle, Washington, USA,
          <year>1993</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>L.</given-names>
            <surname>Hardman</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. van Ossenbruggen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. S.</given-names>
            <surname>Mullender</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Rutledge</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D. C. A.</given-names>
            <surname>Bulterman</surname>
          </string-name>
          .
          <article-title>Do you have the time? Composition and linking in time-based hypermedia</article-title>
          .
          <source>In HYPERTEXT'99: 10th ACM Conference on Hypertext and Hypermedia</source>
          , pages
          <fpage>189</fpage>
          -
          <lpage>196</lpage>
          , New York, NY, USA,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M.</given-names>
            <surname>Hausenblas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Bailer</surname>
          </string-name>
          ,
          <string-name>
            <surname>T.</surname>
          </string-name>
          <article-title>Bu¨rger, and</article-title>
          <string-name>
            <given-names>R.</given-names>
            <surname>Troncy</surname>
          </string-name>
          .
          <article-title>Deploying Multimedia Metadata on the Semantic Web (Poster)</article-title>
          .
          <source>In 2nd International Conference on Semantics And digital Media Technologies (SAMT'07)</source>
          , Genova, Italy,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>M.</given-names>
            <surname>Hausenblas</surname>
          </string-name>
          and
          <string-name>
            <given-names>W.</given-names>
            <surname>Halb</surname>
          </string-name>
          .
          <article-title>Interlinking Multimedia Data</article-title>
          .
          <source>In Linking Open Data Triplification Challenge at the International Conference on Semantic Systems (I-Semantics'08)</source>
          ,
          <year>2008</year>
          . http: //triplify.org/Challenge/Nominations?v=
          <fpage>51f</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>M.</given-names>
            <surname>Hausenblas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Halb</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Raimond</surname>
          </string-name>
          .
          <article-title>Scripting User Contributed Interlinking</article-title>
          . In 4th Workshop on
          <article-title>Scripting for the Semantic Web (SFSW'08)s</article-title>
          , Tenerife, Spain,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15] M. Hausenblas eds.
          <source>Multimedia Vocabularies on the Semantic Web, W3C Incubator Group Report</source>
          . http://www.w3.org/2005/Incubator/mmsem/ XGR-vocabularies/,
          <source>24 July</source>
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>I.</given-names>
            <surname>Jacobs</surname>
          </string-name>
          and
          <string-name>
            <given-names>N.</given-names>
            <surname>Walsh</surname>
          </string-name>
          .
          <source>Architecture of the World Wide Web</source>
          , Volume One, W3C Recommendation. http://www.w3.org/TR/webarch/,
          <source>15 December</source>
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>M.</given-names>
            <surname>Kro</surname>
          </string-name>
          ¨tzsch,
          <string-name>
            <given-names>D.</given-names>
            <surname>Vrandecic</surname>
          </string-name>
          , M. Vo¨lkel, H. Haller, and
          <string-name>
            <given-names>R.</given-names>
            <surname>Studer</surname>
          </string-name>
          . Semantic Wikipedia.
          <source>Journal of Web Semantics</source>
          ,
          <volume>5</volume>
          :
          <fpage>251</fpage>
          -
          <lpage>261</lpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Lafon</surname>
          </string-name>
          and
          <string-name>
            <given-names>B.</given-names>
            <surname>Boss</surname>
          </string-name>
          .
          <article-title>Describing and retrieving photos using RDF and HTTP</article-title>
          , W3C Note. http://www.w3.org/TR/photo-rdf/, 19 April
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Davis</surname>
          </string-name>
          , and
          <string-name>
            <given-names>W.</given-names>
            <surname>Hall</surname>
          </string-name>
          .
          <source>Hypermedia Links and Information Retrieval. 14th Information Retrieval Colloqium</source>
          , Lancaster University, http://eprints.ecs.soton.ac.uk/772/,
          <year>1992</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <source>[20] MPEG-21. Part</source>
          <volume>17</volume>
          :
          <article-title>Fragment Identification of MPEG Resources</article-title>
          . Standard No.
          <source>ISO/IEC 21000-17</source>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>L.</given-names>
            <surname>Nixon</surname>
          </string-name>
          and
          <string-name>
            <given-names>E.</given-names>
            <surname>Simperl</surname>
          </string-name>
          .
          <article-title>Makna and MultiMakna: towards semantic and multimedia capability in wikis for the emerging web</article-title>
          .
          <source>In Semantics'06</source>
          , Vienna, Austria,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>S.</given-names>
            <surname>Pfeiffer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Parker</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Pang</surname>
          </string-name>
          .
          <article-title>Specifying time intervals in URI queries and fragments of time-based Web resources</article-title>
          . http://www.annodex.net/TR/ draft
          <article-title>-pfeiffer-temporal-fragments-03</article-title>
          .html,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>S.</given-names>
            <surname>Pfeiffer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Parker</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Pang</surname>
          </string-name>
          .
          <article-title>The Continuous Media Markup Language (CMML), Version 2</article-title>
          .1. http://www.annodex.net/TR/ draft-pfeiffer-cmml-
          <volume>03</volume>
          .html,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>N.</given-names>
            <surname>Popitsch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Schandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Amiri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Leitich</surname>
          </string-name>
          , and
          <string-name>
            <given-names>W.</given-names>
            <surname>Jochum</surname>
          </string-name>
          .
          <article-title>Ylvi - Multimediaizing the Semantic Wiki</article-title>
          .
          <source>In 1st International Workshop on Semantic Wikis (SemWiki'06)</source>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Raimond</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Abdallah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sandler</surname>
          </string-name>
          , and
          <string-name>
            <given-names>F.</given-names>
            <surname>Giasson</surname>
          </string-name>
          .
          <article-title>The Music Ontology</article-title>
          .
          <source>In 8th International Conference on Music Information Retrieval (ISMIR'07)</source>
          , pages
          <fpage>417</fpage>
          -
          <lpage>422</lpage>
          , Vienna, Austria,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Raimond</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Sandler</surname>
          </string-name>
          .
          <article-title>A Web of Musical Information</article-title>
          .
          <source>In 9th International Conference on Music Information Retrieval (ISMIR'08)</source>
          , Philadelphia, USA,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Raimond</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Sutton</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Sandler</surname>
          </string-name>
          .
          <article-title>Automatic Interlinking of Music Datasets on the Semantic Web</article-title>
          .
          <source>In International Workshop on Linked Data on the Web (LDOW'08)</source>
          , Beijing, China,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>S.</given-names>
            <surname>Schaffert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Baumeister</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Bry</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Kiesel</surname>
          </string-name>
          . Semantic Wikis.
          <source>IEEE Software</source>
          ,
          <volume>25</volume>
          (
          <issue>4</issue>
          ):
          <fpage>8</fpage>
          -
          <lpage>11</lpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>R.</given-names>
            <surname>Schroeter</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Hunter</surname>
          </string-name>
          .
          <article-title>Annotating Relationships between Multiple Mixed-media Digital Objects by Extending Annotea</article-title>
          .
          <source>In 4th European Semantic Web Conference (ESWC'07)</source>
          , pages
          <fpage>533</fpage>
          -
          <lpage>548</lpage>
          , Innsbruck, Austria,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>R.</given-names>
            <surname>Troncy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Hardman</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. van Ossenbruggen</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Hausenblas</surname>
          </string-name>
          .
          <article-title>Identifying Spatial and Temporal Media Fragments on the Web</article-title>
          .
          <source>W3C Video on the Web Workshop</source>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31] V. Tzouvaras eds.
          <source>Multimedia Annotation Interoperability Framework</source>
          , W3C Incubator Group Editor's Draft. http://www.w3.org/2005/Incubator/ mmsem/XGR-interoperability/,
          <source>14 August</source>
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>L. von</given-names>
            <surname>Ahn</surname>
          </string-name>
          .
          <article-title>Games with a Purpose</article-title>
          .
          <source>IEEE Computer</source>
          ,
          <volume>39</volume>
          (
          <issue>6</issue>
          ):
          <fpage>92</fpage>
          -
          <lpage>94</lpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>