<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Cross-media document annotation and enrichment</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ajay Chakravarthy Fabio Ciravegna</string-name>
          <email>a.chakravarthy@dcs.shef.ac.u</email>
          <email>f.ciravegna@dcs.shef.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vitaveska Lanfranchi</string-name>
          <email>v.lanfranchi@dcs.shef.ac.uk</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Sheffield University of Sheffield</institution>
          ,
          <addr-line>Regent Court, 211 Portobello Street, Regent Court, 211 Portobello Street, Sheffield, S1 4P</addr-line>
          ,
          <country>United Kingdom Sheffield</country>
          ,
          <addr-line>S1 4P</addr-line>
          ,
          <country country="UK">United Kingdom</country>
          ,
          <addr-line>+44-114-2221945 +44-114-2221945</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Sheffield</institution>
          ,
          <addr-line>Regent Court, 211 Portobello Street, Sheffield, S1 4P</addr-line>
          ,
          <country country="UK">United Kingdom</country>
          ,
          <addr-line>+44-114-2221945</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2002</year>
      </pub-date>
      <abstract>
        <p>Annotation of documents is a complex and labour intensive task. So far, research has focused on supporting the annotation of documents in single media, e.g. texts or images. Much less attention has been paid to the issue of annotating documents across media, especially useful for web documents that usually contain both text and images. In this paper we describe AKTiveMedia, a tool which supports human-centric annotation of documents across media. It offers a number of features to support different types of annotations, from ontology-based ones to free comments. We discuss what we believe are the main requirements for annotating Web documents, from support of annotator communities, to the reduction of the annotation burden, to the support of document lifecycle and how they have been implemented inside AKTiveMedia. The tool has applications in annotation of web pages, personal memories and knowledge management.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Semantic Web Annotations</kwd>
        <kwd>Image Annotation</kwd>
        <kwd>Text Annotation</kwd>
        <kwd>Knowledge Management</kwd>
        <kwd>Document generation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. INTRODUCTION</title>
      <p>
        The amount of multimedia information stored every day in both
companies and personal archives is growing. Together with the
Web, these archives are reaching sizes unimaginable even until
some years ago. For example, in August 2005 Yahoo claimed to
cover 20 billion pages1, of which 19.2 billion web documents, 1.6
billion images, 50 million audio and video files. It is also
calculated that almost 375 petabytes or 787.5 billion
photographs) are produced each year (almost 2 times all printed
material) with a yearly growth rate of 5%, which is the highest
growth rate among different data types [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. At the same time,
large organizations’ intranets have reached the size of mini Webs,
connecting thousands of computers and having reached
dimensions of dozens of millions of documents; it is expected that
soon they will reach hundreds of millions of pages, i.e. a size
comparable to the Internet at the end of the 90s. Keyword
matching based systems are struggling to keep pace with such an
amount of information. While currently there is not an issue in
indexing and retrieving documents on the Web, in large
organizations and in personal archives the issue of efficiently and
effectively retrieving material is becoming pressing, due to the
density of information.
      </p>
      <p>Also, keyword-based methods are unable to put information in
context. This is a problem for knowledge management, where
very often it is the context that determines the importance of a
document. The context is very difficult to model in keyword based
queries. This becomes impossible when part of the context is
spread across different media (e.g. in images).</p>
      <p>For this reason there is a growing interest in applying
methodologies able to capture the content and the context of
multimedia documents, in order to enable effective searching (and
document-based knowledge management in general).</p>
      <p>
        A common and successful approach to organise and manage huge
quantities of information is to enrich documents with metadata.
Previous research in personal image management [
        <xref ref-type="bibr" rid="ref8">14, 8</xref>
        ] and text
annotation [
        <xref ref-type="bibr" rid="ref10 ref3 ref7">3, 7, 13, 10</xref>
        ] demonstrated how annotating images or
documents could be a way to organize information and transform
it into knowledge that can be used easily later. Metadata enables
the creation of a knowledge base which can then be queried as a
way both to retrieve documents (via content and context) and to
query the structured data (e.g. creating charts illustrating trends).
In this paper we are first of all trying to identify the main
requirements for cross-media annotation, introducing then
AKTiveMedia, a tool that supports cross-media (image/text)
document annotation. We show how AKTiveMedia supports
different types of annotations, from ontology-based to free
comments, how it supports communities of annotators, and a
document lifecycle, allowing users to both create and annotate
documents.
      </p>
      <p>The aim of AKTiveMedia is to address one of the main problems
of document annotation: the task complexity. In general
AKTiveMedia is a tool that fosters knowledge reuse.
.Finally we describe the underlying architecture and draw some
conclusions and future work.</p>
    </sec>
    <sec id="sec-2">
      <title>2. CROSS-MEDIA ANNOTATION</title>
    </sec>
    <sec id="sec-3">
      <title>REQUIREMENTS</title>
      <p>We identified some initial requirements for cross-media document
enrichment. Firstly, we highlight the dimensions of the content
that can be enriched with annotations, then we will proceed
discussing how to reduce the complexity of the annotation task,
and how to support community sharing.</p>
    </sec>
    <sec id="sec-4">
      <title>2.1 Annotation levels and types</title>
      <p>We identify five main dimensions of information that can be
associated to a document via annotation:</p>
      <p>Resource Metadata, like creation date, time, author, etc.;
this type of information is generally provided in a
structured form, for example via EXIF data for images,
document creation time for texts, HTML metadata for
author, etc. It is quite easy to automatically capture
metadata and it provides an important knowledge about
the context in which the document was produced.</p>
      <p>Content annotation: which makes content available for
retrieval; typically, in literature, content has been
represented using ontology-based annotation. This is the
most common type of annotation in the semantic web
and is generally used to mark-up contingency situations
that can change in time. Annotations can be performed
across documents and media, i.e. they may relate the
text content with part of an image, as mentioned in the
examples above.</p>
      <sec id="sec-4-1">
        <title>Immutable knowledge about instances (e.g.); this</title>
        <p>information is generally stored outside the large
majority of documents; it will be described in the
ontology. Some documents, e.g. descriptive or
normative documents, such as dictionaries, etc. can
contain immutable knowledge.</p>
      </sec>
      <sec id="sec-4-2">
        <title>Informal knowledge about the document or its content.</title>
        <p>
          This is generally stored using free text comments that
integrate the document content, adding information and
knowledge not explicitly mentioned within the
document. For example, a user could explain in the
comments why a specific formulation was chosen or
why a specific hypothesis was pursued; i.e. comments
are used to complement the knowledge in the document
with knowledge about the process that generated it.
Another possible way for annotating documents and
images are folksonomies. In our opinion, folksonomies
are more interesting for personal use e.g. to annotate
pictures to share with friends than for use in knowledge
management. In this case the social dimension of
sharing is more important than retrieval; there is no
need of formal classification of concepts; folksonomies
are more a way to attach emotions and memories. In
these cases, free annotation (tags and textual
descriptions) proves to be more interesting for the users
[
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], as demonstrated by the success of community-based
image annotation websites as Flickr2 or social bookmark
managers as De.li.ci.ous3.
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>2.2 Annotation and document lifecycle</title>
      <p>
        In our opinion, annotation of document should follow the whole
document lifecycle, from production to use and be flexible to
support the needs of different types of users. In previous research
work [
        <xref ref-type="bibr" rid="ref2 ref3">3, 2, 14</xref>
        ] the annotation task was considered associated
mainly to the document production task. However, annotation can
happen every time a document is accessed. This is because:
1.
2.
3.
      </p>
      <sec id="sec-5-1">
        <title>The author may want to make available the document content via ontology-based annotation. The author has generally a specific view on the reasons why a document is produced and successively retrieved.</title>
      </sec>
      <sec id="sec-5-2">
        <title>The reader may need a different (level of) annotation than the one provided by the author (i.e. to use a different ontology for marking up content or need more details).</title>
      </sec>
      <sec id="sec-5-3">
        <title>All users may want to comment on the document in</title>
        <p>itself or on other comments.</p>
        <p>Not all annotations must necessarily be widely available. Some
annotations can be personal, others may stay within specific
boundaries (e.g. the department or the company), and others can
be made available.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>2.3 Complexity in annotation</title>
      <p>
        Manually annotating data is a labour intensive and tedious task
[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] for a user. It can increase both the time needed for producing a
document and the information overload.
      </p>
      <p>
        Previous literature studies have highlighted the importance of
cooperative systems able to ease the annotation process [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] and
reduce the information overload. While it is difficult to
completely automate the annotation task, because annotations can
refer to subjective opinions and memories, it is possible to help
users on many fronts, for example automatically extracting
metadata from documents using Information Extraction
methodologies
      </p>
    </sec>
    <sec id="sec-7">
      <title>2.4 Community Contribution</title>
      <p>As experience shows, the importance of user voluntary
contributions is fundamental for the creation of a base of
knowledge. This makes the difference between success and failure
of applications on the Web [12].</p>
      <p>We believe that the social perspective is fundamental as it enables
the explicitation of implicit knowledge. Such explicitation of
knowledge comes very often in informal comments. This
information is generally gladly volunteered by both authors and
readers (as the experience of GoogleMaps shows, where a gigantic
database of information is created by Web users).</p>
      <sec id="sec-7-1">
        <title>2 http://www.flickr.com/</title>
      </sec>
      <sec id="sec-7-2">
        <title>3 http://del.icio.us/</title>
      </sec>
    </sec>
    <sec id="sec-8">
      <title>2.5 Ontology Complexity</title>
      <p>Ontologies for annotations can be quite complex. Most of the
current annotation tools provide a side panel where the ontology
is displayed in the form of a tree. Annotation is done by selecting
an element from the tree. This is clearly an impossible strategy
with a very large ontology, as the user would have to scroll over a
very large tree.</p>
      <p>
        Moreover, a large ontology (even in terms of 100s of concepts) is
difficult to use because users find difficult to remember all the
available concepts and to use them properly. As previous
literature proved [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], when dealing with vast quantities of
information users may want to zoom and visualise only the
sections they are interested in, or filter out what is not relevant for
the current task.
      </p>
      <p>For this reason it is important to find a way to represent the
ontology making it manageable when annotating documents.</p>
    </sec>
    <sec id="sec-9">
      <title>3. AKTIVEMEDIA</title>
      <p>AKTive Media is a user centric system for document enrichment
across media; it uses Semantic Web and language technologies for
acquiring, storing and reusing knowledge. The aim is to provide a
seamless interface that guides users through the annotation
process, reducing the complexity of their task.</p>
      <p>In the following paragraph we will detail how AKTiveMedia
answers to the previously outlined requirements.</p>
    </sec>
    <sec id="sec-10">
      <title>3.1 Overview</title>
      <p>AKTiveMedia supports the annotation of text, images and HTML
documents (containing both text and images) using both
ontologybased and free-text annotations.</p>
      <p>Support is provided both for author and reader annotations, giving
the possibility to load different ontologies accordingly to the task.
Moreover the annotations are stored separately from the
document, alongside with the authorship. This enables controlling
the privacy of annotations and the display.</p>
      <p>In order to support community sharing, AKTiveMedia allows the
user to insert comments and annotations and share them with
other members of the community through a centralised server
(more details is Section 4).</p>
      <p>
        Human Language technologies have been employed to ease the
annotation task: an underlying Information Extraction systems
(T-Rex) [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] has been integrated, that learns from previous
annotations (both user and community ones) and suggests new
annotation to the user, that can accept or reject them, thus
retraining T-Rex and improving the learning process. This is a route
we already successfully explored in Melita[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], OntoMat[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] and
MnM[14], where it was found that the annotation time could
decrease by 80% and interannotator agreement could double [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
But AKTiveMedia goes beyond the single media annotation
suggestions and moves towards cross-media strategies. When an
annotation in inserted in the text, the system autonmatically
inserts it in a knowledge base that will be used to suggest new
annotations when dealing with images (more details in Section
3.1).
      </p>
      <p>Moreover some metadata are automatically captured via the
automatic extraction of EXIF data in images and by extracting
meta-tags from HTML documents. In one application, we also
integrated GPS and calendar information [13].</p>
      <p>The way in which AKTiveMedia deals with the problem of the
ontology, is by using what we call “disappearing ontology” i.e. we
try to hide the unnecessary complexity of the ontology. On the
one hand, users can adopt specific views on the ontology to
annotate their documents: a user may not need to use the complete
ontology all the time.</p>
      <p>Therefore a very high level description of the ontology is
displayed and the details are hidden till the user needs to use
them. Concepts that are not displayed directly in the graph are
retrieved using the search mechanism associated to the ontology
based annotation. Inputting a textual description (e.g. “Director of
KMI”) and selecting form a short list of potential ontology
concepts becomes quite easy for a user (e.g. selecting between
“Position” and “Person”). This reduces the necessity of displaying
a large ontology, while maintaining its details.</p>
      <p>AKTiveMedia supports also editing documents or creating new
ones: this is particularly relevant when wanting to create a new
Semantic Web Website.</p>
      <p>
        In a previous research project, AKTiveDoc [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], we provided
facilities for annotating documents while editing. This enabled
semi-automatic searching of relevant knowledge to insert into the
document. We decided to avoid this feature (annotating while
editing) in AKTiveMedia at this point in time, because in the use
of AKTiveDoc we found that it was not an easily manageable
feature. On the one hand there is an objective complexity in the
software implementation (necessity of aligning annotation and
document during modifications).
      </p>
      <p>On the other hand, users found it difficult to manage the
simultaneous tasks of editing, annotating and retrieval of further
knowledge. It implied a high cognitive complexity which was
difficult to manage in a single environment. Users found easier to
perform the three tasks using three separate environments.This is
why AKTiveMedia separates the editing and the annotation task:
a user can firstly create a new HTML document containing both
text and images, and afterwards the annotation can be performed
as for a normal document.</p>
      <p>The editing functionality has been implemented integrating an
HTML editor based on EKIT4 (Figure 1).</p>
      <sec id="sec-10-1">
        <title>4http://www.hexidec.com/ekit.php</title>
        <p>The document writing task is also supported via functionalities for
retrieving the content of existing documents that operate on the
document annotations. This enables reuse of related knowledge.
As an example, while the user is writing a document, if he’s
writing “John Domingue” the system can start retrieving all John
Domingue’s pictures: the user can then decide to insert one of the
picture in the document. This facility enables knowledge reuse
making easier for the user to write a document.</p>
        <p>Moreover while the user perceives to be writing a simple HTML
document, more information is automatically added by the system,
in terms of metadata and structure. The resultant file will have an
associated RDF annotations file that will contain annotation that
Permission to make digital or hard copies of all or part of this work for
personal or classroom use is granted without fee provided that copies are
not made or distributed for profit or commercial advantage and that
copies bear this notice and the full citation on the first page. To copy
otherwise, or republish, to post on servers or to redistribute to lists,
requires prior specific permission and/or a fee.</p>
        <p>Conference’04, Month 1–2, 2004, City, State, Country.</p>
        <p>Copyright 2004 ACM 1-58113-000-0/00/0004…$5.00.
the system inserted in a transparent way: for example when
creating a H1 heading, the system will match the HTML tag to the
document ontology and will automatically insert an annotation
“title” with the value inserted by the user.</p>
        <p>When the editing is finished, the document can then be annotated
as described in Section 3.1.</p>
        <p>In the following section we will detail AKTiveMedia interface,
using a sample scenario that will allow showing the
functionalities. The chosen scenario is the annotation of the KMI
news corpus, which is a set of news (published on a website)
about the various people visiting the KMI institute and there
contributions. The documents are HTML files containing both
images and text: content of text and images are related, as they are
produced contextually and often describe the same person or
event</p>
      </sec>
    </sec>
    <sec id="sec-11">
      <title>3.2 Interface</title>
      <p>AKTiveMedia supports three main modalities of annotation:</p>
      <p>
        text
All modalities share the same functionalities and very similar
interfaces. They all offer ontology-based enrichment through a
graphical interface, following the paradigm of other annotation
tools like Cream or Melita [
        <xref ref-type="bibr" rid="ref2 ref3">3, 2</xref>
        ]: portions of text or images can be
associated with concepts in the ontology with a straightforward
point&amp;click interface.
      </p>
      <p>Free-text annotations can also be added on top of the
ontologybased ones, to insert more information.</p>
      <p>When the user is starting to annotate a new HTML document,
they can simply annotate by clicking on the concept in the
ontology and then highlighting a sequence of words (see Figure
2).</p>
      <p>It is possible to associate relations among highlighted parts (in the
figure below we declare again that the “John Domingue” has a
visiting entity “Theresa May”). This is done by clicking on the
“John Domingue” instance in the text and then on its relation
(“hasVisitingEntity”, mid-left of figure) and then clicking on the
instance of the “Theresa May” highlighted in the text (see Figure
2).</p>
      <p>When the text has all been annotated, the user can decide to
annotate also the corresponding image(s). When right clicking on
the image, AKTiveMedia switches to the image annotation mode,
without losing the context of the document (the image is opened
in another tab).</p>
      <p>In AKTiveMedia images can be annotated as a whole or in part.
First of all, a title and a description can be inserted for each
image, alongside with free text comments related to the whole
image. This metadata can be further annotated, using the text
based annotation strategy described before. Portions of the image
are identified using the mouse (e.g. by drawing a square) and can
be annotated via the ontology
When a portion of the image is annotated, a popup window
appears (centre right of figure) which enables to describe the
content of the annotation in natural language. A facility is
provided to search for a complete unique description given the
user description. This accesses a triple store of descriptions (as
offered by a gazetteer, or by part of the ontology not shown for
usability reasons). The facility is used for example to input “
John” as the visited person and retrieve complete descriptions of
all the persons with the name John working at the KMI institute.
The selection has a side effect the allocation of the URI in order to
uniquely identify the object.</p>
      <p>Ontological relations between instances (i.e. parts of the image)
can also be inserted. For example it is possible to declare that the
“John Domingue” has a visiting entity “Theresa May”. This is
done by clicking on the instance bearing the relation (“John
Domingue”); the system will then show all the relations possible
for that instance (middle, left in the figure); clicking on the
selected relation and then on the other instance (the “Theresa
May”) will fill the relation with the latter.</p>
      <p>More important, the system allows to establish relations between
text and images, allowing the user to assert that the “John
Domingue” in the picture is the same John “Domingue” that was
annotated as “VistitedPerson” in the text.</p>
      <p>Also these implicit relations can me used by the system to make
the annotation task easier, using a contextual annotation
mechanisms that analyse user’s or system’s annotations in the text
or in the image to suggest new annotations: for example if the text
has been annotated first, when the user annotates an area in the
image as a visiting person, the system will suggest as description
the finding previously inserted in the text (“Theresa May”) – (see
Figure 3)
The user can accept or reject the suggestion. If accepted, an
identity is established between the instances in the text and the
ones in the image (same URI).
In case the image is annotated first, the system will search the text
for descriptions compatible with those in the image. Matching is
done using string distance metrics5</p>
    </sec>
    <sec id="sec-12">
      <title>4. SYSTEM ARCHITECTURE</title>
      <p>The system is based on a configurable plug-in model in which the
different components (e.g. ontology loader, annotation modalities,</p>
      <sec id="sec-12-1">
        <title>5 http://www.dcs.shef.ac.uk/~sam/simmetrics.html</title>
        <p>web services etc.) are independent sub-models that can be
plugged in for creating a custom application.</p>
        <p>This is because AKTiveMedia is more than just an annotation
tool. It is designed to be inserted into user applications. Its
architecture focuses on RDF as a way to store and query data and
to communicate between components and web services as a way
to distribute the architecture. All the annotations are stored as
RDF triples inside a local store and periodically updated into a
central triple store using web services.</p>
        <p>This modular architecture implements the knowledge sharing
scenario, allowing different users to see other people’s
annotations and reuse them (see Figure 4).</p>
        <p>When annotations are saved, they are associated to the document
through a unique URI or hash code, thus enabling retrieval of
annotations performed by other users on the same document.
A plug-in (AKTiveSearch) enables searching and reuse of
knowledge while creating or annotating the document.
AKTiveSearch enables simultaneous multiple queries to different
archives and sources of information, the integration of the
returned information and the filtering of the results based upon
the context of use.</p>
        <p>
          As mentioned before, AKTiveMedia tries to maximise the amount
of resource metadata that can be automatically collected. For this
reason an EXIF Extractor and an Information Extraction system,
T-Rex, are integrated. They both work in the background,
extracting possible metadata and annotations that are later
presented to the user and saved in RDF format. In particular
TRex has been implemented for background training and
annotations, using a separate thread, so to not interfere with the
user’s activity (as the training process can be very long) and to
maximise the efficiency. The schema followed is the same as
Melita’s[
          <xref ref-type="bibr" rid="ref2">2</xref>
          ].
The ontology loader, based on Jena6, is used to load the users
preferred ontology and to implement the selective view
mechanism (disappearing ontology). The interaction between
the user and the system is realised through the user interface that
is also modular to allow different modalities of annotations.
Currently AKTiveMedia supports four annotation modalities
(text annotation, image annotation, 3D annotation and Editing)
and it is possible to mix and match these modalities in order to
6 http://jena.sourceforge.net/
facilitate cross media annotation. The interface component was
designed using the MVC (Model View Component) architecture
in Java. This enables separation of data and visualization,
enabling efficient flow of information across different
modalities, while keeping the user interface simple and easy to
use.
        </p>
      </sec>
    </sec>
    <sec id="sec-13">
      <title>5. EVALUATION</title>
      <p>We have performed a detailed evaluation of AKTive Media
during the Fourth Summer School for Ontological Engineering
in the Semantic Web in Cercedilla, Spain.</p>
      <p>Over 60 students divided in groups of 3-4 persons each were
testing the system. The task involved first of all annotating 10
documents from the KMI corpus and then starting the
Information Extraction system (T-Rex); after the training the
system would start suggesting possible annotations in a semi
supervised way. Log file were collected, recording all the user
activities and students were asked to fill in a questionnaire at the
end of the session.</p>
      <sec id="sec-13-1">
        <title>The results of the evaluation are still under study.</title>
      </sec>
    </sec>
    <sec id="sec-14">
      <title>6. POTENTIAL USE CASES</title>
      <p>In the following sections potential use case in
AKTiveMedia can contribute will be outlined.
which</p>
    </sec>
    <sec id="sec-15">
      <title>6.1 (PhotoCopain) Memories for Life</title>
      <p>Memories for Life is a Grand Challenge for Computing Science
proposed by the UK Computing Research Committee.
Individuals are usually storing an enormous amount of
information about themselves on their computers (documents,
images, web browsing logs, etc). The challenge for computing
researchers is to develop ideas and techniques that help people
get the maximum benefit from their memories, while at the same
time giving them complete control over memories so as to
preserve their privacy. Memories for Life is also regarded as a
Grand Challenge by the UK Foresight Cognitive Systems
project.</p>
      <p>Digital memories clearly offer tremendous potential for science
and technology. We must also ensure that they help society by
widening access to information technology, so that everyone,
not just well-educated people with no disabilities in rich
countries, could benefit from the information revolution. The
challenge is to develop detailed models of an individual’s
abilities, skills, and preference by analysing his or her digital
memories; and to use these models to optimise computer
systems for individuals. A longer-term challenge might be
presenting a story extracted from memories in different
modalities according to ability and preference; for example, as
an oral narrative in the user’s native language, or as a purely
visual virtual reality reconstruction for people such as aphasics
who have problems understanding language. Limited examples
of such systems can be built now; the challenge is in mining the
wealth of information latent in digital memories so that fully
competent systems could be in use in fifteen years.</p>
      <p>AKTive Media is being extensively being used for the
PhotoCopain project which is a part of the memories for life.
Images can be semantically annotated and the narratives linked
to the images. This is possible due to the cross media annotation
capability of AKTive Media. We are currently focusing on auto
narrative generation technology given a set of images in a
timeline [12].</p>
    </sec>
    <sec id="sec-16">
      <title>6.2 E-Response AKT Project</title>
      <p>AKTive Media is currently being ported to integrate with
Compendium7 tool for the AKT E-Response project. The
project aims to use Semantic web agents to automatically deal
with an emergency event. Examples of which can be include:
taking photographs of the incident and sending them to a
semantic web service, locating and notifying the nearest fire
stations about the incidents, etc.</p>
      <p>AKTive Media will serve as the interface for photographs taken
in an emergency situation and there annotations. Further it will
also act as a search interface for the photographs using the
SPARQL search facility.</p>
    </sec>
    <sec id="sec-17">
      <title>7. FUTURE WORK AND CONCLUSIONS</title>
      <p>In this paper, we have described and discussed AKTiveMedia, a
tool for editing and annotating multimedia document containing
images and text. The annotation can be performed within and
across the different media. Annotation is mainly manual, but a
number of strategies are used to reduce the burden of
annotation. We have shown and discussed how the system
satisfies a range of user requirements for use to support
knowledge management and personal archives. In particular the
requirements we analysed are: (1) annotation types and levels,
(2) annotation as community activity, (3) annotation and
document lifecycle, (4) annotation complexity, (5) ontology
complexity and (6) knowledge reuse.</p>
      <p>
        The current applications of AKTiveMedia are in both personal
memory management [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] and knowledge management.
Concerning the latter, AKTiveMedia is the basis for a real world
application under definition within IPAS, a project co-funded by
the UK Department of Trade and Industry and Rolls Royce plc
(www.3worlds.org). The application concerns the editing and
annotation of diagnostic reports on jet engines; the examples
used in this paper are derived from the user requirement analysis
in IPAS.
      </p>
      <p>In the future, we will explore further levels of community
annotations, by addressing in particular issues such as privacy of
data and ownership of annotation. Moreover, we will explore in
more details the use of folksonomies in an industrial
environment, and study their impact on knowledge retrieval and
reuse. For this reason we plan to introduce in AKTiveMedia
facilities for the direct manipulations of folksonomies. Another
venue of development is the annotation of 3D images. The
currently available facility implemented in the system is quite
limited and needs extensions. In 3D annotations, there is an
inherent HCI complexity in annotating an image that can be
rotated.</p>
    </sec>
    <sec id="sec-18">
      <title>8. ACKNOWLEDGMENTS</title>
      <p>This research was partially funded by the Advanced Knowledge
Technologies (AKT) Interdisciplinary Research Collaboration</p>
      <sec id="sec-18-1">
        <title>7 http://www.aktors.org/technologies/compendium/</title>
        <p>(IRC). AKT is sponsored by EPSRC, grant number
GR/N15764/01. This project is also funded by the European
project IST X-Media, funded as part of Framework 6, grant
number FP6-26978.</p>
        <p>AKTiveMedia is an open source project, available under
Academic Free License, Educational Community License,
General Public License. A copy of the binary files and the
source code can be downloaded at
http://sourceforge.net/projects/aktivemedia/.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Brilakis</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Content based integration of construction site images in aec/fm model based systems</article-title>
          ,
          <source>PhD thesis</source>
          , University of Illinois at Urbana-Champaign,
          <year>2005</year>
          , http://wwwpersonal.umich.edu/~brilakis/webdata/Brilakis_Thesis.pdf
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Ciravegna</surname>
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dingli</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Petrelli</surname>
            <given-names>D.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Wilks</surname>
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>UserSystem Cooperation in Document Annotation based on Information Extraction</article-title>
          .
          <source>In Proceedings of the 13th International Conference on Knowledge Engineering and Knowledge Management (EKAW02)</source>
          ,
          <fpage>1</fpage>
          -
          <lpage>4</lpage>
          October 2002 - Siguenza (Spain),
          <source>Lecture Notes in Artificial Intelligence 2473</source>
          , Springer V.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Handschuh</surname>
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Staab</surname>
            <given-names>S.</given-names>
          </string-name>
          , and Ciravegna F..
          <article-title>S-CREAMSemi-automatic CREAtion of Metadata</article-title>
          .
          <source>In Proceedings of the 13th International Conference on Knowledge Engineering and Knowledge Management (EKAW02)</source>
          ,
          <fpage>1</fpage>
          -
          <lpage>4</lpage>
          October 2002 - Siguenza (Spain),
          <source>Lecture Notes in Artificial Intelligence 2473</source>
          , Springer Verlag
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Harris</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          :
          <article-title>Mathematical Structures of Language</article-title>
          . WileyInterscience, New York,
          <year>1968</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Iria</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ireson</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Ciravegna</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>An Experimental Study on Boundary Classification Algorithms for Information Extraction using SVM</article-title>
          .
          <source>Proceedings of the EACL Workshop on Adaptive Text Extraction and Mining (ATEM</source>
          <year>2006</year>
          ),
          <year>April 2006</year>
          , Trento, Italy.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Kang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <article-title>and</article-title>
          <string-name>
            <surname>Shneiderman</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <article-title>Visualization Methods for Personal Photo Collections: Browsing and Searching in the PhotoFinder In Proc</article-title>
          .
          <source>IEEE International Conference on Multimedia and Expo (ICME2000)</source>
          , New York City, New York.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Kuchinsky</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pering</surname>
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Creech</surname>
            <given-names>M.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Freeze</surname>
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Serra</surname>
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gwizdka</surname>
            <given-names>J.,</given-names>
          </string-name>
          <article-title>FotoFile: A Consumer Multimedia Organization and Retrieval System</article-title>
          ,
          <source>Proceedings of ACM CHI99 Conference on Human Factors in Computing Systems</source>
          ,
          <volume>496</volume>
          -
          <fpage>503</fpage>
          ,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Lanfranchi</surname>
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ciravegna</surname>
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Petrelli</surname>
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Semantic Webbased</surname>
          </string-name>
          <article-title>Document: Editing and Browsing in AktiveDoc</article-title>
          ,
          <source>Proceedings of the 2nd European Semantic Web Conference</source>
          , Heraklion, Greece, May 29-June 1, 2005
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>O</given-names>
            <surname>'Really</surname>
          </string-name>
          ,
          <string-name>
            <surname>T.</surname>
          </string-name>
          :
          <article-title>What is Web 2.0</article-title>
          ,
          <string-name>
            <given-names>Design</given-names>
            <surname>Patterns</surname>
          </string-name>
          and
          <article-title>Business Models for the Next Generation of Software.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Petrelli</surname>
            ,
            <given-names>D</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lanfranchi</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ciravegna</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Working Out a Common Task: Design and Evaluation of User-Intelligent System Collaboration</article-title>
          .
          <source>In Proceedings of Tenth IFIP TC13 International Conference on Human-Computer Interaction (INTERACT</source>
          <year>2005</year>
          ), Rome,
          <year>September 2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Thorsten</surname>
            <given-names>B.</given-names>
          </string-name>
          <article-title>Inter-annotator agreement for a German Newspaper Corpus</article-title>
          .
          <source>In the Proceedigs of 2nd international conference of Language Resources and Evaluation (LREC</source>
          <year>2000</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>