=Paper=
{{Paper
|id=Vol-233/paper-34
|storemode=property
|title=An Approach to Self-Annotating Content
|pdfUrl=https://ceur-ws.org/Vol-233/p69.pdf
|volume=Vol-233
|dblpUrl=https://dblp.org/rec/conf/samt/MatellanesSS06
}}
==An Approach to Self-Annotating Content==
An Approach to Self-Annotating Content
Adrian Matellanes, Freddy Snijder and Barbara Schmidt-Belz
Abstract—aceMedia content analysis capabilities are centered II. MOTIVATION
around the concept of the ACE. The ACE is composed of a
content layer, a metadata layer and an intelligence layer. In this A. Content flow
paper we show one application of the ACE Intelligence layer and
Nowadays, content flows from one device to the other.
how its proactive conduct can help in the complex task of adding
semantic metadata to multimedia content. Devices are connected to Personal Area Networks (PANs),
Local Area Networks (LANs) and Wide Area Networks
Index Terms—multimedia content analysis, proactive content, (WANs) allowing multimedia content to be easily transferred
self-annotation and shared. Devices supporting multimedia content range
from powerful media servers to desktop PCs, set-top-boxes
I. INTRODUCTION (thin and thick) and small devices such as mobile phones and
Digital multimedia content management is a very complex PDAs. These devices have different characteristics, specific
task. Huge efforts in research and development are being uses and processing capabilities.
carried out in industry and academia to alleviate this This fact about content flowing from one device to the other
complexity and bring solutions to help end-users and motivates the idea of a content item (an ACE in the aceMedia
professionals to easily manage their collections of multimedia context) self-annotating itself whenever it reaches a target
content. device, given that that device has annotation capabilities.
The aceMedia project tries to help tackle this problem with There are different scenarios where content can enrich its
a wide range of technologies, from multimedia knowledge metadata as it moves from one device to the other. Different
representation [6], multimedia content analysis [3], annotation capabilities can be found in different devices, e.g.
personalized search and browsing [4] to content adaptation [8] device A does not have a certain content classification module
to cite a few references. Fundamental to aceMedia’s approach that is present in device B, but also the same annotation
is the introduction of the Autonomous Content Entity (ACE) modules can have different capabilities depending on the
[1]. The ACE is a multimedia object comprising three layers : device on which they reside, e.g. device A's face recognition
the first layer is the multimedia content itself, the second layer module may know a different set of persons than device B.
is the metadata layer, which includes manual and automatic B. User participation
annotations, and the third layer is a programmable layer called
Purely automatic annotations have a long way to go to
“Intelligence layer” that provides proactiveness to the ACE.
provide the user with accurate semantic annotations. The
The intelligence layer is envisaged to help in the complex
semantic metadata associated with the content can be
problem of multimedia content management by enabling the
improved with the help of the user. Some of our studies,
content items to perform actions on behalf of the user,
contrary to some common beliefs, showed that users are
wherever they reside.
willing to "help the system” with their manual annotations [2].
This paper briefly describes one of the applications of
To incorporate users’ manual annotations we will create
content proactiveness enabled by aceMedia technologies,
proactive content that analyzes its own automatic semantic
namely, the creation of self-annotating content.
annotations and asks the user pertinent questions to solve
Content autonomy is not limited to self-annotation, other
some ambiguities or add some unknown information, e.g. a
activities based on the ACE intelligence layer are also carried
face that has been detected is not known to the face
out [9].
recognition module and the ACE asks the user “Who is the
person whose face is inside the bounding box?" The user,
always in control, can obviously ignore these questions as the
system is not strictly depending on them.
Manuscript received 18 September 2006. This research was supported by III. THE SELF-ANNOTATING PROCESS
the European IST project aceMedia (http://www.acemedia.org) under contract
FP6-001765. A. Proactiveness
Adrian Matellanes is with Motorola Labs, Basingstoke, UK (phone: +44
1256 484794; e-mail: adrian.matellanes@motorola.com). In the previous section we have seen some motivations for
Freddy Snijder is with Philips Research, Eindhoven, The Netherlands (e- giving autonomy and proactiveness to multimedia content
mail: freddy.snijder@philips.com).
Barbara Schmidt-Belz is with Fraunhofer-FIT, Sankt Augustin, Germany
when we try to add manual or automatic semantic annotations
(e-mail: Barbara.Schmidt-Belz@fit.fraunhofer.de). to an ACE. It is important to emphasize here that the whole
process of self-annotation and the ultimate decision to add resulting semantic annotations are then added to the ACE
semantic annotation to an ACE resides in the ACE itself. metadata layer and finally, as explained before, the
We will not go into detail about the software architecture Multimedia Reasoning module is called.
that enables ACEs to run their programmable Intelligence We have just outlined a very simple implementation of a
layer in order to give them autonomy; a brief description of self-annotation ACE. The ACE self-annotating intelligence
this can be found in [1]. layer can indeed be programmed to perform more complex
tasks and take other decisions such as raising questions to the
B. The AnnotationManager runs the analysis
user (see Section II) or prevent certain analysis to be
Content is analyzed by different content analysis modules performed (because of privacy issues for example).
that produce semantic metadata which in turn is added to the
ACE metadata layer. A typical, application driven, annotation IV. CONCLUSION
process is described in [2].
One of the objectives of aceMedia is to explore advanced
In our case of self-annotation, it is important to clarify that
content management techniques through the concept of the
the ACE intelligence layer is in charge of starting/stopping the
ACE and its Intelligence layer. aceMedia has successfully
annotation process and decides which, if any, content analysis
created a framework for the deployment of Autonomous
needs to be run. The ACE programmable intelligence layer
Content Entities (ACEs). These ACEs can have proactive
does not include the analysis algorithms that analyze and
behavior that helps users in their digital media management.
produce new metadata.
We have briefly outlined in this paper one of the
As explained in the previous section, the modules in charge
applications of the ACE Intelligence layer, the creation of
of analyzing the content and adding new metadata can differ
self-annotating content. We have presented the motivation
from one device to the other and are offered to the ACE
which led us to make ACEs self-annotating as opposed to
intelligence layer through a common framework called the
being annotated passively. Finally we have outlined the
Annotation-Manager.
process and workflow of self-annotation.
This AnnotationManager interacts with the ACE
Within aceMedia we are investigating other applications
intelligence layer and runs the requested analysis modules in
of the ACE Intelligence layer that are outside the scope of
the appropriate order. The AnnotationManager also deals with
this paper such as self-organizing ACEs and self-governing
dependencies, e.g. a face recognition module may depend on a
ACEs.
face detection module.
Once the AnnotationManager has called the analysis
REFERENCES
modules requested by the ACE, it will always run the
[1] A.Matellanes, T.May, F.Snijder, P.Villegas, E.O.Dijk and A.Kobzhev,
Multimedia Reasoning module to ensure metadata
“An architecture for multimedia content management” in EWIMT 2005,
consistency, remove ambiguities and derive new annotations London, UK.
if possible. [2] A.Matellanes, A.Evans, B.Erdal “Creating an application for automatic
annotation of images and video” in SWAMM 2006, Edinburgh, UK.
C. Self-annotation process [3] N.Dalal and B.Triggs “Histograms of Oriented Gradients for Human
Detection” in IEEE CVPR 2005, San Diego, USA, June 2005
In this section we will describe a typical self-annotation [4] S.Bloehdorn, K.Petridis, C.Saathoff, N.Simou, V.Tzouvaras, Y.Avrithis,
process. As explained in the previous section, the Self- S.Handschuh, Y.Kompatsiaris, S.Staab and M.G.Strintzis, “Semantic
Annotating ACE is in control of the annotation process but it Annotation of Images and Videos for Multimedia Analysis” in ESWC
2005, Heraklion, Greece, May 2005
does not perform the analysis nor the annotations itself. This
[5] D. Vallet, M. Fernández, P. Castells, P. Mylonas and Y. Avrithis,
way, the ACE can benefit from the different capabilities “Personalized Information Retrieval in Context” in MRC 2006 at AAAI
offered by different devices and contexts, see section II. 2006, Boston, USA, July 2006
When an ACE is transferred to a different device, its self- [6] K. Petridis, S. Bloehdorn, C. Saathoff, N. Simou, S. Dasiopoulou, V.
Tzouvaras, S. Handschuh, Y. Avrithis, I. Kompatsiaris and S. Staab,
annotation process is started. “Knowledge Representation and Semantic Annotation of Multimedia
The self-annotating Intelligence layer, asks the device what Content” in IEE Proceedings on Vision Image and Signal Processing,
semantic annotation capabilities are present in the device. This Special issue on Knowledge-Based Digital Media Processing, Vol. 153,
No. 3, pp. 255-262, June 2006.
request is received by the AnnotationManager. [7] J.Malobabic, H.Le Borgne, N.Murphy, N.O'Connor, “Detecting The
The AnnotationManager will analyze the kind of content Presence Of Large Buildings” in Natural Images 4th International
stored in the ACE, i.e. whether it is a still image, a video clip, Workshop on Content-Based Multimedia Indexing, CBMI 2005, Riga,
Latvia, 21-23 June 2005
or any other type of media. Based on this analysis the [8] N. Sprljan, M. Mrak, G. C. K. Abhayaratne, E. Izquierdo, “A Scalable
AnnotationManager decides what analysis capabilities it can Coding Framework for Efficient Video Adaptation” in Workshop on
offer, e.g. face detection, face recognition, speech recognition, Image Analysis for Multimedia Interactive Services (WIAMIS 2005),
Montreux, Switzerland, April 13-15, 2005
knowledge-assisted analysis, etc.
[9] P.Charlton and J.Teh, “A self-governance approach to supporting
The ACE checks if this type of annotation has already been privacy preference-based content sharing in distributed environments” in
performed, and creates a list of the missing annotation SOAS 2006, Erfurt, Germany, 18-20 September 2006.
categories it is interested in. The ACE sends this list to the
AnnotationManager that calls, in the appropriate order, the
analysis modules and solves dependencies if needed. The