=Paper= {{Paper |id=Vol-1177/CLEF2011wn-ImageCLEF-BudikovaEt2011 |storemode=property |title=MUFIN at ImageCLEF 2011: Success or Failure? |pdfUrl=https://ceur-ws.org/Vol-1177/CLEF2011wn-ImageCLEF-BudikovaEt2011.pdf |volume=Vol-1177 |dblpUrl=https://dblp.org/rec/conf/clef/BudikovaBZ11 }} ==MUFIN at ImageCLEF 2011: Success or Failure?== https://ceur-ws.org/Vol-1177/CLEF2011wn-ImageCLEF-BudikovaEt2011.pdf
MUFIN at ImageCLEF 2011: Success or Failure?

                Petra Budikova, Michal Batko, and Pavel Zezula

                     Masaryk University, Brno, Czech Republic
                      {budikova,batko,zezula}@fi.muni.cz



      Abstract. In all fields of research it is important to discuss and com-
      pare various methods that are being proposed to solve given problems.
      In image retrieval, the ImageCLEF competitions provide such compar-
      ison platform. We have participated in the Photo Annotation Task of
      the ImageCLEF 2011 competition with a system based on the MUFIN
      Annotation Tool. Our approach is, in contrast to typical classifier solu-
      tions, based on a general annotation system for web images that provides
      general keywords for arbitrary image. However, the free-text annotation
      needs to be transformed into the 99 concepts given by the competition
      task. The transformation process is described in detail in the first part
      of this paper. In the second part, we discuss the results achieved by our
      solution. Even though the free-text annotation approach was not as suc-
      cessful as the classifier-based approaches, the results are competitive es-
      pecially for the concepts involving real-world objects. On the other hand,
      our approach does not require training and is scalable to any number of
      concepts.


1   Introduction
In the course of the past few decades, multimedia data processing has become
an integral part of many application fields, including medicine, art, security, etc.
This poses a number of challenges to the computer science – we need techniques
for efficient data representation, storing and retrieval. In many applications, it is
also necessary to understand the semantic meaning of a multimedia object, i.e. to
know what is represented in a picture or what a video is about. Such information
is usually expressed in a textual form, e.g. as a text description that accompanies
the multimedia object. The semantic description of an object can be obtained
manually or (semi)-automatically. Since the manual annotation is an extremely
labor-intensive and time-consuming task for larger data collections, automatic
annotation or classification of multimedia objects is of high importance. Perhaps
the most intensive is the research on automatic annotation of images, which is
essential for semantic image retrieval [4].
    In all fields of research it is important to discuss and compare various meth-
ods that are being proposed to solve given problems. In image retrieval, the
ImageCLEF competitions provide such comparison platform. Each year, a set
of specific tasks is defined that reflects the most challenging problems of current
research. In 2011 as well as in the two previous years, one of the challenges was
the Photo Annotation Task.
2       Petra Budikova, Michal Batko, and Pavel Zezula

    This paper presents the techniques used by the MUFIN group to handle the
Annotation Task. We believe that our methods will be interesting for the com-
munity since our approach is different from the solutions presented in previous
years [10]. Rather than creating a solution tailored for this specific task we em-
ployed a general purpose annotation system and used the returned keywords to
identify the relevant concepts. The results show that while this approach logically
lags behind the precision of the more finely tuned solutions, it is still capable of
solving some instances pretty well.
    The paper is organized as follows. First, we briefly review the related work on
image annotation and distinguish two important classes of annotation methods.
Next, the ImageCLEF Annotation Task is described in more detail and the
quality of the training data is discussed. Finally, we present the methods used
by the MUFIN group, analyze their performance and discuss the results.


2    Related Work

The reason why we are interested in annotation is to simplify access to the
multimedia data. Depending on a situation, different types of metadata may be
needed, both in content and form. In [6], three forms of annotation are discussed:
free text, keywords chosen from a dictionary, and concepts from some ontology.
While the free text annotation does not require any structure, the other two
options pose some restrictions on the terminology used, and in particular make
the selection of keywords smaller. On certain conditions, we then begin to call
the task classification or categorization rather than annotation.
    Even though the conditions are not strictly defined, the common understand-
ing is that classification task works with a relatively small number of concepts
and typically uses machine learning to create specialized classifiers for the given
concepts. To train the classifiers, a sufficiently large training dataset with labeled
data needs to be available. The techniques that can be engaged in the learning
are numerous, including SVMs, kNN classifiers, or probabilistic approaches [7].
Study [10] describes a number of classification setups using different types of
concept learning.
    On the contrary, annotation usually denotes a task where a very large or un-
limited number of concepts is available and typically no training data is given.
The solution to such task needs to exploit some type of data mining, in case of
image annotation it is often based on content-based image retrieval over collec-
tions of images with rich metadata. Such system is described for example in [8].
Typically, the retrieval-based annotation systems exploit tagged images from
photo-sharing sites.


3    ImageCLEF Photo Annotation Task

In the ImageCLEF Photo Annotation Task, the participants were asked to as-
sign relevant keywords to a number of test images. The full setup of the contest
                            MUFIN at ImageCLEF 2011: Success or Failure?             3

is described in the Task overview [11]. From our perspective, two facts are im-
portant: (1) the keywords to be assigned were chosen from a fixed set of 99
concepts, and (2) a set of labeled training images was available. Following our
definition of terms from the previous section, the task thus qualifies as a classifi-
cation problem. As such, it is most straightforwardly solved by machine learning
approaches, using the training data to tune the parameters of the model. The
quality of the training data is then crucial for the correctness of the classification.
    As explained in [10], it is difficult to obtain a large number of labeled images
both for training and contest evaluation. It is virtually impossible to gather such
data only with the help of a few domain experts. Therefore, only a part of the
data was labeled by domain experts. The rest was annotated in a crowdsourcing
way, using workers from the Amazon Mechanical Turk portal. Even though the
organizers of the contest did their best to ensure that only sane results would be
accepted, the gathered data still contain some errors. In the following, we would
like to comment on some of them that we have noticed, so that they could be
corrected in the future. Also, we will discuss later how these errors may have
influenced the performance of the annotation methods.

3.1   Training data deficiencies
During the preparation of our solution for the Photo Annotation Task, we have
identified to following types of errors in the labeled training data:
 – Logical nonsense: Some annotations in the training dataset contradict the
   laws of the real world. The most significant nonsense we found was a num-
   ber of images with the following triplet of concepts: single person, man,
   woman. Such combination appeared for 784 images. Similarly, still life
   and active are concepts that do not match together. Though the emotional
   annotations are more subjective and cannot be so easily discarded as non-
   sense, we also believe that annotating an image by both cute and scary
   concepts is an oxymoron.
 – Annotation inconsistence: Since the contest participants were not provided
   by any explanation of the concepts, it was not always clear to us what counts
   as relevant for a given concept and what does not. One such unclear concept
   was single person. Should a part of a body be counted as a person or
   not? It seems that this question was also unclear to the trainset annotators
   as the concept bodypart sometimes co-occurred with single person and
   sometimes with no person.
 – Concept overuse: The emotional and abstract concepts are definitely diffi-
   cult to assign. Even within our research group, we could not decide what
   determines the “cuteness” of an image or what is the definition of a “nat-
   ural” image. Again, the labeled training data were not of much help as we
   could not discover any inner logic in them. We suspect that the Amazon
   Turk workers solved this problem by using the emotional and abstract terms
   as often as possible. For illustration, from among the 8000 training images
   3910 were labeled cute, 3346 were considered to be visual arts and 4594
   images were perceived as natural.
4        Petra Budikova, Michal Batko, and Pavel Zezula

Each error type is illustrated by a few examples in Figure 1.




Fig. 1. Trainset labeling errors: a) logical nonsense, b) annotation inconsistence, c)
concept overuse.




4     Our solution

The MUFIN Annotation Tool came into existence in the beginning of this year as
an extension of the MUFIN Image Search1 , a content-based search engine that
we have been developing for several years. Our aim was to provide an online
annotation system for arbitrary web images. The first prototype version of the
system is available online2 and was presented in [3]. As the first experiments with
the tool provided promising results, we decided to try our tool in the ImageCLEF
Annotation Task.
   The fundamental difference in the basic orientation of MUFIN Annotation
Tool and the Annotation Task is that our system provides annotations while the
task asks for classification. Our system provides free-text annotation of images,
1
    http://mufin.fi.muni.cz/imgsearch/
2
    http://mufin.fi.muni.cz/annotation/
                          MUFIN at ImageCLEF 2011: Success or Failure?        5

using any keywords that seem relevant using the content-based searching. To be
able to use our tool for the task, we needed to transform the provided keywords
into the restricted set of concepts given by the task. Moreover, even though
the MUFIN tool is quite good at describing the image content it does not give
much information about emotions and technical-related concepts (we will discuss
the reasons later). Therefore, we also had to extend our system and add new
components for specialized processing of some concepts. The overall architecture
of the processing engaged in the Annotation Task is depicted in Figure 2. In the
following sections, we will describe the individual components in more detail.




                        Fig. 2. Concept retrieval schema.
6       Petra Budikova, Michal Batko, and Pavel Zezula

4.1   MUFIN Annotation Tool

The MUFIN Annotation Tool is based on the MUFIN Image Search engine,
which retrieves the nearest neighbors of a given image based on visual and text
similarity. The image search system, described in full detail in [1], enables fast
retrieval of similar images from very large collections. The visual similarity is
defined by five MPEG-7 global descriptors – Scalable Color, Color Structure,
Color Layout, Edge Histogram, and Region Shape – and their respective dis-
tance functions [9]. If some text descriptions of images are available, their tf-idf
similarity score is also taken into consideration. In the ImageCLEF contest, free-
text image descriptions and EXIF tags were available for some images. Together
with the image, these were used as the input of the retrieval engine.
    To obtain an annotation of some input image, we first evaluate the near-
est neighbor query over a large collection of high-quality images with rich and
trustworthy text metadata. In particular, we are currently using the Profimedia
dataset, which contains 20M images from a microstock site [2]. When the query
is processed by the MUFIN search, we obtain a set of images with their respec-
tive keywords. In the Profimedia dataset, each image is accompanied by a set
of title words (typically 3 to 10 words) and keywords (about 20 keywords per
image in average). Both the title words and the keywords of all images in the
result set are merged together (the title words receiving a higher weight) and
the frequencies of individual lemmas are identified. A list of stopwords and the
WordNet lexical database [5] are used to remove irrelevant word types, names,
etc. The Annotation Tool then returns the list of the most frequent keywords
with the respective frequencies, which express the confidence of the annotation.


4.2   Annotation to concept transformation

To transform the free-text annotation into the ImageCLEF concepts it was nec-
essary to find the semantic relations between the individual keywords and con-
cepts. For this purpose, we used the WordNet lexical database, which provides
structured semantic information for English nouns, verbs, adjectives, and ad-
verbs. The individual words are grouped into sets of cognitive synonyms (called
synsets), each expressing a distinct concept. These synsets are interlinked by
different semantic relations, such as hypernym/hyponym, synonym, meronym,
etc. It is thus possible to find out whether any two synsets are related and how.
In our case, we were interested in the relationships between the synsets of our
annotation keywords and the synsets of the ImageCLEF concepts. To obtain
them, we first needed to find the relevant synsets for our keywords and the task
concepts.
    The WordNet synset is defined as a set of words with the same meaning,
accompanied by a short description of the respective semantic concept. Very
often, a word has several meanings and therefore is contained in several synsets.
For instance, the word “cat” appears both in a synset describing the domestic
animal and a synset describing an attractive woman. If we have only a keyword
and no other information about the word sense or context, we need to consider
                           MUFIN at ImageCLEF 2011: Success or Failure?         7

all synsets that contain this keyword. This is the case of keywords returned by
the MUFIN Annotation Tool. We do not try here to determine whether the
synsets are really relevant but rely on the majority voting of a large number of
keywords that are processed.
    The situation is however different in case of the ImageCLEF concepts where
it is much more important to know the correct synsets. Fortunately, we have
actually two possible ways of determining the relevant synsets. First, we can
sort them out manually since the number of concepts is relatively small. The
other, more systematic solution, will run the whole annotation process with all
the candidate synsets of the concepts, log the contributions of the respective
individual synsets, evaluate their performance, and rule out those with a low
success rate.
    Once the synsets are determined, we can look for the relationships. Again,
there are different types of relations and some of them are relevant for one con-
cept but irrelevant for another. Again, we have the same two possibilities of
choosing the relevant relationships as in the case of synsets. In our implemen-
tation, we have used the manual cleaning approach for synsets and automatic
selection approach for relationships.
    With the relevant synsets and relationships, we count a relevance score of
each ImageCLEF concept during the processing of each image. The score is
increased each time a keyword-synset is related to concept-synset. The increase
is proportional to the confidence score of the keyword as produced by the MUFIN
Annotation Tool.
    Finally, the concepts are checked against the OWL ontology provided within
the Annotation Task. The concepts are visited in a decreasing order of their
scores and whenever a conflict between two concepts is detected, the concept
with a lower score is discarded.

4.3   Additional image processing
The mining in keywords of similar images allows us to obtain such information
as is usually contained in the image descriptions. This is most often related to
image content, so the concepts related to nature, buildings, vehicles, etc. can be
identified quite well. However, the Annotation Task considers also concepts that
are less often described in the text. To get some more information about these,
we employed the following three additional information sources:

 – Face recognition: The face recognition algorithms are well-known in the im-
   age processing. We employ face recognition to determine the number of per-
   sons in an image.
 – EXIF tag processing: Some of the input photos are accompanied by EXIF
   tags that provide information about various image properties. When avail-
   able, we use these tags to decide the relevance of concepts related to illumi-
   nation, focus, and time of the day.
 – MUFIN visual search: Apart from the large Profimedia collection, we also
   have the training dataset that can be searched with respect to the visual
8         Petra Budikova, Michal Batko, and Pavel Zezula

      similarity of images. However, since the trainset is rather small and some
      concepts are represented by only a few images, there is quite a high prob-
      ability that the nearest neighbors will not be relevant. Therefore, we only
      consider neighbors within a small range of distances (determined by experi-
      ments).

4.4     Trainset statistics input
Definitely the most difficult concepts to assign are the ones related to user’s
emotions and also the abstract concepts such as technical, overall quality,
etc. As discussed in Section 3.1, it is not quite clear either to us or to the people
who annotated the trainset what these concepts precisely mean. Therefore, it is
very difficult to determine their relevance using the image visual content. The
text provided with the images is also not helpful in most cases.
    We finally decided to rely on the correlations between image content and
the emotions it most probably evokes. For example, images of babies or nature
are usually deemed cute. A set of such correlation rules was derived from the
trainset and used to choose the emotional and abstract concepts.

4.5     Our submissions at ImageCLEF
In our primary run, all the above-described components were integrated as de-
picted in Figure 2. In addition, we submitted three more runs where we tried
various other settings to find out whether the proposed extensions really pro-
vided some added quality to the search results. The components that were left
out in some experiments are depicted by dashed lines in Figure 2. We also ex-
perimented with the transformation of our concept scores into the confidence
values expressed as percentage, which was done either as concept-specific or
concept-independent. The individual run settings were as follows:
    – MufinSubmission100: In this run, all the available information was exploited
      including photo tags, EXIF tags, visual image descriptors, and trainset
      statistics. Concept-specific mapping of annotation scores to confidence values
      was applied.
    – MufinSubmission101: The same settings were used for this run as in the
      previous case but concept-independent mapping of annotation scores was
      applied.
    – MufinSubmission110: In this run, we did not use the EXIF tags for the pro-
      cessing of concepts related to daytime and illumination as described in Sec-
      tion 4.3. The MUFIN visual search in the trainset was omitted as well. How-
      ever, the textual EXIF tags were used as a part of the input for the MUFIN
      Annotation Tool. Concept-independent mapping of annotation scores was
      applied.
    – MufinSubmission120: In this run, the EXIF tags were not applied at all,
      neither as a part of the text-and-visual query in the basic annotation step nor
      in the additional processing. Again, the MUFIN visual search was skipped.
      Concept-independent mapping of annotation scores was applied.
                            MUFIN at ImageCLEF 2011: Success or Failure?            9

5     Discussion of results

As detailed in [11], the following three quality metrics were evaluated to compare
the submitted results: Mean interpolated Average Precision (MAP), F-measure
(F-ex), and Semantic R-Precision (SR-Precision). As we expected, our best sub-
mission was MufinSubmission100 which achieved 0.299 MAP, 0.462 F-ex, and
0.628 SR-precision. The other submissions that we have tried received slightly
worse scores. After the task evaluation and the release of the algorithm for com-
puting the MAP metric, we also re-evaluated our system with better settings of
the MUFIN Annotation Tool that we have improved since the ImageCLEF task
submission. Using these, we were able to gain one or two percent increase of the
MAP score. With respect to the MAP measure, our solution ranked at position
13 among the 18 participating groups.
   Apart from the overall results, it is also interesting to take a closer look at the
performance of the various solutions for individual concepts. The complete list of
concept results for each group is available on the ImageCLEF web pages3 . Here
we focus on the categories and particular examples of concepts where MUFIN
annotation performed either well or poorly and discuss the possible reasons.
   First of all, we need to define what we consider a good result. Basically, there
are two ways: either we only look at the performance, e.g. following the MAP
measure, or we consider the performance in relation to the difficulty of assigning
the given concept. The assigning difficulty can be naturally derived from the
competition results – when no group was able to achieve high precision with
some concept, then the concept is problematic. Since we believe that the second
way is more suitable, we express our results as a percentage of the best MAP
achieved for the given concept. Table 1 and Figure 3 summarize the results of
MufinSubmission100 expressed in this way.




               Fig. 3. MUFIN relative MAP performance per concept.

3
    http://imageclef.org/2011/Photo
10      Petra Budikova, Michal Batko, and Pavel Zezula

                                Landscape elements      58.2 %
            Content element     Pictured objects        63.0 %    62.4 %
                                Urban elements          73.3 %
                                Expressed impression    49.5 %
            Impression                                            52.6 %
                                Felt impression         58.6 %
                                Aesthetics              60.0 %
            Quality                                               56.3 %
                                Blurring                54.6 %
                                Art                     58.6 %
                                Impression              55.1 %
            Representation      Macro                   54.4 %    56.4 %
                                Portrait                62.1 %
                                Still life              42.0 %
                                Abstract categories     64.0 %
                                Activity                49.4 %
                                Events                  24.7 %
            Scene description                                     60.8 %
                                Place                   72.5 %
                                Seasons                 69.4 %
                                Time of day             67.9 %

Table 1. MUFIN average relative MAP performance per category – averages over the
concepts in the respective category are shown.


    Table 1 shows the results averages in groups of semantically close categories
as specified by the ontology provided for ImageCLEF. We can observe that the
MUFIN approach is most successful in categories that are (1) related to visual
image content rather than higher semantics, and (2) probable to be reflected
in image tags. These are, in particular, the categories describing the depicted
elements, landscape, seasons, etc. Categories related to impressions, events, etc.
represent the other end of the spectrum; they are difficult to decide using only
the visual information and (especially the impressions) are rarely described via
tags.
    However, the average MAP values do not differ that much between categories.
The reason for this is revealed if we take a closer look at the results for individual
concepts, as depicted in Figure 3. Here we can notice low peaks in otherwise
well performing categories and vice versa. For instance, the clouds concept in
the landscapes category performs rather poorly. This is caused by the fact that
clouds appear in many images but only as a part of a background, which is
not important enough to appear in the annotation. On the contrary, airplanes
are more interesting and thus regularly appear in the annotations. In fact, we
again encounter the difference between the annotation and classification tasks –
in annotation we are usually interested in the most important/interesting tags
while in classification all relevant tags are wanted.
    Several more extremes are pointed out in Figure 3. For instance, the concept
cute performs well because of its high frequency in the dataset. On the other
hand, for the concept overexposed a specialized classifier is much more suitable
than the annotation mining. The detailed discussion of the best fitting methods
for individual categories is beyond the scope of this paper. However, we believe
                            MUFIN at ImageCLEF 2011: Success or Failure?             11

that it is worth further studies to sort out the different types of concepts as well
as annotation approaches and try to establish some relationships between them.


6   Conclusions
In this study, we have described the MUFIN solution of the ImageCLEF Photo
Annotation Task, which is based on free-text annotation mining, and compared
it to more specialized, classifier-based approaches. The method we presented has
its pros and cons. Mining information from annotated web collections is compli-
cated by a number of features related to the way the annotations are created.
As discussed in [12], we need to expect errors, typing mistakes, synonyms, etc.
However, there are also ways of overcoming these difficulties. In our approach,
we have exploited a well-annotated collection, the semantical information pro-
vided by WordNet, and a specialized ontology. Using these techniques, we have
been able to create an annotation system that shows precision comparable to
average classifiers, which are usually trained for specific purposes only. The main
advantage of our solution lies in the fact that it requires minimum training (and
is therefore less dependent on the availability of high-quality training data) and
is scalable to any number of concepts.


Acknowledgments
This work has been partially supported by Brno PhD Talent Financial Aid and by
the national research projects GAP 103/10/0886 and VF 20102014004. The hardware
infrastructure was provided by the METACentrum under the programme LM 2010005.


References
 1. Batko, M., Falchi, F., Lucchese, C., Novak, D., Perego, R., Rabitti, F., Sedmidub-
    sky, J., Zezula, P.: Building a web-scale image similarity search system. Multimedia
    Tools and Applications 47, 599–629 (2010), http://dx.doi.org/10.1007/s11042-009-
    0339-z, 10.1007/s11042-009-0339-z
 2. Budikova, P., Batko, M., Zezula, P.: Evaluation platform for content-based image
    retrieval systems. In: To appear in Theory and Practice of Digital Libraries (TPDL
    2011) (26-28 September 2011)
 3. Budikova, P., Batko, M., Zezula, P.: Online image annotation. In: 4th Interna-
    tional Conference on Similarity Search and Applications (SISAP 2011). pp. 109–
    110 (2011)
 4. Datta, R., Joshi, D., Li, J., Wang, J.Z.: Image retrieval: Ideas, influences, and
    trends of the new age. ACM Comput. Surv. 40(2) (2008)
 5. Fellbaum, C. (ed.): WordNet: An Electronic Lexical Database. The MIT Press
    (1998)
 6. Hanbury, A.: A survey of methods for image annotation. J. Vis. Lang. Comput.
    19(5), 617–627 (2008)
 7. Kwasnicka, H., Paradowski, M.: Machine learning methods in automatic image
    annotation. In: Advances in Machine Learning II, pp. 387–411 (2010)
12      Petra Budikova, Michal Batko, and Pavel Zezula

 8. Li, X., Chen, L., 0001, L.Z., Lin, F., Ma, W.Y.: Image annotation by large-scale
    content-based image retrieval. In: Nahrstedt, K., Turk, M., Rui, Y., Klas, W.,
    Mayer-Patel, K. (eds.) ACM Multimedia. pp. 607–610. ACM (2006)
 9. MPEG-7: Multimedia content description interfaces. Part 3: Visual. ISO/IEC
    15938-3:2002 (2002)
10. Nowak, S., Huiskes, M.J.: New strategies for image annotation: Overview of the
    photo annotation task at imageclef 2010. In: Braschler, M., Harman, D., Pianta,
    E. (eds.) CLEF (Notebook Papers/LABs/Workshops) (2010)
11. Nowak, S., Nagel, K., Liebetrau, J.: The CLEF 2011 Photo Annotation and
    Concept-based Retrieval Tasks. In: CLEF 2011 working notes (2011)
12. Trant, J.: Studying social tagging and folksonomy: A review and framework. J.
    Digit. Inf. 10(1) (2009)