<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>The Wroclaw University of Technology Participation at ImageCLEF 2010 Photo Annotation Track?</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Michal Stanek</string-name>
          <email>michal.stanek@pwr.wroc.pl</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Oskar Maier</string-name>
          <email>oskar.maier@student.pwr.wroc.pl</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Halina Kwasnicka</string-name>
          <email>halina.kwasnicka@pwr.wroc.pl</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Wrocław University of Technology, Institute of Informatics</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this paper we present three methods for image autoannotation used by the Wroclaw University of Technology group at ImageCLEF 2010 Photo Annotation track. All of our experiments focus on robustness of the global color and texture image features in connection with different similarity measures. To annotate training set we use two version of PATSI algorithm which searches for the most similar images and transferring annotations from them to the target image by applying transfer function. We use both the simple version of the algorithm working only on single similarity matrix, as well as multi-PATSI which uses many similarity measures in order to obtain the final annotations. As third approach to image auto-annotation we use Penalized Discriminant Analysis to train multi class classifier in One-vs-All manner. During training and optimization process of all annotators we use F-measure as evaluation measure trying to achieve its highest value on a training set. Obtained results indicate that our approach achieved a high quality measure only for a small group of terms and it is necessary to take into account also local image characteristics.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Recently, Makadia et. al. [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] proposed a family of image annotation baseline
methods that are build on the hypothesis that visually similar images are likely
to share the same annotations. They treat image annotation as a process of
transferring labels from nearest neighbours. Makadia’s method does not solve the
fundamental problem of determining the number of annotations that should be
assigned to the target image. Thus they assume a constant number of annotations
per image. The transfer is performed in two steps: all annotations from the most
similar image are rewritten and the most frequent words are chosen from the
whole neighbourhood until a given annotation length has been achieved.
      </p>
      <p>
        We extend Makadia’s approach by constructing PATSI (Photo Annotation
through Similar Images) annotator which introduce transfer function [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] as well
? This work is partially financed from the Ministry of Science and Higher
Education Republic of Poland resources in 2008–2010 years as a Poland–Singapore joint
research project 65/N-SINGAPORE/2007/0.
as optimization algorithm which can be used to find optimal number of
neighbours and the best transfer threshold according to the specified quality
measure [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. During our experiments with different similarity metrics we extend this
algorithm to multi-PATSI which perform annotation transfer process based onto
many similarity matrices calculated using different feature sets and similarity
measures and combine results into final annotation based on the quality of each
annotator for specific words.
      </p>
      <p>
        At ImageCLEF 2010 photo annotation track [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] we evaluate PATSI and
multi-PATSI approach with global image features. During experiments we use
grid segmentation and statistical color informations as well as features extracted
using LIRE package [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. As third type of automatic image annotator we train
PDA [
        <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
        ] classifier onto CEDD [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] and Jpeg Coefficient Histogram [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] in
Onevs-All manner.
      </p>
      <p>This paper is organized as follows. In the next section we describe used
automatic image annotation methods with explanation of used features, distance
measures and details of annotation algorithm. The third section describes the
experiments and achieved results. The paper is finished with conclusions and
remarks on possible further improvements of the method.
2</p>
      <p>
        Annotation process
In this section we describe automatic image annotation methods used during
ImageCLEF 2010 Photo Annotation track [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] by our team. First we focus on
types of visual features extracted from images and similarity measures used to
build similarity matrices then we describe the annotation transfer process.
2.1
      </p>
    </sec>
    <sec id="sec-2">
      <title>Visual Features</title>
      <p>The image I in a training dataset D is represented by a n-dimensional vector
of visual features vI = (v1I ; ; vnI ). All visual features are a m-dimensional
vector of low level attributes viI = (xi1;I ; ; xim;I ). The visual features must be
extracted from the image and can represent information about color and texture
for the entire image, or only selected area of the image I.</p>
      <p>
        For all images in both training and tasting dataset we performed visual
feature extraction using self made feature extractor and the image descriptors
contained in the LIRE package [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. We focused mainly on global image
characteristics, but we use also more local information obtained after splitting image by
rectangular 5-by-5 and 20-by-20 grid. The list of extracted features include:
1. From MPEG-7 standard [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] we use following image descriptors calculated
for the whole image:
– Fuzzy Color Histogram – 125 dimensions
– JPEG Coefficient Histogram – 192 dimensions
– General Color Layout – 18 561 dimensions
– Color and Edge Directivity Descriptor (CEDD)[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] – 120 dimensions
– Fuzzy color and texture histogram (FCTH)[
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] – 192 dimensions
2. Tamura features first three from six texture features corresponding to
human visual perception [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]:
– coarseness – size of the texture elements,
– contrast – contrast stands for picture quality,
– directionality – texture orientation.
      </p>
      <p>
        Tamura features vector has 16 dimensions.
3. Auto Color Correlogram features defined in [
        <xref ref-type="bibr" rid="ref12 ref13">12, 13</xref>
        ] – 256 dimensions
4. Gabor texture features [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] – 60 dimensions
5. Statistical color and edges information of image regions (5-by-5 and 20-by-20
grid) in two color spaces RGB and HSV:
– x and y coordinates of the segment center – 2 dimensions,
– the mean value of color in each channel of the color space – 3 dimensions,
– standard deviations of color changes in each channel for a given color
space – 3 dimensions,
– mean eigenvalues of color Hessian in each channel for a given color space
– 3 dimensions.
6. CoOccurance Matrix [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] calculated for each segment of 5-by-5 and
20by-20 segmentation – 21 dimensions
2.2
      </p>
    </sec>
    <sec id="sec-3">
      <title>Distance Metrics</title>
      <p>To obtain the similarity or rather dissimilarity between two images, we measure
the distance between vectors in metric space and divergence between
distributions build onto visual vectors. In our experiments we use distance measures
described below.</p>
      <p>Minkowski distance The Minkowski distance is widely used for measuring
similarity between objects (e.g., images). The Minkowski metric between image
A and B is defined as:
dCos(A; B) =
1</p>
      <p>(vA)(vB)T
kvAk2kvBk2</p>
      <p>:
dMK(A; B) =
n
X viA
i=1
viB p
!1=p
where p is the Minkowski factor for the norm. Particularly, when p is equal one
and two, it is the well known L1 and Euclidean distance respectively.
Cosine distance is a measure of similarity between two vectors of n
dimensions by finding the cosine of the angle between them, often used to compare
documents in text mining
(1)
(2)
Manhattan distance also called cityblock distance or the taxicab metric is
the metric of the Euclidean plane defined by:
dManh(A; B) =</p>
      <p>X (viA
i
viB)
Correlation distance measures the similarity in shape between vectors defined
by
dCorr(u; v) =
1
k(vA
(vA</p>
      <p>vA)(vB
vA)k2k(vB
vB)T
vB)k2T
;
where k(u</p>
      <p>u)k2 is L2 distance between vector u and mean vector u.</p>
      <p>Jensen–Shannon Divergence Based on the visual feature vectors vI one
can build a model M I for the image I. We can assume that M I is a
multidimensional random variable described by multi-variate normal distribution and
all vectors viI are realizations of this model. The probability density function
(PDF) for the model M I is defined as:</p>
      <p>M I (x; ; ) =</p>
      <p>
        1
(2 )N=2j j1=2
exp
where x is the observation vector, the mean vector, and the covariance
matrix. Both and are parameters of the model calculated using
ExpectationMaximization -algorithm [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] on all visual features [v1I ; ; vnI ] of the image I. In
order to avoid problems of inverting covariance matrix (avoid matrix singularity)
one may perform regularization of the covariance matrix. Models of images are
build for all images in the training set, as well as for the query image.
      </p>
      <p>Distance between the models can be computed as Jensen–Shannon
divergence, which is a symmetrized version of Kullback–Leibler divergence:</p>
      <p>DKL(M BkM A);
1
2
where M A, M B are models (PDF) for the images A and B, and DKL is the
Kullback-Leibler distance which for multivariate-normal distribution takes the
form:</p>
      <p>DKL(MA
kMB) =
+
1
2 loge
1
2 ( B
det
det</p>
      <p>B
A
+ tr
where A, B and A, B are covariance matrices and mean vectors from the
respective image models A and B.
2.3</p>
      <p>Automatic Image Annotation Methods
We use three methods of automatic image annotation, such as PATSI (Photo
Annotation through Similar Images) annotator, multi-PATSI annotator and
multiclass PDA classificator. Details of all of those methods are described below.
PATSI Annotator In the PATSI (Photo Annotation through Finding Similar
Images) approach, for a query image Q, a vector of the k most similar images
from the training dataset D needs to be found based on the similarity distance
measure d. Let [r1; ; rk] be the ranking of k the most similar images ordered
decreasingly by similarity. Based on the hypothesis that images similar in
appearance are likely to share the same annotation, keywords from the nearest
neighbours are transferred to the query image. All labels for the image on
position r in the ranking are transferred with a value designated by the transfer
function '(ri).</p>
      <p>To assure that labels from more similar images have a larger impact on
resulting annotation we define ' as
1
'(ri) = ; (8)
i
where ri is an image on position i in the ranking. All words associated with
image ri are then transferred to the resulted annotation with the associated
transfer value 1=i. If the words has been transferred before the transfer values
are summed.</p>
      <p>The resulting query image annotation consists of all the words whose transfer
values were greater than a specified threshold t. The threshold value t has an
impact on the resulting annotation length and its optimal value as well as the
optimal number of neighbours k which should be taken into account during
the annotation process must both be found using an optimization process. The
outline of the PATSI annotation method is presented in the figure 1 and is
summarized in the Algorithm 1.</p>
      <p>The optimal parameters k and t differ greatly not only for different databases,
but also between feature sets, methods of distance measure and transfer
functions. There exists no optimal choice of them that would be suitable in all cases.
We need to adjusting them in each explicit case.</p>
      <p>Algorithm 1 PATSI image annotation algorithm
Require: D – training dataset
d – distance function
Q – annotation quality function
' - transfer function
1: fPreparation Phaseg</p>
      <p>calculate and store visual features all images in training dataset D
2: calculate similarity matrix using distance function d between all images in training
dataset D
3: fOptimization Phaseg</p>
      <p>choose values for k and t maximizing quality function Q on a training dataset
4: fQuery Phaseg</p>
      <p>calculate the visual features for query image Q
5: calculate the distance from query image to all other images in training database</p>
      <p>D.
6: take k images with the smallest distances between the models and create a ranking
of those images.
7: transfer all words from the images in the ranking with the value '(r), where r is
the position of the image in the ranking.
8: as a final annotation take the words which transfer values sum is greater or equal
to the provided threshold t value.</p>
      <p>
        Finding t and k proves to be a non-trivial task. The commonly used
optimization solvers are inapplicable due to the non-linear character of quality
function Q (discrete domain on k and continues on t). To efficiently find t and
k we propose and use the iterative refinement algorithm which is described
in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>Multi-PATSI Annotator During experiments we spot that some of the
features as well as distance metrics are more suitable to detect some groups of
words, while showing a weak performs for others. By combining them together
we can increase overall annotation performance. We propose the multi-PATSI
method that take advantage of this observation by joining together the strengths
of a number of annotation techniques.</p>
      <p>The overall schema of multi-PATSI approach is presented in figure 2. In the
first step we run PATSI algorithm separately for each features sets and distance
functions to obtain annotation vectors. Each element of those vectors represent
whether word should be assigned or not to the query image Q (class f 1; 1g).</p>
      <p>For each of the PATSI annotators at learning stage the performance vector is
calculated. The performance vector corresponds to the efficiency of the
PATSIannotator for each of the annotated words on the testing set.</p>
      <p>
        For each PATSI-annotator the resulted annotation vector is multiplied by
a performance vector to obtain weighted annotator response. All weighted
responses are then summed together creating final annotation. All concepts which
obtain value greater than a threshold tmulti are treated as a final annotation for
a query image Q. Optimal threshold value tmulti can be calculated using
crossvalidation method and optimization technique such as iterative refinement [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
Multi-Class Classification As third annotation method we use Penalized
Discriminant Analysis classifier[
        <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
        ] from Python Machine Learning Module –
MLPY [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] in One vs. All scenario.
      </p>
      <p>In this approach for each concept we train a separate PDA classifier using
the extracted image features. We use all features from images annotated by a
specific concept as positive examples and others as negative. In training we use
four fold cross-validation.
3</p>
      <p>Experimental Results
We submitted five runs for the annotators and features sets described in previous
section:
1. PATSI with Kullback Leibler divergence - hsv color space and grid 20-by-20,
2. PATSI with Kullback Leibler divergence - rgb color space and grid 20-by-20,
3. Multi-PATSI with features presented in Table 2,
4. PDA classifier with CEDD features,
5. PDA classifier with Jpeg Coefficient Histogram features.</p>
      <p>The official results of the five runs in terms of Average Precision (AP ),
Average Equal Error Rate (Avg. EER), Average Area Under Curve (Avg. AUC)
are reported in Table 1. A detailed overview of the quality of annotations for
each of the submitted methods for the 30 best-annotated words is presented in
Table 3.
4</p>
      <p>Conclusion
During the training and optimization process the parameters of the classifiers
was tuned using the F-measure (harmonic mean of precision and recall) instead
of the Average Precision. F-measure resulted that in all submitted annotations
results we optimize annotation length by providing annotation vectors contained
only f 1; 1g values. Using vectors prepared in such a way results in low Average
Precision quality.</p>
      <p>The published results show that the highest measure of quality according to
AP measure, reached the multi-class PDA classifier with CEDD features. On the
other hand the worst in comparison was multi-PATSI annotator.
P S</p>
      <p>H
e (
g
a I
r S
e</p>
      <p>T
v</p>
      <p>A
A P
.
3
e
l
b
a
T
)
B
G
R
(
I
S
T
A
P
t
p
e
c
n
o
s
n
o r
u ru rs l o</p>
      <p>e ra o
S s</p>
      <p>n
a o
u s r r
r
u oo
s
n
o
S s</p>
      <p>n
a o r
u s r l o
r
e lu a o
e
im ed</p>
      <p>r
T r</p>
      <p>u
la l
u B
s
e
m
i
T
l
a
u
e
m
i
a
u
s
T s
l t
e
r
e u</p>
      <p>t
im a
T N
l e
a p
e
r
u e
t
a im
N T
e l
p a
a u
5 6 3 7 4 4 9 5 3 0 2 2 5 1 6 8 5 0 4 9 7 4 6 0 8 1 3 0 5 6
) P 6 9 3 6 5 4 2 6 6 3 6 3 7 5 4 3 2 2 9 8 8 8 7 7 5 4 3 3 0 9
t A ,9 ,8 ,8 ,7 ,6 ,6 ,6 ,5 ,5 ,5 ,4 ,4 ,3 ,3 ,3 ,3 ,3 ,3 ,2 ,2 ,2 ,2 ,2 ,2 ,2 ,2 ,2 ,2 ,2 ,1
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
3 1 6 0 2 9 3 1 4 4 6 7 5 3 4 0 6 2 5 4 2 1 9 4 3 9 6 5 3 4
P 7 2 7 6 7 6 6 7 5 0 4 8 6 6 5 3 1 0 9 9 7 6 4 4 4 3 3 3 3 2
7 ,6 ,6 ,6 ,5 ,5 ,5 ,4 ,3 ,3 ,3 ,3 ,3 ,3 ,3 ,2 ,2 ,2 ,2 ,2 ,2 ,2 ,2 ,2 ,2 ,2 ,2
A ,09 ,09 ,07 ,0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
l i y</p>
      <p>l
t
e oV oB oP tau tud ay te V t
u</p>
      <p>t
u o ra ky lea lan o lo</p>
      <p>V u
u A
s is s r fe l
d o li
o y</p>
      <p>s
d it i
n
e sc
m a i
e
a l
u a d l
n
e lt li
g u</p>
      <p>l t t
m i
n d a u a r</p>
      <p>r
d e ro rca tse ie u</p>
      <p>i
t
a o teh lce yn
i
h n
C</p>
      <p>r
u ru rs l o</p>
      <p>e ra o
t
p
e l
c t e V d
o e oV oB oP tau tud ay ky tu o a l
n u n
u a
is sc ts s l t</p>
      <p>d a e
n u
a lo is
u s e o
nu la
r fe
o li l</p>
      <p>y t
d it</p>
      <p>B u</p>
      <p>s
y i t s y o eh</p>
      <p>V h e
r
a o ig e
r
nn rc t</p>
      <p>G r l
a se rak tea ie
e
c lt le
h u g</p>
      <p>d in
u
n
C N N N N n O D S c N L P C V S m I C P N N T S M A P WV A S
7 3 6 3 5 8 8 2 1 0 8 1 8 5 5 5 4 6 4 1 5 1 8 5 7 8 5 9 9 6
P 4 8 2 2 3 0 9 4 1 8 3 3 0 2 1 1 0 9 8 8 7 7 6 5 4 3 3 2 1 1
A ,9 ,8 ,7 ,
7 ,6 ,
6 ,5 ,
5 ,5 ,
4 ,4 ,
4 ,4 ,
3 ,3 ,3 ,3 ,
2 ,2 ,
2 ,
2 ,2 ,2 ,
2 ,2 ,2 ,
2 ,
2 ,2 ,2
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
r</p>
      <p>V P B tu td y y t
u
e o o o a u a k u l
N N N N n O D S c
e n d
a n</p>
      <p>V u
a o lo is
ts sc is s l</p>
      <p>t r
d a e o
u s
n d
u n
r
a i</p>
      <p>u
fe s
o le ly lty th iV yn se le r</p>
      <p>i
a it</p>
      <p>n e
g o u r
am tea rk
e
m C P N N S T f
C
.
.p 1 2 3 4 5 6 7 8 9 01 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30
L
e
c
a
l
P
l
a
e
lca ed</p>
      <p>r
P r</p>
      <p>u
la l
u B
s
i
y
e
lca ed</p>
      <p>r
P r</p>
      <p>u
la l
u B
e
s
i
ts rn
r u
A S
e
s
i
ts rn
r u
A S
s
t
r
s
t
r
e
r
u
t
a
N
e
p
e
r
u
t
a</p>
      <p>n
N o
e rs
p e
a P
s ts
d h
n g
e i
i
r S
F g
y in
n
o
s
r
e
s
t
h
g
i</p>
      <p>S
e
s
i
r
n
u
S
t fe r</p>
      <p>y
n t
u i</p>
      <p>e
ed lca
r
r P
lu l</p>
      <p>a
d
e
r
r
u
l
B
e
c
a
l
P
l
a
t
ep e l r is ts
c
o e oV oP oB tud ay tau teu oV k l
n u
l
a</p>
      <p>l
n le V t</p>
      <p>r
y a a o a lo is</p>
      <p>A a e P
d a ro scd lt li t l l
s l f
u u o h a g
nd a
n du ity i
g
e e y i</p>
      <p>h r i
n ld t t
em in u u e o am rak tea
n i s r
a ly G r
7 3 7 6 5 5 5 1 3 9 5 7 4 1 6 9 4 9 6 5 8 2 9 5 1 1 0 3 1 9
P 4 8 1 1 3 5 3 1 4 6 2 0 9 9 8 7 7 6 6 5 2 2 1 1 1 1 1 0 0 9
A ,9 ,8 ,7 ,
7 ,6 ,
5 ,5 ,
5 ,4 ,
3 ,3 ,3 ,
2 ,
2 ,2 ,2 ,
2 ,
2 ,
2 ,2 ,
2 ,2 ,
2 ,
2 ,
2 ,2 ,
2 ,
2 ,2 ,1
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
S s</p>
      <p>n
a o r
u s r l o
r
u a o</p>
      <p>r
t l
p
e e l i
c t
o e oV oP oB tau tud ay te V
n u</p>
      <p>y s
u o k i
r
A
l
a
u le u
sd is</p>
      <p>l
V t</p>
      <p>r
a lo o a la ig
y ts t e il
n h s
o lt ro i</p>
      <p>u c li
o
d d a
n
g t</p>
      <p>P
n i
du trrao trea le le li
a g</p>
      <p>h
d t
m n s
em in a a e</p>
      <p>F ap itc</p>
      <p>e
y sc
7 3 9 8 5 6 8 4 1 1 9 9 9 5 2 2 9 7 3 5 0 2 0 5 4 4 9 1 5 7
P 4 8 1 1 3 8 7 1 1 3 0 8 6 2 1 0 9 9 8 7 6 5 5 3 3 3 2 2 1 9
A ,9 ,
8 ,7 ,
7 ,6 ,
5 ,5 ,
5 ,
5 ,
4 ,4 ,
3 ,3 ,
3 ,3 ,3 ,
2 ,
2 ,
2 ,2 ,
2 ,2 ,
2 ,
2 ,
2 ,2 ,
2 ,
2 ,2 ,1
0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0
n
o
i
s
s
ts re
h p
g m
i I
S</p>
      <p>c
gn ite it
n
o
i
s
s
e
r
p
Im en</p>
      <p>d
c r
i
t a
n
o
i
s
s
e
r
p
m
I
c
s
d
n n
e e
i
r rd
F a</p>
      <p>The results show that the method of transferring annotations seems to be
very interesting concept. However, it will be necessary to use outside the global
characteristics of the image also the local features as well as adaptive metric
functions.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Makadia</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pavlovic</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kumar</surname>
            ,
            <given-names>S.:</given-names>
          </string-name>
          <article-title>A new baseline for image annotation</article-title>
          .
          <source>In: ECCV '08</source>
          , Berlin, Heidelberg, Springer-Verlag (
          <year>2008</year>
          )
          <fpage>316</fpage>
          -
          <lpage>329</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Stanek</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Broda</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kwasnicka</surname>
          </string-name>
          , H.:
          <article-title>Patsi - photo annotation through finding similar images with multivariate gaussian models</article-title>
          .
          <source>Lecture Notes in Computer Science, International Conference on Computer Vision and Graphics</source>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Stanek</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Maier</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kwasnicka</surname>
          </string-name>
          , H.:
          <article-title>PATSI - photo annotation through similar images with annotation length optimization</article-title>
          .
          <source>In: Intelligent information systems</source>
          . Publishing House of University of Podlasie (
          <year>2010</year>
          )
          <fpage>219</fpage>
          -
          <lpage>232</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Nowak</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huiskes</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>New strategies for image annotation: Overview of the photo annotation task at imageclef 2010</article-title>
          .
          <source>In: In the Working Notes of CLEF</source>
          <year>2010</year>
          .
          <article-title>(</article-title>
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Lux</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chatzichristofis</surname>
            ,
            <given-names>S.A.</given-names>
          </string-name>
          :
          <article-title>Lire: lucene image retrieval: an extensible java cbir library</article-title>
          .
          <source>In: MM '08: Proceeding of the 16th ACM international conference on Multimedia</source>
          , New York, NY, USA, ACM (
          <year>2008</year>
          )
          <fpage>1085</fpage>
          -
          <lpage>1088</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Ghosh</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Penalized discriminant methods for the classification of tumors from gene expression data</article-title>
          .
          <source>Biometrics</source>
          <volume>59</volume>
          (
          <issue>4</issue>
          ) (
          <year>2003</year>
          )
          <fpage>992</fpage>
          -
          <lpage>1000</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Hastie</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Buja</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tibshirani</surname>
          </string-name>
          , R.:
          <article-title>Penalized discriminant analysis</article-title>
          .
          <source>Annals of Statistics</source>
          <volume>23</volume>
          (
          <year>1995</year>
          )
          <fpage>73</fpage>
          -
          <lpage>102</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Chatzichristofis</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Boutalis</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Cedd: Color and edge directivity descriptor: A compact descriptor for image indexing and retrieval</article-title>
          .
          <source>Computer Vision Systems</source>
          (
          <year>2008</year>
          )
          <fpage>312</fpage>
          -
          <lpage>322</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>S.F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sikora</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Puri</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Overview of the MPEG-7 Standard</article-title>
          .
          <source>IEEE Trans. Circuits and Systems for Video Technology</source>
          <volume>11</volume>
          (
          <issue>6</issue>
          ) (
          <year>2001</year>
          )
          <fpage>688</fpage>
          -
          <lpage>695</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Chatzichristofis</surname>
            ,
            <given-names>S.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Boutalis</surname>
            ,
            <given-names>Y.S.</given-names>
          </string-name>
          : Fcth:
          <article-title>Fuzzy color and texture histogram - a low level feature for accurate image retrieval. Image Analysis for Multimedia Interactive Services</article-title>
          , International Workshop on 0 (
          <year>2008</year>
          )
          <fpage>191</fpage>
          -
          <lpage>196</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Tamura</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mori</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yamawaki</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Texture features corresponding to visual perception</article-title>
          .
          <source>IEEE Transactions on System, Man and Cybernatic</source>
          <volume>6</volume>
          (
          <year>1978</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Goodrum</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Image information retrieval: An overview of current research</article-title>
          .
          <source>Informing Science</source>
          <volume>3</volume>
          (
          <year>2000</year>
          ) 2000
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kumar</surname>
            ,
            <given-names>S.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mitra</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhu</surname>
            ,
            <given-names>W.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zabih</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          :
          <article-title>Image indexing using color correlograms</article-title>
          .
          <source>In: CVPR '97: Proceedings of the 1997 Conference on Computer Vision and Pattern Recognition (CVPR '97)</source>
          , Washington, DC, USA, IEEE Computer Society (
          <year>1997</year>
          )
          <fpage>762</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wong</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Indrawan</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lu</surname>
          </string-name>
          , G.:
          <article-title>Content-based image retrieval using gabor texture features</article-title>
          .
          <source>In: IEEE Transactions PAMI</source>
          . (
          <year>2000</year>
          )
          <fpage>13</fpage>
          -
          <lpage>15</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Haralick</surname>
            ,
            <given-names>R.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shanmugam</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dinstein</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Textural features for image classification</article-title>
          .
          <source>IEEE Transactions on Systems, Man, and Cybernetics</source>
          <volume>3</volume>
          (
          <issue>6</issue>
          ) (
          <year>November 1973</year>
          )
          <fpage>610</fpage>
          -
          <lpage>621</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>McLachlan</surname>
            ,
            <given-names>G.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Krishnan</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <source>The EM Algorithm and Extensions (Wiley Series in Probability and Statistics)</source>
          .
          <volume>2</volume>
          edn. Wiley-Interscience (
          <year>March 2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Albanese</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Merler</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jurman</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Visintainer</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Furlanello</surname>
            ,
            <given-names>C.:</given-names>
          </string-name>
          <article-title>Mlpy machine learning py (</article-title>
          <year>2010</year>
          ) http://mloss.org/software/view/66/.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>