=Paper= {{Paper |id=Vol-1171/CLEF2005wn-ImageCLEF-HoiEt2005 |storemode=property |title=CUHK Experiments with ImageCLEF 2005 |pdfUrl=https://ceur-ws.org/Vol-1171/CLEF2005wn-ImageCLEF-HoiEt2005.pdf |volume=Vol-1171 |dblpUrl=https://dblp.org/rec/conf/clef/HoiZL05a }} ==CUHK Experiments with ImageCLEF 2005== https://ceur-ws.org/Vol-1171/CLEF2005wn-ImageCLEF-HoiEt2005.pdf
      CUHK Experiments with ImageCLEF 2005∗
                        Steven C.H. Hoi, Jianke Zhu and Michael R. Lyu
                        Department of Computer Science and Engineering
                             The Chinese University of Hong Kong
                                  Shatin, N.T., Hong Kong
                           {chhoi, jkzhu, lyu}@cse.cuhk.edu.hk



                                              Abstract


          This paper describes the empirical studies of cross-language and cross-media re-
      trieval for the ImageCLEF competition in 2005. It reports the empirical summary
      of the work of CUHK (The Chinese University of Hong Kong) at ImageCLEF 2005.
      This is the first participation of our group at ImageCLEF. The task we participated
      this year is the “Bilingual ad hoc retrieval” task. There are three major focuses and
      contributions in our participation. The first is the empirical evaluations of language
      models and the smoothing strategies for cross-language image retrieval. The second is
      the evaluations of cross-media image retrieval, i.e., combining text and visual content
      for image retrieval. The last one is the evaluation of the bilingual image retrieval be-
      tween English and Chinese. We provide empirical analysis on the experimental results.
      From the official testing results of the Bilingual ad hoc retrieval task, we achieve the
      highest MAP result (0.4135) in the monolingual query among all organizations.

Categories and Subject Descriptors
H.3 [Information Storage and Retrieval]: H.3.1 Content Analysis and Indexing; H.3.3 Infor-
mation Search and Retrieval; H.3.4 Systems and Software; H.3.7 Digital Libraries; H.2.3 [Database
Managment]: Languages—Query Languages

General Terms
Measurement, Performance, Experimentation

Keywords
Language Models, Text Based Image Retrieval, Multimodal Image Retrieval, Cross-Language
Retrieval, Cross-Media Retrieval, Smoothing Strategy


1    Introduction
Visual information retrieval has been an active research topic for many years. Although content-
based image retrieval (CBIR) has been received considerable studies in the community [9], there
is so far few benchmark image dataset available. The CLEF (Cross Language Evaluation Forum)
organization [7] began the ImageCLEF campaign from 2003 for benchmark evaluation of cross-
language image retrieval [4]. ImageCLEF 2005 offers four different tasks: bilingual ad hoc retrieval,
  ∗ The work described in this paper was fully supported by two grants: Innovation and Technology Fund

ITS/105/03, and the Research Grants Council Earmarked Grant CUHK4205/04E.
interactive search, medical image retrieval and automatic image annotation task. This is the first
participation of our CUHK group (The Chinese University of Hong Kong) at ImageCLEF. The
task we participated this year is the “Bilingual ad hoc retrieval”.
    In the past decade, traditional information retrieval mainly focused on the document retrieval
problems [3]. Along with more and more attentions in multimedia information retrieval in recent
years, the cross-language and cross-media retrieval have been put forward as an important research
topic in the community [4]. The cross-language image retrieval is to tackle the multimodal in-
formation retrieval task by unifying the techniques from traditional information retrieval, natural
language processing (NLP), and traditional CBIR solutions.
    In this participation, we offer the main contributions in three aspects. The first is the empirical
evaluation of language models and the smoothing strategies for cross-language image retrieval. The
second is the evaluation of cross-media image retrieval, i.e., combining text and visual content for
image retrieval. The last one is the methodology and empirical evaluation of the bilingual image
retrieval between English and Chinese.
    The rest of this paper is organized as follows. Section 2 introduces the TF-IDF retrieval
model and the language model based retrieval methods. Section 3 describes the details of our
implementation for this participation, and outlines our empirical study on the cross-language and
cross-media retrieval system. Finally section 4 concludes our work.


2     Language Models for Text Based Image Retrieval
In this participation, we conducted extensive experiments to evaluate the performance of Language
Models and the influences of different smoothing strategies. More specifically, two kinds of retrieval
models are studied in our experiments: (1) The TF-IDF retrieval model (2) The KL-divergence
language models based method. The smoothing strategies for Language Models are evaluated in
our experiments [11]: (1) Jelinek-Mercer (JM), (2) Dirichlet prior (DIR), (3) Absolute discounting
(ABS).

2.1    TF-IDF Similarity Measure for Information Retrieval
We incorporate the Language Models (LM) with the TF-IDF similarity measure[3]. TF-IDF is
widely used in information retrieval, which is a way of weighting the relevance of a query to a
document. The main idea of TF-IDF is to represent each document by a vector in the size of the
overall vocabulary. Each document Di is then represented as a vector (wi1 , wi2 ), · · · , win if n is
the size of the vocabulary. The entry wi,j is calculated as:

                                      wij = T Fij × log(IDFj )                                     (1)
where T Fij is the term frequency of the jth word in the vocabulary in the document Di , i.e. the
number of occurrences. IDFj is the inverse document frequency of the jth term, given as

                                             #documents
                          IDFj =                                                                   (2)
                                   #documents containing the jth term
The similarity between two documents is then defined as the cosine of the angle between the two
vectors.

2.2    Language Modeling for Information Retrieval
A statistical language model, or more simply a language model, is a probabilistic mechanism
for generating text. The first serious statistical language modeler was Claude Shannon [8]. In
exploring the application of his newly founded theory of information to human language, thought of
purely as a statistical source, Shannon measured how well simple n-gram models did at predicting,
or compressing, natural text. In the past several years there has been significant interest in the
use of language modeling methods for a variety of text retrieval and natural language processing
tasks [10].

2.2.1   The KL-divergence Measure
Given two probability mass functions p(x) and q(x), D(p||q), the Kullback-Leibler (KL) divergence
(or relative entropy) between p and q is defined as
                                                X             p(x)
                                    D(p||q) =       p(x)log                                      (3)
                                                x
                                                              q(x)

     One can show that D(p||q) is always non-negative and is zero if and only if p = q. Even though
it is not a true distance between distributions (because it is not symmetric and does not satisfy the
triangle inequality), it is still often useful to think of the KL-divergence as a ”distance” between
distributions [5].

2.2.2   The KL-divergence based Retrieval Model
For the language modeling approach, we assume a query q is generated by a generative model
p(q|θQ ), where θQ denotes the parameters of the query unigram language model. Similarly, we
assume that a document d is generated by a generative model p(q|θD ), where θQ denotes the
parameters of the document unigram language model. Let θ̂Q and θ̂D be the estimated query
and document language models respectively. The relevance value of d with respect to q can be
measured by the following negative KL-divergence function [10]:
                              X                            X
             −D(θ̂Q ||θ̂D ) =   p(w|θ̂Q )logp(w|θ̂D ) + (−   p(w|θ̂Q )logp(w|θ̂Q ))       (4)
                              w                               w

    In the above formula, the second term on the right-hand side of the formula is a query-
dependent constant, i.e., the entropy of the query model θ̂Q . It can be ignored for the ranking
purpose. In general, we consider the smoothing scheme for the estimated document model as
follows:
                                       ½
                                          ps (w|d)   if word w is seen
                         p(w|θ̂D ) =                                                         (5)
                                          αd p(w|C) otherwise
where ps (w|d) is the smoothed probability of a word seen in the document, p(w|C) is the collection
language model, and αd is a coefficient controlling the probability mass assigned to unseen words,
so that all probabilities sum to one [10]. In the subsequent section, we discuss several smoothing
techniques in details.

2.3     Several Smoothing Techniques
A smoothing method may be as simple as adding an extra count to every word, or words of
different count are treated differently. In order to solve the problem efficiently, we select three
representative methods that are popular and relatively efficient. The three methods are described
below.

2.3.1   Jelinek-Mercer (JM)
This method involves a linear interpolation of the maximum likelihood model with the collection
model, using a coefficient λ to control the influence of each model.

                               pλ (ω|d) = (1 − λ)pml (ω|d) + λp(ω|C)                             (6)
  Thus, this is a simple mixture model (but we preserve the name of the more general Jelinek-
Mercer method which involves deleted-interpolation estimation of linearly interpolated n-gram
models.
2.3.2    Dirichlet prior (DIR)
A language model is a multinomial distribution, for which the conjugate prior for Bayesian analysis
is the Dirichlet distribution with parameters (µ(ω1 |C), µp(ω2 |C), . . . , µp(ωn |C)). Thus, the model
is given by

                                              c(ω; d) + µp(ω|C)
                                    pµ (ω|d) = P                                                   (7)
                                                  ω c(ω; d) + µ
The Laplace method is a special case of the technique.

2.3.3    Absolute discounting (ABS)
The idea of the absolute discounting method is to lower the probability of seen words by subtracting
a constant from their counts. It is similar to the Jelinek-Mercer method, but differs in that it
discounts the seen word probability by subtracting a constant instead of multiplying it by 1 − λ.
The model is given by
                                         max(c(ω; d) − δ, 0)
                              pδ (ω|d) =    P                + δp(ω|C)                           (8)
                                              ω c(ω; d)

where δ ∈ [0, 1] is a discount constant and σ = δ|d|µ /|d|, so that all probabilities sum to one.
Here |d|µ is the number ofPunique terms in document d, and |d| is the total count of words in the
documents, so that |d| = ω c(ω; d).


         Table 1: Summary of three primary smoothing methods used in our submission

          Method                      ps (ω|d)                           αd         parameter

           JM               (1 − λ)pml (ω|d) + λp(ω|C)                λ                λ
                                   c(ω;d)+µp(ω|C)
                                    P                             Pµp(ω|C)
           DIR                          c(ω;d)+µ                         c(ω;d)+µ
                                                                                       µ
                                      ω                              ω
                                max(c(ω;d)−δ,0)
                                  P                                      δ|d|µ
           ABS       pδ (ω|d) =       c(ω;d)
                                                + δ|d|µ δp(ω|C)           |d|           δ
                                     ω



    The three methods are summarized in Table 1 in terms of ps (ω|d) and αd in the general form.
It is easy to see that a larger parameter value means smoothing in all cases. Retrieval using any
of the three methods can be very efficiently, when the smoothing parameter is given in advance.
It is as efficient as scoring using a TF-IDF model.


3       Cross-Language and Cross-Media Image Retrieval
In this section, we describe the experimental setup and our experimental development at the
ImageCLEF 2005. In addition, we analyze the results of our submission.

3.1     Experimental Setup
The bilingual ad hoc retrieval task is to find as many relevant images as possible for each given
topic. The St. Andrew collection is used as the benchmark dataset in the campaign. The collection
consists of 28,133 images, all of which associate with textual captions written in British English
(the target language). The caption consists of 8 fields including title, photographer, location,
date, and one or more pre-defined categories (all manually assigned by domain experts). In the
ImageCLEF 2005 campaign, there are totally 28 queries for each language. For each query, two
image samples are given. Figure 1. shows a query example of images, title and narrative texts in
the campaign.
                                                    

                                                     Number: 1 

                                                     aircraft on the ground 

                                                     Relevant images will show one or more
                                                    airplanes positioned on the ground. Aircraft do not
                                                    have to be the focus of the picture, although it should
                                                    be possible to make out that the picture contains
                                                    aircraft. Pictures of aircraft flying are not relevant and
                                                    pictures of any other flying object (e.g. birds) are not
                                                    relevant. 

                                                    

                 Figure 1: A query example in the ImageCLEF 2005 campaign.


3.2    Overview of Our Development
For the Bilingual ad hoc retrieval task, we studied the query tasks in English and Chinese (sim-
plified). Both text and visual information are used in our experiments. To study the language
models, we employ the Lemur toolkit [2] in our experiments. A list of standard stopwords is used
in the parsing step.
    To evaluate the influence on the performance by different schemes, we produced the results by
using different configurations. Tables 2 shows the configurations and the experimental results in
detail. In total, 36 runs with different configurations are submitted in our submission.

3.3    Analysis on the Experimental Results
In this part, we empirically analyze the experimental results of our submission. The goal of our
evaluation is to check whether the language model is effective for cross-language image retrieval
and what kinds of smoothing techniques achieve better performance. Moreover, we like to know
the performance comparison between the Chinese query and the monolingual query.
                                   0.8
                                                                                    CUHK−ad−eng−t−kl−ab3
                                                                                    CUHK−ad−eng−t−kl−jm2
                                   0.7
                                                                                    CUHK−ad−eng−t−kl−di2
                                                                                    CUHK−ad−eng−t−tf−idf2
                                                                                    CUHK−ad−eng−tv−kl−ab3
                                                                                    CUHK−ad−eng−tv−kl−jm2
                                   0.6



                                   0.5
                       Precision




                                   0.4



                                   0.3



                                   0.2



                                   0.1


                                         0   0.1   0.2   0.3   0.4     0.5    0.6    0.7    0.8   0.9       1
                                                                     Recall


       Figure 2: Experimental Result of Precision vs. Recall with Selected Configuration
              Table 2: The configurations and testing results of our submission

  Run ID                      Language       QE       Modality     Method         MAP

  CUHK-ad-eng-t-kl-ab1         english     without      text       KL-LM-ABS      0.3887
  CUHK-ad-eng-t-kl-ab2         english      with        text       KL-LM-ABS      0.4055
  CUHK-ad-eng-t-kl-ab3         english      with        text       KL-LM-ABS      0.4082
  CUHK-ad-eng-t-kl-jm1         english     without      text       KL-LM-JM       0.3844
  CUHK-ad-eng-t-kl-jm2         english      with        text       KL-LM-JM       0.4115
  CUHK-ad-eng-t-kl-di1         english     without      text       KL-LM-DIR      0.382
  CUHK-ad-eng-t-kl-di2         english      with        text       KL-LM-DIR      0.3999
  CUHK-ad-eng-t-tf-idf1        english     without      text       TF-IDF         0.351
  CUHK-ad-eng-t-tf-idf2        english      with        text       TF-IDF         0.3574
  CUHK-ad-eng-tn-kl-ab1        english     without      text       KL-LM-ABS      0.3877
  CUHK-ad-eng-tn-kl-ab2        english      with        text       KL-LM-ABS      0.3838
  CUHK-ad-eng-tn-kl-ab3        english      with        text       KL-LM-ABS      0.4083
  CUHK-ad-eng-tn-kl-jm1        english     without      text       KL-LM-JM       0.3762
  CUHK-ad-eng-tn-kl-jm2        english      with        text       KL-LM-JM       0.4018
  CUHK-ad-eng-tn-kl-di1        english     without      text       KL-LM-DIR      0.3921
  CUHK-ad-eng-tn-kl-di2        english      with        text       KL-LM-DIR      0.399
  CUHK-ad-eng-tn-tf-idf1       english     without      text       TF-IDF         0.3475
  CUHK-ad-eng-tn-tf-idf2       english      with        text       TF-IDF         0.366
  CUHK-ad-eng-v                english     without       vis       Moment-DCT     0.0599
  CUHK-ad-eng-tv-kl-ab1        english     without    text+vis     KL-LM-ABS      0.3941
  CUHK-ad-eng-tv-kl-ab3        english      with      text+vis     KL-LM-ABS      0.4108
  CUHK-ad-eng-tv-kl-jm1        english     without    text+vis     KL-LM-JM       0.3878
  CUHK-ad-eng-tv-kl-jm2        english      with      text+vis     KL-LM-JM       0.4135
  CUHK-ad-eng-tnv-kl-ab2       english      with      text+vis     KL-LM-ABS      0.3864
  CUHK-ad-eng-tnv-kl-ab3       english      with      text+vis     KL-LM-ABS      0.4118
  CUHK-ad-eng-tnv-kl-jm1       english     without    text+vis     KL-LM-JM       0.3787
  CUHK-ad-eng-tnv-kl-jm2       english      with      text+vis     KL-LM-JM       0.4041
  CUHK-ad-chn-t-kl-ab1         chinese     without      text       KL-LM-ABS      0.1815
  CUHK-ad-chn-t-kl-ab2         chinese      with        text       KL-LM-ABS      0.1842
  CUHK-ad-chn-t-kl-jm1         chinese     without      text       KL-LM-JM       0.1821
  CUHK-ad-chn-t-kl-jm2         chinese      with        text       KL-LM-JM       0.2027
  CUHK-ad-chn-tn-kl-ab1        chinese     without      text       KL-LM-ABS      0.1758
  CUHK-ad-chn-tn-kl-ab2        chinese      with        text       KL-LM-ABS      0.1527
  CUHK-ad-chn-tn-kl-ab3        chinese      with        text       KL-LM-ABS      0.1834
  CUHK-ad-chn-tn-kl-jm1        chinese     without      text       KL-LM-JM       0.1843
  CUHK-ad-chn-tn-kl-jm2        chinese      with        text       KL-LM-JM       0.2024

LM denotes Language Model, KL denotes Kullback-Leibler divergence based, DIR denotes the
smoothing using the Dirichlet priors, ABS denotes the smoothing using Absolute discounting,
                         JM denotes the Jelinek-Mercer smoothing.
3.3.1   Empirical Analysis of Language Models
Figure 2 and Figure 3 plot the curves of Precision vs. Recall and the curves of Precision vs.
Number of Returned Documents respectively. From the experimental results in Figure 2 and
Figure 3 as well as Table 2, one can observe that the KL-divergence language model outperforms
the simple TF-IDF retrieval model importantly (around 5%). In evaluation of the smoothing
techniques, we observe that the Jelinek-Mercer smoothing and Absolute Discounting Smoothing
yield better results than the Dirichlet prior (DIR).

                                                                                         CUHK−ad−eng−t−kl−ab3
                                                                                         CUHK−ad−eng−t−kl−jm2
                                     0.55                                                CUHK−ad−eng−t−kl−di2
                                                                                         CUHK−ad−eng−t−tf−idf2
                                                                                         CUHK−ad−eng−tv−kl−ab3
                                      0.5                                                CUHK−ad−eng−tv−kl−jm2


                                     0.45


                                      0.4
                         Precision




                                     0.35


                                      0.3


                                     0.25


                                      0.2


                                     0.15


                                      0.1
                                            5   50   100   150   200   250   300   350        400     450        500
                                                                 # of documents


Figure 3: Experimental Result of Precision vs. Number of Returned Documents with Selected
Configuration


3.3.2   Cross-Language Retrieval: Chinese-To-English Query Translation
To deal with the Chinese queries for retrieving English documents, we first adopt a Chinese
segmentation tool from the Linguistic Data Consortium (LDC) [1], i.e., the “LDC Chinese seg-
menter” 1 , to extract the Chinese words from the given query sentences. The segmentation step
is important toward effective query translation. Figure 4 shows the Chinese segmentation results
of part queries. We can see that the results can still be improved.
    For the bilingual query translation, the second step is to translate the extracted Chinese words
into English words using a Chinese-English dictionary. In our experiment, we employ the LDC
Chinese-to-English Wordlist [1] for the translations. The final translated queries are obtained by
combining the translation results.
    From the experimental results shown in Table 2, we can observe that the mean average precision
of Chinese-To-English Queries is about the half of the monolingual queries. There are a lot of
ways to improve the performance. One is to improve the Chinese segmentation algorithm. Some
post-processing tricks may be effective for improving the performance. Moreover, the translation
results can be further refined. One can tune better results by adopting some Natural Language
Processing techniques [6].

3.3.3   Cross-Media Retrieval: Re-Ranking Scheme with Text and Visual Content
In this participation, we study the combination of text and visual content for cross-media image
retrieval. In our development, we suggest the re-ranking scheme in combination with text and
visual content. For a given query, we first rank the images by using the language modeling
techniques. On the top ranking images, we then re-rank the images by measuring the visual
similarity to the query.
  1 It can be downloaded from: http://www.ldc.upenn.edu/Projects/Chinese/seg.zip .
   1. 地面上的飞机                                            地面 上 的 飞机
      Aircraft on the ground
   2. 演奏台旁聚集的群众                                         演奏 台 旁 聚集 的 群众
      People gathered at bandstand
   3. 狗的坐姿                                              狗 的 坐 姿
      Dog in sitting position
   4. 靠码头的蒸汽船                                           靠 码头 的 蒸汽 船
      Steam ship docked
   5. 动物雕像                                              动物 雕像
      Animal statue
   6. 小帆船                                               小 帆船
      Small sailing boat
   7. 在船上的渔夫们                                           在 船上 的 渔夫 们
      Small sailing boat
   8. 被雪覆盖的建筑物                                          被 雪 覆盖 的 建筑物
      Fishermen in boat
   9. 马拉动运货车或四轮车的图片                                     马拉 动 运 货车 或 四 轮 车 的 图片
      Horse pulling cart or carriage
   10. 苏格兰的太阳                                           苏格兰 的 太阳
      Sun pictures, Scotland

            Figure 4: Chinese segmentation results of part Chinese (Simplified) queries



    In our experiment, two kinds of visual features are used: texture and color features. For
the texture feature, the discrete cosine transform (DCT) is engaged to calculate coefficients that
multiply the basis functions of the DCT. Applying the DCT to an image yields a set of coefficients
to represent the texture of the image. In our implementation, a block-DCT (block size 8x8) is
applied on the normalized input images which generate a 256-dimensional DCT feature. For the
color feature, 9-dimensional color moment is extracted for each image. In total, each image is
represented by a 265-dimensional feature vector.
    As shown in Table 2, the MAP of query results using only the visual information is about 6%,
which is much lower than the text information with over 40%. From the experimental results, we
can observe the re-ranking scheme only produce a marginal improvement compared with the text
only approaches. Some reasons can be explained for the results. One is the engaged visual features
not effective enough to discriminate the images. Another possible reason is that the ground truth
images in the given query may not be quite different in visual content. It is interesting to study
more effective features and learning methods for improving the performance.

3.3.4    Query Expansion for Information Retrieval
From the experimental results in Table 2, we observe that all the queries are greatly enhanced
by adopting Query Expansion 2 (QE). The average improvement for all the queries is around
1.71% which accounts %4.12 of the maximum MAP of 41.35%. It is interesting to find that the
QE especially benefits a lot for the Jelinek-Mercer smoothing method, the mean gain with QE is
about 2.49% which accounts %6.02 of the maximum MAP of 41.35%.
   2 Query expansion refers to adding further terms to a text query (e.g. through PRF or thesaurus) or images to

a visual query
4    Conclusions
In this paper, we reported our empirical studies of cross-language and cross-media image retrieval
at the ImaegCLEF 2005 campaign. We addressed three major focuses and contributions in our
participation. The first is the empirical evaluations of Language Models and the smoothing strate-
gies for Cross-Language image retrieval. The second one is the evaluation of Cross-Media image
retrieval, i.e., combining text and visual content for image retrieval. The last one is the evaluation
of the Bilingual image retrieval between English and Chinese. We conducted empirical analysis
on the experimental results and provided the empirical summary of our participation.


References
 [1] Http://www.ldc.upenn.edu/projects/chinese/.
 [2] Http://www.lemurproject.org/.

 [3] Ricardo Baeza-Yates and Berthier Ribeiro-Neto. Modern Information Retrieval. Addison
     Wesley, 1999.
 [4] P. Clough, H. Mueller, and M. Sanderson. The clef cross language image retrieval track
     (imageclef) 2004. In In the Fifth Workshop of the Cross-Language Evaluation Forum (CLEF
     2004) (LNCS). Springer, 2004.
 [5] T. M. Cover and J. A. Thomas. Elements of Information Theory. Wiley, 1991.
 [6] C. Manning and H. Schütze. Foundations of Statistical Natural Language Processing. The
     MIT Press, 1999.
 [7] J. Savoy. Report on clef-2001 experiments (cross language evaluation forum). In LNCS 2406,
     pages 27–43, 2002.
 [8] C. E. Shannon. Prediction and entropy of printed english. Bell Sys. Tech. Jour., 30:51–64,
     1951.
 [9] A. W. M. Smeulders, M. Worring, S. Santini, A. Gupta, and R. Jain. Content-based image
     retrieval at the end of the early years. IEEE Transactions on Pattern Analysis and Machine
     Intelligence, 22(12):1349–1380, 2000.
[10] Chengxiang Zhai and John Lafferty. Model-based feedback in the kl-divergence retrieval
     model. In In Tenth International Conference on Information and Knowledge Management
     (CIKM2001), pages 403–410, 2001.

[11] Chengxiang Zhai and John Lafferty. A study of smoothing methods for language models
     applied to ad hoc information retrieval. In ACM SIGIR Conference on Research and Devel-
     opment in Information Retrieval (SIGIR’01), pages 334–342, 2001.