=Paper= {{Paper |id=Vol-1179/CLEF2013wn-ImageCLEF-BottcherEt2013 |storemode=property |title=BTU DBIS' Personal Photo Retrieval Runs at ImageCLEF 2013 |pdfUrl=https://ceur-ws.org/Vol-1179/CLEF2013wn-ImageCLEF-BottcherEt2013.pdf |volume=Vol-1179 |dblpUrl=https://dblp.org/rec/conf/clef/BottcherZS13 }} ==BTU DBIS' Personal Photo Retrieval Runs at ImageCLEF 2013== https://ceur-ws.org/Vol-1179/CLEF2013wn-ImageCLEF-BottcherEt2013.pdf
    BTU DBIS’ Personal Photo Retrieval Runs at
               ImageCLEF 2013

              Thomas Böttcher, David Zellhöfer, and Ingo Schmitt

    Brandenburg Technical University, Database and Information Systems Group,
                      Walther-Pauer-Str. 2, 03046 Cottbus
         tboettcher@tu-cottbus.de, david.zellhoefer@tu-cottbus.de,
                            schmitt@tu-cottbus.de



       Abstract. This paper summarizes the results of the BTU DBIS research
       group’s participation in the Personal Photo Retrieval subtask of Image-
       CLEF 2013. In order to solve the subtask, a self-developed multimodal
       multimedia retrieval system, PythiaSearch, is used. The discussed re-
       trieval approaches focus on two different strategies. First, two automatic
       approaches that combine visual features and meta data are examined.
       Second, a manually assisted relevance feedback approach is presented.
       All approaches are based on a special query language, CQQL, which
       supports the logical combination of different features.
       Considering only automatic runs without relevance feedback that have
       been submitted to the subtask, DBIS reached the best overall results,
       while the relevance feedback-assisted approach is placed second amongst
       all participants of the subtask.

       Keywords: Content-Based Image Retrieval, Preference Based Learn-
       ing, Relevance Feedback, Polyrepresentation, Experiments


1    Introduction

This paper summarizes the results of the BTU DBIS research group’s participa-
tion in the Personal Photo Retrieval subtask of ImageCLEF 2013 [3].
    As in DBIS’ participations in various ImageCLEF tasks between 2011 and
2012, the discussed approaches rely on the commuting quantum query language
(CQQL) [16]. CQQL is capable of combining similarity predicates as found in
information retrieval (IR) as well as relational predicates common in databases
(DB) and has been on of the main research fields of the database and information
systems work group at the Brandenburg Technical University (BTU).
    CQQL is an extension of the relational domain calculus, i.e., it can be di-
rectly executed within a relational DB system [11]. To combine both data access
paradigms, CQQL relies on the mathematical foundations of quantum mechan-
ics and logic. For the sake of brevity, the theoretical background of the query
language is omitted. For further details, please refer to the central CQQL pub-
lication [16]. Additional information, e.g., the relation of CQQL to fuzzy logic
can be found in [17]. Its relation to probabilistic IR models is discussed in [26].
2

   In the scope of this paper, CQQL is used for the matching within the used
multi-modal multimedia retrieval system PythiaSearch [24, 23], which has been
developed by DBIS. The system consists of an extraction module for both vi-
sual features and meta data that supports various image formats and PDF,
a matching component relying on CQQL, and a full-featured GUI supporting
graded relevance feedback. In order to carry out the matching between query
documents and a document collection, CQQL combines various features with
the help of logical connectors.

1.1   Personal Photo Retrieval Subtask
The personal photo retrieval subtask 2013 is an extension of 2012’s pilot task.
The current subtask uses 5,555 image documents that have been sampled from
personal photo collections. In contrast to the pilot phase of the task, 2013’s
focus lies on the evaluation of retrieval algorithms using different search strate-
gies and user groups. One objective of the task is to assess whether a retrieval
algorithm’s effectiveness is stable for different user groups [28, 10]. To test the
effectiveness for different users, multiple ground truths are provided reflecting
relevance assessments of CBIR/MIR experts, laypersons or the like.
    The subtask does not provide any training data. Hence, it has to be solved
ad-hoc. The participants are given multiple query-by-example (QBE) documents
and/or browsed documents and are asked to find the best matching documents
illustrating an event or depicting a visual concept. In total, 74 topics are avail-
able. In contrast to the last year, the topics are no longer separated into visual
concepts or events [10]. Furthermore, the information need (IN) for each topic is
not explicitly given. Instead, the IN is concealed inside the query or/and browsed
documents. To infer an IN, the participants get 0-1 QBE document and up to 3
browsed documents. For some topics, there are no QBE documents in order to
model the following usage behavior: a user browsed a personal photo collection
and toggled an action to show more similar images without stating an explicit
preference [10]. The provided browsed documents can sometimes be irrelevant or
have only a low degree of relevance. A more detailed description of the subtask’s
experimental setup, objective, and participation is available in [28].


2     PythiaSearch - an Interactive and Multi-modal
      Multimedia Retrieval System
The interactive retrieval system PythiaSearch [24, 23] forms the core for both the
interactive and non-interactive retrieval experiments that are described in this
paper. In order to express their IN, users can input images (following the QBE
paradigm that is used in the subtask), (multilingual) texts, or PDF documents.
Additionally, it supports a relevance feedback (RF) process that can be used to
personalize the query results based on the user’s interaction with the system.
The interactive parts rely on a common code base for feature extraction and
similarity calculation with the baseline system [28] that has been provided by
                                                                                  3

the organizer of the subtask. Figure 1 shows the GUI of the system. A full
description of the GUI and its conceptual model has been published before [23].




                  Fig. 1. PythiaSearch - graphical user interface




2.1   Evaluation of CQQL
As said before, we will neglect the theoretical foundations of CQQL to facilitate
the understanding of this paper. Mathematically interested readers are recom-
mended to refer to [16] and [13]. The arithmetic evaluation of a CQQL statement
which consists of multiple conditions that are connected by logical connectors is
directly derived from the mathematical framework of quantum mechanics and
logic. In this section, we will sketch the arithmetic evaluation of CQQL as far as
it is necessary for the understanding of this paper.
     Let fϕ (d) be the evaluation of a document d w.r.t. a CQQL query. To con-
struct a CQQL query, various conditions ϕ can be linked in an arbitrary manner
using the conjunction (1), disjunction (2), or negation (3). If ϕ is atomic, fϕ (d)
can be directly evaluated yielding a value out of the interval [0; 1]. For the scope
of this paper, an atomic condition is the result of a similarity measure, e.g., the
similarity of the QBE document’s color histogram and d’s color histogram; or a
Boolean evaluation calculated by a DB system or the like.
After a necessary syntactical normalization step [27], the evaluation of a CQQL
query is performed by recursively applying the succeeding formulas until the
atomic base case is reached:

                           fϕ1 ∧ϕ2 (d) = fϕ1 (d) ∗ fϕ2 (d)                      (1)

                fϕ1 ∨ϕ2 (d) = fϕ1 (d) + fϕ2 (d) − (fϕ1 (d) ∧ fϕ2 (d))           (2)
                               f¬ϕ (d) = 1 − fϕ (d)                             (3)
4

    An example of the arithmetic evaluation of the query that is used in this
paper is given in Section 3. In accordance with the Copenhagen interpretation
of quantum mechanics, the result of an evaluation of a document d yields the
probability of relevance of d w.r.t. the query. This probability value is then used
for the ranking of the result list of documents.


Weighting in CQQL In order to reflect the need for the personalization of a
query and as a necessary step for the support of relevance feedback (RF), CQQL
has been extended with a weighting scheme [15]. The weights in CQQL can be
used to steer the impact of a condition on the overall evaluation result. Weighting
is a crucial part of the machine-based learning supported RF mechanism that is
used in Section 3 and discussed in more detail in [27] and [23].
    The weighting in CQQL is fully embedded into the logical query. That is,
a query maintains its logical properties while weights are used. To illustrate,
Equation 4 denotes a weighted conjunction, whereas Equation 5 states a weighted
disjunction. A weight θi is directly associated with a logical connector and steers
the influence of a condition ϕi on the evaluation. To evaluate a weighted CQQL
query, the weights are syntactically replaced by constant values according to the
following rules:
                      ϕ1 ∧θ1 ,θ2 ϕ2   (ϕ1 ∨ ¬θ1 ) ∧ (ϕ2 ∨ ¬θ2 )                 (4)
                      ϕ1 ∨θ1 ,θ2 ϕ2   (ϕ1 ∧ θ1 ) ∨ (ϕ2 ∧ θ2 )                  (5)


2.2   Result Personalization and Relevance Feedback

As implied before, the relevance judgement of a query’s results is very subjective
with respect to the user’s IN. To refine a subjective IN, PythiaSearch supports
a gradual relevance feedback on the basis of partially ordered sets (posets) [27].
Users can input a poset of documents which contains an arbitrary amount of doc-
uments at various relevance levels. For instance, a poset can define a preference
expressing that a document Di is better than a document Dj . This form of user
input requires no background information of the underlying features and is based
on the subjective qualitative perception of the user alone. Figure 2 illustrates
the mechanism as it is implemented in PythiaSearch’s GUI. In this example,
the second ring contains documents considered more relevant than those on the
third etc., while the center contains the current QBE document.
    Internally, a machine-based learning algorithm (a downhill simplex variant)
is used to find appropriate weight values for a given CQQL query fulfilling the
input preferences. The actual algorithm and its properties is described separately
in [27].


3     Experimental Setup and Results

Motivated by CQQL’s support for formulating multi-modal queries, DBIS par-
ticipated in the 2011 Wikipedia Retrieval task at ImageCLEF [25] combining
                                                                                      5

textual and visual features. This year’s participation in the Personal Photo Re-
trieval subtask focuses on a CQQL-based combination of visual features and
the accompanying meta data. This poses a new challenge for the working group
because the studies carried out before were not relying very much on meta data.
     Our experiments for the Personal Photo Retrieval subtask can be subdivided
into two types of runs. First, fully automatic runs which demonstrate the effec-
tiveness of a CQQL-based logical combination of features with different origin,
i.e., visual and meta data features. Second, the performance of the aforemen-
tioned RF mechanism is investigated (see Section 3.4).


3.1     Used Features

Over the last years, the DBIS working group conducted a lot of experiments on
various image collections, ranging from the Caltech collections [7, 8] to MSRA-
MM [20] in order to assess the retrieval effectiveness of different low-level visual
features. This investigation of single features forms the basis for the decision
which features to combine with CQQL.
    PythiaSearch supports the extraction of low-level global and local visual fea-
tures, e.g., color, edges and texture features or local features like SIFT and
SURF. In total, the extraction component offers more than 30 visual features.
Additionally, the system allows the extraction of common image meta data such
as Exif, IPTC, or XMP. This meta data, e.g., GPS coordinates, the camera
model, or the image orientation extends the variety of features that can be used
for the matching of documents. In accordance with the rules of the subtask,
IPTC-based data is ignored in the following experiments. Table 1 lists all fea-
tures that are used in the experiments.
    The feature extraction and similarity calculation functionality used by Pythia-
Search resembles the baseline system that is provided by the subtask organizer1 .
For a description, see [28]. The main differences between the baseline system
and the system used for the described experiments are the CQQL support, the
supplementary RF mechanism, and the GUI.


3.2     Examined CQQL Query

Based on preparatory study on the retrieval effectiveness of various visual low-
level features and an examination of the subtask’s thematic orientation on both
visual concepts and events, an appropriate CQQL query had to be defined. The
core idea of the examined CQQL query, which is shown in Equation 6, is to use
low-level features that did show a good performance over all six test collections
1
    Please note that the organizer of the subtask did not actively participate in the
    experiments described in this paper nor did he release additional information to the
    working group that other participants could not obtain. Alas, he carried out many
    of the pre-studies including the investigation of generally effective CQQL queries.
    Furthermore, he had a major impact on the development of PythiaSearch and the
    underlying learning algorithm.
6

         Table 1. Features and type, * denotes features in the scope of MPEG-7 [12]

                      Name                           Type              Origin
          Auto Color Correlogram (ACC)       color-related, global                [9]
                       BIC                   color-related, global                [18]
                      CEDD               texture/color-related, global            [4]
          Color Histogram (region based) color-related, pseudo-local              [1]
                 Color Histogram                     global            256 bin RGB histogram
                                                                       (own implementation)
                  Color Layout*              color-related, global                [6]
                 Color Structure*            color-related, global                [6]
                Dominant Color*              color-related, global                [6]
                Edge Histogram*              edge-related, global                 [6]
                      FCTH               texture/color-related, global            [5]
                 Scalable Color*             color-related, global                [6]
                     Tamura                 texture-related, global               [19]
              Region-based Shape*                    global                       [6]
                Person Detection                facial features        Own     implementation
                                                                       (based on OpenCV)
                 Time of creation                  temporal                      Exif
                 GPS coordinate                     spatial                      Exif
                  Camera model                     metadata                      Exif


    Table 2. Overview over the examined test collections in the preparatory study

                          Collection     Collection Size Collection Type
                         Caltech 101 [7]           9,197 Object recognition
                         Caltech 256 [8]          30,607 Object recognition
                         MSRA-MM [20]             65,443 Web Image Sample
                           UCID [14]                 904 Personal photos
                           Wang [21]               1,000 Stock photography
                          Pythia [22]              5,555 Personal photos




that are listed in Table 2 to cope with the visual content of a document. In order
to retrieve similar events, we assume that the presence (or absence) of persons in
a picture, spatial and temporal proximity as well as a similar camera model are
valid indicators. Hence, the core CQQL query is enriched by a person presence
condition in form of a Boolean predicate and the aforementioned features derived
from Exif meta data.
    This concept results in the following CQQL query that uses a weighted con-
junction of 18 conditions, whereas all weights are set to 1 initially to express the
equal importance of all conditions.
  ^
     (ACCsim , BICsim , CEDDsim , ColorHistBordersim , ColorHistCentersim ,
    θi

         ColorHistsim ColorLayoutsim , ColorStructuresim , DominantColorsim ,                   (6)
         EdgeHistsim , FCTHsim , Regionshapesim , ScalableColorsim , Tamurasim ,
         GPSsim , modelsim , timesim , Personsim )

    The value of each condition is determined by a distance measure such as the
Euclidean distance of the corresponding feature between the QBE document and
the retrieved document which is then transformed into a similarity measure in the
interval of [0; 1]. Boolean conditions are evaluated traditionally. The calculation
                                                                                 7

of the GPS coordinate similarity is carried out as we did for the ImageCLEF
2012 Plant Identification task [2]:
                      p
                         (71.5 · (longx − longy ))2 + (111.3 · (latx − laty ))2
      GP S sim = 1 −                                                            (7)
                                             6378.388
whereas long stands for longitude and lat for latitude.
   Following the transformation rules that have been described in Section 2.1,
the arithmetic evaluation of the presented CQQL is as follows:

                         (ACCsim + ¬θ1 − ACCsim ∗ ¬θ1 )∗                       (8)
                           (BICsim + ¬θ2 − BICsim ∗ ¬θ2 )∗                     (9)
                     (CEDDsim + ¬θ3 − CEDDsim ∗ ¬θ3 )∗                        (10)
                                                          ...                 (11)




3.3   Automatic Runs
The 2013 Personal Photo Retrieval subtask provides QBE documents as well as
browsed documents. For the approaches without RF, we use the provided data
in two ways.
    First, we use only the QBE defined documents (run1 ) because of the fact,
that the provided browsed documents can contain misleading information, re-
spectively contain images that did not fulfill the users IN. For the topics for
which no QBE document is available, we use all browsed documents instead.
Whether the browsed documents are really relevant for the user’s IN cannot be
determined automatically. Thus, this approach might be affected by irrelevant
input to the retrieval system.
    Second, we assume that all documents (no matter if they are QBE or browsed
documents) are equally meaningful regarding the user’s IN. Hence, we use all
documents as QBE documents with no special ranking for the labelled QBE
document. This approach is labeled run2 in Figure 4.

3.4   Manual Relevance Feedback Run
The main objective for the manually assisted approach (run 3 ) is to remove
misleading browsed image documents from the initial query and to improve the
retrieval quality using the graded RF approach that is described in Section 2.2.
The experiment is carried out interactively with the PythiaSearch GUI (see Sec-
tion 2). Using the aforementioned preference-based approach, irrelevant, relevant
documents and the relationship between them can be expressed as a poset. To
simplify the user interaction, the GUI offers 3 levels of relevance and a “garbage
can” to collect completely irrelevant documents. Figure 2 shows the three levels
where the center (level 1) contains the query document(s). All documents in
level 2 are more relevant than documents in level 3 and 4, whereas documents
8

in level 3 are more relevant than documents in level 4. All preferences together
create a poset. Documents marked as completely irrelevant are removed from
the query results and have a negative impact on the machine-based learning
algorithm similar to negative QBE documents.




              Fig. 2. Graded Relevance Feedback - Preference Levels



    In order to keep control over the time consumption of submitting 74 queries
manually, we have defined some restrictions for the RF-based experiment.
    First, at most one RF iteration is carried out, i.e., we model the behavior of
an impatient user.
    Second, the assessment of the quality of the results is based on the top-30
results only. This number of documents can be easily inspected without scrolling
and requires significantly less time than inspecting the top-50 or top-100. Because
of this strategy, it may happen that no RF is carried out all because the top-30
results seem relevant to the interacting user.
    Third, to simulate a user that avoids a large amount of interaction with the
system, a total of 6 images is used to define the preferences used during RF.
    In general, obviously irrelevant images from the given IN specification were
removed from the input. Nevertheless, during the submission of all 74 sample
queries it was not always possible to identify the IN without background knowl-
edge. In these cases, the RF process is skipped.


4   Results

With reference to the official results (see Figure 4), our best run, i.e., run 3,
achieves rank 5 in the overall ranking for the average user. Compared to the
results of all participants of the subtask, DBIS is ranked second. Focussing on
the NCDCG cut 5 we reach about 97%, on NDCG cut 100 about 87% and on
MAP cut 100 78% of the best obtained retrieval score.
                                                                                9

    When only automatic runs are considered, i.e., runs without RF, DBIS
achieves the best results. Unfortunately no other runs without RF that use all
available modalities (visual and meta data features) and IN information (QBE
and browsed documents) were submitted. Due to these circumstances, a resilient
interpretation of our results is hardly possible. Anyhow, we assume that the in-
clusion of meta data helps to distance our from the other approaches. In any
case, further information about the techniques used by the other participants is
needed.
    Generally speaking, the outcome of the presented experiments is fully satis-
fying. Though, we acknowledge room for improvements for the RF-based run.
We assume that an inspection of more than the top-30 results and the inclusion
of more preferences might have an impact on the retrieval effectiveness. Further-
more, a specification of the actual IN in textual form will help human assessors
during RF because it will enable them to provide RF for every topic. As said
before, we could not provide RF for all topics because of the lack of this kind of
information. In consequence, we expect an improvement of the RF effectiveness
when this information can be used.
    One objective of this subtask is to examine the robustness of a retrieval ap-
proach with respect to different user groups (e.g. IT experts, non-IT users or
gender-specific groups). Figure 3 shows the variance of the MAP cut 100 scores
of our three submitted runs between the different user groups. The differences
between all groups is relatively small, e.g., 0.395 vs. 0.425 for run 3. The dif-
ference in MAP cut 100 between the best and the worst run is about 7-9%.
Interestingly, the results for female users tend to be the best, whereas the aver-
age user score tends to be worst. This effect can also be observed in the results
of the other groups. Alas, we are not sure why this effect is present amongst all
groups.




           Fig. 3. MAP cut 100 score comparison on different user types
10




           Fig. 4. Overall results considering MAP cut 100 (average user)




     Fig. 5. Results excluding RF runs considering MAP cut 100 (average user)


5    Conclusions and Future Work
The results of our participation in the ImageCLEF 2013 Personal Photo Retrieval
subtask are motivating. Although, DBIS achieved a good effectiveness rank, there
are areas that need further research.
    First, we plan to analyze single features and meta data in more detail to find
out which features or meta data contributes most to the retrieval quality. As said
before, for the RF-supported approach there are various optimizations possible.
In particular, the restriction to one RF iteration seems to limit the retrieval
quality. First informal experiments show that up to three iterations can give a
great performance boost. Furthermore, we assume that the inclusion of RF on
all topics will lead to a performance improvement. Another interesting research
question is the development of the weight values during the RF iterations in order
to reveal whether some features do not contribute to the retrieval effectiveness
at all.


References
 1. Balko, S., Schmitt, I.: Signature Indexing and Self-Refinement in Metric Spaces.
    Cottbus (2012)
 2. Böttcher, T., Schmidt, C., Zellhöfer, D., Schmitt, I.: Btu dbis’ plant identification
    runs at imageclef 2012. In: CLEF (Online Working Notes/Labs/Workshop) (2012)
                                                                                       11

 3. Caputo, B., Mueller, H., Thomee, B., Villegas, M., Paredes, R., Zellhoefer, D.,
    Goeau, H., Joly, A., Bonnet, P., Martinez Gomez, J., Garcia Varea, I., Cazorla,
    M.: ImageCLEF 2013: the Vision, the Data and the Open Challenges, CLEF 2013
    Working Notes, Valencia, Spain (2013)
 4. Chatzichristofis, A.S., Boutalis, S.Y.: CEDD: color and edge directivity descrip-
    tor: a compact descriptor for image indexing and retrieval. In: Proceedings of the
    6th international conference on Computer vision systems. pp. 312–322. ICVS’08,
    Springer-Verlag (2008), http://dl.acm.org/citation.cfm?id=1788524.1788559
 5. Chatzichristofis, A.S., Boutalis, S.Y.: FCTH: Fuzzy Color and Texture Histogram
    - A Low Level Feature for Accurate Image Retrieval. In: Proceedings of the
    2008 Ninth International Workshop on Image Analysis for Multimedia Interac-
    tive Services. pp. 191–196. WIAMIS ’08, IEEE Computer Society (2008), http:
    //dx.doi.org/10.1109/WIAMIS.2008.24
 6. Cieplinski, L., Jeannin, S., Ohm, J.R., Kim, M., Pickering, M., Yamada, A.:
    MPEG-7 Visual XM version 8.1. Pisa, Italy (2001)
 7. Fei-Fei, L., Fergus, R., Perona, P.: Learning generative visual models from few
    training examples an incremental Bayesian approach tested on 101 object cate-
    gories. In: Proceedings of the Workshop on Generative-Model Based Vision (2004)
 8. Griffin, G., Holub, A., Perona, P.: Caltech-256 Object Category Dataset (2007),
    http://authors.library.caltech.edu/7694
 9. Huang, J., Kumar, R.S., Mitra, M., Zhu, W.J., Zabih, R.: Image Indexing Using
    Color Correlograms. In: Proceedings of the 1997 Conference on Computer Vision
    and Pattern Recognition (CVPR ’97). pp. 762–. CVPR ’97, IEEE Computer So-
    ciety (1997), http://dl.acm.org/citation.cfm?id=794189.794514
10. ImageClef: Personal Photo Retrieval 2013. http://www.imageclef.org/2013/
    photo/retrieval, 6. June 2013.
11. Lehrack, S., Schmitt, I.: QSQL: Incorporating Logic-Based Retrieval Conditions
    into SQL. In: Kitagawa, H., Ishikawa, Y., Li, Q., Watanabe, C. (eds.) Database
    Systems for Advanced Applications, 15th International Conference, DASFAA 2010,
    Tsukuba, Japan, April 1-4, 2010, Proceedings, Part I, Lecture Notes in Computer
    Science, vol. 5981, pp. 429–443. Springer (2010)
12. Manjunath, B., Salembier, P., Sikora, T.: Introduction to MPEG-7: Multimedia
    Content Description Interface. John Wiley & Sons, Inc., New York, NY, USA
    (2002)
13. van Rijsbergen, C.: The Geometry of Information Retrieval. Cambridge University
    Press, Cambridge, England (2004)
14. Schaefer, G., Stich, M.: UCID - An Uncompressed Colour Image Database. In:
    Proc. SPIE, Storage and Retrieval Methods and Applications for Multimedia, pp.
    472–480. San Jose, USA (2004)
15. Schmitt, I.: Weighting in CQQL. Cottbus (2007)
16. Schmitt, I.: QQL: A DB&IR Query Language. The VLDB Journal 17(1), 39–56
    (2008)
17. Schmitt, I., Zellhöfer, D., Nürnberger, A.: Towards quantum logic based multi-
    media retrieval. In: IEEE (ed.) Proceedings of the Fuzzy Information Processing
    Society (NAFIPS). pp. 1–6. IEEE (2008), 10.1109/NAFIPS.2008.4531329
18. Stehling, O.R., Nascimento, A.M., Falcão, X.A.: A compact and efficient image re-
    trieval approach based on border/interior pixel classification. In: Proceedings of the
    eleventh international conference on Information and knowledge management. pp.
    102–109. CIKM ’02, ACM (2002), http://doi.acm.org/10.1145/584792.584812
12

19. Tamura, H., Mori, S., Yamawaki, T.: Texture features corresponding to visual
    perception. IEEE Transactions on System, Man and Cybernatic 8(6), 460–472
    (1978)
20. Wang, M., Yang, L., Hua, X.S.: MSRA-MM: Bridging Research and Industrial
    Societies for Multimedia Information Retrieval (2009)
21. Wang, Z.J., Li, J., Wiederhold, G.: SIMPLIcity: Semantics-sensitive Integrated
    Matching for Picture Libraries. In: Proceedings of the 4th International Conference
    on Advances in Visual Information Systems. pp. 360–371. VISUAL ’00, Springer-
    Verlag (2000), http://portal.acm.org/citation.cfm?id=647061.714442
22. Zellhöfer, D.: An Extensible Personal Photograph Collection for Graded Relevance
    Assessments and User Simulation. In: Proceedings of the ACM International Con-
    ference on Multimedia Retrieval. ICMR ’12, ACM (2012)
23. Zellhöfer, D.: A permeable expert search strategy approach to multimodal retrieval.
    In: Proceedings of the 4th Information Interaction in Context Symposium. pp. 62–
    71. IIIX ’12, ACM, New York, NY, USA (2012), http://doi.acm.org/10.1145/
    2362724.2362739
24. Zellhöfer, D., Bertram, M., Böttcher, T., Schmidt, C., Tillmann, C., Schmitt, I.:
    PythiaSearch – A Multiple Search Strategy-supportive Multimedia Retrieval Sys-
    tem. In: Proceedings of the 2nd ACM International Conference on Multimedia
    Retrieval. p. to appear. ICMR ’12, ACM (2012)
25. Zellhöfer, D., Böttcher, T.: BTU DBIS’ Multimodal Wikipedia Retrieval Runs at
    ImageCLEF 2011. In: Vivien Petras, Pamela Forner and Paul D. Clough (eds.)
    CLEF 2011 Labs and Workshop, Notebook Papers, 19-22 September 2011, Ams-
    terdam, The Netherlands (2011)
26. Zellhöfer, D., Frommholz, I., Schmitt, I., Lalmas, M., van Rijsbergen, K.: Towards
    Quantum-Based DB+IR Processing Based on the Principle of Polyrepresentation.
    In: Clough, P., Foley, C., Gurrin, C., Jones, G., Kraaij, W., Lee, H., Murdoch,
    V. (eds.) Advances in Information Retrieval - 33rd European Conference on IR
    Research, ECIR 2011, Dublin, Ireland, April 18-21, 2011. Proceedings, Lecture
    Notes in Computer Science, vol. 6611, pp. 729–732. Springer (2011)
27. Zellhöfer, D., Schmitt, I.: A Preference-based Approach for Interactive Weight
    Learning: Learning Weights within a Logic-Based Query Language. Distributed
    and Parallel Databases (2009), doi:10.1007/s10619-009-7049-4
28. Zellhöfer, D.: Overview of the ImageCLEF 2013 Personal Photo Retrieval Subtask.
    CLEF 2013 working notes, Valencia, Spain, 2013 (2013)