=Paper=
{{Paper
|id=Vol-1175/CLEF2009wn-ImageCLEF-GranadosEt2009
|storemode=property
|title=MIRACLE (FI) at ImageCLEFphoto 2009
|pdfUrl=https://ceur-ws.org/Vol-1175/CLEF2009wn-ImageCLEF-GranadosEt2009.pdf
|volume=Vol-1175
|dblpUrl=https://dblp.org/rec/conf/clef/GranadosBAGGGVDA09
}}
==MIRACLE (FI) at ImageCLEFphoto 2009==
MIRACLE (FI) at ImageCLEFphoto 2009 R. Granados1, X. Benavent2, R. Agerri1, A. García-Serrano3, J.M. Goñi1, J. Gomar2, E. de Ves2, J. Domingo2, G. Ayala2 1 Universidad Politécnica de Madrid, UPM 2 Universidad de Valencia 3 Universidad Nacional de Educación a Distancia, UNED {rgranados@fi.upm.es, xaro.benavent@uv.es, rodrigo.agerri@upm.es, agarcia@lsi.uned.es} Abstract The Miracle-FI participation at ImageCLEF 2009 photo retrieval task main goal was to improve the merge of content-based and text-based techniques in our experiments. The global system includes our own implemented tool IDRA (InDexing and Retrieving Automatically), and the Valencia University CBIR system. Analyzing both “topics_part1.txt” and “topics_part2.txt” task topics files, we have built different queries files, eliminating the negative sentences with the text from title and clusterTitle or clusterDescription, one query for each cluster (or not) of each topic from 1 to 25 and one for each of the three images of each topic from 26 to 50. In the CBIR system the number of low-level features has been increased from the 68 component used at ImageCLEF 2008 up to 114 components, and in this edition only the Mahalanobis distance has been used in our experiments. Three different merging algorithms were developed in order to fuse together different results lists from visual or textual modules, different textual indexations, or cluster level results into a unique topic level results list. For the five runs submitted we observe that MirFI1, MirFI2 and MifFI3 obtain quite higher precision values than the average ones. Experiment MirFI1, our best run for precision metrics (very similar to MirFI2 and MirFI3), appears in the 16th position in R-Precision classification and in the 19th in MAP one (from a total of 84 submitted experiments). MirFI4 and MirFI5 obtain our best diversity values, appearing in position 11th (over 84) in cluster recall classification, and being the 5th best group from all the 19 participating ones. Categories and subject descriptors H.3 [Information Storage and Retrieval]: H.3.1 Content Analysis and Indexing; H.3.2 Information Storage; H.3.3 Information Search and Retrieval; H.3.4 Systems and Software; H.3.7 Digital libraries. H.2 [Database Management]: H.2.5 Heterogeneous Databases; E.2 [Data Storage Representations]. Keywords Information Retrieval, Content-Based Image Retrieval, Merged result lists, Indexing, Named Entities Recognition 1 Introduction The Miracle-FI participation at ImageCLEF 2009 photo retrieval task [17] main goal was to improve the merge of content-based and text-based techniques in our experiments. The global system includes our own implemented tool IDRA (InDexing and Retrieving Automatically), and the Valencia University CBIR system. The BELGA Collection image captions were preprocessing to build a semi-structured XML description similar to the ImageCLEFphoto08 task [11]. Also the images have been preprocessed for CBIR module because some of them have some bands on the frame of the image with color pixels of the RGB and MCY system colors. As we realize that they don’t follow any established format, the solution adopted was to reduce all the images to the 90% of his real size in order to eliminate the different bands and the white pixels frames. Analyzing both “topics_part1.txt” and “topics_part2.txt” task topics files, we have built different queries files: [qf2] “BELGAtopics-tctcd-(q-cl)-fQ.txt”: one query for each cluster of each topic with the text from title, clusterTitle and clusterDescription (eliminating the negative sentences and negative clusters), [qf3] “BELGAtopics-tct-(q-cl)-fQ.txt”: the same as above but just with the text from title and clusterTitle, [qf4] “BELGAtopics-topEnt(1..25)+capEnt(26..50)-(q-cl)-fQ.txt”: one query for each cluster (except negatives ones) of each topic from 1 to 25 and one for each of the three images of each topic from 26 to 50, [qf5] “BELGAtopics-cap(title+desc)-(26..50)-(q-cl)-fQ.txt”: one query for each one of the three images of each topic from 26 to 50, obtained from the title and description fields of the XML captions. In the CBIR system the number of low-level features has been increased from the 68 component used at ImageCLEF 2008 up to 114 components, mainly due to the use of local color histogram descriptors that were not use last year. This edition only the Mahalanobis distance has been used in our experiments. Three different merging algorithms were developed in order to fuse together different results lists from visual or textual modules, different textual indexations, or cluster level results into a unique topic level results list: MAXmerge (the algorithm selects the results from the N lists which have a higher relevance value), EQUImerge (the algorithm selects the first result of each query (cluster), not selected yet), and ENRICH (this merging uses two results lists, a main list and a support list, and when a concrete result appears in both lists, the relevance will be increased). The five runs submitted were: [run1] “MirFI1_T-CT-I_TXT-IMG”: launching [qf3], reordering textual results list with CBIR system, and merging both lists with the ENRICH algorithm, [run2] “MirFI2_T-CT-CD- I_TXT-IMG”: the same as above, but launching [qf2], [run3] “MirFI3_T-CT-CD_TXT”: textual experiment launching [qf2], [run4] “MirFI4_T-CT-CD-I_TXT”: topics 1 to 25 as [run3], and topics 26 to 50, launching [qf5], [run5] “MirFI5_T-CT-CD-I_TXT”: textual experiments with NER. [qf4]. All the queries files in the submitted experiments were built from different queries of different clusters of a topic, so the EQUImerge algorithm was applied to fuse the different clusters-based result lists into an unique one. In the results, we observe that experiment MirFI1, our best run for precision metrics (very similar to MirFI2 and MirFI3), appears in the 16th position in R-Precision classification and in the 19th in MAP one (from a total of 84 submitted experiments). MirFI4 and MirFI5 obtain our best diversity values, appearing in position 11th (over 84) in cluster recall classification, and being the 5th best group from all the 19 participating ones. captions.txt (Part1, part2 BELGA captions TO XML XML fields Selection IDRA docs Index XML captions Text PreProcess Extractor QUERIES queries IDRA IDRA file Search Result TEXTUAL List queries files CONSTRUCTION BELGA topics.txt NE list MERGING NER tagger module Feature VISUAL BELGA Extraction Result Visual Search Images DB List Result List VISUAL Fig. 1. System overview 2 System Description The global system (shown at Fig. 1) includes our own implemented tool IDRA (InDexing and Retrieving Automatically), and the Valencia University CBIR system. The main goal, using IDRA [12] with such a large collection, was to analyze how the obtained results from the textual module could be improved using information from the content-based module. In this year, a global strategy for all experiments has been that the Content- Based module always starts working with a selected textual results list as part of his input data (different from our participation at ImageCLEF 2008 [11]). 2.1 Collection preprocessing ImageCLEFphoto09 task uses the so-called “BELGA Collection” which contains 498,920 images from Belga News Agency. Each photograph is accompanied by a caption composed of English text up to a few sentences in length [3]. Image captions are provided without a specific format. Because of this, we preprocess the captions file to build a semi-structured XML description for each image, similar to the used in the ImageCLEFphoto08 task [1]. This format includes 8 tags (docno, title, description, notes, location, date, image and thumbnail), which we try to fill preprocessing the captions texts. This preprocess consists on trying to identify in each caption the appropriate part of the text to fill the XML tags. We can see an example of this transformation in Fig. 2. caption 1012315|UK OUT NO MAGS NO SALES NO ARCHIVES NO INTERNET MAN02 - 20020802 - MANCHESTER, UNITED KINGDOM : England's Sarah Price emerges out of the water, during the Women's 50m backstroke heats at the Manchester Aquatic Centre, as part of the Commonwealth Games in Manchester, Friday 02 August 2002. EPA PHOTO PA-RUI VIEIRA XMLcaptionFig. 2. “caption TO XML” example The images of the database have been pre-processed for the Content-Based Image module because some of them have extra-information on the image itself. This extra-information consists of some bands on the frame of the image with color pixels of the RGB and MCY system colors. This kind of information is often used for color calibration. So that, the first attempt was to use this extra-information in order to calibrate the color images of the database. But, after a visual analysis of different images we realize that they don’t follow an established format. At fig. 3 different images formats are shown: two vertical color bands, two horizontal color bands, only one color band, some color bands have the two color systems (RGB and MCY), others only one of the color systems, others extra white frame of different sizes. Therefore, the solution adopted was to reduce all the images to the 90% of his real size in order to eliminate the different bands and the white pixels frames. Fig. 3. Different format images from the Belga News Agency database. 2.2 Queries files construction Analyzing both “topics_part1.txt” and “topics_part2.txt” task topics files, we built different queries files to be launched against the IDRA indexation as the first step in the generation of our experiments. The IDRA queries format separated by a blank. The different queries files constructed are explained in the following using the example in Fig. 4 of this year’s topics (from ‘Topics – part 1’) shown in the official website of the task [14]. 1012315 England's Sarah Price emerges out of the water during the Women's 50m backstroke heats at the Manchester Aquatic Centre, as part of the Commonwealth Games in Manchester, Friday 02 August 2002 UK OUT NO MAGS NO SALES NO ARCHIVES NO INTERNET MAN02 - 20020802 - MANCHESTER, UNITED KINGDOM MANCHESTER, UNITED KINGDOM 20020802 1012315.jpg Number: 0 soccer soccer belgium Relevant images contain photographs of the Belgium team in a soccer match. belga38/00704995.jpg spain soccer Relevant images contain photographs of the Spain team in a soccer match. belga6/00110574.jpg beach soccer Relevant images contain photographs of a soccer beach match. belga33/06278068.jpg italy soccer Relevant images contain photographs of the Italy team in a soccer match. belga20/1027435.jpg soccer netherlands Relevant images contain photographs of the Netherlands team in a soccer match or the teams in Netherlands' league. belga10/01214810.jpg soccer -belgium -spain -beach -italy –Netherlands Relevant images contain photographs of any aspects or subtopics of soccer which are not related to the above clusters. belga20/01404831.jpg Fig. 4. Example of ImageCLEFphoto09 topics (from ‘Topics – part 1’) - [qf1] “BELGAtopics-all-(q)-fQ.txt”: one query per topic containing one stream with all the text from all the clusters (not used for runs). [qf1] would contain the query shown in Fig. 5. 0 soccer soccer belgium Relevant images contain photographs of the Belgium team in a soccer match. spain soccer Relevant images contain photographs of the Spain team in a soccer match. beach soccer contain Relevant images contain photographs of a soccer beach match. italy soccer Relevant images contain photographs of the Italy team in a soccer match. soccer netherlands Relevant images contain photographs of the Netherlands team in a soccer match or the teams in Netherlands' league. soccer -belgium -spain -beach -italy –netherlands Relevant images contain photographs of any aspects or subtopics of soccer which are not related to the above clusters. Fig 5. [qf1] example - [qf2] “BELGAtopics-tctcd-(q-cl)-fQ.txt”: one query for each cluster of each topic with the text from title, clusterTitle and clusterDescription. We eliminate the negative sentences (those containing words “not” or “irrelevant”). We do not include the negative clusters as “soccer -belgium -spain -beach -italy - netherlands”. [qf2] would contain the queries shown in Fig. 6 for the topic in the example of Fig. 4: 0-1 soccer soccer belgium Relevant images contain photographs of the Belgium team in a soccer match. 0-2 soccer spain soccer Relevant images contain photographs of the Spain team in a soccer match. 0-3 soccer beach soccer contain Relevant images contain photographs of a soccer beach match. 0-4 soccer italy soccer Relevant images contain photographs of the Italy team in a soccer match. 0-5 soccer soccer netherlands Relevant images contain photographs of the Netherlands team in a soccer match or the teams in Netherlands' league. 1-1 … Fig 6. [qf2] example - [qf3] “BELGAtopics-tct-(q-cl)-fQ.txt”: the same as above but just with the text from title and clusterTitle. Fig 7. shows the constructed queries from the example topic for [qf3]. 0-1 soccer soccer Belgium 0-2 soccer spain soccer 0-3 soccer beach soccer 0-4 soccer italy soccer 0-5 soccer soccer Netherlands 1-1 … Fig 7. [qf3] example - [qf4] “BELGAtopics-topEnt(1..25)+capEnt(26..50)-(q-cl)-fQ.txt”: one query for each cluster (except negatives ones) of each topic from 1 to 25 and for each one of the three images of each topic from 26 to 50. The associated text of each query is obtained extracting the named entities (with the NER tagger module) from the clusterTitle and clusterDescription fields of the corresponding topic, in the case of topics 1 to 25, and from the associated XML files for each of the three images in the case of topics 26 to 50. In the case of the topic example in Fig. 4, the corresponding constructed queries in [qf4] would be: 0-1 soccer belgium belgium belgium soccer soccer match 0-2 spain spain soccer spain spain team soccer soccer match 0-3 beach soccer soccer soccer beach match beach match 0-4 italy soccer soccer italy italy team soccer soccer match. 0-5 soccer netherlands netherlands netherlands team soccer soccer match netherlands league 1-1 … Fig 8. [qf4] example - [qf5] “BELGAtopics-cap(title+desc)-(26..50)-(q-cl)-fQ.txt”: one query for each one of the three images of each topic from 26 to 50. The text for each query is obtained from the concatenation of the TITLE and DESCRIPTION fields of the XML files for these captions. 2.3 IDRA text-based index and retrieval IDRA textual retrieval is based on the VSM approach using weighted vectors based on the TF-IDF weight. Applying this approach, a representing vector will be calculated for each one of the image captions in the collection. The components of the vectors will be the weight values for the different words in the collection. When a query is launched, a vector for that query is also calculated and compared with all the vectors stored during the index process. This comparison will generate the ranked results list for the launched query. The textual retrieval task architecture can be seen in the Fig. 1. Each one of the components takes care of a specific task. These tasks will be sequentially executed: - Text Extractor. Is in charge of extracting the text from the different files. It uses the JDOM Java API to identify the content of each of the tags of the captions XML files. This API has problems with some special characters, so it is needed to carry out a pre-process of the text to eliminate them. - Preprocess. This component process the text in two ways: o special characters deletion: characters with no statistical meaning, like punctuation marks, are eliminated. o stopwords detection: exclusion of semantic empty words from a new constructed list, different from last year one. - XML Fields Selection. With this component, it is possible to select the desired XML tags of the captions files, which will compound the associated text describing each image. In the captions XML files there are eight different tags (DOCNO, TITLE, DESCRIPTION, NOTES, LOCATION, DATE, IMAGE and THUMBNAIL). In the index process, the selected tags from the captions XML files had been three: TITLE, DESCRIPTION, and LOCATION. - IDRA Index. This module indexes the selected text associated with each image (its XML caption). The approach consists in calculate the weights vectors for each one of the images selected texts. Each vector is compounded by the TF-IDF weights values [16] of the different words in the collection. TF-IDF weight is a statistical measure used to evaluate how important a word is to a text in a concrete collection. ⎛N⎞ ti,j: number of occurrences of the word tj in caption text Ti. TF − IDF = ti , j * log 2 ⎜ ⎟ N: total number of images captions in the collection. (1) ⎝ ni ⎠ ni: number of captions in which appears the word ti. All the weights values of each vector will be then normalized using the Euclidean distance between the elements of the vector. Therefore, the IDRA Index process update the next values for each one of the words appearing in the XML captions collection: ni: number of captions in which appears the word ti, Ti: identifier of the image XML caption, ti,j: number of occurrences of the word tj in a caption text Ti, idfj: inverse document frequency ( log2(N/ni) ) in Ti, Ei: Euclidean distance in the corresponding vector used to normalize, wj,i: weight of word tj in Ti. - IDRA Search. For the query text is also calculated his weights vector in the same way as above. Now, the similarity between the query and an image caption will depend on the proximity of their associated vectors. To measure the proximity between two vectors we use the cosine. sim (T i , q ) = cos( Ö ) = ∑ w *w j, i j, i (2) ∑w w * ∑w j, i * j, i q, i * wq, i This value of similarity will be calculated between the query and all the images captions indexed, and the images will be ranked in descending order as the IDRA result list. To index the collection, the system needs approximately 2 days to index each one of the 5 parts in which the collection was divided to be indexed. These 5 indexations processes can be executed concurrently. Queries file response time depends on the concrete queries file launched (on the large of the queries texts), but it takes over 10 hours to obtain a results file for 119 queries (119 queries at cluster level). 2.4 Named Entity Recognition (NER) functionality The general aim of using Named Entity Recognition (NER) in our approach was to perform retrieval using the named entities extracted from both the documents (XML-structured captions as discussed in section 2.1) and the topics. Due to time constraints, only the later was used in one of the runs submitted ([run5]). Considering the nature of the text in the topics, namely, all words were lowercase, topics in the part 2 file, did not contain enough text, etc., would not make it easy to use an off-the-self named entity tagger, we decided instead to tag (tokenize, Part-Of-Speech and NER) the captions document as released by the imageCLEF organisers. The C&C taggers [6] were used off-the-self. After that, our task was reduced to extract those unique linguistic expressions that were tagged as Location, Person or Organization. However, we soon realized that the resulting annotation was not correctly picking up information which seems to be crucial to determine the topic of the images. More specifically, "Time" would refer to an Organization whereas the description "Time magazine correspondent" would refer to a person and, as such, the modifier correspondent seems relevant to describe the situation captured by a given image. Most of these modifiers would not be picked-up by a NER tagger because they are not in uppercase. This turned out to be a symptom of a more general problem as presented in [9]. For tasks such as IR, it is not sufficient to have low level tools that produce high quality linguistic annotation and analysis, but it is also required that the results of the various levels of annotation be consistent with each other Our proposal aims to exploit the interaction between the various levels of annotations (POS, NER and Chunks) provided by the C&C taggers in order to obtain a better bracketing of named entities. The general idea is to create foci consisting of those words or expressions marked-up as named entities. Whenever the C&C tagger annotates a word as a named entity, a chunk/phrase is built around it by attaching those surrounding/satellites terms that act as modifiers of the named entity according to their POS and membership to a particular chunk. We are currently able to deal with periods and abbreviations, with prepositions, adjectives and nouns. This approach allows us to extract entities such as Paris-Roubaix race, princess Mathilde, Leonardo da Vinci international airport (instead of Leonardo da Vinci), District of Columbia, Royal Palace of Brussels, etc. The following caption (id 1470132) is illustrative of our procedure: “American photojournalist James Nachtwey in a file photograph from May 18 2003 as he is awarded the Dan David prize in Tel Aviv for his out standing contribution to photography. It was announced by Time magazine on Thurs day, 11 December 2003 that Nachtwey was injured in Baghdad along with Time magazine senior correspondent Michael Weisskopf when a hand grenade was thrown into a Humvee they were traveling in with the US Army. Both journalists are reported in stable condition and are being evacuated to a US military hospital in Germany.” On the one hand, the named entities annotated by the tagger for this text are: American James Nachtwey, Dan David, Tel Aviv, Baghdad, Michael Weisskopf, US Army, US, Germany, and Nachtwey. On the other, the descriptions extracted by our system are: American photojournalist James Nachtwey, Dan David prize, Tel Aviv, Baghdad, Time magazine senior correspondent Michael Weisskopf, US Army, US military hospital, Germany and Natchtwey. It is particularly noticeable that our system was able to recognize Time magazine, and that the topic of the caption is about the Dan David prize and a US military hospital in Germany. This approach was used to extract both the named entities and descriptions of all the captions. We then used the lowercase version of the entities found in the captions that were present in the part 1 file of the topics. This method was also applied to the captions corresponding to the images that formed the clusters in part 2 of the topics. They constituted the queries file [qf4] for the [run5] (see section 2.2 and 3). The total process of annotation, analysis, extraction of entities/descriptions took 87 machine hours (on a standard Pentium 4 PC). 2.5 Visual Retrieval The VISION-Team at the Computer Science Department of the University of Valencia has its own CBIR system mainly used for relevance feedback algorithms evaluation [8, 15], and that was used for ImageCLEF 2008 for the first time. The low-level features of the CBIR system have been adapted for the images of the new image database (2009) taking into account the results of the last year. As in most CBIR systems, a feature vector represents each image. The first step at the Visual Retrieval system is extracting these features for all the images on the database as for each of the cluster query topic images for each question. We use different low-level features describing color and texture to build a vector of features. The number of low-level features has been increased from the 68 component (ImageCLEF 2008) up to 114 components at the current edition. This increment is mainly due to the use of local color histogram descriptors that were not use last year. • Color information: Color information has been extracted calculating both local and global histograms of the images using a bin of size 10x3 on a HSV color system. Local histograms have been calculated dividing the images in four fragments of the same size. For this database, only the H (hue) component has been used so that the other values where almost zero for as it happened at the IAPR database. Therefore, a feature vector of 10 components for the global histogram, and 40 components for the local histograms represent the color information of the image. • Texture information: As it was done for the IAPR data base, six feature textures have been computed for this repository respectively. The first three ones use code from the implementation done by Smith and Burn in Meastex [19]; the rest have been implemented by the authors. The total of texture features builds a vector of 64 components: o Gabor Convolution Energies [10]. o Gray Level Coocurrence Matrix also known as Spatial Gray Level Dependence [7]. o Gaussian Random Markov Fields [4]. o The granulometric distribution function, first proposed by Dougherty [5]. We have used here not the raw distribution but the coefficients that result of fitting its plot with a B-spline basis. o Finally, the Spatial Size Distribution [2]. We have used two different versions of it by using as the structuring elements for the morphological operation that get size both a horizontal and a vertical segment. The second step is to calculate the similarity distance between the features vectors from each image on the database to each the cluster images. Last edition (ImageCLEF 2008), we tested two different metrics to calculate this distance: the Euclidean and the Mahalanobis. In all experiments better results were obtained with the Mahalanobis distance due to the fact that this measure takes into account the correlations of the data set and is scale-invariant being this feature very useful because the broad differences between the different low-level feature values. Therefore, this edition only the Mahalanobis distance has been used in our experiments. These sorted lists are passed to the merging module of the global system. 2.6 Merging algorithms Different merging algorithms were developed in order to fuse together different results lists from visual or textual modules, different textual indexations, or cluster level results into a unique topic level results list. All the 3 merging algorithms detailed here require trec_eval format [19] for input results lists, which is the format required by the ImageCLEF organization to submit the definitive runs. MAXmerge. This algorithm is used to fuse together the N (configurable value) results lists obtained when a concrete queries file is launched against the N indexations corresponding to the N parts in which the collection was divided to be indexed. For each query, the algorithm selects the results from the N lists which have a higher relevance value for the corresponding query, independently of the list the results appears in. The maximum number of results per query in the resulting list is set up to 1000 (‘max’ is a configurable parameter). EQUImerge. When a results list contains results for queries which references clusters (not topics), this algorithm is used to select the formers represents of each cluster to build a unique results list with results for queries which references topics. The algorithm selects the first result of each query-cluster, not selected yet, to build the results list for the corresponding query. A preliminary step is carried out to separate the results into several lists depending on the number of cluster (inside each topic) the results belong to. The relevance value will be decremented (configurable value ‘decr’) for each result starting with the original relevance value of the first selected result. The maximum number of results per query is set up to 1000 (‘max’ is a configurable parameter). ENRICH. This merging uses two results lists, a main list and a support list. The merged results list will have a maximum of 1000 results per query (configurable). If a concrete result appears in both lists for the same query, the relevance of this result in the merged list will be increased in the following way: newRel: new relevance v. in the merged list supRel mainRel: relevance value in the main list newRel = mainRel + ( posRel + 1) supRel: relevance value in the support list posRel: position in the support list (6) Relevance values will be then normalized from 0 to 1. Every result appearing in the support list but not in the main one (for each query), will be added at the end of the corresponding list. In this case, relevance values will be normalized according with the lower value in the main list. In the last year implementation of this algorithm, this addition didn’t work correctly, so this year it has been modified in a proper way. In this experiment, main and support lists are compound by a maximum of 1000 results for each query. Also the merged lists resulting will be limited to the same number of results per query. 3 Experiments (submitted runs) This campaign 5 different runs were submitted to the ImageCLEFphoto2009. All of them were based on the 5- parts indexation carried out using the TITLE, DESCRIPTION and LOCATION XML tags. Runs start from the launch of some of the queries files described in section 2.2. When one of these files is launched, it is against these 5 indexations and later, the 5 obtained results lists over each indexation are merged using the MAXmerge algorithm. - [run1] “MirFI1_T-CT-I_TXT-IMG”: mixed (textual/visual) experiment launching [qf3], reordering this textual results list with content-based results, and merging both lists with the ENRICH algorithm. - [run2] “MirFI2_T-CT-CD-I_TXT-IMG”: the same as above, but launching [qf2]. - [run3] “MirFI3_T-CT-CD_TXT”: textual experiment launching [qf2]. - [run4] “MirFI4_T-CT-CD-I_TXT”: the first part (topics 1 to 25) doing exactly the same as [run3], and the second one (26 to 50) launching [qf5]. - [run5] “MirFI5_T-CT-CD-I_TXT”: textual experiments treating with NER. [qf4] is launched to obtain the results list. As it can be observed, all the queries files used in the submitted experiments were built at cluster level, that is, with different queries for the different clusters of a topic, as explained in the queries files construction (section 2.2). The obtained results lists will have results for each one of the clusters of each topic, and EQUImerge algorithm will be applied to fuse together the different clusters-based lists into a unique query-based one. 4 Results and concluding remarks After the evaluation by the task organizers, obtained results for each of the submitted experiments are presented in Table 1. The table shows for each run: the identifier, the mean average precision (MAP), the R-Precision, the precision at 10 and 20 first results, the number of relevant images retrieved (out of a total of 34887 relevant images in the collection), and the cluster recall at 10 and 20. Average values from all the experiments presented to the task for these metrics are also shown in the table, as well as the best value obtained for each of the metrics. MAP, R-Precision and CR@10 values are all shown in Fig. 9 in order to be visually compared. Table 1. Results for the submitted experiments. Run Identifier MAP R-Prec Prec@10 Prec@20 RelRet CR@10 CR@20 MirFI1_T-CT-I_TXT-IMG 43.78 51.39 82.00 81.80 16547 64.51 72.00 MirFI2_T-CT-CD-I_TXT-IMG 42.25 50.11 80.00 81.00 16301 63.24 73.41 MirFI3_T-CT-CD_TXT 42.33 50.12 80.80 81.40 16301 63.51 73.31 MirFI4_T-CT-CD-I_TXT 27.84 36.96 47.40 48.90 13627 69.83 76.76 MirFI5_T-CT-CD-I_TXT 17.33 29.53 23.60 26.30 13498 68.49 72.82 average 29.08 34.09 -- -- -- 54.67 62.35 best 50.64 56.43 -- -- -- 82.39 86.07 At first sight, we can observe that MirFI1, is our best run for precision metrics (very similar to MirFI2 and MirFI3), and appears in the 16th position in R-Precision classification and in the 19th in MAP one (from a total of 84 submitted experiments). 19 groups participate in the task and only 6 of them obtain better precision results than our best experiment. Regarding the diversity metrics (cluster recall at 10 an 20, CR@10 and CR@20), MirFI4 and MirFI5 obtain our best diversity values, appearing in position 11th (over 84) in cluster recall classification, and being the 5th best group from all the 19 participating ones. best 90 MirFI4 MirFI5 80 MirFI1 MirFI3 MirFI2 average 70 best MirFI1 MirFI2 MirFI3 60 best MirFI1 MirFI3 MirFI2 average 50 MirFI4 average % MirFI5 40 MirFI4 30 MirFI5 20 10 0 MAP R-Prec CR@10 Metrics Fig. 9. Comparison of own experiments with best and average values Comparing obtained results from experiments MirFI1 and MirFI2, we can see that not using CD (cluster description) tag from the topics is a quite better for precision results and very similar in diversity ones. So we can say that the addition of this field in the queries construction step was not very useful. Obtained results for experiments MirFI2 and MirFI3 are almost the same. So we can conclude that the use of the ENRICH merging algorithm with the visual re-ranked results list, does not affect the results in a significant way. MirFI3 and MirFI4 are differentiated in the way of constructing the second half of the queries (topics from 26 to 50). As explained in sections 2.2 and 3, MirFI4 extracts the text from the captions corresponding to the example images included in the second part of the topics. The evaluation of the results shows that MirFI3 obtains better precision results than MirFI4, but worse diversity ones. One reason is that the use of the captions text adds more information to the queries, which is useful for the diversity aim, is noise for the precision one. The goal of experiment MirFI5 was to analyze if results could be improved with the use of NER techniques for the construction of the queries. The obtained precision results for this experiment was our worse ones (seems as a lot of noise was introduced) but the addition of entities information to the queries, improves the diversity results. Experiments MirFI4 and MirFI5, also show how this additional information improves the diversity results, but makes the precision ones worse. Acknowledgements This work has been partially supported by the Spanish R+D National Plan, by means of the project BRAVO (Multilingual and Multimodal Answers Advanced Search – Information Retrieval), TIN2007-67407-C03-03; by the Madrid’s R+D Regional Plan, by means of the project MAVIR (Enhancing the Access and the Visibility of Networked Multilingual Information for the Community of Madrid), S-0505/TIC/000267; and by the Spanish Ministry of Education and Science, by means of the project MCYT, TIC2002-03494. References 1. Arni, T., Clough, P., Sanderson, M., Grubinger, M.: Overview of the ImageCLEFphoto 2008 Photographic Retrieval Task. In: Working Notes for the CLEF 2008 Workshop, 17-19 September, Aarhus, Denmark (2008) 2. Ayala, G., Domingo, J.: Spatial Size Distributions. Applications to Shape and Texture Analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 23, n. 4, pp. 1430--1442. (2001) 3. Belga News Agency, http://www.belga.be. 4. Chellapa, R., Chatterjee, S.: Classification of Textures using Gaussian Markov Random Fields. IEEE Transactions on Acoustics Speech and Signal Processing, vol. 33, pp. 959--963. (1985) 5. Chen, Y., Dougherty, E.R.: Gray-scale morphological granulometric texture classification. Optical Engineering, vol. 33, n. 8, pp. 2713--2722. (1994) 6. Clark, S., Curran, J.: Wide-coverage efficient statistical parsing with CCG and log-linear models. Computational Linguistics, vol. 33, n. 4, pp. 493--553. (2007) 7. Conners, R.W., Trivedi, M.M., Harlow, C. A.: Segmentation of high-resolution urban scene using texture operators. Computer Vision, Graphics, and Image Processing, vol. 25, n. 3, pp 273--310. (1984) 8. de Ves, E., Domingo, J., Ayala G., Zuccarello, P.: A novel Bayesian framework for relevance feedback in image content- based retrieval systems. Pattern Recognition, vol. 39, pp. 1622--1632. (2006) 9. Finkel, J.R., Manning, C.D.: Joint parsing and named entity recognition. In: Proceedings of Human Language Technologies: The 2009 Annual Conference of the North American Chapter of the Association for Computational Linguistics, pp: 326--334, Boulder, Colorado. Association for Computational Linguistics. (2009) 10.Fogel, I., Sagi, D.: Gabor filters as texture discriminator. Biological Cybernetics, vol. 61, n. 2, pp. 103--113. (1989) 11.Granados, R., Benavent, X., Garcia-Serrano, A., Goñi, J.M.: MIRACLE-FI at ImageCLEFphoto 2008: Experiences in merging text-based and content-based retrievals. In: Working Notes for the CLEF 2008 Workshop, 17-19 September, Aarhus, Denmark (2008) 12.Granados, R., García-Serrano, Ana., Goñi, J.M.: La herramienta IDRA (Indexing and Retrieving Automatically). Demostración en la XXV edición del Congreso Anual de la Sociedad Española para el Procesamiento del Lenguaje Natural 2009 (SEPLN´09). (2009) 13.Grubinger, M., Clough, P., Müller, H., Deselaers, T.: The IAPR TC-12 Benchmark: A New Evaluation Resource for Visual Information Systems. In: Proceedings of International Workshop OntoImage’2006 Language Resources for Content-Based Image Retrieval, held in conjuction with LREC'06, pp. 13--23, Genoa, Italy, 22 May 2006 (2006) 14.ImageCLEF 2009 Photo Retrieval Task, http://www.imageclef.org/2009/photo 15.Leon, T., Zuccarello, P., Ayala, G., de Ves, E., Domingo, J.: Applying logistic regression to relevance feedback in image retrieval systems, Pattern Recognition, vol. 40, pp. 2621--2632. (2007) 16.Manning, C.D., Raghavan, P., Schtze, H.: Introduction to information retrieval. Cambridge Univ Press New York, NY, USA. (2008) 17.Paramita, M., Sanderson, M., Clough, P.: Diversity in photo retrieval: overview of the ImageCLEFPhoto task 2009. CLEF working notes 2009, Corfu, Greece. (2009) 18.Smith, G., Burns, I.: Measuring texture classification algorithms. Pattern Recognition Letters, vol. 18, n. 14, pp. 1495-- 1501. (1997) 19.Text REtrieval Conference (TREC), http://trec.nist.gov/