=Paper=
{{Paper
|id=Vol-1178/CLEF2012wn-ImageCLEF-StathopoulosEt2012
|storemode=property
|title=IPL at CLEF 2012 Medical Retrieval Task
|pdfUrl=https://ceur-ws.org/Vol-1178/CLEF2012wn-ImageCLEF-StathopoulosEt2012.pdf
|volume=Vol-1178
|dblpUrl=https://dblp.org/rec/conf/clef/StathopoulosSK12
}}
==IPL at CLEF 2012 Medical Retrieval Task ==
IPL at CLEF 2012 Medical Retrieval Task Spyridon Stathopoulos, Nikolaos Sakiotis, Theodore Kalamboukis Department of Informatics Athens University of Economics and Business, Athens, Greece spstathop@aueb.gr tzk@aueb.gr Abstract. This article presents an experimental evaluation on using La- tent Semantic Analysis (LSA) for searching very large image databases. It also describes IPL’s participation to the image CLEF ad-hoc tex- tual and visual retrieval for the medical task in 2012. We report on our approaches and methods and present the results of our extensive experi- ments applying data fusion on the results obtained from LSA on several low-level visual features. Key words: LSA, LSI, CBIR 1 Introduction The continuous advances of internet and digital technologies, as well as the rapidly increasing multimedia content used by modern information systems, have imposed a need for an efficient system for organizing and retrieving content from large multimedia collections. However, the performance in image retrieval is still very far from being effective for several reasons: computational cost, scalability and performance. In our runs this year we have experimented with Latent Semantic Analysis (LSA), a technique that, although has been used successfully in many applica- tions in the domain of text retrieval, [1] it has not experienced similar success in CBIR. The main reason being the density of the f eatures × images ma- trix, C, generated in image retrieval, in contrast to textual retrieval where the term × document matrix is sparse. As a result, complexity cost of SVD is raised to prohibitively high levels for both, space and computational time. In this article we give an overview of the application of our methods to ad-hoc medical retrieval and present the results of our submitted runs. Our efforts this year were concentrated on applying the LSA method to a number of low-level visual features and then using data fusion techniques on the SVD transformed low rank approximation of images to enhance retrieval. We explore a what we call SVD-bypass technique to factor the feature matrix by solving a much smaller in size eigenproblem of the term correlation matrix CC T instead of solving the SVD of matrix C. This method proved to be a much more efficient and scalable solution for large data sets. In the next section, we describe our approach and in the following sections we present the submitted IPL’s runs on textual and visual retrieval with their 2 corresponding descriptions and results. Finally, in the last section we conclude the remarks of this work with propositions for future research. 2 Visual Retrieval According to the traditional use of LSA in information retrieval a term-by- document matrix, C, is first constructed and an SVD analysis is then performed on this matrix. However as stated before, the feature matrix in the case of image retrieval, is a dense matrix. This increases the computational costs of the SVD analysis to prohibitive levels for large image databases. A typical example for our database this year, for the color layout feature will produce a matrix of size 11288 × 305000 ≈ 30GB (in double precision) which makes the SVD impossible to solve with our computer resources. In our LSA implementation we solve the eigenproblem of the feature correlation matrix CC T instead. This matrix, for a suitable representation of the images is of a moderate size, demanding less storage and the eigenvalue problem of CC T can be solved much faster than the SVD factorization of the matrix C. We then approximate the feature matrix taking only the k-largest eigenvalues and corresponding eigenvectors of matrix C, for a suitable value of k. 2.1 Preprocessing of the data It is well known that the representation of a digital image depends on several fac- tors, from its resolution to color models etc. In a collection of images, it is highly possible that there will be important variations considering these characteristics. Thus, each image undergoes through several transformations before the feature extraction step. In our case we have applied the following transformations. 1. Size normalization. All images are re-scaled to the same size. 2. Transformation to gray-scale images. 3. Tile splitting. Each image is split into equal-sized, non-overlapping cells we reference to as tiles. 2.2 Feature Extraction and Selection The vector representation of the images was based on three low-level features of MPEG-7, Scalable Color (SC) with 64 coefficients per tile, Color Layout (CL) having 192 coefficients per tile and the Edge Histogram (EH) feature. Experi- ments on CLEF 2011 image collection showed that, the extraction of the edge histogram per tile had a negative impact on retrieval performance, thus this fea- ture was extracted from the whole image instead. All the features were extracted using the Java library Caliph&Emir of the Lire CBIR system [2]. Finally, a simple histogram with 32 levels of gray colors was extracted from each tile. To increase the discriminating power of the histogram, we remove the levels with high frequency and normalize the remaining histogram values for all 3 images. At the same time, all histogram levels with a total frequency above 80% are considered stop-words and thus, they are removed. We refer to this feature as Color Selection Histogram (CSH). 2.3 Construction of the feature-correlation matrix CC T As we have already mentioned the matrix C is full in the case of CBIR and so it is the matrix CC T . This matrix multiplication is the most intensive part of the computations and memory demanding. In our implementation we overcome all these problems by splitting the matrix C into a number of blocks, such that each block can be accommodated into the memory (C = (C1 , C2 , ..., Cp )) and calculate CC T by: X p T CC = Ci CiT (1) i=1 After solving the eigenproblem of the feature-correlation matrix CC T , the k largest eigenvectors, say Uk , and the corresponding eigenvalues, are selected. The original feature vectors are then projected into the k − th dimensional space using the transformation yk = UkT y (2) on the original vector representation of an image y. 2.4 Data Fusion For the data fusion task we used a weighted linear combination of the results obtained by using the LSA method on different features, as defined by : X SCORE(Q, Image) = wi scorei (Q, Image) (3) i were scorei denotes the similarity score of an Image with respect to a feature i. The weight of each feature type is determined as a function of its performance [3]. The wi ’s were estimated by the square of the Mean Average Precision (MAP) values attained by the corresponding feature on the CLEF ’11 collection [4]. These values are listed in Table 1. Table 1. MAP Values of each feature for the CLEF ’11 collection. Feature MAP Scalable Color 0.0043 Color Layout 0.0133 Color Selection Histogram 0.0023 Edge Histogram 0.0111 4 3 Textual Retrieval This year’s collection contains a subset of PubMed of 305000 images from PubMed. A detailed description of the collection in given in the overview paper in [5]. Since records have the same structure as in previous CLEF collections over the past years, we followed the same steps in the textual retrieval task as in CLEF 2011. Each record is identified by a unique figureID, which is associated with: the title of the corresponding article, the article URL, the caption of the figure, the pmid and the figure URL. From the pmid we downloaded the MeSH terms assigned to each article. Our retrieval system was based on the Lucene 1 search engine. For index- ing we removed stop-words and applied Porter’s stemmer. For the multi-field retrieval the weights of the fields were assigned at indexing time. We kept the same structure of the database as in CLEF 2009, 2010 and 2011. This year we used only the default scoring function 2 which was best performing at the 2011 CLEF Ad-Hoc retrieval. Also we use the same weights for the fields as in the last three years [6, 7]. We used two sets of weights: one, that was estimated empirically on the CLEF 2009 collection and a second set where the weights were estimated by the value of the Mean Average precision estimated on the CLEF 2010 collection. 1 http://lucene.apache.org/ 2 http://lucene.apache.org/core/old versioned docs/versions/3 5 0/api/core/org/ apache/lucene/search/Similarity.html 5 4 Experimental Results 4.1 Results from Textual Retrieval This year we’ve submitted a total of six runs using different combinations of fields and corresponding weights. In Table 2 we give the definitions of our textual runs and in Table 3 their corresponding results. Table 2. Definitions of IPL’s runs on textual retrieval. Run ID Description IPL ATCM Abstract, Title, Caption and Mesh terms all in one field. IPL TCM Title, Caption and Mesh terms all in one field. IPL A10T10C60M2 Abstract, Title, Caption and Mesh in 4 fields with weights 1, 1, 6, 0.2 respectively IPL A1T113C335M1 Abstract, Title, Caption and Mesh in 4 fields with weights 0.1, 0.113, 0.335, 0.1 respectively IPL T10C60M2 Title, Caption and Mesh in 3 fields with weights 1, 6, 0.2 respectively IPL T113C335M1 Title, Caption and Mesh in 3 fields with weights 0.113, 0.335, 0.1 respectively Table 3. IPL’s performance results from textual retrieval. Run ID MAP GM-MAP bpref p10 p30 IPL A1T113C335M1 0.2001 0.0752 0.1944 0.2955 0.2091 IPL A10T10C60M2 0.1999 0.0714 0.1954 0.3136 0.2076 IPL T10C60M2 0.188 0.0694 0.1957 0.3364 0.2076 IPL TCM 0.1853 0.0755 0.1832 0.3091 0.2152 IPL T113C335M1 0.1836 0.0706 0.1868 0.3318 0.2061 IPL ATCM 0.1616 0.0615 0.1576 0.2773 0.1742 6 4.2 Results from Visual Retrieval For the visual retrieval task, we’ve also submitted a total of six runs, using differ- ent values for the parameter k which defines the selected number of eigen-values and vectors to use for indexing and retrieval. For all runs a data fusion on vari- ous features was applied, using the weights acquired from runs of each individual feature with the CLEF’s 2011 collection. In Table 4 we give the definitions of our runs and in Table 5 their corresponding results. Table 4. Definitions of IPL’s runs on visual retrieval. Run ID Description R1: IPL AUEB DataFusion EH LSA Edge Histogram and LSA with k=20 on 64 tiles SC CL CSH 64seg 20k for Scalable Color, Color layout Color Selection Histogram R2: IPL AUEB DataFusion EH LSA Edge Histogram and LSA with k=50 on 64 tiles SC CL CSH 64seg 50k for Scalable Color, Color layout Color Selection Histogram R3: IPL AUEB DataFusion EH LSA Edge Histogram and LSA with k=100 on 64 tiles SC CL CSH 64seg 100k for Scalable Color, Color layout Color Selection Histogram R4: IPL AUEB DataFusion LSA LSA with k=20 on 64 tiles SC CL CSH 64seg 20k for Scalable Color, Color layout Color Selection Histogram R5: IPL AUEB DataFusion LSA LSA with k=50 on 64 tiles SC CL CSH 64seg 50k for Scalable Color, Color layout Color Selection Histogram R6: IPL AUEB DataFusion LSA LSA with k=100 on 64 tiles SC CL CSH 64seg 100k for Scalable Color, Color layout Color Selection Histogram Table 5. IPL’s performance results from visual retrieval. Run ID MAP GM-MAP bpref p10 p30 R4 0.0021 0.0001 0.0049 0.0273 0.0242 R3 0.0018 0.0001 0.0053 0.0364 0.0258 R1 0.0017 0.0001 0.0053 0.0227 0.0273 R6 0.0017 0.0002 0.0046 0.0364 0.0212 R2 0.0011 0.0001 0.004 0.0136 0.0136 R5 0.0011 0.0001 0.0039 0.0091 0.0121 7 5 Conclusions and Further work We have presented a new approach to LSA for CBIR replacing the SVD analysis of the feature the matrix C (m × n) by the solution of the eigenproblem for the matrix CC T (m × m). The method overcomes the high cost of SVD in terms of memory and computing time. In addition, in all the experiments, of which only a small part was submitted officially this year, the optimal value of the approximation parameter was less than 50 which makes the method attractive for fusion with several low level features. Certainly our approach is promising and has created new research directions that need further investigation. The image representation has an impact on LSA performance and a more systematic research on that direction is currently under progress. Also the eigenvalues of the matrix CC T follow a Zipfian distribution with the k-th largest values been well separated giving small residual vectors to machine accuracy, which give us an evidence on the stability of the calculated eigenvectors. More work is currently underway in order to determine the stability of the proposed method. References 1. Deerwester, S.C., Dumais, S.T., Landauer, T.K., Furnas, G.W., Harshman, R.A.: Indexing by latent semantic analysis. JASIS 41(6) (1990) 391–407 2. Lux, M., Chatzichristofis, S.A.: Lire: lucene image retrieval: an extensible java cbir library. In El-Saddik, A., Vuong, S., Griwodz, C., Bimbo, A.D., Candan, K.S., Jaimes, A., eds.: ACM Multimedia, ACM (2008) 1085–1088 3. Wu, S., Bi, Y., Zeng, X., Han, L.: Assigning appropriate weights for the linear combination data fusion method in information retrieval. Inf. Process. Manage. 45 (July 2009) 413–426 4. Kalpathy-Cramer, J., Müller, H., Bedrick, S., Eggel, I., de Herrera, A.G.S., Tsikrika, T.: Overview of the clef 2011 medical image classification and retrieval tasks. In: CLEF (Notebook Papers/Labs/Workshop). (2011) 5. Müller, H., de Herrera, A.G.S., Kalpathy-Cramer, J., Fushman, D.D., Antani, S., Eggel, I.: Overview of the imageclef 2012 medical image retrieval and classification tasks. In: CLEF 2012 working notes, Rome, Ital,2012 (2012) 6. Gkoufas, Y., Morou, A., Kalamboukis, T.: Ipl at imageclef 2011. In: CLEF (Note- book Papers/LABs/Workshops). (2011) 7. Gkoufas, Y., Morou, A., Kalamboukis, T.: Combining textual and visual information for image retrieval in the medical domain. The Open Medical Informatics Journal 5 (2011) 50–57