=Paper=
{{Paper
|id=Vol-1178/CLEF2012wn-ImageCLEF-SimpsonEt2012
|storemode=property
|title=ITI's Participation in the ImageCLEF 2012 Medical Retrieval and Classification Tasks
|pdfUrl=https://ceur-ws.org/Vol-1178/CLEF2012wn-ImageCLEF-SimpsonEt2012.pdf
|volume=Vol-1178
}}
==ITI's Participation in the ImageCLEF 2012 Medical Retrieval and Classification Tasks ==
ITI’s Participation in the ImageCLEF 2012 Medical Retrieval and Classification Tasks Matthew S. Simpson, Daekeun You, Md Mahmudur Rahman, Dina Demner-Fushman, Sameer Antani, and George Thoma Lister Hill National Center for Biomedical Communications, U. S. National Library of Medicine, NIH, Bethesda, MD, USA Abstract. This article describes the participation of the Image and Text Integration (ITI) group in the 2012 ImageCLEf medical retrieval and classification tasks. We present our methods for each of the three tasks and discuss our submitted textual, visual, and mixed runs as well as their results. While our methods generally perform well for each task, our best ad-hoc image retrieval submission was ranked first among all the submissions from the participating groups. Keywords: Image Retrieval, Case-based Retrieval, Image Modality 1 Introduction This article describes the participation of the Image and Text Integration (ITI) group in the ImageCLEF 2012 medical retrieval and classification tasks. Our group is from the Communications Engineering Branch of the Lister Hill National Center for Biomedical Communications, which is a division of the U. S. National Library of Medicine. The medical track [15] of ImageCLEF 2012 consists of an image modality classification task and two retrieval tasks. For the classification task, the goal is to classify a given set of medical images according to thirty-one modalities (e.g., “Computerized Tomography,” “Electron Microscopy,” etc.). The modalities are organized hierarchically into meta-classes such as “Radiology” and “Microscopy,” which are themselves types of “Diagnostic Images.” In the first retrieval task, a set of ad-hoc information requests is given, and the goal is to retrieve the most relevant images from a collection of biomedical articles for each topic. Finally, in the second retrieval task, a set of case-based information requests is given, and the goal is to retrieve the most relevant articles describing similar cases. In the following sections, we describe the textual and visual features that comprise our image and article representations (Sections 2–3) and our methods for the modality classification (Section 4) and medical retrieval tasks (Sections 5–6). Our textual approaches primarily utilize the Unified Medical Language System R (UMLS R ) [11] synonymy to identify concepts in topic descriptions and article text, and our visual approaches rely on computed distances between descriptors of various low-level visual features. In developing mixed approaches, we explore the use of clustered visual features that can be represented using text, attribute selection, and ranked list merging strategies. In Section 7 we describe our submitted runs, and in Section 8 we present our results. For the modality classification task, our best submission achieved a classification accuracy of 63.2% and was ranked within the submissions from the top three participating groups. Our best submission for the ad-hoc image retrieval task was ranked first overall, achieving a mean average precision of 0.2377, which is a statistically significant improvement over the second ranked submission. For the case-based article retrieval task, our best submission achieved a mean average precision of 0.1035 and was ranked within the submissions from the top four participating groups, but this submission is statistically indistinguishable from our other case-based submissions. 2 Image Representation for Ad-hoc Retrieval We represent the images contained in biomedical articles using a combination of the textual and visual features described below. 2.1 Textual Features We represent each image in the collection as a structured document of image- related text called an enriched citation. Our representation includes the title, abstract, and MeSH R terms1 of the article in which the image appears as well as the image’s caption and “mentions” (snippets of text within the body of an article that discuss an image). These features can be indexed with a traditional text-based information retrieval system, or they may be exposed as term vectors and combined with the visual feature vectors described below. 2.2 Visual Features In addition to the above textual features, we also represent the visual content of images using various low-level visual descriptors. Table 1 summarizes the descriptors we extract and their dimensionality. Due to the large number of these features, we forego describing them in any detail. However, they are all well-known and discussed extensively in existing literature. Cluster Words. To avoid the computational complexity of computing distances between the above visual descriptors, we create a textual representation of visual features that is easily integrated with our existing textual features. For each visual descriptor listed in Table 1, we cluster the vectors assigned to all images using the k-means++ [3] algorithm. We then assign each cluster a unique “cluster word” and represent each image as a sequence of these words. We add an image’s cluster words to its enriched citation as a “global image feature” field, which can be searched using a traditional text-based information retrieval system. Attribute Selection. An orthogonal approach to transforming our visual de- scriptors into a computationally manageable representation is attribute selection. By eliminating unneeded or redundant information, these techniques can also improve our modality classification and image retrieval methods. We perform attribute selection using the WEKA [8] data mining software. First, we group all our visual descriptors into a single combined vector, and we then perform attribute selection to reduce the dimensionality of this combined feature. 1 MeSH is a controlled vocabulary created by U. S. National Library of Medicine to index biomedical articles. Table 1: Extracted visual descriptors. Descriptor Dimensionality Autocorrelation 25 Color and edge directivity? (CEDD) [5] 144 Color layout? (CLD) [4] 16 Color moment 3 Edge frequency 25 Edge histogram? (EHD) [4] 80 Fuzzy color and texture histogram? (FCTH) [6] 192 Gabor moment? 60 Gray-level co-occurrence matrix moment (GLCM) [19] 20 Local binary pattern (LBP1 ) [14] 256 Local binary pattern (LBP2 ) [14] 256 Local color histogram (LCH) 1024 Primitive length 5 Scale-invariant feature transformation? (SIFT) [12] 256 Semantic concept (SCONCEPT) [16] 30 Shape moment 5 Tamura moment? [20] 18 Combined 2415 ? Feature computed using the Lucene Image Retrieval library [13]. 3 Article Representation for Case-based Retrieval We represent articles using the textual features of each image appearing in the article. Thus, each article enriched citation consists of its title, abstract, and MeSH terms as well as the caption and mention of each contained image. 4 Modality Classification Task We experimented with both flat and hierarchical modality classification methods. Below we describe our flat classification strategy, an extension of this approach that exploits the hierarchical structure of the classes, and a post-processing method for improving the classification accuracy of illustrations. 4.1 Flat Classification Figure 1a provides an overview of our basic classification approach. We utilize multi-class support vector machines (SVMs) as our flat modality classifiers. First, we extract our visual and textual image features from the training images (representing the textual features as term vectors). Then, we perform attribute selection to reduce the dimensionality of the features. We construct the lower- dimensional vectors independently for each feature type (textual or visual) and combine the resulting attributes into a single, compound vector. Finally, we use the lower-dimensional feature vectors to train multi-class SVMs for producing textual, visual, or mixed modality predictions. 4.2 Hierarchical Classification Unlike the flat classification strategy described above, it is possible to exploit the hierarchical organization of the modality classes in order to decompose the task Modality classfication Illustration General Single COMP COMP Single GTAB Radiology_3D Photo Microscopy GPLI DRUS DVDM DMLI GFIG DRMR DVEN DMEL GSCR DRCT DVOR DMTR GFLO DRAN GNCP DMFL GSYS DRXR GGEN DRPE GGEL DRCO GCHE D3DR GMAT GHDR DSEE DSEC DSEM (a) Classifier organization (b) Revised modality hierarchy Fig. 1: Overview of our flat and hierarchical modality classification strategies. into several smaller classification problems that can be sequentially applied. Based on our visual observation of the training samples and our initial experiments, we modified the original modality hierarchy [15] proposed for the task. The hierarchy we used for our experiments is shown in Figure 1b. We train flat multi-class SVMs, as shown in Figure 1a, for each meta-class. For recognizing compound images, we utilize the algorithm proposed by Apostilova et al. [1], which detects sub-figure labels and the border of each sub-figure within a compound image. To arrive at a final class label, an image is sequentially classified beginning at the root of the hierarchy until a leaf class can be determined. 4.3 Illustration Post-processing Because our initial classification experiments resulted in only modest accuracy for the fourteen “Illustration” classes shown in Figure 1b, we concluded that our current textual and visual features may not be sufficient for representing these figures. Therefore, in addition to the aforementioned machine learning modality classification methods, we also developed several complimentary rule- based strategies for increasing the classification accuracy of “Illustration” classes. A majority of the training samples contained in the “Illustration” meta-class, unlike other images in the collection, consist of line drawings or text superimposed on a white background. For example, program listings mostly consist of text; thus, the use of text and line detection methods may increase the classification accuracy of Class GPLI. Similarly, polygons (e.g., rectangles, hexagons, etc.) contained in flowcharts (GFLO), tables (GTAB), system overviews (GSYS), and chemical structures (GCHE) are a distinctive feature of these modalities. We utilize the methods of Jung et al. [10] and OpenCV2 functions to assess the presence of text and polygons, respectively. 5 Ad-Hoc Image Retrieval Task In this section we describe our textual, visual and mixed approaches to the ad-hoc image retrieval task. Descriptions of the submitted runs that utilize these methods are presented in Section 7. 2 http://opencv.willowgarage.com/wiki/ 5.1 Textual Approaches To allow for efficient retrieval and to compare their relative performance, we index our enriched citations with the Essie [9] and Lucene/SOLR3 search engines. Essie is a search engine developed by the U.S. National Library of Medicine and is particularly well-suited for the medical retrieval task due to its ability to automatically expand query terms using the UMLS synonymy. Lucene/SOLR is a popular search engine developed by the Apache Software Foundation that employs the well-known vector space model of information retrieval. We have extracted the UMLS synonymy from Essie and use it for term expansion when indexing enriched citations with Lucene/SOLR. We organize each topic description into a frame-based (e.g., PICO4 ) repre- sentation following the method similar to that described by Demner-Fushman and Lin [7]. Extractors identify concepts related to problems, interventions, age, anatomy, drugs, and modality. We also identify modifiers of the extracted con- cepts and a limited number of relationships among them. We then transform the extracted concepts into queries appropriate for either Essie or Lucene/SOLR. 5.2 Visual Approaches Our visual approaches to image retrieval are based on retrieving images that appear visually similar to the given topic images. We compute the visual similarity between two images as the Euclidean distance between their visual descriptors. For the purposes of computing this distance, we represent each image as a combined feature vector composed of a subset of the visual descriptors listed in Table 1 after attribute selection. 5.3 Mixed Approaches We explore several methods of combing our textual and visual approaches. One such approach involves the use of our image cluster words. For performing multimodal retrieval using cluster words, we first extract the visual descriptors listed in Table 1 from each example image of a given topic. We then locate the clusters to which the extracted descriptors are nearest in order to determine their corresponding cluster words. Finally, we combine these cluster words with words taken from the topic description to form a multimodal query appropriate for either Essie or Lucene/SOLR. While the use of cluster words allows us to create multimodal queries, we can instead directly combine the independent outputs of our textual and visual approaches. In a score merging approach, we apply a min-max normalization to the ranked lists of scores produced by our textual and visual retrieval strategies. We then linearly combine the normalized scores given to each image to produce a final ranking. Similarly, a rank merging approach combines the results of our textual and visual approaches using the ranks of the retrieved images instead of their normalized scores. To produce the final image ranking using this strategy, we re-score each retrieved image as the reciprocal of its rank and then repeat the above procedure for combining scores. 3 http://lucene.apache.org/ 4 PICO is a mnemonic for structuring clinical questions in evidence-based practice and represents Patient/Population/Problem, Intervention, Comparison, and Outcome. Another means of incorporating visual information with our retrieval ap- proaches is through the use of a modality classifier. Using our hierarchical modality classification approach, we can first determine the most probable modal- ities for a topic’s example images. After retrieving a set of images using either our textual or visual methods, we can eliminate retrieved images that are not of the same modality as the topic images. An advantage of performing hierarchical classification is that we can filter the retrieved results using the meta-classes within the hierarchy (e.g., “Radiology”). Finally, we often combine the retrieval results produced by several queries into a single ranked list of images. We perform this query combination, or padding, by simply appending the ranked list of images retrieved by a subsequent query to the end of the ranked list produced by the preceding query. 6 Case-Based Retrieval Task Our method for performing case-based retrieval is analogous to our textual approaches for ad-hoc image retrieval. Here, we index the enriched citations described in Section 3 using the Essie and Lucene/SOLR search engines (for performance comparison). We generate textual and mixed queries appropriate for both search engines according to the approaches described in Sections 5.1. As a form of query expansion for case-based topics, we also explore the possibility of determining relevant disease names to correspond with signs and symptoms found in a topic case. To determine a set of potential diseases, we first use the Google Search API5 to search the World Wide Web using a topic case as a query. We then process the top five documents with MetaMap [2] to extract terms having the UMLS semantic type “Disease or Syndrome.” Finally, we select the top three most frequent diseases for query expansion. 7 Submitted Runs In this section we describe each of our submitted runs for the modality classifica- tion, ad-hoc image retrieval, and case-based article retrieval tasks. Each run is identified by its submission file name or trec_eval run ID and mode (textual, visual or mixed). All submitted runs are automatic. 7.1 Modality Classification Task We submitted the following nine runs for the modality classification task: M1. Visual only Flat.txt (visual): A flat multi-class SVM classification using selected attributes from a combined visual descriptor of 15 features (all descriptors in Table 1 except LCH and SCONCEPT). M2. Visual only Hierarchy.txt (visual): Like Run M1 but classification is per- formed hierarchically. M3. Text only Flat.txt (textual): A flat multi-class SVM classification using selected attributes from a combined term vector created from four textual features (article title, MeSH terms, and image caption and mention). M4. Text only Hierarchy.txt (textual): Like Run M3 but classification is per- formed hierarchically. 5 https://developers.google.com/custom-search/v1/overview M5. Visual Text Flat.txt (mixed): A flat multi-class SVM classification combin- ing the feature representations used in Runs M1–3. M6. Visual Text Hierarchy.txt (mixed): Like Run M5 but classification is per- formed hierarchically. M7. Visual Text Flat w Postprocessing 4 Illustration.txt (mixed): Like Run M5 but additional post-processing is applied for “Illustration” classes. M8. Visual Text Hierarchy w Postprocessing 4 Illustration.txt (mixed): Like Run M7 but classification is performed hierarchically. M9. Image Text Hierarchy Entire set.txt (mixed): Like Run M6 but applied to all the images contained in the retrieval collection. 7.2 Ad-hoc Image Retrieval Task We submitted the following ten runs for the ad-hoc image retrieval task: A1. nlm-se (mixed): A combination of three queries using Essie. (A1.Q1) A disjunction of modality terms extracted from the query topic must occur within the caption or mention fields of an image’s enriched citation; a dis- junction of the remaining terms is allowed to occur in any field. (A1.Q2) A lossy expansion of the verbatim topic is allowed to occur in any field. (A1.Q3) A disjunction of the query images’ cluster words must occur within the global image feature field. A2. nlm-se-cw-mf (mixed): A combination of Query A1.Q1 with the additional query below using Essie. (A2.Q2) A lossy expansion of the verbatim topic is allowed occur in any field of an image’s enriched citation and a disjunction of the query images’ cluster words can optionally occur within the global image feature field. Additionally, the retrieved images are filtered so that they share a least common ancestor modality with the query images, as determined by the modality classifier used in Run M9. Query A2.Q2 is distinct from Queries 1.Q2–3 in that the occurrence of a lossy expansion of the topic is not necessarily weighted more heavily than the occurrence of image cluster words. A3. nlm-se-scw-mf (mixed): Like Run A2 but image cluster words are only considered if the modality classifier used in Run M9 identically labels all the example images of a topic. A4. nlm-lc (mixed): A combination of three queries using Lucene with BM25 similarity and UMLS synonymy. (A4.Q1) A fuzzy phrase-based occurrence of the verbatim topic is allowed in any field of an image’s enriched citation. (A4.Q2) A disjunction of the topic words is allowed to occur in any field. (A4.Q3) A disjunction of the query images’ cluster words must occur within the global image feature field. A5. nlm-lc-cw-mf (mixed): A combination of Query A4.Q1 with the additional query below using Lucene with BM25 similarity and UMLS synonymy. (A5.Q2) A disjunction of the topic words is allowed occur in any field of an image’s enriched citation and a disjunction of the query images’ cluster words can optionally occur within the global image feature field. Additionally, the retrieved images are filtered so that they share a least common ancestor modality with the query images, as determined by the modality classifier used in Run M9. A6. nlm-lc-scw-mf (mixed): Like Run A5 but image cluster words are only considered if the modality classifier used in Run M9 identically labels all the example images of a topic. A7. Combined Selected Fileterd Merge (visual): Similarity matching using 62 min-max normalized attributes selected from a combined visual descriptor of 15 features (all descriptors in Table 1 except LCH and SCONCEPT). Retrieval is performed separately for each query image, and the retrieved results are filtered, according to the modality classifier used in Run M9, so that they share the top two modality levels with the query. Images are scored according the query image resulting in the maximum score. A8. Combined LateFusion Fileterd Merge (visual): Like Run A8 but similarity matching is performed separately for seven features (CLD, GLCM, SCON- CEPT, and the color, Gabor, shape, and Tamura moments from Table 1) whose scores are linearly combined with predefined weights. A9. Txt Img Wighted Merge (mixed): A combination of visual Run A7 with a textual run consisting solely of Query A1.Q2 using score merging. A10. Merge RankToScore weighted (mixed): A combination of visual Run A8 with a textual run consisting solely of Query A1.Q2 using rank mering. 7.3 Case-based Article Retrieval Task We submitted the following eight runs for the case-based article retrieval task: C1. nlm-se-max (textual): A combination of three queries for each topic sentence using Essie. (C1.Q1) A disjunction of modality terms extracted from the sentence must occur within the caption or mention fields of an article’s enriched citation; a disjunction of the remaining terms is allowed to occur in any field. (C1.Q2) A lossy expansion of the verbatim sentence is allowed to occur in any field. (C1.Q3) A disjunction of all extracted words and discovered diseases in the sentence is allowed to occur in any field. Articles are scored according to the sentence resulting in the maximum score. C2. nlm-se-sum (textual): Like Run C1 but articles are scored according to the sum of the scores produced for each sentence. C3. nlm-se-frames-max (textual): A combination of the query below with Query C1.Q2 for each topic sentence using Essie. (C3.Q1) An expansion of the frame-based representation of the sentence is allowed to occur in any field of an article’s enriched citation. Articles are scored according to the sentence resulting in the maximum score. C4. nlm-se-frames-sum (textual): Like Run C3 but articles are scored according to the sum of the scores produced for each sentence. C5. nlm-lc-max (textual): A combination of two queries for each topic sentence using Lucene with language model similarity, Jelinek-Mercer smoothing, and UMLS synonymy. (C5.Q1) A fuzzy phrase-based occurrence of the verbatim sentence is allowed in any field of an article’s enriched citation. (C5.Q2) A disjunction of all words and discovered diseases in the sentence is allowed to occur in any field. Articles are scored according to the sentence resulting in the maximum score. C6. nlm-lc-sum (textual): Like Run C5 but articles are scored according to the sum of the scores produced for each sentence. Table 2: Accuracy results for the modality classification task. File Name Mode Accuracy (%) Visual Text Hierarchy w Postprocessing 4 Illustration.txt Mixed 63.2 Visual Text Flat w Postprocessing 4 Illustration.txt Mixed 61.7 Visual Text Hierarchy.txt Mixed 60.1 Visual Text Flat.txt Mixed 59.1 Visual only Hierarchy.txt Visual 51.6 Visual only Flat.txt Visual 50.3 Image Text Hierarchy Entire set.txt Mixed 44.2 Text only Hierarchy.txt Textual 41.3 Text only Flat.txt Textual 39.4 C7. nlm-lc-total-max (textual): A combination of the query below with Queries C5.Q1–2 (as C7.Q2–3) using Lucene with language model similarity, Jelinek- Mercer smoothing, and UMLS synonymy. (C7.Q1) A fuzzy phrase-based occurrence of the entire verbatim topic is allowed in any field of an article’s enriched citation. Articles are scored according to the sentence resulting in the maximum score. C8. nlm-lc-total-sum (textual): Like Run C7 but articles are scored according to the sum of the scores produced for each sentence. 8 Results We present and discuss the results of our modality classification, ad-hoc image retrieval, and case-based article retrieval task submissions below. 8.1 Modality Classification Task Table 2 presents the classification accuracy of our submitted runs for the modality classification task. Visual Text Hierarchy w Postprocessing 4 Illustration.txt, a mixed approach, achieved the highest accuracy (63.2%) of our submitted runs and was ranked fifth overall. However, it ranked within the submissions from the top three participating groups. This result validates our post-processing method used to improve the recognition of “Illustration” classes, and provides, with our previous experience [17], further evidence that hierarchical classification is a successful strategy. Each of our hierarchical classification methods outperforms the corresponding flat approach having the same feature representation. While our submitted runs were only judged on their ability to identify each of the thirty-one modality classes [15], Table 3 presents the classification accuracy of the intermediate classifiers we used for our hierarchical approaches. For each meta-class in the hierarchy shown in Figure 1b, Table 3 gives the number of classes they contain; the classification accuracy associated with the textual, visual, and mixed feature representations; and the dimensionality of the mixed feature representation after attribute selection. These results demonstrate that the accuracies of the intermediate classifiers generally improve as the number of class labels decreases. Given the limited amount of training data in relation to the number of total modalities, the smaller number of labels per classifier likely is significant for explaining why our hierarchical classification approaches consistently outperform their corresponding flat approaches. Table 3: Accuracy results for our intermediate modality classifiers. ID Number of Classes Feature Accuracy (%) Dims.? Mixed Visual Textual 1 2 (Illustration, General) 96.3 95.6 78.6 159 2 3 (Radiology 3D, Microscopy, Photo) 93.5 87.4 83.8 112 3 8 (DRUS, DRMR, . . . , D3DR) 75.9 64.4 71.3 89 4 4 (DMLI, DMEL, . . . , DMFL) 85.0 83.6 69.4 58 5 4 (DVDM, DVEN, . . . , GNCP) 77.6 62.3 89.1 108 6 14 (GTAB, GPLI, . . . , DSEM) 63.5 53.0 41.2 69 ? Feature dimensionality is given for mixed mode classifiers only. Table 4: Retrieval results for the ad-hoc image retrieval task ID Mode MAP bpref P@10 nlm-se Mixed 0.2377 0.2542 0.3682 Merge RankToScore weighted Mixed 0.2166 0.2198 0.3682 nlm-lc Mixed 0.1941 0.1871 0.2727 nlm-lc-cw-mf Mixed 0.1938 0.1924 0.2636 nlm-lc-scw-mf Mixed 0.1927 0.1940 0.2636 nlm-se-scw-mf Mixed 0.1914 0.2062 0.2864 Txt Img Wighted Merge Mixed 0.1846 0.2039 0.3091 nlm-se-cw-mf Mixed 0.1774 0.1868 0.2909 Combined LateFusion Fileterd Merge Visual 0.0046 0.0107 0.0318 Combined Selected Fileterd Merge Visual 0.0009 0.0028 0.0227 8.2 Ad-hoc Image Retrieval Task Table 4 presents the mean average precision (MAP), binary preference (bpref), and early precision (P@10) of our submitted runs for the ad-hoc image retrieval task. nlm-se achieved the highest MAP (0.2377) among our submitted runs and was ranked first overall. Merge RankToScore weighted, the run achieving our second highest MAP (0.2166), was ranked second overall. Comparing these two runs using Fisher’s paired randomization test [18], a recommended statistical test for evaluating information retrieval systems, we find that nlm-se achieved a statistically significant increase (9.7%, p = 0.0016) over the performance of Merge RankToScore weighted. That the two highest ranked runs were multimodal, as apposed to textual, is an encouraging result, and provides evidence that our ongoing efforts at integrat- ing textual and visual information will be successful. In particular, the use by nlm-se and other runs of cluster words, which are indexed and retrieved using a traditional text-based information retrieval system, is an effective way, not only of incorporating visual information with text, but of avoiding the computa- tional expense common among content-based retrieval methods. Furthermore, Merge RankToScore weighted demonstrates the value of rank merging when com- bining textual and visual retrieval results. Some of our other mixed runs, in utilizing the results of our modality classifiers, may have been weakened due to the modest performance of our classification methods. Table 5: Retrieval results for the case-based article retrieval task ID Mode MAP bpref P@10 nlm-lc-total-sum Textual 0.1035 0.1053 0.1000 nlm-lc-total-max Textual 0.1027 0.1055 0.0923 nlm-se-sum Textual 0.0929 0.0738 0.0769 nlm-se-max Textual 0.0914 0.0736 0.0769 nlm-lc-sum Textual 0.0909 0.0933 0.1231 nlm-lc-max Textual 0.0840 0.0886 0.0923 nlm-se-frames-sum Textual 0.0771 0.0693 0.0692 nlm-se-frames-max Textual 0.0672 0.0574 0.0538 8.3 Case-based Article Retrieval Task Table 5 presents the MAP, bpref, and P@10 of our submitted runs for the case- based article retrieval task. nlm-lc-total-sum, a textual approach using language model similarity, achieved the highest MAP (0.1035) among our submitted runs and was ranked seventh overall. However, it ranked within the submissions from the top four participating groups. Using Fisher’s paired randomization test, we find that there is no statistically significant difference (p < 0.05) in MAP among any of our submitted runs. The relatively low performance of most of the ImageCLEF 2012 case-based submissions may be due, in part, to the existence in the collection of only a small number of case reports, clinical trials, or other types of documents relevant for case-based topics. 9 Conclusion This article describes the methods and results of the Image and Text Integration (ITI) group in the ImageCLEF 2012 medical retrieval and classification tasks. For the modality classification task, our best submission was ranked within the submissions from the top three participating groups. Our best submission for the ad-hoc image retrieval task was ranked first overall. Finally, for the case-based article retrieval task, our best submission was ranked within the submissions from the top four participating groups, though we found no statistical significance between this run and our other case-based submissions. The effectiveness of our multimodal approaches are encouraging and provide evidence that our ongoing efforts at integrating textual and visual information will be successful. Acknowledgments. We would like to thank Antonio Jimeno-Yepes for assisting in expanding case-based topics with disease names, Russell Loane for proving source code for converting frame-based topics to Essie queries, and Srinivas Phadnis for constructing enriched citations and extracting visual features. References 1. Apostolova, E., You, D., Xue, Z., Antani, S., Demner-Fushman, D., Thoma, G.: Image retrieval from scientific publications: Text and image content processing to separate multi-panel figures. Journal of the American Society for Information Science and Technology (To appear) 2. Aronson, A.R.: Effective mapping of biomedical text to the UMLS metathesaurus: The MetaMap program. In: Proc. of the Annual Symp. of the American Medical Informatics Association (AMIA). pp. 17–21 (2001) 3. Arthur, D., Vassilvitskii, S.: k-means++: The advantages of careful seeding. In: Pro- ceedings of the Eighteenth Annual ACM-SIAM Symposium on Discrete Algorithms. pp. 1027–1035. SODA ’07 (2007) 4. Chang, S.F., Sikora, T., Puri, A.: Overview of the MPEG-7 standard. IEEE Transactions on Circuits and Systems for Video Technology 11(6), 688–695 (2001) 5. Chatzichristofis, S.A., Boutalis, Y.S.: CEDD: Color and edge directivity descriptor: A compact descriptor for image indexing and retrieval. In: Gasteratos, A., Vincze, M., Tsotsos, J.K. (eds.) Proceedings of the 6th International Conference on Computer Vision Systems. Lecture Notes in Computer Science, vol. 5008, pp. 312–322. Springer (2008) 6. Chatzichristofis, S.A., Boutalis, Y.S.: FCTH: Fuzzy color and texture histogram: A low level feature for accurate image retrieval. In: Proceedings of the 9th International Workshop on Image Analysis for Multimedia Interactive Services. pp. 191–196 (2008) 7. Demner-Fushman, D., Lin, J.: Answering clinical questions with knowledge-based and statistical techniques. Computational Linguistics 33(1), 63–103 (Mar 2007) 8. Hall, M., Frank, E., Holmes, G., Pfahringer, B., Reutemann, P., Witten, I.H.: The WEKA data mining software: An update. SIGKDD Explorations 11(1) (2009) 9. Ide, N.C., Loane, R.F., Demner-Fushman, D.: Essie: A concept-based search en- gine for structured biomedical text. Journal of the American Medical Informatics Association 1(3), 253–263 (2007) 10. Jung, K., Kim, K.I., Jain, A.K.: Text information extraction in images and video: A survey. Pattern Recognition 37(5), 977–997 (2004) 11. Lindberg, D., Humphreys, B., McCray, A.: The unified medical language system. Methods of Information in Medicine 32(4), 281–291 (1993) 12. Lowe, D.: Object recognition from local scale-invariant features. In: Proceedings of the Seventh IEEE International Conference on Computer Vision. vol. 2, pp. 1150–1157 (1999) 13. Lux, M., Chatzichristofis, S.A.: LIRe: Lucene image retrival—an extensible java CBIR library. In: Proceedings of the 16th ACM International Conference on Multi- media. pp. 1085–1088 (2008) 14. Mäenpää, T.: The Local Binary Pattern Approach to Texture Analysis—Extensions and Applications. Ph.D. thesis, University of Oulu (2003) 15. Müller, H., de Herrara, A.G.S., Kalpathy-Cramer, J., Demner-Fushman, D., An- tani, S., Eggel, I.: Overview of the ImageCLEF 2012 medical image retrieval and classification tasks. In: CLEF 2012 Working Notes (2012) 16. Rahman, M.M., Antani, S., Thoma, G.: A medical image retrieval framework in correlation enhanced visual concept feature space. In: Proceedings of the 22nd IEEE International Symposium on Computer-Based Medical Systems (2009) 17. Simpson, M., Rahman, M.M., Phadmis, S., Apostolova, E., Demner-Fushman, D., Antani, S., Thoma, G.: Text- and content-based approaches to image modality classification and retrieval for the ImageCLEF 2011 medical retrieval track (2011) 18. Smucker, M.D., Allan, J., Carterette, B.: A comparison of statistical significance tests for information retrieval evaluation. In: Proceedings of the Sixteenth ACM Conference on Information and Knowledge Management. pp. 623–632 (2007) 19. Srinivasan, G.N., G., S.: Statistical texture analysis. In: Proceedings of World Academy of Science, Engineering and Technology. vol. 36, pp. 1264–9 (2008) 20. Tamura, H., Mori, S., Yamawaki, T.: Textural features corresponding to visual perception. IEEE Transactions on Systems, Man, and Cybernetics 8(6), 460–73 (1978)