<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Simplified Quadtree Image Segmentation for Image Annotation</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Gerardo Rodmar Conde Ma ́rquez Beneme ́rita Universidad Auto ́noma de Puebla 14 sur y Av. San Claudio</institution>
          ,
          <addr-line>San Manuel, 72570,Puebla, Mexico Hugo Jair Escalante</addr-line>
          ,
          <institution>Universidad Auto ́noma de Nuevo Leo ́n, Graduate Program in Systems Engineering San Nicola ́s de los Garza</institution>
          ,
          <addr-line>NL 66450</addr-line>
          ,
          <country country="MX">Mexico</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2011</year>
      </pub-date>
      <volume>1</volume>
      <issue>1</issue>
      <fpage>24</fpage>
      <lpage>34</lpage>
      <abstract>
        <p>This paper introduces a Quadtree image segmentation technique to be used for image annotation. The proposed method is able to efficiently divide the image in homogeneous segments by merging adjacent regions using border and color information. Our method is highly efficient and provides segmentations of acceptable performance; the segments generated with the proposed technique can be used for automatic image annotation and related tasks (e.g., object recognition). A benefit of our proposal is that it allows us finishing the segmentation at any time by controlling the desired level of the detail for the segmentation; hence, the method is suitable for time restricted scenarios. We compare the performance of an image annotation technique trained on hand labeled images and tested in images segmented with different segmentation algorithms. We found that the best results were obtained when the annotation method was tested in images segmented with the quadtree formulation. Our results give evidence of the efficiency and effectiveness of the proposed method.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>
        Although it is possible to annotate images without a previous segmentation, in general region-level
image annotation results in labels of better quality [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. The latter is due to the fact that region-level
information brings additional information (e.g., shape information or spatial context knowledge) that is
not present in image-level annotation [
        <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4">1, 2, 3, 4</xref>
        ]. Of course, the better the segmentation quality the better
is expected to be the performance of region labeling techniques. However, despite the fact that important
advances have been reported on image segmentation, most segmentation algorithms tend to fall in one
of two extremes. On the one hand, there are sophisticated techniques that tend to give better results, but
are too time consuming and thus not appropriate for some applications such as image retrieval [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. On
the other hand, there are methods that are simple and fast (e.g., grid segmentation) although they do not
provide a good support for automatic annotation [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. An attractive alternative consists of methods that
offer a tradeoff between efficiency and segmentation quality. This paper presents one of such
alternative segmentation techniques that aims at give support to automatic image annotation and content-based
image retrieval (CBIR).
      </p>
      <p>The proposed method for image segmentation is a simplified quadtree technique based on the
following guidelines: recursively divide the image using a quadtree approach, merge homogeneous and similar
quadtrees’ regions based on borders and color information, process and discard large image regions as
fast as possible, prefer detection of large objects, avoid noise in result segments and provide any-time
segmentation. Our proposal implements the mentioned guidelines by dividing the segmentation process
in five steps: (1) edge detection, (2) border processing, (3) color discretization, (4) quadtree scanning
and (5) segmentation enhancement. Although our formulation is intended for automatic image
annotation usage, it can be used for other tasks that require a fast segmentation implementation. We present
preliminary results on the application of the proposed segmentation algorithm in the task of image
annotation. Experimental results show that the proposed technique can be very useful for this task and
motivates research in several directions.</p>
      <p>The rest of this paper is organized as follows. The next section presents background information
on quadtree image segmentation. Section 3 introduces the proposed method. Section 4 describes the
experimental settings we adopted for the evaluation of the proposed segmentation method. Section 5
reports experimental results obtained from our formulation. Finally, Section 6 presents some conclusions
and outlines future work directions.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Preliminaries</title>
      <p>This section presents background segmentation on quadtree image segmentation that will be helpful for
understanding the rest of the paper.
2.1</p>
      <sec id="sec-2-1">
        <title>Image segmentation</title>
        <p>
          According to T. Pavlidis the segmentation of an image I into a set of regions S = fS1; : : : ; Skg should
satisfy the following four main rules [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]: S is a partition of I that covers the whole image, no regions
intersect, homogeneity predicate is satisfied by each region and union of adjacent regions does not satisfy
it. This rules can be expressed as follows:
[ik=1Si = S
        </p>
        <p>
          Si \ S j = f ; 8i 6= j
8Si; P(Si) = true
P(Si [ S j) = f alse; 8i 6= j; Siad jacentS j
(1)
(2)
(3)
(4)
where P(X ) represents the fulfilment of the homogeneous predicate under X . However, achieving
this type of segmentation involves many difficulties. The main issues are due to borders, textures and
illumination changes that make hard to distinguish where objects start and end. Moreover,
segmenting an image is a highly subjective process. Therefore, it is difficult to make everybody agree on one
segmentation or another. Aside these difficulties, most image segmentation algorithms are very time
consuming. For these reasons there is a research trend on image annotation that studies annotation
methods that avoid the use image segmentation [
          <xref ref-type="bibr" rid="ref11 ref12">11, 12</xref>
          ]. Nevertheless, previous work on image annotation
has revealed that region-level image annotation can result in labels of better quality than those generated
with image-level techniques [
          <xref ref-type="bibr" rid="ref2 ref3 ref7 ref8">2, 3, 7, 8</xref>
          ]. Therefore, we think it is crucial to improve the performance of
image segmentation methods in support of annotation techniques.
2.2
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Weak segmentation</title>
        <p>
          L. T. Lan and A. Boucher presented a simplified image segmentation process called Weak
Segmentation [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] . They introduce this method as an easy way of getting over common difficulties of the
segmentation process. It avoids generation of perfect segments. Instead, it tries to figure out -in a general
way- where and what are the main objects of the image. This approach reduces segmentation costs in
effort and time. Figure 1 shows and example of weak segmentation. Based on the weak segmentation
formulation, we propose a quadtree based algorithm capable of segmenting images very quickly, yet able
to provide segments which can be used for automatic image annotation.
The Quadtree structure was first introduced by Hanan Samet in [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. A Quadtree is a data structure
concept that refers to a hierarchical collection of maximal blocks that partition a region. All blocks are
disjoint and have predefined sizes according to the base quad size. The item to be partitioned is the root
quadtree which is recursively partitioned according to predefined criteria. Each step of decomposition
produces four new quadtrees of the same size that are hierarchically associated with their parent quadtree.
Decomposition finishes whenever there are no more quadtrees to be partitioned or when the quadtrees
have reached their minimum size. Quadtree based approaches are commonly used due to their ability to
discard very quickly large amounts of information.
3
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Simplified quadtree image segmentation method</title>
      <p>This section describes the proposed technique for image segmentation, which is based on the idea of
Dividing the image following a quadtree structure and merging similar adjacent regions. The proposed
algorithm for simplified quadtree image segmentation requires the specification of three parameters i)
minimum object size (mos), ii) minimum quad size (mqs) and iii) homogeneity threshold (ht). These
parameters can be adjusted by the user according to their own needs, although below we provide default
values for those parameters. The proposed formulation is divided into the following five stages: (1) edge
detection, (2) border processing, (3) color discretization, (4) quadtree scanning and (5) segmentation
enhancement. Figure 2 shows the implementation design model. The rest of this section provides a
detailed description of each step.
3.1</p>
      <sec id="sec-3-1">
        <title>Border Detection</title>
        <p>
          The first step consists of detecting borders of objects in the image. Borders are important for image
segmentation because they can provide information about object contours. In our proposal, border
information has a crucial influence over the final segmentation results. There exist many border detection
algorithms that provide very good results but they are complex and time consuming. Therefore, we
decided to use the Sobel operator which is a simple and straight forward method to detect image borders [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ].
A threshold is applied to the results obtained with the Sobel operator in order to keep only the highest
values.
3.2
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>Border Selection</title>
        <p>Border selection consists of building an image that contains only the most relevant borders obtained from
step 1. Border relevance is measured by its continuity and length. To be able to evaluate these measures,
we first apply the connected components algorithm to the Sobel result and then we calculate the area
of each component. We keep all the components whose area suggest the they belong to an object that
reaches the mos parameter entered to the algorithm. We call this image Constraints Image.
3.3</p>
      </sec>
      <sec id="sec-3-3">
        <title>Color Discretization</title>
        <p>Color discretization creates a 6-bit color copy of the image. 2 bits per each RGB component. We choose
this small number to speed ups histograms calculations as normalizing and comparing.
3.4</p>
      </sec>
      <sec id="sec-3-4">
        <title>Quadtree Segmentation</title>
        <p>The main step of our method is the quadtree segmentation part, which is described in this section. A
quadtree scanning of the image is the core step of the segmentation. The image is divided into four
regions, and each of these regions is compared with their adjacent 4 neighbors using a comparison
operator. If two regions are evaluated as similar they are merged. Regions that are not merged with any
other region are divided in four new regions and the same comparison operation with their new neighbors
is done. This process is performed recursively until there are no more regions to divide or the region size
has reached the mqs parameter.</p>
        <p>The comparison operator takes two adjacent regions and considers the Constraints Image built in
step 2. It counts pixel borders on each region and use predefined rules to check if there is a border
compatibility between these regions. If no border compatibility is found the operator evaluates them as
non similar, otherwise it constructs color histograms for each region and then normalize and eliminate
noise in them. Next, it calculates the Euclidean distance between these histograms. If this distance is
less than the ht parameter, the regions are evaluated as similar. Regions merged during this step give the
shape of the segmentation result. Regions that could not been assigned to any segment due to absence of
homogeneity or similarity are ignored and left out of the final segmentation result. This behavior helps
our algorithm to produce more homogeneous (noise free) segments.</p>
        <p>Due to the progressive approach of this segmentation procedure, we can take advantage of the
parameter mqs to shorten the segmentation time according to our needs, although we will always have an
approximation to the final segmentation result. This is a desirable property of segmentation methods for
time restricted applications such as video processing or CBIR. Figure 3 depicts the quadtree
processing of a sample image, it also shows how can different values for the mqs parameter support anytime
segmentation.
3.5</p>
      </sec>
      <sec id="sec-3-5">
        <title>Segmentation Enhancement</title>
        <p>The final step of the proposed methodology is the enhancement of segmentation result described in last
section. Segmentation enhancement attempts to improve segmentation results in two steps. First, it
applies the comparison operator to all segments that intersected during step 4 and merge them accordingly.
Second, it looks for segments that do not meet the mos parameter and merges them with the most similar
adjacent segment that does. Figure 4 shows an example of image segmentation using our proposal.
3.6</p>
      </sec>
      <sec id="sec-3-6">
        <title>Complexity analysis</title>
        <p>
          The complexity of the whole process of our proposal is O(N) = N pN, where N = w h is the number
of pixels of the image. Where the complexity of each step of the proposed methodology is as follows: (1)
7 N; (2) 4N; (3) N; (4) 2log4(N 1)6N +C, with C the number of operations required to divide a quadtree;
(5) 2K + K2 + N with K the number of segments found. A segmentation algorithm with complexity of
O(N) = N pN can be considered efficient as most segmentation algorithms are of order O(N2) and
higher [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]. Further, one should note that the proposed method is an any-time technique.
4
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Experimental settings</title>
      <p>This section describes the experimental setting we adopted for the evaluation of the proposed
segmentation method. The goal of our experiments is: 1) to evaluate the segmentation performance of the
proposed method and 2) to evaluate the performance of an annotation method with segments generated
with our segmentation technique in terms of segmentation and annotation accuracy.</p>
      <p>
        For the evaluation of the segmentation performance and efficiency we considered three datasets of
images. The first one consists of 18 heterogeneous manually selected images; the second one consists
of 100 randomly selected images from the Berkeley Segmentation Dataset [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]; and the third dataset
consists of 14 images randomly selected from “Image of the Day” Wikipedia section [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. These datasets
where used to progressively enhance the algorithm attempting to produce acceptable weak segmentation
results for more images as possible in all the three datasets. The parameters of the segmentation algorithm
were fixed by trial an error. The values of those parameters are mos = 1:25%; mqs = 2; ht = 0:5, with
these values we have obtained acceptable results for teh different datasets, hence they can be considered
default values.
      </p>
      <p>
        The annotation performance was evaluated with a dataset consisting of 500 images taken from the
SAIAPR TC-12 [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], which is a benchmark of manually segmented and annotated images. We compare
our proposal with two different segmentation algorithms used in state of the art image annotation:
normalized cuts and grid segmentation. Normalized cuts consists of building a weighted graph from the
image in which each node represents a pixel and the arcs’ values are the similarities between connected
pixels using a function that measures more than 30 features on each pixel [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Grid segmentation
consists of cutting the image by grid of n rows and m columns of equal size. After segmenting the 500
images dataset with the three mentioned algorithms, segmentation results from each segmentation
algorithm where given to an external automated annotation engine [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Then, we compared the annotations
obtained for each image with each of the segmentation algorithms.
      </p>
      <p>The evaluation process for image annotation is as follows. For each set of segmentation results given
by each segmentation algorithm for the 500 images, we take each image and match their automatically
produced segments with the corresponding manually produced segments. Then we check if the
annotation in both segments is the same; more specifically, we estimate the percentage of pixels that were
correctly labeled. Intuitively, images receive points for each correct annotation, maximum grade for a
perfectly annotated image is 1. The points given for each correct annotation are weighted according to
segment’s size. This means that if an image has 3 annotations in its manually annotated version, and
one annotation corresponds to a segment that occupies 75% of the total image area, then that segment
will give 0.75 points in case it is correctly annotated. The execution time that spent each algorithm to
segment the 500 images was also recorded.</p>
      <p>Our testing environment was an AMD Turion 64 bits 1.6GHz 1GB RAM laptop running Windows
XP. The reason for this choice was that we wanted to assure the performance of our proposal within a
common and accessible environment.
5</p>
    </sec>
    <sec id="sec-5">
      <title>Results</title>
      <p>Sample segmentation results obtained with the proposed method are shown Figure 5. From this figure we
can see that segmentations produced with the proposed method are more accurate than those generated
with normalized cuts. It is common for normalized cuts to partition regions that belong to the same object
in more than a single region, the proposed method, on the other hand, correctly identifies homogeneous
segments. One should note that, as stated above, segmentation is a highly subjective process, hence
evaluating the quality of segmentations produced with different methods is a subjective task as well.
Nevertheless, we believe the segmentations produced with our method provide very useful information
about the objects present in the image.</p>
      <p>
        From Table 1 we can see that the Quadtree Segmentation Algorithm produced segments that were
annotated 50% better than those produced with grid segmentation and 35% better than those generated
with normalized cuts. In addition, our proposal segments about 17 times faster than normalized cuts.
Despite the fact that the proposed method was less efficient than the grid approach, the 50% improvement
makes worthwhile using our method. Whereas the improvement over the grid approach is somewhat
expected, the improvement over the normalized cuts technique is an important result, as normalized cuts
has been the most used image segmentation algorithm in the context of automatic image annotation [
        <xref ref-type="bibr" rid="ref1 ref11 ref16 ref2 ref3">1,
2, 3, 11, 16</xref>
        ].
      </p>
      <p>We introduced a simple and efficient method for image segmentation, which resulted very helpful for
image annotation, as evidenced by experimental results. This is specially true for images with big objects
and simple textures. The any time segmentation feature of our proposal in combination with its order of
complexity makes it attractive to be used in time restricted environments. For the experiments reported
in this paper, parameters were set by trial an error, it would be interesting to develop methods that
automatically can tune parameters according the type of images. Future work also includes improving the
selection of relevance borders and extending the color and texture features used in the comparison
operator, this in order to improve the segmentation results in particular for complex images. The multilevel
annotation taking advantage of the hierarchical nature of quadtree is also encouraging. In addition, a
parallel implementation of the algorithm can be done in order speed even more our formulation.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>K.</given-names>
            <surname>Barnard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Fan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Swaminathan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hoogs</surname>
          </string-name>
          , R. Collins,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rondot</surname>
          </string-name>
          , and
          <string-name>
            <surname>J. Kaufhold.</surname>
          </string-name>
          “
          <article-title>Evaluation of Localized Semantics: Data, Methodology,</article-title>
          and Experiments”,
          <source>International Journal of Computer Vision</source>
          , Vol.
          <volume>77</volume>
          (
          <issue>1-3</issue>
          ), pp.
          <fpage>199</fpage>
          -
          <lpage>217</lpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>H. J.</given-names>
            <surname>Escalante</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Montes</surname>
          </string-name>
          , and
          <string-name>
            <surname>L. E. Sucar.</surname>
          </string-name>
          “
          <article-title>Word Co-occurrence and Markov Random Fields for Improving Automatic Image Annotation”</article-title>
          ,
          <source>Proceedings of the 18th British Machine Vision Conference</source>
          , Vol.
          <volume>2</volume>
          , pp.
          <fpage>600</fpage>
          -
          <lpage>609</lpage>
          , Warwick, UK,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>H. J.</given-names>
            <surname>Escalante</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Montes</surname>
          </string-name>
          and
          <string-name>
            <surname>L. E. Sucar.</surname>
          </string-name>
          “
          <article-title>An Energy-based Model for Region Labeling”</article-title>
          ,
          <source>Computer Vision</source>
          and Image Understanding, in press,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>H. J.</given-names>
            <surname>Escalante</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Grubinger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. A.</given-names>
            <surname>Herna</surname>
          </string-name>
          <article-title>´ndez</article-title>
          ,
          <string-name>
            <given-names>J. A.</given-names>
            <surname>Gonza</surname>
          </string-name>
          <article-title>´lez, A</article-title>
          . Lo´pez,
          <string-name>
            <given-names>M.</given-names>
            <surname>Montes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Morales</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. E.</given-names>
            <surname>Sucar</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L.</given-names>
            <surname>Villasen</surname>
          </string-name>
          <article-title>˜or. “The Segmented and Annotated IAPR TC-12 Benchmark”</article-title>
          ,
          <source>Computer Vision and Image Understanding</source>
          , Vol.
          <volume>114</volume>
          , pp.
          <fpage>419</fpage>
          -
          <lpage>428</lpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Shi</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Malik</surname>
          </string-name>
          . “
          <article-title>Normalized Cuts and Image Segmentation”</article-title>
          ,
          <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          , Vol.
          <volume>22</volume>
          (
          <issue>8</issue>
          ), pp.
          <fpage>888</fpage>
          -
          <lpage>905</lpage>
          ,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>L. T.</given-names>
            <surname>Lan</surname>
          </string-name>
          and
          <string-name>
            <surname>A. Boucher. “</surname>
          </string-name>
          <article-title>An Interactive Image Retrieval System: from Symbolic to Semantic”</article-title>
          <source>Proceedings of the International Conference on Electronics, Informations and Communications</source>
          , Hanoi, Vietnam,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>G.</given-names>
            <surname>Papadopoulos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Mezaris</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Kompatsiaris</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Strintzis</surname>
          </string-name>
          . “
          <article-title>Combining Global and Local Information for Knowledge-Assisted Image Analysis and Classification”</article-title>
          ,
          <source>EURASIP Journal on Advances in Signal Processing</source>
          , Vol.
          <year>2007</year>
          ,
          <string-name>
            <surname>Article</surname>
            <given-names>ID</given-names>
          </string-name>
          45842, 15 pages,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>C.</given-names>
            <surname>Saathoff</surname>
          </string-name>
          and
          <string-name>
            <given-names>S.</given-names>
            <surname>Staab</surname>
          </string-name>
          . “
          <article-title>Exploiting Spatial Context in Image Region Labelling using Fuzzy Constraint Reasoning”</article-title>
          ,
          <source>Proceedings of the 9th International Workshop on Image Analysis for Multimedia Interactive Services</source>
          , pp.
          <fpage>16</fpage>
          -
          <lpage>19</lpage>
          , IEEE, Klagenfurt, Austria,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>D. A.</given-names>
            <surname>Forsyth</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Ponce</surname>
          </string-name>
          . “
          <source>Computer Vision</source>
          a Modern Approach”, Prentice Hall,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>T.</given-names>
            <surname>Pavlidis</surname>
          </string-name>
          . “
          <article-title>Segmentation of Pictures and Maps through Functional Approximation”</article-title>
          ,
          <source>Computer Graphics and Image Processing</source>
          , Vol.
          <volume>1</volume>
          (
          <issue>4</issue>
          ), pp.
          <fpage>360</fpage>
          -
          <lpage>372</lpage>
          ,
          <year>1972</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>K.</given-names>
            <surname>Barnard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Duygulu</surname>
          </string-name>
          , N. de Freitas,
          <string-name>
            <given-names>D. A.</given-names>
            <surname>Forsyth</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Blei</surname>
          </string-name>
          , and
          <string-name>
            <surname>M. I. Jordan.</surname>
          </string-name>
          “
          <article-title>Matching Words and Pictures”</article-title>
          ,
          <source>Journal of Machine Learning Research</source>
          , Vol.
          <volume>3</volume>
          , pp.
          <fpage>1107</fpage>
          -
          <lpage>1135</lpage>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>C.</given-names>
            <surname>Galleguillos</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Belongie</surname>
          </string-name>
          . “
          <article-title>Context Based Object Categorization: A Critical Survey”</article-title>
          ,
          <source>Computer Vision</source>
          and Image Understanding,
          <source>(CVIU)</source>
          , Vol.
          <volume>114</volume>
          , pp.
          <fpage>712</fpage>
          -
          <lpage>722</lpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>H.</given-names>
            <surname>Samet</surname>
          </string-name>
          . “
          <article-title>The Quadtree and Related Hierarchical Data Structures”</article-title>
          ,
          <source>ACM Computing Surveys</source>
          , Vol.
          <volume>16</volume>
          (
          <issue>2</issue>
          ), pp.
          <fpage>187</fpage>
          -
          <lpage>260</lpage>
          ,
          <year>1984</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>[14] http://www.eecs.berkeley.edu/Research/Projects/CS/vision/grouping/segbench/</mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>[15] http://en.wikipedia.org/wiki/Wikipedia:Picture_of_the_day/Archive/</mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>P.</given-names>
            <surname>Carbonetto</surname>
          </string-name>
          , N. de Freitas, and
          <string-name>
            <given-names>K.</given-names>
            <surname>Barnard</surname>
          </string-name>
          .
          <article-title>“A Statistical Model for General Context Object Recognition”</article-title>
          ,
          <source>Proceedings of the 8th European Conference on Computer Vision</source>
          , LNCS Vol.
          <volume>3021</volume>
          , pp.
          <fpage>350</fpage>
          -
          <lpage>362</lpage>
          , Springer, Prague, Czech Republic,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>