<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Handwritten Script Identification from a Bi-Script Document at Line Level using Gabor Filters</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>G.G. Rajput</string-name>
          <email>ggrajput@yahoo.co.in</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Anita H.B.</string-name>
          <email>anitahb@yahoo.com</email>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, Gulbarga University</institution>
          ,
          <addr-line>Gulbarga 585106, Karnataka</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In a country like India where more number of scripts are in use, automatic identification of printed and handwritten script facilitates many important applications including sorting of document images and searching online archives of document images. In this paper, a Gabor feature based approach is presented to identify different Indian scripts from handwritten document images. Eight popular Indian scripts are considered here. Features are extracted from pre-processed images, consisting of portion of a line extracted manually from a handwritten document, using Gabor filters. Script classification performance is analyzed using the k-nearest neighbor classifier (KNN). Experiments are performed using five-fold cross validation method. Excellent recognition rate of 100% is achieved for data set size of 100 images for each script.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>In present information technology era, document processing has become an inherent
part of office automation process. Many of the documents in Indian environment are
multi-script in nature. A document containing text information in more than one
script is called a multi-script document. Many of the Indian documents contain two
scripts, namely, the state’s official language (local script) and English. An automatic
script identification technique is useful to sort document images, select appropriate
script-specific OCRs and search online archives of document images for those
containing a particular script. Handwritten script identification is a complex task due
to following reasons; complexity in pre-processing, complexity in feature extraction
and classification, sensitivity of the scheme to the variation in handwritten text in
document (font style, font size and document skew) and performance of the scheme.
Existing script identification techniques mainly depend on various features extracted
from document images at block, line or word level. Block level script identification
identifies the script of the given document in a mixture of various script documents.
In line based Script identification, a document image can contain more than one script
but it requires the same script on a single line. Word level script identification allows
the document to contain more than one script and the script of every word is
identified. A brief description of the existing pieces of work at line level is given
below.</p>
      <p>
        To discriminate between printed text lines in Arabic and English, three techniques
are presented in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Firstly, an approach based on detecting the peaks in the
horizontal projection profile is considered. Secondly, another approach based on the
moments of the profiles using neural networks for classification is presented. Finally,
approach based on classifying run length histogram using neural networks is
described. An automatic scheme to identify text lines of different Indian scripts from
a printed document is attempted in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Features based on water reservoir principle,
contour tracing, profile etc. are employed to identify the scripts. Twelve Indian
scripts have been explored to develop an automatic script recognizer at text line level
in [
        <xref ref-type="bibr" rid="ref3 ref4">3,4</xref>
        ]. Script recognizer has been designed to classify using the characteristics and
shape based features of the script. Devanagari was discriminated through the headline
feature and structural shapes were designed to discriminate English from the other
Indian script. Further this has been extended with Water Reservoirs to accommodate
more scripts rather than triplets. Using the combination of shape, statistical and
Water Reservoirs, an automatic line-wise script identification scheme from printed
documents containing five most popular scripts in the world, namely Roman,
Chinese, Arabic, Devnagari and Bangla has been introduced [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. This has been further
extended to accommodate 12 different Indian scripts in the same document instead of
assuming the document to contain three scripts (triplets). Here various structural
features, horizontal projection profiles, Water reservoirs (top, bottom, left and right
reservoirs), Contour tracing (left and right profiles) were employed as features with a
decision tree classifier for script identification. In [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], a model to identify the script
type of a trilingual document printed in Kannada, Hindi and English scripts is
proposed. The distinct characteristic features of these scripts are thoroughly studied
from the nature of the top and bottom profiles and the model is trained to learn
thoroughly the distinct features of each script. Some background information about
the past researches on both global based approach as well as local based approach for
script identification in document images is reported in [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Thus, all the reported
studies, accomplishing script recognition at the line level, work for printed
documents. Script identification from handwritten documents is a challenging task
due to large variation in handwriting as compared to printed documents. Some pieces
of work of handwritten script identification of Indian scripts at block and word level
can be found in the literature [
        <xref ref-type="bibr" rid="ref10 ref11 ref8 ref9">8-11</xref>
        ]. To the best of our knowledge, script
identification at line level for Indian handwritten scripts has not been reported in the
literature as compared to non Indian scripts [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. This motivated us to design a robust
system for Indian script identification from handwritten documents at line level for
bilingual scripts. The method proposed in this paper employs analysis of portion of a
line comprising at least two words, for script identification, extracted manually from
the scanned document images. Consequently, the script classification task is
simplified and performed faster as compared to the analysis of the entire line
extracted from the handwritten document. Gabor filter bank is used for feature
extraction and classification is done using KNN classifier.
      </p>
    </sec>
    <sec id="sec-2">
      <title>Properties of Scripts</title>
      <p>A brief description of the properties of the scripts considered in our study is given
below. All these scripts are written from left to right.</p>
      <p>English Script. The modern English (Roman) alphabet is a Latin-based alphabet
consisting of 26 letters each of upper and lower case characters. In addition, there are
some special symbols and numerals. The letters A, E, I, O, U are considered vowel
letters and the remaining letters are considered consonant letters. The structure of the
English alphabet contains more vertical and slant strokes.</p>
      <p>Indian Scripts. The scripts considered in this paper are Devanagari, Kannada, Tamil,
Bangla, Telugu, Punjabi, and Malayalam. All the Indian languages do not have the
unique scripts. Some of them use the same script. Devanagari script is used to write
the languages Hindi, Bhojpuri, Marathi, Mundari, Nepali, Pali, Sanskrit, Sindhi and
many more. Devanagari is recognizable by a distinctive horizontal line running along
the tops of the letters that links them together. Assamese and Bangla languages are
written using the Bangla script; Urdu and Kashmiri are written using the same script
and Telugu and Kannada use the same script. Like Kannada and Telagu, Tamil and
Malayalam belong to southern group of Dravidian languages. The Gujarati script is
derived from the Devanagari script. The major difference between Gujarati and
Devanagari is the lack of the top horizontal bar in Gujarati. The Gurmukhi (Punjabi)
alphabet is modeled on the Landa alphabet. Similar to Devanagari script, in
Gurumukhi most of the characters have a horizontal lines at the upper part called
headline and primarily the characters of words in these scripts are connected by a
these headlines. The image blocks of these scripts are shown in Fig.1. The details
about these scripts can be found elsewhere [http://en.wikipedia.org/wiki/Languages_of_India].
3</p>
    </sec>
    <sec id="sec-3">
      <title>The Proposed Method</title>
      <p>The proposed method is inspired by the observation that in Indian context,
handwritten script identification from multilingual/multi-script documents images is
very promising and is still in emerging status.
3.1</p>
      <sec id="sec-3-1">
        <title>Data collection and Preprocessing</title>
        <p>
          At present, standard database of handwritten Indian scripts are not available. Hence,
we created our own database of handwritten documents. The document pages for the
database have been collected by different persons on request under our supervision.
The writers were asked to write few text lines inside A-4 size pages. Restrictions were
not imposed regarding the content of the text and use of pen. Handwritten documents
were written in English, Devanagari, Kannada, Tamil, Bangla, Telugu, Punjabi, and
Malayalam scripts by persons belonging to different professions. The document pages
were scanned at 300 dpi resolution and stored as gray scale images. The scanned
image is then deskewed using the method defined in [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. Noise is removed by
applying median filter. The portion of lines of width 512 pixels and height equal to
that of the height of the largest character appearing in that line were then manually
cropped out from different areas of the document image, and stored as data set
(Fig. 1). It should be noted that the handwritten text line (actually, portion of the line
arbitrarily chosen) may contain two or more words with variable spaces between
words and characters. Numerals that may appear in the text were not considered. It is
ensured that at least 50% of the cropped text line contains text. A sample of line
images representing different scripts is shown in Fig. 1. These lines, representing a
small segment of the handwritten document images are then binarized using well
known Otsu’s global thresholding approach [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] (Fig. 2(b)). The binary images are
then inverted so that text pixels represent value 1 and background pixels represents
value 0(Fig. 2(c)). The salt and pepper noise around the boundary is removed using
morphological opening. This operation also removes discontinuity at pixel level (Fig.
2(d)). However, we do not try to eliminate dots and punctuation marks appearing in
the text line, since these contribute to the features of respective scripts. A total of 800
handwritten line images containing text are created, with 100 lines per scripts.
        </p>
        <sec id="sec-3-1-1">
          <title>Kannada</title>
        </sec>
        <sec id="sec-3-1-2">
          <title>Gujarati</title>
        </sec>
        <sec id="sec-3-1-3">
          <title>Hindi</title>
        </sec>
        <sec id="sec-3-1-4">
          <title>Tamil</title>
        </sec>
        <sec id="sec-3-1-5">
          <title>English</title>
        </sec>
        <sec id="sec-3-1-6">
          <title>Telugu</title>
          <p>Features are the representative measures of a signal which distinguish it from other
signals. A bank of Gabor filters are chosen for the task under consideration. The
features are extracted by using two dimensional Gabor functions by transforming the
image in time domain to the image in frequency domain. The frequency information
of image is needed to see information that is not obvious in time-domain. Inherent
advantages offered by Gabor function include (i) it is the only function for which the
lower bound of space bandwidth product is achieved, (ii) the shapes of Gabor filters
resemble the receptive field profiles of the simple cells in the visual pathway, and (iii)
they are direction specific band-pass filters. Gabor filters are formed by modulating a
complex sinusoid by a Gaussian function with different frequencies and orientations.
A two dimensional Gabor function consists of a sinusoidal plane wave of some
frequency</p>
          <p>⎛ 1 ⎟⎟⎞ exp⎜⎜⎜⎛ − 1 ⎜⎜⎛ x′2
g(x,y) = ⎝⎜⎜⎜ 2πσ xσ y ⎠⎟ ⎜⎜⎝ 2 ⎝⎜⎜⎜ σ x2</p>
          <p>y′2 ⎟⎞ ⎟⎞
+ 2 ⎟⎟ ⎟⎟ exp(2π jWx′)
σ y ⎟⎟⎠ ⎟⎟
⎠
(1)
where σx2 and σy2 control the spatial extent of the filter, θ is the orientation of the
filter and w is the frequency of the sinusoid.</p>
          <p>We employ two dimensional Gabor filters to extract the features from input text
line image to identify the script type from a bi-script document. The preprocessed
input binary image is convolved with Gabor filters considering six different
orientations (0º, 30º, 60º, 90º, 120º, and 150º) and three different frequencies
(a=0.125, b=0.25, c=0.5) with σx = 2 and σy = 4. The values of these parameters are
fixed empirically. From the 18 output images we compute the features of dimension
54. These features are then fed to the K-NN classifier to identify the script. The
feature extraction algorithm is given below (Algorithm-1). An example of Gabor
filtered images for 00 and 300 degree orientations and frequencies a, b, and c is shown
in Fig. 3.</p>
          <p>(a)
(b)
(c)
(a)
(b)
(c)
7. For each output image of step 6 (out of total 18), perform following steps.</p>
          <p>a. Extract cosine part and compute the standard deviation (18 features).
b Extract sine part and compute the standard deviation(18 features).
c. Compute the standard deviation of the entire output image (18 features).</p>
          <p>This forms feature vector of length 54.
3.3</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>Script Recognition</title>
        <p>The KNN classifier is adopted for recognition purpose. This method is well-known
non-parametric classifier, where posterior probability is estimated from the frequency
of nearest neighbors of the unknown pattern. The key idea behind k-nearest neighbor
classification is that similar observations belong to similar classes. The test image
feature vector is classified to a class, to which its k-nearest neighbor belongs to.
Feature vectors stored priori are used to decide the nearest neighbor of the given
feature vector. The recognition process is described below.</p>
        <p>During the training phase, features are extracted from the training set by
performing feature extraction algorithm given in the feature extraction section. These
features are input to KNN classifier to form a knowledge base that is subsequently
used to classify the test images. During test phase, the test image, which is to be
recognized, is processed in a similar way and features are computed as per the
algorithm described in feature extraction section. The classifier computes the
Euclidean distances between the test feature vector with that of the stored features and
identifies the k-nearest neighbor. Finally, the classifier assigns the test image to a
class that has the minimum distance with voting majority. The corresponding script is
declared as recognized script.
3.4</p>
      </sec>
      <sec id="sec-3-3">
        <title>Experimental Results</title>
        <p>We evaluated the performance of the proposed bi-script identification system on a
dataset of 800 pre-processed images obtained as described in section 3.1. Each
biscript document contains one Indian script and an English script. Further, we have
assumed that the bi-script document contains uniscript text in the portion of line
extracted for experimentation. Samples of one script are input to our system and
performance is noted in terms of recognition accuracy. For each data set of 100 line
images of a particular script, 60 images are used for training and remaining 40 images
are used for testing. Identification of the test script is done using KNN classifier. The
results were found to be optimal for k=1 as compared to other values of k. To test the
robustness of the proposed method k-fold cross validation was carried out with k=5.
The proposed method is implemented using Matlab 6.1 software. The recognition
results of all the scripts are tabulated in Table 1 and Table 2. The results clearly shows
that features extracted by using Gabor function yield very good results. The
recognition accuracy of 100% (nearly) is achieved demonstrating the fact that Gabor
filters provide good features for the text images at line level as compared to other
methods found in the literature. However, the results obtained have certain limitation
as explained below. Firstly, the process of extracting the portion of a line, ensuring
that it consists of at least two words, is manual. Secondly, we have assumed that the
extracted portion of the line is uniscript in text. Thirdly, as with the many other
researchers, we have assumed that the documents are text only. Lastly, we need to
validate our proposed system on a larger database. Experimentation is underway to
take care of these limitations and propose the system in general to recognize the script
at word level.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Conclusion</title>
      <p>In this paper, a robust algorithm for script identification from multi script handwritten
documents is presented. Gabor filters are used for feature extraction. Experiments are
performed at line level by considering only a portion of the line. KNN classifier is
used in recognition phase. Recognition rate of 100% is achieved. The proposed
method is independent of style of hand writing. The novelty of the proposed method
lies in the use of Gabor features on a portion of the line for script recognition, instead
of entire line. We have assumed that such a portion of line contains uniscript text.
Though this assumption is valid for many of the multi-lingual documents, in a general
case we need to recognize the script at word level. Hence, our further study involves
extending the proposed method for the remaining Indian scripts and also for script
type identification at word level.
Acknowledgement. The authors thank the reviewers for their helpful comments.
Also, the authors are very grateful to Dr. P. S. Hiremath, Professor, Department of
Computer Science, Gulbarga University, Gulbarga and Dr. Peeta Basa Pati,
Bangalore, for their valuable suggestions during this work.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Elgammmal</surname>
          </string-name>
          . A. M and
          <string-name>
            <surname>Ismail. M.A</surname>
          </string-name>
          , “
          <article-title>Techniques for Language Identification for Hybrid Arabic-English Document Images”</article-title>
          ,
          <source>Proc. Sixth Int'l Conf. Document Analysis and Recognition</source>
          , pp.
          <fpage>1100</fpage>
          -
          <lpage>1104</lpage>
          , (
          <year>2001</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>U.</given-names>
            <surname>Pal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sinha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. B.</given-names>
            <surname>Chaudhuri</surname>
          </string-name>
          ,
          <article-title>"Multi-Script Line identification from Indian Documents"</article-title>
          , ICDAR,
          <source>Seventh International Conference on Document Analysis and Recognition (ICDAR'03) - vol. 2</source>
          , pp.
          <volume>880</volume>
          (
          <year>2003</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Pal</surname>
          </string-name>
          . U and
          <string-name>
            <surname>Chaudhury.B.B</surname>
          </string-name>
          , “
          <article-title>Identification of Different Script Lines from Multi-Script Documents”</article-title>
          ,
          <source>Image and Vision Computing</source>
          , vol.
          <volume>20</volume>
          , no.
          <fpage>13</fpage>
          -
          <issue>14</issue>
          , pp.
          <fpage>945</fpage>
          -
          <lpage>954</lpage>
          (
          <year>2002</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Pal</surname>
          </string-name>
          . U and
          <string-name>
            <surname>B.B. Chaudhuri</surname>
          </string-name>
          , “
          <string-name>
            <surname>Script Line Separation From Indian Multi-Script</surname>
            <given-names>Documents</given-names>
          </string-name>
          ,
          <source>” 5th ICDAR</source>
          , pp.
          <fpage>406</fpage>
          -
          <lpage>409</lpage>
          (
          <year>1999</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Pal</surname>
            <given-names>U.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Chaudhuri</surname>
          </string-name>
          . B. B, “
          <article-title>Automatic identification of English, Chinese, Arabic, Devnagari and Bangla script line”</article-title>
          ,
          <source>Proc. 6th Intl. Conf: Document Analysis and Recognition (ICDAR'OI)</source>
          , pp
          <fpage>790</fpage>
          -
          <lpage>794</lpage>
          (
          <year>2001</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>M. C.</given-names>
            <surname>Padma</surname>
          </string-name>
          and
          <string-name>
            <given-names>P. A.</given-names>
            <surname>Vijaya</surname>
          </string-name>
          ,
          <article-title>Script Identification From Trilingual Documents Using Profile Based Features</article-title>
          ,
          <source>International Journal of Computer Science and Applications</source>
          , Technomathematics Research Foundation, Vol.
          <volume>7</volume>
          No.
          <issue>4</issue>
          , pp.
          <fpage>16</fpage>
          -
          <lpage>33</lpage>
          (
          <year>2010</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>S.</given-names>
            <surname>Abirami</surname>
          </string-name>
          , Dr. D. Manjula, “
          <article-title>A Survey of Script Identification techniques for Multi-Script Document Images”</article-title>
          ,
          <source>International Journal of Recent Trends in Engineering</source>
          , Vol.
          <volume>1</volume>
          , No. 2, May
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>K.</given-names>
            <surname>Roy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Banerjee</surname>
          </string-name>
          and U. Pal,
          <article-title>“A System for Wordwise Handwritten Script Identification for Indian Postal Automation”</article-title>
          ,
          <source>In Proc. IEEE India Annual Conference</source>
          <year>2004</year>
          ,(
          <issue>INDICON04</issue>
          ), pp.
          <fpage>266</fpage>
          -
          <lpage>271</lpage>
          (
          <year>2004</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>Ram</given-names>
            <surname>Sarkar</surname>
          </string-name>
          ,
          <string-name>
            <surname>Nibaran Das</surname>
          </string-name>
          ,
          <string-name>
            <surname>Subhadip Basu</surname>
          </string-name>
          , Mahantapas Kundu,Mita Nasipuri and Dipak Kumar Basu, “
          <article-title>Word level Script Identification from Bangla and Devanagri Handwritten Texts mixed with Roman Script</article-title>
          ,
          <source>Journal of Computing”</source>
          , volume
          <volume>2</volume>
          ,
          <string-name>
            <surname>Issue</surname>
            <given-names>2</given-names>
          </string-name>
          ,
          <year>February 2010</year>
          , ISSN 2151-
          <fpage>9617</fpage>
          (
          <year>2010</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <given-names>B. V.</given-names>
            <surname>Dhandra</surname>
          </string-name>
          and
          <string-name>
            <given-names>Mallikarjun</given-names>
            <surname>Hangarge</surname>
          </string-name>
          .
          <article-title>Article: “Offline Handwritten Script Identification in Document Images</article-title>
          .
          <source>International Journal of Computer Applications”</source>
          ,
          <volume>4</volume>
          (
          <issue>5</issue>
          ):
          <fpage>1</fpage>
          -
          <lpage>5</lpage>
          ,
          <year>July 2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <given-names>G. G.</given-names>
            <surname>Rajput and Anita H B.</surname>
          </string-name>
          , “
          <article-title>Handwritten Script Recognition using DCT and Wavelet Features at Block Level”</article-title>
          , IJCA, Special Issue on RTIPPR (
          <volume>3</volume>
          ):
          <fpage>158</fpage>
          -
          <lpage>163</lpage>
          (
          <year>2010</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Judith</surname>
            <given-names>Hochberg</given-names>
          </string-name>
          , Kevin Bowers,
          <string-name>
            <given-names>Michael</given-names>
            <surname>Cannon</surname>
          </string-name>
          and Patrick Keely, “
          <article-title>Script and language identification for handwritten document images</article-title>
          ,
          <source>” IJDAR</source>
          , vol.
          <volume>2</volume>
          , pp.
          <fpage>45</fpage>
          -
          <lpage>52</lpage>
          (
          <year>1999</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13. G. G. Rajput,
          <string-name>
            <surname>Anita H</surname>
          </string-name>
          . B.,
          <article-title>“A Two Step Approach for Deskewing Handwritten and Machine Printed Document Images using Histograms and Geometric features”</article-title>
          ,
          <source>Proc. of Second Intl. Conf. on Signal and Image Processing(ICSIP-</source>
          <year>2009</year>
          ), Editors:
          <string-name>
            <given-names>D. S.</given-names>
            <surname>Guru</surname>
          </string-name>
          and T. Vasudev, pp
          <fpage>414</fpage>
          -
          <lpage>417</lpage>
          (
          <year>2009</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14. N. Otsu, “
          <article-title>A Threshold Selection Method from Gray-Level Histogram”</article-title>
          ,
          <source>IEEE Transaction Systems, Man and Cybernetics</source>
          , Vol
          <volume>9</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>62</fpage>
          -
          <lpage>66</lpage>
          (
          <year>1979</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>