<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Co-training based Framework for Writer Identification in Offline Handwriting</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Utkarsh Porwal</string-name>
          <email>utkarshp@buffalo.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Venu Govindaraju</string-name>
          <email>govind@buffalo.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Deptt. of Computer Science and Engg., University at Buffalo - SUNY</institution>
          ,
          <addr-line>Amherst, NY - 14228</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <fpage>36</fpage>
      <lpage>40</lpage>
      <abstract>
        <p>-Traditional forensic document analysis methods have focused on feature-classification paradigm where a machine learning based classifier is used to learn discrimination among multiple writers. However, usage of such techniques is restricted to availability of a large labeled dataset which is not always feasible. In this paper, we propose a Cotraining based approach that overcomes this limitation by exploiting independence between multiple views (features) of data. Two learners are initially trained on different views of a smaller labeled training data and their initial hypothesis is used to predict labels on larger unlabeled dataset. Confident predictions from each learner are used to add such data points back to the training data with predicted label as the ground truth label, thereby effectively increasing the size of labeled dataset and improving the overall classification performance. We conduct experiments on publicly available IAM dataset and illustrate the efficacy of proposed approach.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Keywords-Writer Identification, Co-training, Classifier,
Views, Labeled and Unlabeled data</p>
    </sec>
    <sec id="sec-2">
      <title>I. INTRODUCTION</title>
      <p>Writer Identification is a well studied problem in forensic
document analysis where the goal is to correctly label
the writer of an unknown handwriting sample. Existing
research in this area has sought to address this problem
using Machine Learning techniques, where a large labeled
dataset is used to learn a model (supervised learning) that
efficiently discriminates between various different writer
classes. The key advantage of such learning approaches
is their ability to generalize well over unknown test data
distributions. However, such generalization provides greater
performance only when used with a large labeled data.
In real-world scenarios, generating large labeled datasets
requires manual annotation which is not always practical.
The absence of such datasets also leads to inefficient
usage of available unlabeled data that can be exploited to
provide a greater classification performance. To address
these issues, we propose a Co-training based learning
framework that learns multiple classifiers on different views
(features) of smaller labeled data and uses them to predict
labels for unlabeled dataset which are further bootstrapped
to the labeled data for enhancing the prediction performance.</p>
      <p>
        Existing literature on writer identification can be broadly
classified into two categories. First category is of text
dependent features which capture properties of writer
based on the text written. In this writer identification is
done by modeling similar content written by different
writers. This reliance on text dependent features poses
challenges of scalability. In real world application such
data is seldom available which limits the usability of these
techniques for practical purposes. said et al. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] extracted
text dependent features using Gabor filters but the main
limitation was to have a full page of document written by
different writers for identification. Second category is based
on text independent features. They capture writer specific
properties such as slant and loops which are independent
of any text written. These techniques are better suited for
real life scenarios as they directly model writers as opposed
to previous category. Feature selection plays an important
role in such techniques. Several features capturing different
aspects of handwriting has been tried. zois et al. [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] used
morphological features and needed only single word for
identification and niels et al. [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] used allographic features
to compare using Dynamic Time Warping(DTW). All of
this work was focused on better feature selection which
would result in better accuracy. They did not lay stress on
the techniques used and made an assumption that sufficient
amount of such data is available for the system to learn
      </p>
      <p>
        Likewise, writer identification can also be divided
under two major approaches. First is statistical analysis
of several features such as edge hinge distribution. Edge
hinge distribution captures the change in the direction of
writing samples. Second approach is model based writer
identification. In this predefined models of strokes of
handwriting are used. Prime focus of these techniques was
on making a better system for identification using different
techniques for modeling and analysis. Various techniques
such as Latent Dirichlet Allocation(LDA) were proposed
for higher accuracy for identification[
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] but it was based
on the assumption that sufficient training data is available.
      </p>
      <p>
        Existing techniques and methods did not make use of
unlabeled data for the identification. Information tapped in
the unlabeled data can make a significant improvement in
the performance of the system. This information can be
extracted using different techniques such as transductive
SVMs[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] or graph based methods using EM algorithm.
They are used to label unlabeled data in a semi supervised
framework. nigam et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] later proved that Co-training
performs better than these methods in semi supervised
framework. It uses small snippet of labeled data and
iteratively labels some part of unlabeled data. System
retrains itself after every iteration which results in better
accuracy. Co-training has been successfully used for semi
supervised learning in different areas but never been
used for labeling data for writer identification to the best
of our knowledge. Co-training has been used for web
page classification[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], object detection[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] and for visual
trackers[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] . It has been used extensively in NLP for tasks
like named entity recognition[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>The organization of the paper is as follows. Section
2 provides an overview of Co-training based framework.
Multiple data views in form of writer features are described
in Section 3. Section 4 illustrates the proposed approach.
Experimental results are described in Section 5. Section 6
outlines the conclusion.</p>
    </sec>
    <sec id="sec-3">
      <title>II. CO-TRAINING</title>
      <p>
        Co-training is a semi supervised learning algorithm
which needs small amount of training data to start. It
reiteratively labels some unlabeled data points and again
learns from it. blum et al. [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] proposed co-training to
classify web pages on the internet into faculty web pages
and non-faculty web pages. Initially they used small amount
of web pages of faculty members to train a classifier and
were able to correctly classify most of the unlabeled pages
correctly in the end. Co-training requires two separate
views of the data and two learners. blum et al. [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] proved
that co-training works best if the two views are orthogonal
to each other and each of them is capable of classification
independently. They showed that if the two views are
conditionally independent then the accuracy of classifiers
can be increased significantly. This is because system is
using more information to classify data points. Since both
views are sufficient for classification, this brings redundancy
which in turns gives more information. nigam et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]
later proved that completely independent views are not
required for co-training. It works well even if two views
are not completely uncorrelated.
      </p>
      <p>
        Co-training is an iterative bootstrapping method
which increases the confidence of the learner in each
round. It boosts the confidence of score like Expectation
Maximization method but it works better than EM[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. In
EM all the data points are labeled in each round while in
Co-training few of the data points are labeled each round
and then classifiers are retrained. This helps building a
better learner in each iteration which in take would make
better decision and hence the overall accuracy of system
will increase.
      </p>
      <sec id="sec-3-1">
        <title>A. Selection Algorithm</title>
        <p>
          Selection of data points is crucial in the performance
of the algorithm. New points added in each round should
make learner more confident in making decisions about
the labels. Hence, several selection algorithms have been
tried to make a better system as system’s performance can
vary if selection method is changed. Different methods
out performs each other depending on the kind of data
and application. One approach to select points was based
on performance[
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. In this method, some points were
selected randomly and added to the labeled set. System was
retrained and its performance was tested on the unlabeled
data. This process was repeated for some iterations and
performance of every set of points was recorded. Set of
points resulting in best performance were selected to be
added in the labeled set and rest were discarded. This
method was based on the degree of agreement of both
learners over unlabeled data in each round.
        </p>
        <p>
          Some other methods has also been employed like
choosing the top k elements from the newly labeled cache.
This is an intuitive approach as those points were labeled
with the highest confidence by the learner. However, hwa
et al. [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] in their work showed that adding samples with
best confidence score not necessarily results in better
performance of classifiers. So, wang et al. [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] used a
different approach in which some data points with lowest
scores were also added along with the data points with
highest confidence scores. This method was called max-t,
min-s method and t and s were optimized for the best
performance. So, several different selection methods have
been employed as selecting data point in each round is key
to the performance of Co-training.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>III. FEATURE SELECTION</title>
      <p>
        Selection of uncorrelated views is important in the
working of Co-training. blum et al. [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] proposed that both
views should be sufficient for classification. Each learner
trained on the views should be a low error classifier. They
proved that error rates of both the classifiers decreases
during Co-training because of the extra information added
to the system. This extra information directly depends on
the degree of uncorrelation. However, abney et al. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] later
reformulated the explanation given by [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] for the working
of Co-training in terms of measure of agreement between
learners over unlabeled data. abney et al. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] gave an upper
bound on the error rates of learners based on the measure
of their disagreement. Hence, independence of both views
is crucial for the performance of the system. We chose
contour angle features[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] as a first view and we combined
structural and concavity features (SC)[
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] as a second
view. These features can be considered independent as both
captures different properties of style of writing.
      </p>
    </sec>
    <sec id="sec-5">
      <title>IV. PROPOSED METHOD</title>
      <p>Co-training fits naturally for the task of writer
identification as any piece of writing can have different
views. Contour angle features and structural and concavity
features are two such different views for any handwritten
text. They can be considered uncorrelated enough to fit
the task of writer identification in Co-training framework.
Co-training also needs to have two learners to learn over
two views. We used two different instances of Random
Forest as learners to normalize the effect of learner over
views.</p>
      <p>Angle features were used to train first classifier and SC
were used to train the other one. Then in each round a
cache will be extracted from unlabeled data. This cache
would be labeled by both learners and some data points will
be picked from newly labeled cache by selection algorithm.
Selected data points will be added to the training set and
the learners are retrained while remaining data points in
the cache are discarded. This process is repeated unless the
unlabeled set is empty. Below is the pseudo code for the
Co-training algorithm.</p>
      <p>Algorithm 1 Co trainingAlgo
Require:</p>
      <p>L1 Labeled View One
L2 Labeled View Two
U Unlabeled Data
H1 First Classifier
H2 Second Classifier
Train H1 with L1
Train H2 with L2
repeat</p>
      <p>Extract cache C from U
U U C
Label C using H1 and H2
d selection algo(C) where d
add labels(d,H1,H2)
L1 L1 [ view one of d
L2 L2 [ view two of d
Retrain H1 on L1</p>
      <p>Retrain H2 on L2
until U is empty
C</p>
      <sec id="sec-5-1">
        <title>A. Selection Algorithm</title>
        <p>
          Selection algorithm used for selecting data points was
based on agreement of both learners over data points. Points
on which the confidence of both learners was above certain
threshold were selected. In case of documents accuracy
of classifier would be high if two different views will
indicate same label for any data point. Selection method
based on randomly selecting data points and checking their
performance as used in [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] was not good as randomly
checking takes time. The approach is not scalable as there
are several rounds of processing of subset of cache every
time a new cache is retrieved. Below is the pseudo code for
the selection algorithm. Score function in the algorithm gives
the highest value of the confidence scores of the learner for
one data point over all writers.
        </p>
        <p>We used IAM dataset which has total of 4075 line
images written by 93 unique writers. We conducted two
experiments to test the performance of Co-training against
the baseline systems. In first we compared the accuracy
of classifiers after Co-training against baseline methods by
adding the scores of both learners. In this scores of the
class distribution of the two learners were added for each
data point and a joint class distribution score was generated.
Class label with the highest score was assigned to that data
point. Second experiment was based on the maximum of
the confidence score of the label assigned by each learner.
In this each classifier assigns a class label to the data
point. This assignment is based on the highest value of the
confidence score distribution over all classes. Class label
with the higher score between the two is assigned to the
data point.</p>
        <p>Our goal is to show that Co-training can be used to label
unlabeled data even if a small amount of labeled data is
present in the beginning. Therefore experiments were run
on dataset of different sizes. We conducted experiments
with four different settings of data. System was initially
trained over full, half, one fourth and one tenth of the total
training data. In one tenth training data only three samples
per class were present. Table shows that after Co-training
accuracy of classifiers is better than the baseline system
with all sizes of datasets in both experimental settings.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>VI. CONCLUSION</title>
      <p>In this paper we presented a Co-training based
framework for labeling a large dataset of unlabeled document
with the correct writer identities. Previous work in writer
identification was focused on either on developing a better
feature selection algorithm or to use different techniques for
modeling the text of the document. All the work was based
on a assumption that sufficient amount of labeled data is
available for training a system. In our work we address the
problem of limited amount of labeled data present in real
life applications. Our method tries to iteratively generate
more labeled data from unlabeled data. Experimental studies
show that accuracy of learners on the dataset labeled by
Cotraining was better than the baseline system. This proves the
effectiveness of Co-training for labeling a large dataset of
unlabeled documents. In future we would like to address
this problem of limited data by using other semi supervised
learning methods.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Blum</surname>
          </string-name>
          and T. Mitchell,
          <article-title>Combining labeled and unlabeled data with co-training</article-title>
          ,
          <source>In Proceedings ofCOLT '98</source>
          , pp.
          <fpage>92</fpage>
          -
          <lpage>100</lpage>
          .
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Clark</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Curran</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Osborne</surname>
          </string-name>
          ,
          <article-title>Bootstrapping POS taggers using unlabelled data</article-title>
          ,
          <source>In Proceedings of CoNLL</source>
          , Edmonton, Canada, pp.
          <fpage>4955</fpage>
          .
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Abney</surname>
          </string-name>
          , Bootstrapping.
          <source>In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics</source>
          .
          <year>2002</year>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>O.</given-names>
            <surname>Javed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ali</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Shah</surname>
          </string-name>
          ,
          <article-title>Online detection and classification of moving objects using progressively improving detectors</article-title>
          ,
          <source>In Computer Vision and Pattern Recognition</source>
          , pp.
          <fpage>696</fpage>
          -
          <lpage>701</lpage>
          .
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.</given-names>
            <surname>Levin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Viola</surname>
          </string-name>
          and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Freund</surname>
          </string-name>
          ,
          <article-title>Unsupervised improvement of visual detectors using cotraining</article-title>
          ,
          <source>Proceedings of the Ninth IEEE International Conference on Computer Vision</source>
          ,ICCV '
          <fpage>03</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>M.</given-names>
            <surname>Collins</surname>
          </string-name>
          and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Singer</surname>
          </string-name>
          ,
          <article-title>Unsupervised Models for Named Entity Classification, Empirical Methods in Natural Language Processing -</article-title>
          EMNLP.
          <year>1999</year>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>K.</given-names>
            <surname>Nigam</surname>
          </string-name>
          and
          <string-name>
            <given-names>R.</given-names>
            <surname>Ghani</surname>
          </string-name>
          ,
          <article-title>Understanding the Behavior of Cotraining</article-title>
          ,
          <source>In Proceedings of KDD Workshop on Text Mining</source>
          ,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>K.</given-names>
            <surname>Nigam</surname>
          </string-name>
          and
          <string-name>
            <given-names>R.</given-names>
            <surname>Ghani</surname>
          </string-name>
          ,
          <article-title>Analyzing the effectiveness and applicability of co-training</article-title>
          ,
          <source>Proceedings of the Ninth International Conference on Information and Knowledge Management</source>
          , pp.
          <fpage>86</fpage>
          -
          <lpage>93</lpage>
          .
          <year>2000</year>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>R.</given-names>
            <surname>Hwa</surname>
          </string-name>
          ,
          <article-title>Sample selection for statistical grammar induction</article-title>
          ,
          <source>In Proceedings of Joing SIGDAT Conference on EMNLP and VLC</source>
          ,
          <string-name>
            <surname>Hongkong</surname>
          </string-name>
          , China, pp.
          <fpage>4552</fpage>
          .
          <year>2000</year>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>W.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Huang</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Harper</surname>
          </string-name>
          ,
          <article-title>Semi-Supervised Learning for Part-of-Speech Tagging of Mandarin Transcribed Speech</article-title>
          , In IEEE International Conference on Acoustics,
          <source>Speech and Signal Processing</source>
          ,
          <year>2007</year>
          .
          <source>ICASSP</source>
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>T.</given-names>
            <surname>Joachims</surname>
          </string-name>
          ,
          <article-title>Transductive Inference for Text Classification using Support Vector Machines</article-title>
          .
          <source>In Proceedings of the Sixteenth International Conference on Machine Learning</source>
          . pp.
          <fpage>200</fpage>
          -
          <lpage>209</lpage>
          .
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bhardwaj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Reddy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Setlur</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Govindaraju</surname>
          </string-name>
          and
          <string-name>
            <given-names>S.</given-names>
            <surname>Ramachandrula</surname>
          </string-name>
          ,
          <source>Latent Dirichlet allocation based writer identification in offline handwritingIn Proceedings of the 9th IAPR International Workshop on Document Analysis Systems</source>
          . pp.
          <fpage>357</fpage>
          -
          <lpage>362</lpage>
          ,
          <year>2010</year>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>M.</given-names>
            <surname>Bulacu</surname>
          </string-name>
          and
          <string-name>
            <given-names>L.</given-names>
            <surname>Schomaker</surname>
          </string-name>
          ,
          <article-title>Text-Independent Writer Identification and Verification Using Textural and Allographic Features</article-title>
          ,
          <source>In IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          . pp
          <fpage>701</fpage>
          -
          <lpage>717</lpage>
          .
          <year>2007</year>
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>H. E. S.</given-names>
            <surname>Said</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. S.</given-names>
            <surname>Peake</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. N.</given-names>
            <surname>Tan</surname>
          </string-name>
          and
          <string-name>
            <surname>K. D. Baker</surname>
          </string-name>
          ,
          <article-title>Personal identification based on handwriting</article-title>
          .
          <source>Pattern Recognition</source>
          ,
          <volume>33</volume>
          , pp.
          <fpage>149</fpage>
          -
          <lpage>160</lpage>
          .
          <year>2000</year>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>E. N.</given-names>
            <surname>Zois</surname>
          </string-name>
          and
          <string-name>
            <given-names>V.</given-names>
            <surname>Anastassopoulos</surname>
          </string-name>
          ,
          <article-title>Morphological waveform coding for writer indentification</article-title>
          .
          <source>Pattern Recognition</source>
          ,
          <volume>33</volume>
          (
          <issue>3</issue>
          ), pp.
          <fpage>385</fpage>
          -
          <lpage>398</lpage>
          .
          <year>2000</year>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>L.</given-names>
            <surname>Schomaker</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Bulacu</surname>
          </string-name>
          ,
          <article-title>Automatic writer identification using connected-component contours and edge-based features of uppercase Western script</article-title>
          .
          <source>In IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          , pp.
          <fpage>787</fpage>
          -
          <lpage>798</lpage>
          .
          <year>2004</year>
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>R.</given-names>
            <surname>Niels</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Vuurpijl</surname>
          </string-name>
          and
          <string-name>
            <given-names>L.</given-names>
            <surname>Schomaker</surname>
          </string-name>
          ,
          <string-name>
            <surname>Introducing</surname>
            <given-names>TRIGRAPH</given-names>
          </string-name>
          -
          <article-title>Trimodal writer identification</article-title>
          .
          <source>In Proceedings of European Network of Forensic Handwriting Experts</source>
          ,
          <year>2005</year>
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>J.T.</given-names>
            <surname>Favata</surname>
          </string-name>
          , G. Srikantan,
          <string-name>
            <given-names>S.N.</given-names>
            <surname>Srihari</surname>
          </string-name>
          ,
          <article-title>Handprinted character/digit recognition using a multiple feature/resolution philosophy</article-title>
          ,
          <source>In Proceedings of Fourth International Workshop Frontiers of Handwriting Recognition</source>
          .
          <year>1994</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>