<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Oriented Local Binary Patterns for Writer Identification</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Anguelos Nicolaou</string-name>
          <email>anguelos.nicolaou@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marcus Liwicki and Rolf Ingolf</string-name>
          <email>firstname.lastname@unifr.ch</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Document, Image and Voice Analysis (DIVA) Group, University of Fribourg</institution>
          ,
          <addr-line>Bde des Perolles 90, Fribourg</addr-line>
          <country country="CH">Switzerland</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Institute of Computer Science and, Applied Mathematics University of Bern</institution>
          ,
          <addr-line>Neubru ̈ckstrasse 10, 3012 Bern</addr-line>
          ,
          <country country="CH">Switzerland</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>-In this paper we present an oriented texture feature set and apply it to the problem of offline writer identification. Our feature set is based on local binary patterns (LBP) which were broadly used for face recognition in the past. These features are inherently texture features. Thus, we approach the writer identification problem as an oriented texture recognition task and obtain remarkable results comparable to the state of the art. Our experiments were conducted on the ICDAR 2011 and ICHFR 2012 writer identification contest datasets. On these datasets we investigate the strengths of our approach as well its limitations.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>I. INTRODUCTION</title>
      <p>
        Local binary patterns (LBP) were broadly popularized in
2002 with the work of Ojala et al [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] as a texture feature set
extracted directly on grayscale images. As well demonstrated
by Ojala, the histogram of some specific binary patterns is a
very important feature-set. LBP are inherently texture features,
but they have been used in a very broad range of applications
in Computer Vision (CV), many of which exceed the typical
texture recognition tasks. In 2004, Ahonen et al [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] used
successfully LBP for face recognition. In 2007, Zhao et al [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]
extended the operator as a 2D plus time voxel version of LBP,
called VLBP, and used them successfully for facial gesture
recognition. In 2009, Whang et al [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] combined LBP features
with HOG features to address the problem of partial occlusions
in the problem of human detection.
      </p>
      <p>
        While graphology, i.e. the detection of personality traits
based on handwriting, has been associated with bad science [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]
and has failed to provide experimentally sound significant
results [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], handwriting style can be considered an invariant
attribute of the individual. Writer identification has
traditionally been performed by Forensic Document Examiners using
visual examination. In recent decades there is an attempt
to automate the process and codify this knowledge in to
automated methods. In 2005, Bensefia et al [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] successfully
used features derived from statistical analysis of graphemes,
bigrams, and trigrams. In 2008, He et al [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] used Gabor
filter derived features and in 2010 Du et al [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] introduced
LBP on the wavelet domain. Even-though the method of Du
uses LBP for feature extraction in writer identification, the
similarities end there. Our method makes no assumptions
specific to handwriting and treats the problem as a generic
oriented binary texture classification problem. The extent to
which handwriting contains invariant characteristics of the
writer is an open question. While forensic document examiners
have been tested in detecting disguised handwriting by Bird
et al [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], Malik et al [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] have started to address the
issue of different writing styles for automated offline writer
identification systems. It remains an open question whether
handwriting style can provide us with real biometric markers,
invariant to the sample acquisition conditions. By preserving
the generic attributes of our method, we can safely avoid
addressing many complications that are specific to handwriting
analysis and writer detection.
      </p>
    </sec>
    <sec id="sec-2">
      <title>II. LBP FEATURE SET</title>
      <p>Although writer identification seems to require scale
invariant features, scale sensitive features might be suited as well.
Writers tend to write with a specific size, therefore the scale of
the texture tends to be directly dependent on the sampling rate.
The task of writer identification is almost always done with
respect to a dataset, where the sampling rate is defined or at
least known when performing feature extraction. It is feasible
and probably worth the effort of resampling all text images to
a standard sampling resolution, rather than improvising a scale
invariant feature-set. Our feature-set as is the norm, is derived
from the histogram of occurring binary patterns.</p>
      <sec id="sec-2-1">
        <title>A. The LBP operator</title>
        <p>
          LBP were defined in [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ] as a local structural operator,
operating on the periphery of a circular neighborhood. LBP
are encoded as integers, which in binary notation would map
each sample on the periphery to a binary digit. As can be
seen in Fig. 1 and (2), LBP are defined by the radius of the
circular neighborhood and the number of pixels sampled on
the periphery. The sampling neighborhood Nr;b is formally
defined in (1).
        </p>
        <p>8n; : n 2 [0::b</p>
        <p>1] ^
8f (x1; x2) : R2 =) f0; 1g
= (n
2
)=b
Nr;b(I(x; y); n) = I(x + sin( ) r; y + cos( ) r)</p>
        <p>When defined on grayscale images, LBP are obtained by
thresholding each pixel on the periphery by the central pixel.
Because we worked on binary images as input, a lot more
operations than greater or equal (thresholding) were possible
as a binary operation. We generalized our definition of the LBP
in (2), to consider the boolean operator marked as f a third
defining characteristic of the LBP operator LBPr;b;f along
with the radius r and the number of samples b.</p>
        <p>We took into account several factors for selecting the
appropriate LBP binary operator. In what concerns the bit
count, a bit-count of 8 presents us with many benefits.
Implementation wise, the LBP transform is an image that uses
one byte per pixel. Its histogram has 256 bins providing
a high feature-vector dimentionality and good discriminative
properties. Additionally, containing the distinct LBP count to
256, guaranties highly representative sampling in relatively
small surfaces of text.</p>
      </sec>
      <sec id="sec-2-2">
        <title>B. The LBP function</title>
        <p>While LBP are traditionally derived from grayscale images,
when dealing with text, its better to use binarized text images
as input, thus avoiding all information coming from the text
background. We considered many different binary operations
and chose the binary operator ”equals” (3) as f () in (2 ).
f (xceter; xperiphery) =
1 : xceter = xperiphery
0 : xceter 6= xperiphery
(3)
”Equals” as a boolean function on an image means true for any
background pixel in the peripheral neighborhood of a
background pixel, true for any foreground pixel in the peripheral
neighborhood of a foreground pixel, and false for everything
else. When using the ”equals” function as the binary function
in a 8 bit-count LBP, all pixels with only foreground and
(2)</p>
      </sec>
      <sec id="sec-2-3">
        <title>C. Redundant Features</title>
        <p>(a)
(b)
only background have an LBP value of 255. By suppressing
(ignoring) the 255 bin, we make the LBP histogram surface
invariant. All occurrences left in the histogram represent the
pixels in the border between foreground and background. The
core of the feature set comprises of the 255 histogram bins
normalized to a sum of 1. This normalization renders the
features derived from the histogram invariant to the number
of signal pixels in the image.</p>
        <p>Having the normalized 255 bins from the histogram as
the core of the feature set, we calculate some redundant
features that will amplify some aspects of the LBP we consider
significant in the writer identification task. Our goal is to have
a feature-set discriminative enough to work well with naive
classifiers such as nearest neighbor or, even more, classify
writers by clustering the samples without any training.</p>
        <p>The first redundant feature group we use, is edge
participation. We consider each pattern to have a specific probability of
belonging to an edge of a specific orientation; from now on we
call that contribution. The sum of the number of occurrences
of each pattern, multiplied by its contribution factor makes up
the oriented edge occurrences. In Fig. 2a all top-edge patterns
can be seen along with their probability, in 2b we can see the
patterns of the top-left-edge patterns and their probabilities
which are derived from the top-edge patterns by rotating them
counter-clock-wise. By rotating the contributing patterns of
the top-edge, we can obtain the contributing patterns of all
eight edge-orientations. We also add the more general
edgeorientations: horizontal, vertical, ascending, and descending as
separate features which are calculated as the sum of the
respective pair of edge-orientations. In the end we calculate the two
final aggregations of perpendicular and diagonal, which are the
sum of horizontal and vertical and respectively ascending and
descending. In total we obtain 14 edge-features, which we then
normalize to a sum of 1. One of our aims when introducing
these redundant features is to enhance characteristics that have
been associated with writer identification such as text slant.</p>
        <p>
          The second redundant feature-group we implemented are
the rotation invariant hashes. We grouped all patterns, so
that each pattern in a group can be transformed in to any
other pattern in that group by rotating. When having an 8
sample LBP, the distinct rotation invariant patterns are 36
in total [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. Some pattern groups contain only one pattern
eg. pattern 0, while other groups contain up to 8 patterns,
such as all one bit true patterns 1,2,4,8,16,32,64,128. We
took the number of occurrences for each group in the input
image and normalized them to a sum of 1, thus providing
36 rotation invariant features. A complementary feature-group
to the rotation invariant patterns is what we named rotation
phase. For each group of rotation invariant features, we took
the minimum, with respect to the numeric value, pattern in
the group and designated it as group-hash. The number of
clockwise rotations each pattern needs to become its
groupshash, is what we call the rotation phase. By definition, the
distinct phases in an LBP image, are as many as the number
of samples of the LBP. The frequency of all phases normalized
to the sum of 1, provides us with 8 more redundant features
that are complementary to the rotation invariant hashes.
        </p>
        <p>A third group of redundant features we introduced to our
feature-set is what we called beta-function as defined in (4)
along with the bit-count of every pattern.</p>
        <p>8n 2 [1::bitcount]
8lbp 2 [0::2bitcount 1]
d(lbp; n) =
(lbp) =</p>
        <p>X d(lbp; n)
n
( 1 : bit n is set in lbp^</p>
        <p>bit n 1 is not set in lbp
0 : otherwise
(4)
When the sample count is 8, the function, has up to 5 distinct
values. The histogram of the function (5 bins) normalized to
a sum of 1 and the histogram of the bit-count of every pattern
normalized to 1 as well, are the last redundant feature-group
we defined. The function becomes an important feature when
the LBP radius is greater than pen stroke thickness. In those
situations, e.g. a count of one, would indicate the ending
of a line, and a count of three or four would indicate lines
crossing.</p>
        <p>If we put it all together, we have 255 bins of the histogram,
plus 36 rotation invariant features, plus 8 rotation phase
features, plus 14 edge-features, plus 5 function features, plus
9 sample-count features, makes a total of 327 features; these
are the proposed feature-set. The redundant features make the
features well suited for naive classifiers. By setting the 255
histogram bin to 0, the feature set ignores all non signal areas
in the image. The normalization of all bins to a sum of 1, as
well as the nullification of the last bin, renders our feature set
invariant with respect to non signal areas (white).</p>
      </sec>
      <sec id="sec-2-4">
        <title>D. The Classifier</title>
        <p>Once we transform a given set of images into feature
vectors, we can either use them as a nearest neighbor classifier
or perform clustering on them. While clustering seems to be
a more generic approach, it is constrained by the need to
process all samples at the same time. Such a constraint makes
the clustering approach very well suited for research purposes
but hard to match any real world scenarios. The construction
of the classifier consists of four steps. In the first step, we
extract the image features. In the second step, we rebase the
features along the principal components of a given dataset by
performing principal components analysis. This step might, in
a very broad sense of the term, be considered training because
our method acquires information from a given dataset. In the
third step we scale the rebased features by a scaling vector
which was defined by evolutionary optimization on the
trainset. The optimization process is also performed on a given
dataset and should also be considered as a training stage. While
it is not required, it makes more sense that both training steps
are performed on the same dataset. The fourth and last step
is to calculate the L1 norm on the scaled and rebased feature
vectors. Steps two and three can be combined in to a linear
operation on the feature space and in many aspects should be
viewed as a statistically derived heuristic matrix. Our classifier,
as was implemented, has two inputs, a subject dataset and a
reference dataset. The output consists of a table where each
row refers to a sample in the subject dataset and contains all
samples in the reference dataset ranked by similarity to the
specific sample. When benchmarking classification rates of our
method, we can simply run our classifier with an annotated
dataset as both object dataset and reference dataset. In this
case, the first column contains the object sample and the
second column contains the most similar sample in the dataset
other than its self. The rate at which the classes in the first
column agree to the classes in second column, is the nearest
neighbor classification rate.</p>
      </sec>
      <sec id="sec-2-5">
        <title>E. Scale Vector Optimisation</title>
        <p>
          Describing in detail the optimization process of the scaling
vector would go beyond the scope of this paper. In brief
we optimized using an evolutionary algorithm. We used as
input the 125 most prominent components of the features
and the id of the writer of each sample. We optimized using
the ICHFR 2012 writer identification competition dataset [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]
which contains 100 writers contributing 4 writing samples
each. Individuals of the algorithm were modeled as vector
of continuous scaling factors for each feature in the feature
space. The fitness function was based on the classification rate
a nearest neighbor classifier obtains when the feature space
is scaled by each individual. The stoping criteria was set to
2000 generations, and each generation had 20 individuals.
Suitable parents were determined by the rank they obtained
in the generation.
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>III. EXPERIMENTAL PROCEDURE</title>
      <p>
        In order to have a proper understanding of our methods
performance, its robustness, and its limitations, we conducted
a series of experiments. We used two datasets for our
experiments: the dataset from the ICDAR 2011 writer identification
contest [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], hereafter 2011 dataset and the dataset from the
ICHFR 2012 writer identification challenge [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], hereafter
2012 dataset. The 2011 dataset has 26 writers contributing
samples in Greek, English, German, and French with 8 samples
per writer. The 2012 dataset has 100 writers, contributing
samples in Greek and English with 4 samples per writer.
While the 2011 dataset was given as the train set for the 2012
contest, we used them in the opposite manner. In order to
avoid overfitting during the optimization step, we deemed the
”harder” dataset, containing more classes and less samples per
class, was better suited for training.
      </p>
      <sec id="sec-3-1">
        <title>A. Performance</title>
        <p>
          As previously described, our method consist of four stages:
feature extraction, principal components analysis, scaling
vector optimization, and L1 distance estimation. Steps two and
three require a training dataset, while steps one and four are
totally independent of any data. In TABLE I analytical scores
of our method in various modalities can be seen. Apart from
the nearest neighbor accuracy we also add the hard TOP-N
and soft TOP-N criteria [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ], [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. The soft TOP-N criterium
is calculated by estimating the percentage of samples in the
test set that have at least one sample of the same class in their
N nearest neighbors. The hard TOP-N criterium is calculated
by estimating the percentage of samples in the test set that have
only samples of the same class in their N nearest neighbors.
More in detail, in TABLE I we can see various versions of
our method and their performance as well as some state of
the art methods for reference. Methods Tsinghua, MCS-NUST
and Tebessa [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] are the top performing methods from the
ICDAR 2011 writer identification contest. We must point out
that our method had a vastly superior train set, consisting of
400 samples, and we had access to the test set while working.
Our method has two parts that were optimized on our train
set, the 2012 dataset. The first is the principal components of
the train set and the second is the scaling of the feature space.
No PC, No train is the raw feature space without any training,
just the features in an L1 nearest neighbor setup. PC, No train
is the feature space rebased along the principal components of
the the train set in a L1 nearest neighbor setup. PC, Train is
the feature space rebased along the principal components of
the the train-set and scaled along the optimized vector in a
L1 nearest neighbor setup. As we can see our method almost
reaches the overall performance of the state of the art when
it incorporates the full trained heuristics but it also provides
very good results in its untrained form.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>B. Qualitative Experiments</title>
        <p>Apart from providing a comprehensive accuracy score that
is comparable to other methods, in order to describe the
strength and limitations of our method, we performed a series
of experiments that simulate frequently appearing distortions
to the data.</p>
        <p>1) Rotation: Text orientation, is a text image characteristic
that is definitely affected by the writer. Under controlled
homogeneous experimental conditions of data acquisition, text
orientation should depend only on the writer. Quite often in
real life scenarios we have no way of knowing whether an
image has been rotated or not and to which extent. One of the
important characteristics of a writer identification system is the
robustness against moderate rotation. We address this issue by
an experiment where we try to recognize samples of a dataset
with rotated versions of the database. More specifically we
took the 2012 dataset and we rotated its samples by 1 from
20 to 20 . We obtained our measurements by classifying the
original 2012 dataset with the the rotated versions. In Fig. 3
the rotation sensitivity of our method can be demonstrated.
Two different measurements can be seen. The first one, noted
as Sample Self Recognition, is the the nearest neighbor
including the test sample. Sample Self Recognition rate will
be by definition 100% when no rotation occurs. The second
measurement, marked as Nearest Neighbor is the accuracy
of nearest neighbor excluding the first occurrence. Nearest
Neighbor is by definition the accuracy when no rotation occurs.
As can be seen in Fig. 3 our method demonstrates some
rotation tolerance from 5 to +5 with sustainable accuracy
rates, but performance drops significantly beyond this limit1.
It is also worth noticing the fact that 1 and +1 rotations
perform slightly worst than 2 to +2 ; a possible explanation
for this could be aliasing phenomena.</p>
        <p>2) Downsampling: As we stated previously, in most real
world scenarios, the sampling resolution will be known to
a writer identification system, but not always controlled as
sometimes the data are acquired by external sources or at
different times. We devised an experiment that demonstrate
the behavior and limitations of our method in what concerns
the resolution. We took the ICDAR 2011 Writer Identification
dataset and rescaled it to various scales from 100% down to
10%. As can be seen in Fig. 4 we obtained three measurements.
The first, marked as Self Recognition Unscaled Sample, is
the nearest neighbor when classifying the initial dataset with
the subsampled dataset as a database. The second, marked
as Nearest Neighbor Unscaled Sample, is the second nearest
neighbor when classifying the initial dataset with the
subsampled dataset as a database. We presume that the first nearest
neighbor will always be the same sample in different scales
and therefore disregard it for this measurement. The third
measurement, named Nearest Neighbor Scaled Sample, is the
accuracy of the second nearest neighbor when classifying the
scaled dataset with the scaled dataset a database. The first two
measurements describe the sensitivity our method has in
comparing samples of different sampling resolution and therefore
scale as well, while the third measurement demonstrates how
well our method would work on datasets of lower resolution.
We should also point out that the optimization process was
performed on the original resolution. As we expected and can
be seen in Fig. 4, we find that our method has no tolerance in
comparing samples from different sampling rates. We can also
1samples rotated by more than 5 could be manually corrected during
sample aquisition
conclude that our method has tolerance to lower than standard
resolutions, but benefits mostly from higher resolutions. The
out of the norm measurement in Nearest Neighbor Scaled
Sample posed us with a puzzle. The most probable explanation
is that it is related to aliasing but is worth investigating more.</p>
        <p>3) Removing Graphemes: A very important characteristic
of writer identification methods is how much text is required
to in order to reach the claimed accuracy. We conducted an
experiment to answer specifically this question. Our strategy
was to create group datasets that vary only on the amount
of signal (text) and then compare results on these datasets.
As the primary dataset we took the ICDAR 2011 writer
identification dataset, because it provides us with relatively
large text samples. In order to quantify the available signal,
we took the 2011 dataset and for each image in the dataset,
we produced 20 images with different amounts of connected
components from the original image. Due to the very high
locality of our feature set, the fact that we removed connected
components instead of text lines should be negligible and
at the same time it gave us quite higher control over the
signal quantity. As can be seen in Fig. 5 the results are quite
surprising. Instead of having a gradual drop in performance,
the performance is unaffected down to 30% of the graphemes,
bellow that point, performance drops linearly.</p>
        <p>4) Writer vs Writing Style: We submitted an earlier version
of our method to the SigWiComp2013 competition. The goal
of the writer identification part of the competition, is to
measure the performance of writer identification systems, when the
handwriting style has been altered. A sample dataset was made
available by the organizers of the competition. The dataset
contained 55 writers contributing 3 text samples each and
each sample written a different writing style. Having access
to the sample dataset, we performed a simple experiment to
determine whether our features encapsulate writer biometrical
information or simply the writing style. We separated the
dataset of 165 samples in to left and right halves. We then
performed a pair matching of the left halves to the right halves
based on the nearest neighbor classification. We obtained two
measurements, first the percentage of left-samples having an
assigned right-sample written by the same writer (55 classes),
and second the percentage of left-samples having the
specific sample’s complementary right-half as the nearest
neighbor (165 classes). The writer identification rate was 87.27%,
while the specific sample recognition rate was 86.06%. By
definition the writer identification rate is greater or equal to
the sample recognition rate. We performed a one tail t-test
on the results on 165 sample-classifications and obtained a
pvalue of 0.3734, which by all standards make the
recognitionrates indistinguishable. This experiment indicates that for our
method any two samples written in different writing styles
are as different regardless of whether they were written by
the same individual or not. From a forensic perspective, these
measurements imply that our method does not distinguish
between disguised writing style and natural writing style.</p>
        <p>IV.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>DISCUSSION</title>
      <p>
        In this paper we presented a powerful feature set that
summarizes any texture on a binary image as a vector of 327.
We use our feature extraction method to produce a database
from any given dataset with handwriting samples and use it as a
nearest neighbor classifier. In order to improve our classifier we
performed PCA on a specific dataset and linearly transformed
the feature space. We also scaled the features by a scaling
vector in order to increase the impact of the features that
contribute to correct classifications on our test set. Both these
improvements can be combined in to single matrix with which
we multiply all feature vectors. This single matrix should be
viewed as a heuristic matrix statistically derived from the 2012
dataset. It is also valid to think of this matrix as the result
of a supervised learning process. The idea is to calculate
this matrix once per type of texture we want to classify. In
the context of western script handwriting, we obtained the
matrix from the 2012 dataset and used it in benchmarking
our method on the 2011 dataset, our qualitative experiments,
and our submission to SigWiComp2013 [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. When comparing
the experimental results to the state of the art, we can not
obtain a perfectly fair comparison. The state of the art methods
participated in a blind competition with a very small
trainset, although we could maybe assume that participants had
access to larger third-party datasets as well. Since datasets
of competitions are published after the competitions, the only
way to have a perfectly fair comparison to the state of the art
is to participate in those competitions. A comparison of the
untrained classifier (96.63%) to the state of the art (99.5%)
is quite unfair towards our method. On the other hand, a
comparison of our trained classifier (98.56%) to the state of
the art (99.5%) is a bit unfair towards the state of the art. In
the authors opinion, a fair comparison would be a lot closer to
the trained classifier than to the untrained. The performance of
the untrained classifier demonstrates clearly the potency of our
feature set. The qualitative experiments were not performed
with forensics in mind, except for the last one, writer vs
writing style. In writer vs writing style we tried to determine
the extent to which our feature set can deal with disguising
writers; the quick answer is, no our method can not deal with
disguising writers. There are many subtleties in the conclusions
that can be drawn from the writer vs writing style experiment
about what phenomena is that our features model. One could
even say that our method is more about texture similarity than
about writer similarity; assuming there are biometric features
in handwriting, the proposed feature set does not seem to
encapsulate them. From a software engineering perspective the
approach of treating writer identification as a distance metric
instead of a classifier [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] seems more efficient and modular, it
allows for simplification and standardization of benchmarking.
The fact that the proposed features encapsulate no structural
information what so ever, makes them a very good candidate
for fusion with other feature sets.
      </p>
    </sec>
    <sec id="sec-5">
      <title>ACKNOWLEDGMENT</title>
      <p>The first author of this paper would like to thank Georgios
Louloudis for his precious insights on the subject of writer
identification and performance evaluation. The first author
would also like to thank Muhammad Imran Malik for his
effort and assistance in the participation of this method on
the SigWiComp2013 competition.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>T.</given-names>
            <surname>Ojala</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pietikainen</surname>
          </string-name>
          , and T. Maenpaa, “
          <article-title>Multiresolution gray-scale and rotation invariant texture classification with local binary patterns,”</article-title>
          <source>Pattern Analysis and Machine Intelligence</source>
          , IEEE Transactions on, vol.
          <volume>24</volume>
          , no.
          <issue>7</issue>
          , pp.
          <fpage>971</fpage>
          -
          <lpage>987</lpage>
          ,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>T.</given-names>
            <surname>Ahonen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hadid</surname>
          </string-name>
          , and M. Pietika¨inen, “
          <article-title>Face recognition with local binary patterns,” in Computer Vision-ECCV</article-title>
          <year>2004</year>
          . Springer,
          <year>2004</year>
          , pp.
          <fpage>469</fpage>
          -
          <lpage>481</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>G.</given-names>
            <surname>Zhao and M. Pietikainen</surname>
          </string-name>
          , “
          <article-title>Dynamic texture recognition using local binary patterns with an application to facial expressions,”</article-title>
          <source>Pattern Analysis and Machine Intelligence</source>
          , IEEE Transactions on, vol.
          <volume>29</volume>
          , no.
          <issue>6</issue>
          , pp.
          <fpage>915</fpage>
          -
          <lpage>928</lpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. X.</given-names>
            <surname>Han</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Yan</surname>
          </string-name>
          , “
          <article-title>An hog-lbp human detector with partial occlusion handling</article-title>
          ,” in
          <source>Computer Vision</source>
          ,
          <source>2009 IEEE 12th International Conference on. IEEE</source>
          ,
          <year>2009</year>
          , pp.
          <fpage>32</fpage>
          -
          <lpage>39</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>G. A.</given-names>
            <surname>Dean</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I. W.</given-names>
            <surname>Kelly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. H.</given-names>
            <surname>Saklofske</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Furnham</surname>
          </string-name>
          , “
          <article-title>Graphology and human judgment</article-title>
          .”
          <year>1992</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Furnham</surname>
          </string-name>
          , “
          <article-title>Write and wrong: The validity of graphological analysis,” The Hundreth Monkey</article-title>
          and
          <article-title>Other Paradigms of the Paranormal</article-title>
          , pp.
          <fpage>200</fpage>
          -
          <lpage>205</lpage>
          ,
          <year>1991</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bensefia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Paquet</surname>
          </string-name>
          , and L. Heutte, “
          <article-title>Handwritten document analysis for automatic writer recognition,” Electronic letters on computer vision and image analysis</article-title>
          , vol.
          <volume>5</volume>
          , no.
          <issue>2</issue>
          , pp.
          <fpage>72</fpage>
          -
          <lpage>86</lpage>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Z.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>You</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y. Y.</given-names>
            <surname>Tang</surname>
          </string-name>
          , “
          <article-title>Writer identification of chinese handwriting documents using hidden markov tree model,” Pattern Recognition</article-title>
          , vol.
          <volume>41</volume>
          , no.
          <issue>4</issue>
          , pp.
          <fpage>1295</fpage>
          -
          <lpage>1307</lpage>
          ,
          <year>2008</year>
          . [Online]. Available: http://www.sciencedirect.com/science/article/pii/S0031320307004037
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>L.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>You</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Gao</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Tang</surname>
          </string-name>
          , “
          <article-title>Wavelet domain local binary pattern features for writer identification,” in Pattern Recognition (ICPR</article-title>
          ),
          <year>2010</year>
          20th International Conference on. IEEE,
          <year>2010</year>
          , pp.
          <fpage>3691</fpage>
          -
          <lpage>3694</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>C.</given-names>
            <surname>Bird</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Found</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Rogers</surname>
          </string-name>
          , “
          <article-title>Forensic document examiners skill in distinguishing between natural and disguised handwriting behaviors</article-title>
          ,
          <source>” Journal of forensic sciences</source>
          , vol.
          <volume>55</volume>
          , no.
          <issue>5</issue>
          , pp.
          <fpage>1291</fpage>
          -
          <lpage>1295</lpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M. I.</given-names>
            <surname>Malik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Liwicki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Alewijnse</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Ohyama</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Blumenstein</surname>
          </string-name>
          , and
          <string-name>
            <given-names>B.</given-names>
            <surname>Found</surname>
          </string-name>
          , “
          <article-title>Signature verification and writer identification competitions for on- and offline skilled forgeries (sigwicomp2013),”</article-title>
          <source>in 12th Int. Conf. on Document Analysis and Recognition</source>
          , Washigton,
          <string-name>
            <surname>DC</surname>
          </string-name>
          , USA,
          <year>2013</year>
          , p.
          <source>n.A.</source>
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>G.</given-names>
            <surname>Louloudis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Stamatopoulos</surname>
          </string-name>
          , and
          <string-name>
            <given-names>B.</given-names>
            <surname>Gatos</surname>
          </string-name>
          , “
          <article-title>Icdar 2011 writer identification contest,” in Document Analysis and Recognition (ICDAR</article-title>
          ),
          <source>2011 International Conference on. IEEE</source>
          ,
          <year>2011</year>
          , pp.
          <fpage>1475</fpage>
          -
          <lpage>1479</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>G.</given-names>
            <surname>Louloudis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Gatos</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N.</given-names>
            <surname>Stamatopoulos</surname>
          </string-name>
          , “
          <article-title>Icfhr 2012 competition on writer identification challenge 1: Latin/greek documents,” in Frontiers in Handwriting Recognition (ICFHR</article-title>
          ),
          <source>2012 International Conference on. IEEE</source>
          ,
          <year>2012</year>
          , pp.
          <fpage>829</fpage>
          -
          <lpage>834</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>D.</given-names>
            <surname>Chawki</surname>
          </string-name>
          and S.
          <string-name>
            <surname>-M. Labiba</surname>
          </string-name>
          , “
          <article-title>A texture based approach for arabic writer identification and verification</article-title>
          ,”
          <source>in Machine and Web Intelligence (ICMWI)</source>
          ,
          <source>2010 International Conference on. IEEE</source>
          ,
          <year>2010</year>
          , pp.
          <fpage>115</fpage>
          -
          <lpage>120</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>