<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>K. Nowicki and T. A. B. Snijders, “Estimation and prediction for
stochastic blockstructures,” Journal of the American Statistical Asso-
ciation</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>An agent-driven semantical identifier using radial basis neural networks and reinforcement learning</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Christian Napoli, Giuseppe Pappalardo, and Emiliano Tramontana Department of Mathematics and Informatics University of Catania</institution>
          ,
          <addr-line>Viale A. Doria 6, 95125 Catania</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>1990</year>
      </pub-date>
      <volume>96</volume>
      <issue>455</issue>
      <fpage>1077</fpage>
      <lpage>1087</lpage>
      <abstract>
        <p>-Due to the huge availability of documents in digital form, and the deception possibility raise bound to the essence of digital documents and the way they are spread, the authorship attribution problem has constantly increased its relevance. Nowadays, authorship attribution, for both information retrieval and analysis, has gained great importance in the context of security, trust and copyright preservation.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>I. INTRODUCTION</title>
      <p>Nowadays, the automatic attribution of a text to an author,
assisting both information retrieval and analysis, has become
an important issue, e.g. in the context of security, trust and
copyright preservation. This results from the availability of
documents in digital form, and the raising deception
possibilities bound to the essence of the digital reproducible contents,
as well as the need for new mechanical methods that can
organise the constantly increasing amount of digital texts.</p>
      <p>During the last decade only, the field of text classification
and attribution has undergone new developement due to the
novel availability of computational intelligence techniques,
such as natural language processing, advanced data mining
and information retrieval systems, machine learning and
artificial intelligence techniques, agent oriented programming,
etc. Among such techniques, Computer Intelligence (CI) and
Evolutionary Computation (EC) methods have been largely
used for optimisation and positioning problems [1], [2]. In [3],
agent driven clustering has been used as an advanced solution
for some optimal management problems, whereas in [4] such
problems are solved for mechatronical module controls. Agent
driven artificial intelligence is often used in combination with
advanced data analysis techniques in order to create intelligent
control systems [5], [6] by means of multi resolution
analysis [7]. CI and parallel analysis systems have been proposed
in order to support developers, as in [8], [9], [10], [11],
where such a classification and analysis was applied to assist
refactoring in large software systems [12], [13], [14], [15].</p>
      <p>Moreover, CI and techniques like neural networks (NNs)
have been used in [16], [17] in order to model electrical
networks and the related controls starting by classification
strategies, as well as for other complex physical systems
by using several kinds of hybrid NN-based approaches [18],
[19], [20], [21]. All the said works use different forms of
agent-based modeling and clustering for recognition purposes,
and these methods efficiently perform very challenging tasks,
where other common computational methods failed or had
low efficiency or, simply, resulted as inapplicable due to
complicated model underlying the case study. In general,
agent-driven machine learning has been proven as a promising
field of research for the purpose of text classification, since
it allows building classification rules by means of automatic
learning, taking as a basis a set of known texts and trying to
generalise for unknown ones.</p>
      <p>While machine learning and NNs are a very promising
field, the effectiveness of such approaches often lies on the
correct and precise preprocessing of data, i.e. the definition
of semantic categories, affinities and rules used to generate a
set of numbers characterising a text sample, to be successively
given as input to a classifier. Typical text classification, e.g.
by using NNs, takes advantage of topics recognition, however
results are seldom appropriate when it comes to classify people
belonging to the same social group or who are involved
in a similar business (e.g. the classification of: texts from
different scientists in the same field of research, the politicians
belonging to the same party, texts authored by different people
using the same technical jargon).</p>
      <p>In our approach we devise a solution for extracting from the
analysed texts some characteristics that can express the style of
a specific author. Obtaining this kind of information abstraction
is crucial in order to create a precise and correct classification
system. On the other hand, while data abound in the context
of text analysis, a robust classifier should rely on input sets
that are compact enough to be apt to the training process.
Therefore, some data have to reflect averaged evaluations that
concern some anthropological aspects such as the historical
period, or the ethnicity, etc. This work satisfies the above
conditions of extracting compact data from texts since we use
a preprocessing tool for word-grouping and time-period related
analysis of the common lexicon. Such a tool computes a bias
reference system for the recurrence frequency of the word used
in the analysed texts. The main advantage of this choice lies in
the generality of the implemented semantical identifier, which
can be then applied to different contexts and lexical domains
without requiring any modification. Moreover, in order to have
continuous updates or complete renewals of the reference data,
a statically trained NN would not suffice to the purpose of the
work. For these reasons, the developed system is able to
selfcorrect by means of continuous learning reinforcement. The
proposed architecture also diminishes the human intervention
over time thanks to its self-adaption properties.</p>
      <p>Our solution comprises three main collaborating agents:
the first for preprocessing, i.e. to extract meaningful data from
texts; the second for classification by means of a proper Radial
Basis NN (RBNN); and finally, one for adapting by means of
a feedforward NN.</p>
      <p>The rest of this paper is as follows. Section II gives
the details of the implemented preprocessing agent based on
lexicon analysis. Section III describes the proposed classifier
agent based on RBNNs, our introduced modifications and
the structure of the reinforcement learning agent. Section IV
reports on the performed experiments and the related results.
Finally, Section V gives a background of the existing related
works, while Section VI draws our conclusions.</p>
      <p>II. EXTRACTING SEMANTICS FROM LEXICON</p>
      <p>Figure 1 shows the agents for our developed system: a
preprocessing agent extracts characteristics from given text
parts (see text database in the Figure), according to a known
set of words organised into groups (see reference database);
a RBPNN agent takes as input the extracted characteristics,
properly organised, and performs the identification on new
data, after appropriate training. An additional agent, dubbed
adaptive critic, shown in Figure 6, dynamically adapts the
behaviour of the RBPNN agent when new data are available.</p>
      <p>Firstly, preprocessing agent analyses a text given as input
by counting the words that belong to a priori known groups of
mutually related words. Such groups contain words that pertain
to a given concern, and have been built ad-hoc and according
to the semantic relations between words, hence e.g. assisted
by the WordNet lexicon1.</p>
    </sec>
    <sec id="sec-2">
      <title>1http://wordnet.princeton.edu</title>
      <p>Algorithm 1: Find the group a word belongs to and count
occurrences</p>
      <p>Start,
Import a speech into T ext,
Load dictionary into W ords,
Load group database into Groups,
thisW ord = Text.get();
while thisW ord do
thisGroup = Words.search(thisW ord);
if !thisGroup then</p>
      <p>Load a different Lexicon;
if Lexicon.exist(thisW ord) then</p>
      <p>W ords.update();</p>
      <p>Groups.update();
end
else
end</p>
      <p>break;
end
while Text.search(thisW ord) do</p>
      <p>Groups.count(thisGroup);
end
thisW ord = Text.get();
end
Export W ords and Groups,
Stop.</p>
      <p>The fundamental steps of the said analysis (see also
Algorithm 1) are the followings:
1)
2)
3)
4)
5)
6)
7)
8)
9)
import a single text file containing the speech;
import word groups from a predefined database, the
set containing all words from each group is called
dictionary;
compare each word on the text with words on the
dictionary;
if the word exists on the dictionary then the relevant
group is returned;
if the word has not been found then search the
available lexicon;
if the word exists on the lexicon then the related
group is identified;
if the word is unkown, then a new lexicon is loaded
and if the word is found then dictionary and groups
are updated;
search all the occurrences of the word in the text;
when an occurrence has been found, then remove it
from the text and increase the group counter.</p>
      <p>Figure 2 shows the UML class diagram for the software
system performing the above analysis. Class Text holds a text
to be analysed; class Words represents the known dictionary,
i.e. all the known words, which are organised into groups given
by class Groups; class Lexicon holds several dictionaries.</p>
    </sec>
    <sec id="sec-3">
      <title>III. THE RBPNN CLASSIFIER AGENT</title>
      <p>For the work proposed here, we use a variation on Radial
Basis Neural Networks (RBNN). RBNNs have a topology
similar to common FeedForward Neural Networks (FFNN) with
BackPropagation Training Algorithms (BPTA): the primary
get()
get()
exist()</p>
      <p>Filter
service()
search()
update()</p>
      <p>Groups
search()
count()
update()
difference only lies in the activation function that, instead of
being a sigmoid function or a similar activation function, is a
statistical distribution or a statistically significant mathematical
function. The selection of transfer functions is indeed decisive
for the speed of convergence in approximation and
classification problems [22]. The kinds of activation functions used
for Probabilistic Neural Networks (PNNs) have to meet some
important properties to preserve the generalisation abilities of
the ANNs. In addition, these functions have to preserve the
decision boundaries of the probabilistic neural networks.</p>
      <p>The selected RBPNN architecture is shown in Figure 4 and
takes advantage from both the PNN topology and the Radial
Basis Neural Networks (RBNN) used in [23].</p>
      <p>Each neuron performs a weighted sum of its inputs and
passes it through a transfer function f to produce an output.
This occurs for each neural layer in a FFNN. The network
can be perceived as a model connecting inputs and outputs,
with the weights and thresholds being free parameters of the
model, which are modified by the training algorithm. Such
networks can model functions of almost arbitrary complexity
with the number of layers and the number of units in each
layer determining the function complexity. A FFNN is capable
to generalise the model, and to separate the input space in
various classes (e.g. in a 2D variable space it is equivalent to
the separation of the different semi-planes). In any case, such
a FFNN can only create a general model of the entire variable
space, while can not insert single set of inputs into categories.</p>
      <p>On the other hand, a RBNN is capable of clustering
the inputs by fitting each class by means of a radial basis
function [24], while the model is not general for the entire
variable space, it is capable to act on the single variables (e.g.
in a 2D variable space it locates closed subspaces, without any
inference on the remaining space outside such subspaces).</p>
      <p>Another interesting topology is provided by PNNs, which
are mainly FFNNs also functioning as Bayesian networks
with Fisher Kernels [25]. By replacing the sigmoid activation
function often used in neural networks with an exponential
function, a PNN can compute nonlinear decision boundaries
approaching the Bayes optimal classification [26]. Moreover,
a PNN generates accurate predicted target probability scores
with a probabilistic meaning (e.g. in the 2D space it is
equivalent to attribute a probabilistic score to some chosen
points, which in Figure 3 are represented as the size of the
points).</p>
      <p>Finally, in the presented approach we decided to combine
FFNN</p>
      <p>RBNN
PNN</p>
      <p>Our RBPNN
the advantages of both RBNN and PNN using the so called
RBPNN. The RBPNN architecture, while preserving the
capabilities of a PNN, due to its topology, then being capable
of statistical inference, is also capable of clustering since the
standard activation functions of a PNN are substituted by radial
basis functions still verifying the Fisher kernel conditions
required for a PNN (e.g. such an architecture in the 2D variable
space can both locate subspace of points and give to them a
probabilistic score). Figure 3 shows a representation of the
behaviour for each network topology presented above.</p>
      <sec id="sec-3-1">
        <title>A. The RBPNN structure and topology</title>
        <p>In a RBPNN both the input and the first hidden layer
exactly match the PNN architecture: the input neurones are
used as distribution units that supply the same input values
to all the neurones in the first hidden layer that, for historical
reasons, are called pattern units. In a PNN, each pattern unit
performs the dot product of the input pattern vector v by a
weight vector W(0), and then performs a nonlinear operation
on the result. This nonlinear operation gives output x(1) that
is then provided to the following summation layer. While a
common sigmoid function is used for a standard FFNN with
BPTA, in a PNN the activation function is an exponential, such
that, for the j-esime neurone the output is
x(j1) / exp
jjW(0) vjj
2 2
where</p>
        <p>represents the statistical distribution spread.</p>
        <p>The given activation function can be modified or substituted
while the condition of Parzen (window function) is still
satisfied for the estimator N^ . In order to satisfy such a condition
some rules must be verified for the chosen window function in
order to obtain the expected estimate, which can be expressed
as a Parzen window estimate p(x) by means of the kernel K
of f in the d-dimensional space Sd</p>
        <p>n
pn(x) = n1 iP=1 h1dn K
...
...
...
where hn 2 N is called window width or bandwidth parameter
and corresponds to the width of the kernel. In general hn 2 N
depends on the number of available sample data n for the
estimator pn(x). Since the estimator pn(x) converges in mean
square to the expected value p(x) if
lim hpn(x)i
n!1
lim var (pn(x))
n!1
=
=
p(x)
0
where hpn(x)i represents the mean estimator values and
var (pn(x)) the variance of the estimated output with respect
to the expected values, the Parzen condition states that such
convergence holds within the following conditions:
sup K(x)</p>
        <p>x
lim
jxj!1
xK(x)
lim hn
n!1 d
lim nhdn
n!1
&lt;
=
=
=
1
0
0
1
In this case, while preserving the PNN topology, to obtain
the RBPNN capabilities, the activation function is substituted
with a radial basis function (RBF); an RBF still verifies all
the conditions stated before. It then follows the equivalence
between the W(0) vector of weights and the centroids vector of
a radial basis neural network, which, in this case, are computed
as the statistical centroids of all the input sets given to the
network.</p>
        <p>We name f the chosen radial basis function, then the new
output of the first hidden layer for the j-esime neurone is
x(j1) , f
jjv</p>
        <p>W(0)jj
where is a parameter that is intended to control the
distribution shape, quite similar to the used in (1).</p>
        <p>The second hidden layer in a RBPNN is identical to a PNN,
it just computes weighted sums of the received values from the
preceding neurones. This second hidden layer is called indeed
summation layer: the output of the k-esime summation unit is
x(k2) = X Wjkx(j1)</p>
        <p>j
where Wjk represents the weight matrix. Such weight matrix
consists of a weight value for each connection from the j-esime
pattern units to the k-esime summation unit. These summation
P*
(3)
(4)
(5)
(6)
units work as in the neurones of a linear perceptron network.
The training for the output layer is performed as in a RBNN,
however since the number of summation units is very small
and in general remarkably less than in a RBNN, the training
is simplified and the speed greatly increased [27].</p>
        <p>The output of the RBPNN (as shown in Figure 4) is given
to the maximum probability selector module, which effectively
acts as a one-neuron output layer. This selector receives as
input the probability score generated by the RBPNN and
attributes to one author only the analysed text, by selecting
the most probable author, i.e. the one having the maximum
input probability score. Note that the links to this selector are
weighted (with weights adjusted during the training), hence the
actual input is the product between the weight and the output
of the summation layer of the RBPNN.</p>
      </sec>
      <sec id="sec-3-2">
        <title>B. Layer size for a RBPNN</title>
        <p>The devised topology enables us to distribute to different
layers of the network different parts of the classification task.
While the pattern layer is just a nonlinear processing layer, the
summation layer selectively sums the output of the first hidden
layer. The output layer fullfills the nonlinear mapping such as
classification, approximation and prediction. In fact, the first
hidden layer of the RBPNN has the responsibility to perform
the fundamental task expected from a neural network [28].</p>
        <p>In order to have a proper classification of the input dataset,
i.e. of analysed texts to be attributed to authors, the size of the
input layer should match the exact number NF of different
lexical groups given to the RBPNN, whereas the size of the
pattern units should match the number of samples, i.e. analysed
texts, NS .</p>
        <p>The number of the summation units in the second hidden
layer is equal to the number of output units, these should match
the number of people NG we are interested in for the correct
recognition of the speakers (Figure 5).</p>
      </sec>
      <sec id="sec-3-3">
        <title>C. Reinforcement learning</title>
        <p>In order to continuously update the reference database for
our system, a statically trained NN would not suffice for the
purpose of the work. Since the aim of the presented system is
having an expanding database of text samples for classification
and recognition purpose, the agent driven identification should
dynamically follow the changes in such a database. When
a new entry is made then the related feature set and biases
change, it implies that also the RBPNN should be properly
managed in order to ensure a continuous adaptive control for
reinforcement learning. Moreover, for the considered domain
it is desirable that a human supervisor supply suggestions,
expecially when the system starts working. The human activities
are related to the supply of new entries into the text sample
database, and to the removal of misclassifications made by the
RBPNN.</p>
        <p>We used a supervised control configuration (see Figure 6),
where the external control is provided by the actions and
choices of a human operator. While the RBPNN is trained with
a classical backpropagation learning algorithm, it is also
embedded into an actor-critic reinforcement learning architecture,
which back propagates learning by evaluating the correctness
of the RBPNN-made choices with respect to the real word.</p>
        <p>Let be the error function, i.e. = 0 for the results
supported by human verification, or the vectorial deviance for
the results not supported by a positive human response. This
assessment is made by an agent named Critic. We consider
the filtering step for the RBPNN output, to be both: Critic,
i.e. a human supervisor acknowledging or rejecting RBPNN
classifications; or Adaptive critic, i.e. an agent embedding a
NN that in the long run simulates the control activity made by
the human Critic, hence decreasing human control over time.
Adaptive critic needs to learn, and this learning is obtained by
a modified backpropagation algorithm using just as error
function. Hence, Adaptive critic has been implemented by
a simple feedforward NN trained, by means of a traditional
gradient descent algorithm so that the weight modification
wij is
wij =
The f~i is the activation of i-esime neuron, u~i is the i-esime
input to the neurone weighted as
u~i = X wij f~j ( i)
j
(7)
(8)</p>
        <p>The result of the adaptive control determines whether to
continue the training of the RBPNN with new data, and
whether the last training results should be saved or discarded.
At runtime this process results in a continuous adaptive
learning, hence avoiding the classical problem of NN polarisation
and overfitting. Figure 6 shows the developed learning system
reinforcement. According to the literature [29], [30], [31], [32],
straight lines represent the data flow, i.e. training data fed
to the RBPNN, then new data inserted by a supervisor, and
the output of the RBPNN sent to the Critic modules also
by means of a delay operator z 1. Functional modifications
operated within the system are represented as slanting arrows,
i.e. the choices made by a human supervisor (Critic) modify
the Adaptive critic, which adjust the weight of its NN; the
combined output of Critic and Adaptive critic determines
whether the RBPNN should undergo more training epochs and
so modify its weights.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>IV. EXPERIMENTAL SETUP The proposed RBPNN architecture has been tested using several text samples collected from public speeches of different</title>
      <p>Features
DATA
z-1
RBPNN</p>
      <p>Critic
Adaptive
critic
Fig. 6. The adopted supervised learning model reinforcement. Slanting arrows
represent internal commands supplied in order to control or change the status
of the modules, straight arrows represent the data flow along the model, z 1
represents a time delay module which provides 1-step delayed outputs.
people both from the present and the past era. Each text sample
has been given to the preprocessing agent that extract some
characteristics (see Section II), then such results have been
given to the classification agent. The total number of text
samples was 344, and we used 258 of them for training the
classification agent and 86 for validation. The text samples,
both for training and validation, were from different persons
that have given a speech (from A. Einstein to G. Lewis), as
shown in Figure 7.</p>
      <p>Given the flexible structure of the implemented learning
model, the word groups are not fixed and can be modified,
added or removed over time by an external tuning activity. By
using the count of words in a group, instead of a word-by-word
counts, the multi-agent system realises a statistically driven
classifier that identifies the main semantic concerns regarding
the text samples, and then, attributes such concerns to the most
probable person.</p>
      <p>The relevant information useful in order to recognise the
author of the speech is usually largely spread over a certain
number of word groups that could be indication of the cultural
extraction, heritage, field of study, professional category etc.
This implies that we can not exclude any word group, a priori,
while the RBPNN could learn to automatically enhance the
relevant information in order to classify the speeches.</p>
      <p>Figure 7-left shows an example of the classifier
performances for results generated by the RBPNN (before the filter
implemened by the probabilistic selector). Since the RBPNN
results have a probability between 0 and 1, then the shown
performance is 0 when a text was correctly attributed (or
not attributed) to a specific person. Figure 7-right shows the
performances of the system when including the probabilistic
selector. For this case, a boolean selection is involved, then
correct identifications are represented as 0, false positive
identifications as 1 (black marks), and missed identifications
as +1 (white marks).</p>
      <p>For validation purposes, Figure 7-(left and right) shows
results according to e:
e = y
y~
(9)
where e identifies the performance, y~ the classification result,
and y the expected result. Lower e (negative values) identify
an excess of confidence in the attribution of a text to a person,
while greater e (positive values) identify a lack of confidence
in that sense.
and classes, given a subset of already selected features used as
a classifier independent criterion for evaluating feature subsets.</p>
      <p>The system was able to correctly attribute the text to the
proper author with only a 20% of missing assignments.</p>
      <p>V.</p>
      <p>RELATED WORKS</p>
      <p>Several generative models can be used to characterise
datasets that determine properties and allow grouping data
into classes. Generative models are based on stochastic block
structures [33], or on ‘Infinite Hidden Relational Models’ [34],
and ‘Mixed Membership Stochastic Blockmodel’ [35]. The
main issue of class-based models is the type of relational
structure that such solutions are capable to describe. Since
the definition of a class is attribute-dependent, generally the
reported models risk to replicate the existing classes for each
new attribute added. E.g. such models would be unable to
efficiently organise similarities between the classes ‘cats’ and
‘dogs’ as child classes of the more general class ‘mammals’.
Such attribute-dependent classes would have to be replicated as
the classification generates two different classes of ‘mammals’:
the class ‘mammals as cats’ and the class ‘mammals as dogs’.
Consequently, in order to distinguish between the different
races of cats and dogs, it would be necessary to further
multiply the ‘mammals’ class for each one of the identified
race. Therefore, such models quickly lead to an explosion of
classes. In addition, we would either have to add another class
to handle each specific use or a mixed membership model, as
for crossbred species.</p>
      <p>Another paradigm concerns the ’Non-Parametric Latent
Feature Relational Model’ [36], i.e. a Bayesian nonparametric
model in which each entity has boolean valued latent features
that influence the model’s relations. Such relations depend on
well-known covariant sets, which are neither explicit or known
in our case study at the moment of the initial analysis.</p>
      <p>In [37], the authors propose a sequential forward feature
selection method to find the subset of features that are relevant
to a classification task. This approach uses novel estimation of
the conditional mutual information between candidate feature
In [38], data from the charge-discharge simulation of
lithium-ions battery energy storage are used for
classification purposes with recurrent NNs and PNNs by means of a
theoretical framework based on signal theory. While showing
the effectiveness of the neural network based approaches, in
our case study classification results are given by means of a
probability, hence the use of a RBPNN, and an on-line training
achieved by reinforcement learning.</p>
      <p>VI.</p>
    </sec>
    <sec id="sec-5">
      <title>CONCLUSION</title>
      <p>This work has presented a multi-agent system, in which an
agent analyses fragments of texts and another agent consisting
of a RBPNN classifier, performs probabilistic clustering. The
system has successfully managed to identify the most probable
author among a given list for the examined text samples. The
provided identification can be used in order to complement
and integrate a comprehensive verification system, or other
kinds of software systems trying to automatically identify
the author of a written text. The RBPNN classifier agent
is continuously trained by means of reinforcement learning
techniques in order to follow a potential correction provided by
an human supervisor, or an agent that learns about supervision.
The developed system was also able to cope with new data
that are continuously fed into the database, for the adaptation
abilities of its collaborating agents and their reasoning based
on NNs.</p>
    </sec>
    <sec id="sec-6">
      <title>ACKNOWLEDGMENT</title>
      <p>
        This work has been supported by project PRIME funded
        <xref ref-type="bibr" rid="ref22">within POR FESR Sicilia 2007</xref>
        -2013 framework and project
PRISMA PON04a2 A/F funded by the Italian Ministry of
University and Research
        <xref ref-type="bibr" rid="ref22">within PON 2007</xref>
        -2013 framework.
      </p>
      <p>
        G. Capizzi, F. Bonanno, and C. Napoli, “Hybrid neural networks
architectures for soc and voltage prediction of new generation batteries
storage,” in Proceedings of International Conference on Clean
Electrical Power (ICCEP).
        <xref ref-type="bibr" rid="ref13">IEEE, 2011</xref>
        , pp. 341–344.
      </p>
      <p>C. Napoli, F. Bonanno, and G. Capizzi, “An hybrid neuro-wavelet
approach for long-term prediction of solar wind,” IAU Symposium, no.
274, pp. 247–249, 2010.</p>
      <p>Prentice
S. Mika, G. Ratsch, W. Jason, B. Scholkopft, and K.-R. Muller,
“Fisher discriminant analysis with kernels,” in Proceedings of the Signal
Processing Society Workshop. Neural networks for signal processing IX.
IEEE, 1999.</p>
      <p>H. Deshuang and M. Songde, “A new radial basis probabilistic neural
network model,” in Proceedings of Conference on Signal Processing,
vol. 2. IEEE, 1996.</p>
      <p>W. Zhao, D.-S. Huang, and L. Guo, “Optimizing radial basis
probabilistic neural networks using recursive orthogonal least squares algorithms
combined with micro-genetic algorithms,” in Proceedings of Neural
Networks, vol. 3. IEEE, 2003.
H. Javaherian, D. Liu, and O. Kovalenko, “Automotive engine torque
and air-fuel ratio control using dual heuristic dynamic programming,”
in Proceedings of International Joint Conference on Neural Networks
(IJCNN), 2006, pp. 518–525.</p>
      <p>B. Widrow and M. Lehr, “30 years of adaptive neural networks:
perceptron, madaline, and backpropagation,” Proceedings of the IEEE,
vol. 78, no. 9, pp. 1415–1442, Sep 1990.
[32] J.-W. Park, R. Harley, and G. Venayagamoorthy,
“Adaptive-criticbased optimal neurocontrol for synchronous generators in a power
system using mlp/rbf neural networks,” IEEE Transactions on Industry
Applications, vol. 39, no. 5, pp. 1529–1540, Sept 2003.</p>
      <p>Z. Xu, V. Tresp, K. Yu, and H. peter Kriegel, “Infinite hidden relational
models,” in In Proceedings of International Conference on Uncertainity
in Artificial Intelligence (UAI), 2006.</p>
      <p>E. M. Airoldi, D. M. Blei, E. P. Xing, and S. E. Fienberg, “Mixed
membership stochastic block models,” in Advances in Neural
Information Processing Systems (NIPS). Curran Associates, 2009.</p>
      <p>K. Miller, T. Griffiths, and M. Jordan, “Nonparametric latent feature
models for link prediction,” in Advances in neural information
processing systems (NIPS). Curran Associates, Inc., 2009, vol. 22, pp.
1276–1284.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Pappalardo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Tramontana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Marszałek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Połap</surname>
          </string-name>
          , and M. Woz´niak, “
          <article-title>Simplified firefly algorithm for 2d image key-points search,”</article-title>
          <source>in IEEE Symposium Series on Computational Intelligence.</source>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          7269. Springer,
          <year>2012</year>
          , pp.
          <fpage>206</fpage>
          -
          <lpage>213</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <given-names>F.</given-names>
            <surname>Bonanno</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Capizzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gagliano</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          , “
          <article-title>Optimal management of various renewable energy sources by a new forecasting method</article-title>
          ,”
          <source>in Proceedings of International Symposium on Power Electronics</source>
          , Electrical Drives,
          <article-title>Automation and Motion (SPEEDAM)</article-title>
          . IEEE,
          <year>2012</year>
          , pp.
          <fpage>934</fpage>
          -
          <lpage>940</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <given-names>A.</given-names>
            <surname>Nowak</surname>
          </string-name>
          and
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Woz´niak, “Analysis of the active module mechatronical systems</article-title>
          ,” in Proceedings of Mechanika - ICM. Kaunas, Lietuva: Kaunas University of Technology Press,
          <year>2008</year>
          , pp.
          <fpage>371</fpage>
          -
          <lpage>376</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          , G. Pappalardo, and E. Tramontana, “
          <article-title>A hybrid neuro-wavelet predictor for qos control and stability</article-title>
          ,”
          <source>in Proceedings of AI*IA: Advances in Artificial Intelligence</source>
          . Springer,
          <year>2013</year>
          , pp.
          <fpage>527</fpage>
          -
          <lpage>538</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <given-names>F.</given-names>
            <surname>Bonanno</surname>
          </string-name>
          , G. Capizzi,
          <string-name>
            <given-names>G. L.</given-names>
            <surname>Sciuto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          , G. Pappalardo, and E. Tramontana, “
          <article-title>A novel cloud-distributed toolbox for optimal energy dispatch management from renewables in igss by using wrnn predictors and gpu parallel solutions,” in Power Electronics, Electrical Drives, Automation and Motion (SPEEDAM</article-title>
          ),
          <source>2014 International Symposium on. IEEE</source>
          ,
          <year>2014</year>
          , pp.
          <fpage>1077</fpage>
          -
          <lpage>1084</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <given-names>A.</given-names>
            <surname>Nowak</surname>
          </string-name>
          and
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Woz´niak, “Multiresolution derives analysis of module mechatronical systems</article-title>
          ,” Mechanika, vol.
          <volume>6</volume>
          , no.
          <issue>74</issue>
          , pp.
          <fpage>45</fpage>
          -
          <lpage>51</lpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          , G. Pappalardo, and E. Tramontana, “
          <article-title>Using modularity metrics to assist move method refactoring of large systems</article-title>
          ,”
          <source>in Proceedings of International Conference on Complex, Intelligent, and Software Intensive Systems (CISIS)</source>
          . IEEE,
          <year>2013</year>
          , pp.
          <fpage>529</fpage>
          -
          <lpage>534</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          ,
          <year>December 2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <given-names>E.</given-names>
            <surname>Tramontana</surname>
          </string-name>
          , “
          <article-title>Automatically characterising components with concerns and reducing tangling</article-title>
          ,”
          <source>in Proceedings of Computer Software and Applications Conference</source>
          (COMPSAC)
          <article-title>Workshop QUORS</article-title>
          . IEEE,
          <year>July 2013</year>
          . DOI:
          <volume>10</volume>
          .1109/COMPSACW.
          <year>2013</year>
          .
          <volume>114</volume>
          , pp.
          <fpage>499</fpage>
          -
          <lpage>504</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          , G. Papplardo, and E. Tramontana, “
          <article-title>Improving files availability for bittorrent using a diffusion model,”</article-title>
          <source>in IEEE 23nd International Workshop on Enabling Technologies: Infrastructure for Collaborative Enterprises - WETICE</source>
          <year>2014</year>
          ,
          <year>June 2014</year>
          , pp.
          <fpage>191</fpage>
          -
          <lpage>196</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          ,
          <year>December 2011</year>
          , pp.
          <fpage>306</fpage>
          -
          <lpage>314</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <given-names>A.</given-names>
            <surname>Calvagna</surname>
          </string-name>
          and E. Tramontana, “
          <article-title>Delivering dependable reusable components by expressing and enforcing design decisions</article-title>
          ,”
          <source>in Proceedings of Computer Software and Applications Conference</source>
          (COMPSAC)
          <article-title>Workshop QUORS</article-title>
          . IEEE,
          <year>July 2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <source>DOI: 10.1109/COMPSACW</source>
          .
          <year>2013</year>
          .
          <volume>113</volume>
          , pp.
          <fpage>493</fpage>
          -
          <lpage>498</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <given-names>R.</given-names>
            <surname>Giunta</surname>
          </string-name>
          , G. Pappalardo, and E. Tramontana, “AODP:
          <article-title>refactoring code to provide advanced aspect-oriented modularization of design patterns,”</article-title>
          <source>in Proceedings of Symposium on Applied Computing (SAC)</source>
          .
          <source>ACM</source>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <given-names>E.</given-names>
            <surname>Tramontana</surname>
          </string-name>
          , “
          <article-title>Detecting extra relationships for design patterns roles</article-title>
          ,” in Proceedings of AsianPlop,
          <year>March 2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <given-names>G.</given-names>
            <surname>Capizzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          , and
          <string-name>
            <surname>L.</surname>
          </string-name>
          <article-title>Patern o`, “An innovative hybrid neurowavelet method for reconstruction of missing data in astronomical photometric surveys</article-title>
          ,”
          <source>in Proceedings of Artificial Intelligence and Soft Computing (ICAISC)</source>
          . Springer,
          <year>2012</year>
          , pp.
          <fpage>21</fpage>
          -
          <lpage>29</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <surname>LNCS</surname>
          </string-name>
          , vol.
          <volume>8467</volume>
          . Springer,
          <year>2014</year>
          , pp.
          <fpage>22</fpage>
          -
          <lpage>33</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Bonanno</surname>
          </string-name>
          , and G. Capizzi, “
          <article-title>Exploiting solar wind time series correlation with magnetospheric response by using an hybrid neuro-wavelet approach</article-title>
          ,
          <source>” Proceedings of the International Astronomical Union</source>
          , vol.
          <volume>6</volume>
          , no.
          <issue>S274</issue>
          , pp.
          <fpage>156</fpage>
          -
          <lpage>158</lpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <string-name>
            <given-names>G.</given-names>
            <surname>Capizzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Bonanno</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          , “
          <article-title>A new approach for lead-acid batteries modeling by local cosine,” in Power Electronics Electrical Drives Automation and Motion (SPEEDAM</article-title>
          ), 2010 International Symposium on,
          <year>June 2010</year>
          , pp.
          <fpage>1074</fpage>
          -
          <lpage>1079</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <string-name>
            <given-names>W.</given-names>
            <surname>Duch</surname>
          </string-name>
          , “
          <article-title>Towards comprehensive foundations of computational intelligence,” in Challenges for Computational Intelligence</article-title>
          . Springer,
          <year>2007</year>
          , pp.
          <fpage>261</fpage>
          -
          <lpage>316</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          <string-name>
            <given-names>G.</given-names>
            <surname>Capizzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Bonanno</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          , “
          <article-title>Recurrent neural networkbased control strategy for battery energy storage in generation systems with intermittent renewable energy sources</article-title>
          ,”
          <source>in Proceedings of International Conference on Clean Electrical Power (ICCEP)</source>
          . IEEE,
          <year>2011</year>
          , pp.
          <fpage>336</fpage>
          -
          <lpage>340</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          <string-name>
            <surname>Hall</surname>
          </string-name>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>