=Paper= {{Paper |id=Vol-1638/Paper65 |storemode=property |title=Intelligent psycho-diagnostic forecasting system |pdfUrl=https://ceur-ws.org/Vol-1638/Paper65.pdf |volume=Vol-1638 |authors=Alena N. Danilenko }} ==Intelligent psycho-diagnostic forecasting system == https://ceur-ws.org/Vol-1638/Paper65.pdf
Mathematical Modeling


           INTELLIGENT PSYCHO-DIAGNOSTIC
                FORECASTING SYSTEM

                                     A.N. Danilenko

                   Samara National Research University, Samara, Russia



       Abstract. In this article we have considered a problem of psychotherapy effi-
       ciency of. We offer a method of assessing and forecasting efficiency of thera-
       peutic influence. The principle of creating intelligence diagnostic forecasting
       system is described.

       Keywords: decision support systems, neural networks, fuzzy logic, efficiency
       of psychotherapy.


       Citation: Danilenko AN. Intelligent psycho-diagnostic forecasting system.
       CEUR Workshop Proceedings, 2016; 1638: 527-535. DOI: 10.18287/1613-
       0073-2016-1638-527-535


1      Introduction
The main objective of any psychotherapeutic treatment lies in helping patients to
introduce necessary changes to their lives. Success or efficiency of psychotherapy is
assessed depending on how resistant and, in a broader sense - salutary; these changes
are for the patient. Nowadays psychotherapeutic measures providing long-lasting
positive effect are considered to be the most suitable. In other words, overall perfor-
mance of a psychotherapist is evaluated by a result which cannot be subjected to reli-
able forecasting.
The problem of psychotherapeutic forecast efficiency at the initial stage of treatment
has not been solved and the choice of a method which would be adequate to a pa-
tient’s condition and their personal conditions is carried out by psychotherapists intui-
tively or on the basis of their professional experience. Therefore, this aspect of psy-
chotherapeutic treatment has a subjective character while the majority of psychother-
apeutic models do not even consider measuring dynamics of the patient’s condition
and assessing treatment efficiency.




Information Technology and Nanotechnology (ITNT-2016)                                    527
Mathematical Modeling                            Danilenko A.N. Intelligent psycho-diagnostic…


2      Method of efficiency predictive assessment of
       psychotherapeutic influence

We believe that efficiency predictive assessing of psychotherapeutic influence can be
based on a series of measurements conducted for indicators of mental and, naturally,
emotional condition of the patient. If we exclude diagnostics methods of a functional
state by removing psycho-physiological indicators, e.g., questionnaires, aimed at self-
assessment of the mental state of examinees we will arrive at the most widespread
method of forecasting.
One of the diagnostic methods which would meet the requirements of serial indicators
measurement is the color test by M. Lyusher. This method has been chosen by us
because it is not bound to lexical or graphic questions and terms which result in the
so-called "effect of learning" at repeated tests and seriously distort the results of test-
ing. Long-term applying Lyusher’s test (the essence of which is detecting examinee’s
choice of the preferred and non-preferred colors out of the eight shown colors)
demonstrates that color preferences depend on a set of basic personal characteristics
as well as on the actual state caused by a particular situation.
The offered method of assessment and forecast of the examinee’s condition is based
on calculating the described criteria. Since the testing procedure is brief and simple
and its contents bear unconscious character, it can be repeated several times, with
short time intervals without the loss of results reliability. It provides the opportunity
of forecasting patient’s condition dynamics by means of trends creating methods.


3      Intelligent emotional condition dynamic forecasting system

In order to forecast the change in the examinee’s emotional state of we apply the in-
distinct neural production network by Vanga-Mendel. Let us examine the forecast
task in more details in the example below.
Throughout the psychotherapeutic course the patient takes tests with a certain fre-
quency which corresponds to the frequency of carrying out psychotherapeutic ses-
sions (usually not more often than once a week). As a result, the data file with indica-
tors of this patient’s emotional condition is being collected. By carrying out each test
we will receive the value of seven indicators of an emotional state. These indicators
form a complex assessment of an emotional state. For the purpose of maximum sav-
ing of the examinee’s information we have decided to consider these indicators inde-
pendently without carrying out a regression generalization. In this regard, the forecast
for each indicator becomes irrespective of its own neural network. The forecast of one
of the indicators will be described further in this paper.
For the actual forecast to be successful it is necessary to know the values of at least
the same number of test results, as the number of the accesses into the neural network.
The values of the latest tests are used for the forecast below. While implementing the
forecast, the predicted values are used instead of the missing values. Therefore, the
process of forecasting becomes interactive.



Information Technology and Nanotechnology (ITNT-2016)                                     528
Mathematical Modeling                                                  Danilenko A.N. Intelligent psycho-diagnostic…


For implementing the forecast function in indistinct neural networks we have applied
an indistinct neural production network by Vanga-Mendel. It implements the indis-
tinct production model based on rules of the following type:
 If x1 there is A &  & x there is A &  & x there is A ,
                 i1        j             ij          m im
                    THEN y there is Bi , i  1, , n
Moreover, the algorithm of an indistinct conclusion which is carried out by this indis-
tinct neural production network is based on the following provisions:
─ The accessory functions of all the indistinct sets are represented by a Gauss func-
  tion;
─ Indistinct implication - indistinct work;
─ T-norm - indistinct work;
─ Accumulation of the active conclusions of rules is not carried out;
─ The defusification method - the average center.
Thus, the indistinct production model and the mechanism of an indistinct conclusion
for this indistinct neural production network can be displayed by the following ex-
pression:
                              m
 B 'i ( y)   Bi ( y )  Aij ( x' j )                                                                        (1)
                              j 1

As accumulation of rules is not carried out and the method of a defusification is the
method of the average center, a defusified value is determined by the output change in
the formula:
                                                          
         arg max   y   x' 
        n                                m

                            B 'i                 Aij   j

y' 
       i 1       y
                                         j 1                                                                  (2)
                    x' 
                   n    m

                                   Aij       j
                  i 1 j 1


In our case the accessory functions of all indistinct sets are represented by the Gauss
functions. Therefore, expression (2) is put in the following way:
                                                                  
                                                             
                                                                2

           arg max  exp  y  ci    exp   x' j aij   
       n                                     m

     
     i 1 
                                       
                                                    bij   
                             d i   j 1                                                                  (3)
          
y'                                       y                       
                          n   m         x' a  2 
                       
                        i 1 j 1
                                  exp   j ij  
                                        bij  
                                                   

Where ci and d i respectively are the centers and the width of the Gaussian functions
representing the function of accessory indistinct sets Bi of the conclusions of rules;
a i and bi - respectively are the centers and the width of the Gaussian functions rep-
resenting the function of accessory indistinct sets Ai of the prerequisites of rules.
In the final equation we receive:


Information Technology and Nanotechnology (ITNT-2016)                                                           529
Mathematical Modeling                                                         Danilenko A.N. Intelligent psycho-diagnostic…


                                            
      n          m         x' a  2  
         c                                
         i 
                      exp             
                                 j    ij

    i 1         j 1       bij                                                                                                   (4)
                                          
y'                                         
                             y

           n   m          x' a  2 
                  exp             
                               j   ij

         i 1 j 1        bij  
                                         

Expression (4) fully describes the procedure of an indistinct conclusion proceeding
from the above provisions. In figure 1 we have represented a multi-layered structure
of an indistinct network which layers elements realize the corresponding components
of this expression.
The structure of this indistinct network consists of four layers.
Layer 1 carries out a fusification of the variables x' j ( j  1,, n) . Elements of this
layer calculate values of the accessory functions  Aij ( x' j ) set by Gaussian functions

with parameters a ij and bij .
Layer 2 in which the number of the elements is equal to the number of rules in the
base carries out the aggregation of degrees of prerequisites validity of the correspond-
ing rules.
In layer 3 the first element serves as a layer for activation of the conclusions of rules,
ci according to the values of the degrees of the prerequisites of rules aggregated in the
previous layer. The second element of the layer carries out auxiliary calculations for
the subsequent defusification of the result.
Layer 4 which consisting of one element carries out a defusification of an output vari-
able.
Parametrical layers of a network are the first and the third layers where the adjusted
parameters are aij , bij and ci .
The indistinct neural production network of Vanga-Mendel has a multi-layered struc-
ture with direct distribution of a signal, the ouput value which can be changed, by
correcting parameters of the elements of the layers that allow using the algorithm of
the return distribution of a mistake for training the network.
In the set training selection ( x1
                                                 (k )
                                                        ,   x2( k ) ,  , xm( k ) , y ( k ) ), k  1,, K , x1( k ) , x2( k ) , xm( k ) are the
                                                                               (k )
values of accessing variables x1 , x2 ,, xm , y                                      is the reference value of an output
variable y in k .
We will consider procedure of the correction of values, ci , aij and bij .
Stage 1. For each example of the training selection of values of input variables
                                                                                                         (k )
( x1( k ) , x2( k ) ,  , xm( k ) , y ( k ) ), k  1,, K values of an output variable, y ' , are calculated.
Stage 2. Function of a mistake for all examples of the training selection is calculated:
E (k ) 
           2
             y'  y  , k  1,  , K
           1 (k ) (k ) 2
                                                            (5)




Information Technology and Nanotechnology (ITNT-2016)                                                                                    530
Mathematical Modeling                                                                                      Danilenko A.N. Intelligent psycho-diagnostic…



Stage 3. Values ci , aij and                                        bij are corrected for every i provided that every exam-
ple k of the training selection, proceeds from the ratios:
                                                   m         x' ( k ) a  2 
                           ( y ' ( k )  y ( k ) ) exp                       
                                                                      j      ij

                                                             bij  
                                                                                                                                                    (6)
                                                  j 1
ci (t  1)  ci (t )                                                          
                                        n   m            x' ( k ) a  2 
                                               exp                     
                                                               j        ij

                                      i 1 j 1          bij  
                                                                            
                                                                                                     x' ( k ) a  2 
                                                                                   
                                                                                         m
                             2  x' (jk ) aij  y ' ( k )  y ( k )  ci  y ' ( k )   exp  
                                                                                                           j      ij
                                                                                                                                                  (7)
                                                                                        j 1         b             
                                                                                                                      
aij (t  1)  aij (t )                                                                          
                                                              n    m        x' ( k ) a      2

                                                  bij2   exp                               
                                                                                     j       ij

                                                            i 1 j 1       bij               
                                                                                                 
                                                                          
                                                                                                  x '( k )  a  2 
                                                                             
                                                                                       m
                             2  x'(jk ) aij  y '( k )  y ( k )  ci  y '( k )   exp  
                                             2                                                          j       ij
                                                                                                                                                  (8)
                                                                                      j 1        b              
                                                                                                                    
bij (t  1)  bij (t )                                                                       
                                                           n m            x '( k )  a  2 
                                                bij   exp                               
                                                  2                               j       ij

                                                         i 1 j 1        bij  
                                                                                              

Where t is the number of training iterations;   0,1 is the training speed coefficient.
                                                                                                                                       (k )
Stage 4. Values of an output variable for each example of selection y ' are repeat-
edly calculated and the average total error, taking into account all examples of the
training selection, is calculated and compared to some established threshold:

E 
             1 K
             K k 1
                          
                y'( k )  y ( k )  
                                  2
                                                                                                                                                   (9)

In the case of a successful performance of a condition the network is considered to be
correctly trained. Otherwise, transition to stage 3 is carried out and adjustment re-
peats.
Functionality of the automated system of the forecast of dynamics of an emotional
condition of the personality includes the indicators forecasting function of an emo-
tional condition of the person. To comprehend this function an indistinct neural-
production network by Vanga-Mendel is used.
Effective functioning of a neural network requires an empirical selection of some
parameters, defining structure and networks, and parameters of the algorithm of its
training. The forecast window (i.e. a number of accesses into a neural network) is
defined in a system development task. The parameters of a neural network available
for control are a training coefficient.
To check the working capacity and the selection of necessary parameters the function
below is used:
f(x)  0.1  sin(0.02  x) + 0.2  sin(0.03  x)  0.6  sin(0.9  x) 
                                                                                                                                                   (10)
 1.11 sin(0.19  x) + 2.3  sin(0.37  x)
Its schedule and results of the forecast for a neural network are represented in figure
1.



Information Technology and Nanotechnology (ITNT-2016)                                                                                               531
Mathematical Modeling                               Danilenko A.N. Intelligent psycho-diagnostic…


In the function highlighted in blue, on the basis of which training was made, the red
color corresponds to the values predicted by a neural network after training. High-
lighted green portion of the area of the function has also been used for training. Other
area has been used as a test selection for determining the forecast quality of the un-
known data. Training was made before achieving the average total error training
(formula 4) of 0.01.




                 Fig. 1. Function graph of testing and results of the forecast

For the selection of the training coefficient value, a study of dependence of the num-
ber of iterations to achieve an error of training units (training speed) and errors on test
selection from the training coefficient value has been conducted at 0.01. Results are
given in table 1.

                        Table 1. Selection of the training coefficient f

Training coefficient       Number of iterations              Error
    0.1                      319                          0.0126977
    0.2                      154                          0.0115617
    0.3                      113                          0.0122632
    0.4                      87                           0.0126165
    0.5                      73                           0.0120938
    0.6                      63                           0.0117784
    0.7                      61                           0.0116801
    0.8                      60                           0.0134587
    0.9                      56                           0.0122873
The graph on which the value of an error on test has reached 10,000 for descriptive
reasons is given in figure 2.
It is evident from the graph that the number of iterations considerably decreases as the
coefficient value of training increases. Thus, the error slightly increases at great coef-
ficient values of training. Also if the values of the training coefficient are too large the
instability of work of the algorithm of training is possible. Therefore, it is expedient to
choose coefficient of training equal 0.7.




Information Technology and Nanotechnology (ITNT-2016)                                        532
Mathematical Modeling                               Danilenko A.N. Intelligent psycho-diagnostic…


For studying the quality of the forecast, research of dependence of an error upon test
selection from number of steps of the forecast has been conducted. Results are pre-
sented in table 2.




                         Fig. 2. Selection of the training coefficient

                        Table 2. Selection of the training coefficient

 Number of steps of the forecast                      Error

                    1                      0.004398821
                    2                      0.017341442
                    3                      0.059260252
Training was conducted until an error of 0.02 was achieved on the training selection.
It is evident from the table that the transition from the forecast for step one to the
forecast for step two the error increases by 4 times, upon transition to step three – 14
times. Taking into account that the network was trained for the forecast for one step
and the main objective is the forecast for one step, these results can be considered
satisfactory, since an error even at the forecast for step 3 keeps the same order as at
the forecast for step 1.


4      A control variable
For a control variable tests to be taken by several people were held to collect data for
training a neural network. The structure of a control variable included taking the test,
viewing the results of the test and analyzing the graph of change of indicators of an
emotional state with the forecast.
The testee is offered to choose the color which is mostly preferred by them by press-
ing on the desired selection. After the choice is made, the chosen color is removed
from the choice panel. After all colors are eslected, a message on a successful com-
pletion of the test is demonstrated to the testee.




Information Technology and Nanotechnology (ITNT-2016)                                        533
Mathematical Modeling                             Danilenko A.N. Intelligent psycho-diagnostic…


In figure 4 the graph displays the results of choices one after another, a standard color,
a code for a standard color, the corresponding tags, stress and alarm tags. In a separate
field, the psychologist can leave comments to this test in the form of a saved text.




              Fig. 3. Main window of the psychologist, browse mode of results

In order to view the graph of dynamics of emotional indicators, it is necessary to press
‘Graph’ button. Then there will pop-up a corresponding window (figure 5). In the left
part of the graph the log of changes of the chosen indicator is represented. In the right
part, green color indicates the forecast of change of this indicator for three steps.




                      Fig. 4. Graph of changes of emotional indicators

For passing a control test, basic training of a neural network on the given tests of the
three testees has been provided. The total number of tests was 19, 16 of them had
been taken by other testees in the training selection, 3 – in the test mode.


Information Technology and Nanotechnology (ITNT-2016)                                      534
Mathematical Modeling                               Danilenko A.N. Intelligent psycho-diagnostic…


Conclusion
The proposed method shows the real possibility of designing an intelligent system to
estimate and to predict the psychotherapeutic influence at early stages of interaction
between a psychotherapist and a patient. The method based on the classic psycho-
diagnostic procedures is easy to use by the patients and therapists.


References
 1. Wallerstein R, Robbins L, Sargent H, Luborsky L. The psychotherapy Research Project of
    the Menninger Foundation: Rationale, Method and Sample Use. First Report. Bull. of
    Menninger Clin., 1956; 20: 221-278.
 2. Nemchin TA. Conditions of psychological tension. L., 1983.
 3. Hofler M, Gloster AT, Hoyer J. Causal effects in psychotherapy: Counterfactuals counter-
    act overgeneralization. Psychotherapy Research, 2010.
 4. Kazdin A. Methodology, design, and evaluation in psychotherapy research. Handbook of
    psychotherapy and behavioral change. 4th ed., 1994: 19-71.
 5. Spence D. Traditional case studies and prescriptions for improving them. Psychodynamic
    treatment research. A handbook for clinical practice. Wiley, 1993.
 6. Luborsky L, Crits-Christoph P, Mintz J, Auerbach A. Who Will Benefit From Psychother-
    apy. N.Y.: Basic Books, 1988.
 7. Spirtes P, Glymour CN, Scheines R. Causation, Prediction, and Search. Springer-Verlag,
    New York, 1993.
 8. Zinchenko EA. Metod of expert visual definition of working emotional conditions on pro-
    duction, 1983: 59-63.
 9. Karmanov AA. Methodology of diagnostics of key parameters of a mental state in Lyush-
    er's test. URL: http://koob.ru.
10. Prokhorov AO. Mental states and their manifestations in educational process. Kazan, 1991.
11. Sobchik LN. Metod of color elections – modification of eight-color test of Lyusher. Practi-
    cal guidance. SPb.: Speech, 2010; 128 p.
12. Manski C. Identification for Prediction and Decision. Harvard University Press, Cam-
    bridge, Massachusetts, 2007.
13. Bareinboim E, Brito C, Pearl J. Local characterizations of causal bayesian networks. Lec-
    ture Notes in Artificial Intelligence, 2012; 7205: 1-17.
14. Pearl J. The deductive approach to causal inference. Journal of Causal Inference, 2014;
    2(2): 115-130.
15. Osovsky S. Neural networks for information processing. Finance and Statistics, 2002; 344
    p.
16. Borisov VV, Kruglov VV, Fedulov AS. Indistinct models and networks. The hot line –
    Telecom, 2007; 284 p.
17. Didelez V, Kreiner S, Keiding N. Graphical models for inference under outcome-
    dependent sampling. Statistical Science, 2010; 25(3): 368-387.
18. Bareinboim E, Pearl J. A general algorithm for deciding transportability of experimental
    results. Journal of Causal Inference, 2013; 1(1): 107-134.
19. Bareinboim E, Pearl J. Controlling selection bias in causal inference. In Proceedings of the
    15th International Conference on Artificial Intelligence and Statistics (AISTATS), JMLR,
    April 21-23, 2012: 100-108.



Information Technology and Nanotechnology (ITNT-2016)                                        535