=Paper=
{{Paper
|id=Vol-2280/paper-20
|storemode=property
|title=Some Suggestions for Graduate Students and Scholars Undertaking Quantitative Interdisciplinary
Research: Remarks from The Practice
|pdfUrl=https://ceur-ws.org/Vol-2280/paper-20.pdf
|volume=Vol-2280
|authors=Dode Prenga,Safet Sula
|dblpUrl=https://dblp.org/rec/conf/rtacsit/PrengaS18
}}
==Some Suggestions for Graduate Students and Scholars Undertaking Quantitative Interdisciplinary
Research: Remarks from The Practice==
Some suggestions for graduate students and scholars undertaking quantitative interdisciplinary research: remarks from the practice Dode Prenga Safet Sula Department of Physics, Faculty of natural Sciences, University Department of Physics, Faculty of natural Sciences, University of Tirana of Tirana dode.prenga@fshn.edu.al safetsula@yahoo.com. Abstract literature. However, the proper analysis and personalized view on concrete problem remain always Quantitative researches encompass large a in the heart of research work. In this case a good field of works that embrace mathematics, strategy could be based on avoiding inappropriate engineering, physics, other natural science , approaches too. It worth to capitulate some aspects of economy and finance, quantitative sociology this process and below we will discuss it by and so on. Concerning differences and commenting concrete situation encountered. Notice benefiting from similarities is a state of art for that in interdisciplinary studies, the quantitative researchers and it is more highlighted for approach usually starts by assuming a model which young scholars working in interdisciplinary poses an additional question about it validity. In applications. Elements of classical and practice many of such aspects would be addressed and advanced statistics as seen from computing managed by the research team leader and would surely perspective, simulations, special and general subject of detailed expertise, although it happen that if techniques and models are the frontward of following a courageously a more independent path of the start for a successful analysis. In this the research, young scientist might run in inappropriate aspect there are many challenges for young analysis, and problematic interpretation of the results. scientist that must be addressed carefully. This In this regard, in the case of interdisciplinary became more imperative in the framework of researches there are always space for better practice applied informatics. and strategies. We want to illustrate some such cases in 1. Introduction the following. Researches in natural science affront students and 2. Standardized models and software scholars with a permanent challenge, how to shorten Usually in the preparatory work for a concrete the path from data to the appropriate results. In recent research, scholars try to apply a known model. It is a years many methods and techniques from natural common advice from mentors and team leaders to use science have been successfully used on other discipline models that are proven to work on the analysis of the as econometrics, sociology etc., giving rise to the systems under study. Occasionally the level and routine interdisciplinary branches as econo-physics, socio of research drop down to the application and therefore physics and so on. Computing and quantitative analysis the work would be further directed in some verification have been recommended as initial step of research in of the results; secondary estimation etc. This is such fields by many guides or books as described in common in spectral or time series analysis, [Rob16]. Young researchers need to run-through the investigating the reaction coefficients in a model, simulation techniques that might affront them with measurement or data elaboration and so on. Being more complicated situations. Some of them mimic those activities so common in data analysis steps, they physical system as annealing simulation (cooling are considered in dedicated software that attends Monte Carlo) or use biological behavior to speed up standard models and various thematic issues. Likely, in procedures of numerical convergences. In this case, physics or chemistry measurement, the instruments are one needs more solicited knowledge on such natural accompanied with an interface and the software that science too. However, many calculation problems have perform directly the data analysis. Next, in been addressed through advanced and specified econometrics, this is realized using more general tools, techniques that are available as discussed for example the standard statistical software as SPSS, EVIEW, in [Suz13], [Jan12], [Ott17] etc. So, newly debuting SAS, LISREL, and ONYX. They offer adequate researcher in interdisciplinary field necessitating modeling and calculation capacities, as described in the quantitative analysis, computing techniques, algorithms appropriate web pages. Other software wants the user or simulation procedures would probably find fine to be more active as is the case of R, PYTHON etc., solution by carefully reviewing computational and some others software introduce many randomness is a crucial advice for a good start. In mathematical and computational tools as MATLAB (or literature there many arguments in testing randomness its LINUX counterpart, OCTAVE), MATHEMATICA as Dieharder test or other recommended alternatives etc. Notice that basic languages as C, C++, [Wan03] FORTRAN, BASIC, PASCAL etc., are plenipotentiary for every computation requirements but they need a. Frequency Test: Monobit professional skills in programming and computation. b. Frequency Test: Block Detailed remarks on how to use them are largely c. Runs Test elaborated and easy reached in open sources. For this d. Test for the Longest Runs of Ones in a Block reason, in the first tentative, students are advised to e. Binary Matrix Rank Test apply the preprogramed ones models or to use f. Discrete Fourier Transform (Spectral Test) functions form the software libraries. No need to spend g. Non-Overlapping Template Matching Test time for something that others have perfectly done, but h. Overlapping Template Matching Test one has to know that they exist however. Nevertheless, i. Maurer's Universal Statistical Test it is crucial that when using dedicated softs and j. Linear Complexity Test models, each assumption and every condition of the k. Serial Test models should be completely fulfilled. Sometimes this l. Approximate Entropy Test is not rigorously possible. So, a good strategy for m. Cumulative Sums Test valuable quantitative analysis could be the building of n. Random Excursions Test algorithm from the researcher in a more interactive Each type of tests a-n above and others not included environment, aside of dedicated software application. herein can be implemented in specific subroutines, but In following we want to lists some precautions on such comparison between generated random arrays which cases. have been confirmed by tests, seems to be not an easy 2.1 First stop on random numbers tasks. Moreover it needs detailed knowledge on each. Random numbers are a key element in calculation To improve the above mentioned calculation, we algorithms. The process of obtaining unaffected should realize ourselves a better PRN generator. In this simulated systems outcomes is realized by the help of case we must apply step by step testing to fix the better random numbers that drive the algorithm to the new generation. To visualize the un-randomness for PRN unconditioned value. So we pick up a random value for generated in computers let’s start from the evident fact the variable and calculate the modeled value. In other that in generating random normally distributed application the probability to select between numbers we would expect that the outcome should be alternatives is fixed by comparison a given number to a normally distributed. We can test directly for the random one. It is clear that the quality or randomness Gaussianty as suggested in many textbooks of statistics for random numbers could be crucial for the unbiased using the kurtosis (as desired) outcome. Also, it is of great interests in cryptographic security, where it is necessary to E (x − µ )4 K ( x) = − 3 (1) examine the real randomness of various “random” σ4 number generators. Next, the randomness of casted Relation (1) is easy to apply but has many number is decisive for Monte Carlo simulations complementary assumptions that are difficult to be numerical integration. Some young researchers believe tested. Therefore we have applied another idea by that the machine random number generators are quite direct measuring the distance of the distribution under accurate, but in reality this is not the case: it is difficult analysis from normal distribution using q-functions to get a computer to do something by chance. Remark introduced in [Uma10] that a computer follows instructions blindly and is 2 1 therefore predictable. In practical simulations, the 1 (1 − q) ⎛ x − µ ⎞ 1−q random numbers are taken from pseudo-random- p( x) = [(1 + ⎜ ⎟ ] (2) Z 5 − 3q ⎝ σ ⎠ number generators (PRNG) but in more sensitive application randomness is based on a very un- Equation (2) reproduce the Gaussian for q=1 so the predictable physical event as radioactivity or difference q-1 estimate directly the distance from atmospheric noise. Up here we admit that testing normal distribution. In Fig 1 we show the fit of the distribution of machine PRN. As routinely practices by 2.1 Avoiding distribution’s assumption physicist, we use log-log presentation which highlights misuse the differences in the extremities of the graphs. The Statistical analysis is so common in interdisciplinary deviance of the generated number’s distribution from modeling and fitting procedures. So, it happens that the normal ones in the large values limit is easy noticeable assumed theoretical distribution is accepted without by naked eye. By using (2) for an array of 106 proof as describing the system or process under study. generated normally distributed random numbers we Or the assumption of the normally distributed obtained q~1.020 in the generation using for …randn() deviances in the fitting process was not put in doubt end loop in MATLAB; q~1.017 in the array generation too. However, under some specific circumstances, using normrnd() command whereas by a simple there are sufficient arguments that the final error BoxMuller algorithm using rand() as starting points, induced by the violation of normality assumption is not we had q~1.015. determinant [Gen72]. Other view as in [Hu13] suggest to the researchers to go deeper in error analysis. In practice other common assumption are the homogeneity paradigm; time-invariant processes and so on. Here one needs a careful evidence for distributions and other herein mentioned assumption which in turn result in a quite an easy task, but the benefit could be remarkable. 2.1.1 Some worked example for real systems Intriguingly the intuitive assumption that distribution for values of variables arising from a long time natural process would be a lognormal, has remained in the basis of many regulation and predictions. So, the famous Black-Shoe derivation for the distribution of p − pt −1 the return of prices namely r = t has been pt −1 found un-applicable even being very attractive in its Figure 1: Log-log plot of normal random number first appearance. It is suggested that in this case the distribution could be q-Gaussian of the form (2) Based on the arguments of [Tsa09] or [Uma10], the [Bor04]. Following this idea, a lognormal analogue of distribution for numbers generated by the last q-Gaussian (2) has been verified with good statistical algorithm is more Gaussian. By nature we cannot significance even for exchange rate of ALL as we measure the randomness directly, but judging from the represented in [Pre14]. In another such analysis resulting distribution, the random numbers produced in presented in [Sul16] we showed that the probability the second is expected to be better. To this end, we that an extreme flood in Drini cascade calculated using suggest to the young researcher to construct themselves lognormal distribution is as 8 times smaller than the random number generators and hence they would one calculated from the empiric fit distribution always have a profit from the machine ability to obtained using 20 year daily side floods as registered. produce in it own the PRN and method perfection to Practically the discharges from the lakes as response of generate PRN sequences. Next they’d better do near to extreme raining had occurred so frequently last • test the randomness before application years that coincides to the calculation of expected • pre-calculate the overall effect of non- occurrence of one time over more than 100 years. In randomness many other real systems we observed that the best fitted functions are in the parametric form like (2) or its lognormal q-counterpart 1 2.1.2 Measurement and data analysis 2 ⎛ 1 ⎜ ⎛ x1− q − 1 ⎞ ⎞⎟ 1− q assumptions pq ( x) ~ q ⎜1 − β (1 − q)⎜ ) − µ ⎟ ⎟ (3) Another inadequacy in the data elaboration stage could x ⎜ ⎜ 1− q ⎟ ⎟ ⎝ ⎝ ⎠ ⎠ be the assumption that the distribution is stable. This is usually fits better than expected functions say worse in the case of real systems with limited number Gaussian, lognormal, Weibull etc. of points and characteristic heterogeneity. We specifically mention here • Data gathered from measurement process in engineering, natural sciences researches etc. • Data gathered via inquires in social end economical sciences We observed that in some more detailed analysis the un-verified distributions assumption leads to speculative conclusions or even in wrong measurement practice. Scholars report the level of contamination in an area without offering supporting arguments for stationary of the state where measurements have been performed. It seems that mathematically is taken N x= ∑ x ↔ E (x ) ≡ i =1 i xρ (x )dx (4) Figure 2: Illustration of differences between standard and alternative distributions approach N ∫ x − sup port and thus the mean is the best representative of variable So, un-proofed assumption that the distribution on the x in its population. Notice that the right side of (4) data would be Gaussian or lognormal or Weibull etc. exists only if dhe probability density function (the should be avoided in applications until a test would distribution) of variable x is finite that is the case of confirmed it. Otherwise it could happen than stationary distribution. If not, value E does not exist at oversimplification of the systems or tendencies to all, so we cannot perform any statistical report on the confirm generalized expectation would leads to the measurement. In (4) the variable x could be the direct following conclusions seen in a paper recently. It is value measured or an output parameter as error in not surprise if a real erroneous use of normal regression procedures. Hence, in those cases the distribution paradigm would produce the result of Fig.3 verification stationary for the distribution ρ(x) is compulsory. Otherwise, the mean could be refereed as the best value of the sample measured, but not representative for the population. Mathematically the stability for distribution would be measured by parameter α-Levy but in calculation procedures it requires the fit for a complicated t-Student to the empirical data. Instead on can suggested an easy way- out from this situation making use of relations (2) above and testing parameter q. It is related to the α- Levy and there is e simple relationship with degree of 3− q freedom in the T-student by the rule ν = . q −1 Figure 3: Misuse of standard distribution. Fortunately, from the computing point of view, the form (2) can be fitted easy with standard nonlinear To this end, we highlight the logic step of distribution fitting algorithm, whereas T-student is more analysis to test them starting form un-stationary ones complicated. Next one can perform the evaluation of which are most likely to be found in real systems. the stability for the distribution under analysis by simply using the condition of variance finiteness that deviations from the real data were normally 1 distributed which should be analyzed as we discussed making use of the formula σ = calculated in in the preceding paragraph. We have noticed that in (5 − 3q )β practice, neglecting (6) unfortunately is not an isolated 5 error and in some cases young researchers have no idea [Uma10]. Stability requirement is 1 ≤ q ≤ but a 3 about it importance. To complicate things, related to broader rule say 1 ≤ q ≤ 2 has been suggested therein. relations (6) some programs offered themselves a bin Moreover, if q>3 there is no distribution at in statistical number (usually 20) or clearly ask to the user to input sense. In this case the relation (4) became meaningless the bin number. Statistical softs applies directly (6) or hence the arithmetical average value has to be declared similar formula without signaling us, and so avoiding as the mean of the data from the measurement and the subjective bin-size. But again, (6) is valid if never should be confounded with population’s mean deviances are normally distribution that might not be which does not exist. true. Thereof, a very good suggestion for correctness in data analysis is the optimization of the bin size. 2.1.3 Bin optimization procedures Finding the appropriate distribution should not be 2.2 More Flexible when working with considered as a trivial task. Usually the regressions are too easy in the first sight. But here is another point to models and conditions of validity Using well known models is a good practice but in this step in. The trick entails the way we approach the case pre-programmed ones could mislead to wrong underlying distribution for given data frequencies. So interpretation. Many dedicated softs as SPSS, SAS, in practice, the set of data series is ordered in J EVIEW in statistical analysis or LISREL, ONYX etc., categories or classes that (as a rule) are of equal size in structural equation studies, offers various solutions x −x for econometric, socio-dynamic problems and related J = max min h (5) subjects. Their routine includes many preparatory steps d ( j ) = n( xi ) ∈ [xmin + ( j − 1)h, xmin + jh,] and assumptions (again some of them need to be tested separately by the user). In this case a good advice is to build algorithms ourselves. Here is an example what The process (5) is called histogram or discretization of can happen. In the calculation of the informal economy the data distribution. But a (hidden) question remains as a hidden variable, we had in disposal a small portion mostly unanswered and unreported as well: how is of data, only 18 series (years 1998-2016). The model chosen the parameter h in (5)? Mathematically known as MIMIC (multi cause, multi indicators) speaking, the underlying (natural) distribution d should adopted by EVIEW or LISREL have been used by not be affected from the binning procedure (5) and in other researchers consequently and widely analytic view one request that moments of variable x recommended in such calculation. But specifically have not to be affected. So far, this has been considered those programs request a sufficient number of data straightforwardly and optimization rules have been series for statistical analysis (at least above 50 points in included in software or programs, but again, there exist our knowledge). Second they apply directly the unit cases that those steps have not been performed. A roots removing procedures. Next the result obtained as detailed analysis on methods and techniques for output needs further elaboration. If one tries to program histogram optimization is provided in [Shi10]. Correct the routine by ourselves, a detailed description is binning step should use Stokes rules or Friedman- provided in [Jor75]. In our example, we observed that Diaconincs formula the result obtained using deferent methods does not match. This was the result of not fulfillment of 1 presumed assumption by our data set. In particular the − h ~ (3.49 ÷ 3.73)σN 3 (6) use of differences to remove unit roots as recommended, from the other side has reduced significantly the data series from 14 to 12, and for where σ is standard deviation and N is the total number some variables included in model the stationary has not of values in the data set. However in (6) it is assumed been verified! However the very small number of points led to high uncertainty on statistical test. To So we write the algorithm in MATLAB as direct overcome the problem we preferred the calculation application of the model elaborated in [Jor72], [Gol64] using our routine that performed those additional steps. including preparatory steps (a-c). The results we a. analysis of tiny data (monthly records) by which obtained using deferent approach (currency approach the dynamics of the quantities has been identified and MIMIC model in the concrete work) matched (in an high level); especially there have been tow much better. Moreover the reproduced variables fit regimes in the interval considered so we used data very well with original ones confirming the goodness that belongs to the same regime for the fit of the calculation in this case as seen in the Fig 4. In b. accounting for those two effects fitting has been another calculation related to the consumer behavior accepted for lower confidence level we observed inadequate outcomes when using c. number of factor variables , responses and latent variables directly as from the measured. Calculation ones have been calculated using factor analysis was performed basing on standard logistic model used in econometrics and generally in the models involving categorical variables [Kus18]. Again it is preferred to construct the program considering specifics of the system and its variables using the same idea as above. a. Informal Economy by MIMIC 8-1-3 model : Figure 4: Another example of normalizing models: Fitting logistic model in consumer behavior So, the same result has been reported using the logistics and probit approach, that signify an improvement of the calculation. In general we can underline and highlight the importance of carefulness with models and especially a. Detailed verification of all dedicated softs assumption. Avoiding any non-logical operation on the data series b. Constructing single-purpose algorithm instead of b. Reproduction of the indicators: yellow line, unemployment rate, using multi-purpose pre-programed ones blue line, ln(GDP); red line logarithm of narrow money, c. Going deep in the mathematic of the problem before applying retunes Figure 4: Example of easy step by step analysis d. Analyzing the overall state of the system (case study: Informal Economy estimated by MIMIC model) 2.3 Trying Calculation challenges the old solution is permanent. In such cases it is very Some non-linear function or equations cause headache important for the researcher to explore many specific to the practitioners. Let consider for example the techniques and trying again to challenge the problem problem of fitting parametrical functions like the ones by madding up routines. including non-homogenous unit if variable as 3. Non-neglecting Calculation and y~ (7) Simulation Performance y0 + (x − xc )m [A + B cos(log ω (x − xc ) + ϕ )...] Advanced studies include simulation and hard calculation even in the graduate level. Students can try The form (7) is verified as underlying bobble directly in open sources as Wolfram Alpha to calculate dynamics in financial asset or indexes, failures, difficult integrals or they can use MATHEMATICA, explosions etc. [Sor01]. Regressions including MATLAB services etc. In numerical calculus including nonlinear ones do not work in this case. Taboo search integration many method exists and with little effort is not reported as effective too [Sor01]. Moreover, the nearly all problems for not advanced studies could be deviation is an Uhlenberg process that cannot be tested answered using each of them. But choosing the as we do for chi–deviances; hence the statistics for a fit appropriate method or algorithm might result in is not available by standard procedures. To deal with consuming time and energy for students. Clearly there numerical analysis of near to characteristic behavior we exist no general receipt in these cases and it is just the used recently [Pre16] a more complicated form of (7) duty of the research leading, but again some advices by extending relations (5) P(t ) = could help. For many purposes the two above 1− q mentioned software (and surely many others) are really ⎡ (t − t c )1−q − 1 ⎤ (8) mines with opportunities. Just needs to explore them. a + b(t − t c ) m + c(t − t c ) m cos(ω ⎢ ⎥ + φ) ⎣⎢ 1− q ⎦⎥ But again statistical and mathematical tools are 1− q indispensable. Here are some considerations from a ⎡ (t − t c )1−q − 1 ⎤ recent work. + d (t − t c ) m cos(2ω ⎢ ⎥ + φ )... ⎢⎣ 1− q ⎥⎦ 3.1.1 More effort in analytic relations To solve those problems it is suggested a genetic Analytic solutions are always the most desired outcome algorithm model which is detailed in [Sor01]. It is in the study of systems. Let mention here a simple based on two step calculation or ‘slaving parameters”. physical system containing two vectors (magnets). We write an ad hoc such a routine and the fit has been Later on it is proposed to model opinion formation in a found very accurately in the case of the dynamics of pair of individuals. Statistical mechanics calculation exchange rates [Pre16], [Pre14] or anxious-like start with partition function that in this case reads behavior in the water level during intensive floods in Komani Lake [Sul16]. Genetic algorithm is found ⎛ H ⎞ Z = exp⎜ − ⎟dΓ (9) successful for many such fitting difficulties. For ∫ ⎝ kT ⎠ interest of the readers we motioned that genetic Γ algorithm mimics the Darwinian evolution. So in the Where core of the program, one impose by a given probability a mutation in the solution vectors v = m, xc , ω , ϕ , [ ] H =− ∑ Jm m + ∑ µBm (10) i, j i j j i and if the result is not good, one changes the distribution of the random numbers used to impose the the Hamiltonian, m is magnet vectors and B is the mutation. We realized that by using beta distribution to magnetic inductions. Here m2=1. Physical quantities in produce random numbers, the convergence of our ad- principle will be calculated using appropriate formula hoc algorithm has been realized even for more of physics, once the partition function Z is evaluated in complicated forms of (7) resulting in (8) which we analytic form. Calculation of (9) having H given by called near to characteristic behavior in [Pre16]. (10) a genuine trick proposed in [Cif99] just to replace Similarly, the taboo searches can work for other m1m2 = (m1 + m2 )2 − 2 for 2-continus spin magnet situation especially where the possibility of returning in system rends (10) the form J H =− ( 2 ) M 2 − 2 − BM cos( B, M ) (11) that turns calculation (9) to be in analytic form! In statistical physics analytic forms of Z are the most “wanted” cases! Here M=m1+m2 is the sum of tow vectors. All calculation has been performed in [Cif16]. So, in a case-application in socio-dynamics, and practically for calculating of the opinion using an ad- hoc model, we used a more complicated inter-coupled Hamiltonian in the form proposed in [Pre18] O2 J U=J− 2 ⎡ ⎣ 2 ( ⎤ ⎦ ) − FO cos(F , O )⎢1 − α O 2 − 2 ⎥ (12) where O=O1+O2 is the resulting vector of opinion and U is the utility function using terms proposed in Figure 5: Illustration of the calculation of opinion (13) performed by directly using Matlab routines. [Sta09]. Here making use of properties of Bessel function, some adding integrals realized in [Cif16], one realized to find analytic form of the Z integral and 3.1.2 Actual softs offer near everything: symbolic following statistical mechanics formula finally we operations might help significantly obtained the average opinion per individuals as A full analysis of the system having utility (12) following 2 Similar calculation could strain the researcher because ⎛ β JO 2 ⎞ I1 (βFA(O) ) the calculation of Hessian needs differencing (20) and exp⎜⎜ ∫ ⎟ ⎟ A(O)dO analyzing the behavior of parametric equation, 1 0 ⎝ 2 ⎠ 4 − O2 studying the logic solution, imposing constraints etc. Ox = 2 2 ⎛ β JO 2 ⎞ I 0 (β FA(O) ) Fortunately this is not a case: by using symbolic exp⎜⎜ ∫ ⎟ dO equation and differencing in MATLAB ⎝ 2 ⎟⎠ 4 − O 2 0 (13) (MATHEMATICA etc.) we easily identified fixed points, null clines and everything from nonlinear ⎛ αJ 2 where A(O) = O⎜1 − ⎝ 2 [⎞ O −2 ⎟. ⎠ ] dynamics analysis of the system. So if we try to obtain the solution of Next we proceeded with numeric integration of (13) (Oc , φc )i = concerning in the zeros and infinite values. For the ⎧∂ ⎡ J 2 J 2 ⎤⎤ interest of the reader, we mention that the MATLAB ( ) ⎡ ( ⎪ ⎢− O − 2 − FO cos φ ⎢1 − α O − 2 ⎥ ⎥ = 0; ∂ϕ 2 2 ) offer adding facilities when dealing with integrands so ⎪ ⎣ ⎣ ⎦⎦ (13) have been calculated numerically and the result is ⎪⎪ ∂ ⎡ J J 2 ⎤⎤ represented in the Fig.7. This problem solved by using Arg ⎨ ∂O ( 2 ) ⎡ ( ) ⎢− O − 2 − FO cos φ ⎢1 − α O − 2 ⎥ ⎥ = 0; 2 2 ⎪ ⎣ ⎣ ⎦⎦ Matlab and some adding knowledge about functions ⎪0 ≤ ϕ ≤ π c involved in there, is a good argument for suggesting ⎪ crossing of the methods and techniques. We observe ⎪⎩0 ≤ Oc ≤ 2 that without mathematical the trick offered in [Cif99] analytic forms weren’t impossible and so the following which give null clines and in the analysis of second calculation in (14). However, exploring about solutions order derivatives involved in the Hessian, we observe of specific problems even very difficult could result that traditional effort are very likely to fail. Moreover, successful because there is always someone that can symbolic operation in this case facilitate remarkably solve easy our problem. the analysis by giving the opportunity of solving complicated systems including inequalities, plotting complicated graphs. In the Fig.8 is shows such a step 4. Calculate the energy in new configuration, if on searching for stationary state for the system (12) at it is smaller, the move is accepted, else it is zero degree temperature. accepted with Metropolis probability. 5. Stop if no more improvement could be done Basically this algorithm is fruitful for complexes calculation and it worked for some simplified case of equation (12). Other alternatives are available too. But if one use the simplified Metropolis-Hasting method we observe a non-sufficient convergence for simples XY2D model. Notice that new researchers want to follow the simplified MH procedure (11) instead of taking care of full detailed balance assumption. In this case a good advice is to measure directly the acceptance ratio. In many Monte Carlo algorithm would have acceptance ratio around 0.5 or lower, but it is not a receipt however. The suggestion on those cases is to explore patiently on possible specific algorithms rather using general algorithm. In our example in first Figure 6: Qualitative Analysis of stationary state tentative we had an acceptation ratio as high as 0.8, and This example suggest that a better knowledge about by using right formula on probability detailed balance particular programs would be a very helpful when dealing this ratio was decreased to 0.5. Later on a modified with complicated algebra in calculation. version called MALA (Metropolis-Adjusted Langeven 3.1.3 Exploration on simulation platforms Algorithm) as detailed in [Jan12], [Suz13] etc. However In many applications, the first idea coming in mind the solution of the problem was not finally concluded could be speeding up the study, so practically one start until we used the Wolf algorithm. Surely this could be with general algorithm and easiest ones. Not a common circumstance for many students or new surprisingly this can lead the research on some valley researcher therefore we insist in the suggestion of of the solution, making every effort to amend properly being real careful in implementation of every specific the algorithm, useless. As routinely used in numerical of quantitative methods. This is very useful for such simulation, Monte Carlo technique is the broadest researcher dealing in the interdisciplinary studies and method used. In those similar cases it very important to especially for them that have in their basic background explore as many as possible algorithms and methods. do not have a solid mathematical programming Typically algorithm might slow down or might never formation. converge due to the number of states around particular point in the solution space. We will explain in short 4. Conclusions this idea by just evoking the calculation of the average Successful quantitative studies needs for important and opinion of system (12). According to the literature individual efforts in informatics, programing, applied suggestions, we used the WOLF algorithm. The core mathematics and computation techniques. In the data algorithm has the following steps: analysis researches, graduate students and new scientist must pay much effort for a prior deep knowledge of the 1. one start from a random configuration of system, its characteristics and the nature of the state magnets assimilated in the angles between a where the measurements have been made. In general, vector and exterior field (φ ) using preprogramed algorithm or programs would not 2. pick a magnet (i) and calculate the energy of be the first choice and the benefit of building algorithm the cell involving all surrounding magnets themselves could be apparent for new researchers. Investing in a deeper mathematical model analysis 3. randomly select a direction θ, and turn all would be a very good start in the case of young spins upward to this direction researchers with solid natural science background. New scientist dealing with interdisciplinary studies would Comput Neurosci (2010) 29:171–182. DOI have better results if exploring patiently on the 10.1007/s10827-009-0180-4 possibilities of modern engineering programs, [Cif16] Orion Ciftja, Dode Prenga. Magnetic properties of a including forums as well classical XY spin dimer in a “planar” magnetic field. Journal of Magnetism and Magnetic References Materials 416 (2016) 220-225. [Cif99] Orion Ciftja and Marshall Luban. Equation of state [Rob16] M.A. Robinson. Quantitative research principles and and spin-correlation functions of ultra-small methods for human-focused research in engineering classical Heisenberg magnets. Physical Review B design. 'Research methods' publications. May 2016. Volume 60, Number 14 1 October 1999-II. DOI: 10.1007/978-3-319-33781-4_3 [Sta12] Dietrich Stauffer. A Biased Review of [Wan03] Ying Wang. Nonparametric Tests for Randomness. Sociophysics. Journal of Statistical Physics. 151. Ece 461 Project Report, MAY 2003 10.1007/s10955-012-0604-9. [Tsa09] Constantino Tsallis .Computational applications of [Gol64] Goldberger, Arthur S. (1964). Econometric non-extensive statistical mechanics. Journal of Theory. New York: John Wiley & Sons. Pp. 238– Computational and Applied Mathematics 227 243 (2009) p 51-58. [Jor75] Kalr Jorskog, Arthur Goldbwerg. Estimation of a [Uma10] Sabir Umarov, Constantino Tsallis, Murray Gell- model with multiple indicator and multiple causes Mann, Stanly Steinberg. Generalization of of ingle latent variable. Journal of the American symmetric -stable Lévy distributions for q>1. statistical association. Volume 70, issue 351(Sep. Journal of mathematical physics 51, 033502 2010 1975) [Gen72] Gene V Glass Percy D. Peckham, James R. Sanders [Kus18] Elmira Kushta, Dode Prenga, Fatmir Memaj. Consequences of failure to meet assumptions Analysis of consumer behavior in a small size underlying the fixed effects analyses of variance market unit: case study for Vlora District, Albania. and covariance . Review of education research vol. IJSRM,2018 42, no. 3. 1972 [Ott17] Michela Ottobre, Natesh S. Pillai, Konstantinos [Hu13] Shu-Ping Hu PRT-152 Fit, Rather Than Assume, a Spiliopoulosar. Optimal Scaling of the MALA CER Error Distribution. ICEAA Annual algorithm with Irreversible Proposals for Gaussian Conference and Training Workshop New Orleans, targets. 2017 aXiv:1702.01777 [stat.ME] LA.2013 [Jan12] Wolfhard Janke Monte Carlo Simulations of Spin [Bor04] Liza Borland. A Theory of Non-Gaussian option Systems Institut f¨ur Physik, Johannes Gutenberg- Pricing: capturing the smile and the skew (2004) Universit¨at, D-55099 Mainz, Germany. www. Spiedl. org/data/ conference s/SPIEP https://pdfs.semanticscholar.org [Pre14] Dode Prenga. Dinamika dhe vetorganizimi në disa [Suz13] Hideyuki Suzuki. Monte Carlo simulation of sisteme komplekse. Monografi. Pegi, 2014 classical spin models with chaotic billiards. [Sul16] Safet Sula, Dode Prenga. Improving the Analysis of Physical Review E 88, 052144 (2013) Hydrologic Time Data Series in our basins and [Pre18] Unpublished. Possible use of XY model in economic impact of hydro-industries. ISTI Faculty calculations of early stage of opinion formation. of Economy, University of Tirana. 2016. Submitted in the Journal of Modern Physics B. [Pre16] D. Prenga, M. Ifti. Complexity Methods Used in 2018 the Study of Some Real Systems with Weak Characteristic Properties. AIP Conf. Proc. 1722, 080006 (2016) [Sor01] D. Sornette, A. Johansen.“Significance of Log- periodic Precursors to Financial Crashes,” Quantitative Finance, 1: 452. 2001 [Shi10] Hideaki Shimazaki · Shigeru Shinomoto . Kernel bandwidth optimization in spike rate estimation. J