Parameter Identification of the Input Nonlinear Systems with the Colored Noise 1 Jingfan Liu*, Xiangqun Li, Darong Gao School of Electrical Engineering, Northwest University for Nationalities, China Abstract Nonlinear systems widely exist in practical applications, like communication systems, chemical processes, biomedical systems and so on. Therefore, nonlinear systems identificati Jingfan Liu on is quite significant both in theory and application. This thesis presents the identification algorithms for a class of nonlinear systems based on the Youth Project of Central University. Considering the identification of the input nonlinear systems with the colored noise, An extended Newton recursive algorithm are derived for comparison. In the simulation, the results show that the Newton recursive algorithm can get better accurate parameter estimates, The simulation results show the effectiveness of the proposed algorithms. Keywords Newton recursion, Newton method, input nonlinear systems, Hammerstein models 1 Introduction System identification plays an important role in system modeling, for example,signal processingand control engineering [1,3]. Parameter estimation is basic for system modeling and analysis [3-5]. Its purpose is to estimate system parameters according to a certain criterion function by using the observed input and output data. There are many effective methods in identifying systems such as the gradient search based algorithms [6–9], least squares approximation [10–13]. System identification is advancing at a fast speed, and new ideas and methods are emerging in many fields. This paper mainly studies the identification of the Hammerstein nonlinear systems with color noise. The system is widely used in practice, for example, the description of PH value, the process with nonlinear characteristics such as curtain function, dead zone, switching and so on. Therefore, it is of great significance to study this kind of time series model. Many scholars have done a lot of work on the parameter identification problem of Hammerstein system. Narendra and Gallman proposed a selection identification algorithm for nonlinear systems, referred to as NG algorithm [14], Stoica pointed out that there was a convergence problem in this algorithm [15]. Later, Rangan, Wolodkin and Poolla proved that when the linear part of the Hammerstein system was a finite impulse response model (FIR) and the system input was white noise, NG algorithm converges [16]. Chang and Luus proposed to identify Hammerstein system with colored noise by generation selection algorithm, but failed to prove the convergence of the algorithm [17]. Ba,Cerone and Regruto deduced that D( z ) A( z ) =1 (Figure 1) Hammerstein model output error is bounded, then parameters are also bounded [18]. In recent years, Bai proved the convergence of the generation selection algorithm in succession, but the generation selection algorithm wasn’t suitable for online identification [19]. Ding and Chen proposed a recursive method of Hammerstein system that could be used for online identification by using the principle of least squares, and proved its convergence by using the martingale convergence method [20,21]. Vanbeylen, Pintelon and Schoukens researched the maximum likelihood identification method for Hammerstein systems with Gaussian noise, invertible nonlinearity and zero output error [22]. Giri, Rochdi, Chaoui and Brouri identified the AIoTC2022@International Conference on Artificial Intelligence, Internet of Things and Cloud Computing Technology EMAIL: * Corresponding author: e-mail address: liujingfan_2006@126.com (Jingfan Liu) © 2022 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). CEUR Workshop Proceedings (CEUR-WS.org) 67 linear and nonlinear subsystems of the hysteresis Hammerstein system respectively using least squares [23]. Xiang Wei and Zonghai Chen proposed a new identification method for Volterra sequence (Laguerre function), a specific dynamic model of the nonlinear system. Many new methods are also used to identify nonlinear systems such as neural networks and genetic algorithms [24,25]. 2 System description and identification model Hammerstein model nonlinear system is the input nonlinear system, it consists of a static nonlinear segment and a dynamic linear segment(Figure 1) [26], where y (t ) is output, u (t ) is input, u (t ) is the output of the nonlinear part and the noise v(t ) is assumed to be i.i.d. random sequences with zero mean, A( z ) , B( z ) and D( z ) are polynomials in the unit backward shift operator z −1 [ z −1 y (t ) = y (t − 1)] , with Figure 1 Hammerstein system A( z) = 1+ α1z −1 + α2 z −2 ++ αna z −na , B ( z ) = β1 z −1 + β 2 z −2 + β 3 z −3 +  + β nb z − nb , D( z ) = 1 + d1 z −1 + d 2 z −2 +  + d nd z − nd . The intermediate variables μ (t ) , x(t ) and h(t ) are immesurable, and g (⋅) is static nonlinear function of state. The nonlinear part is an unknown polynomial that can be expressed as [27] : μ (t ) = g ( μ (t )) = c1 g1 ( μ (t )) + c2 g 2 ( μ (t )) +  + cnc g nc ( μ (t )) (1) = g ( μ (t ))c We can write the Hammerstein-CARMA model (Figure 1) into the following formula: A( z ) y (t ) = B( z ) μ (t ) + D( z )v(t ) (2) The unknown parameters needed to be estimated are: the linear subsystem parameters α , β , d and the nonlinear part parameters c .Let the superscript T represent the matrix transpose. α := [α1 , α 2 , , α na ]T ∈  na , β := [ β1 , β 2 , , β nb ]T ∈  nb , d := [d1 , d 2 , , d nd ]T ∈  nd , α  γ :=   ∈  na + nd , d  γ  θ :=  β  ∈  n , n := na + nb + nc + nd .  c   γˆt  γ      Let θˆ (t ):=  βˆt  denote the estimate of θ := β at time t .Define the information vector φ(t ) as:    cˆ   c   t 68 φ(t ) := [− y (t − 1), − y (t − 2),, − y(t − na ), (3) v(t − 1), v(t − 2),, v(t − nd )]T ∈  na + nd The output sequence can be written as: y (t ) = ζ T (t ) J + v(t ) (4) Thereinto J is the parameterized vector, ζ (t ) is the information vector, and they are defined as:  α  J :=   ∈  na + nb nc + nd ,  β ⊗ c ζ (t ) := [φT (t ), g ( μ (t − 1)), g ( μ (t − 2)), g ( μ (t − nb ))]T ∈  na + nb nc + nd . β ⊗ c is the Kronecker product of β and c . ln most of the existing papers, the combined parameter β ⊗ c is identified, and the combined parameter needs to be decomposed after the identification result is obtained [28], which increases the computational burden. The goal of this paper is to identify the parameters through the extended Newton recursive algorithm, obtain the parameter vectors α , β , c and d . 3 The extended Newton recursive algorithm 3.1 The algorithm description In this section, Newton method will be used to derive the augmented Newton recursive identification algorithm based on Hammerstein-CARMA model, its basic idea is to introduce stacking output vector and stacking information matrix. Define the input information:  g ( μ (t − 1))   g ( μ (t − 2)  G (t ) :=   ∈  nb ×nc (5)       g ( μ (t − nb ))  From Eq.(1) and Eq. (2), we have: y (t ) = φT (t )γ + β TG (t )c + v(t ) (6) We define a quadratic criterion function as follows: J1 (θ ) = J1 (γ , β , c ) = [ y (t ) − φT (t )γ − β TG (t )c ]2 (7) Since the Hessian matrix H [ J1 (γ , β , c )] of the criterion function J1 is singular, it is useful to introduce the stacked data, so Newton algorithm is used to solve the identification optimization problem.  φ(t )φT (t ) φ(t )c TG T (t ) φ(t ) β TG (t )    H [ J1 (γ , β , c )]=2  G (t )cφ (t ) G (t )cφ (t )G (t ) h23 (γ , β , c, t )  ∈  n×n T T T G (t ) βφ (t ) h32 (γ , β , c, t ) G (t ) ββ G (t )  T T T T   Where: ∂ h23 (γ , β , c, t ):= − {G (t )c[ y (t ) − φT (t )γ + β T G (t )c ]} ∂ ∂c h32 (γ , β , c , t ):= − {G T (t ) β[ y (t ) − φT (t )γ + β T G (t )c ]} ∂b = − G (t )[ y (t ) − φT (t )γ + β T G (t )c ] = − G T (t )[ y (t ) − φT (t )γ + β T G (t )c ]+G T (t ) βc T G T (t ) +G (t )cβ T G (t ) ∈  nb × nc =h23T (γ , β , c, t ) ∈  nc × nb 69 Consider the newest p data and define stacked output vector Y ( p ,t ) and stacked matrices Φ0 ( p , t ) , Φ (c , t ) and Ψ ( β , t ) .  y (t )   y (t − 1)  Y ( p, t ): = ∈p       y (t − p + 1)   φT (t )     φT (t -1)  Φ0 ( p, t ): = ∈  p×( na + nd )     T  φ (t − p + 1)   c TG T (t )    c TG T (t -1)  Φ (c, t ):= ∈  p×nb     T T  c G (t − p + 1)   β TG (t )     β TG (t -1)  Ψ ( β , t ): = ∈  p×nc     T   β G (t − p + 1)  Then define a new criterion function: J 2 (θ ) = J 2 (γ , β , c ) 2 := Y ( p, t ) − Φ0 ( p, t )γ − Ψ ( β , t )c (8) 2 = Y ( p, t ) − Φ0 ( p, t )γ − Φ ( β , t )γ Eq.(8) is equivalent to the following criterion function constructed from the data in a dynamical window with length p . t J 3 (γ , β , c ) =  [ y(i) − φ (i)γ − β G (i)c] i = t − p +1 T T 2 (9) That is J 2 (θ ) = J 3 (γ , β , c ) . If we take t = N and p = N ( N is the data length), then Eq.(8) and Eq.(9) are the least squares criterion functions [29]. Computing the gradient of J 2 (γ , β , c ) gives: Φ0 T ( p, t )    gradθ [ J 2 (γ , β , c )] = −2  Φ T (c, t )  [Y ( p, t ) − Φ0 ( p, t )γ − Ψ ( β , t )c ] Ψ T ( β , t )    Φ0 T ( p, t )    = − 2  Φ T (c, t )  [Y ( p, t ) − Φ0 ( p, t )γ − Φ (c, t ) β ] Ψ ( β , t )  T   Define the extended generalized information matrix Ξ (t ) and expanding innovation into innovation vector E ( p, t ) as:  Φ0 T ( p, t )    =  Φ T (cˆt −1 , t )  ∈  n× p Ξ (t ): Ψ T ( βˆ , t )   t −1  70 =Y ( p, t ) − Φ0 ( p, t )γˆt −1 −Ψ ( βˆt −1, t )cˆt −1 E( p, t ): (10) = Y ( p, t ) − Φ ( p, t )γˆ − Φ(cˆ , t ) βˆ ∈  p 0 t −1 t −1 t −1 Thus, we have:  Φ0T ( p, t )    gradθ [ J 2 (γ , β, c)] = −2  ΦT (cˆt −1, t )  Y ( p, t ) − Φ0 ( p, t )γˆt −1 −Ψ ( βˆt −1, t )cˆt −1 Ψ T ( βˆ , t )  t −1   Φ0T ( p, t )  (11)   = − 2  ΦT (cˆt −1, t )  Y ( p, t ) − Φ0 ( p, t )γˆt −1 − Φ(cˆt −1, t ) βˆt −1 Ψ T ( βˆ , t )  t −1  = − 2Ξ (t ) E( p, t ) Computing the Hessian matrix of the criterion function J 2 (γ , β , c ) . ∂gardθ [ J 2 (γ , β , c )] H [ J 2 (γ , β , c )]= ∂θ T Φ0 ( p, t )Φ0 ( p, t ) Φ0 T ( p, t )Φ (c , t ) Φ0 T ( p, t )Ψ ( β , t )  T   = 2  Φ T (c , t )Φ0 ( p, t ) Φ T (c, t )Φ (c, t ) H 23 (γ , β , c, t )  ∈  n×n  Ψ ( β , t )Φ0 ( p, t ) H 23 (γ , β , c, t ) Ψ ( β , t )Ψ ( β , t )  T T T   Where: ∂ =− H 23 (γ , β , c, t ): ∂c {Φ T (c, t ) [Y ( p, t ) − Φ0 ( p, t )γ − Ψ ( β , t )c ]} ∂ = − {[G (t )c, G (t -1)c,  , G (t -p + 1)c ][Y ( p, t ) − Φ0 ( p, t )γ − Ψ ( β , t )c ]} ∂c ∂  p −1  = −  G (t − i )c  y (t − i ) − φT (t − i )γ − β TG (t − i )c   ∂c  i =0  p −1 { = −  G (t − i )  y (t − i ) − φT (t − i )γ − β TG (t − i )c  − G (t − i )cβ TG (t − i ) i =0 } p −1 { =  G (t − i)  − y (t − i ) + φT (t − i )γ + β TG (t − i )c  + G (t − i )cβ TG (t − i ) i =0 } p −1 { } =  G (t − i)  − y (t − i ) + φT (t − i )γ + β TG (t − i )c  + Φ T (c, t )Ψ ( β , t ) ∈  nb ×nc i =0 Using the Newton method to minimize J 2 (θ ) , we can obtain the following recursive relation of computing θ (t ) : { } −1 θ (t ) = θ (t −1) − H[ J 2 (ηˆt −1, bˆt −1 , cˆt −1 )] gradθ [ J 2 (η t −1, b t −1, c t −1 )] (12) { } Ξ (t)E( p, t) −1 = θ (t −1) + 2 H[ J 2 (ηˆt −1, bˆt −1, cˆt −1 )] On the right side of the equation(12) containing the unknown Hessian matrix H [ J 2 (γˆt −1 , βˆt −1 , cˆt −1 )] , extended generalized information matrix Ξ (t ) and innovation vector E ( p, t ) , and φ(t ) contains the unpredictable noise v(t − i ), i = 1, 2, , nd .In order to solve these difficulties, according to the principle of recursive identification, replacing H [ J 2 (γˆt −1 , βˆt −1 , cˆt −1 )] , Ξ (t ) and E ( p, t ) in the above Eq.(12) with H [ J 2 (γˆt −1 , βˆt −1 , cˆt −1 )] , Ξˆ (t ) and Eˆ ( p, t ) , let vˆ(t − i ) denote the estimate of v(t − i ) to define the estimate of φ(t ) as follows: φˆ (t ) := [− y (t − 1), − y (t − 2), , − y (t − na ), vˆ(t − 1), vˆ(t − 2), , vˆ(t − nd )]T From Eq.(6), v (t ) can be written as:  T (t )γ − β TG (t )c v (t ) = y (t ) − φ We can summarize the Newton extended recursive algorithm (the H-ENR algorithm) for the Hammerstein-CARMA models as follows: θ (t ) = θ (t − 1) + Π ˆ −1 (t ) Ξ (t ) E ( p, t ) (13)  T (t )γ − β TG (t )c v (t ) = y (t ) − φ (14) 71 1 { } Πˆ (t )= Hˆ [ J2 (γˆt −1, βˆt −1, cˆt −1 )] 2 (15)  Φˆ 0T ( p, t )Φˆ 0 ( p, t ) Φˆ T ( p, t )Φ(c , t ) Φˆ 0T ( p, t )Ψ ( βt −1, t )   0 t −1  =  ΦT (ct −1, t )Φˆ 0 ( p, t ) ΦT (ct −1, t )Φ(ct −1, t ) Πˆ 23 (t )   T ˆ ( p, t ) ˆ T (t ) Ψ T ( βt −1, t )Ψ ( βt −1, t ) Ψ ( β t −1 , t )Φ 0 Π 23  { } p −1 Πˆ 23 (t )= G(t − i) − y(t − i) + φˆ T (t − i)γˆt −1 + β t −1 G(t − i)cˆt −1  T    (16) i =0 + ΦT (cˆt −1, t )Ψ ( β t −1, t ) T Ξˆ (t )= Φˆ 0 T ( p, t ) Φ T (cˆt −1 , t ) Ψ T ( βˆt −1 , t )  (17) Eˆ ( p, t ) = Y ( p, t ) − Φˆ 0 ( p, t )γˆt −1 − Φ (cˆt −1 , t ) βˆt −1 (18) Y ( p, t )= [ y (t ),y (t − 1), ,y (t − p + 1) ] T (19) Φˆ 0 ( p, t )= [ φˆ (t ),φˆ (t -1), ,φˆ (t − p + 1) ] T (20) T Ψ ( βˆt −1, t )= GT (t ) βˆt −1, GT (t −1) βˆt −1,GT (t − p +1) βˆt −1  (21) Φ(cˆt −1 , t )= [G(t )cˆt −1 , G(t −1)cˆt −1 ,, G(t − p + 1)cˆt −1 ] T (22) φˆ (t ) = [− y (t − 1), − y (t − 2), , − y (t − na ), (23) vˆ(t − 1), vˆ(t − 2), , vˆ(t − nd )]T [θˆ (t )]( na + nb + nd + 1 : n ) cˆtt = sgn[θˆn + n + n +1 (t )] (24) a b d  [θˆ (t )]( n + n + n + 1 : n )  a b d [θˆ (t )](na + nb + nd + 1: n) = cˆtt (25) The inverse matrix Ωˆ −1 (t ) in Eq. (13), for all t ,the stacked data length p in the nonsingular matrix ˆ (t ) should be large enough to invert the nonsingular matrix. The process of computing θˆ (t ) by the Ω H-ENR algorithm is summarised as follows: (1) Choose the stacked data length p and initialize: let t = 1 , θ( ˆ 0) be an arbitrary real vector with cˆ0 = 1 . (2) Collect the measured data u (t ) and y (t ) ,form stacked vector Y ( p, t ) by Eq.(19) , G (t ) by Eq.(5) the information vector φˆ (t ) by Eq.(23) and Φˆ 0 ( p, t ) by Eq.(20). (3) Compute and form Ψ (bˆt −1 , t ) by Eq. (21) and Φ (cˆt −1 , t ) by Eq. (22). (4) Form information matrix Ξˆ (t ) by Eq. (17) and compute innovation vector Eˆ ( p, t ) by Eq. (18). (5) Compute Πˆ 23 (t ) by Eq. (16) and Πˆ (t ) by Eq. (15). (6) Update the parameter estimation θ̂(t)by Eq. (13). (7) Normalize ĉt by Eq. (24) and Eq. (25) with the first positive element. (8) Increase t by 1 and go to Step 2. 3.2 Example Consider the following Hammerstein nonlinear system: A( z ) y (t ) = B( z ) μ (t ) + D( z )v(t ) A( z ) = 1 + α1 z −1 + α 2 z −2 = 1 − 1.07 z −1 + 0.675 z −2 , B( z ) = β1 z −1 + β 2 z −2 = 1.55 z −1 + 1.20 z −2 , 72 D( z ) = 1 + d1 z −1 = 1 + 0.13z −1 , μ (t ) = g ( μ (t )) = c1μ (t ) + c2 μ 2 (t ) + c3η 3 (t ) = 0.80 μ (t ) + 0.50 μ 2 (t ) + 0.33166 μ 3 (t ), θ = [α1 , α 2 , d1 , β1 , β 2 , c1 , c2 , c3 ]T . In simulation, the input {u (t )} is taken as a persistent excitation signal sequence,the noise {v(t )} is taken as a white noise sequence with zero mean and and variance σ 2 = 0.302 , and the data length is taken as p=100 and p=160 .Adopting the the Newton extended recursive algorithm(H-ENR) to estimate the paramelers of this Hammerstein-CARMA process,the corresponding noise-to-signal ratio is δns = 8.76% ,where the noise-to-signal ratio δns is as follows( h(t) and x(t) in Figure 1): var [ h(t ) ] δ ns = × 100% , var [ x(t ) ] D(z) B(z) h(t)= v(t) , x(t)= μ(t) . A(z) A(z) The parameter estimates and their errors are shown in Table 1 , the estimation errors versus t are shown in Figure 2. From Table 1 and Figure 2, we can get the following conclusions: (1) For the confirmed Hammerstein-CARMA model( v(t ) =0 or σ 2 =0 ),the Newton extended recur- sive algorithm can converge to the true value faster than the extended projection algorithm. For the stochastic Hammerstein-CARMA model ( σ 2 ≠ 0, σ 2 =0.302 ),the parameter estimation of Newton extended recursive algorithm fluctuates greatly, especially for the small stacked data length p , and its estimation errors cannot converge to zero even if the data length t tends to infinity. The reason is that the increment of the extended recursive algorithm does not approach zero. However, when the length of stacked data length p increases, the parameter estimation will become getting more stationary, as shown Table 1 and Figure 2 with p=100 and p=160 . (2) The parameter estimation errors rapidly converges to a small constant as the data length t in- creases. As the data length t goes to infinity, this constant is going to get very small and close to zero. This shows that the extended Newton recursive algorithm is effective. 4 Conclusions This paper studies the parameter estimation methods for the nonlinear Hammerstein-CARMA model. A extended Newton recursive(H-ENR) algorithms are derived based on the Newton method. Aiming at the difficulty that the information vector of Hammerstein-CARMA model contains unmeasured noiseterms, the principle of recursion identification is applied. The unknown noise terms contained in the information vector are replaced by its estimated value, and the estimated value is calculated by the parameterestimated value of the previous time or the previous time. Compared with the extended stochastic gradient algorithms the H-ENR algorithm has improved parameter estimation accuracy.The numerical example shows that the parameter estimates for the proposed H-ENR algorithm converge to their true values.At present, there is a lot of work tobe done in the study of nonlinear systems. ln this paper, only single-input single-output nonlinearsystems are studied. How to extend it to multi-inputmulti-output nonlinearsystems and apply it in the field is the next problem to be considered. 73 Table 1 The H-ENR estimates and errors with p =100 and p = 160 ( σ 2 = 0.302 ) p t α1 α2 d1 β1 β2 c1 c2 c3 δ(100%) 100 1 -1.06888 0.67413 3.04450 1.34687 0.90049 0.80049 0.48525 0.35177 115.55488 2 -1.05271 0.65084 0.48409 1.50838 0.94563 0.80375 0.49613 0.32839 17.27314 5 -1.08051 0.66901 -0.47730 1.48392 1.23121 0.79981 0.50128 0.33017 24.07139 10 -1.07007 0.67512 0.05824 1.54652 1.19742 0.80015 0.49985 0.33151 2.82854 15 -1.07022 0.67525 0.20194 1.54906 1.19976 0.80017 0.49991 0.33140 2.83091 20 -1.07026 0.67529 0.15804 1.54935 1.19966 0.79982 0.50015 0.33187 1.10366 50 -1.07013 0.67531 0.17327 1.55137 1.20127 0.80093 0.49948 0.33020 1.70550 160 1 -1.06684 0.67191 3.65162 1.34688 0.89635 0.80869 0.47924 0.34109 139.30408 2 -1.05152 0.64952 0.51147 1.52817 0.94528 0.79505 0.50071 0.34233 18.11615 5 -1.08051 0.66901 -0.47730 1.48392 1.23121 0.79981 0.50128 0.33017 24.07139 10 -1.07007 0.67512 0.05824 1.54652 1.19742 0.80015 0.49985 0.33151 2.82854 15 -1.07022 0.67525 0.20194 1.54906 1.19976 0.80017 0.49991 0.33140 2.83091 20 -1.07026 0.67529 0.15804 1.54935 1.19966 0.79982 0.50015 0.33187 1.10366 50 -1.07013 0.67531 0.17327 1.55137 1.20127 0.80093 0.49948 0.33020 1.70550 True values -1.07000 0.67500 0.13000 1.55000 1.20000 0.80000 0.50000 0.33166 Figure 2 The H-ENR estimation errors t versus( σ 2 = 0.30 2 ) 5 Acknowledgments This work is supported by the Fundamental Research Funds for the Central Universities under Grant 31920210075. 6 References [1] L. Ljung, System Identification: Theory for the User, 2nd ed., Prentice-Hall, Englewood Cliffs, NJ, 1999. [2] F. Ding, System Identification - New Theory and Methods, Science Press, Beijing, 2013. [3] Y.Cao, P. Li,Y. Zhang, Parallel processing algorithm for railway signal fault diagnosis data based on cloud computing,Future Gen. Comput. Syst. 88 (2018)279-283. [4] Abd-Elrady, E.(2008). A recursive prediction error algorithm for digital predistortion of FIR Wie- ner systems. In Communication systems , networks and digital signal processing (CSNDSp08)(pp.698-701). [5] L. Xu and F. Ding, “Parameter estimation for control systems based on impulse responses,” Inter- national Journal of Control Automation and Systems, vol. 15, no. 6, pp. 2471-2479, December 2017. [6] Lijuan Wan, Feng Ding, Ximei Liu, Chunping Chen. "A New Iterative Least Squares Parameter Estimation Approach for Equation-error Autoregressive Systems", International Journal of Con- trol, Automation and Systems, volume 18, pages 780–790, November 2019 . 74 [7] Abd-Elrady, E. (2009). Adaptive predistortion of Wiener and Hammerstein systems using spectral magnitude matching. In Proceedings of the 11th IASTED conference on signal and image pro- cessing (SIP'09), Honolulu, Hawaii, USA (pp. 17-19). [8] Golub G, Pereyra V. Separable nonlinear least squares: the variable projection method and its ap- plications [J]. Inverse Problems, 2003, 19(2): R1-R26. [9] I. Mansouri, A. Gholampour, O. Kisi, and T. Ozbakkaloglu, “Evaluation of peak and residual con- ditions of actively confined concrete using neuro-fuzzy and neural computing techniques,” Neu- ral Computing and Applications, vol. 29, no. 3, pp. 873-888, February 2018. [10] Tulsyan A, Huang B, Gopaluni R B, Fraser F J. On simultaneous on-line state and parameter es- timation in non-linear state-space models. Journal of Process Control, 2013,23(4):516-526. [11] Y. B. Gong, S. X. Yang, H. L. Ma, and M. Ge, “Fuzzy regression model based on geometric co- ordinate points distance and application to performance evaluation,” Journal of Intelligent & Fuzzy Systems, vol. 34, no. 1, pp. 395-404, January 2018. [12] J. J. Rubio, “SOFMLS: online self-organizing fuzzy modified least square network,” IEEE Transactions on Fuzzy Systems, vol. 17, no. 6, pp. 1296-1309, December 2009. [13] Y. P. Wang, L. C. Wang, D. H. Kong, and B. C. Yin, “Extrinsic least squares regression with closed-form solution on product grassmann manifold for video-based recognition,” Mathematical Problems in Engineering, vol. 2018, no. 1, pp. 1-7, March 2018. [14] Narendra K S, Gallman P G. An iterative method for the identification of nonlinear systems us- ing a Hammerstein model [J]. IEEE Transactions on Automatic Control, 1996, 11(3):546-550. [15] Stoica P. On the convergence of an iterative algorithm used for Hammerstein system identifica- tion [J]. IEEE Transactions on Automatic Control, 1981, 26(4):967-969. [16] Rangan S, Wolodkin G, Poolla K. Identification methods for Hammerstein systems [C]. Proceed- ings of Control and Decision, New Orleans, 1995, 697-702. [17] Chang F, Luus R. A noniterative method for identification using Hammerstein model [J]. IEEE Transactions on Automatic Control, 1971, 16(5):464-468. [18] Cerone V, Regruto D. Parameter bounds for discrete-time Hammerstein models with bounded output errors [J]. IEEE Transaction on Automatic Control, 2003, 48(10):1855-1860. [19] Bai E W, Li D. Convergence of the iterative Hammerstein system identification algorithm [J]. IEEE Transactions on Automatic Control, 2004, 49(11):1929-1940. [20] Ding F, Chen T. Identification of Hammerstein nonlinear ARMAX systems [J]. Automatica, 2005, 41(9):1479-1489. [21] Ding F, Shi Y, Chen T. Gradient-based identification methods for Hammerstein nonlinear AR- MAX models [J].Nonlinear Dynamics, 2006, 45(1-2):31-43. [22] Vanbeylen L, Pintelon R, Schoukens J. Blind maximum likelihood identification of Hammerstein systems [J]. Automatica, 2008, 44(12):3139-3146. [23] Giri F, Rochdi Y, Chaoui F.Z, Brouri A. Identification of Hammerstein systems in presence of hysteresis-backlash and hysteresis-relay nonlinearities [J]. Automatica, 2008, 44(3):767-775. [24] Jing Dahai, Liu Xiao-Ping. Online Identification Method for Nonlinear Models[J]. Control Engi- neering of China, 2007, 14(5):482-484. [25] Yuan Xiaolei, Bai Yan, Dong Ling. Nonlinear System Identification Based on Genetic Pro- gramming [J]. Control Engineering of China, 2009, 16(1):52-55. [26] Liu Y. Bai E W. Iterative identification of Hammerstein systems [J]. Automatica, 2007, 43(2):346-354. [27] Zhang Xueqing, Wang Xiangfeng, Jiang Jie, Wu Zhaoming, Wang Fang. Research of Manipula- tor Vibration Controlling Based on Fuzzy Identification [J]. Control Engineering of China, 2007,14(s1):57-59. [28] Feng Ding, Peter Xiaoping Liu, Guangjun Liu. ldentification methods for Hammerstein nonlinear systems. Digit. Signal Process.21(2):215-238(2011) . [29] Liu, Y.,& Bai,E.W. (2007). Iterative identification of Hammerstein systems. Automatica, 43(2),346-354. 75