Neural Network Training Using Unscented and Extended Kalman Filter

Neural Network Training Using Unscented and Extended Kalman Filter

Research Article Robot Autom Eng J Volume 1 Issue 4 - November 2017 Copyright © All rights are reserved by Denis Pereira de Lima DOI: 10.19080/RAEJ.2017.01.555568 Neural Network Training Using Unscented and Extended Kalman Filter Denis Pereira de Lima*, Rafael Francisco Viana Sanches, Emerson Carlos Pedrino Federal University of São Carlos (UFSCar), Brazil Submission: August 03, 2017; Published: November 10, 2017 *Corresponding author: Denis Pereira de Lima, Federal University of São Carlos (UFSCar), São Carlos, Brazil, Email: Abstract This work demonstrates the training of a multilayered neural network (MNN) using the Kalman filter variations. Kalman filters estimate the weights of a neural network, considering the weights as a dynamic and upgradable system. The Extended Kalman Filter (EKF) is a tool that has been used by many authors for the training of Neural Networks (NN) over the years. However, this filter has some inherent drawbacks such as instability caused by initial conditions due to linearization and costly calculation of the Jacobian matrices. Therefore, the Unscented Kalman Filter variant has been demonstrated by several authors to be superior in terms of convergence, speed and accuracy when compared to the EKF. Training using this algorithm tends to become more precise compared to EKF, due to its linear transformation technique known as the Unscented literature.Transform (UT). The results presented in this study validate the efficiency and accuracy of the Kalman filter variants reported by other authors, with superior performance in non-linear systems when compared with traditional methods. Such approaches are still little explored in the Keywords : Genetic algorithm (GA); Evolutionary strategy (ES); Evolutionary programming (EP); Genetic programming (GP); Differential evolution (DE); Bacterial foraging optimisation (BFO); Artificial bee colony (ABC); Particle swarm optimisation (PSO); Ant colony optimization (ACO);Abbreviations Artificial :immune system (AIS); Bat Algorithm (BA) MNN: Multilayered Neural Network; EKF: Extended Kalman Filter; NN: Neural Networks; UT: Unscented Transform; ANNs: Artificial Neural Networks; SD: Steepest Decent; UKF: Unscented Kalman Filter; GRV: Gaussian Random Variables Introduction the weights of a neural network are adjusted adaptively. In the early 1940s, the pioneers of the field, McCulloch have been widely used in identifying and modeling nonlinear And because of their strong ability to learn, neural networks & Pitts [1], studied the characterization of neural activity, the neural events and the relationships between them. Such systems. Artificial Neural Networks (ANNs) are currently an a neural network can be treated by a propositional logic. additional tool which the engineer can use for a variety of Others, like Hebb [2], were concerned with the adaptation purposes. Classification and regression are the most common laws involved in neural systems. Rosenblatt [3] coined the tasks; however, control, modeling, prediction and forecasting name Perceptron and devised an architecture which has are other common tasks undertaken. For more than three subsequently received much attention. Minsky & Seymour [4] many properties and pointed out limitations of several models decades, the field of ANNs has been a focus for researchers. introduced a rigorous analysis of the Perceptron; they proved of many different software tools used to train these kinds of As a result, one of the outcomes has been the development [5]. Multilayer perceptrons are still the most popular artificial networks, making the selection of an adequate tool difficult for neural net structures being used today. Hopfield [6] tried to a new user [8]. The conventional algorithm for Multilayered contextualize the biological model of addressable memory of Neural Networks (MNN) is Back propagation (BP), which is biological organisms in comparison to the integrated circuits of an algorithm that uses the Steepest Decent method (SD), an a computer system, thus generalizing the collective properties extension of the Laplace method, which tries to approximate, that can be used for both the neurons and the computer systems. as close as possible, to the points of recurring non-stationary According to Zhan & Wan [7], neural networks are good tools point. BP in neural networks works to adjust the weights of the problems, in which you want to predict an event at some future for working with approximation in some nonlinear systems neural network, minimizing the error between the inputs and in order to obtain a desired degree of accuracy in predicting or even in non-stationary processes in which changes occur expected outputs. Problems related to BP are that it does not constantly and fast. As these activities happen in real time, Robot Autom Eng J 1(4): RAEJ.MS.ID.555568 (2017) 00100 Robotics & Automation Engineering Journal converge very quickly and the prediction for non-stationary the mean and the covariance, the estimator yields the same also used to solve non-linear problems that are not linearly generalizes to nonlinear systems without the linearization problems tends to be inaccurate. Another algorithm that is performance as a KF for linear systems however it elegantly separable is the Newton method. The difference between the steps required by the EKF [17,20]. We were motivated by faster than BP. However the Newton method requires more Newton method and BP is that the Newton method converges previous studies demonstrating the best performance and apply two methodologies in training of neural networks in efficiency of UKF related to EKF and, therefore, we tried to computing power and, also, large-scale problems are not feasible [7]. Recently, several authors have shown in studies order to prove of the efficiency of methods. that the Extended Kalman Filter (EKF), already well-discussed The discussion of the subject is distributed according to the in the literature in engineering circles, can also be used for following sections: 2- Training Neural Networks using Kalman the purpose of training networks to perform expected input- TrainingFilters, 3- Numerical Neural Network Testing Results, Using 4-Kalman Conclusion. Filters output mappings [9]. The publications by [10-15] are important with respect to EKF. Several authors such as [13,16-18], have EKF Algorithm described a variant, the Unscented Kalman Filter (UKF). Its performance is better than that of the EKF in terms of numerical Wk and stability and accuracy, and so a comparison of the two methods To understand the EKF, assume a non-linear dynamic Vk in the training of neural networks is needed. The basic system can be described by a state space model where Rk and Qk . difference between the EKF and UKF systems is the approach found in equations 1 and 2 are white Gaussian noise process f kZ, in which Gaussian Random Variables (GRV) are represented with zero mean independent( k ) covariance matrices for propagation through the system dynamics. The GRVs are The function is a transition matrix of a non-linear hkZ( , k ) propagated analytically through the first-order Taylor series function, the matrix possibly varying over a period of time. varying over time [13]. expansion of the nonlinear system. A set of sigma points is The function is a non-linear measurement matrix also ^− X+ = f( kX,) + W #(1) chosen so that their sample mean and sample covariance are k1 kk X k the true mean and true covariance respectively [19] (Figure 1). Y= hkX, + V #(2) k( kk) The basic idea of EKF according to [13] is linearizing the state-space model which is^ defined^− by equations 1 and 2 its This update is made by X or X current values are close to kthe estimatedk ones for each state. function depends on the particular problem to be considered since the linear model is obtained by applying the EKF equations. Also it is possible to say that the approximation made by the EKF occurs in two distinct phases [13]: 4: The first∂fkx(,) phase is the^− construction of the matrices in 3 and F= , XX= #(3) kk+1, ∂X k ∂hk(, X ) ^− H= k /,XX= #(4) kk∂X In this phase, the ijth entry Fkk+1, F kX, is identical to the partial Figure 1: Difference between EKF and UKF [9]. Hk derivative of component ith with respect to the jth order the partial derivative i part of HkX(, ) component of X. Similarly, the ijth entry is identical to with respect to the jth We were motivated by previous studies demonstrating therefore we tried to apply two methodologies in training of component of X. k are evaluated, so the best performance and efficiency of UKF related to EKF ^− ^− X k X k In the first case, the derivatives in ^X . Fkk+1, and Hk are both known as ^ ^− neural networks in order to prove of the efficiency of methods. that in the latter case the derivativesX are evaluated at X k and k are evaluated at time k [13,16]. Filter proposed by Julier et al. [16] is the propagation of a The entries of the matrices The vital operation performed in the Unscented Kalman computable because Fkk+1, and Hk Gaussian random variable through the system dynamics, using In the second phase of the EKF as the matrix a deterministic sampling approach. Using the principle in which are evaluated in the sequence they are used in a function of a set of discretely sampled points can be used to parametrize a first order Taylor series expansion to perform nonlinear How to cite this article: Denis P d L, Rafael F V S, Emerson C P. Neural Network Training Using Unscented and Extended Kalman Filter. Robot Autom 00101 Eng J. 2017; 1(4): 555568. DOI: 10.19080/RAEJ.2017.01.555568. Robotics & Automation Engineering Journal ^ ^− FkX(, ) HkX(,k ) k and round X k and X k T T 12T TT 1 FkX(,k ) and HkX(,k ) are estimated as follows: XxxxYyy= [12 , ,...,km ,] ,= [12 , ,..., yww ] ,= [( ) ,( w ) ] , W= 1Where,TT 1 1TT 2 2TT 2 2TT 1 11 1T 2 functions ^^ , respectively.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    6 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us