990 ieee transactions on ultrasonics, ferroelectrics, and frequency control, vol. 51, no. 8, august 2004
An Incremental Learning Algorithm with Confidence Estimation for Automated Identification of NDE Signals Robi Polikar, Member, IEEE, Lalita Udpa, Senior Member, IEEE, Satish Udpa, Fellow, IEEE, and Vasant Honavar, Member, IEEE
Abstract—An incremental learning algorithm is intro- • applications calling for analysis of large volumes of duced for learning new information from additional data data; and/or that may later become available, after a classifier has al- • applications in which human factors may introduce ready been trained using a previously available database. The proposed algorithm is capable of incrementally learn- significant errors. ing new information without forgetting previously acquired knowledge and without requiring access to the original Such NDE applications are numerous, including but are database, even when new data include examples of previ- not limited to, defect identification in natural gas trans- ously unseen classes. Scenarios requiring such a learning mission pipelines [1], [2], aircraft engine and wheel compo- algorithm are encountered often in nondestructive evalua- nents [3]–[5], nuclear power plant pipings and tubings [6], tion (NDE) in which large volumes of data are collected in batches over a period of time, and new defect types may [7], artificial heart valves, highway bridge decks [8], optical become available in subsequent databases. The algorithm, components such as lenses of high-energy laser generators named Learn++, takes advantage of synergistic general- [9], or concrete sewer pipelines [10] just to name a few. ization performance of an ensemble of classifiers in which A rich collection of classification algorithms has been each classifier is trained with a strategically chosen subset of the training databases that subsequently become avail- developed for a broad range of NDE applications. However, able. The ensemble of classifiers then is combined through the success of all such algorithms depends heavily on the a weighted majority voting procedure. Learn++ is inde- availability of an adequate and representative set of train- pendent of the specific classifier(s) comprising the ensem- ing examples, whose acquisition is often very expensive ble, and hence may be used with any supervised learning and time consuming. Consequently, it is not uncommon for algorithm. The voting procedure also allows Learn++ to estimate the confidence in its own decision. We present the the entire data to become available in small batches over algorithm and its promising results on two separate ultra- a period of time. Furthermore, new defect types or other sonic weld inspection applications. discontinuities may be discovered in subsequent data col- lection episodes. In such settings, it is necessary to update a previously trained classifier in an incremental fashion to I. Introduction accommodate new data (and new classes, if applicable) without compromising classification performance on pre- n increasing number of nondestructive evaluation ceding data. The ability of a classifier to learn under these A(NDE) applications resort to pattern recognition- constraints is known as incremental learning or cumula- based automated-signal classification (ASC) systems for tive learning. Formally, incremental learning assumes that distinguishing signals generated by potentially harmful de- the previously seen data are no longer available, and cu- fects from those generated by benign discontinuities. The mulative learning assumes that all data are cumulatively ASC systems are particularly useful in: available. In general, however, the terms cumulative and incremental learning are often used interchangeably. Scenarios requiring incremental learning arise often in • accurate, consistent, and objective interpretation of NDE applications. For example, in nuclear power plants, ultrasonic, eddy current, magnetic flux leakage, acous- ultrasonic and eddy current data are collected in batches tic emission, thermal or a variety of other NDE signals; from various tubings or pipings during different outage periods in which new types of defect or nondefect indi- Manuscript received October 3, 2003; accepted April 23, 2004. This cations may be discovered in aging components in sub- material is based upon work supported by the National Science Foun- dation under Grant No: ECS-0239090 for R. Polikar and Grant No: sequent inspections. The ASC systems developed using ITR-0219699 for V. Honavar. previously collected databases then would become inad- R. Polikar is with the Department of Electrical and Com- equate in successfully identifying new types of indications. puter Engineering, Rowan University, Glassboro, NJ 08028 (e-mail: [email protected]). Furthermore, even if no additional defect types are added L. Udpa and S. Udpa are with the Department of Electrical and to the database, certain applications may inherently need Computer Engineering, Michigan State University, East Lansing, MI an ASC system capable of incremental learning. Gas trans- 48824. V. Honavar is with the Department of Computer Science, Iowa mission pipeline inspection is a good example. The network State University, Ames, IA 50011. in the United States consists of over 2 million kilometers of
0885–3010/$20.00 c 2004 IEEE polikar et al.: learn++ and identification of nde signals 991 gas pipelines, which are typically inspected by using mag- access to previously available data. This requires that the netic flux leakage (MFL) techniques, generating 10 GB of knowledge formerly acquired from previous data should data for every 100 km of pipeline [2]. The sheer volume of be retained while new information is being learned, which data generated in such an inspection inevitably requires an raises the so-called stability-plasticity dilemma [13]; some incremental learning algorithm, even if the entire data are information may have to be lost to learn new information, available all at once. This is because current algorithms as learning new information will tend to overwrite formerly running on commercially available computers are simply acquired knowledge. Thus, a completely stable classifier not capable of analyzing such immense volumes of data at can preserve existing knowledge but cannot accommodate once due to memory and processor limitations. new information, but a completely plastic classifier can Another issue that is of interest in using the ASC sys- learn new information but cannot retain prior knowledge. tems is the confidence of such systems in their own de- The problem is further complicated when additional data cisions. This issue is of particular interest to the NDE introduce new classes. The challenge then is to design an community [11] because ASC systems can make mistakes, algorithm that can acquire a maximum amount of new by either missing an existing defect or incorrectly classi- information with a minimum loss of prior knowledge by fying a benign indication as a defect (false alarm). Both establishing a delicate balance between stability and plas- types of mistakes have dire consequences; missing defects ticity. can cause unpredicted and possibly catastrophic failure of The typical procedure followed in practice for learning the material, and a false alarm can cause unnecessary and new information from additional data involves discarding premature part replacement, resulting in serious economic the existing classifier and retraining a new classifier us- loss. An ASC system that can estimate its own confidence ing all data that have been accumulated thus far. How- would be able to flag those cases in which the classification ever, this approach does not conform to the definition of may be incorrect, so that such cases then can be further incremental learning, as it causes all previously acquired analyzed. Against this background, an algorithm that can: knowledge to be lost, a phenomenon known as catastrophic • learn from new data without requiring access to pre- forgetting (or interference) [14], [15]. Not conforming to viously used data, the incremental learning definition aside, this approach • retain the formerly acquired knowledge, is undesirable if retraining is computationally or finan- • accommodate new classes, and cially costly, but more importantly it is unfeasible for pro- • estimate the confidence in its own classification hibitively large databases or when the original dataset is lost, corrupted, discarded, or otherwise unavailable. Both would be of significant interest in a number of NDE ap- scenarios are common in practice; many applications, such plications. In this paper, we present the Learn++ algo- as gas transmission pipeline analysis, generate massive rithm that satisfies these criteria by generating an ensem- amounts of data that renders the use of entire data at ble of simple classifiers for each additional database, which once impossible. Furthermore, unavailability of prior data are then combined using a weighted majority voting algo- is also common in databases of restricted or confidential rithm. An overview of incremental learning as well as en- access, such as in medical and military applications in gen- semble approaches will be provided. Learn++ is then for- eral, and the Electric Power Research Institute’s (EPRI) mally introduced along with suitable modifications and im- NDE Level 3 inspector examination data in particular. provements for this work. We present the promising clas- Therefore, several alternative approaches to incremen- sification results and associated confidences estimated by tal learning have been developed, including online learning Learn++ in incrementally learning ultrasonic weld inspec- algorithms that learn one instance at a time [16], [17], and tion data for two different NDE applications. partial memory and boundary analysis algorithms that Although the theoretical development of such an algo- memorize a carefully selected subset of extreme examples rithm is more suitable, and therefore reserved for a journal that lie along the decision boundaries [18]–[22]. However, on pattern recognition or knowledge management [12], the such algorithms have limited applicability for realworld authors feel that this algorithm may be of specific interest NDE problems due to restrictions on classifier type, num- to the audience of this journal as well as to the general ber of classes that can be learned, or the amount of data NDE community. This is true in part because the algo- that can be analyzed. rithm was originally developed in response to the needs of two separate NDE problems on ultrasonic weld inspec- In some studies, incremental learning involves con- tion for defect identification, but more importantly due to trolled modification of classifier weights [23], [24], or incre- countless number of other related applications that may mentally growing/pruning of classifier architecture [25]– benefit from this algorithm. [28]. This approach evaluates current performance of the classifier and adjusts the classifier architecture if and when the present architecture is not sufficient to represent the II. Background decision boundary being learned. One of the most success- A. Incremental Learning ful implementations of this approach is ARTMAP [29]. However, ARTMAP has its own drawbacks, such as clus- A learning algorithm is considered incremental if it can ter proliferation, sensitivity of the performance to the se- learn additional information from new data without having lection of the algorithm parameters, to the noise levels in 992 ieee transactions on ultrasonics, ferroelectrics, and frequency control, vol. 51, no. 8, august 2004 the training data, or to the order in which training data weighted majority voting of the classes predicted by the are presented. Various approaches have been suggested to individual hypotheses. The hypotheses are obtained by overcome such difficulties [30]–[32], along with new algo- training a base classifier, typically a weak learner, using rithms, such as growing neural gas networks [33] and cell instances drawn from iteratively updated distributions of structures [34], [35]. A theoretical framework for the de- the training database. The distribution update rule used sign and analysis of incremental learning algorithms is pre- by Learn++ is strategically designed to accommodate ad- sented in [36]. ditional datasets, in particular those featuring new classes. Each classifier added to the ensemble is trained using a set B. Ensemble of Classifiers of examples drawn from a weighted distribution that gives higher weights to examples misclassified by the previous Learn++, the proposed incremental learning algorithm, ensemble. The Learn++ algorithm is explained in detail is based on the ensemble of classifiers approach. Ensemble below, and a block diagram appears in Fig. 1. approaches typically aim at improving the classifier accu- For each database Dk, k =1,...,K that becomes avail- racy on a complex classification problem through a divide- able, the inputs to Learn++ are labeled training data and-conquer approach. In essence, a group of simple classi- Sk = {(xi,yi) | i =1,...,mk} where xi and yi are train- fiers is generated typically from bootstrapped, jackknifed, ing instances and their correct classes, respectively; a or bagged replicas of the training data (or by changing weak-learning algorithm BaseClassifier; and an integer Tk, other parameters of the classifier), which then are com- the maximum number of classifiers to be generated. For bined through a classifier combination scheme, such as the brevity we will drop the subscript k from all other internal weighted majority voting [37]. The ensemble generally is variables. BaseClassifier can be any supervised algorithm formed from weak classifiers to take advantage of their so- that achieves at least 50% correct classification on Sk af- called instability [38]. This instability promotes diversity ter being trained on a subset of Sk. This is a fairly mild in the classifiers by forcing them to construct sufficiently requirement. In fact, for a two-class problem, this is the different decision boundaries (classification rules) for mi- least that can be expected from a classifier. nor modifications in their training datasets, which in turn At each iteration t, Learn++ first initializes a distri- causes each classifier to make different errors on any given bution Dt, by normalizing a set of weights, wt, assigned instance. A strategic combination of these classifiers then to instances based on their individual classification by the eliminates the individual errors, generating a strong clas- current ensemble (Step 1): sifier. Formal definitions of weak and strong classifiers can be found in [39]. m Learn++ is in part inspired by the AdaBoost (adap- Dt = wt wt(i). (1) tive boosting) algorithm, one of the most successful im- i=1 plementations of the ensemble approach. Boosting creates Learn++ then dichotomizes Sk by drawing a training a strong learner that can achieve an arbitrarily low er- subset TRt and a test subset TEt according to Dt (Step ror rate by combining a group of weak learners that can 2). Unless there is prior reason to choose otherwise, Dt do barely better than random guessing [40], [41]. Ensem- is initially set to be uniform, giving equal probability to ble approaches have drawn much interest and hence have each instance to be selected into TR1. Learn++ then calls been well researched. Such approaches, include but are not th BaseClassifier to generate the t classifier, hypothesis ht limited to, Wolpert’s stacked generalization [42], and Jor- (Step 3). The error of ht is computed on Sk = TRt + dan’s hierarchical mixture of experts (HME) model [43], TEt by adding the distribution weights of all misclassified as well as Schapire’s AdaBoost. Excellent reviews of vari- instances (Step 4): ous methods for combining classifiers can be found in [44], [45], and an overall review of the ensemble approaches can mk be found in [46]–[49]. εt = Dt(i)= Dt(i)[|ht(xi) = yi|] , Research in ensemble systems has predominantly con- i:ht(xi)=yi i=1 (2) centrated on improving the generalization performance in |•| complex problems. Feasibility of ensemble classifiers in in- where [ ] is 1 if the predicate is true, and 0 otherwise. 1 cremental learning, however, has been largely unexplored. If εt > /2, the current ht is discarded and a new ht is 1 Learn++ was developed to close this gap by exploring the generated from a fresh set of TRt and TEt.Ifεt < /2, prospect of using an ensemble systems approach specifi- then the normalized error βt is computed as: cally for incremental learning [12]. βt = εt (1 − εt), 0 <βt < 1. (3)
III. Learn++ as an Ensemble Approach for All hypotheses generated in the previous t iterations Incremental Learning then are combined using weighted majority voting (Step 5) to construct the composite hypothesis H : A. The Learn++ Algorithm t 1 In essence, Learn++ generates a set of classifiers Ht =argmax log . (4) y∈Y βt (henceforth hypotheses) and combines them through t:ht(x)=y polikar et al.: learn++ and identification of nde signals 993
Fig. 1. The block diagram of the Learn++ algorithm for each database Sk that becomes available.
Ht decides on the class that receives the highest total of the main differences between the two algorithms. Ad- vote from individual hypotheses. This voting is less than aBoost uses the previously created hypothesis ht to update democratic, however, as voting weights are based on the the weights, and Learn++ uses Ht and its performance on normalized errors βt: hypotheses with lower normalized er- weight update. The focus of AdaBoost is only indirectly rors are given larger weights, so that the classes predicted based on the performance of the ensemble, but more di- by hypotheses with proven track records are weighted more rectly on the performance of the previously generated sin- heavily. The composite error Et made by Ht then is com- gle hypothesis ht; and Learn++ focuses on instances that puted as the sum of distribution weights of instances mis- are difficult to classify—instances that have not yet been classified by Ht (Step 6) as: properly learned—by the entire ensemble generated thus far. This is precisely what allows Learn++ to learn in- mk crementally, especially when additional classes are intro- Et = Dt(i)= Dt(i)[|Ht(xi) = yi|] . duced in the new data; concentrate on newly introduced i:Ht(xi)= yi i=1 (5) instances, particularly those coming from previously un- 1 If Et > /2,anewht is generated using a new training seen classes, as these are precisely the instances that have subset. Otherwise, the composite normalized error is com- not been learned yet by the ensemble, and hence most dif- puted as: ficult to classify. After Tk hypotheses are generated for each database Bt = Et (1 − Et), 0