Xem mẫu

  1. Adaptive Filtering and Change Detection Fredrik Gustafsson Copyright © 2000 John Wiley & Sons, Ltd ISBNs: 0-471-49287-6 (Hardback); 0-470-84161-3 (Electronic) Part I: Introduction
  2. Adaptive Filtering and Change Detection Fredrik Gustafsson Copyright © 2000 John Wiley & Sons, Ltd ISBNs: 0-471-49287-6 (Hardback); 0-470-84161-3 (Electronic) Extended summary 1.1. About the book ....................... 3 1.1.1. Outlook . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3 1.1.2. Aim . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 1.1.3. Background knowledge .................... 6 1.1.4. Outline reading and advice ................. 7 1.2. Adaptive linearfiltering .................. 8 1.2.1. Signal estimation ....................... 9 1.2.2. Parameter estimation using adaptive filtering . . . . . . . 11 1.2.3. State estimation using Kalman filtering . . . . . . . . . . 13 1.3. Change detection ...................... 17 1.3.1. Filters residual as generators ................ 17 1.3.2. Stoppingrules ........................ 18 1.3.3. One-model approach ..................... 19 1.3.4. Two-model approach ..................... 22 1.3.5. Multi-model approach .................... 23 1.4. Evaluation and formal design ............... 26 1.4.1. Generalconsiderations . . . . . . . . . . . . . . . . . . . . 26 1.4.2. Performance measures .................... 28 1.1. Aboutthebook 1.1.1. Outlook The areas of adaptive filtering and change (fault) detection are quite active fields. both in research and applications . Some central keywords of the book are listed in Table 1.1, and the figures. illustrated in Figure 1.1, give an idea of the relative activity in the different areas . For comparison. the two related and well established areas of adaptive control and system identification are included in the table . Such a search gives a quick idea of the size of the areas. but there are of course many shortcomings. and the comparison may be unfair at several instances . Still. it is interesting to see that the theory has reached many successful applications. whichis directly reflected in the
  3. 4 Extended summary Table 1.1. Keywords andnumber of hits (March 2000) in different databases. For Sci- enceDirect the maximum numberof hits is limited to 2000. On some of the rows, the logical ‘or’ is used for related keywords like ‘adaptive signal processing or adaptive estimation or adaptive filter’. Keyword IEL ScienceDirect IBM patent Adaptive filter/estimation/SP 4661 952 871 Kalman filter 1921 1642 317 Adaptive equalizer (Eq) 479 74 291 Target tracking 890 124 402 Fault diagnosis (FDI) 2413 417 74 Adaptive control (AC) 4563 2000 666 (System) Identification (SI) 8894 2000 317 Total number of items 588683 856692 2582588 I EL ScienceDirect IBM patents Kalman Kalman AC AC Figure 1. l . Relative frequency of keywords in different databases. number of patents. Browsing the titles also indicates that many journal and conference publications concern applications. Figure 1.2 reveals the, perhaps well known, fact that thecommunication industry is more keen to hold patents (here: equalization). Algorithms aimed at real-time implementation are also, of course, more often subject to patents, compared to, for instance, system identification, which is a part of the design process. Table 1.2 lists a few books in these areas. It is not meant to be compre- hensive, only to show a few important monographs in the respective areas. 1.1.2. Aim The aim of the book is to provide theory, algorithms and applications adap- of tive filters with or without support from change detection algorithms. Appli- cations in these areas can be divided into the the following categories:
  4. 1. l About the book 5 Patents per publication SI Adaptive AC Figure 1.2. Relative ratio of number of patents found in the IBM database compared to publications in IEL for different keywords. Table 1.2. Related books. Keyword Books Adaptive filters Haykin (1996), Mulgrew and Cowan (1988), Widrow and Stearns (1985), Cowan and Grant (1985) Kalman filter Kailath et al. (1998), Minkler and Minkler (1990), Anderson and Moore (1979), Brown and Hwang (1997), Chui and Chen (1987)7 Adaptive equalizer Proakis (1995), Haykin (1994), Gardner (1993), Mulgrew and Cowan (1988) Target tracking Bar-Shalom and Fortmann (1988), Bar-Shalom and Li (1993), Blackman (1986) Fault diagnosis Basseville and Nikiforov (1993), Gertler (1998), Chen and Patton (1999), Mangoubi (1998) Adaptive control Wstrom and Wittenmark (1989), Goodwin and Sin (1984) System identification Ljung (1999), Soderstrom and Stoica (1989), Johansson (1993) 0 Surveillance and parameter tracking. Classical surveillanceproblems consist in filteringnoisy measurements of physical variables asflows, tem- peratures, pressures etc, which will be called signal estimation. Model- based approaches, where (time-varying) parameters in a model of a non- stationary signal need to be estimated, is a problem of parameter track- ing. Adaptive control belongs tothisarea.Anotherexample is blind equalization in digital communication. 0 State estimation. The Kalman filter provides the best linear state es- timate, and change detection support can be used to speed up the re-
  5. 6 Extended summary sponse after disturbances and abrupt state changes. Feedback control using state feedback, such as Linear Quadratic Gaussian LQG control, belongs to this area. Navigation and target tracking are two particular application examples. 0 Fault detection. Faults can occur in almost all systems. Change detec- tion here has the role of locating the fault occurrence intime and togive a quick alarm. After the alarm, isolation is often needed to locate the faulty component. The combined task of detection and isolation is com- monly referred to as diagnosis. Fault detection can be recast to one of parameter or state estimation. Faults in actuators and sensors are most easily detected in a state space context, while system dynamic changes often require parametric models. These problems are usually treated separately in literature in the areas of sig- nal processing,mathematical statistics, automatic control, communication sys- tems and quality control. However, the tools for solving these problems have much in common, and the same type of algorithms can be used (C.R. John- son, 1995). The close links between these areas are clearly under-estimated in literature. The main difference of the problem areas above lies in the evaluation cri- teria. In surveillance the parameter estimate should be as close as possible to the true value, while in fault detection it is essential to get an alarm from the change detector as soon as possible after the fault, and at the same time generating few false alarms. In fault detection, isolation of the fault is also a main task. The combination of fault detection and isolation is often abbre- viated to FDI, and the combined task can be referred to as diagnosis. More terminology used in this area is found in Appendix B. The design usually consists of the following steps: 1. Modeling the signal or system. 2. Implementing an algorithm. 3. Tuning the algorithm with respect to certain evaluation criteria, either using real or simulated data. The main focus is on algorithms and their properties, implementation, tuning and evaluation. Modeling is covered only briefly, but the numerous examples should give an idea of the possibilities of model-based signal processing. 1.l .3. Background knowledge The derivations and analysis can be divided into following areas, and some the prior knowledge, or at least orientation, of these is required:
  6. 1. l About the book 7 0 Statistical theory: maximum likelihood, conditional distributions etc. 0 Calculus: integrations, differentiations, equation solving etc. 0 Matrix algebra: projections, subspaces, matrix factorizations etc. 0 Signal modeling: transfer functions and state space models. 0 Classical filter theory: the use of a low-pass filter for signal conditioning, poles and zeros etc. Transforms and frequency domain interpretations occur, but are relatively rare. To use the methods, it is essential to understand themodel and the statistical approach. These are explained in each chapter in a section called ‘Basics’. These sections should provide enough information for understanding and tun- ing the algorithms. A deeper understanding requires the reader to go through the calculus and matrix algebra in the derivations. The practitioner who is mainly interested in what kind of problems can be addressed is advised to start with the examples and applications sections. 1.l .4. Outline and reading advice There are certain shortcuts to approaching the book, and advice on how to read the book is appropriate. Chapter 1 is a summary and overview of the book, while Chapter 2 overviews possible applications and reviews the ba- sic mathematical signal models. These first two chapters should serve as an overview of the field, suitable for those who want to know what can be done rather than how it is done. Chapters 3, 5 and 8 - the first chapter in each part are the core chapters of the book, where standard approaches to adap- ~ tive filtering are detailed. These can be used independently of the rest of the material. The other chapters start with a section called ‘Basics’, which can also be considered as essential knowledge. Part V is a somewhat abstract pre- sentation of filter theory in general, without using explicit signal models. It is advisable to check the content at an early stage, but the reader should in no way spend too much time trying to digest all of the details. Instead, browse through and return to the details later. However, the ideas should be familiar before starting with the other parts. The material can be used as follows: 0 Chapters 1 and 2 are suitable for people from within industry who want an orientation in what adaptive filtering is, and what change detection can add to performance. An important goal is to understand what kind of practical problems can be solved. 0 Chapters 5, 8 and 13 are suitablefor an undergraduatecourse in adaptive filtering.
  7. 8 Extended summary Table 1.3. Organization of the book chapters. Estimation of Approach Parameter Signal State Adaptive filtering and whiteness based change Chapter 3 Chapter 5 Chapter 8 detection Maximum likelihood based change detection Chapter 3 Chapter 6 Chapter 9 Multiple-model based change detection Chapter 4 Chapter 7 Chapter 10 Algebraic (parity space) change detection Chapter 11 0 Chapters 1, 2, 3, 5, 8, 12, 13 andthe ‘Basics’ sectionsin theother chapters can be included in a graduate course on adaptive filtering with orientation of change detection, while a more thorough course for stu- dents specializing in the area would include the whole book. This matrix organization is illustrated in Table 1.3. Part I1 on signal estima- tion has many interesting signal processing applications, but it also serves as a primer on the change detection chapters in Parts I11 and IV. The approach in Chapter 11 is algebraic rather than statistical, and can be studied separately. Appendix A overviews the signal models used in the book, and presents the main notation, while Appendix B summarizes notation used in the literature on fault detection. The only way in which the book should not be approached is probably a reading from cover to cover. The theory in the last part is im- portant to grasp at an early stage, and so are the basics in change detection. Some of the parts on change detection will appear rather repetitive, since the basic ideas are quite similar for signal, parameter and state estimation. More specifically, Part I1 can be seen as a special case (or an illustrative first order example) of Part 1 1 1. 1.2. Adaptive linearfiltering Three conceptually different (although algorithmically similar) cases exist: 0 Signal estimation. 0 Parameter estimation in an unknown model. 0 State estimation in a known model. The following sections will explain the basic ideas of theseproblems, and introduce one central example to each of them that will be used throughout the chapter.
  8. linear 1.2 AdaDtive 9 1.2.1. Signal estimation The basic signal estimation problem is to estimate the signal part Ot in the noisy measurement yt in the model An example of an adaptive algorithm is Here At will be referred to as the forgetting fuctor. It is a design parameter that affects the tracking speed of the algorithm. Aswill become clear from the examples to follow, it is a trade-off between trackingspeed and noise attenuation. The archetypical example is to use At = X, when this also has the interpretation of the pole in a first order low-pass filter. More generally, any(low-pass)filtercanbe used. If it is known that the signal level has undergone an abrupt change, as might be indicated by a change detection algorithm, then there is a possibility to momentarily forget all old information by setting At = 0 once. This is an example of decision feedback in an adaptive filter, which will play an important role in change detection. An illustrative surveillance problem is given below. Example 7.7 Fuelconsumption The following application illustrates the use of change detection for im- proving signal quality. The data consist of measurements of instantaneous fuel consumption available from the electronic injection system ina Volvo 850 GLT used as a test car. The raw data are pulse lengths of a binary signal, called t,, which is the control signal from the electronic injection system to the cylinders. When t , = 1, fuel is injected with roughly constant flow, so the length of the t , pulses is a measure of fuel consumption. The measured signal contains a lot of measurement noise and needs some kind of filtering before being displayed to the driver on the dashboard. Intuitively, the actual fuel consumption cannot change arbitrarily fast, and the measured signal must be smoothed by a filter. There are two requirements on the filter: 0 Good attenuation of noise is necessary to be able to tune theaccelerator during cruising. 0 Goodtracking ability. Tests show that fuel consumption very often changes abruptly, especially in city traffic.
  9. 10 Extended summary 25 - Measurement Y I - Slow filter Time [S] Figure 1.3. Measurements of fuel consumption and two candidate filters. Data collected by the author in a collaboration with Volvo. Theserequirements arecontradictory for standard linear filters. The thin lines in Figure 1.3 show measurements of fuel consumption for a test in city traffic. The solid lines show the result of (1.2) for two particular values of the forgetting factor X. The fast filter follows the abrupt changes well, but attenuates the noise unsatisfactorily, and it is the other way around for the slow filter. The best compromise is probably somewhere in between these filters. The fundamental trade-off between speed and accuracy is inherent in all linear filters. Change detectors provide a tool to design non-linear filters with better performance for the type of abruptly changing signal in Figure 1.3. Figure 1.4 shows the raw data, together witha filter implemented by Volvo (not exactly the samefilter, but the principal functionality is the same). Volvo uses a quite fast low-pass filter to get good tracking ability and then quantizes the result to a multiple of 0.3 to attenuatesome of the noise. To avoid a rapidly changing value in the monitor, they update the monitored estimate only once a second. However, the quantization introduces a problem when trying to minimize fuel consumption manually, and the response time to changes of one second makes the feedback information to the driver less useful.
  10. linear 1.2 AdaDtive 11 25 - Measurement 20 - n - Volvo's filter 20 40 60 80 101 Time [S] Figure 1.4. Measured fuel consumption and a filtered signal similar to Volvo's implemented filter. 1.2.2. Parameter estimation using adaptive filtering A quite general parametric model of a linear system is Here G(q;0) and H ( q ;0) are two filters expressed in the delay operator q defined by qut = u + . The parameters I9 in the model are assumed to be tl time-varying, and are to be estimated recursively. Here and in the sequel u t denotes measured inputs to the system, if available, and et is an unknown input supposed to be modeled as a stochastic noise disturbance. The generic form of an adaptive algorithm is et,, = et + KtEt, Et = Yt - !?to The outputfrom the estimatedmodel is compared to thesystem, which defines the residual E t . The adaptive filter acts as a system inverse, as depicted in Figure 1.5. One common filterthat does this operation for linear in parameter models is the recursive least squares (RLS) filter. Other well known filters are Least Mean Square (LMS) (unnormalized or normalized), the Kalman filter and least squares over sliding window. This book focuses on parametric models that are linear in the parameters (notnecessarily linear in the measurements). The reason for this is that the statistical approaches become optimal in this case. How to obtain sub-optimal algorithms for the general linear filter model will be discussed.
  11. 12 Extended summarv Noise et o u t p u t yt Yt Et M et W W W Input ut System Parameter Bt ut Ad. filter 8 6 M Bt W W Figure 1.5. The interplay between the parameter estimator and the system. Example 7.2 Frictionestimation This example introduces a case study thatwill be studied inSection 5.10.3. It has been noted (Dieckmann, 1992) that the dependence between the so- called wheel slip and traction force is related to friction. The slip is defined as the relative difference of a driven wheel’s circumferential velocity, w,r,, and its absolute velocity, v: , wr, - v , S= 7 V, where r, is the wheel radius. We also define the normalized traction force, p, (sometimes referred to as the friction coefficient) as the ratio of traction force ( F f ) at the wheel and normal force ( N ) on one driven wheel, p = - .f F N Here p is computed from measurements, in this case a fuel injection signal, via an engine model. In the sequel we will consider S and p as measurements. Define the slip slope k as The hypothesis is that k is related to friction, and theproblem is to adaptively estimate it. The goal is to derive a linear regression model, i.e. a model linear in the parameters. The slip slope k we want to compute is defined in (1.7), which for small p reads (including an offset term S) p = k ( s - S), where also S is unknown. Although this is a model linear in the parameters, there are two good reasons for rewriting it as 1 s=p-+S. k
  12. linear 1.2 AdaDtive 13 That is, we consider S to be a function of p rather than the otherway around. The reasons are that S, contains more noise than pm, and that the parameter S is varying much slower as compared to kS. Both these arguments facilitate a successful filter design. Note that all quantities in the basic equation (1.9) are dimensionless. We will apply a filter where 1/k and S are estimated simultaneously. The design goals are 0 to get accurate values on k while keeping the possibility to track slow variations in both k and S, and at the same time 0 to detect abrupt changes in k rapidly. This case study will be approached by a Kalman filter supplemented by a change detection algorithm, detailed in Section 5.10.3. A notationally simplified model that will be used is given by where 154 = (Oil), 0i2))T is the parameter vector consisting of inverse slope and the offset, u is the input to the model (the engine torque) and yt is the t measured output (the wheel slip). An example of measurements of yt, ut from a test drive is shown in Figure 1.6(a). The car enters a low friction area at sample 200 where the slope pa- rameter Oil)increases. Figure 1.6(b) shows a scatter plot of the measurements together with a least squares fit of straight lines, accordingto the model 1.10, to the measurements before and after a friction change. Two adaptive filters were applied, one fast and one slow. Figure 1.7 il- lustrates the basic limitations of linear filters: the slow filter gives a stable estimate, but is too slow in tracking the friction change, while the fast filter gives an output that is too noisy for monitoring purposes. Even the fast filter has problem in tracking the friction change in this example. reason is poor The excitation, as can be seen in Figure 1.6(b). There is only a small variation in u after the change (circles), compared to the first 200 samples. A thorough t discussion on this matter is found in Section 5.10.3. 1.2.3. State estimation using Kalman filtering For state estimation, we essentially run a model in parallel with the system and compute the residual as the difference. The model is specified in state
  13. 14 Extended summary 0.015 0.35 0.01 0.3 - a U1 0.005 - 80.25 e o'2 ' 0 O ; lA0 05 - 51 g 8 0.4 e g 0.3- S e 0.2 B E 0.1 4 0 0 50 150 100 200 250 300 350 "0 0.015 0.005 0.01 0.02 Time [samples] Slip Figure 1.6. Measurements of yt and ut from a test drive going from dry asphalt to ice (a). A scatter plot (b) reveals the straight line friction model. The data were collected by the author in collaboration with Volvo. 0.05 True parameters 0.045 - - Estimated parameters, fast filter - Estimated parameters, slow filter 0.04 - - -- - - - - - - - - - - 50 250 100 200 150 300 Time [samples] Figure 1.7. Estimated slope and offset as a function of sample number. space form: (1.11) Here yt is a measured signal, A, B , C,D are known matrices and xt is an un- known vector (the state). There are three inputs to the system: the observable (and controllable) ut, the non-observable process noise and the measurement noise et.
  14. linear 1.2 AdaDtive 15 Noise e t , wt Output Y t Yt Et M et c c c Input ut System State zt ut Kalman filter 5 % M zt c c c Figure 1.8. The interplay between the state estimator and the system. In a statistical setting, state estimation is achieved via a Kalman $filter, while inadeterministicsetting, where the noises are neglected, the same device is called an observer. A generic form of a Kalman filter and an observer is (1.12) where Kt is the filter gainspecifying the algorithm. The Kalmanfilter provides the equations for mapping the model onto the gain as a function The noise covariance matrices Q, R are generally considered to have some degrees of freedom that act as the design parameters. Figure 1.8 illustrates how the Kalman filter aims at ‘inverting’ the system. Example 1.3 Target tracking The typical application of target tracking is to estimate the position of air- craft. A civil application is air traffic control (ATC) where the traffic surveil- lance system at each airport wants to get the position and predicted positions of all aircraft at a certaindistancefromtheairport.Thereareplenty of military applications where, for several reasons, the position and predicted position of hostile aircraft are crucial information. There are also a few other applications where the target is not an aircraft. As an illustration of Kalman filtering, consider the problemof target track- ing where a radar measures range R and bearing 8 to an aircraft. The mea- surements are shown in Figure 1.9 as circles. These are noisy and the purpose of the Kalman filter is firstly to attenuate the noise and secondly to predict future positions of the aircraft. This application is described in Section 8.12.2. One possible model for these tracking applications is the following state
  15. 16 Extended summary X 104 Initial filter design 104 3.5 3.5 3 3 2.5 2.5 2 2 > 1.5 1 0.5 0 -0.5 0 1 2 3 4 5 0 1 2 3 4 5 X X X 104 104 (4 Figure 1.9. Radar measurements (a) and estimates from a Kalman filter (b) for an aircraft in-flight manoeuvre. space model /l 0 T 0) xt+T = 1 0 0 1 OIXttl T 0 IWt (1.14) \o 0 0 1) Yt = + et. The state vector used here is X = ( ~ 1 ~ ~ 2 , 7 1 1 , 7 1 2 ) ~ ~ xi is the position where in 2D, zthe corresponding velocity. The state equation is one example of i l a motion model describing the dynamics of the object to be tracked. More examples are given in Chapter 8. The measurements are transformedto this coordinate system andKalman a filter is applied. The resulting estimates are marked with stars. We can note that the tracking is poor, and as will be demonstrated, thefilter can be better tuned. To summarize what has been said, conventional adaptive filters have the following well-known shortcoming: Fundamental limitation of linear adaptive filters The adaptation gain in a linear adaptive filter is a compromise between noise attenuation and tracking ability.
  16. 1.3 Change detection 17 1.3. Changedetection Algorithmically, all proposed change detectorscanbeputinto one of the following three categories: 0 Methods using one filter, where a whiteness test is applied to the resid- uals from a linear filter. 0 Methods using two filters, one slow and one fast one, in parallel. 0 Methods using multiple filters in parallel, each one matched to certain assumption on the abrupt changes. In the following subsections, these will be briefly described. Let us note that the computational complexity of the algorithm is proportional to how many filters are used. Before reviewing these methods, we first need to define what is meant by a residual in this context, and we also need a tool for deciding whether a result is significant or not - a stopping rule. 1.3.1. Filtersasresidualgenerators A good understanding of the Kalman and adaptivefilters requires a thorough reading of Chapters 5 and 8. However, as a shortcut to understanding sta- tistical change detection, we only need to know the following property, also illustrated in Figure 1.10. Residual generation Under certain model assumptions, the Kalman and adaptive filters take the measured signals and transform them to a sequence of residuals that resemble white noise before the change occurs. From a change detection point of view, it does not matter which filter we use and the modeling phase can be seen as a standard task. The filters also computes other statistics that are used by some change detectors, but more on this later. 4 Filter Figure 1.10. A whitening filter takes the observed input u and output yt and transforms t them to a sequence of residuals E t .
  17. 18 Extended summary In aperfect world, the residuals would be zero before a changeand non-zero afterwards. Since measurement noise and process disturbances are fundamen- tal problems in the statistical approach to change detection, the actual value of the residuals cannot be predicted. Instead, we have to rely on their average behavior. If there isno change in the system, and the model is correct, then the residuals are so-called white noise, that is a sequence of independent stochastic variables with zero mean and known variance. After the change either the mean or variance or both changes, that is, the residuals become ‘large’ in some sense. The main problem in statistical change detection is to decide what ‘large’ is. Chapter 11 reviews how state space models can be used for filtering (or residual generation as it will be referred to in this context). The idea is to find a set of residuals that is sensitive to the faults, such that a particular fault will excite different combinations of the residuals. The main approach taken in that chapter is based on parity spaces. The first step is to stack all variables into vectors. The linear signal model can then be expressed as where Yt is a vector of outputs, Ut is a vector of inputs, Dt the disturbances and Ft the faults. The residual is then defined as a projection With proper design of W , the residual will react to certain faults in specific patters, making fault isolation possible. A simple example is when the mea- surement is two-dimensional, and the state disturbance and the fault are both scalar functions of time. Then, under certain conditions, it is possible to lin- early transform the measurement to ~t = (dt, f t ) T . A projection that keeps only the second component can now be used as the residual to detect faults, and it is said that the disturbance is decoupled. It should also be noted that the residual is not the only indicator of a change (that is, it is not a sufficient statistic) inall cases. So even though residual based change detection as outlined below is applicable in many cases, there might be improved algorithms. The simplified presentation in this chap- ter hides the fact that the multi-model approaches below actually use other statistics, but the residual still plays a very important role. 1.3.2. Stopping rules Many change detection algorithms, among these algorithms in the classes of one-model and two-model approaches below, can be recast into the problem
  18. 1.3 Chanae detection 19 of deciding on the following two hypotheses: H0 : E(st) = 0, H1 :E(st) > 0. A stopping rule is essentially achieved by low-pass filtering st and comparing this value to a threshold. Below, two such low-pass filters are given: 0 The Cumulative SUM (CUSUM) test of Page (1954): gt = max(gt-1 + st - v,0), alarm if gt > h. The drift parameter U influences the low-pass effect, and the threshold h (and also v ) influences the performance of the detector. 0 The Geometric Moving Average (GMA) test in Roberts (1959) gt = Xgt-1 + (1- X)st, alarm if gt > h. Here, the forgetting factor X is used to tune thelow-pass effect, and the threshold h is used to tune the performance of the detector. Using no forgetting at all (X = 0), corresponds to thresholding directly, which is one option. 1.3.3. One-model approach Statistical whiteness tests can be used to test if the residuals are white noise as they should be if there is no change. Figure 1.11 shows the basic structure, where the filter residuals are trans- formed to a distance measure, that measures the deviation from the no-change hypothesis. The stopping rule decides whether the deviation is significant or not. The most natural distance measures are listed below: 0 Change in the mean. The residual itself is used in the stopping rule and St = Et. 0 Change invariance. The squaredresidual subtracted by a known residual variance X is used and st = E; - X. Data Yt,Ut c Filter Et c Distance meas. -St Stopping rule Alarm k,ta c Figure 1.11. Change detection based on a whiteness test from filter residuals.
  19. 20 Extended summary 0 Change in correlation. The correlation between the residual and past outputs and/or inputs are used and st = Etyt-k or st = ~ t u t - k for some k. 0 Change in sign correlation. For instance, one can use the fact that white residuals should change sign every second sample in theaverage and use st = sign(etet-l). A variant of this sign test is given in Section 4.5.3. Example 7.4 Fuelconsumption To improve on the filter in Example 1.1, the CUSUM test is applied to the residuals of a slow filter, like that in Figure 1.3. For the design parameters h = 5 and v = 0.5, the response in Figure 1.12 is obtained. The vertical lines illustrate the alarm times of the CUSUM algorithm. The lower plot shows how the test statistic exceeds the threshold level h at each alarm. The adaptive filter in this example computes mean of the signal fromthe latest alarm to the the current time. With a bad tuning of the CUSUM algorithm, we get either the total mean of the signal if there are no alarms at all, or we get the signal back as the estimate if the CUSUM test gives an alarm at each time instant. These are the two extreme points in the design. Note that nothing worse can happen, so the stability of the filter is not an issue here. To avoid the first situation where the estimate will converge to theoverall signal mean, a better 30 E g 20 z 1 I 3 10 U 0 0 100 200 300 400 500 ~~ Time [samples] Figure 1.12. Response of an adaptive filter restarted each time a CUSUM test ( h = 5 , v = 0.5), fed with the filter residuals, gives an alarm. The lower plot shows the test statistic of the CUSUM test.
  20. 1.3 Chanae detection 21 design is to use a slow adaptive filter of the type illustrated in Figure 1.3. To avoid the second degenerate case, an alarm can trigger a fast filter instead of a complete filter restart. That is, the algorithm alternates between slow and fast forgetting instead of complete or no forgetting. In contrast to the example above, the next example shows a case where the user gets important information from the change detector itself. Example 1.5 frictionestimation Figure 1.13 shows how a whiteness test, used for restarting the filter, can improve the filtering result in Example 1.2 quite considerably. Here the CUSUM test statistics from the residuals is used, and the test statistics gt are shown in the lower plot in Figure 1.13. Note that the test statistics start to grow at time 200, but that the conservative threshold level of 3 is not reached until time 210. - Estimated parameters 0.04- -- True parameters I I I v-- 100 200 150 250 300 41 I '4 I "0 50 100 150 200 250 300 Time [samples] Figure 1.l 3. Estimated friction parameters (upper plot) and test statistics from the CUSUM test (lower plot). Note the improved tracking ability compared to Figure 1.7. In the following example, change detection is a tool for improved tracking, and the changes themselves do not contain much information for the user. Example 1.6 Target tracking This example is a continuation of Example 1.3, and the application of the CUSUM test is analogous to Example 1.5. Figure 1.14(a) shows the estimated
nguon tai.lieu . vn