Advertisement

Machine Learning

, Volume 79, Issue 3, pp 283–306 | Cite as

Fast adaptive algorithms for abrupt change detection

  • Daniel Nikovski
  • Ankur Jain
Article

Abstract

We propose two fast algorithms for abrupt change detection in streaming data that can operate on arbitrary unknown data distributions before and after the change. The first algorithm, \(\textsf{MB-GT}\) , computes efficiently the average Euclidean distance between all pairs of data points before and after the hypothesized change. The second algorithm, \(\textsf{MB-CUSUM}\) , computes the log-likelihood ratio statistic for the data distributions before and after the change, similarly to the classical CUSUM algorithm, but unlike that algorithm, \(\textsf{MB-CUSUM}\) does not need to know the exact distributions, and uses kernel density estimates instead. Although a straightforward computation of the two change statistics would have computational complexity of O(N 4) with respect to the size N of the streaming data buffer, the proposed algorithms are able to use the computational structure of these statistics to achieve a computational complexity of only O(N 2) and memory requirement of O(N). Furthermore, the algorithms perform surprisingly well on dependent observations generated by underlying dynamical systems, unlike traditional change detection algorithms.

Keywords

Event detection Distribution monitoring CUSUM 

References

  1. Atkeson, C. G., Moore, A. W., & Schaal, S. (1997). Locally weighted learning. Artificial Intelligence Review, 11(1–5), 11–73. CrossRefGoogle Scholar
  2. Basseville, M., & Nikiforov, I. V. (1993). Detection of abrupt changes: theory and application. Englewood Cliffs: Prentice Hall. Google Scholar
  3. Brodsky, B. E., & Darkhovsky, B. S. (1993). Nonparametric methods in change-point problems. Dordrecht: Kluwer. Google Scholar
  4. Fawcett, T., & Provost, F. (1999). Activity monitoring: noticing interesting changes in behavior. In Proceedings of the fifth international conference on knowledge discovery and data mining (KDD-99) (pp. 53–62). Google Scholar
  5. Guha, S., McGregor, A., & Venkatasubramanian, S. (2006). Streaming and sublinear approximation of entropy and information distances. In Proceedings of SODA’06 (pp. 733–742). New York: ACM Press. Google Scholar
  6. Hastie, T., Tibshirani, R., & Friedman, J. H. (2001). The elements of statistical learning. Berlin: Springer. zbMATHGoogle Scholar
  7. Jain, A., & Nikovski, D. (2007). Memory-based change detection algorithms for sensor streams (Technical Report TR2007-079). Mitsubishi Electric Research Laboratories. http://www.merl.com/publications/TR2007-079.
  8. Massey, F. J., Jr. (1951). The Kolmogorov-Smirnov test for goodness of fit. Journal of the American Statistical Association, 46(253), 68–78. zbMATHCrossRefGoogle Scholar
  9. Page, E. S. (1954). Continuous inspection schemes. Biometrika, 41, 100–115. zbMATHMathSciNetGoogle Scholar
  10. Provost, F. J., & Fawcett, T. (1997). Analysis and visualization of classifier performance: comparison under imprecise class and cost distributions. In Proceedings of KDD’97 (pp. 43–48). Menlo Park: AAAI Press. Google Scholar

Copyright information

© Springer Science+Business Media, LLC 2009

Authors and Affiliations

  1. 1.Mitsubishi Electric Research LaboratoriesCambridgeUSA

Personalised recommendations