Abstract
In this paper we discuss new adaptive proposal strategies for sequential Monte Carlo algorithms—also known as particle filters—relying on criteria evaluating the quality of the proposed particles. The choice of the proposal distribution is a major concern and can dramatically influence the quality of the estimates. Thus, we show how the long-used coefficient of variation (suggested by Kong et al. in J. Am. Stat. Assoc. 89(278–288):590–599, 1994) of the weights can be used for estimating the chi-square distance between the target and instrumental distributions of the auxiliary particle filter. As a by-product of this analysis we obtain an auxiliary adjustment multiplier weight type for which this chi-square distance is minimal. Moreover, we establish an empirical estimate of linear complexity of the Kullback-Leibler divergence between the involved distributions. Guided by these results, we discuss adaptive designing of the particle filter proposal distribution and illustrate the methods on a numerical example.
Similar content being viewed by others
References
Anderson, B.D.O., Moore, J.B.: Optimal Filtering. Prentice Hall, New York (1979)
Andrieu, C., Davy, M., Doucet, A.: Efficient particle filtering for jump Markov systems. Application to time-varying autoregressions. IEEE Trans. Signal Process. 51(7), 1762–1770 (2003)
Arouna, B.: Robbins-monro algorithms and variance reduction in finance. J. Comput. Finance 7(2) (2004)
Bollerslev, T., Engle, R.F., Nelson, D.: ARCH models. In: Engle, R.F., McFadden, D. (eds.) Handbook of Econometrics, pp. 2959–3038. North-Holland, Amsterdam (1994)
Cappé, O., Moulines, E., Rydén, T.: Inference in Hidden Markov Models. Springer, Berlin (2005), http://www.tsi.enst.fr/~cappe/ihmm/
Cappé, O., Douc, R., Guillin, A., Marin, J.M., Robert, C.P.: Adaptive importance sampling in general mixture classes. Stat. Comput. (2008, this issue)
Carpenter, J., Clifford, P., Fearnhead, P.: An improved particle filter for non-linear problems. IEE Proc. Radar Sonar Navig. 146, 2–7 (1999)
Chen, M., Chen, G.: Geometric ergodicity of nonlinear autoregressive models with changing conditional variances. Can. J. Stat. 28(3), 605–613 (2000)
Cover, T.M., Thomas, J.A.: Elements of Information Theory. Wiley, New York (1991)
de Boer, P.T., Kroese, D.P., Mannor, S., Rubinstein, R.Y.: A tutorial on the cross-entropy method. Ann. Oper. Res. 134, 19–67 (2005)
Douc, R., Moulines, E.: Limit theorems for weighted samples with applications to sequential Monte Carlo. Ann. Stat. 36 (2008, to appear), arXiv:math.ST/0507042
Douc, R., Moulines, E., Olsson, J.: On the auxiliary particle filter. Probab. Math. Stat. 28(2) (2008)
Doucet, A., Godsill, S., Andrieu, C.: On sequential Monte-Carlo sampling methods for Bayesian filtering. Stat. Comput. 10, 197–208 (2000)
Doucet, A., De Freitas, N., Gordon, N. (eds.): Sequential Monte Carlo Methods in Practice. Springer, New York (2001)
Eickhoff, J.C., Zhu, J., Amemiya, Y.: On the simulation size and the convergence of the Monte Carlo EM algorithm via likelihood-based distances. Stat. Probab. Lett. 67(2), 161–171 (2004)
Evans, M., Swartz, T.: Methods for approximating integrals in Statistics with special emphasis on Bayesian integration problems. Stat. Sci. 10, 254–272 (1995)
Fearnhead, P.: Computational methods for complex stochastic systems: a review of some alternatives to mcmc. Stat. Comput. 18, 151–171 (2008)
Fearnhead, P., Liu, Z.: On-line inference for multiple changepoint problems. J. R. Stat. Soc. Ser. B 69(4), 590–605 (2007)
Fort, G., Moulines, E.: Convergence of the Monte Carlo expectation maximization for curved exponential families. Ann. Stat. 31(4), 1220–1259 (2003)
Fox, D.: Adapting the sample size in particle filters through KLD-sampling. Int. J. Rob. Res. 22(11), 985–1004 (2003)
Geweke, J.: Bayesian inference in econometric models using Monte-Carlo integration. Econometrica 57(6), 1317–1339 (1989)
Givens, G., Raftery, A.: Local adaptive importance sampling for multivariate densities with strong nonlinear relationships. J. Am. Stat. Assoc. 91(433), 132–141 (1996)
Gordon, N., Salmond, D., Smith, A.F.: Novel approach to nonlinear/non-Gaussian Bayesian state estimation. IEE Proc. F Radar Signal Process. 140, 107–113 (1993)
Ho, Y.C., Lee, R.C.K.: A Bayesian approach to problems in stochastic estimation and control. IEEE Trans. Autom. Control 9(4), 333–339 (1964)
Hu, X.L., Schon, T.B., Ljung, L.: A basic convergence result for particle filtering. IEEE Trans. Signal Process. 56(4), 1337–1348 (2008)
Hürzeler, M., Künsch, H.R.: Monte Carlo approximations for general state-space models. J. Comput. Graph Stat. 7, 175–193 (1998)
Kailath, T., Sayed, A., Hassibi, B.: Linear Estimation. Prentice Hall, New York (2000)
Kong, A., Liu, J.S., Wong, W.: Sequential imputation and Bayesian missing data problems. J. Am. Stat. Assoc. 89(278–288), 590–599 (1994)
Künsch, H.R.: Recursive Monte-Carlo filters: algorithms and theoretical analysis. Ann. Stat. 33(5), 1983–2021 (2005), arXiv:math.ST/0602211
Legland, F., Oudjane, N.: A sequential algorithm that keeps the particle system alive. Tech. rep., Rapport de recherche 5826, INRIA (2006), ftp://ftp.inria.fr/INRIA/publication/publi-pdf/RR/RR-5826.pdf
Levine, R.A., Casella, G.: Implementations of the Monte Carlo EM algorithm. J. Comput. Graph. Stat. 10(3), 422–439 (2001)
Levine, R.A., Fan, J.: An automated (Markov chain) Monte Carlo EM algorithm. J. Stat. Comput. Simul. 74(5), 349–359 (2004)
Liu, J.: Monte Carlo Strategies in Scientific Computing. Springer, Berlin (2004)
Oh, M.S., Berger, J.O.: Adaptive importance sampling in Monte Carlo integration. J. Stat. Comput. Simul. 41(3–4), 143–168 (1992)
Oh, M.S., Berger, J.O.: Integration of multimodal functions by Monte Carlo importance sampling. J. Am. Stat. Assoc. 88(422), 450–456 (1993)
Olsson, J., Moulines, E., Douc, R.: Improving the performance of the two-stage sampling particle filter: a statistical perspective. In: Proceedings of the IEEE/SP 14th Workshop on Statistical Signal Processing, Madison, USA, pp. 284–288 (2007)
Pitt, M.K., Shephard, N.: Filtering via simulation: Auxiliary particle filters. J. Am. Stat. Assoc. 94(446), 590–599 (1999)
Ristic, B., Arulampalam, M., Gordon, A.: Beyond Kalman Filters: Particle Filters for Target Tracking. Artech House, Norwood (2004)
Rubinstein, R.Y., Kroese, D.P.: The Cross-Entropy Method. Springer, Berlin (2004)
Shen, C., van den Hengel, A., Dick, A., Brooks, M.J.: Enhanced importance sampling: unscented auxiliary particle filtering for visual tracking. In: AI 2004: Advances in Artificial Intelligence. Lecture Notes in Comput. Sci., vol. 3339, pp. 180–191. Springer, Berlin (2004)
Shephard, N., Pitt, M.: Likelihood analysis of non-Gaussian measurement time series. Biometrika 84(3), 653–667 (1997), erratum in 91, 249–250 (2004)
Soto, A.: Self adaptive particle filter. In: Kaelbling, L.P., Saffiotti, A. (eds.) Proceedings of the 19th International Joint Conferences on Artificial Intelligence (IJCAI), Edinburgh, Scotland, pp. 1398–1406 (2005), http://www.ijcai.org/proceedings05.php
Stephens, M., Donnelly, P.: Inference in molecular population genetics. J. R. Stat. Soc. Ser. B Stat. Methodol. 62(4), 605–655 (2000), with discussion and a reply by the authors
Straka, O., Simandl, M.: Particle filter adaptation based on efficient sample size. In: Proceedings of the 14th IFAC Symposium on System Identification, Newcastle, Australia, pp. 991–996 (2006)
Van der Vaart, A.W.: Asymptotic Statistics. Cambridge University Press, Cambridge (1998)
Wei, G.C.G., Tanner, M.A.: A Monte-Carlo implementation of the EM algorithm and the poor man’s Data Augmentation algorithms. J. Am. Stat. Assoc. 85, 699–704 (1991)
Author information
Authors and Affiliations
Corresponding author
Additional information
This work was partly supported by the National Research Agency (ANR) under the program “ANR-05-BLAN-0299”.
Rights and permissions
About this article
Cite this article
Cornebise, J., Moulines, É. & Olsson, J. Adaptive methods for sequential importance sampling with application to state space models. Stat Comput 18, 461–480 (2008). https://doi.org/10.1007/s11222-008-9089-4
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11222-008-9089-4