Abstract
We develop methods for performing maximum a posteriori (MAP) sequence estimation in non-linear non-Gaussian dynamic models. The methods rely on a particle cloud representation of the filtering distribution which evolves through time using importance sampling and resampling ideas. MAP sequence estimation is then performed using a classical dynamic programming technique applied to the discretised version of the state space. In contrast with standard approaches to the problem which essentially compare only the trajectories generated directly during the filtering stage, our method efficiently computes the optimal trajectory over all combinations of the filtered states. A particular strength of the method is that MAP sequence estimation is performed sequentially in one single forwards pass through the data without the requirement of an additional backward sweep. An application to estimation of a non-linear time series model and to spectral estimation for time-varying autoregressions is described.
Similar content being viewed by others
References
Aguilar, O. and West, M. (2000). Bayesian dynamic factor models and variance matrix discounting for portfolio allocation, J. Bus. Econom. Statist., 18.
Bar-Shalom, Y. and Li, X. (1995). Multitarget-Multisensor Tracking: Principles and Techniques, Academic Press, New York.
Doucet, A., Godsill, S. J. and Andrieu, C. (2000a). On sequential Monte Carlo sampling methods for Bayesian filtering, Statist. Comput., 10, 197-208.
Doucet, A., Godsill, S. J. and West, M. (2000b). Monte Carlo filtering and smoothing with application to time-varying spectral estimation, Proc. IEEE International Conference on Acoustics, Speech and Signal Processing, Vol. II, 701-704.
Doucet, A., De Freitas, J. F. G. and Gordon, N. J. (eds.) (2001). Sequential Monte Carlo Methods in Practice, Springer, New York.
Forney, G. D. (1973). The Viterbi algorithm, Proceedings of the IEEE, 61(3), 268-278.
Godsill, S. J. (1997). Bayesian enhancement of speech and audio signals which can be modelled as ARMA processes, International Statistical Review, 65(1), 1-21.
Godsill, S. J. and Clapp, T. C. (2001). Improvement strategies for Monte Carlo particle filters, Sequential Monte Carlo Methods in Practice (eds. A. Doucet, J. F. G. De Freitas and N. J. Gordon), Springer-Verlag, New York.
Godsill, S. J. and Rayner, P. J. W. (1998a). Digital Audio Restoration: A Statistical Model-Based Approach, Springer, Berlin.
Godsill, S. J. and Rayner, P. J. W. (1998b). Robust reconstruction and analysis of autoregressive signals in impulsive noise using the Gibbs sampler, IEEE Transactions on Speech and Audio Processing, 6(4), 352-372.
Godsill, S. J., Doucet, A. and West, M. (2000). Monte Carlo smoothing for non-linear time series (Submitted).
Gordon, N. J., Salmond, D. J. and Smith, A. F. M. (1993). Novel approach to nonlinear/non-Gaussian Bayesian state estimation, IEE Proceedings-F, 140(2), 107-113.
Higuchi, T. (1997). Monte Carlo filtering using the genetic algorithm operators, J. Statist. Comput. Simulation, 59, 1-23.
Hürzeler, M. and Künsch, H. R. (2000). Monte Carlo approximations for general state space models, J. Comput. Graph. Statist., 7(2), 175-193.
Kitagawa, G. (1987). Non-Gaussian state-space modeling of nonstationary time series, J. Amer. Statist. Assoc., 82(400), 1032-1063.
Kitagawa, G. (1996). Monte Carlo filter and smoother for nonlinear non-Gaussian state space models. J. Comput. Graph. Statist., 5, 1-25.
Kitagawa, G. and Gersch, W. (1996). Smoothness Priors Analysis of Time Series, Lecture Notes in Statist. No. 116, Springer, New York.
Levinson, S. E., Rabiner, L. R. and Sondhi, M. M. (1983). An introduction to the application of the theory of probabilistic functions of a Markov process to automatic speech recognition, Bell System Technical Journal, 62(4), 1035-1074.
Liu, J. and West, M. (2001). Combined parameter and state estimation in simulation-based filtering, Sequential Monte Carlo Methods in Practice (eds. A. Doucet, J. F. G. De Freitas and N. J. Gordon), Springer, New York.
Liu, J. S. and Chen, R. (1998). Sequential Monte Carlo methods for dynamical systems, J. Amer. Statist. Assoc., 93, 1032-1044.
Mendel, J. (1990). Maximum-Likelihood Deconvolution: A Journey into Model-based Signal Processing, Springer, New York.
Pitt, M. K. and Shephard, N. (1999). Filtering via simulation: Auxiliary particle filters, J. Amer. Statist. Assoc., 94, 590-599.
Prado, R., West, M. and Krystal, A. (1999). Evaluation and comparison of EEG traces: Latent structure in non-stationary time series, J. Amer. Statist. Assoc., 94, 1083-1095.
Vermaak, J., Andrieu, C., Doucet, A. and Godsill, S. J. (1999). On-line Bayesian modelling and enhancement of speech signals, Tech. Report CUED/F-INFENG/TR.361, Cambridge University, Engineering Department Cambridge, England.
Viterbi, A. J. (1967). Error bounds for convolutional codes and an asymptotically optimum decoding algorithm, IEEE Trans. Inform. Theory, IT-13, 260-269.
Author information
Authors and Affiliations
About this article
Cite this article
Godsill, S., Doucet, A. & West, M. Maximum a Posteriori Sequence Estimation Using Monte Carlo Particle Filters. Annals of the Institute of Statistical Mathematics 53, 82–96 (2001). https://doi.org/10.1023/A:1017968404964
Issue Date:
DOI: https://doi.org/10.1023/A:1017968404964