Skip to main content

Discrete sequence prediction with commented Markov models

  • Session: Interference of Stochastic Models 1
  • Conference paper
  • First Online:
Grammatical Interference: Learning Syntax from Sentences (ICGI 1996)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 1147))

Included in the following conference series:

  • 140 Accesses

Abstract

This paper introduces the concept of Commented Markov Models (CMMs), an extension of the well known Markov Models, together with the relevant induction mechanisms. Given a discrete alphabet Σ and a source producing an input sequence (s 1,s 2,...,s n) with s iΣ, the task of sequence prediction is to guess the successive sequence element s n+1. Here each element s i may represent an object, a discrete event or any other discrete entity. Prediction with CMM is analogy-based. It is assumed that the final part of the input sequence describes the current state of the source. This final part is matched with earlier subsequences of the input, assuming that it will be continued the same way as was the ‘most similar’ subsequence. CMM learning involves the induction of objects, variables and object classes. While object and class creation are similar to the notions of chunking and merging in other grammatical inference approaches, the use of variables is a novel feature of CMM. It not only generalized the way subsequences can be matched, it also turns CMM from a pure sequence prediction algorithm into a computational model. I will show that CMM has sufficient expressiveness to represent any primitive recursive function. Thus it is not only capable of predicting e.g. the character ‘u’ to follow the sequence ‘seq’, but it can also extrapolate a sequence like ‘45+13=’ by calculating the sum ‘58’.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Aho, A.V.: Algorithms for Finding Patterns in Strings. Handbook of Theoretical Computer Science, J. van Leeuwen, ed., Elsevier Science Publishers B.V. (1990)

    Google Scholar 

  2. Brown, P.F., Della Pietra, V.J., DeSouza, P.V., Lai, J.C., Mercer, R.L.: Class-based n-gram models of natural language. Computational Linguistics 18 (1992) 467–479

    Google Scholar 

  3. Crutchf ield, J.P.: The Calculi of Emergence: Computation, Dynamics and Induction. Physica D 75 (1994) 11–54

    Google Scholar 

  4. Hofstadter, D.R.: Fluid Concepts and Creative Analogies. Basic Books, NY (1995)

    Google Scholar 

  5. Hopcroft, J.E., Ullman, J.D.: Introduction to Automata Theory, Languages and Computation. Addison-Wesley, Reading, Mass. (1979)

    Google Scholar 

  6. Laird, P.: Discrete sequence prediction and its applications. Proceedings of the 9th National Conference on Artificial Intelligence, AAAI (1992)

    Google Scholar 

  7. Laird, P. and Saul, R.: Sequence Extrapolation. Proc. of the IJCAI (1993)

    Google Scholar 

  8. Michell, M.: Analogy-Making as Perception. Bradford Books, Cambr., MA (1993)

    Google Scholar 

  9. Ron, D., Singer, Y., Tishby, N.: The power of Amnesia. Advances in NIPS 6, Cowan, J., Tesauro, G., Alspector, J., eds. Morgan Kaufmann, San Mateo, (1994)

    Google Scholar 

  10. Stolcke, A.: Bayesian Learning of Probabilistic Language Models. University of California dissertation, Berkeley, CA (1994)

    Google Scholar 

  11. Williams, R.: Dynamic history predictive compression. Inf. Syst. 13(1) (1988)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Laurent Miclet Colin de la Higuera

Rights and permissions

Reprints and permissions

Copyright information

© 1996 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Blasig, R. (1996). Discrete sequence prediction with commented Markov models. In: Miclet, L., de la Higuera, C. (eds) Grammatical Interference: Learning Syntax from Sentences. ICGI 1996. Lecture Notes in Computer Science, vol 1147. Springer, Berlin, Heidelberg. https://doi.org/10.1007/BFb0033354

Download citation

  • DOI: https://doi.org/10.1007/BFb0033354

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-61778-5

  • Online ISBN: 978-3-540-70678-6

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics