Abstract
This paper introduces the concept of Commented Markov Models (CMMs), an extension of the well known Markov Models, together with the relevant induction mechanisms. Given a discrete alphabet Σ and a source producing an input sequence (s 1,s 2,...,s n) with s i ∈ Σ, the task of sequence prediction is to guess the successive sequence element s n+1. Here each element s i may represent an object, a discrete event or any other discrete entity. Prediction with CMM is analogy-based. It is assumed that the final part of the input sequence describes the current state of the source. This final part is matched with earlier subsequences of the input, assuming that it will be continued the same way as was the ‘most similar’ subsequence. CMM learning involves the induction of objects, variables and object classes. While object and class creation are similar to the notions of chunking and merging in other grammatical inference approaches, the use of variables is a novel feature of CMM. It not only generalized the way subsequences can be matched, it also turns CMM from a pure sequence prediction algorithm into a computational model. I will show that CMM has sufficient expressiveness to represent any primitive recursive function. Thus it is not only capable of predicting e.g. the character ‘u’ to follow the sequence ‘seq’, but it can also extrapolate a sequence like ‘45+13=’ by calculating the sum ‘58’.
Preview
Unable to display preview. Download preview PDF.
References
Aho, A.V.: Algorithms for Finding Patterns in Strings. Handbook of Theoretical Computer Science, J. van Leeuwen, ed., Elsevier Science Publishers B.V. (1990)
Brown, P.F., Della Pietra, V.J., DeSouza, P.V., Lai, J.C., Mercer, R.L.: Class-based n-gram models of natural language. Computational Linguistics 18 (1992) 467–479
Crutchf ield, J.P.: The Calculi of Emergence: Computation, Dynamics and Induction. Physica D 75 (1994) 11–54
Hofstadter, D.R.: Fluid Concepts and Creative Analogies. Basic Books, NY (1995)
Hopcroft, J.E., Ullman, J.D.: Introduction to Automata Theory, Languages and Computation. Addison-Wesley, Reading, Mass. (1979)
Laird, P.: Discrete sequence prediction and its applications. Proceedings of the 9th National Conference on Artificial Intelligence, AAAI (1992)
Laird, P. and Saul, R.: Sequence Extrapolation. Proc. of the IJCAI (1993)
Michell, M.: Analogy-Making as Perception. Bradford Books, Cambr., MA (1993)
Ron, D., Singer, Y., Tishby, N.: The power of Amnesia. Advances in NIPS 6, Cowan, J., Tesauro, G., Alspector, J., eds. Morgan Kaufmann, San Mateo, (1994)
Stolcke, A.: Bayesian Learning of Probabilistic Language Models. University of California dissertation, Berkeley, CA (1994)
Williams, R.: Dynamic history predictive compression. Inf. Syst. 13(1) (1988)
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 1996 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Blasig, R. (1996). Discrete sequence prediction with commented Markov models. In: Miclet, L., de la Higuera, C. (eds) Grammatical Interference: Learning Syntax from Sentences. ICGI 1996. Lecture Notes in Computer Science, vol 1147. Springer, Berlin, Heidelberg. https://doi.org/10.1007/BFb0033354
Download citation
DOI: https://doi.org/10.1007/BFb0033354
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-61778-5
Online ISBN: 978-3-540-70678-6
eBook Packages: Springer Book Archive