Abstract
We propose new methods for goal and plan recognition based on prefix and infix probability computation in a probabilistic context-free grammar (PCFG) which are both applicable to incomplete data. We define goal recognition as a task of identifying a goal from an action sequence and plan recognition as that of discovering a plan for the goal consisting of goal-subgoal structure respectively. To achieve these tasks, in particular from incomplete data such as sentences in a PCFG that often occurs in applications, we introduce prefix and infix probability computation via parse trees in PCFGs and compute the most likely goal and plan from incomplete data by considering them as prefixes and infixes.
We applied our approach to web session logs taken from the Internet Traffic Archive whose goal and plan recognition is important to improve websites. We tackled the problem of goal recognition from incomplete logs and empirically demonstrated the superiority of our approach compared to other approaches which do not use parsing. We also showed that it is possible to estimate the most likely plans from incomplete logs. All prefix and infix probability computation together with the computation of the most likely goal and plan in this paper is carried out using logic-based modeling language PRISM.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
In this paper, we distinguish goal recognition and plan recognition; the former is a task of identifying a goal from actions but the latter means to discover a plan consisting of goal-subgoal structure to achieve the goal.
- 2.
- 3.
\(\mathrm{expl}_0(G)\) is equivalent to G in view of the distribution semantics. When convenient, we treat \(\mathrm{expl}_0(G)\) as a bag \(\{e_1, e_2, \cdots , e_k\}\) of explanations.
- 4.
This is justified because we assume the consistency of PCFGs [16] that implies the probability of remaining nonterminals in R yielding some terminal sequences is 1.
- 5.
probf/1 is a PRISM’s built-in predicate and displays an explanation graph.
- 6.
\(\mathtt{W} = 1\) because pre_pcfg([a],[a],[]) is logically proved without involving msws.
- 7.
Clustering was done by PRISM. We used a small CFG for clustering, containing 30 rules and 12 nonterminals, because clustering by a mixture of large PCFGs tends to suffer from very high memory usage. To build this grammar, we merged similar symbols such as InternalSearch and Search in the universal session grammar shown in Table 2.
- 8.
It is conducted on a PC with Core i7 Quad 2.67Â GHz, OpenSUSE 11.4 and 72Â GB main memory.
- 9.
We applied a PCFG to prefixes by pretending them to be sentences. In this experiment, we found that the universal session grammar fails to parse at most two sequences for each data set, so we can ignore these sequences.
- 10.
We used a left-to-right HMM where the number of states is varied from 2 to 8. In Fig. 4, only the highest accuracy is plotted for each k. Since logistic regression only accepts fixed length data, we prepare 19 logistic regression models, one for each length k \((2 \le k \le 20)\).
- 11.
We used PRISM to implement a mixture of HMMs and that of PCFG and also to compute prefix probability. For the implementation of logistic regression we used the ‘nnet’ package of R.
- 12.
The entropy is defined as \(- \sum _{\tau } P(\tau )\log P(\tau )\) where \(\tau \) is a possible parse tree [2]. In our setting, a common grammar, the universal session grammar, is used for all data sets. So the entropy only depends on the parameters of a PCFG learned from the data set.
- 13.
This is to simulates a state transition of \(\mathsf{FA}\) made by a string derived from the nonterminal A using \(A \rightarrow B C\).
- 14.
References
Arlitt, M.F., Williamson, C.L.: Web server workload characterization: the search for invariants. ACM SIGMETRICS Perform. Eval. Rev. 24, 126–137 (1996)
Chi, Z.: Statistical properties of probabilistic context-free grammars. Comput. Linguist. 25(1), 131–160 (1999)
De Raedt, L., Kimmig, A., Toivonen, H.: Problog: a probabilistic prolog and its application in link discovery. IJCAI 7, 2462–2467 (2007)
Goodman, B.A., Litman, D.J.: On the interaction between plan recognition and intelligent interfaces. User Model. User-Adapt. Interact. 2(1–2), 83–115 (1992)
Horvitz, E., Breese, J., Heckerman, D., Hovel, D., Rommelse, K.: The lumiere project: Bayesian user modeling for inferring the goals and needs of software users. In: Proceedings of the Fourteenth Conference on Uncertainty in Artificial Intelligence, pp. 256–265. Morgan Kaufmann Publishers Inc. (1998)
Huang, L.: Advanced dynamic programming in semiring and hypergraph frameworks. In: COLING (2008)
The Internet Traffic Archive (2001). http://ita.ee.lbl.gov/
Jelinek, F., Lafferty, J.D.: Computation of the probability of initial substring generation by stochastic context-free grammars. Comput. Linguist. 17(3), 315–323 (1991)
Lesh, N., Rich, C., Sidner, C.L.: Using plan recognition in human-computer collaboration. In: Kay, J. (ed.) UM99 User Modeling. CISM International Centre for Mechanical Sciences - Courses and Lectures, pp. 23–32. Springer, Vienna (1999)
Nederhof, M.J., Satta, G.: Computation of infix probabilities for probabilistic context-free grammars. In: Proceedings of the Conference on Empirical Methods in Natural Language Processing, pp. 1213–1221. Association for Computational Linguistics (2011)
Sato, T.: A statistical learning method for logic programs with distribution semantics. In: Proceedings of Internationall Conference on Logic Programming, ILCP 1995 (1995)
Sato, T., Kameya, Y.: Parameter learning of logic programs for symbolic-statistical modeling. J. Artif. Intell. Res. 15, 391–454 (2001)
Sato, T., Kameya, Y.: New advances in logic-based probabilistic modeling by PRISM. In: De Raedt, L., Frasconi, P., Kersting, K., Muggleton, S.H. (eds.) Probabilistic Inductive Logic Programming. LNCS (LNAI), vol. 4911, pp. 118–155. Springer, Heidelberg (2008)
Sato, T., Meyer, P.: Infinite probability computation by cyclic explanation graphs. Theor. Pract. Logic Program. 15, 1–29 (2013)
Stolcke, A.: An efficient probabilistic context-free parsing algorithm that computes prefix probabilities. Comput. Linguist. 21(2), 165–201 (1995)
Wetherell, C.S.: Probabilistic languages: a review and some open questions. ACM Comput. Surv. (CSUR) 12(4), 361–379 (1980)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this paper
Cite this paper
Kojima, R., Sato, T. (2015). Goal and Plan Recognition via Parse Trees Using Prefix and Infix Probability Computation. In: Davis, J., Ramon, J. (eds) Inductive Logic Programming. Lecture Notes in Computer Science(), vol 9046. Springer, Cham. https://doi.org/10.1007/978-3-319-23708-4_6
Download citation
DOI: https://doi.org/10.1007/978-3-319-23708-4_6
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-23707-7
Online ISBN: 978-3-319-23708-4
eBook Packages: Computer ScienceComputer Science (R0)