Skip to main content

Goal and Plan Recognition via Parse Trees Using Prefix and Infix Probability Computation

  • Conference paper
  • First Online:
Inductive Logic Programming

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 9046))

  • 416 Accesses

Abstract

We propose new methods for goal and plan recognition based on prefix and infix probability computation in a probabilistic context-free grammar (PCFG) which are both applicable to incomplete data. We define goal recognition as a task of identifying a goal from an action sequence and plan recognition as that of discovering a plan for the goal consisting of goal-subgoal structure respectively. To achieve these tasks, in particular from incomplete data such as sentences in a PCFG that often occurs in applications, we introduce prefix and infix probability computation via parse trees in PCFGs and compute the most likely goal and plan from incomplete data by considering them as prefixes and infixes.

We applied our approach to web session logs taken from the Internet Traffic Archive whose goal and plan recognition is important to improve websites. We tackled the problem of goal recognition from incomplete logs and empirically demonstrated the superiority of our approach compared to other approaches which do not use parsing. We also showed that it is possible to estimate the most likely plans from incomplete logs. All prefix and infix probability computation together with the computation of the most likely goal and plan in this paper is carried out using logic-based modeling language PRISM.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    In this paper, we distinguish goal recognition and plan recognition; the former is a task of identifying a goal from actions but the latter means to discover a plan consisting of goal-subgoal structure to achieve the goal.

  2. 2.

    The probability of a prefix in a PCFG is defined to be the sum of probabilities of infinitely many sentences extending it and computed by solving a set of linear equations derived from the CFG [8]. Also there is prefix probability computation based on probabilistic Earley parsing [15].

  3. 3.

    \(\mathrm{expl}_0(G)\) is equivalent to G in view of the distribution semantics. When convenient, we treat \(\mathrm{expl}_0(G)\) as a bag \(\{e_1, e_2, \cdots , e_k\}\) of explanations.

  4. 4.

    This is justified because we assume the consistency of PCFGs [16] that implies the probability of remaining nonterminals in R yielding some terminal sequences is 1.

  5. 5.

    probf/1 is a PRISM’s built-in predicate and displays an explanation graph.

  6. 6.

    \(\mathtt{W} = 1\) because pre_pcfg([a],[a],[]) is logically proved without involving msws.

  7. 7.

    Clustering was done by PRISM. We used a small CFG for clustering, containing 30 rules and 12 nonterminals, because clustering by a mixture of large PCFGs tends to suffer from very high memory usage. To build this grammar, we merged similar symbols such as InternalSearch and Search in the universal session grammar shown in Table 2.

  8. 8.

    It is conducted on a PC with Core i7 Quad 2.67 GHz, OpenSUSE 11.4 and 72 GB main memory.

  9. 9.

    We applied a PCFG to prefixes by pretending them to be sentences. In this experiment, we found that the universal session grammar fails to parse at most two sequences for each data set, so we can ignore these sequences.

  10. 10.

    We used a left-to-right HMM where the number of states is varied from 2 to 8. In Fig. 4, only the highest accuracy is plotted for each k. Since logistic regression only accepts fixed length data, we prepare 19 logistic regression models, one for each length k \((2 \le k \le 20)\).

  11. 11.

    We used PRISM to implement a mixture of HMMs and that of PCFG and also to compute prefix probability. For the implementation of logistic regression we used the ‘nnet’ package of R.

  12. 12.

    The entropy is defined as \(- \sum _{\tau } P(\tau )\log P(\tau )\) where \(\tau \) is a possible parse tree [2]. In our setting, a common grammar, the universal session grammar, is used for all data sets. So the entropy only depends on the parameters of a PCFG learned from the data set.

  13. 13.

    This is to simulates a state transition of \(\mathsf{FA}\) made by a string derived from the nonterminal A using \(A \rightarrow B C\).

  14. 14.

    http://rjida.meijo-u.ac.jp/prism/.

References

  1. Arlitt, M.F., Williamson, C.L.: Web server workload characterization: the search for invariants. ACM SIGMETRICS Perform. Eval. Rev. 24, 126–137 (1996)

    Article  Google Scholar 

  2. Chi, Z.: Statistical properties of probabilistic context-free grammars. Comput. Linguist. 25(1), 131–160 (1999)

    MathSciNet  Google Scholar 

  3. De Raedt, L., Kimmig, A., Toivonen, H.: Problog: a probabilistic prolog and its application in link discovery. IJCAI 7, 2462–2467 (2007)

    Google Scholar 

  4. Goodman, B.A., Litman, D.J.: On the interaction between plan recognition and intelligent interfaces. User Model. User-Adapt. Interact. 2(1–2), 83–115 (1992)

    Article  Google Scholar 

  5. Horvitz, E., Breese, J., Heckerman, D., Hovel, D., Rommelse, K.: The lumiere project: Bayesian user modeling for inferring the goals and needs of software users. In: Proceedings of the Fourteenth Conference on Uncertainty in Artificial Intelligence, pp. 256–265. Morgan Kaufmann Publishers Inc. (1998)

    Google Scholar 

  6. Huang, L.: Advanced dynamic programming in semiring and hypergraph frameworks. In: COLING (2008)

    Google Scholar 

  7. The Internet Traffic Archive (2001). http://ita.ee.lbl.gov/

  8. Jelinek, F., Lafferty, J.D.: Computation of the probability of initial substring generation by stochastic context-free grammars. Comput. Linguist. 17(3), 315–323 (1991)

    Google Scholar 

  9. Lesh, N., Rich, C., Sidner, C.L.: Using plan recognition in human-computer collaboration. In: Kay, J. (ed.) UM99 User Modeling. CISM International Centre for Mechanical Sciences - Courses and Lectures, pp. 23–32. Springer, Vienna (1999)

    Chapter  Google Scholar 

  10. Nederhof, M.J., Satta, G.: Computation of infix probabilities for probabilistic context-free grammars. In: Proceedings of the Conference on Empirical Methods in Natural Language Processing, pp. 1213–1221. Association for Computational Linguistics (2011)

    Google Scholar 

  11. Sato, T.: A statistical learning method for logic programs with distribution semantics. In: Proceedings of Internationall Conference on Logic Programming, ILCP 1995 (1995)

    Google Scholar 

  12. Sato, T., Kameya, Y.: Parameter learning of logic programs for symbolic-statistical modeling. J. Artif. Intell. Res. 15, 391–454 (2001)

    MathSciNet  MATH  Google Scholar 

  13. Sato, T., Kameya, Y.: New advances in logic-based probabilistic modeling by PRISM. In: De Raedt, L., Frasconi, P., Kersting, K., Muggleton, S.H. (eds.) Probabilistic Inductive Logic Programming. LNCS (LNAI), vol. 4911, pp. 118–155. Springer, Heidelberg (2008)

    Chapter  Google Scholar 

  14. Sato, T., Meyer, P.: Infinite probability computation by cyclic explanation graphs. Theor. Pract. Logic Program. 15, 1–29 (2013)

    MATH  Google Scholar 

  15. Stolcke, A.: An efficient probabilistic context-free parsing algorithm that computes prefix probabilities. Comput. Linguist. 21(2), 165–201 (1995)

    MathSciNet  Google Scholar 

  16. Wetherell, C.S.: Probabilistic languages: a review and some open questions. ACM Comput. Surv. (CSUR) 12(4), 361–379 (1980)

    Article  MathSciNet  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ryosuke Kojima .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2015 Springer International Publishing Switzerland

About this paper

Cite this paper

Kojima, R., Sato, T. (2015). Goal and Plan Recognition via Parse Trees Using Prefix and Infix Probability Computation. In: Davis, J., Ramon, J. (eds) Inductive Logic Programming. Lecture Notes in Computer Science(), vol 9046. Springer, Cham. https://doi.org/10.1007/978-3-319-23708-4_6

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-23708-4_6

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-23707-7

  • Online ISBN: 978-3-319-23708-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics