Abstract
Meta-Interpretive Learning (MIL) is an ILP technique which uses higher-order meta-rules to support predicate invention and learning of recursive definitions. In MIL the selection of meta-rules is analogous to the choice of refinement operators in a refinement graph search. The meta-rules determine the structure of permissible rules which in turn defines the hypothesis space. On the other hand, the hypothesis space can be shown to increase rapidly in the number of meta-rules. However, methods for reducing the set of meta-rules have so far not been explored within MIL. In this paper we demonstrate that irreducible, or minimal sets of meta-rules can be found automatically by applying Plotkin’s clausal theory reduction algorithm. When this approach is applied to a set of meta-rules consisting of an enumeration of all meta-rules in a given finite hypothesis language we show that in some cases as few as two meta-rules are complete and sufficient for generating all hypotheses. In our experiments we compare the effect of using a minimal set of meta-rules to randomly chosen subsets of the maximal set of meta-rules. In general the minimal set of meta-rules leads to lower runtimes and higher predictive accuracies than larger randomly selected sets of meta-rules.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
\(H^{i}_{j}\) consists of definite datalog logic programs with predicates of arity at most i and at most j atoms in the body of each clause.
- 2.
It should be noted that MIL uses example driven test-incorporation for finding consistent programs as opposed to the generate-and-test approach of clause refinement.
- 3.
- 4.
References
Muggleton, S.H., Fidjeland, A., Luk, W.: Scalable acceleration of inductive logic programs. In IEEE international conference on field-programmable technology, pp. 252–259. IEEE (2002)
Blockeel, H., Dehaspe, L., Demoen, B., Janssens, G., Ramon, J., Vandecasteele, H.: Improving the efficiency of inductive logic programming through the use of query packs. J. Artif. Intell. Res. 16(1), 135–166 (2002)
Blumer, A., Ehrenfeucht, A., Haussler, D., Warmuth, M.K.: Learnability and the Vapnik-Chervonenkis dimension. J. ACM 36(4), 929–965 (1989)
Carlson, A., Betteridge, J., Kisiel, B., Settles, B., Hruschka Jr., E.R., Mitchell, T.M.: Toward an architecture for never-ending language learning. In: Proceedings of the Twenty-Fourth Conference on Artificial Intelligence (AAAI 2010) (2010)
Cohen, W.: Grammatically biased learning: learning logic programs using an explicit antecedent description language. Artif. Intell. 68, 303–366 (1994)
De Raedt, L.: Declarative modeling for machine learning and data mining. In: Bshouty, N.H., Stoltz, G., Vayatis, N., Zeugmann, T. (eds.) ALT 2012. LNCS, vol. 7568, pp. 12–12. Springer, Heidelberg (2012)
Hinton, G.E.: Learning distributed representations of concepts. Artif. Intell. 40, 1–12 (1986)
Lin, D., Dechter, E., Ellis, K., Tenenbaum, J.B., Muggleton, S.H.: Bias reformulation for one-shot function induction. In: Proceedings of the 23rd European Conference on Artificial Intelligence (ECAI 2014), pp. 525–530. IOS Press, Amsterdam (2014)
Muggleton, S.H.: Inverse entailment and progol. New Gener. Comput. 13, 245–286 (1995)
S.H. Muggleton and D. Lin. Meta-interpretive learning of higher-order dyadic datalog: Predicate invention revisited. In: Proceedings of the 23rd International Joint Conference Artificial Intelligence (IJCAI 2013), pp. 1551–1557 (2013)
Muggleton, S.H., Lin, D., Pahlavi, N., Tamaddoni-Nezhad, A.: Meta-interpretive learning: application to grammatical inference. Mach. Learn. 94, 25–49 (2014)
Muggleton, S.H., Lin, D., Tamaddoni-Nezhad, A.: Meta-interpretive learning of higher-order dyadic datalog: predicate invention revisited. Mach. Learn. 100(1), 49–73 (2015)
Nienhuys-Cheng, S.-H., de Wolf, R.: Foundations of Inductive Logic Programming. LNCS (LNAI), vol. 1228. Springer, Heidelberg (1997)
G.D. Plotkin. Automatic methods of inductive inference. PhD thesis, Edinburgh University, August 1971
Shapiro, E.Y.: Algorithmic Program Debugging. MIT Press, Cambridge (1983)
A. Srinivasan. A study of two probabilistic methods for searching large spaces with ilp. Technical report PRG-TR-16-00, Oxford University Computing Laboratory, Oxford (2000)
Srinivasan, A.: The ALEPH manual, Machine Learning at the Computing Laboratory. Oxford University (2001)
Acknowledgements
The first author acknowledges the support of the BBSRC and Syngenta in funding his PhD Case studentship. The second author would like to thank the Royal Academy of Engineering and Syngenta for funding his present 5 year Research Chair.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this paper
Cite this paper
Cropper, A., Muggleton, S.H. (2015). Logical Minimisation of Meta-Rules Within Meta-Interpretive Learning. In: Davis, J., Ramon, J. (eds) Inductive Logic Programming. Lecture Notes in Computer Science(), vol 9046. Springer, Cham. https://doi.org/10.1007/978-3-319-23708-4_5
Download citation
DOI: https://doi.org/10.1007/978-3-319-23708-4_5
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-23707-7
Online ISBN: 978-3-319-23708-4
eBook Packages: Computer ScienceComputer Science (R0)