Pertinent Background Knowledge for Learning Protein Grammars
- Cite this paper as:
- Bryant C.H., Fredouille D.C., Wilson A., Jayawickreme C.K., Jupe S., Topp S. (2006) Pertinent Background Knowledge for Learning Protein Grammars. In: Fürnkranz J., Scheffer T., Spiliopoulou M. (eds) Machine Learning: ECML 2006. ECML 2006. Lecture Notes in Computer Science, vol 4212. Springer, Berlin, Heidelberg
We are interested in using Inductive Logic Programming (ILP) to infer grammars representing sets of protein sequences. ILP takes as input both examples and background knowledge predicates. This work is a first step in optimising the choice of background knowledge predicates for predicting the function of proteins. We propose methods to obtain different sets of background knowledge. We then study the impact of these sets on inference results through a hard protein function inference task: the prediction of the coupling preference of GPCR proteins. All but one of the proposed sets of background knowledge are statistically shown to have positive impacts on the predictive power of inferred rules, either directly or through interactions with other sets. In addition, this work provides further confirmation, after the work of Muggleton et al., 2001 that ILP can help to predict protein functions.
Unable to display preview. Download preview PDF.