Abstract
A central problem in knowledge-based tasks is to provide a collection of reusable knowledge samples extracted from a textual corpus. Often, such corpora are structured into different documents or topics, respectively. The samples need to be proven for usability and adapted by a domain expert requiring a certain processing time for each sample taken. The goal is to achieve an optimal retrieval and adaptation success meeting the time budget of the domain expert. In this work, we formulate this task as a constrained multi-armed bandit model. We combine it with the model of a configurable data-driven case-based learning agent. A case study evaluates the theoretical considerations in a scenario of regulatory knowledge acquisition. Therefore, a data set is constructed out of a corpus of nuclear safety documents. We use the model to optimize the evaluation process of sample generation of adaptational knowledge. The corresponding knowledge graph has been created in an information extraction step by automatically identifying semantic concepts and their relations.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
International Atomic Energy Association. https://www.iaea.org
Alon, N., Cesa-Bianchi, N., Gentile, C., Mannor, S., Mansour, Y., Shamir, O.: Nonstochastic multi-armed bandits with graph-structured feedback. SIAM J. Comput. 46, 1785–1826 (2014)
Auer, P.: Using confidence bounds for exploitation-exploration trade-offs. J. Mach. Learn. Res. 3, 397–422 (2002)
Auer, P., Cesa-Bianchi, N., Freund, Y., Schapire, R.: The nonstochastic multiarmed bandit problem. SIAM J. Comput. 32, 48–77 (2003)
Banko, M., Cafarella, M.J., Soderland, S., Broadhead, M., Etzioni, O.: Open information extraction from the web. In: Proceedings of the 20th International Joint Conference on Artificial Intelligence, IJCAI 2007, pp. 2670–2676. Morgan Kaufmann Publishers Inc., San Francisco (2007)
Bengs, V., Hüllermeier, E.: Multi-armed bandits with censored consumption of resources. Mach. Learn. 112(1), 217–240 (2023). https://doi.org/10.1007/s10994-022-06271-z
Bergmann, R.: Experience Management. Springer, Heidelberg (2002). https://doi.org/10.1007/3-540-45759-3
Brändle, F., Binz, M., Schulz, E.: Exploration beyond bandits, pp. 147–168. Cambridge University Press, Cambridge (2022)
Carlsson, E., Dubhashi, D.P., Johansson, F.D.: Thompson sampling for bandits with clustered arms. In: IJCAI International Joint Conference on Artificial Intelligence (2021)
Hartig, O.: Foundations of RDF* and SPARQL*: (an alternative approach to statement-level metadata in RDF). In: Alberto Mendelzon Workshop on Foundations of Data Management (2017)
Hearst, M.A.: Automatic acquisition of hyponyms from large text corpora. In: COLING 1992 Volume 2: The 15th International Conference on Computational Linguistics, pp. 539–545 (1992)
Hong, J., Kveton, B., Katariya, S., Zaheer, M., Ghavamzade, M.: Deep hierarchy in bandits. In: ICML International Conference on Machine Learning (2022)
Korger, A., Baumeister, J.: The SECCO ontology for the retrieval and generation of security concepts. In: Cox, M.T., Funk, P., Begum, S. (eds.) ICCBR 2018. LNCS (LNAI), vol. 11156, pp. 186–201. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01081-2_13
Korger, A., Baumeister, J.: Case-based generation of regulatory documents and their semantic relatedness. In: Arai, K., Kapoor, S., Bhatia, R. (eds.) FICC 2020. AISC, vol. 1130, pp. 91–110. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-39442-4_9
Kumar, S., Gao, H., Wang, C., Chang, K., Sundaram, H.: Hierarchical multi-armed bandits for discovering hidden populations. In: ASONAM 2019: International Conference on Advances in Social Networks Analysis and Mining, pp. 145–153, August 2019
Langford, J., Zhang, T.: The epoch-greedy algorithm for contextual multi-armed bandits. In: Proceedings of the 20th International Conference on Neural Information Processing Systems, NIPS 2007, pp. 817–824. Curran Associates Inc., Red Hook (2007)
Losada, D.E., Parapar, J., Barreiro, Á.: Multi-armed bandits for adjudicating documents in pooling-based evaluation of information retrieval systems. Inf. Process. Manag. 53, 1005–1025 (2017)
Perotto, F.S., Verstaevel, N., Trabelsi, I., Vercouter, L.: Combining bandits and lexical analysis for document retrieval in a juridical corpora. In: Bramer, M., Ellis, R. (eds.) SGAI 2020. LNCS (LNAI), vol. 12498, pp. 317–330. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-63799-6_24
Racharak, T., Suntisrivaraporn, B., Tojo, S.: sim\(^\pi \): a concept similarity measure under an agent’s preferences in description logic ELH. In: 8th International Conference on Agents and Artificial Intelligence, pp. 480–487, January 2016
Rahman, M.M., Kutlu, M., Lease, M.: Constructing test collections using multi-armed bandits and active learning. In: The Web Conference, San Francisco, May 2019
Robbins, H.E.: Some aspects of the sequential design of experiments. Bull. Am. Math. Soc. 58, 527–535 (1952)
Schelling, T.C.: Dynamic models of segregation. J. Math. Sociol. 1(2), 143–186 (1971)
Schulz, E., Franklin, N., Gershman, S.: Finding structure in multi-armed bandits. Cogn. Psychol. 119, 101261 (2020)
Sen, R., et al.: Top-\(k\) extreme contextual bandits with arm hierarchy, February 2021
Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction, 2nd edn. The MIT Press, Cambridge (2018)
Thompson, W.R.: On the likelihood that one unknown probability exceeds another in view of the evidence of two samples. Biometrika 25, 285–94 (1933)
Vaswani, A., et al.: Attention is all you need. In: Guyon, I., et al. (eds.) Advances in Neural Information Processing Systems, vol. 30. Curran Associates, Inc. (2017)
W3C: SKOS Simple Knowledge Organization System Reference, August 2009. http://www.w3.org/TR/skos-reference
Wood, D., Lanthaler, M., Cyganiak, R.: RDF 1.1 concepts and abstract syntax, February 2014. http://www.w3.org/TR/2014/REC-rdf11-concepts-20140225/
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Korger, A., Baumeister, J. (2023). Case-Based Sample Generation Using Multi-Armed Bandits. In: Massie, S., Chakraborti, S. (eds) Case-Based Reasoning Research and Development. ICCBR 2023. Lecture Notes in Computer Science(), vol 14141. Springer, Cham. https://doi.org/10.1007/978-3-031-40177-0_8
Download citation
DOI: https://doi.org/10.1007/978-3-031-40177-0_8
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-40176-3
Online ISBN: 978-3-031-40177-0
eBook Packages: Computer ScienceComputer Science (R0)