Abstract
We consider belief change in a context where information comes from reports, and the reporting agents may not be honest. In order to capture this process, we introduce an extended class of epistemic states that includes a history of past reports received. We present a set of postulates that describe how new reports should be incorporated. The postulates describe a new kind of belief change operator, where reported information can either be believed or ignored. We then provide a representation result for these postulates, which characterizes report revision in terms of an underlying set of agents that are perceived to be honest. We then extend our framework by adding observations. In this framework, observations are understood to be highly reliable. As such, when an observation conflicts with a report, we must question the honesty of the agent that provided the report. We introduce a flexible framework where we can set a threshold for the number of false reports an agent can send before they are labelled dishonest. Fundamental results are provided, along with a discussion on key future problems to be addressed in trust and belief revision.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Alchourrón, C.E., Gärdenfors, P., Makinson, D.: On the logic of theory change: partial meet functions for contraction and revision. J. Symb. Log. 50(2), 510–530 (1985)
Booth, R., Hunter, A.: Trust as a precursor to belief revision. J. Artif. Intell. Res. 61, 699–722 (2018)
Darwiche, A., Pearl, J.: On the logic of iterated belief revision. Artif. Intell. 89(1–2), 1–29 (1997)
Dong, X., et al.: Knowledge-based trust: estimating the trustworthiness of web sources. In: Proceedings of the VLDB Endowment, vol. 8 (2015)
Hunter, A.: Belief revision with dishonest reports. In: Aziz, H., Corrêa, D., French, T. (eds.) AI 2022. LNCS, vol. 13728, pp. 397–410. Springer, Cham (2022). https://doi.org/10.1007/978-3-031-22695-3_28
Huynh, T.D., Jennings, N.R., Shadbolt, N.R.: An integrated trust and reputation model for open multi-agent systems. Auton. Agent. Multi-Agent Syst. 13(2), 119–154 (2006)
Katsuno, H., Mendelzon, A.: Propositional knowledge base revision and minimal change. Artif. Intell. 52(2), 263–294 (1992)
Krukow, K., Nielsen, M.: Trust structures. Int. J. Inf. Secur. 6(2–3), 153–181 (2007)
Liu, F., Lorini, E.: Reasoning about belief, evidence and trust in a multi-agent setting. In: An, B., Bazzan, A., Leite, J., Villata, S., van der Torre, L. (eds.) PRIMA 2017. LNCS (LNAI), vol. 10621, pp. 71–89. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-69131-2_5
Schwind, N., Konieczny, S., Perez, R.P.: Darwiche and Pearl’s epistemic states are not total preorders. In: Proceedings of the International Conference on Principles of Knowledge Representation and Reasoning (KR 2022) (2022)
Singleton, J., Booth, R.: Who?s the expert? on multi-source belief change. In: Proceedings of the International Conference on Principles of Knowledge Representation and Reasoning (KR 2022) (2022)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Hunter, A. (2024). Reports, Observations, and Belief Change. In: Liu, T., Webb, G., Yue, L., Wang, D. (eds) AI 2023: Advances in Artificial Intelligence. AI 2023. Lecture Notes in Computer Science(), vol 14472. Springer, Singapore. https://doi.org/10.1007/978-981-99-8391-9_5
Download citation
DOI: https://doi.org/10.1007/978-981-99-8391-9_5
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-8390-2
Online ISBN: 978-981-99-8391-9
eBook Packages: Computer ScienceComputer Science (R0)