Skip to main content

An Approach to Human-Level Commonsense Reasoning

  • Chapter
  • First Online:
Book cover Paraconsistency: Logic and Applications

Part of the book series: Logic, Epistemology, and the Unity of Science ((LEUS,volume 26))

Abstract

Commonsense reasoning has proven exceedingly difficult both to model and to implement in artificial reasoning systems. This paper discusses some of the features of human reasoning that may account for this difficulty, surveys a number of reasoning systems and formalisms, and offers an outline of active logic, a non-classical paraconsistent logic that may be of some use in implementing commonsense reasoning.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 109.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    Note that we do not make the converse claim—that formal logic tends to aim at modelling reasoning processes (i.e., psychologism).

  2. 2.

    While details are complicated, most such approaches aim at the inference of special additional “normal” or “typical” formulas—such as Flies(x) from Bird(x)—when not ruled out by axioms.

  3. 3.

    Let us call a logic that is used by a real-world reasoning agent (human or otherwise) as it goes about its business an “on-board” logic (as opposed to a specification—“spec”—logic that characterises limiting behaviours such as the set of all sentences that (eventually) can be proven). Thus we are using the term “logic” quite broadly, to include any systematic method for drawing conclusions from premises.

  4. 4.

    So-called tense logics and temporal logics express propositions about past, present, and future, but the present is not represented as evolving: Now does not change as theorems are proved, in contrast with the above Clock Rule. In other words, tense logics are also spec logics, rather than on-board logics.

  5. 5.

    For an overview of the various findings reported in this paragraph, see Evans (1982).

  6. 6.

    Although this preference was reversed when the initial statement was a definition such as: if a mineral is a diamond then it is made of compressed carbon.

References

  • Anderson, J. 1990. The adaptive character of thought. Hillsdale: Lawrence Erlbaum.

    Google Scholar 

  • Anderson, J. 1991. Is human cognition adaptive? Behavioral and Brain Sciences 14(3): 471–517.

    Article  Google Scholar 

  • Anderson, J.R., D. Bothell, M.D. Byrne, S. Douglass, C. Lebiere, and Y. Qin. 2004. An integrated theory of mind. Psychological Review 111(4): 1036–1060.

    Article  Google Scholar 

  • Anderson, M.L., W. Gomaa, J. Grant, and D. Perlis. 2008. Active logic semantics for a single agent in a static world. Artificial Intelligence 172: 1045–1063.

    Article  Google Scholar 

  • Arieli, O. 2007. Paraconsistent reasoning and preferential entailments by signed quantified boolean formulae. ACM Transactions on Computational Logic 8(3): Article 18.

    Google Scholar 

  • Arieli, O., and A. Avron. 1998. The value of four values. Artificial Intelligence 102(1): 97–141.

    Article  Google Scholar 

  • Byrne, R. 1989. Suppressing valid inferences with conditionals. Cognition 31: 61–83.

    Article  Google Scholar 

  • Cosmides, L. 1989. The logic of social exchange: Has natural selection shaped how humans reason? Studies with the Wason selection task. Cognition 31: 187–276.

    Article  Google Scholar 

  • Elio, R. 1997. What to believe when inferences are contradicted: The impact of knowledge type and inference rule. In Proceedings of the nineteenth annual conference of the cognitive science society, 211–216. Hillsdale: Lawrence Erlbaum.

    Google Scholar 

  • Elio, R. 1998. How to disbelieve p → q: Resolving contradictions. In Proceedings of the twentieth annual conference of the cognitive science society, 315–320. Mahwah: Lawrence Erlbaum.

    Google Scholar 

  • Evans, J.S.B. 1982. The psychology of deductive reasoning. London: Routledge Keegan Paul.

    Google Scholar 

  • Gentner, D., and A. Stevens. 1983. Mental models. Hillsdale: Lawrence Erlbaum.

    Google Scholar 

  • Gigerenzer, G. 1994. Why the distinction between single-event probabilities and frequencies is important for psychology (and vice versa). In Subjective probability, ed. G. Write and P. Ayton, 129–161. New York: Wiley.

    Google Scholar 

  • Gray, W., C. Sims, W.T. Fu, and M. Schoelles. 2006. The soft constraints hypothesis: A rational analysis approach to resource allocation for interactive behavior. Psychological Review 113(3): 461–482.

    Article  Google Scholar 

  • Johnson-Laird, P. 1983. Mental models. Cambridge: Cambridge University Press.

    Google Scholar 

  • Johnson-Laird, P., and P. Wason. 1970. A theoretical insight into a reasoning task. Cognitive Psychology 1: 134–148.

    Article  Google Scholar 

  • Kahneman, D., and A. Tversky. 1979. Prospect theory: An analysis of decision under risk. Econometrica 47(2): 263–291.

    Article  Google Scholar 

  • Laird, J., A. Newell, and P. Rosenbloom. 1987. Soar: An architecture for general intelligence. Artificial Intelligence 33(1): 1–64.

    Article  Google Scholar 

  • Lebiere, C., and J.R. Anderson. 1993. A connectionist implementation of the ACT-R production system. In Proceedings of the fifteenth annual conference of the cognitive science society, 635–640. Hillsdale: Lawrence Erlbaum.

    Google Scholar 

  • Lenat, D., and R.V. Guha. 1990. Cyc: A midterm report. AI Magazine 11(3): 32–59.

    Google Scholar 

  • Lenat, D., R.V. Guha, K. Pittman, D. Pratt, and M. Shepherd. 1990. Cyc: Toward programs with commonsense. Communications of the ACM 33(8): 30–49.

    Article  Google Scholar 

  • Martins, A.T., M. Pequeno, and T. Pequeno. 2002. A multiple worlds semantics to a paraconsistent nonmonotonic logic. In Paraconsistency, the logical way to the inconsistent, ed. W.A. Carnielli, M.E. Coniglio, and I.M.L. D’Ottavio, 187–212. New York: Marcel Dekker.

    Google Scholar 

  • McCarthy, J., and P. Hayes. 1969. Some philosophical problems from the standpoint of artificial intelligence. In Machine intelligence, ed. B. Meltzer and D. Michie, 463–502. Edinburgh: Edinburgh University Press.

    Google Scholar 

  • Newell, A. (1990). Unified theories of cognition. Cambridge: Harvard University Press.

    Google Scholar 

  • Oaksford, M., and N. Chater. 1994. A rational analysis of the selection task as optimal data selection. Psychological Review 101: 608–631.

    Article  Google Scholar 

  • Oaksford, M., and N. Chater (ed.). 1999. Rational models of cognition. Oxford: Oxford University Press.

    Google Scholar 

  • Oaksford, M., and N. Chater. 2007. Bayesian rationality: The probabilistic approach to human reasoning. Oxford: Oxford University Press.

    Google Scholar 

  • Pollock, J. 1987. Defeasible reasoning. Cognitive Science 11: 481–518.

    Article  Google Scholar 

  • Pollock, J. 1989. How to build a person. Cambridge: Bradford/MIT.

    Google Scholar 

  • Pollock, J. 1991a. A theory of defeasible reasoning. International Journal of Intelligent Systems 6: 33–54.

    Article  Google Scholar 

  • Pollock, J. 1991b. Self-defeating arguments. Minds and Machines 1: 367–392.

    Article  Google Scholar 

  • Pollock, J. 1992. How to reason defeasibly. Artificial Intelligence 57: 1–42.

    Article  Google Scholar 

  • Pollock, J. 1995. Cognitive carpentry. Cambridge: Bradford/MIT.

    Google Scholar 

  • Pollock, J. 1999. Rational cognition in OSCAR. In Proceedings of ATAL-99, ed. N. Jennings and Y. Lesperance. Berlin: Springer.

    Google Scholar 

  • Priest, G. 1989. Reasoning about truth. Artificial Intelligence 39(2): 231–244.

    Article  Google Scholar 

  • Rosenbloom, P.S. (1989). A symbolic goal-oriented perspective on connectionism and soar. In Connectionism in perspective, ed. R. Pfeifer, Z. Schreter, F. Fogelman-Soulie, and L. Steels. Amsterdam: Elsevier.

    Google Scholar 

  • Rosenbloom, P.S., A. Newell, and J.E. Laird. 1990. Towards the knowledge level in soar: The role of the architecture in the use of knowledge. In Architectures for intelligence, ed. K. VanLehn. Hillsdale: Lawrence Erlbaum.

    Google Scholar 

  • Rosenbloom, P., J. Laird, A. Newell, and R. McCarl. 1991. A preliminary analysis of the soar architecture as a basis for general intelligence. Artificial Intelligence 47: 289–325.

    Article  Google Scholar 

  • Shapiro, S.C. 1979. The SNePS semantic network processing system. In Associative networks: The representation and use of knowledge by computers, ed. N. Findler, 179–203. New York: Academic.

    Google Scholar 

  • Shapiro, S.C. 1993. Belief spaces as sets of propositions. Journal of Experimental and Theoretical Artificial Intelligence 5: 225–235.

    Article  Google Scholar 

  • Shapiro, S.C. 2000. SNePS: A logic for natural language understanding and commonsense reasoning. In Natural language processing and knowledge representation: Language for knowledge and knowledge for language, ed. Ł. Iwańska and S.C. Shapiro, 175–195. Menlo Park: AAAI/MIT.

    Google Scholar 

  • Shapiro, S.C., and W. Rapaport. 1987. SNePS considered as a fully intensional propositional semantic network. In The knowledge frontier, ed. N. Cercone and G. McCalla, 263–315. New York: Springer.

    Google Scholar 

  • Shapiro, S.C., and W. Rapaport. 1992. The SNePS family. Computers and Mathematics with Applications 23: 243–275.

    Article  Google Scholar 

  • Shapiro, S.C., G.H. Woodmansee, and M.W Kreuger. 1968. A semantic associational memory net that learns and answers questions (SAMENLAQ). Technical report, Computer Science Department, University of Wisconsin, Madison, WI.

    Google Scholar 

  • Sharpe, D., and G. Lacroix. 1999. Reasoning about apparent contradictions: Resolution strategies and positive–negative asymmetries. Journal of Child Language 26(2): 477–490.

    Article  Google Scholar 

  • Taatgen, N.A., C. Lebiere, and J.R Anderson. 2006. Modeling paradigms in ACT-R. In Cognition and multi-agent interaction: From cognitive modeling to social simulation, ed. R. Sun, 29–52. Cambridge: Cambridge University Press.

    Google Scholar 

Download references

Acknowledgements

We would like to thank the referees for many helpful comments and suggestions.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Michael L. Anderson .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2013 Springer Science+Business Media Dordrecht.

About this chapter

Cite this chapter

Anderson, M.L., Gomaa, W., Grant, J., Perlis, D. (2013). An Approach to Human-Level Commonsense Reasoning. In: Tanaka, K., Berto, F., Mares, E., Paoli, F. (eds) Paraconsistency: Logic and Applications. Logic, Epistemology, and the Unity of Science, vol 26. Springer, Dordrecht. https://doi.org/10.1007/978-94-007-4438-7_12

Download citation

Publish with us

Policies and ethics