Meaning in Artificial Agents: The Symbol Grounding Problem Revisited
- 432 Downloads
The Chinese room argument has presented a persistent headache in the search for Artificial Intelligence. Since it first appeared in the literature, various interpretations have been made, attempting to understand the problems posed by this thought experiment. Throughout all this time, some researchers in the Artificial Intelligence community have seen Symbol Grounding as proposed by Harnad as a solution to the Chinese room argument. The main thesis in this paper is that although related, these two issues present different problems in the framework presented by Harnad himself. The work presented here attempts to shed some light on the relationship between John Searle’s intentionality notion and Harnad’s Symbol Grounding Problem.
KeywordsChinese room argument Symbol grounding problem
- Brentano, F. C. (1874). Psychology from an empirical standpoint. UK: Routledge.Google Scholar
- Davidsson, P. (1993). Toward a general solution to the symbol grounding problem: Combining machine learning and computer vision. In Machine learning and computer vision, in AAAI fall symposium series (pp. 157–161). Machine learning in computer vision: What, why and how, AAAI Press.Google Scholar
- Davidsson, P. (1996). Autonomous agents and the concept of concepts. Ph.D. thesis, Department of Computer Science, Lund University.Google Scholar
- Descartes, R. (2010). Prinicples of philosophy. Whitefish: Kessinger Publishing.Google Scholar
- Harnad, S. (1990). The symbol grounding problem. Physica D, 42, 335–346. http://cogprints.org/615..
- Harnad, S. (1992). There is only one mind/body problem. In Symposium on the perception of intentionality, XXV world congress of psychology, Brussels, Belgium.Google Scholar
- Harnad, S. (1999). The symbol grounding problem. CoRR cs.AI/9906002.Google Scholar
- Harnad, S. (2003). Symbol-grounding problem (Vol. LXVII). MacMillan: Nature Publishing Group. http://cogprints.org/3018.
- Honderich, T. (1995). The Oxford companion to philosophy. Oxford: Oxford University Press.Google Scholar
- Mayo, M. J. (2003). Symbol grounding and its implications for artificial intelligence. In ACSC ’03: Proceedings of the 26th Australasian computer science conference, Australian Computer Society, Inc., Darlinghurst, Australia, Australia, pp. 55–60.Google Scholar
- Rosenstein, M. T., & Cohen, P. R. (1998). Symbol grounding with delay coordinates. In In AAAI technical report WS-98-06, the grounding of word meaning: Data and models (pp. 20–21). Online. http://www.orl.co.uk/omniORB/omniORB.html.
- Searle, J. R. (1980). Minds, brains, and programs. The Behavioral and Brain Sciences, 3, 417–457+.Google Scholar
- Steels, L. (2008). The symbol grounding problem has been solved. so what’s next? Symbols, embodiment and meaning. New Haven: Academic Press. http://www.csl.sony.fr/downloads/papers/2007/steels-07a.pd.
- Witkowski, M. (2002). Anticipatory learning: The animat as discovery engine. In In M. V. Butz, P. G6rard, & O. Sigaud (Eds.), Adaptive Behavior in Anticipatory Learning Systems (ABiALS’02).Google Scholar