Abstract
This paper outlines the non-behavioral Algorithmic Similarity criterion for machine intelligence, and assesses the likelihood that it will eventually be satisfied by computers programmed using Machine Learning (ML). Making this assessment requires overcoming the Black Box Problem, which makes it difficult to characterize the algorithms that are actually acquired via ML. This paper therefore considers Explainable AI’s prospects for solving the Black Box Problem, and for thereby providing a posteriori answers to questions about the possibility of machine intelligence. In addition, it suggests that the real-world nurture and situatedness of ML-programmed computers constitute a priori reasons for thinking that they will not only learn to behave like humans, but that they will also eventually acquire algorithms similar to the ones that are implemented in human brains.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
This appeal to brain-implemented algorithms is in line with cognitive science orthodoxy. Analogous non-behavioral criteria may invoke other theoretical posits.
- 2.
Those impressed by Searle’s (1980) Chinese Room would of course disagree. For them, the algorithms being executed have no bearing on the intelligence that may or may not be possessed.
- 3.
This is an admittedly strong empiricist thesis. Lest it be considered too strong, it is worth considering recent attempts to increase the psychological plausibility of ML methods. For example, Lake et al. (2017) review several ways in which such methods can be modified to incorporate known principles of human learning and development. This research may obviate the need for (overly) strong empiricism.
References
Block, N.: Psychologism and behaviorism. Philos. Rev. 90(1), 5–43 (1981)
Lake, B., Ullman, T., Tenenbaum, J., Gershman, S.: Building machines that learn and think like people. Behavi. Brain Sci. 40, E253 (2017). https://doi.org/10.1017/S0140525X16001837
Ribeiro, M.T., Singh, S., Guestrin, C.: “Why Should I Trust You?”: Explaining the Predictions of Any Classifier. arXiv 1602.04938v3 (2016)
Ritter, S., Barrett, D.G.T., Santoro, A., Botvinick, M.M.: Cognitive psychology for deep neural networks: a shape bias case study. In: Proceedings of the 34th International Conference on Machine Learning (2017)
Searle, J.: Minds, brains and programs. Behav. Brain Sci. 3(3), 417–457 (1980)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this paper
Cite this paper
Zednik, C. (2018). Will Machine Learning Yield Machine Intelligence?. In: Müller, V. (eds) Philosophy and Theory of Artificial Intelligence 2017. PT-AI 2017. Studies in Applied Philosophy, Epistemology and Rational Ethics, vol 44. Springer, Cham. https://doi.org/10.1007/978-3-319-96448-5_23
Download citation
DOI: https://doi.org/10.1007/978-3-319-96448-5_23
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-96447-8
Online ISBN: 978-3-319-96448-5
eBook Packages: Computer ScienceComputer Science (R0)