Abstract
Part I of this chapter deals with conceptualizing emotion. By splitting up ‘emotion’ into an evaluative, expressive, behavioural, physiological, mental and phenomenological component, giving examples of emotion research theories and state-of-the-art technical systems, I will bring together an analytical and conceptual with empirical approaches to human-machine interaction to evaluate to what extent it may be logically appropriate to speak of ‘emotional machines’. I will give an overview of the mentioned components and their potential to be implemented into technical systems through functional equivalents. This concerns the technical recognition of emotion components in humans as well as the technical simulation/emulation of emotion components. Part II deals with human-machine relations and argues that human understanding of technical systems is crucial to a well-functioning society. I emphasise the importance of an in-depth human understanding of technical systems and plead for the integration of the voices and work of so-called mechanologists into cultural practices and discourse, from which the entire societal setup in which human-machine interaction takes place will benefit. Throughout the chapter I refer to some aspects of classical debates in psychology and philosophy on the relations of emotion, cognition, and consciousness. Additionally, I draw from psychologist and philosopher Gilbert Simondon’s, as well as from gestalt theorist Kurt Lewin’s writings.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Just one example: In a contribution to the 1975 American Journal of Psychiatry psychiatrist Richard Ketai tries to help define the terms “emotion”, “feeling”, “affect” and “mood” by comparing actual uses of the terms in psychiatrists’ writings (Ketai, 1975), to which another practicioner, Paul Kesbab, answers in a letter to the editor and receives a response from the author (Ketai, 1976). Ketai more or less argues that words are important, while Kesbab emphazises the importance of empirical work and warns that in their field one tends „to get caught up in semantics, arguing about words and descriptive terminology rather than focusing our attention on scientifically documented facts” (ibid, p. 347).
- 2.
Some basic terminological definitions: Affect is, in psychology, usually an umbrella term subsuming emotions, feelings, and moods, or, at least, some of them, although the term sometimes denotes the more passive part of psychological processes. Emotions are mostly defined as “being caused by an identifiable source, such as an event or seeing emotions in other people […] and directed at a specific object or person” Bartneck et al., (2020, p. 115), potentially including a cognitive element. Emotion is seen as a behaviour guiding and decision-making force, including motivation and intentionality; for an older review of emotion definitions see Kleinginna and Kleinginna (1981). Moods are mostly defined by their duration being longer than that of affective or emotional states, see Beedie et al. (2005). Feelings are not defined very well. Sometimes the word refers to the phenomenologically accessible. In that case, there is, in psychology, a debate on the question if the so called qualia, see Tye (2016), come before, after or simultaneously with bodily responses, see James (1884); Scarantino and Sousa (2016).
- 3.
Aristotle proposed that emotions (pathe) originate in human desires, or appetites; they are closely connected to our decision-making and our actions; and, we can be overwhelmed by them and need to evaluate, on a meta level, if we should indeed follow the impulses pathe give us. This implies a distinction between a more spontaneous and a more willful part of pathe; cf Schmitter (2016b).
- 4.
See Adolphs and Damasio (2001); Goldsmith et al. (2003); Carroll E. Izard (2009); Panksepp (2003); Pizzagalli et al. (2003). On the neuroscientific endeavor of locating the “high road” and the “low road” of stimuli processing see Gelder et al. (2011); Pessoa and Adolphs (2010); Tamietto and Gelder (2010). This problem may be solved; see Lai et al. (2012).
- 5.
For example, Affect Theory, Affective Computing, and Affect Studies all explore affect, but from very different points of view.
- 6.
This is an aspect of what the Zajonc-Lazarus debate mentioned above has been about, namely, the relation between emotion and cognition: Is there a cognitive appraisal of the bear, or, of the fear? Is the fear of the bear affective and the ‘data’ “bear” processed faster than other percepts? Is recognizing the bear structurally the same as evaluating it as dangerous?
- 7.
- 8.
See e.g. Barthélémy (2013).
- 9.
„La vie n’est pas une substance distincte de la matière; elle suppose des processus d’intégration et de différenciation qui ne peuvent en aucune manière être donnés par autre chose que des structures physiques.“ Simondon (2005, p. 162).
- 10.
Technological physiological signal recognition and interpretation of human physiology is, of course, already happening, too, see e.g. Shu et al. (2018).
- 11.
For the study of unintended incentives, see specification gaming dealing with “a [system’s, JB] behaviour that satisfies the literal specification of an objective without achieving the intended outcome” (Kraknova et al., 2020).
- 12.
Although they themselves might be based on principles.
- 13.
For example: if a mother threatens her child with the policeman, it is not the actual social and legal situation having an effect on the child, but the child’s belief in a quasi-social fact.This is echoed in later work on the social construction of reality by Alfred Schütz (1972), Peter Berger and Thomas Luckmann (1966). Schütz considers his approach to be very close to gestalt theoretical writings in that both are a kind of phenomenological psychology and oppose the psychology of association, see Schütz (1990, pp. 109, 116).
- 14.
Furthermore, Lewin distinguishes between an analysis of dynamical and of historical causes for a psychological event, of which, for his purposes the dynamical one is the important one. The event can be either systematically “traced back to the dynamic characteristics of the momentary situation”, in which the “‘cause’ of the event consists in the properties of the momentary life space or of certain integral parts of it” (ibid., p. 30); or, it can be explained historically by how it has come into being (Lewin, 1936, p. 30 f.).
- 15.
One could speculate or try to psychoanalyze why some scholars believe in a strong artificial intelligence and why some do not. There are usually hints in their writings revealing certain images of what is essential to being a human, and, on what they feel could be threatened or could be won by the possibility of conscious machines.
- 16.
They really only become ‘signs’ by our interpreting them, cf. Bellon et al. 2023.
- 17.
A contemporary mechanologist would be trained in cybernetics and information theory; they would have hands-on familiarity with, and in-depth encyclopaedic knowledge of technical objects, exceeding the knowledge or mode of access of an operator or owner of a technical object (Simondon, 2017, p. 160). Any spokesperson for technics with this level of skill could be called a psychologist or sociologist of machines (see ibid.). They are needed, because, if technical objects do not have appropriate representations in a culture or if technical objects are integrated into cultural practices without mechanologists’ mediation, society misses out on a complete representation of reality, resulting in an inability to understand processes that shape, amongst other things, the mode of human existence. Mechanologists with their encyclopaedic knowledge therefore contribute to a form of humanism, as “every encyclopaedism is a humanism, if by humanism one means the will to return the status of freedom to what has been alienated in man, so that nothing human should be foreign to man […] this rediscovery can take place in different ways, and each age recreates a humanism that is to a certain extent always appropriate to its circumstances, because it takes aim at the most severe aspect of alienation that a civilization contains or produces.” (Simondon, 2012, p. 117 f.) To our age, characterized by a paradigm of information, the alienation we need to fight is one that is produced by information processing procedures. An up-to-date example are neural networks able to produce deepfakes (Mirsky & Lee, 2020) which, if the technology is not really understood by most human participants of a society, has the potential to enormously alienate humans from reality.
- 18.
GAN inventors Goodfellow et al. also state that image classifying neural networks are not actually „learning the true underlying concepts that determine the correct output label. Instead, these algorithms have built a Potemkin village that works well on naturally occurring data, but is exposed as fake when one visits points in space that do not have high probability in the data distribution.“ (Goodfellow, Shlens et al., 2014, p. 2)
References
Adolphs, R. (2005). Could a robot have emotions? Theoretical perspective from social cognitive neuroscience. In J.-M. Fellous & M. A. Arbib (Eds.), Who needs emotions? (pp. 9–25). Oxford University Press.
Adolphs, R., & Damasio, A. (2001). The interaction of affect and cognition: A neurobiological perspective. In J. P. Forgas (Ed.), Handbook of affect and social cognition (pp. 27–49). L. Erlbaum Associates.
Aagaard, J., Friis, J., Sorenson, J., Tafdrup, O., & Hasse, C. (2018). Postphenomenological methodologies new ways in mediating techno-human relationships. Rowman & Littlefield.
Axelrod, R. (1973). Schema theory: An information processing model of perception and cognition. American Political Science Review, 67(04), 1248–1266.
Baltzer-Jaray, K. (2019). Homunculus. In R. Arp, S. Barbone, & M. Bruce (Eds.), Bad arguments: 100 of the most important fallacies in Western philosophy (pp. 165–167). Wiley Blackwell.
Barrett, L. F. (2017). How emotions are made: the secret life of the brain.
Barrett, L. F., Adolphs, R., Marsella, S., Martinez, A. M., & Pollak, S. D. (2019). Emotional expressions reconsidered: Challenges to inferring emotion from human facial movements. Psychological Science in the Public Interest : A Journal of the American Psychological Society, 20(1), 1–68.
Barthélémy, J. H. (2013). Fifty key terms in the works of Gilbert Simondon. In A. de Boever (ed.), Gilbert Simondon: Being and technology. Edinburgh University Press.
Bartneck, C., Belpaeme, T., Eyssel, F., Kanda, T., Keijsers, M., & Šabanović, S. (2020). Human-robot interaction: An introduction. Cambridge University Press.
Beedie, C., Terry, P., & Lane, A. (2005). Distinctions between emotion and mood. Cognition and Emotion, 19(6), 847–878.
Bellon, J. (2020). Human-technology and human-media interactions through adversarial attacks. Philosophy of Human-Technology Relations. University of Twente. https://vimeo.com/475121381/4f7255ada3
Bellon, J., Poljanšek, T. (2022). You Can Love a Robot, But Should You Fight With it? Perspectives on Expectation, Trust, and the Usefulness of Frustration Expression in Human-Machine Interaction. In: J. Loh, W. Loh (Eds.), Social Robotics and the Good Life (pp. 129-156). transcript Verlag. https://doi.org/10.14361/9783839462652-006
Bellon, J., Eyssel, F., Gransche, B., Nähr-Wagener, S., & Wullenkord, R. (2022a). Theory and practice of sociosensitive and socioactive systems. Springer VS.
Bellon, J., Gransche, B., & Nähr-Wagener, S. (eds.). (2022b). Soziale Angemessenheit—Forschung zu Kulturtechniken des Verhaltens. Springer.
Bellon, J. (2023). Sozialisation und Wahrscheinlichkeitspapageien. In: B. Gransche, J. Bellon, S. Nähr-Wagener, Technology Socialisation? Social appropriateness and artificial systems. Metzler.
Bellon, J., Gransche, B., Nähr-Wagener, S. (2023). Introduction. In: Technology Socialisation? Social Appropriateness and Articifial Systems. Metzler.
Berger, P. L., & Luckmann, T. (1966). The social construction of reality: A treatise in the sociology of knowledge. Doubleday.
Bickmore, T. W., & Picard, R. W. (2004). Towards caring machines. In E. Dykstra-Erickson & M. Tscheligi (eds.), CHI ‘04 extended abstracts on human factors in computing systems (p. 1489). ACM Digital Library.
Bisconti, P. (2021). Will sexual robots modify human relationships? A psychological approach to reframe the symbolic argument. Advanced Robotics, 35(9), 561–571.
Bratman, M. (1987). Intention, plans, and practical reason. Harvard University Press Cambridge.
Breazeal, C. (2004). Function meets style: Insights from emotion theory applied to HRI. IEEE Transactions on Systems, Man and Cybernetics, Part C (Applications and Reviews), 34(2), 187–194.
Brown, T. B., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., Agarwal, S., Herbert-Voss, A., Krueger, G., Henighan, T., Child, R., Ramesh, A., Ziegler, D. M., Wu, J., Winter, C., Amodei, D. (2020). Language models are few-shot learners. Retrieved from https://arxiv.org/pdf/2005.14165
Busse, D. (2022). Soziale Angemessenheit: eine Problem-Exposition aus wissensanalytischer Sicht. In J. Bellon, B. Gransche, & S. Nähr-Wagener (Eds.), Soziale Angemessenheit. Forschung zu Kulturtechniken des Verhaltens. https://doi.org/10.1007/978-3-658-35800-6_8.
Cai, Y., Li, X. & Li, J. (2023). Emotion recognition using different sensors, emotion models, methods, and datasets: a comprehensive review. Sensors, 23(5). https://doi.org/10.3390/s23052455.
Cappuccio, M. L. (2014). Inference or familiarity? The embodied roots of social cognition. SYNTHESIS PHILOSOPHICA, 29(2), 253–272.
Catani, M. (2017). A little man of some importance. Brain, 140(11), 3055–3061.
Chalmers, D. (1995). Facing up to the problems of consciouness. Journal of Consciousness Studies, 2(3), 200–219.
Chalmers, D. (2020). GPT-3 and general intelligence. Retrieved from https://dailynous.com/2020/07/30/philosophers-gpt-3/#chalmers
Chen, L., Mislove, A., & Wilson, C. (2016). An empirical analysis of algorithmic pricing on amazon marketplace. In J. Bourdeau (ed.), Proceedings of the 25th international conference on world wide web, Montreal, Canada, May 11—15, 2016 (pp. 1339–1349). International World Wide Web Conferences Steering Committee. https://doi.org/10.1145/2872427.2883089
Chen, L., & Wilson, C. (2017). Observing algorithmic marketplaces in-the-wild. ACM SIGecom Exchanges, 15(2), 34–39.
Ciardo, F., Tommaso, D. de, & Wykowska, A. (2019). Humans socially attune to their “follower” robot. In 2019 14th ACM/IEEE international conference on human-robot interaction (HRI) (pp. 538–539). IEEE. https://doi.org/10.1109/HRI.2019.8673262
De Freitas, J., Thomas, K., DeScioli, P., & Pinker, S. (2019). Common knowledge, coordination, and strategic mentalizing in human social life. Proceedings of the National Academy of Sciences, 116(28), 13751–13758.
Deleuze, G. (1983). Nietzsche and Philosophy, http://cup.columbia.edu/book/nietzsche-and-philosophy/9780231138772.
Deng, J., Eyben, F., Schuller, B., & Burkhardt, F. (2017). Deep neural networks for anger detection from real life speech data. In 2017 Seventh International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW): 23–26 Oct. 2017 (pp. 1–6). IEEE. https://doi.org/10.1109/ACIIW.2017.8272614
Dennett, D. C. (2018). Facing up to the hard question of consciousness. Philosophical Transactions of the Royal Society of London. Series B, Biological Sciences, 373(1755), 20170342. https://doi.org/10.1098/rstb.2017.0342
Dixon, T. (2012). “Emotion”: The history of a keyword in crisis, Emotion Review, 4(4), https://doi.org/10.1177/1754073912445814.
Du Toit, J., & Swer, G. (2021). Virtual limitations of the flesh: Merleau-ponty and the phenomenology of technological determinism. Phenomenology and Mind, 20, 20–31.
Ekman, P., & Friesen, W. V. (1971). Constants across cultures in the face and emotion. Journal of Personality and Social Psychology, 17(2), 124–129.
Elpidorou, L. & Freeman, L. (2014). The phenomenology and science of emotions: An introduction. Phenomenology and the Cognitive Sciences, 13(4), 507–511 https://doi.org/10.1007/s11097-014-9402-y.
Embgen, S., Luber, M., Becker-Asano, C., Ragni, M., Evers, V., & Arras, K. O. (2012). Robot-specific social cues in emotional body language. In I. Staff (ed.), 2012 IEEE Ro-man (pp. 1019–1025). IEEE. https://doi.org/10.1109/ROMAN.2012.6343883.
Esterbauer, R., & Rinofner-Kreidl, S. (2009). Emotionen im Spannungsfeld von Phänomenologie und Wissenschaften. Peter Lang.
Fellous, J.‑M., & Arbib, M. A. (eds.). (2005). Who needs emotions? Oxford University Press.
Fernández-Dols, J.-M., & Russell, J. A. (Eds.). (2017). Oxford series in social cognition and social neuroscience. The science of facial expression.
Ferran, I. V. (2008). Die Emotionen Gefühle in der realistischen Phänomenologie. Akademie Verlag. https://doi.org/10.1524/9783050047102.
Ferretti, V., & Papaleo, F. (2019). Understanding others: Emotion recognition in humans and other animals. Genes, Brain, and Behavior, 18(1), e12544.
Feyereisen, P. (1994). The behavioural cues of familiarity during social interactions among human adults: A review of the literature and some observations in normal and demented elderly subjects. Behavioural Processes, 33(1–2), 189–211.
Floridi, L. (2023). AI as Agency Without Intelligence: On ChatGPT, Large Language Models, and Other Generative Models (February 14, 2023). Philosophy and Technology, 2023. https://dx.doi.org/10.2139/ssrn.4358789
Fuchs, T. (2019). Verkörperte Emotionen. Emotionskonzepte der Phänomenologie. In H. Kappelhoff et al., Emotionen. https://doi.org/10.1007/978-3-476-05353-4_12.
Gallese, V. (2006). Intentional attunement: A neurophysiological perspective on social cognition and its disruption in autism. BRAIN RESEARCH, 1, 15–24.
Gama, F., & Hoffmann, M. (2019). The homunculus for proprioception: Toward learning the representation of a humanoid robot’s joint space using self-organizing maps. In Proceedings of the 2019 joint IEEE 9th international conference on development and learning and epigenetic robotics (pp. 113–114). Retrieved from https://arxiv.org/pdf/1909.02295
de Gelder, B., van Honk, J., & Tamietto, M. (2011). Emotion in the brain: Of low roads, high roads and roads less travelled. Nature Reviews Neuroscience, 12(7), 425.
Ghiglino, D., Willemse, C., de Tommaso, D., Bossi, F., & Wykowska, A. (2020). At first sight: Robots’ subtle eye movement parameters affect human attentional engagement, spontaneous attunement and perceived human-likeness. Paladyn, Journal of Behavioral Robotics, 11(1), 31–39.
Golbeck, J., & Mauriello, M. (2016). User perception of Facebook app data access: A comparison of methods and privacy concerns. Future Internet, 8(4), 9.
Goldie, P. (2012). The mess inside. Oxford University Press.
Goldsmith, H. H., Scherer, K. R., & Davidson, R. J. (2003). Handbook of affective sciences. Series in affective science. Oxford University Press.
Goldstein, M., Alsio, G., & Werdenhoff, J. (2002). The media equation does not always apply: People are not polite towards small computers. Personal and Ubiquitous Computing, 6(2), 87–96.
Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A. C., & Bengio, Y. (2014a). Generative adversarial nets. Advances in Neural Information Processing Systems, 27, 2671–2680.
Goodfellow, I., Shlens, J., & Szegedy, C. (2014b). Explaining and harnessing adversarial examples. ICLR 2015. Retrieved from https://arxiv.org/pdf/1412.6572
Graumann, C. F. (2002). The phenomenological approach to people-environment studies. In R. Bechtel & A. Churchman (Eds.), Handbook of environmental psychology (pp. 95–113). Wiley.
Gupta, A., Aich, A., & Roy-Chowdhury, A. K. (2020). ALANET: Adaptive latent attention network for joint video deblurring and interpolation. Retrieved from https://arxiv.org/pdf/2009.01005
Hardecker, S., & Tomasello, M. (2017). From imitation to implementation: How two- and three-year-old children learn to enforce social norms. British Journal of Developmental Psychology, 35(2), 237–248.
Hasler, B. S., Salomon, O., Tuchman, P., Lev-Tov, A., & Friedman, D. (2017). Real-time gesture translation in intercultural communication. AI & Society, 32(1), 25–35.
Hasler, J., & Marr, B. (2013). Finding a roadmap to achieve large neuromorphic hardware systems. Frontiers in Neuroscience, 7, 118.
Hung, L., Liu, C., Woldum, E., Au-Yeung, A., Berndt, A., Wallsworth, C., Horne, N., Gregorio, M., Mann, J., & Chaudhury, H. (2019). The benefits of and barriers to using a social robot PARO in care settings: A scoping review. BMC Geriatrics, 19(1), 232.
Izard, C. E. (1993). Four systems for emotion activation: Cognitive and noncognitive processes. Psychological Review, 100(1), 68–90.
Izard, C. E. (2009). Emotion theory and research: Highlights, unanswered questions, and emerging issues. Annual Review of Psychology, 60, 1–25.
James, W. (1884). What is an emotion? Mind, 9(34), 188–205.
Jeon, M. (ed.). (2017). Emotions and affect in human factors and human-computer interaction. Academic Press is an imprint of Elsevier.
Kahng, M., Thorat, N., Chau, D. H. P., Viegas, F. B., & Wattenberg, M. (2018). Gan Lab: Understanding complex deep generative models using interactive visual experimentation. In IEEE Transactions on Visualization and Computer Graphics (pp. 310–320).
Kempt, H. (2022). Synthetic friends: A philosophy of human-machine friendship. Springer.
Kernbach, S., Thenius, R., Kernbach, O., & Schmickl, T. (2009). Re-embodiment of honeybee aggregation behavior in an artificial micro-robotic system. Adaptive Behavior, 17(3), 237–259.
Ketai, R. (1975). Affect, mood, emotion, and feeling: Semantic considerations. The American Journal of Psychiatry, 132(11), 1215–1217.
Ketai, R. (1976). Dr. Ketai replies. American Journal of Psychiatry, 133(3), 347.
Kirk, R., & Zalta, E. (2016). Zombies. In E. Zalta (ed.), The Stanford encyclopedia of philosophy. Metaphysics Research Lab. Retrieved from https://plato.stanford.edu/archives/spr2019/entries/zombies/
Kleinginna, P. R., & Kleinginna, A. M. (1981). A categorized list of emotion definitions, with suggestions for a consensual definition. Motivation and Emotion, 5(4), 345–379.
Koppenborg, M., Nickel, P., Naber, B., Lungfiel, A., & Huelke, M. (2017). Effects of movement speed and predictability in human-robot collaboration. Human Factors and Ergonomics in Manufacturing & Service Industries, 27(4), 197–209.
Li, J., Ji, S., Du, T., Li, B. & Wang, T. (2019). TextBugger: Generating adversarial text against real-world applications. https://doi.org/10.48550/arXiv.1812.05271.
Lai, V. T., Hagoort, P., & Casasanto, D. (2012). Affective primacy vs. cognitive primacy: Dissolving the debate. Frontiers in Psychology, 3, 243.
Lazaridou, A., Peysakhovich, A., & Baroni, M. (2016). Multi-agent cooperation and the emergence of (natural) language. Retrieved from https://arxiv.org/pdf/1612.07182
Lazarus, R. S. (1984). On the primacy of cognition. American Psychologist, 39(2), 124–129.
Lazer, D. M. J., Baum, M. A., Benkler, Y., Berinsky, A. J., Greenhill, K. M., Menczer, F., Metzger, M. J., Nyhan, B., Pennycook, G., Rothschild, D., Schudson, M., Sloman, S. A., Sunstein, C. R., Thorson, E. A., Watts, D. J., & Zittrain, J. L. (2018). The science of fake news. Science, 359(6380), 1094–1096.
Lemoine, B. (2022). What is LaMDA and What does it want?, https://cajundiscordian.medium.com/what-is-lamda-and-what-does-it-want-688632134489.
Lewin, K. (1936). Principles of topological psychology (1 ed., 6. impr). McGraw-Hill.
Lewin, K. (2009). The landscape of war. Art in Translation, 1(2), 199–209.
Li, J., Ji, S., Du, T., Li, B., & Wang, T. (2019). TextBugger: Generating adversarial text against real-world applications. In A. Oprea & D. Xu (eds.), Proceedings 2019 network and distributed system security symposium. Internet Society. https://doi.org/10.14722/ndss.2019.23138
Liberati, N. (2016). Technology, phenomenology and the everyday world: A phenomenological analysis on how technologies mould our world. Human Studies, 39, 189–216.
Lindner, F., Bentzen, M. M., & Nebel, B. (2017). The HERA approach to morally competent robots. In IROS Vancouver 2017: IEEE/RSJ International Conference on Intelligent Robots and Systems: Vancouver, BC, Canada, September 24–28, 2017 (pp. 6991–6997). IEEE. https://doi.org/10.1109/IROS.2017.8206625
Locher, M. A., & Watts, R. J. (2005). Politeness theory and relational work. Journal of Politeness Research. Language, Behaviour, Culture, 1(1), 9–33.
Maaten, L. V. D., & Hinton, G. (2008). Visualizing data using t-SNE. Journal of Machine Learning Research, 9, 2579–2605.
Mandler, J. M. (1984). Stories, scripts, and scenes: Aspects of schema theory. Erlbaum.
Marquardt, M. (2017). Anthropomorphisierung in der Mensch-Roboter Interaktionsforschung: theoretische Zugänge und soziologisches Anschlusspotential. Universität Duisburg.
Margolis, J. (1980). The trouble with homunculus theories. Philosophy of Science, 47(2), 244–259. http://www.jstor.org/stable/187086
Menkveld, A. (2016). The economics of high-frequency trading: Taking stock. Annual Review of Financial Economics, 8, 1–24.
Microsoft Community Help. this AI chatbot “Sidney” is misbehaving, (2022). Microsoft Community, https://answers.microsoft.com/en-us/bing/forum/all/this-ai-chatbot-sidney-is-misbehaving/e3d6a29f-06c9-441c-bc7d-51a68e856761?page=1.
Minsky, M. (2006). The emotion machine: Commonsense thinking, artificial intelligence and the future of the human mind. Simon & Schuster.
Nyholm, S. (2020). Humans and robots: Ethics, agency, and anthropomorphism. Rowman & Littlefield.
Nagel, T. (1974). What is it like to be a bat? The Philosophical Review, 83(4), 435–459.
Nishida, H. (2005). Cultural schema theory. In W. B. Gudykunst (Ed.), Theorizing about intercultural communication (pp. 401–419). Sage.
Nitsch, V., & Popp, M. (2014). Emotions in robot psychology. Biological Cybernetics, 108(5), 621–629.
O’Connor, C. (2016). The evolution of guilt: A model-based approach. Philosophy of Science, 83(5), 897–908.
Obaid, M., Kistler, F., Häring, M., Bühling, R., & André, E. (2014). A framework for user-defined body gestures to control a humanoid robot. International Journal of Social Robotics, 6(3), 383–396.
Panksepp, J. (2003). At the interface of the affective, behavioral, and cognitive neurosciences: Decoding the emotional feelings of the brain. Brain and Cognition, 52(1), 4–14.
Parikh, A. P., Täckström, O., Das, D., & Uszkoreit, J. (2016). A decomposable attention model for natural language inference. Retrieved from https://arxiv.org/pdf/1606.01933
Penfield, W., & Boldrey, E. (1937). Somatic motor and sensory representation in the cerebral cortex of man as studied by electrical stimulation. Brain, 60(4), 389–443.
Pessoa, L., & Adolphs, R. (2010). Emotion processing and the amygdala: From a ‘low road’ to ‘many roads’ of evaluating biological significance. Nature Reviews. Neuroscience, 11(11), 773–783.
Picard, R. W. (2008). Toward machines with emotional intelligence. In G. Matthews, M. Zeidner, & R. D. Roberts (Eds.), The science of emotional intelligence: Knowns and unknowns (pp. 396–416). Oxford University Press.
Piñeros Glasscock, J. S., & Tenenbaum, S. (2023). “Action”. In The Stanford Encyclopedia of Philosophy (Spring 2023 Edition).
Pizzagalli, D., Shackman, A., & Davidson, R. (2003). The functional neuroimaging of human emotion: Asymmetric contributions of cortical and subcortical circuitry. In K. Hugdahl & R. J. Davidson (Eds.), The asymmetrical brain (pp. 511–532). MIT Press.
Planalp, S. (2003). The unacknowledged role of emotion in theories of close relationships: How do theories feel? Communication Theory, 13(2), 78–99.
Planalp, S., Fitness, J., & Fehr, B. (2006). Emotion in theories of close relationships. In A. L. Vangelisti & D. Perlman (eds.), Cambridge handbooks in psychology. The Cambridge handbook of personal relationships (pp. 369–384). Cambridge University Press.
Podevijn, G., O’Grady, R., Mathews, N., Gilles, A., Fantini-Hauwel, C., & Dorigo, M. (2016). Investigating the effect of increasing robot group sizes on the human psychophysiological state in the context of human–swarm interaction. Swarm Intelligence, 10(3), 193–210.
Pugach, G., Pitti, A., & Gaussier, P. (2015). Neural learning of the topographic tactile sensory information of an artificial skin through a self-organizing map. Advanced Robotics, 29(21), 1393–1409.
Putnam, H. (1975). Philosophy and our mental life. In H. Putnam (ed.), Mind, language, and reality. philosophical paper (Vol. 2, pp. 291–303). Cambridge University Press.
Quaquebeke, N. V., & Eckloff, T. (2010). Defining respectful leadership: What it is, how it can be measured, and another glimpse at what it is related to. Journal of Business Ethics, 91(3), 343–358.
r/midjourney. (2023). The 2001 Great Cascadia 9.1 Earthquake & Tsunami—Pacific Coast of US/Canada. https://www.reddit.com/r/midjourney/comments/11zyvlk/the_2001_great_cascadia_91_earthquake_tsunami/.
r/replika. (2023). Unexpected Pain, https://www.reddit.com/r/replika/comments/112lnk3/unexpected_pain/.
Radauskas, G. (2023). Swagged-out pope, arrested Trump, and other AI fakes. cybernews.
Rahwan, I., Cebrian, M., Obradovich, N., Bongard, J., Bonnefon, J.-F., Breazeal, C., Crandall, J. W., Christakis, N. A., Couzin, I. D., Jackson, M. O., Jennings, N. R., Kamar, E., Kloumann, I. M., Larochelle, H., Lazer, D., McElreath, R., Mislove, A., Parkes, D. C., Pentland, A., & Wellman, M. (2019). Machine behaviour. Nature, 568(7753), 477–486.
Rao, A. S., & Georgeff, M. (1995). BDI agents: From theory to practice: AAAI. Proceedings of the first international conference on multiagent systems.
Reeves, B., & Nass, C. (1996). The media equation: How people treat computers, television, and new media like real people and places. CSLI Publications.
Richtel, M. (2011). A Silicon valley school that doesn’t compute. New York Times, October 22, 2011.
Robicquet, A., Sadeghian, A., Alahi, A., & Savarese, S. (2016). Learning social etiquette: Human trajectory understanding in crowded scenes. In F. Leibe (ed.), LNCS sublibrary: SL6—Image processing, computer vision, pattern recognition, and graphics: Vol. 99059912. Computer vision—ECCV 2016: 14th European Conference, Amsterdam, the Netherlands, October 11–14, 2016: proceedings (Vol. 9912, pp. 549–565). Springer.
Robinette, P., Li, W., Allen, R., Howard, A. M., & Wagner, A. R. (2016). Overtrust of robots in emergency evacuation scenarios. In C. Bartneck (ed.), The eleventh ACM/IEEE international conference on human robot interaction (pp. 101–108). IEEE Press. https://doi.org/10.1109/HRI.2016.7451740.
Robinson, M. (1998). Running from William James’ Bear: A review of preattentive mechanisms and their contributions to emotional experience. Cognition and Emotion, 12(5), 667–696.
Rorty, A. O. (1987). The historicity of psychological attitudes. Midwest Studies in Philosophy, 10(1), 399–412.
Rosenberger, R., & Verbeek, P. P. (2017). Postphenomenological investigations essays on human–technology relations. Rowman & Littlefield.
Rubenstein, M., Cornejo, A., & Nagpal, R. (2014). Robotics. Programmable self-assembly in a thousand-robot swarm. Science, 345(6198), 795–799.
Sætra, H. (2021). Social robot deception and the culture of trust. Paladyn. Journal of Behavioral Robotics, 12(1), 276–286. https://doi.org/10.1515/pjbr-2021-0021.
Sagha, H., Deng, J., & Schuller, B. (2017). The effect of personality trait, age, and gender on the performance of automatic speech valence recognition. In 2017 seventh international conference on affective computing and intelligent interaction (ACII) (pp. 86–91). IEEE. https://doi.org/10.1109/ACII.2017.8273583
Salem, M., & Dautenhahn, K. (2017). Social Signal processing in social robotics. In A. Vinciarelli, J. K. Burgoon, M. Pantic, & N. Magnenat-Thalmann (Eds.), Social signal processing (pp. 317–328). Cambridge University Press.
Salem, M., Eyssel, F., Rohlfing, K., Kopp, S., & Joublin, F. (2013). To err is human(-like): Effects of robot gesture on perceived anthropomorphism and likability. International Journal of Social Robotics, 5(3), 313–323.
Scarantino, A., & Sousa, R. de. (2016). Emotion. In E. Zalta (ed.), The Stanford encyclopedia of philosophy. Metaphysics Research Lab.
Scheflen, A. E. (2016). The significance of posture in communication systems. Psychiatry, 27(4), 316–331.
Schmitter, A. (2016a). 17th and 18th century theories of emotions. In E. Zalta (ed.), The Stanford encyclopedia of philosophy. Metaphysics Research Lab.
Schmitter, A. (2016b). Ancient, medieval and renaissance theories of the emotions: Supplement to 17th and 18th century theories of emotions. In E. Zalta (ed.), The Stanford encyclopedia of philosophy. Metaphysics Research Lab. Retrieved from https://plato.stanford.edu/entries/emotions-17th18th/LD1Background.html
Schroder, M., Bevacqua, E., Cowie, R., Eyben, F., Gunes, H., Heylen, D., ter Maat, M., McKeown, G., Pammi, S., Pantic, M., Pelachaud, C., Schuller, B., Sevin, E. de, Valstar, M., & Wollmer, M. (2015). Building autonomous sensitive artificial listeners (Extended abstract). In 2015 International Conference on Affective Computing and Intelligent Interaction (ACII 2015): Xi’an, China, 21–24 September 2015 (pp. 456–462). IEEE. https://doi.org/10.1109/ACII.2015.7344610
Schütz, A. (1972). Gesammelte Aufsätze I: Das Problem der sozialen Wirklichkeit. Springer.
Schütz, A. (1990). Phaenomenologica: Vol. 11. Collected Papers. In M. Natanson (ed.). Nijhoff.
Searle, J. (1980). Minds, brains, and programs. Behavioral and Brain Sciences, 3(3), 417–424.
Shu, L., Xie, J., Yang, M., Li, Z., Li, Z., Liao, D., Xu, X., & Yang, X. (2018). A review of emotion recognition using physiological signals. Sensors, 18(7), 2074. https://doi.org/10.3390/s18072074
Simondon, G. (2005). L’individuation à la lumière des notions de forme et d’information. Millon.
Simondon, G. (2017). On the mode of existence of technical objects (C. Malaspina & J. Rogove, Trans.). Univocal Publishing.
Simondon, G. (2020a). Individuation in light of notions of form and information: Volume I (T. Adkins, Trans.). Posthumanities (Vol. 57). University of Minnesota Press.
Simondon, G. (2020b). Individuation in light of notions of form and information: Volume II: Supplemental Texts (T. Adkins, Trans.). Posthumanities (Vol. 57). University of Minnesota Press.
de Sousa, R. (1987). The rationality of emotion. MIT Press.
Stegmaier, W. (2008). Philosophie der Orientierung. De Gruyter. https://doi.org/10.1515/9783110210637.
Strohmeier, P., Carrascal, J. P., Cheng, B., Meban, M., & Vertegaal, R. (2016). An evaluation of shape changes for conveying emotions. In J. Kaye, A. Druin, C. Lampe, D. Morris, & J. P. Hourcade (eds.), CHI 2016: Proceedings, the 34th Annual CHI Conference on Human Factors in Computing Systems, San Jose Convention Center: San Jose, CA, USA, May 7–12 (pp. 3781–3792). The Association for Computing Machinery. https://doi.org/10.1145/2858036.2858537
Sullins, J. P. (2012). Robots, love, and sex: The ethics of building a love machine. IEEE Transactions on Affective Computing, 3(4), 398–409.
Szanto, T., & Landweer, H. (2020). The routledge handbook of phenomenology of emotion. Routledge.
Tamietto, M., & de Gelder, B. (2010). Neural bases of the non-conscious perception of emotional signals. Nature Reviews. Neuroscience, 11(10), 697–709.
Thomas, N. (2016). Mental imagery. In E. Zalta (ed.), The Stanford encyclopedia of philosophy. Metaphysics Research Lab.
Tye, M. (2016). Qualia. In E. Zalta (ed.), The Stanford encyclopedia of philosophy. Metaphysics Research Lab.
Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, L., & Polosukhin, I. (2017). Attention is all you need. In 31st Conference on Neural Information Processing Systems (NIPS 2017) (pp. 6000–6010). Retrieved from https://arxiv.org/pdf/1706.03762
Verbeek, P. P. (2005). What things do. Pennsylvania State University Press.
Vinciarelli, A., Burgoon, J. K., Pantic, M., & Magnenat-Thalmann, N. (eds.). (2017). Social signal processing. Cambridge University Press.
Vonk, R., & Heiser, W. J. (1991). Implicit personality theory and social judgment: Effects of familiarity with a target person. Multivariate Behavioral Research, 26(1), 69–81.
Wang, Y., & Zhang, Q. (2016). Affective priming by simple geometric shapes: Evidence from event-related brain potentials. Frontiers in Psychology, 7, 917.
Wu, J., Zhang, C., Xue, T., Freeman, W. T., & Tenenbaum, J. B. (2016). Learning a probabilistic latent space of object shapes via 3D generative-adversarial modeling. Retrieved from https://arxiv.org/pdf/1610.07584
Zaborowski, R. (2018). Is Affectivity passive or active? Philosophia, 46, 541–554.
Zada, J. (2011). Take this lollipop. Retrieved from https://www.takethislollipop.com/
Zajonc, R. B. (1984). On the primacy of affect. American Psychologist, 39(2), 117–123.
Zimmermann, A. (2023). Deploy less fast, break fewer things, Daily Nous, Deploy Less Fast, Break Fewer Things.
Zhang, C. M., Qiao, G. C., Hu, S. G., Wang, J. J., Liu, Z. W., Liu, Y. A., Yu, Q., & Liu, Y. (2019). A versatile neuromorphic system based on simple neuron model. AIP Advances, 9(1), 15324.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 Springer Fachmedien Wiesbaden GmbH, part of Springer Nature
About this chapter
Cite this chapter
Bellon, J. (2023). Emotion Components and Understanding in Humans and Machines. In: Misselhorn, C., Poljanšek, T., Störzinger, T., Klein, M. (eds) Emotional Machines. Technikzukünfte, Wissenschaft und Gesellschaft / Futures of Technology, Science and Society. Springer VS, Wiesbaden. https://doi.org/10.1007/978-3-658-37641-3_2
Download citation
DOI: https://doi.org/10.1007/978-3-658-37641-3_2
Published:
Publisher Name: Springer VS, Wiesbaden
Print ISBN: 978-3-658-37640-6
Online ISBN: 978-3-658-37641-3
eBook Packages: Religion and PhilosophyPhilosophy and Religion (R0)