Power-Posing Robots: The Influence of a Humanoid Robot’s Posture and Size on its Perceived Dominance, Competence, Eeriness, and Threat

When interacting with sophisticated digital technologies, people often fall back on the same interaction scripts they apply to the communication with other humans—especially if the technology in question provides strong anthropomorphic cues (e.g., a human-like embodiment). Accordingly, research indicates that observers tend to interpret the body language of social robots in the same way as they would with another human being. Backed by initial evidence, we assumed that a humanoid robot will be considered as more dominant and competent, but also as more eerie and threatening once it strikes a so-called power pose. Moreover, we pursued the research question whether these effects might be accentuated by the robot’s body size. To this end, the current study presented 204 participants with pictures of the robot NAO in different poses (expansive vs. constrictive), while also manipulating its height (child-sized vs. adult-sized). Our results show that NAO’s posture indeed exerted strong effects on perceptions of dominance and competence. Conversely, participants’ threat and eeriness ratings remained statistically independent of the robot’s depicted body language. Further, we found that the machine’s size did not affect any of the measured interpersonal perceptions in a notable way. The study findings are discussed considering limitations and future research directions.


Introduction
In the field of human nonverbal communication, the effects of assertive body poses remain a popular and hotly debated topic. Adopting certain body postures is suspected to be a key to increasing self-worth, risk tolerance, and dominance in social situations (power-posing, [1][2][3]). A controversial psychological publication of the previous decade initially suggested that these effects might be rooted in bodily feedback loops and hormonal processes [1], but this claim was quickly met with skepticism and empirical objections by other researchers [4][5][6]. Yet, while the physiological aspects of power-posing have been mostly refuted, its cognitive effects on people's self-perceptions remain a much-discussed (and researched) topic to this day [1,[7][8][9].
B Jan-Philipp Stein jan-philipp.stein@uni-wuerzburg.de 1 Psychology of Communication and New Media, University of Würzburg, Würzburg, Germany Apart from the controversial notion of whether people's body language modulates the way they see themselves, however, there is much more scholarly agreement on how power poses influence perceptions by others. In fact, decades worth of research from the field of nonverbal communication clearly demonstrate that body postures, walking patterns, and hand gestures all impact which traits observers attribute to a person [10]. As a particularly prominent finding in this regard, numerous studies have shown that people displaying so-called expansive nonverbal behavior (e.g., a wide stance, sweeping hand movements, direct eye contact) are usually seen as much more persuasive, admirable, and confident than those standing or moving in a constricted manner [11][12][13]. Considering the high social relevance of these perceptions, it comes as no surprise that expansive body language has also been connected to several practical outcomes, such as increased success in job interviews [14], stronger romantic desirability [15], and more favorable ratings for politicians [16].

Applying Principles of Human Interaction to Technology
Due to notable advancements in the areas of computer science and engineering, contemporary digital technology can reach impressive levels of human likeness. Indeed, since many modern-day technologies appear inherently social in nature (for instance by talking back to the user in a human voice), it has been shown that people often apply the same scripts they use with other humans to their interactions with technology [17]. In consequence of this so-called computers are social actors (CASA) phenomenon, technologies such as robots, smart speakers, or even phones are not only ascribed their own personality [18,19], but also regarded with genuine emotional attachment by their owners [20]. Of course, it should be noted that not each and every theory from the area of human-human interaction may be transferred seamlessly to people's interactions with machines; for instance, since emotional bonds can only be developed unilaterally (from the user to the machine), relational expectations may turn out quite different [21]. Along the same lines, several recent publications have cautioned against generalizing social psychological insight to all types of human-robot interaction, instead advocating for a more nuanced perspective [22][23][24]. Regardless of these limitations, however, the CASA paradigm continues to offer most valuable reference points for researchers to make sense of people's approach to intelligent technology [25][26][27][28].
The tendency to treat computers as social actors can be evoked by the most machine-looking, or even completely bodiless technologies, including smart home appliances or online text chatbots [29]. At the same time, research shows that by adding a human-like physical embodiment (e.g., a robotic body) to a digital system, it naturally becomes even easier for people to anthropomorphize it [30,31]. Regarding the actual acceptance of technology, however, increased human likeness does not necessarily equal more liking as well. Instead, research embedded within the impressionistic uncanny valley framework [32,33] has suggested that highly anthropomorphic technologies can also be perceived as eerily imperfect [34][35][36] or downright threatening [37,38]. More specifically, it has been argued that by reaching high (yet not entirely flawless) levels of human likeness, technologies may trigger aversion on both a cognitive and affective level [39], for instance by eliciting cognitive dissonance [40], prompting mortality salience [41], or raising concerns about threatened human uniqueness [42,43]. In turn, these processes may then manifest as an eerie sensation or the expectation of immediate danger. While earlier research mainly attributed the uncanny valley effect to visual features, recent literature shows that eeriness and threat may also be prompted by certain behaviors and mental capacities of artificial beings [38,44,45].
Thus, even if the face or body of a robot might not appear particularly threatening to an observer, its displayed actions and skills may still trigger a negative user response.

The Current Study
Considering the social quality that accompanies many human-machine interactions, a growing body of research has investigated whether nonverbal behavior by embodied digital technologies translates into the same interpersonal perceptions that occur among humans. Serving as the groundwork for this line of research, several studies were able to establish that the body language of social robots and virtual avatars is often recognized accurately by participants [46][47][48]. Subsequently, it was shown that by assuming specific body postures, human-like technologies may indeed appear more or less competent [49], persuasive [50], cooperative [51], authoritative [52], or dominant [53] to their human users.
Based on the reviewed literature, it becomes evident that fundamental principles of human body language might also apply to interactions with social robots. Nevertheless, some notable research gaps remain. For instance, we know of no previous research that has investigated potential interaction effects between a robot's body language and its size on users' evaluations and dispositional attributions. Considering that the popular NAO robot used in most scientific studies is rather small in size-and might therefore trigger associations such as "a toy" or "a child," as well as corresponding cognitive schemas-it stands to reason that different effects might emerge for robots with larger body dimensions. In particular, we expected that perceptions of dominance or competence, which are typically associated more with adult age [54,55], would turn out even stronger for adult-sized than for childsized robots displaying expansive behavior.
H1a: A robot that strikes an expansive pose will appear more dominant than a robot striking a constrictive pose. H1b: This effect will turn out stronger for an adult-sized than for a child-sized robot. H2a: A robot that strikes an expansive pose will appear more competent than a robot striking a constrictive pose. H2b: This effect will turn out stronger for an adult-sized than for a child-sized robot.
Despite being an increasingly prominent construct in the field of human-robot interaction [38,56], we are not familiar with any research that has connected robot postures to the perception of threat. In our opinion, this presents another empirical shortcoming, not least considering that the perceived safety and danger of robots is strongly related to their mass adoption [57]. With our third hypothesis, we therefore scrutinized participants' threat experience in the face of robots displaying different body language. Building upon the effects proposed in H1 and H2, we expected: H3a: A robot that strikes an expansive pose will appear more threatening than a robot striking a constrictive pose. H3b: This effect will turn out stronger for an adult-sized than for a child-sized robot.
Lastly, we strived to situate the current study in the tradition of the uncanny valley framework, which has inspired scholars in many technology-related disciplines for several decades [33,39]. Based on the emerging idea that threat perceptions might be the underlying reason for the eerie sensation often observed in robot experiments [32,58], we hypothesized: H4a: A robot that strikes an expansive pose will appear eerier than a robot striking a constrictive pose. H4b: This effect will turn out stronger for an adult-sized than for a child-sized robot.

Method
The current study was conducted in the form of an online experiment, using self-created sets of robot photographs in a 2 (robot size: child vs. adult) × 2 (robot pose: expansive vs. constrictive) between-subject design. Hypotheses, measures, and analysis strategies were preregistered at https:// aspredicted.org/pa7ta.pdf. Furthermore, we provide all data, codes, and materials of this study in an Open Science Framework repository (https://osf.io/2zx89/).

Participants
An a priori calculation of minimum sample size-assuming a test power of 80% and a medium multivariate effect-resulted in a lower threshold of at least 125 participants. Ultimately, 214 participants (120 female, 93 male, 1 other; age M = 34.47 years, SD = 17.38) were recruited via social media, university mailing lists, and personal contacts for the current study. However, by using a five-point conscientiousness item, we were able to identify seven participants who had responded rather carelessly, leading to their exclusion from the data. Furthermore, three participants were excluded due to technical difficulties during their participation. Although we had initially considered to also exclude all participants who recognized the portrayed robot-which would have led to the additional exclusion of 39 individuals-our statistical analyses showed that removing these participants did not change our findings in any meaningful way. 1 As such, we opted to keep the respective datasets in our study, resulting in a final sample of 204 participants (114 female, 89 male, 1 other) with an average age of M = 34.28 years (SD = 17.13). In terms of professional background, our participants were predominantly students (44.8%), employees (35.8%), and retirees (11.3%). To participate, each person had to give informed consent. Additionally, all participants were offered to join a gift raffle of two e15 shopping vouchers as an incentive.

Procedure and Materials
At the start of our online study, each participant was randomly assigned to one of four experimental groups according to our two-factorial design: child-sized robot in expansive pose, child-sized robot in constrictive pose, adult-sized robot in expansive pose, adult-sized robot in constrictive pose. Subsequently, we presented all participants with two self-created robot images matching their assigned condition (Fig. 1), which had to be viewed for a minimum duration of twenty seconds each before it was possible to proceed to the prepared evaluation questionnaires.
As for the creation of our stimuli, we first took several photographs of the robot NAOv5 in a neutral setting. In order to depict expansive and constrictive body language in a conceptually valid way, we consulted psychological literature [1,4,10,59], as well as a taxonomy proposed by human-computer interaction scholars [51]. By these means, the following criteria were identified for expansive poses: A wide stance (standing) or spread legs (sitting), hands placed on the hips (standing) or behind the head (sitting), head slightly lifted, and direct eye contact. In contrast to this, constrictive poses were characterized by the robot putting its arms together in front of its body, lowering its head, averting its gaze, and assuming a slightly bent posture.
To avoid the mono-stimulus bias often encountered in media-based research [60], we decided to prepare two stimulus pictures for each experimental group: (a) NAO sitting alone in front of a neutral background, and (b) NAO standing next to a group of people. Adobe Photoshop software and license-free stock photos were used to assemble the final stimuli. For the size manipulation, we either portrayed NAO in its original size of 55 cm (≈1 10 ) or depicted a version that appeared approximately 160 cm tall (≈5 3 ). In order to achieve this manipulation in the images with a neutral background, we added electric sockets as a well-known comparison standard so that participants could infer the intended size of the machine.

Dominance
To assess perceptions of dominance, we used five items provided by Straßmann and colleagues [51]. All items (e.g., "dominant", "decisive", "submissive") were presented using 7-point scales (1 = not at all; 7 = completely) and averaged into a composite score after recoding the negatively valenced items. The resulting dominance index achieved good internal consistency, Cronbach's α = 0.83.

Competence
In this study, we strived to assess competence as a relatively general interpersonal impression, i.e., the potential for effective action in any given domain [54]. A suitable instrument was obtained from prior power-posing research [61], which addresses perceived competence rather broadly via seven semantic differentials (e.g., "novice/experienced", "unable/able to compete"). All items were presented with seven gradation points. We observed good internal consistency for the averaged index, Cronbach's α = 0.84.

Threat
As a measure of technology-related threat, we added the tenitem scale (e.g., "This robot gives a peaceful impression.", "I know that this robot would not harm me.", "This robot is up to no good.") developed by Stein and colleagues [38]. All items had to be answered on seven-point Likert scales. Reliability for the averaged scale turned out good, Cronbach's α = 0.84.

Eeriness
The robot's eeriness was assessed using eight semantic differentials (e.g., "bland/uncanny", "boring/shocking") provided by Ho and MacDorman [62]. Again, participants were provided with seven-point scales to express how they perceived the depicted NAO robot. For the resulting eeriness index, we observed acceptable internal consistency, Cronbach's α = 0.79.

Interest in Robots (Covariate)
Previous research shows that positive attitudes towards robotic technology in general strongly affect participants' evaluation of specific robots as well [63,64]. As such, we decided to include participants' interest in robots as a covariate, measuring it with three items ("I would love to interact more with robots.", "I am really interested in robots.", "I don't find robots fascinating at all.") using seven-point scales. The averaged robot interest index showed good internal consistency, Cronbach's α = 0.82. Table 1 gives an overview of the zero-order correlations between the measured variables, whereas Table 2 shows the means and standard deviations obtained in our study. Additionally, Fig. 2 can be used for a graphical inspection of our obtained group differences.

Results
Since we found our four dependent variables to be significantly intercorrelated, we decided to investigate the effects  F(4,196) = 11.81, p < .001, with a large effect size of η p 2 = .19. Conversely, neither the main effect of robot size (p = .458), nor the interaction effect combining both factors (p = .357) turned out significant. The effect of the covariate slightly missed the conventional threshold of significance (p = .056).
Based on the significant multivariate effect of robot pose, we proceeded with univariate analyses regarding this factor. By these means, we encountered significant effects of the robot's body language on perceived dominance, F(1,199) = 35.77, p < .001, η p 2 = .15, and perceived competence, F(1,199) = 31.81, p < .001, η p 2 = .14. Numerically, the large effect sizes are echoed in substantial rating differences between the groups: Participants who viewed expansively posing NAOs rated them much higher in both measures (dominance: M = 3.09, SD = 1.44; competence: M = 4.17, SD = 1.14) than the groups that were shown NAO in constrictive poses (dominance: M = 2.01, SD = 1.05; competence: M = 3.31, SD = 1.03). As such, we confirm hypotheses H1a and H2a, although the corresponding hypotheses H1b and H2b have to be rejected. Lastly, neither the effect of NAO's posing on perceived eeriness (p = .372) nor that on perceived threat (p = .438) was found to be significant, resulting in the rejection of hypotheses H3 and H4.
Concluding our data analysis with an exploratory look into potential gender differences, we carried out separate subgroup analyses using the data of our female and male participants. Doing so, we found similar effect patterns emerging for both examined genders, matching the results described above. The main effect was similar in size for both genders (women: η p 2 = .25; men: η p 2 = .20), and the overall pattern of significant results remained the same. As one notable distinction, we observed that the included covariate-interest in robots-only exerted a significant influence in the MAN-COVA using the female participants' data (p = .009, η p 2 = .12), but not in the analysis focusing on the participating men (p = .955; η p 2 = .01).

Discussion
In order to successfully establish social robots in the roles they are being developed for, positive user experiences are of utmost importance. Acknowledging this, a growing body of literature has discussed how the nonverbal behavior of robots might contribute to smooth human-robot interactions, for instance by evoking certain desirable impressions among users. In our contribution to this emerging research area, we first obtained two noteworthy results: Confident, assertive power poses indeed made robots seem more dominant and competent than more constrictive body language. To our surprise, these effects emerged regardless of the robot's size: Statistically, it did not matter whether the portrayed robot was depicted a mere 55 cm small or in the size of a human adult. From a psychological perspective, this implies that-in contrast to human-to-human contexts-body size might not necessarily elicit different impressions and expectations when engaging a robot. At the same time, we want to caution readers against overgeneralizing our findings, as they are merely based on static, mediated stimuli (i.e., photographs). In fact, we cannot rule out that different effects would occur when presenting participants with video sequences of robots-which include much more nonverbal cues such as timing, sequence, and variation-or even real-life interactions. Especially in the latter case, we would expect participants to feel quite differently in front of an adult-sized than a toy-sized robot, as the machine's sheer mass and physicality would probably suffice to make it seem more dominant. Still, given that most people get acquainted with robots through media instead of real-life contact [66], we suggest that our study offers relevant insight, both to researchers and developers of robotic technology, as well as media producers who strive to portray robots in a certain way.
In the same vein, a worthwhile discussion may be invited by our findings (or lack thereof) concerning negative user perceptions, i.e., threat and eeriness. For both of these study variables, no significant group differences could be observed after manipulating the robot's pose and size. If taken at face value, this suggests that feelings of uncanniness or potential danger might not depend on the posture or dimensions of a social robot, so that engineers might feel free to make use of expansive body language without having to worry about backfiring effects. However, we again have to consider that different results could emerge once people actually stood in front of a robot; in that case, observers might think much more about the machine's ability to grab or hurt them, which would likely prompt more negative reactions. Similarly, the uncanny valley hypothesis in its original form includes the assumption that moving stimuli evoke stronger creepiness and aversion than static ones [32]-a possibility that also needs to be acknowledged in the interpretation of the observed null findings.
In terms of future research directions, we believe that studies focusing on natural human-robot interactions constitute the most important next step to alleviate our study's limitations. At the same time, this might turn out quite difficult, as identical robots of different sizes would be required to secure internal validity. Given the necessary resources, however, any scientific efforts that investigate the effects of robotic body language in live interactions will certainly be of great value. Additionally, we believe that follow-up studies could advance the current line of research by going beyond our dichotomous differentiation between expansive and constrictive poses-looking into the effects of more nuanced nonverbal behaviors instead. After all, it stands to reason that even minuscule details in the presented stimulus pictures might have affected participants' perceptions; for instance, portraying the robot in a relaxed seating position might have subverted any notion of threat, as the machine may not have seemed ready to execute relevant actions in this situation. Likewise, in the stimulus image that depicted the robot among a group of humans, the shown individuals looked rather happy and shared close personal space with the machine-potentially biasing participants towards competence and against threat perceptions. Hence, further studies (e.g., based on insight from the field of proxemics) are all but needed to elucidate on the intricate interactions that determine users' reactions to different robot poses.
Pointing out another methodological limitation of our study, we would like to note that the current experiment made use of only one specific robot design, namely the robot NAO, which is often described as "cute" by study participants [67]. Further research could therefore try to replicate the reported findings using more serious-looking robots, which might inherently appear more threatening or eerie. Then again, as studies have shown that even relatively similar types of social robots can trigger different dispositional attributions [68], it might also be worthwhile to repeat the current study with only slight visual modifications (e.g., NAO robots with different coloration). Of course, all replication and follow-up efforts will need to consider sample characteristics-as literature clearly shows that age, gender, cultural norms, and personality all affect the level of comfort people experience in the face of autonomous technology [63,[69][70][71]. While our exploratory subgroup analysis of potential gender differences suggested rather similar effects for both men and women, it should be noted that our findings are still based on a single convenience sample. As such, more diverse groups of participants should be recruited in order to understand whether the observed results are consistently observed among people from different age brackets, sociodemographic backgrounds, or countries.
Last but not least, researchers might also pursue conceptual extensions of the presented work. As a specific recommendation in this regard, we encourage scholars to look deeper into users' perception of intentionality and trust, two aspects that have been highlighted as most crucial for the successful incorporation of robots into human society [72][73][74]. In all probability, trusting a robot (and its intentions) may shield observers against threat-inducing cues, so that measuring this variable will certainly result in a more complete picture in future studies. Along the same lines, we anticipate that the intelligence ascribed to a robot might constitute another crucial moderator of how people evaluate robotic body language; after all, assuming a robot to be genuinely competent will likely change the impression it conveys with an assertive, powerful pose. As such, including user's mind attributions concerning robots may offer another ideal starting point for further research.

Conclusion
Conveying information about one's emotions, attitudes, and dispositions constitutes one of the primary functions of using body language. Robot engineers can make use of this fundamental principle of human communication to improve the acceptance (and thematic fit) of their creations. According to our study, a robot placing its hands confidently on its hips may indeed seem more competent and powerful than a machine displaying constrictive body language, without necessarily evoking stronger unease among observers. Furthermore, our results suggest that similar effects of nonverbal behavior might apply to robots of different sizes-at least if they remain at a safe (or mediated) distance. long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecomm ons.org/licenses/by/4.0/.