1 Introduction

In recent years, digital technologies have been used and developed to support the learning of English as a Second Language (ESL) (Golonka et al., 2014). Social robots dominated by artificial intelligence (AI) have appeared as an exciting innovation in the field of ESL. AI-based social robots that can interact with an individual in multilingual contexts and provide immediate feedback using speech, movement, and facial expressions have the potential to become excellent resources for foreign language learner (Neumann, 2020; Tuna & Tuna, 2019). Recent reviews on the application of AI-based social robots (e.g., Cheng et al., 2018) indicate that they are very promising for formal and informal language learning in educational contexts ranging from preschool to college. In the ESL context, researchers have acknowledged the learning advantages of AI-based social robots for children in terms of their performance while learning English (e.g., Alemi & Haeri, 2020; Tolksdorf et al., 2021; van den Berghe et al., 2021), the creation of enthusiasm and excitement while learning English (e.g., Crompton et al., 2018), and the growth of sustained interest and engagement (e.g., Kim et al., 2014).

While English is often considered a global language used for communicating and socializing with people from around the world (Sharifian, 2009), most Taiwanese parents think that children can gain a competitive academic advantage and possibly a better career through mastering English (Lan et al., 2012; Lin & Chen, 2016; Oladejo, 2006). Hence, most parents in Taiwan are motivated to have their children learn English and provide immense financial resources for their children to study English language starting from an early age. When the opportunities that parents create to facilitate their child’s English acquisition through English practice at home are crucial for success in achieving English learning goals (Li, 2006), AI-based social robots can play an important role in English practices that parents provide at home to promote their children’s ESL learning. Given that parents have the final decision on whether to adopt AI-based social robots to assist their children to learn English at home, it is important to investigate parental acceptance of AI-based social robots to better understand the factors that drive parental behavioral intention to support the use of these robots for their children’s ESL learning.

To examine how parents intend to behave regarding the use of AI-based social robots, previous studies on technology acceptance that highlight individual motivations, personality traits and perceptions of technology are essential. For example, some researchers have explored the influences of utilitarian and hedonic motivations on adopting or using technology (e.g., Akdim et al., 2022; Keszey, 2020); others have focused on the impact of personality traits (e.g., conscientiousness: Ma Dangi & Mohamed Saat, 2021; Forgas-Coll et al., 2022; personal innovativeness: Jackson et al., 2013; Kim, 2016; Sung & Jeon, 2020) . These studies have verified the significance of the aforementioned variables in the adoption or use of technology. Regarding personality traits, the current study focuses on conscientiousness as one dimension of the Big Five personality traits, and personal innovativeness, both of which are relevant to the adoption of technological innovation. Earlier studies have identified that acceptance intention toward robots is significantly influenced by conscientious (e.g., Forgas-Coll et al., 2022) and personal innovativeness (e.g., Sung & Jeon, 2020). Additionally, Yeh et al. (Yeh 2021) reported that among the Big Five personality traits, conscientious was the only predictor for motivations associated with the acceptance of learning technology adoption. Furthermore, the risk perception factor (e.g., physical risk) is seen as being capable of influencing behavioral intention to adopt or use types of technology (Gunawan et al., 2022). Past work has demonstrated that physical risks are likely to lower individuals’ adoption level of a technology (e.g., Ikhsan & Sunaryo, 2020). However, few studies have integrated physical risk to predict behavioral intentions with respect to the adoption of AI-based social robots. Additionally, given that AI-based social robots are emerging technologies with humanlike features (e.g., intelligence), theses robots have various abilities (i.e., autonomy, adaptability, reactivity, ability to cooperate, humanlike interaction, and personality) that are referred to as product smartness (Rijsdijk & Hultink, 2009; Rijsdijk et al., 2007). It is currently unclear how product smartness affects parental intention to use AI-based social robots for their children’s ESL learning.

To fully understand the relationships between parental behavioral intention to support the use of AI-based social robots and its influence factors, the present study theorizes a model that shows how parents decide to support the use of AI-based social robots for their children (age 12 and under) within the ESL context. Inspired by aforementioned studies, this paper incorporates eleven influencing factors into the analysis of parental behavioral intention to support the use of AI-based social robots, including motivations (i.e., utilitarian and hedonic motivations), product smartness (i.e., autonomy, adaptability, reactivity, ability to cooperate, humanlike interaction, and personality), personality traits (i.e., conscientiousness and innovativeness), and physical risk.

2 Theoretical background

2.1 AI-based social robots

Traditionally, a robot is a machine capable of automatically performing programmed actions in sequence order to complete particular tasks with intelligence (Neumann, 2020), which is not designed for interacting with humans (van den Berghe et al., 2021). Nowadays, robots are often equipped with social intelligence and thereby able to have social interaction and intelligent communication with humans (Neumann, 2020) in a way that follows “the behavioral norms expected by the people with whom the robot is intended to interact” (Bartneck & Forlizzi, 2004, p. 592). Social robots as intelligent and social machines can be either semiautonomous or autonomous (Bartneck & Forlizzi, 2004), and take on different physical forms that exist in the real world, such as human or animals (van den Berghe et al., 2021). To facilitate robot-human interaction, AI-based social robots can use visual recognition, speech production and recognition, body movements, and facial expressions (Chang et al., 2010; Neumann, 2020).

AI-based social robots can be programmed to take on the role of a teacher, tutor, or friend according to various learning task goals (van den Berghe et al., 2021) to support pedagogical purposes related to the language and literacy skills of learners (Neumann, 2020). Compared to other technologies, AI-based social robots with humanoid or pet-like appearances often have more natural interactions with children due to their ability to employ nonverbal cues (e.g., eye gazes, pointing, and other types of gestures) (van den Berghe et al., 2021). Moreover, AI-based social robots can manipulate real-life objects and use whole-body movements and gestures to integrate physical exercises or objects into learning tasks so that children can interact with their physical surroundings, and thereby increase sensorimotor experiences (Neumann, 2020; Wellsby & Pexman, 2014), which benefits vocabulary learning (Wellsby & Pexman, 2014) and appeals to many types of learners (Causo et al., 2016). Moreover, AI-based social robots also can connect with other devices such as computers through wireless internet connections to promote sharing and preserving of learning activities and results (Chang et al., 2010).

To date, very few studies have investigated parents’ perceptions on AI-based social robots’ usage for language learning (e.g., Lin et al., 2021; Louie et al., 2021; Tolksdorf et al., 2021) and most have used qualitative approaches. Those studies have shown that parents found educational robots to be useful and fun (Smakman et al., 2020) for their children’s ESL learning. However, no studies have been conducted that have used quantitative analysis to identify which factors are most influential for determining parental intentions to support the use of AI-based social robots in ESL learning.

2.2 Behavioral intention to support the use of AI-based social robots

Behavioral intention refers to the possibility that individuals believe that they will engage in a specific behavior (Fishbein & Ajzen, 1975). The existing information system literature has assumed that behavioral intention can be transformed into the actual use or adoption of a system or technology (Ajzen, 1991; Venkatesh et al., 2003, 2012). Many studies have provided research evidence that the execution of system use behavior depends on the strength of the behavioral intention to use (e.g., Almaiah et al., 2019; Fang et al., 2021; Gansser & Reich, 2021). For example, Gansser and Reich (2021) found that behavioral intention predicted the use behavior of AI-containing products (e.g., cleaning robots, robots for people with health needs, intelligent personal assistants and care robots) across all three application segments (i.e., mobility, household, and health) in everyday life environments. Given that behavioral intention to use has long been employed as a reliable predictor of actual use behavior, this study assumes that parents’ willingness to use AI-based social robots for their children’s ESL learning at home can predict their actual practical application of AI-based social robots. Thus, the current study focused on investigating the behavioral intention to use AI-based social robots for their children’s ESL learning at home.

2.3 Utilitarian and hedonic motivations

While individual’s technology adopting behavior can be directed by motivations related to use of technology (Ajzen, 1991), most technology adopting studies have focused on utilitarian motivations that reflect users’ desire for pragmatic reasons and benefits to decide to adopt new technologies (e.g., Aboelmaged, 2018; Kritzinger & Petzer, 2021). Conversely, there is a recent growing interest in hedonic motivation rooted in marketing research. Hedonic motivation refers to the extent to which individuals feel fun or pleasure acquired from using a technology (Ramírez-Correa et al., 2019; Venkatesh et al., 2012). Researchers have recognized that hedonic motivation plays a critical role in determining technology adoption (e.g., Brown & Venkatesh, 2005; Venkatesh & Davis, 2000).

Prior research has examined the effects of hedonic and utilitarian motivations in robot settings. For utilitarian motivations, most studies have chosen to study utilitarian motivators such as perceived usefulness and performance expectancy, instead of directly studying the impact of utilitarian motivations, and have proven that utilitarian motivators act as stronger predictors of behavioral intention to use AI-containing products (Gansser & Reich, 2021), hotel robot assistants (Lee et al., 2021), assistive social robots (Heerink et al., 2010) and smart home service robots (Hung et al., 2020). However, Ghazali et al. (2020) found perceived usefulness did not predict intention to use social robots as persuasive agents when making decisions about donating to charities. Regarding the association between hedonic motivation to behavioral intention, the existing study results are conflicting. Hedonic motivation has been demonstrated to be a significant influencer of behavioral intention to adopt AI-containing products (Gansser & Reich, 2021) and hotel robot assistants (Lee et al., 2021). In contrast, Hung et al. (2020) found hedonic motivation did not determine behavioral intention to use smart home service robots.

Hence, AI-based social robots can be instrumentalized to improve young children’s ESL learning quality and efficiency through fun and enjoyable learning experiences. Utilitarian and hedonic motivations are expected to impact parents’ intention to use the AI-based social robots. Thus, this study hypothesizes the following:

  • H1: Utilitarian motivation positively impacts parental intention to use AI-based social robots.

  • H2: Hedonic motivation positively impacts parental intention to use AI-based social robots.

2.4 Product smartness

Intelligent products that contain microchips, software, and sensors can collect, process, and produce information like humans (Rijsdijk & Hultink, 2009; Rijsdijk et al., 2007). AI-based social robots with humanlike senses are intelligent products, since robots can detect and recognize users’ voices, speech, gestures, and facial expressions through cameras and sensors; differentiate their intent, facial expressions, and feedback; track their location within their surroundings; and categorize human social activities and behaviors to decode them (Tung & Au, 2018). According to Rijsdijk et al. (2007) and Rijsdijk and Hultink (2009), product smartness refers to intelligent products, such as robots, displaying various abilities that non-intelligent products do not possess. The dimensions of product smartness include autonomy, adaptability, reactivity, ability to cooperate, humanlike interaction, and personality.

The autonomy dimension represents the ability of an intelligent product to act in pursuit of certain goals, and work independently (Rhiu & Yun, 2018; Rijsdijk & Hultink, 2009; Rijsdijk et al., 2007). Intelligent products with high autonomy often reduce the time and energy users expend on tasks related to operation and control (Baber, 1996). Thus, AI-based social robots forego the need for human intervention as they independently perform multiple functions and ESL tasks with younger children in support of their English learning. Concerning utilitarian benefits, the autonomy of AI-based social robots provides convenience to all users with less need for operation and control. Moreover, the higher level of automation associated with AI-based social robots may lead users to concentrate on emotional benefits (e.g., enjoyment) induced by interacting with these robots (Frank et al., 2021). Maria and Christian (2019) found that the perceived autonomy level of a smart product (i.e., smart washing machine) positively influences the perceived functional value (i.e., features, performance, and quality) and emotional value (i.e., emotional experiences and sentiments regarding the perception and usage of a product). Moreover, Chen and Lin (2022) found the automation of leisure-sports appliances positively influences perceived usefulness. Thus, when parents perceive a higher autonomy level of AI-based social robots, this can lead to higher utilitarian and hedonic motivation. In turn, the following hypotheses are posited:

  • H3: The autonomy of AI-based social robots positively impacts parental utilitarian motivation.

  • H4: The autonomy of AI-based social robots positively impacts parental hedonic motivation.

The adaptability dimension is defined as the ability of an intelligent product to respond and adapt to new conditions (Rhiu & Yun, 2018; Rijsdijk et al., 2007; Rijsdijk & Hultink, 2009). AI can learn from and adapt future actions by analyzing the influence of their previous actions on the environment (High-Level Expert Group on Artificial Intelligence, 2019) based on continuously collected information regarding the users and their usage of the product (Lee & Kim, 2016), which can in turn generate personal meaning and emotional value (Park & Lee, 2014), but also lead to privacy concerns regarding user information collection and use (Lee & Kim, 2016). Similarly, the adaptability of a robot is defined as the perceived ability of a robot to customize its functionality to fit users’ changing needs, preferences, and personality (Heerink et al., 2010). In the ESL learning context, it is important that AI-based social robots can recognize ESL learners’ levels and their learning progress to choose appropriate ESL learning contents that suit the needs and preferences of those learners. In this way, learners can learn at their own pace and benefit from customized learning based on their strengths and needs (Tuna & Tuna, 2019). In a previous study by Park and Lee (2014), perceived adaptability level of a smart product (i.e., smart phones) was identified as positively predicting functional and emotional values. Thus, parental perceptions of AI-based social robots’ adaptability are expected to increase both utilitarian and hedonic motivation. In turn, the following are posited:

  • H5: The adaptability of AI-based social robots positively impacts parental utilitarian motivation.

  • H6: The adaptability of AI-based social robots positively impacts parental hedonic motivation.

The reactivity dimension represents the ability of an intelligent product to make different and instant responses to changes in its environment in a stimulus/response manner (Rijsdijk & Hultink, 2009; Rijsdijk et al., 2007). Reactive behaviors of AI-based social robots can include offering appropriate responses to learners’ performance (e.g., responding with exclamations of joy, or dealing with mistakes by providing hints) when conversing with ESL learners in real time by collecting environmental information about the learners (Chang et al., 2010). While Maria and Christian (2019) noted that the perceived reactivity level of a smart product (i.e., smart washing machine) negatively affects functional value and has no impact on emotional value, Park and Lee (2014) found that the perceived reactivity level of a smart product (i.e., smart phones) positively affects emotional value, but has no impact on functional value. Thus, parental perceptions of AI-based social robots’ reactivity are expected to increase utilitarian motivation and hedonic motivation, leading to the following hypotheses:

  • H7: The reactivity of AI-based social robots positively impacts parental utilitarian motivation.

  • H8: The reactivity of AI-based social robots positively impacts parental hedonic motivation.

The ability to cooperate dimension reflects the ability of an intelligent product to work cooperatively with other devices to complete a task with a common goal (Rijsdijk & Hultink, 2009; Rijsdijk et al., 2007). For example, AI-based social robots need to communicate with computers or other devices that users already own through a wireless channel (e.g., Bluetooth or Wi-Fi) in order to use their software or programs (Chang et al., 2010). The more other products an intelligent product can cooperate with, the more user needs the product is congruent with (Rijsdijk et al., 2007). This ability to cooperate may enhance the functionality of the connected devices (Rijsdijk et al., 2007), create convenience for users (Rhiu & Yun, 2018), and encourage positive emotions in users (Rhiu & Yun, 2018). Until now, no study has investigated the predicting effects of a smart products’ ability to cooperate in terms of utilitarian motivation and hedonic motivation. Even so, parental perception of AI-based social robots’ ability to cooperate is expected to increase both motivation types, as posited below:

  • H9: AI-based social robots’ ability to cooperate positively impacts parental utilitarian motivation.

  • H10: AI-based social robots’ ability to cooperate positively impacts parental hedonic motivation.

The humanlike interaction dimension concerns the ability of an intelligent product to communicate and interact with users in a humanlike way (Rijsdijk & Hultink, 2009; Rijsdijk et al., 2007). In the ESL learning context, it is necessary for AI-based social robots to engage in social interaction by talking, showing facial expressions, exhibiting social behaviors, and being able to act and respond in culturally appropriate manners, all to assist users to practice real-life conversations (Tuna & Tuna, 2019). Meanwhile, users are more likely to become emotionally attached to AI-based social robots that offer humanlike interactions and in turn experience additional emotional benefits (e.g., enjoyment) from these interactions (van Straten et al., 2020). No study has investigated the predicting effect of the perceived humanlike interaction level of a smart product on utilitarian motivation; however, Park and Lee (2014) found that the perceived humanlike interaction level of a smart product (i.e., smart phone) positively affects emotional values. Thus, parental perceptions of AI-based social robots’ humanlike interaction are expected to increase utilitarian motivation and hedonic motivation. In response, the following hypotheses are posited:

  • H11: Humanlike interaction of AI-based social robots positively impacts parental utilitarian motivation.

  • H12: Humanlike interaction of AI-based social robots positively impacts parental hedonic motivation.

Finally, the final dimension of personality refers to the ability to make an overall impression by showing the properties (Govers, 2004; Rijsdijk & Hultink, 2009; Rijsdijk et al., 2007) or emotional states (Rijsdijk & Hultink, 2009; Rijsdijk et al., 2007) of a credible character. Mou et al.’s (2020) systematic review of the personality of robots found that they are manifested according to visual appearance (e.g., anthropomorphic, zoomorphic, caricatured, and functional), language style, vocal features (e.g., volume, speaking speed, pitch, and the amount of speech), movement (e.g., kinesics, movement angles, movement speeds, and movemen t patterns), countenance (e.g., eye contact and gaze behavior), touch sensation of a robot part, the role that the user plays in the human–robot interaction, and proxemics (i.e., the distance between the robot and an object). Robots that display a particular personality have been found to affect users’ behaviors and task effort (Park et al., 2017), as well as evoke enjoyable and favorable emotions (Hwang et al., 2013). However, to date, no study has looked into the predicting effect of perceived personality of a smart product on utilitarian motivation and hedonic motivation. In the current study, parental perception of AI-based social robots’ personality is expected to increase both motivation types, leading to the following hypotheses:

  • H13: The personality of AI-based social robots positively impacts parental utilitarian motivation.

  • H14: The personality of AI-based social robots positively impacts parental hedonic motivation.

2.5 Conscientiousness

Conscientiousness is a primary Big five personality trait that reflects being well-organized, responsible, careful, and efficient (Goldberg, 1990). Highly conscientious individuals tend to be self-disciplined, have concrete goals with a strong desire to fulfil them, and possess a life direction. This implies that they carefully examine whether using a technology can help them effectively reach their goals (Devaraj et al., 2008) and thus may magnify their desire to use the technology. Moreover, highly conscientious individuals who are self-controlled and responsible tend to excel at controlling their desire to express their own emotions and experience pleasure (Taufik et al., 2019). Previous studies have reported conflicting results regarding the predicting effect of conscientiousness on utilitarian and hedonic motivations. For example,Punnoose (2012) found thatconscientiousness has a significant positive direct effect on utilitarian motivation (i.e., perceived usefulness) and a negative direct effect on hedonic motivation (i.e., perceived enjoyment) of using e-learning in the future. However, other researchers have shown conscientiousness to be irrelevant in predicting utilitarian motivation (i.e., perceived usefulness) (e.g., Denden et al., 2022) and hedonic motivation (e.g., Busch, 2020). As such, conscientious parents are more likely to carefully consider the utilitarian benefits of using AI-based social robots to enhance their kids’ ESL learning performance, and they will also be more utilitarian-motivated to use AI-based social robots by virtue of this deep consideration. However, the hedonic benefits that AI-based social robots can bring to their children during the ESL learning process may not reflect goals that highly conscientious parents want to pursue. Therefore, more conscientious parents are more likely to be utilitarian-motivated, and less likely to be hedonism-motivated, as posited below:

  • H15: Conscientiousness positively impacts parental utilitarian motivation.

  • H16: Conscientiousness negatively impacts parental hedonic motivation.

2.6 Personal innovativeness

Personal innovativeness refers to the level of willingness an individual holds to try out any newly introduced technology at an early stage (Agarwal & Prasad, 1998). While individuals respond differently to new innovations, personal innovativeness could be classified as a type of personality trait (Midgley & Dowling, 1978) . Individuals could be considered innovative if they are early adopters of new types of technology (Agarwal & Prasad, 1998). High innovation individuals may tend to recognize the utilitarian and hedonic benefits of adopting new types of technology (Oliveira et al., 2016). Previous studies have demonstrated that personal innovativeness has a positive significant role in enhancing the level of performance expectancy (e.g., Chayomchai, 2020) and perceived usefulness (e.g., Jackson et al., 2013; Shanmugavel & Micheal, 2022) from a utilitarian motivation perspective as well as perceived enjoyment (e.g., Alalwan et al., 2018; Rouibah et al., 2016) from a hedonic motivation perspective. Indeed, AI-based social robots can be seen as novel and new types of learning technology. Parents should have an extent level of personal innovativeness as a means to really be utilitarian- and hedonism-motivated to adopt AI-based social robots. In other words, highly innovative parents will not only be motivated to employ AI-based social robots for utilitarian benefits, but also perceive hedonic benefits associated with using them. Thus, the following hypotheses are proposed:

  • H17: Personal innovativeness positively impacts parental utilitarian motivation.

  • H18: Personal innovativeness positively impacts parental hedonic motivation.

2.7 Perceived physical risk

Physical risk refers to perceptions of risk associated with physical harm from the adoption of a technology (Gunawan et al., 2022). Such harm is not necessarily limited to technology adopters themselves—it may also affect others close to them (e.g., family members) who adopt the technology. Perceptions of risk associated with the adoption of a technology (e.g., physical risk) may negatively influence the motivation to use a technology to achieving one’s goals and perceptions of enjoyment while adopting the technology (Cocosila et al., 2009). Thus, physical risk has become an increasingly popular consideration in the technology adoption literature (Faqih, 2022). However, limited studies have discussed whether perceived physical risk has negative implications for utilitarian and hedonic motivations. One study conducted by Cocosila et al. (2007) showed a negative effect of physical risk on a utilitarian motivator (i.e., perceived usefulness) with respect to the adoption of 3G cell phones. Currently, no study exists that investigates the relationship between physical risk and hedonic motivation in terms of technology adoption. However, Zhong et al. (2021) revealed that hedonic motivation (i.e., enjoyment gained from dining-out behaviors) can be negatively influenced by perceived physical risk.

There are various reasons to perceive AI-based social robots as physically risky. For example, users may suspect that robots possessing intelligence and decision-making capabilities may behave automatically and unpredictably, which might endanger their lives (Delgosha & Hajiheydari, 2021). Similarly, Lin et al. (2021) found that parents perceived high autonomy storytelling robots as much more threatening low autonomy storytelling robots. Moreover, as robots are electrical machines, users could get electrically shocked during interacting with robots (Woo et al.,2021). In sum, parents who believe robots represent a low physical risk tend to be utility-motivated and hedonism-motivated. Hence, the following are posited:

  • H19: Perceived physical risk negatively impacts parental utilitarian motivation.

  • H20: Perceived physical risk negatively impacts parental hedonic motivation.

2.8 Research model

On the basis of a review of the literature, the study proposed a research model involving motivations (i.e., utilitarian and hedonic motivations), product smartness (i.e., autonomy, adaptability, reactivity, ability to cooperate, humanlike interaction, and personality), personality traits (i.e., conscientiousness and innovativeness), and physical risk. In Fig. 1, the conceptual model used in this study and the relationships between the constructs are illustrated. Based on this model, the aim of this study was to identify the factors that influence parental behavioral intention to support the use of AI-based social robots.

Fig. 1
figure 1

Research model

3 Methodology

3.1 Participants

A total of 315 valid anonymous survey responses were collected in two ways: through an online survey distributed to parents who were members of various online communities related to teaching English to children through social networked websites, and a paper survey distributed to school-run parents of children in kindergarten and elementary schools in Taiwan. Participation in the study was voluntary. To encourage participation, all participants were entered into a draw where they could win one of 100 gift cards valued at NT$100.

As Table 1 shows, participants were parents of children aged under 12 years: most were female (60%). The majority of parents (54.3%) were aged 31 to 40 years old; 25.7% were aged 41 to 50; 17.1% were aged 21 to 30, and 2.8 were 51 or above. Respondents ranged in terms of monthly income, with most earning NT$40,001–60,000 (33.3%), followed by NT$20,001–40,000 (33%), above NT60,000 (22.5%), and under 20,000 (11.1%). In terms of employment, 89.5% of these parents worked full- or part-time, 3.2% were students, and 7.3% identified as homemakers. Moreover, 15.6% reported having a child or children aged between 0 and 2 years, 27.9% reported having a child or children aged between 3 and 6 years, 32.4% reported having a child or children who attended grade 1 or 2, 23.5% reported to having a child or children who attended grade 3 or 4, and 16.2% reported having a child or children who attended grade 5 or 6.

Table 1 Demographic data of participants (n = 315)

3.2 Measures

The study relied on previously validated scales for all the research constructs in this particular research model, adjusted to the context of AI-based social robot usage (please see the Appendix for more information). The study also consulted three experts in information systems and/or English teaching systems to examine the intelligibility and appropriateness of all the survey items before distributing the survey and modified items based on the expert's suggestion. Parents' intention to use an AI robot was assessed with three items modified from Zhu et al. (2022). Five items were used to measure utilitarian motivation of parents, and three items to measure parents’ hedonic motivation, all adapted from To et al. (2007). Parents’ perceptions of an AI-based social robot’s autonomy, adaptability, reactivity, ability to cooperate, humanlike interaction, and personality were adapted from Rijsdijk et al.’s (2007) product intelligence inventory. Parents' personality traits pertaining to conscientiousness were assessed with the eight-item conscientiousness subscale in the Big Five inventory, developed by John and Srivastava (1999). Parents’ personal innovativeness was assessed using three items adopted from Jackson et al. (2013). Three items from Wiedmann et al. (2011) were modified to assess parents’ perceptions of physical risk. All research constructs were measured on a seven-point Likert scale ranging from 1 (totally disagree) to 7 (totally agree).

3.3 Data analysis

Partial least squares structural equation modeling (PLS-SEM) was used to test the research model and hypotheses. When evaluating the PLS-SEM results, the study followed a two-step process involving separate assessment of the measurement models and the structural model (Hair et al., 2017). The first step focused on the reliability and validity of the construct measures, while the second tested the hypotheses with the coefficients obtained. Both the measurement model and the structural model were assessed using Smartpls software.

4 Results

4.1 Measurement model

Convergent validity of the measures was assessed by examining the individual item loadings between an item and its corresponding underlying factor, as well as the average variance extracted (AVE). After deleting one conscientiousness item, one ability of learning item, and three utilitarian motivation items due to cross-loadings or weak loadings, the results of the final analysis show that all factor loadings exceeded 0.7 (Hair et al., 2017), and all AVE values exceeded 0.5 (Hair et al., 2017) as shown in Table 2, suggesting good convergent validity. Internal consistency was assessed using Cronbach's Alpha and composite reliability (CR). As depicted in Table 2, the Cronbach's Alpha and CR values of all constructs exceeded the recommended level of 0.70, demonstrating good internal consistency. These results support the convergent validity of the measures. Discriminant validity was assessed using the Fornell-Larcker criteria and the Hetrotrait-Monotrait Ratio (HTMT). The Fornell and Larcker criteria requires that discriminant validity be established by confirming that the square root of the AVE of each construct is larger than its correlation with other constructs. Further, the Hetrotrait-Monotrait Ratio (HTMT) must not exceed the threshold of 0.9 (Henseler et al., 2015). As seen Table 3, the model had good discriminant validity.

Table 2 Cronbach’s alpha, composite reliability, and average variance extracted
Table 3 Discriminant validity

Furthermore, in addition to the reliability and validity assessments, multicollinearity was checked to make sure all inner variance inflation factor (VIF) values were below 5 (Hair et al., 2017), to ensure no significant intercorrelation between the constructs. The study found that the inner VIF values in the structural model ranged from 1.036 to 3.027, indicating that the research constructs did not suffer from multicollinearity issues.

4.2 Structural model

As seen in Table 4 and Fig. 2., seven of the 20 hypotheses are confirmed. Utilitarian motivation (H1) and hedonic motivation (H2) were significant in explaining intention to use AI robots (β = 0.391, p < 0.001 and β = 0.426, p < 0.001, respectively). H3 and H4 concern the relationships between autonomy and both utilitarian and hedonic motivation: both were significant (β = 0.189, p < 0.05 and β = 0.189, p < 0.01, respectively) supporting H3 and H4. However, Hypotheses 5 and 6, which proposed the effect of autonomy on both utilitarian (H5: β = 0.005, p = 0.95) and hedonic motivation (H6: β = 0.047, p = 0.575), were not supported. Moreover, the direct paths from reactivity to both utilitarian (H7: β = 0.091, p = 0.249) and hedonic motivation (H8: β = 0.068, p = 0.404) were not significant. The direct effects of the ability to cooperate on both utilitarian (H9: β = 0.124, p = 0.088) and hedonic motivation (H10: β = 0.059, p = 0.414) were not significant. No significant relationship existed between humanlike interaction and utilitarian motivation (H11: β = 0.078, p = 0.298), whereas humanlike interaction (H12: β = 0.19, p < 0.01) positively affected hedonic motivation. Hypotheses 13 and 14, which suggest that personality has positive effects on utilitarian motivation (H13: β = 0.055, p = 0.383) and hedonic motivation (H14: β = 0.108, p = 0.073), were not supported. Hypotheses 15 and 16 were not supported: the standardized path coefficients from conscientiousness to utilitarian motivation (H15: β = 0.118, p = 0.055) and to hedonic motivation (H16: β = 0.067, p = 0.267) were not statistically significant. H17 and H18 concerned the effect of personal innovativeness on utilitarian (β = 0.160, p < 0.01) and hedonic motivation (β = 0.201, p < 0.001), respectively: both were supported. The direct links from physical risk to both utilitarian (H19: β = -0.048, p = 0.355) and hedonic motivation (H20: β = -0.031, p = 0.540) were not supported. These results show that the relationship between hedonic motivation and intention to use is the strongest one in the model. The research model explains 57.3% of the variance in use intention, 37.6% of the variance in utilitarian motivation, and 47.6% of the variance in hedonic motivation.

Table 4 Summary of structural model
Fig. 2
figure 2

Results of PLS analysis. Note: Solid lines represent significant predictive paths and dashed lines represent nonsignificant predictive paths. *p < .05, **p < .01, ***p < .001

5 Discussion

To our knowledge, the present study is the first to theorize a research model for a sample of parents in order to better understand parental intention to support the use of AI-based social robots to assist in the ESL learning of children under 12 through relationships among motivations (i.e., utilitarian and hedonic motivations), product smartness, personality traits (i.e., conscientiousness and innovativeness), and physical risk. The current study tests the framework with a sample of parents who are prone to using AI-based social robots to help with ESL learning. Understanding the determinants of parental behavioral intention to use educational robots is important for instructional designers, AI-based social robot engineers, and AI-based social robot designers to better understand the needs and wants of parents in an effort to determine whether AI-based social robots fit the context of ELS learning at home.

The results demonstrate that parents’ intention to use AI-based social robots for their kids is not only directly affected by utilitarian motivation—their rational side—but also by their hedonic motivation—their emotional side. Therefore, utilitarian and hedonic motivations are real drivers behind parental intention to use these robots. The findings reveal that utilitarian motivation is significant for the behavioral intention to adopt AI-based social robots, consistent with previous research (e.g., Gansser & Reich, 2021; Heerink et al., 2010; Hung et al., 2020; Lee et al., 2021). This implies that the extent to which AI-based social robots provide utilitarian benefits (i.e., being helpful, effective, functional, necessary, and practical) for kids learning English is significant to the adoption rate of AI-based social robots. Moreover, the results suggest that higher hedonic motivation in terms of using AI-based social robots may result in higher intention to adopt AI-based social robots, similar to studies conducted by Gansser and Reich (2021) and Lee et al. (2021). This may suggest that hedonic reasons such as fun, delight, and enjoyment may lead to the adoption of AI-based social robots.

Moreover, the six smartness attributes of AI-based social robots (i.e., autonomy, adaptability, reactivity, ability to cooperate, humanlike interaction, and personality) were used to determine the influence of parental utilitarian and hedonic motivations. While not all these attributes were found to influence utilitarian and hedonic motivations, some notable results are noted here. 1) Autonomy influences parental utilitarian and hedonic motivations directly, confirming the findings of Maria and Christian (2019) and Chen and Lin (2022). This means that this autonomy attribute can increase parents’ feelings of convenience and provide them with hedonic experiences, because AI-based social robots can perform tasks independently. The results show that the autonomy of AI-based social robots is a source of both parental utilitarian and hedonic motivation, which in turn leads to an increase in the intention to use AI-based social robots. 2) Humanlike interaction significantly predicts parental hedonic motivation, confirming Park and Lee (2014)’s findings, but insignificantly predicts parental utilitarian motivation, contrary to Park and Lee (2014)’s findings. This means that the humanlike interaction attribute can strengthen parental hedonic motivation because AI-based social robots communicate and interact with their children in a humanlike way during the ESL learning process. However, parents tend to perceive that humanlike interaction provides no utilitarian benefit (i.e., being helpful, effective, functional, necessary, and practical). These findings suggest that the higher the level of humanlike interaction the AI-based social robots can achieve, the more hedonically motivated parents will be with respect to their adoption. Further, the study does not support the impacts of adaptability, reactivity, ability to cooperate, or personality on parental utilitarian and hedonic motivation. Apparently, parents perceived that the adaptability, reactivity, ability to cooperate, and personality level of the robots do not provide utilitarian benefits (i.e., being helpful, effective, functional, necessary, and practical) as their children learn English, so they did not become utilitarian-motivated. Moreover, parents are not hedonically motivated to adopt AI-based social robots while they do not perceive that the higher adaptability, reactivity, ability to cooperate, and personality level of the robots leads to more fun, delight, and enjoyment for their children as they learn English. The insignificant results pertaining to adaptability, reactivity, and personality might be, to some degree, due to the uncanny valley effect: parents may exhibit negative emotional reactions to a robot’s human resemblance shown through its adaptive and affective capabilities (Lin et al., 2021). The statistically non-significant effects regarding the ability to cooperate may be explained by the higher expectations that parents put on the robots. It is possible that the widespread dissemination of high-technology products has resulted in a situation where the ability of a robot to work cooperatively with other devices and systems may already be regarded as standard, rather than as a special feature.

This study explored the influence of two types of personality traits on utilitarian and hedonic motivation: conscientiousness and personal innovativeness. The results indicate that the influence of personal innovativeness on parental utilitarian and hedonic motivations to adopt AI-based social robots was validated, but did not validate the effect of conscientiousness on these motivations. This study assumed participants would view AI-based social robots as innovative products, and in turn highlighted the role of personal innovativeness as a means to increase parental utilitarian and hedonic motivation. These findings support the results of Chayomchai (2020), Jackson et al. (2013), and Shanmugavel and Micheal (2022), linking personal innovativeness to utilitarian motivators (i.e., performance expectancy and perceived usefulness); they are also in line with research conducted by Alalwan et al. (2018) and Rouibah et al. (2016) in terms of relating personal innovations to hedonic motivators (i.e., perceived enjoyment). These results suggest that as parents become more innovative, their ability to perceive utilitarian benefits and hedonic appeal associated with the adoption of AI-based social robots increases. That is, utilitarian benefits and hedonic appeal may increase in situations where parents pursue innovation and exploration. However, the hypothesis that conscientiousness has a positive relationship with utilitarian and hedonic motivations was not supported. Some possible reasons for this finding include parental concerns regarding robots, such as that they: lack empathy, may give out their children’s data or information through manipulation or hacking, and increase children’s amount of screen time, as noted in Louie et al.’s (2021) study.

The findings of this study, nevertheless, showed no support for the effects of perceived physical risk on parental utilitarian and hedonic motivations, which contradicts the results obtained by Cocosila et al. (2007) and Zhong et al. (2021). One possible reason is related to the cute, likeable appearance of AI-based social robots: they are usually designed with the goals of increasing children’s learning engagement and motivation, and reducing any fears associated with robots (Randall,2019). Moreover, in Louie et al.’s (2021) study, parents chose robots that had an appealing, smart-looking appearance. This type of choice may reduce parental physical risk perceptions and alleviate the negative effects of perceived physical risk on parental utilitarian and hedonic motivations toward adopting AI-based social robots.

5.1 Theoretical implications

This study is an early attempt to use PLS-SEM to examine determinants of parental behavioral intention to adopt AI-based social robots. The findings provide several essential contributions to the existing technology adoption research. First, past studies investigating AI-based social robots in educational contexts were conducted in school settings with a predominant focus on whether using social robots could improve affective learning outcomes, cognitive learning outcomes, and motivation. Little work has been conducted on parental willingness to adopt AI-based social robots to support children’s ESL learning at home. A research model was proposed and validated in this study to address this gap. The empirical evidence of the present study broadens our understanding of the factors that influence parental intention to use these types of robots. Second, the findings highlight the importance of utilitarian motivation and hedonic motivation in affecting parental willingness to adopt AI-based social robots to support children’s ESL learning at home. That said, hedonic motivation was found to be the strongest predictor of parental intention to adopt AI-based social robots, indicating that parents’ prioritized fun, feelings of delight, and the enjoyment associated with using AI-based social robots. Third, comparatively little is known about what role perceived smartness features of AI-based social robots play in parental intention to adopt such robots for ESL learning, which is an important antecedent of parental adoption of these robots. Evidence for the effect of product smartness, therefore, is critical not only to establish a link between smartness features and motivation, but also to demonstrate that smartness features can affect how users are motivated to adopt the product, subsequently influencing behavioral intention of adoption. The results show that the effect of autonomy on utilitarian motivation surpassed that of personal innovativeness. This suggests that autonomy promotes utilitarian motivation in the context of parental acceptance of AI-based social robots better than personal innovativeness does. Finally, this study has successfully established links from personal innovativeness to motivations. This expands our understanding of parental willingness to adopt AI-based social robots by considering the role of innovativeness in this research context, as well as examining the impact of innovativeness on utilitarian and hedonic motivations. The influence of innovativeness on hedonic motivation was superior to those of autonomy and humanlike interaction. This may indicate that personal innovativeness increases hedonic motivation better than autonomy and humanlike interaction do in the context of parental acceptance of AI-based social robots.

5.2 Practical implications

From a practical perspective, this study offers empirical evidence that needs to be taken into account by AI-based social robot designers, developers, and marketers. First, the results support that utilitarian motivation and hedonic motivation have positive effects on parental intention to adopt AI-based social robots. Thus, designers, developers, and marketers of AI-based social robots should consider emphasizing the utilitarian benefits and hedonic appeal of these robots in their design, development, and marketing activities. The stronger impact of hedonic appeal indicates that greater efforts should be targeted on the fun, feelings of delight, and enjoyment associated with using AI-based robots to encourage parental acceptance. Second, the results clearly show that personal innovativeness is a significant predictor in terms of predicting parents' utilitarian and hedonic motivation to adopt AI-based social robots. Therefore, further marketing efforts should promote these robots as a pioneering and emerging technology that adds value to children’s ESL learning at home. In addition, when introducing AI-based robots to the marketplace, marketers should try to identify parents with high personal innovativeness, which may maximize and expedite product acceptance. Third, the six smartness features of AI-based social robots have different impacts on utilitarian and hedonic motivation. The results provide evidence that autonomy and humanlike interaction positively enable increased hedonic motivation, while autonomy increases utilitarian motivation. For these reasons, designing, developing, and marketing AI-based social robots in a way that facilitates autonomy and humanlike interaction is suggested.

5.3 Limitations and future studies

Several limitations should be mentioned. First, the study utilized a convenience sample where all respondents came from Taiwan. Moreover, half of participants in the study were mothers aged between 30 and 40. Hence, the findings may not be generalizable to other nations and settings. Future researchers can consider utilizing a more diverse sample and random sampling techniques to verify and generalize the findings across multiple nations. Second, given that this study focused on the Big five personality trait of conscientiousness, future researchers can include the other traits (e.g., agreeableness, extroversion, openness, and neuroticism) to elaborate on the results. Third, this study specifically focused on the effect of physical risks on utilitarian and hedonic motivation. Future studies may consider incorporating privacy risk perceptions, which has been identified as an issue of concern for some parents with respect to using robots in children ESL learning contexts (e.g., Louie et al., 2021). Finally, future researchers should attempt to add other factors that may influence utilitarian motivation (e.g., ease of use: Chang et al., 2023) or hedonic motivation (e.g., flow: Zhao & Bacao, 2021), or integrate variables from the Hedonic—Motivation System Adoption Model proposed by Lowry et al. (2013), which may produce a better explanation of the variance in parental willingness to adopt AI-based social robots.